SWG: an architecture for sparse weight gradient computation
On-device training for deep neural networks (DNN) has become a trend due to various user preferences and scenarios. The DNN training process consists of three phases, feedforward (FF), backpropagation (BP), and weight gradient (WG) update. WG takes about one-third of the computation in the whole tra...
Saved in:
| Published in: | Science China. Information sciences Vol. 67; no. 2; p. 122405 |
|---|---|
| Main Authors: | , , , , |
| Format: | Journal Article |
| Language: | English |
| Published: |
Beijing
Science China Press
01.02.2024
Springer Nature B.V |
| Subjects: | |
| ISSN: | 1674-733X, 1869-1919 |
| Online Access: | Get full text |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Be the first to leave a comment!