SWG: an architecture for sparse weight gradient computation

On-device training for deep neural networks (DNN) has become a trend due to various user preferences and scenarios. The DNN training process consists of three phases, feedforward (FF), backpropagation (BP), and weight gradient (WG) update. WG takes about one-third of the computation in the whole tra...

Full description

Saved in:
Bibliographic Details
Published in:Science China. Information sciences Vol. 67; no. 2; p. 122405
Main Authors: Wu, Weiwei, Tu, Fengbin, Li, Xiangyu, Wei, Shaojun, Yin, Shouyi
Format: Journal Article
Language:English
Published: Beijing Science China Press 01.02.2024
Springer Nature B.V
Subjects:
ISSN:1674-733X, 1869-1919
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first