Scalpel: Customizing DNN pruning to the underlying hardware parallelism

As the size of Deep Neural Networks (DNNs) continues to grow to increase accuracy and solve more complex problems, their energy footprint also scales. Weight pruning reduces DNN model size and the computation by removing redundant weights. However, we implemented weight pruning for several popular n...

Full description

Saved in:
Bibliographic Details
Published in:2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA) pp. 548 - 560
Main Authors: Jiecao Yu, Lukefahr, Andrew, Palframan, David, Dasika, Ganesh, Das, Reetuparna, Mahlke, Scott
Format: Conference Proceeding
Language:English
Published: ACM 01.06.2017
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first