Understanding and optimizing asynchronous low-precision stochastic gradient descent

Stochastic gradient descent (SGD) is one of the most popular numerical algorithms used in machine learning and other domains. Since this is likely to continue for the foreseeable future, it is important to study techniques that can make it run fast on parallel hardware. In this paper, we provide the...

Full description

Saved in:
Bibliographic Details
Published in:2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA) pp. 561 - 574
Main Authors: De Sa, Christopher, Feldman, Matthew, Re, Christopher, Olukotun, Kunle
Format: Conference Proceeding
Language:English
Published: ACM 01.06.2017
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first