Sparse coding with memristor networks

Sparse representation of information provides a powerful means to perform feature extraction on high-dimensional data and is of broad interest for applications in signal processing, computer vision, object recognition and neurobiology. Sparse coding is also believed to be a key mechanism by which bi...

Full description

Saved in:
Bibliographic Details
Published in:Nature nanotechnology Vol. 12; no. 8; pp. 784 - 789
Main Authors: Sheridan, Patrick M., Cai, Fuxi, Du, Chao, Ma, Wen, Zhang, Zhengya, Lu, Wei D.
Format: Journal Article
Language:English
Published: London Nature Publishing Group UK 01.08.2017
Nature Publishing Group
Subjects:
ISSN:1748-3387, 1748-3395, 1748-3395
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Sparse representation of information provides a powerful means to perform feature extraction on high-dimensional data and is of broad interest for applications in signal processing, computer vision, object recognition and neurobiology. Sparse coding is also believed to be a key mechanism by which biological neural systems can efficiently process a large amount of complex sensory data while consuming very little power. Here, we report the experimental implementation of sparse coding algorithms in a bio-inspired approach using a 32 × 32 crossbar array of analog memristors. This network enables efficient implementation of pattern matching and lateral neuron inhibition and allows input data to be sparsely encoded using neuron activities and stored dictionary elements. Different dictionary sets can be trained and stored in the same system, depending on the nature of the input signals. Using the sparse coding algorithm, we also perform natural image processing based on a learned dictionary. The implementation of bio-inspired sparse coding algorithms aimed at image processing is demonstrated by exploiting 32 × 32 crossbar arrays of analogue memristors.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:1748-3387
1748-3395
1748-3395
DOI:10.1038/nnano.2017.83