Modeling object recognition in visual cortex using multiple firing k-means and non-negative sparse coding

The HMAX model developed by Serre et al. imitates the process of visual recognition in primates’ visual cortex. However, it has some limits in modeling the V2 neurons or higher level of visual cortex. We extend the model in some biologically plausible ways and construct a five-layer computational mo...

Full description

Saved in:
Bibliographic Details
Published in:Signal processing Vol. 124; pp. 198 - 209
Main Authors: Wang, Yanjiang, Deng, Limiao
Format: Journal Article
Language:English
Published: Elsevier B.V 01.07.2016
Subjects:
ISSN:0165-1684, 1872-7557
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The HMAX model developed by Serre et al. imitates the process of visual recognition in primates’ visual cortex. However, it has some limits in modeling the V2 neurons or higher level of visual cortex. We extend the model in some biologically plausible ways and construct a five-layer computational model, denoted as Sparse-HMAX model. First we use Gabor filters to describe the response properties of V1 neurons as in original HMAX model and describe C1 image patches with HOG descriptors. Then we integrate multiple firing k-means into the HMAX model to emulate the V2 neural responses and non-negative sparse coding to model V4 neurons. To investigate the efficacy of our proposed model, we perform experiments on three public databases: Caltech101, Caltech256 and GRAZ-01. Experimental results demonstrate that Sparse-HMAX model displays great improvements over the original HMAX model both in recognition accuracy and processing speed for object recognition. Our proposed method is also comparable to the prevalent approaches in recognition performance. •We construct a five-layer bio-inspired computational model to imitate the process of visual stream.•Multiple firing k-means is applied for emulating the V2 neural responses.•Non-negative sparse coding is used to reproduce the V4 neural responses.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:0165-1684
1872-7557
DOI:10.1016/j.sigpro.2015.08.017