JOURNAL ARTICLE

Learning Optimized Features for Hierarchical Models of Invariant Object Recognition

Heiko WersingEdgar Körner

Year: 2003 Journal:   Neural Computation Vol: 15 (7)Pages: 1559-1588   Publisher: The MIT Press

Abstract

There is an ongoing debate over the capabilities of hierarchical neural feedforward architectures for performing real-world invariant object recognition. Although a variety of hierarchical models exists, appropriate supervised and unsupervised learning methods are still an issue of intense research. We propose a feedforward model for recognition that shares components like weight sharing, pooling stages, and competitive nonlinearities with earlier approaches but focuses on new methods for learning optimal feature-detecting cells in intermediate stages of the hierarchical network. We show that principles of sparse coding, which were previously mostly applied to the initial feature detection stages, can also be employed to obtain optimized intermediate complex features. We suggest a new approach to optimize the learning of sparse features under the constraints of a weight-sharing or convolutional architecture that uses pooling operations to achieve gradual invariance in the feature hierarchy. The approach explicitly enforces symmetry constraints like translation invariance on the feature set. This leads to a dimension reduction in the search space of optimal features and allows determining more efficiently the basis representatives, which achieve a sparse decomposition of the input. We analyze the quality of the learned feature representation by investigating the recognition performance of the resulting hierarchical network on object and face databases. We show that a hierarchy with features learned on a single object data set can also be applied to face recognition without parameter changes and is competitive with other recent machine learning recognition approaches. To investigate the effect of the interplay between sparse coding and processing nonlinearities, we also consider alternative feedforward pooling nonlinearities such as presynaptic maximum selection and sum-of-squares integration. The comparison shows that a combination of strong competitive nonlinearities with sparse coding offers the best recognition performance in the difficult scenario of segmentation-free recognition in cluttered surround. We demonstrate that for both learning and recognition, a precise segmentation of the objects is not necessary.

Keywords:
Computer science Pooling Artificial intelligence Cognitive neuroscience of visual object recognition Pattern recognition (psychology) Neural coding Feature (linguistics) Convolutional neural network Facial recognition system Feature learning Feature vector Dimensionality reduction Machine learning Feature extraction

Metrics

168
Cited By
12.36
FWCI (Field Weighted Citation Impact)
60
Refs
0.99
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Sparse and Compressive Sensing Techniques
Physical Sciences →  Engineering →  Computational Mechanics
Medical Image Segmentation Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Domain Adaptation and Few-Shot Learning
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Learning Discriminative Hierarchical Features for Object Recognition

Zhen ZuoGang Wang

Journal:   IEEE Signal Processing Letters Year: 2014 Vol: 21 (9)Pages: 1159-1163
BOOK-CHAPTER

Models of invariant object recognition

Edmund T. Rolls

Year: 2001 Pages: 243-322
JOURNAL ARTICLE

Learning hierarchical sparse features for RGB-(D) object recognition

Liefeng BoXiaofeng RenDieter Fox

Journal:   The International Journal of Robotics Research Year: 2014 Vol: 33 (4)Pages: 581-599
BOOK-CHAPTER

Online Learning of Invariant Object Recognition in a Hierarchical Neural Network

Markus LeßmannRolf P. Würtz

Lecture notes in computer science Year: 2014 Pages: 427-434
© 2026 ScienceGate Book Chapters — All rights reserved.