Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/2981562.2981672guideproceedingsArticle/Chapter ViewAbstractPublication PagesnipsConference Proceedingsconference-collections
Article

Sparse deep belief net model for visual area V2

Published: 03 December 2007 Publication History

Abstract

Motivated in part by the hierarchical organization of the cortex, a number of algorithms have recently been proposed that try to learn hierarchical, or "deep," structure from unlabeled data. While several authors have formally or informally compared their algorithms to computations performed in visual area V1 (and the cochlea), little attempt has been made thus far to evaluate these algorithms in terms of their fidelity for mimicking computations at deeper levels in the cortical hierarchy. This paper presents an unsupervised learning model that faithfully mimics certain properties of visual area V2. Specifically, we develop a sparse variant of the deep belief networks of Hinton et al. (2006). We learn two layers of nodes in the network, and demonstrate that the first layer, similar to prior work on sparse coding and ICA, results in localized, oriented, edge filters, similar to the Gabor functions known to model V1 cell receptive fields. Further, the second layer in our model encodes correlations of the first layer responses in the data. Specifically, it picks up both colinear ("contour") features as well as corners and junctions. More interestingly, in a quantitative comparison, the encoding of these more complex "corner" features matches well with the results from the Ito & Komatsu's study of biological V2 responses. This suggests that our sparse variant of deep belief networks holds promise for modeling more higher-order features.

References

[1]
G. E. Hinton, S. Osindero, and Y.-W. Teh. A fast learning algorithm for deep belief nets. Neural Computation, 18(7):1527-1554, 2006.
[2]
M. Ranzato, C. Poultney, S. Chopra, and Y. LeCun. Efficient learning of sparse representations with an energy-based model. In NIPS, 2006.
[3]
Y. Bengio, P. Lamblin, D. Popovici, and H. Larochelle. Greedy layer-wise training of deep networks. In NIPS, 2006.
[4]
H. Larochelle, D. Erhan, A. Courville, J. Bergstra, and Y. Bengio. An empirical evaluation of deep architectures on problems with many factors of variation. In ICML, 2007.
[5]
G. E. Hinton, S. Osindero, and K. Bao. Learning causally linked MRFs. In AISTATS, 2005.
[6]
S. Osindero, M. Welling, and G. E. Hinton. Topographic product models applied to natural scene statistics. Neural Computation, 18:381-344, 2006.
[7]
M. Ito and H. Komatsu. Representation of angles embedded within contour stimuli in area v2 of macaque monkeys. The Journal of Neuroscience, 24(13):3313-3324, 2004.
[8]
J. H. van Hateren and A. van der Schaaf. Independent component filters of natural images compared with simple cells in primary visual cortex. Proc. R. Soc. Lond. B, 265:359-366, 1998.
[9]
A. J. Bell and T. J. Sejnowski. The 'independent components' of natural scenes are edge filters. Vision Research, 37(23):3327-3338, 1997.
[10]
B. A. Olshausen and D. J. Field. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381:607-609, 1996.
[11]
H. Lee, A. Battle, R. Raina, and A. Y. Ng. Efficient sparse coding algorithms. In NIPS, 2007.
[12]
D. Hubel and T. Wiesel. Receptive fields and functional architecture of monkey striate cortex. Journal of Physiology, 195:215-243, 1968.
[13]
R. L. DeValois, E. W. Yund, and N. Hepler. The orientation and direction selectivity of cells in macaque visual cortex. Vision Res., 22:531-544, 1982a.
[14]
H. B. Barlow. The coding of sensory messages. Current Problems in Animal Behavior, 1961.
[15]
P. O. Hoyer and A. Hyvarinen. A multi-layer sparse coding network learns contour coding from natural images. Vision Research, 42(12):1593-1605, 2002.
[16]
Y. Karklin and M. S. Lewicki. A hierarchical bayesian model for learning non-linear statistical regularities in non-stationary natural signals. Neural Computation, 17(2):397-423, 2005.
[17]
A. Hyvarinen and P. O. Hoyer. Emergence of phase and shift invariant features by decomposition of natural images into independent feature subspaces. Neural Computation, 12(7):1705-1720, 2000.
[18]
A. Hyvärinen, P. O. Hoyer, and M. O. Inki. Topographic independent component analysis. Neural Computation, 13(7):1527-1558, 2001.
[19]
A. Hyvarinen, M. Gutmann, and P. O. Hoyer. Statistical model of natural stimuli predicts edge-like pooling of spatial frequency channels in v2. BMC Neuroscience, 6:12, 2005.
[20]
L. Wiskott and T. Sejnowski. Slow feature analysis: Unsupervised learning of invariances. Neural Computation, 14(4):715-770, 2002.
[21]
G. Boynton and J. Hegde. Visual cortex: The continuing puzzle of area v2. Current Biology, 14(13):R523-R524, 2004.
[22]
J. B. Levitt, D. C. Kiper, and J. A. Movshon. Receptive fields and functional architecture of macaque v2. Journal of Neurophysiology, 71(6):2517-2542, 1994.
[23]
J. Hegde and D.C. Van Essen. Selectivity for complex shapes in primate visual area v2. Journal of Neuroscience, 20:RC61-66, 2000.
[24]
G. E. Hinton and R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science, 313(5786):504-507, 2006.
[25]
G. E. Hinton. Training products of experts by minimizing contrastive divergence. Neural Computation, 14:1771-1800, 2002.
[26]
R. Raina, A. Battle, H. Lee, B. Packer, and A. Y. Ng. Self-taught learning: Transfer learning from unlabeled data. In ICML, 2007.

Cited By

View all
  • (2022)Sparse DNN Model for Frequency Expanding of Higher Order Ambisonics Encoding ProcessIEEE/ACM Transactions on Audio, Speech and Language Processing10.1109/TASLP.2022.315326630(1124-1135)Online publication date: 24-Feb-2022
  • (2021)Tactile Perception for Teleoperated Robotic Exploration within Granular MediaACM Transactions on Human-Robot Interaction10.1145/345999610:4(1-27)Online publication date: 14-Jul-2021
  • (2020)Learning representations from audio-visual spatial alignmentProceedings of the 34th International Conference on Neural Information Processing Systems10.5555/3495724.3496121(4733-4744)Online publication date: 6-Dec-2020
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image Guide Proceedings
NIPS'07: Proceedings of the 21st International Conference on Neural Information Processing Systems
December 2007
1736 pages
ISBN:9781605603520

Publisher

Curran Associates Inc.

Red Hook, NY, United States

Publication History

Published: 03 December 2007

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 08 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Sparse DNN Model for Frequency Expanding of Higher Order Ambisonics Encoding ProcessIEEE/ACM Transactions on Audio, Speech and Language Processing10.1109/TASLP.2022.315326630(1124-1135)Online publication date: 24-Feb-2022
  • (2021)Tactile Perception for Teleoperated Robotic Exploration within Granular MediaACM Transactions on Human-Robot Interaction10.1145/345999610:4(1-27)Online publication date: 14-Jul-2021
  • (2020)Learning representations from audio-visual spatial alignmentProceedings of the 34th International Conference on Neural Information Processing Systems10.5555/3495724.3496121(4733-4744)Online publication date: 6-Dec-2020
  • (2020)SRS-DNN: a deep neural network with strengthening response sparsityNeural Computing and Applications10.1007/s00521-019-04309-332:12(8127-8142)Online publication date: 1-Jun-2020
  • (2019)Training auto-encoders effectively via eliminating task-irrelevant input variablesInternational Journal of Computational Science and Engineering10.5555/3337507.333750918:4(332-339)Online publication date: 1-Jan-2019
  • (2019)AttriNetAdjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers10.1145/3341162.3345600(510-517)Online publication date: 9-Sep-2019
  • (2019)A Sparse-denoising Auto-encoder with Generalization Ability on Sub-networkProceedings of the 2019 3rd International Conference on Innovation in Artificial Intelligence10.1145/3319921.3319935(229-233)Online publication date: 15-Mar-2019
  • (2019)DeviceMienProceedings of the International Conference on Internet of Things Design and Implementation10.1145/3302505.3310073(106-117)Online publication date: 15-Apr-2019
  • (2019)Unsupervised Speech Representation Learning Using WaveNet AutoencodersIEEE/ACM Transactions on Audio, Speech and Language Processing10.1109/TASLP.2019.293886327:12(2041-2053)Online publication date: 1-Dec-2019
  • (2019)Contractive Slab and Spike Convolutional Deep Belief NetworkNeural Processing Letters10.1007/s11063-018-9897-249:3(1697-1722)Online publication date: 1-Jun-2019
  • Show More Cited By

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media