Title:

Learning Deep Generative Models

Department: Computer Science
Issue Date: 2-Mar-2010
Abstract (summary): Building intelligent systems that are capable of extracting high-level representations from high-dimensional sensory data lies at the core of solving many AI related tasks, including object recognition, speech perception, and language understanding. Theoretical and biological arguments strongly suggest that building such systems requires models with deep architectures that involve many layers of nonlinear processing. The aim of the thesis is to demonstrate that deep generative models that contain many layers of latent variables and millions of parameters can be learned efficiently, and that the learned high-level feature representations can be successfully applied in a wide spectrum of application domains, including visual object recognition, information retrieval, and classification and regression tasks. In addition, similar methods can be used for nonlinear dimensionality reduction.
Content Type: Thesis

Permanent link

https://hdl.handle.net/1807/19226

Items in TSpace are protected by copyright, with all rights reserved, unless otherwise indicated.