Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
Skip header Section
Deep Learning: Methods and ApplicationsJune 2014
Publisher:
  • Now Publishers Inc.
  • P.O. Box 1024
  • Hanover
  • MA
  • United States
ISBN:978-1-60198-814-0
Published:12 June 2014
Pages:
212
Skip Bibliometrics Section
Reflects downloads up to 09 Nov 2024Bibliometrics
Skip Abstract Section
Abstract

Deep Learning: Methods and Applications provides an overview of general deep learning methodology and its applications to a variety of signal and information processing tasks. The application areas are chosen with the following three criteria in mind: (1) expertise or knowledge of the authors; (2) the application areas that have already been transformed by the successful use of deep learning technology, such as speech recognition and computer vision; and (3) the application areas that have the potential to be impacted significantly by deep learning and that have been benefitting from recent research efforts, including natural language and text processing, information retrieval, and multimodal information processing empowered by multi-task deep learning. Deep Learning: Methods and Applications is a timely and important book for researchers and students with an interest in deep learning methodology and its applications in signal and information processing. "This book provides an overview of a sweeping range of up-to-date deep learning methodologies and their application to a variety of signal and information processing tasks, including not only automatic speech recognition (ASR), but also computer vision, language modeling, text processing, multimodal learning, and information retrieval. This is the first and the most valuable book for "deep and wide learning" of deep learning, not to be missed by anyone who wants to know the breathtaking impact of deep learning on many facets of information processing, especially ASR, all of vital importance to our modern technological society." - Sadaoki Furui, President of Toyota Technological Institute at Chicago, and Professor at the Tokyo Institute of Technology.

Cited By

  1. ACM
    Molokwu B, Shuvo S, Kar N and Kobti Z Node Classification and Link Prediction in Social Graphs using RLVECN Proceedings of the 32nd International Conference on Scientific and Statistical Database Management, (1-10)
  2. Li G, Hu R, Zhang R and Wang X (2020). A mapping model of spectral tilt in normal-to-Lombard speech conversion for intelligibility enhancement, Multimedia Tools and Applications, 79:27-28, (19471-19491), Online publication date: 1-Jul-2020.
  3. ACM
    Adiga D, Bhavsar M, Palan U and Patel S Daily Journals Proceedings of the 14th EAI International Conference on Pervasive Computing Technologies for Healthcare, (305-315)
  4. Uddin M, Zada N, Aziz F, Saeed Y, Zeb A, Ali Shah S, Al-Khasawneh M, Mahmoud M and Stamovlasis D (2020). Prediction of Future Terrorist Activities Using Deep Neural Networks, Complexity, 2020, Online publication date: 1-Jan-2020.
  5. Welchowski T and Schmid M (2019). Sparse kernel deep stacking networks, Computational Statistics, 34:3, (993-1014), Online publication date: 1-Sep-2019.
  6. Li G, Hu R, Wang X and Zhang R (2019). A near-end listening enhancement system by RNN-based noise cancellation and speech modification, Multimedia Tools and Applications, 78:11, (15483-15505), Online publication date: 1-Jun-2019.
  7. Bengio S, Deng L, Morency L and Schuller B Perspectives on predictive power of multimodal deep learning The Handbook of Multimodal-Multisensor Interfaces, (455-472)
  8. Yin P, Xin J and Qi Y (2018). Linear Feature Transform and Enhancement of Classification on Deep Neural Network, Journal of Scientific Computing, 76:3, (1396-1406), Online publication date: 1-Sep-2018.
  9. Becerra A, De La Rosa J and González E (2018). Speech recognition in a dialog system, Multimedia Tools and Applications, 77:12, (15875-15911), Online publication date: 1-Jun-2018.
  10. Ali H, Tran S, Benetos E and D'avila Garcez A (2018). Speaker recognition with hybrid features from a deep belief network, Neural Computing and Applications, 29:6, (13-19), Online publication date: 1-Mar-2018.
  11. Lam A, Nguyen A, Nguyen H and Nguyen T Bug localization with combination of deep learning and information retrieval Proceedings of the 25th International Conference on Program Comprehension, (218-229)
  12. Ghahabi O, Hernando J, Ghahabi O and Hernando J (2017). Deep Learning Backend for Single and Multisession i-Vector Speaker Recognition, IEEE/ACM Transactions on Audio, Speech and Language Processing, 25:4, (807-817), Online publication date: 1-Apr-2017.
  13. Chen Y, Zhang H, Tong Y and Lu M (2017). Diversity Regularized Latent Semantic Match for Hashing, Neurocomputing, 230:C, (77-87), Online publication date: 22-Mar-2017.
  14. ACM
    Sturm B (2017). The “Horse” Inside, Computers in Entertainment, 14:2, (1-32), Online publication date: 30-Dec-2016.
  15. Yamada Y and Morimura T Weight features for predicting future model performance of deep neural networks Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, (2231-2237)
  16. Rere L, Fanany M and Arymurthy A (2016). Metaheuristic Algorithms for Convolution Neural Network, Computational Intelligence and Neuroscience, 2016, (2), Online publication date: 1-Jun-2016.
  17. Welchowski T and Schmid M (2016). A framework for parameter estimation and model selection in kernel deep stacking networks, Artificial Intelligence in Medicine, 70:C, (31-40), Online publication date: 1-Jun-2016.
  18. Kuleshov A and Bernstein A Extended Regression on Manifolds Estimation Proceedings of the 5th International Symposium on Conformal and Probabilistic Prediction with Applications - Volume 9653, (208-228)
  19. Ekpenyong M, Inyang U and Ekong V Intelligent Speech Features Mining for Robust Synthesis System Evaluation Human Language Technology. Challenges for Computer Science and Linguistics, (3-18)
  20. ACM
    Lane N, Bhattacharya S, Georgiev P, Forlivesi C and Kawsar F An Early Resource Characterization of Deep Learning on Wearables, Smartphones and Internet-of-Things Devices Proceedings of the 2015 International Workshop on Internet of Things towards Applications, (7-12)
  21. Kereliuk C, Sturm B and Larsen J (2015). Deep Learning and Music Adversaries, IEEE Transactions on Multimedia, 17:11, (2059-2071), Online publication date: 1-Nov-2015.
  22. Mesnil G, Dauphin Y, Yao K, Bengio Y, Deng L, Hakkani-Tur D, He X, Heck L, Tur G, Yu D and Zweig G (2015). Using recurrent neural networks for slot filling in spoken language understanding, IEEE/ACM Transactions on Audio, Speech and Language Processing, 23:3, (530-539), Online publication date: 1-Mar-2015.
  23. ACM
    Lane N and Georgiev P Can Deep Learning Revolutionize Mobile Sensing? Proceedings of the 16th International Workshop on Mobile Computing Systems and Applications, (117-122)
  24. Schmidhuber J (2015). Deep learning in neural networks, Neural Networks, 61:C, (85-117), Online publication date: 1-Jan-2015.
  25. ACM
    Levy E, David O and Netanyahu N Genetic algorithms and deep learning for automatic painter classification Proceedings of the 2014 Annual Conference on Genetic and Evolutionary Computation, (1143-1150)
Contributors
  • Microsoft Research
  • Tencent

Index Terms

  1. Deep Learning: Methods and Applications

    Reviews

    Charalambos Poullis

    Definitely not for the faint-hearted, this book does not waste any time explaining the basic concepts in machine learning; rather, it assumes that the reader is already moderately knowledgeable in the area. To make things worse, the (understandable) abundance of acronyms and external references throughout the text makes it even harder for anyone unfamiliar with the terminology and the state of the art to follow. However, it does provide a brief history about how deep learning has matured since Hinton's pivotal publications in 2006 [1,2]. What's nice about this monograph is that it introduces each "deep learning" method/architecture through an example application. These example applications are chosen from the published works that popularized the method discussed, in a particular context. For example, the deep autoencoders (a special type of deep neural networks) are introduced in the context of extracting binary speech codes from the raw speech spectrogram data. Similarly, there are case studies of applications of deep learning in the areas of audio, natural language modeling and processing, information retrieval, object recognition and computer vision, and multi-modal and multi-task learning. I found of particular interest the chapter on "Selected Applications in Object Recognition and Computer Vision," where recent successful applications of convolutional neural networks on natural images are analyzed. Over the years, the traditional approach used in computer vision was to use features extracted using scale invariant feature transform (SIFT), histograms of oriented gradients (HOG), and so on, followed by some kind of high-level feature building that then became the input to some trainable classifier, for example support vector machines. A key observation in this traditional approach is that the features are not learned but are rather handcrafted by the researchers. This has worked quite well, especially in recent years with the introduction of SIFT, speeded-up robust features (SURF), HOG, and so on; in fact, one can argue that the quality of the features often determines the success of the entire algorithm. This chapter describes how convolutional neural networks can be used to create a hierarchy of trainable feature extractors (at each layer), which "learn" the features directly from the images; instead, each layer extracts the features from the output of the previous layer. The chapter continues with a detailed description of the architecture of the deep model of the neural network that won the ImageNet 2012 object recognition competition. This is definitely not a textbook for class, but rather a reference book with example applications tailored for researchers and enthusiasts in machine learning interested in learning the details (and finding references) on how to implement and apply deep learning. On a side note, after reading the monograph, I am left wondering whether there is anything new here other than the context of the applications. Convolutional neural networks were proposed in the late 1980s [3] for digit recognition, but failed to launch in other areas due to the computationally intensive calculations. It is evident that the recent advancements in hardware, and in particular programmable graphics processing units (GPUs) and high-performance computing, have resulted in the resurfacing of these methods to make these calculations more tractable, which begs the question: What other methods that were proposed in the past and were ahead of their time should we revisit next__?__ Online Computing Reviews Service

    Access critical reviews of Computing literature here

    Become a reviewer for Computing Reviews.

    Recommendations