Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3340531.3412170acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
tutorial

Neural Bayesian Information Processing

Published: 19 October 2020 Publication History

Abstract

Deep learning is developed as a learning process from source inputs to target outputs where the inference or optimization is performed over an assumed deterministic model with deep structure. A wide range of temporal and spatial data in language and vision are treated as the inputs or outputs to build such a complicated mapping in different information systems. A systematic and elaborate transfer is required to meet the mapping between source and target domains. Also, the semantic structure in natural language and computer vision may not be well represented or trained in mathematical logic or computer programs. The distribution function in discrete or continuous latent variable model for words, sentences, images or videos may not be properly decomposed or estimated. The system robustness to heterogeneous environments may not be assured. This tutorial addresses the fundamentals and advances in statistical models and neural networks, and presents a series of deep Bayesian solutions including variational Bayes, sampling method, Bayesian neural network, variational auto-encoder (VAE), stochastic recurrent neural network, sequence-to-sequence model, attention mechanism, end-to-end network, stochastic temporal convolutional network, temporal difference VAE, normalizing flow and neural ordinary differential equation. Enhancing the prior/posterior representation is addressed in different latent variable models. We illustrate how these models are connected and why they work for a variety of applications on complex patterns in language and vision. The word, sentence and image embeddings are merged with semantic constraint or structural information. Bayesian learning is formulated in the optimization procedure where the posterior collapse is tackled. An informative latent space is trained to incorporate deep Bayesian learning in various information systems.

Supplementary Material

MP4 File (3340531.3412170.mp4)
This is ACM CIKM 2020 Tutorial on Neural Bayesian Information Processing. The tutorial speaker is Prof. Jen-Tzung Chien from National Chiao Tung University. This tutorial focuses on three main themes including domain mapping, probabilistic model and neural network. The solutions of multimedia information processing based on deep Bayesian learning are surveyed and organized in this presentation. Domain mapping between source and target data is complicated with different multimedia contents. A variety of deep models are addressed to handle the flexibility, robustness and heterogeneity in mapping between two domains ranging from text to speech, image and video in different information systems. This tutorial consists of four parts including multimedia information processing, deep learning & modeling, deep Bayesian learning and summarization & future trend.

References

[1]
Emre Aksan and Otmar Hilliges. 2019. STCN: Stochastic Temporal Convolutional Networks. In Proc. of International Conference on Learning Representations. 2019.
[2]
Shiyu Chang, Yang Zhang, Wei Han, Mo Yu, Xiaoxiao Guo, Wei Tan, Xiaodong Cui, Michael Witbrock, Mark A Hasegawa-Johnson, and Thomas S Huang. 2017. Dilated Recurrent Neural Networks. In Advances in Neural Information Processing Systems. 77--87.
[3]
Ricky T. Q. Chen, Yulia Rubanova, Jesse Bettencourt, and David K. Duvenaud. 2018. Neural Ordinary Differential Equations. In Advances in Neural Information Processing Systems. 6571--6583.
[4]
Jen-Tzung Chien. 2015a. Hierarchical Pitman-Yor-Dirichlet Language Model. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 23, 8 (2015), 1259--1272.
[5]
Jen-Tzung Chien. 2015b. Laplace Group Sensing for Acoustic Models. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 23, 5 (2015), 909--922.
[6]
Jen-Tzung Chien. 2016. Hierarchical Theme and Topic Modeling. IEEE Transactions on Neural Networks and Learning Systems, Vol. 27, 3 (2016), 565--578.
[7]
Jen-Tzung Chien. 2018. Bayesian Nonparametric Learning for Hierarchical and Sparse Topics. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 26, 2 (2018), 422--435.
[8]
Jen-Tzung Chien and Yi-Ting Bao. 2018. Tensor-Factorized Neural Networks. IEEE Transactions on Neural Networks and Learning Systems, Vol. 29, 5 (2018), 1998--2011.
[9]
Jen-Tzung Chien and Chuang-Hua Chueh. 2011. Dirichlet Class Language Models for Speech Recognition. IEEE Transactions on Audio, Speech, and Language Processing, Vol. 19, 3 (2011), 482--495.
[10]
Jen-Tzung Chien and Yuan-Chu Ku. 2016. Bayesian Recurrent Neural Network for Language Modeling. IEEE Transactions on Neural Networks and Learning Systems, Vol. 27, 2 (2016), 361--374.
[11]
Jen-Tzung Chien and Chao-Hsi Lee. 2018. Deep Unfolding for Topic Models. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 40, 2 (2018), 318--331.
[12]
Jen-Tzung Chien and Ting-An Lin. 2018. Supportive Attention in End-To-End Memory Networks. In Proc. of IEEE International Workshop on Machine Learning for Signal Processing. 1--6.
[13]
Jen-Tzung Chien and Chun-Wei Wang. 2019 a. Self Attention in Variational Sequential Learning for Summarization. Proc. of Annual Conference of International Speech Communication Association (2019), 1318--1322.
[14]
Jen-Tzung Chien and Chun-Wei Wang. 2019 b. Variational and Hierarchical Recurrent Autoencoder. In Proc. of IEEE International Conference on Acoustics, Speech and Signal Processing. 3202--3206.
[15]
Junyoung Chung, Kyle Kastner, Laurent Dinh, Kratarth Goel, Aaron C Courville, and Yoshua Bengio. 2015. A Recurrent Latent Variable Model for Sequential Data. In Advances in Neural Information Processing Systems. 2980--2988.
[16]
Yann N. Dauphin, Angela Fan, Michael Auli, and David Grangier. 2017. Language Modeling with Gated Convolutional Networks. In Proc. of International Conference on Machine Learning. 933--941.
[17]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv preprint arXiv:1810.04805 (2018).
[18]
Marco Fraccaro, Sø ren Kaae Sø nderby, Ulrich Paquet, and Ole Winther. 2016. Sequential Neural Models with Stochastic Layers. In Advances in Neural Information Processing Systems. 2199--2207.
[19]
Yarin Gal and Zoubin Ghahramani. 2016. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks. In Advances in Neural Information Processing Systems. 1019--1027.
[20]
Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, and Yann N Dauphin. 2017. Convolutional Sequence to Sequence Learning. In Proc. of International Conference on Machine Learning. 1243--1252.
[21]
Anirudh Goyal, Alessandro Sordoni, Marc-Alexandre Côté, Nan Ke, and Yoshua Bengio. 2017. Z-Forcing: Training Stochastic Recurrent Networks. In Advances in Neural Information Processing Systems 30. 6713--6723.
[22]
Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. 2006. Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks. In Proc. of International Conference on Machine Learning. 369--376.
[23]
Karol Gregor, George Papamakarios, Frederic Besse, Lars Buesing, and Theophane Weber. 2019. Temporal Difference Variational Auto-Encoder. In Proc. of International Conference on Learning Representations.
[24]
Eric Jang, Shixiang Gu, and Ben Poole. 2017. Categorical Reparameterization with Gumbel-Softmax. In Proc. of International Conference on Learning Representations.
[25]
Nal Kalchbrenner, Edward Grefenstette, and Phil Blunsom. 2014. A Convolutional Neural Network for Modelling Sentences. In Proc. of Annual Meeting of the Association for Computational Linguistics. 655--665.
[26]
Diederik P. Kingma and Prafulla Dhariwal. 2018. Glow: Generative Flow with Invertible 1x1 Convolutions. In Neural Information Processing Systems. 10215--10224.
[27]
Diederik P. Kingma and Max Welling. 2014. Auto-encoding variational Bayes. In Proc. of International Conference on Learning Representations.
[28]
Che-Yu Kuo and Jen-Tzung Chien. 2018. Markov Recurrent Neural Networks. In Proc. of IEEE International Workshop on Machine Learning for Signal Processing.
[29]
Radford M. Neal. 2012. Bayesian Learning for Neural Networks. Vol. 118. Springer Science & Business Media.
[30]
Danilo Jimenez Rezende and Shakir Mohamed. 2015. Variational inference with normalizing flows. In Proc. of International Conference on Machine Learning. 1530--1538.
[31]
Iulian V. Serban, Alessandro Sordoni, Ryan Lowe, Laurent Charlin, Joelle Pineau, Aaron Courville, and Yoshua Bengio. 2017. A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues. In Proc. of AAAI Conference on Artificial Intelligence. 3295--3301.
[32]
Jakub Tomczak and Max Welling. 2018. VAE with a VampPrior. In International Conference on Artificial Intelligence and Statistics. 1214--1223.
[33]
Youzhi Tu, Man-Wai Mak, and Jen-Tzung Chien. 2020. Variational Domain Adversarial Learning With Mutual Information Maximization for Speaker Verification. IEEE/ACM Transactions on Audio, Speech, and Language Processing, Vol. 28 (2020).
[34]
Aaron Van Den Oord, Oriol Vinyals, and Koray Kavukcuoglu. 2017. Neural Discrete Representation Learning. In Advances in Neural Information Processing Systems. 6309--6318.
[35]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is All you Need. In Advances in Neural Information Processing Systems. 5998--6008.
[36]
Shinji Watanabe and Jen-Tzung Chien. 2015. Bayesian Speech and Language Processing Cambridge University Press.
[37]
Jason Weston, Sumit Chopra, and Antoine Bordes. 2015. Memory Networks. In Proc. of International Conference on Learning Representation.
[38]
Shi Xingjian, Zhourong Chen, Hao Wang, Dit-Yan Yeung, Wai-kin Wong, and Wang-chun Woo. 2015. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In Advances in Neural Information Processing Systems. 802--810.
[39]
Dong Yu, Geoffrey Hinton, Nelson Morgan, Jen-Tzung Chien, and Shigeki Sagayama. 2011. Introduction to the Special Section on Deep Learning for Speech and Language Processing. IEEE Transactions on Audio, Speech, and Language Processing, Vol. 20, 1 (2011), 4--6.

Cited By

View all
  • (2023)Variational Skill Embeddings for Meta Reinforcement Learning2023 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN54540.2023.10191425(1-8)Online publication date: 18-Jun-2023

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
CIKM '20: Proceedings of the 29th ACM International Conference on Information & Knowledge Management
October 2020
3619 pages
ISBN:9781450368599
DOI:10.1145/3340531
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 October 2020

Check for updates

Author Tags

  1. bayesian learning
  2. deep learning
  3. information processing
  4. probabilistic representation
  5. sequential learning

Qualifiers

  • Tutorial

Funding Sources

  • Ministry of Science and Technology, Taiwan

Conference

CIKM '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,861 of 8,427 submissions, 22%

Upcoming Conference

CIKM '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)5
  • Downloads (Last 6 weeks)0
Reflects downloads up to 03 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2023)Variational Skill Embeddings for Meta Reinforcement Learning2023 International Joint Conference on Neural Networks (IJCNN)10.1109/IJCNN54540.2023.10191425(1-8)Online publication date: 18-Jun-2023

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media