Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3447548.3470803acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
abstract

From Deep Learning to Deep Reasoning

Published: 14 August 2021 Publication History

Abstract

The rise of big data and big compute has brought modern neural networks to many walks of digital life, thanks to the relative ease of constructing large models that scale to the real world. Current successes of Transformers and self-supervised pretraining on massive data have led some to believe that deep neural networks will be able to do almost everything once we have sufficient data and computational resources. However, neural networks are fast to exploit surface statistics but fail miserably to generalize to novel combinations. This is because they are not designed for deliberate reasoning -- the capacity to deliberately deduce new knowledge out of the contextualized data. This tutorial reviews recent developments to extend the capacity of neural networks to "learning-to-reason'' from data, where the task is to determine if the data entails a conclusion. This capacity opens up new ways to generate insights from data through arbitrary compositional querying without the need of predefining a narrow set of tasks. The tutorial consists of four parts. The first part covers the learning-to-reason framework, and explains how neural networks can serve as a strong backbone for reasoning through its natural operations such as binding, attention & dynamic computational graphs. The second part goes into more detail on how neural networks perform reasoning over unstructured and structured data, and across modalities. The third part reviews neural memories and their role in reasoning. The last part discusses generalization to novel combinations, under less supervision and with more knowledge.

References

[1]
Somak Aditya, Yezhou Yang, and Chitta Baral. 2019. Integrating knowledge and reasoning in image understanding. In 28th International Joint Conference on Artificial Intelligence, IJCAI 2019 . 6252--6259.
[2]
Saeed Amizadeh, Hamid Palangi, Oleksandr Polozov, Yichen Huang, and Kazuhito Koishida. 2020. Neuro-Symbolic Visual Reasoning: Disentangling" Visual" from" Reasoning". ICML (2020).
[3]
Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and Dan Klein. 2016. Neural module networks. In CVPR. 39--48.
[4]
Dzmitry Bahdanau, Shikhar Murty, Michael Noukhovitch, Thien Huu Nguyen, Harm de Vries, and Aaron Courville. 2019. Systematic generalization: what is required and can it be learned? ICLR (2019).
[5]
Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, et almbox. 2018. Relational inductive biases, deep learning, and graph networks. arXiv preprint arXiv:1806.01261 (2018).
[6]
Léon Bottou. 2014. From machine learning to machine reasoning. Machine Learning, Vol. 94, 2 (2014), 133--149.
[7]
Cameron Buckner and James Garson. 2019. Connectionism. In The Stanford Encyclopedia of Philosophy fall 2019 ed.), Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University.
[8]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT 2019 . 4171--4186.
[9]
Aidan Feeney and Valerie A Thompson. 2014. Reasoning as memory .Psychology Press.
[10]
Marta Garnelo and Murray Shanahan. 2019. Reconciling deep learning with symbolic artificial intelligence: representing objects and relations. Current Opinion in Behavioral Sciences, Vol. 29 (2019), 17--23.
[11]
Alex Graves, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwi'nska, Sergio Gómez Colmenarejo, Edward Grefenstette, Tiago Ramalho, John Agapiou, et almbox. 2016. Hybrid computing using a neural network with dynamic external memory. Nature, Vol. 538, 7626 (2016), 471--476.
[12]
Klaus Greff, Sjoerd van Steenkiste, and Jürgen Schmidhuber. 2020. On the Binding Problem in Artificial Neural Networks. arXiv preprint arXiv:2012.05208 (2020).
[13]
Drew Hudson and Christopher D Manning. 2019. Learning by abstraction: The neural state machine. In NeurIPS. 5901--5914.
[14]
Drew A Hudson and Christopher D Manning. 2018. Compositional Attention Networks for Machine Reasoning. ICLR (2018).
[15]
Daniel Kahneman. 2011. Thinking, fast and slow .Farrar, Straus and Giroux New York.
[16]
Roni Khardon and Dan Roth. 1997. Learning to reason. Journal of the ACM (JACM), Vol. 44, 5 (1997), 697--725.
[17]
Alex Konkel and Neal J Cohen. 2009. Relational memory and the hippocampus: representations and methods. Frontiers in neuroscience, Vol. 3 (2009), 23.
[18]
Hung Le, Truyen Tran, and Svetha Venkatesh. 2020 c. Neural Stored-program Memory. In ICLR 2020: Proceedings of the 8th International Conference on Learning Representations .
[19]
Hung Le, Truyen Tran, and Svetha Venkatesh. 2020 d. Self-Attentive Associative Memory. In ICML, Vol. 119. 5682--5691.
[20]
Thao Minh Le, Vuong Le, Svetha Venkatesh, and Truyen Tran. 2020 a. Dynamic language binding in relational visual reasoning. In IJCAI. 818--824.
[21]
Thao Minh Le, Vuong Le, Svetha Venkatesh, and Truyen Tran. 2020 b. Neural reasoning, fast and slow, for video question answering. In 2020 International Joint Conference on Neural Networks (IJCNN). 1--8. https://doi.org/10.1109/IJCNN48605.2020.9207580
[22]
Xiao Liu, Fanjin Zhang, Zhenyu Hou, Zhaoyu Wang, Li Mian, Jing Zhang, and Jie Tang. 2020. Self-supervised learning: Generative or contrastive. arXiv preprint arXiv:2006.08218, Vol. 1, 2 (2020).
[23]
Rasmus Palm, Ulrich Paquet, and Ole Winther. 2018. Recurrent relational networks. In NeurIPS. 3368--3378.
[24]
Ethan Perez, Florian Strub, Harm De Vries, Vincent Dumoulin, and Aaron Courville. 2018. Film: Visual reasoning with a general conditioning layer. In AAAI .
[25]
Trang Pham, Truyen Tran, Dinh Phung, and Svetha Venkatesh. 2017. Column Networks for Collective Classification. In Proceedings of AAAI Conference on Artificial Intelligence, Vol. 31. 2485--2491.
[26]
Adam Santoro, Ryan Faulkner, David Raposo, Jack Rae, Mike Chrzanowski, Theophane Weber, Daan Wierstra, Oriol Vinyals, Razvan Pascanu, and Timothy Lillicrap. 2018. Relational recurrent neural networks. NIPS (2018).
[27]
Adam Santoro, David Raposo, David G Barrett, Mateusz Malinowski, Razvan Pascanu, Peter Battaglia, and Tim Lillicrap. 2017. A simple neural network module for relational reasoning. In NIPS. 4974--4983.
[28]
Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, and Hannaneh Hajishirzi. 2017. Bidirectional attention flow for machine comprehension. ICLR (2017).
[29]
Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, and Rob Fergus. 2015. End-To-End Memory Networks. NIPS (2015).
[30]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems. 5998--6008.
[31]
Petar Velivc ković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2018. Graph Attention Networks. ICLR (2018).
[32]
Keyulu Xu, Jingling Li, Mozhi Zhang, Simon S Du, Ken-ichi Kawarabayashi, and Stefanie Jegelka. 2019. What Can Neural Networks Reason About? arXiv preprint arXiv:1905.13211 (2019).

Cited By

View all
  • (2022)Machine learning for next‐generation intelligent transportation systemsTransactions on Emerging Telecommunications Technologies10.1002/ett.442733:4Online publication date: 17-Apr-2022

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining
August 2021
4259 pages
ISBN:9781450383325
DOI:10.1145/3447548
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 August 2021

Check for updates

Author Tags

  1. deep learning
  2. dynamic neural networks
  3. learning-to-reason
  4. machine reasoning

Qualifiers

  • Abstract

Conference

KDD '21
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Upcoming Conference

KDD '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)24
  • Downloads (Last 6 weeks)1
Reflects downloads up to 25 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2022)Machine learning for next‐generation intelligent transportation systemsTransactions on Emerging Telecommunications Technologies10.1002/ett.442733:4Online publication date: 17-Apr-2022

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media