Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.5555/3237383.3238107acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Characterizing the Limits of Autonomous Systems

Published: 09 July 2018 Publication History

Abstract

This note answers two central question in the intersection of decision-making and causal inference -- when human input is needed and, if so, how it should be incorporated into an AI system. We introduce the counterfactual agent who proactively considers human input in its decision-making process. We prove that a counterfactual agent dominates the standard autonomous agent who does not consider any human input (i.e., the experimental agent) in terms of performance. These results suggest a trade-off between autonomy and optimality -- while the full autonomy is often preferred, using human input could potentially improve the efficiency of the system. We further characterize under what conditions experimental and counterfactual agents can reach the same level of performance, which elicits the settings where full autonomy can be achieved.

References

[1]
Elias Bareinboim, Andrew Forney, and Judea Pearl. 2015. Bandits with unobserved confounders: A causal approach Advances in Neural Information Processing Systems. 1342--1350.
[2]
E. Bareinboim and J. Pearl. 2016. Causal inference and the data-fusion problem. Proceedings of the National Academy of Sciences Vol. 113 (2016), 7345--7352. Issue 27.
[3]
Anthony Cassandra, Michael L Littman, and Nevin L Zhang. 1997. Incremental pruning: A simple, fast, exact method for partially observable Markov decision processes. In Proceedings of the 13th Conference on Uncertainty in Artificial Intelligence (UAI). Morgan Kaufmann, 54--61.
[4]
Hsien-Te Cheng. 1988. Algorithms for partially observable Markov decision processes. Ph.D. Dissertation. University of British Columbia.
[5]
Nicolás Della Penna, Mark D Reid, and David Balduzzi. 2016. Compliance-Aware Bandits. arXiv preprint arXiv:1602.02852 (2016).
[6]
A. Forney, J. Pearl, and E. Bareinboim. 2017. Counterfactual Data-Fusion for Online Reinforcement Learners Proceedings of the 34th International Conference on Machine Learning (Proceedings of Machine Learning Research), Doina Precup and Yee Whye Teh (Eds.), Vol. Vol. 70. 1156--1164.
[7]
J.Y. Halpern. 1998. Axiomatizing Causal Reasoning. Uncertainty in Artificial Intelligence, bibfieldeditorG.F. Cooper and S. Moral (Eds.). Morgan Kaufmann, San Francisco, CA, 202--210. Also, Journal of Artificial Intelligence Research 12:3, 17--37, 2000.
[8]
Finnian Lattimore, Tor Lattimore, and Mark D Reid. 2016. Causal Bandits: Learning Good Interventions via Causal Inference Advances in Neural Information Processing Systems. 1181--1189.
[9]
Michael L Littman. 1994. The witness algorithm: Solving partially observable Markov decision processes. (1994).
[10]
J. Pearl. 2000. Causality: Models, Reasoning, and Inference. Cambridge University Press, New York.
[11]
Martin L Puterman. 2014. Markov decision processes: discrete stochastic dynamic programming. Wiley & Sons.
[12]
Diederik M Roijers, Peter Vamplew, Shimon Whiteson, Richard Dazeley, et almbox. 2013. A Survey of Multi-Objective Sequential Decision-Making. J. Artif. Intell. Res.(JAIR) Vol. 48 (2013), 67--113.
[13]
David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et almbox. 2016. Mastering the game of Go with deep neural networks and tree search. Nature, Vol. 529, 7587 (2016), 484--489.
[14]
Satinder P Singh, Tommi S Jaakkola, and Michael I Jordan. 1994. Learning Without State-Estimation in Partially Observable Markovian Decision Processes. ICML. 284--292.
[15]
Richard D Smallwood and Edward J Sondik. 1973. The optimal control of partially observable Markov processes over a finite horizon. Operations Research, Vol. 21, 5 (1973), 1071--1088.
[16]
Chris Urmson, Joshua Anhalt, Drew Bagnell, Christopher Baker, Robert Bittner, MN Clark, John Dolan, Dave Duggins, Tugrul Galatali, Chris Geyer, et almbox. 2008. Autonomous driving in urban environments: Boss and the urban challenge. Journal of Field Robotics Vol. 25, 8 (2008), 425--466.
[17]
Junzhe Zhang and Elias Bareinboim. 2017. Transfer Learning in Multi-Armed Bandits: A Causal Approach Proceedings of the 26th International Joint Conference on Artificial Intelligence (IJCAI).
[18]
Shlomo Zilberstein. 2015. Building Strong Semi-autonomous Systems. In Proceedings of the Twenty-Ninth AAAI Conference. AAAI Press, 4088--4092.

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
AAMAS '18: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems
July 2018
2312 pages

Sponsors

In-Cooperation

Publisher

International Foundation for Autonomous Agents and Multiagent Systems

Richland, SC

Publication History

Published: 09 July 2018

Check for updates

Author Tags

  1. human-robot/agent interaction
  2. knowledge representation and reasoning

Qualifiers

  • Research-article

Conference

AAMAS '18
Sponsor:
AAMAS '18: Autonomous Agents and MultiAgent Systems
July 10 - 15, 2018
Stockholm, Sweden

Acceptance Rates

AAMAS '18 Paper Acceptance Rate 149 of 607 submissions, 25%;
Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 91
    Total Downloads
  • Downloads (Last 12 months)3
  • Downloads (Last 6 weeks)0
Reflects downloads up to 25 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media