Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3292500.3330933acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Environment Reconstruction with Hidden Confounders for Reinforcement Learning based Recommendation

Published: 25 July 2019 Publication History
  • Get Citation Alerts
  • Abstract

    Reinforcement learning aims at searching the best policy model for decision making, and has been shown powerful for sequential recommendations. The training of the policy by reinforcement learning, however, is placed in an environment. In many real-world applications, however, the policy training in the real environment can cause an unbearable cost, due to the exploration in the environment. Environment reconstruction from the past data is thus an appealing way to release the power of reinforcement learning in these applications. The reconstruction of the environment is, basically, to extract the casual effect model from the data. However, real-world applications are often too complex to offer fully observable environment information. Therefore, quite possibly there are unobserved confounding variables lying behind the data. The hidden confounder can obstruct an effective reconstruction of the environment. In this paper, by treating the hidden confounder as a hidden policy, we propose a deconfounded multi-agent environment reconstruction (DEMER) approach in order to learn the environment together with the hidden confounder. DEMER adopts a multi-agent generative adversarial imitation learning framework. It proposes to introduce the confounder embedded policy, and use the compatible discriminator for training the policies. We then apply DEMER in an application of driver program recommendation. We firstly use an artificial driver program recommendation environment, abstracted from the real application, to verify and analyze the effectiveness of DEMER. We then test DEMER in the real application of Didi Chuxing. Experiment results show that DEMER can effectively reconstruct the hidden confounder, and thus can build the environment better. DEMER also derives a recommendation policy with a significantly improved performance in the test phase of the real application.

    References

    [1]
    Brenna Argall, Sonia Chernova, Manuela M. Veloso, and Brett Browning. 2009. A survey of robot learning from demonstration. Robotics and Autonomous Systems, Vol. 57, 5 (2009), 469--483.
    [2]
    Elias Bareinboim, Andrew Forney, and Judea Pearl. 2015. Bandits with Unobserved Confounders: A Causal Approach. In Advances in Neural Information Processing Systems 28. 1342--1350.
    [3]
    Chelsea Finn, Paul F. Christiano, Pieter Abbeel, and Sergey Levine. 2016. A Connection between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy-Based Models. arXiv, Vol. abs/1611.03852 (2016).
    [4]
    Andrew Forney, Judea Pearl, and Elias Bareinboim. 2017. Counterfactual Data-Fusion for Online Reinforcement Learners. In Proceedings of the 34th International Conference on Machine Learning. 1156--1164.
    [5]
    Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. 2014. Generative Adversarial Nets. In Advances in Neural Information Processing Systems 27. 2672--2680.
    [6]
    Jonathan Ho and Stefano Ermon. 2016. Generative Adversarial Imitation Learning. In Advances in Neural Information Processing Systems 29. 4565--4573.
    [7]
    Christos Louizos, Uri Shalit, Joris M. Mooij, David Sontag, Richard S. Zemel, and Max Welling. 2017. Causal Effect Inference with Deep Latent-Variable Models. In Advances in Neural Information Processing Systems 30. 6449--6459.
    [8]
    Chaochao Lu, Bernhard Schö lkopf, and José Miguel Herná ndez-Lobato. 2018. Deconfounding Reinforcement Learning in Observational Settings. arXiv, Vol. abs/1812.10576 (2018).
    [9]
    Jacob Menick and Nal Kalchbrenner. 2018. Generating High Fidelity Images with Subscale Pixel Networks and Multidimensional Upscaling. arXiv, Vol. abs/1812.01608 (2018).
    [10]
    Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin A. Riedmiller, Andreas Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. 2015. Human-level control through deep reinforcement learning. Nature, Vol. 518 (2015), 529--533.
    [11]
    Judea Pearl. 2009. Causal inference in statistics: An overview. Statistics surveys, Vol. 3 (2009), 96--146.
    [12]
    Dean Pomerleau. 1991. Efficient Training of Artificial Neural Networks for Autonomous Navigation. Neural Computation, Vol. 3, 1 (1991), 88--97.
    [13]
    Sté phane Ross, Geoffrey J. Gordon, and Drew Bagnell. 2011. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics . 627--635.
    [14]
    Stuart J. Russell. 1998. Learning Agents for Uncertain Environments (Extended Abstract). In Proceedings of the Eleventh Annual Conference on Computational Learning Theory . 101--103.
    [15]
    Stefan Schaal. 1999. Is imitation learning the route to humanoid robots? Trends in cognitive sciences, Vol. 3, 6 (1999), 233--242.
    [16]
    John Schulman, Sergey Levine, Pieter Abbeel, Michael I. Jordan, and Philipp Moritz. 2015. Trust Region Policy Optimization. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6--11 July 2015 . 1889--1897.
    [17]
    Jing-Cheng Shi, Yang Yu, Qing Da, Shi-Yong Chen, and Anxiang Zeng. 2018. Virtual-Taobao: Virtualizing Real-world Online Retail Environment for Reinforcement Learning. arXiv, Vol. abs/1805.10000 (2018).
    [18]
    David Silver, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre, George van den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Vedavyas Panneershelvam, Marc Lanctot, Sander Dieleman, Dominik Grewe, John Nham, Nal Kalchbrenner, Ilya Sutskever, Timothy P. Lillicrap, Madeleine Leach, Koray Kavukcuoglu, Thore Graepel, and Demis Hassabis. 2016. Mastering the game of Go with deep neural networks and tree search. Nature, Vol. 529, 7587 (2016), 484--489.
    [19]
    Richard S. Sutton and Andrew G. Barto. 2018. Reinforcement Learning: An Introduction (2nd Edition) .MIT Press.
    [20]
    Zeyang Ye, Keli Xiao, Yong Ge, and Yuefan Deng. 2019. Applying Simulated Annealing and Parallel Computing to the Mobile Sequential Recommendation. IEEE Transactions on Knowledge and Data Engineering, Vol. 31, 2 (2019), 243--256.
    [21]
    Zeyang Ye, Lihao Zhang, Keli Xiao, Wenjun Zhou, Yong Ge, and Yuefan Deng. 2018. Multi-User Mobile Sequential Recommendation: An Efficient Parallel Computing Paradigm. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining . 2624--2633.

    Cited By

    View all
    • (2024)Cost-aware Offline Safe Meta Reinforcement Learning with Robust In-Distribution Online Task AdaptationProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3662927(743-751)Online publication date: 6-May-2024
    • (2024)Self-Interested Coalitional Crowdsensing for Multi-Agent Interactive Environment MonitoringSensors10.3390/s2402050924:2(509)Online publication date: 14-Jan-2024
    • (2024)A Mixed Generative Adversarial Imitation Learning Based Vehicle Path Planning AlgorithmIEEE Access10.1109/ACCESS.2024.341210912(85859-85879)Online publication date: 2024
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '19: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
    July 2019
    3305 pages
    ISBN:9781450362016
    DOI:10.1145/3292500
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 July 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. environment reconstruction
    2. hidden confounder
    3. imitation learning
    4. recommendation
    5. reinforcement learning

    Qualifiers

    • Research-article

    Funding Sources

    • NSFC
    • National Key R&D Program of China
    • Jiangsu SF

    Conference

    KDD '19
    Sponsor:

    Acceptance Rates

    KDD '19 Paper Acceptance Rate 110 of 1,200 submissions, 9%;
    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Upcoming Conference

    KDD '24

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)50
    • Downloads (Last 6 weeks)11

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Cost-aware Offline Safe Meta Reinforcement Learning with Robust In-Distribution Online Task AdaptationProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3662927(743-751)Online publication date: 6-May-2024
    • (2024)Self-Interested Coalitional Crowdsensing for Multi-Agent Interactive Environment MonitoringSensors10.3390/s2402050924:2(509)Online publication date: 14-Jan-2024
    • (2024)A Mixed Generative Adversarial Imitation Learning Based Vehicle Path Planning AlgorithmIEEE Access10.1109/ACCESS.2024.341210912(85859-85879)Online publication date: 2024
    • (2023)Adversarial counterfactual environment model learningProceedings of the 37th International Conference on Neural Information Processing Systems10.5555/3666122.3669219(70654-70706)Online publication date: 10-Dec-2023
    • (2023)Understanding or Manipulation: Rethinking Online Performance Gains of Modern Recommender SystemsACM Transactions on Information Systems10.1145/363786942:4(1-32)Online publication date: 15-Dec-2023
    • (2023)Deconfounded Causal Collaborative FilteringACM Transactions on Recommender Systems10.1145/36060351:4(1-25)Online publication date: 3-Oct-2023
    • (2023)A Unified Representation Framework for Rideshare Marketplace Equilibrium and EfficiencyProceedings of the 31st ACM International Conference on Advances in Geographic Information Systems10.1145/3589132.3625581(1-11)Online publication date: 13-Nov-2023
    • (2023)Addressing Confounding Feature Issue for Causal RecommendationACM Transactions on Information Systems10.1145/355975741:3(1-23)Online publication date: 7-Feb-2023
    • (2023)Offline Model-Based Adaptable Policy Learning for Decision-Making in Out-of-Support RegionsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2023.331713145:12(15260-15274)Online publication date: Dec-2023
    • (2023)Generative Adversarial Reward Learning for Generalized Behavior Tendency InferenceIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2022.318692035:10(9878-9889)Online publication date: 1-Oct-2023
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media