Abstract
Programming a robot to act intelligently is a challenging endeavor beyond the skill level of most people. Trained roboticists generally program robots for a single purpose. Enabling robots to be programmed by non-experts and to perform multiple tasks are both open challenges in robotics. This paper presents a framework that allows life-long robot task learning from demonstrations. To make that possible, the paper introduces a task representation based on influence diagrams, and a method to transfer knowledge between similar tasks. A novel approach to influence diagram learning is presented along with a demonstration method that allows non-experts to teach tasks to the robot in an intuitive manner. The results from three user studies validate that the approach enables both a simulated and a physical robot to learn complex tasks from a variety of teachers, refining those tasks during on-line performance, successfully completing the tasks in different environments, and transferring knowledge from one task to another.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Argall, B. D., Chernova, S., Veloso, M., & Browning, B. (2009). A survey of robot learning from demonstration. Robotics and Autonomous Systems, 57(5), 469–483.
Atkeson, C., & Schaal, S. (1997). Robot learning from demonstration. In Proceedings of the Fourteenth International Conference on Machine Learning.
Bentivegna, D. (2004). Learning from observation using primitives. PhD thesis, College of Computing, Georgia Institute of Technology, Atlanta, GA.
Billard, A., Calinon, S., Dillmann, R., & Schaal, S. (2008). Robot programming by demonstration, Chapter 59. Cambridge: MIT Press.
Bransford, J. D., Brown, A. L., & Cocking, R. R. (1999). How people learn: Brain, mind, experience, and school. Washington, D.C.: National Academy Press.
Brunner, C., Peynot, T., & Underwood, J. (2009). Towards discrimination of challenging conditions for ugvs with visual and infrared sensors’. In Proceedings of 2009 Australasian Conference on Robotics and Automation (ACRA), Sydney, Australia, Decemeber 2009.
Chernova, S., & M. Veloso. (2007). Confidence-based policy learning from demonstration using gaussian mixture models. In International Conference on Autonomous Agents and Multiagent Systems.
Chernova, S., & Veloso, M. (2008). Teaching collaborative multi-robot tasks through demonstration. In IEEE-RAS International Conference on Humanoid Robots, Daejeon, Korea.
Churchill, W., & Newman, P. (2012). Practice makes perfect? managing and leveraging visual experiences for lifelong navigation. In Proc. IEEE International Conference on Robotics and Automation (ICRA2012), Minnesota, USA.
Ciocarlie, M. (2012). Ros table top perception, 8 2012. http://www.ros.org/wiki/tabletop_object_perception.
Dillmann, R., Kaiser, M. & Ude, A. (1995). Acquisition of elementary robot skills from human demonstration. In International Symposium on Intelligent Robotic Systems.
Endsley, M. R. (1988). Design and evaluation for situation awareness enhancement. In Proceedings of the Human Factors Society 32nd Annual Meeting, Santa Monica, CA: Human Factors Society.
Garland A., & Lesh, N. (2003). Learning hierarchical taskmodels by demonstration. Technical Report Technical Report TR2003-01, Mitsubishi Electric Research Laboratories.
Gold, K. (2009) An information pipeline model of human-robot interaction. In Proceedings of the 4th ACM/IEEE international conference on Human robot interaction (pp. 85–92). New York, NY, USA: ACM. ISBN 978-1-60558-404-1. doi:10.1145/1514095.1514113.
Grollman, D. & Jenkins, O. (2007a). Dogged learning for robots. In Proceedings of the IEEE International Conference on Robotics and Automation.
Grollman, D. H. & Jenkins, O. C. (2007b). Learning robot soccer skills from demon-stration. In International Conference on Development and Learning (pp 276–281), London, UK.
Grollman, D. H. & Jenkins, O. C. (2008). Sparse incremental learning for interactive robot control policy estimation. In International Conference on Robotics and Automation (pp 3315–3320) Pasadena, CA, USA.
Hentschel, M., & Wagner, B. (2011). An adaptive memory model for long-term navigation of autonomous mobile robots. Journal of Robotics, 2011, 506245. doi:10.1155/2011/506245.
Hersch, M., Guenter, F., Calinon, S., & Billard, A. (2008). Dynamical system modulation for robot learning via kinesthetic demonstrations. IEEE Transactions on Robotics, 24(6), 1463–1467.
Ijspeert, J. A., Nakanishi, J. & Schaal, S. (2002). Movement imitation with nonlinear dynamical systems in humanoid robots. In international conference on robotics and automation (icra2002).
Kaelbling, L. P., Littman, M. L., & Moore, A. W. (1996). Reinforcement learning: A survey. Journal of Artificial Intelligence Research, 4, 237–285.
Koenig, N. (2012). Robot life-long task learning from human demonstrations: A bayesian approach. Cres-12-002. Los Angeles, CA: University of Southern California.
Koenig, N. & Howard, A. (2004). Design and use paradigms for gazebo, an open-source multi-robot simulator. In IEEE/RSJ International Conference on Intelligent Robots and Systems (pp 2149–2154) Sendai, Japan.
Koenig, N., Takayama, L., & Matarić, M. (2010). Communication and knowledge sharing in human-robot interaction. Neural Networks, 23, 1104–1112.
Kuniyoshi, Y., Inaba, M., & Inoue, H. (1994). Learning by watching: Extracting reusable task knowledge from visual observation of human performance. IEEE Transaction on Robotics and Automation, 10, 799–822.
Matarić, M. J. (1997). Behavior-based control: Examples from navigation, learning, and group behavior. Journal of Experimental and Theoretical Artificial Intelligence, 9(2), 323–336.
Mohseni-Kabir, A., Rich, C., Chernova, S., Sidner, C. L., & Miller, D. (2015). Interactive hierarchical task learning from a single demonstration. In Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction HRI ’15 (pp 205–212) New York, NY: ACM. ISBN 978-1-4503-2883-8. doi:10.1145/2696454.2696474. http://doi.acm.org/10.1145/2696454.2696474.
Myers, B. A. (1990). Creating user interfaces using programming by example, visual programming, and constraints. ACM Transactions on Programming Languages and Systems, 12(2), 143–177.
Nakanishi, J., Morimoto, J., Endo, G., Cheng, G., Schaal, S., & Kawato., M. (2004). Learning from demonstration and adaptation of biped locomotion. Robotics and autonomous systems, 1(2–3), 79–91.
Nicolescu, M., & Matarić, M. (2003). Methods for robot task learning: Demonstrations, generalization and practice. In Proceedings of the Second International Joint Conference on Autonomous Agents and Multi-Agent Systems.
Pardowitz, M., & Dillmann, R. (2007). Towards life-long learning in household robots: The piagetian approach. In IEEE 6th International Conference on Development and Learning (pp 88–93).
Pastor, P., Hoffmann, H., Asfour, T., & Schaal, S. (2009). Learning and generalization of motor skills by learning from demonstration. In international conference on robotics and automation.
Pearl, J. (1988). Probabilistic reasoning in intelligent systems: Networks of plausible inference. San Francisco, CA: Morgan Kaufmann.
Peters, J., Vijayakumar, S., & Schaal S. (2003). Reinforcement learning for humanoid robotics. In IEEE-RAS International Conference on Humanoid Robots (Humanoids2003).
Pollard, N., & Hodgins, J. K. (2002). Generalizing demonstrated manipulation tasks. In Workshop on the Algorithmic Foundations of Robotics.
Pomerleau, D. (1991). Efficient training of artificial neural networks for autonomous navigation. Neural Computation, 3(1), 88–97.
Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T., Leibs, J., Wheeler, R., & Ng, A. Y. (2009). Ros: An open-source robot operating system. In ICRA Workshop on Open Source Software.
Rosenstein, M., & Barto, A. (2004). Supervised actor-critic reinforcement learning, chapter 14. New York, NY, USA: Wiley.
Russell, S. (1998). Learning agents for uncertain environments. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory.
Schaal, S. (1999). Is imitation learning the route to humanoid robotics? Trends in Cognitive Sciences, 3(6), 233–242.
Schaal, S. Peters, J., Nakanishi, J., & Ijspeert, A. (2003). Control, planning, learning, and imitation with dynamic movement primitives. In Workshop on bilateral paradigms on humans and humanoids, IEEE International Conference on Intelligent Robots and Systems (IROS 2003).
Schaal, S., Ijspeert, A., & A. Billard. (2004). Computational approaches to motor learning by imitation (pp. 199–218). Number 1431. Oxford University Press.
Shachter, R. D. (1988). Probabilistic inference and influence diagrams. Operations Research, 36, 589–604.
Sutton, R. (1998). Reinforcement learning. Cambridge, MA: MIT Press.
Thomaz, A., & Breazeal, C. (2006a). Reinforcement learning with human teachers: Evidence of feedback and guidance with implications for learning performance. In Proceedings of the 21st National Conference on Artificial Intelligence.
Thomaz, A. L., & Breazeal, C. (2006b). Reinforcement learning with human teachers: Understanding how people want to teach robots. In Proceedings of the 15th IEEE International Symposium on Robot and Human Interactive Communication
Thrun, S., & Mitchell, T. (1995). Lifelong robot learning. Robotics and Autonomous Systems, 15, 25–46.
Vijayakumar, S., & Schaal, S. (2000). Locally weighted projection regression: An \(o(n)\) algorithm for incremental real time learning in high dimensional space. In Proceedings of the 17th International Conference on Machine Learning.
Acknowledgments
This work was supported by NSF grant for “SHB: Small: Socially Assistive Human-Machine Interaction for Improved Compliance and Health Outcomes” IIS-1117279; ONR MURI grant “ANTIDOTE: Adaptive Networks for Threat and Intrusian Detection Or TErmination” N00014-09-1-1031; and ONR DURIP Grant “Acquisition of a Personal Robotic Platform for DoD-Sponsored Research in human–robot Interaction, Motor Control, and Perception” N000141210729.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Koenig, N., Matarić, M.J. Robot life-long task learning from human demonstrations: a Bayesian approach. Auton Robot 41, 1173–1188 (2017). https://doi.org/10.1007/s10514-016-9601-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10514-016-9601-1