Abstract
We want to understand how animals can learn to solve complex tasks. To achieve this, it makes sense to first hypothesize learning models and then compare these models to real biological learning data. But how to perform such a comparison is still unclear. We propose that yoking is an important component to such an analysis. In yoking, two agents are made to experience the same inputs, rewards or perform the same actions – possibly in combination. We use yoking as an analytical tool to identify the algorithm that drives learning in a target agent. We evaluate this approach in a synthetic task, where we know the ground truth learning algorithm. Then we apply it to biological data from a physical puzzle task, to identify the learning algorithm behind physical problem solving in Goffin’s cockatoos. Our results show that yoking works, and can be used to identify the target algorithm more reliably, with less variance and assumptions, than a more unconstrained approach to identify learning algorithms.
Funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany’s Excellence Strategy – EXC 2002/1 “Science of Intelligence” – project number 390523135.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Anderson, D.I., et al.: The flip side of perception-action coupling: locomotor experience and the ontogeny of visual-postural coupling. Hum. Mov. Sci. 20(4–5), 461–487 (2001)
Argall, B.D., Chernova, S., Veloso, M., Browning, B.: A survey of robot learning from demonstration. Robot. Auton. Syst. 57(5), 469–483 (2009)
Church, R.M.: Systematic effect of random error in the yoked control design. Psychol. Bull. 62(2), 122 (1964)
Dice, L.R.: Measures of the amount of ecologic association between species. Ecology 26(3), 297–302 (1945)
Gardner, R.A., Gardner, B.T.: Feedforward versus feedbackward: an ethological alternative to the law of effect. Behav. Brain Sci. 11(3), 429–447 (1988)
Held, R., Hein, A.: Movement-produced stimulation in the development of visually guided behavior. J. Comp. Physiol. Psychol. 56(5), 872 (1963)
Lee, D., Gujarathi, P., Wood, J.N.: Controlled-rearing studies of newborn chicks and deep neural networks. Preprint arXiv:2112.06106 (2021)
Ng, A.Y., Russell, S.J.: Algorithms for inverse reinforcement learning. In: Proceedings of the Seventeenth International Conference on Machine Learning, pp. 663–670. Morgan Kaufmann Publishers Inc., San Francisco (2000)
Ostrovski, G., Castro, P.S., Dabney, W.: The difficulty of passive learning in deep reinforcement learning. In: Advances in Neural Information Processing Systems, vol. 34 (2021)
Salkind, N.J.: Encyclopedia of Research Design, vol. 1. Sage, Thousand Oaks (2010)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)
Torabi, F., Warnell, G., Stone, P.: Behavioral cloning from observation. Preprint arXiv:1805.01954 (2018)
Wood, S.M., Wood, J.N.: Using automation to combat the replication crisis: a case study from controlled-rearing studies of newborn chicks. Infant Behav. Dev. 57, 101329 (2019)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Baum, M. et al. (2022). Yoking-Based Identification of Learning Behavior in Artificial and Biological Agents. In: Cañamero, L., Gaussier, P., Wilson, M., Boucenna, S., Cuperlier, N. (eds) From Animals to Animats 16. SAB 2022. Lecture Notes in Computer Science(), vol 13499. Springer, Cham. https://doi.org/10.1007/978-3-031-16770-6_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-16770-6_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-16769-0
Online ISBN: 978-3-031-16770-6
eBook Packages: Computer ScienceComputer Science (R0)