Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Versatile Control of Fluid-directed Solid Objects Using Multi-task Reinforcement Learning

Published: 18 October 2022 Publication History

Abstract

We propose a learning-based controller for high-dimensional dynamic systems with coupled fluid and solid objects. The dynamic behaviors of such systems can vary across different simulators and the control tasks subject to changing requirements from users. Our controller features high versatility and can adapt to changing dynamic behaviors and multiple tasks without re-training, which is achieved by combining two training strategies. We use meta-reinforcement learning to inform the controller of changing simulation parameters. We further design a novel task representation, which allows the controller to adapt to continually changing tasks via hindsight experience replay. We highlight the robustness and generality of our controller on a row of dynamic-rich tasks, including scooping up solid balls from a water pool, in-air ball acrobatics using fluid spouts, and zero-shot transferring to unseen simulators and constitutive models. In all the scenarios, our controller consistently outperforms the plain multi-task reinforcement-learning baseline.

References

[1]
Marcin Andrychowicz, Filip Wolski, Alex Ray, Jonas Schneider, Rachel Fong, Peter Welinder, Bob McGrew, Josh Tobin, OpenAI Pieter Abbeel, and Wojciech Zaremba. 2017. Hindsight experience replay. In Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Long Beach, CA. Retrieved from https://proceedings.neurips.cc/paper/2017/file/453fadbd8a1a3af50a9df4df899537b5-Paper.pdf.
[2]
Markus Becker and Matthias Teschner. 2007. Weakly compressible SPH for free surface flows. In Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation (SCA’07). Eurographics Association, Goslar, DEU, 209–217.
[3]
Kevin Bergamin, Simon Clavet, Daniel Holden, and James Richard Forbes. 2019. DReCon: Data-driven responsive control of physics-based characters. ACM Trans. Graph. 38, 6, Article 206 (Nov. 2019), 11 pages.
[4]
Stelian Coros, Sebastian Martin, Bernhard Thomaszewski, Christian Schumacher, Robert Sumner, and Markus Gross. 2012. Deformable objects alive! ACM Trans. Graph. 31, 4, Article 69 (July 2012), 9 pages.
[5]
Yan Duan, John Schulman, Xi Chen, Peter L. Bartlett, Ilya Sutskever, and Pieter Abbeel. 2016. RL\(^2\): Fast Reinforcement Learning via Slow Reinforcement Learning. Retrieved from https://arXiv:1611.02779.
[6]
Raanan Fattal and Dani Lischinski. 2004. Target-driven smoke animation. ACM Trans. Graph. 23, 3 (Aug. 2004), 441–448.
[7]
Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, Sydney, Australia, 1126–1135. Retrieved from https://proceedings.mlr.press/v70/finn17a.html.
[8]
Abhishek Gupta, Russell Mendonca, YuXuan Liu, Pieter Abbeel, and Sergey Levine. 2018. Meta-reinforcement learning of structured exploration strategies. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS’18). Curran Associates, Red Hook, NY, 5307–5316.
[9]
Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. 2018. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the 35th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 80), Jennifer Dy and Andreas Krause (Eds.). PMLR, Stockholm, Sweden, 1861–1870. Retrieved from https://proceedings.mlr.press/v80/haarnoja18b.html.
[10]
Hado van Hasselt, Arthur Guez, and David Silver. 2016. Deep reinforcement learning with double q-learning. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI’16). AAAI Press, 2094–2100.
[11]
Yuanming Hu, Yu Fang, Ziheng Ge, Ziyin Qu, Yixin Zhu, Andre Pradhana, and Chenfanfu Jiang. 2018. A moving least squares material point method with displacement discontinuity and two-way rigid body coupling. ACM Trans. Graph. 37, 4 (2018), 1–14.
[12]
Yuanming Hu, Tzu-Mao Li, Luke Anderson, Jonathan Ragan-Kelley, and Frédo Durand. 2019a. Taichi: A language for high-performance computation on spatially sparse data structures. ACM Trans. Graph. 38, 6 (2019), 201.
[13]
Yuanming Hu, Tzu-Mao Li, Luke Anderson, Jonathan Ragan-Kelley, and Frédo Durand. 2019b. Taichi: A language for high-performance computation on spatially sparse data structures. ACM Trans. Graph. 38, 6 (2019), 1–16.
[14]
Yuanming Hu, Jiancheng Liu, Andrew Spielberg, Joshua B. Tenenbaum, William T. Freeman, Jiajun Wu, Daniela Rus, and Wojciech Matusik. 2019c. Chainqueen: A real-time differentiable physical simulator for soft robotics. In Proceedings of the International Conference on Robotics and Automation (ICRA’19). IEEE, 6265–6271.
[15]
Byungsoo Kim, Vinicius C. Azevedo, Markus Gross, and Barbara Solenthaler. 2019. Transport-based neural style transfer for smoke simulations. ACM Trans. Graph. 38, 6, Article 188 (Nov. 2019), 11 pages.
[16]
Yunzhu Li, Toru Lin, Kexin Yi, Daniel Bear, Daniel Yamins, Jiajun Wu, Joshua Tenenbaum, and Antonio Torralba. 2020. Visual grounding of learned physical models. In Proceedings of the 37th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 119), Hal Daum III and Aarti Singh (Eds.). PMLR, Vienna, Austria, 5927–5936. Retrieved from https://proceedings.mlr.press/v119/li20j.html.
[17]
Yunzhu Li, Jiajun Wu, Russ Tedrake, Joshua B. Tenenbaum, and Antonio Torralba. 2019. Learning particle dynamics for manipulating rigid bodies, deformable objects, and fluids. In Proceedings of the International Conference on Learning Representations (ICLR’19). Open Publishing, New Orleans, LA.
[18]
Pingchuan Ma, Yunsheng Tian, Zherong Pan, Bo Ren, and Dinesh Manocha. 2018. Fluid-directed rigid body control using deep reinforcement learning. ACM Trans. Graph. 37, 4 (2018), 1–11.
[19]
Antoine McNamara, Adrien Treuille, Zoran Popović, and Jos Stam. 2004. Fluid control using the adjoint method. ACM Trans. Graph. 23, 3 (2004), 449–456.
[20]
Russell Mendonca, Abhishek Gupta, Rosen Kralev, Pieter Abbeel, Sergey Levine, and Chelsea Finn. 2019. Guided meta-policy search. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc., Vancouver, Canada. Retrieved from https://proceedings.neurips.cc/paper/2019/file/d324a0cc02881779dcda44a675fdcaaa-Paper.pdf.
[21]
Sehee Min, Jungdam Won, Seunghwan Lee, Jungnam Park, and Jehee Lee. 2019. SoftCon: Simulation and control of soft-bodied animals with biomimetic actuators. ACM Trans. Graph. 38, 6, Article 208 (Nov. 2019), 12 pages.
[22]
Michael B. Nielsen and Robert Bridson. 2011. Guide shapes for high resolution naturalistic liquid simulation. ACM Trans. Graph. 30, 4, Article 83 (July 2011), 8 pages.
[23]
Zherong Pan and Dinesh Manocha. 2017a. Efficient solver for spacetime control of smoke. ACM Trans. Graph. 36, 4 (2017), 1.
[24]
Zherong Pan and Dinesh Manocha. 2017b. Feedback motion planning for liquid pouring using supervised learning. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS’17). IEEE, Vancouver, BC, Canada, 1252–1259.
[25]
Zherong Pan and Dinesh Manocha. 2018a. Active animations of reduced deformable models with environment interactions. ACM Trans. Graph. 37, 3, Article 36 (Aug. 2018), 17 pages.
[26]
Zherong Pan and Dinesh Manocha. 2018b. Active animations of reduced deformable models with environment interactions. ACM Trans. Graph. 37, 3 (2018), 1–17.
[27]
Zherong Pan, Bo Ren, and Dinesh Manocha. 2019. GPU-based contact-aware trajectory optimization using a smooth force model. In Proceedings of the 18th Annual ACM SIGGRAPH/Eurographics Symposium on Computer Animation (SCA’19). Association for Computing Machinery, New York, NY, Article 4, 12 pages.
[28]
Soohwan Park, Hoseok Ryu, Seyoung Lee, Sunmin Lee, and Jehee Lee. 2019. Learning predict-and-simulate policies from unorganized human motion data. ACM Trans. Graph. 38, 6, Article 205 (Nov. 2019), 11 pages.
[29]
Xue Bin Peng, Pieter Abbeel, Sergey Levine, and Michiel van de Panne. 2018. Deepmimic: Example-guided deep reinforcement learning of physics-based character skills. ACM Trans. Graph. 37, 4 (2018), 1–14.
[30]
Xue Bin Peng, Glen Berseth, KangKang Yin, and Michiel Van De Panne. 2017. Deeploco: Dynamic locomotion skills using hierarchical deep reinforcement learning. ACM Trans. Graph. 36, 4 (2017), 1–13.
[31]
Xue Bin Peng, Ze Ma, Pieter Abbeel, Sergey Levine, and Angjoo Kanazawa. 2021. AMP: Adversarial motion priors for stylized physics-based character control. ACM Trans. Graph. 40, 4, Article 144 (July 2021), 20 pages.
[32]
Jovan Popović, Steven M. Seitz, and Michael Erdmann. 2003. Motion sketching for control of rigid-body simulations. ACM Trans. Graph. 22, 4 (Oct. 2003), 1034–1054.
[33]
Michael Posa, Cecilia Cantu, and Russ Tedrake. 2014. A direct method for trajectory optimization of rigid bodies through contact. Int. J. Robot. Res. 33, 1 (2014), 69–81. arXiv:https://doi.org/10.1177/0278364913506757
[34]
Lukas Prantl, Boris Bonev, and Nils Thuerey. 2019. Generating Liquid Simulations with Deformation-aware Neural Networks. Retrieved from https://arXiv:1704.07854.
[35]
Kate Rakelly, Aurick Zhou, Chelsea Finn, Sergey Levine, and Deirdre Quillen. 2019. Efficient off-policy meta-reinforcement learning via probabilistic context variables. In Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, Long Beach, CA, 5331–5340. Retrieved from https://proceedings.mlr.press/v97/rakelly19a.html.
[36]
Karthik Raveendran, Nils Thuerey, Chris Wojtan, and Greg Turk. 2012. Controlling liquids using meshes. In Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation (SCA’12). Eurographics Association, Goslar, DEU, 255–264.
[37]
Karthik Raveendran, Chris Wojtan, Nils Thuerey, and Greg Turk. 2014. Blending liquids. ACM Trans. Graph. 33, 4, Article 137 (July 2014), 10 pages.
[38]
Syuhei Sato, Yoshinori Dobashi, Theodore Kim, and Tomoyuki Nishita. 2018. Example-based turbulence style transfer. ACM Trans. Graph. 37, 4 (2018), 1–9.
[39]
Tom Schaul, Daniel Horgan, Karol Gregor, and David Silver. 2015. Universal value function approximators. In Proceedings of the 32nd International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 37), Francis Bach and David Blei (Eds.). PMLR, Lille, France, 1312–1320. Retrieved from https://proceedings.mlr.press/v37/schaul15.html.
[40]
Connor Schenck and Dieter Fox. 2017. Visual closed-loop control for pouring liquids. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA’17). IEEE, Singapore, 2629–2636.
[41]
Connor Schenck and Dieter Fox. 2018. SPNets: Differentiable fluid dynamics for deep neural networks. In Proceedings of the 2nd Conference on Robot Learning(Proceedings of Machine Learning Research, Vol. 87), Aude Billard, Anca Dragan, Jan Peters, and Jun Morimoto (Eds.). PMLR, Stockholm, Sweden, 317–335. Retrieved from https://proceedings.mlr.press/v87/schenck18a.html.
[42]
John Schulman, Sergey Levine, Pieter Abbeel, Michael Jordan, and Philipp Moritz. 2015. Trust region policy optimization. In Proceedings of the 32nd International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 37), Francis Bach and David Blei (Eds.). PMLR, Lille, France, 1889–1897. Retrieved from https://proceedings.mlr.press/v37/schulman15.html.
[43]
Weiguang Si, Sung-Hee Lee, Eftychios Sifakis, and Demetri Terzopoulos. 2015. Realistic biomechanical simulation and control of human swimming. ACM Trans. Graph. 34, 1, Article 10 (Dec. 2015), 15 pages.
[44]
David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin Riedmiller. 2014. Deterministic policy gradient algorithms. In Proceedings of the 31st International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 32), Eric P. Xing and Tony Jebara (Eds.). PMLR, Bejing, China, 387–395. Retrieved from https://proceedings.mlr.press/v32/silver14.html.
[45]
Andrew Spielberg, Allan Zhao, Yuanming Hu, Tao Du, Wojciech Matusik, and Daniela Rus. 2019. Learning-in-the-loop optimization: End-to-end control and co-design of soft robots through learned deep latent representations. Adv. Neural Info. Process. Syst. 32 (2019), 8284–8294.
[46]
Sebastian Starke, Yiwei Zhao, Fabio Zinno, and Taku Komura. 2021. Neural animation layering for synthesizing martial arts movements. ACM Trans. Graph. 40, 4, Article 92 (July 2021), 16 pages.
[47]
Richard S. Sutton and Andrew G. Barto. 2018. Reinforcement Learning: An Introduction. A Bradford Book, Cambridge, MA.
[48]
Jie Tan, Yuting Gu, Greg Turk, and C. Karen Liu. 2011. Articulated swimming creatures. In Proceedings of the ACM Special Interest Group on Computer Graphics and Interactive Techniques (SIGGRAPH’11). Association for Computing Machinery, New York, NY, USA, Article 58, 12 pages.
[49]
Jie Tan, Greg Turk, and C. Karen Liu. 2012. Soft body locomotion. ACM Trans. Graph. 31, 4, Article 26 (jul 2012), 11 pages.
[50]
Jingwei Tang, Vinicius C. Azevedo, Guillaume Cordonnier, and Barbara Solenthaler. 2021. Honey, I shrunk the domain: Frequency-aware force field reduction for efficient fluids optimization. Comput. Graph. Forum 40, 2 (2021), 339–353. Retrieved from.
[51]
Yuval Tassa, Tom Erez, and Emanuel Todorov. 2012. Synthesis and stabilization of complex behaviors through online trajectory optimization. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, Vilamoura, Algarve, Portugal, 4906–4913.
[52]
Nils Thuerey. 2016. Interpolations of smoke and liquid simulations. ACM Trans. Graph. 36, 1 (2016), 1–16.
[53]
Jonathan Tompson, Kristofer Schlachter, Pablo Sprechmann, and Ken Perlin. 2017. Accelerating Eulerian fluid simulation with convolutional networks. In Proceedings of the 34th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, Sydney, Australia, 3424–3433. Retrieved from https://proceedings.mlr.press/v70/tompson17a.html.
[54]
Marc Toussaint, Jung-Su Ha, and Danny Driess. 2020. Describing physics for physical reasoning: Force-based sequential manipulation planning. IEEE Robot. Autom. Lett. 5, 4 (2020), 6209–6216.
[55]
Adrien Treuille, Antoine McNamara, Zoran Popović, and Jos Stam. 2003. Keyframe control of smoke simulations. ACM Trans. Graph. 22, 3 (July 2003), 716–723.
[56]
Jack M. Wang, Samuel R. Hamner, Scott L. Delp, and Vladlen Koltun. 2012. Optimizing locomotion controllers using biologically-based actuators and objectives. ACM Trans. Graph. 31, 4 (2012), 1–11.
[57]
Ziyu Wang, Tom Schaul, Matteo Hessel, Hado Hasselt, Marc Lanctot, and Nando Freitas. 2016. Dueling network architectures for deep reinforcement learning. In Proceedings of the 33rd International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 48), Maria Florina Balcan and Kilian Q. Weinberger (Eds.). PMLR, New York, New York, 1995–2003. https://proceedings.mlr.press/v48/wangf16.html.
[58]
Jungdam Won, Deepak Gopinath, and Jessica Hodgins. 2020. A scalable approach to control diverse behaviors for physically simulated characters. ACM Trans. Graph. 39, 4, Article 33 (July 2020), 12 pages.
[59]
You Xie, Erik Franz, Mengyu Chu, and Nils Thuerey. 2018. tempoGAN: A temporally coherent, volumetric GAN for super-resolution fluid flow. ACM Trans. Graph. 37, 4 (2018), 1–15.
[60]
KangKang Yin, Kevin Loken, and Michiel Van de Panne. 2007. Simbicon: Simple biped locomotion control. ACM Trans. Graph. 26, 3 (2007), 105.
[61]
Wenhao Yu, Greg Turk, and C. Karen Liu. 2018. Learning symmetric and low-energy locomotion. ACM Trans. Graph. 37, 4, Article 144 (July 2018), 12 pages.
[62]
Yunbo Zhang, Wenhao Yu, C. Karen Liu, Charlie Kemp, and Greg Turk. 2020. Learning to manipulate amorphous materials. ACM Trans. Graph. 39, 6 (2020), 1–11.
[63]
Rui Zhao, Xudong Sun, and Volker Tresp. 2019. Maximum entropy-regularized multi-goal reinforcement learning. In Proceedings of the 36th International Conference on Machine Learning(Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, Long Beach, CA, 7553–7562. Retrieved from https://proceedings.mlr.press/v97/zhao19d.html.
[64]
Yongning Zhu and Robert Bridson. 2005. Animating sand as a fluid. ACM Trans. Graph. 24, 3 (2005), 965–972.

Cited By

View all
  • (2024)Multi-Level Progressive Reinforcement Learning for Control Policy in Physical Simulations2024 IEEE International Conference on Robotics and Automation (ICRA)10.1109/ICRA57147.2024.10610992(9502-9508)Online publication date: 13-May-2024
  • (2023)DiffFR: Differentiable SPH-Based Fluid-Rigid Coupling for Rigid Body ControlACM Transactions on Graphics10.1145/361831842:6(1-17)Online publication date: 5-Dec-2023

Index Terms

  1. Versatile Control of Fluid-directed Solid Objects Using Multi-task Reinforcement Learning

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Graphics
    ACM Transactions on Graphics  Volume 42, Issue 2
    April 2023
    210 pages
    ISSN:0730-0301
    EISSN:1557-7368
    DOI:10.1145/3563904
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 18 October 2022
    Online AM: 12 August 2022
    Accepted: 12 July 2022
    Revised: 18 May 2022
    Received: 01 August 2021
    Published in TOG Volume 42, Issue 2

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Fluid/solid coupling
    2. optimal control
    3. reinforcement learning

    Qualifiers

    • Research-article
    • Refereed

    Funding Sources

    • Open Project Program of State Key Laboratory of Virtual Reality Technology and Systems, Beihang University

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)226
    • Downloads (Last 6 weeks)24
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Multi-Level Progressive Reinforcement Learning for Control Policy in Physical Simulations2024 IEEE International Conference on Robotics and Automation (ICRA)10.1109/ICRA57147.2024.10610992(9502-9508)Online publication date: 13-May-2024
    • (2023)DiffFR: Differentiable SPH-Based Fluid-Rigid Coupling for Rigid Body ControlACM Transactions on Graphics10.1145/361831842:6(1-17)Online publication date: 5-Dec-2023

    View Options

    Get Access

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media