Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

A Deep Reinforcement Learning-Enabled Dynamic Redeployment System for Mobile Ambulances

Published: 29 March 2019 Publication History

Abstract

Protecting citizens' lives from emergent accidents (e.g. traffic accidents) and diseases (e.g. heart attack) is of vital importance in urban computing. Every day many people are caught in emergent accidents or diseases and thus need ambulances to transport them to hospitals. In this paper, we propose a dynamic ambulance redeployment system to reduce the time needed for ambulances to pick up patients and to increase the probability of patients being saved in time. For patients in danger, every second counts. Specifically, whenever there is an ambulance becoming available (e.g. finishing transporting a patient to a hospital), our dynamic ambulance redeployment system will redeploy it to a proper ambulance station such that it can better pick up future patients. However, the dynamic ambulance redeployment is challenging, as when we redeploy an available ambulance we need to simultaneously consider each station's multiple dynamic factors. To trade off these multiple factors using handcrafted rules are almost impossible. To deal with this issue, we propose using a deep neural network, called deep score network, to balance each station's dynamic factors into one score, leveraging the excellent representation ability of deep neural networks. And then we propose a deep reinforcement learning framework to learn the deep score network. Finally, based on the learned deep score network, we provide an effective dynamic ambulance redeployment algorithm. Experiment results using data collected in real world show clear advantages of our method over baselines, e.g. comparing with baselines, our method can save ~100 seconds (~20%) of average pickup time of patients and improve the ratio of patients being picked up within 10 minutes from 0.786 to 0.838. With our method, people in danger can be better saved.

References

[1]
Kai Arulkumaran, Marc Peter Deisenroth, Miles Brundage, and Anil Anthony Bharath. 2017. Deep Reinforcement Learning: A Brief Survey. IEEE Signal Processing Magazine 34(2017), 26--38.
[2]
ASIRT. 2018. Road Crash Statistics. http://asirt.org/Initiatives/Informing-Road-Users/Road-Safety-Facts/Road-Crash-Statistics.
[3]
Christopher M. Bishop. 2007. Pattern recognition and machine learning, 5th Edition. Springer. http://www.worldcat.org/oclc/71008143.
[4]
Tianjin EMS center. 2015. Emergency medical services center of Tianjin. http://tianjin.emss.cn/hj.htm.
[5]
Albert Y. Chen, Tsung-Yu Lu, Matthew Huei-Ming Ma, and Wei-Zen Sun. 2016. Demand Forecast Using Data Analytics for the Preallocation of Ambulances. IEEE Journal of Biomedical and Health Informatics 20, 4 (July 2016), 1178--1187.
[6]
Richard Church and Charles R. Velle. 1983. The Maximal Covering Location Problem. Regional Science Association 32 (1983), 101--118.
[7]
Cleveland Clinic. 2018. Sudden Cardiac Statistics. https://my.clevelandclinic.org/health/diseases/17522-sudden-cardiac-death-sudden-cardiac-arrest.
[8]
Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein. 2009. Introduction to Algorithms (3. ed.). MIT Press.
[9]
Mark S. Daskin. 1983. A Maximum Expected Covering Location Model: Formulation, Properties and Heuristic Solution. Transportation Science 17, 1 (1983), 48--70.
[10]
Dirk Degel, Lara Wiesche, Sebastian Rachuba, and Brigitte Werners. 2015. Time-dependent ambulance allocation considering data-driven empirically required coverage. Health Care Management Science 18 (2015), 444--458.
[11]
Heart Foundation. 2018. Sudden Cardiac Statistics 2. https://www.heartfoundation.org.au/your-heart/sudden-cardiac-death.
[12]
Michel Gendreau, Gilbert Laporte, and Frédéric Semet. 2001. A dynamic model and parallel tabu search heuristic for real-time ambulance relocation. Parallel Comput. 27, 12 (September 2001), 1641--1653.
[13]
Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep Learning. MIT Press.
[14]
Peter E. Hart, Nils J. Nilsson, and Bertram Raphael. 1968. A Formal Basis for the Heuristic Determination of Minimum Cost Paths. IEEE Transactions on Systems Science and Cybernetics 4 (July 1968), 100--107.
[15]
Matteo Hessel, Joseph Modayil, Hado van Hasselt, Tom Schaul, Georg Ostrovski, Will Dabney, Dan Horgan, Bilal Piot, Mohammad Gheshlaghi Azar, and David Silver. 2018. Rainbow: Combining Improvements in Deep Reinforcement Learning. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, Louisiana, USA, February 2-7, 2018.
[16]
C.J. Jagtenberg, S. Bhulai, and R.D. van der Mei. 2015. An efficient heuristic for real-time ambulance redeployment. Operations Research for Health Care 4 (2015), 27--35.
[17]
C.J. Jagtenberg, S. Bhulai, and R. D. van der Mei. 2017. Optimal Ambulance Dispatching. Operations Research for Health Care 248 (March 2017), 269--291.
[18]
Leslie Pack Kaelbling, Michael L. Littman, and Andrew W. Moore. 1996. Reinforcement Learning: A Survey. Journal of Artificial Intelligence Research 4 (1996), 237--285.
[19]
Yann LeCun, Yoshua Bengio, and Geoffrey E. Hinton. 2015. Deep learning. Nature 521, 7553 (2015), 436--444.
[20]
Sergey Levine, Chelsea Finn, Trevor Darrell, and Pieter Abbeel. 2016. End-to-End Training of Deep Visuomotor Policies. Journal of Machine Learning Research 17 (2016), 39:1--39:40.
[21]
Shuo Ma, Yu Zheng, and Ouri Wolfson. 2015. Real-Time City-Scale Taxi Ridesharing. IEEE Transactions on Knowledge and Data Engineering 27 (2015), 1782--1795.
[22]
Matthew S. Maxwell, Mateo Restrepo, Shane G. Henderson, and Huseyin Topaloglu. 2010. Approximate Dynamic Programming for Ambulance Redeployment. INFORMS Journal on Computing 22, 2 (May 2010), 266--281.
[23]
Richard McCormack and Graham Coates. 2015. A simulation model to enable the optimization of ambulance fleet allocation and base station location for increased patient survival. European Journal of Operational Research 247 (May 2015), 294--309.
[24]
Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. 2015. Human-level control through deep reinforcement learning. Nature 518, 7540 (February 2015), 529--533.
[25]
Meng Qu, Hengshu Zhu, Junming Liu, Guannan Liu, and Hui Xiong. 2014. A cost-effective recommender system for taxi drivers. In The 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 45--54.
[26]
Sandhya Saisubramanian, Pradeep Varakantham, and Hoong Chuin Lau. 2015. Risk Based Optimization for Improving Emergency Medical Systems. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence. 702--708.
[27]
Verena Schmid. 2012. Solving the dynamic ambulance relocation and dispatching problem using approximate dynamic programming. European Journal of Operational Research 219, 3 (April 2012), 611--621.
[28]
Verena Schmid and Karl F. Doerner. 2010. Ambulance location and relocation problems with time-dependent travel times. European Journal of Operational Research 207, 3 (October 2010), 1293--1303.
[29]
David Silver, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre, George van den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Vedavyas Panneershelvam, Marc Lanctot, Sander Dieleman, Dominik Grewe, John Nham, Nal Kalchbrenner, Ilya Sutskever, Timothy P. Lillicrap, Madeleine Leach, Koray Kavukcuoglu, Thore Graepel, and Demis Hassabi. 2016. Mastering the game of Go with deep neural networks and tree search. Nature 529, 7587 (2016), 484--489.
[30]
David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin A. Riedmiller. 2014. Deterministic Policy Gradient Algorithms. In Proceedings of the 31th International Conference on Machine Learning. 387--395.
[31]
David Silver, Julian Schrittwieser, Karen Simonyan, Ioannis Antonoglou, Aja Huang, Arthur Guez, Thomas Hubert, Lucas Baker, Matthew Lai, Adrian Bolton, Yutian Chen, Timothy Lillicrap, Fan Hui, Laurent Sifre, George van den Driessche, Thore Graepel, and Demis Hassabis. 2017. Mastering the game of Go without human knowledge. Nature 550 (2017), 354--359.
[32]
Lawrence V. Snyder and Mark S. Daskin. 2004. Reliability Models for Facility Location: The Expected Failure Cost Case. Transportation Science 39, 3 (August 2004), 400--416.
[33]
Richard S. Sutton and Andrew G. Barto. 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning 8 (May 1992), 229--256.
[34]
Richard S. Sutton and Andrew G. Barto. 2017. Reinforcement Learning: An Introduction (Second Edition, in progress). MIT Press, Cambridge, MA.
[35]
Pieter L. van den Berg, J. Theresia van Essen, and Eline J. Harderwijk. 2016. Comparison of static ambulance location models. In Proceedings of the 3rd International Conference on Logistics Operations Management. 1040--1051.
[36]
Dong Wang, Junbo Zhang, Wei Cao, Jian Li, and Yu Zheng. 2018. When Will You Arrive? Estimating Travel Time Based on Deep Neural Networks. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence.
[37]
Yilun Wang, Yu Zheng, and Yexiang Xue. 2014. Travel time estimation of a path using sparse trajectories. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 25--34.
[38]
Bradford S. Westgate, Dawn B. Woodard, David S. Matteson, and Shane G. Henderson. 2016. Large-network travel time distribution estimation for ambulances. European Journal of Operational Research 252 (2016), 322--333.
[39]
Wikipedia. 2018. Half-normal distribution. https://en.wikipedia.org/wiki/Half-normal_distribution.
[40]
Jing Yuan, Yu Zheng, Chengyang Zhang, Wenlei Xie, Xing Xie, Guangzhong Sun, and Yan Huang. 2010. T-drive: driving directions based on taxi trajectories. In Proceedings of the 18th ACM SIGSPATIAL International Symposium on Advances in Geographic Information Systems. 99--108.
[41]
Yisong Yue, Lavanya Marla, and Ramayya Krishnan. 2012. An Efficient Simulation-Based Approach to Ambulance Fleet Allocation and Dynamic Redeployment. In Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence.
[42]
Siyuan Zhang, Lu Qin, Yu Zheng, and Hong Cheng. 2016. Effective and Efficient: Large-Scale Dynamic City Express. IEEE Transactions on Knowledge and Data Engineering 28 (2016), 3203--3217.
[43]
Lu Zhen, Kai Wang, Hongtao Hu, and Daofang Chang. 2014. A simulation optimization framework for ambulance deployment and relocation problems. Computers & Industrial Engineering 72 (March 2014), 12--23.
[44]
Zhengyi Zhou and David S. Matteson. 2015. Predicting Ambulance Demand: a Spatio-Temporal Kernel Approach. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2297--2303.

Cited By

View all
  • (2024)Time-Constrained Actor-Critic Reinforcement Learning for Concurrent Order Dispatch in On-Demand DeliveryIEEE Transactions on Mobile Computing10.1109/TMC.2023.334281523:8(8175-8192)Online publication date: Aug-2024
  • (2024)SpeedAdv: Enabling Green Light Optimized Speed Advisory for Diverse Traffic LightsIEEE Transactions on Mobile Computing10.1109/TMC.2023.3319697(1-14)Online publication date: 2024
  • (2024)Scalable policies for the dynamic traveling multi-maintainer problem with alertsEuropean Journal of Operational Research10.1016/j.ejor.2024.05.049319:1(121-134)Online publication date: Nov-2024
  • Show More Cited By

Index Terms

  1. A Deep Reinforcement Learning-Enabled Dynamic Redeployment System for Mobile Ambulances

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
    Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies  Volume 3, Issue 1
    March 2019
    786 pages
    EISSN:2474-9567
    DOI:10.1145/3323054
    Issue’s Table of Contents
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 29 March 2019
    Accepted: 01 January 2019
    Revised: 01 November 2018
    Received: 01 August 2018
    Published in IMWUT Volume 3, Issue 1

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Mobile computing
    2. deep reinforcement learning
    3. deep score network
    4. dynamic redeployment system for mobile ambulances
    5. urban computing

    Qualifiers

    • Research-article
    • Research
    • Refereed

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)84
    • Downloads (Last 6 weeks)18
    Reflects downloads up to 03 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Time-Constrained Actor-Critic Reinforcement Learning for Concurrent Order Dispatch in On-Demand DeliveryIEEE Transactions on Mobile Computing10.1109/TMC.2023.334281523:8(8175-8192)Online publication date: Aug-2024
    • (2024)SpeedAdv: Enabling Green Light Optimized Speed Advisory for Diverse Traffic LightsIEEE Transactions on Mobile Computing10.1109/TMC.2023.3319697(1-14)Online publication date: 2024
    • (2024)Scalable policies for the dynamic traveling multi-maintainer problem with alertsEuropean Journal of Operational Research10.1016/j.ejor.2024.05.049319:1(121-134)Online publication date: Nov-2024
    • (2024)RELight: a random ensemble reinforcement learning based method for traffic light controlApplied Intelligence10.1007/s10489-023-05197-w54:1(95-112)Online publication date: 1-Jan-2024
    • (2023)DEAR: Dynamic Electric Ambulance RedeploymentProceedings of the 18th International Symposium on Spatial and Temporal Data10.1145/3609956.3609959(11-20)Online publication date: 23-Aug-2023
    • (2023)CausalSE: Understanding Varied Spatial Effects with Missing Data Toward Adding New Bike-sharing StationsACM Transactions on Knowledge Discovery from Data10.1145/353642717:2(1-24)Online publication date: 20-Mar-2023
    • (2023)FedDSR: Daily Schedule Recommendation in a Federated Deep Reinforcement Learning FrameworkIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2021.313026535:4(3912-3924)Online publication date: 1-Apr-2023
    • (2023)REDE: Exploring Relay Transportation for Efficient Last-mile Delivery2023 IEEE 39th International Conference on Data Engineering (ICDE)10.1109/ICDE55515.2023.00230(3003-3016)Online publication date: Apr-2023
    • (2023)Artificial intelligence and machine learning in prehospital emergency care: A scoping reviewiScience10.1016/j.isci.2023.10740726:8(107407)Online publication date: Aug-2023
    • (2023)Urban Mobility-Driven Crowdsensing: Recent Advances in Machine Learning Designs and Ubiquitous ApplicationsMobile Crowdsourcing10.1007/978-3-031-32397-3_2(33-58)Online publication date: 21-Apr-2023
    • Show More Cited By

    View Options

    Get Access

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media