Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

A Driver-Centric Vehicle Reposition Framework via Multi-agent Reinforcement Learning

  • Conference paper
  • First Online:
Green, Pervasive, and Cloud Computing (GPC 2020)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 12398))

Included in the following conference series:

Abstract

The e-hailing platforms have transformed the way people travel, live, and socialize. The efficiency of the platform is substantially influenced by the distribution differences between demands and supplies in the city. Therefore, an appropriate reposition vehicle strategy can significantly balance this distribution difference, which will promote platform benefits, customer goodwill and greatly alleviate traffic congestions. Due to the complicated relationship between vehicles and the temporal correlation of reposition actions, it is a challenging task to reposition vehicles in the city. Existing studies mostly focus on individual drivers that can hardly capture the relationship between drivers and long-term variations of demands and supplies in the city. In this paper, we introduce the reinforcement learning with geographic information and propose a geographic-based multi-agent deep deterministic policy gradient algorithm (gbMADDPG). The algorithm is driver-centric which takes the passenger searching time as an optimization goal to reduce the idle time of vehicles. We will demonstrate the effectiveness of our proposed algorithm framework through simulation experiments based on real data.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    Didi Chuxing. [n. d.]. ([n. d.]). https://www.didiglobal.com/.

  2. 2.

    Uber. [n. d.]. ([n. d.]). https://www.uber.com/.

  3. 3.

    Comset. https://github.com/Chessnl/COMSET-GISCUP.

  4. 4.

    Nyc taxi and limousine commission. https://www1.nyc.gov/site/tlc/about/data-and-research.page.

References

  1. Arulkumaran, K., Deisenroth, M.P., Brundage, M., Bharath, A.A.: A brief survey of deep reinforcement learning. arXiv preprint arXiv:1708.05866 (2017)

  2. Chen, C., Ding, Y., Wang, Z., Zhao, J., Guo, B., Zhang, D.: Vtracer: when online vehicle trajectory compression meets mobile edge computing. IEEE Syst. J. 14(2), 1635–1646 (2019)

    Article  Google Scholar 

  3. Chen, C., Ding, Y., Xie, X., Zhang, S., Wang, Z., Feng, L.: Trajcompressor: an online map-matching-based trajectory compression framework leveraging vehicle heading direction and change. IEEE Trans. Intell. Transp. Syst. 21(5), 2012–2028 (2019)

    Article  Google Scholar 

  4. Chen, C., et al.: Crowddeliver: planning city-wide package delivery paths leveraging the crowd of taxis. IEEE Trans. Intell. Transp. Syst. 18(6), 1478–1496 (2016)

    Google Scholar 

  5. Foerster, J., Assael, I.A., De Freitas, N., Whiteson, S.: Learning to communicate with deep multi-agent reinforcement learning. In: Advances in Neural Information Processing Systems, pp. 2137–2145 (2016)

    Google Scholar 

  6. Guo, S., et al.: ROD-revenue: seeking strategies analysis and revenue prediction in ride-on-demand service using multi-source urban data. IEEE Trans. Mob. Comput. 19, 2202–2220 (2019)

    Google Scholar 

  7. Jin, J., et al.: Coride: joint order dispatching and fleet management for multi-scale ride-hailing platforms. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 1983–1992 (2019)

    Google Scholar 

  8. Li, B., et al.: Hunting or waiting? Discovering passenger-finding strategies from a large-scale real-world taxi dataset. In: 2011 IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops), pp. 63–68. IEEE (2011)

    Google Scholar 

  9. Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015)

  10. Lin, K., Zhao, R., Xu, Z., Zhou, J.: Efficient large-scale fleet management via multi-agent deep reinforcement learning. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1774–1783 (2018)

    Google Scholar 

  11. Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, O.P., Mordatch, I.: Multi-agent actor-critic for mixed cooperative-competitive environments. In: Advances in Neural Information Processing Systems, pp. 6379–6390 (2017)

    Google Scholar 

  12. Mao, H., Zhang, Z., Xiao, Z., Gong, Z.: Modelling the dynamic joint policy of teammates with attention multi-agent DDPG. In: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, pp. 1108–1116. International Foundation for Autonomous Agents and Multiagent Systems (2019)

    Google Scholar 

  13. Mnih, V., et al.: Asynchronous methods for deep reinforcement learning. In: International Conference on Machine Learning, pp. 1928–1937 (2016)

    Google Scholar 

  14. Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)

    Article  Google Scholar 

  15. Pham, T.H., De Magistris, G., Tachibana, R.: Optlayer-practical constrained optimization for deep reinforcement learning in the real world. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 66–6243. IEEE (2018)

    Google Scholar 

  16. Pinkus, A.: Approximation theory of the MLP model in neural networks. Acta Numerica 8, 143–195 (1999)

    Article  MathSciNet  Google Scholar 

  17. Silver, D., et al.: Mastering the game of go with deep neural networks and tree search. Nature 529(7587), 484 (2016)

    Article  Google Scholar 

  18. Silver, D., Lever, G., Heess, N., Degris, T., Wierstra, D., Riedmiller, M.: Deterministic policy gradient algorithms (2014)

    Google Scholar 

  19. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)

    MATH  Google Scholar 

  20. Sutton, R.S., McAllester, D.A., Singh, S.P., Mansour, Y.: Policy gradient methods for reinforcement learning with function approximation. In: Advances in Neural Information Processing Systems, pp. 1057–1063 (2000)

    Google Scholar 

  21. Tang, H., Kerber, M., Huang, Q., Guibas, L.: Locating lucrative passengers for taxicab drivers. In: Proceedings of the 21st ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, pp. 504–507 (2013)

    Google Scholar 

  22. Wang, S., Li, L., Ma, W., Chen, X.: Trajectory analysis for on-demand services: a survey focusing on spatial-temporal demand and supply patterns. Transp. Res. Part C: Emerg. Technol. 108, 74–99 (2019)

    Article  Google Scholar 

  23. Wen, J., Zhao, J., Jaillet, P.: Rebalancing shared mobility-on-demand systems: a reinforcement learning approach. In: 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), pp. 220–225. IEEE (2017)

    Google Scholar 

  24. Yang, Y., Luo, R., Li, M., Zhou, M., Zhang, W., Wang, J.: Mean field multi-agent reinforcement learning. arXiv preprint arXiv:1802.05438 (2018)

  25. Yuan, N.J., Zheng, Y., Zhang, L., Xie, X.: T-finder: a recommender system for finding passengers and vacant taxis. IEEE Trans. Knowl. Data Eng. 25(10), 90–2403 (2012)

    Google Scholar 

  26. Zhang, R., Ghanem, R.: Demand, supply, and performance of street-hail taxi. IEEE Trans. Intell. Transp. Syst. 21, 4123–4132 (2019)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chao Chen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, C., Deng, M., Chen, C., Xiang, C. (2020). A Driver-Centric Vehicle Reposition Framework via Multi-agent Reinforcement Learning. In: Yu, Z., Becker, C., Xing, G. (eds) Green, Pervasive, and Cloud Computing. GPC 2020. Lecture Notes in Computer Science(), vol 12398. Springer, Cham. https://doi.org/10.1007/978-3-030-64243-3_17

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-64243-3_17

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-64242-6

  • Online ISBN: 978-3-030-64243-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics