Abstract
The e-hailing platforms have transformed the way people travel, live, and socialize. The efficiency of the platform is substantially influenced by the distribution differences between demands and supplies in the city. Therefore, an appropriate reposition vehicle strategy can significantly balance this distribution difference, which will promote platform benefits, customer goodwill and greatly alleviate traffic congestions. Due to the complicated relationship between vehicles and the temporal correlation of reposition actions, it is a challenging task to reposition vehicles in the city. Existing studies mostly focus on individual drivers that can hardly capture the relationship between drivers and long-term variations of demands and supplies in the city. In this paper, we introduce the reinforcement learning with geographic information and propose a geographic-based multi-agent deep deterministic policy gradient algorithm (gbMADDPG). The algorithm is driver-centric which takes the passenger searching time as an optimization goal to reduce the idle time of vehicles. We will demonstrate the effectiveness of our proposed algorithm framework through simulation experiments based on real data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Didi Chuxing. [n. d.]. ([n. d.]). https://www.didiglobal.com/.
- 2.
Uber. [n. d.]. ([n. d.]). https://www.uber.com/.
- 3.
- 4.
Nyc taxi and limousine commission. https://www1.nyc.gov/site/tlc/about/data-and-research.page.
References
Arulkumaran, K., Deisenroth, M.P., Brundage, M., Bharath, A.A.: A brief survey of deep reinforcement learning. arXiv preprint arXiv:1708.05866 (2017)
Chen, C., Ding, Y., Wang, Z., Zhao, J., Guo, B., Zhang, D.: Vtracer: when online vehicle trajectory compression meets mobile edge computing. IEEE Syst. J. 14(2), 1635–1646 (2019)
Chen, C., Ding, Y., Xie, X., Zhang, S., Wang, Z., Feng, L.: Trajcompressor: an online map-matching-based trajectory compression framework leveraging vehicle heading direction and change. IEEE Trans. Intell. Transp. Syst. 21(5), 2012–2028 (2019)
Chen, C., et al.: Crowddeliver: planning city-wide package delivery paths leveraging the crowd of taxis. IEEE Trans. Intell. Transp. Syst. 18(6), 1478–1496 (2016)
Foerster, J., Assael, I.A., De Freitas, N., Whiteson, S.: Learning to communicate with deep multi-agent reinforcement learning. In: Advances in Neural Information Processing Systems, pp. 2137–2145 (2016)
Guo, S., et al.: ROD-revenue: seeking strategies analysis and revenue prediction in ride-on-demand service using multi-source urban data. IEEE Trans. Mob. Comput. 19, 2202–2220 (2019)
Jin, J., et al.: Coride: joint order dispatching and fleet management for multi-scale ride-hailing platforms. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 1983–1992 (2019)
Li, B., et al.: Hunting or waiting? Discovering passenger-finding strategies from a large-scale real-world taxi dataset. In: 2011 IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops), pp. 63–68. IEEE (2011)
Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015)
Lin, K., Zhao, R., Xu, Z., Zhou, J.: Efficient large-scale fleet management via multi-agent deep reinforcement learning. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1774–1783 (2018)
Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, O.P., Mordatch, I.: Multi-agent actor-critic for mixed cooperative-competitive environments. In: Advances in Neural Information Processing Systems, pp. 6379–6390 (2017)
Mao, H., Zhang, Z., Xiao, Z., Gong, Z.: Modelling the dynamic joint policy of teammates with attention multi-agent DDPG. In: Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, pp. 1108–1116. International Foundation for Autonomous Agents and Multiagent Systems (2019)
Mnih, V., et al.: Asynchronous methods for deep reinforcement learning. In: International Conference on Machine Learning, pp. 1928–1937 (2016)
Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529–533 (2015)
Pham, T.H., De Magistris, G., Tachibana, R.: Optlayer-practical constrained optimization for deep reinforcement learning in the real world. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 66–6243. IEEE (2018)
Pinkus, A.: Approximation theory of the MLP model in neural networks. Acta Numerica 8, 143–195 (1999)
Silver, D., et al.: Mastering the game of go with deep neural networks and tree search. Nature 529(7587), 484 (2016)
Silver, D., Lever, G., Heess, N., Degris, T., Wierstra, D., Riedmiller, M.: Deterministic policy gradient algorithms (2014)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)
Sutton, R.S., McAllester, D.A., Singh, S.P., Mansour, Y.: Policy gradient methods for reinforcement learning with function approximation. In: Advances in Neural Information Processing Systems, pp. 1057–1063 (2000)
Tang, H., Kerber, M., Huang, Q., Guibas, L.: Locating lucrative passengers for taxicab drivers. In: Proceedings of the 21st ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, pp. 504–507 (2013)
Wang, S., Li, L., Ma, W., Chen, X.: Trajectory analysis for on-demand services: a survey focusing on spatial-temporal demand and supply patterns. Transp. Res. Part C: Emerg. Technol. 108, 74–99 (2019)
Wen, J., Zhao, J., Jaillet, P.: Rebalancing shared mobility-on-demand systems: a reinforcement learning approach. In: 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), pp. 220–225. IEEE (2017)
Yang, Y., Luo, R., Li, M., Zhou, M., Zhang, W., Wang, J.: Mean field multi-agent reinforcement learning. arXiv preprint arXiv:1802.05438 (2018)
Yuan, N.J., Zheng, Y., Zhang, L., Xie, X.: T-finder: a recommender system for finding passengers and vacant taxis. IEEE Trans. Knowl. Data Eng. 25(10), 90–2403 (2012)
Zhang, R., Ghanem, R.: Demand, supply, and performance of street-hail taxi. IEEE Trans. Intell. Transp. Syst. 21, 4123–4132 (2019)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, C., Deng, M., Chen, C., Xiang, C. (2020). A Driver-Centric Vehicle Reposition Framework via Multi-agent Reinforcement Learning. In: Yu, Z., Becker, C., Xing, G. (eds) Green, Pervasive, and Cloud Computing. GPC 2020. Lecture Notes in Computer Science(), vol 12398. Springer, Cham. https://doi.org/10.1007/978-3-030-64243-3_17
Download citation
DOI: https://doi.org/10.1007/978-3-030-64243-3_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-64242-6
Online ISBN: 978-3-030-64243-3
eBook Packages: Computer ScienceComputer Science (R0)