A Machine Learning-Based Approach for Multi-AGV Dispatching at Automated Container Terminals
Abstract
:1. Introduction
2. Literature Review
2.1. Scheduling Problem Research
2.2. Machine Learning Methods for Optimization
2.3. Research on Contract Net Protocol
3. Problem Description and Model Formulation
3.1. Problem Description
3.2. Model Formulation
4. Approach Design
4.1. Agent
4.2. State
4.3. Action
4.4. Reward Function
4.5. Algorithm Implementation
- (1)
- Initialize the Q-table recording the state, action, reward, and agent;
- (2)
- Observe the environment of multiagent system, including states and agents;
- (3)
- Release the transportation tasks by the manger agent;
- (4)
- Submit the requirement proposal by agents;
- (5)
- Choose an action for the agent;
- (6)
- Calculate the reward of the state and action;
- (7)
- Update the Q-table;
- (8)
- Announce the results of the bid by the manager agent;
- (9)
- Contract the protocol between the manager and bidder agent;
- (10)
- Cycle in turn until all task assignments end.
5. Case Study
5.1. Experiment Design
5.2. Experimental Results
5.2.1. Small-Scale Case
5.2.2. Large-Scale Case
5.3. Comparison of Different Approaches
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A
Task ID | Container Block | QC | Earliest Time | Latest Time |
---|---|---|---|---|
15151050 | 5 | 2 | 0 | 2100 |
15150726 | 4 | 1 | 0 | 720 |
15151322 | 3 | 2 | 60 | 3600 |
15152214 | 1 | 1 | 60 | 8400 |
15155329 | 4 | 2 | 120 | 26520 |
15151520 | 2 | 1 | 180 | 4560 |
15151632 | 3 | 1 | 240 | 4740 |
15153108 | 2 | 1 | 240 | 13020 |
15151248 | 1 | 1 | 240 | 3240 |
15153185 | 3 | 1 | 300 | 13560 |
References
- Soylu, M.; Özdemirel, N.E.; Kayaligil, S. A self-organizing neural network approach for the single AGV routing problem. Eur. J. Oper. Res. 2000, 121, 124–137. [Google Scholar] [CrossRef]
- Yang, Y.; Zhong, M.; Dessouky, Y.; Postolache, O. An integrated scheduling method for AGV routing in automated container terminals. Comput. Ind. Eng. 2018, 126, 482–493. [Google Scholar] [CrossRef]
- Nishi, T.; Hiranaka, Y.; Grossmann, I.E. A bilevel decomposition algorithm for simultaneous production scheduling and conflict-free routing for automated guided vehicles. Comput. Oper. Res. 2011, 38, 876–888. [Google Scholar] [CrossRef]
- Xin, J.; Negenborn, R.R.; Corman, F.; Lodewijks, G. Control of interacting machines in automated container terminals using a sequential planning approach for collision avoidance. Transp. Res. Part C Emerg. Technol. 2015, 60, 377–396. [Google Scholar] [CrossRef]
- Zhong, M.; Yang, Y.; Dessouky, Y.; Postolache, O. Multi-AGV scheduling for conflict-free path planning in automated container terminals. Comput. Ind. Eng. 2020, 142, 106371. [Google Scholar] [CrossRef]
- Yue, L.; Fan, H.; Ma, M. Optimizing configuration and scheduling of double 40 ft dual-trolley quay cranes and AGVs for improving container terminal services. J. Clean. Prod. 2021, 292, 126019. [Google Scholar] [CrossRef]
- Roy, D.; Gupta, A.; De Koster, R.B. A non-linear traffic flow-based queuing model to estimate container terminal throughput with AGVs. Int. J. Prod. Res. 2016, 54, 472–493. [Google Scholar] [CrossRef]
- Salman, S.; Alaswad, S. Alleviating road network congestion: Traffic pattern optimization using markov chain traffic assignment. Comput. Oper. Res. 2018, 99, 191–205. [Google Scholar] [CrossRef]
- Małopolski, W. A sustainable and conflict-free operation of AGVs in a square topology. Comput. Ind. Eng. 2018, 126, 472–481. [Google Scholar] [CrossRef]
- Li, K.; Zhou, T.; Liu, B.H.; Li, H. A multi-agent system for sharing distributed manufacturing resources. Expert Syst. Appl. 2018, 99, 32–43. [Google Scholar] [CrossRef]
- Framiñán Torres, J.M.; Pérez González, P.; Fernández-Viagas Escudero, V.; González, V. Assessing the potential of decentralised scheduling: An experimental study for the job shop case. In 10th IFAC Conference on Manufacturing Modelling, Management and Control, MIM 2022. IFAC PapersOnLine 2022, 55, 2617–2622. [Google Scholar]
- Singh, N.; Dang, Q.V.; Akcay, A.; Adan, I.; Martagan, T. A matheuristic for AGV scheduling with battery constraints. Eur. J. Oper. Res. 2022, 298, 855–873. [Google Scholar] [CrossRef]
- Yu, N.N.; Li, T.K.; Wang, B.L.; Yuan, S.P.; Wang, Y. Reliability oriented multi-AGVs online scheduling and path planning problem of automated sorting warehouse system. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1043, 22035. [Google Scholar] [CrossRef]
- Li, Q.; Pogromsky, A.; Adriaansen, T.; Udding, J.T. A control of collision and deadlock avoidance for automated guided vehicles with a fault-tolerance capability. Int. J. Adv. Robot. Syst. 2016, 13, 64. [Google Scholar] [CrossRef] [Green Version]
- Zhong, M.; Yang, Y.; Sun, S.; Zhou, Y.; Postolache, O.; Ge, Y.E. Priority-based speed control strategy for automated guided vehicle path planning in automated container terminals. Trans. Inst. Meas. Control 2020, 42, 3079–3090. [Google Scholar] [CrossRef]
- Li, C.; Zhang, L.; Zhang, L. A route and speed optimization model to find conflict-free routes for automated guided vehicles in large warehouses based on quick response code technology. Adv. Eng. Inform. 2022, 52, 101604. [Google Scholar] [CrossRef]
- Xin, J.; Wei, L.; Wang, D.; Xuan, H. Receding horizon path planning of automated guided vehicles using a time-space network model. Optim. Control. Appl. Methods 2020, 41, 1889–1903. [Google Scholar] [CrossRef]
- Xin, J.; Wei, L.; D’Ariano, A.; Zhang, F.; Negenborn, R. Flexible time-space network formulation and hybrid metaheuristic for conflict-free and energy-efficient path planning of automated guided vehicles. J. Clean. Prod. 2023, 398, 136472. [Google Scholar] [CrossRef]
- Luo, J.; Wu, Y. Modelling of dual-cycle strategy for container storage and vehicle scheduling problems at automated container terminals. Transp. Res. Part E Logist. Transp. Rev. 2015, 79, 49–64. [Google Scholar] [CrossRef]
- Yang, Y.; He, S.; Sun, S. Research on the cooperative scheduling of ARMGs and AGVs in a sea-rail automated container terminal under the rail-in-port model. J. Mar. Sci. Eng. 2023, 11, 557. [Google Scholar] [CrossRef]
- Ji, S.; Luan, D.; Chen, Z.; Guo, D. Integrated scheduling in automated container terminals considering AGV conflict-free routing. Transp. Lett. 2021, 13, 501–513. [Google Scholar]
- Wang, Q.; Tang, C. Deep reinforcement learning for transportation network combinatorial optimization: A survey. Knowl. Based Syst. 2021, 233, 107526. [Google Scholar] [CrossRef]
- Gumuskaya, V.; van Jaarsveld, W.; Dijkman, R.; Grefen, P.; Veenstra, A. Integrating stochastic programs and decision trees in capacitated barge planning with uncertain container arrivals. Transp. Res. Part C Emerg. Technol. 2021, 132, 103383. [Google Scholar] [CrossRef]
- Peng, Y.; Liu, H.; Li, X.; Huang, J.; Wang, W. Machine learning method for energy consumption prediction of ships in port considering green ports. J. Clean. Prod. 2020, 264, 121564. [Google Scholar] [CrossRef]
- Gao, Y.; Chang, D.; Chen, C.H. A digital twin-based approach for optimizing operation energy consumption at automated container terminals. J. Clean. Prod. 2023, 385, 135782. [Google Scholar] [CrossRef]
- Kintsakis, A.M.; Psomopoulos, F.E.; Mitkas, P.A. Reinforcement learning based scheduling in a workflow management system. Eng. Appl. Artif. Intell. 2019, 81, 94–106. [Google Scholar] [CrossRef]
- Kim, T.; Kim, Y.W.; Lee, D.; Kim, M. Reinforcement learning approach to scheduling of precast concrete production. J. Clean. Prod. 2022, 336, 130419. [Google Scholar] [CrossRef]
- Lin, C.C.; Chen, K.Y.; Hsieh, L.T. Real-time charging scheduling of automated guided vehicles in cyber-physical smart factories using feature-based reinforcement learning. IEEE Trans. Intell. Transp. Syst. 2023, 24, 4016–4026. [Google Scholar] [CrossRef]
- Cals, B.; Zhang, Y.; Dijkman, R.; van Dorst, C. Solving the online batching problem using deep reinforcement learning. Comput. Ind. Eng. 2021, 156, 107221. [Google Scholar] [CrossRef]
- James, J.Q.; Yu, W.; Gu, J. Online vehicle routing with neural combinatorial optimization and deep reinforcement learning. IEEE Trans. Intell. Transp. Syst. 2019, 20, 3806–3817. [Google Scholar]
- Ahamed, T.; Zou, B.; Farazi, N.P.; Tulabandhula, T. Deep reinforcement learning for Crowdsourced urban delivery. Transp. Res. Part B Methodol. 2021, 152, 227–257. [Google Scholar] [CrossRef]
- Zhang, K.; He, F.; Zhang, Z.; Lin, X.; Li, M. Multi-vehicle routing problems with soft time windows: A multi-agent reinforcement learning approach. Transp. Res. Part C Emerg. Technol. 2020, 121, 102861. [Google Scholar] [CrossRef]
- Jelen, G.; Babic, J.; Podobnik, V. A multi-agent system for context-aware electric vehicle fleet routing: A step towards more sustainable urban operations. J. Clean. Prod. 2022, 374, 134047. [Google Scholar] [CrossRef]
- Choe, R.; Kim, J.; Ryu, K.R. Online preference learning for adaptive dispatching of AGVs in an automated container terminal. Appl. Soft Comput. 2016, 38, 647–660. [Google Scholar] [CrossRef]
- Hu, H.; Yang, X.; Xiao, S.; Wang, F. Anti-conflict agv path planning in automated container terminals based on multi-agent reinforcement learning. Int. J. Prod. Res. 2023, 61, 65–80. [Google Scholar] [CrossRef]
- Smith, R.G. The contract net protocol: High-level communication and control in a distributed problem solver. IEEE Trans. Comput. 1980, 29, 1104–1113. [Google Scholar] [CrossRef]
- Gao, J.; Wong, T.; Wang, C. Coordinating patient preferences through automated negotiation: A multiagent systems model for diagnostic services scheduling. Adv. Eng. Inform. 2019, 42, 100934. [Google Scholar] [CrossRef]
- Jégou, D.; Kim, D.W.; Baptiste, P.; Lee, K.H. A contract net based intelligent agent system for solving the reactive hoist scheduling problem. Expert Syst. Appl. 2006, 30, 156–167. [Google Scholar] [CrossRef]
- Cardon, A.; Vacher, J.P. Multi-objective genetic agents based on a contract-net system for job-shop scheduling problems. IFAC Proc. 2000, 33, 951–956. [Google Scholar] [CrossRef]
- Liang, H.; Kang, F. A novel task optimal allocation approach based on contract net protocol for agent-oriented UUV swarm system modeling. Optik 2016, 127, 3928–3933. [Google Scholar] [CrossRef]
- Zhen, Z.; Wen, L.; Wang, B.; Hu, Z.; Zhang, D. Improved contract network protocol algorithm based cooperative target allocation of heterogeneous UAV swarm. Aerosp. Sci. Technol. 2021, 119, 107054. [Google Scholar] [CrossRef]
- González-Briones, A.; De La Prieta, F.; Mohamad, M.S.; Omatu, S.; Corchado, J.M. Multi-agent systems applications in energy optimization problems: A State-of-the-Art Review. Energies 2018, 11, 1928. [Google Scholar] [CrossRef] [Green Version]
- Chen, H.; Ji, Y.; Niu, L. Reinforcement learning path planning algorithm based on obstacle area expansion strategy. Intell. Serv. Robot. 2020, 13, 289–297. [Google Scholar] [CrossRef]
- Qiu, L.; Hsu, W.J.; Huang, S.Y.; Wang, H. Scheduling and routing algorithms for AGVs: A survey. Int. J. Prod. Res. 2002, 40, 745–760. [Google Scholar] [CrossRef]
- Soltani, A.R.; Tawfik, H.; Goulermas, J.Y.; Fernando, T. Path planning in construction sites: Performance evaluation of the Dijkstra, A∗, and GA search algorithms. Adv. Eng. Inform. 2002, 16, 291–303. [Google Scholar] [CrossRef]
- Sun, P.; Shan, R. Predictive control with velocity observer for cushion robot based on PSO for path planning. J. Syst. Sci. Complex. 2020, 33, 988–1011. [Google Scholar] [CrossRef]
Problems | Solutions | Authors |
---|---|---|
AGV scheduling considering battery to minimize travel and tardiness costs | A metaheuristic-based adaptive large neighborhood search algorithm | Singh et al. [12] |
Multi-AGV online scheduling in warehouses | Improved path planning algorithm with grid blocking degree | Yu et al. [13] |
AGV conflict-free path planning to minimize distances | Dijkstra depth-first search algorithm | Zhong et al. [5] |
AGV conflict-free path planning to minimize energy consumption | A branch and bound algorithm | Li et al. [16] |
AGV path planning in MVPD problem | A receding horizon planning methodology | Xin et al. [17] |
AGV dispatching and container storage allocation | GA | Luo & Wu [19] |
Integrated scheduling considering AGV conflicts | An approach based on the conflict resolution strategy and bilevel adaptive GA | Ji et al. [21] |
AGV online dispatching | An online preference learning algorithm | Choe et al. [34] |
AGV anti-conflict path at ACT | MADDPG combining reinforcement learning | Hu et al. [35] |
Sets and indices | ||
Parameters | and | |
in the time window | ||
in the time window | ||
The giant number | ||
Decision variables | ||
; | ||
Parameters | ||
The acceleration of AGV | ||
The deceleration time and the acceleration time of AGV | ||
The safe distance of AGV | ||
The distance of the path p formed by two nodes | ||
Variables | AGV k transporting the task i chooses the path p | |
AGV k and k′ meets at the path p | ||
The waiting time of AGV during the transportation | ||
The total transportation time of AGV passing paths |
Initialize the environment for agents transporting Create a new m × n-dimensional environment matrix Construct a reward matrix Set parameters of QL |
For episode = 1 to M |
Obtain the state of tasks and agents Select all possible actions of the state Calculate Q-value of all possible actions Select one action and take it as the next state Find the best Q-value which is the best state for the next action Update the Q-table |
If to the end Break; End if |
End for |
Action | 1 | 2 | 3 | 4 | |
---|---|---|---|---|---|
State | |||||
1 | 0.00 | 0.00 | 0.00 | 0.00 | |
2 | 0.86 | 0.35 | 0.46 | 0.86 | |
3 | 0.96 | 0.28 | 0.14 | 0.90 | |
4 | 0.61 | 0.40 | 0.10 | 0.79 | |
5 | 0.81 | 0.08 | 0.03 | 0.94 |
AGV Number | Task Sequence | Passed Nodes |
---|---|---|
AGV 1 | Task 2–Task 4–Task 8 | [59, 58, 57, 56, 55, 49, 37, 25, 13, 14, 2, 3, 4]–[55, 49, 37, 38, 26, 14, 2, 3, 4]–[57, 56, 55, 49, 37, 38, 26, 14, 2, 3, 4] |
AGV 2 | Task 3–Task 5–Task 6–Task 10 | [57, 56, 53, 45, 33, 21, 22, 10]–[59, 58, 57, 51, 41, 42, 30, 18, 6, 7, 8, 9, 10]–[57, 56, 55, 49, 37, 25, 26, 14, 2, 3, 4]–[59, 58, 57, 56, 55, 49, 37, 38, 39, 27, 15, 3, 4] |
AGV 3 | Task 1–Task 7–Task 9 | [60, 59, 58, 57, 51, 41, 29, 30, 18, 6, 7, 8, 9, 10]–[57, 56, 55, 49, 37, 25, 13, 1, 2, 3, 4]–[55, 49, 37, 25, 26, 14, 2, 3, 4] |
Experiment ID | Number of Tasks | Number of AGVs | Delay Time (s) | Congestion Rate |
---|---|---|---|---|
1 | 20 | 3 | 1754.83 | 3.14% |
2 | 30 | 4 | 3339.53 | 3.06% |
3 | 40 | 4 | 4222.47 | 2.74% |
4 | 50 | 5 | 4811.17 | 2.50% |
5 | 60 | 5 | 5750.67 | 2.44% |
6 | 80 | 6 | 8068.1 | 2.24% |
7 | 100 | 7 | 8773.6 | 1.93% |
8 | 120 | 8 | 10,346.24 | 1.82% |
9 | 150 | 9 | 12,860.38 | 1.75% |
10 | 200 | 10 | 12,994.32 | 1.63% |
QL-CNP | CNA | Dijkstra | ||||
---|---|---|---|---|---|---|
Task Scale | Congestion Rate | Time | Congestion Rate | Time | Congestion Rate | Time |
50 | 2.50% | 444 | 3.99% | 4738 | 4.02% | 3049 |
100 | 1.93% | 529 | 3.06% | 7680 | 3.12% | 4738 |
150 | 1.75% | 1020 | NA | NA | 2.87% | 4918 |
Average | 2.06% | 664.33 | 3.525% | 6209 | 3.34% | 4235 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gao, Y.; Chen, C.-H.; Chang, D. A Machine Learning-Based Approach for Multi-AGV Dispatching at Automated Container Terminals. J. Mar. Sci. Eng. 2023, 11, 1407. https://doi.org/10.3390/jmse11071407
Gao Y, Chen C-H, Chang D. A Machine Learning-Based Approach for Multi-AGV Dispatching at Automated Container Terminals. Journal of Marine Science and Engineering. 2023; 11(7):1407. https://doi.org/10.3390/jmse11071407
Chicago/Turabian StyleGao, Yinping, Chun-Hsien Chen, and Daofang Chang. 2023. "A Machine Learning-Based Approach for Multi-AGV Dispatching at Automated Container Terminals" Journal of Marine Science and Engineering 11, no. 7: 1407. https://doi.org/10.3390/jmse11071407