Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3394810.3394815acmotherconferencesArticle/Chapter ViewAbstractPublication PagesrtnsConference Proceedingsconference-collections
research-article

Adaptive Routing with Guaranteed Delay Bounds using Safe Reinforcement Learning

Published: 12 June 2020 Publication History

Abstract

Time-critical networks require strict delay bounds on the transmission time of packets from source to destination. Routes for transmissions are usually statically determined, using knowledge about worst-case transmission times between nodes. This is generally a conservative method, that guarantees transmission times but does not provide any optimization for the typical case. In real networks, the typical delays vary from those considered during static route planning. The challenge in such a scenario is to minimize the total delay from a source to a destination node, while adhering to the timing constraints. For known typical and worst-case delays, an algorithm was presented to (statically) determine the policy to be followed during the packet transmission in terms of edge choices.
In this paper we relax the assumption of knowing the typical delay, and we assume only worst-case bounds are available. We present a reinforcement learning solution to obtain optimal routing paths from a source to a destination when the typical transmission time is stochastic and unknown. Our reinforcement learning policy is based on the observation of the state-space during each packet transmission and on adaptation for future packets to congestion and unpredictable circumstances in the network. We ensure that our policy only makes safe routing decisions, thus never violating pre-determined timing constraints. We conduct experiments to evaluate the routing in a congested network and in a network where the typical delays have a large variance. Finally, we analyze the application of the algorithm to large randomly generated networks.

References

[1]
J. A. Clouse and P. E. Utgoff. A teaching method for reinforcement learning. In Proceedings of the Ninth International Conference on Machine Learning, pages 92--110, 12 1992.
[2]
P. Abbeel, A. Coates, and A. Y. Ng. Autonomous helicopter aerobatics through apprenticeship learning. Int. J. Rob. Res., 29(13):1608--1639, Nov. 2010.
[3]
S. Baruah. Rapid routing with guaranteed delay bounds. In 2018 IEEE Real-Time Systems Symposium (RTSS), pages 13--22, December 2018.
[4]
D. P. Bertsekas and J. N. Tsitsiklis. An analysis of stochastic shortest path problems. Math. Oper. Res., 16(3):580--595, Aug. 1991.
[5]
J. Carlström. Decomposition of reinforcement learning for admission control of self-similar call arrival processes. In Proceedings of the 13th International Conference on Neural Information Processing Systems, NIPS'00, pages 989--995. MIT Press, 2000.
[6]
E. W. Dijkstra. A note on two problems in connexion with graphs. Numerische Mathematik, 1(1):269--271, Dec 1959.
[7]
K. Driessens and S. Džeroski. Integrating guidance into relational reinforcement learning. Machine Learning, 57(3):271--304, Dec 2004.
[8]
J. Garcia and F. Fernandez. Safe exploration of state and action spaces in reinforcement learning. CoRR, abs/1402.0560, 2014.
[9]
J. García and F. Fernández. A comprehensive survey on safe reinforcement learning. Journal on Machine Learning Research, 16(1):1437--1480, Jan. 2015.
[10]
A. A. Hagberg, D. A. Schult, and P. J. Swart. Exploring network structure, dynamics, and function using networkx. In G. Varoquaux, T. Vaught, and J. Millman, editors, Proceedings of the 7th Python in Science Conference, pages 11--15, Pasadena, CA USA, 2008.
[11]
R. P. Loui. Optimal paths in graphs with stochastic or multidimensional weights. Commun. ACM, 26(9):670--676, Sept. 1983.
[12]
K. Mehlhorn and P. Sanders. Algorithms and Data Structures: The Basic Toolbox. Springer Publishing Company, Incorporated, 1 edition, 2008.
[13]
L. Peshkin and V. Savova. Reinforcement learning for adaptive routing. CoRR, abs/cs/0703138, 2007.
[14]
G. H. Polychronopoulos. Stochastic and Dynamic Shortest Distance Problems. PhD thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1992.
[15]
W. D. Smart and L. P. Kaelbling. Practical reinforcement learning in continuous spaces. In Proceedings of the Seventeenth International Conference on Machine Learning, ICML '00, pages 903--910, San Francisco, CA, USA, 2000. Morgan Kaufmann Publishers Inc.
[16]
R. S. Sutton and A. G. Barto. Reinforcement learning: An Introduction. Adaptive computation and machine learning. MIT Press, 2018.
[17]
J. Tang, A. Singh, N. Goehausen, and P. Abbeel. Parameterized maneuver learning for autonomous helicopter flight. In 2010 IEEE International Conference on Robotics and Automation, pages 1142--1148, May 2010.
[18]
G. Tesauro. Temporal difference learning and td-gammon. Commun. ACM, 38(3):58--68, Mar. 1995.
[19]
A. L. Thomaz and C. Breazeal. Reinforcement learning with human teachers: Evidence of feedback and guidance with implications for learning performance. In Proceedings of the 21st National Conference on Artificial Intelligence - Volume 1, AAAI'06, pages 1000--1005. AAAI Press, 2006.
[20]
H. Tong and T. X. Brown. Reinforcement learning for call admission control and routing under quality of service constraints in multimedia networks. Machine Learning, 49(2):111--139, Nov 2002.
[21]
P. Q. Vidal, R. I. Rodríguez, M. Ángel Rodríguez González, and C. V. Regueiro. Learning on real robots from experience and simple user feedback. Journal of Physical Agents, 7(1):57--65, 2013.

Cited By

View all
  • (2024)Deep Distributional Reinforcement Learning-Based Adaptive Routing With Guaranteed Delay BoundsIEEE/ACM Transactions on Networking10.1109/TNET.2024.342565232:6(4692-4706)Online publication date: Dec-2024
  • (2023)A robust routing strategy based on deep reinforcement learning for mega satellite constellationsElectronics Letters10.1049/ell2.1282059:11Online publication date: 6-Jun-2023
  • (2022)IQoR-LSE: An Intelligent QoS On-Demand Routing Algorithm With Link State EstimationIEEE Systems Journal10.1109/JSYST.2022.314999016:4(5821-5830)Online publication date: Dec-2022
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
RTNS '20: Proceedings of the 28th International Conference on Real-Time Networks and Systems
June 2020
177 pages
ISBN:9781450375931
DOI:10.1145/3394810
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

In-Cooperation

  • INRIA: INRIA Saclay Île-de-France

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 12 June 2020

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

RTNS 2020

Acceptance Rates

Overall Acceptance Rate 119 of 255 submissions, 47%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)31
  • Downloads (Last 6 weeks)0
Reflects downloads up to 28 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Deep Distributional Reinforcement Learning-Based Adaptive Routing With Guaranteed Delay BoundsIEEE/ACM Transactions on Networking10.1109/TNET.2024.342565232:6(4692-4706)Online publication date: Dec-2024
  • (2023)A robust routing strategy based on deep reinforcement learning for mega satellite constellationsElectronics Letters10.1049/ell2.1282059:11Online publication date: 6-Jun-2023
  • (2022)IQoR-LSE: An Intelligent QoS On-Demand Routing Algorithm With Link State EstimationIEEE Systems Journal10.1109/JSYST.2022.314999016:4(5821-5830)Online publication date: Dec-2022
  • (2022)Reinforcement Learning assisted Routing for Time Sensitive NetworksGLOBECOM 2022 - 2022 IEEE Global Communications Conference10.1109/GLOBECOM48099.2022.10001630(3863-3868)Online publication date: 4-Dec-2022
  • (2022)MALOC: Building an adaptive scheduling and routing framework for rate-constrained TSN traffic2022 IEEE 27th International Conference on Emerging Technologies and Factory Automation (ETFA)10.1109/ETFA52439.2022.9921474(1-4)Online publication date: 6-Sep-2022
  • (2021)DRL-ER: An Intelligent Energy-Aware Routing Protocol With Guaranteed Delay Bounds in Satellite Mega-ConstellationsIEEE Transactions on Network Science and Engineering10.1109/TNSE.2020.30394998:4(2872-2884)Online publication date: 1-Oct-2021
  • (2021)RECCE: Deep Reinforcement Learning for Joint Routing and Scheduling in Time-Constrained Wireless NetworksIEEE Access10.1109/ACCESS.2021.31149679(132053-132063)Online publication date: 2021
  • (2020)Hard-Real-Time Routing in Probabilistic Graphs to Minimize Expected Delay2020 IEEE Real-Time Systems Symposium (RTSS)10.1109/RTSS49844.2020.00017(63-75)Online publication date: Dec-2020

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media