Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Application of Reinforcement Learning for Intelligent Support Decision System: A Paradigm Towards Safety and Explainability

  • Conference paper
  • First Online:
Artificial Intelligence in HCI (HCII 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14050))

Included in the following conference series:

  • 1874 Accesses

Abstract

Artificial Intelligence (AI) offers the potential to transform our lives in radical ways. In particular, when AI is combined with the rapid development of mobile communication and advanced sensors, this allows autonomous driving (AD) to make a great progress. In fact, Autonomous Vehicles (AVs) can mitigate some shortcomings of manual driving, but at the same time the underlying technology is not yet mature enough to be widely applied in all scenarios and for all types of vehicles. In this context, the traditional SAE-levels of automation (J3016B: Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles—SAE International. Available online: https://www.sae.org/standards/content/j3016_201806/) can lead to uncertain and ambiguous situations, so yielding to a great risk in the control of the vehicle. In this context, the human drivers should be supported to take the right decision, especially on those edge-cases where automation can fail. A decision-making system is well designed if it can augment human cognition and emphasize human judgement and intuition. It is worth to noting here that such systems should not be considered as teammates or collaborators, because humans are responsible for the final decision and actions, but the technology can assist them, reducing workload, raising performances and ensuring safety. The main objective of this paper is to present an intelligent decision support system (IDSS), in order to provide the optimal decision, about which is the best action to perform, by using an explainable and safe paradigm, based on AI techniques.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 159.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    In other words, while AVs aims at revolutionizing our “consolidated concept” of transportation, at the same time they introduce new challenges. One of them is the takeover transitions in conditionally automated driving (SAE-L3), where drivers are no longer required to actively monitor the driving environment and can be allowed to fully engage in non-driving-related tasks (NDRTs), but at the same time they are still regarded as a fallback mechanism for the automation, requiring to take the control of the vehicle back, when the automation reaches the limits of its ODD (Operational Design Domain), considering that the situational understanding and prediction capabilities of AVs are at the moment far less sophisticated than the capabilities of human drivers.

  2. 2.

    Taking into account that there is also the risk that humans lose some skills, thus fundamental changes can occur to what humans are expected to learn.

References

  1. Chu, D., Li, H., Zhao, C., Zhou, T.: Trajectory tracking of autonomous vehicle based on model predictive control with pid feedback. IEEE Trans. Intell. Transp. Syst. 23, 1–12 (2022). https://doi.org/10.1109/TITS.2022.3150365

  2. Marcano, M., et al.: From the concept of being “the Boss” to the idea of being “a Team”: the adaptive Co-Pilot as the enabler for a new cooperative framework. Appl. Sci. 11(15), 6950 (2021). https://doi.org/10.3390/app11156950

    Article  Google Scholar 

  3. Huang, C., Lv, C., Hang, P., Hu, Z., Xing, Y.: Human–machine adaptive shared control for safe driving under automation degradation. IEEE Intell. Transp. Syst. Mag. 14(2), 53–66 (2021)

    Article  Google Scholar 

  4. Deng, H., Zhao, Y., Feng, S., Wang, Q., Lin, F.: Shared control for intelligent vehicle based on handling inverse dynamics and driving intention. IEEE Trans. Veh. Technol. 71(3), 2706–2720 (2022)

    Article  Google Scholar 

  5. Russell, H.E.B., Harbott, L.K., Nisky, I., Pan, S., Okamura, A.M., Christian Gerdes, J.: Motor learning affects car-to-driver handover in automated vehicles. Sci. Robot. 1(1), eaah5682 (2016). https://doi.org/10.1126/scirobotics.aah5682

    Article  Google Scholar 

  6. Flemisch, F., Schieben, A., Schoemig, N., Strauss, M., Lueke, S., Heyden, A.: Design of human computer interfaces for highly automated vehicles in the EU-project HAVEit. In: Stephanidis, C. (ed.) UAHCI 2011. LNCS, vol. 6767, pp. 270–279. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21666-4_30

  7. Flemish, F.O., Goodrich, K.H., Adams, A.A., Conway, S.R., Palmer, M.T., Schutte, P.C.: The H-Metaphor as a guideline for vehicle automation and interaction. University of Munich: Munich, Germany (2003). http://www.sti.nasa.gov. Accessed 24 May 2021

  8. Bainbridge, L.: Ironies of automation. Automatica 19, 775–779 (1983). https://doi.org/10.1016/0005-1098(83)90046-8

    Article  Google Scholar 

  9. Benloucif, A., Nguyen, A.-T., Sentouh, C., Popieul, J.-C.: Cooperative trajectory planning for haptic shared control between driver and automation in highway driving. IEEE Trans. Industr. Electron. 66(12), 9846–9857 (2019)

    Article  Google Scholar 

  10. Wang, W., et al.: Decision-making in driver-automation shared control: a review and perspectives. IEEE/CAA J. Automatica Sinica 7(5), 1289–1307 (2020)

    Google Scholar 

  11. Castellano, A., Karimshoushtari, M., Novara, C., Tango, F.: A supervisor agent-based on the markovian decision process framework to optimize the behavior of a highly automated system. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) Augmented Cognition. Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence), vol. 12776, pp. 351–368. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-78114-9_24

    Chapter  Google Scholar 

  12. Shneiderman, B.: Human-centered artificial intelligence: reliable, safe & trustworthy. Int. J. Hum. Comput. Inter. 36(6), 495–504 (2020). https://doi.org/10.1080/10447318.2020.1741118

    Article  Google Scholar 

  13. Madl, T., Baars, B.J., Franklin, S.: The timing of the cognitive cycle. PLoS ONE 6(4), e14803 (2011). https://doi.org/10.1371/journal.pone.0014803

    Article  Google Scholar 

  14. Rasmussen, J.: Skills, rules, and knowledge; signals, signs, and symbols, and other distinctions in human performance models. IEEE Trans. Syst. Man Cybern. SMC-13(3), 257–266 (1983). https://doi.org/10.1109/TSMC.1983.6313160

    Article  Google Scholar 

  15. Shneiderman, B.: Human-centered artificial intelligence: three fresh ideas. AIS Trans. Hum. Comput. Inter. 12, 109–124 (2020). https://doi.org/10.17705/1thci.00131

  16. Poler, R., Mula, J., Díaz-Madroñero, M.: Dynamic Programming. In: Operations Research Problems, pp. 325–374. Springer, London (2014). https://doi.org/10.1007/978-1-4471-5577-5_9

    Chapter  MATH  Google Scholar 

  17. Barto A.G.: Reinforcement learning: an introduction (Adaptive Computation and Machine Learning), 3rd ed. The MIT press (1998 ) (cit. on p. 18)

    Google Scholar 

  18. Olivier, P., Tango, F.: A reinforcement learning approach to optimize the longitudinal behavior of a partial autonomous driving assistance system. In: ECAI 2012. IOS Press, pp. 987–992 (2012) (cit. on p. 20)

    Google Scholar 

  19. Melo, F.S.: Convergence of q-learning: a simple proof. in: institute of systems and robotics. Tech. Rep, pp. 1–4 (2001) (cit. on p. 20)

    Google Scholar 

  20. Dayan, P., Watkins, C.J.C.H.: Q-learning. Mach. Learn. 8(3), 279–292 (1992) 

    Google Scholar 

  21. ERTRAC: “Connected automated driving roadmap.” https://www.ertrac.org/uploads/documentsearch/id57/ERTRAC-CAD-Roadmap-2019.pdf (2019)

  22. Jerry, W.:  By what Hubris? The readiness of the human operator to take over when the automation fails or hands over control. In: Proceedings of the DDI2018 6th International Conference on Driver Distraction and Inattention, Gothenburg, Sweden, 15–17 October 2018, pp. 182–184 (2018)

    Google Scholar 

  23. Turchetti, C.: Stochastic models of Neural Networks. IOS Press (2004) (cit. on p. 28)

    Google Scholar 

  24. Insurance Institute for Highway Safety. Self-driving vehicles could struggle to eliminate most crashes (2020).  https://www.iihs.org/news/detail/self-driving-vehicles-could-struggle-to-eliminate-most-crashes

  25. Bonnefon, J.F., Shariff, A., Rahwan, I.: The social dilemma of autonomous vehicles. Science 352(6293), 1573–1576 (2016). https://doi.org/10.1126/science.aaf2654

    Article  Google Scholar 

Download references

Acknowledgment

This work was supported by the NewControl project, within the Electronic Components and Systems For European Leadership Joint Undertaking (ESCEL JU) in collaboration with the European Union’s Horizon2020 Framework Programme and National Authorities, under grant agreement N° 826653–2.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Carlo Novara .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Maiuri, C., Karimshoushtari, M., Tango, F., Novara, C. (2023). Application of Reinforcement Learning for Intelligent Support Decision System: A Paradigm Towards Safety and Explainability. In: Degen, H., Ntoa, S. (eds) Artificial Intelligence in HCI. HCII 2023. Lecture Notes in Computer Science(), vol 14050. Springer, Cham. https://doi.org/10.1007/978-3-031-35891-3_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-35891-3_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-35890-6

  • Online ISBN: 978-3-031-35891-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics