Abstract
Artificial Intelligence (AI) offers the potential to transform our lives in radical ways. In particular, when AI is combined with the rapid development of mobile communication and advanced sensors, this allows autonomous driving (AD) to make a great progress. In fact, Autonomous Vehicles (AVs) can mitigate some shortcomings of manual driving, but at the same time the underlying technology is not yet mature enough to be widely applied in all scenarios and for all types of vehicles. In this context, the traditional SAE-levels of automation (J3016B: Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles—SAE International. Available online: https://www.sae.org/standards/content/j3016_201806/) can lead to uncertain and ambiguous situations, so yielding to a great risk in the control of the vehicle. In this context, the human drivers should be supported to take the right decision, especially on those edge-cases where automation can fail. A decision-making system is well designed if it can augment human cognition and emphasize human judgement and intuition. It is worth to noting here that such systems should not be considered as teammates or collaborators, because humans are responsible for the final decision and actions, but the technology can assist them, reducing workload, raising performances and ensuring safety. The main objective of this paper is to present an intelligent decision support system (IDSS), in order to provide the optimal decision, about which is the best action to perform, by using an explainable and safe paradigm, based on AI techniques.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
In other words, while AVs aims at revolutionizing our “consolidated concept” of transportation, at the same time they introduce new challenges. One of them is the takeover transitions in conditionally automated driving (SAE-L3), where drivers are no longer required to actively monitor the driving environment and can be allowed to fully engage in non-driving-related tasks (NDRTs), but at the same time they are still regarded as a fallback mechanism for the automation, requiring to take the control of the vehicle back, when the automation reaches the limits of its ODD (Operational Design Domain), considering that the situational understanding and prediction capabilities of AVs are at the moment far less sophisticated than the capabilities of human drivers.
- 2.
Taking into account that there is also the risk that humans lose some skills, thus fundamental changes can occur to what humans are expected to learn.
References
Chu, D., Li, H., Zhao, C., Zhou, T.: Trajectory tracking of autonomous vehicle based on model predictive control with pid feedback. IEEE Trans. Intell. Transp. Syst. 23, 1–12 (2022). https://doi.org/10.1109/TITS.2022.3150365
Marcano, M., et al.: From the concept of being “the Boss” to the idea of being “a Team”: the adaptive Co-Pilot as the enabler for a new cooperative framework. Appl. Sci. 11(15), 6950 (2021). https://doi.org/10.3390/app11156950
Huang, C., Lv, C., Hang, P., Hu, Z., Xing, Y.: Human–machine adaptive shared control for safe driving under automation degradation. IEEE Intell. Transp. Syst. Mag. 14(2), 53–66 (2021)
Deng, H., Zhao, Y., Feng, S., Wang, Q., Lin, F.: Shared control for intelligent vehicle based on handling inverse dynamics and driving intention. IEEE Trans. Veh. Technol. 71(3), 2706–2720 (2022)
Russell, H.E.B., Harbott, L.K., Nisky, I., Pan, S., Okamura, A.M., Christian Gerdes, J.: Motor learning affects car-to-driver handover in automated vehicles. Sci. Robot. 1(1), eaah5682 (2016). https://doi.org/10.1126/scirobotics.aah5682
Flemisch, F., Schieben, A., Schoemig, N., Strauss, M., Lueke, S., Heyden, A.: Design of human computer interfaces for highly automated vehicles in the EU-project HAVEit. In: Stephanidis, C. (ed.) UAHCI 2011. LNCS, vol. 6767, pp. 270–279. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21666-4_30
Flemish, F.O., Goodrich, K.H., Adams, A.A., Conway, S.R., Palmer, M.T., Schutte, P.C.: The H-Metaphor as a guideline for vehicle automation and interaction. University of Munich: Munich, Germany (2003). http://www.sti.nasa.gov. Accessed 24 May 2021
Bainbridge, L.: Ironies of automation. Automatica 19, 775–779 (1983). https://doi.org/10.1016/0005-1098(83)90046-8
Benloucif, A., Nguyen, A.-T., Sentouh, C., Popieul, J.-C.: Cooperative trajectory planning for haptic shared control between driver and automation in highway driving. IEEE Trans. Industr. Electron. 66(12), 9846–9857 (2019)
Wang, W., et al.: Decision-making in driver-automation shared control: a review and perspectives. IEEE/CAA J. Automatica Sinica 7(5), 1289–1307 (2020)
Castellano, A., Karimshoushtari, M., Novara, C., Tango, F.: A supervisor agent-based on the markovian decision process framework to optimize the behavior of a highly automated system. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) Augmented Cognition. Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence), vol. 12776, pp. 351–368. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-78114-9_24
Shneiderman, B.: Human-centered artificial intelligence: reliable, safe & trustworthy. Int. J. Hum. Comput. Inter. 36(6), 495–504 (2020). https://doi.org/10.1080/10447318.2020.1741118
Madl, T., Baars, B.J., Franklin, S.: The timing of the cognitive cycle. PLoS ONE 6(4), e14803 (2011). https://doi.org/10.1371/journal.pone.0014803
Rasmussen, J.: Skills, rules, and knowledge; signals, signs, and symbols, and other distinctions in human performance models. IEEE Trans. Syst. Man Cybern. SMC-13(3), 257–266 (1983). https://doi.org/10.1109/TSMC.1983.6313160
Shneiderman, B.: Human-centered artificial intelligence: three fresh ideas. AIS Trans. Hum. Comput. Inter. 12, 109–124 (2020). https://doi.org/10.17705/1thci.00131
Poler, R., Mula, J., DĂaz-Madroñero, M.: Dynamic Programming. In: Operations Research Problems, pp. 325–374. Springer, London (2014). https://doi.org/10.1007/978-1-4471-5577-5_9
Barto A.G.: Reinforcement learning: an introduction (Adaptive Computation and Machine Learning), 3rd ed. The MIT press (1998 ) (cit. on p. 18)
Olivier, P., Tango, F.: A reinforcement learning approach to optimize the longitudinal behavior of a partial autonomous driving assistance system. In: ECAI 2012. IOS Press, pp. 987–992 (2012) (cit. on p. 20)
Melo, F.S.: Convergence of q-learning: a simple proof. in: institute of systems and robotics. Tech. Rep, pp. 1–4 (2001) (cit. on p. 20)
Dayan, P., Watkins, C.J.C.H.: Q-learning. Mach. Learn. 8(3), 279–292 (1992)Â
ERTRAC: “Connected automated driving roadmap.” https://www.ertrac.org/uploads/documentsearch/id57/ERTRAC-CAD-Roadmap-2019.pdf (2019)
Jerry, W.: By what Hubris? The readiness of the human operator to take over when the automation fails or hands over control. In: Proceedings of the DDI2018 6th International Conference on Driver Distraction and Inattention, Gothenburg, Sweden, 15–17 October 2018, pp. 182–184 (2018)
Turchetti, C.: Stochastic models of Neural Networks. IOS Press (2004) (cit. on p. 28)
Insurance Institute for Highway Safety. Self-driving vehicles could struggle to eliminate most crashes (2020).  https://www.iihs.org/news/detail/self-driving-vehicles-could-struggle-to-eliminate-most-crashes
Bonnefon, J.F., Shariff, A., Rahwan, I.: The social dilemma of autonomous vehicles. Science 352(6293), 1573–1576 (2016). https://doi.org/10.1126/science.aaf2654
Acknowledgment
This work was supported by the NewControl project, within the Electronic Components and Systems For European Leadership Joint Undertaking (ESCEL JU) in collaboration with the European Union’s Horizon2020 Framework Programme and National Authorities, under grant agreement N° 826653–2.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Maiuri, C., Karimshoushtari, M., Tango, F., Novara, C. (2023). Application of Reinforcement Learning for Intelligent Support Decision System: A Paradigm Towards Safety and Explainability. In: Degen, H., Ntoa, S. (eds) Artificial Intelligence in HCI. HCII 2023. Lecture Notes in Computer Science(), vol 14050. Springer, Cham. https://doi.org/10.1007/978-3-031-35891-3_15
Download citation
DOI: https://doi.org/10.1007/978-3-031-35891-3_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-35890-6
Online ISBN: 978-3-031-35891-3
eBook Packages: Computer ScienceComputer Science (R0)