1. Introduction
Industry 4.0 was characterized by the increase in automation of processes and the adoption of artificial intelligence (AI) in a variety of industries, such as manufacturing, the process industry, healthcare, and the transport industry. AI-driven automation has certainly demonstrated its value by increasing productivity, reducing cost, and adding a new layer of safety; however, the rapid and widespread adoption of these intelligent systems without proper consideration of the ethical, societal, and safety implications has quickly brought to light some drawbacks.
Smart automated systems contribute to the overall system performance and safety by automating routine tasks, identifying and anticipating hazards and failure events, and supporting operator awareness and decision-making in high-complexity troubleshooting situations; even so, the human still has the important task of monitoring possible automation failures, carrying out manual procedures if needed, and performing decision-making. Failure in the proper integration between them has previously led to disastrous consequences (such as the accident of Air France flight 447 in 2009) due to operator loss of expertise, reduced vigilance and situational awareness, complacency, reduced adaptability, or information overload [
1]. Particularly for safety-critical systems, such as those in transport, nuclear, process, and infrastructure industries, it is expressly dangerous to over-rely and overestimate the capabilities of intelligent systems. Human experts should not be replaced but integrated into the loop so that humans and systems can collaborate dynamically to mitigate each other’s limitations and enhance each other’s strengths. Therefore, the current trend in Industry 5.0 is collaboration and cooperation between humans and intelligent systems to achieve optimal performance. This approach is not novel but, as Sendhoff and Wersing (2020) have noted in their work ’Cooperative Intelligence—A Humane Perspective’ [
2], a ’human-centric’ view of AI-based systems that has recently been taken to the forefront.
Despite the potential benefits of human–machine collaboration for high-stakes industries, there are specific challenges to its adoption, such as the complexity of integrating human and machine/AI roles and feedback [
3], safety issues related to low model transparency and explainability, vulnerability to adversarial attacks and human bias [
4], and system safety and robustness certification [
5]. Following the publication of the Ethics Guidelines for Trustworthy AI presented in 2020 by the High-Level Expert Group on Artificial Intelligence [
6], the new essential health and safety requirements for machinery with AI (Machinery Regulation (EU) 2023/1230) and the EU AI Act (Regulation (EU) 2024/1689), major concerns were raised regarding the risks of harm to the health, safety, or fundamental rights of people in their interaction with autonomous systems/machines, particularly those with self-evolving behavior, with close interaction with humans, or those used as a safety component or product. Innovation and the use of state-of-the-art CI technology may be hampered by these regulatory constraints, low digitization levels, and lack of AI expertise, specifically for small and medium enterprises (SMEs) in industry [
7].
Collaborative intelligence (CI) is a complex and interdisciplinary field, leveraging advances in AI and machine learning (ML) to achieve higher levels of synergy between humans and machines. As such, we aim to outline the large diversity of AI methods and collaboration paradigms explored in the literature. We intend to assist engineers and practitioners in the related fields of robotics, human–computer interaction (HCI), ergonomics, AI, safety engineering, and behavioral sciences, to narrow down CI solutions for domain-specific problems and reduce the barrier to adoption, by offering a broad overview of potential applications of collaborative intelligence, with safety in mind.
The systematic categorization proposed next is based on the main drive for the collaboration, either assistance to the human or to the machine, enabling the capture of a diversity of solutions and the discovery of how different AI methods can be employed for the same problem. Collaborative intelligence, also known as cooperative intelligence [
2], or Hybrid Intelligent Systems [
8], can be defined in different ways depending on the goals and modes of collaboration/interaction. Akata et al. (2020) [
9] has defined hybrid intelligence as a combination of human and machine intelligence that enhances and leverages human capabilities, and achieves goals unattainable by the human or machine alone, through human–machine teaming. While in [
2], a wider definition was presented, considering cooperative intelligence as a state of the system required to establish a relationship between multiple agents (human or not) and that beyond working together to achieve a common goal, it could be leveraged just to benefit from each other and live together synergistically. Here, we use a general interaction-based definition, where the human and the machine/algorithm intelligence are interdependent and interact to achieve a goal (contrary to being used independently for the same goal), identifying two main types of purposeful interactions in human–machine collaboration: the machine assists the human, or the human assists the machine.
In the human–machine collaboration scenario, in which the main goal is for the machine to assist the human-in-the-loop (HIL), the objective of the intelligent system is to monitor and support the human operator in their tasks:
By monitoring the human and the task context and using the knowledge to detect critical conditions, support, or adapt to the operator’s cognitive status for optimal system performance;
By amplifying the operator’s physical capabilities, as in the case of exoskeletons that can be worn by the user and enhance their physical performance, or telerobots that allow the human to perform a variety of complex tasks using the capabilities of robots;
By embodying human physical capabilities as in the case of cobots, that can work alongside humans performing complementary tasks or substituting the operator in a dynamic way.
In the type of human–machine collaboration in which the main goal is for the human-in-the-loop to assist the machine during a learning process, a test process, or a deployment process, the human can provide assistance:
By annotating data to be employed by AI algorithms. This step is crucial for the performance of supervised machine learning (ML); however, due to the effort and cost of manual labeling of large amounts of data, more efficient learning strategies have been developed, such as those using active learning that tries to maximize the model’s performance while querying a human to annotate a data sample as few times as possible.
By the direct demonstration of tasks, as in the robot learning from a demonstration paradigm or by direct intervention on the automated process.
By using expert knowledge to validate and explain intelligent machine behavior that, despite the emergence of explainable AI techniques, is still required to ensure the outcomes and that the generated explanations match the expected behavior in a reliable and unbiased way.
2. Related Works
Safety in industry is regulated by legal requirements and standards, in particular for the development and deployment of systems that interact with humans. For instance, in human–robotic collaboration applications, the new essential health and safety requirements under the EU Machinery Regulation 2023/1230 apply and have to be considered during the design stage and throughout the lifecycle of the robotic system. A commonly applied standard is the harmonized standard EN ISO 12100:2010 [
10], regarding risk assessment and risk reduction principles, described by three steps: (1) the implementation of inherent safe design measures that aim to eliminate or reduce the risk of hazards, (2) the implementation of safeguarding and/or complementary protective measures when a hazard cannot be eliminated or its risk is reduced at the design stage, and (3) if risks remain, the disclosure of information for use that shall recommend safety procedures to be implemented by the user. A more specific harmonized standard for the context human–robot collaboration is the EN ISO 10218-2:2011 [
11], including safety requirements and protective measures.
For the safety of railway systems, specifically for control software with impact on safety, the recent EN 50716:2023 [
12] provides lifecycle development processes and technical requirements, including formal methods and the integration of ML/AI techniques.
In the case of highly automated vehicles, the complexity of the environment and intended functionality requires safety guidance on multiple system levels. General functional safety can be addressed by the requirements of EN ISO 26262:2018 [
13], while SOTIF EN ISO 21448:2022 [
14] focuses on design, verification and validation guidance for the specification of the intended functionality, which can apply to hazards caused by the limitations of the implemented ML/AI in advanced driver-assistance systems. In the case of fully autonomous vehicles, the standard UL 4600 (2023) [
15] proposes an approach for assessing and validating system safety, while for systems requiring human interaction and supervision, other standards specific to safe and ergonomic design should be applied (such as EN ISO 9241-210:2019 [
16]).
However, with the fast development and introduction to the market of CI technology, other current legislation and standards may require updating to deal with the emergent challenges of new digital technologies, such as the artificial intelligence, Internet of Things and robotics domains. To support the operationalization of compliance to new safety requirements, in addition to leveraging existing and emerging standards, state-of-the-art research efforts can shed light on domain-specific solutions.
Previous surveys and literature reviews have approached some of these topics in a more focused manner, mostly covering two main general lines of research: intelligent human–machine interactions, namely human–robot and human–computer interactions, and human-in-the-loop learning.
Hua et al. (2021) [
17] surveyed state-of-the-art deep learning, reinforcement, imitation, and transfer learning AI methods for robot control and adaptation to diverse complex environments and tasks, including their application to human–robot collaboration. In the recent work of Borboni et al. (2023) [
18], the potential role of AI in the use of cobots for industrial applications was explored, and the state-of-the-art research on AI-based collaborative robotic applications was analyzed. Liu and Wang (2018) [
19] also covered human–robot collaboration in their review, specifically, gesture recognition used for communication between human workers and robots. The work reports on the most important technologies and algorithms of gesture recognition existing in the current research. A model and classification scheme of gesture recognition for human–robot collaboration is proposed, with four technical components: sensor technologies, gesture identification, gesture tracking, and gesture classification.
Another possible effective communication channel between robots, machines, and humans is the human gaze. Zhang et al. (2020) [
20] performed a literature review of human gaze modeling and its potential applications. Human gaze data can be used by AI to develop the intelligent attentional selection of information, or for the AI agents to be aware of the human cognitive and emotional state, fostering more natural communication and interactions between them. The work reviewed human-gaze-assisted AI agents in multiple fields, such as computer vision, natural language processing, imitation, and reinforcement learning, as well as robotics. In a more high-level and broad mapping study by Šumak et al. (2022) [
21], state-of-the-art AI methods for intelligent human–computer interaction using sensor data were reviewed. The mapping found that studies have mostly focused on recognizing the emotion and stress of HCI users, followed by gestures and facial expressions identification, using, more frequently, deep learning algorithms, including CNNs, and from the classical machine learning algorithms, SVMs.
An alternative type of collaboration is human-in-the-loop learning, typically employing the human to deal with sparse data, lack of training data, or improve the performance of machine learning methods with expert task knowledge. Wu et al. (2022) [
22] analyzed the literature from a data perspective, categorizing methods based on the stage at which the human was added to the loop—in the data processing stage, model training stage, or in the design and application of the system. From the point of view of who is in control during learning, human-in-the-loop machine learning methods can be mainly divided into active learning, where the system is responsible for the learning process and interacts with the human for data annotation; interactive machine learning, where the interaction is less structured, more frequent, and incremental; and machine teaching, where the human expert is the responsible for the learning process by transferring knowledge [
23]. From a safety perspective, human–machine collaboration has been leveraged for safe learning and anomaly detection by incorporating human knowledge, demonstration, supervision, or feedback in the learning process [
24]. Due to the black-box nature of AI/DL-based systems and their weakness to adversarial attacks and out-of-distribution inputs [
25], the unique skills of humans such as pattern discrimination, high-level conceptualization, and hazard identification can be used by the system to mitigate safety risks.
To our knowledge, the current literature explorations of collaborative intelligence and AI problems, techniques, and challenges have not taken into account the common aspects between the previously mentioned disparate topics of work and how collaborative interactions are at the core of the targeted technologies. With this survey, we intend to give a general and unified overview of recent works employing AI methods for collaborative intelligence problems, with an additional focus on safety or application to safety-critical industries. For these industries, safety is put on the forefront of performance; however, both the probabilistic nature of AI and the uncertainty of human behavior make the regulatory certification of CI methods a challenge [
24]. The work of [
5] provided a general overview of potential safety issues in human–machine collaboration from the machine side, the human side, and the interaction side, and possible countermeasures to be considered early on in the development of human–machine teaming, lacking, however, a direct link to specific tasks and application domains. Safety is an emergent property of the combination of components of a system [
24], and as such, the risk assessment process should involve the setting of the operating conditions and identification of potential hazards for the individual components according to their role in the system, highly dependent on the application domain. Our aim is to promote the progress and implementation of collaborative intelligence in safety-critical industries by providing a wide array of application examples, their limitations, and insight into outstanding safety concerns. We analyze the trends and gaps in the AI problems addressed, the interaction tasks, and techniques used in the latest collaborative intelligence research landscape to set the directions for future research on safer human–machine collaboration. Furthermore, we uncover from the reviewed papers a sub-categorization of collaborative intelligence tasks (
Figure 1) and link it to the previously identified types of CI interactions (Machine assists human-in-the-loop or human-in-the-loop assists the machine).
As a wide survey of AI methods for CI solutions, we cover a broad range of techniques that aim to emulate human intelligence through computer algorithms [
26]. AI techniques generally fit into three main groups: machine reasoning, machine learning, and robotics [
27]. These three groups are, however, a very coarse classification, blending categorizations of learning style, similarity in form or function, and the AI problem it tackles. For this review, a more comprehensive taxonomy will be used expanding on the AI Knowledge Map developed by Corea (2019) [
28]. We categorize AI solutions based on the intelligence problem that is solved (AI problem domain) and the type of AI approach used (AI paradigms).
Below, we summarize the standard domain categories that AI problems typically fall into:
Perception domain: Tasks of sensing and understanding signals from the physical world, transforming those data into valuable and relevant information for the specific application.
Reasoning domain: Tasks that use logic and existing knowledge to solve new problems.
Knowledge representation domain: Tasks that try to represent knowledge in a way that can be manipulated and understood.
Planning domain: Tasks that try to find the best way to achieve goals within certain spatial, temporal, or resource constraints.
Communication domain: Tasks regarding the understanding of language and communication.
Control domain: Tasks regarding the monitoring and regulation of a system to achieve a desired result.
Next, we specify the categories of AI paradigms that will be employed:
Logic-based approaches: Approaches that use logic and abstract high-level representations of knowledge to solve problems. Logic-based approaches are the foundation of classical symbolic AI and are used for three main problems: knowledge representation, reasoning, and model-checking and verification [
29].
Knowledge-based approaches: Approaches that use knowledge representation systems, such as large knowledge bases, to represent explicitly and declaratively known notions, information, and rules in order to infer new implicit symbolic knowledge [
30].
Probabilistic approaches: Approaches that use probabilistic knowledge and scenarios, such as Naïve Bayesian Networks, Markov Models, and Restricted Boltzmann Machines [
31].
Machine learning approaches: Approaches that allow a machine/computer to learn from data, i.e., data-driven methods, such as Artificial Neural Networks and Ensemble Learning algorithms [
32,
33]. These approaches can be sub-classified based in the degree of supervision they have during learning (unsupervised, supervised, or semi-supervised learning) or the prior knowledge (inductive, transductive or transfer learning).
Neuro-symbolic approaches: Approaches that integrate symbolic and sub-symbolic models at any stage of an intelligent system (model design, input, output, reasoning, or learning stage) [
34].
Search and Optimization approaches: Approaches that allow to intelligently search through many solutions to achieve the target objective function value while satisfying the problem’s constraints [
35]. In general, optimization methods can be divided into either local or global algorithms.
Embodied intelligence approaches: Approaches that allow for an agent to have higher intelligence, such as movement, perception, interaction, and visualization abilities. Approaches like reinforcement learning and learning (programming) from demonstration include not only the intelligent agent but also its “body” that interacts with the world according to some constraints, and the specific environment it is situated in. A broader definition can be taken to include simulation, where the embodied intelligence agent can purposefully exchange energy and information with a simulated physical environment [
36].
Section 3 contains the reviewed works organized according to the general type of CI interaction, the sub-categorization of CI tasks found during the analysis (displayed in
Figure 1), and the AI taxonomy mentioned above. In addition, a summary of the papers is given for each CI task sub-category. Lastly, a discussion of the most important insights discovered through the review and analysis of the research questions is presented in
Section 4, also providing recommendations and future research opportunities for the field of collaborative intelligence.
3. Results
To conduct the review, the following databases were searched and used as the principal research systems: ACM digital library, ISI Web of Knowledge, Wiley Inter Science, Scopus, and IEEE Xplore. Works published between 2006 and 2023, in conferences of class A or journals of the first or second quartile, related to the main topics of the review were identified: collaborative intelligence, artificial intelligence, and safety. Articles not related to industrial applications or safety-critical industries were not included.
After the definition of the classification strategy, the papers were categorized according to the general CI task sub-category, the AI paradigm employed, the AI problem domain addressed, the collaborative intelligence goal achieved, and the application domain (if mentioned). In addition, the research goals, research methodology, and results were analyzed for each category. To guide the discovery of trends and gaps in the field, we studied the type of CI tasks addressed and the common AI methods and techniques employed for collaborative intelligence problems, the industries most targeted by the research, how the safety of human–machine interactions is addressed, and identified common future lines of research/work.
The final 91 articles retrieved from the literature are presented and described by CI interaction type and in chronological order in
Table 1,
Table 2,
Table 3,
Table 4,
Table 5,
Table 6,
Table 7,
Table 8,
Table 9,
Table 10 and
Table 11. From an initial analysis, it was observed that most of the articles retrieved that fit our focused research topics were published recently, between 2019 and 2023 (
Figure 2), while no article published between 2006 and 2014 was found that matched our search queries. The most addressed collaborative interactions involved the machine assisting the human with human–robot collaboration tasks in the manufacturing industry domain. The second largest portion of the papers provided AI solutions for CI tasks independently of the application domain, not having specified it in the paper, or mentioning only potential target applications. However, the survey was able to retrieve research works focused on safety-critical industries, such as automotive, maritime, aviation, nuclear, and rail industries (
Figure 3).
The distribution of the papers following the defined AI problem and the paradigm taxonomy is depicted in
Figure 4 and
Figure 5, respectfully. A tendency is clear for the type of AI problems that are approached in collaborative intelligence research, with most of the articles including perception, planning, and/or control problems. In terms of AI paradigms, machine learning is the most popular technique in the retrieved papers, followed by probabilistic approaches. It is worthy of note that a large portion of work has employed either hybrid techniques or made use of more than one type of AI method to solve multiple AI problems, often necessary to achieve complex CI solutions.
In the selected research works, a variety of common collaborative intelligence tasks were found (
Figure 6). The identified subtypes of CI tasks in which the machine assists the human were
Cobot motion planning and task scheduling,
Robot control,
Design optimization,
Human/object detection and tracking,
Human intention prediction and motion recognition,
Human state recognition,
Human mental model estimation,
CI safety assessment, and
Accident/error prediction; the tasks related to humans assisting the machine were
Human assistance during learning, including the safe training of RL agents, interactive alarm flood reduction, personalized autonomous driving, language-based assistance for navigation, cobot motion planning and coordination learning, and robot control learning, and
Human assistance during deployment, addressing worker-assisted emergency detection, worker-assisted active learning for real-time safety assessment, and worker-assisted robot visual grounding.
4. Discussion
This survey analyzed recent research work based on the type of CI interaction studied, the CI task performed, and the type of AI techniques used to address the task. It was clear that due to the extensiveness of the CI domain, only a sample of the available relevant literature was analyzed. Simultaneously, we have to consider that “collaborative intelligence” is still an emerging concept and that it has yet to be broadened to other domains. Next, we present the key takeaways from our review, insights into future research trends and the developed safety recommendations for practitioners in the CI field.
As a big portion of the reviewed methods were developed for the manufacturing context or safety-critical applications, such as aviation and the automotive industry, the majority of the proposed solutions consider a single component of a whole complex system. The evaluation and validation stage of the methods was varied, ranging from testing in simulated environments and tasks, testing with standard test sets, lab-based experimental testing using representative use-cases of the target applications, and assessment in multiple real-world scenarios and environments (performed by a minority of the works). Even still, the quality and representativeness of the training and testing data were rarely assessed, and when mentioned, it was to highlight the need for more diverse datasets. Several papers made use of high-fidelity digital twins to accelerate testing time and improve the safety of testing procedures, particularly for solutions that involve humans.
Establishing benchmarks for such complex and fast-developing AI technology is challenging but particularly for safety-critical and human–machine collaborative systems, at the minimum, the already existing industrial standards for system safety and methods specific for model safety, robustness, and dependability should be used at the different stages of model development. The recent review of Mohseni et al. (2022) [
142] can be used to aid the selection of specific ML safety techniques, which can be applied at the design stage for inherently safe design specification, the development stage for increased performance and robustness, and/or the deployment stage for run-time error detection. In addition, the integration of different system modules should account for how errors propagate downstream from the initial perception component to the final decision component, requiring the use of AI methods and metrics to quantify model uncertainty and transmit it through the system.
There is still a clear need for general guidelines, with comprehensive paradigm-agnostic metrics and specific procedures, that can be followed by CI researchers and engineers to ensure the reliability and dependability of the whole system before deployment in open-world settings. As a multitude of AI paradigms and models can be applied to solve one specific problem, and as it is typically difficult to determine what type of approach is more suitable to which problem, we recommend that these guidelines follow and build upon the CI task categorization here proposed that groups multiple similar tasks and high-level collaborative goals, in order to develop procedures that are independent of the model chosen and instead conditional on the type of data used and expected output:
- -
The proposed solutions for cobot motion planning and task scheduling are typically limited by the type of workspace/environment that the methods are applicable to, the number of workers accounted for by the solution, the types of tasks that they can be applied to, and by the uncertainties of human behavior. Dynamic and flexible approaches, such as the one by Zhu et al. (2019) [
40], can better deal with variations in the expected operations but are usually harder to implement and test for compliance with safety requirements. Future research for robot motion planning and scheduling for collaborative human–robot interactions should be focused on dynamic/adaptive planning methods, online learning for the continuous improvement of the models, personalized assistance to the human collaborator, and more natural communication channels between human and robots. A particularly interesting and novel avenue to deal with the uncertainty of human behavior, and increase the comfort and safety of human–robot interactions is to estimate and match the human’s expected behavior of the robot. The exploration of this problem will be key to develop systems that can resolve conflicts between humans and machines, reach a common ground, and adjust tasks and goals accordingly [
9].
- -
The use of human demonstrations to aid the machine or AI learning process during training has a wide range of applications in robotics and automation, including for manipulation tasks in manufacturing domains, and for navigation tasks of mobile robots and vehicles [
143]. Learning from demonstrations is a promising and versatile approach for robots to learn complex tasks by learning to imitate a human expert. There are, however, several limitations of the current techniques, such as the assumption that all demonstrations are optimal and the reliance on human data, which have pushed the research into investigating and developing solutions that (a) quantify the human factors in LfD and try to increase the quality of demonstrations [
129]; (b) optimize the learning process while minimizing the human time cost of requesting demonstrations [
135]; and (c) learn to generalize to novel scenarios, estimate the suitability to the new scenario, and determine when it needs human assistance/intervention [
131,
143]. As some of the reviewed works proposed, human assistance/intervention can be applied with less effort by training a model to perform the human’s intervention job (can be either to reward or block agent actions), or by estimating, during real-time deployment, the confidence/doubt of the learned policy model relative to the execution risk, further improving safety [
24]. Future applications should employ methods for human feedback quality assessment, human error detection, or outlier/out-of-distribution human sample detection, such as reducing data labeling ambiguity [
144], improving demonstration quality scoring [
145], or improving the human’s response time in interventions that can lead to undesirable effects.
- -
The main limitation of the current human/object detection and tracking techniques is still robustness. As such, we suggest the use of (a) transfer learning, (b) ensemble classifiers, or (c) multiple sensor modalities to reduce occlusions and overlap effects, or (d) to take into consideration target variants/target changes of appearance as possible strategies to increase the robustness of detection systems. Future research should also focus on developing novel methods to deal with unaccounted variability in the data after deployment, such as continual learning.
- -
The proposed solutions for human intention prediction in driving applications were limited by the fact that the developed improvements or sub-systems need to be integrated with an autonomous driving system/ADAS, and tested for robustness with real vehicles and environments. On the other hand, most of the motion recognition works proposed for human–robot collaboration applications were limited by the type of task or case study data used to train and test the models, as most focused only on single-person detection and upper limb movement data. Compared to the works that used image/video data, the papers that employed physiological data for motion intention recognition had comparatively small training datasets with a limited number of participants, even though they allowed to detect the intention up to 513 ms–2 s before the execution of the motion. One possible future work line for human motion intention prediction is the adaptation and personalization of the models to specific subjects when it becomes possible to have access to large amounts of big data that physiological sensors can provide. Fusing multiple modalities can also help to better distinguish between assistive or disruptive operator intentions, and intentional or incidental contact between humans and robots as in the work of Amin et al. (2020) [
81], by combining vision and tactile perception data. With these limitations in mind, we recommend, as a future work line, going beyond the prediction of the driver’s intentions for real-time assistance, and developing human-inspired safety metrics that can be retrieved from the trained models and serve as a prior for an autonomous driving policy.
- -
Ground truth labels for human state recognition can be obtained through subjective questionnaires or objective indicators. When these methods are not available during the online stage of the model implementation, behavioral indicators of the state of interest can be used instead (as in [
96]) for continuous learning or model performance assessment. In addition, we recommend that systems that are already using wearable sensors to monitor human data for other purposes, such as using human posture and muscle activity for hand-over intention recognition [
76], or human skeletal data for determining the optimal tool delivery position [
48] could add a worker physical state recognition module for real-time robot behavior adaptation or to improve the ergonomic design of collaborative systems. Safety applications of human state recognition were addressed by the set of reviewed papers, demonstrating how these techniques can be used to prevent critical situations when the operator/user is not performing at the required level, and can cause harm to himself or others.
The challenges of building Hybrid Intelligent Systems include collaboration between humans and machines, adaptation to humans and the environment, systems that behave responsibly, and systems that can explain and share knowledge and strategies with each other [
9]. Research has mainly focused on collaboration and adaptation, but more work is needed to develop explainable systems that take into account the ethical, legal, and societal responsibilities, as well as the consequences of these systems to better manage conflicts between intelligent systems and human experts.
Some of the reviewed CI solutions have shown signs of moving in this direction, such as the work of Oh et al. (2021) [
55] and Li et al. (2022) [
57], that proposed shared-control methods for the blending of the intelligent system’s control commands and human commands, with the possibility to assign control authority to a human in high-risk scenarios or in cases of considerable disagreements between the two policies. This type of control solution, that is a middle-ground between direct and autonomous control, may be a suitable option for new teleoperation and autonomous driving systems in Industry 5.0, to comply with human agency and oversight requirements recently entered into force by the EU AI Act (Regulation (EU) 2024/1689) for high-risk AI systems. Still, other current machinery safety requirements, such as those of the Machinery Directive update (Regulation (EU) 2023/1230), require that any machinery, or related product with self-evolving behavior or logic that operates with varying levels of autonomy should communicate its planned actions (such as what it is going to do and why) to operators in a comprehensible manner. This topic is rarely addressed in shared-control research, and methods from the explainable AI domain [
146] and human–machine interaction domain [
58] should be integrated and adopted. As an example, the presented work of Kottinger et al. (2022) [
114] provided an explainability method for robot trajectory planning oversight, by generating human-understandable path visualizations.
Other CI solutions for conflict resolution may include, for when the intelligent system detects human decision-making bias or model disparities between the human and system, generating explanations that persuade the worker to take the most beneficial action or to repair their policy [
105]. Alternatively, the estimate of the system’s confidence/uncertainty on their outputs can be communicated to the human stakeholders to support informed decision-making [
109,
110].
An aspect of collaborative intelligence that was missed in the reviewed works is the fact that in many cyber–physical systems, the intelligent machine might have to interact with many humans in different contexts, and the humans might interact with different system instantiations, sharing the same knowledge base [
2]. Computational techniques such as federated learning (also referred as collaborative learning) need to come into play to achieve this multi-human multi-agent collaboration scenario.
We provide a summary of recommended actions (
Figure 7) taken from the analysis of the literature on the safe application of AI to collaborative intelligence problems.