The vaccine administration scenario detailed in Section
2 relies on the availability of a uniform access layer that sits on top of several different systems (e.g., data management, services, and IoT platforms). The Web provides the necessary infrastructure to integrate and make accessible all of these systems, effectively becoming an application architecture for the proposed MAS [
39], on top of which autonomous agents may interact and cooperate to achieve common goals. In the following subsections, we present the relevant background in multiagent systems, the Semantic Web, and the WoT, followed by a discussion of the related work in norms, policies, and preferences, with a focus on the governance of autonomous agents, both within and spanning those communities.
3.1 Multiagent Systems
A multiagent system (MAS) is composed of a (dynamic) set of
agents interacting inside a shared, possibly distributed,
environment which itself comprises a dynamic set of
artefacts. Agents are goal-oriented autonomous entities, encapsulating a logical thread of control, that pursue their tasks by communicating with other agents and by perceiving and acting upon artefacts within the environment. In essence, an MAS addresses the challenges of how agents may coordinate their efforts and cooperate in light of their autonomy [
145]. Artefacts model any kind of (non-autonomous) resource or tool that agents can use and possibly share to achieve their goals. An agent perceives the observable state of an artefact, reacts to events related to state changes, and performs actions that correspond to operations provided by the artefact’s interface. The coordinated and organised activities taking place in the system result from the concurrent and complex tasks handled by
groups of agents interacting with each other or acting within an environment. Such activities may lead to recurrent patterns of cooperation captured by agent
organisations. Changes in the state of the environment may also lead agents to react and possibly affect the state of the organisation.
Research into multiagent systems has led to a number of concrete programming models.
3 These models
4 are concerned with agent-oriented programming [
20], interaction and protocol languages [
123], environment infrastructures [
146], and agent organisation model and management systems [
58]. The results produced so far have clearly demonstrated the importance of these concepts and abstractions for the development of multiagent applications. Additionally, a variety of languages, tools, and platforms for multiagent-oriented programming (MAOP) have been developed and application success stories exists (e.g., [
49]). This type of research is often referred to under the umbrella of
Engineering Multiagent Systems (EMAS). An overview and comparative analysis of several prominent MAOPs can be found in [
91]. One of the most prominent underlying architectures used by many agent-oriented programming systems is the
Belief-Desire-Intention (BDI) architecture, which models knowledge (i.e.,
beliefs) that the agent knows about, either through observation of the environment or interaction with other agents; goals (i.e.,
desires) that the agent would like to bring about; and goals and plans of action (i.e.,
intentions) that the agent is currently focused on.
From an agent development environment perspective, the Jade platform [
12] provides a variety of behaviours (one-shot, cyclic, contract net) and is still available, although the last release dates back to 2017. Although Jade does not directly provide support for BDI-based agents, they can be added through extensions such as Jadex [
22]. Jack [
27] is an example of a closed-source BDI architecture, whereas the practical Agent Programming Language (2APL) is another open-source language that retains BDI semantics [
47]. GOAL [
75] offers a further BDI architecture which is actively maintained, whereas SPADE
5 is a recently introduced Python-based BDI platform. The JaCaMo MAOP framework, based on the JaCaMo conceptual meta-model [
17], offers first-class abstractions to program the agents’ working environment and their organisation in addition to offering the Jason interpreter for the BDI-based
AgentSpeak language [
20].
Whilst MAOP is thriving within the academic community, industrial adoption of MAOP technologies is in its infancy, and standardisation efforts such as FIPA [
63] (that superseded KQML) have received little attention in recent years [
99].
3.2 Agents and the Semantic Web
Attempts to tightly integrate autonomous agents and Web technologies date back to the vision of the
Semantic Web of the early 2000s. Berners-Lee et al. [
14] originally envisioned “a web of data that can be processed directly and indirectly by machines” in which intelligent agents act on behalf of humans by searching for and understanding relevant information published on the Web or acquired via services. Such information could potentially be made available by multiple sources, using alternative ontologies, often with different provenance. Autonomous agents rely on communication languages and protocols to exchange data and coordinate their behaviour and, thus, collaborate. Early approaches were based on speech acts [
8], focused on message types or
performatives (e.g.,
request,
inform, and
promise) based on a folk categorisation of the intended meaning of the communication. This evolved through the DARPA-funded Knowledge Sharing Effort (KSE) resulting in a communication language, the
Knowledge Query Manipulation Language (KQML), defining the mechanism by which agents communicated and an ontology language, the
Knowledge Interchange Format (KIF) describing the knowledge that the performative referred to [
62]. Although agents could perform services on behalf of their peers, discovered through capability registries [
51], service invocation occurred as a by-product of requesting information. This contrasts with the notion of
web services and
things, which use web-based communication protocols, whereby the invocation of services could be requested explicitly (in a similar manner to calling methods or functions within a programming language) by providing the relevant input parameters as data or knowledge fragments.
The prominent view from a Semantic Web perspective is that multiagent systems operate on the Web through the provision of services, using HTTP as the de facto standard transport protocol. Additionally, the Semantic Web community has developed standards, protocols, vocabularies, ontologies, and knowledge representation formalisms to facilitate the integration of machine-processible data from diverse sources at scale, using the existing web infrastructure. As such, the two communities diverged due to different priorities, though there is increasing recognition [
39] that the Web is a natural application architecture for MAS and can support different types of interactions between agents and resources.
From a knowledge representation perspective, standards such as RDFS [
24] and OWL [
69] facilitate the representation of complex knowledge about agents, services, things, and their relationship in an explicit and processable way. An example is the Provenance ontology (PROV-O), a data model for workflows expressed using agents, their actions, and other assets.
6 Additionally, reasoning engines have been developed that are capable of reasoning over OWL ontologies, albeit often with some restrictions (see Pellet,
7 HermiT,
8 FACT++,
9 Racer,
10 and RDFox
11). However, the use of ontologically grounded annotations for services within agent communication predates the Semantic Web [
57,
77] and, in some cases, the Web itself [
76]. Semantic Web service research exploited both F-Logic [
84] as used by WSMO [
119] and DAML-S [
6] (based on the DARPA Agent Markup Language) which evolved into OWL-S [
98]. Other approaches to support service utilisation were developed using OWL, e.g., the OWL ontology for protocols, OWL-P [
52], or using federated service discovery mechanisms such as the semantically annotated version of UDDI [
109]. These frameworks and ontologies were key in facilitating the discovery and use of services by autonomous agents and provided an alternative communication paradigm built on web-based infrastructure. In addition, from the knowledge perspective, bespoke protocols were developed to support the decentralised management and exchange of knowledge and information amongst networks of agents or peers [
131].
Other efforts include the provision of infrastructures for supporting the cleaning and validation of the data published on Linked Open Data Platforms, e.g., LOD Laundromat [
11]
12 and OOPS [
113].
13 Such techniques help detect errors in the data exchanged between agents and things. The SPARQL [
72] query language facilitates federated querying over distributed data sources accessible via the Web, whereas the Linked Data Platform [
130] can be used to manipulate RDF data via HTTP operations. Approaches have also been proposed to enrich SPARQL with qualitative and quantitative preferences [
70,
111] to select query results that satisfy user-defined criteria.
In recent years, the Semantic Web community has broadened its focus beyond knowledge representation, reasoning, and querying to include knowledge extraction, discovery, search, and retrieval. However, many of the proposed tools and techniques have yet to be used extensively within MAS or by the MAS community. A recent survey [
85] identified several open-research challenges and opportunities in relation to the suitability of existing proposals for autonomous agent use cases, the combination of symbolic and sub-symbolic AI techniques for enhancing agent learning, and the development of tools and techniques for validation and verification.
3.3 Agents and the Web of Things
The
Web of Things (WoT) [
90] refers to the Internet of Things (IoT) with an application of Web standards and technologies for improving interoperability of IoT devices and infrastructure. Things are resources that can be acted upon or queried via APIs (e.g., WoT scripting API [
88]);
autonomous goal-driven agents14 thus can make use of a WoT environment via WoT technologies and become part of the WoT ecosystem. Indeed, bringing agents to the Web requires more than simply exploiting Web protocols (such as HTTP [
61]) and data formats (e.g., XML [
23], RDF [
44]). The communication infrastructure used by agents should comply with an architectural style based on well-defined principles, such as
Representational State Transfer (REST) [
60] as instantiated in the Architecture of the World Wide Web [
79].
15 Furthermore, for things to be used without human intervention, they must be formally described. To this end, the W3C published the
Thing Description [
80] standard, which specifies how a JSON-LD representation of thing affordances (i.e., properties or actions) via Web APIs can be provided. In addition, the
WoT Discovery [
35] standard provides a mechanism for the automatic discovery of thing descriptions (obviating the need to hard-code the location of such descriptions beforehand). These standards support improved heterogeneity by decoupling agents from thing implementation details.
The WoT activity highlights the importance of metadata with clear semantics, and made their standards, especially thing descriptions, compatible with RDF and Semantic Web technologies. In fact, even before a standardisation effort for the WoT started, multiple initiatives suggested the use of the Semantic Web to improve IoT systems [
118]. More precisely, in REST-style hypermedia systems such as the WoT, things and agents are resources that interact by producing and consuming hypermedia about their state and the artefacts surrounding them [
38]. All resources are identified through IRIs
16 to support global referencing irrespective of contextual information. Therefore, resources can be represented through semantic descriptions that are expressed in a uniform data exchange format such as RDF using terms from some standardised and interlinked vocabulary expressed in OWL [
69]. This standardised knowledge model hides the specifics of the implementation and facilitates interconnected resources that can be queried by exposing SPARQL endpoints. Of particular interest to WoT environments are the vocabularies that describe sensors and actuators (SOSA/SSN [
71]), provenance (PROV [
92]), and temporal entities (OWL-Time [
40]).
The WoT provides a natural substrate for multiagent systems based on the vision that systems of interconnected things should be open and easily reconfigurable and, therefore, such systems should comprise autonomous and collaborative components. This notion was supported by Singh and Chopra [
125] who argue that IoT systems need the kind of decentralised intelligence that an MAS provides. Likewise, Ciortea et al. [
39] recommend integrating the Web and an MAS to leverage the proven benefits of hypermedia systems for MASs. Importantly, these articles emphasise governance as a major challenge.
The technologies that emerge from the WoT community are often industry oriented and paralleled by standardisation efforts. A recent example is the abstract
WoT architecture design document [
90], supported by the
Thing Description [
80] and the
WoT Scripting API [
88] specifications, for which a reference implementation is provided.
17 Although these technologies are more mature than MAOP technologies from an engineering perspective and have a clear path to industry adoption, they lack the rich abstractions related to agents and autonomy that MAOP technologies provide. For example, the notion of a
servient, as introduced in the WoT architecture design document, can be considered an evolutionary step from a stricter server-client separation — a notion that is considered simplistic within the MAS community. Recent approaches have sought to form a bridge between the MAOP and WoT technology ecosystems [
36,
37]. However, this line of research is young and the corresponding technologies are nascent.
3.4 Norms, Policies, and Preferences
Norms, policies, and preferences can help govern autonomous agent behaviour. The term
norm has several meanings in natural language and is used widely in economics and social science. In MASs, the term “norm” typically expresses a deontic concept (e.g., a prohibition, permission, obligation, or dispensation). A coherent set of norms, i.e., created and evaluated as a unit, is referred to as an
institution [
103]. The same understanding of norms is found in the Semantic Web literature, where there is also a body of work focusing on policy specification and enforcement. Here,
policy is an overarching term used to refer to a variety of system constraints, whereas the term
preferences is primarily used in connection with privacy and personal data protection.
The study of norms is a long-running and active line of research within the MAS community, as evidenced by numerous Dagstuhl seminars [
5,
48] and a handbook on the topic [
32]. Normative MASs [
16] are realised and characterised in multiple ways, including those based on (1) the agents’ reasoning capabilities, (2) whether norms are implicit or explicit, and (3) whether the architecture includes monitoring and enforcement mechanisms.
Agent capabilities vis-à-vis norms typically fall into three categories:
•
norm unaware, whereby agents may be regimented by external agencies to enforce norm compliance [
7];
•
norm aware, whereby agents may choose whether to comply with norms depending on the alignment of their goals with those norms, the penalties for non-compliance, and the likelihood of enforcement [
122]; and
•
value aware, whereby agents, in addition to being norm aware, are able to participate in norm creation and norm revision by reasoning about the values supported (or not) by particular norms [
41].
Thus, compliance in normative systems depends on how individual agents reason and adapt to norms at both design and runtime [
93,
136,
141]. Implicit norms that reside within the agents themselves are expressed through agent behaviour but are not otherwise externally discernible, whereas explicit or referenceable norms may have an abstract representation involving variables and a grounded (detached) representation in an entity such as a contract [
124], institution [
53,
56,
103], or organisation [
17,
139]. Agents that are norm aware or value aware should be able to
•
decide whether they want to follow them, and
•
adapt their behaviour according to the norms if they decide to do so.
Such agents may additionally be able to engage in norm revision processes. Norms, and more broadly
conventions or
social norms [
94], are established in an agent society in one of two ways: top-down and bottom-up [
101,
149].
In
top-down systems, norms are identified as part of the MAS design process and are either hard-coded into the agents’ behaviour (implicit representation), eschewing any form of normative reasoning and narrowing the scope for behavioural adaptation, or are prescriptive and explicitly represented and, thus, external to the agents, typically represented in the form of abstract regulations (e.g., ungrounded terms over variables) that, as a result of agent actions, become detached (e.g., grounded terms over literals). The n-BDI variant [
43] is a BDI-based agent architecture that allows for the internalisation of norms in which the design suggests an agent-internal process that synthesises norm-style rules based on observed behaviour. N-Jason [
93] agents perceive institutional facts, which they internalise as beliefs and, hence, incorporate in their reasoning. Norms designed offline, however thoughtfully crafted for the long term, are at risk of losing relevance in open, always-on, environments such as the Web because it is not possible to anticipate all eventualities at design time. Furthermore, drift in the agent demographic or in systems goals are likely to make norm revision essential over any sufficiently long system lifetime. With explicit norms, any norm change will affect the entire population. Such changes can be effected through a human-in-the-loop approach in which human designers revise the norms and then switch the system over at some suitable point, such as through a shutdown/reboot sequence or the use of norm-aware planning [
122]. In the latter case, an agent must manage a plan sequence that, although initially compliant, may cease to be part of the way through the plan due to the change in norms. Such an agent must also be able to check that its learned way of achieving a goal is compliant with the new norms, perhaps by means of some oracle [
107] or by being able to acquire a fresh plan that is compliant.
In
bottom-up systems, an individual agent decides whether to adopt a norm: with implicit norms, it may seek advice from others or apply indirect reinforcement learning over its observations as a basis for prediction, possibly in combination with a
strategy update function [
149]. In such systems, norms are deemed to have emerged once they have been adopted by a sufficiently large fraction of the population. This is typically 90% in most of the literature and 100% in some cases (which is hard to achieve), or assumes a simple majority, which can risk oscillatory outcomes. However,
convergence (this term appears to be used interchangeably with emergence in the literature [
102]) is a function of the
capabilities of the agents. Emergence with explicit norms depends on agent reasoning capabilities. An agent might inform the regulator that it wants to take a particular action in a particular state (without sanction)—the agent knows what it wants but not how to get it—as a request to change the norms without having to reason about norm representation. A more difficult approach is that an agent might propose a new (abstract) norm—the agent knows how to define a new norm to get what it wants [
73,
102]. As above, changes have to be actioned, which could be as outlined previously, although pluralist approaches are possible, as put forward by Ostrom [
106] or by using one of the many voting mechanisms. The challenge for an agent then becomes how to decide which way to vote, which depends on their reasoning capabilities: are they able to evaluate the consequences of the norm change and are they selfish (i.e., vote “yes” if the change is individually beneficial, e.g., increases their utility) or altruistic (i.e., vote “yes” if the change is collectively beneficial)? More sophisticated still would be the use of argumentation to determine whether the revision is consistent with the population’s values [
122,
134].
In the early days, Semantic Web researchers proposed general policy languages, such as KAoS [
21], Rei [
82], and Protune [
19], which cater to a variety of different constraints (access control, privacy preferences, regulatory requirements, etc.). A prominent early attempt to provide a semantic model of policies as soft constraints for agents was OWL Polar [
121], an OWL DL explicit policy representation language. OWL Polar aims to fulfil the essential requirements of policy representation, reasoning, and analysis, where policies are system-level principles of ideal activity that are binding upon the components of that system and, thus, are used to regulate the behaviour of agents [
121]. Over the years, the Semantic Web community has also proposed policy languages that are tailored to better cater to access control, privacy preferences, licensing, and regulatory governance requirements, including detailed surveys, for example, of the various policy languages and the different access control enforcement strategies for RDF [
87]. From a privacy perspective, the Platform for Privacy Preferences Project (P3P) [
42] specification, deemed obsolete in 2018, aimed to allow websites to express their privacy preferences in a machine-readable format that could be interpreted by agents that could automate decision making on behalf of humans. The P3P initiative, despite having failed, inspired subsequent work on representing and reasoning over privacy preferences, such as using OWL [
65], catering to more expressive privacy preferences [
89], and representing consent for personal data processing [
18].
Many existing proposals rely on WebID [
120], a community-driven specification that offers an identification mechanism making use of Semantic Web technologies to provide password-less authentication. An extension of WebID (specifically WebID-OIDC that relies on OpenID Connect
18) is used in the
Solid project. Solid
19 is an ongoing initiative, lead by Tim Berners-Lee, aimed at deploying a distributed Linked Data infrastructure for governing one’s personal data, which is built on top of Linked Data Platforms. Additionally, there has been work on
usage control in the form of licensing [
28,
66,
67,
68,
143] and, more recently, policy languages have been used as a means to represent regulatory constraints [
50,
108]. The Open Digital Rights Language [
64,
78], although primarily designed for licensing, has been extended to cater to access policies [
133]; requests, data offers and agreements [
132]; and regulatory policies [
50]. Usage control, however, often proves challenging for organisations and users, and any constraints imposed on the use of data need to ensure that policies are applied consistently across organisations and that there are robust propagation mechanisms preventing policies from becoming invalid [
45,
46]. The notion of FAIR ICT Agents [
86] is based on FAIR (Findable, Accessible, Interoperable and Reusable) principles [
147], where ICT denotes
interactive intelligent agents that are constrained via goals, preferences, norms and usage restrictions. Thus far, the WoT standards offer only limited support for norms, policies and preferences, which are currently described in guidelines targeted at human developers rather than as declarative, machine-readable statements usable by agents [
117].
Although research on norm-aware agents has made reasonable progress to date, much remains to be done to elevate human oversight to align with the three categories [
74]:
human-in-the-loop, where there may be human intervention in each decision cycle;
human-on-the-loop, where there is human intervention in the design cycle and operation monitoring; and
human-in-command, where there is human oversight of the overall system, including the means to decide when and how to engage the AI system. The motivated scenario presented herein draws on human-on-the-loop and human-in-command. Indeed, it is these levels of abstraction that inspire the governance framework introduced in Section
4, since those are the system characteristics we aim to facilitate.