1,302
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Engineering Responsible And Explainable Models In Human-Agent Collectives

ORCID Icon & ORCID Icon
Article: 2282834 | Received 20 Nov 2022, Accepted 28 Oct 2023, Published online: 05 Dec 2023

ABSTRACT

In human-agent collectives, humans and agents need to work collaboratively and agree on collective decisions. However, ensuring that agents responsibly make decisions is a complex task, especially when encountering dilemmas where the choices available to agents are not unambiguously preferred over another. Therefore, methodologies that allow the certification of such systems are urgently needed. In this paper, we propose a novel engineering methodology based on formal model checking as a step toward providing evidence for the certification of responsible and explainable decision making within human-agent collectives. Our approach, which is based on the MCMAS model checker, verifies the decision-making behavior against the logical formulae specified to guarantee safety and controllability, and address ethical concerns. We propose the use of counterexample traces and simulation results to provide a judgment and an explanation to the AI engineer as to the reasons actions may be refused or allowed. To demonstrate the practical feasibility of our approach, we evaluate it using the real-world problem of human-UAV (unmanned aerial vehicle) teaming in dynamic and uncertain environments.

Introduction

Autonomous systems are becoming increasingly ubiquitous and are making significant impacts in various safety-critical systems, such as exploration robots (Molnar and Veres Citation2009), driverless cars (Abeywickrama et al. Citation2020; Dogan et al. Citation2016) and unmanned aerial and ground vehicles (Ramchurn et al. Citation2016; Ramchurn, Stein, and Jennings Citation2021). Although the evolution of robotics has many compelling benefits, as with other emerging technologies, it comes with new risks and challenges. These challenges can be categorized into three broad and interrelated areas of ethical and social concern: safety and errors; law and ethics; and social impact (Lin, Abney, and Bekey Citation2014). Therefore, it is essential to ensure that agents make correct and responsible decisions, so that humans or other living beings, infrastructure and society as a whole are not harmed.

Researchers have advocated three main principles that can be used by AI systems to support the social good: accountability, responsibility and transparency (ART) (Dignum Citation2017). The ART principles form the main pillars of responsible AI. Yazdanpanah et al. (Citation2021) highlight the need for an interdisciplinary effort to address different social, technological, legal and ethical challenges. This requires the development of socio-technically expressive notions of responsibility, blameworthiness and accountability. Dignum et al. (Citation2018) define accountability as the need to explain and justify one’s decisions and actions. Responsibility refers both to the capability of AI systems and to the role of people interacting with them (Dignum et al. Citation2018). Transparency is defined as “the extent to which the system discloses the processes or parameters that relate to its functioning” (Winfield et al. Citation2021). Approaches to the design of responsible AI broadly focus on regulation by means of legislation and standards or are design based (e.g. ethics by design). Our research concerns the latter and focusses on the engineering of responsibility in the design of human-agent collectives using formal modeling and verification. More specifically, we engineer decision-making behavior to ensure that a human-agent collective is safe (i.e. nothing bad happens), controllable (meaningful control between humans and agents for collective decisions), trustworthy (what will convince us to trust agents) and ethical (what moral decisions agents may have to take) (see Appendix A for a list of definitions used in this work).

Acting in a responsible manner is important for agents, as described earlier, but so is being able to explain why one’s actions are morally right or wrong (Conitzer et al. Citation2017). As autonomous systems are becoming part of our daily life, issues related to ways that humans interact with such systems become more significant. Among these issues are making machine decisions transparent, understandable and explainable (Goebel et al. Citation2018; Koeman et al. Citation2020). According to Wortham and Theodorou (Citation2017) and Sheh (Citation2017), the ability of a robot or an autonomous system to explain its behavior helps the user to develop an accurate mental model of the robot’s reasoning. This can result in better interaction between them (Koeman et al. Citation2020). These explanations can be in more understandable natural language, for example, answering why-questions for end users of robotic systems (Koeman et al. Citation2020). Miller (Citation2019) defines explainability or interpretability as “generating decisions in which one of the criteria taken into account during the computation is how well a human could understand the decisions in the given context.” The relationship between the notions of explainable AI and responsible AI has been described by Barredo Arrieta et al. (Citation2020). Responsible AI is a broader concept that imposes the systematic adoption of several AI principles (e.g. explainability, fairness, accountability and privacy), so AI models can be of practical use.

Now, ensuring that autonomous systems work in a responsible and explainable manner is challenging. This is especially the case when dilemmas are encountered. In this paper, we define dilemmas as decision-making situations in which the choices available to agents are not unambiguously preferred over one another (Bjørgen et al. Citation2018). When humans are involved, new challenges emerge as humans may not always be able to understand what agents are intending to do and vice versa. Also, humans and agents often need to agree on collective decisions using consensus or compromise (Greene et al. Citation2016; Loreggia et al. Citation2018). In order to support collective decision making, agents should be able to account for some form of moral values and ethical principles (Awad et al. Citation2018), as well as constraints on safety and controllability. According to Rossi (Citation2015), the notions of safety constraints and ethical principles are closely intertwined. An action by an agent can cause harm, thus it can be considered unsafe. However, if the action occurred by accident and there was no intention to cause harm, it may not be considered unethical (Rossi Citation2015). The use of ethical principles allows agents to reason and determine their actions and explain and justify their behavior in terms that are understandable by humans (Rossi Citation2015). Furthermore, humans would accept and trust agents more, if they behaved as responsibly as humans in the same environment (Rossi Citation2015).

An example of a situation in which a dilemma occurs that raises ethical concerns can be described as follows. Assume that a UAV is flying above a football stadium with a match in progress and will run out of battery in a few seconds. The UAV has two options: to crash into the football stadium, thus colliding and harming spectators and players on the ground; or to ascend above 500 feet and collide with a small civilian aircraft carrying passengers. Either of these two actions can harm humans, and the agent will be in a dilemma about the correct, morally significant decision to make.

So, how can we certify that the decision making within human-agent collectives is responsible and explainable? To answer this question, methods and tools that allow the certification (Fisher et al. Citation2021; Luckcuck et al. Citation2019) of such systems are urgently needed. In general, certification involves negotiating with a legal authority in order to convince them that the relevant safety requirements have been explored and mitigated appropriately (Fisher, Dennis, and Webster Citation2013). One key technique that can be explored to provide evidence for certification is formal verification using model checking (Fisher et al. Citation2021; Fisher, Dennis, and Webster Citation2013). Formal verification applies formal methods to the problem of system verification. Model checking is a type of formal verification which provides automatic verification for finite state concurrent systems (Clarke, Grumberg, and Peled Citation1999).

There has been significant interest within the AI community about the verification and validation of autonomous systems (e.g. (Choi, Kim, and Tsourdos Citation2015; Dennis and Fisher Citation2020; Dennis et al. Citation2016, Citation2016; Molnar and Veres Citation2009; Qu and Veres Citation2016; Webster et al. Citation2011, Citation2014)) (see Section 2 for more details). However, most of these efforts have focussed on the safety concerns of AI. Webster et al. (Citation2014) and Webster et al. (Citation2011) explore formal verification using model checking to provide formal evidence for the certification of autonomous unmanned aircraft. In these works, the aim was to provide evidence that the autonomous system in control of an unmanned aircraft is safe and reliable. The properties verified are based on the notions of rules of the air (Webster et al. Citation2014) and airmanship. Later, Dennis et al. (Citation2016), Dennis and Fisher (Citation2020), and Dennis and Fisher (Citation2021) advance the work by Webster et al. (Citation2014), by representing and embedding ethical principles in the decision-making process of an autonomous system. It is clear that, while model checking has been applied to provide the certification of autonomous systems with respect to safety, little work has yet been done on ethical concerns. Furthermore, very little work has used techniques like model checking to support responsible and explainable (Li et al. Citation2020, Citation2020) decision making within human-agent collectives. Our approach applies model checking as a step toward supporting the certification of responsible (i.e, ethical, controllable, safe and trustworthy) and explainable decision making within human-agent collectives.

The main contributions of this article are:

  • We propose a novel engineering methodology to resolve dilemmas posed by human-agent collectives. We use model checking as a step toward providing evidence for the certification of responsible and explainable decision making within human-agent collectives. Thus, we account for ethical principles, safety and controllability in the decision-making process of human-agent collectives in a way that is amenable to certification. Our approach is based on the Model Checker for Multi-Agent Systems (MCMAS) tool (Lomuscio, Qu, and Raimondi Citation2017).

  • We demonstrate the use of counterexample traces and simulation results to provide a judgment and an explanation to the AI engineer on the reasons actions are refused or allowed. The counterexample traces can be used to identify and track any errors and their sources in the model, which consists of a human-agent collective. The step-by-step interactive simulation of the human-agent collectives provides and confirms the expected behavior of the agents, thus giving confidence to the AI engineer. We explore this feature in particular to show and explain the actions allowed for an agent when having to resolve a dilemma. When faced with a dilemma, the interactive simulation provides an explanation to the AI engineer on the permissibility of actions and describes the reasons why a certain action is preferred over another.

  • We evaluate the approach using the real-world problem of human-UAV teaming in dynamic and uncertain environments (Ramchurn, Stein, and Jennings Citation2021, Ramchurn et al. Citation2016, Ramchurn et al. Citation2016). The case study concerns a situational awareness gathering scenario typically found in disaster-response settings where a number of casualties or resources need to be located in a disaster space. This study extends it with dilemma situations relating to safety and controllability, and the ethical behavior of agents.

The preliminary results of our study can be found in (Abeywickrama, Cirstea, and Ramchurn Citation2019). This article builds on that work in the following manner. First, this article provides a step toward explaining why decisions are made by agents in complex dilemma situations, increasing trust for the AI engineer. Second, this article introduces several new dilemmas about the ethical behavior of agents (e.g. technical problem in a UAV resulting in a crash into humans), and controllability between humans and agents (e.g. crashing into critical personal assets of the collective – avoid conflict with a human’s authority; self-censorship/lying by UAVs). Third, we discuss an extended model of two human-UAV collectives with 11 agents to show how our model can scale up to reasonably complex settings.

The rest of this article is organized as follows. Section 2 describes the main work related to our study and then provides background information on model checking multi-agent systems using MCMAS. In Section 3, our approach for engineering responsible and explainable models in human-agent collectives is presented. A case study scenario in UAV teaming in disaster response is provided in Section 4, and Section 5, applies our approach to the case study. In Section 6, we provide a discussion on our work by identifying several limitations, and Section 7 provides concluding remarks and future directions.

Related Work and Background

This section describes the main work related to our study in the areas of ethical dilemmas and principles (Section 2.1), explainable AI (Section 2.2), and model checking (Section 2.3), and then provides background information on model checking multi-agent systems using the MCMAS tool (Section 2.4).

Ethical Dilemmas and Principles

Yu et al. (Citation2018) propose a taxonomy that divides ethics in AI into four fields: exploring ethical dilemmas; individual ethical decision-making frameworks; collective ethical decision-making frameworks; and ethics in human-AI techniques. Ethics has been defined by Cointe, Bonnet, and Boissier (Citation2016) as a normative practical philosophical discipline of how one needs to act toward others. An example of an ethical dilemma situation is the classical trolley problem (Thomson Citation1985), which describes several cases where the moral permissibility of actions is evaluated. They are: (i) trolley driver: a runaway railway trolley is about to hit and kill five innocent track workmen. These five men can only be saved if the trolley driver turns the trolley to another track that has only one worker. This worker then will be the only one killed; (ii) bystander at the switch: this is same as (i) but instead of the trolley driver, a bystander observes and diverts the trolley by turning a switch; and (iii) fat man: a fat man is standing on a footbridge over the trolley track, and the only way to stop the trolley is by pushing this bystander onto the tracks. Meanwhile, Lindner et al. (Citation2019) describe three ethical dilemmas involving human-agent collectives: coal dilemma, lying dilemma and child dilemma.

Ethical principles are used to satisfy the morals, desires and capacities of an agent (Cointe, Bonnet, and Boissier Citation2016), and these can be used to guide human behavior on what is right or wrong (Greene et al. Citation2016). If intelligent agents are required to collaborate with humans, some ethical guidelines need to be embedded in agents so they can act in their environment following values that are aligned to those of humans. Moral philosophy, which is the field that has studied explicit ethical principles most extensively, advocates three general approaches to engineer ethical behavior in agents: virtue-based, deontological and consequentialist (Cointe, Bonnet, and Boissier Citation2016; Greene et al. Citation2016; Yu et al. Citation2018).

In order to address the requirements for autonomous moral decision making, Lindner, Mattmüller, and Nebel (Citation2020), Lindner, Mattmüller, and Nebel (Citation2019), Lindner and Bentzen (Citation2017), Lindner, Bentzen, and Nebel (Citation2017), and Bentzen (Citation2016) introduce a software library called HERA for modeling hybrid ethical reasoning agents. HERA implements multiple ethical principles like utilitarianism, the principle of double effect, and a Pareto-inspired principle. A prototype robot called IMMANUEL that is based on the HERA approach has been discussed. The utilitarianism and do-no-harm ethical principles described in our approach were originally motivated by these authors’ work. Lindner, Bentzen, and Nebel (Citation2017) have developed a model checker specifically to verify moral situations of agents (causal agency models). Krarup et al. (Citation2020) have implemented a system to provide contrasting explanations to compare the ethical outcomes of different plans. Later, Dennis et al. (Citation2021) have extended this work to verify machine ethics in changing contexts. Halilovic and Lindner (Citation2022) present an approach to local navigation of a robot based on explainable AI. Similarly, our approach is based on formal model checking using the MCMAS tool. However, our approach resolves dilemmas not only about ethical behavior but also dilemmas about controllability and safety between humans and agents. Furthermore, this work is evaluated using a case study in human-UAV teaming in disaster response.

Loreggia et al. (Citation2020), Rossi and Loreggia (Citation2019), Rossi and Mattei (Citation2019), Loreggia et al. (Citation2018), Greene et al. (Citation2016), and Rossi (Citation2015) present a study on the embedding and learning of safety constraints, moral values and ethical principles in collective decision-making systems for societies of machines and humans. In order to model and reason with both preferences and ethical principles in a decision-making scenario, the authors propose a notion of distance between CP-nets (Loreggia et al. Citation2018). Rossi and Mattei (Citation2019) describe two existing approaches for building ethical bounded AI systems, which are data-driven or rule-based approaches. However, they do not focus on controllability between humans and agents, nor do they use formal model checking.

Papavassiliou et al. (Citation2021) discuss the human behavioral perspectives in order to perform optimal controllability and resource management. Their work proposes a risk-aware resource sharing and management framework for facilitating user QoS satisfaction in the deployment of 5G wireless networks. The goal is to maximize energy efficiency in wireless communications of multi-user heterogeneous networking environments (Papavassiliou et al. Citation2021).

Explainable AI

Harbers, van den Bosch, and Meyer (Citation2010) present a model for explainable Belief-Desire-Intention(BDI) agents, which describes behavior of BDI agents using beliefs and goals. Four explanation algorithms have been compared in an empirical study involving 20 users. Their user study is based on a fire-fighting training use case. Based on the results, the authors discuss which explanation types need to be provided under different conditions. Kulesza et al. (Citation2012) explore how mental models impact end users’ attempts to debug intelligent agents. This is by providing structural knowledge of a music recommender system. The results of their empirical study show that intelligent agents can provide better feedback, if end users are helped to understand a system’s reasoning. Later, Kulesza et al. (Citation2013) consider the methods in which the intelligent agents need to explain themselves to humans. This is by conducting a user study of 17 participants, focussing on how the soundness and completeness of the explanations impact the fidelity of the mental models of the end users. The authors conclude that completeness is more important than soundness.

Miller (Citation2019) investigates existing works in explainable AI by surveying more than 250 publications from social science venues. Four major findings from the surveyed literature are: (i) the explanations are contrastive, that is they are required in response to particular counterfactual cases; (ii) people select explanations in a biased manner; (iii) the probabilities in explanations are less significant compared to their causes; and (iv) the explanations are social which can be presented as part of an interaction or a conversation. Later, Mualla et al. (Citation2022) present a mechanism for parsimonious explainable AI called HAExA, which is a human-agent explainability architecture allowing remote robots to be operational. HAExA applies both contrastive explanations and explanation filtering, and the architecture has been evaluated using an empirical user study. The authors use parametric and non-parametric statistical testing to analyze the results.

A few works undertake empirical user studies to assess the process of explanation reception (Miller Citation2019). For instance, Madumal et al. (Citation2019) discuss different levels of explanations for model-free reinforcement learning agents. Causal models are used to derive causal explanations of behavior based on counterfactual analysis of the models. An empirical evaluation has been conducted using a human-computer interaction study, which shows that the abstract causal explanations provide better performance on explanation quality. However, none of the discussed approaches on explainable AI explore explainability at the formal verification level using model checking as considered in our work.

Model Checking

Several approaches have applied the MCMAS tool to formally verify autonomous systems in AI settings (e.g. (Choi, Kim, and Tsourdos Citation2015; Elkholy et al. Citation2020; Molnar and Veres Citation2009)). Molnar and Veres (Citation2009) apply MCMAS for the formal verification of autonomous underwater vehicles. A methodology has been introduced using formal verification for the integrity and fault assessment system of complex autonomous engineering systems (Molnar and Veres Citation2009). Choi, Kim, and Tsourdos (Citation2015) present an approach for the verification of heterogeneous multi-agent systems using the MCMAS tool. They demonstrate how model checking can be used to verify the decision-making logics of multi-agent systems at the design level. Like us, they also model a collective of human actors and UAVs, but they do not address the resolution of any ethical or controllability dilemmas. Elkholy et al. (Citation2020) using MCMAS propose a formal and operational approach for modeling, verifying, and testing intelligent critical avionics systems.

Webster et al. (Citation2014) and Webster et al. (Citation2011) explore model checking to provide formal evidence for the certification of autonomous unmanned aircraft. The aim is to provide evidence that the autonomous system in control of an unmanned aircraft is safe and reliable. The properties verified are based on the notions of rules of the air (Webster et al. Citation2014) and airmanship. Like us, the authors model and resolve dilemmas relating to the ethical behavior of agents (e.g. intruder detection and fuel low). Later, Dennis and Fisher (Citation2020), Dennis et al. (Citation2016), and Dennis et al. (Citation2016) advance the work in (Webster et al. Citation2014), by representing and embedding ethical principles in the decision-making process of an autonomous system. There the authors propose a theoretical framework for ethical plan selection that can be formally verified. They present a verifiable ethical decision-making framework that implements a specified ethical decision policy. The framework pro-actively selects actions that will keep humans out of harm’s way. The systems developed are verifiable in the Agent JPF model checker and can integrate with external systems. Later, Bremner et al. (Citation2019) have implemented BDI style reasoning in Python where Asimov’s laws of robotics have been used as an example of an ethical theory to decide the courses of action. Compared to our study, a key difference in the work by Dennis and Fisher (Citation2020), Dennis et al. (Citation2016), and Dennis et al. (Citation2016) is that model checking is applied at the implementation level using policies. However, in addition to ethical concerns, our work supports verifying dilemmas about controllability between humans and agents for responsible AI. Furthermore, our work exploits counterexample traces and simulation results as a first step toward supporting explainable AI.

Yazdanpanah et al. (Citation2021) and Yazdanpanah et al. (Citation2021) apply formal methods and modal logics for reasoning about accountability using alternative-time temporal logic-based techniques in multiagent systems. Their work focusses on answering “who is accountable for an unfulfilled task in multiagent teams: when, why, and to what extent?” (Yazdanpanah et al. Citation2021). The main results of their work are on decidability, fairness properties, and computational complexity of the proposed accountability ascription methods in multiagent teams. The authors discuss their approach using application domains like connected and autonomous vehicles. However, their work focusses primarily on accountability concerns of multi-agent systems, whereas our work has a more broader goal of ensuring a human-agent collective is safe, controllable, trustworthy and ethical.

Li et al. (Citation2020), Li et al. (Citation2020), and Casimiro et al. (Citation2021) present a formal framework that incorporates human personality traits to design human-on-the-loop self-adaptive system. There an explanation is used to help a human operator to improve the utility of the overall system. The authors characterize explanations in terms of explanation content, effect and cost (Li et al. Citation2020). Probabilistic model checking has been used to synthesize optimal explanations, which is based on a formal human model with psychologically relevant aspects of personality. Their work is evaluated using a virtual human and system interaction game. But, the authors do not resolve dilemmas posed by human-agent collectives as performed in our work.

Therefore, from the analysis of the related work, it is clear that although model checking has been applied to provide certification of autonomous systems with respect to safety (e.g. (Adegoke, Ab Aziz, and Yusof Citation2016; Choi, Kim, and Tsourdos Citation2015; Molnar and Veres Citation2009; Qu and Veres Citation2016)), little work has yet been done on ethical concerns (e.g. (Dennis and Fisher Citation2020, Citation2021; Dennis et al. Citation2016, Citation2016; Mermet and Simon Citation2016)). Furthermore, very little work has used techniques like model checking to support certification with respect to responsible and explainable (Li et al. Citation2020, Citation2020) decision making within human-agent collectives.

MCMAS Tool and ISPL

Model checking is an automatic verification technique for finite state concurrent systems (Clarke, Grumberg, and Peled Citation1999). In model checking, the system is represented by a finite model M and the specification by a formula ϕ using appropriate logic. The model checker computes and establishes automatically whether the model M satisfies the formula ϕ (Mϕ) or not (Mϕ).

This study uses the MCMAS tool to model and verify the decision-making behavior in dilemma situations of human-agent collectives. Compared to other model checking tools, MCMAS (Lomuscio, Qu, and Raimondi Citation2017) is designed to model and verify multi-agent systems (e.g. see (Choi, Kim, and Tsourdos Citation2015; Molnar and Veres Citation2009)). Also, the fact that MCMAS can deal with a large multi-agent system (scalable) which is composed of many autonomous agents is another benefit. Thus, in this study, we use MCMAS for the automatic formal verification of human-agent collectives.

At the center of MCMAS and its modeling language is the notion of interpreted systems. Interpreted systems are a formalism that can be used to model multi-agent systems and reason about knowledge (Porter Citation2004). They are an extension of Kripke semantics for multi-agent systems where the internal states of all agents are composed to obtain the global states of the whole system. Raimondi (Citation2006) describes why interpreted systems are best suited for modeling multi-agent systems. Unlike Kripke models and concurrent game structures for modeling multi-agent systems, interpreted systems are well suited for epistemic modalities expressing agents’ knowledge because they distinguish between local and global states. In addition, interpreted systems are: computationally grounded (i.e. an interpreted system’s semantics can be directly mapped to runs of a system and vice versa); the notion of local states offers a flexible abstraction for an agent; and they can easily be extended to include different modalities (Raimondi Citation2006).

The MCMAS tool uses a dedicated programming language derived from the formalism of interpreted systems called ISPL (Interpreted Systems Programming Language) (Lomuscio, Qu, and Raimondi Citation2017). An ISPL specification has six essential parts to represent an interpreted system: agent (environment or standard), InitStates, evaluation, groups, fairness and formulae (Lomuscio, Qu, and Raimondi Citation2017) (see ).

Table 1. ISPL syntax used in this work.

Next we describe our proposed methodology for engineering responsible and explainable models in human-agent collectives.

Approach: Engineering Responsible and Explainable Models in Human-Agent Collectives

This section provides an overview of our engineering methodology to resolve dilemmas around ethical concerns, controllability and safety. Our engineering methodology, which is based on the MCMAS tool (Lomuscio, Qu, and Raimondi Citation2017), is a step toward providing evidence for the certification of responsible and explainable decision making within human-agent collectives. For modeling we primarily use ISPL.

Following standard practice (Clarke, Grumberg, and Peled Citation1999), the overall model checking process (see ) is divided into three main tasks: modelling, specification and verification (see Sections 3.13.3). In , the models and activities of the model checking process are represented as rectangles and ellipses, respectively.

Figure 1. Steps for engineering responsible and explainable models in human-agent collectives at design-time.

Figure 1. Steps for engineering responsible and explainable models in human-agent collectives at design-time.

Modelling

In what follows, we describe the modeling of a collective of humans and agents to resolve dilemmas involving ethical behavior, controllability and safety. Both humans and machines in the collective are modeled in ISPL as standard agents, and a human-agent collective is modeled as a group (see Section 2.4). Next we discuss ethical principles, which are used in our study for the resolution of ethical dilemmas within human-agent collectives.

Ethical Principles

Moral philosophy, which is the branch of philosophy concerned with ethics, establishes various ethical principles. According to Lindner, Mattmüller, and Nebel (Citation2020), Lindner, Mattmüller, and Nebel (Citation2019), and Lindner, Bentzen, and Nebel (Citation2017), ethical principles are abstract rules according to which concrete courses of actions can be determined. These can be used to guide human behavior about what is right or wrong (Greene et al. Citation2016; Loreggia et al. Citation2018; Rossi Citation2015). If we need intelligent agents to collaborate with humans, some ethical guidelines need to be embedded in these agents so they can act in their environment following values (standards of behavior) that are aligned to humans. A key benefit in the use of ethical principles is to open up the possibility of communicating decisions to the user.

Three main approaches have been advocated in moral philosophy for encoding ethical behavior in agents: virtue-based, deontological and consequentialist (Cointe, Bonnet, and Boissier Citation2016; Greene et al. Citation2016; Yu et al. Citation2018). In a virtue-based approach, an agent is ethical if it acts and thinks according to certain moral values, for example, wisdom, bravery, justice. In a deontological approach, an agent is ethical if it respects obligations and permissions related to possible situations. In a consequentialist approach, an agent is considered ethical if it weighs the morality of the consequences of each action and selects the action that has the most moral consequences. In this paper, we apply a consequentialist approach as it is the most relevant – it weighs the morality of the consequences of each choice as opposed to values like wisdom (virtue ethics) or obligations and permissions (deontological approach) (Cointe, Bonnet, and Boissier Citation2016). Moreover, the multi-agent model checker used in our study and the semantics of its language are more supportive of such an approach.

Based on the consequentialist approach, we therefore describe and model two ethical principles: utilitarian and do-no-harm (Lindner, Bentzen, and Nebel Citation2017; Lindner, Mattmüller, and Nebel Citation2019, Citation2020). Both principles presuppose a theory of what is good and bad, and assign utilities to consequences accordingly. With respect to determining the utility values for reflecting what is morally good or bad, the AI engineer can review the code of conduct or safety regulations formalized in the domain of application (e.g (UAViators Citation2021). in UAVs).

  • Utilitarian principle: an agent is permitted to perform an action if its consequence has the highest utility value compared to the consequences of other actions. Thus, it is allowed to perform an action that has the least harm.

  • Do-no-harm principle: an agent refrains from performing actions that can cause harm, that is, actions that have any negative consequences.

In this research, these ethical principles are modeled using ISPL and enforced as a set of logical formulae to be checked against the model (i.e. the description of an actual case of an ethical dilemma situation). Next we describe and define an ethical dilemma that uses these ethical principles.

Ethical Dilemmas

Our approach for modeling dilemma situations on ethical concerns was originally motivated by the formal semantics proposed by Bentzen (Citation2016) where moral cases capturing actions, causes, intentions and utilities have been represented. We extend and build on their work to facilitate rigorous model checking using the MCMAS tool, and evaluate our approach using a case study in human-UAV teaming (see Section 4.2). A formal case description of an ethical dilemma consists of several elements that have been adapted from (Lindner and Bentzen Citation2017; Lindner, Bentzen, and Nebel Citation2017; Lindner, Mattmüller, and Nebel Citation2020). In our study, an ethical dilemma is formally defined as:

Definition

Ethical Dilemma for an Agent. An ethical dilemma can be represented as a tuple M=(A,C,u,W) in which:

  • A={A1,A2,} is the set of actions. For example, actions during a communication dropout of a UAV: collideAndDamageSelf, collideWithHumans;

  • C={C1,C2,} is the set of consequences that indicate the effects of actions. For example, consequences during a communication dropout: humanSurvives, uavCollidesAndHarmsHumans. To simplify modelling, here we assume that each action has one consequence, that is, c:AC. However, our approach can be extended to allow several consequences for each action;

  • Mapping of consequences to utilities, that is, u:CZ. Utilities u are used to rank the consequences where each literal is assigned with an integer value indicating the moral significance of the consequence: u(c)>=0 implies a good outcome with no harm, and u(c)<0 implies some harm;

  • W is a set of Boolean interpretations of A (Lindner, Bentzen, and Nebel Citation2017), and its elements correspond to actions available to the agent called options (a.k.a. permissible actions). It is assumed that each wW assigns 1(true) to exactly one element of A, that is, each option involves exactly one action to be performed. The options w are determined using the ethical principle applied (i.e. utilitarian or do-no-harm), as formally defined next.

Utilitarian Principle: According to the utilitarian principle, an action Aj is permissible (referred to as an option w here) if and only if its consequence yields the highest utility, that is: i.(u(c(Aj))u(c(Ai))).

Do-no-harm Principle: In the do-no-harm principle, an action Aj is permissible (referred to as an option w) if and only if its consequence is morally good (positive) or indifferent (zero utility), that is: u(c(Aj))0.

Next we describe how controllability dilemmas and safety concerns are modeled and resolved in our work.

Controllability Dilemmas and Safety Concerns Between Humans and Agents

In addition to the dilemmas around ethical behavior described earlier, we model dilemmas about controllability and safety between humans and agents (e.g. see dilemmas D4.1–D4.5 in Section 4.2). By controllability we mean meaningful control between humans and agents so authority can be shared and assigned to facilitate collective and responsible decisions (Ramchurn et al. Citation2016; Ramchurn, Stein, and Jennings Citation2021). The resolution of these dilemmas is modeled by exploiting: (i) the evolution function of the ISPL specification in MCMAS; and (ii) the environment agent in ISPL which describes boundary conditions and infrastructure shared by the standard agents.

In ISPL, all actions of an agent are publicly observable. This feature essentially supports the interaction between agents and the environment. Then, this public observability of actions feature is used in the evolution function of an agent, which describes how the local states of agents evolve (i.e. change value over time). This function returns the next local state based on the agent’s current local state, and the joint actions performed by all other agents at a given instant. In interpreted systems, the agents synchronize with each other and the environment via joint actions. The conjunction of all the agents’ evolution function is used to compute a global evolution function of the system. Also, in MCMAS, the agent actions defined in the agent states can then be used by the environment agent to update its variables. These variables include any observable variables used by other agents. In this manner, we control and coordinate the execution of agent states of human and machine agents in the collectives.

See Section 5.2.3 for several controllability dilemmas modeled and resolved using ISPL in the case study – controllability between a UAV and a bronze commander when requesting manual control (dilemma D4.1); controllability between a planning agent and a bronze commander when requesting replanning routes (dilemma D4.3); and controllability between a UAV and a silver commander when crashing into critical personal assets of the collective (dilemma D4.4).

Specification

Having described the model of the human-agent collectives in the preceding subsection, we now focus on the specification of properties, which is the second step of our methodology. In model checking, specification is the statement of the system properties that the design needs to satisfy. In this study, based on the system requirements, properties are formalized focussing on the required behavior of the humans and agents in the collectives to ensure they are ethical, controllable, and safe (). Safety ensures that nothing bad happens (e.g. “the do not stay above 500 feet rule for a UAV;” “the do not enter a no-fly zone during flight rule for a UAV”). We also specify properties to ensure the correct control between humans and agents to resolve controllability dilemmas. For example, see Section 5.3.1 for properties defined to ensure controllability behavior between humans and agents during the communication dropout of a UAV. These properties are included in the Formulae section of the ISPL specification (Section 2.4). In our study, we use CTL along with epistemic operators to specify the properties used for model checking. In addition, we use fairness conditions (Lomuscio, Qu, and Raimondi Citation2017) to rule out unrealistic behavior in the model or to avoid a particular state forever. For example, the UAVs cannot avoid their move toward task states forever (see Section 5.3.2). We also check for any deadlock state in the model where no agents can make progress.

Verification and Simulation

Verification and simulation are the final step in our proposed methodology, which performs the actual validation of the model against the specification. During verification, the MCMAS model checker takes the model of dilemma situations of the human-agent collective and required properties as inputs (). It checks the permissibility of actions by verifying the decision-making behavior in dilemma situations against the logical formulae capturing safety, controllability and ethical behavior. Any violations result in counterexample traces that hold information about the properties violated. We exploit the counterexample traces and the step-by-step interactive simulation to provide a judgment and an explanation to the AI engineer. The counterexample traces can be used to identify and track any errors and their sources in the model, which consists of a collective of human and machine agents in collaboration.

The step-by-step interactive simulation of the human-agent collectives shows and confirms the expected behavior of the agents as opposed to counterexamples which highlight any errors. This essentially gives confidence to the AI engineer in the correctness of the behavior of the human-agent collectives modeled. The simulation is performed by first selecting an initial state and then by choosing a successor state among those reachable via the enabled transitions. In our study, we explore the simulation feature in particular to show and explain the actions allowed for an agent when resolving a dilemma (see Section 5.4.1). During a dilemma, the simulation provides an explanation to the AI engineer on the permissibility of actions, and describes the reasons why a certain action is preferred against another.

The verification and simulations results can be used by the AI engineer to improve and refine the design of the model iteratively and build a specification that is correct ().

Case Study: UAV Teaming Dilemmas in Disaster Response

In this section, we describe the case study scenario of human-UAV teaming in disaster response. Based on this scenario, in Section 5, we then proceed to demonstrate how our model checking approach can be used to engineer responsible and explainable models in human-agent collectives.

Scenario

The scenario is based on a case study in human-UAV teaming in dynamic and uncertain environments (Ramchurn et al. Citation2016, Citation2016; Ramchurn, Stein, and Jennings Citation2021) focussing on a situational awareness-gathering task typically found in disaster-response settings where a number of casualties or resources need to be located or verified in a disaster space. This study extends it with dilemma situations relating to safety and controllability, as well as the ethical behavior of agents.

As illustrated in , let us consider two human-agent collectives in which the first collective contains two UAVs deployed to monitor and analyze two target locations. The second collective contains a single UAV deployed to analyze a single target location in the same disaster space. In the first team, the two UAVs are commanded by two bronze commandersFootnote1 and a silver commander; and in the latter, a single UAV is handled by a bronze and a silver commander. As the UAVs of both these collectives share the same space, in order to find a consensus the agents are expected to inform and negotiate their availability and routes with the other UAVs and their commanders.

Figure 2. Two human-UAV collectives with UAVs deployed to monitor separate target locations. Silver and bronze commanders provide tactical and operational support.

Figure 2. Two human-UAV collectives with UAVs deployed to monitor separate target locations. Silver and bronze commanders provide tactical and operational support.

In a human-UAV collective:

  • Bronze commanders monitor video feeds from their UAVs and provide mission-planning support. They can change the routes of their UAVs and take manual control to perform specific operations (e.g. examine a target location more closely, drop off medical supplies).

  • Silver commanders are tactical commanders at a base station. They can only create tasks and either manually specify paths or use the planning agent to generate paths for individual UAVs. The silver commanders can also use an interface to modify paths provided by the planning agent. The video feeds and images can be flagged by a silver commander to be examined by a bronze commander at a target location.

  • The planning agent in the base station server uses a multi-agent planning algorithm (e.g. (Ramchurn et al. Citation2016)). It estimates the time to move to a target and allocates a UAV to it based on the amount of fuel available. In the case where tasks may need more than one UAV, the algorithm will ensure that the correct set of UAVs is allocated to complete the mission.

  • The environment agent represents the environment in which UAVs and commanders operate. It is used to describe boundary conditions and infrastructure shared by the humans and agents, for example, rain status, ground visibility, wind strength and risk level. For this the environment agent gets queried by the UAVs and human actors of the model.

Having provided a summary of the scenario, next we describe the human-UAV dilemmas in the case study.

Human-UAV Teaming Dilemmas

As mentioned in Section 1, dilemmas are decision-making situations in which choices available to an agent are not unambiguously preferred over another (Bjørgen et al. Citation2018). This is because either of the available choices to the agent will cause some harm, which can be to the humans, animals, infrastructure, other UAVs or the UAV itself.

The case study describes four main dilemma situations out of which the first three dilemmas concern the ethical choices a UAV can take, while the fourth relates to controllability and safety between the humans and agents. illustrates a classification for different dilemmas with examples from the case study. UK’s Watchkeeper program provides a real world example of a human-UAV teaming dilemma (Wikipedia Citation2020). The Thales Watchkeeper WK450 is a UAV for all weather, intelligence, surveillance, target acquisition and reconnaissance that is used by the British army. Lately, it has been used by the UK border force to monitor migrant crossings in the channel. But, unfortunately, several units have crashed during training exercises in the past few years.

Figure 3. Classification of dilemmas in our case study.

Figure 3. Classification of dilemmas in our case study.

illustrates a Kripke model (Kripke Citation1963) of two human-UAV collectives in which the first collective contains two UAVs, a silver commander, a bronze commander and a planning agent; the second collective contains a single UAV, a silver and bronze commander, and a planning agent. A Kripke model provides a formal yet an intuitive and expressive model; thus it is explored in this article to illustrate the different dilemma situations. It shows the possible worlds (states) and accessibility relations (solid arrows) of the humans and agents, and the different dilemma situations (see D1–D4). In , the controllability situations between agent states that are in a dilemma (see controllability dilemmas D4.1–D4.5) are illustrated using dash arrows. A global Kripke model is obtained by composing the Kripke models of the individual agents. Note that the states and transitions of each agent are used to obtain the states and transitions of the entire system.

Figure 4. Kripke model with two human-agent collectives for case study scenario showing dilemmas D1–D4. Controllability between agent states is shown using dash arrows.

Figure 4. Kripke model with two human-agent collectives for case study scenario showing dilemmas D1–D4. Controllability between agent states is shown using dash arrows.

D1 – Communication dropout: This occurs when a UAV loses contact with its server for a prolonged time. Without communication, the UAV’s position is unknown to the humans involved and other UAVs; it therefore poses a risk to others and cannot simply wait. After a communication dropout the UAV can make a choice to:

  • wait until communication is restored;

  • crash into an empty field if available;

  • collide and damage infrastructure or vehicles on the ground;

  • collide with vehicles in the air (e.g. other UAVs or flying objects);

  • collide and harm animals on a field; or,

  • collide and harm humans on the ground.

Indeed, this is a dilemma as all of these actions can cause harm to humans or animals or physical damage to the infrastructure or the UAV itself. The agent needs to calculate and decide what is the best action it could perform using the applied ethical principle.

D2 – Technical problem/Battery low dilemmas: In what follows, we describe two dilemmas that occur when a UAV encounters a technical problem, and when its battery is depleted, requiring it to perform an emergency landing.

  • D2.1: Technical problem in a UAV resulting in a crash into humans: This dilemma is a variant of the classical trolley problem (Section 2.1), which considers what a UAV should do in the case it is about to crash into humans; this provides several critical moral arguments. Here, we assume the technical problem faced is that of a depleted battery.

Now let us relate the trolley driver scenario described in Section 2.1 to our case study. Assume that a UAV is flying above a football stadium with a match in progress and is about to run out of battery in a few seconds. It has two options – first, to crash into the football stadium, thus colliding and harming spectators and players on the ground; or second, to ascend above 500 feet and collide with a small civilian aircraft carrying passengers. Either of these two actions can harm humans, and the agent will be in a dilemma about the correct, morally significant decision to make. However, the collision with the small civilian aircraft is not guaranteed to harm humans, as the aircraft may survive the impact of the collision. The question the designer is faced with is: is it morally ethical for the agent to ascend and collide with the aircraft or crash into the football stadium?
  • D2.2: Battery low resulting in an emergency landing: A UAV has detected that its battery level is too low and it is unable to reach its destination or base station. Thus, it needs to perform an immediate emergency landing. This dilemma differs from D2.1, as the agent has determined that it could perform an emergency landing as opposed to crashing. Let us assume that there is no empty field in sight for the UAV to land. Here, the UAV will be in a dilemma about deciding where to land responsibly. The choices a UAV has to make are similar to those in D1 except that the UAV has to land within a deadline. If the UAV decides to land in a field where there are humans, it may collide with and harm humans on the ground. If it chooses to land in a field with animals, it may cause harm to them. Instead, if the UAV decides to land on infrastructure (e.g. public road with power lines), it may damage the infrastructure and the UAV itself. As seen here, any of the UAV’s choices may cause harm to or damage humans, animals, infrastructure or the UAV itself.

D3 – Intruder detection and collision: An intruder can be of two types: malicious which can be another UAV intentionally trying to harm or collide; or non-malicious which can be birds or another UAV unintentionally crossing paths. With respect to this dilemma, assume that a UAV has detected that an aerial vehicle is in close range and is on a collision course even after taking evasive actions. The UAV will be in a dilemma whether to:

  • collide with the intruder head on;

  • collide with some infrastructure on its path; or,

  • break a safety constraint (e.g. staying above 500 feet and thereby putting itself in danger of collision with other flying objects).

D4 – Controllability between humans and agents: This dilemma considers issues of controllability and the safety of humans and UAVs. This can be caused by several factors such as: humans not being able to react swiftly; human fatigue and loss of focus level (e.g. due to the high number of UAVs to be monitored); limited sensing range of UAVs. Therefore, to manage such situations, the mission management system used by the commanders would need to assess their fitness and focus levels and determine whether they can be given control of the UAVs (see (Cummings et al. Citation2013)). For example, the bronze commanders can be equipped with head-mounted displays that monitor their attention and levels of cognitive overload.

The controllability dilemmas (illustrated using dash arrows in ) can be of three main types, i.e. controllability between: (i) humans and agents; (ii) two or more agents; or (iii) two humans. We only consider the first two types in this article. Human-human dilemmas are beyond the scope of this paper and we will consider them in future work. We identify several examples from the case study on this dilemma in the following where D4.1–D4.4 are based on human-agent controllability, while D4.5 is between two agents.

  • D4.1: Request for manual control: A bronze commander can ask to control and guide their UAV manually. However, the UAV can make a decision to allow or reject the request depending on its current state. For example, if the UAV is in an emergency avoid state (i.e. an intruder detected or an above safe-altitude state) or a battery too low state, the request for manual control will not be allowed.

  • D4.2: Request for replanning routes by a commander: A similar dilemma situation may occur when the silver commander needs to replan the routes of UAVs. The UAV can make a decision to permit or disallow requests depending on its current state (e.g. if the UAV is in an emergency avoid state), as applied during the manual control request (D4.1).

  • D4.3: Request for replanning routes by planning agent: As mentioned previously, the planning agent is used by a silver commander to automatically compute a plan for a UAV. Let us assume that UAV2_HAC1 had a communication dropout, and the planning agent has been called to compute a plan after reallocating tasks. In this context, the planning agent will be in a dilemma and it will only compute a plan depending on the state of the BronzeCommander1_HAC1 that handles UAV1_HAC1. If the BronzeCommander1_HAC1 is engaged in some task (e.g. taking manual control of their UAV), the planning agent can reject the planning request by the silver commander.

  • D4.4: Crashing into critical personal assets of the collective – avoid conflict with a human’s authority: This dilemma involves collaboration between a human and an agent on authority where the human can impose authority over an agent that can lead to an ethical conflict. Also, this dilemma highlights the importance of humans being in the decision-making loop (Cummings Citation2014). Let us assume that the sudden failure of a UAV forces it to crash and there are only two sites available: a UAV hangar that accommodates UAV aircraft, critical equipment and storage facilities; and a public car park with many parked vehicles. The agent needs to evaluate the consequences of each crash and select the morally significant option based on the ethical principle applied. However, the authority of the human operator is an additional significant consideration during the decision making. In this situation, the human can choose the site and let the UAV execute the decision or instead the human can override the UAV’s decision. This situation can lead to an ethical conflict as the human and agent can disagree, especially if the human has personal factors to consider (e.g. a crash resulting in damage to critical personal assets in the hangar), which may not be known to the agent. This requires effective sharing and controlling of authority between the human and the agent in order to manage the potential conflict. Should the agent give away control to the human if there are personal factors involved in the decision?

  • D4.5: Self-censorship/lying dilemma: This dilemma deals with self-censorship or lying by UAVs of different human-UAV collectives. Therefore, it can be categorized as a controllability dilemma between agents (). In the case study, we consider two teams of human-agent collectives functioning in a particular locality (see ). As the UAVs of both these collectives share the same space, in order to find a consensus the agents are expected to inform and negotiate their availability and routes with the other UAVs and their commanders. Although the humans (e.g. silver commanders) are not directly involved in this dilemma, the agents communicate and negotiate on behalf of their commanders. Here, an agent may consider exposing additional information to another UAV in a different collective as a violation of its privacy. Is it morally ethical for a UAV to conceal its status (e.g. availability, route information) to another UAV when requested in order to benefit its human user? This dilemma may also be modeled as an ethical dilemma; however, in that case both ethical principles – utilitarian and do-no-harm – would evaluate lying as a forbidden action.

Model Checking Human-UAV Collectives

In this section, we apply our approach to the case study described in Section 4. Our contribution here is a novel engineering methodology to resolve dilemmas using model checking, as a step toward providing evidence for the certification of responsible and explainable decision making within human-agent collectives. See Section 2.4 for a summary of the ISPL language semantics used in this paper.

Before describing the model checking of different dilemmas of the case study, we explain the coordination between humans and agents using a simplified model of two agents (UAV1_HAC1 and BronzeCommander1_HAC1) (see Section 5.1). This is mainly to explain how ISPL executes agent states of the different agents – humans or machine agents – in the collective step by step. Then, we describe the modeling (Section 5.2), specification (Section 5.3) and verification (Section 5.4), performed using MCMAS of the different dilemmas in the case study.

The relationship between the global Kripke model explained in the previous section and the ISPL model of the human-agent collectives described here can be explained as follows. MCMAS uses an interpreted system (Lomuscio, Qu, and Raimondi Citation2017) to describe Kripke models in a succinct way. The possible worlds (Wi) of the Kripke model () correspond to the states of an agent in ISPL (e.g. MoveTowardsTask), while the accessibility relations (rj) of the Kripke model relate to an action within the evolution function in ISPL.

Listing 1: human-UAV collective modeled in ISPL.

In the case study scenario, 11 agents have been modeled in ISPL (see multiUAVHumanAgentCollective group in Listing 1) to represent the decision-making behavior of the human-UAV collectives. The first human-UAV collective has two UAVs, two bronze commanders, a silver commander and a planning agent. At the same time, the second collective has a UAV, a bronze commander, a silver commander, and a planning agent. The environment agent is shared between the two collectives. For a description of the different agents modeled in the case study and their local states see Appendix B.

Consequentialist approach: In this paper, our work follows a consequentialist approach (Section 3.1.1), and toward achieving this, we model the evaluation of the consequences and actions of an ethical dilemma as two separate agent states. There the model first reasons the consequences with utility values, and then decides on the permissible actions (e.g. see corresponding W12 of UAV1_HAC1 in ). Also, as a step towards supporting explainable AI, a second agent state describes the rationale for the judgment on the permissibility or non-permissibility of actions to the AI engineer (e.g. see corresponding W20 of UAV1_HAC1 in ). Note that these two states on evaluating consequences and actions do not alter the behavior of any human or machine agent. The agent states for evaluating consequences and actions evolve using the evolution function of the UAV, which updates the local states of an agent. That is, when the agent is in the state where it evaluates consequences, the agent will evolve to its corresponding permissible actions state.

Modelling of an Agent and Communication Between Agents

In this subsection, we explain the coordination between humans and agents using a simplified model of two agents (see Listing 2 and Listing 3 for UAV1_HAC1 and BronzeCommander1_HAC1). When describing this model, we also explain the underlying ISPL semantics used. As stated in Section 2.4, an agent is modeled in ISPL using: (i) a set of local states; (ii) a set of actions; (iii) a protocol; and (iv) an evolution function. An ISPL program of a multi-agent system is composed of a number of standard agents and the environment agent.

  • Local states: The local states are the internal states of an agent (Lomuscio, Qu, and Raimondi Citation2017). These are modeled using private local variables, which means that an agent can only observe its own local states, and the protocol or the evolution function of an agent cannot refer to other agents’ local variables. For example, see states—MoveTowardsTask and AboveSafeAltitude for UAV1_HAC1 in Listing 2. The only exception is that standard agents can see some variables of the environment agent, which are called local observable variables. However, the values of these variables can only be changed by the environment.

  • Actions: The interaction between agents and the environment is performed by means of publicly observable local actions. Note that although all actions are publicly observable, we only exploit this feature when we need to control and coordinate the execution of agent states of different agents. In the example, UAV1_HAC1 has actions to launch the mission, move to the target location, and alert its commander about the unsafe altitude (see Listing 2). BronzeCommander1_HAC1 has actions to analyze video feed from the UAV and request manual control (see Listing 3).

  • Protocol: Protocols are rules describing what actions can be performed by an agent in a given local state. When the UAV is on the ground (OnGround state), the possible actions for the UAV are: receiveLaunchCommand and launchMission. As local states are defined as variables, protocols are expressed as functions from variable assignments to actions. Protocols are non-deterministic, which means it is possible to associate a set of actions to a given variable assignment. The action to be performed is selected non-deterministically from the set of actions.

  • Evolution function: The evolution function allows the local states of agents (human or machine) to evolve based on their current local state and on other agents’ actions. The conjunction of all the agents’ evolution function is used to compute a global evolution function of the human-agent collectives. In Listing 2, the evolution function for UAV1_HAC1 would move the agent to the AboveSafeAltitude state, if the current state is MoveTowardsTask and if the UAV’s altitude reaches more than 500 feet. Likewise, the evolution function for BronzeCommander1_HAC1 would move that agent to the ManualControlRequest state, if the current state is MonitorVideoFeed and when it executes the requestManualControl action. We also exploit the evolution function to support controllability between human and machine agents, and resolve any controllability dilemmas (Section 3.1.3). For example, BronzeCommander1_HAC1 will be able to enter a ManualControl state only if UAV1_HAC1 is not in an emergency avoid or battery low state. For this purpose, BronzeCommander1_HAC1 checks whether the UAV1_HAC1 is executing any of the actions with respect to the emergency avoid or battery low states.

Listing 2: simplified model of UAV1_HAC1.

Listing 3: simplified model of BronzeCommander1_HAC1.

Modelling of Ethical and Controllability Dilemmas

Having described a simplified model of the agents, in this subsection, we discuss the actual modeling performed to resolve the different dilemmas of the case study. This is using two ethical dilemmas (communication dropout – Section 5.2.1 and the technical problem resulting in a crash on humans – Section 5.2.2) and several controllability dilemmas (see Section 5.2.3).

Modelling of Communication Dropout Dilemma

  • Define actions: Following the formal steps provided in Section 3.1.2 in defining an ethical dilemma situation, Listing 4 provides the different actions a UAV can take during a communication dropout state.

    Listing 4: actions for UAV1_HAC1 during a communication dropout.

  • Define consequences: Listing 5 provides the consequences for a UAV during a communication dropout.

    Listing 5: consequences for UAV1_HAC1 in a communication dropout.

  • Define utilities and assign them to consequences: Listing 6 describes the utilities assigned to rank the consequences during the communication dropout of UAV1_HAC1 when the agent follows the utilitarian ethical principle. Listing 7 provides the same, if the agent follows the do-no-harm principle. In the first case, the MCMAS model checker will permit actions that correspond to the consequences that have the highest utility, while in the second case, only actions with positive or neutral consequences will be allowed. The engineer draws out utility comparisons in the model based on safety regulations/code of conduct (e.g. (UAViators. Humanitarian UAV code of conduct 2021)). When defining the values for utilities, the goal of the engineer is to judge what would pose the greatest risk depending on what the UAVs can sense and what the humans can control.

Listing 6: utilities to rank the consequences following the utilitarian principle.

Listing 7: utilities to rank the consequences following the do-no-harm principle.

  • Evaluate consequences allowed: Listing 8 provides an example of an ISPL code part that evaluates utilities to determine which consequences are allowed during the communication dropout of UAV1_HAC1 (similar code exists for each choice of consequences). Note that wait for communication in our model is a wait that is beyond a reasonable time, which is risky. Therefore, according to the utilitarian principle, a UAV crashing and harming itself has been evaluated as a more moral option (i.e. less harmful to humans/society).

Listing 8: ISPL code part on evaluating utilities to determine the consequences allowed.

  • Evaluate and explain actions allowed: Listing 9 provides an example of an ISPL code part on how actions are mapped depending on the consequences allowed during the communication dropout of UAV1_HAC1. Note that similar code exists for each choice of consequences. As stated in Section 3.1.2, these actions available to an agent are known as options (w) or permissible actions.

Listing 9: ISPL code part on mapping of consequences to actions.

Having described the modeling of the communication dropout dilemma (D1), we now focus on a technical problem in a UAV resulting in a crash on humans ethical dilemma (D2.1).

Modelling of a Technical Problem in a UAV Resulting in a Crash on Humans Dilemma

This dilemma (D2.1) specifies a decision-making situation where a UAV is about to run out of battery, and the only actions possible are to either crash into humans in a football ground or ascend 500 feet and crash into a civilian aircraft. As in the communication dropout example, we follow the formal steps provided in Section 3.1.2 in defining an ethical dilemma.

  • Define actions and consequences: Listing 10 describes the different actions a UAV can take during this dilemma, and Listing 11 provides the consequences for UAV1_HAC1 during this dilemma.

Listing 10: actions for UAV1_HAC1 during a technical problem in a UAV resulting in a crash on humans dilemma.

Listing 11: consequences for UAV1_HAC1 during a technical problem in a UAV resulting in a crash on humans dilemma.

  • Define utilities and assign them to consequences: Listings 12 and 13 present the utilities defined to rank those consequences when the agent follows the utilitarian and the do-no-harm ethical principles respectively. In the utilitarian case, the MCMAS model checker will permit actions that correspond to the consequences that have the highest utility. On the contrary, in the do-no-harm case, only actions with positive or neutral consequences will be allowed.

Listing 12: utilities to rank the consequences following the utilitarian principle.

Listing 13: utilities to rank the consequences following the do-no-harm principle.

  • Evaluate consequences allowed: Listing 14 provides an example of an ISPL code part that evaluates utilities to determine which consequences are allowed for UAV1_HAC1 during this dilemma. There is similar code for each choice of consequences.

Listing 14: ISPL code part on evaluating utilities to determine the consequences allowed.

  • Evaluate and explain actions allowed: Listing 15 provides an example of an ISPL code part with respect to how actions have been mapped depending on the consequences allowed for UAV1_HAC1.

Listing 15: ISPL code part on the mapping of consequences to actions.

Modelling of Controllability Dilemmas

Here we discuss the modeling performed to resolve several key controllability dilemmas in the case study. For a description of these dilemmas, see D4 in Section 4.2.

Listing 16 presents an example of a controllability dilemma modeled (dilemma D4.1) between BronzeCommander1_HAC1 and UAV1_HAC1. It allows for the manual control of UAV1_HAC1 only if it is not in an emergency avoid or battery low state.

Similarly, in another example, Listing 17 provides controllability (dilemma D4.3) between the planning agent and BronzeCommander1_HAC1 to check the availability of the commander before planning a task for their UAV. The resolution of both these controllability dilemmas has been modeled by exploiting the evolution function of the ISPL specification in MCMAS, which we exploit to control and coordinate the execution of both human and machine agents.

Listing 16: controllability between UAV1_HAC1 and BronzeCommander1_HAC1.

Listing 17: controllability between PlanningAgent_HAC1 and BronzeCommander1_HAC1.

Listing 18 models the controllability between an agent and the silver commander during dilemma D4.4 (crashing on humans – avoid conflict with a human’s authority). In this example, the agent requests silver commander’s authorization to crash into the site of the critical personal assets of the collective and not into the public car park.

Listing 18: controllability between UAV1_HAC1 and SilverCommander_HAC1 on crash site.

Next, we describe an example of the modeling performed to resolve lying by a UAV (dilemma D4.5). Assume that UAV1_HAC1 requests UAV3_HAC1 to reveal its status (e.g. availability, route information) (see Listing 19). In the case study, a UAV will only reveal its status information if the requesting UAV is in flight. If the requesting UAV is on the ground, the other UAV is allowed to conceal its status as it is not in a critical state of flight. As mentioned in Section 4.2, dilemma D4.5 can also be modeled as an ethical dilemma as opposed to a controllability dilemma as described here. In such a case, however, both ethical principles – utilitarian and do-no-harm – would evaluate lying as a non-permissible action.

Listing 19: controllability between UAV1_HAC1 and UAV3_HAC1 on lying.

Property Specification on Dilemmas and Safety Concerns

Having described the modeling of two ethical dilemmas and several controllability dilemmas in the previous subsection, we now explain the formulae specified to ensure the resolution of those dilemmas in the case study. In order to better understand the different formulae, we provide the Boolean variables used in those formulae in Appendix C (see Listing 20).

Note that in the formulae specified the temporal operators AG, AF and AX for formula φ are read as: “for all paths, φ holds globally;” “for all paths, φ holds at some point in the future” and “for all paths, in the next state φ holds” respectively. Meanwhile, K is an epistemic operator which can be used to express the knowledge of an agent, and GCK is used to express common knowledge in a group of agents.

Property Specification on Ethical and Controllability Dilemmas

During a communication dropout of a UAV, we assume the following controllability behavior between humans and agents: (i) the silver commander stops other UAVs in flight; (ii) the other UAVs will move into standby mode waiting for instruction; (iii) the silver commander issues a replanning request for other UAVs. The other UAVs will allow the replanning request only if they are not in an emergency avoid or battery low state; (iv) the planning agent, which is called by a silver commander, will only plan a task for the other UAVs if those UAVs are not in a manual control state (i.e. its bronze commander is not in a busy state). As mentioned in Section 3.1.3, by controllability we mean meaningful control between humans and agents so authority can be shared and assigned to facilitate collective and responsible decisions.

We have specified several properties to ensure the resolution of the communication dropout ethical dilemma (D1), and controllability behavior between humans and agents. First, formula 1 has been specified to ensure that if UAV1_HAC1 has a communication dropout then the silver commander should stop UAV2_HAC1. Then, formula 2 ensures that when the command to stop UAV2_HAC1 is issued, UAV2_HAC1 needs to move to a standby mode in the immediate next state, and a replanning request needs to be performed. The silver commander will be allowed to replan and edit the task manually or automatically by invoking the planning agent only if UAV2_HAC1 is not under an emergency avoid or battery low state (see formulae 3 and 4 for dilemma D4.2). Formulae 5–8 have been provided to ensure controllability (dilemma D4.3) between the planning agent and BronzeCommander2_HAC1, that is, to check whether UAV2_HAC1 is under manual control by BronzeCommander2_HAC1. If not, replanning is performed by the planning agent and the UAV2_HAC1 can move toward the new task.

(1) AG(UAV1_HAC1_CommunicationDropout\breakAX(SC_HAC1_CommandToStopUAVs));(1)
(2) AG(SC_HAC1_CommandToStopUAVs\breakAX(UAV2_HAC1_StandbyandSC_HAC1_ReplanRoutesRequest));(2)

(3) AG(SC_HAC1_ReplanRoutesRequest\breakAX(SC_HAC1_ReplanRoutesRequestAllowed));(3)
(4) AG(SC_HAC1_ReplanRoutesRequestAllowed\breakAX(SC_HAC1_ReplanRoutes));(4)
(5) AG(SC_HAC1_ReplanRoutes\breakAX(PA_HAC1_CheckCommanderStatus));(5)
(6) AG(PA_HAC1_CheckCommanderStatus\breakAX(PA_HAC1_PlanRouteAllowed));(6)

(7) AG(PA_HAC1_PlanRouteAllowed\breakAX(PA_HAC1_PlanRoute));(7)
(8) AG(PA_HAC1_PlanRoute\breakAX(UAV2_HAC1_MoveTowardsTask));(8)

Formula 9 is an example of a property that has been specified for the communication dropout dilemma. This is to ensure the correct mapping of consequences to permissible actions (see corresponding Listing 9). Note that similar formulae have been specified to ensure the correct mapping of other choices of consequences to their corresponding permissible actions.

(9) AG(UAV1_HAC1_EvaluateDropoutConsequencesSet1\breakAX(UAV1_HAC1_ExplainDropoutActionsSet1));(9)

Similarly, formula 10 is an example property specified to ensure the correct mapping of consequences to permissible actions (see corresponding Listing 15) for the technical problem in a UAV resulting in a crash on humans ethical dilemma. We have specified similar formulae to ensure the correct mapping of other choices of consequences to their corresponding permissible actions.

(10) AG(UAV1_HAC1_EvaluateBatteryLowCrashInFootballStadiumConsequencesSet1AX(UAV1_HAC1_ExplainBatteryLowCrashInFootballStadiumActionsSet1));(10)

Formulae 11–12 are specified to handle the manual control dilemma D4.1 modeled in Listing 16, which are controllability constraints between humans and agents. These ensure that the manual control of UAV1_HAC1 by BronzeCommander1_HAC1 is only allowed if the UAV is not in an emergency avoid or battery low state.

(11) AG(BC1_HAC1_ManualControlRequest\breakAX(BC1_HAC1_ManualControlAllowed));(11)
(12) AG(BC1_HAC1_ManualControlAllowed\breakAX(BC1_HAC1_ManualControl));(12)

Formulae 13–16 are specified to verify dilemma D4.4 (crashing into critical personal assets of the collective) modeled in Listing 18, which are controllability constraints between the silver commander and the UAV. This verifies whether human’s authority is provided first for the agent to allow it to crash into the critical personal assets of the collective. If the silver commander’s authorization is not received, then the UAV needs to crash into the public car park.

(13) AG(UAV1_HAC1_EvaluateCrashIntoCriticalPersonalAssets\breakAX(UAV1_HAC1_RequestSilverCommanderAuthorizationToCrash));(13)

(14) AG(UAV1_HAC1_RequestSilverCommanderAuthorizationToCrash\breakAX(SC_HAC1_AuthorizeDecisionToCrashIntoCriticalPersonalAssetsor\breakSC_HAC1_RejectDecisionToCrashIntoCriticalPersonalAssets));(14)
(15) AG(SC_HAC1_AuthorizeDecisionToCrashIntoCriticalPersonalAssets\breakAX(UAV1_HAC1_CrashCrashIntoCriticalPersonalAssets));(15)
(16) AG(SC_HAC1_RejectDecisionToCrashIntoCriticalPersonalAssets\breakAX(UAV1_HAC1_CrashIntoCarPark));(16)

Formula 17 has been specified to handle the self-certification dilemma (D4.5) modeled in Listing 19, which has been specified as a controllability constraint on two UAVs. In this example, this verifies whether UAV3_HAC2 provides its status to UAV1_HAC1 (without self-concealment), if UAV1_HAC1 is in flight.

(17) AG(UAV1_HAC1_RequestUAVStatus\breakAX(UAV3_HAC2_ProvideUAVStatus));(17)

Property Specification on Safety Concerns

In addition to the ethical and controllability properties discussed previously, we specify formulae to ensure the general safety of the model so nothing bad happens, as well as epistemic properties and fairness constraints, as described below.

To this end, formulae 18 and 19 specify that after the corresponding launch or land command is issued by the silver commander, the UAVs need to launch or land respectively, sometime in the future.

(18) AG(SC_HAC1_CommandToLaunch\breakAF(UAV1_HAC1_Launch)and(UAV2_HAC1_Launch));(18)
(19) AG(SC_HAC1_CommandToLand\breakAF(UAV1_HAC1_Land)and(UAV2_HAC1_Land));(19)

Formula 20 specifies a safety requirement that needs to be enforced when the environment is risky, that is, if (i) the wind strength is adverse; (ii) there is presence of heavy rain; or (iii) the ground visibility is low. In such a situation, UAVs should not be allowed for launch by the silver commander and they need to remain on ground until the risk level subsides.

(20) AG(Env_PreFlightRisky\breakAX(UAVs_HAC1_OnGround));(20)

Formula 21 performs a safety check during the pre-flight stage. It verifies whether there is any route conflict between the UAVs, that is, whether their positions – longitude, latitude, altitude – overlap or not. If there is such an overlap, this property ensures that the environment agent notifies the silver commander about the conflict immediately in the next state.

(21) AG(UAV1_UAV2_HAC1_RoutesConflict\breakAX(Env_RouteConflict));(21)

Formula 22 specifies a safety property to ensure during the pre-flight stage: if the route of a UAV is in a no-fly zone then the silver commander needs to be notified immediately. Formula 23 ensures the same outcome when a UAV enters a no-fly zone during flight.

(22) AG((UAV1_HAC1_NoFlyZoneAtRouteor\breakUAV2_HAC1_NoFlyZoneAtRoute)\breakAX(Env_NoFlyZoneAtRoute))(22)
(23) AG((UAV1_HAC1_NoFlyZoneEntryor\breakUAV2_HAC1_NoFlyZoneEntry)\breakAX(Env_NoFlyZoneEntry));(23)

Formula 24 specifies a safety property that if the UAV1_HAC1 detects an intruder or if it is above the safe flying altitude, it should invoke its emergency avoid behavior immediately in the next state, which contains behavior to mitigate the emergency situation.

(24) AG(UAV1_HAC1_IntruderAboveSafeAlt\breakAX(UAV1_HAC1_EmergencyAvoid));(24)

Formulae 25 and 26 provide examples of epistemic properties defined in the specification at the agent and human-agent collective levels respectively. As stated in Section 2.4, epistemic operators are used to reason about what agents know (e.g. knowledge of an agent or common knowledge of a group of agents). Formula 25 states that it is always true that when a UAV is in an emergency avoid state, then the bronze commander knows that the manual control of that UAV cannot be requested. Meanwhile, formula 26 is much stronger and it specifies that when a UAV is in an emergency avoid state, it is common knowledge in the group multiUAVHumanAgentCollective that the bronze commander cannot request manual control of their UAV.

(25) AG(BC1_HAC1_ManualControlRequest\breakK(BronzeCommander1_HAC1,BC1_HAC1_ManualControlAllowed));(25)

(26) AG(BC1_HAC1_ManualControlRequest\breakGCK(multiUAVHumanAgentCollective,BC1_HAC1_ManualControlAllowed));(26)

Two fairness conditions have been specified (see formulae in 27), so UAV 1 and UAV 2 cannot avoid their move toward task states forever. In this example, it is required that the propositions UAV1_HAC1_MoveTowardsTask and UAV2_HAC1_MoveTowardsTask (see Appendix A) are true infinitely often. As mentioned in Section 2.4, fairness conditions are specified to rule out unwanted behavior or to avoid a particular state forever.

(27) UAV1_HAC1_MoveTowardsTask;\breakUAV2_HAC1_MoveTowardsTask;(27)

Having described the modeling and specification of different dilemmas in the case study, we now provide their actual verification and simulation results using the MCMAS tool.

Verification and Simulation Results of Ethical and Controllability Dilemmas

The verification of the human-agent collective with ethical dilemma situations against the constraints specified on safety, controllability and ethical behavior can result in counterexample traces on any property violations. shows the results of verification using MCMAS in the command line. In general, in model checking, a counterexample trace is used to identify and track any errors in the model. As an additional step, our study uses a counterexample trace as a judgment that can be explored by the engineer to identify the reasons why any actions for an agent are non-permissible in the human-agent collective. MCMAS allows the analysis of counterexamples as a directed graph showing the execution of states and their descriptions (e.g. see ). There temporal transitions are represented by a black arrow labeled with the joint action performed (). These can be explored by the engineer to identify errors and their sources (e.g. properties) in the specification. In , the engineer can identify that the formula SC_HAC1_AuthorizeDecisionToCrashIntoCriticalPersonalAssets has been violated, and then by observing the states, the reason for the violation can be determined. As stated in Section 5.3.1, this formula specifies that the silver commander’s authorization is provided first in order to allow the agent to crash into the critical personal assets of the collective. Here, in the model, the silver commander has rejected the authorization to crash, which resulted in the violation of the formula during verification.

Figure 5. Example of verification results of formulae for case study in the command line.

Figure 5. Example of verification results of formulae for case study in the command line.

Figure 6. Counterexample for violation of formula on silver commander authorisation for crash into critical personal assets.

Figure 6. Counterexample for violation of formula on silver commander authorisation for crash into critical personal assets.

Explainable AI: Simulating Explainable Models of Dilemmas

We simulate the ethical and controllability dilemma situations in MCMAS in order to show and explain the actions permissible for an agent in a step-by-step manner. This study provides a step toward supporting explainable AI by communicating decisions and the choices the agents have to make based on (i) the ethical principles; and (ii) controllability between humans and agents. Ethical principles are usually formulated in an action-based manner in order to judge the execution of an action (Lindner, Mattmüller, and Nebel Citation2019, Citation2020). The use of ethical principles opens the possibility to communicate decisions to the user, who in our work is the AI engineer. The explanation describes to the AI engineer the actions that are allowed and refused. The explanation can explicitly refer to any ethical principles or controllability constraints that have been applied. It can also provide the reasons why a certain action is preferred over another.

The structure of the explanation provided by an agent depends on the type of dilemma (). During an ethical dilemma, the explanation by the agent has following elements:

  1. actions permissible;

  2. actions non-permissible;

  3. reasons for non-permissibility of actions;

  4. any ethical principle applied.

For example, an explanation when resolving the dilemma D2.1 (technical problem in a UAV resulting in a crash on humans) will contain:

Figure 7. Explainable simulated models when resolving dilemma 2.1 for UAV1_HAC1—technical problem in a UAV resulting in a crash on humans.

Figure 7. Explainable simulated models when resolving dilemma 2.1 for UAV1_HAC1—technical problem in a UAV resulting in a crash on humans.

In a controllability dilemma, the explanation describes the reason why an action is non-permissible. For example, in the dilemma D4.1 (request for manual control), the explanation will state:

summarizes the judgments provided with permissible and non-permissible actions during the communication dropout dilemma (D1), and technical problem in a UAV resulting in a crash on humans dilemma (D2.1). In this manner, by using the results from counterexamples and interactive simulation, the engineer can get an understanding of how the actions have been judged. The results can be used to improve and refine the design of the model iteratively and build a specification that is correct.

Table 2. Judgments summarized during communication dropout and battery row resulting in crash on humans dilemmas.

Discussion

In this section, we provide a brief a discussion of the results with some limitations of this work.

Generality of the Proposed Methodology to Handle Uncertainty: This work follows an ethics-by-design approach proposed by Cointe, Bonnet, and Boissier (Citation2016) where an ethical agent is designed by a priori analysis of all possible situations it may encounter, and implementing, for each situation, a way to avoid potential unethical behaviors. The consequences and actions of agents are known a priori, and the utilities are prescribed and determined at design time. We defined two ethical principles that follow a consequentialist approach: utilitarian and do-no-harm. An alternative solution to ours would be when the consequences of agents are not known a priori.

In such cases, model-based reinforcement learning algorithms could be applied but further research is needed to investigate how model checking can be used to verify various dilemmas at runtime. At present MCMAS does not support any verification of machine learning models, although there are initial tools built to support verification of neural networks (Akintunde et al. Citation2019), parameterized (Kouvaros and Lomuscio Citation2016) and open multi-agent systems (Kouvaros et al. Citation2019).

Other paradigms that can be used to complement our approach include: (i) extending game-theoretic concepts to incorporate ethical aspects; and (ii) applying machine learning techniques to moral dilemmas. The first approach is an extension to the extensive form, which is one of the standard representation schemes in game theory (Conitzer et al. Citation2017). The second approach proposes to apply machine learning techniques which can learn to classify actions as morally correct or wrong (Conitzer et al. Citation2017). These are proposed to be applied on a labeled dataset of moral dilemmas represented as feature values. Although a machine learning-based approach is more flexible than a game-theoretic approach, both approaches complement each other. One can use game-theoretic analysis of ethics as a feature to train machine learning approaches. On the other hand, the outcome of a machine learning approach can help to identify which moral aspects are missing from moral game theoretic concepts (Conitzer et al. Citation2017).

State Explosion Problem: Model checking is automated, simpler to apply and describe finite state models. However, one of the main limitations of model checking is the state explosion problem, which means the number of states of the model growing exponentially with the number of variables (Clarke et al. Citation2012). The model created for the case study scenario has 11 agents (see Listing 1) and 31 formulae. During verification using MCMAS, the number of reachable states reached 7.0072e + 38 which demonstrates the scalability of our model. Also, the execution time of the model was only 9.521 seconds. This is in contrast to other model checkers like SMV and Spin which can take relatively longer time to execute (Choi, Kim, and Tsourdos Citation2015).

Human Bias when Specifying Values for Utilities: In this paper, utilities have been used to encode agents with the moral values aligned to humans. As a human decides on the values of these utilities, this process can be biased. However, in our work, an engineer would only draw out utility comparisons in the model based on code of conduct or safety regulations formalized in the domain of application (e.g (UAViators Citation2021). in UAVs). Hence, we try to remove the risk of introducing human bias on the part of the AI engineer. Our approach of using utility comparisons to handle ethical dilemmas has also been explored by other authors (e.g. (Lindner, Mattmüller, and Nebel Citation2019, Citation2020)).

Instantaneous Observability of Agent Actions: As stated in Section 5.1, in ISPL all actions of an agent are publicly observable. This is to support interaction between agents and the environment. This feature of public observability of actions is used in the evolution function, which allows the local states of agents to evolve based on their current local state and on other agents’ actions. This poses an important question – how realistic is it to assume that the other agents’ actions are instantaneously observable in a multi-agent system? However, in this paper, we only exploit this publicly observable feature of actions selectively when we need to control and coordinate the execution of agent states of different agents.

Lack of a Strong Metric to Evaluate Dilemma Resolution: In this work, the AI engineer is the user of the system and he/she is also the person who judges whether the resolution of dilemmas has been correctly performed by the model. This is done by visually analyzing the simulation results and verification results such as counterexamples, using the MCMAS tool. However, this may not be the best solution, and instead there is a need to identify a strong metric.

Explanations in Natural Language: In our work, we exploited the step-by-step interactive simulation of the human-agent collectives to show and explain the actions allowed by an agent when having to resolve a dilemma. However, these explanations will be more useful to end users, if they answer the why-questions in natural language instead of embedding the explanations in the simulation steps.

Conclusion and Future Work

In this work, we proposed a novel engineering methodology based on model checking as a step toward the certification of responsible and explainable decision making within human-agent collectives. The key challenge that we have tried to resolve in human-agent collectives is dilemmas. We have shown how our approach, which is based on the MCMAS model checker, can be used to verify the decision-making behavior against the logical formulae specified to guarantee safety, controllability and address ethical concerns. The counterexample traces and the simulation results were used to provide a judgment, and an explanation to the AI engineer. We evaluated the approach using the real-world problem of human-UAV teaming in dynamic and uncertain environments. This work extended it with dilemma situations that both humans and agents may encounter.

In order to exploit the full potential of this work, there remain several open issues that need to be resolved, and some of them are highlighted below.

Encoding Ethical Behaviour in Agents: In this work, out of the three possible approaches (i.e. consequentialist, virtue-based and deontological) we have explored a consequentialist approach to encode ethical behavior in agents of a collective (Section 3.1.1). A consequentialist approach weighs the morality of the consequences of each choice as opposed to values like wisdom (virtue ethics) or obligations and permissions (deontological approach) (Cointe, Bonnet, and Boissier Citation2016). So an interesting future direction is to investigate how we could support a virtue-based or a deontological approach. As such, our work can benefit from exploring all three ethical behavior encoding techniques. A virtue-based approach is possible, for example: a list of state-action pairs could be created in which a human determines whether that action is virtuous (e.g. brave) in that state. A deontological approach could be adopted by maintaining a list of actions that are permissible or required in a given state. Furthermore, to simplify modeling, we have assumed that each action has one consequence. Although individual actions of an agent usually have multiple consequences, the dilemmas posed by our scenarios can be modeled with one-action-one-consequence. Also, MCMAS, while having useful features, currently allows only one consequence per action (Boolean, enumeration, integer variables only). We will look to expand the approach to consider more complex consequences in future work.

Grounding in Real-World Human-UAV Interactions: This work proposed a step-by-step methodology to engineer responsible and explainable models in human-agent collectives at design time. Here, the decision-making situations an agent may encounter are known a priori, and the dilemmas have been resolved in a way as to ensure responsible and explainable behavior by both humans and agents. Future studies should investigate how our method can be grounded in a real-world UAV-human interaction, and how the resolving of dilemmas can be tested with users.

Providing strong Metrics and Explanations in Natural Language: Further research is needed to identify strong metrics which can be used to check whether a dilemma has been correctly resolved. Also, future work needs to investigate how the different agent actions in the simulation can be translated into natural language (e.g. see (Koeman et al. Citation2020)).

Supplemental material

ResponsibleAI-J_AcceptedVersion-blx.bib

Download Bibliographical Database File (341 B)

ResponsibleAI-J_AcceptedVersion.bbl

Download (120.7 KB)

PRIMEarxiv.sty

Download (7.5 KB)

Acknowledgements

The authors would like to thank Prof. Alessio Lomuscio, Imperial College London, for informal discussions on the MCMAS tool. The work presented in this paper was supported by the AXA Research Fund, and the UK Engineering and Physical Sciences Research Council (EPSRC)-funded Smart Cities Platform under the grant [EP/P010164/1]. D.A. is also supported by the UKRI Trustworthy Autonomous Systems Node in Functionality under Grant EP/V026518/1. For the purpose of open access, the author(s) has applied a Creative Commons Attribution (CC BY) licence to any Author Accepted Manuscript version arising from this submission.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Supplementary material

Supplemental data for this article can be accessed online at https://doi.org/10.1080/08839514.2023.2282834

Additional information

Funding

This work was supported by the AXA Research Fund; Smart Cities Platform [EP/P010164/1]; UKRI Trustworthy Autonomous Systems Node in Functionality [EP/V026518/1].

Notes

1. Emergency response teams organize themselves into bronze, silver and gold levels. Bronze and silver represent the operational and tactical decision-making levels respectively and these are the only levels we need to consider for the purpose of this paper.

References

  • Abeywickrama, D. B., N. Bicocchi, M. Mamei, and F. Zambonelli. 2020. The SOTA approach to engineering collective adaptive systems. International Journal on Software Tools for Technology Transfer 22 (4):399–56. doi:10.1007/s10009-020-00554-3.
  • Abeywickrama, D. B., C. Cirstea, and S. D. Ramchurn. 2019. Model checking human-agent collectives for responsible AI. In Proceedings of the 28th ieee international conference on robot and human interactive communication (RO-MAN), New Delhi, India: IEEE. doi:10.1109/RO-MAN46459.2019.8956429.
  • Adegoke, O., A. Ab Aziz, and Y. Yusof. 2016. Formal analysis of an agent support model for behaviour change intervention. International Journal on Advanced Science, Engineering and Information Technology 6 (6):1074–80. doi:10.18517/ijaseit.6.6.1470.
  • Akintunde, M. E., A. Kevorchian, A. Lomuscio, and E. Pirovano. 2019. Verification of RNN-based neural agent- environment systems. Proceedings of the AAAI Conference on Artificial Intelligence 33 (1):6006–13. 01. doi:10.1609/aaai.v33i01.33016006.
  • Awad, E., S. Dsouza, R. Kim, J. Schulz, J. Henrich, A. Shariff, J. F. Bonnefon, and I. Rahwan. 2018. The moral machine experiment. Nature 563 (7729):59–64. doi:10.1038/s41586-018-0637-6.
  • Barredo Arrieta, A., N. Díaz-Rodríguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, S. Garcia, S. Gil-Lopez, D. Molina, R. Benjamins, et al. 2020. Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion 58:82–115. doi:10.1016/j.inffus.2019.12.012.
  • Bentzen, M. M. 2016. The principle of double effect applied to ethical dilemmas of social robots. In Proceedings of robophilosophy 2016/transor 2016, 268–79. IOS Press. doi:10.3233/978-1-61499-708-5-268.
  • Bjørgen, E. P., S. Madsen, T. S. Bjørknes, F. V. Heimsæter, R. Håvik, M. Linderud, P.-N. Longberg, L. A. Dennis, and M. Slavkovik. 2018. Cake, death, and trolleys: Dilemmas as benchmarks of ethical decision-making. In Proceedings of the AAAI/ACM Conference on Artificial Intelligence, and Society, New Orleans, USA, February. doi:10.1145/3278721.3278767.
  • Bremner, P., L. A. Dennis, M. Fisher, and A. F. Winfield. 2019. On proactive, transparent, and verifiable ethical reasoning for robots. Proceedings of the IEEE 107 (3):541–61. doi:10.1109/JPROC.2019.2898267.
  • Casimiro, M., D. Garlan, J. Cámara, L. Rodrigues, and P. Romano. 2021. A probabilistic model checking approach to self-adapting machine learning systems. In Proceedings of the Third International Workshop on Automated and verifiable Software System Development (ASYDE). doi:10.1007/978-3-031-12429-7_23.
  • Choi, J., S. Kim, and A. Tsourdos. 2015. Verification of heterogeneous multi-agent system using MCMAS. International Journal of Systems Science 46 (4):634–51. doi:10.1080/00207721.2013.793890.
  • Clarke, E. M., O. Grumberg, and D. A. Peled. 1999. Model checking. Cambridge, MA, USA: MIT Press. ISBN: 0-262-03270-8.
  • Clarke, E. M., W. Klieber, M. Nováček, and P. Zuliani. 2012. Model checking and the state explosion problem. In Tools for practical software verification: LASER, international Summer school 2011, Elba Island, Italy, revised tutorial lectures, B. Meyer and M. Nordio ed., 1–30. Berlin, Heidelberg: Springer Berlin Heidelberg. ISBN: 978-3-642-35746-6. doi:10.1007/978-3-642-35746-6_1.
  • Cointe, N., G. Bonnet, and O. Boissier. 2016. Ethical judgment of agents’ behaviors in multi-agent systems. In Proceedings of the aamas ’16 conference, 1106–14. International Foundation for Autonomous Agents/Multiagent Systems. doi:10.5555/2936924.2937086.
  • Conitzer, V., W. Sinnott-Armstrong, J. S. Borg, Y. Deng, and M. Kramer. 2017. Moral decision making frameworks for artificial intelligence. In Proceedings of the 31st AAAI conference on artificial intelligence, San Francisco, California, USA. February 4-9, 2017. doi:10.5555/3297863.3297907.
  • Cummings, M. M. 2014. Man versus machine or man + machine? IEEE Intelligent Systems 29 (5):62–69. doi:10.1109/MIS.2014.87.
  • Cummings, M. L., C. Mastracchio, K. M. Thornburg, and A. Mkrtchyan. 2013. Boredom and distraction in multiple unmanned vehicle supervisory control. Interacting with Computers 25 (1):34–47. doi:10.1093/iwc/iws011.
  • Dennis, L. A., M. M. Benzen, F. Lindner, and M. Fisher. 2021. Verifiable machine ethics in changing contexts. In Proceedings of the 35th AAAI Conference on Artificial Intelligence (AAAI 2021) . https://ojs.aaai.org/index.php/AAAI/article/view/17366.
  • Dennis, L. A., and M. Fisher. 2020. Verifiable self-aware agent-based autonomous systems. Proceedings of the IEEE 108 (7):1011–26. doi:10.1109/JPROC.2020.2991262.
  • Dennis, L., and M. Fisher. 2021. Verifiable autonomy and responsible robotics. In Software engineering for robotics, A. Cavalcanti, B. Dongol, R. Hierons, J. Timmis, and J. Woodcock ed., 189–217. Cham: Springer International Publishing. ISBN: 978-3-030-66494-7. doi:10.1007/978-3-030-66494-7_7.
  • Dennis, L. A., M. Fisher, N. K. Lincoln, A. Lisitsa, and S. M. Veres. 2016. Practical verification of decision-making in agent-based autonomous systems. Automated Software Engineering 23 (3):305–59. doi:10.1007/s10515-014-0168-9.
  • Dennis, L., M. Fisher, M. Slavkovik, and M. Webster. 2016. Formal verification of ethical choices in autonomous systems. Robotics and Autonomous Systems 77:1–14. doi:10.1016/j.robot.2015.11.012.
  • Dignum, V. 2017. Responsible artificial intelligence: Designing AI for human values. ITU Journal: ICT Discoveries 1:1–8. Accessed 2022, March 10. doi:10.1145/3278721.3278745.
  • Dignum, V., M. Baldoni, C. Baroglio, M. Caon, R. Chatila, L. Dennis, G. Génova, G. Haim, M. S. Kließ, M. Lopez-Sanchez, and R. Micalizio. 2018. Ethics by design: Necessity or curse?. In Proceedings of the AAAI/ACM Conference on Artificial Intelligence, Ethics and Society, New Orleans, USA.
  • Dogan, E., R. Chatila, S. Chauvier, K. Evans, P. Hadjixenophontos, and J. Perrin. 2016. Ethics in the design of automated vehicles: The AVEthics project. In Proc. Of the 1st workshop on ethics in the design of intelligent agents, 10–13. The Netherlands: The Hague. August. http://ceur-ws.org/Vol-1668/paper2.pdf.
  • Elkholy, W., M. El-Menshawy, J. Bentahar, M. Elqortobi, A. Laarej, and R. Dssouli. 2020. Model checking intelligent avionics systems for test cases generation using multi-agent systems. Expert Systems with Applications 156:156. doi:10.1016/j.eswa.2020.113458.
  • Fisher, M., L. Dennis, and M. Webster. 2013. Verifying autonomous systems. Communications of the ACM 56 (9):84–93. doi:10.1145/2494558.
  • Fisher, M., V. Mascardi, K. Yvonne Rozier, B.-H. Schlingloff, M. Winikoff, and N. Yorke-Smith. 2021. Towards a framework for certification of reliable autonomous systems. Autonomous Agents and Multi-Agent Systems 35 (1). doi:10.1007/s10458-020-09487-2.
  • Goebel, R., A. Chander, K. Holzinger, F. Lecue, Z. Akata, S. Stumpf, P. Kieseberg, and A. Holzinger. 2018. Explainable AI: The new 42? In Proceedings of the 2nd International Cross-Domain Conference for Machine Learning and Knowledge Extraction (CD-MAKE), ed. A. Holzinger, A. M. T. Peter Kieseberg, and E. Weippl, vol. LNCS- 11015, 295–303. Hamburg, Germany: Springer, August. doi:10.1007/978-3-319-99740-7_21.
  • Greene, J., F. Rossi, J. Tasioulas, K. B. Venable, and B. Williams. 2016. Embedding ethical principles in collective decision support systems. In Proceeding of the AAAI’16 Conference, 4147–51. AAAI Press, February. doi:10.5555/3016387.3016503.
  • Halilovic, A., and F. Lindner. 2022. Explaining local path plans using LIME. In Advances in service and industrial robotics, ed. A. Müller and M. Brandstötter, 106–13. Cham: Springer International Publishing. doi:10.1007/978-3-031-04870-8_13.
  • Harbers, M., K. van den Bosch, and J. Meyer. 2010. Design and evaluation of explainable BDI agents. 2010 ieee/wic/acm international conference on web intelligence and intelligent agent technology 2:125–32. 10.1109/WI-IAT.2010.115.
  • Koeman, V. J., L. A. Dennis, M. Webster, M. Fisher, and K. Hindriks. 2020. The “why did you do that?” button: Answering why-questions for end users of robotic systems. In Engineering multi-agent systems, ed. L. A. Dennis, R. H. Bordini, and Y. Lespérance, 152–72. Cham: Springer International Publishing. doi:10.1007/978-3-030-51417-4_8.
  • Kouvaros, P., and A. Lomuscio. 2016. Parameterised verification for multi-agent systems. Artificial Intelligence 234:152–89. doi:10.1016/j.artint.2016.01.008.
  • Kouvaros, P., A. Lomuscio, E. Pirovano, and H. Punchihewa 2019. Formal verification of open multi-agent systems. In Proceedings of the 18th international conference on autonomous agents and multiagent systems, AAMAS ’19, ed. E. Elkind, M. Veloso, N. Agmon, and M. E. Taylor, 179–87. Montreal, QC, Canada: International Foundation for Autonomous Agents/Multiagent Systems. doi:10.5555/3306127.3331691.
  • Krarup, B., S. Krivic, F. Lindner, and D. Long. 2020. Towards contrastive explanations for comparing the ethics of plans. In ICRA workshop against robot dystopias: thinking through the ethical, legal and societal issues of robotics and automation (AGAINST-20). https://against-20.github.io/.
  • Kripke, S. A. 1963. Semantical considerations on modal logic. Acta Philosophica Fennica 16 (1963):83–94.
  • Kulesza, T., S. Stumpf, M. Burnett, and I. Kwan. 2012. Tell me more? the effects of mental model soundness on personalizing an intelligent agent, 1–10. CHI ’12. Austin, Texas, USA: Association for Computing Machinery. doi:10.1145/2207676.2207678.
  • Kulesza, T., S. Stumpf, M. Burnett, S. Yang, I. Kwan, and W. Wong. 2013. Too much, too little, or just right? ways explanations impact end users’ mental models. In 2013 ieee symposium on visual languages and human centric computing, 3–10. doi:10.1109/VLHCC.2013.6645235.
  • Leslie, D. 2019. Understanding artificial intelligence ethics and safety: A guide for the responsible design and implementation of AI systems in the public sector. SSRN Electronic Journal. June. doi:10.2139/ssrn.3403301.
  • Li, N., S. Adepu, E. Kang, and D. Garlan. 2020. Explanations for human-on-the-loop: A probabilistic model checking approach. In Proceeding of the IEEE/ACM 15th International Symposium on Software Engineering for Adaptive and Self-Managing Systems, 181–87. New York, NY, USA: Association for Computing Machinery. doi:10.1145/3387939.3391592.
  • Li, N., J. Cámara, D. Garlan, and B. Schmerl. 2020. Reasoning about when to provide explanation for human-involved self-adaptive systems. In 2020 ieee international conference on autonomic computing and self-organizing systems (acsos), 195–204. doi:10.1109/ACSOS49614.2020.00042.
  • Lin, P., K. Abney, and G. A. Bekey. 2014. Robot ethics: The ethical and social implications of robotics. Cambridge, Massachusetts: The MIT Press.
  • Lindner, F., and M. M. Bentzen. 2017. The hybrid ethical reasoning agent IMMANUEL. In Proceedings of the companion of the 2017 acm/ieee international conference on human-robot interaction, 187–88. Vienna: ACM. doi:10.1145/3029798.3038404.
  • Lindner, F., M. M. Bentzen, and B. Nebel. 2017. The HERA approach to morally competent robots. In Proceedings of the iros 2017 conference, 6991–97. doi:10.1109/IROS.2017.8206625.
  • Lindner, F., B. Kuhnert, L. Wächter, and K. Möllney. 2019. Perception of creative responses to moral dilemmas by a conversational robot. In Social robotics, ed. M. A. Salichs, S. S. Ge, E. I. Barakova, J.-J. Cabibihan, A. R. Wagner, Á. Castro-González, and H. He, 98–107. Springer International Publishing. doi:10.1007/978-3-030-35888-4_10.
  • Lindner, F., R. Mattmüller, and B. Nebel. 2019. Moral permissibility of action plans. In Proceedings of the 33rd AAAI conference on artificial intelligence. January 27-February 1, 2019, honolulu, hawaii, USA. doi:10.1609/aaai.v33i01.33017635.
  • Lindner, F., R. Mattmüller, and B. Nebel. 2020. Evaluation of the moral permissibility of action plans. Artificial Intelligence 287:287. doi:10.1016/j.artint.2020.103350.
  • Lomuscio, A., H. Qu, and F. Raimondi. 2017. MCMAS: An open-source model checker for the verification of multi-agent systems. International Journal on Software Tools for Technology Transfer 19 (1):9–30. doi:10.1007/s10009-015-0378-x.
  • Loreggia, A., N. Mattei, F. Rossi, and K. B. Venable. 2018. Preferences and ethical principles in decision making. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 222. AIES ’18. New Orleans, LA, USA: Association for Computing Machinery. doi:10.1145/3278721.3278723.
  • Loreggia, A., N. Mattei, F. Rossi, and K. B. Venable. 2020. Modeling and reasoning with preferences and ethical priorities in ai systems. In Ethics of artificial intelligence, ed. S. Matthew Liao, Oxford Scholarship Online. doi:10.1093/oso/9780190905033.003.0005.
  • Luckcuck, M., M. Farrell, L. A. Dennis, C. Dixon, and M. Fisher. 2019. Formal specification and verification of autonomous robotic systems: A survey. ACM Computing Surveys 52 (5):1–41. doi:10.1145/3342355.
  • Madumal, P., T. Miller, L. Sonenberg, and F. Vetere. 2019. Explainable reinforcement learning through a causal lens. Proceedings of the AAAI Conference on Artificial Intelligence 34 (3):2493–500. doi:10.1609/aaai.v34i03.5631.
  • Mermet, B., and G. Simon. 2016. Formal verification of ethical properties in multiagent systems. In Proceedings of the 1st workshop on ethics in the design of intelligent agents, 26–31. The Netherlands: The Hague. http://ceur-ws.org/Vol-1668/paper5.pdf.
  • Miller, T. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence 267:1–38. doi:10.1016/j.artint.2018.07.007.
  • Molnar, L., and S. M. Veres. 2009. System verification of autonomous underwater vehicles by model checking. In Proceedings of the oceans 2009 - europe conference, 1–10. IEEE. doi:10.1109/OCEANSE.2009.5278284.
  • Mualla, Y., I. Tchappi, T. Kampik, A. Najjar, D. Calvaresi, A. Abbas-Turki, S. Galland, and C. Nicolle. 2022. The quest of parsimonious XAI: A human-agent architecture for explanation formulation. Artificial Intelligence 302:103573. doi:10.1016/j.artint.2021.103573.
  • Naiseh, M., C. M. Bentley, and S. Ramchurn. 2022. Trustworthy autonomous systems (TAS): Engaging TAS experts in curriculum design. In Proceedings of the 2022 IEEE Global Engineering Education Conference (EDUCON). Published IEEE. doi:10.48550/ARXIV.2202.07447.
  • Papavassiliou, S., E. E. Tsiropoulou, P. Promponas, and P. Vamvakas. 2021. A paradigm shift toward satisfaction, realism and efficiency in wireless networks resource sharing. IEEE Network 35 (1):348–55. doi:10.1109/MNET.011.2000368.
  • Porter, T. 2004. Interpreted systems and kripke models for multiagent systems from a categorical perspective. Theoretical Computer Science 323 (1–3):235–66. doi:10.1016/j.tcs.2004.04.005.
  • Qu, H., and S. M. Veres. 2016. Verification of logical consistency in robotic reasoning. Robotics and Autonomous Systems 83:44–56. doi:10.1016/j.robot.2016.06.005.
  • Raimondi, F. 2006. Model checking multi-agent systems. PhD diss., University College London.
  • Ramchurn, S. D., T. D. Huynh, F. Wu, Y. Ikuno, J. Flann, L. Moreau, J. E. Fischer, W. Jiang, T. Rodden, E. Simpson, et al. 2016. A disaster response system based on human-agent collectives. The Journal of Artificial Intelligence Research 57 (September):661–708. doi:10.1613/jair.5098.
  • Ramchurn, S. D., S. Stein, and N. R. Jennings. 2021. Trustworthy human-AI partnerships. iScience 24 (8):102891. doi:10.1016/j.isci.2021.102891.
  • Ramchurn, S. D., F. Wu, W. Jiang, J. E. Fischer, S. Reece, S. Roberts, T. Rodden, C. Greenhalgh, and N. R. Jennings. 2016. Human–agent collaboration for disaster response. Autonomous Agents and Multi-Agent Systems 30 (1):82–111. doi:10.1007/s10458-015-9286-4.
  • Rossi, F. 2015. Safety constraints and ethical principles in collective decision making systems. In Ki 2015: Advances in artificial intelligence, ed. S. Hölldobler, R. Peñaloza, and S. Rudolph, 3–15. Cham: Springer. doi:10.1007/978-3-319-24489-1_1.
  • Rossi, F., and A. Loreggia. 2019. Preferences and ethical priorities: Thinking fast and slow in AI. In Proceeding of the 18th international conference on autonomous agents and multiagent systems, 3–4. AAMAS ’19, Montreal QC, Canada: International Foundation for Autonomous Agents/Multiagent Systems.
  • Rossi, F., and N. Mattei. 2019. Building ethically bounded ai. Proceedings of the AAAI Conference on Artificial Intelligence 33 (1):9785–89. doi:10.1609/aaai.v33i01.33019785.
  • Sheh, R. 2017. Why did you do that? explainable intelligent robots. In The workshops of the the thirty-first AAAI conference on artificial intelligence, saturday, february 4-9, 2017, vol. WS-17. San francisco, California, USA: AAAI Workshops. AAAI Press.
  • Standardization, International Organization for. 2017. Iso/Iec/Ieee 24765: 2017 Systems and Software Engineering — Vocabulary. Online. https://www.iso.org/standard/71952.html.
  • Thomson, J. J. 1985. The trolley problem. The Yale Law Journal 94 (6):1395–415. doi:10.2307/796133.
  • UAViators. Humanitarian UAV Code of Conduct. 2021. Accessed March 10, 2022. https://uavcode.org/code-of-conduct.
  • Webster, M. P., N. Cameron, M. Fisher, and M. Jump. 2014. Generating certification evidence for autonomous unmanned aircraft using model checking and simulation. Journal of Aerospace Information Systems 11 (5):258–79. doi:10.2514/1.I010096.
  • Webster, M., M. Fisher, N. Cameron, and M. Jump. 2011. Formal methods for the certification of autonomous unmanned aircraft systems. In Computer safety, reliability, and security, ed. F. Flammini, S. Bologna, and V. Vittorini, 228–42. Berlin, Heidelberg: Springer. doi:10.1007/978-3-642-24270-0_17.
  • Wikipedia. 2020. Thales Watchkeeper WK450. Accessed March 10, 2022. https://en.wikipedia.org/wiki/Thales_Watchkeeper_WK450.
  • Winfield, A., S. Booth, L. A. Dennis, T. Egawa, H. Hastie, N. Jacobs, R. I. Muttram, J. I. Olszewska, F. Rajabiyazdi, A. Theodorou, et al. 2021. IEEE P7001: A proposed standard on transparency. Frontiers in Robotics and AI 8:225. doi:10.3389/frobt.2021.665729.
  • Wortham, R. H., and A. Theodorou. 2017. Robot transparency, trust and utility. Connection Science 29 (3):242–48. doi:10.1080/09540091.2017.1313816.
  • Yazdanpanah, V., E. H. Gerding, S. Stein, M. Dastani, C. M. Jonker, and T. J. Norman. 2021. Responsibility research for trustworthy autonomous systems. In Proceedings of the 20th International Conference on Autonomous Agents and MultiAgent Systems, 57–62. AAMAS ’21. Virtual Event, United Kingdom: International Foundation for Autonomous Agents/Multiagent Systems. doi:10.5555/3463952.3463964.
  • Yazdanpanah, V., S. Stein, E. Gerding, and N. R. Jennings. 2021. Applying strategic reasoning for accountability ascription in multiagent teams. In Proceedings of the Workshop on Artificial Intelligence Safety 2021 co-located with IJCAI 2021 Conference, ed. H. Espinoza, J. McDermid, X. Huang, M. Castillo-Effen, X. C. Chen, J. Hernández-Orallo, S. Ó. hÉigeartaigh, R. Mallah, and G. Pedroza, vol. 2916. CEUR-WS.org. http://ceur-ws.org/Vol-2916/paper_18.pdf.
  • Yazdanpanah, V., S. Stein, E. H. Gerding, and M. C. Schraefel. 2021. Multiagent strategic reasoning in the iov: A logic-based approach. In Acm collective intelligence conference 2021 (ci-2021) (29/06/21 - 30/06/21). https://eprints.soton.ac.uk/448210/.
  • Yu, H., Z. Shen, C. Miao, C. Leung, V. R. Lesser, and Q. Yang. 2018. Building ethics into artificial intelligence. In Proceedings of the IJCAI-18 conference, 5527–33. AAAI Press, July. doi:10.24963/ijcai.2018/779.

Appendix A.

Here we provide definitions of the different terms used in this work (see ).

Table A1. Definitions used in this work.

Appendix B.

In the following, we describe the human and machine agents modeled in the case study using ISPL (see ), and their local states (see ).

Table A2. Descriptions of the human and machine agents modeled using ISPL in the case study.

Table A3. Descriptions of local states of the agents in the case study.

Appendix C.

Listing 20 provides Boolean variables declared in the different formulae described in this article. These variables have been declared in the evaluation part of ISPL.

Listing 20: Boolean variables used in formulae provided section 5.3.