1,907
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Dual use concerns of generative AI and large language models

Article: 2304381 | Received 15 May 2023, Accepted 08 Jan 2024, Published online: 29 Jan 2024

ABSTRACT

We suggest the implementation of the Dual Use Research of Concern (DURC) framework, originally designed for life sciences, to the domain of generative AI, with a specific focus on Large Language Models (LLMs). With its demonstrated advantages and drawbacks in biological research, we believe the DURC criteria can be effectively redefined for LLMs, potentially contributing to improved AI governance. Acknowledging the balance that must be struck when employing the DURC framework, we highlight its crucial political role in enhancing societal awareness of the impact of generative AI. As a final point, we offer a series of specific recommendations for applying the DURC approach to LLM research.

Introduction

High-ranking authorities in Europe and the United States have recently expressed their concern regarding the societal implications of generative artificial intelligence (Coulter and Mukherjee Citation2023; White House Citation2023a). These concerns have quickly found their way into regulatory documents, e.g. the Executive Order signed by the United States President J. Biden (The White House Citation2023) or the recently amended and agreed-upon European legislative proposal for the AI Act (Council of Europe Citation2023; European Parliament Citation2023b). In academic circles, such worries have also been echoed by distinguished scientists, e.g. computer scientist Geoffrey Hinton compared the risks associated with generative AI to the regulation of biological and chemical weapons by emphasizing that, although not ‘foolproof’, this regulation generally prevents their use (Heaven Citation2023). Perhaps the best examples of publicly declared concerns were, in early 2023, the call for a moratorium initiated by the Future of Life Institute and backed by figures such as Elon Musk; in late 2023, the ‘Bletchley Declaration’ signed at the UK summit on AI (AI Safety Summit Citation2023). A temporary halt on the development of large language models (LLMs) ‘more powerful than GPT-4’ (Future of Life Institute Citation2023) is not an isolated call, but part of a historical lineage of similar appeals in the face of emerging technologies, many falling under the concept of Dual Use Research of Concern (DURC). In this paper, we argue that DURC is relevant and applicable to the domain of generative AI, especially in relation to LLMs.

In Section ‘What is Dual Use Research of Concern?’, we introduce the concept of dual-use research, originally conceived in relation to chemical and biological weapons. Despite its roots, we illustrate that DURC fundamentally differs from this early interpretation of dual-use. Between 2005 and 2021, DURC has been employed in areas such as gain-of-function biological research and gene editing, maintaining its relevance as a utilitarian governance approach for potentially high-risk innovative biotechnological research. In Section ‘Applying DURC framework to generative AI and LLMs’, we propose the application of the DURC framework to generative AI, including Large Language Models (LLMs). We redefine the traditional DURC categories, typically used in biological research, to address dual-use concerns specific to generative AI and LLMs. In Section ‘Dual-use and Shared Responsibility’, we examine how DURC frameworks could play a role in establishing the shared responsibility of complex stakeholder networks that are often involved in the development and deployment of AI technologies. In Section ‘Benefits and limitations of the DURC framework’, we explore the advantages and potential limitations of the DURC framework. Applying DURC to generative AI could heighten political awareness of its significant role as a societal transformation force. This symbolic value of DURC should not be underestimated, knowing that scientific research and ambitious technological innovation are not going to be put to a halt. We conclude in Section ‘Specific DURC recommendations for generative AI and LLMs’ by providing a series of specific recommendations for LLM research as an application of the DURC framework.

What is Dual Use Research of Concern?

In some regulatory frameworks (e.g. US Department of Commerce, Citation2023; Regulation (EU) 2021/821 Citation2021), the notions of ‘dual-use’ and ‘misuse’ refer to the interplay between civil and military research. Dual use concerns initially came to the public eye at the time of the Manhattan Project, which led both to nuclear energy production and to the atomic bomb. Already at the time, the dual nature of the implications of their research plagued scientists and raised questions about what we now call ‘open science’ (Schweber Citation2000). In biotechnology, the awareness of the dual use situation began with the recombinant DNA technology in the 1970s. In 1975, the Asilomar conference proposed a moratorium on genetic engineering (Berg et al. Citation1975). As we show below, this short-lived moratorium in biological research is similar to the current proposals for generative AI. In the later years, e.g. during the 2001 anthrax attacks in the United States, the dual-use debate went on to address concerns about potential use of research for bioterrorism (Atlas Citation2002). Once more, this is similar to the current debate on the dual use of AI (Urbina et al. Citation2022).

There exists a more general sense in which technologies have been called ‘dual use’, the widest scope being ‘technologies that could be used for either good or bad purposes’ (Koplin Citation2023). This very broad definition seems to include numerous techniques and devices with multiple uses. For example, a kitchen knife is a necessary tool in everyday life but it also accounts for at least half of knife crimes (Hern, Glazebrook, and Beckett Citation2005). The mere possibility of using a technology for bad purposes is not enough for ethical analysis; one needs to consider the responsibility of the innovator and the imputability of the outcome. In the case of knife crime, it would be very strange to assign the responsibility to the manufacturer of the knife. Thus, the concept of dual use needs to be defined more precisely.

One possible criterion for classifying a technology as dual-use would be to require a demonstrable potential for large-scale harm via malevolent or negligent use of this technology. This introduces the aspects of scale and risk, as opposed to a mere possibility of one-shot effect. For instance, prior to World War I, the deployment of biological agents in warfare was rare and sporadic, typically involving immediate tactical measures such as well poisoning, launching diseased cadavers into besieged cities, or distributing infected blankets (Geissler, Moon, and van Citation1999). The specific worry about the military use of toxins and pathogens emerged later, with research and application of typhoid and anthrax bacilli, cholera vibriones, dysentery bacteria, paratyphoid bacteria, and botulinus toxin, demonstrating the potential for large-scale harm. These biological and toxin weapons have led to the Biological and Toxin Weapons Convention (BTWC) and various verification projects (e.g. VEREX).

The considerations of scope and risk make the notion of dual-use more precise, and they point to specific frameworks of evaluating and addressing the risks of harm as well as responsibilities involved. In this paper, we choose to pursue an analogy between LLMs and an existing dual-use technology with a considerable history of DURC governance.

The current phase of the dual-use debate in biology was initiated in 2005 with the publication of a study reconstructing the 1918 Spanish influenza virus. In a note added in a late-stage revision of this publication, the authors inserted a new statement clarifying the purpose of their research. They stated that their work was driven by ‘historical curiosity’ but also added a pragmatic goal: ‘The fundamental purpose of this work was to provide information critical to protect public health and to develop measures effective against future influenza pandemics’ (Tumpey et al. Citation2005). The belated addition of such a utilitarian objective gave rise to a lively debate and a new wave of the DURC debate in biological research in the early 2000s.

A few years later, a similar scenario unfolded regarding gain-of-function (GOF) research on the H5N1 bird influenza virus. Searching for the genetic signature of pandemic transmission capabilities, scientists altered the H5N1 virus to render it airborne transmissible among mammals (Herfst et al. Citation2012; Imai et al. Citation2012). Once more, they justified their creation of an entirely new and potentially pandemic virus as essential for public health protection – a utilitarian argument that sparked intense debate in academic circles, with many scholars questioning its validity (Evans Citation2013; Lipsitch and Bloom Citation2012; Lipsitch and Galvani Citation2014). Critics voiced concerns about the creation of high and unprecedented risks in the pursuit of scientific knowledge, leading to suggestions and implementation of a research moratorium and publishing restrictions (Collins and Fauci Citation2012). Nonetheless, these publication restrictions only lasted a few months and were lifted after the original article had been revised. The federal funding freeze endured until 2017 and was lifted following the publication of the US regulatory framework on DURC (US HHS Citation2017).

This debate reveals a conflict between two paramount values in responsible research and innovation: the pursuit of knowledge and the safeguarding of public safety. At its core, the concept of DURC encapsulates a utilitarian dilemma, wherein a single scientific endeavor can simultaneously pose significant security threats and yield vital societal benefits. This dilemma is not exclusive to the fields of chemistry or biology; it can extend to other research areas, including artificial intelligence.

Nowadays, the most commonly accepted definition of DURC comes from the Fink Report:

Research that, on the basis of ‘state of the art and knowledge’, could reasonably lead to knowledge, products or technologies that could be directly diverted and/or pose a threat to public health; agriculture, wildlife, flora, the environment and/or national security. (National Research Council Citation2004)

This definition has been adopted in 2007 by the National Security Advisory Board for Biosecurity (NSABB), a body under the National Institutes of Health governing high risk in biological research (NSABB Citation2007). While the Fink Report and the NSABB guidelines are primarily oriented towards life sciences, there is no inherent aspect in the definition that would preclude its application to information technology and AI ().

Table 1. Distinction between the civil-military duality and dual use concerns in research.

Both the ‘civil – military’ dilemma and the ‘high risk – high benefit’ utilitarian dilemma are called ‘dual use’ in the literature. Here, we use the latter concept in line with recent scientific literature on DURC (Korn et al. Citation2019). There is no shortage of discussion of military applications of the life sciences, e.g. for making chemical weapons or bioterrorism. Digital technologies, including AI, have also been widely discussed in the context of civil vs. military application (Sanger Citation2023). There are concerns that AI-fueled cyberspace skirmishes could ‘escalate into conventional warfare’ (Taddeo and Floridi Citation2018), as well as numerous scenarios explored for autonomous weapons and their role in warfare (Christie et al. Citation2023; Scharre Citation2018). While the military applications debate is broad and rich, there is little analysis of DURC aspects of generative AI research, and in particular of LLMs. However, in 2018, the Future of Humanity Institute found that then-current AI technologies expand existing threats, introduces new threats, and changes the type of threats to the digital, physical, and political security of individuals and nations (Brundage et al. Citation2018). Five years later, OpenAI published AI governance commitments that heavily resound with concerns typical for dual-use research. Their first commitment is to safety ‘in areas including misuse, societal risks, and national security concerns, such as bio, cyber, and other safety areas’, while they mention examples of ways in which systems ‘can lower barriers to entry for [bio, chemical, and radiological] weapons development, design, acquisition, or use’ (OpenAI Citation2023a). Similarly, Anthropic made a dedicated study on the possibility of using LLMs for the design of biological weapons, highlighting growing biosecurity risks (Anthropic Citation2023). This discussion of risks can be seen as setting the stage for considering generative AI as dual use.

Applying DURC framework to generative AI and LLMs

In 2022 and 2023, the DURC approach is getting increasingly relevant to generative AI as LLMs are getting easier to build using standardized tools. This is similar to what happened in biology in the years following the publication of the Registry of Standard Biological Parts (Galdzicki et al. Citation2011). This modular approach promoted do-it-yourself construction of biological systems and gave easy access to the tools of synthetic biology. In generative AI, a group of researchers at Stanford recently published a method of obtaining a highly functioning model, known as Stanford Alpaca, which is ‘surprisingly small and easy/cheap to reproduce’ (Taori et al. Citation2023). Another group of researchers managed to bring down the price of fine-tuning LLMs even further (Zhang et al. Citation2023). Alphabet (formerly Google) has published a small LLM, called Gemini Nano (Pichai and Hassabis Citation2023), with the goal of running it on EDGE devices with little computational power.

There is a clear trend indicating that high-performance, finely-tuned LLMs are becoming increasingly available. Similar to the case of modular biological parts, LLMs can now be created inexpensively by a growing number of researchers. This increased accessibility amplifies the potential for misuse, necessitating the regulation of foundation models (Bommasani et al. Citation2022) due to their potential to spread misinformation, manipulate, influence, perpetrate scams, or generate toxic language. Specific social risks have been identified by LLM developers (Weidinger et al. Citation2021). While some of the identified risks can be emergent, i.e. unintended by the initial designers and arising from training, others can stem from poor design, e.g. insufficient mitigation against bias or poor elimination of toxic outputs. Further, output toxicity may be highly context-dependent, e.g. generating a phrase calling someone a dog might be an offense, or not, depending on the circumstances. This dependency on context is an intrinsic characteristic of several types of harmful language, including insults, medical advice, etc. In this sense, LLMs present risks that are highly contextual.

Some areas for potential misuse of LLMs carry a malevolent intention by design, e.g. the user’s intention to make disinformation cheaper and more effective, or an intention to facilitate fraud, scams and more targeted manipulation, or to assist code generation for cyber-attacks, creating weapons, or illegitimate surveillance and censorship (Weidinger et al. Citation2021). These intentions can only come true in virtue of the LLM’s capacity to empower such types of uses, which underscores the LLM potential for misuse. This potential is always present in the model itself, even if the user is unaware of it. Furthermore, an even more concerning aspect of LLMs lies with the ‘unknown unknowns’, i.e. potential misuses that are not foreseen by current foresight and risk evaluation benchmarks (Tamkin et al. Citation2021). This means that while recent efforts to mitigate the risks associated with LLMs have concentrated on specific types of harm, more harmful behavior may emerge from future use of LLMs. Thus we believe that powerful generative AI research, including on LLMs, should be classified under DURC.

Before exploring the parallels between DURC in biological research and generative AI, we address an obvious difference between the two. Harmful biological agents usually exist in nature, whereas AI systems are works of human ingenuity. The former exist autonomously, independent of human influence, whereas the latter can be understood as agents only by projection. One might be tempted to infer that the risks stem from different sources: human intent versus naturally occurring life. However, the case of gain-of-function (GOF) research shows that this distinction is not necessarily valid: the variants of H5N1 virus manufactured for increasing scientific knowledge did not previously exist in nature. These artificially engineered forms of bird flu were intentionally created for research purposes, yet their creation has escalated pandemic risks. This example suggests that the boundary between life and technique does not imply an insurmountable difference between DURC categories in biology and in computer science.

Moreover, LLMs possess unique characteristics that distinguish them from other digital systems in terms of risk. First of all, they are considered a ‘generalist technology’ (Vannuccini and Prytkova Citation2021), so instead of serving one purpose, they can generate an extensive range of outputs, which span across contexts of application. This generalization, combined with their autonomy, can create risks that are far less predictable and more pervasive compared to traditional digital systems. Secondly, LLMs have an exceptional ability to mimic human-like behavior (speech and appearances, in case of multimodal systems) and cognition. This poses unique risks for impersonation and individual exploitation. Lastly, LLMs accrue increasingly important roles in decision-making processes (logistics, healthcare, finance, etc.), which puts immense pressure on their reliability.

Research on artificial agents, be they biological or digital, could be subject to DURC. As part of the definition and regulation of DURC, NSABB identified a set of seven research categories as criteria for DURC (NSABB Citation2007). Although these categories have been devised for the life sciences, we adapt and rephrase them for the use with generative AI ().

Table 2. NSABB dual use categories applied to generative AI, including LLMs.

Based on the application of these criteria, LLMs and generative AI research can be considered as DURC. This categorization, however, does not imply that LLM research should be prohibited or that the benefits of the technology should not be exploited. Like in biology, where GOF research continues to this day, DURC raises awareness of risks and provides guidelines on how to encourage safety when applying LLM research.

Nuclear energy is another technological dual-use area that can be compared to LLMs (Koplin Citation2023). Apocalyptic projections apply in both cases: some estimate that nuclear risks could ‘set civilization back centuries’ (Scouras Citation2019), while others claim that AI brings about existential risks (Bostrom and Yudkowsky Citation2014). However, the parallel between LLMs and nuclear science is more difficult to establish than the parallel with biological research. The biggest gap belongs with the problem of scarcity and accessibility of resources. A motivated and complex government effort is needed for nuclear science to be exploited with potential harm. Materials, such as uranium and plutonium, are rare, heavily regulated, and monitored by international bodies. Meanwhile, LLMs are either open source or can be developed with growing ease by private companies or even by individual researchers, as recent LLaMA-2 models show well (Touvron et al. Citation2023). While nuclear research requires political will at the level of countries, LLMs present a multi-stakeholder dilemma involving individuals, research labs, and business. This makes AI governance a more complex task requiring specific measures not seen in the nuclear sphere.

Dual-use and shared responsibility

The DURC framework is not a panacea but also not a pharmakon: it is a necessary pragmatic step in the governance of LLMs, similarly to what occurred in GOF research. Even if one agrees that LLM research is DURC, it does not by itself require setting a rigid regulatory framework; rather, it may suggest self-regulation measures like voluntary commitments (White House Citation2023b) or industry-wide self-governance bodies (OpenAI Citation2023b). In bioethics, there is a debate on whether a well-developed system of self-regulation could strike a better balance between respecting scientific openness and protecting society from harm. Some argue that it will work provided that scientists engage with the system in good faith (Resnik Citation2010); others oppose it, saying that it might be an overestimation of the scientists’ competences in assessing security risks (Selgelid Citation2007). It remains to be seen if self-regulation can be efficient in the AI sector, not least because manufacturers are private companies which also need to promote their risky products.

One of the crucial functions of a DURC framework is to distribute shared responsibility along a complex chain of stakeholders in the case of malfunctioning or damage. In the case of LLMs, stakeholders include research and innovation actors such as individual researchers and engineering teams, industrial actors such as manufacturers or deployers of AI systems, ecosystem members such as open-source platforms and independent model evaluation groups, governance bodies including regulators and standardization agencies, and final users. To arrive at a working notion of responsibility, criteria need to be established for how it should be shared among the different actors in the value chain, e.g. the programmers who design a foundation model and those who design control layers, the trainers who select training data, the manufacturer of the AI system and that of possible plugins, an intermediary entity using the API supplied by the manufacturer, and the final user (Grinbaum et al. Citation2017). The HHS P3CO framework in biological or chemical research introduced a set of criteria and norms governing shared responsibility (US HHS Citation2017). Similarly, shared responsibility needs to be actively promoted in the AI sector. For example, the proposal for an ‘AI Liability Directive’ from the European Commission establishes the responsibility of the manufacturer for each individual output of an LLM, even if the manufacturer could not possibly intend the system to produce such an outcome (European Commission Citation2022). A careful consideration of DURC should remove the burden of duality from the manufacturer and distribute it along the chain of actors involved in building up the dual or damaging aspect of a particular outcome, rather than the model itself. Similarly to GOF in biotechnology, researchers should not be liable for anything and everything that the system they have built might do in the future. Such indefinite precaution may lead to the ‘infantilization of technology’ (Grinbaum and Groves Citation2013), which continues to treat technologies as direct ‘organs’ or ‘extensions’ of their creators, imputing them with unlimited responsibility. Instead, responsibility should be limited, much like the responsibility of parents for their children: LLMs should not be fixed in a position of eternal childspeak. This would set the stage for more equitable and accountable research from the societal perspective. With an accepted DURC framework, emergent issues of responsibility for the use, misuse, or outcomes of generative AI could be addressed systematically rather than ad hoc.

Benefits and limitations of the DURC framework

There are potential pitfalls to categorizing a research domain as DURC. One such risk is the erosion of open science benefits. Specifically, the implementation of DURC could obstruct or even preclude the online publication of AI models, disrupting the principles of open source and open data (LAION e.V Citation2023). If DURC constraints on the advancement of science become excessively restrictive, then less established or unverified researchers and research teams may struggle to pursue their work in a fully compliant way. Therefore, any proposed DURC framework for AI systems should strive to balance regulatory measures with the promotion of open source and open data in AI model development. While the application of DURC may inevitably curtail certain benefits, e.g. reproducibility, such limitations should be kept to a minimum.

The need to introduce the right amount of limitations and constraints leads to a known problem in DURC, namely excessive formalization and bureaucratization. When the NSABB published the final recommendations, they were adopted in policy documents, e.g. in the US Department of Health and Human Services ‘Framework for Guiding Funding Decisions about Proposed Research Involving Enhanced Potential Pandemic Pathogens’ (Evans Citation2020; US HHS Citation2017), but overall they failed to make a lasting impact on science. One reason for this belongs with the use of language that can hardly, if at all, be implemented on the operational level: the guidelines suggested that scientists should consider scenarios that are ‘credible’, ‘realistic’, or ‘plausible’, while also being ‘highly unlikely but still credible’, based on fourteen different categories of possible damage. These requirements proved to be more suited for regulatory purposes or legal proceedings, than scientific work of biologists. As Evans puts it, ‘ … mandating scientists [to] conduct research only [on] certain issues would be an unjustifiable burden on their freedom (in addition to any utilitarian assertions about the role of scientific freedom in promoting health outcomes)’ (Evans Citation2020). Freedom of research, which is enshrined, e.g. in the German (Art. 5 Absatz 3 GG) or Austrian (Art. 17 StGG) constitutions, can be seen as a permanent counterweight to the utilitarian DURC arguments.

Moreover, as debates on GOF showed, the utilitarian analysis of risks and benefits is not clear-cut and can be manipulated. Some scientists think that ‘gain-of-function research can come in handy’, while others admit that ‘their practical importance wasn’t […] very extraordinary’ (Dance Citation2021). In GOF research, the risks and benefits are subject to expert controversy and cannot be agreed upon consensually. The same applies to LLMs and generative AI models. An established analysis provides twenty-one identified types of harm (Weidinger et al. Citation2021), while the benefits of LLMs are difficult to quantify.

The fact that generative AI models are not designed with a specific purpose further complicates their benefit assessment within the DURC framework. As an example, Sam Altman, the CEO of OpenAI, has admitted to overlooking the ‘problem-solving’ criterion when building a business around a powerful technology that was not developed to provide a specific benefit or solve a particular problem (Mollman Citation2023). However, such limitations of the utilitarian approach are not unprecedented in the context of emerging technologies (Grinbaum and Groves Citation2013). It is important to consider the individual desire and the ambition of groundbreaking scientists. These qualities form an integral part of the virtue ethics analysis, which complements the utilitarian approach by considering the values of a scientist on an individual level.

Another type of limitation of the DURC framework is its focus on short- or, at most, medium-term horizon. Making realistic risk evaluations long into the future is not feasible due to uncertainty. LLMs, however, will have long-term effects on language (Grinbaum et al. Citation2021) which can hardly, if ever, be addressed via governance frameworks. For example, a language always carries, implicitly or explicitly, a particular set of cultural values that express civilizational choices and a particular mode of life. Over time, these values will influence the users of LLMs. Many such effects will remain invisible to the user, but their longer-term influence on language and culture as a whole will eventually emerge and therefore should not be ignored.

Moreover, DURC frameworks usually focus on state-funded research that can be misused to threaten public health or national security. DURC largely relies on policy levers directed towards government-funded research. These levers are, however, not in place for LLMs. Generative AI developers are typically funded independently, often through private corporations, and do not rely heavily on government support. The usual funding-related incentives and policy measures are not applicable to major LLM developers. Furthermore, the global and diffuse nature of LLM development means that regulatory efforts in one jurisdiction might not prevent misuse in another, not to mention the potential for clandestine development that evade law enforcement. On the whole, the implementation of DURC meets here new challenges that indicate a lack of ‘teeth’ in any governance framework due to low entrance barriers and highly international character of generative AI.

Despite all these limitations, DURC has a positive, and sometimes necessary, role to play in research. One major role of the DURC framework is to facilitate the relationship between science and politics. The application of DURC to generative AI would reflect broad political awareness that LLMs are playing a major role in the life of society as a whole. LLMs are a powerful tool that can influence all aspects of life, from private to professional and political, and therefore create risks with far-reaching implications. The symbolic and political value of explicitly treating LLMs as DURC should not be overlooked. The DURC framework would set the stage and rules for reflecting upon, and anticipating, the influence of technology on society while addressing the inevitable conflicts that will arise.

Specific DURC recommendations for generative AI and LLMs

The benefits and limitations of DURC outlined in the previous section demonstrate the need to adapt this framework to generative AI and LLMs. Principles of risk governance should not be too general for DURC to be impactful and operational in generative AI. We suggest the following recommendations.

  • 1a) Develop standardized benchmarks to evaluate foundation models and generative AI systems for intentional abuse

  • 1b) Evaluate foundation models for unintentional harm via ‘red-teaming’ by independent human testers

Foundation models or general-purpose AI (GPAI) models, including openly accessible ones, have recently been included in the European AI Act (European Parliament Citation2023a) as requiring specific compliance measures. This makes them subject to regulation by public authorities. Other provisions of the AI Act only apply to AI systems understood as marketable products. Although the debates continue on which GPAI models should be regulated directly, involved parties agree on the importance of evaluating the AI systems that are accessible to the public (e.g. ChatGPT using GPT-3 or GPT-4). Evaluation is a cornerstone of DURC. For LLMs, it includes a variety of testing techniques (automatically computed benchmarks on standardized datasets, penetration testing, human ‘red teaming’, etc.). A DURC framework should begin with testing and evaluation requirements; the results of such testing should be published alongside the model.

2)

Evaluate datasets and use high-quality data for training

Many emergent risks of LLMs stem from existing bias in training datasets, often replicating or amplifying harmful statistical associations in their training data (Caliskan, Bryson, and Narayanan Citation2017; Qian et al. Citation2022). It is a problem that is particularly striking for historically marginalized groups, languages, and cultures (Field et al. Citation2021). Beyond bias, the epistemic quality of the training data is also an important factor, e.g. training on books vs. online forums leads to the outputs of vastly different quality. LLMs may also have the ability to ‘know what they don’t know’ (Osband et al. Citation2022). Depending on specific applications of the LLMs, they may be given additional requirements of both unfair bias mitigation, and adversarial quality control.

3)

Enforce the human-machine distinction via watermarks

At a societal level, the use of nudging and deception can lend itself to political manipulation (Reisach Citation2021). LLMs can be used to create disinformation at scale (Xu Citation2020). A scalable production of fake content has the potential to create ‘filter bubbles’ or ‘echo chambers’, whereby media consumers rely only on unverified content (Colleoni, Rozza, and Arvidsson Citation2014). Moreover, chatbots can be designed to achieve optimal rankings in recommendation algorithms that supply the content to the end users, emphasizing specific political views. Risks arising from fraud or manipulation, which comprise a large spectrum of societal risks, imply that users should not mistake an output produced by a machine for an output created by a human author (Aaronson Citation2022; Grinbaum and Adomaitis Citation2022b; Kirchenbauer et al. Citation2023). The purpose of algorithmically designed watermarks is to maintain the possibility of distinguishing between machine and human authorship. Yet, the use of watermarking techniques for LLMs should remain unintrusive. The user’s experience, e.g. in obtaining medical or legal advice, should not be perturbed by irrelevant disclaimers in the outputs. We recommend that watermarks be sufficiently hidden from the user but detectable only with a minor effort, as well as sufficiently robust to resist adversarial attempts to blur the origin of the text by editing. Even if watermark efficiency cannot be absolutely guaranteed (H. Zhang et al. Citation2023), the introduction of watermarks is a necessary regulatory step from the societal point of view.

4)

Avoid overpolicing and sterilization of language

Any implemented controls must be proportional to the risks, while unnecessary limitations can distort or impoverish the generated language. The current iterations of LLMs encourage a stale and repetitive use of literary devices (Smith-Ruiu Citation2023) as well as an arbitrary avoidance of critical topics (Weidinger et al. Citation2021). Since humans imitate language abilities of their interlocutors, be they machines or other humans (Grinbaum Citation2023), this creates risks for the cognitive and cultural developments of the individual and of society as a whole. In general, output filtering is useful, for example generalist chatbots should not provide medical or legal advice. However, this has collateral effects, namely, excluding all types of ‘toxic’ language results in a sterilized language. Human users may find that their own language gets less esthetically pleasing and more banal as a result of their interaction with chatbots.

5)

Introduce a set of criteria and norms governing shared responsibility

LLMs and LLM-based chatbots are machines, yet they acquire qualities by projection (Grinbaum and Adomaitis Citation2022a). Their emergent ‘conduct’ may lead to morally significant consequences. For example, chatbots can be perceived as lying, misleading, hurting, manipulating or insulting human beings (Davis Citation2016). Such effects usually induce respective projections of moral responsibility on machines. However, digital agents are not moral agents and cannot assume responsibility. A chatbot should never be perceived by the user as a responsible person, even by projection (Grinbaum Citation2019). To remove such projections, a comprehensive set of criteria and norms that clearly outline the shared responsibilities between AI developers, users, and other stakeholders should be created (Adomaitis, Grinbaum, and Lenzi Citation2022; Dignum Citation2019). The DURC framework can help to phrase such criteria in a language understandable to legal professionals and non-experts in general. Shared responsibility should emphasize the collective dimension of design and deployment of AI systems.

In conclusion, the application of the Dual Use Research of Concern (DURC) framework to the field of generative AI and Large Language Models (LLMs) brings a new perspective on the rapidly expanding influence of these technologies. It serves as a call for reflective governance allowing researchers, policymakers, and the public to conscientiously address the broad-reaching implications of LLMs. The recommendations provided in this article are meant to offer tangible starting points for shaping the ethical trajectory of generative AI research, thereby ensuring that the balance between innovation and security is maintained. It is our hope that this perspective will stimulate further dialogue, leading to the emergence of robust strategies that uphold the integrity and potential of AI, while safeguarding societal interests.

Authors contribution

All authors contributed to the study conception and design. Sections ‘Introductions’ and ‘Dual-use and Shared Responsibility’ were written with equal contribution. Sections ‘What is Dual Use Research of Concern?’ and ‘Applying DURC framework to generative AI and LLMs’ were conceived by Adomaitis and later edited by Grinbaum. Sections ‘Benefits and limitations of the DURC framework’ and ‘Specific DURC recommendations for generative AI and LLMs’ were conceived by Grinbaum and later edited by Adomaitis. All authors read and approved the final manuscript.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This research was supported through projects TechEthos (grant number 101006249) and MultiRATE (grant number 101073929) funded by the European Commission Horizon program.

Notes on contributors

Alexei Grinbaum

Alexei Grinbaum is senior research scientist at CEA-Saclay with a background in quantum information theory. He writes on ethical questions of emerging technologies, including robotics and AI. In AI ethics, his scientific interest is in the field of watermarking language outputs of generative AI. Grinbaum is the chair of the CEA Operational Ethics Committee for Digital Technologies and member of the French National Digital Ethics Committee (CNPEN). At CNPEN, he was co-rapporteur of Opinion 3 'Ethical issues of conversational agents' and Opinion 7 'Generative AI systems: ethical issues'. He contributes to several EU projects on technology ethics, including AI ethics-by-design and Ethics Readiness Levels procedures, and serves as ethics expert to the European Commission. His books include 'Mécanique des étreintes' (2014), 'Les robots et le mal' (2019), and 'Parole de machines' (2023).

Laurynas Adomaitis

Laurynas Adomaitis is a postdoctoral researcher at CEA within the H-2020 and Horizon Europe projects TechEthos and MultiRATE. Initially educated in Lithuania, he graduated with a PhD in Philosophy from Scuola Normale Superiore di Pisa. He is currently working on the ethical analysis and ethical readiness of high socio-economic impact technologies, with a special focus on AI and digital technologies. He also teaches ethics at Institut d'optique Graduate School, CentraleSupélec, and Emlyon Business School. Before joining CEA, he worked as an AI Ethics Researcher at Nord Security, a leading internet privacy and cybersecurity company. In 2020, he completed the Regulation of AI and Robotics in Europe certificate at the EURA Centre of Excellence. His publications focus on ethical and conceptual issues imposed by the emergence of new technologies, especially AI and virtual reality.

References