This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bridging Deliberative Democracy and Deployment of Societal-Scale Technology

Ned Cooper [email protected] 0000-0003-1834-279X College of Engineering, Computing & Cybernetics, The Australian National UniversityCanberraAustralia
(2018)
copyright: nonecopyright: acmcopyrightjournalyear: 2018doi: XXXXXXX.XXXXXXX

This position paper encourages the Human-Computer Interaction (HCI) community to focus on designing deliberative processes to inform and coordinate technology and policy design for large language models (LLMs)—a ‘societal-scale technology’. First, I propose a definition for societal-scale technology and locate LLMs within this definition. Next, I argue that existing processes to ensure the safety of LLMs are insufficient and do not give the systems democratic legitimacy. Instead, we require processes of deliberation amongst users and other stakeholders on questions about the safety of outputs and deployment contexts. This shift in AI safety research and practice will require the design of corporate and public policies that determine how to enact deliberation and the design of interfaces and technical features to translate the outcomes of deliberation into technical development processes. To conclude, I propose roles for the HCI community to ensure deliberative processes inform technology and policy design for LLMs and other societal-scale technology.

1. Societal-Scale Technology

In this position paper, I define ‘societal-scale technology’ as an artefact or system that is:

  • developed based on interactions with society, or

  • impacts society once deployed.

I define society in this position paper as groupings of people across political, economic, geographical, and cultural boundaries. For example, I consider LLMs to be societal-scale technology, as LLMs interact with training data that represent groupings of people across such boundaries, and, at least in recent deployment cases (e.g., the ChatGPT research preview), LLMs have impacted groups of people across such boundaries. Within such societies, ChatGPT has impacted direct stakeholders (those who interact directly with ChatGPT, such as school students) and indirect stakeholders (those who may or may not have interacted directly with ChatGPT but are affected by the interaction of direct stakeholders with ChatGPT, such as school teachers) (Friedman and Hendry, 2019).

2. Processes of Deliberation for LLMs

While LLMs may interact with representations of societies during development in the form of training datasets, the group of people actively developing LLMs rarely, if ever, reflect the boundary-spanning nature of the societies in which LLMs are deployed. For example, the workforce of OpenAI is not (and could never be) as diverse as the societies in which they deployed ChatGPT. The values and preferences of those groups of people actively developing an LLM are reflected in any system they develop, regardless of the representation of diverse values and preferences in a training dataset—through the specification of system behaviour once deployed. In the case of ChatGPT, for example, this was achieved through content filters, among other mechanisms. The content filters specified specific values and preferences that OpenAI intended ChatGPT to reflect. However, I contend that the specification of values and preferences by one group of developers is insufficient to ensure the safety of LLMs and does not give such systems democratic legitimacy. A small group of expert developers cannot adequately foresee the safety risks of technology deployed across societies—composed of multiple professions and stakeholders—let alone speak for those stakeholders during development and deployment (Joyce et al., 2021).

Instead, I encourage those considering safe development and deployment strategies for LLMs to review the emphasis of political science literature over recent decades on deliberation as the essence of democratic legitimacy (Fishkin, 1991; Curato et al., 2017; Knight and Johnson, 1994). Deliberative democracy involves an association of members who govern their affairs by deliberation among the members (Cohen, 2005). In other words, deliberative democracy is about making collective decisions determined by those subject to the decisions: not only their preferences, interests, and choices but also their reasoning (Elster and Przeworski, 1998). If algorithmic fairness is primarily a political task, as Wong (2020) argues, rather than solely a technical task, we must consider how to resolve issues politically rather than technically. In the societal-scale context of LLMs, instead of groups of developers within individual organisations resolving such political questions for us, I argue that we must design processes that provide people using a system (i.e., direct stakeholders) or people affected by a system (i.e., indirect stakeholders) with the opportunity to deliberate with others (including developers) on how the system functions, and in what contexts to deploy those systems.

3. Roles for HCI in Deliberative AI

OpenAI accompanied the release of ChatGPT with a request for feedback from users on the outputs of the system to improve its safety. As stated in the blog post announcing the release of ChatGPT:

”We are interested in feedback regarding harmful outputs that could occur in real-world, non-adversarial conditions, as well as feedback that helps us uncover and understand novel risks and possible mitigations.” (OpenAI, 2022)

Yet, the scope of feedback accepted through the ChatGPT interface is severely limited—to indications of approval or disapproval of an individual output of the system and some indication by the user of what an ‘ideal’ answer might have been. This feedback process remains tightly controlled by those deploying the system and focuses on aligning outputs to individual user preferences. It does not facilitate deliberation amongst stakeholders regarding the proper outputs of LLMs nor the appropriate deployment contexts for LLMs.

HCI researchers and practitioners are well-placed to build deliberative capacity into AI safety research and practice—the HCI community takes users’ preferences seriously and interrogates the methods through which we elicit user preferences. To this end, I envisage three roles for the HCI community for the safe development and deployment of LLMs:

  1. (1)

    Designing corporate and public policies that define criteria for and conditions of membership of associations for deliberation

  2. (2)

    Developing platforms or interfaces that facilitate bidirectional communication among developers and members of associations, and multi-directional communication amongst the members of an association

  3. (3)

    Designing processes for documentation of collective decisions, and research on how to link documented decisions to technical development processes (e.g., how could the processes of Reinforcement Learning from Human Feedback expand to include deliberation amongst stakeholders or to facilitate feedback from collectives of users?).

References

  • (1)
  • Cohen (2005) Joshua Cohen. 2005. Deliberation and Democratic Legitimacy. In Debates in contemporary political philosophy. Routledge, 352–370.
  • Curato et al. (2017) Nicole Curato, John S Dryzek, Selen A Ercan, Carolyn M Hendriks, and Simon Niemeyer. 2017. Twelve Key Findings in Deliberative Democracy Research. Daedalus 146, 3 (2017), 28–38.
  • Elster and Przeworski (1998) Jon Elster and Adam Przeworski. 1998. Deliberative Democracy. Vol. 1. Cambridge University Press.
  • Fishkin (1991) James S Fishkin. 1991. Democracy and Deliberation: New Directions for Democratic Reform. Yale University Press.
  • Friedman and Hendry (2019) Batya Friedman and David G Hendry. 2019. Value Sensitive Design: Shaping Technology with Moral Imagination. Mit Press.
  • Joyce et al. (2021) Kelly Joyce, Laurel Smith-Doerr, Sharla Alegria, Susan Bell, Taylor Cruz, Steve G Hoffman, Safiya Umoja Noble, and Benjamin Shestakofsky. 2021. Toward a Sociology of Artificial Intelligence: A Call for Research on Inequalities and Structural Change. Socius 7 (2021). https://doi.org/10.1177/2378023121999581
  • Knight and Johnson (1994) Jack Knight and James Johnson. 1994. Aggregation and Deliberation: On the Possibility of Democratic Legitimacy. Political theory 22, 2 (1994), 277–296.
  • OpenAI (2022) OpenAI. 2022. ChatGPT: Optimizing language models for dialogue. https://openai.com/blog/chatgpt/
  • Wong (2020) Pak-Hang Wong. 2020. Democratizing Algorithmic Fairness. Philosophy & Technology 33 (2020), 225–244.