When and Why is Persuasion Hard? A Computational Complexity Result
Abstract
As generative foundation models improve, they also tend to become more persuasive, raising concerns that AI automation will enable governments, firms, and other actors to manipulate beliefs with unprecedented scale and effectiveness at virtually no cost. The full economic and social ramifications of this trend have been difficult to foresee, however, given that we currently lack a complete theoretical understanding of why persuasion is costly for human labor to produce in the first place. This paper places human and AI agents on a common conceptual footing by formalizing informational persuasion as a mathematical decision problem and characterizing its computational complexity. A novel proof establishes that persuasive messages are challenging to discover (NP-Hard) but easy to adopt if supplied by others (NP). This asymmetry helps explain why people are susceptible to persuasion, even in contexts where all relevant information is publicly available. The result also illuminates why litigation, strategic communication, and other persuasion-oriented activities have historically been so human capital intensive, and it provides a new theoretical basis for studying how AI will impact various industries.
Introduction
Advocates argue that artificial intelligence will simultaneously accelerate the pace of scientific discovery and make existing knowledge more accessible (e.g., by translating medical information into people’s native languages; Vieira et al., 2021). However, the generality of artificial intelligence raises dual-use concerns: similar to how a predictive model intended to avoid drug toxicity can be “inverted” to increase the potency of chemical weapons (Urbina et al., 2022), systems that are capable of explanation and creativity will also generally be potent instruments of manipulation and disinformation. A comprehensive assessment of the technology requires that we anticipate both its potential epistemic costs and benefits, ideally maximizing the latter while minimizing the former.
Recent work has demonstrated that large language models are highly effective at persuading people across a variety of tasks and domains (Matz et al., 2024; Durmus et al., 2024; Burtell and Woodside, 2023; Shin and Kim, 2023; Ahn et al., 2021; Karinshak et al., 2023; Carrasco-Farre, 2024; Breum et al., 2023), raising concerns about their potential use for widespread misinformation, manipulation, and deception (Allen and Weyl, 2024; Kreps et al., 2022; Zellers et al., 2019). These concerns are especially acute for people whose identifiable characteristics—such as race, gender, or sexual identity—subject them to higher rates of algorithmic persuasion and bias (Bar-Gill et al., 2023; Speicher et al., 2018; Hannak et al., 2014; Mikians et al., 2012).
Large language models and other innovations have significantly enhanced the persuasive capacities of machines in recent years, but humans have been refining their powers of persuasion since antiquity (Aristotle’s Rhetoric dates from the 4th century BCE). The large-scale manufacture of persuasion is, perhaps unsurprisingly, a well-developed fixture of modern economies. One frequently-cited estimate holds that “one-quarter of GDP is persuasion” (McCloskey and Klamer, 1995): law, advertising, politics, science, public relations, and many other professions revolve—in whole or in part—around changing other people’s minds. The advent of targeted digital advertising has, moreover, expanded the specificity and scope of persuasion (Wu, 2017; Zuboff, 2019).
The economics of persuasion raise three important questions: (1) what does persuasion accomplish to justify such enormous expenditures; (2) why has persuasion historically commanded so much brain power; and (3) how will AI automation reshape persuasion and, with it, the broader epistemic landscape of society?
A variety of authors have advanced theories of the benefits of persuasion—i.e., what persuaders can achieve in various circumstances (e.g., Milgrom and Roberts, 1986; Grossman and Hart, 1980; Crawford and Sobel, 1982; Kamenica and Gentzkow, 2011; Schwartzstein and Sunderam, 2021; Aina, 2021). Relatively few, however, have focused on the cost of generating persuasive messages. In practice, such costs play an important role in determining when, where, and how persuasion takes place. The purpose of this paper is to formally establish a key driver of these costs—namely, the computational resources (natural or artificial) required to generate persuasive messages.
Given a suitable formalization of the persuasion problem, computational complexity theory can be used to characterize its production function. This approach has previously been applied to explain a variety of economic and social phenomena, such as incomplete uptake of public information (Aragones et al., 2005), slow convergence to Nash equilibrium (Daskalakis et al., 2009), narrow choice bracketing (Camara, 2022), and persistent market inefficiencies (Spear, 1989). Closest to the present paper is that of Dughmi and Xu (2016), who characterize the complexity of the model introduced by Kamenica and Gentzkow (2011).
Informational Persuasion
“1.11 The world is determined by the facts, and by these being all the facts.” (emphasis added)
—Ludwig Wittgenstein, 1922
The present paper studies informational persuasion: the selective disclosure of private information or the use of models (narratives, stories, etc.) to “frame” public information with the intention of increasing a counter-party’s belief in a focal claim. This formalization of persuasion is highly general in that it assumes little structure beyond the standard probability axioms and naturally captures a wide variety of applications. It can be cast as a formal decision problem as follows.
Definition 1 (Informational Persuasion).
Let a probability space , focal event , set of facts , and threshold of belief be given. Is there a subset of the facts that induces the receiver to raise their belief in the focal event above the threshold, ?
Although some instances of data withholding are nefarious (e.g., a scientist omitting disconfirmatory evidence from their report), attention and other constraints mean that informational persuasion is a normalized feature of many interactions (e.g., “please submit no more than three letters of recommendation”). The use of models (narratives, stories, etc.) to characterize publicly available information is a frequent practice among lawyers, politicians, academics, and other professional persuaders, as has been documented extensively elsewhere (Shiller, 2017; Spiegler, 2016; Eliaz and Spiegler, 2020; Roos and Reccius, 2021; Andre et al., 2023; Bénabou et al., 2018; Graeber et al., 2024). Informational persuasion is isomorphic to the following problem, which asks whether a model can be used to “frame” public information (see, e.g., Schwartzstein and Sunderam, 2021).
Definition 2 (Model Selection).
Let an informational persuasion problem be given. Define the space of models , each of which induces a liklihood function over the focal event as . Let a prior be given. Is there a model such that ?
Our main result, Theorem 1, shows that informational persuasion is NP-Complete. Problems in this complexity class share two principle features: proposed solutions can be quickly verified but they cannot, in general, be quickly discovered.111The idea that this class of problems cannot be solved quickly (i.e., in polynomial time) is the famous conjecture. Although open, it is widely believed to be true among computer scientists (Gasarch, 2019) on account of how strange the implications of its converse would be, among other reasons (Aaronson, 2016). Informational persuasion is therefore structurally similar to mathematical proof in the sense that solutions are categorically easier to check than to construct. Indeed, proving theorems in an axiomatic system such as Zermelo-Frankel is likewise NP-Complete, which implies that a fast algorithm for informational persuasion, were it to exist, could be be used to quickly prove arbitrary theorems, and vice-versa.
A central puzzles in the persuasion literature is why people are susceptible to persuasion in the first place, especially in contexts where all information is publicly available? More specifically, why do people accept externally supplied persuasion rather than coming up with their own models, narratives, and stories?
Many existing theories do not address this question, but rather directly assume that people are persuadable. The proof of Theorem 1 provides some indication of why this might be the case. The essential insight is that, in unrestricted inferential contexts, each new fact casts every other fact in a new light, shifting the credence it lends to the focal event. Translating this idea into Kolmogorov’s set-theoretic model of probability theory reveals that informational persuasion is, in the worst case, a highly non-convex optimization problem.
Formal Result
Let be a probability space. We study communication between a sender and receiver . The sender has access to a collection of private facts for , each of which is a subset of . The sender selectively reports a subset of the facts to the receiver. The receiver updates beliefs on the basis of the newly reported facts.
There is a focal event . The sender is concerned with maximizing the receiver’s posterior belief in . Let denote the sender’s expectation of the receiver’s posterior conditional on a report .
Theorem 1.
Informational persuasion is NP-Complete.
Proof.
We show that persuasion is both NP and NP-Hard.
NP: A valid report constitutes a certificate for the informational persuasion decision problem. To verify such a report, one evaluates
(1) |
Set intersection, function lookup (to evaluate ), and arithmetic operations can all be implemented in linear time by a deterministic Turing machine. Hence, informational persuasion is in NP.
NP-Hard: We prove that informational persuasion is NP-Hard using a reduction of the exact cover problem, which is known to be NP-Complete (Garey and Johnson, 1979). The exact cover problem can be stated as follows: For a set and collection of subsets such that , is there a collection such that for all and ? In other words, within a given cover, can we find an exact cover?
Let a set and cover be given as above. Create a duplicate copy of each and denote them and , respectively. Construct an event space , a -algebra , and prior where for all . Furthermore, let , , and .
Now, , , , and define an informational persuasion problem. We claim that the this decision problem has the same truth-value as the original exact cover problem. Note that, per our construction, the informational persuasion evaluates to TRUE if and only if there exists a subset such that . Let denote the set of indices included in . Per the definitions of and , this condition is equivalent to . But this is the same as saying that
(2) |
which, is, in turn, is equivalent to two statements: first, that , i.e., the set is disjoint; and second, that
(3) |
which is equivalent, in turn, to the assertion that . But and are just copies of and , so this implies that is also a cover of . Hence is an exact cover of . In other words, an index solves the informational persuasion problem if and only if it solves the corresponding exact cover problem. Hence, informational persuasion is in NP-Hard.
∎
Discussion
Rational theories of belief formation (such as those typically assumed in statistical decision theory and economics) assume that people instantaneously update their internal state of belief to reflect the sum total of information they possess.222Attention is generally conceived of as a “bottleneck” interposed between an economic agent and the external world (Loewenstein and Wojtowicz, 2023). Even models which incorporate attentional frictions typically assume that people update instantaneously once information makes it through the bottleneck, however. The ramifications of even readily available information are not always apparent, however: it takes time to notice patterns, generate good explanations, rule out bad explanations, separate signal from noise, and generally transform the “raw material” of information into the “finished products” of knowledge and understanding.
Characterizing the computational complexity of informational persuasion directly informs our understanding of this process by characterizing the production technology of informational persuasion. Theorem 1 shows that the computational resource cost of informational persuasion can, in the worst case, grow exponentially in the number of facts considered. This helps explain why industries centered around persuasion, such as those identified by (McCloskey and Klamer, 1995), are human-capital intensive. It also informs conversations about how artificial intelligence, social media, and other information technologies will impact our shared “epistemic commons.”
Not all persuasion problems are equally difficult, however. There are clearly situations where it is easy to “cherry-pick” examples to support a particular conclusion—for example, if each “fact” is a Gaussian draw and the focal belief is a one-tailed hypothesis test concerning the distribution’s mean. Although left to future work, the present approach frames informational persuasion directly in terms of the Kolmogorov probability axioms, and therefore naturally bridges persuasion with statistical theory. The simplicity of persuasion in the above example, for example, follows directly from statistical sufficiency.
References
- Aaronson (2016) Scott Aaronson. P np. Open problems in mathematics, pages 1–122, 2016.
- Ahn et al. (2021) Jungyong Ahn, Jungwon Kim, and Yongjun Sung. Ai-powered recommendations: the roles of perceived similarity and psychological distance on persuasion. International Journal of Advertising, 40(8):1366–1384, 2021.
- Aina (2021) Chiara Aina. Tailored stories. Technical report, Mimeo, 2021.
- Allen and Weyl (2024) Danielle Allen and E Glen Weyl. The real dangers of generative ai. Journal of Democracy, 35(1):147–162, 2024.
- Andre et al. (2023) Peter Andre, Ingar Haaland, Christopher Roth, and Johannes Wohlfart. Narratives about the macroeconomy. 2023.
- Aragones et al. (2005) Enriqueta Aragones, Itzhak Gilboa, Andrew Postlewaite, and David Schmeidler. Fact-free learning. American Economic Review, 95(5):1355–1368, 2005.
- Bar-Gill et al. (2023) Oren Bar-Gill, Cass R Sunstein, and Inbal Talgam-Cohen. Algorithmic harm in consumer markets. Journal of Legal Analysis, 15(1):1–47, 2023.
- Bénabou et al. (2018) Roland Bénabou, Armin Falk, and Jean Tirole. Narratives, imperatives, and moral reasoning. Technical report, National Bureau of Economic Research, 2018.
- Breum et al. (2023) Simon Martin Breum, Daniel Vædele Egdal, Victor Gram Mortensen, Anders Giovanni Møller, and Luca Maria Aiello. The persuasive power of large language models. arXiv preprint arXiv:2312.15523, 2023.
- Burtell and Woodside (2023) Matthew Burtell and Thomas Woodside. Artificial influence: An analysis of ai-driven persuasion. arXiv preprint arXiv:2303.08721, 2023.
- Camara (2022) Modibo K Camara. Computationally tractable choice. In EC, page 28, 2022.
- Carrasco-Farre (2024) Carlos Carrasco-Farre. Large language models are as persuasive as humans, but why? about the cognitive effort and moral-emotional language of llm arguments. arXiv preprint arXiv:2404.09329, 2024.
- Crawford and Sobel (1982) Vincent P Crawford and Joel Sobel. Strategic information transmission. Econometrica: Journal of the Econometric Society, pages 1431–1451, 1982.
- Daskalakis et al. (2009) Constantinos Daskalakis, Paul W Goldberg, and Christos H Papadimitriou. The complexity of computing a nash equilibrium. Communications of the ACM, 52(2):89–97, 2009.
- Dughmi and Xu (2016) Shaddin Dughmi and Haifeng Xu. Algorithmic bayesian persuasion. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 412–425, 2016.
- Durmus et al. (2024) Esin Durmus, Liane Lovitt, Alex Tamkin, Stuart Ritchie, Jack Clark, and Deep Ganguli. Measuring the persuasiveness of language models. 2024. URL https://www.anthropic.com/news/measuring-model-persuasiveness.
- Eliaz and Spiegler (2020) Kfir Eliaz and Ran Spiegler. A model of competing narratives. American Economic Review, 110(12):3786–3816, 2020.
- Garey and Johnson (1979) Michael R Garey and David S Johnson. Computers and Intractability: A Guide to the Theory of NP-Completeness. 1979.
- Gasarch (2019) William I Gasarch. Guest column: The third p=? np poll. ACM SIGACT News, 50(1):38–59, 2019.
- Graeber et al. (2024) Thomas Graeber, Christopher Roth, and Florian Zimmermann. Stories, statistics, and memory. The Quarterly Journal of Economics, page qjae020, 2024.
- Grossman and Hart (1980) Sanford J Grossman and Oliver D Hart. Disclosure laws and takeover bids. The Journal of Finance, 35(2):323–334, 1980.
- Hannak et al. (2014) Aniko Hannak, Gary Soeller, David Lazer, Alan Mislove, and Christo Wilson. Measuring price discrimination and steering on e-commerce web sites. In Proceedings of the 2014 conference on internet measurement conference, pages 305–318, 2014.
- Kamenica and Gentzkow (2011) Emir Kamenica and Matthew Gentzkow. Bayesian persuasion. American Economic Review, 101(6):2590–2615, 2011.
- Karinshak et al. (2023) Elise Karinshak, Sunny Xun Liu, Joon Sung Park, and Jeffrey T Hancock. Working with ai to persuade: Examining a large language model’s ability to generate pro-vaccination messages. Proceedings of the ACM on Human-Computer Interaction, 7(CSCW1):1–29, 2023.
- Kreps et al. (2022) Sarah Kreps, R Miles McCain, and Miles Brundage. All the news that’s fit to fabricate: Ai-generated text as a tool of media misinformation. Journal of experimental political science, 9(1):104–117, 2022.
- Loewenstein and Wojtowicz (2023) George Loewenstein and Zachary Wojtowicz. The economics of attention. Available at SSRN 4368304, 2023.
- Matz et al. (2024) SC Matz, JD Teeny, Sumer S Vaid, H Peters, GM Harari, and M Cerf. The potential of generative ai for personalized persuasion at scale. Scientific Reports, 14(1):4692, 2024.
- McCloskey and Klamer (1995) Donald McCloskey and Arjo Klamer. One quarter of gdp is persuasion. The American Economic Review, 85(2):191–195, 1995.
- Mikians et al. (2012) Jakub Mikians, László Gyarmati, Vijay Erramilli, and Nikolaos Laoutaris. Detecting price and search discrimination on the internet. In Proceedings of the 11th ACM workshop on hot topics in networks, pages 79–84, 2012.
- Milgrom and Roberts (1986) Paul Milgrom and John Roberts. Relying on the information of interested parties. The RAND Journal of Economics, pages 18–32, 1986.
- Roos and Reccius (2021) Michael Roos and Matthias Reccius. Narratives in economics. Journal of Economic Surveys, 2021.
- Schwartzstein and Sunderam (2021) Joshua Schwartzstein and Adi Sunderam. Using models to persuade. American Economic Review, 111(1):276–323, 2021.
- Shiller (2017) Robert J Shiller. Narrative economics. American economic review, 107(4):967–1004, 2017.
- Shin and Kim (2023) Minkyu Shin and Jin Kim. Enhancing human persuasion with large language models. arXiv preprint arXiv:2311.16466, 2023.
- Spear (1989) Stephen E Spear. Learning rational expectations under computability constraints. Econometrica: Journal of the Econometric Society, pages 889–910, 1989.
- Speicher et al. (2018) Till Speicher, Muhammad Ali, Giridhari Venkatadri, Filipe Nunes Ribeiro, George Arvanitakis, Fabrício Benevenuto, Krishna P Gummadi, Patrick Loiseau, and Alan Mislove. Potential for discrimination in online targeted advertising. In Conference on fairness, accountability and transparency, pages 5–19. PMLR, 2018.
- Spiegler (2016) Ran Spiegler. Bayesian networks and boundedly rational expectations. The Quarterly Journal of Economics, 131(3):1243–1290, 2016.
- Urbina et al. (2022) Fabio Urbina, Filippa Lentzos, Cédric Invernizzi, and Sean Ekins. Dual use of artificial-intelligence-powered drug discovery. Nature Machine Intelligence, 4(3):189–191, 2022.
- Vieira et al. (2021) Lucas Nunes Vieira, Minako O’Hagan, and Carol O’Sullivan. Understanding the societal impacts of machine translation: a critical review of the literature on medical and legal use cases. Information, Communication & Society, 24(11):1515–1532, 2021.
- Wu (2017) Tim Wu. The attention merchants: The epic scramble to get inside our heads. Vintage, 2017.
- Zellers et al. (2019) Rowan Zellers, Ari Holtzman, Hannah Rashkin, Yonatan Bisk, Ali Farhadi, Franziska Roesner, and Yejin Choi. Defending against neural fake news. Advances in neural information processing systems, 32, 2019.
- Zuboff (2019) Shoshana Zuboff. The Age of Surveillance Capitalism: The Fight for a Human Future at the New Frontier of Power. Routledge, 2019.