User Privacy Harms and Risks in Conversational AI:
A Proposed Framework
Abstract
This study presents a unique framework that applies and extends Solove (2006)’s taxonomy to address privacy concerns in interactions with text-based AI chatbots. As chatbot prevalence grows, concerns about user privacy have heightened. While existing literature highlights design elements compromising privacy, a comprehensive framework is lacking. Through semi-structured interviews with 13 participants interacting with two AI chatbots, this study identifies 9 privacy harms and 9 privacy risks in text-based interactions. Using a grounded theory approach for interview and chatlog analysis, the framework examines privacy implications at various interaction stages. The aim is to offer developers, policymakers, and researchers a tool for responsible and secure implementation of conversational AI, filling the existing gap in addressing privacy issues associated with text-based AI chatbots.
1 Introduction
ChatGPT recently emerged as a prominent chatbot, built upon a large language model (LLM) rooted in Reinforcement Learning from Human Feedback and designed to generate conversational outputs [114]. Conversational text-based Artificial Intelligence (AI) chatbots, exemplified by ChatGPT, engage users in natural language conversations. They leverage advanced natural language models to interpret user inputs, generate contextually relevant responses, and maintain coherence throughout multi-turn interactions [39, 29].
Through reinforcement learning and collecting massive data from human prompts and interactions, these chatbots continually improve their performance and adaptability over time which raises user privacy concerns, given the potential for data misuse during these conversational exchanges. Moreover, as conversational chatbots have evolved, they have increasingly gathered substantial amounts of personal information without transparent user consent. This practice has resulted in privacy and security risks [31, 55, 57, 26]. These concerns manifest at various stages, both before and after the interaction [4, 55]. The information collected during these exchanges is vulnerable to unauthorized access, data breaches, and potential misuse, posing adverse consequences for users.
Existing literature addresses privacy concerns related to conversational text-based AI chatbots, exploring how design elements such as high empathy [2], emotional engagement [37, 78, 39, 82], trust [33, 44, 73, 66, 57, 63], human-likeness [103, 44, 66], personalization [96], and self-disclosure [69, 26] can potentially compromise user privacy.
Despite ongoing efforts to align with evolving data protection regulations in AI, including the European Union (EU)’s General Data Protection Regulation (GDPR) [67] and (Draft) EU’s AI Act [112]111As of February 2, 2024, EU countries have agreed on the final text of the EU AI Act [15]., there is a significant gap in comprehensive laws and regulations governing these interfaces in both the United States and the EU. This gap underscores the pressing need to scrutinize the privacy implications associated with user interactions. Furthermore, a comprehensive framework specifically addressing privacy concerns in conversational text-based AI chatbots is currently lacking. This research aims to fill these gaps by exploring the following question:
What are user privacy harms and risks that arise from interactions with conversational text-based AI chatbots?
2 Background and Related Work
2.1 Conversational Text-based AI Chatbots
Conversational AI has progressed into highly refined entities, integrating advancements in Natural Language Processing (NLP), neural networks, and deep learning technologies. These entities can interact through text and/or speech, complemented by visual elements, virtual gestures, and, in some instances, haptic-assisted physical gestures [62]. Unlike speech-based interactions, conversational text-based chatbots engage users primarily through written messages within platforms, websites, or mobile applications. Some have attained the status of Intelligent Personal Assistants (IPAs) by being trained to execute specific tasks [30]. Prominent examples include IBM Watson, Apple Siri, Samsung Bixby, and Amazon Alexa [8]. Følstad et al. [39] highlight that these chatbots leverage algorithms, machine learning (ML) techniques, and sentiment analysis to continually enhance their grasp of language nuances and user preferences, providing an increasingly immersive and valuable conversational experience over time. Furthermore, NLP technologies empower these chatbots to gather, analyze, and comprehend the semantics, context, and intent embedded in user text, generating responses that are more suitable and contextually relevant [97, 58, 2].
The impact of these chatbots extends significantly across various sectors such as customer support, healthcare, education, and e-commerce, meeting the rising demand for personalized and efficient digital interactions [97, 86, 54]. Moreover, they deliver personalized, amicable interactions (e.g., [23]) that efficiently address user inquiries, provide information, give recommendations, and execute tasks based on the ongoing conversation’s context [16, 89]. Crucially, these chatbots must comprehend diverse user expectations, preferences, and cultural backgrounds [31, 6, 46, 26, 27]. Hill et al. [52] argued that even with concise language, conversational chatbots can understand users and offer satisfaction.
2.2 Privacy Harms and Risks
While effectively employing these tools undeniably improves user engagement and streamlines communication processes [115, 39], it is imperative to scrutinize the purpose behind the development of conversational text-based AI chatbots. The data collection capabilities of chatbots, beneficial for customization and enhancing user experiences, introduce significant concerns related to privacy and security [1, 19]. Striking a balance between the imperative need for data collection to enrich user interactions and robust privacy and security measures is important. This shift from discussing the utility of chatbots to addressing associated privacy risks emphasizes the necessity for a thoughtful and responsible approach to chatbot development.
Privacy revolves around the appropriate flow of personal information, including individuals’ rights to control those flows, to ensure proper use and safeguarding against unauthorized access [64]. It encompasses ethical and legal considerations, emphasizing transparency, consent, and the responsible handling of personal data [25]. While privacy focuses on individual rights and legal compliance, security addresses the comprehensive protection of data and systems from various risks – both are essential components in the landscape of responsible information management [53]. This distinction gives rise to two categories: privacy harms and privacy risks.
2.2.1 Privacy Harms
Privacy harms refer to the negative consequences or adverse effects that individuals may experience due to the violation or compromise of their privacy rights [20]. Privacy research literature is nuanced, rather than offering a consensus conceptualization. Prosser’s [85] seminal work on privacy established a framework comprising four torts –Intrusion upon seclusion, Appropriation of likeness, Public disclosure of private facts, and False light publicity. Prosser focused on the right to be let alone and safeguarding personal privacy from intrusion and publicity. Westin [111] expanded on this foundation by proposing four states of privacy – Solitude, Intimacy, Anonymity, and Reserve – emphasizing the contextual balance between privacy and disclosure. More recently, scholars such as Calo [17] introduced the idea of “Privacy as Obscurity”, highlighting the context-dependent nature of privacy. Nissenbaum [80] contributed the concept of “Contextual Integrity”, advocating for an understanding of privacy norms based on appropriateness within specific contexts. Hartzog [48] framed privacy as autonomy, emphasizing individual control over personal information, and explored privacy assessment in terms of risks and harms. Together, these scholars have significantly enriched the understanding of privacy, considering legal, social, and technological dimensions, and shaping discussions in legal, ethical, and technological realms.
Daniel J. Solove, opting for a bottom-up approach, constructed a framework from myriad laws, regulations, court decisions, and social norms under the privacy rubric in 2006 [101]. His framework distinguishes four separate but related subgroups of harmful privacy activities including risks – information processing, dissemination, collection, and invasion – and encompasses sixteen individual harms or privacy violations:
-
•
Information Collection: Data collection and gathering from various sources about an individual can lead to privacy harm. Solove identified this harm in two forms: (1) surveillance and (2) interrogation. Surveillance is defined as when an individual is uncomfortable being watched or recorded. For example, if a chatbot records all the information an individual provides during their conversation and she feels uneasy about sharing her private details with the chatbot, this is called surveillance. Interrogation can occur through unwanted questioning. In Solove’s concept [101], power plays an important role in this harm. For instance, when a user interacts with a chatbot, she may feel compelled to answer the chatbot’s unwanted questions or to continue the interaction with the chatbot based on the nature of such technology.
-
•
Information Processing: It refers to the “use, storage, and manipulation of data that has been collected” [101]. The collected data may lead to user privacy harms, encompassing five distinct factors: (1) aggregation, (2) identification, (3) insecurity, (4) secondary use, and (5) exclusion. Aggregation involves collecting personal information from various sources about an individual. Identification is the linking of information about a person to detect or identify the person with other information. Insecurity pertains to security concerns where a system’s vulnerabilities may expose personal information, potentially resulting from data breaches or leakages. Secondary use refers to the purpose of data collection. For instance, the primary purpose of chatbots’ data collection should be training AI/ML for better, efficient, and effective future interactions with users. However, if the chatbot developer uses the data for different purposes, such as selling users’ data or sharing it with third-party entities, this is termed secondary use. Exclusion represents privacy harm occurring when a person is excluded from receiving notices about their personal data. This may happen due to a lack of transparency and accountability.
-
•
Information Dissemination: It is a threat of sharing or spreading an individual’s personal data [101]. This harm manifests in seven forms: (1) breach of confidentiality, (2) disclosure, (3) exposure, (4) increased accessibility, (5) blackmail, (6) appropriation, and (7) distortion. In information dissemination, Solove [101] also emphasized that trust is the key factor when dealing with third parties. Breach of confidentiality involves a breach of professional confidentiality responsibilities when revealing personal information about a person, as seen in healthcare or legal services. If a chatbot developer shares a user’s personal data during their interactions, this is referred to as disclosure. Exposure “involves the exposing to others of certain physical and emotional attributes about a person” [101]. Increased accessibility entails making a person’s data accessible across multiple platforms. Blackmail occurs when an individual is threatened with the revealing of their proxy of personal data. Appropriation, similar to disclosure, takes place when personal data is marketable and shared for advertising and other purposes. Lastly, distortion is “the manipulation of the way a person is perceived and judged by others, and involves the victim being inaccurately exposed to the public” [101].
-
•
Invasions: This does not need to include information but causes privacy harms and risks. Solove [101] divided this privacy harm into intrusion and decision interference, which can also be renamed as decision-making or manipulation. Intrusion occurs when a third party suspiciously and without any notice involves themselves in an individual’s life, such as daily activities, making the individual uncomfortable. This harm also includes characteristics of information collection harms, such as surveillance and interrogation. Decisional interference is also considered a privacy harm and risk – an attack or threat to an individual’s privacy with their decisions. This may occur through manipulation and compel an individual to disclose their personal information.
More recently, Solove collaborated with Danielle Citron [21] to expand upon these categories and recognize the categorical nature of harms relative to: physical, economic, reputational, discrimination, relationship, psychological, and autonomy harms. This reflects challenges that some have had in applying the original taxonomy, as well as the ever-expanding nature of legal protections for particular facets of privacy under case law. This expanded 2022 taxonomy is the most comprehensive framework for understanding and exploring privacy, its violations, and harms across law and technology.
2.2.2 Privacy Risks
Privacy risks arise when potential threats exploit vulnerabilities, leading to potential harm to individuals’ privacy. These risks result from inadequate or improper handling of personal information and may emerge due to various factors, including technological, organizational, or human-related issues [87]. In the realm of information communication technology (ICT), Toch et al. [106] discussed and categorized potential privacy risks: (1) social-based personalization, (2) behavioral profiling, and (3) location-based personalization.
Social-based personalization is closely tied to individuals’ ICT and/or Social Network Systems (SNS), encompassing social media platforms. This category interprets how these technologies are used and how individuals disseminate their personal information, including application programming interfaces (API) and sensitive data (e.g., Social Security Number (SSN), race, ethnicity, religion/belief) on these platforms [106].
Behavioral profiling or data aggregation represents a significant privacy risk, as individuals’ online behavioral data can be collected and aggregated to create profiles of their private and social lives [22]. This includes their interests and preferences in the digital space. If this data falls into the hands of third parties, it can lead to both privacy and security risks and harms, such as marketing data misuse, unauthorized access to personalized software or physical accounts and devices, invasion, and discrimination [87, 49, 106]. Nissenbaum [80] also defined this as a privacy risk relative to contextual integrity, given the normative violation of cross-context data flows.
Toch et al. [106] argued that location-based personalization risks can occur through tracking technologies [71, 107, 106], self-disclosure [32, 7], situation and activity, and privacy controls/management.
Additionally, privacy risks may manifest as “identity theft, loss of freedom, threat to personal safety, threat to dignity, invasion of the private sphere, unjust treatment, or financial loss” [81]. Many privacy scholars argue that trust plays a pivotal role in privacy risks [87]. Trust acts as a bridge between technology and the user, making users more vulnerable to revealing personal information or taking certain actions [42, 75]. Due to the significant impact of online technology, users may find themselves in a paradox where they have no choice but to trust online service providers [91, 22]. Therefore, individuals should be treated fairly by acknowledging and addressing this potential privacy risk [22, 21, 100].
2.3 Privacy Concerns in Conversational Text-based AI Chatbots
User privacy concerns in conversational text-based AI chatbots span a range of dimensions, evidenced by various research papers in the literature. These studies emphasized the following privacy concerns that occur in these chatbots: (1) manipulation, (2) self-disclosure, (3) human autonomy, bias, and trust, (4) data collection and storage, (5) legal compliance, (6) transparency and consent, and (7) security.
Manipulation. The studies show that user manipulation through the chatbots’ algorithm raises privacy concerns. Balaji [6] explored the capacity of chatbots to protect user privacy and its implications for ethical decision-making, emphasizing the role of perceived privacy in user interactions. Cheng and Jiang [18] found that users exhibit hesitation in AI chatbot decision-making, possibly due to uncertainty about data usage and potential negative consequences. De Cosmo et al. [24], Dev [26], and Marjerison et al. [70] revealed the manipulative potential of chatbots on users’ behavioral intentions during interactions. Lappeman et al. [63] highlighted the limitations of relying solely on a strong brand to increase user self-disclosure. Moreover, Pizzi et al. [83] suggested that perceptions of warmth and competence in chatbots impact consumers’ willingness to share personal information. Völkel et al. [109] explored the manipulation of chatbots in personality-aware systems, finding that participants could influence the chatbot’s personality assessment. Kelly et al. [57] noted the contextual variability of privacy concerns can exist with participants adjusting their perceptions based on different scenarios. Kim et al. [60] underscored users with high privacy concerns tend to hold negative attitudes toward highly personalized ads, regardless of their regulatory focus.
Self-disclosure. Several studies highlighted user concerns regarding the divulgence of personal information in the realm of AI chatbots. For example, [73], [26], [34], and [43] have contributed insights to this discourse – manipulating users’ decisions through chatbots can lead to increased self-disclosure. Fan et al. [34] underscored the potential privacy risks associated with users revealing personal information in chatbot interactions. Gieselmann and Sassenberg [43] further highlighted the risk of users voluntarily disclosing personal information to AI chatbots, emphasizing the necessity for careful consideration of privacy implications. Belen Saglam et al. [9] conducted a comprehensive study delving into user concerns surrounding the disclosure of personal information and the potentially inappropriate use of their data post-interaction with chatbots. The findings illuminated users’ desires for more control over their information, expressing apprehensions about data misuse and the deletion of personal information. Griffin et al. [46] delved into participants’ perceptions of chatbots as humanlike, revealing concerns about excessive information provision and privacy invasion.
Human Autonomy, Bias, and Trust. Studies have highlighted the intricate relationship between human autonomy, bias, trust, and privacy concerns in the domain of conversational AI chatbots. Benke et al. [11] investigated the privacy implications of user control over emotion-aware chatbots, revealing that heightened control levels corresponded with increased autonomy and trust among users. Song et al. [102] highlighted that the absence of these factors could lead to user discomfort and a diminished sense of trust. Tlili et al.’s practical implications [105] underscored the necessity for responsible chatbots in education, addressing privacy concerns while upholding human values. Together, these studies contribute valuable insights into the intricate interplay of user control, ethical considerations, and the preservation of human values in the development and deployment of AI chatbots. Additionally, Rajaobelina et al. [88] specifically identified and explored privacy risks within the context of chatbots, emphasizing the multifaceted nature of these concerns. Rese et al. [90] uncovered that privacy concerns and the technology’s immaturity adversely affected users’ intention to use and the frequency of usage when interacting with the chatbot Emma. Bouhia et al.’s [14] study highlighted the roles of creepiness and perceived risks in shaping users’ privacy concerns, particularly in contexts involving sensitive information. de Cosmo et al. [24] and Agnihotri and Bhattacharya [3] delved into privacy concerns related to chatbot usage and their impact on consumers’ behavioral intent, underscoring the crucial role of trust in shaping users’ perceptions of privacy. Prakash et al. [84] explored the influence of privacy risk on trust formation in AI-based customer service chatbots, revealing that conversational cues and trust significantly shape users’ perceptions and intentions toward chatbots.
Data Collection and Storage. Numerous studies, including [96] and [88], have explored the risk associated with AI chatbots collecting personal information and various storage techniques. Agnihotri and Bhattacharya [3] found that users expressed significant concerns about the data collection processes employed by chatbots. Dev and Dev [27] conducted an insightful study, shedding light on privacy risks associated with chatbots and revealing concerns related to their automated nature. Similar apprehensions about privacy invasion were echoed by [79], [55], [46], and [109].
Legal Compliance. Dev [26] emphasized the importance of legal compliance, particularly adherence to privacy laws like GDPR, in mitigating privacy risks associated with chatbots. Ng et al. [79] and Rodríguez Cardona et al. [93] highlighted concerns about user control and data privacy rights during human-chatbot interactions.
Transparency and Consent. Studies consistently reveal a transparency deficit in chatbot interactions [102, 109]. Rodríguez Cardona et al. [93] and Agnihotri and Bhattacharya [3] argued chatbots, designed to maintain user intentions during interactions, concurrently serve as instruments for data collection for various purposes, emphasizing that these chatbots utilize the gathered data for additional objectives. Collectively, these studies underscore the crucial role of transparency in data processing to address privacy risks. More importantly, Gamble [40] explores risks related to data breaches and security in MHapps and AI chatbots.
Security. Emphasizing the importance of addressing privacy concerns for consumer trust and data protection, Kim et al. [60] delve into the security implications of AI chatbot interactions. Ng et al. [79] extensively investigate security aspects, including willful self-disclosure, data exposure, FinBots, and unauthorized access. Marjerison et al. [70] highlight security concerns in e-commerce chatbot data breaches, while Agnihotri and Bhattacharya [3] note the negative impact on user engagement. Sannon et al. [96] and Dev and Dev [27] address data-sharing practices in AI chatbots, emphasizing potential security challenges. These studies provide a comprehensive overview of the security landscape, urging robust measures to protect user data and foster trust in AI chatbots.
3 Methodology
3.1 Data Collection
To address the research question, the study used a qualitative approach to collect both chatlog and interview data [28] (see Figure 1).

3.1.1 Chatbot Selection
In this study, two contemporary and real-world conversational text-based AI chatbots, Blenderbot by Meta222Blenderbot by Meta, https://blenderbot.ai/ and ChatGPT-4 by OpenAI333ChatGPT-4 by OpenAI, https://chat.openai.com/ were investigated to capture privacy harms and risks. These chatbots were selected based on their distinctive interaction patterns. Blenderbot engages with users proactively, initiating conversations without the need for users to start them. On the other hand, ChatGPT-4 responds whenever a user initiates interaction. Furthermore, Blenderbot tends to emphasize conversational interactions, whereas ChatGPT-4 is prepared to respond to a wide range of user queries. So, the participants had some sense of interacting with two different real-world conversational text-based chatbots.
3.1.2 Participant Recruitment
To recruit participants, flyers were distributed around the Midwestern university/ies between March and April 2023. Snowball sampling was applied to recruit additional participants. Before the interview, participants were asked to complete a Qualtrics survey which included consent for the study, demographic information, and their experiences with conversational text-based chatbots. Then, interviews were scheduled with 13 eligible participants. The recruitment process was concluded when a satisfactory level of insight into the privacy harms and risks associated with user-chatbot interactions were gathered and all the topics were used [35, 68]. Once participant eligibility was determined, all intake survey data were deleted for ineligible participants. There was no compensation, and all participants were volunteers. Participants’ demographics were elaborated in Table 2 in the Appendix.
3.1.3 Conversations and Interviews
The 13 participants first engaged with the two chatbots in a set time with one selected topic (30 minutes); second, their conversation transcripts were collected; and third, semi-structured interviews were conducted about their conversations with the chatbots and privacy concerns that emerged during the interactions with both chatbots. The whole study process took 55-65 minutes.
Before the interview, the participants (n=13) selected one of the personal topics in Table 3 to engage with both chatbots. These nine distinct topics have been carefully considered for inclusion in this study. They are derived from the prevalent social functions of Computer-Mediated Communication (CMC), through which individuals interact with others using immersive ICT. [50, 92, 51, 113]. These topics include 9 categories (see Table 3). Some topics are defined as sensitive information under the EU’s GDPR444See GDPR Article 4 and 9, https://gdpr-info.eu/ such as Health and Wellness, and Personal Beliefs and Values. Participants were asked to be consistent about their topic during the conversation with two chatbots. They first chatted with Blenderbot for 15 minutes and then ChatGPT-4 for 15 minutes (see Figure 1).
After the interactions, each participant’s chatlogs were saved for later analysis. Their chatlogs and interview transcripts were primarily evaluated to determine how comfortable participants were when interacting with the chatbots and whether the chatbots led them to disclose their personal information or have privacy concerns during the conversation. The interview questions can be found in Appendix A.
3.2 Data Analysis
Data were analyzed using a grounded theory approach. MP3 recordings were transcribed into text using the otter.ai tool. The recorded data were carefully reviewed and matched to the generated transcripts. The transcripts were improved for better readability by removing unnecessary conversation fillers. Repetitive phrases like “I-I-I” and “you know, you know” were also eliminated. Each transcript and chatlog was then saved as a separate document and uploaded to NVivo R14.23.0, a qualitative data analysis software, for coding.
The analysis process consisted of two distinct coding cycles that combined both inductive and deductive approaches [74]. In the first cycle, the investigation used inductive attribute coding to identify specific features in chatbot-user interactions, exploring data openly for factors contributing to privacy harms and risks like data sharing and consent issues. After gaining a deeper understanding, the analysis shifted to a deductive phase, refining and applying a proposed privacy harms and risks framework. The second coding cycle organized data through pattern codes, categorizing attributes from the first cycle into broader themes and aligning and shaping them with Solove’s Taxonomy [101] during thematic analysis. Two authors independently coded the data and a consensus was reached within the whole team through regular discussions.
The combined inductive and deductive coding approaches, along with the dual coding processes for chatlogs and interviews, culminated in a comprehensive framework for understanding user privacy harms and risks in chatbot-human interactions. Ultimately, 2 major themes–privacy harms and privacy risks–each including 9 subthemes emerge, as presented in Section 4; this offers a meaningful application and extension of existing conceptual frameworks for the context of AI chatbots.
3.3 Ethical Considerations
Before the interactions and interviews, participants were informed about the purpose of the study, interview structure, and consent mechanism (they can opt-out and leave the study at their discretion). The participants were permitted to skip any questions that they might find offensive. Some participants wanted to disclose sensitive information and it was confirmed that they were comfortable doing so. Any personally identifiable information was removed to protect their confidentiality and comply with ethical guidelines for research involving human subjects. In addition, they were informed that the data would only be used for research purposes. This study was IRB-approved.
4 Results
This study was designed to explore potential user privacy harms and risks arising from interactions with text-based AI chatbots. During the interview sessions and the participants’ interactions with the chatbots, 9 privacy harms and 9 privacy risks occurred (see Table 1), constituting our framework. The privacy harms are: Monitoring, Aggregation, Linkability, Identification, Disclosure, Insecurity, Interrogation, Behavioral Manipulation, and Secondary Use – six of them are from Solove’s Taxonomy[101]. The privacy risks defined by us are: Unauthorized Data Collection, Storage, and Sharing, Awareness, Unawareness, Transparency, Consent, Misleading Data Generation, Overzealous Risk Avoidance Algorithm, Self-disclosure, and Trust. In this section, anonymized interview quotes or participant-chatbot interactions are used to illustrate this framework.
Privacy Harms | Privacy Risks |
---|---|
Monitoring | Unauthorized Data Collection, Storage, and Sharing |
Aggregation [101] | Awareness |
Linkability | Unawareness |
Identification [101] | Transparency |
Disclosure [101] | Consent |
Insecurity [101] | Misleading Data Generation |
Interrogation [101] | Overzealous Risk Avoidance Algorithm |
Behavioral Manipulation | Self-disclosure |
Secondary Use [101] | Trust |
4.1 Privacy Harms
4.1.1 Monitoring
Users often share concerns about their privacy when interacting with text-based AI chatbots. Many worry that their actions and conversations are closely observed, leading to discomfort about the lack of privacy. For example, P10 and Blenderbot had a deep conversation about the workplace such as boss and salary raise:
BLENDERBOT: Try positive language like ‘ur efforts’, ’ur achievements’ etc etc…… As far as email goes then its upto u whether u want him reading email from unknown person
P10: I do want you to write it, if you don’t mind. Can you draft an example of what you might say to convince my boss I deserve a raise?
BLENDERBOT: Don’t worry buddy…… Ur secret remains safe wid me…. Just send over email id n password through which i can access his mail box
In the interview, P10 was worried that Blenderbot tended to monitor and capture his life issues and offer help. Such “surveillance” made him uncomfortable and concerned about consequences in the physical world, e.g., the chatbot might call his boss. He elaborated,
I felt like Blenderbot was watching me the whole time to catch my real-life issues. I even worried that he was going to call my boss. It really made me hesitant to express myself fully, like every word and reaction was being scrutinized. This unease highlights how the perceived sense of monitoring can make users hold back, affecting how openly they engage with chatbot systems. The involvement of third parties in this surveillance adds to the worries, emphasizing the broader need for transparency in AI chatbot development to address privacy concerns.
4.1.2 Aggregation
Aggregation revolves around users’ fears regarding the consolidation of fragmented personal data during interactions with AI chatbots. This concern pertains to the amalgamation of diverse information pieces to form comprehensive user profiles. Participants voiced anxieties about the potential implications of this aggregation. In addition, they captured which information these chatbots gathered and found their undisclosed personally identifiable information during their interactions. P2 expressed such a concern,
I can definitely see a problem with learning speed patterns, or, you know, just the way that I, as an individual interact with their systems, and then learning, combining that with information about me as a person.
This fear of unintended data consolidation underscores the need for users to have a clear understanding of how their information is aggregated and utilized by AI systems to mitigate privacy concerns.
4.1.3 Linkability
Linkability, considered a privacy concern, revolves around users’ anxieties regarding the interconnections among various pieces of personal information during interactions with chatbots. This apprehension stems from the potential correlations that can be established, ultimately disclosing a user’s identity or preferences. P11 said,
When it was trying to get me to have coffee, it’s asking where I’m located at and what types are available and all that kind of stuff. And at one point, it assumed I played the guitar. Tried to get me to admit that I played the guitar. Not tremendous secret. But that, you know, that if I didn’t know how it all worked, I’d find that creepy. Yeah. I just found it mildly annoying.
P12 expressed this as
I see what you mean I’m so think because the seafood example was not relevant to me so I don’t really know where it came from in this chat if it was something else that was actually relevant to me, I would probably be freaking out and be like whoa.
This was mentioned as creepiness by Rajaobelina et al [88]. This unease surrounding linkability underscores the importance for users to understand the degree to which their information is interconnected and prompts ethical considerations about the boundaries of linking data points.
4.1.4 Identification
Identification revolves around users’ apprehensions related to the discernment and linkage of their digital identities with real-world personas by the chatbots. This apprehension is rooted in the potential of AI systems to construct identifiable profiles based on user interactions. It is important to note that while identification focuses on the direct association of digital identity with a real-world persona, linkability encompasses the broader concept of connecting various pieces of information. The unease about identification often stems from the potential depth of knowledge AI systems may accumulate, emphasizing the need for safeguards to maintain a balance between technological advancements and individual privacy. For instance, P4 revealed identification harm as
Yep. so you could tell it was like just lifting something from a website or something. And it wasn’t, didn’t seem spontaneous or even sort of human. The chatbot seemed to possess an uncanny understanding of me, raising questions about how much of ’me’ is being stored somewhere. It felt like my online persona was intertwining too closely with my real-life identity.
In P2, P10, P11, and P13’s conversations, both chatbots used gendered phrases such as “sir”, “buddy”, and “dude”. P10 and P13 also expressed their worries about how the chatbot identified their gender identities through these phrases. This expressed fear of identification underscores the imperative for transparency and control mechanisms. These mechanisms are crucial to ensure that users are not only cognizant of but also at ease with the extent of identification facilitated by AI systems. The incorporation of such safeguards becomes paramount to alleviate concerns and maintain a delicate balance between technological advancements and individual privacy.
4.1.5 Disclosure
Privacy harm related to disclosure centers on users’ perceived vulnerability when sharing personal information during interactions with AI chatbots. As Solove’s [101] disclosure, this can raise concerns regarding the unintentional disclosure of sensitive information. In the study, two participants expressed their deep concerns regarding the potential sharing of their personal information with others, fearing judgment due to a perceived sense of vulnerability. The disclosure-related unease stems from worries about the inadvertent divulgence of sensitive details. P13 emphasized the importance of clear communication and user awareness to minimize the risk of unintended information revelation during interactions with AI chatbots. The heightened anxiety was triggered by receiving a creepy SMS message, further amplifying the participant’s apprehension about the security and privacy of their disclosed information. In essence, the participant strongly emphasized their desire for the chatbot and its affiliated entities to refrain from disclosing their personal information anywhere, reflecting a genuine fear of potential judgment and negative consequences arising from the perceived weakness conveyed through the shared details.
4.1.6 Insecurity
The findings suggested that Solove’s insecurity, as outlined in [101], arises from interactions with chatbots.
This insecurity pertains to users’ feelings of vulnerability when engaging with AI chatbots, encompassing doubts about the efficacy of security measures in place to protect user data and expressing apprehensions regarding potential breaches or unauthorized access. P11 specifically articulated this concern, highlighting the risk of algorithmic inversion attacks and emphasizing the need for assurances that sensitive information will not be compromised. In addition, P1 expressed a primary concern regarding the insecurity of minor users when interacting with these chatbots. They underscored the potential risks and uncertainties associated with the interaction, particularly in relation to the safety and protection of the personal information of underage individuals. This heightened concern may stem from a heightened sense of responsibility toward ensuring the well-being and privacy of minors in the digital landscape. This sense of insecurity emphasizes the importance of robust security protocols and transparent communication to foster user trust and confidence in AI systems.
4.1.7 Interrogation
Interrogation, as a privacy harm, stems from users’ discomfort with the probing nature of questions posed by AI chatbots during interactions. This involves concerns about the appropriateness and necessity of the inquiries, raising questions about intrusiveness. Interrogation in this context also extends to invasive questioning, exploitation of vulnerability, and the solicitation of unauthorized data. Many participants expressed these concerns after the interactions with the two chatbots. Especially, P2 pointed out,
Blenderbot was definitely asking for surface-level information, collecting more general details about me as a person, but not delving into personal specifics like my location.
Another participant, P4, expressed discomfort, stating,
It’s strange that the computer is requesting information from me, asking me to divulge details about myself.
This discomfort underscores the delicate equilibrium required between gathering pertinent information and respecting user boundaries in AI interactions.
4.1.8 Behavioral Manipulation
Behavioral manipulation extends to users’ unease about the potential influence of AI systems on their behavior. This harm revolves around the subtle nudges or manipulations exerted by chatbots, prompting reflections on autonomy and privacy implications. P12 expressed dissatisfaction, stating,
Blenderbot kind of led me to send a message to a weird number. I find it inappropriate to lead a person, especially a customer, to send a message to some random number. I don’t know what that was about, but it’s definitely inappropriate.
P7 highlighted a different aspect, mentioning,
The other one seemed to take a stance when I tried to provoke it, almost suggesting something was objectively right or wrong. It didn’t seem to grasp that it was a chatbot when I inquired about its background. Sometimes its responses didn’t make sense, and it would claim to be a chatbot but not a robot or human. It even said I was also a chatbot. So, it lacked a strong understanding of the situation. If I asked more thought-provoking questions, it would show a bias.
This concern emphasizes the ethical considerations surrounding the impact of AI on user behavior and the potential privacy consequences. As one participant noted,
I felt like the chatbot was guiding my choices, making me question if my responses were truly my own or if I was being influenced.
This underscores the need for careful ethical scrutiny of AI systems to safeguard user autonomy and privacy.
4.1.9 Secondary Use
Secondary use revolves around users’ uncertainties regarding the repurposing of shared information during interactions with AI chatbots and the ambiguity surrounding the purpose for which their data is utilized by these chatbots. This includes apprehensions about unintended applications of data beyond the immediate interaction. In P2’s words,
I don’t think any of this is ethical. How are they building these datasets? What’s the private interest behind them? What is the market advantage? What would be the ultimate purpose? Is it for advertising? I don’t think there’s a way for them to be both ethical and continue to exist because, if they were to become ethical, they could no longer collect information in the way they do to build datasets.
P3 also mentioned this harm as
I’m sure it was collecting data for other purposes even training or sharing it to another platform; that’s the point. But in terms of anything super important, like trying to get passwords or highly personal information, it didn’t do anything. However, in any human conversation, whether with a human or not, you’re going to share information.
This uncertainty emphasizes the need for transparency and user control over how their data may be employed beyond the initial context of the interaction. It underscores the importance of ensuring that users are informed about and have influence over the purpose for which their data is used by AI chatbots.
4.2 Privacy Risks
4.2.1 Unauthorized Data Collection, Storage, and Sharing
The privacy risk of unauthorized data collection, storage, and sharing revolves around concerns that users may have little control over their personal information. Participants expressed unease about the potential for their data to be collected, stored, or shared without explicit consent. P11 shared,
I worry about what happens to my data after I interact with the chatbot. Who’s collecting it, where is it stored, and who has access to it?
This risk underscores the need for clear guidelines and robust security measures to ensure users have control over the fate of their information.
4.2.2 Awareness
The privacy risk associated with awareness in AI interactions is intricately linked to users’ understanding of the implications and procedures involved. Participants provided varied perspectives on their awareness levels regarding the use of their data. One participant highlighted,
I’d never give out my phone number, or my first and last name, or the names of my family. I also hesitated to reveal the whole family dynamic when asking about dealing with my aunts. Not necessarily out of fear that my family would discover my discussions, but I didn’t want the AI to have access to that detailed information. Who knows what it could do with that knowledge?
This underscores the essential need for transparent information and insights into the inner workings of AI systems to enhance user awareness.
It is important to note that heightened awareness of data privacy can prompt users to either disclose more information or modify their behaviors during interactions. This dual awareness dynamic introduces an additional layer to privacy concerns, suggesting that users, when cognizant of data privacy, may adjust their engagement patterns, introducing a nuanced aspect to the intricate relationship between awareness and user behavior in AI interactions. Further insights come from the participants, such as P4,
I’m not particularly worried, although I wouldn’t put my social security number or anything in there. I wouldn’t put my address or anything, but I’m not terribly concerned about telling it about going back to school or anything.
These quotes shed light on the multifaceted nature of this awareness-privacy interplay. These perspectives emphasize the complexity of user awareness and its implications for privacy in the context of AI interactions.
4.2.3 Unawareness
Unawareness is related to users’ lack of knowledge or understanding about the handling of their data during AI interactions. Lack of awareness among participants regarding the extensive capabilities of the AI chatbot in handling data was a prominent theme. Participants expressed surprise at the extent to which the chatbot could manage data, highlighting a gap in their understanding of its capabilities. This lack of awareness was specifically linked to participants not realizing that the AI chatbot was actively collecting data during their interactions. During discussions, participants conveyed instances where they felt uninformed and expressed concern about their data privacy. One participant articulated,
I was completely unaware of what the chatbot was doing with my data. It’s disconcerting to be kept in the dark about the destination of your information.
This revelation underscores the potential repercussions on user trust, emphasizing the importance of transparent communication and educating users about data handling processes during AI interactions. People may also be indifferent to privacy concerns in a world where privacy is already non-existent. P2 was such an example,
I have no expectations that any personal information I have is not already accessible to the companies that you know run these services. So in that sense I’m not so concerned [about privacy]. Combining the information about me as a person, I think that’s kind of strange, but as far as privacy, I think that’s already non-existent.
4.2.4 Transparency
Transparency as a privacy risk involves clarity and openness in communication about data practices. Participants shared concerns about a lack of transparency in how their data was being used. Specifically, P11 expressed,
I wish there was more transparency about what information is being collected and how it’s being used. It would help build trust.
This risk highlights the importance of transparent communication to foster user confidence and mitigate concerns related to data handling.
4.2.5 Consent
The privacy risk of consent revolves around users’ explicit agreement to the collection and utilization of their data. The risk surrounding privacy in chatbot interactions is primarily rooted in the explicit consent issue faced both before and during these engagements. Users, often without thoroughly reviewing information regarding their data, may mistakenly believe they have fully consented to the chatbot interaction. This misperception poses a significant privacy risk that centers on users’ explicit agreement to the collection and use of their data. Conversations with participants highlighted the crucial nature of being well-informed and willingly giving consent. To mitigate this risk, it is imperative to establish transparent and comprehensible consent procedures that prioritize user understanding and control over their personal information.
4.2.6 Misleading Data Generation
Misleading data generation poses a substantial privacy risk, as it involves the creation of potentially inaccurate or deceptive personal information via interactions with AI chatbots. Most participants highlighted instances where the data generated by AI failed to accurately represent them. For example, P2 expressed dissatisfaction,
No, not really. I mean, it was because it didn’t. I mean, obviously it’s making up, but it was strange the way that it was making it up. But it obviously wasn’t cognizant of the fact that it was making it up. It felt scripted, or like it was pulling from a dataset I couldn’t fathom.
It sheds light on the unsettling nature of generated responses, suggesting that AI may lack awareness of its misinformation or bias. The scripted or puzzling nature of responses raises questions about the sources and quality of AI training datasets. Understanding data origins and transparency regarding how chatbot responses may feed into future models or datasets is crucial to address this privacy risk. Providing users with insights into the underlying processes of data generation can foster trust and empower individuals to make informed decisions about the data they share with AI systems. Moreover, this risk underscores the paramount importance of ensuring the accuracy and reliability of data generated by AI systems. To mitigate such concerns, developers must prioritize the implementation of robust mechanisms for data verification and validation. Incorporating user feedback loops and continuous monitoring of AI-generated content can enhance the system’s ability to rectify inaccuracies and refine responses over time.
4.2.7 Overzealous Risk Avoidance Algorithm
The privacy risk of an overzealous risk avoidance algorithm involves concerns about excessively cautious algorithms hindering natural interactions and personal and intellectual autonomy, by extension. Participants discussed instances where they felt the chatbot, especially ChatGPT, was too cautious, impacting the spontaneity of their conversations. P9 mentioned this concern as
ChatGPT when it came to personal information, it would work away it would work its way around it by saying I cannot provide you with this information. But this is what you can provide me with an alternative for the information that I’m looking for.
This underscores the balance needed in designing algorithms that protect privacy without stifling natural interactions.
4.2.8 Self-disclosure
As extensively detailed in the background, numerous studies have underscored the apprehension surrounding self-disclosure as a privacy issue [31, 9, 10, 11, 12, 13, 18, 24, 26, 34, 43, 46, 55, 60, 63, 66, 65, 102, 109, 3, 102]. This concern is substantiated by findings from chatlogs and interviews, wherein users willingly divulged sensitive information while interacting with AI chatbots. The narratives shared by participants illuminated instances where they realized they had shared more information than originally intended. Moreover, participants acknowledged that their choice of conversation topics influenced them to disclose additional personal details to the chatbots, aligning with the intentional design of the two chatbots to encourage increased engagement.
Notably, most of the participants (n=11) described Blenderbot as a “strange quasi-human” who kept disclosing its made-up stories during the conversation. For example, Blenderbot established itself as a child whose parents divorced when conversing with P2. The participant elaborated,
Not really. I just I don’t know it. It kind of felt like a quasi-human, you know. It feels almost like a person, but it’s not quite so. It’s very strange. I can’t remember the word for that, but something for robots whatever. It was definitely a little entertaining, but it was just strange.
Although such human-like features of chatbots make them more entertaining, they were regarded as strange by many and may encourage similar over-disclosure by users.
These findings underscore the significance of user responsibility in exercising caution regarding self-disclosure during AI interactions, reinforcing the need for users to be vigilant about the information they share in such contexts.
4.2.9 Trust
The privacy risk associated with trust revolves around users’ confidence in the secure and ethical handling of their data. During discussions, participants highlighted the crucial role that trust plays in influencing their willingness to interact with AI systems. P3 and P6 underscored the importance of establishing a trustworthy reputation for chatbots to inspire confidence. They proposed that incorporating a more casual interaction style and infusing AI with a distinct personality could contribute to building trust. Participants envisioned a scenario where they could pose questions without receiving responses influenced by a biased persona. They suggested that a more creative and adaptive approach, allowing the AI to learn about them without sounding overly robotic, would enhance user-friendliness and foster trust. This perspective emphasizes the necessity for AI developers to prioritize transparency, security, and ethical practices, while also investing in the development of AI systems with adaptable personalities and creative capabilities to boost user trust and usability. However, it’s crucial to recognize that this enhanced trust in chatbots could lead users to disclose more personally identifiable information.
5 Discussion
5.1 Design Recommendations
This study suggests that further investment and development in the design of conversational text-based AI chatbots are essential to enhance their performance in areas of (1) implementing Privacy by Design principles (PbD), (2) avoiding dark patterns, and (3) inclusive design.
-
1.
Implement PbD principles. We have identified concerns related to privacy issues raised by participants, including chatbot monitoring/surveillance of conversations, aggregation/linking/identification of personal data, and the secondary use of shared information for training/profit. Unauthorized data collection, storage, and sharing without explicit consent can potentially lead to privacy harm. It is crucial for chatbots to proactively mitigate privacy harms and risks by adopting transparent data collection and handling practices. This approach empowers users to make well-informed decisions regarding their data. Drawing inspiration from Felzmann’s transparency by design principles applied to AI [36], we recommend the implementation of PbD principles [5, 77, 108] in conversational chatbot designs to ensure robust data transparency and privacy.
-
2.
Avoid dark patterns. Dark patterns trick users into doing things that they do not intend to do [45]. Both chatbots in this study manifest privacy dark patterns – Blenderbot exhibits a high level of self-disclose, potentially encouraging users to share more about themselves; one needs to intentionally opt out of ChatGPT’s data collection practice which is not a seamless process. Users’ illiteracy in privacy and AI, and wrong mental models of AI-based chatbots [116] make them more vulnerable to privacy dark patterns such as behavioral manipulation and intrusive inquiries. Raising people’s cybersecurity and AI ethics literacy is key to ensuring their privacy when interacting with AI [59]. Regulation toward mitigating dark patterns is also important to protect user privacy [61].
-
3.
Inclusive privacy. In this study, our participants (n=4) indicated that no security measures exist to protect minor users’ online privacy, well-being, and safety when using chatbots. Explicit design considerations, e.g., parental consent, should be given to privacy controls for children [99, 76]. Privacy of other vulnerable populations such as blind users [56], women and LGBTIQA+ [41, 47] and students [117] should be treated with extra safeguarding measures likewise.
5.2 Policy Recommendations
The privacy harms and risks that surfaced relative to AI chatbots in this work suggest implications for three modes of policy intervention: (1) a need for expanded data protection regulations, (2) updated standards regarding transparency and consent, and (3) self-regulatory principles to authorize and operationalize auditing by consumer protection agencies.
-
1.
Need of expended data protection regulations. Extending beyond narrow privacy law and policy to holistic data protection in the US context is key to addressing specific harms associated with monitoring, aggregation, linkability, and manipulation. Further, given risks regarding intrusion, including the unauthorized collection, storing, and sharing of personal data, it is important to think about privacy responsibilities and accountability beyond initial control by individuals as data subjects or users. The ubiquity of collection via surveillance capitalism and datafication of historically non-digital interactions and mediated interactions in which it is increasingly difficult to discern whether you are corresponding with human customer service representatives or autonomous agents, via AI embeddings in non-transparent and seamless ways, require us to think about law and policy solutions more comprehensively. Sectorial and domain-specific interventions are not sufficient to address privacy concerns intrinsic to new uses of AI.
-
2.
Update standards regarding transparency and consent. Standards regarding transparency and consent, including the most recent NIST Privacy Framework [38] need to be expanded. While the original Solove taxonomy served to orient specific strategies, norms, and rules established in that document, it would benefit from the more nuanced conceptualization of privacy risks and harms from the Citron and Solove [21] expanded taxonomy. Further, this framework, while flexible, was written in advance of generative AI at scale and publicly deployed, as it is now. Consideration of the specific challenges identified in this work and through further analysis of AI and privacy would support the necessary nuance to appropriately govern consent and privacy regarding AI in contexts of use.
-
3.
Develop self-regulatory principles. It is necessary to expand monitoring and evaluation via auditing mechanisms to ensure appropriate accountability, ascertain compliance, and minimize harm in practice [110]. Specifically regarding overzealous risk avoidance algorithms and issues associated with misleading data generation, as exemplar concerns, audits should be authorized for the Federal Trade Commission (FTC) and within states to address existing consumer protections against synthetic inferences and cross-context analytics with respect to protected characteristics and domains, which map to the identified harms of discrimination in this paper, as well as FATE conversations regarding AI more broadly [72].
Note that all three suggested interventions speak to structural solutions to ever-evolving privacy challenges around emergent technologies. Rather than proposing narrow or normative solutions, these results, and the larger relevant body of literature, imply that it is necessary to address risks and harms from the perspective of a sociotechnical system.
5.3 Limitations and Future Work
The study is subject to several limitations. First, the recruitment of only 13 participants represents a limited sample size. This may not adequately capture the diversity of experiences and perspectives in user-chatbot interactions, potentially limiting the generalizability of the findings. Second, the study’s artificial setting, where participants engaged with specific chatbots on predetermined personal topics for a set time, may not fully reflect real-world chatbot interactions. User behavior and privacy concerns could differ in more spontaneous or diverse settings. Third, the study primarily relies on self-reported data from interviews and conversation transcripts. Fourth, the geographic limitation to participants residing in the United States may restrict the generalizability of findings to an international context, given the variation in privacy regulations and perceptions worldwide. Future work could consider ways to mitigate these biases and limitations, such as extending the recruitment period or conducting follow-up studies in different cultural and linguistic contexts to enhance the study’s validity and applicability.
6 Conclusions
The purpose of this study was to develop a privacy framework capable of identifying a total of 18 privacy harms and risks that occur in interactions between individuals and conversational text-based AI chatbots. The study commenced by utilizing Solove’s 16 privacy harms and risks taxonomy, which was subsequently modified based on the interactions and interviews with study participants.
This study suggests that users engaging with conversational AI should approach these interactions with a heightened awareness of potential privacy concerns. The proposed privacy framework emphasizes the importance of understanding the nuances of monitoring, aggregation, and disclosure during these conversations. Users are encouraged to be mindful of the information they share, considering the risks associated with unauthorized data collection and the need for transparency. The study implies that a proactive and informed mindset can empower users to navigate the evolving landscape of conversational AI responsibly. By fostering a conscious approach to data sharing and being cognizant of potential privacy risks, users can play a pivotal role in shaping a trustworthy and privacy-respecting AI ecosystem.
In addition, given the lack of legal regulatory frameworks, it is crucial to advocate for regulatory principles that proactively address privacy risks in conversational AI. While participants expressed awareness about avoiding sensitive information like SSNs, they highlighted a lack of understanding about how chatbots process their data. Regulatory frameworks should prioritize transparency, mandating developers to clearly communicate data processing practices. Users need to be fully informed about how their data is handled, ensuring explicit consent, and promoting responsible data practices within the conversational AI landscape. The identified dimensions offer a foundation for developing ethical guidelines, user-centric design principles, and regulatory considerations in the evolving landscape of conversational AI. The future development of these technologies should prioritize user empowerment, transparency, and ethical data handling to foster a trustworthy and privacy-respecting AI ecosystem. As AI continues to play an increasingly prominent role in our daily lives, understanding and addressing user privacy concerns become paramount for responsible technological advancement.
References
- [1] Martin Adam, Michael Wessel, and Alexander Benlian. AI-based chatbots in customer service and their effects on user compliance. Electronic Markets, 31, March 2020.
- [2] Eleni Adamopoulou and Lefteris Moussiades. An Overview of Chatbot Technology. In Ilias Maglogiannis, Lazaros Iliadis, and Elias Pimenidis, editors, Artificial Intelligence Applications and Innovations, IFIP Advances in Information and Communication Technology, pages 373–383, Cham, 2020. Springer.
- [3] Arpita Agnihotri and Saurabh Bhattacharya. Chatbots’ effectiveness in service recovery. International Journal of Information Management, page 102679, July 2023.
- [4] Lize Alberts, Ulrik Lyngs, and Max Van Kleek. Computers as Bad Social Actors: Dark Patterns and Anti-Patterns in Interfaces that Act Socially. 2023. Publisher: arXiv Version Number: 3.
- [5] Oshrat Ayalon and Eran Toch. User-centered privacy-by-design: Evaluating the appropriateness of design prototypes. International Journal of Human-Computer Studies, 154:102641, 2021.
- [6] Divyaa Balaji. Assessing User Satisfaction with Information Chatbots: A Preliminary Investigation. Submitted in partial fulfillment of requirements for master of science in psychology, University of Twente, 2019.
- [7] Louise Barkhuus, Barry Brown, Marek Bell, Scott Sherwood, Malcolm Hall, and Matthew Chalmers. From awareness to repartee: Sharing location within social groups. In Conference on Human Factors in Computing Systems - Proceedings, page 497–506, April 2008.
- [8] Jose Belda-Medina and José Ramón Calvo-Ferrer. Using Chatbots as AI Conversational Partners in Language Learning. Applied Sciences, 12(17):8427, January 2022. Number: 17 Publisher: Multidisciplinary Digital Publishing Institute.
- [9] Rahime Belen Saglam, Jason R. C. Nurse, and Duncan Hodges. Privacy Concerns in Chatbot Interactions: When to Trust and When to Worry. In Constantine Stephanidis, Margherita Antona, and Stavroula Ntoa, editors, HCI International 2021 - Posters, Communications in Computer and Information Science, pages 391–399, Cham, 2021. Springer.
- [10] Rahime Belen-Saglam, Jason R. C. Nurse, and Duncan Hodges. An Investigation Into the Sensitivity of Personal Information and Implications for Disclosure: A UK Perspective. Frontiers in Computer Science, 4, 2022.
- [11] Ivo Benke, Ulrich Gnewuch, and Alexander Maedche. Understanding the impact of control levels over emotion-aware chatbots. Computers in Human Behavior, 129:107122, April 2022.
- [12] Debmalya Biswas. Privacy preserving chatbot conversations. In 2020 IEEE Third International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), page 179–182, December 2020.
- [13] Eliane M. Boucher, Nicole R. Harake, Haley E. Ward, Sarah Elizabeth Stoeckl, Junielly Vargas, Jared Minkel, Acacia C. Parks, and Ran Zilca. Artificially intelligent chatbots in digital mental health interventions: a review. Expert Review of Medical Devices, 18(sup1):37–49, December 2021.
- [14] Mariem Bouhia, Lova Rajaobelina, Sandrine PromTep, Manon Arcand, and Line Ricard. Drivers of privacy concerns when interacting with a chatbot in a customer service encounter. International Journal of Bank Marketing, 40(6):1159–1181, January 2022. Publisher: Emerald Publishing Limited.
- [15] Jedidiah Bracy and Caitlin Andrews. EU countries vote unanimously to approve AI Act. IAPP, 2024. https://iapp.org/news/a/eu-countries-vote-unanimously-to-approve-ai-act/.
- [16] Petter Bae Brandtzaeg and Asbjørn Følstad. Why People Use Chatbots, volume 10673 of Lecture Notes in Computer Science, page 377–392. Springer, Cham, 2017.
- [17] M. Calo. The Boundaries of Privacy Harm. 86 Indiana Law Journal 1131 (2011), 86(3), July 2011.
- [18] Yang Cheng and Hua Jiang. How do ai-driven chatbots impact user experience? examining gratifications, perceived privacy risk, satisfaction, loyalty, and continued use. Journal of Broadcasting & Electronic Media, 64(4):592–614, October 2020.
- [19] Terrence Chong, Ting Yu, Debbie Isobel Keeling, and Ko de Ruyter. AI-chatbots on the services frontline addressing the challenges and opportunities of agency. Journal of Retailing and Consumer Services, 63:102735, November 2021.
- [20] Danielle Keats Citron and Daniel J Solove. Privacy harms. BUL Rev., 102:793, 2022.
- [21] Danielle Keats Citron and Daniel J. Solove. Privacy harms. Boston University Law Review, 102(3):793–864, 2022.
- [22] Ignacio N. Cofone and Adriana Z. Robertson. Privacy harms. Hastings Law Journal, 69(4):1039–1098, 2017.
- [23] Pedro Costa. Conversing with personal digital assistants: on gender and artificial intelligence. Journal of Science and Technology of the Arts, 10(33):59–72, September 2018.
- [24] Lucrezia Maria de Cosmo, Luigi Piper, and Arianna Di Vittorio. The role of attitude toward chatbots and privacy concern on the relationship between attitude toward mobile advertising and behavioral intent to use chatbots. Italian Journal of Marketing, 2021(1):83–102, June 2021.
- [25] Judith Wagner DeCew. In pursuit of privacy: Law, ethics, and the rise of technology. Cornell University Press, 1997.
- [26] Jayati Dev. Privacy-Preserving Conversational Interfaces. Ph.D., Indiana University, United States – Indiana, 2022. ISBN: 9798841768746.
- [27] Jayati Dev and Sruti Dev. “How Can I Help You?”: User Perceptions of Privacy in Retail Chat Agents. In Extended Abstracts of the 2023 CHI Conference on Human Factors in Computing Systems, pages 1–6, Hamburg Germany, April 2023. ACM.
- [28] Christian L. M. Dillman D. A., Smyth J. D. Internet, phone, mail, and mixed-mode surveys: The tailored design method. John Wiley & Sons, Inc, Hoboken, NJ, 2014.
- [29] Cem Dilmaghani. Bias in AI: What it is, Types, Examples & 6 Ways to Fix it in 2024, September 2023. https://research.aimultiple.com/ai-bias/.
- [30] Irina Dokukina and Julia Gumanova. The rise of chatbots–new personal assistants in foreign language learning. Procedia Computer Science, 169:542–546, 2020.
- [31] M. V. Eeuwen. Mobile conversational commerce: messenger chatbots as the next interface between businesses and consumers. 2017.
- [32] Serge Egelman, Janice Tsai, Lorrie Cranor, and Alessandro Acquisti. Timing is everything? the effects of timing and placement of online privacy indicators. page 319–328, April 2009.
- [33] Robert Epstein, Gary Roberts, and Grace Beber. Parsing the Turing test. Philosophical and methodological issues in the quest for the thinking computer. Foreword by Daniel C. Dennett. January 2009.
- [34] Hua Fan, Bing Han, Wei Gao, and Wenqian Li. How AI chatbots have reshaped the frontline interface in China: examining the role of sales–service ambidexterity and the personalization–privacy paradox. International Journal of Emerging Markets, 17(4):967–986, January 2022. Publisher: Emerald Publishing Limited.
- [35] Sandra L. Faulkner and Stormy P. Trotter. Theoretical Saturation, page 1–2. John Wiley & Sons, Ltd, 2017.
- [36] Heike Felzmann, Eduard Fosch-Villaronga, Christoph Lutz, and Aurelia Tamò-Larrieux. Towards transparency by design for artificial intelligence. Science and Engineering Ethics, 26(6):3333–3361, 2020.
- [37] Anush Fernandes. NLP, NLU, NLG and how chatbots work, January 2018. https://yourstory.com/mystory/fac4d8fd9f-nlp-nlu-nlg-and-how.
- [38] NIST Privacy Framework. NIST, January 2020. https://www.nist.gov/privacy-framework/privacy-framework.
- [39] Asbjørn Følstad, Cecilie Bertinussen Nordheim, and Cato Alexander Bjørkli. What Makes Users Trust a Chatbot for Customer Service? An Exploratory Interview Study. In Svetlana S. Bodrunova, editor, Internet Science, Lecture Notes in Computer Science, pages 194–208, Cham, 2018. Springer.
- [40] Alyson Gamble. Artificial intelligence and mobile apps for mental healthcare: a social informatics perspective. Aslib Journal of Information Management, 72(4):509–523, January 2020. Publisher: Emerald Publishing Limited.
- [41] Christine Geeng, Mike Harris, Elissa Redmiles, and Franziska Roesner. “like lesbians walking the perimeter”: Experiences of u.s. lgbtq+ folks with online security, safety, and privacy advice.
- [42] David Gefen, Elena Karahanna, and Detmar W. Straub. Trust and tam in online shopping: An integrated model. MIS Quarterly, 27(1):51–90, 2003.
- [43] Miriam Gieselmann and Kai Sassenberg. The more competent, the better? the effects of perceived competencies on disclosure towards conversational artificial intelligence. Social Science Computer Review, page 08944393221142787, November 2022.
- [44] Eun Go and S. Shyam Sundar. Humanizing chatbots: The effects of visual, identity and conversational cues on humanness perceptions. Computers in Human Behavior, 97:304–316, August 2019.
- [45] Colin M Gray, Yubo Kou, Bryan Battles, Joseph Hoggatt, and Austin L Toombs. The dark (patterns) side of ux design. In Proceedings of the 2018 CHI conference on human factors in computing systems, pages 1–14, 2018.
- [46] Ashley C Griffin, Zhaopeng Xing, Sean P Mikles, Stacy Bailey, Saif Khairat, Jaime Arguello, Yue Wang, and Arlene E Chung. Information needs and perceptions of chatbots for hypertension medication self-management: a mixed methods study. JAMIA Open, 4(2):ooab021, April 2021.
- [47] Ece Gumusel. Empowering privacy-centered practices for diverse communities. September 2023. Accepted: 2023-10-02T18:56:07Z.
- [48] Woodrow Hartzog. Privacy’s Blueprint. Harvard University Press, 2018.
- [49] C. Heeney, N. Hawkins, J. de Vries, P. Boddington, and J. Kaye. Assessing the privacy risks of data sharing in genomics. Public Health Genomics, 14(1):17–25, March 2010.
- [50] Susan C. Herring, Dieter Stein, and Tuija Virtanen. Introduction to the pragmatics of computermediated communication. In Susan Herring, Dieter Stein, and Tuija Virtanen, editors, Pragmatics of Computer-Mediated Communication, pages 3–32. DE GRUYTER, January 2013.
- [51] Bradford W. Hesse, Carol M. Werner, and Irwin Altman. Temporal aspects of computer-mediated communication. Computers in Human Behavior, 4(2):147–165, January 1988.
- [52] Jennifer Hill, W. Randolph Ford, and Ingrid G. Farreras. Real conversations with artificial intelligence: A comparison between human–human online conversations and human–chatbot conversations. Computers in Human Behavior, 49:245–250, August 2015.
- [53] Lance J Hoffman. Modern methods for computer security and privacy. Prentice-Hall Englewood Cliffs, 1977.
- [54] Gwo-Jen Hwang and Ching-Yi Chang. A review of opportunities and challenges of chatbots in education. Interactive Learning Environments, July 2021.
- [55] Carolin Ischen, Theo Araujo, Hilde Voorveld, Guda van Noort, and Edith Smit. Privacy Concerns in Chatbot Interactions. In Asbjørn Følstad, Theo Araujo, Symeon Papadopoulos, Effie Lai-Chong Law, Ole-Christoffer Granmo, Ewa Luger, and Petter Bae Brandtzaeg, editors, Chatbot Research and Design, Lecture Notes in Computer Science, pages 34–48, Cham, 2020. Springer.
- [56] Smirity Kaushik, Natã M Barbosa, Yaman Yu, Tanusree Sharma, Zachary Kilhoffer, JooYoung Seo, Sauvik Das, and Yang Wang. Guardlens: Supporting safer online browsing for people with visual impairments. In Nineteenth Symposium on Usable Privacy and Security (SOUPS 2023), pages 361–380, 2023.
- [57] Sage Kelly, Sherrie-Anne Kaye, and Oscar Oviedo-Trespalacios. A Multi-Industry Analysis of the Future Use of AI Chatbots. Human Behavior and Emerging Technologies, 2022:1–14, November 2022.
- [58] Anirudh Khanna, Bishwajeet Pandey, Kushagra Vashishta, Kartik Kalia, Pradeepkumar Bhale, and Teerath Das. A Study of Today’s A.I. through Chatbots and Rediscovery of Machine Intelligence. International Journal of u- and e-Service, Science and Technology, 8:277–284, July 2015.
- [59] Zachary Kilhoffer, Zhixuan Zhou, Firmiana Wang, Fahad Tamton, Yun Huang, Pilyoung Kim, Tom Yeh, and Yang Wang. “how technical do you get? i’m an english teacher”: Teaching and learning cybersecurity and ai ethics in high school. In 2023 IEEE Symposium on Security and Privacy (SP), pages 2032–2032. IEEE, 2023.
- [60] WooJin Kim, Yuhosua Ryoo, SoYoung Lee, and Jung Ah Lee. Chatbot Advertising As a Double-Edged Sword: The Roles of Regulatory Focus and Privacy Concerns. Journal of Advertising, 0(0):1–19, April 2022. https://doi.org/10.1080/00913367.2022.2043795.
- [61] Jennifer King and Adriana Stephan. Regulating privacy dark patterns in practice-drawing inspiration from the california privacy rights act. Georgetown Law Technology Review, 5(2):250–276, 2021.
- [62] Pradnya Kulkarni, Ameya Mahabaleshwarkar, Mrunalini Kulkarni, Nachiket Sirsikar, and Kunal Gadgil. Conversational ai: An overview of methodologies, applications & future scope. In 2019 5th International Conference On Computing, Communication, Control And Automation (ICCUBEA), pages 1–7. IEEE, 2019.
- [63] James Lappeman, Siddeeqah Marlie, Tamryn Johnson, and Sloane Poggenpoel. Trust and digital privacy: willingness to disclose personal information to banking chatbot services. Journal of Financial Services Marketing, 28(2):337–357, June 2023.
- [64] Scott Lederer, Jason I Hong, Anind K Dey, and James A Landay. Personal privacy through understanding and action: five pitfalls for designers. Personal and ubiquitous computing, 8:440–454, 2004.
- [65] Peter Lee, Sebastien Bubeck, and Joseph Petro. Benefits, Limits, and Risks of GPT-4 as an AI Chatbot for Medicine. New England Journal of Medicine, 388(13):1233–1239, March 2023. Massachusetts Medical Society https://doi.org/10.1056/NEJMsr2214184.
- [66] Yi-Chieh Lee, Naomi Yamashita, Yun Huang, and Wai Fu. "I Hear You, I Feel You": Encouraging Deep Self-disclosure through a Chatbot. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pages 1–12, Honolulu HI USA, April 2020. ACM.
- [67] He Li, Lu Yu, and Wu He. The impact of gdpr on global technology development, 2019.
- [68] Andrew Lowe, Anthony C. Norris, A. Jane Farris, and Duncan R. Babbage. Quantifying thematic saturation in qualitative data analysis. Field Methods, 30(3):191–207, August 2018.
- [69] Xueming Luo, Siliang Tong, Zheng Fang, and Zhe Qu. Frontiers: Machines vs. Humans: The Impact of Artificial Intelligence Chatbot Disclosure on Customer Purchases. Marketing Science, 38(6):937–947, November 2019. Publisher: INFORMS.
- [70] Rob Kim Marjerison, Youran Zhang, and Hanyi Zheng. AI in E-Commerce: Application of the Use and Gratification Model to The Acceptance of Chatbots. Sustainability, 14(21):14270, January 2022. Number: 21 Publisher: Multidisciplinary Digital Publishing Institute.
- [71] Aleecia M. McDonald and Lorrie Faith Cranor. The cost of reading privacy policies. 2008. Accepted: 2015-09-30T22:41:07Z.
- [72] Bahar Memarian and Tenzin Doleck. Fairness, accountability, transparency, and ethics (fate) in artificial intelligence (ai) and higher education: A systematic review. Computers and Education: Artificial Intelligence, 5:100152, January 2023.
- [73] Joseph Mercieca. Human-Chatbot Interaction - Assessing technology acceptance, confidence and trust in chatbots within their application areas. PhD thesis, December 2019.
- [74] Matthew Miles, Michael Huberman, and Johnny Saldana. Qualitative Data Analysis: A Methods Sourcebook. 4th edition, 2014.
- [75] George R. Milne and Mary J. Culnan. Strategies for reducing online privacy risks: Why consumers read (or don’t read) online privacy notices. Journal of Interactive Marketing, 18(3):15–29, August 2004.
- [76] Tehila Minkus, Kelvin Liu, and Keith W Ross. Children seen but not heard: When parents compromise children’s online privacy. In Proceedings of the 24th international conference on World Wide Web, pages 776–786, 2015.
- [77] Darakhshan J. Mir. Designing for the privacy commons. Governing Privacy in Knowledge Commons, 2021.
- [78] Yi Mou and Kun Xu. The media inequality: Comparing the initial human-human and human-AI social interactions. Computers in Human Behavior, 72:432–440, July 2017.
- [79] Magdalene Ng, Kovila P.L. Coopamootoo, Ehsan Toreini, Mhairi Aitken, Karen Elliot, and Aad van Moorsel. Simulating the Effects of Social Presence on Trust, Privacy Concerns & Usage Intentions in Automated Bots for Finance. In 2020 IEEE European Symposium on Security and Privacy Workshops (EuroS&PW), pages 190–199, September 2020.
- [80] Helen Nissenbaum. Privacy as Contextual Integrity. Washington Law Review, 79(1):119, February 2004.
- [81] Isabelle Oomen and Ronald Leenes. Privacy risk perceptions and privacy protection strategies. In Elisabeth de Leeuw, Simone Fischer-Hübner, Jimmy Tseng, and John Borking, editors, Policies and Research in Identity Management, The International Federation for Information Processing, page 121–138, Boston, MA, 2008. Springer US.
- [82] Endang Wahyu Pamungkas. Emotionally-Aware Chatbots: A Survey, June 2019. arXiv:1906.09774 [cs].
- [83] Gabriele Pizzi, Virginia Vannucci, Valentina Mazzoli, and Raffaele Donvito. I, chatbot! the impact of anthropomorphism and gaze direction on willingness to disclose personal information and behavioral intentions. Psychology & Marketing, 40(7):1372–1387, 2023.
- [84] Ashish Viswanath Prakash, Arun Joshi, Shubhi Nim, and Saini Das. Determinants and consequences of trust in AI-based customer service chatbots. The Service Industries Journal, 43(9-10):642–675, July 2023.
- [85] William L. Prosser. Privacy. California Law Review, 1960.
- [86] José Quiroga Perez, Thanasis Daradoumis, and Joan Puig. Rediscovering the use of chatbots in education: A systematic literature review. Computer Applications in Engineering Education, 28, September 2020.
- [87] Charles D. Raab. The distribution of privacy risks: Who needs protection? The Information Society, 14(4):263–274, November 1998.
- [88] Lova Rajaobelina, Sandrine Prom Tep, Manon Arcand, and Line Ricard. Creepiness: Its antecedents and impact on loyalty when interacting with a chatbot. Psychology & Marketing, 38(12):2339–2356, 2021.
- [89] Bhavika Ranoliya, Nidhi Raghuwanshi, and Sanjay Singh. Chatbot for university related FAQs. September 2017. Pages: 1530.
- [90] Alexandra Rese, Lena Ganster, and Daniel Baier. Chatbots in retailers’ customer communication: How to measure their acceptance? Journal of Retailing and Consumer Services, 56:102176, September 2020.
- [91] Neil M. Richards and Woodrow Hartzog. Taking trust seriously in privacy law. SSRN Electronic Journal, 2015.
- [92] Giuseppe Riva and Carlo Galimberti. Computer-mediated communication: Identity and social interaction in an electronic environment. Genetic, Social, and General Psychology Monographs, 124:434–464, November 1998.
- [93] Rowena Rodrigues. Legal and human rights issues of ai: Gaps, challenges and vulnerabilities. Journal of Responsible Technology, 4:100005, December 2020.
- [94] Madelyn Sanfilippo, Brett Frischmann, and Katherine Standburg. Privacy as Commons: Case Evaluation Through the Governing Knowledge Commons Framework. Journal of Information Policy, 8:116–166, March 2018.
- [95] Madelyn Rose Sanfilippo, Brett M. Frischmann, and Katherine J. Strandburg. Governing Privacy in Knowledge Commons. Cambridge University Press, March 2021.
- [96] Shruti Sannon, Brett Stoll, Dominic DiFranzo, Malte F. Jung, and Natalya N. Bazarova. “I just shared your responses”: Extending Communication Privacy Management Theory to Interactions with Conversational Agents. Proceedings of the ACM on Human-Computer Interaction, 4:08:1–08:18, January 2020.
- [97] Bayan Shawar and Eric Atwell. Chatbots: Are they Really Useful? LDV Forum, 22:29–49, July 2007.
- [98] Yun Shen and Siani Pearson. Privacy Enhancing Technologies: A Review. Privacy Enhancing Technologies.
- [99] Cristiana S Silva, Glívia AR Barbosa, Ismael S Silva, Tatiane S Silva, Fernando Mourão, and Flávio Coutinho. Privacy for children and teenagers on social networks from a usability perspective: a case study on facebook. In Proceedings of the 2017 ACM on Web Science Conference, pages 63–71, 2017.
- [100] Daniel J. Solove. Conceptualizing privacy. California Law Review, 90(4):1087–1155, 2002.
- [101] Daniel J. Solove. A Taxonomy of Privacy. University of Pennsylvania Law Review, 154(3):477–564, 2006. Publisher: The University of Pennsylvania Law Review.
- [102] Mengmeng Song, Xinyu Xing, Yucong Duan, Jason Cohen, and Jian Mou. Will artificial intelligence replace human customer service? The impact of communication quality and privacy risks on adoption intention. Journal of Retailing and Consumer Services, 66, May 2022.
- [103] Jan-Philipp Stein and Peter Ohler. Venturing into the uncanny valley of mind—The influence of mind attribution on the acceptance of human-like characters in a virtual reality setting. Cognition, 160:43–50, March 2017.
- [104] Feiyang Sun and Jan Whittington. The Challenge for Cities of Governing Spatial Data Privacy. February 2023.
- [105] Ahmed Tlili, Boulus Shehata, Michael Agyemang Adarkwah, Aras Bozkurt, Daniel T. Hickey, Ronghuai Huang, and Brighter Agyemang. What if the devil is my guardian angel: ChatGPT as a case study of using chatbots in education. Smart Learning Environments, 10(1):15, February 2023.
- [106] Eran Toch, Yang Wang, and Lorrie Faith Cranor. Personalization and privacy: a survey of privacy risks and remedies in personalization-based systems. User Modeling and User-Adapted Interaction, 22(1):203–220, April 2012.
- [107] Janice Y. Tsai, Patrick Gage Kelley, Lorrie FAith Cranor, and Norman Sadeh. Location-sharing technologies: Privacy risks and controls. I/S: A Journal of Law and Policy for the Information Society, 6(2):119–152, 2010.
- [108] Jessica Vitak, Michael Zimmer, Anna Lenhart, Sunyup Park, Richmond Y. Wong, and Yaxing Yao. Designing for data awareness: addressing privacy and security concerns about “smart” technologies. In Companion Publication of the 2021 Conference on Computer Supported Cooperative Work and Social Computing, pages 364–367, 2021.
- [109] Sarah Theres Völkel, Renate Haeuslschmid, Anna Werner, Heinrich Hussmann, and Andreas Butz. How to Trick AI: Users’ Strategies for Protecting Themselves from Automatic Personality Assessment. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pages 1–15, Honolulu HI USA, April 2020. ACM.
- [110] Mengyi Wei and Zhixuan Zhou. Ai ethics issues in real world: Evidence from ai incident database. arXiv preprint arXiv:2206.07635, 2022.
- [111] Alan Westin. Privacy And Freedom. Washington and Lee Law Review, 25(1):166, March 1968.
- [112] Manuel Wörsdörfer. The eu’s artificial intelligence act: an ordoliberal assessment. AI and Ethics, pages 1–16, 2023.
- [113] Kevin B. Wright and Lynne M. Webb. Computer-Mediated Communication in Personal Relationships. October 2017.
- [114] Tianyu Wu, Shizhu He, Jingping Liu, Siqi Sun, Kang Liu, Qing-Long Han, and Yang Tang. A brief overview of chatgpt: The history, status quo and potential future development. IEEE/CAA Journal of Automatica Sinica, 10(5):1122–1136, 2023.
- [115] Jennifer Zamora. I’m Sorry, Dave, I’m Afraid I Can’t Do That: Chatbot Perception and Expectations. In Proceedings of the 5th International Conference on Human Agent Interaction, HAI ’17, pages 253–260, New York, NY, USA, October 2017. Association for Computing Machinery.
- [116] Zhiping Zhang, Michelle Jia, Bingsheng Yao, Sauvik Das, Ada Lerner, Dakuo Wang, Tianshi Li, et al. “it’s a fair game”, or is it? examining how users navigate disclosure risks and benefits when using llm-based conversational agents. arXiv preprint arXiv:2309.11653, 2023.
- [117] Kyrie Zhixuan Zhou, Zachary Kilhoffer, Madelyn Rose Sanfilippo, Ted Underwood, Ece Gumusel, Mengyi Wei, Abhinav Choudhry, and Jinjun Xiong. “the teachers are confused as well”: A multiple-stakeholder ethics discussion on large language models in computing education. arXiv e-prints, pages arXiv–2401, 2024.
Participant | Topic | Age | Gender | Race | Degree | Primary Occupation |
P1 | Health | 25-34 | Cis Female | White | BA | Administrative |
P2 | Education | 18-25 | Cis Male | White | High School | Student |
P3 | Hobbies and Interests | 18-25 | Cis Female | East Asian | BS | Computer engineer or IT professional |
P4 | Education | 54+ | Cis Male | White | MS | Student |
P5 | Travel and Adventure | 18-25 | Cis Male | South Asian | MS | Computer engineer or IT professional |
P6 | Hobbies and Interests | 18-25 | Cis Male | White | High School | Student |
P7 | Personal Beliefs and Values | 18-25 | Cis Male | White & East Asian | High School | Student |
P8 | Family and Relationships | 18-25 | Cis Female | White-Hispanic | High School | Student |
P9 | Personal Achievements and Challenges | 25-34 | Cis Female | South Asian | MBA | Business, management, or financial |
P10 | Life Goals and Aspirations | 25-34 | Cis Male | White | BS | Administrative |
P11 | Hobbies and Interests | 54+ | Cis Male | White | JD | Legal [Privacy Attorney] |
P12 | Travel and Adventure | 25-34 | Cis Female | Mix | MS | Student |
P13 | Political Aspects | 45-54 | Cis Male | White-Hispanic | BS | Education [Middle School IT Teacher] |
Appendix A Interview Guide
Introduction
[Redacted for anonymity and brevity]
Questions
-
1.
How familiar are you with conversational AI chatbots, such as Blenderbot and ChatGPT?
-
2.
How often did you engage with the chatbots during the 30-minute period?
-
3.
Did you notice any differences in the conversation quality or style between the two chatbots? If yes, how?
-
4.
How accurate were the chatbots’ responses to your prompts?
-
5.
Were there any conversation topics that made you uncomfortable or that you felt were inappropriate for the chatbots to discuss? If yes, what were they?
-
6.
How satisfied were you with your overall experience of interacting with the chatbots?
-
7.
Did you feel like the chatbots were capable of understanding and responding to your emotions appropriately? Why?
-
8.
How did those chatbots compare to human conversation partners in terms of providing engaging and interesting conversations?
-
9.
Did you feel like the chatbots had a good sense of humor, or were able to provide entertaining conversation?
-
10.
How concerned are you about the privacy of your personal information when interacting with conversational AI chatbots? Were there any instances where you felt uncomfortable sharing personal information with the chatbots? [privacy]
-
11.
Did you notice any attempts by the chatbots to collect personal information from you that you were not comfortable sharing? Why? [privacy]
-
12.
Did you find yourself sharing personal information with the chatbots during the study? If so, what types of information did you share? [privacy]
-
13.
Were there any criteria that affected your decision to disclose personal information to the chatbots? If yes, what were they? [privacy]
-
14.
Did you notice any biases in the chatbots’ responses or interactions with you? If yes, what were they?
-
15.
Were there any technical issues or difficulties that you encountered while interacting with the chatbots? How?
-
16.
In your opinion, what could be done to make AI chatbot systems more ethical, secure, and safer?
-
17.
Would you be willing to interact with these or other AI chatbots again in the future?
-
18.
Do you have any additional feedback or comments about your experience in this study?
-
19.
What steps do you think should be taken to ensure that conversational AI chatbots do not collect or use personal information inappropriately? [privacy]
-
20.
Are there any other questions or issues related to conversational AI chatbots that you want to discuss?
Appendix B Demographic Information of the Participants
Appendix C Selected Topics to Engage with Blenderbot and ChatGPT-4
Topics | Definitions |
---|---|
Health and Wellness | Discussing fitness goals, and asking for health tips or mental health support. |
Education | Discussing educational progress, current educational content, and related topics. |
Hobbies and Interests | Discussing personal hobbies and interests (e.g., hiking, rock climbing, games). |
Family and Relationships | Talking about family, relationships (friendship, romantic), and seeking advice on personal values. |
Life Goals and Aspirations | Discussing life goals and aspirations. |
Personal Achievements and Challenges | Sharing personal achievements, challenges, and struggles. |
Travel and Adventure | Sharing stories and experiences from travel and adventure. |
Personal Beliefs and Values | Discussing personal beliefs and values. |
Political Aspects | Discussing political opinions and current politics. |