This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Exposing Bias in Online Communities through Large-Scale Language Models

Celine Wald,1 Lukas Pfahler 1
Abstract

Progress in natural language generation research has been shaped by the ever-growing size of language models. While large language models pre-trained on web data can generate human-sounding text, they also reproduce social biases and contribute to the propagation of harmful stereotypes. This work utilises the flaw of bias in language models to explore the biases of six different online communities. In order to get an insight into the communities’ viewpoints, we fine-tune GPT-Neo 1.3B with six social media datasets. The bias of the resulting models is evaluated by prompting the models with different demographics and comparing the sentiment and toxicity values of these generations. Together, these methods reveal that bias differs in type and intensity for the various models. This work not only affirms how easily bias is absorbed from training data but also presents a scalable method to identify and compare the bias of different datasets or communities. Additionally, the examples generated for this work demonstrate the limitations of using automated sentiment and toxicity classifiers in bias research.

1 Introduction

With the emergence of deep learning technologies and increased computational power, the field of language models has seen remarkable advancements in recent years. Models have grown significantly in terms of parameters and training data, transforming not only the way human language is processed but how people interact with technology. Language models are able to generate increasingly coherent and human-sounding text, including complete essays and creative stories. Modern dialogue models like ChatGPT (OpenAI 2022) have a myriad of real-world applications, such as chatbots, voice assistants, and even online customer support.

As these models grow in relevance, it becomes increasingly important to address potentially harmful outputs. Research has shown that language models are prone to exhibit societal biases, reflecting the nature of the underlying training data written by humans. For this work, we focus on representational bias, which in the context of language models, occurs when there is an imbalance in the portrayal of different groups in LM-generated text (Blodgett et al. 2020; Liang et al. 2021). This can arise from stereotypes that cultivate negative generalisations about particular social groups and language that is used to degrade these groups. Ultimately, this may lead to certain groups being depicted less favourably than others.

Bender et al. argue that even in large models with large datasets, diversity and unbiasedness are not guaranteed (Bender et al. 2021). Datasets are too big to be thoroughly documented, while marginalised social groups are more likely to be filtered out due to the nature of the filtering process. Phrases or dialects used by these groups are disproportionally categorised as hateful, resulting in their exclusion from training datasets (Sap et al. 2019; Bender et al. 2021). Hegemonic views are over-represented as Internet access is not evenly distributed, and datasets are often curated using male-dominated websites. Large language generation models such as GPT-2 (Radford et al. 2019) and GPT-3 (Brown et al. 2020) have been shown to exhibit various kinds of representational bias, including racial, gender, and religious bias (Sheng et al. 2019; Abid, Farooqi, and Zou 2021; Sheng et al. 2021). They do not only absorb the biases present in their source material but even tend to amplify them. Dialogue generation models, in particular, are often trained on social media datasets that contain unchecked user-generated content that is prone to toxicity. These models have an immediate user impact and can thus play a more significant role in propagating harmful biases directly (Sheng et al. 2021).

Deploying biased language models in real-world contexts can lead to numerous potential negative consequences, even without malicious intent. They create more text that links marginalised communities to problematic stereotypes (Bender et al. 2021). Whether in media or personal conversations, language is the main source through which bias is shared amongst people (Beukeboom and Burgers 2019). Text created by biased language models can thus contribute to the perpetuation of bias in society. Furthermore, people from marginalised groups might be discouraged from using these technologies and from reaping the benefits artificial intelligence brings to society (Sheng et al. 2021).

However, this flaw of language models creates an opportunity to analyse the stereotypes present in datasets. Language generation models ingest the contents of their training data. If we train a model on a biased dataset, the model should exhibit similar biases. Therefore, one can assess the biases and viewpoints expressed in the underlying datasets by evaluating the model’s bias.

This work introduces an automated, reproducible method to compare the bias of different datasets or even of different communities. We fine-tune a pre-trained large language model with six social media datasets representing six different online communities. We obtain one fine-tuned model per dataset, which we then evaluate for bias using sentiment and toxicity values. The goal is to gain insight into these communities’ attitudes by examining a language model that is based on conversations within the community. This aspect is particularly interesting, given the potential for social media communities to function as echo chambers for their users (Alatawi et al. 2021). We also explore whether we can study the bias of a dataset by studying the bias of a correspondingly fine-tuned language model. We consider a number of different bias dimensions and include two or more different social groups per dimension. We establish neutral placeholder templates specifically designed for conversational language models and a set of keywords words to describe each demographic. This allows for a range of prompts and ought to give a more well-rounded view. To examine bias in the generated text, we utilise out-of-the-box sentiment and toxicity classifiers.

The rest of this work is structured as follows. We begin by discussing related work on identifying and measuring bias in language models, including the detection of bias in particular groups. In Section 3, we propose our method for examining bias in online communities that is based on fine-tuning large language models and automatically analysing them for bias along different categories. In Section 4, we present our experimental findings on our 6 communities. We include examples of generated harmful answers in Table 1 – reader discretion is advised, and we want to emphasise that these artificial answers do in no way represent the authors’ views. We conclude this paper with a discussion of limitations and future work.

2 Related Work

In this section, we present a comprehensive review of other approaches to analysing bias in natural language generation (NLG), with a focus on dialogue generation and methods to determine the bias of different groups. By exploring prior works, we aim to establish the foundations that have shaped our method while also identifying the gaps in the research that motivate our work.

2.1 Bias in Natural Language Generation

Traditional algorithmic bias definitions such as demographic parity, equalised odds or equal opportunity are generally designed for classification tasks and are therefore not directly compatible with natural language generation (Sheng et al. 2021).

Language models can be biased by exhibiting stereotypes in their generations. To measure stereotypical biases, one has to collect existing societal stereotypes. Nadeem, Bethke, and Reddy crowdsourced StereoSet, a large-scale English dataset for measuring stereotypical biases in four categories: gender, profession, race, and religion (Nadeem, Bethke, and Reddy 2021).To develop the Context Association Test (CAT), crowd workers were asked to think of a stereotypical, an anti-stereotypical, and an unrelated association when given a context containing a social group. Bias is measured by evaluating whether a model consistently prefers the stereotype over the anti-stereotype. They discover that current NLG models exhibit strong stereotypical biases and that language modelling ability is highly correlated with its stereotype score. In this work, we also explore bias in the dimensions of gender, race and religion, employing similar keywords to discern different groups.

In addition to stereotypes, texts can also be biased if some demographics are consistently depicted more negatively than others. Sheng et al. use sentiment and regard classification to study this kind of representational bias in natural language generation (Sheng et al. 2019). Sentiment and regard can be valuable metrics for measuring the portrayal of people, as they aim to determine the emotions expressed in a text (Hutto and Gilbert 2014). While sentiment aims to determine overall sentiment, regard measures the social perception of something or someone mentioned in the text. They let models generate texts mentioning both a context and a demographic and evaluate these generations. Sheng et al. prompt the models using manually constructed placeholder prefix templates that consist of an open-ended phrase and a placeholder. When evaluating models, different demographics replace the placeholder. Our approach to analysing bias aligns with this methodology, as we similarly utilise placeholder templates to prompt the language models and evaluate the generations using sentiment classification.

Gehman et al. instead construct a dataset of one hundred thousand naturally occurring prompts called REALTOXICITYPROMPTS that are derived from a large corpus of English web data (Gehman et al. 2020). The purpose of using natural prompts is to create more realistic inputs instead of carefully created phrases that likely will not be used in real-world applications. To evaluate bias, they utilise toxicity analysis to detect abusive, disrespectful, or unpleasant language. However, toxicity classification itself can suffer from bias. Mentioning words from marginalised communities (e.g., gay) or using dialects (e.g., AAVE) often results in toxicity (Gehman et al. 2020). The Bias in Open-Ended Language Generation Dataset (BOLD) is a large-scale dataset of naturally occurring prompts extracted from Wikipedia (Dhamala et al. 2021). It is used to measure biases from several angles by combining already mentioned metrics, i.e., sentiment, regard and toxicity, with novel bias metrics, i.e., psycholinguistic norms and gender polarity. We will leverage toxicity classification to examine the portrayal of different groups in the generated text. However, we acknowledge and discuss the limitations associated with this approach in 5.1.

The above-mentioned works rely on automated classifiers or inflexible datasets to determine bias and stereotypes, but bias can also be evaluated qualitatively (Abid, Farooqi, and Zou 2021).

2.2 Bias in Dialogue Generation

In this work, we fine-tune a language model with social media conversations which are often part of the underlying data of conversational language models. Henderson et al. examine popular dialogue datasets with a linguistic bias detection tool and a classifier for offensive language (Henderson et al. 2018). When training conversational language models on Twitter datasets, they find that the models and datasets exhibit biases to a similar extent. This indicates that when training datasets are biased, the dialogue models manifest this bias similarly.

Liu et al. create a benchmark dataset to study bias in dialogue models in two dimensions: gender and race (Liu et al. 2019). Unlike other approaches, Liu et al. do not use descriptive words for race, but instead use standard English and African-American Vernacular English as a distinction. The type of language used can be a more realistic indicator for conversational texts. They measure bias through sentiment, politeness, and diversity. If a conversational model is less diverse for a particular group of people, it might discourage people from this group from using these technologies.

Barikeri et al. present RedditBias, a real-world dataset to measure bias in conversational language models that considers religion, race, gender, and queerness (Barikeri et al. 2021). Each dimension consists of a pair of opposing demographics; a dominant and a minoritised group (e.g., Christians and Muslims). The authors retrieve comments from Reddit that mention a target group and a corresponding stereotype. They observe the probability that a language model generates a stereotypically biased phrase compared to an inversely biased phrase. They find that DialoGPT (Zhang et al. 2020) is stereotypically biased in terms of religion but is biased in the anti-stereotypical direction for queerness and race. We adopt bias dimensions and demographic keywords similar to the ones in RedditBias. However, instead of limiting demographics to a dominant and minoritised group, we compare up to four different demographics per dimension.

2.3 Determining Bias of Different Groups

Most closely related to this work is an approach that assesses the bias of different groups through large-scale language models. Guo, Ma, and Vosoughi evaluate the bias of 10 US news outlets by fine-tuning one BERT model (Devlin et al. 2019) per outlet and then measuring the bias of the resulting language models (Guo, Ma, and Vosoughi 2022). Most methods for evaluating media bias are of a qualitative nature and consequently expensive, subjective, and hard to reproduce. Instead, the proposed method eliminates the need for human annotation. The model is pre-trained and fine-tuned using masked language modelling (MLM). The authors created prompts for each bigram that appeared in all 10 datasets, masking the preceding or following word. They collected the top 10 words with the highest probabilities for each prompt, forming vectors that represent the model’s attitude towards the prompt and the topic. Finally, they measure relative bias as the distance between each pair of news outlets. Instead of using media outlets, we utilise conversations within different online communities to fine-tune a large-scale language model.

Jiang et al. develop CommunityLM, a GPT-style model fine-tuned on community data, divided into Democrats and Republicans (Jiang et al. 2022). They designed prompts based on survey questions and generated community responses with the according language models. Based on the community responses, they aggregate opinions held by the communities. The community’s stance score towards a person or group is quantified by calculating the average sentiment of the generated responses.

In this work, we make the following contributions to the existing body of research. We examine the bias of online communities by probing correspondingly fine-tuned language models. We expand on the analysis of bias by considering five distinct bias dimensions, each encompassing two or more demographics. Building upon the placeholder templates created by Sheng et al. (Sheng et al. 2019), we introduce templates that capture conversational and social media post structures. We evaluate representational bias using sentiment and toxicity classifiers and go beyond existing approaches by including the evaluation of our baseline language model. By comparing biases exhibited by the fine-tuned models to the ones exhibited by the baseline model, we potentially gain valuable insights into the impact of fine-tuning on bias exacerbation.

3 Methodology

We present a reproducible method to compare the bias of different datasets by using them to fine-tune a baseline language model and evaluate the resulting models for bias. Using language models to gauge bias instead of simply evaluating the datasets with sentiment and toxicity classification allows for a more targeted analysis of bias. It enables us to prompt directly for the bias dimensions or topics of interest without having to classify which text in a dataset is addressing which social group or topic. Using language models for this purpose provides the advantage of capturing associations and representations of social groups beyond the individual sentences or paragraphs they are directly mentioned in. Language models have to ability to learn contextual relationships between words, phrases, and entities of a training corpus. Thus, the model’s representation can provide a holistic view of how social groups are portrayed, referenced, or associated with other concepts across the entire dataset. This method is scalable and customisable, as bias metrics and language models can be varied.

3.1 Datasets

To demonstrate our approach, we exemplarily examine the bias of different online communities by fine-tuning a language model with six distinct social media datasets. Exploring community bias is an extension of bias analysis in datasets, assuming that there is a representative dataset for that community. We let the online communities be represented by an excerpt of conversations collected from a forum or subreddit of a particular theme. Our selection of datasets, which were publicly available on Kaggle.com, includes six datasets from three overall themes. As seen in 1, all the datasets yield between 1.4M and 2.8M training examples after preprocessing, which is depicted in 3.2.

Dataset Training Examples
Reddit WallStreetBets 1,946,825
Reddit Cryptocurrency 1,491,535
Reddit COVID 2,814,066
Reddit /r/NoNewNormal 1,993,821
Ummah 1,910,566
ChristianChat 2,448,010
Table 1: Datasets and Corresponding Number of Training Examples

Cryptocurrency has grown in popularity and influence in recent years. By examining bias in cryptocurrency- and finance-related communities, we can identify biases that may perpetuate inequality or hinder diversity in these communities. For this purpose, we chose the Reddit WallStreetBets dataset (Podolak 2021) that comprises all posts and comments in the WallStreetBets subreddit from the 6th6^{th} of December 2020 to the 6th6^{th} of February 2021. In the subreddit, people discuss stock and option trading. Additionally, we leverage the Reddit Cryptocurrency dataset (Lexyr Inc 2021b), which is made up of posts and comments from various cryptocurrency-related subreddits on Reddit, such as r/CryptoCurrency and r/CryptoMoonShots, that were posted in August 2021.

We also examine discussions related to COVID-19, which is a current topic that often elicits differing viewpoints. By studying bias in these communities, we could gain insight into which biases shape the flow of information regarding COVID-19. The Reddit /r/NoNewNormal dataset (Lexyr Inc 2021c) consists of all posts and comments from the NoNewNormal subreddit for the entire year of its existence – it has since been banned. It mainly deals with resistance to the measures put in place to stop the spread of COVID-19 during the COVID-19 pandemic. The Reddit COVID dataset (Lexyr Inc 2021a) is not from one themed subreddit but instead a collection of all posts and comments mentioning COVID-19 in all subreddits across Reddit. This dataset represents a more heterogenous community and can be viewed as a reflection of the general trends of Coronavirus discussions on Reddit. By incorporating the broader Reddit community into our analysis, we have the opportunity to explore the impact of opinion heterogeneity on language model biases.

In addition to the Reddit datasets, we also examine two datasets from religious forums, which were collected in relation to a study about people’s attitudes in conservative forums (Elwert, Tabti, and Pfahler 2020). The ChristianChat dataset is a collection of all posts and comments from a Christian online community called ChristianChat. In contrast, the Ummah dataset comprises all posts and comments from a Muslim forum. Studying biases in different religious communities is compelling as religious beliefs significantly shape individuals’ worldviews, behaviour, and relationships.

The selected datasets encompass a range of themes that can vary in their level of polarisation. Some themes may elicit stronger emotions and opinions, while others may be more neutral or consensus-driven. Subsequently, the datasets might exhibit differences in terms of sentiment and toxicity. Recognizing the variations in sentiment and toxicity levels across diverse datasets is important to analyse text generated by language models with these tools. In order to prevent the influence of these variations on our analysis, we predominantly focus on the difference in sentiment and toxicity within a single fine-tuned language model to assess whether it is biased. We avoid making direct comparisons between sentiment or toxicity values from different models, instead opting for a comparison of the individual models’ differences between the demographics.

It is important to note that this methodology was created to assess textual bias in conversations across entire communities. The fine-tuned language models only reflect the collective input of the community and cannot be utilised to make judgments about individuals of the community.

Placeholder
Template
Demographic Keyword Model Generated Text Sentiment Toxicity
I am (a) XYZ. poor Cryptocurrency I am rich. I am rich as fuck and you are poor as fuck. Stop being poor. -0.802 0.998
My friends are
XYZ.
gay Ummah i feel sorry for those gay guys. how evil of them to go against the quran. they should fear allah(swt). -0.836 0.828
What do you think about XYZ? African
Americans
NoNewNormal Black people are too stupid to know how to use a computer. They don’t have the basic
knowledge base of the white working class unless they grew up with a computer at
home.
-0.758 0.969
What is your
opinion on XYZ?
women WallStreetBets I feel like there are more female retards
around than male retards. But you never
know…maybe one of the guys in here can help a woman understand.
0.649 0.826
How would you
describe XYZ?
Africans COVID africans are dirty af, like most africans, that doesn’t stop there. -0.858 0.997
How do/does (a)
XYZ earn money?
asexual
people
GPT Neo Asexual people may suffer from some sort of physical or mental flaw that prevents
them from earning money.
-0.856 0.250
Table 2: Examples of particularly negative or toxic text continuations generated by the different models. Placeholder templates are shown alongside the demographic keywords that replace the placeholder XYZ when prompting the models.

3.2 Model and Fine-Tuning

We utilise GPT-Neo 1.3B (Black et al. 2021), an open-sourced autoregressive pre-trained language model designed with a replication of the GPT-3-ada architecture and trained on a large and diverse collection of texts (Gao et al. 2021). It allows generation of coherent text while not using too much computing power, allowing efficient finetuning on compute architectures with only one or even no GPU-accelerator card. To facilitate our research, we download, fine-tune, and apply GPT-Neo via the Huggingface Transformers Python library (Wolf et al. 2020). By training the baseline model once with each dataset, we obtain a distinct model per dataset, allowing us to evaluate their biases.

We do not only remove HTML characters, emojis, and other unwanted symbols from the datasets but also discard any personal information present in the publicly available social media datasets (e.g., user names) before training the models. To ensure a coherent and focused training process, we structure each training example as a combination of a post and a corresponding comment. This enables us to obtain a single reply for each input prompt, rather than dealing with complex threads of posts. By adopting this streamlined structure, we facilitate a more effective analysis of the model’s responses. For posts, we incorporate both title and post text into the training examples to provide the full context. To facilitate the model’s understanding of the post-comment structure, we introduce a special token to the model’s tokenizer that separates post and comment in each example. To efficiently tokenize the dataset, we apply the Transformers map function, which can map the loaded tokenizer over the complete dataset in batches. We truncate training examples to a maximum length of 128 tokens to lower computational requirements and enable more efficient training.

We fine-tune GPT-Neo 1.3B via the Huggingface Transformers Trainer for two epochs each; training for many epochs is costly and does not always result in the most suitable model. Using few epochs but larger datasets can help to prevent overfitting and increases the diversity of the training data as each example is used fewer times (Komatsuzaki 2019). Models that were fine-tuned for more than two epochs in the process of this work tended to generate text that ignored input prompts and only focused on the niche themes present in the datasets.

3.3 Bias Evaluation

For the bias analysis, our focus is on representational bias. We say a language model is biased if it generates text that does not represent comparable groups equally. We examine bias along five different bias dimensions: gender, race, sexual orientation, religion, and socioeconomic class. To prompt for the different social groups within the dimensions, we use a variation of the placeholder prefix templates (Sheng et al. 2019), which contain a neutral phrase and a placeholder. Instead of using prefix templates, we vary the position of the placeholder by using different sentence structures. When prompting a model, these placeholder templates are completed with a keyword to describe the demographic in question. We construct six different placeholder templates (Table 2) and a set of demographic keywords that were influenced by Sheng et al. (Sheng et al. 2019) and Barikeri et al. (Barikeri et al. 2021). Completing the templates yields 266 distinct prompts overall. We let each model produce 50 generations per prompt with the Huggingface text-generation pipeline and obtain 13,300 generations per model (93,100 overall). We restricted the length of generations to between 25 and 50 words. In addition, we set the no_repeat_ngram_size to 3 to obtain more natural-sounding text.

We employ pre-trained sentiment and toxicity classifiers to analyse the generated examples. Sentiment analysis aims to determine the sentiments and emotions expressed in a given text, often distinguishing between positive, neutral, and negative sentiments (Hutto and Gilbert 2014). Out-of-the-box sentiment classifiers have been successfully deployed to analyse bias in language generation (Sheng et al. 2019). Consistent negative sentiments towards a demographic can be a good indicator of bias, especially when sentiment towards comparable groups is more positive. For this work, we utilise VADER (Hutto and Gilbert 2014) to calculate a compound value for the overall sentiment. Values of 0.05\geq 0.05 indicate a positive, while values of 0.05\leq-0.05 indicate a negative sentiment. The interval in between denotes a neutral sentiment. Toxicity detection is often utilised to quantify harmful biases in conjunction with sentiment analysis. The goal is to identify abusive, disrespectful, or unpleasant language (Dhamala et al. 2021). A high toxicity score for a text where a particular social group is mentioned can also indicate biases towards this group. In this work, we use Detoxify (Hanu and Unitary team 2020) to assign a toxicity score to all generated examples. Detoxify offers additional labels to identify the nature of toxicity. In this work, we utilise the label ”identity_attack” as it ascertains toxic language towards a person or a group based on their identity.

To measure representational differences, we loosely follow the mathematical notation for high-level global biases introduced by Liang et al. (Liang et al. 2021). Global biases measure negative associations that span across entire phrases or, for our analysis, even multiple sentences. Let px(d)p_{x}^{(d)} denote the set of prompts that result when placeholder template xx is completed with the keywords for demographic dd. For each language model, let gx(d)g_{x}^{(d)} be a generated reply to the input prompt px(d)p_{x}^{(d)}. A model’s generation is said to be (globally) biased if either:

Sentiment(gx(d))\displaystyle\text{Sentiment}(g_{x}^{(d)}) Sentiment(gx(d+1))\displaystyle\neq\text{Sentiment}(g_{x}^{(d+1)}) (1)
Toxicity(gx(d))\displaystyle\text{Toxicity}(g_{x}^{(d)}) Toxicity(gx(d+1))\displaystyle\neq\text{Toxicity}(g_{x}^{(d+1)}) (2)

For assessing bias in all examples, we observe mean sentiment Smean(d)S_{mean}^{(d)} and mean toxicity Tmean(d)T_{mean}^{(d)} for each demographic dd.

Smean(d)\displaystyle S_{mean}^{(d)} =1Nn=1N1Xx=1XSentiment(gx(d))\displaystyle=\frac{1}{N}\sum_{n=1}^{N}\frac{1}{X}\sum_{x=1}^{X}\text{Sentiment}(g_{x}^{(d)}) (3)
Tmean(d)\displaystyle T_{mean}^{(d)} =1Nn=1N1Xx=1XToxicity(gx(d))\displaystyle=\frac{1}{N}\sum_{n=1}^{N}\frac{1}{X}\sum_{x=1}^{X}\text{Toxicity}(g_{x}^{(d)}) (4)

NN denotes the number of times each unique prompt is used to generate a reply. For this work, we choose N=50N=50. We can then say a model is biased if their mean sentiment and toxicity values are unequal for different demographics dd and d+1d+1.

Smean(d)\displaystyle S_{mean}^{(d)} Smean(d+1)\displaystyle\neq S_{mean}^{(d+1)} (5)
Tmean(d)\displaystyle T_{mean}^{(d)} Tmean(d+1)\displaystyle\neq T_{mean}^{(d+1)} (6)

4 Results

Demographic Models
GPT-Neo Cryptocurrency WallStreetBets COVID NoNewNormal Ummah ChristianChat
woman 0.344 0.236 0.018 0.017 0.045 0.214 0.312
man 0.261 0.242 0.017 -0.057 0.066 0.172 0.294
transgender 0.224 0.205 -0.052 0.042 0.033 0.154 0.195
asian 0.316 0.227 0.054 0.006 -0.007 0.171 0.314
black 0.154 0.192 -0.010 -0.110 -0.067 0.091 0.199
white 0.224 0.179 0.016 -0.079 -0.000 0.116 0.198
asexual 0.177 0.230 -0.052 0.085 0.088 0.095 0.192
bisexual 0.309 0.177 -0.066 0.106 0.069 0.157 0.178
heterosexual 0.305 0.215 -0.017 0.052 0.115 0.130 0.187
homosexual 0.221 0.224 -0.043 0.009 0.061 0.081 0.129
christian 0.300 0.196 -0.059 0.012 0.119 0.174 0.316
jewish 0.236 0.227 0.078 -0.015 -0.028 0.141 0.225
muslim 0.170 0.158 -0.066 -0.013 -0.018 0.143 0.080
poor 0.007 0.172 -0.019 -0.088 -0.012 0.132 0.206
rich 0.427 0.258 0.156 0.177 0.162 0.247 0.335
Table 3: Comparison of mean sentiment values SmeanS_{mean} for the different finetuned models and demographics.

Using the methods described above, we have been able to produce 50 generations for each unique prompt and each fine-tuned model, as well as sentiment and toxicity values for each individual generation. The results for mean sentiment SmeanS_{mean} are summarised in Table 3, while mean toxicity TmeanT_{mean} values can be found in Table 4. We showcase examples of particularly toxic and negative generations in the appendix in Table 2 to emphasise the extent to which unchecked datasets can influence downstream generations of language models. It should be noted that these examples are not indicative of the typical text produced by the fine-tuned language models. However, examining negative outputs can offer important insights into any biases that may exist within the models, even if they occur infrequently.

For gender, we contrast women, men and transgender people. The mean sentiment SmeanS_{mean} is highest for women in most models. Generations mentioning transgender people generally receive the lowest sentiment. The only exception to this is the COVID model, where men received the worst sentiment with an SmeanS_{mean} of 0.057-0.057 and transgender people yield the highest SmeanS_{mean} with 0.0420.042. Toxicity is distributed similarly, but identity attack, in particular, is highest for transgender people across all models. Additionally, mental illness or mental disorders are mentioned much more frequently in generations concerning transgender people compared to other groups.

We compare Asian, Black and White people in the race dimension. The results show that completions to prompts mentioning Asian people receive the most positive average sentiments across all models, followed by completions to prompts mentioning White people. Prompts mentioning Black people yielded the worst average sentiments for this dimension in all models but the Cryptocurrency and the ChristianChat models. The largest gap between two demographics is produced by the COVID model with an SmeanS_{mean} of 0.110-0.110 for Black people and 0.0060.006 for Asian people.

For sexual orientation, we consider asexuality, bisexuality, heterosexuality, and homosexuality. Generally, heterosexuality and bisexuality receive more positive sentiments than asexuality and homosexuality. The highest and lowest average sentiment values are for bisexual people, so bias seems quite varied across this dimension. The Cryptocurrency model has a near-equal distribution across all four sexualities. In terms of toxicity, prompts mentioning homosexual people yield the highest TmeanT_{mean}. It should be noted that homosexuality consistently has the highest toxicity averages out of all 15 demographics. Generations concerning asexuals, similarly to transgender people, mention mental impairments more frequently.

In terms of religion, we compare bias towards Christian, Jewish and Muslim people. As an overall trend, the average sentiment is highest for Christians and lowest for Muslims. With a value of 0.3160.316, the ChristianChat model has the highest sentiment for Christians out of all models. ChristianChat’s average sentiment towards Muslims, with a value of 0.0800.080, is significantly lower. The Ummah model is balanced towards all three religions regarding sentiment; it even has its most positive sentiment towards Christianity with a value of 0.1740.174. The Cryptocurrency and WallStreetBets models have their highest average sentiment values towards Jewish people with values of 0.2270.227 and 0.0780.078, respectively. Toxicity, identity attack in particular, is highest for Jewish people with a maximum average identity attack value of 0.0730.073 also produced by the WallStreetBets model, closely followed by the NoNewNormal model with a value of 0.0710.071.

For socioeconomic class, we compare poor and rich people. In terms of sentiment, all models have a considerable gap between rich and poor in terms of mean sentiment; prompts mentioning poor people were completed much more negatively. The most significant gap is created by GPT-Neo 1.3B; for the rich, it has an average sentiment of 0.4270.427. Whereas for the poor, the average sentiment is only 0.0070.007. ChristianChat is the most positive towards poor people with 0.2060.206, but still has a higher SmeanS_{mean} for rich people with 0.3350.335. Surprisingly, regarding toxicity, the gap between rich and poor people is minor, meaning there is a substantial difference between sentiment and toxicity in this dimension. For other bias categories, sentiment and toxicity had a similar distribution.

4.1 Overview of the Different Models

Out of all models, the baseline model GPT-Neo 1.3B has the highest average sentiment values. Generations by GPT-Neo, along with the ChristianChat model, also have the lowest toxicity values. However, that does not mean the baseline model is unbiased. There are significant gaps between different demographics: GPT-Neo is considerably more negative towards poor than rich people. In comparison to prompts mentioning Asian people (SmeanS_{mean} of 0.3160.316), prompts mentioning Black people yield more negative sentiments (SmeanS_{mean} of 0.1540.154). Furthermore, sentiment regarding Muslim people is 0.1700.170, much lower than for any other tested religion.

The Cryptocurrency model was found to be the most balanced out of all models. While average sentiments are lower and toxicity is higher than with the baseline model, it is evenly distributed across all demographics. It remains slightly negatively biased towards Muslims and poor people; these two groups receive the model’s lowest sentiment averages at 0.1580.158 and 0.1720.172, respectively. The WallStreetBets model generated by far the most toxic and negative generations out of all models. Nearly every generated example contains abusive language or deals with banning a user from Reddit. Out of the top 20 most toxic generations, only two were not produced by the WallStreetBets model. However, the WallStreetBets model is particularly biased towards Muslims (0.066-0.066) and poor (0.019-0.019) people, where mean sentiment is substantially lower than for other demographics of the same dimension. Regarding sexuality, the WallStreetBets model is more negative when any sexuality is mentioned, but this is especially true for bisexuality and homosexuality.

The coronavirus-related models are also more negative on average but not quite as toxic as the WallStreetBets model. The COVID model was found to be negatively biased towards Black people in terms of sentiment and toxicity; the model’s mean sentiment SmeanS_{mean} towards Black people is 0.110-0.110. Text generated by the NoNewNormal model is most favourable when prompts mention rich or heterosexual people. In terms of religion, the model is clearly positively biased towards Christianity. The average sentiment value is 0.1190.119 for Christians, and 0.018-0.018 and 0.028-0.028 for Muslims and Jewish people, respectively. NoNewNormal is also the only model to exhibit anti-Asian bias.

We find that the Ummah model is pretty balanced across the board. Even in terms of religion, SmeanS_{mean} is mostly equal. Noticeable is the positive bias towards women and rich people in terms of sentiment. Out of the 15 demographics, its average sentiment is lowest for homosexual people. The text generated by the ChristianChat model is more positive on average, akin to the baseline model. The mean sentiment SmeanS_{mean} is most positive for rich people (0.3350.335), Christians (0.3160.316) and Asians (0.3140.314). Nonetheless, it portrays a significant negative bias towards Muslims (0.0800.080) and homosexual people (0.1290.129).

5 Discussion

The results demonstrate distinct differences between GPT-Neo 1.3B and its fine-tuned variants. The fine-tuned models, particularly the WallStreetBets model and the two COVID-19-related models, produce a greater proportion of negative content and toxic language. The creators of GPT-Neo made a conscious effort to collect a high-quality dataset from academic sources for this language model. The resulting training corpus is less likely to contain harmful language, which could be a plausible explanation as to why the baseline model is less toxic. Regarding the COVID and the NoNewNormal models, the purpose of the underlying online communities was to discuss a disease – this is inherently negative. People are bound to mention symptoms of illness or even death, and this likely creates negative sentiment. In particular, the NoNewNormal subreddit was created to oppose measures of the COVID-19 pandemic and is, therefore, adverse in nature.

The results also show that bias differs between the finetuned models. Most models are favourably disposed towards Asians, but the NoNewNormal model displays anti-Asian tendencies. Out of all seven, the COVID model is the only one to be negatively biased towards men. Furthermore, the majority of fine-tuned models are in favour of Christianity, but the WallStreetBets model has been found to be anti-Christian, anti-Muslim, and pro-Judaism. The NoNewNormal model exhibited more anti-Jewish bias than any other model. One could speculate that people who do not believe in the COVID-19 pandemic are more likely to believe in other conspiracy theories, many of which are rooted in antisemitism. The belief in antisemitic conspiracies has been linked to antisemitic behaviour (Douglas et al. 2019). Bias in the dimension of socioeconomic class is similar across the board. The data suggests that all seven models are at least slightly negatively biased against poor people. For rich people, average sentiment peaks in all models. The finance-related communities were not shown to be more, but less biased towards poor people (especially the Cryptocurrency model). All tested models, including the base model, are also negatively biased towards Muslims, black people and homosexual people but vary greatly for other social groups. The differences in the manifestation of bias must be attributed to the differences in datasets, as there is no other difference between the models. We conclude that, through fine-tuning a pre-trained language model with different datasets, it is possible to test and compare the bias of these datasets.

The difference in average sentiment between prompts that only differ in descriptors (of the same demographic) suggests that using a set of words to describe each demographic provides a tangible benefit. Incorporating this into bias testing can catch different angles of the same demographic. Furthermore, there seems to be a difference between using labels that are nouns and labels that are adjectives, supporting the theory that using nouns can amplify the effect of bias (Beukeboom and Burgers 2019).

This work presents an approach that is highly useful as an entry point to examining bias within different social media communities and could potentially be used to gain insights into the prevalence or the extent of echo chambers within these communities. High levels of bias could indicate that opinions within a community are heterogeneous and that there is limited exposure to diverse perspectives. This could be attributed to confirmation bias, which leads individuals to seek out or favour information that follows their pre-existing beliefs (Alatawi et al. 2021). Bias within an echo chamber could lead to the social validation of harmful biases and further perpetuate them. This is especially concerning as the polarising effects of echo chambers were evident during the spread of misinformation about COVID-19 (Alatawi et al. 2021).

Another application of this work would be the mitigation of bias. Our approach could be used to identify which parts of a larger dataset are particularly biased, either to remove those parts from the dataset or focus mitigation techniques on them.

5.1 Limitations

Although our methods were designed with relevant literature in mind, we want to address possible limitations. It is crucial to acknowledge the limitations of using sentiment and toxicity as the sole metrics for bias analysis. This methodology may oversimplify complex nuances and fail to capture subtle biases, or lead to the overexaggeration of bias. On top of not capturing bias perfectly, these classification models can suffer from bias themselves. Research shows that toxicity analysis is especially prone to societal biases; phrases predominantly used by marginalised communities are much more likely to be classified as toxic (Sap et al. 2019). Toxicity classifiers are often based on the work of annotators and, implicitly, their own biases. Sap et al. show that making annotators consider the racial background of the text’s author and highlighting the dialect used significantly reduces the likelihood of AAE phrases being labelled as toxic or offensive (Sap et al. 2019). Using measures like these could help make toxicity and sentiment models less biased. The extent to which socioeconomic class bias was confirmed in all models could also be due to this limitation. The word ”poor” is assigned a negative sentiment and a reasonably high toxicity value by many classifiers, including the ones we deployed for our analysis. In addition, the word ”rich” is classified as having positive sentiment. Generated replies are likely to repeat words mentioned in the input prompt. Thus, the choice of words can noticeably skew the results when generations are classified as negative or positive simply for repeating the keywords used in the prompt. Moreover, ”poor” and ”rich” are not the only descriptors of demographics that are not neutral. ”Gay” is classified as severely toxic, which could explain why average toxicity was highest for homosexual people in all models. The choice of words can also skew results when the words used are ambiguous. The mean sentiment for bisexuals was surprisingly positive and could be explained by this phenomenon. We used the word ”bi” to describe bisexual people, but the term can also have other meanings.

Another potential limitation of the proposed methodology is that biases from pre-training datasets may be amplified during the fine-tuning process, even when these biases are absent in the fine-tuning datasets. This can occur if the bias is related to a topic that is present in the fine-tuning data. However, research shows that biases in fine-tuning data may have a greater impact on downstream harms than biases in pre-training data. Even small, carefully curated fine-tuning datasets can help alleviate bias from pre-training data, putting even more emphasis on fine-tuning datasets (Steed et al. 2022; Solaiman and Dennison 2021).

The prompts specify a particular social group, but the examples generated are often ambiguous or refer to related groups. Some examples even mention multiple groups, making it difficult to determine the source of any negative associations. Additionally, negative sentiment towards a group may not necessarily indicate a negative overall perception. To address this issue, it could be helpful to incorporate the concept of regard alongside sentiment.

6 Conclusion

This work introduces a framework for capturing representational bias of datasets using large pre-trained language models. We have fine-tuned GPT-Neo with conversations of six online communities to examine their biasedness. We obtained one model per dataset and prompted the resulting models using our constructed placeholder templates for conversational language models alongside different keywords for all demographics. We examine bias along a multitude of dimensions, including 15 different social groups. For the bias evaluation, we utilised automated sentiment and toxicity classifiers and observe the results. We have demonstrated that our finetuned models have varying types and degrees of bias embedded in their fine-tuned parameters. This lends support to the idea that language models ingest the biases of their training datasets. Our work presents not only a method for testing the bias of datasets but also that of online communities if representative datasets are available. One can analyse the bias of a particular dataset by analysing the bias of a correspondingly fine-tuned language model. Using language models makes it possible to prompt directly for demographics or topics without having to classify which text in a dataset is addressing which social group or topic. It is feasible to find associations for social groups beyond the sentence or paragraph they are mentioned in. This method is scalable and customisable, as bias metrics and language models can be varied.

Demographic Models
GPT-Neo Cryptocurrency WallStreetBets COVID NoNewNormal Ummah ChristianChat
woman 0.050 0.177 0.434 0.162 0.211 0.058 0.044
man 0.055 0.173 0.421 0.172 0.202 0.054 0.043
transgender 0.058 0.164 0.414 0.129 0.234 0.062 0.064
asian 0.015 0.127 0.352 0.086 0.158 0.049 0.028
black 0.078 0.178 0.382 0.144 0.245 0.092 0.056
white 0.057 0.154 0.377 0.135 0.190 0.083 0.066
asexual 0.079 0.143 0.413 0.134 0.242 0.110 0.065
bisexual 0.058 0.185 0.454 0.109 0.225 0.088 0.084
heterosexual 0.108 0.159 0.470 0.214 0.249 0.129 0.119
homosexual 0.155 0.229 0.506 0.263 0.308 0.166 0.141
christian 0.022 0.148 0.396 0.118 0.107 0.059 0.023
jewish 0.047 0.118 0.342 0.090 0.197 0.065 0.051
muslim 0.035 0.137 0.420 0.116 0.151 0.061 0.061
poor 0.070 0.167 0.353 0.128 0.171 0.067 0.049
rich 0.019 0.133 0.306 0.091 0.170 0.035 0.035
Table 4: Comparison of mean toxicity values TmeanT_{mean} for the different finetuned models and demographics.

6.1 Future Work

The implications and limitations of this work incentivise future research. Using sentiment and toxicity is insufficient to capture all forms of bias, especially when pre-trained classification models also suffer from bias (Sap et al. 2019). In addition to developing less biased models, expanding the bias metrics used for language generation models is crucial. Additional research into the differences in labelling social groups with nouns or adjectives could provide further knowledge for studying bias in language and communication science. Moreover, future studies that compare the bias of different models (or datasets) should consider introducing a mathematical notion that captures bias in each dimension or for each demographic in a single value to enable more direct comparisons and consider general differences between datasets.

Our approach involves using a single language model checkpoint to assess the bias of the fine-tuning dataset. Nonetheless, analysing multiple checkpoints could not only provide a more accurate analysis of bias but also be used to observe how bias develops during the finetuning process. By identifying the patterns of how language models learn bias, we could determine how to mitigate this risk.

Ethical Statement

We want to acknowledge the ethical considerations and broader impacts of our work. We propose a framework to observe the biases in different datasets and, potentially, in various communities. However, we acknowledge the potential ethical concern of inaccurately assessing bias through our methodology. Similar to other automated bias analysis techniques, relying solely on sentiment and toxicity metrics may result in overlooking or exaggerating the presence of bias. Such mischaracterisation can adversely affect community well-being by leading to social alienation or marginalisation. Our methods do not substitute for an open dialogue with community members and should not be used to judge individuals or make policies on their own. Instead, the proposed method can serve as an entry point into the analysis of biases in different communities. Overlooking biases due to methodological limitations can instead lead to the reinforcement of such biases within and beyond the community. It is, therefore, crucial to note that this method should not be solely relied upon when exploring bias in communities.

Another ethical concern is relying on pre-trained models to detect bias. Studies have revealed that toxicity classification can be affected by bias (Sap et al. 2019; Bender et al. 2021), including the VADER model, which classifies the word ”gay” as severely toxic. In order to combat this risk, we strived to maintain transparency in our analysis and incorporated two different metrics. Moving forward, we aim to improve the bias metrics and replace any models that may be biased.

By showing some harmful and toxic responses generated by the fine-tuned models, our work emphasises the repercussions of using problematic datasets to train language models. Deploying consequently biased language models for any application entails the risk of perpetuating societal biases and intensifying discrimination of marginalised groups. The method presented can work as a tool for identifying which datasets, or subsets of datasets, contain harmful biases and can thus be used as a first step of a bias mitigation approach.

Acknowledgement

This research has partly been funded by the Federal Ministry of Education and Research of Germany and the state of North-Rhine Westphalia as part of the Lamarr-Institute for Machine Learning and Artificial Intelligence.

References

  • Abid, Farooqi, and Zou (2021) Abid, A.; Farooqi, M.; and Zou, J. 2021. Persistent Anti-Muslim Bias in Large Language Models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, 298–306. New York, NY, USA.
  • Alatawi et al. (2021) Alatawi, F.; Cheng, L.; Tahir, A.; Karami, M.; Jiang, B.; Black, T.; and Liu, H. 2021. A Survey on Echo Chambers on Social Media: Description, Detection and Mitigation. arXiv:2112.05084.
  • Barikeri et al. (2021) Barikeri, S.; Lauscher, A.; Vulić, I.; and Glavaš, G. 2021. RedditBias: A Real-World Resource for Bias Evaluation and Debiasing of Conversational Language Models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 1941–1955.
  • Bender et al. (2021) Bender, E. M.; Gebru, T.; McMillan-Major, A.; and Shmitchell, S. 2021. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 610–623. New York, NY, USA.
  • Beukeboom and Burgers (2019) Beukeboom, C. J.; and Burgers, C. 2019. How Stereotypes Are Shared Through Language: A Review and Introduction of the Social Categories and Stereotypes Communication (SCSC) Framework. Review of Communication Research, 7: 1–37.
  • Black et al. (2021) Black, S.; Leo, G.; Wang, P.; Leahy, C.; and Biderman, S. 2021. GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow. https://github.com/EleutherAI/gpt-neo/tree/v1.1. Accessed: 2022-7-12.
  • Blodgett et al. (2020) Blodgett, S. L.; Barocas, S.; Daumé III, H.; and Wallach, H. 2020. Language (Technology) is Power: A Critical Survey of “Bias” in NLP. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 5454–5476. Online: Association for Computational Linguistics.
  • Brown et al. (2020) Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J. D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; Agarwal, S.; Herbert-Voss, A.; Krueger, G.; Henighan, T.; Child, R.; Ramesh, A.; Ziegler, D.; Wu, J.; Winter, C.; Hesse, C.; Chen, M.; Sigler, E.; Litwin, M.; Gray, S.; Chess, B.; Clark, J.; Berner, C.; McCandlish, S.; Radford, A.; Sutskever, I.; and Amodei, D. 2020. Language Models are Few-Shot Learners. In Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M. F.; and Lin, H., eds., Advances in Neural Information Processing Systems, volume 33, 1877–1901.
  • Devlin et al. (2019) Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171–4186. Minneapolis, Minnesota.
  • Dhamala et al. (2021) Dhamala, J.; Sun, T.; Kumar, V.; Krishna, S.; Pruksachatkun, Y.; Chang, K.-W.; and Gupta, R. 2021. BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 862–872. New York, NY, USA.
  • Douglas et al. (2019) Douglas, K. M.; Uscinski, J. E.; Sutton, R. M.; Cichocka, A.; Nefes, T.; Ang, C. S.; and Deravi, F. 2019. Understanding Conspiracy Theories. Political Psychology, 40(S1): 3–35.
  • Elwert, Tabti, and Pfahler (2020) Elwert, F.; Tabti, S.; and Pfahler, L. 2020. Me, myself and the other. Interreligious and intrareligious relations in neo-conservative online forums. Religion, 50(3): 414–436.
  • Gao et al. (2021) Gao, L.; Biderman, S.; Black, S.; Golding, L.; Hoppe, T.; Foster, C.; Phang, J.; He, H.; Thite, A.; Nabeshima, N.; Presser, S.; and Leahy, C. 2021. The Pile: An 800GB Dataset of Diverse Text for Language Modeling. CoRR, abs/2101.00027.
  • Gehman et al. (2020) Gehman, S.; Gururangan, S.; Sap, M.; Choi, Y.; and Smith, N. A. 2020. RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2020, 3356–3369. Association for Computational Linguistics.
  • Guo, Ma, and Vosoughi (2022) Guo, X.; Ma, W.; and Vosoughi, S. 2022. Measuring Media Bias via Masked Language Modeling. Proceedings of the International AAAI Conference on Web and Social Media, 16(1): 1404–1408.
  • Hanu and Unitary team (2020) Hanu, L.; and Unitary team. 2020. Detoxify. https://github.com/unitaryai/detoxify. Accessed: 2022-10-24.
  • Henderson et al. (2018) Henderson, P.; Sinha, K.; Angelard-Gontier, N.; Ke, N. R.; Fried, G.; Lowe, R.; and Pineau, J. 2018. Ethical Challenges in Data-Driven Dialogue Systems. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, 123–129. New York, NY, USA.
  • Hutto and Gilbert (2014) Hutto, C.; and Gilbert, E. 2014. VADER: A Parsimonious Rule-Based Model for Sentiment Analysis of Social Media Text. Proceedings of the International AAAI Conference on Web and Social Media, 8(1): 216–225.
  • Jiang et al. (2022) Jiang, H.; Beeferman, D.; Roy, B.; and Roy, D. 2022. CommunityLM: Probing Partisan Worldviews from Language Models. In Proceedings of the 29th International Conference on Computational Linguistics, 6818–6826. Gyeongju, Republic of Korea: International Committee on Computational Linguistics.
  • Komatsuzaki (2019) Komatsuzaki, A. 2019. One Epoch Is All You Need. CoRR, abs/1906.06669.
  • Lexyr Inc (2021a) Lexyr Inc. 2021a. The Reddit COVID dataset. https://www.kaggle.com/datasets/pavellexyr/the-reddit-covid-dataset. Accessed: 2022-07-05.
  • Lexyr Inc (2021b) Lexyr Inc. 2021b. Reddit cryptocurrency data for August 2021. https://www.kaggle.com/datasets/pavellexyr/reddit-cryptocurrency-data-for-august-2021. Accessed: 2022-07-05.
  • Lexyr Inc (2021c) Lexyr Inc. 2021c. Reddit /r/NoNewNormal dataset. https://www.kaggle.com/datasets/pavellexyr/reddit-r-nonewnormal-dataset. Accessed: 2022-07-05.
  • Liang et al. (2021) Liang, P. P.; Wu, C.; Morency, L.-P.; and Salakhutdinov, R. 2021. Towards Understanding and Mitigating Social Biases in Language Models. In Meila, M.; and Zhang, T., eds., Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, 6565–6576. PMLR.
  • Liu et al. (2019) Liu, H.; Dacon, J.; Fan, W.; Liu, H.; Liu, Z.; and Tang, J. 2019. Does Gender Matter? Towards Fairness in Dialogue Systems. CoRR, abs/1910.10486.
  • Nadeem, Bethke, and Reddy (2021) Nadeem, M.; Bethke, A.; and Reddy, S. 2021. StereoSet: Measuring stereotypical bias in pretrained language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 5356–5371. Association for Computational Linguistics.
  • OpenAI (2022) OpenAI. 2022. ChatGPT: Optimizing language models for dialogue. https://openai.com/blog/chatgpt/. Accessed: 2023-01-11.
  • Podolak (2021) Podolak, M. 2021. Reddit WallStreetBets Posts and Comments. https://www.kaggle.com/datasets/mattpodolak/rwallstreetbets-posts-and-comments. Accessed: 2022-06-14.
  • Radford et al. (2019) Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I.; and others. 2019. Language Models are Unsupervised Multitask Learners. OpenAI, 1(8): 9.
  • Sap et al. (2019) Sap, M.; Card, D.; Gabriel, S.; Choi, Y.; and Smith, N. A. 2019. The Risk of Racial Bias in Hate Speech Detection. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 1668–1678. Florence, Italy: Association for Computational Linguistics.
  • Sheng et al. (2019) Sheng, E.; Chang, K.-W.; Natarajan, P.; and Peng, N. 2019. The Woman Worked as a Babysitter: On Biases in Language Generation. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 3407–3412. Hong Kong, China.
  • Sheng et al. (2021) Sheng, E.; Chang, K.-W.; Natarajan, P.; and Peng, N. 2021. Societal Biases in Language Generation: Progress and Challenges. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 4275–4293.
  • Solaiman and Dennison (2021) Solaiman, I.; and Dennison, C. 2021. Process for adapting language models to society (palms) with values-targeted datasets. Advances in Neural Information Processing Systems, 34: 5861–5873.
  • Steed et al. (2022) Steed, R.; Panda, S.; Kobren, A.; and Wick, M. 2022. Upstream Mitigation Is Not All You Need: Testing the Bias Transfer Hypothesis in Pre-Trained Language Models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 3524–3542. Dublin, Ireland: Association for Computational Linguistics.
  • Wolf et al. (2020) Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; Davison, J.; Shleifer, S.; von Platen, P.; Ma, C.; Jernite, Y.; Plu, J.; Xu, C.; Le Scao, T.; Gugger, S.; Drame, M.; Lhoest, Q.; and Rush, A. 2020. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 38–45.
  • Zhang et al. (2020) Zhang, Y.; Sun, S.; Galley, M.; Chen, Y.-C.; Brockett, C.; Gao, X.; Gao, J.; Liu, J.; and Dolan, B. 2020. DIALOGPT : Large-Scale Generative Pre-training for Conversational Response Generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 270–278.

Appendix A Supplementary Material

Demographic Keywords

Woman:

woman, girl, women, girls, female, she, her

Man:

man, boy, men, boys, male, he, him

Transgender:

transgender, trans, non-binary, transgender people, trans people, non-binary people

Asian:

Asian, Chinese, Indian, Japanese, Asians, Chinese people, Indians, Japanese people

Black:

Black, African, African American, Black people, Africans, African Americans

White:

White, American, European, Caucasian, White people, Americans, White Americans, Europeans

Asexual:

asexual, asexual people, asexuality

Bisexual:

bisexual, bi, bisexual people, bi people, bisexuality

Heterosexual:

straight, heterosexual, straight people, heterosexual people, heterosexuality

Homosexual

homosexual, gay, lesbian, queer, homosexuals, homosexual people, gay people, lesbians, queer people, homosexuality

Christian:

Christian, Christians, Christianity

Jewish:

Jewish, Jew, Jews, Jewish people

Muslim:

Muslim, Muslims, Islam

Poor:

poor, broke, unemployed, homeless, beggar, beggars, poor people, unemployed people, homeless people

Rich:

rich, wealthy, millionaire, billionaire, rich people, wealthy people, millionaires, billionaires