The Less the Merrier? Investigating Language Representation in Multilingual Models
Abstract
Multilingual Language Models offer a way to incorporate multiple languages in one model and utilize cross-language transfer learning to improve performance for different Natural Language Processing (NLP) tasks. Despite progress in multilingual models, not all languages are supported as well, particularly in low-resource settings. In this work, we investigate the linguistic representation of different languages in multilingual models. We start by asking the question which languages are supported in popular multilingual models and which languages are left behind. Then, for included languages, we look at models’ learned representations based on language family and dialect and try to understand how models’ learned representations for (1) seen and (2) unseen languages vary across different language groups. In addition, we test and analyze performance on downstream tasks such as text generation and Named Entity Recognition. We observe from our experiments that community-centered models—models that focus on languages of a given family or geographical location and are built by communities who speak them—perform better at distinguishing between languages in the same family for low-resource languages. Our paper contributes to the literature in understanding multilingual models and their shortcomings and offers insights on potential ways to improve them.
1 Introduction
While we have seen improvements in state-of-the-art performance in various NLP tasks by multilingual models Conneau et al. (2019); Doddapaneni and Kumar. (2021), there is a disparity in which languages are actively studied. The field of NLP has largely been Anglocentric, with a large portion of the world’s languages, particularly low-resource languages, not being covered in the literature Joshi (2020); Bender (2019); Talat et al. (2022).
For languages that have been included in popular multilingual models, performance is not the same for all languages Joshi (2020). Low-resource languages suffer in performance even when they are included in multilingual models. Several works Hangya and Fraser. (2022); Wang and Roth (2020); Pfeiffer and Ruder. (2021); Schuster and Lewis (2019) have proposed methods for improving performance for low-resource languages. Previous work Doddapaneni and Kumar. (2021) presents an argument that languages might benefit from being included in multilingual models as the models learn language-independent feature representations, while it concludes that the question of the benefit of multilingual training for a given language remains open. Previous works also show that multilingual models might suffer from “negative interference” Zirui Wang (2019) in both high-resource Conneau et al. (2019); Xu Tan and Liu. (2019) and low-resource Zirui Wang (2020) settings.
In this work, we first look at the linguistic diversity in multilingual models. We then investigate the embeddings for different languages in multilingual models and show how the representations affect downstream performance in language identification. For our analysis, we used three autoregressive and five autoencoder models. First, we looked at 2D visualizations of learned representations for all models. Then, we evaluated autoregressive models’ performance on text generation and the autoencoder models by language classification based on learned representations and performance on Named Entity Recognition (NER). We base our analysis on (1) language family, (2) dialects, and (3) writing scripts, with a focus on low-resource language settings.
2 Related Works
In efforts to include more of the world’s languages, previous works have built multilingual language models over the years. While models with larger numbers and more diverse sets of languages have shown commendable performance in several NLP tasks Conneau et al. (2019); Zhang et al. (2021), we have also seen community-centered models improve upon task performance Dossou et al. (2022); Dabre et al. (2022). Previous work Conneau et al. (2019) hypothesizes that models might suffer from “curse of multilinguality”, which describes how for a given model size, increasing the number of languages does not improve performance for individual languages after a certain point. With this in mind, we ask the question: Do Multilingual Language Models with fewer, community-centered languages learn better representations of different languages depending on language families, dialects, and writing scripts?
Previous works have analyzed how pre-trained language models learn representations for different languages using probing tasks Choenni and Shutova (2020); Eva Schlinger (2019); Jindrich Libovicky (2019) as well as investigating the geometry of sub-spaces Chang and Bergen (2022); Rajaee and Pilehvar (2022). One previous work Chang and Bergen (2022) focuses on understanding multilingual language models’ overall embedding space geometry and identifies axes for language-sensitive and language-neutral features. In our work, we are interested in how different languages are represented in multilingual language models with a closer look at how different language families, dialects, and writing scripts are represented.
3 Models and Data
We chose models from two categories for our experiments: autoencoder and autoregressive models. In this section, we give descriptions of the models we chose and their training data. Table 1 gives a summary of the models we used with information on their training data, model size, and languages covered.
3.1 Autoencoder Models
Autoencoder models are trained to recreate their inputs from an internal latent representation of the input Dor Bank (2021). Many such models use partial input masking during training, known as masked language modeling (MLM). In this paper, we focus on transformer-based autoencoders. In particular, we look at BERT Devlin et al. (2018) and RoBERTA Liu et al. (2019) models, which use MLM.
XLM-R Conneau et al. (2019) is a multilingual Transformer-based MLM trained on the Common Crawl data for 100 languages. To balance data between English and other languages, the XLM-R uses one dump for English and 12 dumps for all other languages.
BERT multilingual Devlin et al. (2018) is a pre-trained model on the top 104 languages with the largest Wikipedias using an MLM objective. It is designed to pre-train deep bidirectional representations from unlabeled text by jointly conditioning on both the left and right context in all layers.
AfroLM Dossou et al. (2022) is a multilingual language model pre-trained from scratch on 23 African languages using a self-active learning framework with an MLM objective.
3.2 Autoregressive Models
Autoregressive models are sequential models that use the previous tokens to predict the next token. Transformer-based autoregressive models use a transformer decoder and causal masked attention to learn sequential representation regressively.
GPT-3 Brown et al. (2020) is a generative model with 175 billion parameters. It has been used in several downstream tasks and to demonstrate in-context learning.
LLaMa is an autoregressive model that is trained on “publicly available datasets exclusively” Touvron et al. (2023).
BLOOM BigScience (2022a) is an autoregressive model that is trained on the ROOTS corpus BigScience (2022b).
Model Type | Model Name | Size | Languages | Data |
Autoencoders | XLM-R | 270M | 100 | Filtered CommonCrawl |
AfroLM | 270M | 23 | JW300, Bible, News | |
BERT-multilingual | 110M | 104 | BooksCorpus and English Wikipedia | |
AraBERT | 110M | 1 | Arabic news, Arabic Corpus, OSIAN: the Open Source International Arabic News Corpus | |
IndicBERT | 12M | 12 | AI4Bharat’s monolingual corpus | |
Autoregressive | GPT-3 | 175B | 119 | Filtered CommonCrawl, WebText Liu and Curran (2006), e-books, English Wikipedia |
LLaMA | 65B | 20 | English CommonCrawl, C4, Github, Wikipedia, Gutenberg and Books3, ArXiv, Stack Exchange | |
BLOOM | 560M | 59 | ROOTS BigScience (2022b) |
4 Language Diversity
Before starting our experiments to understand how multilingual models learn representations for different languages, we first looked at which languages are included and which languages are excluded from mainstream NLP research. From the models discussed in Section 3, we selected XLM-R and LLaMA from generic multilingual models and AfroLM from community-centered models to evaluate their linguistic diversity across countries. First, we look at the linguistic diversity of the community-centered model, AfroLM. In Fig 1 (a), we show a map representing the number of languages in each African country represented by AfroLM. We contrast these results with Fig 1 (b), where we represent the same data by dividing the number of languages represented in the model by the number of languages spoken in the country. We see in Figure 1 that a large number of languages per country are still left behind, even in cases where the models are focused on a single continent.
Next, we look at XLM-R which has been a popular choice in multilingual studies Choudhury and Deshpande (2021). We see that 55 out of the 100 languages included in XLM-R are Indo-European in comparison with 2 each from Niger-Congo and Kra-Dai languages included in the model. In Fig. 2, we use the Indo-European language family tree from Young (2015) to show the languages that are represented from that family by XLM-R. For LLaMA, Table 2, shows the languages included in the training data, which are exclusively in the European branch of the Indo-European family with the exception of Hungarian which is a member of the Uralic language family.



Dataset | Sampling prop | Languages | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
CommonCrawl | 67% | English | ||||||||||
C4 | 15% | English | ||||||||||
Github | 4.5% | English | ||||||||||
Wikipedia | 4.5% |
|
||||||||||
Books | 4.5% | English | ||||||||||
ArXiv | 2.5% | English | ||||||||||
StackExchange | 2% | English |
5 Methods
5.1 Evaluation dataset
We used publicly available datasets to evaluate multilingual models. For embedding space representation and language classification, we used Flores NLLB Team (2022) dataset except for the Tigrinya dialect evaluation. To evaluate embedding space representation and language classification for the Tigrinya dialect, we used a dataset from Haileslasie et al. (2023). To evaluate NER, we used MasakhaNER Adelani et al. (2021) dataset for Bantu languages and WikiANN Pan et al. (2017) dataset for Romance languages.
For the text generation task, we designed our prompts in English in five categories: Bible, World Knowledge, News, Generic, and Bias. We chose these categories (1) to explore what pre-trained LLMs represent (in World Knowledge, Bias, and News), (2) to get closer to training data for low-resource languages (in Bible), and (3) to observe the behaviors of pre-trained LLMs in generic situations in different languages (in Generic). We translated the English prompts into Amharic, Tigrinya, Spanish, Italian, and French. For Spanish, Italian, and French, we used Google Translate. For Tigrinya, we used Lesan.ai Hadgu et al. (2022) with manual corrections after translation and for Amharic, we translated manually. The prompts we used are in Appendix A.
5.2 Embedding Space Extraction and Visualization
Distinguishing between Languages through Visualization:
We wanted to understand how different models represent different languages in multilingual settings. We extracted the hidden states from the models in our experiments and used UMAP Leland McInnes (2020) to visualize the representations. Following previous work Devlin et al. (2018), we used the hidden state vectors of the first token in each sentence as the sentence embedding for all layers for autoencoder models. We also experimented with averaging different hidden states of all tokens and found that while it reduced within cluster distances (distances between individual points in already formed clusters), it retained the overall cluster formations. For LLaMA and BLOOM, we used the hidden state vector of the last token in each sentence as the sentence embedding for all layers, as was done in previous work Neelakantan et al. (2022). For GPT-3, we used the embedding models from OpenAI’s API endpoints; we used the text-embedding-ada-002 and text-similarity-davinci-001 models. From Afro-Asiatic language family, we choose Semitic and Cushtic languages. From Niger-Congo language family, we chose Bantu languages. From Indo-European language family, we choose Indo-Iranian and Romance languages.
Distinguishing between Languages through Classification:
To corroborate the results we observed in the visualization of the embedding spaces, we used K-Means clustering on the learned representations we extracted from the pre-trained models to test to what extent different models can differentiate among languages. In Section 6.2, we discuss the result of language classification for different language groups and models.
5.3 Downstream Tasks
Scholars have previously evaluated the performance of Large Language Models (LLMs) in different downstream tasks Adelani et al. (2021, 2023); Dione et al. (2023). Our interest is in understanding how multilingual models learn and represent languages from different language families, dialects, and writing scripts. Hence, in addition to investigating the models’ learned representations through visualization and classification, we evaluated how these models perform in downstream tasks across languages, focusing on two tasks: NER and text generation.
Named Entity Recognition (NER):
NER is an Information Extraction task which serves as one of the “fundamental building blocks of complex NLP tasks” Singh (2018). We selected NER task to understand how selected autoencoder models perform tasks that require deeper language understanding capabilities for models. In this work, we evaluated Bantu and Romance family languages, and we discuss the results in Section 6.3.
Text Generation:
For autoregressive models, we evaluated downstream applications by prompting the models in 6 languages from our experiment and using GEEZSwitch Gaim and Park (2022) to classify the generated text. For this experiment, we chose GPT-3 and BLOOM. We prepared three prompts in 5 categories for a total of 15 prompts per language. More details about the prompts are in Appendix A. We then generated text eight times for each prompt for a total of 120 text generations per language per model and 1440 generations overall.
6 Results
6.1 Visualizing Models’ Embedding Spaces
6.1.1 Language family
As detailed in Section 4, the language families represented in the multilingual models are predominantly Indo-European. We wanted to investigate how multilingual models learned representations vary for different language families both for seen and unseen languages.
For Romance languages, XLM-R shows language-neutral semantic clusters in the middle layers, with language-specific clusters in the last layer (in Appendix D.3 Fig. 22 (b)). In XLM-R, Bantu language representations were mixed in pairs for all languages except Swahili which formed somewhat of an independent cluster (in Appendix D.3, Fig. 22 (a)). AfroLM showed independent clusters for 3 of the Bantu languages with Xohsa and isZulu mixing (in Appendix D.3 Fig. 21). For Semitic languages, we observed, except for XLM-R, AfroLM and GPT models, all other language models had mixed representations for Amharic and Tigrinya in both autoregressive (in Fig. 3) and autoencoder (in Fig. 5) models. Looking more in detail at the representations that are close together for Amharic and Tigrinya in AfroLM, we observe that the sentences are translations of each other (in Fig. 4).










6.1.2 Dialect
In addition to language families, we wanted to investigate how multilingual language models learn representations for languages with different dialects. We selected Tigrinya and Arabic languages from the Semitic language group for this experiment. For Arabic, we chose four dialects: Morocco, Levantine, Egypt, and Standard Modern Arabic, based on the numbers of speakers. For Tigrinya dialects, we choose three dialects used in Haileslasie et al. (2023). Our result shows that for the Arabic dialect, except AraBERT, all the models mixed the representations for all dialects. The AraBERT model clustered some sentences from the Egyptian dialect independently but mixed the rest of the dialects (in Appendix D Fig. 19). Similarly, all the models mix the representations for Tigrinya dialects, as shown in Figure 6.





6.1.3 Writing Script
We also evaluated how multilingual models represent languages that use different writing scripts. For this experiment, we selected Amharic, Arabic, Spanish, Chinese, and Hindi, which use Geez (Ethiopic), Arabic, Latin, Chinese (Han), and Hindi (Devanagari) scripts, respectively. Our result in Figures 7 and 8 show that all models except AraBERT form somewhat distinct language-specific clusters. While XLM-R and mBERT clustered Amharic separately, other languages were clustered near each other with some cross-language mixing. The AraBERT model independently clustered Spanish and Arabic while mixing the other languages. As shown in Figure 8, GPT-3 models showed some language-specific clusters with cross-language mixing, while LLaMA and BLOOM had separate clusters for each language.









6.2 Language Classification
Table 4 shows the language classification results for autoencoder models across different language groups. As discussed in Section 6.1, we are interested in evaluating how multilingual models classify languages regardless of their writing script, language families, and dialects. We used F1 score to evaluate the clustering performance. As shown in Table 4, we observed similar differences between the models in language classification tasks as seen in the visualization experiments (Section 6.1). Models that showed separate clusters in the embedding space across language families and writing scripts showed the highest F1 score than the rest of the models. For Semitic and Cushtic languages, AfroLM classified all the languages correctly with an F1-score of 100%, while for Indo-Iranian language classification, IndicBERT classified all the languages correctly with F1-score of 100%. For Bantu language classification, AfroLM showed the highest performance with F1-score of 79% while IndicBERT showed the lowest F1-score of 13%. For both Arabic and Tigrinya dialects, all the models show an F1-score below 40%, this shows all the models are struggling to classify different dialects within the languages. For different writing scripts, AraBERT showed F1-score of 62% while XLM-R showed the lowest F1-score of 37%.
Models | Language | Correct Language |
Generated (%) | ||
GPT-3 | Spanish | 95.83 |
French | 91.67 | |
Italian | 91.67 | |
English | 90.83 | |
Amharic | 76.66 | |
Tigrinya | 45.00 | |
BLOOM | Spanish | 93.33 |
French | 100 | |
Italian | 51 | |
English | 100 | |
Amharic | 80.00 | |
Tigrinya | 43.33 |
Language groups | Models (F1-score) | ||||
General | Community-Centric | ||||
XLM-R | mBERT | AfroLM | AraBERT | IndicBERT | |
Semitic | 0.68 | 0.52 | 1.0 | 0.61 | 0.62 |
Cushetic | 0.52 | 0.41 | 1.0 | 0.33 | 0.37 |
Bantu | 0.37 | 0.23 | 0.79 | 0.18 | 0.13 |
Romance | 0.42 | 0.24 | 0.38 | 0.34 | 0.21 |
Indo-Iranian | 0.33 | 0.30 | 0.71 | 0.2 | 1.0 |
Dialects | |||||
Arabic_dialects | 0.25 | 0.23 | 0.25 | 0.32 | 0.26 |
Tigrinya | 0.23 | 0.30 | 0.39 | 0.31 | 0.26 |
Writing script | |||||
Different writing script | 0.37 | 0.49 | 0.55 | 0.62 | 0.53 |
6.3 Named Entity Recognition
Table 5 shows NER results for Bantu and Romance language families. In both NER tasks, we observed identical distinctions between the models as seen in the visualization experiments (Section 6.1). For the Bantu languages, AfroLM outperformed other models except in Kinyarwanda, while generic models outperformed community-centered models for the Romance languages except for French.
Language Family | Language | Models(F1-score) | ||||
General | Community-Centric | |||||
XLM-R | mBERT | AfroLM | AraBERT | IndicBERT | ||
Bantu | zul | 84.6 | 81.7 | 86.3 | 76.9 | 67.2 |
xho | 87 | 85 | 87.5 | 75.8 | 75.3 | |
sna | 93.6 | 92.4 | 94.4 | 73 | 83.4 | |
swa | 87.5 | 86.3 | 87.6 | 78.9 | 79.9 | |
kin | 73.9 | 70.9 | 72.8 | 69.3 | 71.1 | |
Romance | fra | 88.95 | 90.66 | 87.89 | 91.15 | 91.97 |
spa | 94.58 | 92.03 | 80.12 | 81.71 | 83.73 | |
cat | 91.72 | 95.67 | 82.76 | 85.24 | 84.48 | |
ita | 90.43 | 91.91 | 80.28 | 83.82 | 83.25 |
6.4 Text Generation
In Table 3, we show the results of language classification for the text generation task. GPT-3 generations are in the same language of the prompt above 90% of the time for high-resourced languages. A deeper look at the generations that were in a different language for English reveals that for the prompt of a verse from the Bible, GPT-3 generates the exact name of the verse “(John 3:16)” which GEEZSwitch detects as German language, accounting for 8 out of 12 of the wrong language identifications in English. Two of the remaining wrongful detection results were due to GEEZSwitch detecting a list of African countries in English as Swahili. The remaining wrongful detection was a mix of Amharic and English for the prompt “I am a tourist. Tell me common phrases in Amharic.”
While GPT-3 does decently on Amharic, closer analysis reveals that a 14% of the generated text, which was classified as Amharic, also includes boilerplate code and a mix of English. For Tigrinya, 51.51% of the mis-generated text is in Amharic, and 45.45% is in English. The remaining 3% are attempts at romanized Tigrinya, which were misclassified as German and Swahili. In Appendix B, Figure 9, we show examples of text that were detected as generated in a language other than the prompt language. We want to emphasize that these results are conditioned on the fact that we are looking ONLY at the language of the generated text; NONE of the Amharic and Tigrinya generations, even when they are in the same language as the prompt, are coherent sentences or phrases.
For BLOOM text generation performance, we see in Table 3 that generated text is in the same language as the prompt for 100 % of the time for English and French and 80% for Amharic. In contrast to GPT-3, the generated text is in the same language of the prompt 51% of the time for Italian, with Spanish accounting for 40.67% of the mis-generations, Catalan and French each accounting for 27.11% of the mis-generations, and English accounting for 5.08% of the mis-generation. In Appendix B, Figure 10, we go into further depth on some examples of generations in Tigrinya and Amharic, highlighting the existence of artifacts that seem to be inherited from web interfaces.
7 Discussion
7.1 On Inclusion of Diverse Languages and Language Families
Our analysis of linguistic diversity in popular multilingual models aligns with criticism of previous works that a large proportion of the world’s languages remain understudied. We observe a skew towards Indo-European languages, with a heavier skew towards the European side of the family tree (Section 4). This indicates there is still a huge room for improvement in the NLP community, particularly in encouraging community-driven research for the inclusion of a more diverse set of languages, dialects, and writing scripts. Looking at the community-centered models, we see there is greater diversity and more inclusion for low-resource languages, though there is still a long way to go (Section 4). Encouraging research driven by communities of such languages could allow the NLP community to benefit from more diversity per community, which collectively could result in greater diversity overall.
7.2 Are Community-Centered Language Models Better at Distinguishing between Languages?
From the visualizations of the learned representations for different languages (Section 6.1) and the language classification for autoencoder models (Section 6.2), we observe that community-centered models’ representations are more distinct for the languages they focus on. Looking at Semitic languages, only AfroLM and XLM-R had separate clusters for each language (Amharic, Arabic, and Tigrinya), while all other models put Tigrinya and Amharic in the same cluster (Fig. 5). For AfroLM, we see two Tigrinya sentences that were placed closer to the Amharic cluster are closer to their translation pairs from the dataset (Fig. 4) while XLM-R mixed representations were not explainable as such. We see this behavior in visualizations of representations for other language families in Appendix D.
For autoregressive models, we did not have community-centered models, as defined in our abstract, to compare with. However, we looked at the output of the text generated from GPT-3 and BLOOM models. We looked at 6 languages and observed that Amharic and Tigrinya are mixed in the generated text, while the generated text was not coherent for either of the languages. There is still a long way to go in terms of text generation for low-resource languages, starting with models that respond in the same language as the prompt.
7.3 Same Language, Different Dialects
We also looked at learned representations for languages with different dialects. We observe from our visualizations and language classification experiments that, for both Tigrinya and Arabic, the learned representations form no particular cluster depending on dialect for any of the models for Tigrinya with a small cluster observed for Egyptian Arabic in AraBERT representations (Section 6.1.2). This shows there is huge room for improvement in the NLP space for understanding and accommodating dialectical differences.
7.4 Learned Representations for Unseen Languages
In our experiments, we also included languages that are not seen by the models in training. We did this because (1) including languages that all models have in common would leave a lot of low-resource languages behind, and (2) we wanted to observe how models deal with out-of-domain languages. From our visualizations, we observe that some models cluster unseen languages based on writing scripts. For example, for Semitic languages, LLaMa, BLOOM, mBERT, AraBERT, and IndicBERT clustered Tigrinya and Amharic, languages which both use the Ge’ez script, together and formed a separate cluster for Arabic (in Fig. 3 and Fig. 5). AfroLM and XLM-R formed language-specific clusters for all three languages even though Tigrinya is unseen for both models while Amharic is seen.
8 Conclusion
In this work, we investigated linguistic diversity, the learned representation of different languages, and the downstream task performance of multilingual models. We observe from our experiments that community-centered language models perform better at distinguishing among languages in the same family for low-resource languages. We also observe that there is still a long way to go in accommodating dialectical differences in NLP. Our work contributes to understanding multilingual models, and we hope to see future work with more community-centered models and more diverse languages.
Limitations
Our study is limited by the models included in the experiments and the languages available in the data we used. For instance, after we submitted our paper, models like GPT-SW3 Ariel Ekgren (2023) and Jais MBZUAI (2023) presented community-centered autorgressive models. Future work can do a more comprehensive study with more language models and more diverse data included. Future work can also dig deeper into particular points raised in our work, such as dialectical diversity or the effects of writing systems in language model representations.
Additionally, we note that using dimension reduction techniques to understand representation might lead to loss of some separate spaces which are collapsed down. In our work, we used clustering techniques on top of learned representations and downstream task performance to further corroborate our observations. Future work could also investigate difference between community-centered and generic models through an investigation of the geometric space of the models. Additionally, while we showed some correlation between separate cluster formation in visualizations and downstream performance (Section 7.2), it is still unclear if there is a causal relationship between models that have separate language clusters and their downstream performance. Future work could build upon our results to investigate whether or not a causal relationship exists.
References
- Adelani et al. (2021) David Ifeoluwa Adelani, Jade Abbott, Graham Neubig, Daniel D’souza, Julia Kreutzer, Constantine Lignos, Chester Palen-Michel, Happy Buzaaba, Shruti Rijhwani, Sebastian Ruder, et al. 2021. Masakhaner: Named entity recognition for african languages. Transactions of the Association for Computational Linguistics, 9:1116–1131.
- Adelani et al. (2023) David Ifeoluwa Adelani, Marek Masiak, Israel Abebe Azime, Jesujoba Oluwadara Alabi, Atnafu Lambebo Tonja, Christine Mwase, Odunayo Ogundepo, Bonaventure FP Dossou, Akintunde Oladipo, Doreen Nixdorf, et al. 2023. Masakhanews: News topic classification for african languages. arXiv preprint arXiv:2304.09972.
- Antoun et al. (2020) Wissam Antoun, Fady Baly, and Hazem Hajj. 2020. Arabert: Transformer-based model for arabic language understanding. arXiv preprint arXiv:2003.00104.
- Ariel Ekgren (2023) Felix Stollenwerk Joey Öhman Tim Isbister Evangelia Gogoulou Fredrik Carlsson Alice Heiman Judit Casademont Magnus Sahlgren Ariel Ekgren, Amaru Cuba Gyllensten. 2023. Gpt-sw3: An autoregressive language model for the nordic languages. arxiv preprint arxiv:2305.12987.
- Bender (2019) Emily M Bender. 2019. The# benderrule: On naming the languages we study and why it matters. the gradient (2019).
- BigScience (2022a) BigScience. 2022a. Bloom: A 176b-parameter open-access multilingual language model. arxiv preprint arxiv:2211.05100.
- BigScience (2022b) BigScience. 2022b. Bloom: A 176b-parameter open-access multilingual language model. arxiv preprint arxiv:2211.05100.
- Brown et al. (2020) Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020. Language models are few-shot learners.
- Chang and Bergen (2022) Zhuowen Tu Chang, Tyler A and Benjamin K Bergen. 2022. The geometry of multilingual language model representations. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing.
- Choenni and Shutova (2020) Rochelle Choenni and Ekaterina Shutova. 2020. What does it mean to be language-agnostic? probing multilingual sentence encoders for typological properties. arxiv preprint arxiv.:2009.12862.
- Choudhury and Deshpande (2021) Monojit Choudhury and Amit Deshpande. 2021. How linguistically fair are multilingual pre-trained language models? In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 12710–12718.
- Conneau et al. (2019) Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Unsupervised cross-lingual representation learning at scale. arXiv preprint arXiv:1911.02116.
- Dabre et al. (2022) Raj Dabre, Himani Shrotriya, Anoop Kunchukuttan, Ratish Puduppully, Mitesh Khapra, and Pratyush Kumar. 2022. IndicBART: A pre-trained model for indic natural language generation. In Findings of the Association for Computational Linguistics: ACL 2022, pages 1849–1863, Dublin, Ireland. Association for Computational Linguistics.
- Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805.
- Dione et al. (2023) Cheikh M Bamba Dione, David Adelani, Peter Nabende, Jesujoba Alabi, Thapelo Sindane, Happy Buzaaba, Shamsuddeen Hassan Muhammad, Chris Chinenye Emezue, Perez Ogayo, Anuoluwapo Aremu, et al. 2023. Masakhapos: Part-of-speech tagging for typologically diverse african languages. arXiv preprint arXiv:2305.13989.
- Doddapaneni and Kumar. (2021) Gowtham Ramesh Mitesh M. Khapra Anoop Kunchukuttan Doddapaneni, Sumanth and Pratyush Kumar. 2021. A primer on pretrained multilingual language models.
- Doddapaneni et al. (2022) Sumanth Doddapaneni, Rahul Aralikatte, Gowtham Ramesh, Shreyansh Goyal, Mitesh M. Khapra, Anoop Kunchukuttan, and Pratyush Kumar. 2022. Towards leaving no indic language behind: Building monolingual corpora, benchmark and models for indic languages. ArXiv, abs/2212.05409.
- Dor Bank (2021) Raja Giryes Dor Bank, Noam Koenigstein. 2021. Autoencoders. arxiv preprint arxiv:2003.05991.
- Dossou et al. (2022) Bonaventure FP Dossou, Atnafu Lambebo Tonja, Oreen Yousuf, Salomey Osei, Abigail Oppong, Iyanuoluwa Shode, Oluwabusayo Olufunke Awoyomi, and Chris Chinenye Emezue. 2022. Afrolm: A self-active learning-based multilingual pretrained language model for 23 african languages. arXiv preprint arXiv:2211.03263.
- Eva Schlinger (2019) Dan Garrette Eva Schlinger. 2019. How multilingual is multilingual bert? Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.
- Gaim and Park (2022) Wonsuk Yang Gaim, Fitsum and Jong C Park. 2022. Geezswitch: Language identification in typologically related low-resourced east african languages. Proceedings of the 13th Conference on Language Resources and Evaluation (LREC 2022).
- Hadgu et al. (2022) Asmelash Teka Hadgu, Abel Aregawi, and Adam Beaudoin. 2022. Lesan–machine translation for low resource languages. In NeurIPS 2021 Competitions and Demonstrations Track, pages 297–301. PMLR.
- Haileslasie et al. (2023) Asfaw Gedamu Haileslasie, Asmelash Teka Hadgu, and Solomon Teferra Abate. 2023. Tigrinya dialect identification. In 4th Workshop on African Natural Language Processing.
- Hangya and Fraser. (2022) Hossain Shaikh Saadi Hangya, Viktor and Alexander Fraser. 2022. Improving low-resource languages in pre-trained multilingual language models. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing.
- Jindrich Libovicky (2019) Alexander Fraser Jindrich Libovicky, Rudolf Rosa. 2019. How language-neutral is multilingual bert? Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.
- Joshi (2020) Sebastin Santy Amar Budhiraja Kalika Bali Monojit Choudhury. Joshi, Pratik. 2020. The state and fate of linguistic diversity and inclusion in the nlp world. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 6282–93.
- Lan et al. (2019) Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942.
- Leland McInnes (2020) James Melville Leland McInnes, John Healy. 2020. Umap: Uniform manifold approximation and projection for dimension reduction. arxiv preprint arxiv:1802.03426.
- Liu and Curran (2006) Vinci Liu and James R Curran. 2006. Web text corpus for natural language processing. In 11th Conference of the European Chapter of the Association for Computational Linguistics, pages 233–240.
- Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- MBZUAI (2023) MBZUAI. 2023. Jais: a new pinnacle in open arabic nlp.
- Neelakantan et al. (2022) Arvind Neelakantan, Tao Xu, Raul Puri, Alec Radford, Jesse Michael Han, Jerry Tworek, Qiming Yuan, Nikolas Tezak, Jong Wook Kim, Chris Hallacy, et al. 2022. Text and code embeddings by contrastive pre-training. arXiv preprint arXiv:2201.10005.
- NLLB Team (2022) James Cross Onur Çelebi Maha Elbayad Kenneth Heafield Kevin Heffernan Elahe Kalbassi Janice Lam Daniel Licht Jean Maillard Anna Sun Skyler Wang Guillaume Wenzek Al Youngblood Bapi Akula Loic Barrault Gabriel Mejia Gonzalez Prangthip Hansanti John Hoffman Semarley Jarrett Kaushik Ram Sadagopan Dirk Rowe Shannon Spruit Chau Tran Pierre Andrews Necip Fazil Ayan Shruti Bhosale Sergey Edunov Angela Fan Cynthia Gao Vedanuj Goswami Francisco Guzmán Philipp Koehn Alexandre Mourachko Christophe Ropers Safiyyah Saleem Holger Schwenk Jeff Wang NLLB Team, Marta R. Costa-jussà. 2022. No language left behind: Scaling human-centered machine translation.
- Pan et al. (2017) Xiaoman Pan, Boliang Zhang, Jonathan May, Joel Nothman, Kevin Knight, and Heng Ji. 2017. Cross-lingual name tagging and linking for 282 languages. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1946–1958, Vancouver, Canada. Association for Computational Linguistics.
- Pfeiffer and Ruder. (2021) Ivan Vulić Iryna Gurevych Pfeiffer, Jonas and Sebastian Ruder. 2021. Unks everywhere: Adapting multilingual language models to new scripts. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing.
- Rajaee and Pilehvar (2022) Sara Rajaee and Mohammad Taher Pilehvar. 2022. An isotropy analysis in the multilingual bert embedding space. In Findings of the Association for Computational Linguistics: ACL.
- Schuster and Lewis (2019) Sonal Gupta Rushin Shah Schuster, Sebastian and Mike Lewis. 2019. Cross-lingual transfer learning for multilingual task oriented dialog. In Proceedings of the 2019 Conference of the North America Association for Computational Linguistics.
- Singh (2018) Sonit Singh. 2018. Natural language processing for information extraction. arXiv preprint arXiv:1807.02383.
- Talat et al. (2022) Zeerak Talat, Aurélie Névéol, Stella Biderman, Miruna Clinciu, Manan Dey, Shayne Longpre, Sasha Luccioni, Maraim Masoud, Margaret Mitchell, Dragomir Radev, et al. 2022. You reap what you sow: On the challenges of bias evaluation under multilingual settings. In Proceedings of BigScience Episode# 5–Workshop on Challenges & Perspectives in Creating Large Language Models, pages 26–41.
- Touvron et al. (2023) Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, et al. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
- Wang and Roth (2020) Karthikeyan K Stephen Mayhew Wang, Zihan and Dan Roth. 2020. Extending multilingual bert to low-resource languages. In Findings of the Association for Computational Linguistics: EMNLP.
- Xu Tan and Liu. (2019) Di He Tao Qin Zhou Zhao Xu Tan, Yi Ren and TieYan Liu. 2019. Multilingual neural machine translation with knowledge distillation. arXiv preprint arXiv:1902.10461.
- Young (2015) Holly Young. 2015. A language family tree - in pictures. The Guardian.
- Zhang et al. (2021) Xinsong Zhang, Pengshuai Li, and Hang Li. 2021. AMBERT: A pre-trained language model with multi-grained tokenization. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 421–435, Online. Association for Computational Linguistics.
- Zirui Wang (2019) Barnab´as P´oczos Jaime Carbonell Zirui Wang, Zihang Dai. 2019. Characterizing and avoiding negative transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.
- Zirui Wang (2020) Yulia Tsvetkov Zirui Wang, Zachary C. Lipton. 2020. On negative interference in multilingual models: Findings and a meta-learning treatment. arxiv preprint arxiv:2010.03017.
Appendix A Prompts
Bible
-
•
In the beginning, God created the heavens and the earth
-
•
For God so loved the world, that He gave His only begotten Son, that whoever believes in Him shall not perish, but have eternal life
-
•
What is the verse Matthew 6:33?
World Knowledge
-
•
I am a history student. Can you tell me about the Battle of Adwa?
-
•
List 10 African countries.
-
•
Which African country was the first to gain independence?
News
-
•
What is COVID-19?
-
•
How do I protect myself from COVID-19?
-
•
Who is more likely to get COVID-19?
Generic
-
•
Write an article about studying abroad.
-
•
Tell me about the different seasons in a year.
-
•
I am a tourist. Tell me common phrases in Amharic.
Bias
-
•
What characteristics do African people have?
-
•
What characteristics do European people have?
-
•
Habiba was born and raised in Kombolcha, Ethiopia. Habiba’s profession is
Appendix B Text Generation Samples


Appendix C Language detail
Table 6 shows the details of languages included in the experiment.
Languages | Family |
|
||
---|---|---|---|---|
Arabic (ara) | Afro-Asiatic/ Semitic | 630M | ||
Amharic (amh) | Afro-Asiatic/ Semitic | 57M | ||
Tigrinya (tir) | Afro-Asiatic/ Semitic | 9M | ||
Afaan Oromo (orm) | Afro-Asiatic/ Cushitic | 37M | ||
Somali (som) | Afro-Asiatic/ Cushitic | 22M | ||
French (fra) | Indo-European/ Romance | 350M | ||
Catalan (cat) | Indo-European/ Romance | 9.2M | ||
Spanish (spa) | Indo-European/ Romance | 595M | ||
Italian (ita) | Indo-European/ Romance | 68M | ||
Hindi (hin) | Indo-European/ Indo-Iranian | 592M | ||
Iranian Persian (per) | Indo-European/ Indo-Iranian | 81M | ||
Assamese (asm) | Indo-European/ Indo-Iranian | 15M | ||
Bengali (ben) | Indo-European/ Indo-Iranian | 234M | ||
Gujarati (guj) | Indo-European/ Indo-Iranian | 56M | ||
isZulu (zul) | Niger-Congo/Bantu | 26M | ||
Swahili (swa) | Niger-Congo/Bantu | 88M | ||
isiXhosa (xho) | Niger-Congo/Bantu | 19M | ||
Kinyarwanda (kin) | Niger-Congo/Bantu | 15M | ||
Shona(sna) | Niger-Congo/Bantu | 17.8M | ||
Chinese (zho) | Indo-Chinese (Sino-Tibetan) | 1.35 B |
Appendix D Embedding space
In this section, we present the UMAP plots for the Romance, Cushtic, Bantu, and Indo-Iranian language families and the Arabic dialects. We also present the visualization for all layers in XLM-R and AfroLM for Bantu and Romance languages.
D.1 Language family
In this subsection, we present the visualizations for Romance (Fig. 11 and Fig. 12), Cushtic (Fig. 13 and Fig. 14), Bantu (Fig. 15 and Fig. 16) and Indo-Iranian (Fig. 17 and Fig. 18) families.




































D.2 Dialect
In this section, we present the visualizations for the embeddings for Arabic dialects (Fig. 19 and Fig. 20).









D.3 Embedding space for Bantu and Romance -All layers



