[1]\fnmAloka \surFernando
1]\orgdivDept. of Computer Science & Engineering, \orgnameUniversity of Moratuwa, \orgaddress\postcode10400 \countrySri Lanka
2]\orgnameSchool of Mathematical and Computational Sciences, Massey University, \orgaddress\streetPalmerston North, \postcode4443, \countryNew Zealand
Linguistic Entity Masking to Improve Cross-Lingual Representation of Multilingual Language Models for Low-Resource Languages
Abstract
Multilingual Pre-trained Language models (multiPLMs), trained on the Masked Language Modelling (MLM) objective are commonly being used for cross-lingual tasks such as bitext mining. However, the performance of these models is still suboptimal for low-resource languages (LRLs). To improve the language representation of a given multiPLM, it is possible to further pre-train it. This is known as continual pre-training. Previous research has shown that continual pre-training with MLM and subsequently with Translation Language Modelling (TLM) improves the cross-lingual representation of multiPLMs. However, during masking, both MLM and TLM give equal weight to all tokens in the input sequence, irrespective of the linguistic properties of the tokens. In this paper, we introduce a novel masking strategy, Linguistic Entity Masking (LEM) to be used in the continual pre-training step to further improve the cross-lingual representations of existing multiPLMs. In contrast to MLM and TLM, LEM limits masking to the linguistic entity types nouns, verbs and named entities, which hold a higher prominence in a sentence. Secondly, we limit masking to a single token within the linguistic entity span thus keeping more context, whereas, in MLM and TLM, tokens are masked randomly. We evaluate the effectiveness of LEM using three downstream tasks, namely bitext mining, parallel data curation and code-mixed sentiment analysis using three low-resource language pairs English-Sinhala, English-Tamil, and Sinhala-Tamil. Experiment results show that continually pre-training a multiPLM with LEM outperforms a multiPLM continually pre-trained with MLM+TLM for all three tasks.
keywords:
Masked Language Modelling, Translation Language Modelling, multilingual Pre-trained Language Model, bitext mining, sentiment analysis, XLM-R, Sinhala, Tamil1 Introduction
Encoder-based Multilingual Pre-trained Language Models (multiPLMs) such as mBERT [1] and XLM-R [2] produce State-of-the-art results for many Natural Language Processing (NLP) tasks in the context of low-resource languages (LRLs) [3, 4]. One success factor of these multiPLMs is the training objective utilized during the pre-training stage. mBERT and XLM-R were pre-trained using the Masked Language Modeling (MLM) objective. However, the performance of these models for cross-lingual tasks such as bitext mining had been suboptimal [5], due to the lack of an explicit cross-lingual pre-training objective [6]. Translation Language Modeling (TLM) objective [7] was introduced to improve the cross-lingual capability of the existing multiPLMs. In contrast to MLM that uses only monolingual data, TLM uses parallel data across multiple languages in a continual pre-training step to further improve the cross-lingual representations. Conneau and Lample [7] had proven that TLM improved the performance of cross-lingual classification and unsupervised Neural Machine Translation (NMT).
From a linguistic perspective, different words in a sentence have different linguistic properties. Previous work has demonstrated that Pre-trained Language Models (PLMs) capture the notion of syntactic structures and grammatical properties in the language [8, 9, 10]. Named Entities (NEs), Verbs and Nouns significantly contribute to defining the syntactic structure and the semantics of the sentence. Further, these elements play a crucial role in establishing syntactic relationships such as subject-verb agreement, which is generally stronger than those between other words in the sentence. To highlight the prominence of NEs, Verbs and Nouns in a sentence, we visualize the self-attention weight matrix in terms of a heatmap for an English sentence Jack walks towards the road, and its Sinhala translation in Figure 1. In the English sentence, the words ”Jack” (NE) and ”walk” (Verb) get the highest attention from other words. Similarly, the words which gets the highest attention in Sinhala is a Named Entity and a Noun.


Based on this hypothesis, we introduce a linguistically motivated masking strategy named Linguistic Entity Masking (LEM). In LEM, we limit to masking a single token from the linguistic entity span. As linguistic entities, we consider NEs, nouns and verbs in the sentence. Masking a single token contrasts existing span masking techniques [11, 12, 13], which have masked consecutive token spans of the selected n-gram words. We apply LEM on both monolingual (, akin to MLM) and parallel sentences (, akin to TLM), and conduct continual pre-training on the multiPLM.
We use XLM-R as the multiPLM in our experiments. We continually pre-train XLM-R with objectives. The resulting model is referred to as XLM- hereafter. We continually pre-train this same multiPLM with MLM+TLM as well, which serves as our baseline (we term the resulting model XLM-RMLM+TLM). We evaluate these models on bitext mining, parallel data curation, as well as code-mixed sentiment classification downstream tasks on three LRL pairs English-Sinhala, English-Tamil, and Sinhala-Tamil. Our results showcase that XLM- outperforms XLM-RMLM+TLM for these tasks.
We carry out extensive experiments (1) to determine the type of monolingual data that is most effective in the first continual pre-training step. ie. dependent monolingual data (source and target sides of parallel data taken separately as monolingual data) vs independent monolingual data (monolingual data without any explicit translation content between the two languages), (2) to empirically evaluate the existing masking strategies (Sub-word masking [1], whole-word masking [1] and span Masking [12]), (3) to identify the most contributing linguistic entity or combination of linguistic entities for masking, (4) to determine the optimal number of tokens for masking within a linguistic entity and (5) to determine the impact of using noisy parallel sentences in the continual pre-training stage.
Our contributions are as follows:
-
•
We introduce a novel masking strategy, Linguistic Entity Masking (LEM), to improve the cross-lingual representations of existing multiPLMs. We show that XLM- outperforms XLM-RMLM+TLM for the three downstream tasks, considering English and two LRLs Sinhala and Tamil.
-
•
We conduct an empirical study of the existing masking strategies to evaluate the effectiveness of them on the bitext mining task for LRLs.
-
•
We conduct ablation experiments to find the most contributing linguistic entity (among NEs, nouns and verbs) for masking and the number of tokens to mask within the linguistic entity span.
-
•
We show that using dependent monolingual corpora than the independent monolingual corpora during the continual pre-training step is favourable for improving the cross-lingual representations.
The rest of the paper is organized as follows. In Section 2 we discuss the related work in the context of pre-training objectives and masking strategies. In Section 3 we describe our methodology, assumptions along with the theoretical justification for the LEM strategy. We describe the experiments in Section 4 and detail out the experimental setup in Section 5. The results are discussed comprehensively in Section 6, while additional ablation studies are detailed in Section 7. The limitations of our approach and potential future directions are highlighted in the Discussion Section 8, and the paper conclusion is in Section 9.
2 Related Work
2.1 MLM and TLM Objectives
Encoder-based multiPLMs such as mBERT and XLM-R were trained on monolingual data using the Masked Language Modeling (MLM) objective. These models have significantly enhanced the performance of various downstream tasks [14, 15, 16].
In BERT (and its multiPLM variant, mBERT), which was trained with MLM111BERT was also trained using the next sentence prediction task., 15% of the input tokens were randomly selected for corrupting, following a uniform distribution. Out of these, 80% of the time the tokens were replaced with a [MASK] token, 10% of the time the tokens were replaced with a random token and 10% of the time they were left unchanged. Here the MLM objective predicts the corrupted (both masked and replaced) tokens. Successor models such as XLM-R [2] adopt the same 15% masking percentage and 80%-10%-10% corruption rule during pre-training. The contextualized representations produced by these pre-trained models are then used to obtain sentence embeddings for downstream NLP tasks. However, these models have been reported to be suboptimal for cross-lingual tasks such as bitext mining, due to the lack of an explicit objective for improving cross-lingual representations [6].
To address this limitation, Conneau and Lample [7] introduced Translation Language Modeling (TLM), which extended the MLM objective using parallel data. TLM accepts a concatenated pair of parallel sentences as input, and tokens were masked from both sentences. The rationale was to utilize the context of its translation counterpart to accurately predict the masked token, there by strengthening the cross-lingual capability. TLM was applied in a continual pre-training step, on top of the MLM pre-trained model. In this setting, the MLM step was still required to learn the linguistic information inherent to the languages, while the TLM step strengthened the cross-lingual signal across the language pairs.
2.2 Different Masking Strategies
In BERT’s MLM strategy, the sub-words were masked. Subsequent work experimented by varying the type of tokens for masking, as summarized in Table 1. The follow-up work masked consecutive sub-words in text spans [12] and correlated text spans with Point-wise Mutual Information (PMI) masking [13]. Zhuang [17] proposed a heuristic-masking strategy where they considered the unmasked token prediction in addition to the masked token prediction during language model pre-training. Golchin et al. [18] utilized an in-domain keyword masking strategy for domain adaptation of the PLM. Most of these techniques have primarily relied on monolingual data and have been evaluated predominantly on high-resource languages.
Closest to our work is Entity/Phrase masking [11]. This contrasts with our work in three ways. Entity/Phrase masking, selected NEs, noun phrases and verb phrases for masking. As per analysis in Fig 1, we considered only verb and noun words, in addition to NEs for masking. Secondly, while Entity/Phrase masking masked all consecutive tokens within NEs or Noun/Verb phrases identified by a chunking tool, LEM takes a more targeted approach by limiting masking to a single token within the selected linguistic entity span. Finally, they followed a multi-staged pre-training approach. The pre-training stages were, sub-word masking similar to BERT, followed by Phrase masking and Named Entity masking. In comparison, our approach’s two continual pre-training stages apply the same LEM strategy with monolingual data and parallel data. They have evaluated the strategy only on High Resource Languages English and Chinese. Further, this strategy has not been extended with parallel data for cross-lingual improvement.
Masking Strategy | Pre-training | Masked token Type |
---|---|---|
Sub-word Masking | Pre-training | sub-words |
Whole-Word Masking | Pre-training | all sub-words in the word |
Entity/Phrase Masking [11] | Multi-stage Pre-training | all sub-words in the Named Entity/Noun Phrase |
Span Masking (spanBERT) [12] | Pre-training | all sub-words in the word n-gram span |
Point-wise Mutual Information (PMI) Masking [13] | Pre-training | all sub-words in the correlated word-spans |
Wettig et al. [19] conducted an empirical study on what tokens should be masked and in what percentages. However, their study was limited to the English language and focused only on downstream tasks such as classification and question-answering. To date, no empirical study has explored these alternative masking strategies specifically for sentence retrieval tasks, particularly in the context of LRL pairs.
3 Methodology
In this section, we discuss in detail the LEM strategy. A comparison between our masking strategy and existing masking strategies is shown in Figure 2. Instead of pre-training a multiPLM from scratch—a computationally expensive process—we leverage LEM in a continual pre-training step. This is a widely adopted approach to improve multiPLMs with respective to representation improvements [7, 20].

Figure 3 illustrates our two-stage continual pre-training process. Similar to the MLM and TLM training sequence used in XLM [7], on top of the multiPLM, we apply the continual pre-training with monolingual and parallel data respectively.

.
3.1 Theoretical Framework for Linguistic Entity Masking (LEM)
The theoretical framework of LEM in a monolingual setting () can be described as follows:
Let the monolingual sequence be defined as where is a word and is the number of words in the sequence. After tokenization, sequence can be represented as as in Eq. 1.
(1) |
Here, is a token (sub-word) and is the total number of sub-words returned by the tokenizer. From this sequence, the linguistic entities NEs, verbs and nouns are identified, and can now be represented as a collection of linguistic entities as shown in Eq. 2. From these linguistic entities, a single token is sampled over a uniform distribution, up to a total of 15% for masking. If 15% cannot be obtained from linguistic entities, the remainder would be sampled from the remaining tokens. We use the same corruption rule, 80%-10%-10% as BERT.
(2) |
During training, the cross-entropy loss () for masked token prediction, as in Eq. 3 is minimized. In the equation, N is the total number of tokens for prediction and is the expected true label.
(3) |
Finally, we extend the TLM objective with LEM into the parallel data setting (LEMpara). Here we concatenate the source sentence () and target sentence () from the parallel sentence-pair as a single input example and obtain the tokenized output as represented by in Eq. 4. and are the tokenized source and target sentences, respectively. and are the number of tokens (sub-words) in the source and target sentences (respectively) after tokenization.
(4) |
Similar to LEMmono, in this step, a single token from each linguistic entity (NE, verb or noun) from both languages are selected for corruption according to the 80%-10%-10% rule. If 15% of linguistic units were not found in the sequence, the balance is sampled from the remaining tokens. During training, the corrupted token prediction cross-entropy loss, () (Eq. 5) is minimized. S and T correspond to the total number of tokens masked from the source and target side sentences respectively. and are the true tokens to be predicted.
(5) |
Languages such as Sinhala and Tamil exhibit morphological richness, requiring words to be inflected based on attributes such as number, gender, and case category. Table 2 shows examples for such word inflections. Additionally, the presence of out-of-vocabulary words in LRLs often leads to an increased number of sub-words after tokenization. Therefore, approaches like whole-word masking, span masking, or entity/phrase masking tend to mask longer spans. This reduction in context weakens the ability to accurately predict the masked tokens, ultimately hindering representation learning. In contrast, LEM mitigates this issue by masking a single token from a linguistic entity, which we empirically prove in Section 7.1.
![]() |
4 Experiments
4.1 Impact of the type of monolingual data in
We experiment with independent and dependent monolingual data to observe its impact on the continual pre-training step . We sample 60,000 sentences from the MADLAD-400 and all the available 60,000 sentences from SiTa-Trilingual dataset for each language. Next, we continually pre-train XLM-R with those datasets separately with and evaluate on the bitext evaluation dataset.
To assess whether increasing the independent training data would yield in any improvements, we repeat the experiment with a sample size of 100,000 from MADLAD-400. Finally, as an extreme case, we conduct a third experiment using 500,000 sentences for the Sinhala-Tamil language pair. Here, 60,000, 100,000, and 500,000 correspond to the training data size per language, with the full training set size being double the amount specified. We conduct this evaluation for all three language pairs.
4.2 Evaluation of Different Masking Strategies
We empirically evaluate various masking strategies and assess their performance on the bitext mining task. The masking strategies explored in this study are as follows:
Sub-word Masking - Following the BERT MLM, with each sentence, 15% of tokens are selected randomly and corrupted according to 80%-10%-10% rule.
Whole Word Masking - All the sub-words corresponding to the randomly sampled words are masked. A total of 15% tokens are sampled and corrupted according to 80%-10%-10% rule.
Span Masking - Consecutive word spans are sampled over a geometrical distribution and 15% of tokens are masked. The masking is limited to whole-word tokens as defined in the original work.
4.3 Evaluation of LEM Strategy and Ablation Study
This section describes the ablation experiments we conduct to determine the most contributing linguistic entity or their combination in the LEM strategy. We use the baselines as described in Section 5.3.
We identify the NEs in English, Sinhala, and Tamil sentences, using an in-house fine-tuned multilingual NER model [21]. To identify nouns and verbs in the sequences, we employ Flair POS tagger [22] for English, the Sinhala TnT POS Tagger [23, 24] for Sinhala, and ThamizhiUDp [25] for Tamil. Flair reported an F1 score of 98.19% and is the best model for English POS Tagging. The Sinhala POS Tagger had been trained using SVM and has an overall accuracy of 84.68% with a 59.86% accuracy for tagging unknown words. The Tamil POS Tagger is a neural-based model, with a F1 score of 93.27%. For Sinhala and Tamil, these are the models that returned optimal results.
The initial ablation experiment masks a single linguistic entity type, such as only NEs, only verbs, or only nouns. Subsequently, combinations of these linguistic entity types are examined.
In our experiments, 100%NE+15%MLM means, priority is given for sampling from NEs. If it does not produce enough tokens for masking, then the balance is sampled from the remaining tokens. When combining several linguistic entities, e.g. 100%NE+100%VB+15%MLM means, priority is given to sample the tokens for masking from both NEs and verbs.
4.4 Evaluation Tasks
We evaluate the success of our LEM masking strategy on three downstream tasks - bitext mining, parallel data curation and code-mixed sentiment classification.
4.4.1 Bitext Mining
Bitext mining is a sentence retrieval task that retrieves a target language translation for a given source sentence or vice versa from a document-aligned dataset. The performance of bitext mining relies heavily on the quality of cross-lingual embeddings. Recent bitext mining techniques are embedding-based, where they identify the translation pairs based on the semantic distance between the source sentences and candidate target sentences. Improvements in bitext mining techniques can be categorized into two: (1) refining the semantic similarity distance calculation function [26, 27] between sentence embeddings and (2) enhancing cross-lingual sentence representations [28, 29, 20]. Our LEM strategy aims to improve the latter.
We obtain the sentence embeddings from XLM-RLEM as well as from the baselines, and use margin-based cosine similarity [26] function to identify parallel sentence pairs. We choose margin-based cosine similarity over conventional cosine similarity for this task due to its lower rate of false positives. Then we rank the parallel sentences according to their similarity scores. Bitext mining is performed using the three criteria: Forward (FW), Backward (BW), and Intersection (IN) [26]. FW retrieves the target sentence for each source sentence, BW retrieves the source sentence for each target sentence, and IN considers the intersection of the parallel sentences retrieved using FW and BW criteria. The Recall evaluation metric is used to report the bitext mining results.
4.4.2 Parallel Data Curation
Although large-scale bitext mining [30, 31] alleviates the parallel data scarcity problem in NMT, they are mostly noisy [32, 33]. Therefore, a parallel data curation step is crucial to filter out noisy parallel sentences from the corpus. This is done by obtaining sentence representations from a multiPLM and by calculating the semantic similarity using cosine distance [20, 31] between each parallel sentence pair. Then the parallel sentences are sorted in descending order and the top-most ranked parallel sentences are used to train the NMT system.
To further evaluate the effectiveness of these models with improved cross-lingual representations, we conduct a parallel corpus curation task and perform an extrinsic evaluation by training NMT systems with the top-ranked sentences.
First, we rank the parallel sentences for translation quality using the baseline (Section 5.3) and the XLM-RLEM models. Using the top 50,000 sentences from the ranked parallel sentence pairs, we train NMT systems for each language pair. NMT scores are reported on the Flores+ devtest, using sacreBLEU [34], ChrF [35], ChrF++ [36] and spBLEU [37] metrics. We base the discussion of the results using the ChrF metric while we include the full results Table 13 in Appendix B.
4.4.3 Code-Mixed Sentiment Analysis
MultiPLMs fine-tuned on task-specific monolingual datasets has achieved the state-of-the-art performance in sentiment analysis tasks [38]. However, sentiment analysis on code-mixed data remains a challenging task. Code-mixing [39] occurs when linguistic units—such as phrases, words, or morphemes—from one language are embedded into the utterance or sentence of another language. In this setting, the performance of sentiment analysis on code-mixed data largely depends on the quality of the cross-lingual embeddings learned by the MultiPLM.
We conduct the baseline experiments (Section 5.3) and compare them with the results we obtain with the XLM-RLEM model. Here we use each encoder model by fine-tuning them on an English-Sinhala code-mixed task. We follow a two-step fine-tuning, where the first fine-tuning is done using the English Amazon product review sentiment analysis dataset. Here we report the zero-shot scores on the code-mixed EnSi evaluation set. Finally, the intermediate model is further fine-tuned on the English-Sinhala code-mixed dataset and the results are reported. We use Precision, Recall and F1 to report the evaluation scores for the sentiment analysis task.
5 Experiment Setup
5.1 Data Selection
We consider English, Sinhala and Tamil for our experiments. Sinhala and Tamil are the official languages of Sri Lanka and English is used as a link language. They belong to three distinct language families; Indo-European, Indo-Aryan and Dravidian language families, respectively. Further, Sinhala and Tamil are low-resource and medium-resource languages respectively [40, 41]. They are also morphologically rich languages. Sinhala, in particular, is only used in the island nation of Sri Lanka and has seen slow progress in language technologies [41, 42].
Monolingual and Parallel Data: As elaborated in Section 4.1, we carry out an ablation study to determine the type of monolingual data that is most suitable for the first continual pre-training step. We obtain the independent monolingual data from MADLAD-400 [43]. It is a collection of document-level data of 3 Trillion tokens from Common Crawl222https://commoncrawl.org/ for 419 languages. As the dependent-monolingual data, we obtain the monolingual sides from the SiTa-Trilingual parallel dataset [44]. It is a human-curated gold standard three-way parallel dataset between Sinhala-Tamil-English languages with 60,000 training data.
We preprocess the MADLAD-400 data to extract clean sentences for each language as follows: First, the document-level data is segmented into sentences using the nltk333https://www.nltk.org/index.html sentence tokenizer. We then filter these sentences using the LID (Language IDentification) model444https://github.com/gordicaleksa/Open-NLLB. Subsequently, we remove noisy data, including HTML tags, URLs, and sentences with less than 60% textual content. Finally, religious texts were excluded through a keyword filter.555Keywords being bible book names along with common words from the bible. However, for SiTa-Trilingual data, no preprocessing was applied as the data was of high quality.
NLLB/CCAligned Datasets: For the parallel data curation task, we obtain parallel data from NLLB [31] and CCAligned [45] corpora. Both these corpora provide parallel data for the three language pairs: English-Sinhala, English-Tamil, and Sinhala-Tamil. NLLB and CCAligned are known to be noisy parallel data for the considered language pairs [33].
ParaCrawl Dataset: To analyse the performance of LEM strategy with noisy data, we select the English-Sinhala ParaCrawl [46] dataset. It is a web-mined parallel corpus with 217,412 parallel sentences.
Code-mixed Sentiment Classification pairsDataset: For the code-mixed sentiment classification task, we use the English-Sinhala code-mixed dataset [47] of 13,521 sentences. We have experimented with the English-Sinhala language pair.
Amazon Product Review Dataset:
To report the zero-shot scores for the English-Sinhala sentiment classification task, we used the Amazon product review sentiment analysis dataset 666https://www.kaggle.com/datafiniti/consumer-reviews-of-amazon-products. The full dataset has training data of 3.6M and test data 400,000 respectively. During fine-tuning, we sample only 100,000 as training data, in order to avoid catastrophic forgetting [48] of the cross-lingual representations in XLM-
Trilingual Bitext Mining Evaluation Set: For the bitext mining task, we use an existing human-created dataset [27]. It consists of trilingual data obtained from four Sri Lankan news sources Army777https://www.army.lk/, Hiru888https://www.hirunews.lk/, ITN999https://www.itnnews.lk/ and Newsfirst101010https://english.newsfirst.lk/. For each news source, there are human-aligned 300 sentence-pairs.
Flores+ Evaluation Set: For the NMT experiments, we use dev and devtest splits from Flores+111111https://github.com/openlanguagedata/flores as the validation and evaluation sets, respectively.
5.2 multiPLM Selection
We select XLM-R as the base multiPLM for our experiments. Other popular multiPLMs, XLM and mBERT do not cover Sinhala language. XLM-R has already demonstrated promising performance in downstream tasks for the low-resource languages considered in this study [4, 47, 49, 21]. It is a 278M parameter model, pre-trained for 100 languages. However, the amount of Sinhala and Tamil data used during XLM-R pre-training is much lower than that for English (English 55B, Sinhala 243M, Tamil 595M).
5.3 Baselines
In our evaluation of downstream tasks, we set up two baseline experiments.
-
•
XLM-R - Obtain embeddings from the out-of-the-box XLM-R pre-trained model.
-
•
XLM-RMLM+TLM [7] - We continually pre-train the XLM-R with MLM+TLM objectives and use the representation improved encoder to obtain embeddings.
5.4 Implementation and Hyper-parameters
5.4.1 Linguistic Entity Masking (LEM)
We customize the MLM training implementation released with the sentence-transformers121212https://www.sbert.net/ library (built on Huggingface transformers131313https://huggingface.co/docs/transformers/index), to support XLM-R tokenization and implement the LEM strategy. Each continual pre-training experiment is executed for 60 epochs with early stopping. Then the checkpoint with the least validation loss is selected as the best-performing model. The experiments are conducted on Nvidia Quadro RTX 6000 GPU with 24GB VRAM. The hyper-parameters of XLM-R 141414https://huggingface.co/FacebookAI/xlm-roberta-base model and other training parameters used in the continual pre-training experiments are shown in Table 3.
Hyperparameter | Argument value |
---|---|
No of Layers | 12 |
Hidden Size | 768 |
Attention Heads | 12 |
hidden_dropout_prob | 0.1 |
Learning Rate | 5e-3 |
Training batch-size | 32 |
Sequence Length | 120 |
Adam | 1 e-08 |
Adam | 0.9 |
Adam | 0.99 |
5.4.2 NMT Experiments
We obtain the top-ranked sentences for the NMT experiments and train a Sentencepiece151515https://github.com/google/sentencepiece tokenizer with a vocabulary size of 25000. Then we use fairseq toolkit [50] to model and train the vanilla transformer-based Sequence-to-Sequence NMT model. The experiments are conducted on a Nvidia Quadro RTX6000 GPU with 24GB VRAM. The hyper-parameters used during training along with the training parameters are shown in Table 4. Each experiment is run for 100 epochs with the early stopping criteria.
Hyper-parameter | Argument value |
---|---|
encoder/decoder Layers | 6 |
encoder/decoder attention heads | 4 |
encoder-embed-dim | 512 |
decoder-embed-dim | 512 |
encoder-ffn-embed-dim | 2048 |
decoder-ffn-embed-dim | 2048 |
dropout | 0.4 |
attention-dropout | 0.2 |
optimizer | adam |
Adam , Adam | 0.9, 0.99 |
warmup-updates | 4000 |
warmup-init-lr | 1e-7 |
learning rate | 1e-3 |
batch-size | 32 |
patience | 6 |
fp16 | True |
5.4.3 Code-Mixed Sentiment Analysis
For the sentiment classification experiments, we fine-tune both the baseline models and the XLM-RLEM model. A linear layer is added as the classification head to facilitate binary sentiment classification. Full fine-tuning is performed, allowing updates to all model parameters, including those in the classification head, during training. The Hyper-parameters used during these experiments are presented in Table 5. We train for 20 epochs with early stopping.
Hyper-parameter | Argument value |
---|---|
Number of Layers | 12 |
Hidden Size | 768 |
Attention Heads | 12 |
Hidden dropout prob | 0.1 |
Learning Rate | 2e-5 |
Weight Decay | 0.01 |
Training Batch Size | 128 |
Sequence Length | 80 |
Adam | 1e-08 |
Adam | 0.9 |
Adam | 0.99 |
5.4.4 Improving Continual Pre-training Efficiency
Named Entity Recognition (NER) and Part-of-Speech (POS) tagging during training increases training time drastically. We introduce a pre-processing step to mitigate this issue. Specifically, a dictionary is created to store the linguistic entities, such as named entities, verbs, and nouns, for each sentence. We maintain a sub-word-level mapping in the dictionary, allowing for precise token identification while reducing computational overhead during training.
6 Results and Discussion
6.1 Impact of the type of monolingual data in
Figure 4 shows the bitext mining results for the step using independent and dependent monolingual data. The detailed results are available in Table 12 in Appendix A.

The highest performance was observed for the dependent monolingual data, a trend consistent across all three language pairs. Surprisingly, when the independent set was increased to 100,000, this increase did not surpass the scores obtained using the dependent monolingual 50,000 training dataset. This was evident in all three language pairs. When the dataset size was further increased to 500,000, the results further deteriorated.161616Due to resource constraints and the reduced performance for the Sinhala-Tamil direction, the experiment using 500,000 was not conducted for the other two language pairs. Therefore, it is evident that utilizing dependent monolingual data is advantageous during the LEMmono step to enhance cross-lingual representations.
6.2 Evaluation of Different Masking Strategies
Table 6 shows the experimental results for the bitext mining task using XLM-R continually pre-trained with different MLM strategies (Section 4.2). Based on the averaged recall scores, the baseline XLM-R consistently delivered the highest performance in bitext mining tasks. An exception was observed for the Sinhala-Tamil BW criterion, where the sub-word masking strategy outperformed the baseline. However, a key observation was that continual pre-training with the existing masking strategies in general deteriorated the already learnt cross-lingual representations in the XLM-R model.
As outlined in Section 3.1, morphologically rich, low-resource languages tend to generate a higher number of sub-word tokens during the tokenization process due to the presence of infrequent words in the sequences. Consequently, masking strategies like whole-word masking and span masking result in longer masked spans, which reduce the available context for accurate predictions. We hypothesize that this reduced context reduces the prediction accuracy, thus weakening the already learnt representations in the XLM-R model. As a result, the existing masking strategies return degraded results.
Experiment | Army | Hiru | ITN | Newsfirst | Averages | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | |
English - Sinhala | |||||||||||||||
XLM-R | 92.33 | 93.33 | 89.67 | 96.35 | 96.68 | 95.68 | 94.00 | 96.00 | 92.33 | 96.67 | 95.33 | 94.33 | 94.84 | 95.34 | 93.00 |
Sub-word Masking | 88.33 | 93.67 | 85.33 | 92.03 | 93.36 | 89.70 | 91.67 | 96.67 | 93.67 | 91.67 | 95.33 | 90.00 | 90.92 | 94.76 | 89.68 |
Whole-word Masking | 87.33 | 92.67 | 85.33 | 95.02 | 94.01 | 94.02 | 93.00 | 91.67 | 90.33 | 93.67 | 93.67 | 91.67 | 92.25 | 93.00 | 90.34 |
Span Masking | 89.00 | 89.67 | 85.00 | 95.02 | 94.02 | 92.03 | 90.33 | 91.67 | 85.67 | 93.67 | 92.67 | 90.33 | 92.00 | 92.01 | 88.26 |
English - Tamil | |||||||||||||||
XLM-R | 86.67 | 88.33 | 82.00 | 83.00 | 78.33 | 72.67 | 83.22 | 83.56 | 78.86 | 92.33 | 91.33 | 89.33 | 86.31 | 85.39 | 80.71 |
Sub-word Masking | 84.00 | 86.00 | 77.67 | 80.33 | 75.00 | 68.33 | 83.56 | 82.21 | 78.52 | 90.67 | 91.00 | 89.67 | 84.64 | 83.55 | 78.55 |
Whole-word Masking | 83.33 | 87.33 | 77.67 | 78.67 | 73.33 | 64.33 | 80.20 | 80.87 | 75.84 | 85.67 | 91.00 | 83.67 | 81.97 | 83.13 | 75.38 |
Span Masking | 82.67 | 83.00 | 75.33 | 78.67 | 76.67 | 69.33 | 83.22 | 82.22 | 76.85 | 89.67 | 90.00 | 85.67 | 83.56 | 82.97 | 76.79 |
Sinhala-Tamil | |||||||||||||||
XLM-R | 83.44 | 81.46 | 78.15 | 90.67 | 91.00 | 87.33 | 91.33 | 90.00 | 87.00 | 93.67 | 95.33 | 92.33 | 89.78 | 89.45 | 86.20 |
Sub-word Masking | 86.75 | 88.08 | 81.96 | 88.00 | 89.33 | 84.00 | 93.33 | 92.67 | 89.33 | 90.33 | 94.00 | 89.00 | 89.60 | 91.02 | 86.07 |
Whole-word Masking | 85.76 | 89.73 | 81.46 | 88.33 | 91.33 | 84.67 | 90.33 | 90.33 | 86.67 | 90.00 | 91.67 | 87.67 | 88.61 | 90.77 | 85.11 |
spanMasking | 85.78 | 85.10 | 81.79 | 88.67 | 91.00 | 87.00 | 91.00 | 91.00 | 87.33 | 89.00 | 90.67 | 84.33 | 88.61 | 89.44 | 85.11 |
6.3 Evaluation of LEM Strategy and Ablation Study
The ablation study results for considering each linguistic entity along with combinations of them, on the LEM strategy are available in Tables 14, 15, 16 in the Appendix C. In Table 7, we summarise the final gains.
For the Sinhala-Tamil language pair, compared to the scores obtained with XLM-R raw embeddings, XLM-RLEM returned the highest gain of +3.1 Recall points. Compared with XLM-RMLM+TLM, this was a gain of +1.4 points. For the English-Tamil language pair, the LEMmono step produced a reduced score compared to the XLM-R baseline. However, after the second continual pre-training step XLM-RLEM scores surpassed the XLM-R baseline by +1.2. Further, the highest gain produced by our method compared to XLM-RMLM+TLM was for the English-Tamil language pair, which was +2.4 points. With the English-Sinhala language pair, a similar behaviour was observed. Here XLM-RLEM compared to baseline XLM-R has a gain of +0.4 points while the improvement compared to XLM-RMLM+TLM was +1.7 scores.
In all these language pairs, the best scores were produced when the linguistic entity NEs were included in the LEM strategy. Due to the consistent gains across the three language pairs, we can safely conclude that LEM is favourable to improving the cross-lingual representations in existing multiPLMs.
Average Gains | Overall Average | |||
FW | BW | IN | Gain | |
Sinhala-Tamil | ||||
XLM-RLEM vs XLM-R | +2.36 | +4.14 | +2.90 | +3.1 |
XLM-RLEM vs XLM-RMLM+TLM | +1.95 | +0.48 | +1.83 | +1.4 |
English-Tamil | ||||
XLM-RLEM vs XLM-R | +0.75 | +1.59 | +1.17 | +1.2 |
XLM-RLEM vs XLM-RMLM+TLM | +2.34 | +1.84 | +2.92 | +2.4 |
English-Sinhala | ||||
XLM-RLEM vs XLM-R | +0.25 | +0.50 | +0.42 | +0.4 |
XLM-RLEM vs XLM-RMLM+TLM | +1.50 | +1.50 | +2.08 | +1.7 |
6.4 Parallel Data Curation
Table 8 shows the NMT results for training the top 50,000 sentence pairs obtained by ranking the parallel sentences with the baseline and XLM-RLEM models for the NLLB and CCAligned corpora. It can be observed consistently that both XLM-RMLM+TLM and XLM-RLEM improved models outperform the baseline XLM-R scores.
The NMT results show that the XLM-RLEM model produce superior results compared to XLM-RMLM+TLM, for all three language pairs across the two corpora. We believe the magnitude of the gain is dependent on the characteristics of the parallel corpus and the size of the training data sample. For the English-Tamil language pair, the CCAligned corpus produce a significant gain for XLM-RLEM compared to XLM-RMLM+TLM. This justifies the effectiveness of the LEM strategy which was not evident with random masking followed in MLM+TLM. The rest of the gains vary from +0.3 to +0.8 ChrF points. According to metric analysis by Kocmi et al. [51], these gains are equivalent to +0.48 to +1.12 BLEU points with a human accuracy of 54.2% to 66% respectively. This means the improvement in the translation quality in the NMT systems is almost in line with a minimum human accuracy rating of 54.2% to 66%.
The observations are consistent with other metrics, as shown in Table 13 in the Appendix B as well. The results further prove that the scoring from the XLM-RLEM model has managed to identify quality sentence pairs more than the other models. Therefore, improvement in the cross-lingual representations with the LEM strategy benefits the parallel data curation task as well.
Sinhala - Tamil | English - Sinhala | English-Tamil | |
NLLB | |||
XLM-R | 38.6 | 33.1 | 44.00 |
XLM-RMLM+TLM | 41.3 | 43.2 | 50.70 |
XLM-RLEM | (+3.5/+0.8) 42.1 | (+10.8/+0.7) 43.9 | (+7.2/+0.5) 51.2 |
CCAligned | |||
XLM-R | 37.2 | 10.2 | 5.2 |
XLM-RMLM+TLM | 42.3 | 33.9 | 31.5 |
XLM-RLEM | (+5.2/+0.3) 42.6 | (+24.3/+0.6) 34.5 | (+29.1/+2.8) 34.3 |
6.5 Code-Mixed Sentiment Analysis
Table 9 summarizes the evaluation scores for the code-mixed sentiment analysis task on the English-Sinhala dataset. The XLM-RLEM model consistently outperformed both XLM-R and XLM-RMLM+TLM when fine-tuned directly on the code-mixed dataset.
In the zero-shot evaluation (fine-tuned on English-only data), all models showed reduced F1 scores. We suspect that the English monolingual fine-tuning step adversely impacted the cross-lingual alignment between the two languages.
When further fine-tuned on the code-mixed dataset, XLM-RLEM achieved the highest scores, indicating that the cross-lingual enhancement benefits the code-mixed sentiment classification task. These results highlight the promise of LEM in improving multilingual models, particularly for low-resource and code-mixed language applications.
Precision | Recall | F1 | |
Fine-tuning on English-Sinhala Code-mixed Sentiment Analysis dataset | |||
XLM-R | 88.72% | 88.72% | 88.72% |
XLM-RMLM+TLM | 88.88% | 88.88% | 88.88% |
XLM-RLEM | 89.09% | 89.33% | 89.20% |
Fine-tuning on English Sentiment Analysis dataset (Zero-shot scores) | |||
XLM-R | 74.19% | 82.14% | 75.17% |
XLM-RMLM+TLM | 74.74% | 79.57% | 73.92% |
XLM-RLEM | 69.18% | 76.36% | 68.13% |
Fine-tuning on English-Sinhala Code-mixed Sentiment Analysis dataset | |||
XLM-R | 88.72% | 91.00% | 89.77% |
XLM-RMLM+TLM | 89.19% | 90.83% | 89.97% |
XLM-RLEM | 90.63% | 90.63% | 90.63% |
7 Ablation Studies
7.1 The Number of Tokens for Masking in LEM Strategy
To evaluate the impact of the masked token count within linguistic entities, we conducted an ablation study by varying the number of masked tokens. We conducted this for the Sinhala-Tamil language pair. As reported in Table 16, the best result was returned for the 100% NE+15% MLM and 100% NE+15% TLM combinations in the and steps respectively. Therefore, we used this setting and the number of tokens for masking was varied. Results on the bitext mining task are reported in Table 10. It reveals a clear trend of decreasing performance.
When masked only one token per linguistic entity, the average performance across tasks was the highest. This outcome suggested that minimal masking preserved more contextual information, allowing the model to better capture dependencies critical for downstream tasks. As the masked token count increased to two or more, the average performance dropped. This drop was significant when increasing the token count to 3 and 4. This indicated that excessive masking had disrupted the contextual integrity of linguistic entities, which lead to suboptimal representations.
Interestingly, the performance drop became less pronounced when the number of masked tokens increased from 3 tokens to 4 tokens (a decrease of only 0.02). This suggested a potential saturation point where further masking within an entity had diminishing negative effects, as the model might already struggle to leverage the remaining context effectively.
No. of Tokens Masked in Linguistic Entity | FW (Recall) | BW (Recall) | IN (Recall) | Average (Recall) |
---|---|---|---|---|
1 | 92.13 | 93.18 | 89.10 | 91.47 |
2 | 91.02 | 92.52 | 87.78 | 90.44 |
3 | 83.79 | 87.37 | 79.05 | 83.40 |
4 | 84.12 | 87.03 | 78.97 | 83.38 |
7.2 Effect of noise in LEM Strategy
We investigated the impact of applying the LEM strategy to noisy data. This analysis provides critical insights into the robustness and adaptability of LEM when faced with real-world noisy data. We specifically focus on the English-Sinhala language pair and use the ParaCrawl171717https://opus.nlpl.eu/ dataset.
As per Table 14, the current ablation revealed that the best results for English-Sinhala were achieved with the combination of 100%VB+15%MLM and 100%VB+15%TLM during the and steps, respectively. We ran the LEM experiments with ParaCrawl data for the same combinations.
Table 11 presents the final scores. We observed that the results were comparable to those derived from the cleaner SiTa-Trilingual dataset. Further, in BW criteria the scores slightly surpass and in FW criteria the scores are the same as that obtained when using high-quality data. This equivalence underscores the resilience of the LEM strategy to noise in the training data.
Dataset | Quality of the Dataset | FW (Recall) | BW (Recall) | IN (Recall) | Overall Average (Recall) |
---|---|---|---|---|---|
SiTa-Trilingual | High Quality | 95.09 | 94.67 | 93.42 | 94.39 |
ParaCrawl | Noisy | 95.09 | 95.00 | 92.49 | 94.19 |
The ability of LEM to maintain high performance in noisy settings highlights its practical applicability in low-resource scenarios, where parallel data is often noisy or inconsistent. This robustness not only complements our findings but also demonstrates that LEM can effectively mitigate the challenges associated with data quality, a common issue in low-resource language processing.
8 Discussion
The LEM strategy is very much driven by the accuracy of the underlying tools to identify the linguistic entities. The sub-optimal performance of the NER model and POS Taggers can affect the final results.
Although the NER model performs well with English sentences, we observe two main error types with it for Sinhala and Tamil language text, as shown in Table 17 in Appendix D. As False Positives, we observe words which were not a part of the NE tagged as NEs. Secondly, in the category of False Negatives, the NER model fails to identify all the words belonging to the NE sequence, and incorrectly label these words as Other etc. Similar instances are found in PoS Tagging as well as per examples in Table 18 in Appendix D that resulted in False Positives and False Negatives. However for English language the returned PoS tags were mostly accurate.
As future work, we will continually pre-train a single multilingual model for multiple languages using the LEM strategy. This is in contrast to the current approach which yields specialized encoders for each language pair. Additionally, as a single multilingual model, its performance on downstream tasks can be analysed. Secondly, with the advancements in the field, we hope more sophisticated NER models and PoS Tagger tools might be introduced in future. We will re-evaluate the LEM performance upon the availability of such tools. Thirdly, we will investigate the impact of the LEM strategy on language families, examining its effectiveness across a broader linguistic spectrum.
9 Conclusion
Multilingual PLMs trained with masking strategies are less effective for downstream tasks. This research introduced LEM strategy to improve cross-lingual representations of existing multiPLMs. Here the objective is to mask a single token, specifically targeting linguistic entities (NEs, nouns and verbs). Extending this LEM strategy with parallel data yields even better results, as evidenced in low-resource language pairs such as English-Sinhala, English-Tamil, and Sinhala-Tamil. The improved cross-lingual representations showed superior performance on the three evaluation tasks.
Funding Provided upon paper acceptance.
Availability of data and materials We have only used publicly available data in this research. Where applicable, the citations and/or URLs were provided under the relevant sections.
Source Code Will be released upon acceptance.
Appendix A Type of monolingual data parallel data for MLM
Dataset | Dataset Size | Army | Hiru | ITN | Newsfirst | Averages | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
F | B | I | F | B | I | F | B | I | F | B | I | F | B | I | ||
English - Sinhala | ||||||||||||||||
SiTa | 59333 | 88.33 | 91.00 | 85.33 | 92.03 | 93.36 | 89.70 | 91.67 | 92.67 | 88.67 | 91.67 | 95.33 | 90.00 | 90.92 | 93.09 | 88.42 |
MADLAD400 | 60000 | 82.67 | 88.33 | 78.00 | 85.05 | 91.36 | 82.00 | 85.33 | 86.00 | 79.67 | 91.67 | 92.67 | 87.67 | 86.18 | 89.59 | 81.83 |
MADLAD400 | 100000 | 86.67 | 91.67 | 83.33 | 91.69 | 96.01 | 91.03 | 88.00 | 90.33 | 83.00 | 91.33 | 95.00 | 89.00 | 89.42 | 93.25 | 86.59 |
English - Tamil | ||||||||||||||||
SiTa | 59333 | 84.00 | 86.00 | 77.67 | 80.33 | 75.00 | 68.33 | 81.56 | 82.21 | 78.52 | 90.67 | 91.00 | 87.00 | 84.14 | 83.55 | 77.88 |
MADLAD400 | 60000 | 81.67 | 78.67 | 69.33 | 75.33 | 69.67 | 60.67 | 81.18 | 77.15 | 69.77 | 90.00 | 86.67 | 81.67 | 82.05 | 78.04 | 70.36 |
MADLAD400 | 100000 | 81.33 | 79.67 | 71.67 | 77.67 | 71.33 | 62.67 | 78.86 | 76.17 | 68.79 | 88.67 | 88.00 | 82.00 | 81.63 | 78.79 | 71.28 |
Sinhala - Tamil | ||||||||||||||||
SiTa | 59333 | 86.75 | 88.08 | 81.46 | 88.00 | 89.33 | 84.00 | 93.33 | 92.67 | 89.33 | 90.33 | 94.00 | 89.00 | 89.60 | 91.02 | 85.95 |
MADLAD400 | 60000 | 84.77 | 89.73 | 80.46 | 86.00 | 89.00 | 83.00 | 92.67 | 92.00 | 89.00 | 89.00 | 92.67 | 85.67 | 88.11 | 90.85 | 84.53 |
MADLAD400 | 100000 | 84.11 | 88.08 | 78.81 | 86.00 | 89.33 | 81.33 | 90.67 | 93.67 | 87.33 | 88.67 | 92.33 | 85.00 | 87.36 | 90.85 | 83.12 |
MADLAD400 | 500000 | 82.12 | 83.11 | 75.17 | 85.67 | 88.33 | 79.67 | 87.67 | 91.00 | 83.67 | 87.67 | 90.67 | 82.33 | 85.78 | 88.28 | 80.21 |
Appendix B Parallel Data Curation Task NMT extrinsic Evaluation Results
The NMT evaluation scores for the parallel data curation task are reported in Table 13 for the Flores+ benchmark devtest evaluation set. While the discussion on the NMT results has been based using the ChrF++ metric, here we present the scores for the same experiments using NMT evaluation metrics sacreBLEU, multi-bleu, ChrF, ChrF++ and spBLEU.
sacreBLEU | multi-bleu | ChrF | ChrF++ | SpBLEU | |
NLLB | |||||
Sinhala - Tamil | |||||
XLM-R | 2.6 | 2.58 | 38.6 | 33.58 | 11.9 |
XLM-RMLM+TLM | 3.2 | 3.23 | 41.3 | 35.99 | 14.5 |
XLM-RLEM | 3.6 | 3.60 | 42.1 | 36.68 | 15.2 |
English - Tamil | |||||
XLM-R | 6.2 | 6.18 | 44.00 | 38.28 | 18.40 |
XLM-RMLM+TLM | 9.2 | 9.16 | 50.70 | 45.35 | 25.20 |
XLM-RLEM | 9.3 | 9.47 | 51.20 | 45.86 | 25.80 |
English - Sinhala | |||||
XLM-R | 4.9 | 4.91 | 33.1 | 30.37 | 13.6 |
XLM-RMLM+TLM | 9.4 | 9.42 | 43.2 | 39.78 | 23.3 |
XLM-RLEM | 9.9 | 9.85 | 43.9 | 40.31 | 23.8 |
CCAligned | |||||
Sinhala - Tamil | |||||
XLM-R | 2.2 | 2.23 | 37.2 | 32.43 | 10.6 |
XLM-RMLM+TLM | 3.7 | 3.74 | 42.3 | 36.02 | 15.2 |
XLM-RLEM | 3.6 | 3.61 | 42.6 | 36.90 | 14.9 |
English - Tamil | |||||
XLM-R | 0.2 | 0.17 | 5.2 | 5.80 | 1.2 |
XLM-RMLM+TLM | 3.2 | 3.24 | 31.5 | 28.55 | 11.5 |
XLM-RLEM | 3.5 | 3.48 | 34.3 | 30.96 | 12.5 |
English - Sinhala | |||||
XLM-R | 0.4 | 0.37 | 10.2 | 10.13 | 2.3 |
XLM-RMLM+TLM | 5.0 | 5.00 | 33.9 | 31.17 | 14.8 |
XLM-RLEM | 5.1 | 5.09 | 34.5 | 31.71 | 15.3 |
Appendix C Linguistic Entity Masking Ablation Study
In this section, we present the full experiments along with the scores obtained during the ablation study of our LEM masking strategy. The Tables 14, 15 and 16 contain the bitext mining recall scores for English-Sinhala, English-Tamil and Sinhala-Tamil language pairs respectively.
Experiment | Army | Hiru | ITN | Newsfirst | Averages | ||||||||||
F | B | I | F | B | I | F | B | I | F | B | I | F | B | I | |
Baselines | |||||||||||||||
XLM-R | 92.33 | 93.33 | 89.67 | 96.35 | 96.68 | 95.68 | 94.00 | 96.00 | 92.33 | 96.67 | 95.33 | 94.33 | 94.84 | 95.34 | 93.00 |
15%MLM | 88.33 | 91.00 | 85.33 | 92.03 | 93.36 | 89.70 | 91.67 | 92.67 | 88.67 | 91.67 | 95.33 | 90.00 | 90.92 | 93.09 | 88.42 |
15% TLM on 15% MLM | 91.33 | 92.67 | 88.67 | 94.35 | 95.68 | 93.36 | 94.00 | 94.00 | 90.67 | 94.67 | 95.00 | 92.67 | 93.59 | 94.34 | 91.34 |
LEMmono | |||||||||||||||
100%NE+15% MLM | 89.67 | 93.00 | 88.33 | 93.02 | 94.02 | 92.03 | 89.67 | 93.00 | 87.00 | 93.67 | 94.67 | 91.67 | 91.51 | 93.67 | 89.76 |
100% VB+15% MLM | 89.67 | 93.33 | 87.33 | 94.02 | 95.02 | 92.69 | 92.00 | 93.67 | 89.67 | 93.00 | 95.33 | 92.33 | 92.17 | 94.34 | 90.51 |
100% NN+15% MLM | 81.33 | 88.33 | 76.33 | 93.36 | 95.02 | 92.36 | 90.33 | 91.67 | 86.00 | 91.00 | 92.33 | 87.67 | 89.00 | 91.84 | 85.59 |
100% NE+ 100%VB+15% MLM | 91.33 | 91.00 | 87.67 | 95.35 | 94.02 | 93.36 | 92.33 | 94.00 | 89.33 | 93.33 | 94.33 | 90.67 | 93.09 | 93.34 | 90.26 |
100% NE+ 100%NN+15% MLM | 88.00 | 91.00 | 84.00 | 94.02 | 95.35 | 92.69 | 89.33 | 95.67 | 89.00 | 94.00 | 95.67 | 91.67 | 91.34 | 94.42 | 89.34 |
100% NE+ 100%VB+ 100%NN+15% MLM | 89.67 | 92.33 | 87.00 | 94.02 | 94.02 | 91.69 | 92.33 | 95.00 | 91.00 | 94.00 | 92.33 | 90.33 | 92.50 | 93.42 | 90.01 |
MLMmono+TLMpara | |||||||||||||||
100% NE+15% TLM on 15% MLM | 90.00 | 91.67 | 87.33 | 95.02 | 95.35 | 93.36 | 94.00 | 96.67 | 92.67 | 96.67 | 96.67 | 93.33 | 93.92 | 95.09 | 91.67 |
100% VB+15% TLM on 15% MLM | 91.67 | 90.33 | 86.67 | 94.35 | 95.02 | 92.69 | 93.00 | 95.33 | 89.67 | 95.00 | 94.67 | 91.67 | 93.50 | 93.84 | 90.17 |
100% NN+15% TLM on 15% MLM | 89.00 | 92.00 | 85.00 | 93.36 | 95.02 | 91.36 | 94.33 | 96.00 | 92.33 | 94.67 | 95.00 | 92.00 | 92.84 | 94.50 | 90.17 |
100% NE+ 100%VB+15% TLM on 15% MLM | 91.33 | 91.33 | 87.67 | 95.35 | 94.68 | 92.69 | 94.00 | 95.00 | 91.33 | 97.33 | 95.00 | 93.67 | 94.50 | 94.00 | 91.34 |
100% NE+ 100%NN+15% TLM on 15% MLM | 88.67 | 91.00 | 85.00 | 94.35 | 95.35 | 93.02 | 94.00 | 96.00 | 92.00 | 93.67 | 95.00 | 91.33 | 92.67 | 94.34 | 90.34 |
100%NE+100%VB+100%NN+15%TLM on 15% MLM | 90.67 | 91.33 | 87.33 | 94.68 | 97.34 | 94.35 | 93.67 | 95.00 | 91.00 | 94.33 | 96.33 | 92.33 | 93.34 | 95.00 | 91.25 |
15% TLM on (100%NE+15% MLM) | 89.00 | 93.00 | 87.00 | 94.35 | 95.35 | 93.64 | 92.00 | 95.67 | 90.00 | 95.00 | 90.00 | 93.33 | 92.59 | 93.50 | 90.99 |
100% NE+15% TLM on (100%NE+15% MLM) | 91.67 | 95.33 | 89.33 | 94.68 | 96.01 | 94.35 | 92.00 | 96.33 | 92.67 | 94.67 | 95.67 | 92.67 | 93.25 | 95.84 | 92.25 |
100% VB+15% TLM on (100%NE+15% MLM) | 90.00 | 91.67 | 86.00 | 94.02 | 95.02 | 93.36 | 92.67 | 94.67 | 90.00 | 93.33 | 95.00 | 91.67 | 92.50 | 94.09 | 90.26 |
100% NN+15% TLM on (100%NE+15% MLM) | 89.00 | 92.00 | 87.00 | 94.02 | 94.02 | 92.36 | 93.00 | 93.33 | 89.00 | 94.00 | 94.00 | 91.00 | 92.50 | 93.34 | 89.84 |
100% NE+ 100%VB+15% TLM on (100%NE+15% MLM) | 89.67 | 93.33 | 88.00 | 95.02 | 94.68 | 93.36 | 92.00 | 95.33 | 90.00 | 95.67 | 95.33 | 93.33 | 93.09 | 94.67 | 91.17 |
100% NE+ 100%NN+15% TLM on (100%NE+15% MLM) | 89.33 | 93.00 | 87.00 | 94.35 | 94.68 | 93.02 | 93.67 | 94.67 | 90.67 | 95.67 | 96.67 | 94.00 | 93.25 | 94.75 | 91.17 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+15% MLM) | 91.67 | 92.33 | 88.33 | 95.68 | 95.68 | 95.02 | 92.33 | 93.33 | 88.67 | 93.67 | 95.00 | 91.33 | 93.34 | 94.09 | 90.84 |
15% TLM on (100%VB+15% MLM) | 91.67 | 92.00 | 89.00 | 94.35 | 96.01 | 94.02 | 94.33 | 95.00 | 91.67 | 95.67 | 96.00 | 93.33 | 94.00 | 94.75 | 92.00 |
100% NE+15% TLM on (100%VB+15% MLM) | 90.33 | 91.67 | 87.67 | 95.02 | 96.35 | 94.35 | 93.67 | 94.33 | 90.00 | 96.67 | 95.67 | 93.67 | 93.92 | 94.50 | 91.42 |
100% VB+15% TLM on (100%VB+15% MLM) | 91.67 | 93.33 | 90.67 | 96.68 | 95.35 | 95.35 | 95.33 | 94.33 | 93.67 | 96.67 | 95.67 | 94.00 | 95.09 | 94.67 | 93.42 |
100% NN+15% TLM on (100%VB+15% MLM) | 88.33 | 91.67 | 86.00 | 95.02 | 95.02 | 93.36 | 93.00 | 93.67 | 89.67 | 92.67 | 94.67 | 91.33 | 92.25 | 93.75 | 90.09 |
100% NE+ 100%VB+15% TLM on (100%VB+15% MLM) | 90.00 | 94.33 | 88.33 | 94.35 | 95.68 | 93.02 | 94.00 | 95.33 | 91.00 | 96.67 | 95.33 | 94.33 | 93.75 | 95.17 | 91.67 |
100% NE+ 100%NN+15% TLM on (100%VB+15% MLM) | 89.67 | 91.33 | 86.33 | 94.68 | 95.68 | 93.69 | 93.67 | 94.33 | 91.33 | 95.67 | 96.33 | 93.67 | 93.42 | 94.42 | 91.26 |
100%NE+100%VB+100%NN+15%TLM on (100%VB+15% MLM) | 92.00 | 92.33 | 87.33 | 95.35 | 95.68 | 94.02 | 93.00 | 94.00 | 89.67 | 95.67 | 95.00 | 93.00 | 94.00 | 94.25 | 91.00 |
15% TLM on (100%NN+15%MLM) | 90.33 | 93.33 | 87.33 | 94.35 | 94.68 | 93.02 | 94.67 | 95.00 | 92.00 | 94.67 | 95.00 | 92.67 | 93.50 | 94.50 | 91.26 |
100% NE+15% TLM on (100%NN+15%MLM) | 89.00 | 93.67 | 87.00 | 94.35 | 95.35 | 92.36 | 95.00 | 95.33 | 91.33 | 96.00 | 95.33 | 92.67 | 93.59 | 94.92 | 90.84 |
100% VB+15% TLM on (100%NN+15%MLM) | 88.00 | 93.33 | 86.67 | 93.69 | 95.68 | 93.02 | 94.33 | 95.67 | 94.67 | 94.67 | 94.00 | 91.67 | 92.67 | 94.67 | 91.51 |
100% NN+15% TLM on (100%NN+15%MLM) | 91.00 | 92.00 | 87.67 | 95.68 | 95.02 | 94.02 | 94.33 | 96.33 | 92.67 | 95.00 | 95.67 | 92.67 | 94.00 | 94.75 | 91.76 |
100% NE+ 100%VB+15% TLM on (100%NN+15%MLM) | 90.67 | 93.67 | 87.67 | 95.02 | 94.68 | 93.02 | 95.00 | 95.67 | 92.33 | 94.33 | 94.00 | 91.67 | 93.75 | 94.50 | 91.17 |
100% NE+ 100%NN+15% TLM on (100%NN+15%MLM) | 91.67 | 91.33 | 87.67 | 94.68 | 95.68 | 94.02 | 93.00 | 95.33 | 90.67 | 94.33 | 95.00 | 92.33 | 93.42 | 94.34 | 91.17 |
100%NE+100%VB+100%NN+15%TLM on (100%NN+15%MLM) | 88.67 | 92.00 | 86.00 | 96.01 | 96.01 | 95.02 | 94.00 | 95.33 | 91.33 | 94.33 | 94.67 | 91.33 | 93.25 | 94.50 | 90.92 |
15% TLM on (100%NE+100%VB+15%MLM) | 88.67 | 93.00 | 86.67 | 94.35 | 95.02 | 93.02 | 92.33 | 93.67 | 88.67 | 93.00 | 94.33 | 90.67 | 92.09 | 94.00 | 89.76 |
100% NE+15% TLM on (100%NE+100%VB+15%MLM) | 89.67 | 91.33 | 87.00 | 94.68 | 95.68 | 93.69 | 93.67 | 94.33 | 93.67 | 95.67 | 94.33 | 91.33 | 93.42 | 93.92 | 91.42 |
100% VB+15% TLM on (100%NE+100%VB+15%MLM) | 88.00 | 93.33 | 86.67 | 93.69 | 95.68 | 93.02 | 94.33 | 94.33 | 94.67 | 94.67 | 94.00 | 91.67 | 92.67 | 94.34 | 91.51 |
100% NN+15% TLM on (100%NE+100%VB+15%MLM) | 88.67 | 93.00 | 86.67 | 93.67 | 95.02 | 93.02 | 94.00 | 93.67 | 90.00 | 93.00 | 94.33 | 90.67 | 92.33 | 94.00 | 90.09 |
100% NE+ 100%VB+15% TLM on (100%NE+100%VB+15%MLM) | 91.33 | 92.67 | 89.00 | 94.68 | 95.68 | 93.69 | 94.00 | 94.67 | 91.67 | 95.33 | 95.33 | 93.00 | 93.84 | 94.59 | 91.84 |
100% NE+ 100%NN+15% TLM on (100%NE+100%VB+15%MLM) | 91.00 | 91.33 | 87.67 | 94.02 | 94.68 | 92.36 | 94.67 | 94.67 | 91.67 | 95.67 | 95.33 | 93.00 | 93.84 | 94.00 | 91.17 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+15%MLM) | 92.00 | 93.00 | 89.00 | 95.35 | 96.01 | 94.68 | 94.33 | 94.00 | 91.00 | 96.00 | 96.00 | 94.33 | 94.42 | 94.75 | 92.25 |
15% TLM on (100%NE+100%NN+15%MLM) | 91.33 | 94.00 | 88.67 | 94.02 | 95.02 | 92.03 | 95.33 | 95.67 | 93.00 | 94.33 | 97.67 | 94.00 | 93.75 | 95.59 | 91.92 |
100% NE+15% TLM on (100%NE+100%NN+15%MLM) | 87.67 | 90.33 | 93.67 | 94.02 | 95.35 | 93.02 | 96.00 | 94.67 | 92.67 | 93.67 | 94.67 | 91.67 | 92.84 | 93.75 | 92.76 |
100% VB+15% TLM on (100%NE+100%NN+15%MLM) | 91.00 | 92.00 | 87.00 | 94.35 | 94.35 | 92.69 | 93.67 | 96.33 | 93.00 | 95.67 | 95.33 | 93.00 | 93.67 | 94.50 | 91.42 |
100% NN+15% TLM on (100%NE+100%NN+15%MLM) | 88.33 | 91.67 | 84.33 | 95.02 | 95.35 | 93.64 | 94.67 | 94.67 | 91.67 | 94.33 | 95.33 | 92.33 | 93.09 | 94.25 | 90.49 |
100% NE+100%VB+15% TLM on (100%NE+100%NN+15%MLM) | 90.00 | 93.33 | 86.67 | 94.68 | 94.68 | 93.36 | 93.33 | 93.00 | 89.33 | 96.33 | 96.00 | 94.33 | 93.59 | 94.25 | 90.92 |
100% NE+100%NN+15% TLM on (100%NE+100%NN+15%MLM) | 87.67 | 90.33 | 84.00 | 95.68 | 96.01 | 94.35 | 92.00 | 95.00 | 90.33 | 94.67 | 96.33 | 93.00 | 92.50 | 94.42 | 90.42 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%NN+15%MLM) | 88.67 | 91.67 | 85.00 | 95.35 | 95.68 | 94.35 | 94.00 | 93.67 | 90.33 | 95.67 | 95.33 | 93.00 | 93.42 | 94.09 | 90.67 |
15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 93.00 | 91.67 | 88.00 | 95.35 | 96.01 | 94.02 | 94.67 | 95.00 | 92.00 | 93.67 | 94.67 | 91.67 | 94.17 | 94.34 | 91.42 |
100% NE+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 89.00 | 91.00 | 84.67 | 95.35 | 96.35 | 94.68 | 96.00 | 95.33 | 93.00 | 96.00 | 95.67 | 93.33 | 94.09 | 94.59 | 91.42 |
100% VB+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 89.67 | 92.67 | 87.00 | 95.35 | 95.35 | 93.67 | 95.00 | 93.33 | 91.67 | 95.67 | 93.33 | 91.00 | 93.92 | 93.67 | 90.83 |
100% NN+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 89.67 | 91.33 | 86.00 | 95.02 | 95.35 | 93.33 | 93.67 | 94.67 | 91.33 | 93.67 | 94.00 | 90.33 | 93.00 | 93.84 | 90.25 |
100% NE+ 100%VB+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 88.67 | 92.00 | 85.33 | 93.69 | 95.68 | 92.69 | 93.00 | 95.67 | 90.33 | 95.00 | 95.33 | 92.33 | 92.59 | 94.67 | 90.17 |
100% NE+ 100%NN+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 86.67 | 91.67 | 84.67 | 96.35 | 96.01 | 94.68 | 94.33 | 95.67 | 92.33 | 94.33 | 94.67 | 91.33 | 92.92 | 94.50 | 90.75 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+100%NN+15%MLM) | 91.33 | 92.00 | 88.00 | 95.68 | 95.68 | 94.02 | 93.67 | 94.00 | 91.33 | 93.67 | 94.00 | 90.33 | 93.59 | 93.92 | 90.92 |
Experiment | Army | Hiru | ITN | Newsfirst | Average | ||||||||||
FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | |
Baselines | |||||||||||||||
XLM-R | 86.67 | 88.33 | 82.00 | 83.00 | 78.33 | 72.67 | 83.22 | 83.56 | 78.86 | 92.33 | 91.33 | 89.33 | 86.31 | 85.39 | 80.71 |
15%MLM | 84.00 | 86.00 | 77.67 | 80.33 | 75.00 | 68.33 | 81.56 | 82.21 | 78.52 | 90.67 | 91.00 | 87.00 | 84.14 | 83.55 | 77.88 |
15%TLM on 15%MLM | 86.67 | 85.67 | 79.33 | 80.33 | 78.67 | 71.00 | 81.88 | 83.56 | 77.52 | 90.00 | 92.67 | 88.00 | 84.72 | 85.14 | 78.96 |
LEMmono | |||||||||||||||
100% NE+15% MLM | 86.00 | 86.67 | 81.00 | 79.33 | 75.33 | 66.67 | 81.21 | 81.21 | 74.83 | 93.00 | 92.00 | 90.00 | 84.89 | 83.80 | 78.12 |
100% VB+15% MLM | 85.67 | 84.67 | 76.67 | 78.67 | 76.00 | 68.00 | 81.88 | 82.55 | 75.84 | 91.00 | 90.00 | 86.33 | 84.30 | 83.30 | 76.71 |
100% NN+15% MLM | 83.33 | 84.67 | 77.00 | 73.67 | 72.67 | 61.67 | 75.84 | 82.22 | 70.13 | 90.00 | 91.00 | 87.00 | 80.71 | 82.64 | 73.95 |
100% NE+100%VB+15% MLM | 83.00 | 86.67 | 77.67 | 77.67 | 74.33 | 65.00 | 81.21 | 83.56 | 75.84 | 89.00 | 88.67 | 84.00 | 82.72 | 83.31 | 75.63 |
100% NE+100%NN+15% MLM | 82.67 | 85.33 | 75.00 | 75.33 | 72.67 | 62.00 | 80.54 | 84.23 | 74.48 | 90.33 | 90.00 | 86.33 | 82.22 | 83.06 | 74.45 |
100% NE+100% VB +100%NN+15% MLM | 83.00 | 83.33 | 78.00 | 74.67 | 73.67 | 64.67 | 80.87 | 83.89 | 76.17 | 91.33 | 92.67 | 88.67 | 82.47 | 83.39 | 76.88 |
LEMmono+LEMpara | |||||||||||||||
100% NE+15% TLM on 15%MLM | 83.00 | 85.33 | 76.33 | 79.67 | 78.33 | 70.00 | 83.89 | 85.91 | 79.87 | 91.00 | 93.33 | 89.00 | 84.39 | 85.73 | 78.80 |
100% VB+15% TLM on 15%MLM | 87.00 | 86.67 | 81.67 | 80.67 | 79.00 | 72.33 | 83.89 | 85.57 | 79.87 | 91.67 | 92.33 | 88.67 | 85.81 | 85.89 | 80.63 |
100% NN+15% TLM on 15%MLM | 85.00 | 86.67 | 79.67 | 79.33 | 77.00 | 69.00 | 83.89 | 86.24 | 80.54 | 91.33 | 94.00 | 89.67 | 84.89 | 85.98 | 79.72 |
100% NE+100% VB+15% TLM on 15%MLM | 85.67 | 76.67 | 69.33 | 79.67 | 76.67 | 69.33 | 83.22 | 84.23 | 77.85 | 92.00 | 92.00 | 89.33 | 85.14 | 82.39 | 76.46 |
100% NE+100% NN+15% TLM on 15%MLM | 84.67 | 85.00 | 77.67 | 81.00 | 80.00 | 72.33 | 81.98 | 84.23 | 77.85 | 90.00 | 92.00 | 87.67 | 84.41 | 85.31 | 78.88 |
100% NE+100% VB+ 100% NN+ 15% TLM on 15%MLM | 85.00 | 85.33 | 80.00 | 78.67 | 78.33 | 70.00 | 84.23 | 88.59 | 80.87 | 90.00 | 93.67 | 88.33 | 84.47 | 86.48 | 79.80 |
15% TLM on 100%NE+15%MLM | 87.00 | 86.33 | 81.33 | 81.33 | 80.00 | 71.67 | 81.21 | 84.23 | 77.52 | 92.67 | 91.33 | 89.00 | 85.55 | 85.47 | 79.88 |
100%NE+15% TLM on 100%NE+15%MLM | 87.67 | 87.00 | 81.67 | 82.00 | 81.33 | 73.00 | 81.88 | 84.23 | 77.18 | 91.33 | 92.67 | 88.33 | 85.72 | 86.31 | 80.05 |
100% VB+15% TLM on 100%NE+15%MLM | 88.33 | 89.33 | 83.67 | 80.00 | 77.67 | 69.67 | 81.54 | 84.23 | 75.50 | 91.00 | 93.00 | 89.00 | 85.22 | 86.06 | 79.46 |
100% NN+15% TLM on 100%NE+15%MLM | 86.33 | 87.67 | 80.33 | 81.33 | 79.67 | 70.67 | 80.54 | 84.56 | 76.51 | 92.00 | 91.33 | 88.00 | 85.05 | 85.81 | 78.88 |
100% NE+100% VB+15% TLM on 100%NE+15%MLM/ | 84.67 | 85.67 | 78.00 | 82.33 | 76.67 | 70.33 | 80.54 | 83.22 | 76.85 | 89.33 | 92.67 | 87.67 | 84.22 | 84.56 | 78.21 |
100% NE+100% NN+15% TLM on 100%NE+15%MLM/ | 84.67 | 85.67 | 78.00 | 82.33 | 76.67 | 70.33 | 80.54 | 83.22 | 76.85 | 89.33 | 92.67 | 87.67 | 84.22 | 84.56 | 78.21 |
100%NE+100%VB+100%NN+15%TLM on 100%NE+15%MLM/ | 85.00 | 84.33 | 78.33 | 78.00 | 76.67 | 67.33 | 79.53 | 83.89 | 75.84 | 92.33 | 92.00 | 90.00 | 83.72 | 84.22 | 77.88 |
15% TLM on (100%VB+15%MLM) | 88.00 | 88.67 | 83.67 | 82.00 | 79.00 | 72.33 | 84.90 | 84.90 | 80.54 | 93.33 | 93.00 | 91.00 | 87.06 | 86.39 | 81.88 |
100% NE+ 15% TLM on (100%VB+15%MLM) | 84.00 | 87.67 | 79.00 | 78.67 | 81.33 | 71.00 | 82.22 | 85.57 | 78.86 | 90.67 | 93.33 | 88.33 | 83.89 | 86.98 | 79.30 |
100% VB+ 15% TLM on (100%VB+15%MLM) | 86.00 | 88.67 | 80.67 | 81.33 | 78.33 | 70.67 | 82.22 | 84.56 | 76.85 | 91.33 | 92.33 | 87.67 | 85.22 | 85.97 | 78.96 |
100% NN+15% TLM on (100%VB+15%MLM) | 86.33 | 85.33 | 80.33 | 79.67 | 79.00 | 70.33 | 82.22 | 84.90 | 77.52 | 90.33 | 93.67 | 88.00 | 84.64 | 85.72 | 79.05 |
100% NE+ 100% VB+ 15% TLM on (100%VB+15%MLM) | 85.67 | 88.00 | 80.33 | 80.33 | 76.00 | 69.00 | 81.54 | 83.58 | 77.18 | 90.67 | 93.00 | 88.00 | 84.55 | 85.14 | 78.63 |
100% NE+ 100% NN+ 15% TLM on (100%VB+15%MLM) | 87.33 | 87.67 | 81.67 | 78.00 | 78.00 | 68.67 | 81.54 | 83.89 | 76.85 | 91.00 | 92.00 | 87.67 | 84.47 | 85.39 | 78.71 |
100% NE+ 100% NN+ 100%VB+ 15% TLM on (100%VB+15%MLM) | 86.33 | 87.00 | 80.67 | 78.33 | 76.67 | 67.33 | 82.89 | 84.56 | 77.85 | 90.67 | 92.33 | 87.33 | 84.55 | 85.14 | 78.30 |
15% TLM on (100%NN+15%MLM) | 84.67 | 88.33 | 81.00 | 81.00 | 77.33 | 69.67 | 83.22 | 85.91 | 78.86 | 91.67 | 92.33 | 89.67 | 85.14 | 85.98 | 79.80 |
100% NE+ 15% TLM on (100%NN+15%MLM) | 85.33 | 86.67 | 79.00 | 78.33 | 76.33 | 67.33 | 81.88 | 84.56 | 76.85 | 91.00 | 91.33 | 87.67 | 84.14 | 84.72 | 77.71 |
100% VB+15% TLM on (100%NN+15%MLM) | 84.67 | 87.67 | 80.67 | 78.67 | 76.00 | 67.33 | 79.19 | 84.29 | 75.50 | 90.00 | 92.67 | 88.00 | 83.13 | 85.16 | 77.88 |
100% NN+ 15% TLM on (100%NN+15%MLM) | 85.00 | 87.00 | 79.33 | 77.33 | 76.33 | 66.00 | 80.87 | 83.56 | 74.83 | 89.67 | 92.67 | 87.00 | 83.22 | 84.89 | 76.79 |
100% NE+ 100% VB+ 15% TLM on (100%NN+15%MLM) | 82.33 | 86.00 | 77.67 | 78.33 | 74.33 | 65.00 | 81.21 | 85.91 | 76.51 | 62.67 | 65.00 | 61.00 | 76.14 | 77.81 | 70.04 |
100% NE+ 100% NN+ 15% TLM on (100%NN+15%MLM) | 86.00 | 87.00 | 80.00 | 78.00 | 76.67 | 66.67 | 79.53 | 84.23 | 74.16 | 88.67 | 92.33 | 86.67 | 83.05 | 85.06 | 76.87 |
100% NE+ 100% NN+ 100%VB+ 15% TLM on (100%NN+15%MLM) | 86.33 | 90.00 | 82.00 | 76.00 | 78.33 | 66.33 | 79.87 | 85.91 | 76.16 | 90.33 | 92.00 | 87.67 | 83.13 | 86.56 | 78.04 |
15% TLM on (100%NE+100%VB+15%MLM) | 85.33 | 89.33 | 80.00 | 80.00 | 75.33 | 67.33 | 85.34 | 84.29 | 78.86 | 90.00 | 91.67 | 87.00 | 85.17 | 85.16 | 78.30 |
100% NE+ 15% TLM on (100%NE+100%VB+15%MLM) | 86.00 | 87.33 | 80.33 | 80.33 | 78.33 | 71.33 | 82.22 | 81.88 | 75.50 | 89.00 | 93.00 | 88.00 | 84.39 | 85.14 | 78.79 |
100% VB+15% TLM on (100%NE+100%VB+15%MLM) | 84.67 | 87.67 | 79.33 | 78.00 | 75.33 | 67.00 | 83.89 | 84.56 | 77.85 | 88.33 | 92.00 | 86.67 | 83.72 | 84.89 | 77.71 |
100% NN+ 15% TLM on (100%NE+100%VB+15%MLM) | 86.00 | 86.67 | 79.00 | 81.00 | 75.67 | 67.00 | 83.89 | 84.56 | 78.52 | 92.00 | 93.00 | 89.33 | 85.72 | 84.97 | 78.46 |
100% NE+100% VB+ 15% TLM on (100%NE+100%VB+15%MLM) | 84.67 | 87.67 | 78.00 | 78.33 | 75.67 | 68.00 | 90.87 | 84.90 | 76.17 | 89.00 | 92.00 | 86.67 | 85.72 | 85.06 | 77.21 |
100% NE+ 100% NN+ 15% TLM on (100%NE+100%VB+15%MLM) | 83.00 | 86.67 | 78.33 | 84.67 | 77.00 | 72.00 | 81.88 | 84.56 | 77.18 | 89.00 | 93.00 | 87.67 | 84.64 | 85.31 | 78.80 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+15%MLM) | 87.67 | 86.33 | 80.00 | 79.00 | 78.00 | 69.33 | 82.89 | 84.29 | 78.52 | 90.00 | 93.67 | 88.67 | 84.89 | 85.57 | 79.13 |
15% TLM on (100%NE+100%NN+15%MLM) | 86.00 | 89.33 | 80.00 | 80.00 | 76.33 | 69.67 | 85.34 | 85.24 | 78.86 | 90.00 | 92.67 | 88.00 | 85.33 | 85.89 | 79.13 |
100% NE+ 15% TLM on (100%NE+100%NN+15%MLM) | 86.00 | 86.67 | 80.00 | 78.33 | 76.67 | 65.00 | 82.55 | 85.57 | 78.52 | 90.67 | 93.00 | 88.33 | 84.39 | 85.48 | 77.96 |
100% VB+15% TLM on (100%NE+100%NN+15%MLM) | 84.67 | 87.67 | 79.33 | 80.67 | 78.67 | 70.00 | 81.54 | 85.23 | 78.86 | 89.33 | 93.00 | 87.00 | 84.05 | 86.14 | 78.80 |
100% NN+ 15% TLM on (100%NE+100%NN+15%MLM) | 85.67 | 85.00 | 78.00 | 80.00 | 76.33 | 68.67 | 81.21 | 84.56 | 77.18 | 92.33 | 93.00 | 89.33 | 84.80 | 84.72 | 78.29 |
100% NE+100% VB+ 15% TLM on (100%NE+100%NN+15%MLM) | 84.33 | 85.33 | 77.33 | 79.00 | 77.67 | 69.00 | 84.56 | 85.91 | 80.20 | 91.00 | 92.67 | 89.00 | 84.72 | 85.39 | 78.88 |
100% NE+100% NN+ 15% TLM on (100%NE+100%NN+15%MLM) | 86.67 | 83.67 | 78.67 | 76.33 | 78.00 | 66.00 | 82.55 | 85.57 | 78.19 | 91.33 | 91.67 | 87.67 | 84.22 | 84.73 | 77.63 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%NN+15%MLM) | 82.33 | 86.00 | 76.67 | 77.00 | 79.00 | 68.67 | 81.21 | 82.55 | 75.84 | 91.00 | 91.33 | 88.00 | 82.89 | 84.72 | 77.29 |
15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 84.00 | 86.00 | 79.00 | 83.00 | 77.33 | 71.67 | 82.55 | 85.23 | 77.85 | 90.33 | 94.33 | 88.67 | 84.97 | 85.73 | 79.30 |
100% NE+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 82.33 | 86.33 | 77.67 | 80.00 | 77.33 | 68.67 | 81.88 | 84.56 | 76.85 | 88.67 | 91.67 | 86.67 | 83.22 | 84.97 | 77.46 |
100% VB+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 84.67 | 88.00 | 79.67 | 79.00 | 77.00 | 68.00 | 83.89 | 84.90 | 78.52 | 91.00 | 94.00 | 90.00 | 84.64 | 85.97 | 79.05 |
100% NN+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 85.33 | 87.00 | 81.33 | 77.00 | 74.67 | 66.33 | 82.55 | 84.90 | 78.86 | 89.33 | 93.00 | 87.33 | 83.55 | 84.89 | 78.46 |
100% NE+100% VB+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 82.67 | 85.67 | 78.33 | 76.67 | 75.00 | 66.00 | 83.22 | 85.91 | 77.52 | 88.00 | 92.67 | 85.67 | 82.64 | 84.81 | 76.88 |
100% NE+ 100% NN+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 84.33 | 84.00 | 78.00 | 76.67 | 77.00 | 67.33 | 83.21 | 85.57 | 78.19 | 87.00 | 93.00 | 85.00 | 82.80 | 84.89 | 77.13 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+100%NN+15%MLM) | 85.33 | 85.33 | 79.67 | 76.33 | 76.00 | 67.00 | 82.22 | 84.90 | 77.12 | 89.00 | 94.00 | 88.33 | 83.22 | 85.06 | 78.03 |
Experiment | Army | Hiru | ITN | Newsfirst | Average | ||||||||||
FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | FW | BW | IN | |
Baselines | |||||||||||||||
XLM-R | 83.44 | 81.46 | 78.15 | 90.67 | 91.00 | 87.33 | 91.33 | 90.00 | 87.00 | 93.67 | 95.33 | 92.33 | 89.78 | 89.45 | 86.20 |
15%MLM | 86.75 | 88.08 | 81.46 | 88.00 | 89.33 | 84.00 | 93.33 | 92.67 | 89.33 | 90.33 | 94.00 | 89.00 | 89.60 | 91.02 | 85.95 |
15%TLM on 15%MLM | 87.75 | 90.40 | 83.11 | 88.67 | 93.33 | 86.33 | 93.00 | 94.33 | 90.00 | 91.33 | 94.33 | 89.67 | 90.19 | 93.10 | 87.28 |
LEMmono | |||||||||||||||
100% NE+15% MLM | 86.42 | 92.05 | 83.78 | 89.33 | 92.00 | 87.67 | 94.00 | 94.33 | 90.67 | 91.33 | 94.00 | 88.67 | 90.27 | 93.10 | 87.69 |
100% VB+15% MLM | 83.44 | 88.08 | 78.81 | 87.33 | 90.33 | 83.33 | 92.33 | 94.00 | 88.00 | 90.00 | 92.00 | 87.67 | 88.28 | 91.10 | 84.45 |
100% NN+15% MLM | 85.10 | 87.75 | 80.13 | 88.00 | 91.67 | 85.33 | 92.00 | 91.67 | 88.00 | 90.67 | 93.33 | 87.67 | 88.94 | 91.10 | 85.28 |
100% NE+100%VB+15% MLM | 84.43 | 90.73 | 82.12 | 88.67 | 91.00 | 85.33 | 94.00 | 92.33 | 88.33 | 91.00 | 94.33 | 88.00 | 89.53 | 92.10 | 85.95 |
100% NE+100%NN+15% MLM | 85.43 | 88.08 | 79.47 | 88.33 | 89.67 | 85.00 | 95.00 | 94.67 | 91.33 | 92.33 | 93.33 | 89.67 | 90.27 | 91.44 | 86.37 |
100% NE+100% VB +100%NN+15% MLM | 83.11 | 88.41 | 79.80 | 86.67 | 91.33 | 84.33 | 91.67 | 89.67 | 85.33 | 90.33 | 93.67 | 88.33 | 87.94 | 90.77 | 84.45 |
LEMmono+LEMpara | |||||||||||||||
100% NE+15% TLM on 15%MLM | 89.07 | 90.73 | 85.10 | 89.33 | 91.00 | 85.67 | 95.67 | 94.67 | 92.67 | 91.00 | 93.33 | 88.33 | 91.27 | 92.43 | 87.94 |
100% VB+15% TLM on 15%MLM | 88.41 | 91.00 | 84.77 | 87.67 | 91.00 | 85.67 | 93.67 | 93.67 | 90.67 | 92.33 | 93.00 | 90.00 | 90.52 | 92.17 | 87.78 |
100% NN+15% TLM on 15%MLM | 88.74 | 90.07 | 84.44 | 89.67 | 91.67 | 86.67 | 94.67 | 93.33 | 90.67 | 92.00 | 91.67 | 87.33 | 91.27 | 91.68 | 87.28 |
100% NE+100% VB+15% TLM on 15%MLM | 86.75 | 90.73 | 83.11 | 89.67 | 90.33 | 86.33 | 92.67 | 92.67 | 88.67 | 92.33 | 95.33 | 90.67 | 90.36 | 92.27 | 87.19 |
100% NE+100% NN+15% TLM on 15%MLM | 86.42 | 90.73 | 83.11 | 87.33 | 89.33 | 84.00 | 94.67 | 93.33 | 91.67 | 92.00 | 93.67 | 88.33 | 90.11 | 91.77 | 86.78 |
100% NE+100% VB+ 100% NN+ 15% TLM on 15%MLM | 85.43 | 91.39 | 81.79 | 89.00 | 92.33 | 86.67 | 93.67 | 93.33 | 88.67 | 90.33 | 93.00 | 87.00 | 89.61 | 92.51 | 86.03 |
15% TLM on 100%NE+15%MLM | 87.09 | 89.73 | 83.44 | 89.33 | 92.00 | 86.33 | 94.33 | 92.67 | 89.33 | 92.00 | 93.67 | 89.67 | 90.69 | 92.02 | 87.19 |
15% NE+15%TLM on 100%NE+15%MLM | 88.33 | 93.33 | 87.33 | 88.33 | 93.33 | 87.33 | 93.33 | 94.00 | 89.00 | 92.00 | 93.67 | 89.67 | 90.50 | 93.58 | 88.33 |
100% VB+15% TLM on 100%NE+15%MLM | 86.42 | 90.07 | 83.11 | 90.00 | 92.00 | 87.67 | 94.33 | 93.00 | 90.33 | 92.00 | 94.67 | 90.00 | 90.69 | 92.43 | 87.78 |
100% NN+15% TLM on 100%NE+15%MLM | 86.09 | 91.72 | 83.78 | 89.67 | 92.67 | 87.67 | 95.33 | 95.00 | 91.67 | 92.67 | 93.33 | 89.67 | 90.94 | 93.18 | 88.19 |
100% NE+100% VB+15% TLM on 100%NE+15%MLM/ | 87.09 | 90.07 | 84.11 | 89.00 | 91.33 | 86.67 | 95.67 | 94.67 | 91.67 | 90.67 | 92.33 | 88.67 | 90.61 | 92.10 | 87.78 |
100% NE+100% NN+15% TLM on 100%NE+15%MLM/ | 86.09 | 90.73 | 83.11 | 90.00 | 93.67 | 89.33 | 94.33 | 94.00 | 90.33 | 91.00 | 95.33 | 89.33 | 90.36 | 93.43 | 88.03 |
100%NE+100%VB+100%NN+15%TLM on 100%NE+15%MLM/ | 86.09 | 93.05 | 84.11 | 89.67 | 92.00 | 88.00 | 95.67 | 95.00 | 93.33 | 91.00 | 94.00 | 89.33 | 90.61 | 93.51 | 88.69 |
15% TLM on (100%VB+15%MLM) | 89.07 | 88.41 | 83.44 | 89.67 | 92.33 | 87.00 | 93.33 | 93.67 | 90.00 | 91.00 | 91.67 | 87.33 | 90.77 | 91.52 | 86.94 |
100% NE+ 15% TLM on (100%VB+15%MLM) | 87.75 | 88.74 | 83.11 | 90.00 | 91.00 | 86.33 | 95.00 | 94.67 | 91.67 | 93.00 | 91.67 | 88.00 | 91.44 | 91.52 | 87.28 |
100% VB+ 15% TLM on (100%VB+15%MLM) | 88.74 | 90.75 | 84.44 | 89.67 | 92.00 | 86.67 | 92.67 | 94.33 | 89.33 | 92.33 | 93.33 | 89.33 | 90.85 | 92.60 | 87.44 |
100% NN+15% TLM on (100%VB+15%MLM) | 86.42 | 90.73 | 84.11 | 89.67 | 92.33 | 86.33 | 91.33 | 93.33 | 88.00 | 92.00 | 94.00 | 89.67 | 89.86 | 92.60 | 87.03 |
100% NE+ 100% VB+ 15% TLM on (100%VB+15%MLM) | 85.76 | 88.08 | 81.13 | 90.33 | 92.33 | 88.00 | 93.00 | 91.33 | 88.67 | 92.33 | 92.00 | 88.33 | 90.36 | 90.94 | 86.53 |
100% NE+ 100% NN+ 15% TLM on (100%VB+15%MLM) | 87.09 | 89.07 | 82.78 | 88.67 | 92.33 | 85.00 | 93.33 | 93.67 | 89.67 | 92.00 | 91.33 | 87.00 | 90.27 | 91.60 | 86.11 |
100% NE+ 100% NN+ 100%VB+ 15% TLM on (100%VB+15%MLM) | 85.77 | 90.40 | 83.11 | 89.67 | 92.00 | 86.00 | 92.33 | 93.67 | 88.67 | 92.00 | 92.00 | 88.00 | 89.94 | 92.02 | 86.44 |
15% TLM on (100%NN+15%MLM) | 88.41 | 91.39 | 85.76 | 88.33 | 92.67 | 85.67 | 95.67 | 95.67 | 91.67 | 91.00 | 93.67 | 89.33 | 90.85 | 93.35 | 88.11 |
100% NE+ 15% TLM on (100%NN+15%MLM) | 89.40 | 92.72 | 87.42 | 90.13 | 93.33 | 88.67 | 96.67 | 93.00 | 90.67 | 92.33 | 93.67 | 89.67 | 92.13 | 93.18 | 89.10 |
100% VB+15% TLM on (100%NN+15%MLM) | 87.75 | 90.07 | 83.11 | 87.67 | 92.00 | 85.00 | 93.33 | 93.33 | 89.00 | 89.67 | 92.33 | 87.67 | 89.60 | 91.93 | 86.19 |
100% NN+ 15% TLM on (100%NN+15%MLM) | 85.43 | 90.73 | 82.12 | 88.67 | 93.00 | 86.67 | 95.33 | 93.67 | 91.00 | 93.00 | 92.67 | 89.33 | 90.61 | 92.52 | 87.28 |
100% NE+ 100% VB+ 15% TLM on (100%NN+15%MLM) | 86.09 | 90.40 | 82.78 | 91.33 | 92.00 | 88.33 | 94.67 | 94.67 | 91.33 | 91.33 | 93.00 | 88.33 | 90.86 | 92.52 | 87.69 |
100% NE+ 100% NN+ 15% TLM on (100%NN+15%MLM) | 87.75 | 89.40 | 83.11 | 88.33 | 92.00 | 85.67 | 95.00 | 93.67 | 90.67 | 92.00 | 93.00 | 89.33 | 90.77 | 92.02 | 87.19 |
100% NE+ 100% NN+ 100%VB+ 15% TLM on (100%NN+15%MLM) | 85.43 | 92.05 | 83.78 | 89.33 | 93.00 | 87.00 | 94.33 | 93.33 | 90.00 | 90.67 | 92.67 | 88.00 | 89.94 | 92.76 | 87.19 |
15% TLM on (100%NE+100%VB+15%MLM) | 86.09 | 91.06 | 83.44 | 90.33 | 90.67 | 87.33 | 96.33 | 94.33 | 92.33 | 92.33 | 94.33 | 90.00 | 91.27 | 92.60 | 88.28 |
100% NE+ 15% TLM on (100%NE+100%VB+15%MLM) | 86.75 | 90.07 | 83.44 | 89.33 | 91.33 | 86.67 | 93.67 | 94.67 | 91.33 | 92.00 | 93.33 | 89.67 | 90.44 | 92.35 | 87.78 |
100% VB+15% TLM on (100%NE+100%VB+15%MLM) | 84.44 | 91.39 | 81.46 | 89.33 | 91.67 | 85.00 | 95.33 | 93.33 | 91.00 | 92.00 | 93.33 | 89.33 | 90.28 | 92.43 | 86.70 |
100% NN+ 15% TLM on (100%NE+100%VB+15%MLM) | 85.76 | 92.05 | 85.00 | 90.67 | 93.00 | 88.33 | 95.67 | 95.33 | 92.67 | 89.33 | 92.67 | 86.33 | 90.36 | 93.26 | 88.08 |
100% NE+100% VB+ 15% TLM on (100%NE+100%VB+15%MLM) | 87.09 | 89.73 | 83.11 | 90.67 | 92.33 | 88.33 | 94.67 | 92.67 | 89.33 | 92.00 | 93.00 | 90.00 | 91.10 | 91.93 | 87.69 |
100% NE+ 100% NN+ 15% TLM on (100%NE+100%VB+15%MLM) | 85.76 | 92.05 | 85.00 | 90.67 | 91.67 | 87.67 | 95.67 | 94.00 | 91.33 | 90.67 | 93.67 | 88.33 | 90.69 | 92.85 | 88.08 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+15%MLM) | 86.75 | 92.05 | 84.77 | 88.33 | 90.00 | 86.67 | 94.33 | 94.67 | 90.67 | 89.67 | 93.67 | 87.67 | 89.77 | 92.60 | 87.44 |
15% TLM on (100%NE+100%NN+15%MLM) | 86.75 | 91.72 | 83.11 | 90.67 | 92.33 | 88.67 | 95.67 | 93.67 | 91.33 | 92.00 | 94.33 | 90.00 | 91.27 | 93.01 | 88.28 |
100% NE+ 15% TLM on (100%NE+100%NN+15%MLM) | 86.75 | 87.47 | 81.79 | 90.33 | 93.00 | 88.00 | 95.33 | 93.33 | 90.33 | 93.67 | 94.00 | 90.33 | 91.52 | 91.95 | 87.61 |
100% VB+15% TLM on (100%NE+100%NN+15%MLM) | 87.75 | 90.40 | 83.44 | 89.33 | 92.33 | 86.67 | 95.00 | 94.00 | 91.33 | 91.67 | 94.67 | 89.67 | 90.94 | 92.85 | 87.78 |
100% NN+ 15% TLM on (100%NE+100%NN+15%MLM) | 87.75 | 90.40 | 84.77 | 87.67 | 89.67 | 84.00 | 95.33 | 95.33 | 92.00 | 90.00 | 93.67 | 88.67 | 90.19 | 92.27 | 87.36 |
100% NE+100% VB+ 15% TLM on (100%NE+100%NN+15%MLM) | 85.76 | 87.75 | 81.13 | 90.33 | 90.67 | 86.00 | 94.33 | 92.00 | 89.00 | 92.33 | 93.67 | 90.00 | 90.69 | 91.02 | 86.53 |
100% NE+100% NN+ 15% TLM on (100%NE+100%NN+15%MLM) | 87.75 | 90.07 | 84.44 | 90.33 | 92.00 | 86.67 | 96.00 | 94.00 | 91.67 | 93.00 | 93.67 | 89.00 | 91.77 | 92.43 | 87.94 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%NN+15%MLM) | 85.76 | 88.74 | 91.46 | 89.67 | 92.00 | 86.00 | 94.33 | 93.67 | 91.00 | 93.00 | 94.00 | 89.67 | 90.69 | 92.10 | 89.53 |
15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 86.09 | 89.40 | 82.45 | 89.33 | 92.00 | 87.00 | 94.33 | 91.67 | 88.33 | 91.33 | 92.00 | 86.33 | 90.27 | 91.27 | 86.03 |
100% NE+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 88.76 | 89.40 | 84.44 | 89.67 | 91.33 | 87.33 | 95.00 | 94.00 | 90.67 | 90.33 | 92.00 | 87.00 | 90.94 | 91.68 | 87.36 |
100% VB+15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 85.76 | 89.40 | 82.12 | 90.33 | 93.33 | 88.67 | 94.67 | 94.00 | 90.67 | 93.00 | 93.67 | 90.00 | 90.94 | 92.60 | 87.86 |
100% NN+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 85.43 | 90.73 | 82.78 | 89.67 | 91.67 | 87.33 | 95.33 | 92.33 | 90.33 | 90.00 | 93.33 | 86.67 | 90.11 | 92.02 | 86.78 |
100% NE+100% VB+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 86.42 | 91.00 | 82.78 | 88.67 | 91.67 | 86.33 | 94.00 | 92.67 | 89.33 | 93.00 | 93.67 | 90.33 | 90.52 | 92.25 | 87.19 |
100% NE+ 100% NN+ 15% TLM on (100%NE+100%VB+100%NN+15%MLM) | 86.75 | 90.73 | 84.11 | 90.67 | 93.33 | 88.33 | 97.33 | 92.67 | 92.33 | 91.33 | 92.00 | 86.67 | 91.52 | 92.18 | 87.86 |
100%NE+100%VB+100%NN+15%TLM on (100%NE+100%VB+100%NN+15%MLM) | 87.42 | 89.40 | 82.78 | 89.33 | 92.33 | 86.33 | 94.67 | 95.00 | 91.33 | 91.33 | 94.00 | 88.67 | 90.69 | 92.68 | 87.28 |
Appendix D Limitations of the NER Model and Pos Tagger
Examples highlighting the error categories found with the NER model and PoS taggers (Section 8) are shown in Table 17 and Table 18 respectively.
![]() |
![]() |
References
- \bibcommenthead
- Devlin et al. [2019] Devlin J, Chang MW, Lee K, Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); 2019. p. 4171–4186.
- Conneau et al. [2020] Conneau A, Khandelwal K, Goyal N, Chaudhary V, Wenzek G, Guzmán F, et al. Unsupervised Cross-lingual Representation Learning at Scale. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020. p. 8440–8451.
- Ács et al. [2021] Ács J, Lévai D, Kornai A. Evaluating Transferability of BERT Models on Uralic Languages. In: Proceedings of the Seventh International Workshop on Computational Linguistics of Uralic Languages; 2021. p. 8–17.
- Dhananjaya et al. [2022] Dhananjaya V, Demotte P, Ranathunga S, Jayasena S. BERTifying Sinhala-A Comprehensive Analysis of Pre-trained Language Models for Sinhala Text Classification. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference; 2022. p. 7377–7385.
- Hu et al. [2020] Hu J, Ruder S, Siddhant A, Neubig G, Firat O, Johnson M. XTREME: a massively multilingual multi-task benchmark for evaluating cross-lingual generalization. In: Proceedings of the 37th International Conference on Machine Learning; 2020. p. 4411–4421.
- Hu et al. [2021] Hu J, Johnson M, Firat O, Siddhant A, Neubig G. Explicit Alignment Objectives for Multilingual Bidirectional Encoders. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies; 2021. p. 3633–3643.
- Conneau and Lample [2019] Conneau A, Lample G. Cross-lingual language model pretraining. Advances in neural information processing systems. 2019;32.
- Nastase and Merlo [2024] Nastase V, Merlo P. Tracking linguistic information in transformer-based sentence embeddings through targeted sparsification. In: Proceedings of the 9th Workshop on Representation Learning for NLP (RepL4NLP-2024); 2024. p. 203–214.
- Nastase and Merlo [2023] Nastase V, Merlo P. Grammatical information in BERT sentence embeddings as two-dimensional arrays. In: Proceedings of the 8th Workshop on Representation Learning for NLP (RepL4NLP 2023); 2023. p. 22–39.
- Aoyama and Schneider [2022] Aoyama T, Schneider N. Probe-less probing of BERT’s layer-wise linguistic knowledge with masked word prediction. In: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop; 2022. p. 195–201.
- Sun et al. [2019] Sun Y, Wang S, Li Y, Feng S, Chen X, Zhang H, et al. Ernie: Enhanced representation through knowledge integration. arXiv preprint arXiv:190409223. 2019;.
- Joshi et al. [2020] Joshi M, Chen D, Liu Y, Weld DS, Zettlemoyer L, Levy O. SpanBERT: Improving Pre-training by Representing and Predicting Spans. Transactions of the Association for Computational Linguistics. 2020;8:64–77.
- Levine et al. [2020] Levine Y, Lenz B, Lieber O, Abend O, Leyton-Brown K, Tennenholtz M, et al. PMI-Masking: Principled masking of correlated spans. In: International Conference on Learning Representations; 2020. .
- Rajpurkar et al. [2016] Rajpurkar P, Zhang J, Lopyrev K, Liang P. SQuAD: 100,000+ Questions for Machine Comprehension of Text. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing; 2016. p. 2383–2392.
- Lai et al. [2017] Lai G, Xie Q, Liu H, Yang Y, Hovy E. RACE: Large-scale ReAding Comprehension Dataset From Examinations. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing; 2017. p. 785–794.
- Wang et al. [2018] Wang A, Singh A, Michael J, Hill F, Levy O, Bowman S. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding. In: Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP; 2018. p. 353–355.
- Zhuang [2023] Zhuang Y. Heuristic Masking for Text Representation Pretraining. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2023. p. 1–5.
- Golchin et al. [2023] Golchin S, Surdeanu M, Tavabi N, Kiapour A. Do not Mask Randomly: Effective Domain-adaptive Pre-training by Masking In-domain Keywords. In: Proceedings of the 8th Workshop on Representation Learning for NLP (RepL4NLP 2023); 2023. p. 13–21.
- Wettig et al. [2023] Wettig A, Gao T, Zhong Z, Chen D. Should You Mask 15% in Masked Language Modeling? In: Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics; 2023. p. 2977–2992.
- Feng et al. [2022] Feng F, Yang Y, Cer D, Arivazhagan N, Wang W. Language-agnostic BERT Sentence Embedding. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); 2022. p. 878–891.
- Ranathunga et al. [2024] Ranathunga S, Ranasinghea A, Shamala J, Dandeniyaa A, Galappaththia R, Samaraweeraa M. A Multi-way Parallel Named Entity Annotated Corpus for English, Tamil and Sinhala. arXiv preprint arXiv:241202056. 2024;.
- Akbik et al. [2018] Akbik A, Blythe D, Vollgraf R. Contextual String Embeddings for Sequence Labeling. In: COLING 2018, 27th International Conference on Computational Linguistics; 2018. p. 1638–1649.
- Fernando and Ranathunga [2018] Fernando S, Ranathunga S. Evaluation of different classifiers for sinhala pos tagging. In: 2018 Moratuwa Engineering Research Conference (MERCon). IEEE; 2018. p. 96–101.
- Fernando et al. [2016] Fernando S, Ranathunga S, Jayasena S, Dias G. Comprehensive part-of-speech tag set and svm based pos tagger for sinhala. In: Proceedings of the 6th Workshop on South and Southeast Asian Natural Language Processing (WSSANLP2016); 2016. p. 173–182.
- Sarveswaran and Dias [2020] Sarveswaran K, Dias G. ThamizhiUDp: A Dependency Parser for Tamil. In: Proceedings of the 17th International Conference on Natural Language Processing (ICON); 2020. p. 200–207.
- Artetxe and Schwenk [2019] Artetxe M, Schwenk H. Margin-based Parallel Corpus Mining with Multilingual Sentence Embeddings. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2019. p. 3197–3203.
- Fernando et al. [2023] Fernando A, Ranathunga S, Sachintha D, Piyarathna L, Rajitha C. Exploiting bilingual lexicons to improve multilingual embedding-based document and sentence alignment for low-resource languages. Knowledge and Information Systems. 2023;65(2):571–612.
- Artetxe and Schwenk [2019] Artetxe M, Schwenk H. Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond. Transactions of the Association for Computational Linguistics. 2019;7:597–610.
- Yang et al. [2020] Yang Y, Cer D, Ahmad A, Guo M, Law J, Constant N, et al. Multilingual Universal Sentence Encoder for Semantic Retrieval. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: System Demonstrations; 2020. p. 87–94.
- Schwenk et al. [2021] Schwenk H, Wenzek G, Edunov S, Grave É, Joulin A, Fan A. CCMatrix: Mining Billions of High-Quality Parallel Sentences on the Web. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers); 2021. p. 6490–6500.
- Costa-jussà et al. [2022] Costa-jussà MR, Cross J, Çelebi O, Elbayad M, Heafield K, Heffernan K, et al. No language left behind: Scaling human-centered machine translation. arXiv preprint arXiv:220704672. 2022;.
- Kreutzer et al. [2022] Kreutzer J, Caswell I, Wang L, Wahab A, van Esch D, Ulzii-Orshikh N, et al. Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets. Transactions of the Association for Computational Linguistics. 2022;10:50–72. 10.1162/tacl_a_00447.
- Ranathunga et al. [2024] Ranathunga S, De Silva N, Menan V, Fernando A, Rathnayake C. Quality Does Matter: A Detailed Look at the Quality and Utility of Web-Mined Parallel Corpora. In: Graham Y, Purver M, editors. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers). St. Julian’s, Malta: Association for Computational Linguistics; 2024. p. 860–880. Available from: https://aclanthology.org/2024.eacl-long.52.
- Post [2018] Post M. A Call for Clarity in Reporting BLEU Scores. In: Proceedings of the Third Conference on Machine Translation: Research Papers. Belgium, Brussels: Association for Computational Linguistics; 2018. p. 186–191. Available from: https://www.aclweb.org/anthology/W18-6319.
- Popović [2015] Popović M. chrF: character n-gram F-score for automatic MT evaluation. In: Proceedings of the tenth workshop on statistical machine translation; 2015. p. 392–395.
- Popović [2017] Popović M. chrF++: words helping character n-grams. In: Proceedings of the second conference on machine translation; 2017. p. 612–618.
- Goyal et al. [2022] Goyal N, Gao C, Chaudhary V, Chen PJ, Wenzek G, Ju D, et al. The flores-101 evaluation benchmark for low-resource and multilingual machine translation. Transactions of the Association for Computational Linguistics. 2022;10:522–538.
- Barbieri et al. [2022] Barbieri F, Anke LE, Camacho-Collados J. XLM-T: Multilingual Language Models in Twitter for Sentiment Analysis and Beyond. In: Proceedings of the Thirteenth Language Resources and Evaluation Conference; 2022. p. 258–266.
- Myers-Scotton and Jake [1993] Myers-Scotton C, Jake J.: Duelling languages. Grammatical structure in Codeswitching–Clarendon Press. Oxford.
- Joshi et al. [2020] Joshi P, Santy S, Budhiraja A, Bali K, Choudhury M. The State and Fate of Linguistic Diversity and Inclusion in the NLP World. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020. p. 6282–6293.
- Ranathunga and de Silva [2022] Ranathunga S, de Silva N. Some Languages are More Equal than Others: Probing Deeper into the Linguistic Disparity in the NLP World. In: Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing; 2022. p. 823–848.
- de Silva [2023] de Silva N. Survey on Publicly Available Sinhala Natural Language Processing Tools and Research. arXiv preprint arXiv:190602358v20. 2023;.
- Kudugunta et al. [2024] Kudugunta S, Caswell I, Zhang B, Garcia X, Xin D, Kusupati A, et al. Madlad-400: A multilingual and document-level large audited dataset. Advances in Neural Information Processing Systems. 2024;36.
- Fernando et al. [2020] Fernando A, Ranathunga S, Dias G. Data augmentation and terminology integration for domain-specific sinhala-english-tamil statistical machine translation. arXiv preprint arXiv:201102821. 2020;.
- El-Kishky et al. [2020] El-Kishky A, Chaudhary V, Guzmán F, Koehn P. CCAligned: A Massive Collection of Cross-Lingual Web-Document Pairs. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP); 2020. p. 5960–5969.
- Bañón et al. [2020] Bañón M, Chen P, Haddow B, Heafield K, Hoang H, Esplà-Gomis M, et al. ParaCrawl: Web-Scale Acquisition of Parallel Corpora. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics; 2020. p. 4555–4567.
- Rathnayake et al. [2022] Rathnayake H, Sumanapala J, Rukshani R, Ranathunga S. Adapter-based fine-tuning of pre-trained multilingual language models for code-mixed and code-switched text classification. Knowledge and Information Systems. 2022;64(7):1937–1966.
- Koloski et al. [2023] Koloski B, Škrlj B, Robnik-Šikonja M, Pollak S. Measuring Catastrophic Forgetting in Cross-Lingual Transfer Paradigms: Exploring Tuning Strategies. arXiv preprint arXiv:230906089. 2023;.
- Udawatta et al. [2024] Udawatta P, Udayangana I, Gamage C, Shekhar R, Ranathunga S. Use of prompt-based learning for code-mixed and code-switched text classification. World Wide Web. 2024;27(5):63.
- Ott et al. [2019] Ott M, Edunov S, Baevski A, Fan A, Gross S, Ng N, et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling. In: Proceedings of NAACL-HLT 2019: Demonstrations; 2019. p. 48–53.
- Kocmi et al. [2024] Kocmi T, Zouhar V, Federmann C, Post M. Navigating the Metrics Maze: Reconciling Score Magnitudes and Accuracies. In: Ku LW, Martins A, Srikumar V, editors. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Bangkok, Thailand: Association for Computational Linguistics; 2024. p. 1999–2014. Available from: https://aclanthology.org/2024.acl-long.110.