WeChat Neural Machine Translation Systems for WMT21
Abstract
This paper introduces WeChat AI’s participation in WMT 2021 shared news translation task on EnglishChinese, EnglishJapanese, JapaneseEnglish and EnglishGerman. Our systems are based on the Transformer Vaswani et al. (2017) with several novel and effective variants. In our experiments, we employ data filtering, large-scale synthetic data generation (i.e., back-translation, knowledge distillation, forward-translation, iterative in-domain knowledge transfer), advanced finetuning approaches, and boosted Self-BLEU based model ensemble. Our constrained systems achieve 36.9, 46.9, 27.8 and 31.3 case-sensitive BLEU scores on EnglishChinese, EnglishJapanese, JapaneseEnglish and EnglishGerman, respectively. The BLEU scores of EnglishChinese, EnglishJapanese and JapaneseEnglish are the highest among all submissions, and that of EnglishGerman is the highest among all constrained submissions.
1 Introduction
We participate in the WMT 2021 shared news translation task in three language pairs and four language directions, EnglishChinese, EnglishJapanese, and EnglishGerman. In this year’s translation tasks, we mainly improve the final ensemble model’s performance by increasing the diversity of both the model architecture and the synthetic data, as well as optimizing the ensemble searching algorithm.
Diversity is a metric we are particularly interested in this year. To quantify the diversity among different models, we compute Self-BLEU Zhu et al. (2018) from the translations of the models on the valid set. To be precise, we use the translation of one model as the hypothesis and the translations of other models as references to calculate an average BLEU score. A higher Self-BLEU means this model is less diverse.
For model architectures Vaswani et al. (2017); Meng and Zhang (2019); Yan et al. (2020), we exploit several novel Transformer variants to strengthen model performance and diversity. Besides the Pre-Norm Transformer, the Post-Norm Transformer is also used as one of our baselines this year. We adopt some novel initialization methods Huang et al. (2020) to alleviate the gradient vanishing problem of the Post-Norm Transformer. We combine the Average Attention Transformer (AAN) Zhang et al. (2018) and Multi-Head-Attention Vaswani et al. (2017) to derive a series of effective and diverse model variants. Furthermore, Talking-Heads Attention Shazeer et al. (2020) is introduced to the Transformer and shows a significant diversity from all the other variants.
For the synthetic data generation, we exploit the large-scale back-translation Sennrich et al. (2016a) method to leverage the target-side monolingual data and the sequence-level knowledge distillation Kim and Rush (2016) to leverage the source-side of bilingual data. To use the source-side monolingual data, we explore forward-translation by ensemble models to get general domain synthetic data. We also use iterative in-domain knowledge transfer Meng et al. (2020) to generate in-domain data. Furthermore, several data augmentation methods are applied to improve the model robustness, including different token-level noise and dynamic top-p sampling.
For training strategies, we mainly focus on scheduled sampling based on decoding steps Liu et al. (2021b), the confidence-aware scheduled sampling Mihaylova and Martins (2019); Duckworth et al. (2019); Liu et al. (2021a), the target denoising Meng et al. (2020) method and the Graduated Label Smoothing Wang et al. (2020) for in-domain finetuning.
For model ensemble, we select high-potential candidate models based on two indicators, namely model performance (BLEU scores on valid set) and model diversity (Self-BLEU scores among all other models). Furthermore, we propose a search algorithm based on the Self-BLEU scores between the candidate models with selected models. We observed that this novel method can achieve the same BLEU score as the brute force search while saving approximately 95% of search time.
2 Model Architectures
In this section, we describe the model architectures used in the four translation directions, including several different variants for the Transformer Vaswani et al. (2017) .
2.1 Model Configurations
Deeper and wider architectures are used this year since they show strong capacity as the number of parameters increases. In our experiments, we use multiple model configurations with 20/25-layer encoders for deeper models and the hidden size is set to 1024 for all models. Compared to our WMT20 models Meng et al. (2020), we also increase the decoder depth from 6 to 8 and 10 as we find that gives a certain improvement, but deeper depths give limited performance gains. For the wider models, we adopt 8/12/15 encoder layers and 1024/2048 for hidden size. The filter sizes of models are set from 8192 to 15000. Note that all the above model configurations are applied to the following variant models.
2.2 Transformer with Different Layer-Norm
The Transformer Vaswani et al. (2017) with Pre-Norm Xiong et al. (2020) is a widely used architecture in machine translation. It is also our baseline model as its performance and training stability is better than the Post-Norm counterpart.
Recent studies Liu et al. (2020); Huang et al. (2020) show that the unstable training problem of Post-Norm Transformer can be mitigated by modifying initialization of the network and the successfully converged Post-Norm models generally outperform Pre-Norm counterparts. We adopt these initialization methods Huang et al. (2020) to our training flows to stabilize the training of deep Post-Norm Transformer. Our experiments have shown that the Post-Norm model has a good diversity compared to the Pre-Norm Model and slightly outperform the Pre-Norm Model. We will further analyze the model diversity of different variants in Sec. 5.1.
2.3 Average Attention Transformer
We also use Average Attention Transformer (AAN) Zhang et al. (2018) as we used last year to introduce more model diversity. In the Average Attention Transformer, a fast and straightforward average attention is utilized to replace the self-attention module in the decoder with almost no performance loss. The context representation for each input embedding is as follows:
(1) |
where is the input embedding for step and is the current time step. denotes the position-wise feed-forward network proposed by Vaswani et al. (2017).
In our preliminary experiments, we observe that the Self-BLEU Zhu et al. (2018) scores between AAN and Transformer are lower than the scores between the Transformer with different configurations.
2.4 Weighted Attention Transformer
We further explore three weighting strategies to improve the modeling of history information from previous positions in AAN. Compared to the average weight across all positions, we try three methods including decreasing weights with position increasing, learnable weights and exponential weights. In our experiments, We observe exponential weights perform best among all these strategies. The exponential weights context representation is calculated as follows:
(2) |
(3) |
where is a tuned parameter. In our previous experiments, we test different alpha, including 0.3, 0.5, and 0.7, on the valid set and we set the alpha to 0.7 in all subsequent experiments as it slightly outperform the others.

2.5 Mixed-AAN Transformers
Our preliminary experiments show that the decoder structure is strongly related to the model diversity in the Transformer. Therefore, we propose to stack different types of decoder layers to derive different Transformer variants. As shown in Figure 1, we mainly adopt three Mixed-AAN Transformer architectures: a) Alternately mixing the standard self-attention layer and the average attention layer, b) Continuously stacking several average attention layers on the bottom layers and then stacking self-attention layers for the rest layers. c) Stacking both the self-attention layer and average attention layer at each layer and using their average sum to form the final hidden states (named as ‘dual attention layer’).
In the experiments, Mixed-AAN not only performs better but also shows strong diversity compared to the vanilla Transformer. With four Mixed-AAN models, we reach a better ensemble result than the result with ten models which consist of deeper and wider standard Transformer. We will further analyze the effects of different architectures from performance, diversity, and model ensemble in Sec. 5.1
2.6 Talking-Heads Attention
In Multi-Head Attention, the different attention heads perform separate computations, which are then summed at the end. Talking-Heads Attention Shazeer et al. (2020) is a new variation that inserts two additional learned linear projection weights, and , to transform the attention-logits and the attention scores respectively, moving information across attention heads. The calculation formula is as follows:
|
(4) |
We adopt this method in both encoders and decoders to improve information interaction between attention heads. This approach shows the most remarkable diversity among all the above variants with only a slight performance loss.

3 System Overview
In this section, we describe our system used in the WMT 2021 news shared task. We depicts the overview of our NMT system in Figure 2, which can be divided into four parts, namely data filtering, large-scale synthetic data generation, in-domain finetuning, and ensemble. The synthetic data generation part further includes the generation of general domain and in-domain data. Next, we proceed to illustrate these four parts.
3.1 Data Filtering
We filter the bilingual training corpus with the following rules for most language pairs:
-
•
Normalize punctuation with Moses scripts except Japanese data.
-
•
Filter out the sentences longer than 100 words or exceed 40 characters in a single word.
-
•
Filter out the duplicated sentence pairs.
-
•
The word ratio between the source and the target words must not exceed 1:4 or 4:1.
-
•
Filter out the sentences where the fast-text result does not match the origin language.
-
•
Filter out the sentences that have invalid Unicode characters.
Besides these rules, we filter out sentence pairs in which Chinese sentence has English characters in En-Zh parallel data. The monolingual corpus is also filtered with the n-gram language model trained by the bilingual training data for each language. All the above rules are applied to synthetic parallel data.
3.2 General Domain Synthetic Data Generation
In this section, we describe our techniques for constructing general domain synthetic data. The general domain synthetic data is generated via large-scale back-translation, forward-translation and knowledge distillation to enhance the models’ performance for all domains. Then, we exploit the iterative in-domain knowledge transfer Meng et al. (2020) in Sec 3.3, which transfers in-domain knowledge to the vast source-side monolingual corpus, and builds our in-domain synthetic data. In the following sections, we elaborate the above techniques in detail.
3.2.1 Large-scale Back-Translation
Back-translation is the most commonly used data augmentation technique to incorporate the target side monolingual data into NMT Hoang et al. (2018). Previous work Edunov et al. (2018) has shown that different methods of generating pseudo corpus has a different influence on translation quality. Following these works, we attempt several generating strategies as follows:
-
•
Beam Search: Generate target translation by beam search with beam size 5.
-
•
Sampling Top-K: Select a word randomly from top-K (K is set to 10) words at each decoding step.
-
•
Dynamic Sampling Top-p: Selected a word at each decoding step from the smallest set whose cumulative probability mass exceeds p and the p is dynamically changing from 0.9 to 0.95 during data generation.
Note that we also use Tagged Back-Translation Caswell et al. (2019) in EnDe and Right-to-Left (R2L) back-translation in EnJa, as we achieve a better BLEU score after using these methods.
3.2.2 Knowledge Distillation
Knowledge Distillation (KD) has proven to be a powerful technique for NMT Kim and Rush (2016); Wang et al. (2021) to transfer knowledge from the teacher model to student models. In particular, we first use the teacher models to generate synthetic corpus in the forward direction (i.e., EnZh). Then, we use the generated corpus to train our student models.
Notably, Right-to-Left (R2L) knowledge distillation is a good complement to the Left-to-Right (L2R) way and can further improve model performance.
3.2.3 Forward-Translation
Using monolingual data from the source language to further enhance the performance and robustness of the model is also an effective approach. We use the ensemble model to generate high quality forward-translation data and obtain a stable improvement in EnZh and EnDe directions.
3.3 Iterative In-domain Knowledge Transfer
Since in-domain knowledge transfer Meng et al. (2020) delivered a massive performance boost last year, we still use this technique in EnJa and EnDe this year. It is not applied to EnZh because no significant improvement is observed. We guess the reason is that the in-domain finetuning in the EnZh direction does not bring a significant improvement compared to the other directions. And in-domain knowledge transfer is aiming at enhancing the effect of finetuning, so this does not have a noticeable effect in the English-Chinese direction.
We first use normal finetuning in Sec. 3.5 to equip our models with in-domain knowledge. Then, we ensemble these models to translate the source monolingual data into the target language. We use 4 models with different architectures and training data as our ensemble model. Next, we combine the source language sentences with the generated in-domain target language sentences as pseudo-parallel corpus. Afterwards, we retrain our models with both in-domain pseudo-parallel data and general domain synthetic data.
3.4 Data Augmentation
Once the pseudo-data is constructed, we further obtain diverse data by adding different noise. Compared to previous years’ WMT competitions, we implement a multi-level static noise approach for our pseudo corpus:
-
•
Token-level: Noise on every single subword after byte pair encoding.
-
•
Word-level: Noise on every single word before byte pair encoding.
-
•
Span-level: Noise on a continuous sequence of tokens before byte pair encoding.
The different granularities of noise make the data more diverse. The noise types are random replacement, random deletion and random permutation. We apply the three noise types in a parallel way for each sentence. The probability of enabling each of the three operations is 0.2.
Furthermore, an on-the-fly noise approach is applied to the synthetic data. By using on-the-fly noise, the model is trained with different noises in every epoch rather than all the same along this training stage.
3.5 In-domain Finetuning
A domain mismatch exists between the obtained system trained with large-scale general domain data and the target test set. To alleviate this mismatch, we finetune these convergent models on small scale in-domain data, which is widely used for domain adaption Luong and Manning (2015); Li et al. (2019). We take the previous test sets as in-domain data and extract documents that are originally created in the source language for each translation direction Sun et al. (2019). We also explore several advanced finetuning approaches to strengthen the effects of domain adaption and ease the exposure bias issue, which is more serious under domain shift.
Target Denoising Meng et al. (2020).
In the training stage, the model never sees its own errors. Thus the model trained with teacher-forcing is prune to accumulated errors in testing Ranzato et al. (2016). To mitigate this training-generation discrepancy, we add noisy perturbations into decoder inputs when finetuning. Thus the model becomes more robust to prediction errors by target denoising. Specifically, the finetuning data generator chooses 30% of sentence pairs to add noise, and keeps the remaining 70% of sentence pairs unchanged. For a chosen pair, we keep the source sentence unchanged, and replace the -th token of the target sentence with (1) a random token of the current target sentence 15% of the time (2) the unchanged -th token 85% of the time.
Graduated Label-smoothing Wang et al. (2020).
Finetuning on a small scale in-domain data can easily lead to the over-fitting phenomenon which is harmful to the model ensemble. It generally appears as the model over confidently outputting similar words. To further preventing over-fitting of in-domain finetuning, we apply the Graduated Label-smoothing approach, which assigns a higher smoothing penalty for high-confidence predictions, during in-domain finetuning. Concretely, following the paper’s setting, we set the smoothing penalty to 0.3 for tokens with confidence above 0.7, zero for tokens with confidence below 0.3, and 0.1 for the remaining tokens.
Confidence-Aware Scheduled Sampling.
Vanilla scheduled sampling Zhang et al. (2019) simulates the inference scene by randomly replacing golden target input tokens with predicted ones during training. However, its critical schedule strategies are only based on training steps, ignoring the real-time model competence. To address this issue, we propose confidence-aware scheduled sampling Liu et al. (2021a), which quantifies real-time model competence by the confidence of model predictions. At the t-th target token position, we calculate the model confidence as follow:
(5) |
Next, we design fine-grained schedule strategies based on the model competence. The fine-grained schedule strategy is conducted at all decoding steps simultaneously:
(6) |
where is a threshold to measure whether is high enough ( 0.9) to sample the predicted token .
We further sample more noisy tokens at high-confidence token positions, which prevents scheduled sampling from degenerating into the teacher forcing mode.
(7) |
where is a threshold to measure whether is high enough ( 0.95) to sample the random target token .
Scheduled Sampling Based on Decoding Steps.
We propose scheduled sampling methods based on decoding steps from the perspective of simulating the distribution of real translation errors Liu et al. (2021b). Namely, we gradually increase the selection probability of predicted tokens with the growth of the index of decoded tokens. At the -th decoding step, the probability of sampling golden tokens is calculated as follow:
-
•
Linear Decay: , where is the minimum value, and and is respectively the slope and offset of the decay.
-
•
Exponential Decay: , where is the radix to adjust the decay.
-
•
Inverse Sigmoid Decay: , where is the mathematical constant, and is a hyperparameter to adjust the decay.
Following our preliminary conclusions Liu et al. (2021b), we choose the exponential decay and set to 0.99 by default.
3.6 Boosted Self-BLEU based Ensemble (BSBE)
After we get numerous finetuned models, we need to search for the best combination for ensemble model. Ordinary random or greedy search is oversimplified to search for a good model combination and enumerate over all combinations of candidate models is inefficient. The Self-BLEU based pruning strategy Meng et al. (2020) we proposed in last year’s competition achieve definite improvements over the ordinary ensemble.
However, diversity is not the only feature we need to consider but the performance in the valid set is also an important metric. Therefore, we combine Self-BLEU and valid set BLEU together to derive a Boosted Self-BLEU-based Ensemble (BSBE) algorithm. Then, we apply a greedy search strategy in the top N ranked models to find the best ensemble models.
See algorithm 1 for the pseudo-code. The algorithm takes as input a list of n strong single models M, BLEU scores on valid set for each model B, average Self-BLEU scores for each model S, the number of models and the number of ensemble models . The algorithm return a list C consists of selected models. We calculate the weighted score for each model as line 2 in the pseudo-code. The weight calculated in line 3 is a factor to balance the scale of Self-BLUE and valid set BLEU. Then the list C initially contains the model has a highest weighted score. Next, we iteratively re-compute the average Self-BLEU between the remaining models in and selected models in C, based on which we select the model has minimum Self-BLEU score into C.
In our experiments, we save around 95% searching time by using this novel method to achieve the same BLEU score of the Brute Force search. We will further analyze the effect of Boosted Self-BLEU based Ensemble in section 5.2.
4 Experiments And Results
4.1 Settings
The implementation of our models is based on Fairseq111https://github.com/pytorch/fairseq for EnZh and ENDe, and OpenNMT222https://github.com/OpenNMT/OpenNMT-py for EnJa. All the single models are carried out on 8 NVIDIA V100 GPUs, each of which has 32 GB memory. We use the Adam optimizer with = 0.9, = 0.998. The gradient accumulation is used due to the high GPU memory consumption. The batch size is set to 8192 tokens per GPU and we set the “update-freq” parameter in Fairseq to 2. The learning rate is set to 0.0005 for Fairseq and 2.0 for OpenNMT. We use warmup step = 4000. We calculate sacreBLEU333https://github.com/mjpost/sacrebleu score for all experiments which is officially recommended.
EnZh | EnDe | EnJa | |
Bilingual Data | 30.7M | 74.8M | 12.3M |
Source Mono Data | 200.5M | 332.8M | 210.8M |
Target Mono Data | 405.2M | 237.9M | 354.7M |
System | EnZh | EnJa | JaEn | EnDe |
Baseline | 44.53 | 35.78 | 19.71 | 33.28 |
+ Back Translation | 46.52 | 36.12 | 20.82 | 35.28 |
+ Knowledge Distillation | 47.14 | 36.66 | 21.63 | 36.38 |
+ Forward Translation | 47.38 | – | – | 36.78 |
+ Mix BT | 48.17 | 37.22 | 22.11 | – |
+ Finetune | 49.81 | 42.54 | 25.91 | 39.21 |
+ Advanced Finetune | 50.20 | – | – | 39.56 |
+ 1st In-domain Knowledge Transfer | – | 40.32 | 24.49 | 39.23 |
+ Finetune | – | 43.66 | 26.24 | – |
+ Advanced Finetune | – | – | – | 39.87 |
+ 2nd In-domain Knowledge Transfer | – | 43.69 | 25.89 | – |
+ Finetune | – | 44.23 | 26.27 | – |
+ Advanced Finetune | – | 44.42 | 26.38 | – |
+ Normal Ensemble | 50.57 | 45.11 | 28.01 | 40.42 |
+ BSBE | 50.94 | 45.35 | 28.24 | 40.59 |
+ Post-Process | – | – | – | 41.88 |
4.2 Dataset
The statistics of all training data is shown in Table 1. For each language pair, the bilingual data is the combination of all parallel data released by WMT21. For monolingual data, we select data from News Crawl, Common Crawl and Extended Common Crawl, it is then divided into several parts, each containing 50M sentences.
For general domain synthetic data, we use all the target monolingual data to generate back-translation data and a part of source monolingual data (about 80 to 100 million for different languages) to get forward translation data. For the in-domain pseudo-parallel data, we use the entire source monolingual data and bilingual data. All the test and valid data from previous years are used as in-domain data.
We use the methods described in Sec. 3.1 to filter bilingual and monolingual data.
4.3 Pre-processing and Post-processing
English and German sentences are segmented by Moses444http://www.statmt.org/moses/, while Japanese use Mecab555https://github.com/taku910/mecab for segmentation. We segment the Chinese sentences with an in-house word segmentation tool. We apply punctuation normalization in English, German and Chinese data. Truecasing is applied to English Japanese and EnglishGerman. We use byte pair encoding BPE Sennrich et al. (2016b) with 32K operations for all the languages.
For the post-processing, we apply de-truecaseing and de-tokenizing on the English and German translations with the scripts provided in Moses. For the Chinese translations, we transpose the punctuations to the Chinese format.
4.4 EnglishChinese
The results of EnZh on newstest2020 are shown in Table 2. For the EnZh task, filtering out part of sentence pairs containing English characters in Chinese sentences shows a significant improvement in the valid set. After applying large-scale Back-Translation, we obtain +2.0 BLEU score on the baseline. We further gain +0.62 BLEU score after applying knowledge distillation and +0.24 BLEU from Forward-Translation. Surprisingly, we observe that adding more BT data from different shards with different generation strategy can further boost the model performance to 48.17. The finetuned model achieves a 49.81 BLEU score, which demonstrates that the domain of the training corpus is apart from the test set domain. The advanced finetuning further brings about 0.41 BLEU score gains compared to normal finetune. Our best single model achieves a 50.22 BLEU score.
In preliminary experiments, we select the best performing models as our ensemble combinations obtaining +0.4 BLEU score. On top of that, even after searching hundreds of models, no better results are obtained. With BSBE strategies in Sec. 3.6, a better model combination with less number of models are quickly searched, and we finally achieve 50.94 BLEU score. Our WMT2021 EnglishChinese submission achieves a SacreBLEU score of 36.9, which is the highest among all submissions and chrF score of 0.337.
4.5 EnglishJapanese
The results of EnJa on newstest2020 are shown in Table 2. For the EnJa task, we filter out the sentence pairs containing Japanese characters in the English side and vice versa. The Back-Translation and Knowledge Distillation improve the baseline from 35.78 to 36.66. Adding more BT data further brings in 0.56 improvements. The improvement by finetuning is much larger than other directions, which is 5.32 BLEU. We speculate that this is because there is less bilingual data for English and Japanese than for other languages, and the test results for Japanese are char level BLEU so this direction is more influenced by the in-domain finetuning. Two In-domain knowledge transfers improve BLEU score from 37.22 to 43.69. Normal finetune still provides 0.54 improvements after in-domain knowledge transfer. Then, we apply advanced finetuning methods to further get 0.19 BLEU improvements. Our final ensemble result outperforms baseline 9.57 BLEU.
Model | En-Zh | En-Ja | Ja-En | En-De |
Transformer | 49.92 | 44.27 | 26.12 | 39.76 |
Transformer with Post-Norm | 49.97 | - | - | - |
Average Attention Transformer | 49.91 | 44.38 | 26.31 | 39.62 |
Weighted Attention Transformer | 49.99 | - | - | 39.74 |
Average First Transformer | 50.14 | 44.42 | 26.37 | 39.87 |
Average Bottom Transformer | 50.10 | 44.36 | 26.38 | 39.77 |
Dual Attention Transformer | 50.20 | - | - | 39.87 |
Talking-Heads Attention | 49.89 | - | - | 39.70 |
Model | Transformer | Post-Norm | AAN | Weighted | Avg-First | Self-First | Dual | TH |
Transformer | 100 | 78.12 | 76.02 | 75.08 | 74.47 | 74.02 | 73.51 | 72.63 |
Post-Norm | 78.12 | 100 | 76.12 | 75.10 | 74.33 | 74.05 | 73.45 | 72.59 |
AAN | 76.02 | 76.12 | 100 | 79.24 | 74.81 | 74.97 | 73.43 | 72.13 |
Weighted | 75.08 | 75.10 | 79.24 | 100 | 74.72 | 74.93 | 73.55 | 72.21 |
Avg-First | 74.46 | 74.33 | 74.81 | 74.72 | 100 | 75.25 | 74.28 | 72.25 |
Avg-Bot | 74.02 | 74.05 | 74.97 | 74.93 | 75.25 | 100 | 74.21 | 72.33 |
Dual | 73.51 | 73.45 | 73.43 | 73.55 | 74.28 | 74.21 | 100 | 72.23 |
TH | 72.63 | 72.59 | 72.13 | 72.21 | 72.25 | 72.33 | 72.23 | 100 |
4.6 JapaneseEnglish
The JaEn task follows the same training procedure as EnJa. From Table 2, we can observe that Back-Translation can provide 1.11 BLEU improvements from baseline. Knowledge Distillation and more BT data can improve the BLEU score from 20.82 to 22.11. The finetuning improvement is 3.8 which is slightly less than the EnJa direction but still larger than EnZh and EnDe. We also apply two-turn in-domain knowledge transfer and further boost the BLEU score to 25.89. After normal finetuning, the BLEU score achieves 26.27. The advanced finetuning methods provide a slight improvement on JaEn. After ensemble, we achieve 28.24 BLEU in newstest2020.
4.7 EnglishGerman
The results of EnDe on newstest2020 are shown in Table 2. After adding back-translation, we improve the BLEU score from 33.28 to 35.28. Knowledge Distillation further boosts the BLEU score to 36.58. The finetuning further brings in 2.63 improvements. After injecting the in-domain knowledge into the monolingual corpus, we get another 0.31 BLEU gain. We apply a post-processing procedure on EnDe. Specifically, we normalize the English quotations to German ones in German hypotheses, which brings in 1.3 BLEU improvements.
5 Analysis
To verify the effectiveness of our approach, we conduct analytical experiments on model variants, finetune methods, and ensemble strategies in this section.
5.1 Effects of Model Architecture
We conduct several experiments to validate the effectiveness of Transformer Vaswani et al. (2017) variants we used and list results in Table 3. We also investigate the diversity of different variants and the impacts on the model ensemble. The results is listed in Table 4 and Table 5. Here we take EnZh models as examples to conduct the diversity and ensemble experiments. The results in other directions show similar trends.
Performance.
As shown in Table 3, AAN performs slightly worse than other variants in EnZh but Mixed-AAN variants outperform normal Transformer. Weighted Attention Transformer provides noticeable improvement compare to AAN and sometimes better than vanilla Transformer.
Diversity.
The Self-BLEU scores in Table 4 demonstrate the difference between two models, more different models generally have lower scores. As we can see, AAN and all the variants with AAN have an absolutely lower Self-BLEU score with the Transformer. The Talking-Heads Attention has the minimum scores among all the variants.
Ensemble.
In our preliminary experiments, we observe that more diverse models can significantly help the model ensemble. The results are listed in Table 5. We get a more robust ensemble model with only four models using our novel variants than searching from dozens of Deeper and Wider Transformer models. Even these four models are trained with the same training data. After we combine the four models with Deeper and Wider Transformer, we can further get a significant improvement.
Take EnZh as an examble, our final submission consist of 1 Average First Transformer, 1 Average Bottom Transformer, 1 Dual Attention Transformer, 1 Weighted Attention Transformer and 1 Transformer with Post-Norm.
Models | newstest2020 |
Deeper & Wider Transformer | 50.31 |
Weighted & Mixed-AAN | 50.44 |
Ensemble with all models above | 50.62 |
5.2 Effects of Boosted Self-BLEU based Ensemble
To verify the superiority of our Boosted Self-BLEU based Ensemble (BSBE) method, we randomly select 10 models with different architecture and training data. For our submitted system, we search from over 500 models. We use a greedy search algorithm Deng et al. (2018) as our baseline. The greedy search greedily selects the best performance model into candidate ensemble models. If the selected model provides a positive improvement, we keep it in the candidates. Otherwise, it is added to a temporary model list and still has a weak chance to be reused in the future. One model from the temporary list can be reused once, after which it is withdrawn definitely. We compare the results of greedy search, BSBE and Brute Force and list the ensemble model BLEU and the number of searches in Table 6. Note that is the number of models, which is 10 here. For BSBE, we need to get the translation result of every model to calculate the Self-BLEU. After that, we only need to perform the inference process once.
5.3 Effects of Advanced Finetuning
In this section, we describe our experiments on advanced finetuning in the four translation directions. As shown in Table 7, all the advanced finetuning methods outperform normal finetuning. For EnZh, Scheldule Sampling Based on Decoding Steps with Graduated Label Smoothing improves the model performance from 49.81 to 50.20. For EnJa, Target Denoising with Graduated Label Smoothing provides the highest BLEU gain, which are 0.19 and 0.11. For the EnDe direction, Confidence-Aware Scheldule Sampling with Graduated Label Smoothing performs the best, improving from 39.21 to 39.42. These findings are in line with the conclusion of Wang and Sennrich (2020) that links exposure bias with domain shift.
Algorithm | BLEU | Number of Searches |
Greedy | 50.19 | |
Brute Force | 50.44 | |
BSBE | 50.44 |
Finetuning Approach | En-Zh | En-Ja | Ja-En | En-De |
Normal | 49.81 | 44.23 | 26.27 | 39.21 |
Graduated Label Smoothing | 49.95 | 44.32 | 26.35 | 39.32 |
+ Target Denoising | 50.09 | 44.42 | 26.38 | 39.34 |
+ Confidence-Aware Scheldule Sampling | 50.17 | 44.35 | 26.33 | 39.42 |
+ Scheldule Sampling Based on Decoding Steps | 50.20 | 44.36 | 26.33 | 39.40 |
6 Conclusion
We investigate various novel Transformer based architectures to build robust systems. Our systems are also built on several popular data augmentation methods such as back-translation, knowledge distillation and iterative in-domain knowledge transfer. We enhance our system with advanced finetuning approaches, i.e., target denoising, graduated label smoothing and confidence-aware scheduled sampling. A boosted Self-BLEU based model ensemble is also employed which plays a key role in our systems. Our constrained systems achieve 36.9, 46.9, 27.8 and 31.3 case-sensitive BLEU scores on EnglishChinese, EnglishJapanese, JapaneseEnglish and EnglishGerman, respectively. The BLEU scores of EnglishChinese, EnglishJapanese and JapaneseEnglish are the highest among all submissions, and that of EnglishGerman is the highest among all constrained submissions.
Acknowledgements
Yijin Liu and Jinan Xu have been supported by the National Key R&D Program of China (2020AAA0108001) and the National Nature Science Foundation of China (No. 61976015, 61976016, 61876198 and 61370130). The authors would like to thank the anonymous reviewers for their valuable comments and suggestions to improve this paper.
References
- Caswell et al. (2019) Isaac Caswell, Ciprian Chelba, and David Grangier. 2019. Tagged back-translation. In Proceedings of the Fourth Conference on Machine Translation (Volume 1: Research Papers), pages 53–63, Florence, Italy. Association for Computational Linguistics.
- Deng et al. (2018) Yongchao Deng, Shanbo Cheng, Jun Lu, Kai Song, Jingang Wang, Shenglan Wu, Liang Yao, Guchun Zhang, Haibo Zhang, Pei Zhang, Changfeng Zhu, and Boxing Chen. 2018. Alibaba’s neural machine translation systems for WMT18. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers, pages 368–376, Belgium, Brussels. Association for Computational Linguistics.
- Duckworth et al. (2019) Daniel Duckworth, Arvind Neelakantan, Ben Goodrich, Lukasz Kaiser, and Samy Bengio. 2019. Parallel scheduled sampling. arXiv preprint arXiv:1906.04331.
- Edunov et al. (2018) Sergey Edunov, Myle Ott, Michael Auli, and David Grangier. 2018. Understanding back-translation at scale. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 489–500, Brussels, Belgium. Association for Computational Linguistics.
- Hoang et al. (2018) Vu Cong Duy Hoang, Philipp Koehn, Gholamreza Haffari, and Trevor Cohn. 2018. Iterative back-translation for neural machine translation. In Proceedings of the 2nd Workshop on Neural Machine Translation and Generation, pages 18–24, Melbourne, Australia. Association for Computational Linguistics.
- Huang et al. (2020) Xiao Shi Huang, Felipe Pérez, Jimmy Ba, and Maksims Volkovs. 2020. Improving transformer optimization through better initialization. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 4475–4483. PMLR.
- Kim and Rush (2016) Yoon Kim and Alexander M. Rush. 2016. Sequence-level knowledge distillation. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 1317–1327, Austin, Texas. Association for Computational Linguistics.
- Li et al. (2019) Bei Li, Yinqiao Li, Chen Xu, Ye Lin, Jiqiang Liu, Hui Liu, Ziyang Wang, Yuhao Zhang, Nuo Xu, Zeyang Wang, Kai Feng, Hexuan Chen, Tengbo Liu, Yanyang Li, Qiang Wang, Tong Xiao, and Jingbo Zhu. 2019. The NiuTrans machine translation systems for WMT19. In Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1), pages 257–266, Florence, Italy. Association for Computational Linguistics.
- Liu et al. (2020) Xiaodong Liu, Kevin Duh, Liyuan Liu, and Jianfeng Gao. 2020. Very deep transformers for neural machine translation. arXiv preprint arXiv:2008.07772.
- Liu et al. (2021a) Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, and Jie Zhou. 2021a. Confidence-aware scheduled sampling for neural machine translation. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 2327–2337.
- Liu et al. (2021b) Yijin Liu, Fandong Meng, Yufeng Chen, Jinan Xu, and Jie Zhou. 2021b. Scheduled sampling based on decoding steps for neural machine translation. In Proceedings of EMNLP.
- Luong and Manning (2015) Minh-Thang Luong and Christopher D Manning. 2015. Stanford neural machine translation systems for spoken language domains. In Proceedings of the International Workshop on Spoken Language Translation, pages 76–79.
- Meng et al. (2020) Fandong Meng, Jianhao Yan, Yijin Liu, Yuan Gao, Xianfeng Zeng, Qinsong Zeng, Peng Li, Ming Chen, Jie Zhou, Sifan Liu, and Hao Zhou. 2020. WeChat neural machine translation systems for WMT20. In Proceedings of the Fifth Conference on Machine Translation, pages 239–247, Online. Association for Computational Linguistics.
- Meng and Zhang (2019) Fandong Meng and Jinchao Zhang. 2019. DTMT: A novel deep transition architecture for neural machine translation. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, pages 224–231. AAAI Press.
- Mihaylova and Martins (2019) Tsvetomila Mihaylova and André F. T. Martins. 2019. Scheduled sampling for transformers. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pages 351–356, Florence, Italy. Association for Computational Linguistics.
- Ranzato et al. (2016) Marc’Aurelio Ranzato, Sumit Chopra, Michael Auli, and Wojciech Zaremba. 2016. Sequence level training with recurrent neural networks. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings.
- Sennrich et al. (2016a) Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016a. Improving neural machine translation models with monolingual data. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 86–96, Berlin, Germany. Association for Computational Linguistics.
- Sennrich et al. (2016b) Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016b. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1715–1725, Berlin, Germany. Association for Computational Linguistics.
- Shazeer et al. (2020) Noam Shazeer, Zhenzhong Lan, Youlong Cheng, Nan Ding, and Le Hou. 2020. Talking-heads attention. arXiv preprint arXiv:2003.02436.
- Sun et al. (2019) Meng Sun, Bojian Jiang, Hao Xiong, Zhongjun He, Hua Wu, and Haifeng Wang. 2019. Baidu neural machine translation systems for WMT19. In Proceedings of the Fourth Conference on Machine Translation (Volume 2: Shared Task Papers, Day 1), pages 374–381, Florence, Italy. Association for Computational Linguistics.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pages 5998–6008.
- Wang and Sennrich (2020) Chaojun Wang and Rico Sennrich. 2020. On exposure bias, hallucination and domain shift in neural machine translation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 3544–3552, Online. Association for Computational Linguistics.
- Wang et al. (2021) Fusheng Wang, Jianhao Yan, Fandong Meng, and Jie Zhou. 2021. Selective knowledge distillation for neural machine translation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 6456–6466, Online.
- Wang et al. (2020) Shuo Wang, Zhaopeng Tu, Shuming Shi, and Yang Liu. 2020. On the inference calibration of neural machine translation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 3070–3079, Online. Association for Computational Linguistics.
- Xiong et al. (2020) Ruibin Xiong, Yunchang Yang, Di He, Kai Zheng, Shuxin Zheng, Chen Xing, Huishuai Zhang, Yanyan Lan, Liwei Wang, and Tie-Yan Liu. 2020. On layer normalization in the transformer architecture. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 10524–10533. PMLR.
- Yan et al. (2020) Jianhao Yan, Fandong Meng, and Jie Zhou. 2020. Multi-unit transformers for neural machine translation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1047–1059, Online. Association for Computational Linguistics.
- Zhang et al. (2018) Biao Zhang, Deyi Xiong, and Jinsong Su. 2018. Accelerating neural transformer via an average attention network. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1789–1798, Melbourne, Australia. Association for Computational Linguistics.
- Zhang et al. (2019) Wen Zhang, Yang Feng, Fandong Meng, Di You, and Qun Liu. 2019. Bridging the gap between training and inference for neural machine translation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 4334–4343, Florence, Italy. Association for Computational Linguistics.
- Zhu et al. (2018) Yaoming Zhu, Sidi Lu, Lei Zheng, Jiaxian Guo, Weinan Zhang, Jun Wang, and Yong Yu. 2018. Texygen: A benchmarking platform for text generation models. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR 2018, Ann Arbor, MI, USA, July 08-12, 2018, pages 1097–1100. ACM.