LazyFormer: Self Attention with Lazy Update
Abstract
Improving the efficiency of Transformer-based language pre-training is an important task in NLP, especially for the self-attention module, which is computationally expensive. In this paper, we propose a simple but effective solution, called LazyFormer, which computes the self-attention distribution infrequently. LazyFormer composes of multiple lazy blocks, each of which contains multiple Transformer layers. In each lazy block, the self-attention distribution is only computed once in the first layer and then is reused in all upper layers. In this way, the cost of computation could be largely saved. We also provide several training tricks for LazyFormer. Extensive experiments demonstrate the effectiveness of the proposed method.
1 Introduction
Using pre-trained contextual representations (e.g., BERT) Devlin et al. (2018) have become the standard way to improve the performance on the downstream tasks in natural language processing. Transformer Vaswani et al. (2017) is the basic building block for almost all pre-training methods Liu et al. (2019); Clark et al. (2019b); Devlin et al. (2018). A Transformer layer is composed of an efficient densely connected network operated on each position separately and a less-efficient self-attention module, which costs ( is sequence length) time and space. This quadratic cost becomes a bottleneck in Transformer, especially when is large. Many recent works Wang et al. (2020); Beltagy et al. (2020); Zaheer et al. (2020); Kitaev et al. (2020); Choromanski et al. (2020) tried to reduce the cost to or , by sparsifying or approximating the attention matrix.

In this paper, different from previous works, we explore a simple idea to improve the model efficiency by computing the self-attention infrequently. We call it LazyFormer. More specifically, a LazyFormer consists of multiple basic blocks, and each basic block is composed of Transformer layers, as shown in Figure 1. In each block, we only compute Query-Key dot-product attention once in the first layer and then reuse it in the upper layers. In this way, LazyFormer only needs to calculate the attention once in Transformer layers, reducing the computational cost from to .
We conduct extensive experiments to verify the efficiency and effectiveness of LazyFormer in language pre-training. From the results, we observe: 1) compared with a standard Transformer model of the same model capacity, LazyFormer is 1.3x faster without hurting any performance. 2) LazyFormer allows us to train the larger models effectively. Specifically, with the same pre-training cost, the larger LazyFormer can outperform the baseline by 1 point. Besides, it can achieve a better GLUE score by only using 50% costs. 3) LazyFormer can better handle longer sequences. With increasing, the speed-up is more significant.
Related work
LazyFormer is inspired by several recent works which investigate what self-attention module learns in each layer Clark et al. (2019a); Vig and Yonatan (2019); Vig (2019); Xiao et al. (2019). From these previous works, we can easily see that the distributions of self-attention outputs are similar between the adjacent Transformer layers. Such observations motivate us to reuse the self-attention outputs in the lower layers to the upper layers.
There are some works that leverage such observation for better training, like Gong et al. (2019) and Lan et al. (2019). However, these works share the parameters of self-attention in different layers but still need to compute self-attention in all layers. Therefore they cannot save the computational costs. Our work is also orthogonal to the works that modify the self-attention modules to reduce the cost, such as ReformerKitaev et al. (2020) and LinformerWang et al. (2020). These works aim to reduce the computation in each layer, while ours is to reuse the self-attention outputs from lower layers. Both works could be combined and further reduce the cost of language pre-training.
2 LazyFormer
The attention module (Vaswani et al., 2017) in Transformer can be generally formulated as querying a dictionary with key-value pairs, e.g., , where is the dimensionality of the hidden representations. In the self-attention module, (Query), (Key) and (Value) are parameterized from the input , i.e., , where , and are the projection matrices. It is easy to see that the computational complexity of is , since it has to compute the pair-wise correlations for all input tokens, and produces an matrix (attention matrix). When we use stacked Transformer layers, as all attention matrices need to be computed, the total cost of self-attention calculation is .
Many recent works Vig and Yonatan (2019); Vig (2019) show the attention matrices are similar in different layers, especially in the adjacent layers. Therefore, we argue that the attention matrix maybe does not need to be computed in every layer. For example, we can only calculate the attentions in a layer and reuse it for multiple adjacent upper layers. Formally, we define a new Transformer variant called LazyFormer, which composes of several “lazy” blocks. Each lazy block consists of Transformer layers, as shown in Figure 1. In each block, the attention matrix will be only computed in the first layer using the input to the block, and then reused by all upper layers. We can stack lazy blocks to construct a -layer Transformer model. In this way, the cost of computing the attention could be reduced from to .
Based on the design of lazy block, we further use the following two additional methods to improve LazyFormer:
Wider Layers.
In LazyFormer, the number of projection matrices and are reduced. Therefore, the total parameters in LazyFormer are less than the stacked Transformers when using the same width and depth. For example, the BERT-based model has 12 layers. The embedding dimension is set to 768, and the hidden dimension is set to 3072. This configuration leads to a model size of about 110M parameters. If we use for LazyFormer in the same setting, the model only contains about 100M parameters.
As the model capacity plays an important role in language pre-training Raffel et al. (2019), we can slightly increase the hidden/embedding dimension in LazyFormer, to match the same number of parameters. Note that increasing the hidden/embedding dimension only slightly affects the efficiency. Wider LazyFormer is still much faster than a Transformer of the same depth. Besides, we can even increase the width of LazyFormer until its forward/backward speed match the Transformer to achieve better performance when using the same pre-training cost.
Remove dropout in self-attention.
Dropout is used in self-attention by default. The cost of that dropout is also , as it is applied on the attention matrix. Recent work Lan et al. (2019) shows the dropout in self-attention can be safely removed, without hurting the performance. Therefore, for better efficiency, we also remove the dropout in self-attention in LazyFormer.
3 Experiment
To verify the performance of the proposed LazyFormer, we conduct extensive experiments and demonstrate the results in this section. We use BERT-Base (112M parameters) architecture for all experiments. Specifically, BERT-Base is consists of 12 Transformer layers, in which the embedding dimension is 768, the number of attention heads is 12, and the hidden dimension is 3072. Besides absolute positional encoding, we further use the relative positional encoding Raffel et al. (2019) in the self-attention module for better performance. We provide all the experimental details and results in the Appendix.
3.1 Experimental Design
Following Devlin et al. (2018), we use the 16GB corpus (English Wikipedia corpus and BookCorpus (Zhu et al., 2015)) for pre-training. We set the vocabulary size (sub-word tokens) as 32,768, sequence length as 512, and batch size as 256. We use the GLUE (General Language Understanding Evaluation) dataset (Wang et al., 2018) as the downstream tasks to evaluate the performance of the pre-trained models. All codes are implemented based on fairseq (Ott et al., 2019) in PyTorch (Paszke et al., 2017). All models are run on 8 NVIDIA Tesla V100 GPUs with mixed-precision (Micikevicius et al., 2017).
We use Mx to denote LazyFormer structure, where is the number of layers in each lazy block, and is number of total blocks. For example, M2x6 denotes the LazyFormer with 6 blocks, each with 2 Transformer layers.
3.2 Overall Comparison
Steps | MNLI-m/mm | QNLI | QQP | SST | CoLA | MRPC | RTE | STS | Avg. | |
---|---|---|---|---|---|---|---|---|---|---|
BERT | 1 | 85.72/85.67 | 92.07 | 91.21 | 92.78 | 58.80 | 88.73 | 67.51 | 89.23 | 83.52 |
M2x6-S | 1 | 85.69/85.46 | 91.14 | 91.13 | 92.66 | 57.86 | 87.25 | 72.20 | 89.79 | 83.69 |
M2x6-SD | 1 | 85.34/85.50 | 91.65 | 91.22 | 93.23 | 57.85 | 86.76 | 70.40 | 89.69 | 83.52 |
M2x6 | 1 | 86.34/86.31 | 91.82 | 91.38 | 93.00 | 60.77 | 87.75 | 73.65 | 90.00 | 84.56 |
M2x6mid | 500 | 85.81/85.58 | 91.62 | 91.31 | 92.66 | 58.81 | 87.75 | 71.48 | 88.90 | 83.77 |



Params | (,,) | Time 111Time indicates the training wall time of 100 iterations by 8 V100 GPUs. | Speedup | |
---|---|---|---|---|
BERT | 112 | (3072,768,12) | 45 | 1.0x |
M2x6-S | 112 | (3456,768,12) | 35 | 1.3x |
M2x6-SD | 112 | (3456,768,12) | 38 | 1.2x |
M2x6 | 157 | (4480,896,14) | 45 | 1.0x |
Time \footreffootnote1 | GLUE-Avg. | ||
BERT (M1x12-SD) | 3072 | 45 | 83.52 |
M2x6-SD | 3456 | 38 | 83.52 |
M3x4-SD | 3584 | 35 | 82.94 |
M4x3-SD | 3648 | 34 | 83.22 |
M6x2-SD | 3712 | 33 | 82.47 |
M5M3M2M2-SD | 3584 | 35 | 82.89 |
M2M2M3M5-SD | 3584 | 35 | 82.82 |
First, we set up models for the overall comparison. Besides the baseline BERT model, we set up two LazyFormer variants: 1) M2x6-S, which uses six lazy blocks with two Transformer layers in each block, and increases hidden dimension to 3456, to retain the same parameter size as BERT; 2) M2x6, which increases hidden dimension to 4480, embedding dimension to 896, and the number of attention heads to 14, to retain the same pre-training cost as BERT. Both M2x6 and M2x6-S remove dropout in the self-attention module. The detailed settings can be found in Table 2.
The results are shown in Table 6. Firstly, M2x6-S achieves a slight improvement over BERT but is about 1.3x faster. This result indicates that LazyFormer is much more efficient, without hurting any performance.
Furthermore, M2x6 outperforms baselines by 1 point in terms of GLUE average score and is consistently better on almost all GLUE tasks. This demonstrates another important strength of LazyFormer: it allows us to increase the model capacity to achieve better performance while using the same computational cost. Besides, from Table 2, we can also see that the results of M2x6mid (intermediate 500k-step checkpoint of M2x6) are already competitive to that of BERT (trained for 1M steps). This suggests LazyFormer can use a significantly short time to learn a better model. As shown in Figure 2, both M2x6 and M2x6-S converge much faster than BERT in terms of both pre-training validation loss and down-stream task performance.
3.3 Ablation Studies
Remove dropout in self-attention.
As aforementioned, dropout in self-attention brings additional costs. So we empirically study whether using dropout is essential. First, as shown in Table 6, compared M2x6-S with M2x6-SD (M2x6-S with dropout in self-attention), removing dropout in self-attention slightly improves the performance. Besides, as shown in Table 2, removing dropout in self-attention can bring the 10% speed-up. In short, removing dropout in self-attention can improve efficiency without hurting the performance.
Different layouts.
There are many design choices in LazyFormer. For example, one can set low-level blocks with more/fewer layers or set the whole model with more/fewer blocks. We study how different layouts perform and summarize the results in Table 3. First, we find that when the number of blocks decreases, the training is faster, but the acceleration rate is not significant. At the same time, the final performance gets worse. Therefore, we observe using six blocks is a trade-off choice for both efficiency and effectiveness. Second, we find the number of blocks is the key factor for the performance. For all the 4-block models, setting each block with 3 layers (M3x4-SD) achieves similar performance to the models that set different numbers of layers in different blocks (M5M3M2M2-SD 222M5M3M2M2 denotes the model with 4 sequential lazy blocks, each with 5, 3, 2, 2 layers respectively., M2M2M3M5-SD).
Params | (,,) | MNLI-m/mm | GLUE-Avg. | |
---|---|---|---|---|
M2x6-S | 112 | (3456,768,12) | 85.72/85.67 | 83.69 |
M2x6 | 157 | (4480,896,14) | 86.34/86.31 | 84.56 |
M2x6 | 161 | (6144,768,12) | 86.08/86.57 | 84.77 |
M2x7 | 148 | (4480,768,12) | 86.25/86.26 | 84.75 |

Different model settings under the same computational cost.
As aforementioned, LazyFormer allows us to use a larger model capacity within the same computational cost. Therefore, we investigate different settings to increase model parameters and summarize the results in Table 4. Specifically, we study the performance when we increase hidden dimension , embedding dimension , the number of attention heads , or stack more blocks. From the results, we find all these settings can achieve much better performance than the baseline.
Speedup for longer sequences.
If considering the cost from densely connected layer, LazyFormer reduces the cost from to , where is the cost of densely connected layer. Therefore, with increasing, the cost is dominated by the term, and the speedup brought by LazyFormer will be also larger. To study this, we provide speedup ratios given different , shown in Figure 3. It is easy to find the LazyFormer can bring almost 2x speed-up for long sequences.
4 Conclusion
We propose LazyFormer, which lazily updates the self-attention module and lead to an efficient model architecture of the Transformer. Specifically, LazyFormer composes of multiple lazy blocks, each of which contains multiple Transformer layers. In each lazy block, the self-attention distribution is only computed once in the first layer and then is reused in all upper layers. Besides, LazyFormer further removes dropout in the self-attention module for better efficiency and increases model capacity for better performance. Extensive experimental results demonstrate both the efficiency and effectiveness of LazyFormer.
References
- Beltagy et al. (2020) Iz Beltagy, Matthew E Peters, and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.
- Choromanski et al. (2020) Krzysztof Choromanski, Valerii Likhosherstov, David Dohan, Xingyou Song, Andreea Gane, Tamas Sarlos, Peter Hawkins, Jared Davis, Afroz Mohiuddin, Lukasz Kaiser, et al. 2020. Rethinking attention with performers. arXiv preprint arXiv:2009.14794.
- Clark et al. (2019a) Kevin Clark, Urvashi Khandelwal, Omer Levy, and Christopher D Manning. 2019a. What does bert look at? an analysis of bert’s attention. ACL 2019, page 276.
- Clark et al. (2019b) Kevin Clark, Minh-Thang Luong, Quoc V Le, and Christopher D Manning. 2019b. Electra: Pre-training text encoders as discriminators rather than generators. In International Conference on Learning Representations.
- Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
- Gong et al. (2019) Linyuan Gong, Di He, Zhuohan Li, Tao Qin, Liwei Wang, and Tieyan Liu. 2019. Efficient training of bert by progressively stacking. In International Conference on Machine Learning, pages 2337–2346.
- Kingma and Ba (2014) Diederik P. Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. CoRR, abs/1412.6980.
- Kitaev et al. (2020) Nikita Kitaev, Łukasz Kaiser, and Anselm Levskaya. 2020. Reformer: The efficient transformer. arXiv preprint arXiv:2001.04451.
- Koehn et al. (2007) Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris Callison-Burch, Marcello Federico, Nicola Bertoldi, Brooke Cowan, Wade Shen, Christine Moran, Richard Zens, Chris Dyer, Ondrej Bojar, Alexandra Constantin, and Evan Herbst. 2007. Moses: Open source toolkit for statistical machine translation. In ACL.
- Lan et al. (2019) Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. 2019. Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942.
- Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- Micikevicius et al. (2017) Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, et al. 2017. Mixed precision training. arXiv preprint arXiv:1710.03740.
- Ott et al. (2019) Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, and Michael Auli. 2019. fairseq: A fast, extensible toolkit for sequence modeling. arXiv preprint arXiv:1904.01038.
- Paszke et al. (2017) Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. 2017. Automatic differentiation in pytorch.
- Raffel et al. (2019) Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. 2019. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683.
- Sennrich et al. (2015) Rico Sennrich, Barry Haddow, and Alexandra Birch. 2015. Neural machine translation of rare words with subword units. CoRR, abs/1508.07909.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, pages 6000–6010.
- Vig (2019) Jesse Vig. 2019. A multiscale visualization of attention in the transformer model. arXiv preprint arXiv:1906.05714.
- Vig and Yonatan (2019) Jesse Vig and Belinkov Yonatan. 2019. Analyzing the structure of attention in a transformer language model. arXiv preprint arXiv:1906.04284.
- Wang et al. (2018) Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. 2018. GLUE: A multi-task benchmark and analysis platform for natural language understanding. CoRR, abs/1804.07461.
- Wang et al. (2020) Sinong Wang, Belinda Li, Madian Khabsa, Han Fang, and Hao Ma. 2020. Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768.
- Xiao et al. (2019) Tong Xiao, Yinqiao Li, Jingbo Zhu, Zhengtao Yu, and Tongran Liu. 2019. Sharing attention weights for fast transformer. arXiv preprint arXiv:1906.11024.
- Zaheer et al. (2020) Manzil Zaheer, Guru Guruganesh, Kumar Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Ontanon, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, et al. 2020. Big bird: Transformers for longer sequences. Advances in Neural Information Processing Systems, 33.
- Zhu et al. (2015) Yukun Zhu, Ryan Kiros, Richard Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, and Sanja Fidler. 2015. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books. In arXiv preprint arXiv:1506.06724.
Appendix A Experimental Details
Pre-training.
Following BERT (Devlin et al., 2018), we use both English Wikipedia corpus and BookCorpus (Zhu et al., 2015) for language pre-training. By concatenating these two datasets, we can obtain a corpus with 16GB raw text. We adopt the following consecutive pre-processing steps: segmenting documents into sentences by Spacy333https://spacy.io, normalizing, lower-casing, and tokenizing the texts by Moses decoder (Koehn et al., 2007), and finally, applying byte pair encoding (BPE) (Sennrich et al., 2015) setting the vocabulary size to 32,768.
We found the data cleaning is important for language pre-training. To this end, we de-duplicate the documents, normalize the punctuations, concatenate the short sequences, replace the URL and other hyperlinks to special tokens, and filter the low-frequency tokens. Therefore, our re-implemented baselines, like BERT, can achieve a higher average GLUE scores than the original papers.
We use masked language modeling as the objective of pre-training. We remove the next sentence prediction task and use FULL-SENTENCES mode to pack sentences as suggested in RoBERTa (Liu et al., 2019). We train the models for 1000 steps where the batch size is 256 and the maximum sequence length is 512. The masked probability is set to 0.15, with replacing 80% of the masked positions by [MASK], 10% by randomly sampled words, and keep the remaining 10% unchanged. We use Adam (Kingma and Ba, 2014) as the optimizer, and set the its hyperparameter to 1e-6 and to (0.9, 0.999). The peak learning rate is set to 1e-4 with a 10-step warm-up stage. After the warm-up stage, the learning rate decays linearly to zero. We set the dropout probability to 0.1, gradient clip norm to 1.0, and weight decay to 0.01. Besides the final checkpoint, we also save intermediate checkpoints and fine-tune them on downstream tasks, to check the efficiency of different methods.
Fine-tuning.
We use the GLUE (General Language Understanding Evaluation) dataset (Wang et al., 2018) as the downstream tasks to evaluate the performance of the pre-trained models. Specifically, we use nine tasks in GLUE, including CoLA, RTE, MRPC, STS-B, SST, QNLI, QQP, and MNLI-m/mm. For the evaluation metrics, we report Matthews correlation for CoLA, Pearson correlation for STS-B, and accuracy for other tasks. We use the same optimizer (Adam) with the same hyperparameters as in pre-training. Following previous works, we search the learning rates during the fine-tuning for each downstream task. The setting details are listed in Table 5. For a fair comparison, we do not apply any tricks for fine-tuning. Each configuration will be run five times with different random seeds, and the median of these five results on the development set will be used as the performance of one configuration. We will ultimately report the best number over all configurations. We also provide all the detailed results in Table 6.
Pre-training | Fine-tuning | |
Max Steps | 1 | - |
Max Epochs | - | 10 |
Learning Rate | 1e-4 | {2e-5, 3e-5, 4e-5, 5e-5} |
Batch Size | 256 | 32 |
Warm-up Ratio | 0.01 | 0.06 |
Sequence Length | 512 | 512 |
Learning Rate Decay | Linear | Linear |
Adam | 1e-6 | 1e-6 |
Adam (, ) | (0.9, 0.999) | (0.9, 0.999) |
Clip Norm | 1.0 | 1.0 |
Dropout | 0.1 | 0.1 |
Weight Decay | 0.01 | 0.01 |
Params | (,,) | Pre-training Time | Speedup Ratio | Steps | MNLI-m/mm | QNLI | QQP | SST | CoLA | MRPC | RTE | STS | Avg. | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
BERT (M1x12-SD) | 112 | (3072,768,12) | 45 | 1.0x | 100 | 82.31/82.51 | 89.93 | 90.76 | 90.60 | 51.79 | 86.03 | 64.26 | 88.03 | 80.69 |
300 | 84.62/84.70 | 91.43 | 91.09 | 91.86 | 54.69 | 87.50 | 64.62 | 88.26 | 82.09 | |||||
600 | 85.42/85.25 | 91.69 | 91.18 | 92.55 | 55.48 | 88.73 | 65.70 | 88.96 | 82.77 | |||||
1 | 85.72/85.67 | 92.07 | 91.21 | 92.78 | 58.80 | 88.73 | 67.51 | 89.23 | 83.52 | |||||
M2x6-S | 112 | (3456,768,12) | 35 | 1.3x | 100 | 81.84/82.35 | 89.15 | 90.68 | 90.37 | 46.05 | 84.31 | 63.54 | 87.91 | 79.58 |
300 | 84.28/84.05 | 90.77 | 91.03 | 91.97 | 53.18 | 86.52 | 68.23 | 89.18 | 82.14 | |||||
600 | 85.27/85.04 | 91.14 | 91.18 | 92.43 | 57.28 | 87.75 | 70.04 | 89.60 | 83.30 | |||||
1 | 85.69/85.46 | 91.14 | 91.13 | 92.66 | 57.86 | 87.25 | 72.20 | 89.79 | 83.69 | |||||
M2x6-SD | 112 | (3456,768,12) | 38 | 1.2x | 100 | 82.23/82.39 | 89.58 | 90.76 | 90.94 | 51.50 | 84.80 | 64.62 | 87.81 | 80.52 |
300 | 84.74/84.59 | 91.27 | 91.15 | 91.74 | 56.83 | 87.99 | 64.62 | 88.37 | 82.37 | |||||
600 | 85.30/85.25 | 91.49 | 91.22 | 93.00 | 58.80 | 87.75 | 67.87 | 89.19 | 83.32 | |||||
1 | 85.34/85.50 | 91.65 | 91.22 | 93.23 | 57.85 | 86.76 | 70.40 | 89.69 | 83.52 | |||||
M2x6 | 157 | (4480,896,14) | 45 | 1.0x | 100 | 82.85/82.92 | 89.64 | 90.86 | 90.48 | 47.48 | 86.52 | 66.07 | 88.49 | 80.59 |
300 | 85.12/84.94 | 91.12 | 91.23 | 92.09 | 56.63 | 87.75 | 69.68 | 88.89 | 83.05 | |||||
600 | 86.02/85.62 | 91.78 | 91.37 | 92.43 | 59.83 | 87.50 | 72.92 | 89.42 | 84.10 | |||||
1 | 86.34/86.31 | 91.82 | 91.38 | 93.00 | 60.77 | 87.75 | 73.65 | 90.00 | 84.56 | |||||
M3x4-SD | 112 | (3584,768,12) | 35 | 1.3x | 100 | 82.24/82.45 | 89.11 | 90.79 | 90.71 | 48.34 | 85.78 | 66.43 | 88.63 | 80.50 |
300 | 84.26/84.49 | 90.26 | 91.06 | 92.20 | 51.03 | 87.01 | 66.43 | 88.66 | 81.71 | |||||
600 | 85.07/85.18 | 90.54 | 91.14 | 91.74 | 52.87 | 87.99 | 70.40 | 89.13 | 82.67 | |||||
1 | 85.55/85.39 | 91.09 | 91.14 | 92.43 | 54.80 | 87.01 | 69.31 | 89.71 | 82.94 | |||||
M4x3-SD | 112 | (3648,768,12) | 34 | 1.3x | 100 | 82.44/82.82 | 88.82 | 90.86 | 90.71 | 48.84 | 84.80 | 66.07 | 87.80 | 80.35 |
300 | 84.47/84.48 | 90.32 | 91.23 | 92.32 | 54.23 | 87.25 | 67.15 | 88.16 | 82.18 | |||||
600 | 85.07/85.03 | 90.81 | 91.32 | 92.43 | 55.04 | 88.48 | 68.95 | 88.75 | 82.88 | |||||
1 | 85.48/85.31 | 91.01 | 91.39 | 93.46 | 55.76 | 88.24 | 68.95 | 89.35 | 83.22 | |||||
M6x2-SD | 112 | (3712,768,12) | 33 | 1.4x | 100 | 82.24/82.55 | 88.85 | 90.84 | 90.94 | 50.51 | 85.54 | 62.09 | 85.54 | 79.90 |
300 | 83.72/83.93 | 90.35 | 91.14 | 92.32 | 52.34 | 87.75 | 64.26 | 86.30 | 81.34 | |||||
600 | 84.71/84.68 | 90.68 | 91.23 | 92.32 | 54.69 | 87.01 | 67.51 | 86.72 | 82.17 | |||||
1 | 84.93/85.04 | 90.98 | 91.34 | 92.55 | 55.27 | 87.25 | 67.87 | 86.98 | 82.47 | |||||
M5M3M2M2-SD | 112 | (3584,768,12) | 35 | 1.3x | 100 | 82.36/82.87 | 89.11 | 90.86 | 91.06 | 47.62 | 85.54 | 68.59 | 88.37 | 80.71 |
300 | 84.13/84.20 | 90.30 | 91.15 | 92.09 | 51.87 | 87.99 | 66.07 | 88.17 | 81.77 | |||||
600 | 85.07/84.96 | 90.85 | 91.35 | 92.32 | 53.64 | 89.22 | 68.23 | 88.56 | 82.69 | |||||
1 | 85.23/85.03 | 91.09 | 91.29 | 92.55 | 52.36 | 88.24 | 71.12 | 89.10 | 82.89 | |||||
M2M2M3M5-SD | 112 | (3584,768,12) | 35 | 1.3x | 100 | 82.14/82.76 | 88.82 | 90.74 | 90.94 | 49.63 | 87.25 | 63.54 | 87.80 | 80.40 |
300 | 84.30/84.47 | 90.44 | 91.07 | 92.32 | 51.16 | 86.76 | 65.70 | 88.44 | 81.63 | |||||
600 | 85.09/85.25 | 90.66 | 91.15 | 92.89 | 53.41 | 87.99 | 68.59 | 89.02 | 82.67 | |||||
1 | 85.47/85.54 | 90.83 | 91.13 | 92.89 | 53.48 | 88.24 | 68.23 | 89.54 | 82.82 | |||||
M2x6 | 161 | (6144,768,12) | 45 | 1.0x | 100 | 82.76/82.96 | 89.91 | 90.66 | 90.83 | 50.49 | 85.29 | 64.26 | 87.45 | 80.51 |
300 | 85.11/85.14 | 91.12 | 91.14 | 92.09 | 55.21 | 86.76 | 70.04 | 89.12 | 82.86 | |||||
600 | 85.96/85.93 | 91.78 | 91.32 | 93.12 | 58.80 | 87.99 | 70.76 | 89.67 | 83.93 | |||||
1 | 86.08/86.57 | 91.76 | 91.39 | 92.89 | 60.65 | 87.99 | 75.45 | 90.18 | 84.77 | |||||
M2x7 | 148 | (4480,768,12) | 46 | 1.0x | 100 | 82.20/82.46 | 89.36 | 90.63 | 90.94 | 48.19 | 85.54 | 66.43 | 88.04 | 80.42 |
300 | 85.23/84.86 | 90.94 | 91.03 | 92.20 | 56.12 | 87.01 | 68.23 | 89.20 | 82.76 | |||||
600 | 85.89/85.94 | 91.31 | 91.30 | 92.20 | 57.01 | 88.73 | 71.84 | 89.66 | 83.76 | |||||
1 | 86.25/86.26 | 91.71 | 91.28 | 92.55 | 61.34 | 88.24 | 75.09 | 90.02 | 84.75 |