From Two to One: A New Scene Text Recognizer with
Visual Language Modeling Network
Abstract
In this paper, we abandon the dominant complex language model and rethink the linguistic learning process in the scene text recognition. Different from previous methods considering the visual and linguistic information in two separate structures, we propose a Visual Language Modeling Network (VisionLAN), which views the visual and linguistic information as a union by directly enduing the vision model with language capability. Specially, we introduce the text recognition of character-wise occluded feature maps in the training stage. Such operation guides the vision model to use not only the visual texture of characters, but also the linguistic information in visual context for recognition when the visual cues are confused (e.g. occlusion, noise, etc.). As the linguistic information is acquired along with visual features without the need of extra language model, VisionLAN significantly improves the speed by 39% and adaptively considers the linguistic information to enhance the visual features for accurate recognition. Furthermore, an Occlusion Scene Text (OST) dataset is proposed to evaluate the performance on the case of missing character-wise visual cues. The state of-the-art results on several benchmarks prove our effectiveness. Code and dataset are available at https://github.com/wangyuxin87/VisionLAN.
1 Introduction

As a fundamental and pivotal task, scene text recognition (STR) aiming to read the text content from natural images has attracted great interest in computer vision [15, 31, 32, 42, 46]. By taking the text image as input and textual prediction as output, some early methods regard the text recognition as a symbol classification task [31, 19]. However, it is hard to recognize images with confused visual cues (e.g. occlusion, noise, etc.), which are beyond visual discrimination. As the scene text image contains two-level contents: visual texture and linguistic information, inspired by the Natural Language Processing (NLP) methods [23, 5], recent STR works have shifted their research focus to acquiring linguistic information to assist recognition [47, 46, 28, 45]. Thus, the two-step architecture of vision and language models (top left of Fig. 1) is popular in recent methods. Specifically, the vision model only focuses on visual texture of characters without considering the linguistic information. Then, the language model predicts the relationship between characters through the linguistic learning structure (RNN [32], CNN [7] and Transformer [45]).
Though these methods achieve promising results, there are still two problems: 1) the extra huge computation cost. The computation cost of language model increases significantly with the word length getting longer (linear growth for RNN [32]/ CNN [7] and quadratic growth for Transformer [45] in Fig. 1). Furthermore, many methods adopt a deep bi-directional reasoning architecture [38, 45, 32] to capture more robust linguistic information, which further doubles the computation burden and greatly limits their efficiency in the real application. 2) The difficulty of aggregating two independent information. It is difficult to comprehensively consider and effectively fuse the visual and linguistic information from two separate structures for accurate recognition [7, 46]. In this paper, we attribute these two problems to the lack of language ability in the vision model, which only focuses on the visual texture of characters without initiatively learning linguistic information [45]. As shown in bottom of Fig. 1, inspired by the human cognitive process that the language capability can be acquired [21, 11], we use vision model as the basic network, and guide it to reason the occluded character during the training stage. Thus, vision model is trained to initiatively learn linguistic information in the visual context. In the test stage, vision model adaptively considers the linguistic information in the visual space for feature enhancement when the visual cues are confused (e.g. occlusion, noise, etc.), which effectively supplements the features of occluded characters, and correctly highlights the discriminating visual cues of confused characters (shown in Fig. 5). To the best of our knowledge, this is the first work to give vision model the ability to perceive language in scene text recognition. We call this new simple architecture as Visual Language Modeling Network (VisionLAN).

The pipeline of VisionLAN is shown in Fig. 2. VisionLAN contains three parts: backbone network, Masked Language-aware Module (MLM) and Visual Reasoning Module (VRM). In the training stage, visual features V are firstly extracted from the backbone network. Then MLM takes the visual features V and character index P as inputs, and automatically generates the character mask map at corresponding position through a Weakly-supervised Complementary Learning. MLM aims to simulate the case of missing character-wise visual cues by occluding visual messages in V with . In order to consider the linguistic information during the visual texture modeling, we propose a VRM with the ability to capture long-range dependencies in the visual space. VRM takes the occluded feature map as input, and is guided to make the word-level prediction. In the test stage, we remove the MLM and only use VRM for recognition. As the linguistic information is acquired along with visual features without the need of extra language model, VisionLAN introduces ZERO computation cost for capturing linguistic information (top right in Fig. 1) and significantly improves the speed by 39% (Sec. 4.4). Compared with previous methods, the proposed VisionLAN obtains more robust performance on the occluded and low-quality images, and achieves new state-of-the-art results on several benchmarks with a concise pipeline. In addition, an Occlusion Scene Text (OST) dataset is proposed to evaluate the performance on the case of missing character-wise visual cues.
The main contributions of this paper are as follows: 1) A new simple architecture is proposed for scene text recognition. We further visualize the feature maps to illustrate how VisionLAN initiatively uses linguistic information to handle the confused visual cues (e.g. occluded, noise, etc.). 2) We propose a Weakly-supervised Complementary Learning to generate accurate character-wise mask map in MLM with only word-level annotations. 3) A new Occlusion Scene Text (OST) dataset is proposed to evaluate the recognition performance of occluded images. Compared with previous methods, VisionLAN achieves the state-of-the-art performance on seven benchmarks (irregular and regular) and OST with a concise pipeline.
2 Related Work
2.1 Scene Text Recognition
Scene text recognition (STR) has been a long-term research topic in computer vision [42, 47, 7]. With deep learning becoming the most promising machine learning tool [35, 40, 41, 8, 17, 20], significant progress has been made in the past few years for STR research [28, 24]. In this section, we divide these methods into two categories according to whether linguistic rules are used, namely language-free methods and language-aware methods.
Language-free methods [42, 48, 31, 19] view STR as a visual classification task and mainly rely on the visual information for prediction. CRNN [31] extracts sequential visual features through combined CNN and RNN, then a Connectionist Temporal Classication (CTC) [9] decoder is used to maximize the probability of all the paths for final prediction. Patel et al. [27] automatically generate the custom lexicon for an image to greatly boost the performance of text reading systems. Zhang et al. [48] regard text recognition as a visual matching task. They calculate the similarity map between visual features of input image and the pre-defined alphabet to predict the text sequence. Liao et al. [18] regard the text recognition as a pixel-wise classification task. Similarly, Textscanner [36] further proposes an order map to ensure a more accurate transcription from characters to the word. In general, the language-free methods ignore linguistic rules in the recognition process, which usually fail to recognize images with confused visual cues (e.g. blur, occlusion, etc.).
Language-aware methods [16, 4, 47, 43] try to leverage linguistic rules to assist the recognition process. Lee et al. [15] use RNNs to automatically learn the sequential dynamics in word strings without manually defining N-grams. Aster [32] firstly uses a rectification module before recognition, and then adopts RNNs to model the linguistic information by using the character predicted from the last time step. However, such serial and time-dependent operation in RNN limits the computation efficiency and the performance of semantic reasoning [45]. Thus, SRN [45] proposes a global semantic reasoning module based on transformer units [35] for pure language modeling, which takes the prediction of vision model as input and predicts the relationships among characters to refine the recognition results. Fang et al. [7] design a completely CNN-based architecture for both vision and language modeling. Though these methods achieve promising results on scene text recognition task, the additionally introduced language model will significantly increase the computation cost. Furthermore, it is also difficult to comprehensively consider and effectively fuse the independent visual and linguistic information in the two-step architecture for accurate recognition [7, 46]. Different from previous methods considering the visual and linguistic information in two separate structures, we directly endue the vision model with language ability and propose a VisionLAN to view the two information as a union. Thus, it is possible to enhance the confused visual cues by capturing linguistic information in the visual context.
2.2 Masking and Prediction.
BERT [5] introduces a cloze task to mask the tokens of input sentence, which is used to learn a robust bi-directional representation based on the context. Following [5], some works use a similar concept to handle the vision-and-language task [34, 1, 22]. ViLBERT [22] uses a two stream model to process visual and textual inputs, and pre-trains their model through the two proxy tasks. Su et al. [34] propose a general structure to fit for most visual-linguistic downstream tasks, which takes both visual and linguistic features as input. As the STR datasets are weakly labeled with word-level annotations, it is difficult to directly implement these masking approaches in STR task. Different from these methods that mask in token or image patch level, in this paper, we propose a Weakly-supervised Complementary Learning to automatically mask the input image in the feature level. Thus, VisionLAN learns linguistic information from a new perspective by guiding the model to make word-level prediction on the case of missing character-wise visual cues.
3 Proposed Method
The VisionLAN is an end-to-end trainable framework with three parts containing: backbone network, Masked Language-aware Module (MLM) and Visual Reasoning Module (VRM). In this section, we first detail the pipeline of proposed method in Sec. 3.1, and then we introduce MLM and VRM in Sec. 3.2 and Sec. 3.3 respectively.
3.1 Pipeline
The pipeline of VisionLAN is shown in Fig. 2. In the training stage, given an input image, the 2D features V are firstly extracted from backbone network. Then, MLM takes extracted features and character index as inputs, and generates the position-aware character mask map through the Weakly-Supervised Complementary Learning. is used to occlude the character-wise visual messages in V to simulate the case of missing character-wise visual semantics. After that, VRM takes occluded feature map as input and makes prediction under the complete word-level supervision. In the testing stage, we remove MLM and only use VRM for prediction.
3.2 Masked Language-aware Module
In order to occlude the character-wise visual cues for the guidance of linguistic learning, we propose a Masked Language-aware Module (MLM) to automatically generate the character-wise mask map with only original word-level annotations.
As shown in Fig. 3, MLM takes visual features V and the character index as inputs. Character index indicates the index of the occluded character, which is randomly obtained for each input word image with length . Then the transformer unit [35] is used to improve the feature representation ability. Finally, after integrating with character index information, the character mask map is obtained through a sigmoid layer, which is used to generate the occluded feature map in Fig. 2.
To guide the learning process of , two parallel branches are designed based on the Weakly-supervised Complementary Learning (WCL). WCL aims to guide to cover more area of the occluded character, which complementarily makes contain more region of other characters. In the first branch, we implement the element-wise product between and to generate the feature map containing visual semantics of the occluded character (e.g. character “b” in the word “burns” with character index 1 in Fig. 3). In contrast, the element-wise product between and in the second branch is used to generate the feature map containing visual semantics of other characters (e.g. string “urns” in the word “burns” in Fig. 3). By doing these, the complementary learning process guides the to only cover the character at corresponding position without overlapping other characters (shown in Fig. 7). We share the weights of transformer unit and prediction layer (Eq. 1) among two parallel branches for the feature representation enhancement and semantic guidance. is the feature map and is the attention map, where is the channel number, is the max time step, and are the height and width. is positional encoding [35] of character orders. , , are trainable weights and is the time step.

(1) |
(2) |
(3) |
Compared with BERT [5], though both approaches mask out the information in a certain time step, the proposed MLM masks the visual features in the 2d spatial space instead of covering token-level information. Furthermore, as STR datasets are weakly labeled, it is difficult to obtain the accurate character-wise pixel-level annotations. Thus, it is impractical to directly implement BERT-based methods [34, 1, 22] into STR task. Based on these, MLM helps the model to learn linguistic information from a new perspective, which can not be replaced by exiting masking approaches.
The supervisions of WCL are automatically obtained by using the original word-level annotation and randomly generated character index (detailed in Sec. 4). Thus, MLM automatically generates accurate character mask map without the need of additional annotations, making it possible for the real application.
3.3 Visual Reasoning Module

Different from previous methods capturing the visual and linguistic information in the two-step architecture, we propose the Visual Reasoning Module (VRM) to model the two information simultaneously in a unified structure. As a pure vision-based structure, VRM aims to reason the word-level prediction from occluded features by using the character-wise information in the visual context.
The details of VRM is shown in Fig. 4, it contains two parts: Visual Semantic Reasoning (VSR) layer and Parallel Prediction (PP) layer. VSR layer consists of transformer units [35], which are proved to be effective for modeling long-range dependencies in recent computer vision tasks [2, 24]. Specially, position encoding is used to perceive the pixel location information. Different from [45] using transformer units for pure language modeling, the transformer units in the proposed VRM are used for sequence modeling, which will not be influenced by length of the word. Then, the PP layer is designed to predict the characters in parallel, which has identical formulation as Eq. 1.
In order to achieve the language modeling process , the reasoning process of the character needs to purely depend on the information of other characters. As MLM accurately occludes the character information in the training stage, VSR layer is guided to predict the dependencies between visual features of characters to infer the semantics of occluded character. Thus, with the word-level supervision, VSR layer learns to initiatively model the linguistic information in visual context to assist recognition. In the testing stage, VSR layer is able to adaptively consider the linguistic information for visual feature enhancement when the current visual semantics are confused (e.g. occlusion, noise, etc.).

We visualize the feature maps generated from VSR layer in testing to better understand how the learned linguistic information improves the recognition performance. As shown in Fig. 5, VSR layer effectively supplements the semantics of occluded character “r” in the word “better”, and correctly highlights the discriminating visual cues of character “t” in the word “trans” with the help of linguistic information in visual context. Without the initiative linguistic learning guided by MLM, VRM wrongly predicts the input images as “bettep” and “rrans”.
3.4 Training Objective
The final objective function of the proposed method is formulated in Eq. 4. is loss in VRM, and & are losses for predicting masked character and other characters in MLM respectively. and are used to balance the losses. Specially, we set , and use cross-entropy loss formulated in Eq. 5 for , and . and represent the prediction and ground truth. We set N to 25 in our experiments.
(4) |
(5) |
4 Experiment
4.1 Datasets
We conduct experiments following the setup of [45] in the purpose of fair comparison. The training datasets are SynthText (ST) [10] and SynthText90K (90K) [12]. The performance is evaluated on 6 benchmarks containing IIIT 5K-Words (IIIT5K) [25], ICDAR2013 (IC13) [14], ICDAR2015 (IC15) [13], Street View Text (SVT) [37], Street View Text-Perspective (SVTP) [29] and CUTE80 (CT) [30]. Details of above 6 datasets can be found in previous works [45, 28].
In addition, we provide a new Occlusion Scene Text (OST) dataset to reflect the ability for recognizing cases with missing visual cues. This dataset is collected from 6 benchmarks (IC13, IC15, IIIT5K, SVT, SVTP and CT) containing 4832 images. Images in this dataset are manually occluded in weak or heavy degree (shown in Fig. 6). Weak and heavy degrees mean that we occlude the character using one or two lines. For each image, we randomly choose one degree to only cover one character. More examples of OST are shown in the supplementary materials.
4.2 Implementation Details
We use the ResNet45 [32, 38, 28] as our backbone. Particularly, we set the stride to 2 in stage 2,3,4 and initialize the weights by default. Following the most recent works [45, 28], we set the image size to (there is no obvious difference with the size of in our experiments). Data augmentation including random rotation, color jittering and perspective distortion. We conduct the experiments on 4 NVIDIA V100 GPUs with batch size 384. The network is trained end-to-end using Adam optimizer with learning rate 1e-4. The recognition covers 37 characters including a-z, 0-9, and an end-of-sequence symbol.
Following [45], we divide the training process into 2 steps: language-free (LF) step and language-aware (LA) step. It is worth mentioning that we control the total number of training sessions to be consistent with existing methods for fair comparison. 1) In LF step, we split the connection between MLM and VRM ( in Fig. 2) to guarantee a more stable learning process of both modules. VRM in this step will not acquire the language capability and only uses visual texture for prediction. 2) In LA step, generated from MLM is used to occlude the feature map to guide the learning of linguistic rules in VRM. Specifically, we control the ratio of occluded number in a batch, which aims to balance the cases with rich or weak visual information during the training stage.
As all the training images have word-level annotations, we randomly generate the character index based on the length of word, and use this index and the original word-level annotation to generate the labels for MLM (e.g. when index is 4 and word is “house”, the labels are “s” and “houe” respectively). The label generating process is automatic without manual intervention, making it easy to finetune our model on other datasets.
4.3 Ablation Study
We illustrate the effectiveness of proposed modules in this section. To be specific, baseline contains VRM with two transformer units in Tab. 1& 2& 3.
Ratio | IIIT5K | IC13 | SVT | IC15 | SVTP | CT |
---|---|---|---|---|---|---|
Baseline | 94.5 | 94.2 | 89.3 | 79.8 | 81.1 | 85.8 |
1:2 | 95.0 | 94.8 | 90.4 | 80.8 | 83.0 | 88.0 |
1:1 | 95.4 | 95.0 | 91.0 | 81.8 | 83.7 | 88.2 |
2:1 | 95.0 | 94.7 | 90.0 | 81.1 | 82.7 | 88.1 |
Methods | IIIT5K | IC13 | SVT | IC15 | SVTP | CT |
---|---|---|---|---|---|---|
Mas only | 94.8 | 94.7 | 89.8 | 81.7 | 82.3 | 87.2 |
Rem only | 95.2 | 94.8 | 89.9 | 81.1 | 82.2 | 88.0 |
WCL | 95.4 | 95.0 | 91.0 | 81.8 | 83.7 | 88.2 |
Methods | Average accuracy(%) |
---|---|
Baseline | 88.8 |
Dropout [33] | 89.0 |
Cutout [6] | 89.0 |
MLM | 90.2 |
Methods | IIIT5K | IC13 | SVT | IC15 | SVTP | CT |
---|---|---|---|---|---|---|
VRM-2L | 95.4 | 95.0 | 91.0 | 81.8 | 83.7 | 88.2 |
VRM-3L | 95.8 | 95.7 | 91.7 | 83.7 | 86.0 | 88.5 |
The effectiveness of MLM. The proposed MLM aims to guide the linguistic learning process in the VRM. We conduct several experiments to evaluate its effectiveness in Tab. 1. The baseline model is implemented without MLM. We change the ratio of occluded number in a batch to study its influence to the recognition performance (e.g. when the batch size is 128, = 1:3 means that we use to occlude V for only 32 samples in 1 batch, and feature maps of the rest 96 samples remain unchanged). As shown in Tab. 1, the proposed MLM significantly improves the performance of baseline model when the ratio ranges from 1:2 to 2:1. For the irregular datasets (IC15, SVTP, CT) containing amounts of images with confused visual cues (blur, occlusion, noise, etc.), the proposed MLM improves the baseline model at least 2% in accuracy with 1:1, which further demonstrates that the initiative linguistic learning process effectively helps the vision model to handle confused visual cues. For regular datasets, the improvement is also considerable (0.9%, 0.8%, and 1.7% on IIIT5K, IC13 and SVT datasets respectively). When the ratio raises up to 2:1, the performance drops slightly. We infer that the large value of ratio will break the balance between cases with rich and weak visual cues during the training process. Therefore, we set the value of ratio to 1:1 in the rest experiments.
Methods | Training Data | Annos | IIIT5K | IC13 | SVT | IC15 | SVTP | CT | |
---|---|---|---|---|---|---|---|---|---|
Lan-free | CTC [31] | 90K | word | 81.2 | 89.6 | 82.7 | - | - | - |
ACE [42] | 90K | word | 82.3 | 89.7 | 82.6 | 68.9 | 70.1 | 82.6 | |
FCN [19] | ST | word, char | 91.9 | 91.5 | 86.4 | - | - | - | |
Lan-aware | FAN [3] | 90K+ST | word | 87.4 | 93.3 | 85.9 | 70.6 | - | - |
AON [4] | 90K+ST | word | 87.0 | - | 82.8 | 68.2 | 73.0 | 76.8 | |
ASTER [32] | 90K+ST | word | 93.4 | 91.8 | 89.5 | 76.1 | 78.5 | 79.5 | |
ESIR [47] | 90K+ST | word | 93.3 | 91.3 | 90.2 | 76.9 | 79.6 | 83.3 | |
ScRN [43] | 90K+ST | word, char | 94.4 | 93.9 | 88.9 | 78.7 | 80.8 | 87.5 | |
SAR [16] | 90K+ST | word | 91.5 | 91.0 | 84.5 | 69.2 | 76.4 | 83.3 | |
TextScanner [36] | 90K+ST | word, char | 83.9 | 92.9 | 90.1 | 79.4 | 84.3 | 83.3 | |
DAN [38] | 90K+ST | word | 94.3 | 93.9 | 89.2 | 74.5 | 80.0 | 84.4 | |
Wang et al. [39] | 90K+ST | word | 94.4 | 93.7 | 89.8 | 75.1 | 80.2 | 86.8 | |
SRN [45] | 90K+ST | word | 94.8 | 95.5 | 91.5 | 82.7 | 85.1 | 87.8 | |
SEED [28] | 90K+ST | word | 93.8 | 92.8 | 89.6 | 80.0 | 81.4 | 83.6 | |
Ours | Baseline | 90K+ST | word | 94.6 | 94.3 | 89.3 | 81.2 | 81.6 | 86.8 |
VisionLAN | 90K+ST | word | 95.8 | 95.7 | 91.7 | 83.7 | 86.0 | 88.5 |
The effectiveness of WCL. To demonstrate the effectiveness of proposed Weakly-supervised Complementary Learning in MLM, we conduct several experiments implemented with only the first branch (occluded character) or the second branch (remaining string). As shown in Tab. 2, MLM implemented with the complementary learning process obtains better results than the methods only guiding the semantics of occluded character or remaining string during the training stage.
Compared with other masking methods. We compare MLM with [6, 33] to evaluate our effectiveness in language modeling. All the modules only work on for fair comparison. As shown in Tab. 3, the proposed MLM significantly improves the recognition results (1.4% vs 0.2%). As detailed in Sec. 3.3, the reasoning process of the character needs to purely depend on the information of other characters without containing current character-wise information. Thus, randomly masking pixel-wise feature [6, 33] does not have the ability of linguistic learning. Benefiting from the well-designed architecture and ingenious weakly supervised learning, MLM accurately localizes character-wise visual cues, which has the ability to guide the linguistic learning process in VRM.
The effectiveness of VRM. To study the relationship between the recognition performance and the ability of capturing linguistic information, we compare the results of models implemented with different number of transformer units in VSR layer. As shown in Tab. 4, VRM implemented with three transformer units further improves the performance, which has the stronger language capability.
4.4 Comparisons with State-of-the-Arts
We compare our method with previous state-of-the-art methods on 6 benchmarks in Tab. 5. We simply divide the methods into language-free and language-aware methods according to whether linguistic information are used. The language-aware methods perform better than language-free methods in general. Benefiting from adaptively considering the linguistic information for feature enhancement, the proposed VisionLAN achieves state-of-the-art performance across the 6 public datasets compared with both language-free and language-aware methods. Specifically, for regular datasets, the proposed VisionLAN obtains 1%, 0.2% and 0.2% improvement on IIIT5K, IC13 and SVT datasets respectively. For irregular datasets, the increases are 1%, 0.9% and 0.7% on IC15, SVTP and CT respectively.
Methods | Speed | EIPs |
---|---|---|
Baseline | 11.5ms | - |
Baseline + [32] | 43.2ms | 3.0M |
Baseline + [45] | 19ms | 12.6M |
VisionLAN | 11.5ms | 0M |

Methods | Average | Weak | Heavy |
---|---|---|---|
Baseline | 53.0 | 63.2 | 42.7 |
Baseline + [32] | 53.9 | 63.9 | 43.9 |
Baseline + [45] | 58.2 | 68.4 | 48.0 |
VisionLAN | 60.3 | 70.3 | 50.3 |
As VisionLAN adaptively considers the visual and linguistic information in the 2d visual space, our method is less sensitive to the distorted images. Thus, the proposed method can obtain better results than ASTER [32] and ESIR [47] on irregular datasets, which adopt the rectification process before recognition. As shown in Tab. 5, the increases are 7.6%, 7.5% and 9% for [32], and 6.8%, 6.4% and 5.2% for [47] on IC15, SVTP and CT datasets respectively.
We further compare the differences between the existing methods and ours in recognition speed and the extra introduced parameters (EIPs) for capturing linguistic information in Tab. 6. In terms of approaching speed and parameters, we implement one transformer unit in GSRM of [45] (the same goes for Sec. 4.5). As the linguistic information is acquired along with the visual features without the need of extra language model, the proposed VisionLAN significantly improves the speed by at least 39% (11.5ms vs 19ms and 43.2ms) without introducing extra parameters (0M vs 12.6M and 3M). Furthermore, as VisionLAN directly considers the linguistic information in the visual space, its efficiency of capturing linguistic information will not be affected by the word length.
4.5 The Language Capability on OST Dataset
To evaluate the language capability of our VisionLAN in detail, we compare our method with recent most popular language models (RNN [32] and Transformer [45]) on OST dataset to evaluate their performance on the case of missing character-wise visual cues. Specifically, we connect these language models to VRM following the implementation details in their papers. As shown in Tab. 7, though the linguistic information captured by [32] and [45] can assist the prediction of vision model, the proposed VisionLAN significantly outperforms these methods by viewing the visual and linguistic information as a union. Through adaptively aggregating the two information in a unified structure instead of considering them independently, VisionLAN improves the baseline model by 7.3% in average.
4.6 The Generalization Ability on Long Chinese Dataset
We evaluate VisionLAN on non-Latin Long Text (TRW15 [49]) to prove its generalization ability. This dataset contains 2997 cropped images, and we set the max length to 50. We train the proposed VisionLAN following the setup of [45]. As shown in Tab. 8, compared with language-free (CTC) and language-aware (2D-Attention) methods, VisionLAN outperforms these approaches by at least 14.9%. Benefiting from viewing the visual and linguistic information as a union, the proposed VisionLAN achieves a new state-of-the-art result and significantly outperforms SRN [45] by 3.2%. More experiments on other datasets (e.g. MLT [26], etc.) are available in the supplementaries.
4.7 The Qualitative Analysis
MLM in character-wise localization. To qualitatively analyze the effectiveness of MLM, we visualize some examples of generated in Fig. 7. The generated effectively localizes character-wise visual cues at corresponding position with the guidance of character index . Furthermore, MLM is able to handle the distorted images (e.g. the curved word image “nothing”) and the localization of repeated characters (e.g. the character “b” with in word “confabbing”). The quantitative evaluation of character-wise localization performance and more visualizations of are available in the supplementaries.
The effectiveness of VisionLAN. We collect some recognition results to illustrate how the learned linguistic information helps vision model to improve the performance. As shown in Fig. 8 (a), VisionLAN can handle the cases with confusing characters. For example, as the character “e” has the similar visual cues to character “f” in the image with word “before”, the VisionLAN without MLM wrongly gives the prediction “f”, while VisionLAN correctly infers the character “e” with the help of linguistic information. For the samples in Fig. 8 (b), VisionLAN can also use linguistic rules to eliminate the background interference (including occlusion, illumination, background textures, etc.). Furthermore, the accurate recognition of the blurred characters in Fig. 8 (c) also demonstrates the effectiveness of our method.
5 Conclusion
As the first work to endue the vision model with language capability, this paper proposes a concise and effective architecture for scene text recognition. VisionLAN successfully achieves the transformation from two-step to one-step recognition (from Two to One), which adaptively considers both visual and linguistic information in a unified structure without the need of extra language model. Compared with previous language model, VisionLAN shows a stronger language capability while maintaining high efficiency. In addition, a new Occlusion Scene Text dataset is proposed to evaluate the performance on the cases of missing character-wise visual cues. Extensive experiments on seven benchmarks and the proposed OST dataset demonstrate the effectiveness and efficiency of our method. We regard the proposed VisionLAN as a basic step toward more robust and accurate scene text recognition, and we will further explore its potential in the future.


Acknowledgments
This work is supported by the National Nature Science Foundation of China (62121002, 62022076, U1936210), the Fundamental Research Funds for the Central Universities under Grant WK3480000011, the China Postdoctoral Science Foundation (2021M693092) and JD AI research. We also acknowledge the support of GPU cluster built by MCC Lab of Information Science and Technology Institution, USTC.
References
- [1] Chris Alberti, Jeffrey Ling, Michael Collins, and David Reitter. Fusion of detected objects in text for visual question answering. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 2131–2140, 2019.
- [2] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. arXiv preprint arXiv:2005.12872, 2020.
- [3] Zhanzhan Cheng, Fan Bai, Yunlu Xu, Gang Zheng, Shiliang Pu, and Shuigeng Zhou. Focusing attention: Towards accurate text recognition in natural images. In Proceedings of the IEEE international conference on computer vision, pages 5076–5084, 2017.
- [4] Zhanzhan Cheng, Yangliu Xu, Fan Bai, Yi Niu, Shiliang Pu, and Shuigeng Zhou. Aon: Towards arbitrarily-oriented text recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5571–5579, 2018.
- [5] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- [6] Terrance DeVries and Graham W Taylor. Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552, 2017.
- [7] Shancheng Fang, Hongtao Xie, Zheng-Jun Zha, Nannan Sun, Jianlong Tan, and Yongdong Zhang. Attention and language ensemble for scene text recognition with convolutional sequence modeling. In Proceedings of the 26th ACM international conference on Multimedia, pages 248–256, 2018.
- [8] Jiannan Ge, Hongtao Xie, Shaobo Min, and Yongdong Zhang. Semantic-guided reinforced region embedding for generalized zero-shot learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 1406–1414, 2021.
- [9] Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd international conference on Machine learning, pages 369–376, 2006.
- [10] Ankush Gupta, Andrea Vedaldi, and Andrew Zisserman. Synthetic data for text localisation in natural images. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2315–2324, 2016.
- [11] Catherine L Harris. Language and cognition. Encyclopedia of cognitive science, pages 1–6, 2006.
- [12] Max Jaderberg, Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Synthetic data and artificial neural networks for natural scene text recognition. NIPS, 2014.
- [13] Dimosthenis Karatzas, Lluis Gomez-Bigorda, Anguelos Nicolaou, Suman Ghosh, Andrew Bagdanov, Masakazu Iwamura, Jiri Matas, Lukas Neumann, Vijay Ramaseshan Chandrasekhar, Shijian Lu, et al. Icdar 2015 competition on robust reading. In 2015 13th International Conference on Document Analysis and Recognition (ICDAR), pages 1156–1160. IEEE, 2015.
- [14] Dimosthenis Karatzas, Faisal Shafait, Seiichi Uchida, Masakazu Iwamura, Lluis Gomez i Bigorda, Sergi Robles Mestre, Joan Mas, David Fernandez Mota, Jon Almazan Almazan, and Lluis Pere De Las Heras. Icdar 2013 robust reading competition. In 2013 12th International Conference on Document Analysis and Recognition, pages 1484–1493. IEEE, 2013.
- [15] Chen-Yu Lee and Simon Osindero. Recursive recurrent nets with attention modeling for ocr in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2231–2239, 2016.
- [16] Hui Li, Peng Wang, Chunhua Shen, and Guyu Zhang. Show, attend and read: A simple and strong baseline for irregular text recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 8610–8617, 2019.
- [17] Jiaming Li, Hongtao Xie, Jiahong Li, Zhongyuan Wang, and Yongdong Zhang. Frequency-aware discriminative feature learning supervised by single-center loss for face forgery detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6458–6467, 2021.
- [18] Minghui Liao, Pengyuan Lyu, Minghang He, Cong Yao, Wenhao Wu, and Xiang Bai. Mask textspotter: An end-to-end trainable neural network for spotting text with arbitrary shapes. IEEE transactions on pattern analysis and machine intelligence, 2019.
- [19] Minghui Liao, Jian Zhang, Zhaoyi Wan, Fengming Xie, Jiajun Liang, Pengyuan Lyu, Cong Yao, and Xiang Bai. Scene text recognition from two-dimensional perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 8714–8721, 2019.
- [20] Fanchao Lin, Hongtao Xie, Yan Li, and Yongdong Zhang. Query-memory re-aggregation for weakly-supervised video object segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 2038–2046, 2021.
- [21] John L Locke. Why do infants begin to talk? language as an unintended consequence. Journal of child language, 23(2):251–268, 1996.
- [22] Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. In Advances in Neural Information Processing Systems, pages 13–23, 2019.
- [23] Minh-Thang Luong, Hieu Pham, and Christopher D Manning. Effective approaches to attention-based neural machine translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 1412–1421, 2015.
- [24] Pengyuan Lyu, Zhicheng Yang, Xinhang Leng, Xiaojun Wu, Ruiyu Li, and Xiaoyong Shen. 2d attentional irregular scene text recognizer. arXiv preprint arXiv:1906.05708, 2019.
- [25] Anand Mishra, Karteek Alahari, and CV Jawahar. Scene text recognition using higher order language priors. In BMVC, 2012.
- [26] Nibal Nayef, Yash Patel, Michal Busta, Pinaki Nath Chowdhury, Dimosthenis Karatzas, Wafa Khlif, Jiri Matas, Umapada Pal, Jean-Christophe Burie, Cheng-lin Liu, et al. Icdar2019 robust reading challenge on multi-lingual scene text detection and recognition—rrc-mlt-2019. In 2019 International Conference on Document Analysis and Recognition (ICDAR), pages 1582–1587. IEEE, 2019.
- [27] Yash Patel, Lluis Gomez, Marçal Rusinol, and Dimosthenis Karatzas. Dynamic lexicon generation for natural scene images. In European Conference on Computer Vision, pages 395–410. Springer, 2016.
- [28] Zhi Qiao, Yu Zhou, Dongbao Yang, Yucan Zhou, and Weiping Wang. Seed: Semantics enhanced encoder-decoder framework for scene text recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13528–13537, 2020.
- [29] Trung Quy Phan, Palaiahnakote Shivakumara, Shangxuan Tian, and Chew Lim Tan. Recognizing text with perspective distortion in natural scenes. In Proceedings of the IEEE International Conference on Computer Vision, pages 569–576, 2013.
- [30] Anhar Risnumawan, Palaiahankote Shivakumara, Chee Seng Chan, and Chew Lim Tan. A robust arbitrary text detection system for natural scene images. Expert Systems with Applications, 41(18):8027–8048, 2014.
- [31] Baoguang Shi, Xiang Bai, and Cong Yao. An end-to-end trainable neural network for image-based sequence recognition and its application to scene text recognition. IEEE transactions on pattern analysis and machine intelligence, 39(11):2298–2304, 2016.
- [32] Baoguang Shi, Mingkun Yang, Xinggang Wang, Pengyuan Lyu, Cong Yao, and Xiang Bai. Aster: An attentional scene text recognizer with flexible rectification. IEEE transactions on pattern analysis and machine intelligence, 41(9):2035–2048, 2018.
- [33] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1):1929–1958, 2014.
- [34] Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei Lu, Furu Wei, and Jifeng Dai. Vl-bert: Pre-training of generic visual-linguistic representations. In International Conference on Learning Representations, 2019.
- [35] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008, 2017.
- [36] Zhaoyi Wan, Mingling He, Haoran Chen, Xiang Bai, and Cong Yao. Textscanner: Reading characters in order for robust scene text recognition. arXiv preprint arXiv:1912.12422, 2019.
- [37] Kai Wang, Boris Babenko, and Serge Belongie. End-to-end scene text recognition. In 2011 International Conference on Computer Vision, pages 1457–1464. IEEE, 2011.
- [38] Tianwei Wang, Yuanzhi Zhu, Lianwen Jin, Canjie Luo, Xiaoxue Chen, Yaqiang Wu, Qianying Wang, and Mingxiang Cai. Decoupled attention network for text recognition. In AAAI, pages 12216–12224, 2020.
- [39] Yizhi Wang and Zhouhui Lian. Exploring font-independent features for scene text recognition. ECCV, 2020.
- [40] Yuxin Wang, Hongtao Xie, Zilong Fu, and Yongdong Zhang. Dsrn: A deep scale relationship network for scene text detection. In IJCAI, pages 947–953, 2019.
- [41] Yuxin Wang, Hongtao Xie, Zheng-Jun Zha, Mengting Xing, Zilong Fu, and Yongdong Zhang. Contournet: Taking a further step toward accurate arbitrary-shaped scene text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11753–11762, 2020.
- [42] Zecheng Xie, Yaoxiong Huang, Yuanzhi Zhu, Lianwen Jin, Yuliang Liu, and Lele Xie. Aggregation cross-entropy for sequence recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6538–6547, 2019.
- [43] Mingkun Yang, Yushuo Guan, Minghui Liao, Xin He, Kaigui Bian, Song Bai, Cong Yao, and Xiang Bai. Symmetry-constrained rectification network for scene text recognition. In Proceedings of the IEEE International Conference on Computer Vision, pages 9147–9156, 2019.
- [44] Fei Yin, Yi-Chao Wu, Xu-Yao Zhang, and Cheng-Lin Liu. Scene text recognition with sliding convolutional character models. arXiv preprint arXiv:1709.01727, 2017.
- [45] Deli Yu, Xuan Li, Chengquan Zhang, Tao Liu, Junyu Han, Jingtuo Liu, and Errui Ding. Towards accurate scene text recognition with semantic reasoning networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12113–12122, 2020.
- [46] Xiaoyu Yue, Zhanghui Kuang, Chenhao Lin, Hongbin Sun, and Wayne Zhang. Robustscanner: Dynamically enhancing positional clues for robust text recognition. eccv, 2020.
- [47] Fangneng Zhan and Shijian Lu. Esir: End-to-end scene text recognition via iterative image rectification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2059–2068, 2019.
- [48] Chuhan Zhang, Ankush Gupta, and Andrew Zisserman. Adaptive text recognition through visual matching. ECCV, 2020.
- [49] Xinyu Zhou, Shuchang Zhou, Cong Yao, Zhimin Cao, and Qi Yin. Icdar 2015 text reading in the wild competition. arXiv preprint arXiv:1506.03184, 2015.