2022
1]\orgdivInternational Core Business Unit, \orgnameAlibaba Group, \orgaddress\cityHangzhou 310051, \countryChina
2]\orgdivComputer Vision Lab, \orgnameETH Zürich, \orgaddress\cityZürich 8092, \countrySwitzerland
Masked Vision-Language Transformer in Fashion
Abstract
We present a masked vision-language transformer (MVLT) for fashion-specific multi-modal representation. Technically, we simply utilize vision transformer architecture for replacing the BERT in the pre-training model, making MVLT the first end-to-end framework for the fashion domain. Besides, we designed masked image reconstruction (MIR) for a fine-grained understanding of fashion. MVLT is an extensible and convenient architecture that admits raw multi-modal inputs without extra pre-processing models (e.g., ResNet), implicitly modeling the vision-language alignments. More importantly, MVLT can easily generalize to various matching and generative tasks. Experimental results show obvious improvements in retrieval (rank@5: 17%) and recognition (accuracy: 3%) tasks over the Fashion-Gen 2018 winner Kaleido-BERT. Code is made available at https://github.com/GewelsJI/MVLT.
keywords:
Vision-language, masked image reconstruction, transformer, fashion, e-commercial.1 Introduction

The emergence of transformer†† Contributed equally. ∗ Corresponding author. Work was done while Ge-Peng Ji was an research intern in Alibaba Group. is drawing enormous attention from the academic community, facilitating the advancement of computer vision (CV) vit2021ICLR ; liu2021Swin and natural language processing (NLP) vaswani2017attention ; sun2022paradigm . Benefiting from the robustness of transformers, researchers also contribute to the vision-language (VL) field agarwal2021evaluating ; chen2020generative ; lin2021m6 ; ramesh2021zero ; wu2021fashion with zeal. To better utilize the pre-trained models in CV and NLP, existing general VL models are mainly based on the BERT model devlin2019bert or adopt the well-pretrained vision extractors he2016deep ; ren2015faster or both. However, general VL methods qi2020imagebert ; lu2019vilbert ; chen2020uniter still struggle when applied to the fashion domain in e-commerce because they suffer from the two main issues: a) Insufficient Granularity. Unlike the general objects with complex backgrounds, only focusing on coarse-grained semantics is insufficient for a fashion product hsiao2019fashion++ ; vasileva2018learning ; fan2022domain , as it would lead the network to generate sub-optimal results. Contrarily, the fashion-oriented framework requires more fine-grained representations, such as a suit with different materials (e.g., wool, linen, and cotton) or collars (e.g., band, camp, and windsor). b) Bad Transferability. The pre-extracted visual features are not discriminative for fashion-oriented tasks, restricting the cross-modal representations.
To address the above issues, we present a novel VL framework, termed masked vision-language transformer (MVLT). Specifically, we introduce a generative task, masked image reconstruction (MIR), for the fashion-based VL framework. Compared to previous pre-training tasks, such as masked image modeling (regression task) or masked image classification (classification task), MIR enables the network to learn more fine-grained representations via pixel-level visual knowledge (see Fig. 1). Further, inspired by pyramid vision transformer (PVT) wang2021pyramid , we utilize a pyramid architecture for our VL transformer. Then, we introduce the MIR task. These two improvements significantly enhance the ability to adapt to fashion-specific understanding and generative tasks, and can conduct in an end-to-end manner. To this end, MVLT can directly process the raw multi-modal inputs in dense formats (i.e., linguistic tokens and visual patches) without extra (e.g., ResNet) pre-processing models yang2020fashion ; al2020paris . Our main contributions are summarized as follows:
-
•
We introduce a novel masked image reconstruction (MIR) task, which is the first real pixel-level generative strategy utilized in VL pre-training.
-
•
Based on the MIR task, we present an end-to-end VL framework, called MVLT, for the fashion domain, greatly promoting the transferability to the downstream tasks and large-scale web applications.
-
•
Extensive experiments show that MVLT significantly outperforms the state-of-the-art models on matching and generative tasks.
2 Background

In recent years, BERT-based pre-training models have been widely investigated in VL tasks. Many previous attempts, such as LXMERT tan2019lxmert , VL-BERT su2020vl , and FashionBERT gao2020fashionbert , were successful in a wide range of downstream applications. Experiments and discussions show that BERT is a powerful method for learning multi-modal representations, outperforming several previous CNN-based lee2018stacked or LSTM-based niu2017hierarchical ; xia2022skating approaches. Compared to previous studies, this paper aims to develop a more efficient self-supervised objective that can be easily implemented in pre-training and provides better representations for real-world applications. Thus, we review research on masked learning strategies and end-to-end multi-modal schemes that inspired us the most.
2.1 Masked Learning Strategies
Masked modeling is the vital self-supervised task in BERT devlin2019bert and initially demonstrates outstanding abilities in natural language processing. Researchers have replicated its strength in language models because of its utility in multi-modal and vision tasks. Most VL works su2020vl ; lu2019vilbert ; li2020oscar transfer masked modeling into visual tokens and use a regression task to construct the token feature from nonsense-replace or a classification task to predict the token’s attribute. To reduce the difficulty in learning, Kaleido-BERT Zhuge2021KaleidoBERT optimizes masked modeling by employing a Kaleido strategy that facilitates coherent learning for multi-grained semantics. Although this work improves the performance of VL-related tasks in fashion indeed, we argue that the token-patch pre-alignment scheme by using auxiliary tool zhuge2021salient ; xu2015show is still complex and impedes the application to practical settings. Another work arici2021mlim introduces the MLIM approach that strengthens the masked image modeling with an image reconstruction task, which shares a similar idea to ours. However, our experiments showed that requiring a model to reconstruct the entire image without any reminder is too difficult. Recently, BEiT bao2022beit and MAE he2021masked utilize a BERT-style pre-training as part of the visual learner, and they discover that models are effective at learning semantics with such a scheme. These two works strengthen our conviction that converting the original masked image modeling (i.e., a regression task) to a masked image reconstruction task is possible. However, our primary goal is to design a generative pretext task that makes the multi-modal modeling in VL pre-training easier while eliminating the need for using prior knowledge. It will be extremely helpful in our practical application setting with billion-level data.
2.2 End-To-End Multi-Modal Schemes
Pixel-BERT huang2020pixel is the first method to consider end-to-end pre-training. It employs 22 max-pooling layers to reduce the spatial dimension of image features, with each image being downsampled 64 times. Although this work sets a precedent for end-to-end training, such a coarse and rigid method cannot work well in practical settings because it is simply combined with a ResNet he2016deep as part of joint pre-training, without considering the loss in speed and performance. Recently, VX2TEXT lin2021vx2text proposes to convert all modalities into language space and then perform end-to-end pre-training using a relaxation scheme. Though it is exciting to translate all the modalities into a unified latent space, it ignores that the usage of data extracted by pre-trained methods as input to the model cannot be regarded as an end-to-end framework. According to the timeline, ViLT kim2021vilt is the first method that indeed investigates an end-to-end framework via replacing region- or grid-based features with patch-based projections. However, without other designs, it cannot obtain competitive performance since it is just a vanilla extension of ViT vit2021ICLR . Grid-VLP yan2021grid is similar to ViLT, but it takes a further step by demonstrating that using a pre-trained CNN network as the visual backbone can improve performance on downstream tasks. SOHO huang2021seeing takes the entire image as input and creates a visual dictionary to affine the local region. However, this method does not fit fashion-specific applications due to the lack of reliable alignment information. As a result, the vision dictionary may merely learn the location of the background or foreground rather than complex semantics. FashionVLP Goenka2022fashionvlp uses a feedback strategy to achieve better retrieval performance. In practice, they use the well-pretrained knowledge extracted from ResNet and then model the whole, cropped, and landmark representations. Besides, they adopt Faster-RCNN as an object detector for popping out RoI candidates. Besides, some works are designed for end-to-end pre-training lei2021less ; xu2021e2e ; akbari2021vatt , but they are used for specific tasks and are not directly applicable to our research.

Despite existing methods employing different approaches to construct an end-to-end scheme, solutions that forgo pre-trained methods (e.g., ResNet, BERT) and use raw data (i.e., text, image) as inputs remain under-explored and are needed urgently in multi-modal applications.
Remarks. As shown in Fig. 2, similar to the existing two fashion-based approaches, i.e., FashionBERT (a) and Kaleido-BERT (b), the proposed MVLT (c) is also a patch-based VL learner, which extends the pyramid vision transformer wang2021pyramid to an architecture that adaptively extracts hierarchical representations for fashion cross-modal tasks. It is the first model that solves the end-to-end problem of VL pre-training in fashion, which allows us to simplify the implementation of our MVLT in the fashion industry using a twin-tower architecture yi2019sampling .
3 Masked Vision-Language Transformer
Our goal is to build an end-to-end VL framework for the fashion domain. The overall pipeline of our MVLT is depicted in Fig. 3. Like PVT, our architecture inherits four stages’ properties and generates features with different sizes. Two keys of the proposed architecture are the multi-modal encoder (Sec. 3.1) and the pre-training objectives (Sec. 3.2).
3.1 Multi-Modal Encoder
As shown in Fig. 3, MVLT admits visual and verbal inputs. On the language side, we first tokenize the caption of a fashion product and use the specific token [MASK] to randomly mask out the caption tokens with the masking ratio111We follow the default setting in BERT devlin2019bert . . Following the masking procedure, we obtain a sequence of word tokens. Then, we insert a specific [CLS] token at the head of this sequence. Besides, we pad the sequence to a unified length using the [PAD] token if the length is shorter than 128. This procedure generates the language input ids . On the vision side, we treat as visual input, where and denote the height and width of the given input. This input is sliced into multiple grid-like patches , where is the total number of patches and denotes the patch size. Similarly, the split patches are masked out with mask ratio . We provide more details about the above masking strategy for the language and vision parts in Sec. 3.2.
The above multi-modal inputs are embedded and fed into the consequent four VL interaction stages (i.e., ). In the first stage, we generate the vision and language embeddings, and respectively, via the given inputs ( and ). Regarding the subsequent stages, we consider only the -th stage, to have concise illustrations. As shown in the bottom part of Fig. 3, we first embed the language embedding into the language hidden feature , which is formulated as:
(1) |
where and are the learnable linear embedding and position embedding matrices. is the size of the hidden feature embedding.
Hyperparameter | ||||
---|---|---|---|---|
Layer number | 2 | 2 | 2 | 2 |
Hidden size | 64 | 128 | 320 | 512 |
Reduction size | ||||
Kernel size | 4 | 2 | 2 | 2 |
Stride length | 4 | 2 | 2 | 2 |
The visual embeddings are , where denotes the spatial reduction factor of visual embedding. To acquire pyramid visual features, are then embedded and flattened into the visual hidden feature via a two-dimensional projection (i.e., Conv2D block). In particular, this projection enforces the network to reduce the equivalent spatial dimension from to by utilizing the convolutional kernel with kernel size and stride length . This could be formulated as:
(2) |
where denotes the position embedding matrix. We then concatenate these two VL hidden features and feed them into multiple () VL transformer encoders. Each encoder contains the multi-head self-attention layer with spatial reduction (i.e., reduce block), multi-layer perceptron, and layer normalization. Finally, we obtain the encoded multi-modal feature and divide it into a language part and a visual part , where the operation consists in recovering the spatial dimension of the given feature.
After four VL interaction stages, we generate the four text embeddings and four pyramid vision embeddings , respectively. Table 1 presents more detailed hyperparameter settings of our method.
3.2 Pre-Training Objectives
To acquire discriminative multi-modal representations, we adopt three pre-training tasks to establish the inter-and intra-relationships between the most primitive VL modalities, including vision (masked image reconstruction, MIR), language (i.e., masked language modeling, MLM), and VL (image-text matching, ITM) modalities.

Objective 1: Masked Image Reconstruction (MIR). As for the general domain, models are enough to learn the coarse-grained semantics from the patch- or region-based objectives and achieve satisfactory results. However, the fashion-specific models require more fine-grained representations, such as a suit with different materials (e.g., wool) or collars (e.g., Windsor), which needs a pixel-to-pixel vision pre-training objective. Inspired by the masked language modeling devlin2019bert , we attempt to build the pixel-to-pixel relationships from the perspective of generative tasks, which promote the scalability of visual representations. We design the Masked Image Reconstruction (MIR) to accomplish this idea. To help our model learn better by MIR, we utilize the pyramid characteristic of PVT architecture wang2021pyramid to design a flexible masking strategy. Unlike the ViT-based method (a) in Fig. 4, PVT-based architecture (b) masks out the input image according to the masking unit matrix that contains small-grained patches. Given the patch sequence , the masked-out sequence is defined as:
(3) | ||||
where represents a function (or procedure) of our masking strategy, is the randomly selected area of the masking unit, and [ZERO] means that we use a pixel value of zero222In fact, we set to bring better optimization stability and less pattern degradation. to fill the selected areas. The masking units are derived from the indicator function:
(4) |
where each value in a set of integers is randomly selected from range with ratio . is the total number of masking units. For instance in Fig. 4 (b), we can define from 1 to 8. In our default settings, we set to capture more fine-grained semantics333The vanilla masking strategy in Fig. 4 (a) with becomes a special case of our masking strategy in Fig. 4 (b) when ..
Since the smooth- loss is less sensitive to the outliers, we use it as the pre-training objective to reconstruct the whole image via the masked-out sequence . It is defined as:
(5) |
where and denote the pixel at coordinate in the reconstructed image and the input image , respectively. is parameterized by learnable weights . Function denotes a standard four-level U-Net ronneberger2015u decoder, which admits four pyramidal vision embeddings as inputs.
Objective 2: Image-Text Matching (ITM). The appended classification embedding in the last language embedding is used to couple the representations from VL modalities. We utilize the function to denote a full-connected (FC) and softmax layers, parameterized by the weights . outputs a two-class probability vector , representing whether the input fashion image and caption match (i.e., positive pair) or not (i.e., negative pair). The positive pairs are selected from the same fashion product category, whereas the negative pairs are chosen at random from different entries. The binary cross-entropy loss function finally constrains this task:
(6) | ||||
where denotes the ground-truth label, i.e., for matched pairs and for unmatched pairs.
Objective 3: Masked Language Modeling (MLM). Following alberti2019fusion , we randomly use the specific token [MASK] to replace the original text tokens. The target of the MLM is to predict the text content for the masked tokens using the unmasked tokens and patches. Given a tokenized sequence , the masked-out sequence is denoted by . We use the cross-entropy loss to model this objective:
(7) |
where denotes the predicted probability for each masked-out token using . The function represents the parameters of a classifier. The final pre-training objective of the proposed MVLT is a combination of the three objectives:
(8) |
VSE | VSE++ | SCAN | PFAN | ViLBERT | ImageBERT | FashionBERT | VL-BERT | OSCAR | Kaleido-BERT | MVLT | ||||
Task | Metric | arXiv14 | BMVC18 | ECCV18 | arXiv19 | NeurIPS19 | arXiv20 | SIGIR20 | ICLR20 | ECCV20 | CVPR21 | OUR22 | Diff | |
TIR | @1 | 4.350% | 4.600% | 4.300% | 6.200% | 21.12% | 24.78% | 26.75% | 22.63% | 25.10% | 33.88% | 34.60% | ||
@5 | 12.76% | 16.89% | 13.00% | 20.79% | 37.23% | 45.20% | 46.48% | 36.48% | 49.14% | 60.60% | 78.00% | |||
@10 | 20.91% | 28.99% | 22.30% | 31.52% | 50.11% | 55.90% | 55.74% | 48.52% | 56.68% | 68.59% | 89.50% | |||
38.02 | 50.48 | 39.6 | 58.51 | 108.46 | 125.88 | 128.97 | 107.63 | 130.92 | 163.07 | 202.1 | ||||
ITR | @1 | 4.010% | 4.590% | 4.590% | 4.290% | 20.97% | 22.76% | 23.96% | 19.26% | 23.39% | 27.99% | 33.10% | ||
@5 | 11.03% | 14.99% | 16.50% | 14.90% | 40.49% | 41.89% | 46.31% | 39.90% | 44.67% | 60.09% | 77.20% | |||
@10 | 22.14% | 24.10% | 26.60% | 24.20% | 48.21% | 50.77% | 52.12% | 46.05% | 52.55% | 68.37% | 91.10% | |||
37.18 | 43.68 | 47.69 | 43.39 | 109.67 | 115.42 | 122.39 | 105.21 | 120.61 | 156.45 | 201.4 | ||||
M-CR | N/A | N/A | N/A | N/A | N/A | 90.77% | 91.25% | N/A | 91.79% | 95.07% | 98.26% | |||
macro- | N/A | N/A | N/A | N/A | N/A | 0.699 | 0.705 | N/A | 0.727 | 0.714 | 0.896 | |||
N/A | N/A | N/A | N/A | N/A | 160.67 | 161.75 | N/A | 164.49 | 166.47 | 187.86 | ||||
S-CR | N/A | N/A | N/A | N/A | N/A | 80.11% | 85.27% | N/A | 84.23% | 88.07% | 93.57% | |||
macro- | N/A | N/A | N/A | N/A | N/A | 0.575 | 0.620 | N/A | 0.591 | 0.636 | 0.829 | |||
N/A | N/A | N/A | N/A | N/A | 137.61 | 147.27 | N/A | 143.33 | 151.67 | 176.47 |
3.3 Downstream Tasks
For a fair comparison, we follow the same training/inference protocols as in gao2020fashionbert ; Zhuge2021KaleidoBERT and also adopt the Fashion-Gen 2018 rostamzadeh2018fashion benchmark as the base of our experiments. This dataset contains fashion products (i.e., entries for training and entries for testing) and their associated product descriptions. Each product corresponds to an image set (including samples) at various viewing angles. As a result, we utilize and image-text pairs as training and testing partitions, respectively. For a fair comparison, we test MVLT and compared models on Fashion-Gen using the following four fashion-related VL downstream tasks.
Task 1: Text-Image Retrieval (TIR). The TIR task requires the model to find a text with the highest similarity value with different query images. In particular, we take a product title and its corresponding image as a positive image-text pair, while the negative pairs are randomly selected from a pool of mismatched images. To increase our experiment’s difficulty, we constrain a set of image-text candidates (i.e., a positive pair and 100 negative pairs) in the same sub-category, making them as similar as possible.
Task 2: Image-Text Retrieval (ITR). As the reverse process of the TIR task, the ITR task aims to retrieve a matching image given a sequence of text entries of fashion description, where these bidirectional retrieval tasks (i.e., TIR and ITR) become a prominent member of cross-modal research. Similar to the above selection strategy in the TIR, we prepare a set of candidate image-text pairs, including a positive pair and 100 negative pairs from the same sub-category. We evaluate the zero-shot learning ability of our MVLT without further fine-tuning for these two retrieval tasks. We utilize three accuracy metrics (i.e., @1, @5, and @10) for the evaluation by ranking a series of predicted probabilities.
Task 3: Category Recognition (M-CR and S-CR). This task has two parts: main-category recognition (M-CR) and sub-category recognition (S-CR). These tasks act as the fundamental role of practical e-commerce applications that offer the specific category of the queried product. We expect that the model should possess the ability to recognize differences under different granularity levels: main-categories and sub-categories, such as . After the class embedding in the last language embedding , we add two independent FC layers to generate the final probabilities for two different recognition tasks. This procedure requires additional fine-tuning with recognition labels. We utilize two recognition-related metrics to evaluate performance: accuracy () and macro F-measure (macro-).
Task 4: Masked Image Generation (MIG). MIG task can be viewed as a pixel-wise reconstruction task. Each patch in the image is randomly masked with the probability (refer to the pre-training task MIR in Sec. 3.2). Then, we ask the model to recreate the whole image using the uncovered areas as visual clues.

4 Experiments
This section will detail our experiment to determine the factors leading to the success of the proposed MVLT.
4.1 Settings
This part provides the hyperparameter settings for our training procedure: i) Pre-training. We utilize PyTorch to implement our method, which is accelerated by 8 Tesla V100 GPUs. We adopt AdamW optimizer with a momentum value of , a mini-batch size of (i.e., per GPU), a weight decay of . To avoid over-fitting, we initialize MVLT on ImageNet pre-trained weights wang2021pyramid . The learning rate is initially set to and is changed using a cosine learning schedule. For the visual side, the input image is resized to and split into the multiple sub-patches with a size of . For the language side, all the product captions are tokenized and padded to tokens with a unified length of , including classification, caption, and padding tokens. The mask probabilities for vision and language are set to and , respectively. We empirically set weighting factors to balance the orders of magnitude of different loss values. ii) Fine-tuning. We transfer the pre-trained VL representation to each downstream application via fine-tuning in an end-to-end manner, whose settings are consistent with the pre-training process.


4.2 Results
As described in Sec. 3.3, we provide the details of four downstream fashion-related tasks. Experimental results show that our MVLT outperforms all competitors, including VSE kiros2014unifying , VSE++ faghri2017vse++ , SCAN lee2018stacked , PFAN wang2019position , ViLBERT lu2019vilbert , ImageBERT qi2020imagebert , FashionBERT gao2020fashionbert , VL-BERT su2020vl , OSCAR li2020oscar , and Kaleido-BERT Zhuge2021KaleidoBERT , which demonstrate the superiority for handling the VL understanding and generation tasks.
TIR and ITR. As shown in Table 2, our MVLT surpass the best method (i.e., Kaleido-BERT-CVPR21) on the TIR task by margins of +17.40%, +20.91% across the , . As for ITR, our method delivers more competitive results, with improvements of +17.11%, +22.73% on the , metrics, respectively. In any case, these results strongly support that our model is powerful enough to match vision and language. They also show how a) MIR and b) end-to-end pre-training are useful in fashion. We believe that MVLT would set a precedent in many industrial applications because it is a simple, cost-effective, and powerful architecture. Besides, we present the visualization results of these two retrieval tasks in Fig. 5.
M-CR and S-CR. Compared with BERT-based architectures qi2020imagebert ; gao2020fashionbert ; li2020oscar ; Zhuge2021KaleidoBERT , we also achieve top-1 performances in these two tasks, demonstrating our method have an excellent VL understanding capability. Moreover, compared with the best method Kaleido-BERT, our architecture improves by 0.193 in macro- metric for the S-CR task. In addition, the mean improvements in terms of the Sum metric (i.e., M-CR: +21.39 and S-CR: +24.80) are very significant. Since this metric is very sensitive to data distribution, it demonstrates MVLT has super-strong robustness. We also present the recognition results of M-CR and S-CR in Fig. 6.
MIG. As shown in Fig. 7, we showcase reconstructed images on the validation part of Fashion-Gen 2018 (a) and our e-commercial website (b). As seen, the reconstruction performance is truly remarkable. Since it requires our method to learn the fashion semantics truly, such results demonstrate the generative ability of our approach.
4.3 Ablation Studies
Mask Ratio. Table 3 (a) present four variants for different mask probability (i.e., (A1), (A2), (A3), (A4)) and our choice: (Final). The rises steadily with the masking probability until it reaches the sweet spot (%); then it reach performance plummets (%). We argue that increasing the will make MIR more complex, allowing MVLT to learn better semantics in a more restricted situation. However, masking out too much region will naturally result in losing valid visual information, leading to bad results.
(a) Mask Ratio () | (b) Masking Unit Size () | (c) Masking Style | (d) Pre-Training Tasks | (e) Pre-Train | MVLT | ||||||||||||
(A1) | (A2) | (A3) | (A4) | (B1) | (B2) | (B3) | (B4) | (C1) | (C2) | (C3) | (D1) | (D2) | (D3) | (E1) | (Final) | ||
App. | Metric | Grid | Stroke | Center | ITM | ITM+MIR | ITM+MLM | w/o PVT | |||||||||
TIR | @1 | 31.10% | 33.50% | 30.50% | 30.70% | 31.90% | 30.30% | 30.00% | 32.20% | 32.20% | 31.40% | 30.40% | 30.40% | 32.20% | 32.90% | 29.00% | 34.60% |
@5 | 75.70% | 76.00% | 75.50% | 73.80% | 75.30% | 75.60% | 73.90% | 76.90% | 75.30% | 76.10% | 75.10% | 74.10% | 76.00% | 76.20% | 72.20% | 78.00% | |
@10 | 88.60% | 88.70% | 88.00% | 88.60% | 89.60% | 88.60% | 88.20% | 88.60% | 88.50% | 89.20% | 87.20% | 83.50% | 87.20% | 88.60% | 86.60% | 89.50% | |
\hdashline | 195.40 | 198.20 | 194.00 | 193.10 | 196.80 | 194.50 | 192.10 | 197.70 | 196.00 | 196.70 | 192.70 | 188.00 | 195.40 | 197.70 | 187.80 | 202.10 | |
Diff | |||||||||||||||||
ITR | @1 | 30.00% | 29.90% | 29.90% | 28.50% | 29.00% | 29.70% | 29.00% | 28.90% | 31.40% | 31.10% | 30.10% | 29.30% | 30.40% | 28.40% | 25.60% | 33.10% |
@5 | 75.70% | 74.90% | 76.50% | 75.00% | 76.90% | 77.10% | 74.20% | 77.30% | 77.40% | 74.50% | 73.90% | 70.80% | 75.50% | 76.30% | 71.50% | 77.20% | |
@10 | 88.80% | 89.00% | 89.20% | 88.20% | 89.40% | 87.70% | 88.00% | 89.90% | 89.60% | 88.50% | 87.80% | 86.80% | 87.80% | 88.80% | 85.90% | 91.10% | |
\hdashline | 194.50 | 193.80 | 195.60 | 191.70 | 195.30 | 194.50 | 191.20 | 196.10 | 198.40 | 194.10 | 191.80 | 186.90 | 193.70 | 193.50 | 183.00 | 201.40 | |
Diff | |||||||||||||||||
M-CR | 98.16% | 97.87% | 98.09% | 98.06% | 98.03% | 98.04% | 98.11% | 98.01% | 98.12% | 98.07% | 98.04% | 96.49% | 97.11% | 98.08% | 97.92% | 98.26% | |
macro- | 0.870 | 0.860 | 0.890 | 0.870 | 0.870 | 0.880 | 0.850 | 0.870 | 0.869 | 0.877 | 0.870 | 0.806 | 0.853 | 0.876 | 0.879 | 0.896 | |
\hdashline | 185.16 | 183.87 | 187.09 | 185.06 | 185.03 | 186.04 | 183.11 | 185.01 | 185.02 | 185.77 | 185.04 | 177.09 | 182.41 | 185.68 | 185.82 | 187.86 | |
Diff | |||||||||||||||||
S-CR | 93.10% | 93.34% | 93.36% | 93.23% | 93.29% | 93.34% | 93.32% | 93.32% | 93.37% | 93.21% | 93.59% | 89.64% | 90.87% | 93.29% | 92.90% | 93.57% | |
macro- | 0.800 | 0.810 | 0.820 | 0.810 | 0.810 | 0.810 | 0.800 | 0.799 | 0.794 | 0.814 | 0.830 | 0.703 | 0.728 | 0.809 | 0.790 | 0.829 | |
\hdashline | 173.10 | 174.34 | 175.36 | 174.23 | 174.29 | 174.34 | 173.32 | 173.22 | 172.77 | 174.61 | 176.59 | 159.94 | 163.67 | 174.19 | 171.90 | 176.47 | |
Diff |
Masked Unit Size. Thanks to PVT’s flexibility, we can easily try different sizes of masked patches. As shown in Table 3 (b), we derive four variants with masked unit size (i.e., 1 (B1), 2 (B2), 8 (B3), 16 (B4)) to compare with our setting: 4 (Final). We found the performance is sensitive to this factor. It makes sense, revealing how vital it is to learn a robust fashion-related representation with a moderate granularity.

Masking Style. As shown in Fig. 8, we designed four types of masking strategies for the MIR task, whose quantitative differences are presented in Table 3 (c), i.e., grid (C1), stroke (C2), center (C3) and our random grid (Final) masking strategies. As can be seen, the random grid masking (Final) yields the best results, while the other three perform poorly. We believe this is because, in comparison to the grid (C1) and center (C3), random grid masking (Final) can help MVLT construct comprehensive representations. As our strategy (Final) does, the stroke (C2) also randomly masks the image given, yet it more or less leaves unmasked visual cues in the sub-patches. Our strategy enables the model to easily predict the masked region because semantics in the image are well preserved, enhancing the model’s robustness to learning in-sight knowledge.
Pre-Training Objectives. As shown in Table 3 (d), we derive four different variants to investigate the contribution of each objective, including ITM (D1), ITM+MIR (D2), ITM+MLM (D3), and our ITM+MIR+MLM (Final). When comparing D3 to D1 and D2 in the TIR task, we can see that D3 has a better performance in metric: % (D1) < % (D2) < % (D3). We conclude MLM task can help the model thoroughly learn the language knowledge, so it provides a more precise query to recall better-matching images. In the ITR task, we find the similar conclusion when comparing (D2) to (D1) and D3 in metric: % (D1) < % (D2) < % (D3). It indicates that better visual learning leads to an accurate image query to match the most appropriate caption.
Loading Pre-Trained Weight. As seen in Table 4, we add an experiment to demonstrate it is very important to load the PVT’s weight pre-trained on ImageNet deng2009imagenet . If not, it is obvious that our MVLT will suffer fierce drops (i.e., ITR: in , S-CR: in ). It is reasonable because a method pre-trained on large-scale general datasets can be more applicable in a specific field. It has already learned information such as color, texture, shape, etc.
TIR | ITR | M-CR | S-CR | |||||
macro- | macro- | |||||||
w/o PVT | 72.20% | 86.60% | 71.50% | 85.90% | 97.92% | 0.879 | 92.90% | 0.790 |
w/ PVT | 78.00% | 89.50% | 77.20% | 91.10% | 98.26% | 0.896 | 93.57% | 0.829 |
Diff |
4.4 More Discussions
How does MVLT perform in general domains? To further investigate the potential abilities in general settings, we here discuss two extended questions. a) Can the general models be directly transferred to the fashion domain? Inspired by the huge impact of general vision-language models, as in Table 5, we further investigate the zero-shot performance of two typical general models (i.e., ViLBERT lu2019vilbert and CLIP radford2021learning ). This has once again demonstrated the necessity and superiority of MVLT pre-trained on the specific domains. b) Can MVLT also work well in the general domain? We further verify the potential ability of our MVLT on the general domain. Table 6 reports the performance on MS-COCO 2014 dataset lin2014microsoft , where MVLT follows the same training standards as in kim2021vilt . It shows that MVLT achieves promising results compared to the latest models (i.e., Unicoder-VL li2020unicoder , UNITER chen2020uniter , and ViLT kim2021vilt ) without extra training data and special retrieval losses during the training. It indicates that MVLT is also a promising solution when extended to general scenes.
TIR | ITR | |||||
---|---|---|---|---|---|---|
@1 | @5 | @10 | @1 | @5 | @10 | |
ViLBERT (Zero-shot) | 7.18% | 18.73% | 29.84% | 8.99% | 15.34% | 26.14% |
CLIP (Zero-shot) | 16.30% | 40.60% | 55.60% | 13.60% | 43.10% | 57.60% |
MVLT (OUR) | 34.60% | 78.00% | 89.50% | 33.10% | 77.20% | 91.10% |
TIR task (5K Test) | ITR task (5K Test) | |||||
---|---|---|---|---|---|---|
@1 | @5 | @10 | @1 | @5 | @10 | |
Unicoder-VL† | 48.40% | 76.70% | 85.90% | 62.30% | 87.10% | 92.80% |
UNITER-Base† | 50.30% | 78.50% | 87.20% | 64.40% | 87.40% | 93.10% |
ViLT-Base/32 | 41.30% | 72.00% | 82.50% | 61.80% | 86.20% | 92.60% |
MVLT (OUR) | 49.66% | 79.88% | 87.50% | 65.38% | 90.04% | 93.60% |
Why do pyramid architecture and MIR benefit? As mentioned in the introduction, there are two understudied problems in the fashion domain. To solve the transferability problem, pyramidal architecture wang2021pyramid takes raw data as input without complex pre-processing, which essentially alleviates the applied burden in industry. Besides, MIR does not need human annotations like classification tags, bounding boxes, or pixel-wise segmentation labels. For the granularity problem wu2022multi , the pyramidal architecture wang2021pyramid provides multi-scale features with rich semantics. Combined with the MIR task, our framework can represent multi-grained fashion knowledge (e.g., dress, V-neck). These features are helpful and urgently required in this field.
A VL model that performs well for semantic understanding tasks (e.g., retrieval chen2022cross , classification) can serve as a good foundation and be easily applied to downstream tasks (e.g., text-to-image synthesis liu2022verbal , image captioning) by utilizing an additional decoder. We didn’t conduct image captioning experiments because we focused on basic representation learning in fashion this time.
MVLT v.s. MAE he2021masked . MAE learns general representations by allowing the model to explore pixel-to-pixel associations. So MVLT and MAE are similar in this regard. However, our MVLT is the first that introduces the vision reconstruction-alike pre-training for multi-modal research (e.g., fashion domain).
5 Conclusion
We present a vision-language framework named MVLT, which provides two contributions in this field: 1) a newly-designed masked image reconstruction (MIR) objective, and 2) an end-to-end pre-training scheme. The experimental and ablative analysis demonstrates the superiority of various matching and generative tasks. MVLT outperforms the cutting-edge method Kaleido-BERT with large margins on retrieval and recognition tasks, which would catalyze the fashion domain. The designed out-of-box method working end-to-end could simplify the workflow (e.g., data pre-processing and model training) for the actual engineering value, which improves development and business efficiency on large-scale e-commerce websites by approximately 50%.
In the future, we will continue to investigate an extremely efficient method in this field using famous technologies like hashing zhang2022modality , network pruning, and knowledge distil to alleviate the storage and computing limitations in real-world e-commerce applications.
Acknowledgments
This work is funded by Toyota Motor Europe via the research project TRACE-Zürich. The authors also would like to thank the anonymous reviewers and editor for their helpful comments on this manuscript.
Conflicts of Interests
The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
References
- (1) D. Gao, L. Jin, B. Chen, M. Qiu, P. Li, Y. Wei, Y. Hu, and H. Wang, “Fashionbert: Text and image matching with adaptive loss for cross-modal retrieval,” in International ACM SIGIR Conference on Research and Development in Information Retrieval. New York, NY, USA: Association for Computing Machinery, 2020, pp. 2251–2260, DOI: 10.1145/3397271.3401430.
- (2) M. Zhuge, D. Gao, D.-P. Fan, L. Jin, B. Chen, H. Zhou, M. Qiu, and L. Shao, “Kaleido-bert: Vision-language pre-training on fashion domain,” in Conference on computer vision and pattern recognition. Nashville, TN, USA: IEEE, 2021, pp. 12 642–12 652, DOI: 10.1109/CVPR46437.2021.01246.
- (3) A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, “An image is worth 16-16 words: Transformers for image recognition at scale,” in International Conference on Learning Representations. [Online]: PMLR, 2021.
- (4) Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in International conference on computer vision. Montreal, Canada: IEEE, 2021, pp. 9992–10 002, DOI: 10.1109/ICCV48922.2021.00986.
- (5) A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in Advances in neural information processing systems, vol. 30. Long Beach Convention Center, Long Beach, US: Curran Associates, Inc., 2017.
- (6) T.-X. Sun, X.-Y. Liu, X.-P. Qiu, and X.-J. Huang, “Paradigm shift in natural language processing,” Machine Intelligence Research, vol. 19, no. 3, pp. 169–183, 2022, DOI: 10.1007/s11633-022-1331-6.
- (7) S. Agarwal, G. Krueger, J. Clark, A. Radford, J. W. Kim, and M. Brundage, “Evaluating clip: towards characterization of broader capabilities and downstream implications,” [Online], 2021, Available: https://arxiv.org/abs/2108.02818.
- (8) M. Chen, A. Radford, R. Child, J. Wu, H. Jun, D. Luan, and I. Sutskever, “Generative pretraining from pixels,” in International conference on machine learning. PMLR, 2020, pp. 1691–1703.
- (9) J. Lin, R. Men, A. Yang, C. Zhou, M. Ding, Y. Zhang, P. Wang, A. Wang, L. Jiang, X. Jia et al., “M6: A chinese multimodal pretrainer,” in Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. Association for Computing Machinery, 2021, p. 3251–3261, DOI: 10.1145/3447548.3467206.
- (10) A. Ramesh, M. Pavlov, G. Goh, S. Gray, C. Voss, A. Radford, M. Chen, and I. Sutskever, “Zero-shot text-to-image generation,” in International conference on machine learning. PMLR, 2021, pp. 8821–8831.
- (11) H. Wu, Y. Gao, X. Guo, Z. Al-Halah, S. Rennie, K. Grauman, and R. Feris, “Fashion iq: A new dataset towards retrieving images by natural language feedback,” in Conference on computer vision and pattern recognition. Nashville, TN, USA: IEEE, 2021, pp. 11 307–11 317, DOI: 10.1109/CVPR46437.2021.01115.
- (12) J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional transformers for language understanding,” in Annual Conference of the North American Chapter of the Association for Computational Linguistics. Minneapolis, Minnesota: Association for Computational Linguistics, 2019, pp. 4171–4186, DOI: 10.18653/v1/N19-1423.
- (13) K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Conference on computer vision and pattern recognition. Las Vegas, NV, USA: IEEE, 2016, pp. 770–778, DOI: 10.1109/CVPR.2016.90.
- (14) S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” in Advances in neural information processing systems, vol. 28. Montreal, Quebec, Canada: Curran Associates, Inc., 2015.
- (15) D. Qi, L. Su, J. Song, E. Cui, T. Bharti, and A. Sacheti, “Imagebert: Cross-modal pre-training with large-scale weak-supervised image-text data,” [Online], 2020, Available: https://arxiv.org/abs/2001.07966.
- (16) J. Lu, D. Batra, D. Parikh, and S. Lee, “Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks,” in Advances in neural information processing systems, vol. 32. Vancouver, Canada: Curran Associates, Inc., 2019.
- (17) Y.-C. Chen, L. Li, L. Yu, A. El Kholy, F. Ahmed, Z. Gan, Y. Cheng, and J. Liu, “Uniter: Universal image-text representation learning,” in European conference on computer vision. Glasgow, United Kingdom: Springer, 2020, pp. 104–120, DOI: 10.1007/978-3-030-58577-8_7.
- (18) W.-L. Hsiao, I. Katsman, C.-Y. Wu, D. Parikh, and K. Grauman, “Fashion++: Minimal edits for outfit improvement,” in International conference on computer vision. Montreal, Canada: IEEE, 2019, pp. 5046–5055, DOI: 10.1109/ICCV.2019.00515.
- (19) M. I. Vasileva, B. A. Plummer, K. Dusad, S. Rajpal, R. Kumar, and D. Forsyth, “Learning type-aware embeddings for fashion compatibility,” in European conference on computer vision. Munich, Germany: Springer, 2018, pp. 405–421, DOI: 10.1007/978-3-030-01270-0_24.
- (20) D.-P. Fan, M. Zhuge, and L. Shao, “Domain specific pre-training of cross modality transformer model,” 2022, uS Patent App. 17/186,745.
- (21) W. Wang, E. Xie, X. Li, D.-P. Fan, K. Song, D. Liang, T. Lu, P. Luo, and L. Shao, “Pyramid vision transformer: A versatile backbone for dense prediction without convolutions,” in International conference on computer vision. Montreal, Canada: IEEE, 2021, pp. 548–558, DOI: 10.1109/ICCV48922.2021.00061.
- (22) X. Yang, H. Zhang, D. Jin, Y. Liu, C.-H. Wu, J. Tan, D. Xie, J. Wang, and X. Wang, “Fashion captioning: Towards generating accurate descriptions with semantic rewards,” in European conference on computer vision. Glasgow, United Kingdom: Springer, 2020, pp. 1–17, DOI: 10.1007/978-3-030-58601-0_1.
- (23) Z. Al-Halah and K. Grauman, “From paris to berlin: Discovering fashion style influences around the world,” in Conference on computer vision and pattern recognition. Seattle, WA, USA: IEEE, 2020, pp. 10 133–10 142, DOI: 10.1109/CVPR42600.2020.01015.
- (24) H. Tan and M. Bansal, “Lxmert: Learning cross-modality encoder representations from transformers,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. Hong Kong, China: Association for Computational Linguistics, 2019, p. 5100–5111.
- (25) W. Su, X. Zhu, Y. Cao, B. Li, L. Lu, F. Wei, and J. Dai, “Vl-bert: Pre-training of generic visual-linguistic representations,” in International Conference on Learning Representations. [Online]: PMLR, 2020.
- (26) K.-H. Lee, X. Chen, G. Hua, H. Hu, and X. He, “Stacked cross attention for image-text matching,” in European conference on computer vision. Munich, Germany: Springer, 2018, pp. 212–228, DOI: 10.1007/978-3-030-01225-0_13.
- (27) Z. Niu, M. Zhou, L. Wang, X. Gao, and G. Hua, “Hierarchical multimodal lstm for dense visual-semantic embedding,” in International conference on computer vision. Venice, Italy: IEEE, 2017, pp. 1899–1907, DOI: 10.1109/ICCV.2017.208.
- (28) J. Xia, M. Zhuge, T. Geng, S. Fan, Y. Wei, Z. He, and F. Zheng, “Skating-mixer: Multimodal mlp for scoring figure skating,” [Online], 2022, Available: https://arxiv.org/abs/2203.03990.
- (29) X. Li, X. Yin, C. Li, P. Zhang, X. Hu, L. Zhang, L. Wang, H. Hu, L. Dong, F. Wei et al., “Oscar: Object-semantics aligned pre-training for vision-language tasks,” in European conference on computer vision. Glasgow, United Kingdom: Springer, 2020, pp. 121–137, DOI: 10.1007/978-3-030-58577-8_8.
- (30) M. Zhuge, D.-P. Fan, N. Liu, D. Zhang, D. Xu, and L. Shao, “Salient object detection via integrity learning,” Transactions on pattern analysis and machine intelligence, 2022, DOI: 10.1109/TPAMI.2022.3179526.
- (31) K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhudinov, R. Zemel, and Y. Bengio, “Show, attend and tell: Neural image caption generation with visual attention,” in International conference on machine learning. Lille, France: PMLR, 2015, pp. 2048–2057.
- (32) T. Arici, M. S. Seyfioglu, T. Neiman, Y. Xu, S. Train, T. Chilimbi, B. Zeng, and I. Tutar, “Mlim: Vision-and-language model pre-training with masked language and image modeling,” [Online], 2021, Available: https://arxiv.org/abs/2109.12178.
- (33) H. Bao, L. Dong, and F. Wei, “BEiT: BERT Pre-Training of Image Transformers,” in ICLR, 2022, Available: https://openreview.net/forum?id=p-BhZSz59o4.
- (34) K. He, X. Chen, S. Xie, Y. Li, P. Dollár, and R. Girshick, “Masked autoencoders are scalable vision learners,” in Conference on computer vision and pattern recognition. New Orleans, LA, USA: IEEE, 2022, pp. 15 979–15 988, DOI: 10.1109/CVPR52688.2022.01553.
- (35) Z. Huang, Z. Zeng, B. Liu, D. Fu, and J. Fu, “Pixel-bert: Aligning image pixels with text by deep multi-modal transformers,” [Online], 2020, Available: https://arxiv.org/abs/2004.00849.
- (36) X. Lin, G. Bertasius, J. Wang, S.-F. Chang, D. Parikh, and L. Torresani, “Vx2text: End-to-end learning of video-based text generation from multimodal inputs,” in Conference on computer vision and pattern recognition. Nashville, TN, USA: IEEE, 2021, pp. 7001–7011, DOI: 10.1109/CVPR46437.2021.00693.
- (37) W. Kim, B. Son, and I. Kim, “Vilt: Vision-and-language transformer without convolution or region supervision,” in International conference on machine learning. PMLR, 2021, pp. 5583–5594.
- (38) M. Yan, H. Xu, C. Li, B. Bi, J. Tian, M. Gui, and W. Wang, “Grid-vlp: Revisiting grid features for vision-language pre-training,” [Online], 2021, Available: https://arxiv.org/abs/2108.09479.
- (39) Z. Huang, Z. Zeng, Y. Huang, B. Liu, D. Fu, and J. Fu, “Seeing out of the box: End-to-end pre-training for vision-language representation learning,” in Conference on computer vision and pattern recognition. Nashville, TN, USA: IEEE, 2021, pp. 12 971–12 980, DOI: 10.1109/CVPR46437.2021.01278.
- (40) S. Goenka, Z. Zheng, A. Jaiswal, R. Chada, Y. Wu, V. Hedau, and P. Natarajan, “Fashionvlp: Vision language transformer for fashion retrieval with feedback,” in Conference on computer vision and pattern recognition. New Orleans, LA, USA: IEEE, 2022, pp. 14 085–14 095, DOI: 10.1109/CVPR52688.2022.01371.
- (41) J. Lei, L. Li, L. Zhou, Z. Gan, T. L. Berg, M. Bansal, and J. Liu, “Less is more: Clipbert for video-and-language learning via sparse sampling,” in Conference on computer vision and pattern recognition. Nashville, TN, USA: IEEE, 2021, pp. 7327–7337, DOI: 10.1109/CVPR46437.2021.00725.
- (42) H. Xu, M. Yan, C. Li, B. Bi, S. Huang, W. Xiao, and F. Huang, “E2e-vlp: End-to-end vision-language pre-training enhanced by visual learning,” in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing. Association for Computational Linguistics, 2021, p. 503–513.
- (43) H. Akbari, L. Yuan, R. Qian, W.-H. Chuang, S.-F. Chang, Y. Cui, and B. Gong, “Vatt: Transformers for multimodal self-supervised learning from raw video, audio and text,” in Advances in neural information processing systems, vol. 34. Curran Associates, Inc., 2021, pp. 24 206–24 221.
- (44) X. Yi, J. Yang, L. Hong, D. Z. Cheng, L. Heldt, A. Kumthekar, Z. Zhao, L. Wei, and E. Chi, “Sampling-bias-corrected neural modeling for large corpus item recommendations,” in Proceedings of the 13th ACM Conference on Recommender Systems. New York, NY, USA: Association for Computing Machinery, 2019, p. 269–277, DOI: 10.1145/3298689.3346996.
- (45) O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer Assisted Intervention. Munich, Germany: Springer, 2015, pp. 234–241, DOI: 10.1007/978-3-319-24574-4_28.
- (46) C. Alberti, J. Ling, M. Collins, and D. Reitter, “Fusion of detected objects in text for visual question answering,” in 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference. Hong Kong, China: Association for Computational Linguistics, 2019, pp. 2131–2140.
- (47) N. Rostamzadeh, S. Hosseini, T. Boquet, W. Stokowiec, Y. Zhang, C. Jauvin, and C. Pal, “Fashion-gen: The generative fashion dataset and challenge,” in International conference on machine learning Workshops, 2018.
- (48) R. Kiros, R. Salakhutdinov, and R. S. Zemel, “Unifying visual-semantic embeddings with multimodal neural language models,” [Online], 2014, Available: https://arxiv.org/abs/1411.2539.
- (49) F. Faghri, D. J. Fleet, J. R. Kiros, and S. Fidler, “Vse++: Improving visual-semantic embeddings with hard negatives,” in British Machine Vision Conference. Newcastle, UK: BMVA Press, 2018.
- (50) Y. Wang, H. Yang, X. Qian, L. Ma, J. Lu, B. Li, and X. Fan, “Position focused attention network for image-text matching,” [Online], 2019, Available: https://arxiv.org/abs/1907.09748.
- (51) J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Conference on computer vision and pattern recognition. Miami, FL, USA: IEEE, 2009, pp. 248–255, DOI: 10.1109/CVPR.2009.5206848.
- (52) A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark et al., “Learning transferable visual models from natural language supervision,” in International conference on machine learning. PMLR, 2021, pp. 8748–8763.
- (53) T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European conference on computer vision. Zurich, Switzerland: Springer, 2014, pp. 740–755, DOI: 10.1007/978-3-319-10602-1_48.
- (54) G. Li, N. Duan, Y. Fang, M. Gong, D. Jiang, and M. Zhou, “Unicoder-vl: A universal encoder for vision and language by cross-modal pre-training,” in AAAI Conference on Artificial Intelligence. New York, NY, USA: AAAI Press, 2020, pp. 11 336–11 344.
- (55) L. Y. Wu, D. Liu, X. Guo, R. Hong, L. Liu, and R. Zhang, “Multi-scale spatial representation learning via recursive hermite polynomial networks,” in Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence. Messe Wien, Vienna, Austria: International Joint Conferences on Artificial Intelligence Organization, 2022, pp. 1465–1473, DOI: 10.24963/ijcai.2022/204.
- (56) D. Chen and et al., “Cross-modal retrieval with heterogenous graph embedding,” in Proceedings of the 30th ACM International Conference on Multimedia. Lisboa, Portugal: Association for Computing Machinery, 2022, p. 3291–3300, DOI: 10.1145/3503161.3548195.
- (57) D. Liu, L. Wu, F. Zheng, L. Liu, and M. Wang, “Verbal-person nets: Pose-guided multi-granularity language-to-person generation,” Transactions on Neural Networks and Learning Systems, 2022, DOI: 10.1109/TNNLS.2022.3151631.
- (58) Z. Zhang, H. Luo, L. Zhu, G. Lu, and H. T. Shen, “Modality-invariant asymmetric networks for cross-modal hashing,” Transactions on Knowledge and Data Engineering, 2022, DOI: 10.1109/TKDE.2022.3144352.