This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

RaSa: Relation and Sensitivity Aware Representation Learning
for Text-based Person Search

Yang Bai1    Min Cao1 Corresponding author    Daming Gao1    Ziqiang Cao1    Chen Chen2   
Zhenfeng Fan3
   Liqiang Nie4&Min Zhang1,4 1School of Computer Science and Technology, Soochow University
2Institute of Automation, Chinese Academy of Sciences
3Institute of Computing Technology, Chinese Academy of Sciences
4Harbin Institute of Technology, Shenzhen [email protected], [email protected]
Abstract

Text-based person search aims to retrieve the specified person images given a textual description. The key to tackling such a challenging task is to learn powerful multi-modal representations. Towards this, we propose a Relation and Sensitivity aware representation learning method (RaSa), including two novel tasks: Relation-Aware learning (RA) and Sensitivity-Aware learning (SA). For one thing, existing methods cluster representations of all positive pairs without distinction and overlook the noise problem caused by the weak positive pairs where the text and the paired image have noise correspondences, thus leading to overfitting learning. RA offsets the overfitting risk by introducing a novel positive relation detection task (i.e., learning to distinguish strong and weak positive pairs). For another thing, learning invariant representation under data augmentation (i.e., being insensitive to some transformations) is a general practice for improving representation’s robustness in existing methods. Beyond that, we encourage the representation to perceive the sensitive transformation by SA (i.e., learning to detect the replaced words), thus promoting the representation’s robustness. Experiments demonstrate that RaSa outperforms existing state-of-the-art methods by 6.94%, 4.45% and 15.35% in terms of Rank@1 on CUHK-PEDES, ICFG-PEDES and RSTPReid datasets, respectively. Code is available at: https://github.com/Flame-Chasers/RaSa.

Refer to caption
Figure 1: Illustration of (a) two types of positive relations for relation-aware learning, where the noise interference in the weak positive pairs is highlighted in red, (b) replaced token detection for sensitivity-aware learning, in which word replacement is used as the sensitive transformation and the replaced words are marked in bold.

1 Introduction

Text-based person search Li et al. (2017b); Wang et al. (2021a) aims at retrieving the person images in a large-scale person image pool given a query of textual description about that person. This task is related to person re-identification Ji et al. (2021); Wang et al. (2022b) and text-image retrieval Cao et al. (2022); Li et al. (2021a), which have been very active research topics in recent years. It, however, exhibits unique characteristics and challenges. Compared to person re-identification with image queries, text-based person search with more accessible open-form text queries provides a more user-friendly searching procedure while embracing greater challenges due to the cross-modal search. In addition, compared to general image-text retrieval, text-based person search focuses on cross-modal retrieval specific for the person with more fine-grained details, tending to larger intra-class variance as well as smaller inter-class variance, which toughly bottlenecks the retrieval performance.

Targeting learning powerful feature representation and achieving cross-modal alignment for text-based person search, researchers have developed a batch of technologies over the past few years Wu et al. (2021); Shao et al. (2022). It has been proved that the model armed with reasonable tasks tends to learn better representation. In this paper, we propose a representation learning method, namely RaSa, with two novel tasks: relation-aware learning and sensitivity-aware learning for text-based person search.

Relation-aware learning. In existing methods Han et al. (2021); Li et al. (2022b), the de facto optimization objective is to bring image and text representations of the same identity (i.e., positive pairs) together and repel representations of different identities (i.e., negative pairs) away. However, it tends to encounter the following issue. Normally, a textual description is generated by annotating a particular single image in the text-based person search dataset. The text strongly matches the annotated image without a doubt, whereas it is not always well-aligned to other positive images of the same person at the semantic level due to intra-class variation in the image. As shown in Figure 1 (a), the images and texts depict the same person, leading to a positive relation for each image-text pair. However, there exist two different types of positive relations. text1 (resp. text2) is the exact description of image1 (resp. image2), where they are completely matched and form a strong positive pair. Nevertheless, image1 and text2 (resp. image2 and text1) constitute a weak positive pair with the noise interference. For instance, “white t-shirt” and “blue shorts” in text1 correspond to non-existent objects in image2 due to the occlusion. Existing methods endow the strong and weak positive pairs with equal weight in learning representations, regardless of the noise problem from the weak pairs, eventually leading to overfitting learning.

In order to mitigate the impacts of the noise interference from weak positive pairs, we propose a Relation-Aware learning (RA) task, which is composed of a probabilistic Image-Text Matching (pp-ITM) task and a Positive Relation Detection (PRD) task. pp-ITM is a variant of the commonly-used ITM, aiming to distinguish negative and positive pairs with a probabilistic strong or weak positive inputting, while PRD is designed to explicitly makes a distinction between the strong and weak positive pairs. Therein, pp-ITM emphasizes the consistency between strong and weak positive pairs, whereas PRD highlights their difference and can be regarded as the regularization of pp-ITM. The model armed with RA can not only learn valuable information from weak positive pairs by pp-ITM but also alleviate noise interference from them by PRD, eventually reaching a trade-off.

Sensitivity-aware learning. Learning invariant representations under a set of manually chosen transformations (also called insensitive transformations in this context) is a general practice for improving the robustness of representation in the existing methods Caron et al. (2020); Chen and He (2021). We recognize it but there is more. Inspired by the recent success of equivariant contrastive learning Dangovski et al. (2022), we explore the sensitive transformation that would hurt performance when applied to learn transformation-invariant representations. Rather than keeping invariance under insensitive transformation, we encourage the learned representations to have the ability to be aware of the sensitive transformation.

Towards this end, we propose a Sensitivity-Aware learning (SA) task. We adopt the word replacement as the sensitive transformation and develop a Momentum-based Replaced Token Detection (mm-RTD) pretext task to detect whether a token comes from the original textual description or the replacement, as shown in Figure 1 (b). The closer the replaced word is to the original one (i.e., more confusing word), the more difficult this detection task is. When the model is trained to well solve such a detection task, it is expected to have the ability to learn better representation. With these in mind, we use Masked Language Modeling (MLM) to perform the word replacement, which utilizes the image and the text contextual tokens to predict the masked tokens. Furthermore, considering that the momentum model, a slow-moving average of the online model, can learn more stable representations than the current online model Grill et al. (2020) to generate more confusing words, we employ MLM from the momentum model to carry out the word replacement. Overall, MLM and mm-RTD together form a Sensitivity-Aware learning (SA), which offers powerful surrogate supervision for representation learning.

Our contributions can be summarized as follows:

  • We differentiate between strong and weak positive image-text pairs in learning representation and propose a relation-aware learning task.

  • We pioneer the idea of learning representation under the sensitive transformation to the text-based person search and develop a sensitivity-aware learning task.

  • Extensive experiments demonstrate RaSa outperforms existing state-of-the-art methods by 6.946.94%, 4.454.45% and 15.3515.35% in terms of Rank@1 metric on CUHK-PEDES, ICFG-PEDES and RSTPReid datasets, respectively.

Refer to caption
Figure 2: Model architecture of RaSa. It consists of an image encoder, a text encoder and a cross-modal encoder. An intra- and cross-modal CL task is attached after the unimodal encoders for unimodal representation learning. RA and SA tasks are tied after the cross-modal encoders for multi-modal representation learning. The momentum model (a slow-moving of the online model) is used to guide the online model to learn better representations.

2 Related Work

2.0.1 Text-based Person Search

Li et al. Li et al. (2017b) first introduce the text-based person search task and publish a challenging dataset CUHK-PEDES. Following this, a series of methods are proposed to solve this task. Part of methods Zheng et al. (2020a); Wang et al. (2021a) focus on designing a reasonable cross-modal alignment strategy, while others Zhang and Lu (2018); Shao et al. (2022) concentrate on learning powerful feature representation. For cross-modal alignment, it begins with global alignment Zheng et al. (2020b) or local correspondences (e.g., patch-word or region-phrase correspondences) Chen et al. (2022); Niu et al. (2020), and evolves into self-adaptively learning semantic alignment across different granularity Li et al. (2022b); Gao et al. (2021). Beyond that, some works Wang et al. (2020); Zhu et al. (2021) utilize external technologies (e.g., human segmentation, pose estimation or attributes prediction) to assist with the cross-modal alignment. For representation learning, Wu et al. Wu et al. (2021) propose two color-related tasks based on the observation that color plays a key role in text-based person search. Zeng et al. Zeng et al. (2021) develop three auxiliary reasoning tasks with gender classification, appearance similarity and image-to-text generation. Ding et al. Ding et al. (2021) firstly notice the noise interference from weak positive pairs and propose to keep the difference between strong and weak positive pairs by manually assigning different margins in the triplet loss. More recently, some works Han et al. (2021); Shu et al. (2022); Yan et al. (2022) resort to vision-language pretraining models to learn better representations. In this paper, we design two novel tasks: RA and SA. RA detects the type of the positive pair to weaken noise from weak positive pairs, differently from the method Ding et al. (2021) with the sophisticated trick. SA focuses on representation learning by detecting sensitive transformation, which is under-explored in the previous methods.

2.0.2 Equivariant Contrastive Learning

Different from contrastive learning He et al. (2020) that aims to learn transformation-insensitive representations, equivariant contrastive learning Dangovski et al. (2022) is recently proposed by additionally encouraging the learned representations to have the ability to be aware of sensitive transformations. Mathematically, the notions of insensitivity and sensitivity can be inductively summarized as: f(T(x))=T(f(x))f(T(x))=T^{\prime}(f(x)) where TT denotes a group of transformations of an input instance xx, and ff is an encoder to compute the representation of xx. When TT^{\prime} is the identity transformation, it can be said that ff is trained to be insensitive to TT; otherwise, ff is sensitive to TT. Equivariant contrastive learning has shown its successful application in the fields of computer vision (CV) Dangovski et al. (2022) and natural language processing (NLP) Chuang et al. (2022), which inspires us to explore sensitive transformations for learning high-quality representations in the cross-modal retrieval task. In this paper, we develop a sensitivity-aware learning with MLM-based word replacement as the sensitive transformation to encourage the model to perceive the replaced words, thus obtaining more informative and discriminative representations.

3 Method

In this section, we take ALBEF Li et al. (2021a) as the backbone111Experiments on more backbones are shown in Appendix A.4. and elaborate on the proposed method RaSa by introducing the modal architecture in Section 3.1 and the optimization objectives involving the proposed RA and SA tasks in Section 3.2.

3.1 Model Architecture

As illustrated in Figure 2, the proposed RaSa consists of two unimodal encoders and a cross-modal encoder. We adopt 1212-layer and 66-layer transformer blocks for the image and text encoders, respectively. The cross-modal encoder comprises 66-layer transformer blocks, where a cross-attention module is added after the self-attention module in each block. Considering that the textual description usually covers a part of the information in the corresponding image, we employ a text-guided asymmetric cross-attention module in the cross-modal encoder, i.e., using the textual representation as query and the visual one as key and value. Simultaneously, we maintain a momentum version of the online model via Exponential Moving Average (EMA). Specifically, EMA is formulated as θ^=mθ^+(1m)θ\hat{\theta}=m\hat{\theta}+(1-m)\theta, where θ^\hat{\theta} and θ\theta are the parameters of the momentum and online models, respectively, and m[0,1]m\in[0,1] is a momentum coefficient. The momentum model presents a delayed and more stable version of the online model and is used to guide the online model to learn better representations.

Given an image-text pair (I,T)(I,T), we first feed the image II into the image encoder to obtain a sequence of visual representations {vcls,v1,,vM}\{v_{cls},v_{1},\cdots,v_{M}\} with vclsv_{cls} being the global visual representation and viv_{i} (i=1,,M)(i=1,\cdots,M) being the patch representation. Similarly, we obtain a sequence of textual representations {tcls,t1,,tN}\{t_{cls},t_{1},\cdots,t_{N}\} by feeding the text TT into the text encoder, where tclst_{cls} is the global textual representation and tit_{i} (i=1,,N)(i=1,\cdots,N) is the token representation. The visual and textual representations are then fed to the cross-modal encoder to obtain a sequence of multi-modal representations {fcls,f1,,fN}\{f_{cls},f_{1},\cdots,f_{N}\}, where fclsf_{cls} denotes the joint representation of II and TT, and fif_{i} (i=1,,N)(i=1,\cdots,N) can be regarded as the joint representation of the image II and the ii-th token in the text TT. Simultaneously, the momentum model is employed to obtain a sequence of momentum representations.

3.2 Optimization Objectives

3.2.1 Relation-aware Learning

The vanilla widely-used ITM predicts whether an inputted image-text pair is positive or negative, defined as:

Litm=𝔼p(I,T)(yitm,ϕitm(I,T)),L_{itm}=\mathbb{E}_{p(I,T)}{\mathcal{H}}(y^{itm},\phi^{itm}(I,T)), (1)

where \mathcal{H} represents a cross-entropy function, yitmy^{itm} is a 22-dimension one-hot vector representing the ground-truth label (i.e., [0,1]\left[0,1\right]^{\top} for the positive pair, and [1,0]\left[1,0\right]^{\top} for the negative pair), and ϕitm(I,T)\phi^{itm}(I,T) is the predicted matching probability of the pair that is computed by feeding fclsf_{cls} into a binary classifier, a fully-connected layer followed by a softmax function.

However, it is unreasonable to directly adopt the vanilla ITM in text-based person search. On the one hand, there exists noise interference from weak positive pairs, which would hamper the representation learning. On the other hand, the weak positive pairs contain certain valuable alignment information that can facilitate representation learning. As a result, to reach a balance, we retain a proportion of weak positive pairs in ITM by introducing the probabilistic inputting. Specifically, we input the weak positive pair with a small probability of pwp^{w} and the strong positive pair with a probability of 1pw1-p^{w}. To distinguish with the vanilla ITM, we denote the proposed probabilistic ITM as pp-ITM.

Furthermore, we continue to alleviate the noise effect of the weak pairs. We propose a Positive Relation Detection (PRD) pretext task to detect the type of the positive pair (i.e., strong or weak), which is formulated as:

Lprd=𝔼p(I,Tp)(yprd,ϕprd(I,Tp)),L_{prd}=\mathbb{E}_{p({I},{T^{p}})}{\mathcal{H}}(y^{prd},\phi^{prd}({I},{T^{p}})), (2)

where (I,Tp)({I},{T^{p}}) denotes a positive pair, yprdy^{prd} is the ground truth label (i.e., [1,0]\left[1,0\right]^{\top} for the strong positive pair and [0,1]\left[0,1\right]^{\top} for the weak pair), and ϕprd(I,Tp)\phi^{prd}({I},{T^{p}}) is the predicted probability of the pair which is computed by appending a binary classifier to the joint representation fclsf_{cls} of the pair.

Taken together, we define the Relation-Aware learning (RA) task as:

Lra=Lp-itm+λ1Lprd,L_{ra}=L_{p\raisebox{0.0pt}{-}itm}+\lambda_{1}L_{prd}, (3)

where the weight λ1\lambda_{1} is a hyper-parameter.

During the process of the optimization, pp-ITM focuses on the consistency between strong and weak positive pairs, while PRD highlights their difference. In essence, PRD plays a role of a regularized compensation for pp-ITM. As a whole, RA achieves a trade-off between the benefits of the weak pair and the risk of its side effects.

3.2.2 Sensitivity-aware Learning

Learning invariant representations under the insensitive transformation of data is a common way to enhance the robustness of the learned representations. We go beyond it and propose to learn representations that are aware of the sensitive transformation. Specifically, we adopt the MLM-based word replacement as the sensitive transformation and propose a Momentum-based Replaced Token Detection (mm-RTD) pretext task to detect (i.e., being aware of) the replacement.

Refer to caption
Figure 3: Illustration of mm-RTD. It aims to detect whether a token is from the original textual description or the replacement with the aid of the information of the contextual tokens and the paired image. The text with word replacement is obtained by the result of the Masked Language Modeling (MLM) from the momentum model.

Given a strong positive pair (I,Ts)(I,T^{s}), MLM loss is formulated as:

Lmlm=𝔼p(I,Tmsk)(ymlm,ϕmlm(I,Tmsk)),L_{mlm}=\mathbb{E}_{p(I,T^{msk})}{\mathcal{H}}(y^{mlm},\phi^{mlm}(I,T^{msk})), (4)

where TmskT^{msk} is a masked text in which each token in the input text TsT^{s} is randomly masked with a probability of pmp^{m}, ymlmy^{mlm} is a one-hot vector denoting the ground truth of the masked token and ϕmlm(I,Tmsk)\phi^{mlm}(I,T^{msk}) is the predicted probability for the masked token based on the information of the contextual text TmskT^{msk} and the paired image II.

We use the result of MLM from the momentum model as the word replacement, denoted as TrepT^{rep}. The momentum model is a slow-moving of the online model and can learn more stable representations. Therefore, the momentum model is expected to generate more confusing tokens. As mm-RTD detects such challenging tokens well, the model is motivated to learn more informative representations to distinguish the tiny differences. Remarkably, besides serving as a generator for the word replacement, MLM also plays a role of token-level optimization, promoting fine-grained representation learning.

Next, mm-RTD performs a detection of the MLM-based token replacement. Specifically, the pair (I,Trep)(I,T^{rep}) is inputted to the model to obtain a sequence of multi-modal representations {fcls,f1,,fN}\{f_{cls},f_{1},...,f_{N}\}, and a binary classifier works on {f1,,fN}\{f_{1},...,f_{N}\} to predict whether the ii-th token is replaced or not. mm-RTD minimizes a cross-entropy loss:

Lm-rtd=𝔼p(I,Trep)(ym-rtd,ϕm-rtd(I,Trep)),L_{m\raisebox{0.0pt}{-}rtd}=\mathbb{E}_{p(I,T^{rep})}{\mathcal{H}}(y^{m\raisebox{0.0pt}{-}rtd},\phi^{m\raisebox{0.0pt}{-}rtd}(I,T^{rep})), (5)

where ym-rtdy^{m\raisebox{0.0pt}{-}rtd} is a one-hot vector denoting the ground truth of the replaced token and ϕm-rtd(I,Trep)\phi^{m\raisebox{0.0pt}{-}rtd}(I,T^{rep}) is the predicted replacement probability. We illustrate the pipeline of mm-RTD in Figure 3 for clarity.

Overall, Sensitivity-Aware learning (SA) loss is defined as:

Lsa=Lmlm+λ2Lm-rtd,L_{sa}=L_{mlm}+\lambda_{2}L_{m\raisebox{0.0pt}{-}rtd}, (6)

where the weight λ2\lambda_{2} is a hyper-parameter.

In conclusion, RA works on the global representation fclsf_{cls} and mainly focuses on the correlation between the image and text, which can be regarded as a coarse-grained optimization. As a complement, SA acts on the token representations {f1,,fN}\{f_{1},...,f_{N}\} and pays more attention to the interaction between the image and textual tokens, exhibiting a fine-grained optimization. The two complementary tasks effectively facilitate representation learning.

3.2.3 Contrastive Learning

The proposed RA and SA are directly applied on the multi-modal representations from the cross-modal encoder. Furthermore, we introduce an intermediate Contrastive Learning task (CL) on the representations from the unimodal encoders, so as to make the subsequent cross-modal fusion easier to perform multi-modal representation learning.

Given an image-text pair (I,T)(I,T), we feed it into the unimodal encoders and obtain the global visual and textual representations vclsv_{cls} and tclst_{cls}. Then a linear layer is applied to project them to lower-dimensional representations vclsv^{\prime}_{cls} and tclst^{\prime}_{cls}. Meanwhile, we obtain the output of momentum unimodal encoders, denoted as v^cls\hat{v}^{\prime}_{cls} and t^cls\hat{t}^{\prime}_{cls}. We maintain an image queue Q^v\hat{Q}_{v} and a text queue Q^t\hat{Q}_{t} to store the recent RR projected representations v^cls\hat{v}^{\prime}_{cls} and t^cls\hat{t}^{\prime}_{cls}, similarly to MoCo He et al. (2020). The introduction of the queues implicitly enlarges the batch size, and a larger batch will provide more negative samples, thereby facilitating representation learning.

In CL, the general form of InfoNCE loss is formulated as:

Lnce(x,x+,Q)=𝔼p(x,x+)[logexp(s(x,x+)/τ)xiQexp(s(x,xi)/τ)],L_{nce}(x,x_{+},Q)=-\mathbb{E}_{p(x,x_{+})}[\log\frac{\exp(s(x,x_{+})/\tau)}{\sum\limits_{x_{i}\in Q}\exp(s(x,x_{i})/\tau)}], (7)

where τ\tau is a learnable temperature parameter, QQ denotes a maintained queue, and s(x,x+)=xTx+/xx+s(x,x_{+})=x^{\rm T}x_{+}/\lVert x\rVert\lVert x_{+}\rVert measures the cosine similarity between xx and x+x_{+}.

Beyond the widely-used cross-modal image-text contrastive learning (ITC) Li et al. (2021a); Radford et al. (2021), denoted as:

Litc=[Lnce(vcls,t^cls,Q^t)+Lnce(tcls,v^cls,Q^v)]/ 2,L_{itc}=[L_{nce}(v^{\prime}_{cls},\hat{t}^{\prime}_{cls},\hat{Q}_{t})+L_{nce}(t^{\prime}_{cls},\hat{v}^{\prime}_{cls},\hat{Q}_{v})]\ /\ 2, (8)

we additionally explore the intra-modal contrastive learning (IMC). The representations of the same person are supposed to stay closer than those of different persons within each modality. IMC loss is formulated as:

Limc=[Lnce(vcls,v^cls,Q^v)+Lnce(tcls,t^cls,Q^t)]/ 2.L_{imc}=[L_{nce}(v^{\prime}_{cls},\hat{v}^{\prime}_{cls},\hat{Q}_{v})+L_{nce}(t^{\prime}_{cls},\hat{t}^{\prime}_{cls},\hat{Q}_{t})]\ /\ 2. (9)

Taken together, we define the overall loss for CL as:

Lcl=(Litc+Limc)/ 2.L_{cl}=(L_{itc}+L_{imc})\ /\ 2. (10)

3.2.4 Joint Learning

Overall, we formulate the joint optimization objective as:

L=Lra+Lsa+λ3Lcl,L=L_{ra}+L_{sa}+\lambda_{3}L_{cl}, (11)

where λ3\lambda_{3} is a hyper-parameter.

During inference, given a query text and a large-scale image pool, we use the predicted matching probability from pp-ITM to rank all images. Considering the inefficiency of the cross-modal encoder with quadratic interaction operation, we refer to ALBEF Li et al. (2021a) and exclude a large number of irrelevant image candidates prior to the cross-modal encoder, thereby speeding up the inference. Specifically, we first calculate each pair’s similarity s(tcls,vcls)s(t_{cls},v_{cls}) via the unimodal encoders, and then select the first 128128 images with the highest similarities to send them to the cross-modal encoder and compute the pp-ITM matching probabilities for ranking.

4 Experiments

We conduct experiments on three text-based person search datasets: CUHK-PEDES Li et al. (2017b), ICFG-PEDES Ding et al. (2021) and RSTPReid Zhu et al. (2021). The introduction of each dataset and the implementation details of the proposed method are shown in Appendix A.1 and A.2, respectively.

4.1 Evaluation Protocol

We adopt the widely-used Rank@K (R@K for short, K=1,5,101,5,10) metric to evaluate the performance of the proposed method. Specifically, given a query text, we rank all the test images via the similarity with the text and the search is deemed to be successful if top-K images contain any corresponding identity. R@K is the percentage of successful searches. We also adopt the mean average precision (mAP) as a complementary metric.

Method R@1 R@5 R@10 mAP
w/o VLP GNA-RNN Li et al. (2017b) 19.05 - 53.64 -
Dual Path Zheng et al. (2020b) 44.40 66.26 75.07 -
CMPM/C Zhang and Lu (2018) 49.37 71.69 79.27 -
ViTAA Wang et al. (2020) 55.97 75.84 83.52 -
DSSL Zhu et al. (2021) 59.98 80.41 87.56 -
MGEL Wang et al. (2021a) 60.27 80.01 86.74 -
ACSA Ji et al. (2022) 63.56 81.40 87.70 -
SAF Li et al. (2022b) 64.13 82.62 88.40 58.61
TIPCB Chen et al. (2022) 64.26 83.19 89.10 -
CAIBC Wang et al. (2022c) 64.43 82.87 88.37 -
C2A2\rm C_{2}A_{2} Niu et al. (2022) 64.82 83.54 89.77 -
LGUR Shao et al. (2022) 65.25 83.12 89.00 -
w/ VLP PSLD Han et al. (2021) 64.08 81.73 88.19 60.08
IVT Shu et al. (2022) 65.59 83.11 89.21 -
CFine Yan et al. (2022) 69.57 85.93 91.15 -
ALBEF(backbone) Li et al. (2021a) 60.28 79.52 86.34 56.67
RaSa (Ours) 76.51 90.29 94.25 69.38
Table 1: Comparison with other methods on CUHK-PEDES. VLP denotes vision-language pretraining. For a fair comparison, all reported results come from the methods without re-ranking.

4.2 Backbones

Most text-based person search methods Li et al. (2022b); Shao et al. (2022) rely on two feature extractors pre-trained on unaligned images and texts separately, such as ResNet He et al. (2016) or ViT Dosovitskiy et al. (2020) for the visual extractor, Bi-LSTM Hochreiter and Schmidhuber (1997) or BERT Devlin et al. (2018) for the textual extractor. Recently, some works Shu et al. (2022); Yan et al. (2022) have applied vision-language pretraining (VLP) to text-based person search and obtained impressive results. Following this, we adopt VLP models as the backbone.

The proposed RaSa can be plugged into various backbones. To adequately verify the effectiveness, we conduct RaSa on three VLP models: ALBEFF Li et al. (2021a), TCL Yang et al. (2022) and CLIP Radford et al. (2021). We use ALBEF as the backbone by default in the following experiments, which is pre-trained on 1414M image-text pairs and adopts ITC and ITM tasks for image-text retrieval. The details and experiments on TCL and CLIP are shown in Appendix A.4.

Method R@1 R@5 R@10 mAP
w/o VLP Dual Path Zheng et al. (2020b) 38.99 59.44 68.41 -
CMPM/C Zhang and Lu (2018) 43.51 65.44 74.26 -
ViTAA Wang et al. (2020) 50.98 68.79 75.78 -
SSAN Ding et al. (2021) 54.23 72.63 79.53 -
SAF Li et al. (2022b) 54.86 72.13 79.13 32.76
TIPCB Chen et al. (2022) 54.96 74.72 81.89 -
SRCF Suo et al. (2022) 57.18 75.01 81.49 -
LGUR Shao et al. (2022) 59.02 75.32 81.56 -
w/ VLP IVT Shu et al. (2022) 56.04 73.60 80.22 -
CFine Yan et al. (2022) 60.83 76.55 82.42 -
ALBEF(backbone) Li et al. (2021a) 34.46 52.32 60.40 19.62
RaSa (Ours) 65.28 80.40 85.12 41.29
Table 2: Comparison with other methods on ICFG-PEDES.
Method R@1 R@5 R@10 mAP
w/o VLP DSSL Zhu et al. (2021) 32.43 55.08 63.19 -
SSAN Ding et al. (2021) 43.50 67.80 77.15 -
SAF Li et al. (2022b) 44.05 67.30 76.25 36.81
CAIBC Wang et al. (2022c) 47.35 69.55 79.00 -
ACSA Ji et al. (2022) 48.40 71.85 81.45 -
C2A2\rm C_{2}A_{2} Niu et al. (2022) 51.55 76.75 85.15 -
w/ VLP IVT Shu et al. (2022) 46.70 70.00 78.80 -
CFine Yan et al. (2022) 50.55 72.50 81.60 -
ALBEF(backbone) Li et al. (2021a) 50.10 73.70 82.10 41.73
RaSa (Ours) 66.90 86.50 91.35 52.31
Table 3: Comparison with other methods on RSTPReid.

4.3 Comparison with State-of-the-art Methods

We compare the proposed RaSa with the existing text-based person search methods on CUHK-PEDES, ICFG-PEDES and RSTPReid, as shown in Table 1, 2 and 3, respectively. RaSa achieves the highest performance in terms of all metrics, outperforming existing state-of-the-art methods by a large margin. Specifically, compared with the current best-performing method CFine Yan et al. (2022), RaSa gains a significant R@1 improvement of 6.946.94%, 4.454.45% and 15.3515.35% on the three datasets, respectively. The comparison clearly demonstrates the effectiveness of RaSa in text-based person search.

4.4 Ablation Study

We analyze the effectiveness and contribution of each optimization objective in RaSa by conducting a series of ablation experiments on CUHK-PEDES, as shown in Table 4.

4.4.1 Effectiveness of Optimization Objectives

RaSa consists of three optimization objectives. CL provides an explicit alignment before the cross-modal fusion. RA implements the deep fusion by the cross-modal encoder with an alleviation of noise interference. And SA encourages the learned representations to be sensitive to the MLM-based token replacement.

Refer to caption
Figure 4: The impact of the hyper-parameters at R@1 on CUHK-PEDES. pwp^{w} denotes the probability of inputting weak positive pairs in RA. pmp^{m} means the masking ratio of the tokens in a text in SA. λ1\lambda_{1}, λ2\lambda_{2} and λ3\lambda_{3} are the loss weights.

We can see from Table 4, (1) RaSa with a single CL achieves a modest performance of 61.3561.35% and 59.4459.44% in terms of R@1 and mAP, respectively. On account of the modality gap between the image and text and the fine-grained intra-class variation, CL contributes a coarse alignment with a lack of deep interaction across modalities, which is not enough to handle such a challenging retrieval task. (2) When adding RA(pp-ITM + PRD), the performance has a remarkable improvement of 12.8512.85% at R@1 and 8.678.67% at mAP, effectively demonstrating that deep cross-modal fusion with RA is extraordinarily significant to text-based person search. And (3) with the aid of SA(MLM + mm-RTD), RaSa achieves the best performance of 76.5176.51% at R@1 and 69.3869.38% at mAP. SA utilizes the visual information and the contextual token information of the corresponding text to detect whether a token has been replaced or not. In order to handle such a challenging detection task, the learned representations are encouraged to be powerful enough to distinguish the tiny difference between the original token and the replaced one.

Module Setting R@1 R@5 R@10 mAP
CL ITC + IMC 61.35 80.44 86.91 59.44
+RA ITM 71.29 86.70 91.46 67.82
ss-ITM 73.52 88.71 92.98 66.74
pp-ITM 72.58 87.98 92.51 68.29
ITM + PRD 73.03 87.75 92.45 68.45
pp-ITM + PRD 74.20 89.02 92.95 68.11
++SA MLM 74.81 89.85 93.66 68.32
MLM + ff-RTD 75.13 89.93 93.47 69.17
MLM + oo-RTD 75.99 90.21 94.09 69.35
MLM + mm-RTD 76.51 90.29 94.25 69.38
Table 4: Comparison of RaSa with different settings on CUHK-PEDES. ITM learns from all positive pairs without a probabilistic inputting. ss-ITM learns from only strong positive pairs and discards all weak positive pairs. pp-ITM uses a probabilistic inputting of strong and weak positive pairs. ff-RTD adopts DistilBERT Sanh et al. (2019) as a fixed generator to produce the replaced tokens. oo-RTD uses the online model as the generator, while mm-RTD is based on the momentum model.

4.4.2 Analysis of RA

RA contains pp-ITM and PRD, where the former focuses on the consistency between the strong and weak positive pairs, while the latter highlights their difference, serving as a regularization of pp-ITM.

The vanilla ITM learns from all positive pairs without the probabilistic inputting. However, there exists too much noise interference from weak positive pairs. Intuitively, we can discard all weak positives to get rid of the noise. ss-ITM only uses the strong positive pairs and gains a boost of 2.232.23% at R@1 compared to the vanilla ITM. Nevertheless, such a straightforward way ignores the weak supervision from the weak positives which is also beneficial to representation learning. To reach a trade-off between the benefits of the weak supervision and the risk of side effects, pp-ITM resorts to the probabilistic inputting and retains a small proportion of the weak positives. Compared with the vanilla ITM and ss-ITM, pp-ITM achieves an intermediate performance. Not surprisingly at all, the more noise there exists, the more it affects the retrieval performance. In order to alleviate the impact of the noise, we further propose PRD to perform an explicit distinction between the strong and weak positives, which serve as a regularization for pp-ITM. Significantly, no matter whether adding PRD to the vanilla ITM or pp-ITM, PRD can obtain consistent performance improvement, which powerfully demonstrates its effectiveness.

4.4.3 Analysis of SA

SA includes MLM and mm-RTD. MLM not only plays the role of generating the text with word replacement but also performs a token-level optimization. mm-RTD detects the replaced tokens by virtue of the visual information and the contextual token information.

Based on CL and RA, adding a single MLM without the replacement detection task brings a slight boost of 0.610.61% at R@1. Furthermore, we introduce the detection task and use the momentum model as the generator to produce the replaced tokens. In order to adequately investigate the effectiveness of the generator, we compare three different variants. (1) Following DiffCSE Chuang et al. (2022), we use DistilBERT Sanh et al. (2019) as a fixed generator for the word replacement, which is denoted as ff-RTD. From Table 4, RaSa with ff-RTD gains a modest performance of 75.1375.13% at R@1. We argue that the generated tokens from a fixed generator can be easily detected as the training advances and thus provides a limited effect on learning representation. (2) oo-RTD adopts the online model as the generator. RaSa with oo-RTD achieves a better performance of 75.99%75.99\% at R@1. Compared with ff-RTD, oo-RTD resorts to a dynamic generator which is optimized constantly during the whole training process and can produce more confusing tokens with the proceeding of the model’s training, effectively increasing the difficulty of replaced tokens detection and facilitating representation learning. And (3) mm-RTD adopts the momentum model as the generator and reaches the best performance of 76.5176.51% at R@1. The momentum model is a slow-moving of the online model and can obtain more stable representations. As the training goes ahead, the momentum model iteratively bootstraps MLM to generate more challenging tokens for detection, which encourages the learned representations to be powerful enough to distinguish the tiny difference and substantially improve results.

4.4.4 Hyper-parameters

In Section 3.2, we use the inputting probability pwp^{w} to retain a small proportion of weak positive pairs to alleviate the noise, the masking ratio pmp^{m} to randomly mask tokens to perform the replaced token detection, and the loss weights λ1\lambda_{1}, λ2\lambda_{2}, λ3\lambda_{3} to make a trade-off. We show how these hyper-parameters impact the performance of RaSa in Figure 4. (1) The best result is achieved at pw=0.1p^{w}=0.1. The inputting probability pwp^{w} in RA is introduced to seek a balance between the useful information and the noise from weak positives. A larger pwp^{w} may introduce too much noise, while a smaller pwp^{w} hinders the model from making full use of the useful information. (2) RaSa performs best at pm=0.3p^{m}=0.3. A larger pmp^{m} brings more perturbations to the text, making the detection task too difficult to be carried out. In contrast, when pmp^{m} goes smaller, SA will contribute less to representation learning. And (3) for the loss weights λ1\lambda_{1}, λ2\lambda_{2} and λ3\lambda_{3}, they present an overall trend of first increasing and then decreasing. Empirical results show that RaSa performs best when they are set as 0.50.5.

4.5 Extended Experiments and Visualization

To go a step further and validate the effectiveness of RaSa, we perform extended experiments on two coarse-grained image-text retrieval datasets (Flickr30K Plummer et al. (2015) and COCO Lin et al. (2014)), as well as two fine-grained datasets (CUB Reed et al. (2016) and Flowers Reed et al. (2016)). The experimental results are shown in Appendix A.3. Besides, we conduct a series of domain generalization experiments following LGUR Shao et al. (2022) in Appendix A.3 to verify the generalization ability of RaSa. These results clearly demonstrate the effectiveness and the generalization ability of RaSa.

For a qualitative analysis, we also present the retrieval visualization in Appendix A.5, vividly showing the excellent retrieval ability of RaSa.

5 Conclusion

In this paper, we propose a Relation and Sensitivity aware representation learning method (RaSa) for text-based person search, which contains two novel tasks, RA and SA, to learn powerful multi-modal representations. Given that the noise from the weak positive pairs tends to result in overfitting learning, the proposed RA utilizes an explicit detection between strong and weak positive pairs to highlight the difference, serving as a regularization of pp-ITM that focuses on their consistency. Beyond learning transformation-insensitive representations, SA encourages the sensitivity to MLM-based token replacement. Extensive experiments on multiple benchmarks demonstrate the effectiveness of RaSa.

Acknowledgments

This work is supported by the National Science Foundation of China under Grant NSFC 62002252, and is also partially supported by the National Science Foundation of China under Grant NSFC 62106165.

References

  • Cao et al. [2022] Min Cao, Shiping Li, Juntao Li, Liqiang Nie, and Min Zhang. Image-text retrieval: A survey on recent research and development. In Lud De Raedt, editor, Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, pages 5410–5417. International Joint Conferences on Artificial Intelligence Organization, 7 2022. Survey Track.
  • Caron et al. [2020] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33:9912–9924, 2020.
  • Chen and He [2021] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15750–15758, 2021.
  • Chen et al. [2020] Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and Jingjing Liu. Uniter: Universal image-text representation learning. In European conference on computer vision, pages 104–120. Springer, 2020.
  • Chen et al. [2022] Yuhao Chen, Guoqing Zhang, Yujiang Lu, Zhenxing Wang, and Yuhui Zheng. Tipcb: A simple but effective part-based convolutional baseline for text-based person search. Neurocomputing, 494:171–181, 2022.
  • Chuang et al. [2022] Yung-Sung Chuang, Rumen Dangovski, Hongyin Luo, Yang Zhang, Shiyu Chang, Marin Soljacic, Shang-Wen Li, Wen-tau Yih, Yoon Kim, and James Glass. DiffCSE: Difference-based contrastive learning for sentence embeddings. In Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 2022.
  • Dangovski et al. [2022] Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash Srivastava, Brian Cheung, Pulkit Agrawal, and Marin Soljacic. Equivariant self-supervised learning: Encouraging equivariance in representations. In International Conference on Learning Representations, 2022.
  • Devlin et al. [2018] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  • Ding et al. [2021] Zefeng Ding, Changxing Ding, Zhiyin Shao, and Dacheng Tao. Semantically self-aligned network for text-to-image part-aware person re-identification. arXiv preprint arXiv:2107.12666, 2021.
  • Dosovitskiy et al. [2020] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  • Gao et al. [2021] Chenyang Gao, Guanyu Cai, Xinyang Jiang, Feng Zheng, Jun Zhang, Yifei Gong, Pai Peng, Xiaowei Guo, and Xing Sun. Contextual non-local alignment over full-scale representation for text-based person search. arXiv preprint arXiv:2101.03036, 2021.
  • Grill et al. [2020] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271–21284, 2020.
  • Han et al. [2021] Xiao Han, Sen He, Li Zhang, and Tao Xiang. Text-based person search with limited data. In BMVC, 2021.
  • Harris [1954] Zellig S Harris. Distributional structure. Word, 10(2-3):146–162, 1954.
  • He et al. [2016] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • He et al. [2020] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9729–9738, 2020.
  • Hochreiter and Schmidhuber [1997] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  • Ji et al. [2021] Deyi Ji, Haoran Wang, Hanzhe Hu, Weihao Gan, Wei Wu, and Junjie Yan. Context-aware graph convolution network for target re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 1646–1654, 2021.
  • Ji et al. [2022] Zhong Ji, Junhua Hu, Deyin Liu, Lin Yuanbo Wu, and Ye Zhao. Asymmetric cross-scale alignment for text-based person search. IEEE Transactions on Multimedia, 2022.
  • Jia et al. [2021] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In International Conference on Machine Learning, pages 4904–4916. PMLR, 2021.
  • Jing et al. [2021] Ya Jing, Wei Wang, Liang Wang, and Tieniu Tan. Learning aligned image-text representations using graph attentive relational network. IEEE Transactions on Image Processing, 30:1840–1852, 2021.
  • Karpathy and Fei-Fei [2015] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3128–3137, 2015.
  • Klein et al. [2015] Benjamin Klein, Guy Lev, Gil Sadeh, and Lior Wolf. Associating neural word embeddings with deep image representations using fisher vectors. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4437–4446, 2015.
  • Lee et al. [2018] Kuang-Huei Lee, Xi Chen, Gang Hua, Houdong Hu, and Xiaodong He. Stacked cross attention for image-text matching. In Proceedings of the European conference on computer vision (ECCV), pages 201–216, 2018.
  • Li et al. [2017a] Shuang Li, Tong Xiao, Hongsheng Li, Wei Yang, and Xiaogang Wang. Identity-aware textual-visual matching with latent co-attention. In Proceedings of the IEEE International Conference on Computer Vision, pages 1890–1899, 2017.
  • Li et al. [2017b] Shuang Li, Tong Xiao, Hongsheng Li, Bolei Zhou, Dayu Yue, and Xiaogang Wang. Person search with natural language description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1970–1979, 2017.
  • Li et al. [2020] Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang, Houdong Hu, Li Dong, Furu Wei, et al. Oscar: Object-semantics aligned pre-training for vision-language tasks. In European Conference on Computer Vision, pages 121–137. Springer, 2020.
  • Li et al. [2021a] Junnan Li, Ramprasaath Selvaraju, Akhilesh Gotmare, Shafiq Joty, Caiming Xiong, and Steven Chu Hong Hoi. Align before fuse: Vision and language representation learning with momentum distillation. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 9694–9705. Curran Associates, Inc., 2021.
  • Li et al. [2021b] Wei Li, Can Gao, Guocheng Niu, Xinyan Xiao, Hao Liu, Jiachen Liu, Hua Wu, and Haifeng Wang. Unimo: Towards unified-modal understanding and generation via cross-modal contrastive learning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 2592–2607, 2021.
  • Li et al. [2022a] Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In ICML, 2022.
  • Li et al. [2022b] Shiping Li, Min Cao, and Min Zhang. Learning semantic-aligned feature representation for text-based person search. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 2724–2728. IEEE, 2022.
  • Lin et al. [2014] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740–755. Springer, 2014.
  • Loshchilov and Hutter [2019] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In International Conference on Learning Representations, 2019.
  • Mikolov et al. [2013] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems, 26, 2013.
  • Niu et al. [2020] Kai Niu, Yan Huang, Wanli Ouyang, and Liang Wang. Improving description-based person re-identification by multi-granularity image-text alignments. IEEE Transactions on Image Processing, 29:5542–5556, 2020.
  • Niu et al. [2022] Kai Niu, Linjiang Huang, Yan Huang, Peng Wang, Liang Wang, and Yanning Zhang. Cross-modal co-occurrence attributes alignments for person search by language. In Proceedings of the 30th ACM International Conference on Multimedia, pages 4426–4434, 2022.
  • Plummer et al. [2015] Bryan A Plummer, Liwei Wang, Chris M Cervantes, Juan C Caicedo, Julia Hockenmaier, and Svetlana Lazebnik. Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. In Proceedings of the IEEE international conference on computer vision, pages 2641–2649, 2015.
  • Radford et al. [2021] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pages 8748–8763. PMLR, 2021.
  • Reed et al. [2016] Scott Reed, Zeynep Akata, Honglak Lee, and Bernt Schiele. Learning deep representations of fine-grained visual descriptions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 49–58, 2016.
  • Sanh et al. [2019] Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108, 2019.
  • Sarafianos et al. [2019] Nikolaos Sarafianos, Xiang Xu, and Ioannis A Kakadiaris. Adversarial representation learning for text-to-image matching. In Proceedings of the IEEE/CVF international conference on computer vision, pages 5814–5824, 2019.
  • Shao et al. [2022] Zhiyin Shao, Xinyu Zhang, Meng Fang, Zhifeng Lin, Jian Wang, and Changxing Ding. Learning granularity-unified representations for text-to-image person re-identification. 2022.
  • Shu et al. [2022] Xiujun Shu, Wei Wen, Haoqian Wu, Keyu Chen, Yiran Song, Ruizhi Qiao, Bo Ren, and Xiao Wang. See finer, see more: Implicit modality alignment for text-based person retrieval. arXiv preprint arXiv:2208.08608, 2022.
  • Suo et al. [2022] Wei Suo, Mengyang Sun, Kai Niu, Yiqi Gao, Peng Wang, Yanning Zhang, and Qi Wu. A simple and robust correlation filtering method for text-based person search. In European Conference on Computer Vision, pages 726–742. Springer, 2022.
  • Wang et al. [2020] Zhe Wang, Zhiyuan Fang, Jun Wang, and Yezhou Yang. Vitaa: Visual-textual attributes alignment in person search by natural language. In European Conference on Computer Vision, pages 402–420. Springer, 2020.
  • Wang et al. [2021a] Chengji Wang, Zhiming Luo, Yaojin Lin, and Shaozi Li. Text-based person search via multi-granularity embedding learning. In IJCAI, pages 1068–1074, 2021.
  • Wang et al. [2021b] Chengji Wang, Zhiming Luo, Zhun Zhong, and Shaozi Li. Divide-and-merge the embedding space for cross-modality person search. Neurocomputing, 463:388–399, 2021.
  • Wang et al. [2022a] Chengji Wang, Zhiming Luo, Yaojin Lin, and Shaozi Li. Improving embedding learning by virtual attribute decoupling for text-based person search. Neural Computing and Applications, 34(7):5625–5647, 2022.
  • Wang et al. [2022b] Tao Wang, Hong Liu, Pinhao Song, Tianyu Guo, and Wei Shi. Pose-guided feature disentangling for occluded person re-identification based on transformer. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 2540–2549, 2022.
  • Wang et al. [2022c] Zijie Wang, Aichun Zhu, Jingyi Xue, Xili Wan, Chao Liu, Tian Wang, and Yifeng Li. Caibc: Capturing all-round information beyond color for text-based person retrieval. In Proceedings of the 30th ACM International Conference on Multimedia, pages 5314–5322, 2022.
  • Wen et al. [2021] Keyu Wen, Jin Xia, Yuanyuan Huang, Linyang Li, Jiayan Xu, and Jie Shao. Cookie: Contrastive cross-modal knowledge sharing pre-training for vision-language representation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2208–2217, 2021.
  • Wu et al. [2021] Yushuang Wu, Zizheng Yan, Xiaoguang Han, Guanbin Li, Changqing Zou, and Shuguang Cui. Lapscore: Language-guided person search via color reasoning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1624–1633, 2021.
  • Yan et al. [2022] Shuanglin Yan, Neng Dong, Liyan Zhang, and Jinhui Tang. Clip-driven fine-grained text-image person re-identification. arXiv preprint arXiv:2210.10276, 2022.
  • Yang et al. [2022] Jinyu Yang, Jiali Duan, Son Tran, Yi Xu, Sampath Chanda, Liqun Chen, Belinda Zeng, Trishul Chilimbi, and Junzhou Huang. Vision-language pre-training with triple contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15671–15680, 2022.
  • Zeng et al. [2021] Pengpeng Zeng, Shuaiqi Jing, Jingkuan Song, Kaixuan Fan, Xiangpeng Li, Liansuo We, and Yuan Guo. Relation-aware aggregation network with auxiliary guidance for text-based person search. World Wide Web, pages 1–18, 2021.
  • Zhang and Lu [2018] Ying Zhang and Huchuan Lu. Deep cross-modal projection learning for image-text matching. In Proceedings of the European conference on computer vision (ECCV), pages 686–701, 2018.
  • Zheng et al. [2020a] Kecheng Zheng, Wu Liu, Jiawei Liu, Zheng-Jun Zha, and Tao Mei. Hierarchical gumbel attention network for text-based person search. In Proceedings of the 28th ACM International Conference on Multimedia, pages 3441–3449, 2020.
  • Zheng et al. [2020b] Zhedong Zheng, Liang Zheng, Michael Garrett, Yi Yang, Mingliang Xu, and Yi-Dong Shen. Dual-path convolutional image-text embeddings with instance loss. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), 16(2):1–23, 2020.
  • Zhu et al. [2021] Aichun Zhu, Zijie Wang, Yifeng Li, Xili Wan, Jing Jin, Tian Wang, Fangqiang Hu, and Gang Hua. Dssl: Deep surroundings-person separation learning for text-based person retrieval. In Proceedings of the 29th ACM International Conference on Multimedia, pages 209–217, 2021.

Appendix A Appendix

A.1 Datasets

CUHK-PEDES Li et al. [2017b] is the most commonly-used dataset in text-based person search. It consists of 40,20640,206 images and 80,44080,440 texts from 13,00313,003 identities in total, which are split into 34,05434,054 images and 68,12668,126 texts from 11,00311,003 identities in the training set, 3,0783,078 images and 6,1586,158 texts from 1,0001,000 identities in the validation set, and 3,0743,074 images and 6,1566,156 texts from 1,0001,000 identities in the test set. The average length of all texts is 2323.

ICFG-PEDES Ding et al. [2021] is a recently published dataset, which contains 54,52254,522 images from 4,1024,102 identities in total. Each of the images is described by one text. The dataset is split into 34,67434,674 images from 3,1023,102 identities in the training set, and 19,84819,848 images from 1,0001,000 identities in the test set. On average, there are 3737 words for each text.

RSTPReid Zhu et al. [2021] is also a newly released dataset to properly handle real scenarios. It contains 20,50520,505 images of 4,1014,101 identities. Each identity has 55 corresponding images captured from different cameras. Each image is annotated with 22 textual descriptions, and each description is no shorter than 2323 words. There are 3,7013,701/200200/200200 identities utilized for training/validation/testing, respectively.

Method Flickr30K (1K test set) COCO (5K test set)
TR IR TR IR
R@1 R@5 R@10 R@1 R@5 R@10 R@1 R@5 R@10 R@1 R@5 R@10
UNITER Chen et al. [2020] 87.30 98.00 99.20 75.56 94.08 96.76 65.68 88.56 93.76 52.93 79.93 87.95
COOKIE Wen et al. [2021] 89.00 98.90 99.70 75.60 94.60 97.20 71.60 90.90 95.40 54.50 81.00 88.20
Oscar Li et al. [2020] - - - - - - 73.50 92.20 96.00 57.50 82.80 89.80
UNIMO Li et al. [2021b] 89.40 98.90 99.80 78.04 94.24 97.12 - - - - - -
ALIGN Jia et al. [2021] 95.30 99.80 100.00 84.90 97.40 98.60 77.00 93.50 96.90 59.90 83.30 89.80
BLIP Li et al. [2022a] 97.40 99.80 99.90 87.60 97.70 99.00 82.40 95.40 97.90 65.10 86.30 91.80
ALBEF(backbone) Li et al. [2021a] 95.50 99.80 99.90 85.44 97.34 98.70 77.26 94.02 97.04 60.31 84.22 90.51
RaSa (Ours) 96.00 99.80 100.00 85.90 97.54 98.72 77.44 94.12 97.18 61.00 84.49 90.83
Table 5: Results of coarse-grained retrieval on Flickr30K and COCO.
Method CUB Flowers
TR IR TR IR
R@1 AP@50 R@1 AP@50
Bow Harris [1954] 44.1 39.6 57.7 57.3
Word2Vec Mikolov et al. [2013] 38.6 33.5 54.2 52.1
GMM+HGLMM Klein et al. [2015] 36.5 35.6 54.8 52.8
Word CNN Reed et al. [2016] 51.0 43.3 60.7 56.3
Word CNN-RNN Reed et al. [2016] 56.8 48.7 65.6 59.6
Triplet Li et al. [2017a] 52.5 52.4 64.3 64.9
Latent Co-attention Li et al. [2017a] 61.5 57.6 68.4 70.1
CMPM/C Zhang and Lu [2018] 64.3 67.9 68.9 69.7
TIMAM Sarafianos et al. [2019] 67.7 70.3 70.6 73.7
GARN Jing et al. [2021] 69.7 69.4 71.8 72.4
DME Wang et al. [2021b] 69.4 71.8 72.4 74.6
iVAD Wang et al. [2022a] 70.3 72.5 73.0 75.1
RaSa (Ours) 84.3 84.5 87.1 84.3
Table 6: Results of fine-grained retrieval on CUB and Flowers.

A.2 Implementation Details

All experiments are conducted on 44 NVIDIA 30903090 GPUs. We train our model with 3030 epochs and a batch size of 5252. The AdamW optimizer Loshchilov and Hutter [2019] is adopted with a weight decay of 0.020.02. The learning rate is initialized as 1e41e-4 for the parameters of the classifiers in PRD and mm-RTD, and 1e51e-5 for the rest parameters of the model. All images are resized to 384×384384\times 384 and random horizontal flipping is employed for data augmentation. The input texts are set with a maximum length of 5050 for all datasets. The momentum coefficient in the momentum model is set as m=0.995m=0.995. The queue size RR is set as 65,53665,536 and the temperature τ\tau is set as 0.070.07 in CL. The probability of inputting the weak positive pair is set as pw=0.1p^{w}=0.1 in RA, and the probability of masking the word in the text is set as pm=0.3p^{m}=0.3 in SA. The hyper-parameters in the objective function are set as λ1=0.5\lambda_{1}=0.5, λ2=0.5\lambda_{2}=0.5, λ3=0.5\lambda_{3}=0.5.

A.3 Extended Experiments

We conduct extended experiments to verify the effectiveness of RaSa, including coarse-grained retrieval and fine-grained retrieval. Moreover, in order to verify the generalization ability of RaSa, we also conduct a series of domain generalization experiments, following LGUR Shao et al. [2022].

A.3.1 Coarse-grained Retrieval

We consider two datasets for the coarse-grained retrieval task: Flickr30K Plummer et al. [2015] and COCO Lin et al. [2014]. Different from the text-based person search datasets with only one object (i.e., person) in the images and the fine-grained textual sentences, the images in Flickr30K and COCO contain various objects and the corresponding sentences usually present a coarse-grained description. We follow the widely-used Karpathy split Karpathy and Fei-Fei [2015] for both datasets. The images in Flickr30K are split into 2929K/11K/11K and the images in COCO are split into 113113K/55K/55K for training/validation/testing, respectively. Each image in both two datasets is annotated by five sentences.

It should be noted that each image together with the paired texts is a unique class in the two datasets, as a result of which there is no intra-class variation in the images and all of positive image-text pairs belong to the strong positive type. Therefore, the proposed RA, which aims at differentiating between strong and weak positive pairs, no longer applies to the experiments on Flickr30K and COCO. We only perform SA and use the vanilla ITM for the experiments.

As shown in Table 5, RaSa achieves a comparable performance compared with existing methods. Particularly, compared with the backbone model ALBEF Li et al. [2021a]222We report the results reproduced with the released code of ALBEF, where the batch size is set as same as the introduction in Appendix A.2 for a fair comparison., RaSa with only SA still brings consistent improvement in terms of all metrics. We argue that SA constructs a non-trivial pretext task to explicitly endow the model with the ability to perceive the sensitive transformation, which significantly facilitates the representation learning and eventually gains a better performance.

Method R@1 R@5 R@10
C \rightarrow I Dual Path Zheng et al. [2020b] 15.41 29.80 38.19
MIA Niu et al. [2020] 19.35 36.78 46.42
SCAN Lee et al. [2018] 21.27 39.26 48.83
SSAN Ding et al. [2021] 29.24 49.00 58.53
LGUR Shao et al. [2022] 34.25 52.58 60.85
RaSa (Ours) 50.59 67.46 74.09
I \rightarrow C Dual Path Zheng et al. [2020b] 7.63 17.14 23.52
MIA Niu et al. [2020] 10.93 23.77 32.39
SCAN Lee et al. [2018] 13.63 28.61 37.05
SSAN Ding et al. [2021] 21.07 38.94 48.54
LGUR Shao et al. [2022] 25.44 44.48 54.39
RaSa (Ours) 50.70 72.40 79.58
Table 7: Comparison with other methods on domain generalization task. We adopt CUHK-PEDES (denoted as C) and ICFG-PEDES (represented as I) as the source domain and the target domain in turn.
Method R@1 R@5 R@10 mAP
w/o VLP GNA-RNN Li et al. [2017b] 19.05 - 53.64 -
Dual Path Zheng et al. [2020b] 44.40 66.26 75.07 -
CMPM/C Zhang and Lu [2018] 49.37 71.69 79.27 -
ViTAA Wang et al. [2020] 55.97 75.84 83.52 -
DSSL Zhu et al. [2021] 59.98 80.41 87.56 -
MGEL Wang et al. [2021a] 60.27 80.01 86.74 -
ACSA Ji et al. [2022] 63.56 81.40 87.70 -
SAF Li et al. [2022b] 64.13 82.62 88.40 58.61
TIPCB Chen et al. [2022] 64.26 83.19 89.10 -
CAIBC Wang et al. [2022c] 64.43 82.87 88.37 -
C2A2\rm C_{2}A_{2} Niu et al. [2022] 64.82 83.54 89.77 -
LGUR Shao et al. [2022] 65.25 83.12 89.00 -
w/ VLP PSLD Han et al. [2021] 64.08 81.73 88.19 60.08
IVT Shu et al. [2022] 65.59 83.11 89.21 -
CFine Yan et al. [2022] 69.57 85.93 91.15 -
CLIP Radford et al. [2021] 43.05 66.41 76.36 38.91
RaSaCLIP{}_{\text{CLIP}} 57.60 78.09 84.91 55.52
TCL Yang et al. [2022] 57.60 77.14 84.39 53.64
RaSaTCL{}_{\text{TCL}} 73.23 89.20 93.32 66.43
Table 8: Comparison with other methods on CUHK-PEDES. RaSaCLIP{}_{\text{CLIP}} adopts CLIP as the backbone, while RaSaTCL{}_{\text{TCL}} uses TCL as the backbone.

A.3.2 Fine-grained Retrieval

Apart from the fine-grained retrieval task of text-based person search, we furthermore evaluate RaSa on other fine-grained datasets: CUB Reed et al. [2016] and Flowers Reed et al. [2016]. CUB contains 11,78811,788 bird images from 200200 different categories, and each image is annotated with 1010 sentences. The dataset is split into 100100/5050/5050 categories for training/validation/testing, respectively. Flowers consists of 8,1898,189 flower images from 102102 categories, and each image has 1010 descriptions. There are 6262, 2020 and 2020 categories utilized for training, validation and testing, respectively.

Following common settings Reed et al. [2016]; Sarafianos et al. [2019], we take random cropping and horizontal flipping as the data augmentation, and the maximum length of the input texts is set as 3030. Other settings are kept as same as the introduction in Appendix A.2. Therein, we use AP@50 metric for the evaluation of text-to-image retrieval and R@1 for image-to-text retrieval, where AP@50 reflects the average matching percentage of top-50 retrieved images of all test text classes. During inference, existing methods usually compute the metrics according to the similarity between the image embedding and the average of the corresponding text embeddings. However, since RaSa is a one-stream model and its final output is the multi-modal embedding rather than the text embedding, we compute the metrics by averaging the multi-modal embeddings of the same identity.

From Table 6, RaSa outperforms all existing state-of-the-art methods by a large margin. Specifically, compared with iVAD Wang et al. [2022a], the performance of RaSa has 14.014.0% and 12.012.0% improvements on CUB and 14.114.1% and 9.29.2% boosts on Flowers in terms of R@1 and AP@50, respectively. It is worth noting that existing methods ignore the noise interference caused by the weak positive pairs and model all positive relations without distinction. Inevitably, they are vulnerable to overfitting learning. On the contrary, RaSa utilizes RA to explicitly distinguish different types of positive relation and SA to learn more robust representations. As a result, it achieves a decent performance.

A.3.3 Domain Generalization

We conduct a series of domain generalization experiments to investigate the generalization ability of RaSa. Specifically, we use the model trained on the source domain to evaluate the performance on the target domain, where CUHK-PEDES and ICFG-PEDES are adopted as the source domain and the target domain in turn.

As shown in Table 7, RaSa outperforms other methods by a large margin. We conjecture that there exist two factors bringing such a significant improvement. (1) Other methods are inclined to overfitting learning since they neglect the noise interference from the weak positive pairs, while RaSa substantially alleviates the effect of the noise and is able to learn more robust representations. (2) The parameters of RaSa are initialized from the VLP models which contain abundant multi-modal knowledge and eventually facilitate representation learning. Overall, the results on the domain generalization task effectively demonstrate the powerful generalization ability of RaSa.

A.4 Backbones and Experiments

Apart from ALBEF Li et al. [2021a], we also apply RaSa on other backbones: TCL Yang et al. [2022] and CLIP Radford et al. [2021].

TCL has a similar architecture with ALBEF and is pretrained on 44M image-text pairs. CLIP is pretrained on 400400M image-text pairs and is comprised of two unimodal encoders to individually process the images and texts. However, the proposed RaSa works on the multi-modal features from the cross-modal encoder. Therefore, we additionally append a one-layer transformer block on the outputs of CLIP as the cross-modal encoder when adopting CLIP as the backbone.

As shown in Table 8, no matter whether TCL or CLIP is adopted as the backbone, RaSa always brings consistent improvements in terms of all metrics. Meanwhile, a stronger backbone can lead to a better performance. For example, in terms of R@1, RaSaTCL{}_{\text{TCL}} achieves the best performance with 73.2373.23%, while RaSaCLIP{}_{\text{CLIP}} achieves a modest performance of 57.6057.60%. We conjecture that (1) the lack of cross-modal deep fusion in the backbone CLIP makes the model difficult to capture fine-grained details, which tends to have a negative impact to the performance, and (2) the parameters of the one-layer transformer block are randomly initialized, rendering the model inclined to be trapped in the local minimum.

Refer to caption
Figure 5: Visualization of top-10 retrieval results on CUHK-PEDES. The first row in each example presents the retrieval results from the backbone ALBEF, and the second row shows the results from RaSa. Correct/Incorrect images are marked by green/red rectangles.

A.5 Visualization

We exhibit three top-10 retrieval examples of the backbone ALBEF Li et al. [2021a] and RaSa in Figure 5, where the first row and the second row in each example present the retrieval results from ALBEF and RaSa, respectively. It can be seen that RaSa can retrieve the corresponding pedestrian images for a query text more accurately. This is mainly due to the alleviation of the noise interference in RA and the powerful sensitivity-aware learning strategy in SA. The visualization vividly demonstrates the effectiveness of RaSa.