This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Exploring Invariant Representation for Visible-Infrared Person Re-Identification

Lei Tan, Yukang Zhang, Shengmei Shen,  , Yan Wang, Pingyang Dai, Xianming Lin, Yongjian Wu, Rongrong Ji Lei Tan, Pingyang Dai, Xianming Lin and Rongrong Ji are with the Media Analytics and Computing Laboratory, Department of Artificial Intelligence, School of Informatics, Xiamen University, Xiamen 361005, China (e-mail: tanlei@stu.xmu.edu.cn; pydai@xmu.edu.cn; linxm@xmu.edu.cn; rrji@xmu.edu.cn).Yukang Zhang, and Shengmei Shen are with Pensees Pte Ltd., Singapore (e-mail: zhangyk@pensees.ai; jane.shen@pensees.ai).Yan Wang is with the Pinterest, Seattle, WA 98100 USA. (e-mail: yanw@pinterest.com).Yongjian Wu is with the Youtu Laboratory, Tencent Technology (Shanghai) Co. Ltd, Shanghai 361005, China (e-mail: littlekenwu@tencent.com;).
Abstract

Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy. In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM). In particular, we observe that the reflective intensity of the same surface in photos shot in different wavelengths could be transformed using a linear model. Besides, we show the variable linear factor across the different surfaces is the main culprit which initiates the modality discrepancy. We integrate such a reflection observation into an image-level data augmentation by proposing the linear transformation generator (LTG). Moreover, at the feature level, we introduce a cross-center loss to explore a more compact intra-class distribution and modality-aware spatial attention to take advantage of textured regions more efficiently. Experiment results on two standard cross-spectral person re-identification datasets, i.e., RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.

Index Terms:
Person Re-Identification, Cross-Spectral Retrieval, Representation Learning.

I Introduction

Person re-identification (re-id), aiming to address the problem of matching people over a distributed set of non-overlapping cameras, has attracted intensive attention in the last few years due to its wide applications in surveillance systems [1, 2, 3, 4]. Since silicon-based digital cameras are naturally sensitive to near-infrared (NIR), most cameras provide extra infrared (IR) images instead of the visible (VIS) images under poor illumination conditions for better visual quality. This in practice puts the Re-ID problem in a cross-spectral setting and requires the approaches to properly handle both the intra-class variance and the even more significant modality discrepancies between cross-spectral images [5].

Refer to caption
Figure 1: Illustration of the cross-spectral transformation. G refers to the green channel of visible image. Under the same illumination, the cross-spectral transformation could be described as a linear transformation in material-similar surface. Still, in the whole image level, the transformation is nonlinear due to the diversity of materials. Since the Re-ID image pairs is not well aligned, we select the cross-spectral image pairs from [6].

Compared with the single-modality person re-identification problem, the major challenge in such application lies in how to minimize the modality discrepancy between the visible and infrared images. Continuous research efforts have been devoted to addressing this challenge and inspired two typical frameworks.

One is called feature-based approaches [7, 8, 9, 10, 11], which aim to bridge the apparent gap in the feature space. These approaches typically train a single-path or multi-path Deep Neural Network (DNN) from end to end to obtain a spectrum-invariant embedding. Although the core idea of directly end-to-end learning makes these approaches highly effective and often achieves state-of-the-art performance, this body of works still shows limited ability in minimizing the modality discrepancy. To makes things worse, due to the high complexity and lack of interpretability, the models are hard to adjust or improve. The other direction is image-level approaches [5, 12, 13], which aim to bridge the appearance discrepancy by transforming the image from one spectrum to another using DNN-based image processing such as Generative Adversarial Networks (GANs) [14]. These approaches usually give better visual verisimilitude and adjustability. However, the lack of a large-scale database providing image pairs across the spectrum makes GAN training challenging and limits the performance of those approaches.

In this paper, we explores the possibility of using actual physics models behind multi-spectral imaging to provide more interpretability, and thus further push the boundary of cross-spectral Re-ID approaches. Based on the Lambertian reflection model [15, 16], the illuminations of the same region in VIS and NIR photos should be able to be described using a simple linear model, as long as the region is composed of one consistent material (detail disscussed in Sec. III). This is illustrated in Figure 1. Here we use paired VIS-NIR images in the dataset in [6]. For the red and yellow regions in the middle of the image, with a simple linear model, we can accurately predict the pixel values of the NIR image based on the VIS image, as long as the region only has one material. Although the linear transformation exists in the pixel-level of cross-spectral image pairs, the linear factor is determined by the reflection function of the material. It means the linear factor not a constant factor on different surfaces, which causes an image-level non-linear transformation. In Section III, we analyze and visualize the result to confirm whether the different linear factors in different surfaces are the main culprit that induces the modality discrepancy. It is interesting to find that the modality discrepancy occurs when using variable linear factors among different patches in the image.

The above observation provides us a fresh perspective on the cross-spectral Re-ID task. Empirically, adopting the observation in the image generation seems the most intuitive way. As long as we are able to identify regions with different materials and calculate the linear coefficients to transform the input image from one spectrum to another, the modality discrepancy would be easy to bridge. While this direction may work considering the recent progress on semantic segmentation and a large amount of extra-label for materials, it may be over-complicated and thus not the most effective way. Even if semantic segmentation works reasonably well, this type of paradigm is still fragile when facing unseen materials.

Driven by this analysis, we extend the aforementioned observation in a feature-level learning framework named robust feature mining (RFM) network to explore a modality-invariant feature representation. Our RFM starts with an easy and efficient linear transformation generator (LTG). To encourage the network to be robust to this kind of linear transformation, the LTG takes effect on randomly selected image patches by a random linear factor and aims to get invariable feature representation to such a transformation. To further mine the invariable feature representation in the feature space, the RFM utilizes the cross-center loss which penalizes the modality discrepancy residual in each part after the single-path DNN model. The cross-center loss employs the modality center instead of the class-center for optimization and shows promising improvement in minimizing the modality discrepancy. Finally, to overcome the tendency of over-focusing on the low modality discrepancy but texture-lacking regions, RFM adopts a modality-aware spatial attention module (MAM) to strengthen the influence of those highly discriminative regions via high-level global features.

We summarize our main contributions as follows:

  • As an effort to explore the physics model behind the modality discrepancy in the cross-spectral Re-ID task, based on the Lambertian model we demonstrate that the cross-spectral transformation could be described as a linear transformation in regions of the same material and the culprit of modality discrepancy is mainly comes from the image-level non-linear transformation cause by the diversity materials.

  • By extending the aforementioned diffuse reflection model, we propose a robust feature mining network (RFM). The RFM effectively takes advantage of the above transformation and embeds it in an end-to-end fine-grid framework to explore modality-invariable feature representation.

  • Extensive experiments on two standard datasets RegDB and SYSU-MM01 prove the effectiveness and superior performance of the proposed RFM.

II Related Works

Image-level approaches decompose the modality discrepancy and appearance discrepancies aiming at exploring an efficiency transformation between VIS modality and NIR modality as an image preprocessing. Under this condition, the cross-modality discrepancy is considered as an individual problem alongside the Re-ID problem. D2RL [5] makes the first attempt by using variational autoencoders (VAE) for style disentanglement and generates synthesis images from one domain to another. Wang et al. [12] disentangle the features in the encoder and decode from the exchanged features to generate high-quality paired images. AlignGan [17] employs a unified GAN framework together with efficient constraints for both image and feature alignment, achieving significant improvement on the quality of synthesized images and final Re-ID performance. Most recently, X-modality [13] designs a light-weight network to learn an intermediate mediator between visible and infrared images instead of focusing on mAPping visible images to infrared images directly. Similarly, Ye et al. [18] utilize generated grayscale images as an assistant for training a more suitable network. Hi-CMD [19] suggests an efficient generator structure that is possible to change the pose and illumination attributes while maintaining the identity characteristic of a specific person. Although playing a min-max game between the generator and discriminator offers visually impressive results, the generated images are still far from photorealistic, and in turn, limit the final performance.

Feature-level approaches aim at learning modality-invariant feature representations. Wu et al. [20] build the largest VIS-NIR Re-ID dataset SYSU-MM01 and propose a zero-padding framework for shared feature learning. TONE + HCML [9] improves the two-stream cross-modality Re-ID framework by jointly optimizing the modality-specific and modality-shared metrics. Inspired by adversarial learning, cmGan [7] adopts a generative adversarial training method to decrease the feature representation discrepancy for different modality images. Ye et al. [21] propose an attention-based framework to mine the part-level and graph-level contextual cues. Wu et al. [11] propose a modality-gated extractor attached with focal modality-aware similarity preserving loss. Lu et al. [10] try to introduce modality-specific features based on the cross-modality near the neighbor affinity model. DG-VAE[22] leverages the Gaussian based variational auto-encoder to disentangle an identity-discriminable and an identity-ambiguous cross-modality feature subspace. However, most of those methods focus on feature representations on top of the entire image. MPANet [23] joint modality and pattern alignment network to discover cross-modality nuances for visible-infrared person Re-ID. Due to the diversity of materials, it is difficult to learn an efficient representation in the whole image-level to reduce the modality discrepancy.

Refer to caption
Figure 2: Example images from the VIS-NIR scene dataset [6]. After we divide the visible image into the red, green, and blue channels and form chromaticity band-ratios from these three spectra and the NIR image, it is clear that the ratio for pixels from the surface with high material-similarity is nearly constant.
Refer to caption
Figure 3: A example about how modality discrepancy occurs. Feature space visualization of 100 randomly selected images with/without the local linear transformation. (a)\sim(b): A same linear factor take effect on the whole image will bring limited modality discrepancy. (c): Variable linear factors take effect on different part shows a huge modality discrepancy. The ’cross’ and ’dot’ marks indicate the samples from the original one and the generated one respectively.

III Reflection Prior for Cross-Spectral Images

VIS-NIR matching is a longstanding computer vision problem that has been explored for decades [24, 25]. One of the main challenges is to formulate and thus alleviate the modality discrepancy. Using the Lambertian model to analyze the digital image from the multi-sensor cameras is widely applied in some pioneer works[16, 26]. With a light source emitting photons across different wavelengths λ\lambda, the response of each pixel (x,y)(x,y) in the camera sensors can be formulated as:

ρj(x,y)=σ(x,y)λjEj(λ,x,y)S(λ,x,y)Qj(λ)𝑑λ,{\rho_{j}}(x,y)=\sigma(x,y)\int_{\lambda_{j}}{{E_{j}}(\lambda,x,y)}S(\lambda,x,y){Q_{j}}(\lambda)d\lambda, (1)

where λ\lambda is the wavelength, E(λ)E(\lambda) and S(λ)S(\lambda) denote the spectral power distribution (SPD) of incident light and surface spectral reflectance, and Q(λ){Q}(\lambda) is the spectral sensitivity of the camera sensor. j={R,G,B,N}j=\left\{{R,G,B,N}\right\} indicates the channel (spectrum). σ(x,y)\sigma(x,y) is the Lambertian reflection term which an constant factor and can be calculate by the dot product of the surface normal with the illumination direction.

Following Eq. 1, we leverage a mild assumption to derive a representation between the SPD of the light source and incident light. Generally, we could describe the SPD of light source by a relative spectral power distribution F(λ,x,y)F(\lambda,x,y) together with a variable ω\omega which reflects the illumination intensity. We assume that the SPD of incident light in the whole image keeps the same relative spectral power distribution as the light source. Then we could formulate the E(λ,x,y)E(\lambda,x,y) as:

Ej(λ,x,y)=βj(x,y)ωjFj(λ),{E_{j}}(\lambda,x,y)={\beta_{j}}(x,y){\omega_{j}}{F_{j}}(\lambda), (2)

where the β\beta is a parameter to reflect the ratio of intensity between the incident light and the light source. Then from Eq. 1 and Eq. 2, if we now consider the images under different spectra like G channel images and NIR images, it is clear that the transformation of G-NIR could be described as:

ρN(x,y)ρG(x,y)=ωNβN(x,y)λNFN(λ)S(λ,x,y)QN(λ)𝑑λωGβG(x,y)λGFG(λ)S(λ,x,y)QG(λ)𝑑λ.\frac{{{\rho_{N}}(x,y)}}{{{\rho_{G}}(x,y)}}=\frac{{{\omega_{N}}{\beta_{N}}(x,y)\int_{{\lambda_{N}}}{{F_{N}}(\lambda)S(\lambda,x,y){Q_{N}}(\lambda)}d\lambda}}{{{\omega_{G}}{\beta_{G}}(x,y)\int_{{\lambda_{G}}}{{F_{G}}(\lambda)S(\lambda,x,y){Q_{G}}(\lambda)d\lambda}}}. (3)

Under an ideal condition, β\beta is supposed to be a high-order term determined by the distance between the light source and the surface [27]. Now, we could utilize this approximation and regard the β\beta as a constant under the same light source to get a simplified expression:

ρN(x,y)ρG(x,y)=ωNM(x,y,N)ωGM(x,y,G).\frac{{{\rho_{N}}(x,y)}}{{{\rho_{G}}(x,y)}}=\frac{{{\omega_{N}}M(x,y,N)}}{{{\omega_{G}}M(x,y,G)}}. (4)

Since F(λ)F(\lambda), Q(λ){Q}(\lambda), and S(λ)S(\lambda) are three inner functions depending on the SPD of the light source, the sensitivity of the camera sensor, and the reflection function of the surface material, we replace the Riemann integral with a function M(x,y,j)M(x,y,j). In addition, the ωNωG\frac{\omega_{N}}{\omega_{G}} could be considered as a constant factor in two determined spectra. From this representation, one could observe that in those regions of the same material and under the same illumination condition, the cross-spectral transformation is a linear transformation like shown in Figure 1. And if we extend it to the entire image, the factor is only influenced by the S(λ,x,y)S(\lambda,x,y) which is determined by the material.

To verify whether the above equation could be used in various real-life scenarios, we used the paired VIS-NIR scene image dataset introduced by [6]. In Figure 2, we form chromaticity band-ratios between three VIS spectra and NIR spectrum at each pixel and use the color to reflect the ratio. We found that the ratio is nearly constant within a region with a consistent material, which holds across R, G, B, and NIR spectra.

After observing the above linear transformation, we further explore whether the variable linear factor in different surfaces is the main culprit that induced the modality gap in such an application. Due to the lacking of a material-based segmentation method to generate according samples, we uniformly segment 100 randomly selected images into six parts from the top to the bottom and multiply each part by a linear factor. Then, we send the new images and original images into an ImageNet [28] pre-trained Resnet-50 [24]. Although not so well-aligned, benefitting from the body structure prior from the head to toe, we still find that the modality discrepancy occurs when suffering variable linear factors.

Refer to caption
Figure 4: Framework of the proposed RFM. GAP: Global Average Pooling. SAP: Spatial-wise Average Pooling. The RFM is built upon a single-path part model with several improvements to adapt to the reflection model.
Refer to caption
Figure 5: The motivation of linear transformation generator (LTG). Herein, we construct an ideal person with only two different surfaces and ignore the background. (a): As demonstrated above, to obtain a spectral-invariant feature representation, the network should be robust to such a transformation that takes effect upon definite surfaces by definite linear factors. (b): An ideal data augmentation strategy that takes effect upon definite surfaces by random linear factors. However, this method needs a hard-achieved extra material-aware network for segmentation. (c): The idea of LTG. By taking effect upon random surfaces by random linear factors, the LTG encourages the network to be robust to a linear transformation anywhere in the image. Under this condition, the cross-spectral transformation can be considered as an easy state of LTG space.

IV Method

IV-A Model Architecture

As illustrated in Figure 4, the proposed RFM starts with the linear transformation generator part. Input images II will be transformed to II^{*}. The single-path Resnet-based backbone [24] will takes the enhanced images to generate the feature FF with spatial dimension Hf×WfH_{f}\times W_{f}. In this work, we leverage a single-path backbone rather than a multi-path one like in the previous works [9, 7, 10, 21] to extract a modality invariant feature embedding. This structure significantly decreases the computational complexity. To emphasize those textured regions with high discriminability, we also employ a modality-aware spatial attention module to generate the modality-specific features FattvF_{att}^{v} and FattnF_{att}^{n}. Following the PCB model [29], feature FattF_{att} will be uniformly segmented to KK parts with spatial dimension HfK×Wf\frac{H_{f}}{K}\times{W_{f}} from the height, and then the attached global average pooling layer (GAP) will squeeze all the features to a vector 𝑽i\bm{V}_{i} (i=1,2,,Ki=1,2,\cdots,K). Compared with constraining the vectors from each classifier straightly or after the BNNeck [30], we attach the constraints after a fully connected layer to reduce the dimension. The same strategy is also employed by some previous work [10, 12, 17].

Input: II: Input image; CC, HH and WW: Image channel and size; pp: Probability of the LTG; smins_{min} and smaxs_{max}: Area of the selected region; rminr_{min} and rmaxr_{max}: Aspect of the selected region; tmint_{min}: Terminate the LTG;
Output: Enhanced image II^{\ast}
Initialization: p1p_{1}\leftarrow Rand(0,1)(0,1);
if p1pp_{1}\geq p then
      III^{\ast}\leftarrow I;
       return II^{\ast};
else
      M=Ones(C,H,W)M=Ones(C,H,W);
       while TrueTrue do
            SrS_{r}\leftarrowRand(smin,smax)×W×H(s_{min},s_{max})\times W\times H;
             rrr_{r}\leftarrowRand(rmin,rmax)(r_{min},r_{max});
             HrSr×rrH_{r}\leftarrow\sqrt{S_{r}\times r_{r}}; WrSrrrW_{r}\leftarrow\frac{S_{r}}{r_{r}};
             xrx_{r}\leftarrowRand(0,W)(0,W); yry_{r}\leftarrowRand(0,H)(0,H);
             if xr+WrWx_{r}+W_{r}\leq W and yr+HrHy_{r}+H_{r}\leq H then
                  Irle(C,xr,yr,xr+Wr,yr+Hr)I_{rle}\leftarrow(C,x_{r},y_{r},x_{r}+W_{r},y_{r}+H_{r});
                   Mrle(C,xr,yr,xr+Wr,yr+Hr)M_{rle}\leftarrow(C,x_{r},y_{r},x_{r}+W_{r},y_{r}+H_{r});
                   for i0i\leftarrow 0 to CC do
                        Ic(i,Wr,Hr)I_{c}\leftarrow(i,W_{r},H_{r});
                         Mc(i,Wr,Hr)M_{c}\leftarrow(i,W_{r},H_{r});
                         αmax1max(Ic)\alpha_{max}\leftarrow\frac{1}{max(I_{c})};
                         ααmax×fg()\alpha\leftarrow\alpha_{max}\times f_{g}(\cdot);
                         I(Ic)α×IcI(I_{c})\leftarrow\alpha\times I_{c};
                         M(Mc)α×McM(M_{c})\leftarrow\alpha\times M_{c};
                   end for
                  
             end if
            if min(M)tminmin(M)\leq t_{min} then
                  III^{\ast}\leftarrow I;
                   Return II^{\ast}
             end if
            
       end while
      
end if
Algorithm 1 linear transformation generator

IV-B linear transformation generator

Based on the aforementioned observation, to obtain a modality invariant feature representation, it is necessary to encourage the network to overcome such a transformation that takes effect on a determined surface with a determined linear factor as depicted in Fig. 5 (a). Although utilizing random linear factors in different surfaces as shown in Fig. 5 (b) seems like the most intuitive way for data augmentation, it highly relies on the well-trained material-aware segmentation network which is hard to achieve. Therefore, we propose the LTG as shown in Fig. 5 (c). The LTG randomly selects several image patches and multiplies them with a variable linear parameter. Under the LTG, the cross-spectral transformation could be considered as an easy state of the state space.

Concretely, the LTG is conducting with a certain probability pp. For an input image II, the LTG randomly selects a rectangle region IrleI_{rle} and multiplies it with a linear factor α\alpha. In case it may exceed the upper bound, we calculate the maximum feasible linear factor in the IrleI_{rle} as αmax\alpha_{max}. The linear factor α\alpha is calculated by the multiply of αmax\alpha_{max} and a random generator fg()f_{g}(\cdot) on 0 to 1. As shown in Fig. 6, a small linear factor can not provide enough modality gap in the training phase. Therefore, a U-shaped Beta distribution is utilized for the fg()f_{g}(\cdot) to obtain high-quality training samples. Furthermore, we set the smins_{min} and smaxs_{max} to control area of the selected region and the rminr_{min} and rmaxr_{max} to control the aspect ratio. Then, the region will start on a random selected point (xr,yr)(x_{r},y_{r}). The whole processing will be repeated if the rectangle region IrleI_{rle} is out of the boundary.

Different from the data augmentation strategy like random erasing [31] or color-jitter which takes effect on the image only once, the LTG will be repeated several times to obtain a higher modality discrepancy. Therefore, to avoid the vanishing feature, we set a memory matrix MM to store the cumulative changes at each pixel. And we set a tmint_{min} to terminate the LTG when M(c,x,y)<tminM(c,x,y)<t_{min}. Specifically, we show the procedure in Alg. 1

Refer to caption
Figure 6: A example of modality discrepancy. Small linear factors may not so efficient to generate images with modality gap in the training stage..
Refer to caption
Figure 7: Comparisons between center loss (a), hetero-center loss (b) and cross-center loss (c). By exchanging the center of each modality to cluster the samples, the cross-center loss could obtain a more compact distribution.

IV-C Cross-center Loss

Motivation. Adopting a center-based loss function to directly optimize the modality gap is an intuitive strategy. The widely used center loss [32] aims to minimize the intra-class discrepancy by the class center. As illustrated in Figure 7 (a), in center loss, the modality discrepancy is largely ignored which limited the optimization for the modality discrepancy. Recently, HC-Loss [33] is proposed to optimize the modality gap by aligning the center of two modalities. However, the HC-loss only focus on the average of the cross-spectral distribution as illustrated in Figure 7 (b), which shows limited performance to compact the final distribution. To fully utilize the most consistent feature, we introduce a cross-center loss to simultaneously optimize the average and the variance of the cross-spectral distribution as shown in Figure 7 (c).

Definition. The cross-center loss is designed to address the modality discrepancy in the cross-spectral Re-ID issue. As a brief recap, for each input IiI_{i}, the basic center loss can be formulated as:

c=12i=1ND(f(Ii),cyi),{\mathcal{L}}_{c}=\frac{1}{{2}}\sum\limits_{i=1}^{N}{D(f({I_{i}}),{c_{{y_{i}}}})}, (5)

where cyidc_{y_{i}}\in{\mathbb{R}}^{d} refers to the center of class yiy_{i} and dd denotes the dimension of feature. f()f(\cdot) refers to the backbone network and D()D(\cdot) is the distance function, which is Euclidean distance here. NN is the size of the mini-batch. In order to add modality discrepancy into consideration, we divide the IiI_{i} into VIS modality IivI_{i}^{v} and NIR modality IinI_{i}^{n}, then use the center of each modality in single classes instead of the original center for optimization. To simultaneously minimize the intra-class discrepancy as well as the modality discrepancy, we further revise the loss by exchanging the position of two centers:

cc=12(i=1NvD(f(Iiv),cyin)+i=1NnD(f(Iin),cyiv)).{\mathcal{L}}_{cc}=\frac{1}{2}(\sum\limits_{i=1}^{{N_{v}}}{D(f(I_{i}^{v}),{c_{y_{i}^{n}}})}+\sum\limits_{i=1}^{{N_{n}}}{D(f(I_{i}^{n}),{c_{y_{i}^{v}}})}). (6)

Here the cyinc_{y_{i}^{n}} and cyivc_{y_{i}^{v}} refer to the two modality centers of class yiy_{i}. NvN_{v} and NnN_{n} denote the number of VIS and NIR samples respectively in each mini-batch.

IV-D Modality-Aware Spatial Attention

Although using the Cross-Center loss highly improves the performance, the network will easily pay much attention to those textureless regions. These parts provide lower discrepancy but limited discriminative features for re-identification. An intuitive get-around is to concatenate the feature with the higher-level feature, but it will also bring a thorny problem that the higher-level features with better representation ability will dominate the final output. To solve the modality discrepancy under limited samples, the higher-level features tend to overfit in the training stage which in turn decreased the final performance. Therefore, we introduce a modality-aware spatial attention module (MAM) that leverages the high-level feature in every single modality to obtain an attention mAP to strengthen the low-level feature. According to the reflection model, considering the condition that the deep features will be influenced by the modality discrepancy, we use a modality-aware network to match the features with modality discrepancy. Given the output feature of Residual-Block 3 as FF, we divide it into FvF^{v} and FnF^{n} by its modality. The attention mAP FattvF_{att}^{v} and FattnF_{att}^{n} is obtained from the outputs of Residual-Block 4 attached with spatial-wise average pooing layer (SAP) and the Sigmoid function. Specifically, the FattvF_{att}^{v} and FattnF_{att}^{n} could be described as:

Fattm=σ(fS(fm(Fm))),m=v,n,F_{att}^{m}=\sigma({f_{S}}({f_{m}}(F^{m}))),m=v,n, (7)

where we use vv and nn to indicate the VIS and NIR modalities respectively. σ()\sigma(\cdot) represents the sigmoid function, fS()f_{S}(\cdot) represents the spatial-wise average pooing, and fm()f_{m}(\cdot) represents the modality-aware residual blocks.

IV-E Optimization

Besides the cross-center loss, the identity-based softmax cross-entropy loss id\mathcal{L}_{id} and triplet loss tri\mathcal{L}_{tri} with online hard-mining [34] are combined to supervise a more discriminative and robust embedding. Therefore, our loss function for each classifier can be described as:

cls=id+tri+λcc.{\mathcal{L}_{cls}}={\mathcal{L}}_{id}+{\mathcal{L}}_{tri}+\lambda{\mathcal{L}}_{cc}. (8)

In order to balance different parts, we add a parameter λ\lambda to adjust the weight of cross-center loss. To constrain the high-level feature, we utilize the softmax cross-entropy for each modality as idv\mathcal{L}_{id}^{v} for VIS and idn\mathcal{L}_{id}^{n} for NIR. Therefore, the total loss function could be described as:

=idv+idn+k=1Kclsk,\mathcal{L}=\mathcal{L}_{id}^{v}+\mathcal{L}_{id}^{n}+\sum\limits_{k=1}^{K}{\mathcal{L}_{cls}^{k}}, (9)

where the KK refers to the number of parts.

TABLE I: Performance on RegDB and SYSU-MM01 datasets. R-1, 10, 20 denotes the Rank-1, 10, 20 accuracy. * denotes the method using multi-query in the testing phase.
Model Pub. RegDB SYSU-MM01
Visible to Thermal Thermal to Visible All Search Indoor Search
R-1 R-10 R-20 mAP R-1 R-10 R-20 mAP R-1 R-10 R-20 mAP R-1 R-10 R-20 mAP
Zero-Padding[20] ICCV’17 17.8 34.2 44.4 18.9 16.6 34.7 44.3 17.8 14.8 54.1 71.3 15.9 20.6 68.4 85.8 26.9
HCML[9] AAAI’18 24.4 47.5 56.8 20.8 21.7 45.0 55.6 22.2 14.3 53.2 69.2 16.2 24.5 73.3 86.7 30.1
BDTR[8] IJCAI’18 33.6 58.6 67.4 32.8 32.9 58.5 68.4 32.0 17.0 55.4 72.0 19.7 - - - -
cmGAN[7] IJCAI’18 - - - - - - - - 27.0 67.5 80.6 27.8 31.7 77.2 89.2 42.2
HSME[35] AAAI’19 50.9 73.4 81.7 47.0 50.2 72.4 81.1 46.2 20.7 62.8 78.0 23.2 - - - -
AlignGAN[17] ICCV’19 57.9 - - 53.6 56.3 - - 53.4 42.4 85.0 93.7 40.7 45.9 87.6 94.4 54.3
X-Modality[13] AAAI’20 62.2 83.1 91.7 60.2 - - - - 49.9 89.8 96.0 50.7 - - - -
MGE + FMASP[11] IJCV’20 65.1 83.7 - 64.5 - - - - 43.6 86.3 - 45.0 48.6 89.5 - 57.5
DG-VAE[22] MM’20 73.0 86.9 - 71.8 - - - - 59.5 93.8 - 58.5 - - - -
Hi-CMD[19] CVPR’20 70.9 86.4 - 66.0 - - - - 34.9 77.6 - 35.9 - - - -
SSFT[10] CVPR’20 72.3 - - 72.9 71.0 - - 71.7 61.6 89.2 93.9 63.3 70.5 94.9 97.7 72.6
DDAG[21] ECCV’20 69.3 86.2 91.5 63.5 68.1 85.2 90.3 61.8 54.8 90.4 95.8 53.0 61.0 94.1 98.4 68.0
CICL + IAMA[36] AAAI’21 78.8 - - 69.4 77.9 - - 69.4 57.2 94.3 98.4 59.3 66.6 98.8 99.7 74.7
MCLNet[37] ICCV’21 80.3 92.7 96.0 73.1 75.9 90.9 94.6 69.5 65.4 93.3 97.1 62.0 72.6 97.0 99.2 76.6
SMCL[38] ICCV’21 83.9 - - 79.8 83.1 - - 78.6 67.4 92.9 96.8 61.8 68.8 96.6 98.8 75.6
VCD + VML[39] CVPR’21 73.2 - - 71.6 71.8 - - 70.1 60.0 94.2 98.1 58.8 66.1 96.6 99.4 73.0
NFS[40] CVPR’21 80.5 92.0 95.1 72.1 78.0 90.5 93.6 69.8 56.9 91.3 96.5 55.5 62.8 96.5 99.1 69.8
MPANet[23] CVPR’21 82.8 - - 80.7 83.7 - - 80.9 70.6 96.2 98.8 68.2 76.7 98.2 99.6 81.0
MMN[41] MM’21 91.6 97.7 98.9 84.1 87.5 96.0 98.1 80.5 70.6 96.2 99.0 66.9 76.2 97.2 99.3 79.6
RFM This work 93.5 98.4 99.3 87.5 91.0 97.5 98.6 86.6 72.5 97.7 99.6 70.5 81.1 99.6 100.0 84.6

V Experiments

V-A Experimental Setup

Dataset. The proposed RFM is evaluated on two public available cross-spectral Re-ID datasets RegDB [42] and SYSU-MM01 [20]. The Cumulative Matching Characteristic (CMC) and mean Average Precision (mAP) are used as evaluation metrics.

RegDB consists of 412 persons, where each person has 10 RGB images and 10 IR images. We evaluate in both visible-to-infrared and infrared-to-visible settings. Following the evaluation protocol of [9], we repeat 10 trails with a randomly half-half split of the dataset, one-half identities for training and the other for testing. The final results are based on an average of 10 times testing.

SYSU-MM01 is a large-scale dataset collected by four visible cameras and two infrared cameras, including indoor and outdoor environments. The training set contains 22,25822,258 RGB images and 11,90911,909 IR images of 395395 identities, and the testing set contains 9696 identities with 3,8033,803 IR images as the query. The gallery set is determined by the testing mode, which includes all-search and indoor-search. In the all-search mode, all of the images from visible cameras are used as the gallery, while in the indoor-search model, only two indoor visible cameras are used as the gallery.

Implementation details. We employ the Resnet-50 [24] pre-trained on ImageNet [28] as the backbone network. We resize all the input images to 384×192384\times 192 and adopt commonly used horizontal flipping as data augmentation. Following [21], in each mini-batch, we randomly select 7 identities with 8 visible images and 8 infrared images for training. SGD is utilized as the optimizer. During the training stage, the learning rate is initiated from 0.030.03 and is decayed to 0.0030.003 and 0.00030.0003 respectively after 40th and 70th epoch. Following the random erasing [31], the probability of the LTG is set to 0.5, smin=0.02s_{min}=0.02, smax=0.4s_{max}=0.4, rmin=1rmax=0.3r_{min}=\frac{1}{r_{max}}=0.3 and tmin=1e6t_{min}=1e-6. For the CC-Loss, we set its weight λ\lambda as 3.

V-B Comparison with state-of-the-art

In Table I, we evaluate RFM against the previously reported state-of-the-art methods on the RegDB and SYSU-MM01. For all tracks, we could observe that the proposed RFM outperforms the previous methods, which indicates the great adaptability of RFM in the cross-spectral person re-id task. Compared with the SYSU-MM01, samples in the RegDB dataset contain fewer occlusion, less pose variation, and more consistent illumination. As mentioned above, it is much harder to minimize the modality discrepancy in a complex environment. That may be the reason why RFM achieves a larger improvement in RegDB rather than SYSU-MM01. Additionally, compared with the previous methods, the proposed RFM is more efficient with better interpretability and could be used together with other popular approaches.

TABLE II: Ablation study on SYSU-MM01. LTG, CC, and MAM refers to the linear transformation generator, cross-center loss and modality-aware spatial attention.
Model Settings SYSU-MM01
REL CC MAM R-1 R-10 R-20 mAP
Baseline 60.3 93.4 97.5 58.2
+LTG 65.5 95.5 98.4 62.0
+LTG+CC 70.9 97.2 99.3 67.9
+LTG+MAM 66.4 95.8 98.9 63.0
+LTG+CC+MAM 72.5 97.7 99.6 70.5

V-C Ablation Study

Here, we evaluate the influence of the proposed architectural components with different settings and show the quantitative results in Table II. In particular, we select the Resnet-50 backbone with 12 parts PCB strategy as the baseline and conduct the whole experiments in the SYSU-MM01 all-search mode.

Compared to the baseline, add the linear transformation generator could highly improve the Rank-1 accuracy and mAP as 5.2% and 3.8% in respectively. Further adding the cross-center loss (CC Loss) also obtain a great gain to the model. The modality-aware attention module is designed to overcome the tendency of over-focusing on the low modality discrepancy caused by the use of CC Loss. Therefore, it shows better effectiveness when combines with CC Loss.

TABLE III: Comparison of the different types of LTG on SYSU-MM01.
Model Settings (fg()f_{g}(\cdot)) SYSU-MM01
R-1 R-10 R-20 mAP
Random Erasing[31] Constant 70.6 95.6 98.7 68.0
LTG Equal weight 69.2 96.7 99.1 66.3
LTG Beta(0.3, 0.3) 72.3 97.5 99.5 70.2
LTG Beta(0.5, 0.5) 72.5 97.7 99.6 70.5
Refer to caption
Figure 8: Validation results of Cross-Center Loss on SYSU-MM01. Compared with the previous works aforementioned, the CC-Loss shows significant improvements in this task.
TABLE IV: Comparison of the different types of spatial attention strategy on SYSU-MM01. Since the scale of HC-Loss hasn’t been normalized by the batch size, for fair comparison, we test both default and re-implement settings.
Model Settings SYSU-MM01
R-1 R-10 R-20 mAP
Baseline(+LTG) Center-Loss(3.0) 67.5 96.7 99.2 66.0
Baseline(+LTG) HC-Loss(0.5) 68.7 96.3 98.9 65.1
Baseline(+LTG) HC-Loss*(3.0) 69.4 96.5 99.0 65.7
Baseline(+LTG) CC-Loss(3.0) 70.9 97.2 99.3 67.9
TABLE V: Comparison of the different types of spatial attention strategy on SYSU-MM01.
Model Settings SYSU-MM01
R-1 R-10 R-20 mAP
RFM w/o attention 70.9 97.2 99.3 67.9
RFM Spatial-Attention[43] 69.6 96.1 98.8 65.1
RFM MAM 72.5 97.7 99.6 70.5
Refer to caption
Figure 9: Visualization results of LTG. The upper is the input images and the lower is the augmented image. The LTG significantly destroys the modality-relevant information like color while maintaining the modality-irrelevant information like edge.

V-D Discussion

LTG. The LTG uses the linear function to generate image samples with modality discrepancy. Since we utilize a random generator fg()f_{g}(\cdot) to control the transformation, we make a comparison between different fg()f_{g}(\cdot) to show its performance. As shown in Table III, the U-shape beta shape distribution significantly improves the final performance when compared to the uniform distribution. The best performance occurs when the fg()f_{g}(\cdot) is set to Beta(0.5, 0.5). In addition, we also add random erasing into the comparison. Actually, the random erasing could be regarded as a very special case of LTG, when the fg()f_{g}(\cdot) is a constant factor and takes effect only once. Also, the LTG still shows superior performance than the random erasing. We visualize the augmented images in Fig. 9. As is known to all, the modality-relevant information mainly comes from the high similarity in colors, which is considered to be eliminated to train a cross-spectral model. Obviously, the LTG-generated images destroy the color similarity in the single spectral sample pairs by a high modality discrepancy, while maintaining the modality-irrelevant information like edge. Therefore, the LTG largely enforces the network to focus on modality-irrelevant information.

CC-Loss. In Eq. 8, we set a parameter λ\lambda to control the trade-off between the cross-center loss with identity loss and triplet loss. To explore the impact of the hyperparameter, we give an empirical analysis on the SYSU-MM01 datasets and report the results in Figure 8. From the results, we can observe that even adding the cross-center loss with a small weight (0.5), the final accuracy and mAP could improve significantly. The best performance is achieved when the parameter λ\lambda goes to 3. Although a bigger λ\lambda may obtain a more compact feature space for every single identity, the high weight of cross-center loss makes limited optimization for the feature from different identities. In addition, we compare the CC-Loss with some previous works like HC-Loss [33] and center loss [32] to demonstrate the effectiveness of CC-Loss. For a fair comparison, we conduct all the experiments in the LTG-augmented baseline method and show the results in Fig IV. Besides the default setting as HC-Loss (0.5), we conduct another version as HC-Loss(3.0) for a fair comparison. For HC-Loss(3.0), its value is divided by the batch size to keep the same setting with the center loss and the cross-center loss. Although both of HC-Loss and center loss could provide a positive influence in the cross-modality person re-identification, CC-Loss still shows better adaptability in this task.

Modality-aware spatial attention. To demonstrate the effectiveness of the modality-aware spatial attention module (MAM), we compare the modality-aware spatial attention with the widely used spatial attention module [43] and show the results in Table V. We observe that straightly adding the spatial attention module in the current layer can not provide more effective feature representation and decrease the final performance. However, MAM performs much better than the original spatial attention module by combining the higher-layer information as a guidance.

Refer to caption
Figure 10: Visualization results of MAM. (a) and (b) refers to the VIS images and IR images separately. And from the left to right is the input image, original feature map, attention map, and feature mAP after the MAM. Compared with the original feature mAP, MAM helps the features focus more on those discriminative regions like the logos and bag in the above.

VI Limitations

In the foundation of the specific observation in the cross-spectral person re-identification, the proposed LTG may not be a universal data augmentation strategy like random flipping or random erasing. Whether breaking the modality-similarity between the image pairs could make sense in other computer vision tasks still need to be evaluated. On the other hand, the scale of the RegDB and SYSU-MM01 datasets are small. Although the proposed RFM reaches state-of-the-art results in both two datasets, the performance of every component in an open-world scenario hasn’t been verification. Meanwhile, for the reflection model, we only considered a diffuse refection model, without explicit modeling specularities and higher-order reflections. In addition, expressing the E(λ)E(\lambda) by the SPD of the light source includes an assumption that each part gets the same illumination from the light source. It may not be true for shadows, but does not affect the Re-ID task largely, because the regions in the shadow could be treated as a different material. Overall, we believe the reflection model, if used wisely, could still provide a new perspective to effectively solve the modality discrepancy in cross-spectral tasks.

VII Conclusion

In this paper, we address the modality discrepancy in the cross-modality person re-identification task with the robust feature mining network (RFM). We make an attempt to explore how the modality discrepancy occurs in the cross-spectral Re-ID problem and find that the diversity of linear factors across different surfaces is the main culprit. By extending the observation, we introduced the linear transformation generator with good interpretability to mine a more robust feature space for this task. Meanwhile, we adopt the fine-grid backbone with the proposed cross-center loss to minimize the modality discrepancy in the parts. In addition, a modality-aware spatial attention module is attached to guide the network to pay more attention to those textured regions with high discriminability. As a consequence, the proposed RFM reaches a significant improvement compared with the previous state-of-the-art methods in two publicly available datasets RegDB and SYSU-MM01.

Acknowledgments

This work was supported by the National Science Fund for Distinguished Young Scholars (No.62025603), the National Natural Science Foundation of China (No.U1705262, No. 62176222, No. 62176223, No. 62176226, No. 62072386, No. 62072387, No. 62072389, No. 62002305, No. 61772443, No. 61802324 and No. 61702136), Guangdong Basic and Applied Basic Research Foundation (No.2019B1515120049), the Natural Science Foundation of Fujian Province of China (No.2021J01002), and the Fundamental Research Funds for the central universities (No. 20720200077, No. 20720200090 and No. 20720200091).

References

  • [1] C. Eom and B. Ham, “Learning disentangled representation for robust person re-identification,” in Proceedings of the NeurIPS, 2019, pp. 5297–5308.
  • [2] Y. Zhai, S. Lu, Q. Ye, X. Shan, J. Chen, R. Ji, and Y. Tian, “Ad-cluster: Augmented discriminative clustering for domain adaptive person re-identification,” in Proceedings of the CVPR, June 2020.
  • [3] F. Zheng, C. Deng, X. Sun, X. Jiang, X. Guo, Z. Yu, F. Huang, and R. Ji, “Pyramidal person re-identification via multi-loss dynamic training,” in Proceedings of the CVPR, 2019, pp. 8514–8522.
  • [4] Z. Zheng, X. Yang, Z. Yu, L. Zheng, Y. Yang, and J. Kautz, “Joint discriminative and generative learning for person re-identification,” in Proceedings of the CVPR, 2019, pp. 2138–2147.
  • [5] Z. Wang, Z. Wang, Y. Zheng, Y.-Y. Chuang, and S. Satoh, “Learning to reduce dual-level discrepancy for infrared-visible person re-identification,” in Proceedings of the CVPR, 2019, pp. 618–626.
  • [6] M. Brown and S. Süsstrunk, “Multi-spectral sift for scene category recognition,” in Proceedings of the CVPR.   IEEE, 2011, pp. 177–184.
  • [7] P. Dai, R. Ji, H. Wang, Q. Wu, and Y. Huang, “Cross-modality person re-identification with generative adversarial training.” in Proceedings of the IJCAI, 2018, pp. 677–683.
  • [8] M. Ye, Z. Wang, X. Lan, and P. C. Yuen, “Visible thermal person re-identification via dual-constrained top-ranking.” in Proceedings of the IJCAI, 2018, pp. 1092–1099.
  • [9] M. Ye, X. Lan, J. Li, and P. C. Yuen, “Hierarchical discriminative learning for visible thermal person re-identification,” in Proceedings of the AAAI, 2018.
  • [10] Y. Lu, Y. Wu, B. Liu, T. Zhang, B. Li, Q. Chu, and N. Yu, “Cross-modality person re-identification with shared-specific feature transfer,” in Proceedings of the CVPR, June 2020.
  • [11] A. Wu, W.-S. Zheng, S. Gong, and J. Lai, “Rgb-ir person re-identification by cross-modality similarity preservation,” IJCV, pp. 1–21, 2020.
  • [12] G.-A. Wang, T. Z. Yang, J. Cheng, J. Chang, X. Liang, Z. Hou et al., “Cross-modality paired-images generation for rgb-infrared person re-identification,” in Proceedings of the AAAI, 2020.
  • [13] D. Li, X. Wei, X. Hong, and Y. Gong, “Infrared-visible cross-modal person re-identification with an x modality.” in Proceedings of the AAAI, 2020, pp. 4610–4617.
  • [14] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Proceedings of the NeurIPS, 2014, pp. 2672–2680.
  • [15] B. Horn, B. Klaus, and P. Horn, Robot vision.   MIT press, 1986.
  • [16] G. D. Finlayson, S. D. Hordley, C. Lu, and M. S. Drew, “On the removal of shadows from images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 1, pp. 59–68, 2005.
  • [17] G. Wang, T. Zhang, J. Cheng, S. Liu, Y. Yang, and Z. Hou, “Rgb-infrared cross-modality person re-identification via joint pixel and feature alignment,” in Proceedings of the ICCV, 2019, pp. 3623–3632.
  • [18] M. Ye, J. Shen, and L. Shao, “Visible-infrared person re-identification via homogeneous augmented tri-modal learning,” IEEE Transactions on Information Forensics and Security, 2020.
  • [19] S. Choi, S. Lee, Y. Kim, T. Kim, and C. Kim, “Hi-cmd: Hierarchical cross-modality disentanglement for visible-infrared person re-identification,” in Proceedings of the CVPR, 2020, pp. 10 257–10 266.
  • [20] A. Wu, W.-S. Zheng, H.-X. Yu, S. Gong, and J. Lai, “Rgb-infrared cross-modality person re-identification,” in Proceedings of the ICCV, 2017, pp. 5380–5389.
  • [21] M. Ye, J. Shen, D. J. Crandall, L. Shao, and J. Luo, “Dynamic dual-attentive aggregation learning for visible-infrared person re-identification,” in Proceedings of the ECCV, 2020.
  • [22] N. Pu, W. Chen, Y. Liu, E. M. Bakker, and M. S. Lew, “Dual gaussian-based variational subspace disentanglement for visible-infrared person re-identification,” in Proceedings of the ACM Multimedia, 2020.
  • [23] Q. Wu, P. Dai, J. Chen, C.-W. Lin, Y. Wu, F. Huang, B. Zhong, and R. Ji, “Discover cross-modality nuances for visible-infrared person re-identification,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 4330–4339.
  • [24] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the CVPR, 2016, pp. 770–778.
  • [25] C. Peng, X. Gao, N. Wang, and J. Li, “Graphical representation for heterogeneous face recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39, no. 2, pp. 301–312, 2016.
  • [26] J. Chen, D. Yi, J. Yang, G. Zhao, S. Z. Li, and M. Pietikainen, “Learning mappings for face synthesis from near infrared to visual light images,” in Proceedings of the CVPR, 2009, pp. 156–163.
  • [27] B. T. Phong, “Illumination for computer generated pictures,” Communications of the ACM, vol. 18, no. 6, pp. 311–317, 1975.
  • [28] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Proceedings of the CVPR, 2009, pp. 248–255.
  • [29] Y. Sun, L. Zheng, Y. Yang, Q. Tian, and S. Wang, “Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline),” in Proceedings of the ECCV, 2018, pp. 480–496.
  • [30] H. Luo, Y. Gu, X. Liao, S. Lai, and W. Jiang, “Bag of tricks and a strong baseline for deep person re-identification,” in Proceedings of the CVPR Workshops, June 2019.
  • [31] Z. Zhong, L. Zheng, G. Kang, S. Li, and Y. Yang, “Random erasing data augmentation.” in Proceedings of the AAAI, 2020.
  • [32] Y. Wen, K. Zhang, Z. Li, and Y. Qiao, “A discriminative feature learning approach for deep face recognition,” in Proceedings of the ECCV.   Springer, 2016, pp. 499–515.
  • [33] Y. Zhu, Z. Yang, L. Wang, S. Zhao, X. Hu, and D. Tao, “Hetero-center loss for cross-modality person re-identification,” Neurocomputing, vol. 386, pp. 97–109, 2020.
  • [34] F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in Proceedings of the CVPR, 2015, pp. 815–823.
  • [35] Y. Hao, N. Wang, J. Li, and X. Gao, “Hsme: Hypersphere manifold embedding for visible thermal person re-identification,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, 2019, pp. 8385–8392.
  • [36] Z. Zhao, B. Liu, Q. Chu, Y. Lu, and N. Yu, “Joint color-irrelevant consistency learning and identity-aware modality adaptation for visible-infrared cross modality person re-identification,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 4, 2021, pp. 3520–3528.
  • [37] X. Hao, S. Zhao, M. Ye, and J. Shen, “Cross-modality person re-identification via modality confusion and center aggregation,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 16 403–16 412.
  • [38] Z. Wei, X. Yang, N. Wang, and X. Gao, “Syncretic modality collaborative learning for visible infrared person re-identification,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 225–234.
  • [39] X. Tian, Z. Zhang, S. Lin, Y. Qu, Y. Xie, and L. Ma, “Farewell to mutual information: Variational distillation for cross-modal person re-identification,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 1522–1531.
  • [40] Y. Chen, L. Wan, Z. Li, Q. Jing, and Z. Sun, “Neural feature search for rgb-infrared person re-identification,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 587–597.
  • [41] Y. Zhang, Y. Yan, Y. Lu, and H. Wang, “Towards a unified middle modality learning for visible-infrared person re-identification,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 788–796.
  • [42] D. T. Nguyen, H. G. Hong, K. W. Kim, and K. R. Park, “Person recognition system based on a combination of body images from visible light and thermal cameras,” Sensors, vol. 17, no. 3, p. 605, 2017.
  • [43] S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “Cbam: Convolutional block attention module,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 3–19.