Cross-modality paired-images generation and augmentation for RGB-infrared person re-identification

Authors: Wang, G., Yang, Y., Zhang, T., Cheng, J., Hou, Z., Tiwari, P. and Pandey, H.M.

Journal: Neural Networks

Volume: 128

Pages: 294-304

eISSN: 1879-2782

ISSN: 0893-6080

DOI: 10.1016/j.neunet.2020.05.008

Abstract:

RGB-Infrared (IR) person re-identification is very challenging due to the large cross-modality variations between RGB and IR images. Considering no correspondence labels between every pair of RGB and IR images, most methods try to alleviate the variations with set-level alignment by reducing marginal distribution divergence between the entire RGB and IR sets. However, this set-level alignment strategy may lead to misalignment of some instances, which limit the performance for RGB–IR Re-ID. Different from existing methods, in this paper, we propose to generate cross-modality paired-images and perform both global set-level and fine-grained instance-level alignments. Our proposed method enjoys several merits. First, our method can perform set-level alignment by disentangling modality-specific and modality-invariant features. Compared with conventional methods, ours can explicitly remove the modality-specific features and the modality variation can be better reduced. Second, given cross-modality unpaired-images of a person, our method can generate cross-modality paired images from exchanged features. With them, we can directly perform instance-level alignment by minimizing distances of every pair of images. Third, our method learns a latent manifold space. In the space, we can random sample and generate lots of images of unseen classes. Training with those images, the learned identity feature space is more smooth can generalize better when test. Finally, extensive experimental results on two standard benchmarks demonstrate that the proposed model favorably against state-of-the-art methods.

Source: Scopus

Cross-modality paired-images generation and augmentation for RGB-infrared person re-identification.

Authors: Wang, G., Yang, Y., Zhang, T., Cheng, J., Hou, Z., Tiwari, P. and Pandey, H.M.

Journal: Neural Netw

Volume: 128

Pages: 294-304

eISSN: 1879-2782

DOI: 10.1016/j.neunet.2020.05.008

Abstract:

RGB-Infrared (IR) person re-identification is very challenging due to the large cross-modality variations between RGB and IR images. Considering no correspondence labels between every pair of RGB and IR images, most methods try to alleviate the variations with set-level alignment by reducing marginal distribution divergence between the entire RGB and IR sets. However, this set-level alignment strategy may lead to misalignment of some instances, which limit the performance for RGB-IR Re-ID. Different from existing methods, in this paper, we propose to generate cross-modality paired-images and perform both global set-level and fine-grained instance-level alignments. Our proposed method enjoys several merits. First, our method can perform set-level alignment by disentangling modality-specific and modality-invariant features. Compared with conventional methods, ours can explicitly remove the modality-specific features and the modality variation can be better reduced. Second, given cross-modality unpaired-images of a person, our method can generate cross-modality paired images from exchanged features. With them, we can directly perform instance-level alignment by minimizing distances of every pair of images. Third, our method learns a latent manifold space. In the space, we can random sample and generate lots of images of unseen classes. Training with those images, the learned identity feature space is more smooth can generalize better when test. Finally, extensive experimental results on two standard benchmarks demonstrate that the proposed model favorably against state-of-the-art methods.

Source: PubMed

Cross-modality paired-images generation and augmentation for RGB-infrared person re-identification

Authors: Wang, G., Yang, Y., Zhang, T., Cheng, J., Hou, Z., Tiwari, P. and Pandey, H.M.

Journal: NEURAL NETWORKS

Volume: 128

Pages: 294-304

eISSN: 1879-2782

ISSN: 0893-6080

DOI: 10.1016/j.neunet.2020.05.008

Source: Web of Science (Lite)

Cross-modality paired-images generation and augmentation for RGB-infrared person re-identification.

Authors: Wang, G., Yang, Y., Zhang, T., Cheng, J., Hou, Z., Tiwari, P. and Pandey, H.M.

Journal: Neural networks : the official journal of the International Neural Network Society

Volume: 128

Pages: 294-304

eISSN: 1879-2782

ISSN: 0893-6080

DOI: 10.1016/j.neunet.2020.05.008

Abstract:

RGB-Infrared (IR) person re-identification is very challenging due to the large cross-modality variations between RGB and IR images. Considering no correspondence labels between every pair of RGB and IR images, most methods try to alleviate the variations with set-level alignment by reducing marginal distribution divergence between the entire RGB and IR sets. However, this set-level alignment strategy may lead to misalignment of some instances, which limit the performance for RGB-IR Re-ID. Different from existing methods, in this paper, we propose to generate cross-modality paired-images and perform both global set-level and fine-grained instance-level alignments. Our proposed method enjoys several merits. First, our method can perform set-level alignment by disentangling modality-specific and modality-invariant features. Compared with conventional methods, ours can explicitly remove the modality-specific features and the modality variation can be better reduced. Second, given cross-modality unpaired-images of a person, our method can generate cross-modality paired images from exchanged features. With them, we can directly perform instance-level alignment by minimizing distances of every pair of images. Third, our method learns a latent manifold space. In the space, we can random sample and generate lots of images of unseen classes. Training with those images, the learned identity feature space is more smooth can generalize better when test. Finally, extensive experimental results on two standard benchmarks demonstrate that the proposed model favorably against state-of-the-art methods.

Source: Europe PubMed Central