Abstract
Existing vehicle re-identification (Re-ID) methods either extract valuable background information to enhance the robustness of the vehicle model or segment background interference information to learn vehicle fine-grained information. However, these methods do not consider the background information as a trade-off attribute to unite valuable background and background interference. This work proposes the trade-off background joint learning method for unsupervised vehicle Re-ID, which consists of two branches, to exploit the ambivalence of background information. In the global branch, a background focus of the pyramid global branch module is proposed to optimize the sample feature space. The designed pyramid background-aware attention extracts background-related features from the global image and constructs a two-fold confidence metric based on background-related and identity-related confidence scores to obtain robust clustering results during the clustering. In the local branch, a background filtering of the local branch module is proposed to alleviate the background interference. First, the background of each local region is segmented and weakened. Then, a background adaptive local label smoothing is designed to reduce noise in every local region. Comprehensive experiments on VeRi-776 and VeRi-Wild are conducted to validate the performance of the proposed balanced background information method. Experimental results show that the proposed method outperforms the state-of-the-art.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
The datasets analyzed in the current study are available in the following ways: The VeRi-776 dataset can be obtained from https://github.com/VehicleReId/VeRi; the VeRi-Wild dataset can be obtained from https://github.com/PKU-IMRE/VERI-Wild.
References
Zheng, Z., Jiang, M., Wang, Z., Wang, J., Ding, E.: Going beyond real data: a robust visual representation for vehicle re-identification (2020)
Zhuge, C., Peng, Y., Li, Y., Ai, J., Chen, J.: Attribute-guided feature extraction and augmentation robust learning for vehicle re-identification (2020)
Chen, X., Sui, H., Fang, J., Feng, W., Zhou, M.: Vehicle re-identification using distance-based global and partial multi-regional feature learning. IEEE Trans. Intell. Transp. Syst. PP, 1–11 (2020)
Lu, Z., Lin, R., Lou, X., Zheng, L., Hu, H.: Identity-unrelated information decoupling model for vehicle re-identification. IEEE Trans. Intell. Transp. Syst. 23, 19001–19015 (2022)
Fu, Y., et al.: Self-similarity grouping: a simple unsupervised cross domain adaptation approach for person re-identification, pp. 6112–6121 (2019)
Wang, D., Zhang, S.: Unsupervised person re-identification via multi-label classification, pp. 10981–10990 (2020)
Yang, Q., Yu, H.-X., Wu, A., Zheng, W.-S.: Patch-based discriminative feature learning for unsupervised person re-identification, pp. 3633–3642 (2019)
Sun, Y., Zheng, L., Yang, Y., Tian, Q., Wang, S.: Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline), pp. 480–496 (2018)
Yu, J., Oh, H.: Unsupervised vehicle re-identification via self-supervised metric learning using feature dictionary (2021)
Zheng, A., Sun, X., Li, C., Tang, J.: Viewpoint-aware progressive clustering for unsupervised vehicle re-identification. IEEE (2021)
Yu, J., Kim, J., Kim, M., Oh, H.: Camera-tracklet-aware contrastive learning for unsupervised vehicle re-identification (2021)
Zhu, W., Peng, B.: Manifold-based aggregation clustering for unsupervised vehicle re-identification. Knowl.-Based Syst. 235, 107624 (2022)
Munir, A., Martinel, N., Micheloni, C.: Oriented splits network to distill background for vehicle re-identification, pp. 1–8. IEEE (2021)
Wu, M., Zhang, Y., Zhang, T., Zhang, W.: Background segmentation for vehicle re-identification (2020)
Peng, J., et al.: Eliminating cross-camera bias for vehicle re-identification (2019)
Khorramshahi, P., Peri, N., Chen, J.C., Chellappa, R.: The devil is in the details: self-supervised attention for vehicle re-identification (2020)
Zhu, X., Luo, Z., Fu, P., Ji, X.: VOC-ReID: vehicle re-identification based on vehicle-orientation-camera. IEEE (2020)
Chen, J., Lu, Y., Yu, Q., Luo, X., Zhou, Y.: Transunet: Transformers make strong encoders for medical image segmentation (2021)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: Convolutional block attention module, pp. 3–19 (2018)
Liu, X., Liu, W., Mei, T., Ma, H.: A Deep Learning-Based Approach to Progressive Vehicle Re-identification for Urban Surveillance, pp. 869–884. Springer, Berlin (2016)
Lou, Y., Bai, Y., Liu, J., Wang, S., Duan, L.: Veri-wild: a large dataset and a new method for vehicle re-identification in the wild, pp. 3235–3243 (2019)
Cho, Y., Kim, W.J., Hong, S., Yoon, S.E.: Part-based pseudo label refinement for unsupervised person re-identification. arXiv e-prints (2022)
Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605 (2008)
Yang, M., et al.: Learning with twin noisy labels for visible-infrared person re-identification, pp. 14308–14317 (2022)
Huang, Z., et al.: Learning with noisy correspondence for cross-modal matching. Adv. Neural. Inf. Process. Syst. 34, 29406–29419 (2021)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition, pp. 770–778 (2016)
Pan, X., Luo, P., Shi, J., Tang, X.: Two at once: enhancing learning and generalization capacities via ibn-net, pp. 464–479 (2018)
Ge, Y., Chen, D., Li, H.: Mutual mean-teaching: Pseudo label refinery for unsupervised domain adaptation on person re-identification. arXiv preprint arXiv:2001.01526 (2020)
Ding, Y., Fan, H., Xu, M., Yang, Y.: Adaptive exploration for unsupervised person re-identification. ACM Trans. Multimedia Comput. Commun. Appl. (TOMM) 16, 1–19 (2020)
Ge, Y., Chen, D., Zhu, F., Zhao, R., Li, H.: Self-paced contrastive learning with hybrid memory for domain adaptive object Re-ID (2020)
Yu, J., Oh, H.: Unsupervised person re-identification via multi-label prediction and classification based on graph-structural insight (2021)
Hu, Z., Zhu, C., He, G.: Hard-sample guided hybrid contrast learning for unsupervised person re-identification (2021)
Yang, F., Zhong, Z., Luo, Z., Cai, Y., Sebe, N.: Joint noise-tolerant learning and meta camera shift adaptation for unsupervised person re-identification (2021)
Han, X., et al.: Rethinking sampling strategies for unsupervised person re-identification (2021)
Zhang, X., Ge, Y., Qiao, Y., Li, H.: Refining pseudo labels with clustering consensus over generations for unsupervised object re-identification (2021)
Chen, H., Lagadec, B., Bremond, F.: ICE: inter-instance contrastive encoding for unsupervised person re-identification (2021)
Li, M., Li, C.-G., Guo, J.: Cluster-guided asymmetric contrastive learning for unsupervised person re-identification. IEEE Trans. Image Process. 31, 3606–3617 (2022)
Selvaraju, R.R., et al.: Grad-CAM: Visual explanations from deep networks via gradient-based localization, pp. 618–626 (2017)
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant No. 62076117 and 62166026, the Jiangxi Key Laboratory of Smart City under Grant No. 20192BCD40002 and the Jiangxi Provincial Natural Science Foundation under Grant No. 20224BAB212.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have no competing interests to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, S., Wang, Q., Min, W. et al. Trade-off background joint learning for unsupervised vehicle re-identification. Vis Comput 39, 3823–3835 (2023). https://doi.org/10.1007/s00371-023-03034-2
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00371-023-03034-2