pith. machine review for the scientific record. sign in

arxiv: 2604.02808 · v1 · submitted 2026-04-03 · 💻 cs.CV

Recognition: 2 theorem links

· Lean Theorem

CMCC-ReID: Cross-Modality Clothing-Change Person Re-Identification

Guanglin Niu, Hanzi Wang, Haoxuan Xu

Pith reviewed 2026-05-13 20:30 UTC · model grok-4.3

classification 💻 cs.CV
keywords person re-identificationcross-modalityclothing changevisible-infrareddisentangling learningprototype learningsurveillance
0
0 comments X

The pith

A Progressive Identity Alignment Network separates identity cues from clothing and modality variations to match people across visible and infrared images even when outfits change.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Person re-identification must handle both changes in camera type between visible and infrared and changes in clothing across long-term captures. The paper defines this combined challenge as the new Cross-Modality Clothing-Change Re-Identification task and releases the SYSU-CMCC dataset where each identity appears in both modalities wearing different outfits. The proposed Progressive Identity Alignment Network uses Dual-Branch Disentangling Learning to isolate identity features from clothing factors and Bi-Directional Prototype Learning to align embeddings within and across modalities. Experiments on the new dataset show that this network outperforms methods adapted from separate visible-infrared or clothing-change re-identification work.

Core claim

The paper introduces the CMCC-ReID task to match pedestrians when both modality (visible versus infrared) and clothing differ between views. It supplies the SYSU-CMCC benchmark containing paired visible-infrared images of the same identities in distinct outfits. The Progressive Identity Alignment Network progressively mitigates the two sources of variation by applying Dual-Branch Disentangling Learning to produce clothing-agnostic representations and Bi-Directional Prototype Learning to perform intra- and inter-modality contrast that further suppresses clothing interference.

What carries the argument

Progressive Identity Alignment Network (PIA) built from a Dual-Branch Disentangling Learning module that separates identity-related cues from clothing-related factors and a Bi-Directional Prototype Learning module that conducts contrastive alignment across modalities in embedding space.

Load-bearing premise

Identity cues can be reliably isolated from clothing factors in the learned features without losing the discriminative power needed for accurate matching under real surveillance conditions.

What would settle it

If ablation tests on SYSU-CMCC show that removing either the Dual-Branch Disentangling Learning module or the Bi-Directional Prototype Learning module produces no accuracy drop, or if the full PIA network fails to outperform adapted baselines from visible-infrared and clothing-change re-identification.

Figures

Figures reproduced from arXiv: 2604.02808 by Guanglin Niu, Hanzi Wang, Haoxuan Xu.

Figure 1
Figure 1. Figure 1: Motivation of CMCC-ReID. In the long-term surveillance scenarios, two critical issues (i.e., modality discrepancy and clothing variation) are not mutually exclusive but coexist. This observation motivates the CMCC-ReID task, which targets pedestrian matching across both modality and clothing changes. In such long-term surveillance scenario, two critical issues jointly challenge conventional ReID systems: a… view at source ↗
Figure 2
Figure 2. Figure 2: Illustration of the key issues in CMCC-ReID. (a) In the visible modality, the high contrast enables CAL to produce a heatmap with suppressed clothing cues. (b) In the infrared modality, low contrast causes CAL to generate a heatmap dominated by clothing information. (c) Comparison between SAAI and its ResNet-50 baseline shows that direct modality alignment may even degrade performance under clothing variat… view at source ↗
Figure 3
Figure 3. Figure 3: Overview of PIA, which consists of two key components: a Dual-Branch Dis￾entanglement Learning (DBDL) module for clothing-invariant feature extraction in both modalities, and a Bi-Directional Prototype Learning (BPL) module for cross￾modality identity alignment. Through progressive optimization, PIA achieves robust and modality-consistent identity representation. ages of 73 identities. During testing, we a… view at source ↗
Figure 4
Figure 4. Figure 4: (a) Results of parameter sensitivity analysis for key hyperparameters in our model. (b) Visualization of retrieval results. Effectiveness of Linter. From Index-4 to Index-5, adding Linter notably im￾proves Rank-1/mAP by 10.1%/7.0% and 7.7%/8.7% under the V2I and I2V settings, respectively. This indicates that Linter plays a critical role in bridging the cross-modality gap. By enforcing modality-invariant i… view at source ↗
Figure 5
Figure 5. Figure 5: Visualizations of (a) cosine distance distributions and (b) attention maps. achieved when α is set to 0.9, indicating that a moderate momentum facilitates more stable and discriminative prototype updates. 5.4 Visualization Retrieval Results. To qualitatively validate the effectiveness of PIA, Fig. 4b presents Rank-5 retrieval comparisons between PIA and the baseline CAL on SYSU-CMCC. For each query, correc… view at source ↗
read the original abstract

Person Re-Identification (ReID) faces severe challenges from modality discrepancy and clothing variation in long-term surveillance scenario. While existing studies have made significant progress in either Visible-Infrared ReID (VI-ReID) or Clothing-Change ReID (CC-ReID), real-world surveillance system often face both challenges simultaneously. To address this overlooked yet realistic problem, we define a new task, termed Cross-Modality Clothing-Change Re-Identification (CMCC-ReID), which targets pedestrian matching across variations in both modality and clothing. To advance research in this direction, we construct a new benchmark SYSU-CMCC, where each identity is captured in both visible and infrared domains with distinct outfits, reflecting the dual heterogeneity of long-term surveillance. To tackle CMCC-ReID, we propose a Progressive Identity Alignment Network (PIA) that progressively mitigates the issues of clothing variation and modality discrepancy. Specifically, a Dual-Branch Disentangling Learning (DBDL) module separates identity-related cues from clothing-related factors to achieve clothing-agnostic representation, and a Bi-Directional Prototype Learning (BPL) module performs intra-modality and inter-modality contrast in the embedding space to bridge the modality gap while further suppressing clothing interference. Extensive experiments on the SYSU-CMCC dataset demonstrate that PIA establishes a strong baseline for this new task and significantly outperforms existing methods.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper defines a new task called Cross-Modality Clothing-Change Person Re-Identification (CMCC-ReID) that requires matching pedestrians across simultaneous visible-infrared modality shifts and clothing changes. It introduces the SYSU-CMCC benchmark dataset in which each identity appears in both modalities with distinct outfits. The authors propose the Progressive Identity Alignment Network (PIA) whose Dual-Branch Disentangling Learning (DBDL) module separates identity cues from clothing factors and whose Bi-Directional Prototype Learning (BPL) module performs intra- and inter-modality contrastive alignment. Extensive experiments are claimed to show that PIA establishes a strong baseline and significantly outperforms adapted VI-ReID and CC-ReID methods on SYSU-CMCC.

Significance. If the quantitative results and fair baseline comparisons hold, the work is significant because it formalizes and provides the first benchmark for a realistic long-term surveillance scenario that combines two previously studied but separately addressed heterogeneities. The DBDL and BPL modules offer a concrete architectural approach to clothing-agnostic and modality-bridging representations; successful validation would supply both a reproducible starting point and falsifiable performance targets for subsequent CMCC-ReID research.

major comments (3)
  1. [Experiments] Experiments section: the central claim that PIA 'significantly outperforms existing methods' is load-bearing yet unsupported by any reported rank-1, mAP, or CMC numbers, ablation tables, or error bars in the manuscript text. Without these metrics it is impossible to assess the magnitude of improvement or the contribution of DBDL versus BPL.
  2. [Experiments] Baseline adaptation paragraph: the paper must document that all compared VI-ReID and CC-ReID methods were re-trained from scratch on SYSU-CMCC using identical backbone, optimizer, data augmentation, batch size, and loss weighting as PIA. Any deviation in training protocol would render the performance gap non-attributable to the proposed modules.
  3. [Dataset] Dataset construction subsection: the number of identities, total images per modality, train/test splits, and clothing-change statistics for SYSU-CMCC are not stated. These details are required to judge whether the benchmark is sufficiently challenging and whether the reported gains generalize beyond the specific split used.
minor comments (2)
  1. [Method] The abstract states that PIA 'progressively mitigates' clothing and modality issues, but the manuscript never defines the progressive schedule or the order in which DBDL and BPL are applied during training.
  2. [Method] Notation for the prototype embeddings in the BPL module should be introduced with explicit dimensionality and normalization details to avoid ambiguity when readers attempt re-implementation.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive comments, which highlight areas where the manuscript can be strengthened for clarity and reproducibility. We address each major comment below and will revise the manuscript accordingly to incorporate the requested details.

read point-by-point responses
  1. Referee: Experiments section: the central claim that PIA 'significantly outperforms existing methods' is load-bearing yet unsupported by any reported rank-1, mAP, or CMC numbers, ablation tables, or error bars in the manuscript text. Without these metrics it is impossible to assess the magnitude of improvement or the contribution of DBDL versus BPL.

    Authors: We agree that explicit quantitative metrics are essential for evaluating the claims. The manuscript includes experimental tables with rank-1, mAP, CMC curves, and ablation results, but these are not summarized in the main text. In the revision, we will add a concise summary of the key performance numbers (including error bars from multiple runs) directly in the Experiments section, along with explicit discussion of the contributions from DBDL and BPL based on the ablations. revision: yes

  2. Referee: Baseline adaptation paragraph: the paper must document that all compared VI-ReID and CC-ReID methods were re-trained from scratch on SYSU-CMCC using identical backbone, optimizer, data augmentation, batch size, and loss weighting as PIA. Any deviation in training protocol would render the performance gap non-attributable to the proposed modules.

    Authors: We confirm that all baselines were re-implemented and trained from scratch on SYSU-CMCC under identical protocols to PIA. To address the concern, we will expand the baseline adaptation paragraph to explicitly state the shared backbone, optimizer, data augmentation, batch size, and loss weighting details for every compared method, ensuring the performance differences can be attributed to the proposed modules. revision: yes

  3. Referee: Dataset construction subsection: the number of identities, total images per modality, train/test splits, and clothing-change statistics for SYSU-CMCC are not stated. These details are required to judge whether the benchmark is sufficiently challenging and whether the reported gains generalize beyond the specific split used.

    Authors: We acknowledge that these statistics are necessary for assessing the benchmark. In the revised manuscript, we will expand the Dataset construction subsection to report the exact number of identities, total images per modality, train/test splits, and clothing-change statistics (e.g., number of outfit variations per identity). This will allow readers to evaluate the dataset's difficulty and the generalizability of the results. revision: yes

Circularity Check

0 steps flagged

No circularity: new task definition, modules, and empirical claims stand independently

full rationale

The paper defines CMCC-ReID as a new task, constructs the SYSU-CMCC benchmark, and introduces PIA with DBDL (separating identity from clothing cues) and BPL (intra- and inter-modality contrast) modules whose descriptions are functional rather than self-referential. No equations appear that reduce claimed performance or alignment metrics to fitted parameters from the same data by construction, and no self-citation chain is invoked to justify uniqueness or force the architecture. The outperformance statement is presented as an empirical observation on the new dataset, not a renaming or definitional tautology. This is the common case of a self-contained proposal whose central claims do not collapse to their inputs.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

No free parameters, axioms, or invented entities are described in the abstract; the approach relies on standard deep learning contrastive techniques without introducing new physical or mathematical entities.

pith-pipeline@v0.9.0 · 5549 in / 998 out tokens · 51968 ms · 2026-05-13T20:30:05.604261+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Towards Robust Text-to-Image Person Retrieval: Multi-View Reformulation for Semantic Compensation

    cs.CV 2026-04 unverdicted novelty 5.0

    A multi-view semantic reformulation and feature compensation method using LLMs and VLMs improves text-to-image person retrieval accuracy without training and reaches SOTA on three datasets.

Reference graph

Works this paper leans on

64 extracted references · 64 canonical work pages · cited by 1 Pith paper · 1 internal anchor

  1. [1]

    In: CVPR

    Chen, J., Jiang, X., Wang, F., Zhang, J., Zheng, F., Sun, X., Zheng, W.S.: Learning 3d shape feature for texture-insensitive person re-identification. In: CVPR. pp. 8146–8155 (2021)

  2. [2]

    In: CVPR

    Chen, Y., Wan, L., Li, Z., Jing, Q., Sun, Z.: Neural feature search for rgb-infrared person re-identification. In: CVPR. pp. 587–597 (2021)

  3. [3]

    In: CVPR

    Choi, S., Lee, S., Kim, Y., Kim, T., Kim, C.: Hi-cmd: Hierarchical cross-modality disentanglement for visible-infrared person re-identification. In: CVPR. pp. 10257– 10266 (2020)

  4. [4]

    In: AAAI

    Dai,W.,Lu,L.,Li,Z.:Diffusion-basedsyntheticdatagenerationforvisible-infrared person re-identification. In: AAAI. vol. 39, pp. 11185–11193 (2025)

  5. [5]

    In: CVPR

    Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: CVPR. pp. 248–255. Ieee (2009)

  6. [6]

    In: ICCV

    Fang, X., Yang, Y., Fu, Y.: Visible-infrared person re-identification via semantic alignment and affinity inference. In: ICCV. pp. 11270–11279 (2023)

  7. [7]

    In: CVPR

    Feng, J., Wu, A., Zheng, W.S.: Shape-erased feature learning for visible-infrared person re-identification. In: CVPR. pp. 22752–22761 (2023)

  8. [8]

    In: CVPR

    Gu, X., Chang, H., Ma, B., Bai, S., Shan, S., Chen, X.: Clothes-changing person re-identification with rgb modality only. In: CVPR. pp. 1060–1069 (2022)

  9. [9]

    In: CVPR

    Guo, W., Pan, Z., Liang, Y., Xi, Z., Zhong, Z., Feng, J., Zhou, J.: Lidar-based person re-identification. In: CVPR. pp. 17437–17447 (2024)

  10. [10]

    In: CVPR

    Han, K., Gong, S., Huang, Y., Wang, L., Tan, T.: Clothing-change feature aug- mentation for person re-identification. In: CVPR. pp. 22066–22075 (2023)

  11. [11]

    In: ICCV

    Hao, X., Zhao, S., Ye, M., Shen, J.: Cross-modality person re-identification via modality confusion and center aggregation. In: ICCV. pp. 16403–16412 (2021)

  12. [12]

    In: CVPR

    He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR. pp. 770–778 (2016)

  13. [13]

    In: CVPR

    He, W., Deng, Y., Tang, S., Chen, Q., Xie, Q., Wang, Y., Bai, L., Zhu, F., Zhao, R., Ouyang, W., et al.: Instruct-reid: A multi-purpose person re-identification task with instructions. In: CVPR. pp. 17521–17531 (2024)

  14. [14]

    In: CVPR

    Hong, P., Wu, T., Wu, A., Han, X., Zheng, W.S.: Fine-grained shape-appearance mutual learning for cloth-changing person re-identification. In: CVPR. pp. 10513– 10522 (2021)

  15. [15]

    In: ICCV

    Huang, Y., Wu, Q., Xu, J., Zhong, Y., Zhang, Z.: Clothing status awareness for long-term person re-identification. In: ICCV. pp. 11895–11904 (2021)

  16. [16]

    In: AAAI

    Huang, Z., Liu, J., Li, L., Zheng, K., Zha, Z.J.: Modality-adaptive mixup and invariant decomposition for rgb-infrared person re-identification. In: AAAI. vol. 36, pp. 1034–1042 (2022)

  17. [17]

    In: International conference on machine learning

    Ioffe, S., Szegedy, C.: Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning. pp. 448–456. pmlr (2015) 16 H. Xu et al

  18. [18]

    In: ECCV

    Jiang, K., Zhang, T., Liu, X., Qian, B., Zhang, Y., Wu, F.: Cross-modality trans- former for visible-infrared person re-identification. In: ECCV. pp. 480–496 (2022)

  19. [19]

    In: CVPR

    Jin, X., He, T., Zheng, K., Yin, Z., Shen, X., Huang, Z., Feng, R., Huang, J., Chen, Z., Hua, X.S.: Cloth-changing person re-identification from a single image with gait prediction and regularization. In: CVPR. pp. 14278–14287 (2022)

  20. [20]

    NeurIPS33, 18661–18673 (2020)

    Khosla, P., Teterwak, P., Wang, C., Sarna, A., Tian, Y., Isola, P., Maschinot, A., Liu, C., Krishnan, D.: Supervised contrastive learning. NeurIPS33, 18661–18673 (2020)

  21. [21]

    In: CVPR

    Kim, M., Kim, S., Park, J., Park, S., Sohn, K.: Partmix: Regularization strategy to learn part discovery for visible-infrared person re-identification. In: CVPR. pp. 18621–18632 (2023)

  22. [22]

    Adam: A Method for Stochastic Optimization

    Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)

  23. [23]

    In: AAAI

    Li, D., Wei, X., Hong, X., Gong, Y.: Infrared-visible cross-modal person re- identification with an x modality. In: AAAI. vol. 34, pp. 4610–4617 (2020)

  24. [24]

    arXiv preprint arXiv:2005.04966 (2020)

    Li, J., Zhou, P., Xiong, C., Hoi, S.C.: Prototypical contrastive learning of unsuper- vised representations. arXiv preprint arXiv:2005.04966 (2020)

  25. [25]

    IEEE Transactions on Multimedia 25, 8432–8444 (2023)

    Liang, T., Jin, Y., Liu, W., Li, Y.: Cross-modality transformer with modality min- ing for visible-infrared person re-identification. IEEE Transactions on Multimedia 25, 8432–8444 (2023)

  26. [26]

    In: CVPR

    Liang, X., Rawat, Y.S.: Differ: Disentangling identity features via semantic cues for clothes-changing person re-id. In: CVPR. pp. 13980–13989 (2025)

  27. [27]

    TIP32, 5075–5086 (2023)

    Liu, F., Ye, M., Du, B.: Dual level adaptive weighting for cloth-changing person re-identification. TIP32, 5075–5086 (2023)

  28. [28]

    In: ACM MM

    Liu, F., Ye, M., Du, B.: Cloth-aware augmentation for cloth-generalized person re-identification. In: ACM MM. pp. 4053–4062 (2024)

  29. [29]

    In: ICCV

    Liu, F., Kim, M., Gu, Z., Jain, A., Liu, X.: Learning clothing and pose invariant 3d shape representation for long-term person re-identification. In: ICCV. pp. 19617– 19626 (2023)

  30. [30]

    In: CVPR

    Liu, F., Kim, M., Ren, Z., Liu, X.: Distilling clip with dual guidance for learning discriminative human body shape representation. In: CVPR. pp. 256–266 (2024)

  31. [31]

    In: CVPR

    Liu, J., Sun, Y., Zhu, F., Pei, H., Yang, Y., Li, W.: Learning memory-augmented unidirectional metrics for cross-modality person re-identification. In: CVPR. pp. 19366–19375 (2022)

  32. [32]

    In: AAAI

    Lu, H., Zou, X., Zhang, P.: Learning progressive modality-shared transformers for effective visible-infrared person re-identification. In: AAAI. vol. 37, pp. 1835–1843 (2023)

  33. [33]

    In: CVPR

    Lu, Y., Wu, Y., Liu, B., Zhang, T., Li, B., Chu, Q., Yu, N.: Cross-modality person re-identification with shared-specific feature transfer. In: CVPR. pp. 13379–13389 (2020)

  34. [34]

    In: CVPRW

    Luo, H., Gu, Y., Liao, X., Lai, S., Jiang, W.: Bag of tricks and a strong baseline for deep person re-identification. In: CVPRW. pp. 0–0 (2019)

  35. [35]

    In: ACCV (2020)

    Qian, X., Wang, W., Zhang, L., Zhu, F., Fu, Y., Xiang, T., Jiang, Y.G., Xue, X.: Long-term cloth-changing person re-identification. In: ACCV (2020)

  36. [36]

    In: AAAI

    Qiu, L., Chen, S., Yan, Y., Xue, J.H., Wang, D.H., Zhu, S.: High-order struc- ture based middle-feature learning for visible-infrared person re-identification. In: AAAI. vol. 38, pp. 4596–4604 (2024)

  37. [37]

    In: CVPR

    Ren, K., Zhang, L.: Implicit discriminative knowledge learning for visible-infrared person re-identification. In: CVPR. pp. 393–402 (2024)

  38. [38]

    In: CVPRW

    Wan, F., Wu, Y., Qian, X., Chen, Y., Fu, Y.: When person re-identification meets changing clothes. In: CVPRW. pp. 830–831 (2020) CMCC-ReID 17

  39. [39]

    In: AAAI

    Wang, G.A., Zhang, T., Yang, Y., Cheng, J., Chang, J., Liang, X., Hou, Z.G.: Cross-modality paired-images generation for rgb-infrared person re-identification. In: AAAI. vol. 34, pp. 12144–12151 (2020)

  40. [40]

    In: ICCV

    Wang, G., Zhang, T., Cheng, J., Liu, S., Yang, Y., Hou, Z.: Rgb-infrared cross- modality person re-identification via joint pixel and feature alignment. In: ICCV. pp. 3623–3632 (2019)

  41. [41]

    In: CVPR

    Wang, Z., Wang, Z., Zheng, Y., Chuang, Y.Y., Satoh, S.: Learning to reduce dual- level discrepancy for infrared-visible person re-identification. In: CVPR. pp. 618– 626 (2019)

  42. [42]

    In: ICCV

    Wei, Z., Yang, X., Wang, N., Gao, X.: Syncretic modality collaborative learning for visible infrared person re-identification. In: ICCV. pp. 225–234 (2021)

  43. [43]

    In: ECCV

    Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: ECCV. pp. 3–19 (2018)

  44. [44]

    In: ICCV

    Wu, A., Zheng, W.S., Yu, H.X., Gong, S., Lai, J.: Rgb-infrared cross-modality person re-identification. In: ICCV. pp. 5380–5389 (2017)

  45. [45]

    In: ICCV

    Wu, J., Liu, H., Su, Y., Shi, W., Tang, H.: Learning concordant attention via target-aware alignment for visible-infrared person re-identification. In: ICCV. pp. 11122–11131 (2023)

  46. [46]

    In: CVPR

    Wu, Q., Dai, P., Chen, J., Lin, C.W., Wu, Y., Huang, F., Zhong, B., Ji, R.: Discover cross-modality nuances for visible-infrared person re-identification. In: CVPR. pp. 4330–4339 (2021)

  47. [47]

    In: ECCV

    Wu, Y., Meng, L.C., Zichao, Y., Chan, S., Wang, H.Q.: Wrim-net: Wide-ranging information mining network for visible-infrared person re-identification. In: ECCV. pp. 55–72 (2024)

  48. [48]

    In: ICASSP

    Xu, H., Li, B., Niu, G.: Identity-aware feature decoupling learning for clothing- change person re-identification. In: ICASSP. pp. 1–5 (2025)

  49. [49]

    arXiv preprint arXiv:2603.14243 (2026)

    Xu, H., Niu, G.: Bit: Matching-based bi-directional interaction transformation net- work for visible-infrared person re-identification. arXiv preprint arXiv:2603.14243 (2026)

  50. [50]

    In: ICCV

    Xu, P., Zhu, X.: Deepchange: A long-term person re-identification benchmark with clothes change. In: ICCV. pp. 11196–11205 (2023)

  51. [51]

    In: CVPR

    Yang, B., Chen, J., Ye, M.: Shallow-deep collaborative learning for unsupervised visible-infrared person re-identification. In: CVPR. pp. 16870–16879 (2024)

  52. [52]

    IEEE TPAMI43(6), 2029–2046 (2019)

    Yang, Q., Wu, A., Zheng, W.S.: Person re-identification by contour sketch under moderate clothing change. IEEE TPAMI43(6), 2029–2046 (2019)

  53. [53]

    In: CVPR

    Yang, Z., Lin, M., Zhong, X., Wu, Y., Wang, Z.: Good is bad: Causality inspired cloth-debiasing for cloth-changing person re-identification. In: CVPR. pp. 1472– 1481 (2023)

  54. [54]

    In: ICCV

    Ye, M., Ruan, W., Du, B., Shou, M.Z.: Channel augmented joint learning for visible-infrared recognition. In: ICCV. pp. 13567–13576 (2021)

  55. [55]

    Crandall, D., Shao, L., Luo, J.: Dynamic dual-attentive aggre- gation learning for visible-infrared person re-identification

    Ye, M., Shen, J., J. Crandall, D., Shao, L., Luo, J.: Dynamic dual-attentive aggre- gation learning for visible-infrared person re-identification. In: ECCV. pp. 229–247 (2020)

  56. [56]

    IEEE TPAMI44(6), 2872–2893 (2021)

    Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., Hoi, S.C.: Deep learning for person re-identification: A survey and outlook. IEEE TPAMI44(6), 2872–2893 (2021)

  57. [57]

    In: ICCV

    Yu, H., Cheng, X., Peng, W., Liu, W., Zhao, G.: Modality unifying network for visible-infrared person re-identification. In: ICCV. pp. 11185–11195 (2023)

  58. [58]

    In: CVPR

    Yu, S., Li, S., Chen, D., Zhao, R., Yan, J., Qiao, Y.: Cocas: A large-scale clothes changing person dataset for re-identification. In: CVPR. pp. 3400–3409 (2020) 18 H. Xu et al

  59. [59]

    In: CVPR

    Zhang, Q., Lai, C., Liu, J., Huang, N., Han, J.: Fmcnet: Feature-level modality compensation for visible-infrared person re-identification. In: CVPR. pp. 7349– 7358 (2022)

  60. [60]

    In: ECCV

    Zhang, Y., Zhao, S., Kang, Y., Shen, J.: Modality synergy complement learning with cascaded aggregation for visible-infrared person re-identification. In: ECCV. pp. 462–479 (2022)

  61. [61]

    In: CVPR

    Zhang, Y., Wang, H.: Diverse embedding expansion network and low-light cross- modality benchmark for visible-infrared person re-identification. In: CVPR. pp. 2153–2162 (2023)

  62. [62]

    In: ACM MM

    Zhang, Y., Yan, Y., Lu, Y., Wang, H.: Towards a unified middle modality learning for visible-infrared person re-identification. In: ACM MM. pp. 788–796 (2021)

  63. [63]

    In: ICCV

    Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., Tian, Q.: Scalable person re- identification: A benchmark. In: ICCV. pp. 1116–1124 (2015)

  64. [64]

    In: AAAI

    Zhong, Z., Zheng, L., Kang, G., Li, S., Yang, Y.: Random erasing data augmenta- tion. In: AAAI. vol. 34, pp. 13001–13008 (2020)