pith. machine review for the scientific record. sign in

arxiv: 2605.01309 · v1 · submitted 2026-05-02 · 💻 cs.CV

Recognition: unknown

CUE: Concept-Aware Multi-Label Expansion to Mitigate Concept Confusion in Long-Tailed Learning

Authors on Pith no claims yet

Pith reviewed 2026-05-09 14:35 UTC · model grok-4.3

classification 💻 cs.CV
keywords long-tailed learningconcept confusionmulti-label expansionCLIPlarge language modelslogit adjustmentimage classificationclass imbalance
0
0 comments X

The pith

Multi-label concept signals from CLIP visuals and LLM semantics preserve inter-class feature sharing disrupted by single-label training on long-tailed data.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Long-tailed distributions create concept confusion because single-label supervision treats classes as mutually exclusive, which blocks related classes from sharing visual features and lets head classes overpower the model. CUE addresses this by building multi-label concept sets that combine instance-level visual cues extracted from zero-shot CLIP with class-level semantic cues generated by an LLM. These signals enter training as separately weighted Binary Logit-Adjustment auxiliary losses that run alongside the standard Logit-Adjustment loss. The result is a model that maintains better discriminability across head and tail classes while still correcting for sample imbalance.

Core claim

CUE constructs concept sets by extracting instance-level visual cues from zero-shot CLIP and generating class-level semantic cues with an LLM; the two cues are incorporated via separately weighted Binary Logit-Adjustment auxiliary losses and jointly optimized with the baseline Logit-Adjustment loss, which mitigates the concept confusion caused by mutual exclusivity of single-label supervision under long-tailed distributions.

What carries the argument

CUE, the concept-aware multi-label expansion that builds auxiliary concept sets from CLIP visual cues and LLM semantic cues and folds them into weighted Binary Logit-Adjustment losses.

If this is right

  • The auxiliary losses let the model learn shared features among concept-related classes without undoing the tail-class boost from logit adjustment.
  • Performance becomes more balanced because head-class dominance is reduced while tail-class accuracy rises.
  • The same cue construction and loss weighting can be applied on top of other logit-adjustment or re-weighting baselines.
  • Joint optimization of the main loss and the two auxiliary losses ensures the added signals reinforce rather than compete with distribution correction.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If cue quality scales with future improvements in vision-language and language models, long-tailed recognition accuracy would continue to rise without new labeled data.
  • The same multi-label expansion idea could be tested in non-vision long-tailed settings by swapping CLIP for domain-appropriate cue extractors.
  • Manual inspection of the extracted concept sets on a small held-out set would show whether the method's success depends on high cue precision.

Load-bearing premise

The cues pulled from zero-shot CLIP and generated by the LLM correctly identify genuine inter-class relationships and do not add enough noise or wrong multi-label assignments to hurt overall discriminability.

What would settle it

Retraining the identical baseline with the same auxiliary losses but replacing the CLIP and LLM concept sets with random or shuffled labels; if accuracy falls back to or below the plain Logit-Adjustment baseline, the specific content of the cues is required for the reported gains.

Figures

Figures reproduced from arXiv: 2605.01309 by Chikai Shang, Jiacheng Yang, Junlong Gao, Mengke Li, Ruichi Zhang, Yang Lu, Yang Zhou.

Figure 1
Figure 1. Figure 1: Motivation: Handling Concept Confusion, Not Just Class Imbalance. (a) On CIFAR100-LT (IR=100), fine-tuning CLIP with LIFT [33] turns many zero-shot correct predictions into errors, especially on tail classes. (b) Most of these errors are misclassified into semantically related categories, a phenomenon we term as concept confusion. This confusion mainly arises from the fine-tuning process, which disrupts in… view at source ↗
Figure 2
Figure 2. Figure 2: Overall framework. This figure illustrates the overall structure of our approach compared with the existing long-tailed learning methods. (a) Existing LTL methods rely on a single-label balanced loss, which tends to suppress semantically related categories and bias learning toward head classes, often leading to concept confusion. (b) CUE augments the main balanced loss with two concept-aware cues: (I) VLM-… view at source ↗
Figure 3
Figure 3. Figure 3 view at source ↗
Figure 4
Figure 4. Figure 4: Sensitivity analysis of CUE with respect to the weight view at source ↗
Figure 5
Figure 5. Figure 5: Grad-CAM visualization comparing LIFT, CUE and CLIP on several examples. By sharing related cues across correlated categories, CUE produces more precise and coherent localization than both LIFT and CLIP view at source ↗
read the original abstract

Long-tailed distributions are common in real-world recognition tasks, where a few head classes have many samples while most tail classes have very few. Recently, fine-tuning foundation models for long-tailed learning has gained attention due to their excellent performance. However, most existing methods focus solely on mitigating long-tailed distribution bias while overlooking concept confusion caused by the long-tailed distribution. In this paper, we study this problem and attribute it to the mutual exclusivity of single-label supervision under long-tailed distributions, which suppresses feature sharing among related classes and amplifies the dominance of head classes, leading to disrupted inter-class discriminability. To address this, we propose CUE, Concept-aware mUlti-label Expansion, which introduces multi-label concept signals to preserve disrupted inter-class relationships. Specifically, CUE constructs concept sets by (i) extracting instance-level visual cues from zero-shot CLIP and (ii) generating class-level semantic cues with LLM; the two cues are incorporated via separately weighted Binary Logit-Adjustment (BLA) auxiliary losses and jointly optimized with the baseline Logit-Adjustment (LA) loss. Experiments on several long-tailed benchmarks, CUE achieves balanced and strong performance, surpassing recent state-of-the-art methods. Code is available at: https://github.com/zhangruichi/CUE.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper claims that long-tailed distributions cause concept confusion by enforcing mutual exclusivity in single-label supervision, which suppresses feature sharing among related classes and amplifies head-class dominance. To address this, CUE extracts instance-level visual cues via zero-shot CLIP and class-level semantic cues via LLM to expand labels into multi-label concept sets; these are incorporated through separately weighted Binary Logit-Adjustment (BLA) auxiliary losses that are jointly optimized with the baseline Logit-Adjustment (LA) loss. Experiments on long-tailed benchmarks are reported to yield balanced performance that surpasses recent state-of-the-art methods.

Significance. If the central claim holds, the work usefully identifies an overlooked source of inter-class discriminability loss in long-tailed learning and demonstrates a practical way to restore it by leveraging foundation-model cues without extra labeled data. The open-sourced code at the provided GitHub link is a clear strength for reproducibility.

major comments (2)
  1. [§3.2] §3.2 (Concept Set Construction): the central claim that the CLIP and LLM cues accurately recover disrupted inter-class relationships rests on an unverified assumption; the manuscript should report quantitative validation of cue quality (e.g., precision/recall of the generated multi-label assignments on a held-out subset or human annotation study), broken down by head vs. tail classes, because zero-shot CLIP accuracy is known to degrade on tail classes and LLM cues are generic.
  2. [§4] §4 (Experiments): the reported gains on long-tailed benchmarks do not isolate whether improvements stem from faithful concept signals or from the mere addition of auxiliary regularization; an ablation replacing the extracted cues with random or empty multi-label targets while keeping the BLA losses should be added to test this.
minor comments (2)
  1. [Abstract] Abstract: the phrase 'several long-tailed benchmarks' should name the specific datasets (e.g., CIFAR-100-LT, ImageNet-LT) for immediate clarity.
  2. [§3.3] Notation: the weighting factors for the two BLA losses are introduced as free parameters; a brief sensitivity analysis or default-value justification would improve reproducibility.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and insightful comments, which help clarify the presentation of our central claims. We address each major comment below and will incorporate the suggested analyses into the revised manuscript to strengthen the evidence for our approach.

read point-by-point responses
  1. Referee: [§3.2] §3.2 (Concept Set Construction): the central claim that the CLIP and LLM cues accurately recover disrupted inter-class relationships rests on an unverified assumption; the manuscript should report quantitative validation of cue quality (e.g., precision/recall of the generated multi-label assignments on a held-out subset or human annotation study), broken down by head vs. tail classes, because zero-shot CLIP accuracy is known to degrade on tail classes and LLM cues are generic.

    Authors: We agree that direct quantitative validation of cue quality would provide stronger support for the assumption that the CLIP and LLM cues recover meaningful inter-class relationships. While the downstream performance gains on long-tailed benchmarks offer indirect evidence of cue utility, we acknowledge that explicit metrics are needed, especially given known limitations of zero-shot CLIP on tail classes. In the revision, we will add a cue-quality analysis including precision/recall of the generated multi-label assignments against a held-out subset (where ground-truth multi-label annotations can be obtained) and a small-scale human annotation study on randomly sampled images. Results will be broken down by head versus tail classes to quantify any degradation. revision: yes

  2. Referee: [§4] §4 (Experiments): the reported gains on long-tailed benchmarks do not isolate whether improvements stem from faithful concept signals or from the mere addition of auxiliary regularization; an ablation replacing the extracted cues with random or empty multi-label targets while keeping the BLA losses should be added to test this.

    Authors: We concur that an ablation isolating the contribution of the specific concept signals versus generic auxiliary regularization is necessary to substantiate our claims. To address this, we will include a new ablation in Section 4 where the extracted multi-label concept sets are replaced with random or empty targets while retaining the Binary Logit-Adjustment (BLA) losses and their weighting scheme. This will allow direct comparison to the original CUE results and clarify whether the observed balanced performance improvements arise from the faithful visual and semantic cues or from the multi-label loss structure alone. revision: yes

Circularity Check

0 steps flagged

No circularity: method and losses defined independently of results

full rationale

The paper defines CUE by constructing concept sets from external zero-shot CLIP instance cues and LLM class cues, then applies separately weighted BLA auxiliary losses jointly optimized with baseline LA loss. This construction and optimization are specified without reference to the final performance numbers or to any fitted parameter that is later renamed as a prediction. No equations reduce the inter-class discriminability claim to a self-definition, no self-citation chain supplies a uniqueness theorem, and no ansatz is smuggled via prior work by the same authors. Evaluation occurs on standard external long-tailed benchmarks, keeping the derivation self-contained.

Axiom & Free-Parameter Ledger

1 free parameters · 2 axioms · 0 invented entities

The central claim rests on the quality of cues produced by pre-trained foundation models and on the assumption that adding multi-label signals via auxiliary losses will restore inter-class relationships without side effects.

free parameters (1)
  • BLA auxiliary loss weights
    The description of separately weighted Binary Logit-Adjustment auxiliary losses implies tunable scalar weights whose values are not reported in the abstract.
axioms (2)
  • domain assumption Zero-shot CLIP extracts reliable instance-level visual concept cues
    Invoked when constructing concept sets from images.
  • domain assumption LLMs generate accurate class-level semantic concept cues
    Invoked when constructing concept sets from class names.

pith-pipeline@v0.9.0 · 5549 in / 1476 out tokens · 45663 ms · 2026-05-09T14:35:10.732706+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

51 extracted references · 4 canonical work pages · 1 internal anchor

  1. [1]

    Cdul: Clip-driven unsupervised learning for multi-label image classification

    Rabab Abdelfattah, Qing Guo, Xiaoguang Li, Xiaofeng Wang, and Song Wang. Cdul: Clip-driven unsupervised learning for multi-label image classification. InICCV, pages 1348–1357, 2023. 3

  2. [2]

    Application and development of artificial intelligence and intelligent disease diagnosis.Current pharmaceutical design, 26(26):3069–3075, 2020

    Chunyan Ao, Shunshan Jin, Hui Ding, Quan Zou, and Liang Yu. Application and development of artificial intelligence and intelligent disease diagnosis.Current pharmaceutical design, 26(26):3069–3075, 2020. 2

  3. [3]

    On the Opportunities and Risks of Foundation Models

    Rishi Bommasani. On the opportunities and risks of founda- tion models.arXiv preprint arXiv:2108.07258, 2021. 2

  4. [4]

    Lan- guage models are few-shot learners

    Tom Brown, Benjamin Mann, Nick Ryder, Melanie Sub- biah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakan- tan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Lan- guage models are few-shot learners. 33:1877–1901, 2020. 2

  5. [5]

    Learning imbalanced datasets with label- distribution-aware margin loss

    Kaidi Cao, Colin Wei, Adrien Gaidon, Nikos Arechiga, and Tengyu Ma. Learning imbalanced datasets with label- distribution-aware margin loss. InNeurIPS, 2019. 2, 5

  6. [6]

    Softmatch: Addressing the quantity-quality tradeoff in semi- supervised learning

    Hao Chen, Ran Tao, Yue Fan, Yidong Wang, Jindong Wang, Bernt Schiele, Xing Xie, Bhiksha Raj, and Marios Savvides. Softmatch: Addressing the quantity-quality tradeoff in semi- supervised learning. InICLR, 2023. 3

  7. [7]

    Adaptformer: Adapt- ing vision transformers for scalable visual recognition

    Shoufa Chen, Chongjian Ge, Zhan Tong, Jiangliu Wang, Yibing Song, Jue Wang, and Ping Luo. Adaptformer: Adapt- ing vision transformers for scalable visual recognition. In NeurIPS, pages 16664–16678, 2022. 3, 5, 7

  8. [8]

    Reslt: Residual learning for long-tailed recogni- tion.IEEE TPAMI, 45(3):3695–3706, 2022

    Jiequan Cui, Shu Liu, Zhuotao Tian, Zhisheng Zhong, and Jiaya Jia. Reslt: Residual learning for long-tailed recogni- tion.IEEE TPAMI, 45(3):3695–3706, 2022. 7, 8

  9. [9]

    Class-balanced loss based on effective number of samples

    Yin Cui, Menglin Jia, Tsung-Yi Lin, Yang Song, and Serge Belongie. Class-balanced loss based on effective number of samples. InCVPR, pages 9268–9277, 2019. 2

  10. [10]

    LPT: Long-tailed prompt tuning for image classifica- tion

    Bowen Dong, Pan Zhou, Shuicheng Y AN, and Wangmeng Zuo. LPT: Long-tailed prompt tuning for image classifica- tion. InICLR, 2023. 2, 3, 5, 6

  11. [11]

    Dtl: Parameter-and memory-efficient disentangled vision learn- ing.IEEE TPAMI, 2025

    Minghao Fu, Ke Zhu, Zonghao Ding, and Jianxin Wu. Dtl: Parameter-and memory-efficient disentangled vision learn- ing.IEEE TPAMI, 2025. 3

  12. [12]

    Improving diffusion models for class-imbalanced training data via capacity manipulation

    Feng Hong, Jiangchao Yao, Yifei Shen, Dongsheng Li, Ya Zhang, and Yanfeng Wang. Improving diffusion models for class-imbalanced training data via capacity manipulation. In ICLR, 2026. 2

  13. [13]

    Disentangling label dis- tribution for long-tailed visual recognition

    Youngkyu Hong, Seungju Han, Kwanghee Choi, Seokjun Seo, Beomsu Kim, and Buru Chang. Disentangling label dis- tribution for long-tailed visual recognition. InCVPR, pages 6626–6636, 2021. 2

  14. [14]

    Learning to explore sample relationships

    Zhi Hou, Baosheng Yu, Chaoyue Wang, Yibing Zhan, and Dacheng Tao. Learning to explore sample relationships. IEEE TPAMI, 2025. 2

  15. [15]

    Parameter-efficient transfer learning for nlp

    Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin De Laroussilhe, Andrea Gesmundo, Mona Attariyan, and Sylvain Gelly. Parameter-efficient transfer learning for nlp. InICML, pages 2790–2799. PMLR, 2019. 3, 7

  16. [16]

    LoRA: Low-rank adaptation of large language models

    Edward J Hu, yelong shen, Phillip Wallis, Zeyuan Allen- Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. LoRA: Low-rank adaptation of large language models. In ICLR, 2022. 2, 3, 7

  17. [17]

    Simple: Similar pseudo label exploitation for semi- supervised classification

    Zijian Hu, Zhengyu Yang, Xuefeng Hu, and Ram Neva- tia. Simple: Similar pseudo label exploitation for semi- supervised classification. InCVPR, pages 15099–15108,

  18. [18]

    Scaling up visual and vision-language representation learning with noisy text supervision

    Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. InICML, pages 4904– 4916, 2021. 2

  19. [19]

    Vi- sual prompt tuning

    Menglin Jia, Luming Tang, Bor-Chun Chen, Claire Cardie, Serge Belongie, Bharath Hariharan, and Ser-Nam Lim. Vi- sual prompt tuning. InECCV, pages 709–727, 2022. 2, 3, 7

  20. [20]

    Mimic-cxr, a de- identified publicly available database of chest radiographs with free-text reports.Scientific data, 6(1):317, 2019

    Alistair EW Johnson, Tom J Pollard, Seth J Berkowitz, Nathaniel R Greenbaum, Matthew P Lungren, Chih-ying Deng, Roger G Mark, and Steven Horng. Mimic-cxr, a de- identified publicly available database of chest radiographs with free-text reports.Scientific data, 6(1):317, 2019. 2

  21. [21]

    Exploring balanced feature spaces for representation learn- ing

    Bingyi Kang, Yu Li, Sa Xie, Zehuan Yuan, and Jiashi Feng. Exploring balanced feature spaces for representation learn- ing. InICLR, 2020. 7

  22. [22]

    Decou- pling representation and classifier for long-tailed recogni- tion

    Bingyi Kang, Saining Xie, Marcus Rohrbach, Zhicheng Yan, Albert Gordo, Jiashi Feng, and Yannis Kalantidis. Decou- pling representation and classifier for long-tailed recogni- tion. InICLR, 2020. 2

  23. [23]

    Long-tailed visual recognition via gaussian clouded logit adjustment

    Mengke Li, Yiu-ming Cheung, and Yang Lu. Long-tailed visual recognition via gaussian clouded logit adjustment. In CVPR, pages 6929–6938, 2022. 2

  24. [24]

    Improving visual prompt tuning by gaussian neighborhood minimization for long-tailed visual recognition

    Mengke Li, Ye Liu, Yang Lu, Yiqun Zhang, Yiu-ming Che- ung, and Hui Huang. Improving visual prompt tuning by gaussian neighborhood minimization for long-tailed visual recognition. InNeurIPS, pages 103985–104009, 2024. 2

  25. [25]

    Large-scale long-tailed recognition in an open world

    Ziwei Liu, Zhongqi Miao, Xiaohang Zhan, Jiayun Wang, Boqing Gong, and Stella X Yu. Large-scale long-tailed recognition in an open world. InCVPR, pages 2537–2546,

  26. [26]

    Retrieval augmented clas- sification for long-tail visual recognition

    Alexander Long, Wei Yin, Thalaiyasingam Ajanthan, Vu Nguyen, Pulak Purkait, Ravi Garg, Alan Blair, Chunhua Shen, and Anton Van den Hengel. Retrieval augmented clas- sification for long-tail visual recognition. InCVPR, pages 6959–6969, 2022. 2, 5, 6

  27. [27]

    Long-tailed recognition with model re- balancing.NeurIPS, 2025

    Jiaan Luo, Feng Hong, Qiang Hu, Xiaofeng Cao, Feng Liu, and Jiangchao Yao. Long-tailed recognition with model re- balancing.NeurIPS, 2025. 2

  28. [28]

    A simple long- tailed recognition baseline via vision-language model.arXiv preprint arXiv:2111.14745, 2021

    Teli Ma, Shijie Geng, Mengmeng Wang, Jing Shao, Jiasen Lu, Hongsheng Li, Peng Gao, and Yu Qiao. A simple long- tailed recognition baseline via vision-language model.arXiv preprint arXiv:2111.14745, 2021. 5, 6

  29. [29]

    Lessons and insights from a unifying study of parameter-efficient fine-tuning (peft) in visual recognition

    Zheda Mai, Ping Zhang, Cheng-Hao Tu, Hong-You Chen, Quang-Huy Nguyen, Li Zhang, and Wei-Lun Chao. Lessons and insights from a unifying study of parameter-efficient fine-tuning (peft) in visual recognition. InCVPR, pages 14845–14857, 2025. 3

  30. [30]

    Long-tail learning via logit adjustment

    Aditya Krishna Menon, Sadeep Jayasumana, Ankit Singh Rawat, Himanshu Jain, Andreas Veit, and Sanjiv Kumar. Long-tail learning via logit adjustment. InICLR, 2021. 2, 3, 7, 8

  31. [31]

    Learn- ing transferable visual models from natural language super- vision

    Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learn- ing transferable visual models from natural language super- vision. InICML, pages 8748–8763, 2021. 2

  32. [32]

    How re-sampling helps for long-tail learning? InNeurIPS, pages 75669–75687, 2023

    Jiang-Xin Shi, Tong Wei, Yuke Xiang, and Yu-Feng Li. How re-sampling helps for long-tail learning? InNeurIPS, pages 75669–75687, 2023. 2

  33. [33]

    Long-tail learning with foundation model: Heavy fine-tuning hurts

    Jiang-Xin Shi, Tong Wei, Zhi Zhou, Jie-Jing Shao, Xin-Yan Han, and Yu-Feng Li. Long-tail learning with foundation model: Heavy fine-tuning hurts. InICML, pages 45014– 45039, 2024. 1, 2, 3, 5, 6

  34. [34]

    Lift+: Lightweight fine-tuning for long-tail learning.arXiv preprint arXiv:2504.13282, 2025

    Jiang-Xin Shi, Tong Wei, and Yu-Feng Li. Lift+: Lightweight fine-tuning for long-tail learning.arXiv preprint arXiv:2504.13282, 2025. 2

  35. [35]

    Fixmatch: Simplifying semi-supervised learning with consistency and confidence

    Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. InNeurIPS, pages 596–608, 2020. 3

  36. [36]

    Adapters strike back

    Jan-Martin O Steitz and Stefan Roth. Adapters strike back. InCVPR, pages 23449–23459, 2024. 2

  37. [37]

    Rethinking classifier re-training in long-tailed recognition: Label over- smooth can balance

    Siyu Sun, Han Lu, Jiangtong Li, Yichen Xie, Tianjiao Li, Xiaokang Yang, Liqing Zhang, and Junchi Yan. Rethinking classifier re-training in long-tailed recognition: Label over- smooth can balance. InICLR, pages 26681–26697, 2025. 7, 8

  38. [38]

    Vl-ltr: Learning class-wise visual-linguistic repre- sentation for long-tailed visual recognition

    Changyao Tian, Wenhai Wang, Xizhou Zhu, Jifeng Dai, and Yu Qiao. Vl-ltr: Learning class-wise visual-linguistic repre- sentation for long-tailed visual recognition. InECCV, pages 73–91, 2022. 2, 5, 6

  39. [39]

    The inaturalist species classification and de- tection dataset

    Grant Van Horn, Oisin Mac Aodha, Yang Song, Yin Cui, Chen Sun, Alex Shepard, Hartwig Adam, Pietro Perona, and Serge Belongie. The inaturalist species classification and de- tection dataset. InCVPR, pages 8769–8778, 2018. 2, 5

  40. [40]

    Kill two birds with one stone: Rethinking data augmentation for deep long-tailed learning

    Binwu Wang, Pengkun Wang, Wei Xu, Xu Wang, Yudong Zhang, Kun Wang, and Yang Wang. Kill two birds with one stone: Rethinking data augmentation for deep long-tailed learning. InICLR, 2024. 7, 8

  41. [41]

    Debiased learning from naturally imbalanced pseudo-labels

    Xudong Wang, Zhirong Wu, Long Lian, and Stella X Yu. Debiased learning from naturally imbalanced pseudo-labels. InCVPR, pages 14647–14657, 2022. 3

  42. [42]

    Exploring vision-language models for imbalanced learning

    Yidong Wang, Zhuohao Yu, Jindong Wang, Qiang Heng, Hao Chen, Wei Ye, Rui Xie, Xing Xie, and Shikun Zhang. Exploring vision-language models for imbalanced learning. IJCV, 132(1):224–237, 2024. 2, 6

  43. [43]

    Lmpt: Prompt tuning with class-specific embedding loss for long- tailed multi-label visual recognition

    Peng Xia, Di Xu, Ming Hu, Lie Ju, and Zongyuan Ge. Lmpt: Prompt tuning with class-specific embedding loss for long- tailed multi-label visual recognition. InALVR, pages 26–36,

  44. [44]

    Parameter-efficient fine-tuning for pre-trained vision models: A survey.arXiv preprint arXiv:2402.02242, 2024

    Yi Xin, Siqi Luo, Haodi Zhou, Junlong Du, Xiaohong Liu, Yue Fan, Qing Li, and Yuntao Du. Parameter-efficient fine-tuning for pre-trained vision models: A survey.arXiv preprint arXiv:2402.02242, 2024. 3

  45. [45]

    Learning imbalanced data with vision trans- formers

    Zhengzhuo Xu, Ruikang Liu, Shuo Yang, Zenghao Chai, and Chun Yuan. Learning imbalanced data with vision trans- formers. InCVPR, pages 15793–15803, 2023. 5, 6

  46. [46]

    An auto-driving system by interactive driving knowledge acquisition

    Seiji Yasunobu and Ryota Sasaki. An auto-driving system by interactive driving knowledge acquisition. InSICE 2003 Annual Conference (IEEE Cat. No. 03TH8734), pages 2935– 2938, 2003. 2

  47. [47]

    A systematic review on long-tailed learning.IEEE TNNLS, 2025

    Chongsheng Zhang, George Almpanidis, Gaojuan Fan, Bin- quan Deng, Yanbo Zhang, Ji Liu, Aouaidjia Kamel, Paolo Soda, and Jo ˜ao Gama. A systematic review on long-tailed learning.IEEE TNNLS, 2025. 2

  48. [48]

    Otamatch: Optimal transport assign- ment with pseudonce for semi-supervised learning.IEEE TIP, 2024

    Jinjin Zhang, Junjie Liu, Debang Li, Qiuyu Huang, Jiaxin Chen, and Di Huang. Otamatch: Optimal transport assign- ment with pseudonce for semi-supervised learning.IEEE TIP, 2024. 3

  49. [49]

    Distribution alignment: A unified framework for long-tail visual recognition

    Songyang Zhang, Zeming Li, Shipeng Yan, Xuming He, and Jian Sun. Distribution alignment: A unified framework for long-tail visual recognition. InCVPR, pages 2361–2370,

  50. [50]

    Long-tailed diffusion models with oriented calibration

    Tianjiao Zhang, Huangjie Zheng, Jiangchao Yao, Xiangfeng Wang, Mingyuan Zhou, Ya Zhang, and Yanfeng Wang. Long-tailed diffusion models with oriented calibration. In ICLR, 2024. 2

  51. [51]

    Deep long-tailed learning: A survey.IEEE TPAMI, 45(9):10795–10816, 2023

    Yifan Zhang, Bingyi Kang, Bryan Hooi, Shuicheng Yan, and Jiashi Feng. Deep long-tailed learning: A survey.IEEE TPAMI, 45(9):10795–10816, 2023. 2