pith. machine review for the scientific record. sign in

arxiv: 2604.10551 · v1 · submitted 2026-04-12 · 💻 cs.CV

Recognition: unknown

NTIRE 2026 Challenge on Short-form UGC Video Restoration in the Wild with Generative Models: Datasets, Methods and Results

Bingchen Li, Boqi Zhang, Chao Zhou, Chengxi Zeng, Chen Lu, David Bull, Dehao Feng, Fan Zhang, Fengkai Zhang, Guoyi Xu, Hang Song, Hao Chen, Hao Kang, Haoran Bai, Huadong Ma, Huiyuan Fu, Jiachao Gong, Jiachen Tu, Jiahui Liu, Jiajia Liu, Jiangxin Dong, Jiawei Shi, Jing Yang, Jingyi Xu, Jingyu Ma, Jinshan Pan, Jiyuan Zhang, Junyang Chen, Kun Liu, Kun Yuan, Lei Lei, Linfeng Li, Long Sun, Mai Xu, Meisong Zheng, Qi Xu, Radu Timofte, Shang-Quan Sun, Shengxi Li, Shibo Yin, Shinan Chen, Shiyao Xiong, Shuai Liu, Sibin Deng, Suhang Yao, Tianhao Peng, Wei Zhou, Wenqi Ren, Xiaotao Wang, Xiaoxu Chen, Xijun Wang, Xilei Zhu, Xindong Zhang, Xin Li, Xinzhe Zhu, Xiuhao Qiu, Xu Cheng, Yahui Wang, Yanan Xing, Yaokun Shi, Yaoxin Jiang, Yeying Jin, Yibin Huang, Yilian Zhong, Ying Chen, Yixing Yang, Yixin Yang, Yizhen Shao, Yulin Ren, Yushun Fang, Yuxiang Chen, Yuxuan Jiang, Zhaokun Hu, Zhibo Chen, Zhijie Ma, Zhirui Liu, Zhuoya Zou, Zihong Chen

Authors on Pith no claims yet

Pith reviewed 2026-05-10 15:55 UTC · model grok-4.3

classification 💻 cs.CV
keywords short-form video restorationuser-generated contentgenerative modelsNTIRE challengeKwaiVIR benchmarkvideo enhancementsubjective evaluation
0
0 comments X

The pith

The NTIRE 2026 Challenge shows that generative methods restore short-form UGC videos effectively on the new KwaiVIR benchmark.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

This paper overviews the NTIRE 2026 Challenge on restoring short-form user-generated content videos in the wild with generative models. It introduces the KwaiVIR benchmark containing 200 synthetic training videos, 48 wild training videos, plus validation and test sets drawn from real-world sources. The challenge runs two tracks: a primary subjective track scored by user studies and a secondary objective track using standard metrics. Twelve teams submitted final solutions after 95 registrations, and those methods delivered strong results across the benchmark.

Core claim

The submitted methods achieved strong performance on the KwaiVIR benchmark, demonstrating encouraging progress in short-form UGC video restoration in the wild through generative-model approaches evaluated on both synthetic and real degradations.

What carries the argument

The KwaiVIR benchmark dataset paired with a dual-track evaluation that combines user-study subjective scoring and objective metrics to rank generative restoration outputs.

Load-bearing premise

The KwaiVIR mix of synthetic and real-world short videos plus the user-study ranking accurately reflects the complex degradations found in actual short-form UGC content.

What would settle it

New tests on a fresh collection of unseen wild short-form videos where top challenge entries fail to match or exceed prior methods in user preference scores.

Figures

Figures reproduced from arXiv: 2604.10551 by Bingchen Li, Boqi Zhang, Chao Zhou, Chengxi Zeng, Chen Lu, David Bull, Dehao Feng, Fan Zhang, Fengkai Zhang, Guoyi Xu, Hang Song, Hao Chen, Hao Kang, Haoran Bai, Huadong Ma, Huiyuan Fu, Jiachao Gong, Jiachen Tu, Jiahui Liu, Jiajia Liu, Jiangxin Dong, Jiawei Shi, Jing Yang, Jingyi Xu, Jingyu Ma, Jinshan Pan, Jiyuan Zhang, Junyang Chen, Kun Liu, Kun Yuan, Lei Lei, Linfeng Li, Long Sun, Mai Xu, Meisong Zheng, Qi Xu, Radu Timofte, Shang-Quan Sun, Shengxi Li, Shibo Yin, Shinan Chen, Shiyao Xiong, Shuai Liu, Sibin Deng, Suhang Yao, Tianhao Peng, Wei Zhou, Wenqi Ren, Xiaotao Wang, Xiaoxu Chen, Xijun Wang, Xilei Zhu, Xindong Zhang, Xin Li, Xinzhe Zhu, Xiuhao Qiu, Xu Cheng, Yahui Wang, Yanan Xing, Yaokun Shi, Yaoxin Jiang, Yeying Jin, Yibin Huang, Yilian Zhong, Ying Chen, Yixing Yang, Yixin Yang, Yizhen Shao, Yulin Ren, Yushun Fang, Yuxiang Chen, Yuxuan Jiang, Zhaokun Hu, Zhibo Chen, Zhijie Ma, Zhirui Liu, Zhuoya Zou, Zihong Chen.

Figure 1
Figure 1. Figure 1: The overall framework of Team RedMediaTech. [PITH_FULL_IMAGE:figures/full_fig_p003_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: The proposed video restoration framework built upon the [PITH_FULL_IMAGE:figures/full_fig_p006_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: The DiT design of the proposed video restoration frame [PITH_FULL_IMAGE:figures/full_fig_p006_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: The overall framework of Team BVI. combines the high pixel-level fidelity of SeedVR2 with the perceptual quality and temporal consistency offered by FlashVSR. 4.8. IMAG@NJUST This team proposes a diffusion-based video restoration framework centered on CoDiVSR, titled “Rethinking What to Condition and What to Disentangle in Diffusion-based Video Super-Resolution”. For the objective track, they adopt a singl… view at source ↗
read the original abstract

This paper presents an overview of the NTIRE 2026 Challenge on Short-form UGC Video Restoration in the Wild with Generative Models. This challenge utilizes a new short-form UGC (S-UGC) video restoration benchmark, termed KwaiVIR, which is contributed by USTC and Kuaishou Technology. It contains both synthetically distorted videos and real-world short-form UGC videos in the wild. For this edition, the released data include 200 synthetic training videos, 48 wild training videos, 11 validation videos, and 20 testing videos. The primary goal of this challenge is to establish a strong and practical benchmark for restoring short-form UGC videos under complex real-world degradations, especially in the emerging paradigm of generative-model-based S-UGC video restoration. This challenge has two tracks: (i) the primary track is a subjective track, where the evaluation is based on a user study; (ii) the second track is an objective track. These two tracks enable a comprehensive assessment of restoration quality. In total, 95 teams have registered for this competition. And 12 teams submitted valid final solutions and fact sheets for the testing phase. The submitted methods achieved strong performance on the KwaiVIR benchmark, demonstrating encouraging progress in short-form UGC video restoration in the wild.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 3 minor

Summary. The paper presents an overview of the NTIRE 2026 Challenge on Short-form UGC Video Restoration in the Wild with Generative Models. It introduces the KwaiVIR benchmark contributed by USTC and Kuaishou, containing 200 synthetic training videos, 48 wild training videos, 11 validation videos, and 20 test videos. The challenge features two tracks: a primary subjective track evaluated via user study and a secondary objective track. 95 teams registered, with 12 submitting valid final solutions and fact sheets; the paper states that these methods achieved strong performance on the KwaiVIR benchmark, demonstrating encouraging progress in generative-model-based restoration of short-form UGC videos under complex real-world degradations.

Significance. If the results hold, this work is significant for establishing a dedicated benchmark for short-form user-generated content video restoration, an area of growing practical importance for social media and mobile platforms. The inclusion of both synthetic and real-world wild videos, combined with dual subjective-objective evaluation tracks, provides a more comprehensive assessment framework than purely objective metrics alone. The high registration rate and 12 submissions indicate community engagement and can help drive advancements in generative approaches to handling authentic degradations.

major comments (2)
  1. [Abstract] Abstract: The central claim that 'the submitted methods achieved strong performance' and demonstrate 'encouraging progress' is not supported by any quantitative metrics (e.g., user-study mean opinion scores, PSNR/SSIM values, or comparisons against baselines or prior methods). Without these numbers, error bars, or ranking tables, the strength of the performance claim cannot be independently assessed.
  2. [Dataset description / Evaluation] Dataset and evaluation sections: The benchmark relies on only 48 wild training videos and 20 test videos for the in-the-wild track. This limited scale raises questions about whether the reported progress generalizes to the diversity of real-world short-form UGC degradations, directly affecting the reliability of the 'strong performance' conclusion.
minor comments (3)
  1. [Evaluation] The paper should provide explicit details on the user-study protocol (number of participants, rating scale, statistical significance testing) to allow readers to interpret the subjective track results.
  2. [Results] A summary table listing the top teams, their methods, and key scores (both subjective and objective) would improve clarity and allow direct comparison of contributions.
  3. [Throughout] Minor typographical inconsistencies in video counts or track descriptions between the abstract and main text should be reconciled for precision.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments on our NTIRE 2026 challenge overview paper. We address each major comment below and propose targeted revisions where appropriate to strengthen the manuscript.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim that 'the submitted methods achieved strong performance' and demonstrate 'encouraging progress' is not supported by any quantitative metrics (e.g., user-study mean opinion scores, PSNR/SSIM values, or comparisons against baselines or prior methods). Without these numbers, error bars, or ranking tables, the strength of the performance claim cannot be independently assessed.

    Authors: We agree that the abstract, as a high-level summary, presents the performance claim qualitatively without specific numbers. The full manuscript contains dedicated results sections that report quantitative outcomes from both tracks, including mean opinion scores from the user study for the primary subjective track, PSNR/SSIM and other objective metrics for the secondary track, ranking tables of the 12 submitted solutions, and comparisons against standard baselines and prior generative restoration methods. These details support the 'strong performance' description relative to the challenge benchmark. To improve clarity and allow independent assessment from the abstract alone, we will revise it to include key quantitative highlights (e.g., top MOS values and relative improvements) while keeping it concise. revision: yes

  2. Referee: [Dataset description / Evaluation] Dataset and evaluation sections: The benchmark relies on only 48 wild training videos and 20 test videos for the in-the-wild track. This limited scale raises questions about whether the reported progress generalizes to the diversity of real-world short-form UGC degradations, directly affecting the reliability of the 'strong performance' conclusion.

    Authors: We acknowledge the modest scale of the wild video portion (48 training + 20 test videos), which stems from the inherent challenges of sourcing and curating authentic short-form UGC content with diverse, complex real-world degradations while maintaining evaluation feasibility. The videos were deliberately selected by the benchmark contributors (USTC and Kuaishou) to span representative degradation types encountered in social media platforms. The challenge design mitigates scale limitations through its dual-track evaluation: the primary subjective user study provides perceptual assessment beyond objective metrics, and the synthetic track (200 videos) offers complementary controlled data. We will expand the manuscript with an explicit limitations paragraph discussing dataset scale, potential generalization caveats, and plans for future expansions in subsequent NTIRE editions. This does not alter the core benchmark but strengthens transparency around the 'strong performance' interpretation. revision: partial

Circularity Check

0 steps flagged

No significant circularity: descriptive competition overview

full rationale

The paper is a standard NTIRE challenge report that describes the KwaiVIR dataset construction, two evaluation tracks, participation numbers (95 registered, 12 valid submissions), and empirical outcomes of submitted methods. It contains no equations, derivations, fitted parameters presented as predictions, or load-bearing self-citations that reduce any claim to the authors' own inputs by construction. The statement that methods 'achieved strong performance' is a factual summary of competition results rather than a derived quantity. This matches the default expectation for non-derivational papers and warrants score 0.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

This is an empirical challenge overview with no mathematical modeling, so the ledger is empty.

pith-pipeline@v0.9.0 · 5854 in / 1073 out tokens · 32906 ms · 2026-05-10T15:55:43.017941+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

83 extracted references · 17 canonical work pages · 6 internal anchors

  1. [1]

    NT-HAZE: A Benchmark Dataset for Re- alistic Night-time Image Dehazing

    Radu Ancuti, Codruta Ancuti, Radu Timofte, and Cos- min Ancuti. NT-HAZE: A Benchmark Dataset for Re- alistic Night-time Image Dehazing . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  2. [2]

    NTIRE 2026 Nighttime Image Dehazing Challenge Report

    Radu Ancuti, Alexandru Brateanu, Florin Vasluianu, Raul Balmez, Ciprian Orhei, Codruta Ancuti, Radu Timofte, Cos- min Ancuti, et al. NTIRE 2026 Nighttime Image Dehazing Challenge Report . InProceedings of the IEEE/CVF Confer- ence on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  3. [3]

    Vivid-vr: Distilling concepts from text-to-video diffusion transformer for photorealistic video restoration

    Haoran Bai, Xiaoxu Chen, Canqian Yang, Zongyao He, Sibin Deng, and Ying Chen. Vivid-vr: Distilling concepts from text-to-video diffusion transformer for photorealistic video restoration. InInternational Conference on Learning Representations (ICLR), 2026. 7

  4. [4]

    Qwen2.5-vl technical report, 2025

    Shuai Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, Sibo Song, Kai Dang, Peng Wang, Shijie Wang, Jun Tang, Humen Zhong, Yuanzhi Zhu, Mingkun Yang, Zhao- hai Li, Jianqiang Wan, Pengfei Wang, Wei Ding, Zheren Fu, Yiheng Xu, Jiabo Ye, Xi Zhang, Tianbao Xie, Zesen Cheng, Hang Zhang, Zhibo Yang, Haiyang Xu, and Junyang Lin. Qwen2.5-vl technical report, 2025. 5

  5. [5]

    NTIRE 2026 Challenge on Single Image Re- flection Removal in the Wild: Datasets, Results, and Meth- ods

    Jie Cai, Kangning Yang, Zhiyuan Li, Florin Vasluianu, Radu Timofte, et al. NTIRE 2026 Challenge on Single Image Re- flection Removal in the Wild: Datasets, Results, and Meth- ods . InProceedings of the IEEE/CVF Conference on Com- puter Vision and Pattern Recognition (CVPR) Workshops,

  6. [6]

    Investigating tradeoffs in real-world video super-resolution

    Kelvin CK Chan, Shangchen Zhou, Xiangyu Xu, and Chen Change Loy. Investigating tradeoffs in real-world video super-resolution. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5962–5971, 2022. 2

  7. [7]

    Stcdit: Spatio-temporally consistent diffusion transformer for high-quality video super-resolution.arXiv preprint arXiv:2511.18786, 2025

    Junyang Chen, Jiangxin Dong, Long Sun, Yixin Yang, and Jinshan Pan. Stcdit: Spatio-temporally consistent diffusion transformer for high-quality video super-resolution.arXiv preprint arXiv:2511.18786, 2025. 5

  8. [8]

    Simple baselines for image restoration

    Liangyu Chen, Xiaojie Chu, Xiangyu Zhang, and Jian Sun. Simple baselines for image restoration. InEuropean confer- ence on computer vision, pages 17–33. Springer, 2022. 9

  9. [9]

    arXiv preprint arXiv:2505.16239 (2025)

    Zheng Chen, Zichen Zou, Kewei Zhang, Xiongfei Su, Xin Yuan, Yong Guo, and Yulun Zhang. Dove: Efficient one- step diffusion model for real-world video super-resolution. arXiv preprint arXiv:2505.16239, 2025. 2, 4, 6, 7

  10. [10]

    The Fourth Challenge on Image Super-Resolution (×4) at NTIRE 2026: Benchmark Results and Method Overview

    Zheng Chen, Kai Liu, Jingkai Wang, Xianglong Yan, Jianze Li, Ziqing Zhang, Jue Gong, Jiatong Li, Lei Sun, Xi- aoyang Liu, Radu Timofte, Yulun Zhang, et al. The Fourth Challenge on Image Super-Resolution (×4) at NTIRE 2026: Benchmark Results and Method Overview . InProceedings of the IEEE/CVF Conference on Computer Vision and Pat- tern Recognition (CVPR) W...

  11. [11]

    Low Light Image Enhancement Challenge at NTIRE 2026

    George Ciubotariu, Sharif S M A, Abdur Rehman, Fayaz Ali, Rizwan Ali Naqvi, Marcos Conde, Radu Timofte, et al. Low Light Image Enhancement Challenge at NTIRE 2026 . In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition (CVPR) Workshops, 2026. 2

  12. [12]

    High FPS Video Frame Interpolation Challenge at NTIRE 2026

    George Ciubotariu, Zhuyun Zhou, Yeying Jin, Zongwei Wu, Radu Timofte, et al. High FPS Video Frame Interpolation Challenge at NTIRE 2026 . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  13. [13]

    NTIRE 2026 Rip Current Detection and Segmentation (RipDetSeg) Chal- lenge Report

    Andrei Dumitriu, Aakash Ralhan, Florin Miron, Florin Ta- tui, Radu Tudor Ionescu, Radu Timofte, et al. NTIRE 2026 Rip Current Detection and Segmentation (RipDetSeg) Chal- lenge Report . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  14. [14]

    Conde, Zongwei Wu, Yeying Jin, Radu Timofte, et al

    Omar Elezabi, Marcos V . Conde, Zongwei Wu, Yeying Jin, Radu Timofte, et al. Photography Retouching Trans- fer, NTIRE 2026 Challenge: Report . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  15. [15]

    One-step diffusion transformer for control- lable real-world image super-resolution.arXiv preprint arXiv:2511.17138, 2025

    Yushun Fang, Yuxiang Chen, Shibo Yin, Qiang Hu, Jiangchao Yao, Ya Zhang, Xiaoyun Zhang, and Yan- feng Wang. One-step diffusion transformer for control- lable real-world image super-resolution.arXiv preprint arXiv:2511.17138, 2025. 5

  16. [16]

    NTIRE 2026 Challenge on End-to-End Financial Receipt Restoration and Reasoning from Degraded Images: Datasets, Methods and Results

    Bochen Guan, Jinlong Li, Kangning Yang, Chuang Ke, Jie Cai, Florin Vasluianu, Radu Timofte, et al. NTIRE 2026 Challenge on End-to-End Financial Receipt Restoration and Reasoning from Degraded Images: Datasets, Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  17. [17]

    Qmamba: On first exploration of vision mamba for image quality assessment.arXiv preprint arXiv:2406.09546, 2024

    Fengbin Guan, Xin Li, Zihao Yu, Yiting Lu, and Zhibo Chen. Qmamba: On first exploration of vision mamba for image quality assessment.arXiv preprint arXiv:2406.09546, 2024. 2

  18. [18]

    NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: AI Flash Portrait (Track 3)

    Ya-nan Guan, Shaonan Zhang, Hang Guo, Yawen Wang, Xinying Fan, Jie Liang, Hui Zeng, Guanyi Qin, Lishen Qu, Tao Dai, Shu-Tao Xia, Lei Zhang, Radu Timofte, et al. NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: AI Flash Portrait (Track 3) . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  19. [19]

    NTIRE 2026 Challenge on Robust AI-Generated Image Detection in the Wild

    Aleksandr Gushchin, Khaled Abud, Ekaterina Shumitskaya, Artem Filippov, Georgii Bychkov, Sergey Lavrushkin, Mikhail Erofeev, Anastasia Antsiferova, Changsheng Chen, Shunquan Tan, Radu Timofte, Dmitriy Vatolin, et al. NTIRE 2026 Challenge on Robust AI-Generated Image Detection in the Wild . InProceedings of the IEEE/CVF Conference on Computer Vision and Pa...

  20. [20]

    Masked autoencoders are scalable vision learners

    Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Doll´ar, and Ross Girshick. Masked autoencoders are scalable vision learners. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000– 16009, 2022. 9

  21. [21]

    Robust Deepfake De- tection, NTIRE 2026 Challenge: Report

    Benedikt Hopf, Radu Timofte, et al. Robust Deepfake De- tection, NTIRE 2026 Challenge: Report . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  22. [22]

    LoRA: Low-Rank Adaptation of Large Language Models

    Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen- Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. Lora: Low-rank adaptation of large language models.arXiv preprint arXiv:2106.09685, 2021. 8

  23. [23]

    Lora: Low-rank adaptation of large language models.ICLR, 1(2):3, 2022

    Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen- Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. Lora: Low-rank adaptation of large language models.ICLR, 1(2):3, 2022. 7

  24. [24]

    NTIRE 2026 Low-light Enhancement: Twilight Cowboy Challenge

    Aleksei Khalin, Egor Ershov, Artem Panshin, Sergey Ko- rchagin, Georgiy Lobarev, Arseniy Terekhin, Sofiia Doro- gova, Amir Shamsutdinov, Yasin Mamedov, Bakhtiyar Khalfin, Bogdan Sheludko, Emil Zilyaev, Nikola Bani ´c, Georgy Perevozchikov, Radu Timofte, et al. NTIRE 2026 Low-light Enhancement: Twilight Cowboy Challenge . In Proceedings of the IEEE/CVF Con...

  25. [25]

    Rope: Robust position estimation in wireless sensor net- works

    Loukas Lazos, Radha Poovendran, and Srdjan Capkun. Rope: Robust position estimation in wireless sensor net- works. InIPSN 2005. Fourth International Symposium on Information Processing in Sensor Networks, 2005., pages 324–331. IEEE, 2005. 3

  26. [26]

    Hybrid agents for image restoration.arXiv preprint arXiv:2503.10120, 2025

    Bingchen Li, Xin Li, Yiting Lu, and Zhibo Chen. Hybrid agents for image restoration.arXiv preprint arXiv:2503.10120, 2025. 2

  27. [27]

    Test-time preference optimization for image restoration

    Bingchen Li, Xin Li, Jiaqi Xu, Jiaming Guo, Wenbo Li, Ren- jing Pei, and Zhibo Chen. Test-time preference optimization for image restoration. InProceedings of the AAAI Confer- ence on Artificial Intelligence, pages 5973–5981, 2026. 2

  28. [28]

    The First Challenge on Mobile Real-World Image Super- Resolution at NTIRE 2026: Benchmark Results and Method Overview

    Jiatong Li, Zheng Chen, Kai Liu, Jingkai Wang, Zihan Zhou, Xiaoyang Liu, Libo Zhu, Radu Timofte, Yulun Zhang, et al. The First Challenge on Mobile Real-World Image Super- Resolution at NTIRE 2026: Benchmark Results and Method Overview . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  29. [29]

    Ntire 2024 challenge on short-form ugc video qual- ity assessment: Methods and results

    Xin Li, Kun Yuan, Yajing Pei, Yiting Lu, Ming Sun, Chao Zhou, Zhibo Chen, Radu Timofte, Wei Sun, Haoning Wu, et al. Ntire 2024 challenge on short-form ugc video qual- ity assessment: Methods and results. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6415–6431, 2024. 1

  30. [30]

    Diffusion models for image restoration and enhancement: A compre- hensive survey.International Journal of Computer Vision, 133(11):8078–8108, 2025

    Xin Li, Yulin Ren, Xin Jin, Cuiling Lan, Xingrui Wang, Wenjun Zeng, Xinchao Wang, and Zhibo Chen. Diffusion models for image restoration and enhancement: A compre- hensive survey.International Journal of Computer Vision, 133(11):8078–8108, 2025. 2

  31. [31]

    NTIRE 2025 challenge on short-form ugc video quality assessment and enhancement: Kwaisr dataset and study

    Xin Li, Xijun Wang, Bingchen Li, Kun Yuan, Yizhen Shao, Suhang Yao, Ming Sun, Chao Zhou, Radu Timofte, and Zhibo Chen. NTIRE 2025 challenge on short-form ugc video quality assessment and enhancement: Kwaisr dataset and study. InProceedings of the IEEE/CVF Conference on Com- puter Vision and Pattern Recognition (CVPR) Workshops,

  32. [32]

    NTIRE 2025 challenge on short-form ugc video quality assessment and enhancement: Methods and results

    Xin Li, Kun Yuan, Bingchen Li, Fengbin Guan, Yizhen Shao, Zihao Yu, Xijun Wang, Yiting Lu, Wei Luo, Suhang Yao, Ming Sun, Chao Zhou, Zhibo Chen, Radu Timofte, et al. NTIRE 2025 challenge on short-form ugc video quality assessment and enhancement: Methods and results. InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CV...

  33. [33]

    NTIRE 2026 The Second Challenge on Day and Night Raindrop Removal for Dual-Focused Images: Methods and Results

    Xin Li, Yeying Jin, Suhang Yao, Beibei Lin, Zhaoxin Fan, Wending Yan, Xin Jin, Zongwei Wu, Bingchen Li, Peishu Shi, Yufei Yang, Yu Li, Zhibo Chen, Bihan Wen, Robby Tan, Radu Timofte, et al. NTIRE 2026 The Second Challenge on Day and Night Raindrop Removal for Dual-Focused Images: Methods and Results . InProceedings of the IEEE/CVF Con- ference on Computer...

  34. [34]

    Lsdir: A large scale dataset for image restoration

    Yawei Li, Kai Zhang, Jingyun Liang, Jiezhang Cao, Chao Liu, Rui Gong, Yulun Zhang, Hao Tang, Yun Liu, Didier Demandolx, and Radu Timofte. Lsdir: A large scale dataset for image restoration. InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition Work- shops, pages 1775–1787, 2023. 8

  35. [35]

    The First Chal- lenge on Remote Sensing Infrared Image Super-Resolution at NTIRE 2026: Benchmark Results and Method Overview

    Kai Liu, Haoyang Yue, Zeli Lin, Zheng Chen, Jingkai Wang, Jue Gong, Radu Timofte, Yulun Zhang, et al. The First Chal- lenge on Remote Sensing Infrared Image Super-Resolution at NTIRE 2026: Benchmark Results and Method Overview . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  36. [36]

    Conde, et al

    Shuhong Liu, Ziteng Cui, Chenyu Bao, Xuangeng Chu, Lin Gu, Bin Ren, Radu Timofte, Marcos V . Conde, et al. 3D Restoration and Reconstruction in Adverse Conditions: Re- alX3D Challenge Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  37. [37]

    NTIRE 2026 X- AIGC Quality Assessment Challenge: Methods and Results

    Xiaohong Liu, Xiongkuo Min, Guangtao Zhai, Qiang Hu, Jiezhang Cao, Yu Zhou, Wei Sun, Farong Wen, Zitong Xu, Yingjie Zhou, Huiyu Duan, Lu Liu, Jiarui Wang, Siqi Luo, Chunyi Li, Li Xu, Zicheng Zhang, Yue Shi, Yubo Wang, Minghong Zhang, Chunchao Guo, Zhichao Hu, Mingtao Chen, Xiele Wu, Xin Ma, Zhaohe Lv, Yuanhao Xue, Jiaqi Wang, Xinxing Sha, Radu Timofte, et...

  38. [38]

    Styleam: Perception-oriented unsupervised domain adaption for no- reference image quality assessment.IEEE Transactions on Multimedia, 27:2043–2058, 2024

    Yiting Lu, Xin Li, Jianzhao Liu, and Zhibo Chen. Styleam: Perception-oriented unsupervised domain adaption for no- reference image quality assessment.IEEE Transactions on Multimedia, 27:2043–2058, 2024. 2

  39. [39]

    Kvq: Kwai video quality assessment for short-form videos

    Yiting Lu, Xin Li, Yajing Pei, Kun Yuan, Qizhi Xie, Yunpeng Qu, Ming Sun, Chao Zhou, and Zhibo Chen. Kvq: Kwai video quality assessment for short-form videos. InProceed- ings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 25963–25973, 2024. 1

  40. [40]

    Spectral Normalization for Generative Adversarial Networks

    Takeru Miyato, Toshiki Kataoka, Masanori Koyama, and Yuichi Yoshida. Spectral normalization for generative ad- versarial networks.arXiv preprint arXiv:1802.05957, 2018. 9

  41. [41]

    NTIRE 2026 Challenge on Video Saliency Predic- tion: Methods and Results

    Andrey Moskalenko, Alexey Bryncev, Ivan Kosmynin, Kira Shilovskaya, Mikhail Erofeev, Dmitry Vatolin, Radu Timo- fte, et al. NTIRE 2026 Challenge on Video Saliency Predic- tion: Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  42. [42]

    Bvi-aom: A new training dataset for deep video compression optimization

    Jakub Nawała, Yuxuan Jiang, Fan Zhang, Xiaoqing Zhu, Joel Sole, and David Bull. Bvi-aom: A new training dataset for deep video compression optimization. InVCIP, pages 1–5. IEEE, 2024. 8

  43. [43]

    NTIRE 2026 Challenge on Efficient Burst HDR and Restoration: Datasets, Methods, and Results

    Hyunhee Park, Eunpil Park, Sangmin Lee, Radu Timofte, et al. NTIRE 2026 Challenge on Efficient Burst HDR and Restoration: Datasets, Methods, and Results . InProceed- ings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  44. [44]

    NTIRE 2026 Challenge on Learned Smartphone ISP with Unpaired Data: Methods and Results

    Georgy Perevozchikov, Daniil Vladimirov, Radu Timofte, et al. NTIRE 2026 Challenge on Learned Smartphone ISP with Unpaired Data: Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pat- tern Recognition (CVPR) Workshops, 2026. 2

  45. [45]

    NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: Professional Image Quality Assessment (Track 1)

    Guanyi Qin, Jie Liang, Bingbing Zhang, Lishen Qu, Ya-nan Guan, Hui Zeng, Lei Zhang, Radu Timofte, et al. NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: Professional Image Quality Assessment (Track 1) . InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  46. [46]

    The Second Challenge on Cross-Domain Few-Shot Object Detection at NTIRE 2026: Methods and Results

    Xingyu Qiu, Yuqian Fu, Jiawei Geng, Bin Ren, Jiancheng Pan, Zongwei Wu, Hao Tang, Yanwei Fu, Radu Timo- fte, Nicu Sebe, Mohamed Elhoseiny, et al. The Second Challenge on Cross-Domain Few-Shot Object Detection at NTIRE 2026: Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  47. [47]

    NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: Multi-Exposure Image Fusion in Dynamic Scenes (Track2)

    Lishen Qu, Yao Liu, Jie Liang, Hui Zeng, Wen Dai, Ya-nan Guan, Guanyi Qin, Shihao Zhou, Jufeng Yang, Lei Zhang, Radu Timofte, et al. NTIRE 2026 The 3rd Restore Any Image Model (RAIM) Challenge: Multi-Exposure Image Fusion in Dynamic Scenes (Track2) . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  48. [48]

    The Eleventh NTIRE 2026 Efficient Super-Resolution Challenge Report

    Bin Ren, Hang Guo, Yan Shu, Jiaqi Ma, Ziteng Cui, Shuhong Liu, Guofeng Mei, Lei Sun, Zongwei Wu, Fahad Shahbaz Khan, Salman Khan, Radu Timofte, Yawei Li, et al. The Eleventh NTIRE 2026 Efficient Super-Resolution Challenge Report . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  49. [49]

    Moe-diffir: Task-customized diffusion priors for universal compressed image restoration

    Yulin Ren, Xin Li, Bingchen Li, Xingrui Wang, Mengxi Guo, Shijie Zhao, Li Zhang, and Zhibo Chen. Moe-diffir: Task-customized diffusion priors for universal compressed image restoration. InEuropean Conference on Computer Vi- sion, pages 116–134. Springer, 2024. 2

  50. [50]

    Conde, Jeffrey Chen, Zhuyun Zhou, Zongwei Wu, Radu Timofte, et al

    Tim Seizinger, Florin-Alexandru Vasluianu, Marcos V . Conde, Jeffrey Chen, Zhuyun Zhou, Zongwei Wu, Radu Timofte, et al. The First Controllable Bokeh Rendering Challenge at NTIRE 2026 . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  51. [51]

    Very Deep Convolutional Networks for Large-Scale Image Recognition

    Karen Simonyan and Andrew Zisserman. Very deep convo- lutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556, 2014. 9

  52. [52]

    Score-Based Generative Modeling through Stochastic Differential Equations

    Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Ab- hishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equa- tions.arXiv preprint arXiv:2011.13456, 2020. 9

  53. [53]

    The Third Challenge on Image Denoising at NTIRE 2026: Methods and Results

    Lei Sun, Hang Guo, Bin Ren, Shaolin Su, Xian Wang, Danda Pani Paudel, Luc Van Gool, Radu Timofte, Yawei Li, et al. The Third Challenge on Image Denoising at NTIRE 2026: Methods and Results . InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  54. [54]

    The Second Challenge on Event-Based Image Deblurring at NTIRE 2026: Methods and Results

    Lei Sun, Weilun Li, Xian Wang, Zhendong Li, Letian Shi, Dannong Xu, Deheng Zhang, Mengshun Hu, Shuang Guo, Shaolin Su, Radu Timofte, Danda Pani Paudel, Luc Van Gool, et al. The Second Challenge on Event-Based Image Deblurring at NTIRE 2026: Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Wor...

  55. [55]

    NTIRE 2026 The First Challenge on Blind Computational Aberration Correction: Methods and Results

    Lei Sun, Xiaolong Qian, Qi Jiang, Xian Wang, Yao Gao, Kailun Yang, Kaiwei Wang, Radu Timofte, Danda Pani Paudel, Luc Van Gool, et al. NTIRE 2026 The First Challenge on Blind Computational Aberration Correction: Methods and Results . InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  56. [56]

    Learning- Based Ambient Lighting Normalization: NTIRE 2026 Chal- lenge Results and Findings

    Florin-Alexandru Vasluianu, Tim Seizinger, Jeffrey Chen, Zhuyun Zhou, Zongwei Wu, Radu Timofte, et al. Learning- Based Ambient Lighting Normalization: NTIRE 2026 Chal- lenge Results and Findings . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  57. [57]

    Advances in Single- Image Shadow Removal: Results from the NTIRE 2026 Challenge

    Florin-Alexandru Vasluianu, Tim Seizinger, Zhuyun Zhou, Zongwei Wu, Radu Timofte, et al. Advances in Single- Image Shadow Removal: Results from the NTIRE 2026 Challenge . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  58. [58]

    Wan: Open and Advanced Large-Scale Video Generative Models

    Team Wan, Ang Wang, Baole Ai, Bin Wen, Chaojie Mao, Chen-Wei Xie, Di Chen, Feiwu Yu, Haiming Zhao, Jianxiao Yang, et al. Wan: Open and advanced large-scale video gen- erative models.arXiv preprint arXiv:2503.20314, 2025. 3, 5, 10

  59. [59]

    Exploiting diffusion prior for real-world image super-resolution.International Journal of Computer Vision, 132(12):5929–5949, 2024

    Jianyi Wang, Zongsheng Yue, Shangchen Zhou, Kelvin CK Chan, and Chen Change Loy. Exploiting diffusion prior for real-world image super-resolution.International Journal of Computer Vision, 132(12):5929–5949, 2024. 2

  60. [60]

    Seedvr2: One-step video restoration via diffusion adversarial post-training.arXiv preprint arXiv:2506.05301, 2025

    Jianyi Wang, Shanchuan Lin, Zhijie Lin, Yuxi Ren, Meng Wei, Zongsheng Yue, Shangchen Zhou, Hao Chen, Yang Zhao, Ceyuan Yang, et al. Seedvr2: One-step video restora- tion via diffusion adversarial post-training.arXiv preprint arXiv:2506.05301, 2025. 7

  61. [61]

    The Second Challenge on Real-World Face Restoration at NTIRE 2026: Methods and Results

    Jingkai Wang, Jue Gong, Zheng Chen, Kai Liu, Jiatong Li, Yulun Zhang, Radu Timofte, et al. The Second Challenge on Real-World Face Restoration at NTIRE 2026: Methods and Results . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Work- shops, 2026. 2

  62. [62]

    NTIRE 2026 Challenge on 3D Content Super-Resolution: Methods and Results

    Longguang Wang, Yulan Guo, Yingqian Wang, Juncheng Li, Sida Peng, Ye Zhang, Radu Timofte, Minglin Chen, Yi Wang, Qibin Hu, Wenjie Lei, et al. NTIRE 2026 Challenge on 3D Content Super-Resolution: Methods and Results . In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition (CVPR) Workshops, 2026. 2

  63. [63]

    Esrgan: En- hanced super-resolution generative adversarial networks

    Xintao Wang, Ke Yu, Shixiang Wu, Jinjin Gu, Yihao Liu, Chao Dong, Yu Qiao, and Chen Change Loy. Esrgan: En- hanced super-resolution generative adversarial networks. In Proc. Eur. Conf. Comput. Vis. workshops, pages 0–0, 2018. 4

  64. [64]

    Real-esrgan: Training real-world blind super-resolution with pure synthetic data

    Xintao Wang, Liangbin Xie, Chao Dong, and Ying Shan. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. InProceedings of the IEEE/CVF Inter- national Conference on Computer Vision, pages 1905–1914,

  65. [65]

    Liftvsr: Lifting image diffusion to video super-resolution via hybrid temporal modeling with only 4×rtx 4090s.arXiv preprint arXiv:2506.08529, 2025

    Xijun Wang, Xin Li, Bingchen Li, and Zhibo Chen. Liftvsr: Lifting image diffusion to video super-resolution via hybrid temporal modeling with only 4×rtx 4090s.arXiv preprint arXiv:2506.08529, 2025. 2

  66. [66]

    NTIRE 2026 Challenge on Light Field Image Super-Resolution: Methods and Results

    Yingqian Wang, Zhengyu Liang, Fengyuan Zhang, Wending Zhao, Longguang Wang, Juncheng Li, Jungang Yang, Radu Timofte, Yulan Guo, et al. NTIRE 2026 Challenge on Light Field Image Super-Resolution: Methods and Results . In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition (CVPR) Workshops, 2026. 2

  67. [67]

    Qwen-Image Technical Report

    Chenfei Wu, Jiahao Li, Jingren Zhou, Junyang Lin, Kaiyuan Gao, Kun Yan, Sheng-ming Yin, Shuai Bai, Xiao Xu, Yilei Chen, et al. Qwen-image technical report.arXiv preprint arXiv:2508.02324, 2025. 3

  68. [68]

    Exploring video quality assessment on user gener- ated contents from aesthetic and technical perspectives

    Haoning Wu, Erli Zhang, Liang Liao, Chaofeng Chen, Jing- wen Hou, Annan Wang, Wenxiu Sun, Qiong Yan, and Weisi Lin. Exploring video quality assessment on user gener- ated contents from aesthetic and technical perspectives. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 20144–20154, 2023. 5

  69. [69]

    Rethinking diffusion model-based video super-resolution: Leveraging dense guidance from aligned features.arXiv preprint arXiv:2511.16928, 2025

    Jingyi Xu, Meisong Zheng, Ying Chen, Minglang Qiao, Xin Deng, and Mai Xu. Rethinking diffusion model-based video super-resolution: Leveraging dense guidance from aligned features.arXiv preprint arXiv:2511.16928, 2025. 7

  70. [70]

    Codabench: Flexible, easy-to-use, and reproducible meta-benchmark platform.Patterns, 3(7), 2022

    Zhen Xu, Sergio Escalera, Adrien Pav ˜ao, Magali Richard, Wei-Wei Tu, Quanming Yao, Huan Zhao, and Isabelle Guyon. Codabench: Flexible, easy-to-use, and reproducible meta-benchmark platform.Patterns, 3(7), 2022. 3

  71. [71]

    Efficient Low Light Image Enhancement: NTIRE 2026 Challenge Report

    Jiebin Yan, Chenyu Tu, Qinghua Lin, Zongwei WU, Weixia Zhang, Zhihua Wang, Peibei Cao, Yuming Fang, Xiaoning Liu, Zhuyun Zhou, Radu Timofte, et al. Efficient Low Light Image Enhancement: NTIRE 2026 Challenge Report . In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition (CVPR) Workshops, 2026. 2

  72. [72]

    CogVideoX: Text-to-Video Diffusion Models with An Expert Transformer

    Zhuoyi Yang, Jiayan Teng, Wendi Zheng, Ming Ding, Shiyu Huang, Jiazheng Xu, Yuanming Yang, Wenyi Hong, Xiao- han Zhang, Guanyu Feng, et al. Cogvideox: Text-to-video diffusion models with an expert transformer.arXiv preprint arXiv:2408.06072, 2024. 2, 5, 6

  73. [73]

    Sf-iqa: Quality and similarity integration for ai generated im- age quality assessment

    Zihao Yu, Fengbin Guan, Yiting Lu, Xin Li, and Zhibo Chen. Sf-iqa: Quality and similarity integration for ai generated im- age quality assessment. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6692–6701, 2024. 2

  74. [74]

    NTIRE 2026 Challenge on High- Resolution Depth of non-Lambertian Surfaces

    Pierluigi Zama Ramirez, Fabio Tosi, Luigi Di Stefano, Radu Timofte, Alex Costanzino, Matteo Poggi, Samuele Salti, Ste- fano Mattoccia, et al. NTIRE 2026 Challenge on High- Resolution Depth of non-Lambertian Surfaces . InProceed- ings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

  75. [75]

    Efros, Eli Shecht- man, and Oliver Wang

    Richard Zhang, Phillip Isola, Alexei A. Efros, Eli Shecht- man, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 586–595, 2018. 2, 3, 8

  76. [76]

    Time- aware one step diffusion network for real-world image super- resolution.arXiv preprint arXiv:2508.16557, 2025

    Tainyi Zhang, Zheng-Peng Duan, Peng-Tao Jiang, Bo Li, Ming-Ming Cheng, Chun-Le Guo, and Chongyi Li. Time- aware one step diffusion network for real-world image super- resolution.arXiv preprint arXiv:2508.16557, 2025. 8

  77. [77]

    Realviformer: Investigating attention for real-world video super-resolution

    Yuehan Zhang and Angela Yao. Realviformer: Investigating attention for real-world video super-resolution. InEuropean conference on computer vision, pages 412–428. Springer,

  78. [78]

    Md-vqa: Multi-dimensional quality assessment for ugc live videos

    Zicheng Zhang, Wei Wu, Wei Sun, Danyang Tu, Wei Lu, Xiongkuo Min, Ying Chen, and Guangtao Zhai. Md-vqa: Multi-dimensional quality assessment for ugc live videos. In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition, pages 1746–1755, 2023. 5

  79. [79]

    The Trade-off between Texture Preservation and Denoising: A Robust Dual-Branch Diffusion Pipeline for Challenging Short-Form Videos

    Meisong Zheng, Xiaoxu Chen, Jing Yang, Zhaokun Hu, Ji- ahui Liu, Ying Chen, Haoran Bai, Sibin Deng, Shengxi Li, and Mai Xu. The Trade-off between Texture Preservation and Denoising: A Robust Dual-Branch Diffusion Pipeline for Challenging Short-Form Videos . InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshop...

  80. [80]

    NTIRE 2026 Challenge Report on Anomaly Detection of Face Enhancement for UGC Images

    Yan Zhong, Qiufang Ma, Zhen Wang, Tingting Jiang, Radu Timofte, et al. NTIRE 2026 Challenge Report on Anomaly Detection of Face Enhancement for UGC Images . InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2026. 2

Showing first 80 references.