pith. machine review for the scientific record. sign in

arxiv: 2605.02860 · v1 · submitted 2026-05-04 · 💻 cs.AI · cs.LG· cs.SE

Recognition: 2 theorem links

· Lean Theorem

Standing on the Shoulders of Giants: Stabilized Knowledge Distillation for Cross--Language Code Clone Detection

Authors on Pith no claims yet

Pith reviewed 2026-05-08 17:53 UTC · model grok-4.3

classification 💻 cs.AI cs.LGcs.SE
keywords cross-language code clone detectionknowledge distillationresponse stabilizationcompact language modelsreasoning transferbinary classificationmodel fine-tuningsemantic code similarity
0
0 comments X

The pith

Stabilized knowledge distillation transfers reasoning from large models to make compact models reliable for detecting code clones across languages.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper aims to show that distilling reasoning abilities into compact models, together with techniques to stabilize their outputs, overcomes the problems small models have when following complex prompts for code analysis. This matters because large models bring high costs, privacy concerns, and inconsistent formatting while compact models are cheaper and private but often produce unusable or erratic results. The approach builds synthetic training examples that emphasize reasoning steps over code pairs from different languages and adds stabilization such as forcing a final conclusion or attaching a classification layer. Experiments across several language pairs demonstrate that the resulting models become more consistent and sometimes more accurate, especially when the test code comes from a different distribution than the training data.

Core claim

The paper claims that a distillation process which transfers reasoning-oriented capabilities from a large model into compact student models via synthetic cross-language code pairs, followed by response stabilization through forced conclusion prompting and the addition of binary or contrastive classification heads, produces outputs that can be reliably mapped to binary clone labels, improves predictive performance in many cases, and reduces inference time relative to generation-only inference.

What carries the argument

Response stabilization methods consisting of forced conclusion prompting, a binary classification head, and a contrastive classification head, which convert the model's output into consistent binary decisions while preserving the transferred reasoning.

Load-bearing premise

The synthetic reasoning examples generated by the large model supply a generalizable clone-detection logic that transfers cleanly to compact models and to real-world code pairs.

What would settle it

If stabilized compact models show no gain in response consistency or detection accuracy on a fresh collection of cross-language code pairs drawn from actual developer projects, the claimed benefit would not hold.

Figures

Figures reproduced from arXiv: 2605.02860 by Fatemeh H. Fard, Mohamad Khajezade, Mohamed Sami Shehata.

Figure 1
Figure 1. Figure 1: Knowledge distillation and fine-tuning methodology: A seed dataset is generated from the Project CodeNet dataset, where each sample consists of code1, code2, and a label indicating whether the pair is a clone (1) or non-clone (0). A template generator embeds each sample into a user prompt used to query the DeepSeek R1 teacher model. The teacher output includes reasoning and conclusion components, which are… view at source ↗
Figure 2
Figure 2. Figure 2: This figure shows the actual user prompt used to encourage DeepSeek-R1 (the teacher model) to view at source ↗
Figure 3
Figure 3. Figure 3: This figure represents the prompt that we used in the second stage of our forced-conclusion. the view at source ↗
Figure 4
Figure 4. Figure 4: F1 score comparison across language pairs for Phi3 and Qwen with and without knowledge distillation view at source ↗
Figure 5
Figure 5. Figure 5: Line chart comparing F1 scores without forced conclusion (Base model) and with forced conclusion view at source ↗
Figure 6
Figure 6. Figure 6: Line chart comparing F1 scores for base models and the binary classification head under Baseline and view at source ↗
Figure 7
Figure 7. Figure 7: Line chart comparing F1 scores for base models and the contrastive classification head under Baseline view at source ↗
Figure 8
Figure 8. Figure 8: Comparison of different methods (binary head, contrastive head, and forced conclusion) across SD view at source ↗
read the original abstract

Cross-language code clone detection (X-CCD) is challenging because semantically equivalent programs written in different languages often share little surface similarity. Although large language models (LLMs) have shown promise for semantic clone detection, their use as black-box systems raises concerns about cost, reproducibility, privacy, and unreliable output formatting. In particular, compact open-source models often struggle to follow reasoning-oriented prompts and to produce outputs that can be consistently mapped to binary clone labels. To address these limitations, we propose a knowledge distillation framework that transfers reasoning capabilities from DeepSeek-R1 into compact open-source student models for X-CCD. Using cross-language code pairs derived from Project CodeNet, we construct reasoning-oriented synthetic training data and fine-tune Phi3 and Qwen-Coder with LoRA adapters. We further introduce response stabilization methods, including forced conclusion prompting, a binary classification head, and a contrastive classification head, and evaluate model behavior using both predictive metrics and response rate. Experiments on Python--Java, Rust--Java, Rust--Python, and Rust--Ruby show that knowledge distillation consistently improves the reliability of compact models and often improves predictive performance, especially under distribution shift. In addition, classification-head variants substantially reduce inference time compared to generation-based inference. Overall, our results show that reasoning-oriented distillation combined with response stabilization makes compact open-source models more practical and reliable for X-CCD detection.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes a knowledge distillation framework to transfer reasoning capabilities from DeepSeek-R1 to compact open-source models (Phi-3 and Qwen-Coder) for cross-language code clone detection (X-CCD). Using synthetic reasoning-oriented data constructed from Project CodeNet cross-language pairs, the authors fine-tune student models with LoRA and introduce response stabilization techniques (forced conclusion prompting, binary classification head, contrastive classification head). Experiments on Python-Java, Rust-Java, Rust-Python, and Rust-Ruby pairs report improved reliability, predictive performance (especially under distribution shift), and faster inference for classification-head variants compared to generation-based inference.

Significance. If the quantitative results and controls hold, the work demonstrates a practical path to making smaller open-source models reliable for semantic X-CCD tasks, addressing cost, reproducibility, privacy, and output consistency issues with large black-box LLMs. The emphasis on reasoning-oriented distillation plus stabilization is a targeted contribution that could improve deployability in code analysis pipelines.

major comments (2)
  1. [Abstract and Experiments] Abstract and Experiments section: The claims of 'consistent reliability gains' and 'often improves predictive performance' are stated without any numerical metrics, baseline comparisons (e.g., against non-distilled Phi-3/Qwen-Coder or prior X-CCD methods), statistical significance tests, or details on data splits and distribution-shift construction. This absence prevents assessment of effect sizes and undermines support for the central claim of improved practicality.
  2. [Experiments] Data construction and evaluation (Experiments section): The central assumption that synthetic reasoning data from DeepSeek-R1 encodes generalizable clone-detection logic (rather than teacher-specific biases or prompt artifacts) is load-bearing but untested. No ablations isolate reasoning transfer from format mimicry or CodeNet artifacts, and all reported gains use held-out pairs from the same source; this leaves the transferability claim unsupported.
minor comments (2)
  1. [Abstract] The abstract would benefit from one or two key quantitative results (e.g., accuracy deltas or response-rate improvements) to give readers an immediate sense of the gains.
  2. [Method] Notation for the stabilization methods (forced conclusion prompting, binary/contrastive heads) should be defined more explicitly when first introduced, including how the binary label is extracted from generated text.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback, which highlights opportunities to strengthen the clarity of our quantitative claims and the validation of our distillation approach. We address each major comment below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract and Experiments] Abstract and Experiments section: The claims of 'consistent reliability gains' and 'often improves predictive performance' are stated without any numerical metrics, baseline comparisons (e.g., against non-distilled Phi-3/Qwen-Coder or prior X-CCD methods), statistical significance tests, or details on data splits and distribution-shift construction. This absence prevents assessment of effect sizes and undermines support for the central claim of improved practicality.

    Authors: We agree that the abstract would benefit from explicit numerical support for the central claims. In the revised version we will update the abstract to report key metrics (e.g., reliability rates, accuracy deltas under distribution shift, and inference-time reductions) drawn from the Experiments section. The Experiments section already contains comparisons against the non-distilled Phi-3 and Qwen-Coder baselines, details on the Project CodeNet splits, and the construction of the distribution-shift test sets; however, we will add (i) explicit comparisons against representative prior X-CCD methods and (ii) statistical significance tests (paired t-tests across seeds) to make effect sizes and reliability gains easier to assess. revision: yes

  2. Referee: [Experiments] Data construction and evaluation (Experiments section): The central assumption that synthetic reasoning data from DeepSeek-R1 encodes generalizable clone-detection logic (rather than teacher-specific biases or prompt artifacts) is load-bearing but untested. No ablations isolate reasoning transfer from format mimicry or CodeNet artifacts, and all reported gains use held-out pairs from the same source; this leaves the transferability claim unsupported.

    Authors: We acknowledge that isolating the contribution of reasoning-oriented distillation from format mimicry or dataset artifacts is necessary to substantiate the transferability claim. Our current evaluation uses held-out CodeNet pairs to demonstrate within-distribution generalization, and the response-stabilization techniques (forced conclusion prompting and classification heads) were introduced precisely to reduce format artifacts. To strengthen the evidence, we will add two ablations in the revised Experiments section: (1) a direct comparison of reasoning-oriented distillation versus standard supervised fine-tuning on identical CodeNet pairs (to separate reasoning transfer from format mimicry), and (2) an evaluation on an additional cross-language split constructed to increase distribution shift beyond the original held-out sets. These additions will provide a clearer test of whether the distilled logic generalizes beyond teacher-specific or CodeNet-specific patterns. revision: yes

Circularity Check

0 steps flagged

No circularity: empirical KD setup uses external teacher and held-out evaluation

full rationale

The paper presents a standard empirical knowledge-distillation pipeline: synthetic reasoning data is generated by an external model (DeepSeek-R1) on Project CodeNet pairs, students (Phi-3, Qwen-Coder) are fine-tuned with LoRA, and performance is measured on held-out cross-language splits. No equations, self-definitional loops, fitted parameters renamed as predictions, or load-bearing self-citations appear in the provided text. All claims rest on externally generated data and separate test sets rather than reducing to the training inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The framework rests on standard transfer-learning assumptions plus the unverified premise that teacher-generated reasoning traces are high-quality and unbiased for the target task.

axioms (1)
  • domain assumption Synthetic reasoning traces produced by DeepSeek-R1 on Project CodeNet pairs constitute effective supervision for student models on unseen cross-language code.
    Invoked when constructing training data and claiming generalization.

pith-pipeline@v0.9.0 · 5568 in / 1243 out tokens · 40658 ms · 2026-05-08T17:53:33.948888+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

57 extracted references · 31 canonical work pages · 6 internal anchors

  1. [1]

    Marah Abdin, Jyoti Aneja, Hany Awadalla, Ahmed Awadallah, Ammar Ahmad Awan, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Jianmin Bao, Harkirat Behl, et al. 2024. Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone.arXiv preprint arXiv:2404.14219(2024)

  2. [2]

    Amal Alazba, Hamoud Aljamaan, and Mohammad Alshayeb. 2024. Cort: transformer-based code representations with self-supervision by predicting reserved words for code smell detection.Empirical Software Engineering29, 3 (2024), 59

  3. [3]

    Asra Sulaiman Alshabib, Sajjad Mahmood, and Mohammad Alshayeb. 2025. A systematic literature review on cross-language source code clone detection.Computer Science Review58 (2025), 100786

  4. [4]

    Baxter, Andrew Yahin, Leonardo Moura, Marcelo Sant’Anna, and Lorraine Bier

    Ira D. Baxter, Andrew Yahin, Leonardo Moura, Marcelo Sant’Anna, and Lorraine Bier. 1998. Clone Detection Using Abstract Syntax Trees. InProceedings of the International Conference on Software Maintenance (ICSM). IEEE Computer Society, 368–377. https://doi.org/10.1109/ICSM.1998.738528

  5. [5]

    Cristian Bucilă, Rich Caruana, and Alexandru Niculescu-Mizil. 2006. Model Compression. InProceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM, 535–541. https://doi.org/10. 1145/1150402.1150464

  6. [6]

    Giordano d’Aloisio, Luca Traini, Federica Sarro, and Antinisca Di Marco. 2025. On the compression of language models for code: An empirical study on codebert. In2025 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER). IEEE, 12–23

  7. [8]

    Shihan Dou, Junjie Shan, Haoxiang Jia, Wenhao Deng, Zhiheng Xi, Wei He, Yueming Wu, Tao Gui, Yang Liu, and Xuanjing Huang. 2023. Towards Understanding the Capability of Large Language Models on Code Clone Detection: A Survey.arXiv preprint arXiv:2308.01191(2023). arXiv:2308.01191 [cs.SE] https://arxiv.org/abs/2308.01191

  8. [9]

    Yangkai Du, Tengfei Ma, Lingfei Wu, Xuhong Zhang, and Shouling Ji. 2024. AdaCCD: Adaptive Semantic Contrasts Discovery Based Cross Lingual Adaptation for Code Clone Detection. InProceedings of the AAAI Conference on Artificial ACM Trans. Softw. Eng. Methodol., Vol. 1, No. 1, Article . Publication date: May . 36 Khajezade et al. Intelligence, Vol. 38. 17942–17950

  9. [10]

    Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, and Ming Zhou. 2020. CodeBERT: A Pre-Trained Model for Programming and Natural Languages. In Findings of the Association for Computational Linguistics: EMNLP 2020. Association for Computational Linguistics, 1536–1547. https://doi.org/10.1...

  10. [11]

    Suriya Gunasekar, Yi Zhang, Jyoti Aneja, Caio César Teodoro Mendes, Allie Del Giorno, Sivakanth Gopi, Mojan Javaheripi, Piero Kauffmann, Gustavo de Rosa, Olli Saarikivi, et al. 2023. Textbooks are all you need.arXiv preprint arXiv:2306.11644(2023)

  11. [12]

    Daya Guo, Shuai Lu, Nan Duan, Yanlin Wang, Ming Zhou, and Jian Yin. 2022. UniXcoder: Unified Cross-Modal Pre-training for Code Representation. InProceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL). Association for Computational Linguistics, 7212–7225. https://doi.org/10.18653/v1/2022.acl-long.499

  12. [13]

    Daya Guo, Shuo Ren, Shuai Lu, Zhangyin Feng, Duyu Tang, Shujie Liu, Long Zhou, Nan Duan, Alexey Svyatkovskiy, Shengyu Fu, Michele Tufano, Shao Kun Deng, Colin Clement, Dawn Drain, Neel Sundaresan, Jian Yin, Daxin Jiang, and Ming Zhou. 2021. GraphCodeBERT: Pre-training Code Representations with Data Flow. InProceedings of the 9th International Conference o...

  13. [15]

    Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, et al. 2025. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning.arXiv preprint arXiv:2501.12948(2025)

  14. [16]

    Break-Fix

    Emman Haider, Daniel Perez-Becker, Thomas Portet, Piyush Madan, Amit Garg, Atabak Ashfaq, David Majercak, Wen Wen, Dongwoo Kim, Ziyi Yang, et al. 2024. Phi-3 safety post-training: Aligning language models with a" break-fix" cycle.arXiv preprint arXiv:2407.13833(2024)

  15. [17]

    Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the Knowledge in a Neural Network.arXiv preprint arXiv:1503.02531(2015)

  16. [18]

    Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen

    Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen

  17. [19]

    InInternational Conference on Learning Representations

    LoRA: Low-Rank Adaptation of Large Language Models. InInternational Conference on Learning Representations

  18. [20]

    Baizhou Huang, Shuai Lu, Xiaojun Wan, and Nan Duan. 2024. Enhancing Large Language Models in Coding Through Multi-Perspective Self-Consistency. InProceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics, Bangkok, Thailand, 1429–1450. https: //doi.org/10.18653...

  19. [21]

    Binyuan Hui, Jian Yang, Zeyu Cui, Jiaxi Yang, Dayiheng Liu, Lei Zhang, Tianyu Liu, Jiajun Zhang, Bowen Yu, Keming Lu, et al. 2024. Qwen2.5-Coder Technical Report.arXiv preprint arXiv:2409.12186(2024)

  20. [22]

    Lingxiao Jiang, Ghassan Misherghi, Zhendong Su, and Stephane Glondu. 2007. DECKARD: Scalable and Accurate Tree-Based Detection of Code Clones. InProceedings of the 29th International Conference on Software Engineering (ICSE). IEEE Computer Society, 96–105. https://doi.org/10.1109/ICSE.2007.30

  21. [23]

    Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. 2020. TinyBERT: Distilling BERT for Natural Language Understanding. InFindings of the Association for Computational Linguistics: EMNLP

  22. [24]

    TinyBERT: Distilling BERT for natural language understanding

    Association for Computational Linguistics, 4163–4174. https://doi.org/10.18653/v1/2020.findings-emnlp.372

  23. [25]

    Toshihiro Kamiya, Shinji Kusumoto, and Katsuro Inoue. 2002. CCFinder: A Multilinguistic Token-Based Code Clone Detection System for Large Scale Source Code. InIEEE Transactions on Software Engineering, Vol. 28. IEEE, 654–670. https://doi.org/10.1109/TSE.2002.1019480

  24. [26]

    Mohamad Khajezade, Jie JW Wu, Fatemeh Hendijani Fard, Gema Rodríguez-Pérez, and Mohamed Sami Shehata. 2024. Investigating the Efficacy of Large Language Models for Code Clone Detection. InProceedings of the 32nd IEEE/ACM International Conference on Program Comprehension. 161–165

  25. [27]

    Maggie Lei, Hao Li, Ji Li, Namrata Aundhkar, and Dae-Kyoo Kim. 2022. Deep learning application on code clone detection: A review of current knowledge.Journal of Systems and Software184 (2022), 111141

  26. [28]

    Yonglin Li, Shanzhi Gu, and Mingyang Geng. 2025. Symmetry-Aware Code Generation: Distilling Pseudocode Reasoning for Lightweight Deployment of Large Language Models.Symmetry17, 8 (2025), 1325

  27. [29]

    Jiayi Lin, Hande Dong, Yutao Xie, and Lei Zhang. 2024. Scaling Laws Behind Code Understanding Model.arXiv preprint arXiv:2402.12813(2024). arXiv:2402.12813 [cs.SE] https://arxiv.org/abs/2402.12813

  28. [30]

    Jiayi Lin, Yanlin Wang, Yibiao Yang, Lei Zhang, and Yutao Xie. 2026. Towards Better Code Understanding in Decoder- Only Models with Contrastive Learning.arXiv preprint arXiv:2406.12326(2026). arXiv:2406.12326 [cs.SE] https: //arxiv.org/abs/2406.12326

  29. [31]

    Chenyao Liu, Zeqi Lin, Jian-Guang Lou, Lijie Wen, and Dongmei Zhang. 2021. Can neural clone detection generalize to unseen functionalitiesƒ. In2021 36th IEEE/ACM International Conference on Automated Software Engineering (ASE). ACM Trans. Softw. Eng. Methodol., Vol. 1, No. 1, Article . Publication date: May . Standing on the Shoulders of Giants: Stabilize...

  30. [32]

    Min, Yangruibo Ding, Luca Buratti, Saurabh Pujar, Gail Kaiser, Suman Jana, and Baishakhi Ray

    Marcus J. Min, Yangruibo Ding, Luca Buratti, Saurabh Pujar, Gail Kaiser, Suman Jana, and Baishakhi Ray. 2023. Beyond Accuracy: Evaluating Self-Consistency of Code Large Language Models with IdentityChain. InProceedings of the 12th International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=caW7LdAALh

  31. [33]

    Mawal A Mohammed, Mohammad Alshayeb, and Jameleddine Hassine. 2022. A search-based approach for detecting circular dependency bad smell in goal-oriented models.Software and Systems Modeling21, 5 (2022), 2007–2037

  32. [34]

    Micheline Bénédicte Moumoula, Abdoul Kader Kabore, Jacques Klein, and Tegawendé Bissyande. 2024. Large Language Models for cross-language code clone detection.arXiv preprint arXiv:2408.04430(2024)

  33. [35]

    Micheline Bénédicte Moumoula, Abdoul Kader Kaboré, Jacques Klein, and Tegawendé F Bissyandé. 2025. The Struggles of LLMs in Cross-Lingual Code Clone Detection.Proceedings of the ACM on Software Engineering2, FSE (2025), 1023–1045

  34. [36]

    Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe. 2022. Training Language Models to Follow Instructions with Human Fee...

  35. [37]

    Ruchir Puri, David S Kung, Geert Janssen, Wei Zhang, Giacomo Domeniconi, Vladimir Zolotov, Julian Dolby, Jie Chen, Mihir Choudhury, Lindsey Decker, et al. 2021. Codenet: A large-scale ai for code dataset for learning a diversity of coding tasks.arXiv preprint arXiv:2105.12655(2021)

  36. [38]

    Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. 2019. DistilBERT, a Distilled Version of BERT: Smaller, Faster, Cheaper and Lighter.arXiv preprint arXiv:1910.01108(2019). arXiv:1910.01108 [cs.CL] https://arxiv. org/abs/1910.01108

  37. [39]

    Jieke Shi, Zhou Yang, Bowen Xu, Hong Jin Kang, Hongwei He, David Lo Lu, and David Lo. 2024. Greening Large Language Models of Code. InProceedings of the 46th IEEE/ACM International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP ’24). Association for Computing Machinery, 142–153. https://doi.org/10. 1145/3639475.3640097

  38. [40]

    Jieke Shi, Zhou Yang, Bowen Xu, Hong Jin Kang Liu, Hongwei He, David Lo Lu, and David Lo. 2022. Compressing Pre-trained Models of Code into 3 MB. InProceedings of the 37th IEEE/ACM International Conference on Automated Software Engineering (ASE ’22). Association for Computing Machinery. https://doi.org/10.1145/3551349.3556964

  39. [41]

    Tim Sonnekalb, Bernd Gruner, Clemens-Alexander Brust, and Patrick Mäder. 2022. Generalizability of code clone detection on codebert. InProceedings of the 37th IEEE/ACM International Conference on Automated Software Engineering. 1–3

  40. [42]

    Nikita Sorokin, Tikhonov Anton, Dmitry Abulkhanov, Ivan Sedykh, Irina Piontkovskaya, and Valentin Malykh. 2025. CCT-Code: Cross-Consistency Training for Multilingual Clone Detection and Code Search. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Vol...

  41. [43]

    Chenning Tao, Qi Zhan, Xing Hu, and Xin Xia. 2022. C4: Contrastive cross-language code clone detection. InProceedings of the 30th IEEE/ACM International Conference on Program Comprehension. 413–424

  42. [44]

    Hashimoto

    Rohan Taori, Ishaan Gulrajani, Tianyi Zhang, Yann Dubois, Xuechen Li, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. 2023. Alpaca: A Strong, Replicable Instruction-Following Model. https://crfm.stanford.edu/2023/03/13/ alpaca.html

  43. [45]

    Boshi Wang, Sewon Min, Xiang Deng, Jiaming Shen, You Wu, Luke Zettlemoyer, and Huan Sun. 2023. Towards Understanding Chain-of-Thought Prompting: An Empirical Study of What Matters. InProceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (Eds.). A...

  44. [47]

    Ruiqi Wang, Zezhou Yang, Cuiyun Gao, Xin Xia, and Qing Liao. 2025. An Empirical Study of Knowledge Distillation for Code Understanding Tasks.arXiv preprint arXiv:2508.15423(2025). arXiv:2508.15423 [cs.SE] https://arxiv.org/abs/ 2508.15423

  45. [48]

    Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. InAdvances in Neural Information Processing Systems, Vol. 33. 5776–5788. https://proceedings.neurips.cc/paper/2020/hash/3f5ee243547dee91fbd053c1c4a845aa- Abstract.html

  46. [49]

    Le, Ed H

    Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc V. Le, Ed H. Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. 2022. Self-Consistency Improves Chain of Thought Reasoning in Language Models. InProceedings of the 11th ACM Trans. Softw. Eng. Methodol., Vol. 1, No. 1, Article . Publication date: May . 38 Khajezade et al. International Conference on Learn...

  47. [50]

    Yue Wang, Weishi Wang, Shafiq Joty, and Steven C. H. Hoi. 2021. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. InProceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, 8696–8708. https: //doi.org/10.18653/v1/2021....

  48. [51]

    Finetuned Language Models Are Zero-Shot Learners

    Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. 2021. Finetuned Language Models Are Zero-Shot Learners.arXiv preprint arXiv:2109.01652(2021)

  49. [52]

    Violet Xiang, Charlie Snell, Kanishk Gandhi, Alon Albalak, Anikait Singh, Chase Blagden, Duy Phung, Rafael Rafailov, Nathan Lile, Dakota Mahan, et al . 2025. Towards system 2 reasoning in llms: Learning how to think with meta chain-of-thought.arXiv preprint arXiv:2501.04682(2025)

  50. [53]

    Mohammad A Yahya and Dae-Kyoo Kim. 2022. Cross-language source code clone detection using deep learning with infercode.arXiv preprint arXiv:2205.04913(2022)

  51. [54]

    Cheng Yang, Chufan Shi, Siheng Li, Bo Shui, Yujiu Yang, and Wai Lam. 2025. Llm2: Let large language models harness system 2 reasoning. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers). 168–177

  52. [55]

    Hao Yu, Wing Lam, Long Chen, Ge Li, Tao Xie, and Qianxiang Wang. 2019. Neural detection of semantic code clones via tree-based convolution. In2019 IEEE/ACM 27th International Conference on Program Comprehension (ICPC). IEEE, 70–80

  53. [56]

    Dejiao Zhang, Wasi Uddin Ahmad, Ming Tan, Hantian Ding, Ramesh Nallapati, Dan Roth, Xiaofei Ma, and Bing Xiang

  54. [57]

    InProceedings of the 12th International Conference on Learning Representations (ICLR)

    CodeSage: Code Representation Learning at Scale. InProceedings of the 12th International Conference on Learning Representations (ICLR). https://openreview.net/forum?id=vfzRRjumpX

  55. [58]

    Haibo Zhang and Kouichi Sakurai. 2021. A survey of software clone detection from security perspective.IEEE Access9 (2021), 48157–48173

  56. [59]

    Renrui Zhang, Jiaming Han, Chris Liu, Peng Gao, Aojun Zhou, Xiangfei Hu, Shilin Yan, Pan Lu, Hongsheng Li, and Yu Qiao. 2023. LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention.arXiv preprint arXiv:2303.16199(2023)

  57. [60]

    Maksim Zubkov, Egor Spirin, Egor Bogomolov, and Timofey Bryksin. 2022. Evaluation of contrastive learning with various code representations for code clone detection.arXiv preprint arXiv:2206.08726(2022). ACM Trans. Softw. Eng. Methodol., Vol. 1, No. 1, Article . Publication date: May