pith. machine review for the scientific record. sign in

arxiv: 2604.12663 · v1 · submitted 2026-04-14 · 💻 cs.AI

Recognition: unknown

Human-Centric Topic Modeling with Goal-Prompted Contrastive Learning and Optimal Transport

Authors on Pith no claims yet

Pith reviewed 2026-05-10 14:50 UTC · model grok-4.3

classification 💻 cs.AI
keywords human-centric topic modelinggoal-prompted contrastive learningoptimal transporttopic coherencetopic diversityLLM promptinggoal alignment
0
0 comments X

The pith

Integrating human goals via LLM prompting and optimal transport produces topics more aligned with user intent while improving coherence and diversity.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper aims to establish that topic modeling can directly incorporate explicit human goals instead of relying solely on statistical patterns from the data. It does so by first using large language models to extract goal candidates from documents and then applying optimal transport to guide contrastive learning of the topics. A sympathetic reader would care because current methods often generate redundant or off-target topics that fail to address the user's underlying purpose. Experiments on subreddit datasets demonstrate gains in coherence, diversity, and goal alignment over prior approaches.

Core claim

The central claim is that goal candidates extracted from documents via LLM-based prompting can be integrated into topic discovery through optimal transport in a contrastive learning framework, enabling the GCTM-OT model to generate topics that respect explicit human objectives while achieving strong statistical quality.

What carries the argument

GCTM-OT, the goal-prompted contrastive topic model with optimal transport that aligns topic representations to LLM-derived goal candidates during learning.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same prompting-plus-transport pattern could extend to other unsupervised tasks such as document clustering or summarization where user intent matters.
  • Interactive systems become possible in which users review initial topics and refine goals for subsequent rounds.
  • Testing the approach on non-social-media corpora would reveal whether the benefits hold beyond forum-style data.
  • Optimal transport may prove useful more broadly for bridging LLM semantics with classical statistical models.

Load-bearing premise

LLM-based prompting reliably extracts accurate and useful goal candidates from documents without introducing systematic bias or errors.

What would settle it

Replacing the LLM goal extraction step with random or incorrect goal inputs and observing whether the reported gains in goal alignment, coherence, and diversity disappear on the same datasets.

Figures

Figures reproduced from arXiv: 2604.12663 by Dongxin Wang, Haiping Huang, Jialin Yu, Philip Torr, Rui Wang, Yi Zheng, Yuanzhi Yao, Yuxiang Zhou.

Figure 1
Figure 1. Figure 1: Key differences between Traditional Topic Modeling [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: The framework of GCTM-OT (a) and details of Goal-oriented Text Representation Mechanism (b). [PITH_FULL_IMAGE:figures/full_fig_p003_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Details of Goal Summarization Prompt P𝑠 . Detailed Prompt for Text Augmentation Output: <Augmented Text> <s>[INST] messages: [ {role: system, content: You are a helpful assistant that extract critical information, rephrase text and make sentence smooth.}, {role: user, content: I will give you a document . Please rephrase it as little as possible, ensure the output starts with the prefix 'REPHRASED:' and is… view at source ↗
Figure 4
Figure 4. Figure 4: Details of Text Augmentation Prompt P𝑎. Here, 𝑁𝑔𝑥 denotes the number of goals in document 𝑥, ranging from 3 to 5 in our experiments. If a document is deemed irrelevant to the goal H𝑔, it is annotated as ‘irrelevant’ and excluded from the corpus. The details of the prompt P𝑠 are shown in [PITH_FULL_IMAGE:figures/full_fig_p003_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: The comparison of topic coherence values vs. different topic number settings [20, 30, 40, 50]. [PITH_FULL_IMAGE:figures/full_fig_p007_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Comparision on𝐺𝑇 𝑅 and𝐺𝐶𝑅 metrics. 5.2 Topic Evaluation and Results Analysis In this section, we conduct extensive experiments on the ‘Both￾ering’, ‘TeslaModel3’ and ‘AskAcademia’ datasets, and present the results along with an analysis of topic coherence, diversity and goal-relevance metrics. For each dataset, we run the model with four different topic number configurations [20, 30, 40, 50], and we select… view at source ↗
read the original abstract

Existing topic modeling methods, from LDA to recent neural and LLM-based approaches, which focus mainly on statistical coherence, often produce redundant or off-target topics that miss the user's underlying intent. We introduce Human-centric Topic Modeling, \emph{Human-TM}), a novel task formulation that integrates a human-provided goal directly into the topic modeling process to produce interpretable, diverse and goal-oriented topics. To tackle this challenge, we propose the \textbf{G}oal-prompted \textbf{C}ontrastive \textbf{T}opic \textbf{M}odel with \textbf{O}ptimal \textbf{T}ransport (GCTM-OT), which first uses LLM-based prompting to extract goal candidates from documents, then incorporates these into semantic-aware contrastive learning via optimal transport for topic discovery. Experimental results on three public subreddit datasets show that GCTM-OT outperforms state-of-the-art baselines in topic coherence and diversity while significantly improving alignment with human-provided goals, paving the way for more human-centric topic discovery systems.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 1 minor

Summary. The manuscript introduces Human-Centric Topic Modeling (Human-TM) as a new task formulation that directly incorporates a human-provided goal into topic discovery to produce interpretable, diverse, and goal-oriented topics. It proposes GCTM-OT, which extracts goal candidates from documents via LLM-based prompting and then applies semantic-aware contrastive learning regularized by optimal transport. Experiments on three public subreddit datasets report that GCTM-OT outperforms state-of-the-art baselines on topic coherence, diversity, and alignment with human-provided goals.

Significance. If the experimental claims hold after addressing validation concerns, the work would advance topic modeling by moving beyond purely statistical coherence toward explicit human-intent alignment, a timely direction given LLM prevalence. The technical choice to combine contrastive learning with optimal transport for goal incorporation is a clear strength, offering a principled mechanism for semantic alignment that could be adopted more broadly. The use of public datasets also supports reproducibility.

major comments (1)
  1. [Section 3] Section 3 (goal candidate extraction via LLM prompting): No independent validation of the extracted goal candidates is described (e.g., no human inter-annotator agreement, no semantic similarity to explicit human goals, or ablation on prompt quality). Because the headline experimental result (improved alignment with human-provided goals) depends on these candidates feeding into the contrastive + OT stages, any measured gains risk being circular or LLM-artifactual rather than reflective of true human intent. This is load-bearing for the central claim in the abstract and results.
minor comments (1)
  1. [Abstract] Abstract: The claim of outperformance would be clearer if the specific baselines, coherence/diversity metrics (e.g., NPMI, topic diversity formula), and any statistical significance tests were named rather than left at the level of 'outperforms state-of-the-art baselines'.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We are grateful to the referee for the thoughtful review and the recommendation for major revision. The feedback on validating the goal candidate extraction is well-taken, and we will incorporate the necessary changes to address this concern. Below we provide a point-by-point response.

read point-by-point responses
  1. Referee: [Section 3] Section 3 (goal candidate extraction via LLM prompting): No independent validation of the extracted goal candidates is described (e.g., no human inter-annotator agreement, no semantic similarity to explicit human goals, or ablation on prompt quality). Because the headline experimental result (improved alignment with human-provided goals) depends on these candidates feeding into the contrastive + OT stages, any measured gains risk being circular or LLM-artifactual rather than reflective of true human intent. This is load-bearing for the central claim in the abstract and results.

    Authors: We thank the referee for highlighting this important aspect. We concur that without explicit validation, there is a risk that the observed improvements in goal alignment could stem from LLM-specific patterns rather than robust human intent capture. To address this, we will revise the manuscript to include a dedicated validation subsection in Section 3. Specifically, we will conduct a human evaluation where multiple annotators rate the relevance of extracted goal candidates to the original documents and compute inter-annotator agreement using Cohen's kappa. We will also perform an ablation study on different prompting strategies and report their impact on downstream topic quality. Furthermore, we will measure the semantic similarity (using sentence embeddings) between the LLM-extracted candidates and the human-provided goals used in our alignment evaluation, to quantify how well the extraction approximates true human intent. These revisions will be added to strengthen the empirical support for our central claims. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation relies on standard components without self-referential reduction

full rationale

The paper formulates Human-TM as integrating human-provided goals into topic modeling via LLM prompting for goal candidates, followed by contrastive learning and optimal transport. No equations or steps reduce a claimed output (e.g., topic coherence, diversity, or goal alignment) to a fitted parameter or self-citation by construction. The experimental claims rest on comparisons to baselines on public datasets, with no evidence that alignment metrics are tautologically defined from the LLM extraction step itself. Self-citations, if present, are not load-bearing for the core pipeline. This is a standard methodological proposal with independent empirical evaluation.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

Abstract-only review yields no explicit free parameters, axioms, or invented entities; the method description relies on standard components (LLM prompting, contrastive loss, optimal transport) whose assumptions are not detailed here.

pith-pipeline@v0.9.0 · 5500 in / 1185 out tokens · 42441 ms · 2026-05-10T14:50:40.628379+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

56 extracted references · 4 canonical work pages · 3 internal anchors

  1. [1]

    Aly Abdelrazek, Yomna Eid, Eman Gawish, Walaa Medhat, and Ahmed Hassan Yousef. 2023. Topic modeling algorithms and applications: A survey.Information Systems112 (2023), 102131

  2. [2]

    Suman Adhya, Avishek Lahiri, Debarshi Kumar Sanyal, and Partha Pratim Das

  3. [3]

    InPro- ceedings of the 19th International Conference on Natural Language Processing, ICON 2022, New Delhi, India, December 15-18, 2022

    Improving Contextualized Topic Models with Negative Sampling. InPro- ceedings of the 19th International Conference on Natural Language Processing, ICON 2022, New Delhi, India, December 15-18, 2022. Association for Computa- tional Linguistics, 128–138

  4. [4]

    Blei, Andrew Y

    David M. Blei, Andrew Y. Ng, and Michael I. Jordan. 2003. Latent Dirichlet Allocation.Journal of Machine Learning Research3 (2003), 993–1022

  5. [5]

    Tom B Brown. 2020. Language models are few-shot learners.arXiv preprint arXiv:2005.14165(2020)

  6. [6]

    Aaron Carass, Snehashis Roy, Adrian Gherman, Jacob C Reinhold, Andrew Jesson, Tal Arbel, Oskar Maier, Heinz Handels, Mohsen Ghafoorian, Bram Platel, et al

  7. [7]

    Evaluating white matter lesion segmentations with refined Sørensen-Dice analysis.Scientific reports10, 1 (2020), 8242

  8. [8]

    Shuyu Chang, Rui Wang, Peng Ren, Qi Wang, and Haiping Huang. 2024. A Large Language Model Guided Topic Refinement Mechanism for Short Text Modeling. arXiv preprint arXiv:2403.17706(2024)

  9. [9]

    Ching-Yao Chuang, Joshua Robinson, Yen-Chen Lin, Antonio Torralba, and Ste- fanie Jegelka. 2020. Debiased Contrastive Learning. InAdvances in Neural Infor- mation Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual

  10. [10]

    Marco Cuturi. 2013. Sinkhorn Distances: Lightspeed Computation of Optimal Transport. InAdvances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States. 2292–2300

  11. [11]

    Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Associa- tion for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, ...

  12. [12]

    Kennedy Edemacu and Xintao Wu. 2025. Privacy Preserving Prompt Engineering: A Survey.Comput. Surveys57, 10 (2025), 255:1–255:36

  13. [13]

    Zheng Fang, Yulan He, and Rob Procter. 2024. CWTM: Leveraging Contextualized Word Embeddings from BERT for Neural Topic Modeling. InProceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC/COLING 2024, 20-25 May, 2024, Torino, Italy. ELRA and ICCL, 4273–4286

  14. [14]

    Fiona Fui-Hoon Nah, Ruilin Zheng, Jingyuan Cai, Keng Siau, and Langtao Chen

  15. [15]

    277–304 pages

    Generative AI and ChatGPT: Applications, challenges, and AI-human collaboration. 277–304 pages

  16. [16]

    David W Goodall. 1967. The distribution of the matching coefficient.Biometrics (1967), 647–656

  17. [17]

    Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde- Farley, Sherjil Ozair, Aaron C

    Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde- Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. 2014. Generative Adversarial Nets. InAdvances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada. 2672–2680

  18. [18]

    Borgwardt, Malte J

    Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, and Alexander J. Smola. 2012. A Kernel Two-Sample Test.Journal of Machine Learning Research13 (2012), 723–773

  19. [19]

    Kingma and Jimmy Ba

    Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Opti- mization. In3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings

  20. [20]

    Kingma and Max Welling

    Diederik P. Kingma and Max Welling. 2014. Auto-Encoding Variational Bayes. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings

  21. [21]

    Sven Kosub. 2019. A note on the triangle inequality for the Jaccard distance. Pattern Recognition Letters120 (2019), 36–38

  22. [22]

    Jiayu Lin. 2016. On the dirichlet distribution.Department of Mathematics and Statistics, Queens University40 (2016)

  23. [23]

    Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv:1907.11692

  24. [24]

    van der Veer, Lamiece Hassan, Chenghua Lin, and Goran Nenadic

    Yanan Ma, Chenghao Xiao, Chenhan Yuan, Sabine N. van der Veer, Lamiece Hassan, Chenghua Lin, and Goran Nenadic. 2025. CAST: Corpus-Aware Self- similarity Enhanced Topic modelling. InProceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2025 - Volume 1:...

  25. [25]

    Yishu Miao, Lei Yu, and Phil Blunsom. 2016. Neural Variational Inference for Text Processing. InProceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016 (JMLR Workshop and Conference Proceedings, Vol. 48). JMLR.org, 1727–1736

  26. [26]

    Yida Mu, Chun Dong, Kalina Bontcheva, and Xingyi Song. 2024. Large Language Models Offer an Alternative to the Traditional Approach of Topic Modelling. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation, LREC/COLING 2024, 20-25 May, 2024, Torino, Italy. ELRA and ICCL, 10160–10171

  27. [27]

    Giorgio Patrini, Rianne van den Berg, Patrick Forré, Marcello Carioni, Samarth Bhargav, Max Welling, Tim Genewein, and Frank Nielsen. 2019. Sinkhorn AutoEn- coders. InProceedings of the Thirty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI 2019, Tel A viv, Israel, July 22-25, 2019 (Proceedings of Machine Learning Research, Vol. 115). AUAI...

  28. [28]

    Gabriel Peyré and Marco Cuturi. 2019. Computational Optimal Transport.Foun- dations and Trends in Machine Learning11, 5-6 (2019), 355–607

  29. [29]

    Chau Pham, Alexander Miserlis Hoyle, Simeng Sun, Philip Resnik, and Mohit Iyyer. 2024. TopicGPT: A Prompt-based Topic Modeling Framework. InProceed- ings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), NAACL 2024, Mexico City, Mexico, June 16-21, 20...

  30. [30]

    Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. 2019. Language models are unsupervised multitask learners.OpenAI blog 1, 8 (2019), 9

  31. [31]

    Nils Reimers and Iryna Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. InProceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Confer- ence on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019. Association for Computational L...

  32. [32]

    Michael Röder, Andreas Both, and Alexander Hinneburg. 2015. Exploring the Space of Topic Coherence Measures. InProceedings of the Eighth ACM Interna- tional Conference on Web Search and Data Mining, WSDM 2015, Shanghai, China, February 2-6, 2015. ACM, 399–408

  33. [33]

    Pranab Sahoo, Ayush Kumar Singh, Sriparna Saha, Vinija Jain, Samrat Mondal, and Aman Chadha. 2024. A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications.CoRRabs/2402.07927 (2024)

  34. [34]

    Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. 2020. MPNet: Masked and Permuted Pre-training for Language Understanding. InAdvances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual. Human-Centric Topic Modeling with Goal-Prompted Contrastiv...

  35. [35]

    Marek Strong, Rami Aly, and Andreas Vlachos. 2024. Zero-Shot Fact Verification via Natural Logic and Large Language Models. InFindings of the Association for Computational Linguistics: EMNLP 2024, Miami, Florida, USA, November 12-16,

  36. [36]

    Association for Computational Linguistics, 17021–17035

  37. [37]

    George R Terrell and David W Scott. 1992. Variable kernel density estimation. The Annals of Statistics(1992), 1236–1265

  38. [38]

    Hoang Tran Vuong, Tue Le, Tu Vu, Tung Nguyen, Linh Ngo Van, Sang Dinh, and Thien Huu Nguyen. 2025. HiCOT: Improving Neural Topic Models via Optimal Transport and Contrastive Learning. InFindings of the Association for Computational Linguistics: ACL 2025. 13894–13920

  39. [39]

    Wallach, David M

    Hanna M. Wallach, David M. Mimno, and Andrew McCallum. 2009. Rethinking LDA: Why Priors Matter. InAdvances in Neural Information Processing Systems 22: 23rd Annual Conference on Neural Information Processing Systems 2009. Proceedings of a meeting held 7-10 December 2009, Vancouver, British Columbia, Canada. Curran Associates, Inc., 1973–1981

  40. [40]

    White, Longqi Yang, Reid Andersen, Georg Buscher, Dhruv Joshi, and Nagu Rangan

    Mengting Wan, Tara Safavi, Sujay Kumar Jauhar, Yujin Kim, Scott Counts, Jennifer Neville, Siddharth Suri, Chirag Shah, Ryen W. White, Longqi Yang, Reid Andersen, Georg Buscher, Dhruv Joshi, and Nagu Rangan. 2024. TnT-LLM: Text Mining at Scale with Large Language Models. InProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining...

  41. [41]

    Rui Wang, Xuemeng Hu, Deyu Zhou, Yulan He, Yuxuan Xiong, Chenchen Ye, and Haiyang Xu. 2020. Neural Topic Modeling with Bidirectional Adversarial Training. InProceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020. Association for Computational Linguistics, 340–350

  42. [42]

    Rui Wang, Xing Liu, Yanan Wang, Shuyu Chang, Yuanzhi Yao, and Haiping Huang

  43. [43]

    InProceedings of the Eighteenth ACM International Conference on Web Search and Data Mining, WSDM 2025, Hannover, Germany, March 10-14,

    Mining Topics towards ChatGPT Using a Disentangled Contextualized- neural Topic Model. InProceedings of the Eighteenth ACM International Conference on Web Search and Data Mining, WSDM 2025, Hannover, Germany, March 10-14,

  44. [44]

    Rui Wang, Deyu Zhou, and Yulan He. 2019. ATM: Adversarial-neural Topic Model.Information Processing & Management56, 6 (2019)

  45. [45]

    Rui Wang, Deyu Zhou, and Yulan He. 2019. Open Event Extraction from Online Text using a Generative Adversarial Network. InProceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019. Association for ...

  46. [46]

    Rui Wang, Deyu Zhou, Haiping Huang, and Yongquan Zhou. 2025. MIT: Mutual Information Topic Model for Diverse Topic Extraction.IEEE Transactions on Neural Networks and Learning Systems36, 2 (2025), 2523–2537

  47. [47]

    Tongzhou Wang and Phillip Isola. 2020. Understanding Contrastive Represen- tation Learning through Alignment and Uniformity on the Hypersphere. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event (Proceedings of Machine Learning Research, Vol. 119). PMLR, 9929–9939

  48. [48]

    Wenhui Wang, Furu Wei, Li Dong, Hangbo Bao, Nan Yang, and Ming Zhou. 2020. MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre- Trained Transformers. InAdvances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual

  49. [49]

    Zihan Wang, Jingbo Shang, and Ruiqi Zhong. 2023. Goal-Driven Explainable Clustering via Language Descriptions. InProceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023. Association for Computational Linguistics, 10626–10649

  50. [50]

    Chi, Quoc V

    Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed H. Chi, Quoc V. Le, and Denny Zhou. 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. InAdvances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, Novem...

  51. [51]

    Xiaobao Wu, Thong Nguyen, and Anh Tuan Luu. 2024. A survey on neural topic models: methods, applications, and challenges.Artificial Intelligence Review57, 2 (2024), 18

  52. [52]

    Weijie Xu, Xiaoyu Jiang, Srinivasan Sengamedu Hanumantha Rao, Francis Ian- nacci, and Jinjin Zhao. 2023. vONTSS: vMF based semi-supervised neural topic modeling with optimal transport. InFindings of the Association for Computa- tional Linguistics: ACL 2023, Toronto, Canada, July 9-14, 2023. Association for Computational Linguistics, 4433–4457

  53. [53]

    Xiaohao Yang, He Zhao, Weijie Xu, Yuanyuan Qi, Jueqing Lu, Dinh Phung, and Lan Du. 2025. Neural Topic Modeling with Large Language Models in the Loop. InProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2025, Vienna, Austria, July 27 - August 1,

  54. [54]

    Association for Computational Linguistics, 1377–1401

  55. [55]

    Dejian Yu and Bo Xiang. 2023. Discovering topics and trends in the field of Arti- ficial Intelligence: Using LDA topic modeling.Expert Systems with Applications 225 (2023), 120114

  56. [56]

    Yuwei Zhang, Zihan Wang, and Jingbo Shang. 2023. ClusterLLM: Large Language Models as a Guide for Text Clustering. InProceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023. Association for Computational Linguistics, 13903–13920