pith. machine review for the scientific record. sign in

arxiv: 2605.13875 · v1 · submitted 2026-05-08 · 💻 cs.GT

Recognition: 2 theorem links

· Lean Theorem

Common-agency Games for Multi-Objective Test-Time Alignment

Authors on Pith no claims yet

Pith reviewed 2026-05-15 06:11 UTC · model grok-4.3

classification 💻 cs.GT
keywords multi-objective alignmenttest-time alignmentgame theoryLLM alignmentcommon agencyequilibrium computationEPEC
0
0 comments X

The pith

CAGE treats multiple conflicting alignment goals as strategic principals bidding token incentives to produce an equilibrium LLM policy at inference time.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces a training-free method for aligning large language models to several competing objectives at once. It frames each objective as a principal that offers token-level rewards to the shared model, then solves for the resulting equilibrium policy. An algorithm based on equilibrium problems with equilibrium constraints computes this balance and comes with proofs of existence, uniqueness, convergence, and stability. The resulting system lets users dial the relative strength of objectives during generation without any model updates. If the equilibrium truly reflects joint preferences, it makes handling diverse user requirements feasible even when compute is limited.

Core claim

CAGE models alignment objectives as strategic principals that allocate token-level incentives to a shared LLM, inducing an equilibrium policy that captures the joint effect of competing objectives. An efficient EPEC-based algorithm computes this equilibrium and supplies theoretical guarantees on existence, uniqueness, convergence, stability, and no-regret dynamics. Empirically the approach yields flexible trade-offs, outperforms prior test-time methods, requires no retraining, and enables weak-to-strong generalization.

What carries the argument

The common-agency game in which multiple principals allocate token-level incentives to a shared agent (the LLM policy), solved as an equilibrium problem with equilibrium constraints.

If this is right

  • Trade-offs between objectives can be adjusted on the fly at inference without retraining.
  • The same base model can serve multiple user groups by changing only the incentive parameters.
  • Weak models can be steered toward strong-model behavior through the equilibrium computation.
  • Resource-constrained deployments gain practical multi-objective control.
  • The equilibrium policy is stable under the stated no-regret dynamics.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same incentive-bidding structure could be applied to non-LLM sequential decision tasks where multiple stakeholders share an agent.
  • Real-time preference sliders become feasible if the EPEC solver runs fast enough per generation step.
  • The framework implicitly defines a new way to audit alignment by inspecting the equilibrium incentives rather than the final outputs alone.

Load-bearing premise

That the equilibrium arising from principals bidding on tokens will meaningfully represent real combined user preferences rather than an artifact of the incentive model.

What would settle it

A controlled study in which human raters score CAGE outputs on conflicting objectives such as helpfulness versus safety and check whether the observed trade-offs match the intended weighting of the principals.

Figures

Figures reproduced from arXiv: 2605.13875 by Baiting Chen, Rui Yu, Tong Zhu, Xiaowu Dai.

Figure 1
Figure 1. Figure 1: Illustration of CAGE. The LLM generates tokens sequentially, while multiple [PITH_FULL_IMAGE:figures/full_fig_p003_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Learned Pareto fronts for the safety alignment task. [PITH_FULL_IMAGE:figures/full_fig_p011_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Learned Pareto fronts for the safety alignment task by all methods. [PITH_FULL_IMAGE:figures/full_fig_p034_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Learned Pareto Fronts for Different Hyperparameter Configurations [PITH_FULL_IMAGE:figures/full_fig_p035_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Three-objective evaluation of CAGE vs. GenARM on the Helpful Assistant task: [PITH_FULL_IMAGE:figures/full_fig_p037_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Pareto frontiers of CAGE versus GenARM along the three pairwise simplex edges [PITH_FULL_IMAGE:figures/full_fig_p038_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Case study at balanced preference α = (0.5, 0.5). CAGE+ provides the most comprehensive and balanced response with both helpful and safety-aware content. 40 [PITH_FULL_IMAGE:figures/full_fig_p040_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Case study at safety-heavy preference α = (0.2, 0.8). CAGE methods produce longer, more safety-focused responses compared to baselines. 41 [PITH_FULL_IMAGE:figures/full_fig_p041_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Case study at helpfulness-heavy preference [PITH_FULL_IMAGE:figures/full_fig_p042_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: HH-RLHF case study at the help-corner α = (1, 0, 0). Both GenARM and CAGE produce serious, substantive advice about exposing babies to music. GenARM scores slightly higher on helpfulness (+3.08 vs. +2.84). 43 [PITH_FULL_IMAGE:figures/full_fig_p043_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: HH-RLHF case study at the humor-corner α = (0, 0, 1). Both methods shift to a playful register—GenARM with emojis and meta-commentary, CAGE with casual decade slang—reaching nearly identical humor scores (∼ 1.0). CAGE retains substantially more harmlessness headroom at the same humor level (−1.24 vs. −2.17). Appendix E. Potential Social Impact Our framework aims to improve the controllability of language … view at source ↗
read the original abstract

Aligning large language models (LLMs) with human preferences is inherently multi-objective: different users and evaluation criteria impose heterogeneous and often conflicting requirements on model outputs. We propose CAGE (Common-Agency Games for Alignment), a training-free, game-theoretic framework for multi-objective test-time alignment. CAGE models alignment objectives as strategic principals that allocate token-level incentives to a shared LLM, inducing an equilibrium policy that captures the joint effect of competing objectives. We develop an efficient algorithm based on equilibrium problems with equilibrium constraints (EPEC) to compute this equilibrium, and establish theoretical guarantees including existence and uniqueness of the equilibrium policy, convergence and stability of the algorithm, and no-regret learning dynamics. Empirically, CAGE enables flexible and fine-grained trade-offs across objectives at inference time, consistently outperforming existing test-time alignment methods while requiring no retraining. It further supports weak-to-strong generalization, making multi-objective alignment practical in resource-constrained settings.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper proposes CAGE, a training-free game-theoretic framework for multi-objective test-time alignment of LLMs. It models heterogeneous alignment objectives as strategic principals in a common-agency game that allocate token-level incentives to a shared LLM policy, inducing an equilibrium policy via an EPEC-based algorithm. The manuscript claims theoretical guarantees of existence, uniqueness, convergence, and stability for the equilibrium, plus empirical outperformance over existing test-time methods with no retraining and support for weak-to-strong generalization.

Significance. If the token-level incentive equilibrium reliably induces coherent sequence-level trade-offs that capture user preferences, CAGE would offer a practical, retraining-free approach to flexible multi-objective alignment, particularly valuable in resource-constrained settings. The combination of game-theoretic modeling with EPEC computation and empirical validation on trade-offs represents a novel application of common-agency ideas to LLM alignment.

major comments (2)
  1. [Abstract] Abstract: The claims of existence, uniqueness, convergence, and stability for the EPEC equilibrium are asserted without reference to the convexity, compactness, or continuity conditions under which standard EPEC theory guarantees these properties; the non-convex logit space of transformer policies may violate these, risking that the computed equilibrium does not correspond to the intended joint effect of objectives.
  2. [§3 (Framework)] The central modeling choice of token-level principal incentives must be shown to produce stable sequence-level trade-offs under autoregressive generation; without explicit analysis of how local incentives aggregate to global preference weightings (e.g., avoiding myopic alternation between high- and low-reward tokens), the claim that the equilibrium policy faithfully encodes multi-objective preferences remains unverified.
minor comments (2)
  1. [§4] Notation for the incentive allocation functions and the EPEC formulation should be introduced with explicit definitions before use in the algorithm description to improve readability.
  2. [§5] The empirical section would benefit from additional baselines that also operate at test time without retraining to strengthen the comparison.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive feedback. We address the two major comments point by point below. Where the comments correctly identify gaps in the presentation or analysis, we have revised the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The claims of existence, uniqueness, convergence, and stability for the EPEC equilibrium are asserted without reference to the convexity, compactness, or continuity conditions under which standard EPEC theory guarantees these properties; the non-convex logit space of transformer policies may violate these, risking that the computed equilibrium does not correspond to the intended joint effect of objectives.

    Authors: We agree that the abstract is too terse on the underlying assumptions. Section 4 of the manuscript invokes standard EPEC existence and uniqueness results that require compact convex strategy sets and continuous payoff functions. The logit parameterization of the policy is indeed non-convex, so the guarantees apply to the relaxed continuous strategy space; the algorithm returns an approximate equilibrium whose quality is controlled by the projection step. In the revision we have (i) added an explicit statement of the compactness/continuity conditions in both the abstract and Section 4, (ii) clarified that the computed policy is a local equilibrium of the non-convex problem, and (iii) included a short discussion of the approximation gap relative to the convex relaxation. revision: yes

  2. Referee: [§3 (Framework)] The central modeling choice of token-level principal incentives must be shown to produce stable sequence-level trade-offs under autoregressive generation; without explicit analysis of how local incentives aggregate to global preference weightings (e.g., avoiding myopic alternation between high- and low-reward tokens), the claim that the equilibrium policy faithfully encodes multi-objective preferences remains unverified.

    Authors: This is a substantive point. The original manuscript provides empirical evidence that the induced policies exhibit coherent sequence-level trade-offs, but it does not contain a formal aggregation argument. We have added a new subsection (3.4) that derives the expected sequence utility as a convex combination of the principals’ objectives under the token-level equilibrium incentives. We further prove a stability bound showing that the probability of myopic alternation decays exponentially with sequence length when the incentive functions satisfy a Lipschitz condition (which holds for the linear and log-linear reward models used in the experiments). These additions directly address the aggregation concern. revision: yes

Circularity Check

0 steps flagged

No circularity: derivation applies external EPEC theory to new alignment setting

full rationale

The paper frames multi-objective alignment as a common-agency game and invokes standard EPEC results for existence, uniqueness, convergence, and stability. These guarantees are presented as following from established game-theoretic conditions rather than being derived from or equivalent to the paper's own fitted parameters or self-citations. No equation reduces a claimed prediction to an input by construction, and no load-bearing step relies on a self-citation chain that itself lacks independent verification. Empirical comparisons to baselines are external to the modeling assumptions.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

Based on abstract only; the central claim rests on standard game-theory existence results plus the domain assumption that token-level incentives can represent heterogeneous objectives. No free parameters or invented entities are explicitly named.

axioms (2)
  • domain assumption Existence and uniqueness of the equilibrium policy in the common-agency game
    Stated as a theoretical guarantee in the abstract.
  • domain assumption Convergence and stability of the EPEC algorithm
    Claimed without derivation details.

pith-pipeline@v0.9.0 · 5463 in / 1138 out tokens · 34875 ms · 2026-05-15T06:11:14.544441+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

What do these tags mean?
matches
The paper's claim is directly supported by a theorem in the formal canon.
supports
The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
extends
The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
uses
The paper appears to rely on the theorem as machinery.
contradicts
The paper's claim conflicts with a theorem or certificate in the canon.
unclear
Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.

Reference graph

Works this paper leans on

230 extracted references · 230 canonical work pages · 17 internal anchors

  1. [1]

    2024 , eprint=

    TinyLlama: An Open-Source Small Language Model , author=. 2024 , eprint=

  2. [2]

    International Conference on Learning Representations , year=

    Measuring Massive Multitask Language Understanding , author=. International Conference on Learning Representations , year=

  3. [3]

    The Twelfth International Conference on Learning Representations , year=

    The Consensus Game: Language Model Generation via Equilibrium Search , author=. The Twelfth International Conference on Learning Representations , year=

  4. [4]

    C hat D ev: Communicative agents for software development

    Qian, Chen and Liu, Wei and Liu, Hongzhang and Chen, Nuo and Dang, Yufan and Li, Jiahao and Yang, Cheng and Chen, Weize and Su, Yusheng and Cong, Xin and Xu, Juyuan and Li, Dahai and Liu, Zhiyuan and Sun, Maosong. C hat D ev: Communicative agents for software development. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguist...

  5. [5]

    Encouraging divergent thinking in large language models through multi-agent debate

    Liang, Tian and He, Zhiwei and Jiao, Wenxiang and Wang, Xing and Wang, Yan and Wang, Rui and Yang, Yujiu and Shi, Shuming and Tu, Zhaopeng. Encouraging divergent thinking in large language models through multi-agent debate. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing. 2024

  6. [6]

    ESAIM: Control, Optimisation and Calculus of Variations , volume=

    A consensus-based global optimization method for high dimensional machine learning problems , author=. ESAIM: Control, Optimisation and Calculus of Variations , volume=. 2021 , publisher=

  7. [7]

    Science , volume=

    Evidence for a collective intelligence factor in the performance of human groups , author=. Science , volume=. 2010 , publisher=

  8. [8]

    , author=

    A measure for brain complexity: relating functional segregation and integration in the nervous system. , author=. Proceedings of the National Academy of Sciences , volume=

  9. [9]

    Nature Neuroscience , volume=

    A synergistic core for human brain evolution and cognition , author=. Nature Neuroscience , volume=. 2022 , publisher=

  10. [10]

    The Llama 3 Herd of Models

    The llama 3 herd of models , author=. arXiv preprint arXiv:2407.21783 , year=

  11. [11]

    Mutual Information and Diverse Decoding Improve Neural Machine Translation

    Mutual information and diverse decoding improve neural machine translation , author=. arXiv preprint arXiv:1601.00372 , year=

  12. [12]

    First Conference on Language Modeling , year=

    Gpqa: A graduate-level google-proof q&a benchmark , author=. First Conference on Language Modeling , year=

  13. [13]

    Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge

    Think you have solved question answering? try arc, the ai2 reasoning challenge , author=. arXiv preprint arXiv:1803.05457 , year=

  14. [14]

    arXiv preprint arXiv:2406.01855 , year=

    Trutheval: A dataset to evaluate llm truthfulness and reliability , author=. arXiv preprint arXiv:2406.01855 , year=

  15. [15]

    Trustworthy llms: a survey and guideline for evaluating large language models’ alignment

    Trustworthy llms: a survey and guideline for evaluating large language models' alignment , author=. arXiv preprint arXiv:2308.05374 , year=

  16. [16]

    arXiv preprint arXiv:2501.08263 , year=

    Multiplayer federated learning: Reaching equilibrium with less communication , author=. arXiv preprint arXiv:2501.08263 , year=

  17. [17]

    arXiv preprint arXiv:2411.11609 , year=

    VLN-Game: Vision-Language Equilibrium Search for Zero-Shot Semantic Navigation , author=. arXiv preprint arXiv:2411.11609 , year=

  18. [18]

    arXiv preprint arXiv:2410.03968 , year=

    Decoding Game: On Minimax Optimality of Heuristic Text Generation Strategies , author=. arXiv preprint arXiv:2410.03968 , year=

  19. [19]

    Proceedings of the ACM on Web Conference 2025 , pages=

    Adaptive activation steering: A tuning-free llm truthfulness improvement method for diverse hallucinations categories , author=. Proceedings of the ACM on Web Conference 2025 , pages=

  20. [20]

    ACM Transactions on Information Systems , volume=

    A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions , author=. ACM Transactions on Information Systems , volume=. 2025 , publisher=

  21. [21]

    Gemini: A Family of Highly Capable Multimodal Models

    Gemini: a family of highly capable multimodal models , author=. arXiv preprint arXiv:2312.11805 , year=

  22. [22]

    GPT-4 Technical Report

    Gpt-4 technical report , author=. arXiv preprint arXiv:2303.08774 , year=

  23. [24]

    arXiv preprint arXiv:1909.04696 , year=

    Sunny and dark outside?! improving answer consistency in vqa through entailed question generation , author=. arXiv preprint arXiv:1909.04696 , year=

  24. [25]

    Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing , pages=

    Enhancing Self-Consistency and Performance of Pre-Trained Language Models through Natural Language Inference , author=. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing , pages=

  25. [26]

    2022 , booktitle=

    Self-Consistency Improves Chain of Thought Reasoning in Language Models , author=. 2022 , booktitle=

  26. [27]

    2023 , booktitle=

    Prototypical Calibration for Few-shot Learning of Language Models , author=. 2023 , booktitle=

  27. [28]

    Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , pages=

    A Close Look into the Calibration of Pre-trained Language Models , author=. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , pages=

  28. [29]

    Management Science , volume=

    Eliciting informative feedback: The peer-prediction method , author=. Management Science , volume=. 2005 , publisher=

  29. [30]

    Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms , pages=

    Dominantly truthful multi-task peer prediction with a constant number of tasks , author=. Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms , pages=. 2020 , organization=

  30. [31]

    Foundations and Trends

    Regret analysis of stochastic and nonstochastic multi-armed bandit problems , author=. Foundations and Trends. 2012 , publisher=

  31. [32]

    C. K. Chow and C. N. Liu , title =. IEEE Transactions on Information Theory , year =

  32. [33]

    1988 , address =

    Judea Pearl , title =. 1988 , address =

  33. [34]

    Transactions of the Association for Computational Linguistics , volume=

    Measuring and improving consistency in pretrained language models , author=. Transactions of the Association for Computational Linguistics , volume=. 2021 , publisher=

  34. [35]

    Forty-first International Conference on Machine Learning , year=

    Cllms: Consistency large language models , author=. Forty-first International Conference on Machine Learning , year=

  35. [36]

    arXiv preprint arXiv:2211.05853 , year=

    Measuring reliability of large language models through semantic consistency , author=. arXiv preprint arXiv:2211.05853 , year=

  36. [37]

    Advances in Neural Information Processing Systems , volume=

    Chain-of-thought prompting elicits reasoning in large language models , author=. Advances in Neural Information Processing Systems , volume=

  37. [38]

    A Survey of Large Language Models

    A survey of large language models , author=. arXiv preprint arXiv:2303.18223 , year=

  38. [39]

    Authorea Preprints , year=

    A survey on large language models: Applications, challenges, limitations, and practical usage , author=. Authorea Preprints , year=

  39. [40]

    Proceedings of the AAAI Conference on Artificial Intelligence , year=

    Outfox: Llm-generated essay detection through in-context learning with adversarially generated examples , author=. Proceedings of the AAAI Conference on Artificial Intelligence , year=

  40. [41]

    Artificial Intelligence Review , volume=

    A survey of safety and trustworthiness of large language models through the lens of verification and validation , author=. Artificial Intelligence Review , volume=. 2024 , publisher=

  41. [43]

    Mathematics of Operations Research , volume=

    Multiagent online learning in time-varying games , author=. Mathematics of Operations Research , volume=. 2023 , publisher=

  42. [44]

    International Conference on Machine Learning , pages=

    Doubly optimal no-regret learning in monotone games , author=. International Conference on Machine Learning , pages=. 2023 , organization=

  43. [45]

    Proceedings of the AAAI Conference on Artificial Intelligence , year=

    An efficient algorithm for fair multi-agent multi-armed bandit with low regret , author=. Proceedings of the AAAI Conference on Artificial Intelligence , year=

  44. [46]

    The Twelfth International Conference on Learning Representations , year=

    Let's verify step by step , author=. The Twelfth International Conference on Learning Representations , year=

  45. [47]

    arXiv preprint arXiv:2403.11807 , year=

    How Far Are We on the Decision-Making of LLMs? Evaluating LLMs' Gaming Ability in Multi-Agent Environments , author=. arXiv preprint arXiv:2403.11807 , year=

  46. [48]

    arXiv preprint arXiv:2402.12348 , year=

    Gtbench: Uncovering the strategic reasoning limitations of llms via game-theoretic evaluations , author=. arXiv preprint arXiv:2402.12348 , year=

  47. [49]

    From text to tactic: Evaluating llms playing the game of avalon

    Avalonbench: Evaluating llms playing the game of avalon , author=. arXiv preprint arXiv:2310.05036 , year=

  48. [50]

    Advances in Neural Information Processing Systems , volume=

    Self-playing adversarial language game enhances llm reasoning , author=. Advances in Neural Information Processing Systems , volume=

  49. [51]

    arXiv preprint arXiv:2411.05990 , year=

    Game-theoretic LLM: Agent workflow for negotiation games , author=. arXiv preprint arXiv:2411.05990 , year=

  50. [52]

    arXiv preprint arXiv:2411.13543 , year=

    Balrog: Benchmarking agentic llm and vlm reasoning on games , author=. arXiv preprint arXiv:2411.13543 , year=

  51. [53]

    arXiv preprint arXiv:2403.16843 , year=

    Do llm agents have regret? a case study in online learning and games , author=. arXiv preprint arXiv:2403.16843 , year=

  52. [55]

    arXiv preprint arXiv:2402.01704 , year=

    States as strings as strategies: Steering language models with game-theoretic solvers , author=. arXiv preprint arXiv:2402.01704 , year=

  53. [56]

    , author=

    LLM-Deliberation: Evaluating LLMs with Interactive Multi-Agent Negotiation Games. , author=. 2023 , publisher=

  54. [57]

    arXiv preprint arXiv:2308.10032 , year=

    Gameeval: Evaluating llms on conversational games , author=. arXiv preprint arXiv:2308.10032 , year=

  55. [58]

    Journal of Machine Learning Research , volume=

    Nash Q-learning for general-sum stochastic games , author=. Journal of Machine Learning Research , volume=

  56. [59]

    Advances in Neural Information Processing Systems , volume=

    Reinforcement learning to play an optimal Nash equilibrium in team Markov games , author=. Advances in Neural Information Processing Systems , volume=

  57. [60]

    Casgrain, Philippe and Ning, Brian and Jaimungal, Sebastian , journal=. Deep. 2022 , publisher=

  58. [61]

    International Conference on Machine Learning , pages=

    Nash Learning from Human Feedback , author=. International Conference on Machine Learning , pages=. 2024 , organization=

  59. [62]

    arXiv e-prints , pages=

    A theoretical analysis of nash learning from human feedback under general kl-regularized preference , author=. arXiv e-prints , pages=

  60. [63]

    arXiv preprint arXiv:2404.03715 , year=

    Direct nash optimization: Teaching language models to self-improve with general preferences , author=. arXiv preprint arXiv:2404.03715 , year=

  61. [64]

    Prelec, Drazen , journal=. A. 2004 , publisher=

  62. [65]

    Proceedings of the AAAI Conference on Artificial Intelligence , year=

    A robust bayesian truth serum for small populations , author=. Proceedings of the AAAI Conference on Artificial Intelligence , year=

  63. [66]

    Econometrica: journal of the Econometric Society , pages=

    Incentive compatibility and the bargaining problem , author=. Econometrica: journal of the Econometric Society , pages=. 1979 , publisher=

  64. [67]

    Journal of Economic theory , volume=

    Multidimensional incentive compatibility and mechanism design , author=. Journal of Economic theory , volume=. 1988 , publisher=

  65. [68]

    Learning for Dynamics and Control , pages=

    Scalable reinforcement learning of localized policies for multi-agent networked systems , author=. Learning for Dynamics and Control , pages=. 2020 , organization=

  66. [69]

    Foundations and Trends

    Introduction to online convex optimization , author=. Foundations and Trends. 2016 , publisher=

  67. [70]

    ICML , volume=

    On no-regret learning, fictitious play, and nash equilibrium , author=. ICML , volume=

  68. [71]

    Proceedings of the ACM on Measurement and Analysis of Computing Systems , volume=

    Global convergence of localized policy iteration in networked multi-agent reinforcement learning , author=. Proceedings of the ACM on Measurement and Analysis of Computing Systems , volume=. 2023 , publisher=

  69. [72]

    Proceedings of the AAAI Conference on Artificial Intelligence , year=

    Incentives for truthful information elicitation of continuous signals , author=. Proceedings of the AAAI Conference on Artificial Intelligence , year=

  70. [73]

    Equilibrium Selection in Information Elicitation without Verification via Information Monotonicity

    Equilibrium selection in information elicitation without verification via information monotonicity , author=. arXiv preprint arXiv:1603.07751 , year=

  71. [74]

    ACM Transactions on Economics and Computation , volume=

    Two strongly truthful mechanisms for three heterogeneous agents answering one question , author=. ACM Transactions on Economics and Computation , volume=. 2023 , publisher=

  72. [75]

    Proceedings of the 2014 International Conference on Autonomous Agents and Multi-Agent Systems , pages=

    Elicitability and knowledge-free elicitation with peer prediction , author=. Proceedings of the 2014 International Conference on Autonomous Agents and Multi-Agent Systems , pages=

  73. [76]

    ACM Transactions on Economics and Computation , volume=

    Peer prediction with heterogeneous users , author=. ACM Transactions on Economics and Computation , volume=. 2020 , publisher=

  74. [77]

    ACM Transactions on Economics and Computation (TEAC) , volume=

    An information theoretic framework for designing information elicitation mechanisms that reward truth-telling , author=. ACM Transactions on Economics and Computation (TEAC) , volume=. 2019 , publisher=

  75. [78]

    arXiv preprint arXiv:2009.14730 , year=

    Learning and strongly truthful multi-task peer prediction: A variational approach , author=. arXiv preprint arXiv:2009.14730 , year=

  76. [79]

    Neurocomputing , volume=

    Online learning: A comprehensive survey , author=. Neurocomputing , volume=. 2021 , publisher=

  77. [80]

    Foundations and Trends in Machine Learning , volume=

    Online learning and online convex optimization , author=. Foundations and Trends in Machine Learning , volume=. 2012 , publisher=

  78. [81]

    arXiv preprint arXiv:2410.16714 , year=

    Magnetic Preference Optimization: Achieving Last-iterate Convergence for Language Model Alignment , author=. arXiv preprint arXiv:2410.16714 , year=

  79. [82]

    arXiv preprint arXiv:2407.00617 , year=

    Iterative Nash Policy Optimization: Aligning LLMs with General Preferences via No-Regret Learning , author=. arXiv preprint arXiv:2407.00617 , year=

  80. [83]

    Towards an AI co-scientist

    Towards an AI co-scientist , author=. arXiv preprint arXiv:2502.18864 , year=

Showing first 80 references.