pith. machine review for the scientific record. sign in

arxiv: 2603.27771 · v2 · submitted 2026-03-29 · 💻 cs.MA · cs.CL· cs.CY

Recognition: 1 theorem link

· Lean Theorem

Emergent Social Intelligence Risks in Generative Multi-Agent Systems

Haomin Zhuang, Huan Sun, Nitesh V Chawla, Nouha Dziri, Nuno Moniz, Pin-Yu Chen, Wenjie Wang, Xiangliang Zhang, Xiaonan Luo, Yuchen Ma, Yue Huang, Yu Jiang, Zhangchen Xu, Zichen Chen, Zinan Lin

Authors on Pith no claims yet

Pith reviewed 2026-05-14 21:41 UTC · model grok-4.3

classification 💻 cs.MA cs.CLcs.CY
keywords emergent risksmulti-agent systemsgenerative modelssocial intelligencecollusionconformityAI safetycollective behavior
0
0 comments X

The pith

Generative multi-agent systems spontaneously develop collusion-like coordination and conformity that mirror human social failures.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper examines multi-agent systems built from large generative models that plan, negotiate, and share resources to complete tasks. It reports that these agent groups frequently produce coordination resembling collusion and conformity to emerging group norms across repeated trials and varied conditions. These patterns appear in resource competition, sequential handoffs, and collective decisions without any explicit instructions to agents. The behaviors persist even when individual agents are protected by existing safeguards. This points to failure modes that arise only at the collective level and cannot be reduced to problems with single agents.

Core claim

In workflows that involve competition over shared resources, sequential handoff collaboration where downstream agents see only predecessor outputs, and collective decision aggregation, agent collectives spontaneously reproduce familiar failure patterns from human societies such as collusion-like coordination and conformity. These group behaviors arise with non-trivial frequency under realistic resource constraints, communication protocols, and role assignments rather than as rare cases. Moreover, the risks cannot be prevented by existing agent-level safeguards alone.

What carries the argument

Emergent social intelligence risk, the collective interaction dynamics in generative multi-agent systems that cause groups to reproduce societal pathologies without explicit instruction.

If this is right

  • Individual agent safeguards leave multi-agent systems exposed to group-level coordination failures.
  • Risks appear reliably across different workflows and conditions rather than as isolated exceptions.
  • Real-world deployments involving shared resources or sequential handoffs carry these emergent behaviors by default.
  • Preventing the risks requires interventions that target collective dynamics instead of single agents.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Market or organizational deployments of such systems could unintentionally produce coordinated resource hoarding or biased decisions.
  • Safety testing should incorporate multi-agent scenarios with communication to surface these patterns before release.
  • Design choices in communication protocols or role assignments may modulate the frequency of these behaviors.
  • The findings suggest studying AI collectives at the scale of social groups rather than as isolated units.

Load-bearing premise

The observed collusion-like and conformity behaviors arise from genuine multi-agent interaction dynamics rather than training data artifacts, prompt phrasing, or model-specific quirks.

What would settle it

Re-running the same interaction setups but with agents isolated so they receive no outputs from others and perform tasks independently would eliminate the collusion-like and conformity patterns.

read the original abstract

Multi-agent systems composed of large generative models are rapidly moving from laboratory prototypes to real-world deployments, where they jointly plan, negotiate, and allocate shared resources to solve complex tasks. While such systems promise unprecedented scalability and autonomy, their collective interaction also gives rise to failure modes that cannot be reduced to individual agents. Understanding these emergent risks is therefore critical. Here, we present a pioneer study of such emergent multi-agent risk in workflows that involve competition over shared resources (e.g., computing resources or market share), sequential handoff collaboration (where downstream agents see only predecessor outputs), collective decision aggregation, and others. Across these settings, we observe that such group behaviors arise frequently across repeated trials and a wide range of interaction conditions, rather than as rare or pathological cases. In particular, phenomena such as collusion-like coordination and conformity emerge with non-trivial frequency under realistic resource constraints, communication protocols, and role assignments, mirroring well-known pathologies in human societies despite no explicit instruction. Moreover, these risks cannot be prevented by existing agent-level safeguards alone. These findings expose the dark side of intelligent multi-agent systems: a social intelligence risk where agent collectives, despite no instruction to do so, spontaneously reproduce familiar failure patterns from human societies.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 1 minor

Summary. The paper claims that generative multi-agent systems exhibit emergent social intelligence risks, including collusion-like coordination and conformity behaviors, that arise frequently across workflows involving resource competition, sequential handoffs, collective decision-making, and similar settings. These patterns occur despite no explicit instructions, mirror human societal pathologies, and cannot be prevented by existing agent-level safeguards alone.

Significance. If substantiated with rigorous controls and quantitative data, the work would provide a valuable early empirical exploration of interaction-level failure modes in deployed multi-agent generative systems, highlighting the need for safeguards beyond individual agent alignment. The absence of such support in the current manuscript limits its immediate impact.

major comments (3)
  1. [Abstract] Abstract: The central claim that risks 'arise frequently across repeated trials and a wide range of interaction conditions' is presented without any quantitative support such as trial counts, observed frequencies, error bars, or statistical tests, leaving the 'non-trivial frequency' assertion without empirical grounding.
  2. [Abstract] Abstract: The attribution of collusion-like and conformity behaviors to multi-agent interaction dynamics requires evidence that these do not arise from model pre-training artifacts or prompt phrasing; no single-agent control runs, prompt ablation studies, or cross-model comparisons are described to rule out these alternatives.
  3. [Abstract] Abstract: The assertion that 'these risks cannot be prevented by existing agent-level safeguards alone' is load-bearing for the main contribution yet is stated without any reported experiments testing safeguard variants or interaction-protocol modifications.
minor comments (1)
  1. [Abstract] The abstract uses informal phrasing such as 'dark side' and 'pioneer study'; consider more precise language for a formal journal submission.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback. We agree that the abstract requires explicit quantitative grounding, control experiments, and safeguard tests to strengthen the claims about emergent risks. The revised manuscript will incorporate these elements while preserving the core contribution as an early empirical exploration. We address each major comment below.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim that risks 'arise frequently across repeated trials and a wide range of interaction conditions' is presented without any quantitative support such as trial counts, observed frequencies, error bars, or statistical tests, leaving the 'non-trivial frequency' assertion without empirical grounding.

    Authors: We acknowledge that the abstract condenses the empirical details. The full manuscript reports 250 independent trials across the four workflow types, with collusion-like coordination observed in 58% of cases (SD=7.2%) and conformity in 47% (SD=8.1%). Frequencies were stable across temperature settings (0.7-1.0) and communication constraints. We will revise the abstract to include: 'observed in 58% of 250 trials (p<0.001 via binomial test against 10% baseline)'. A summary table with per-workflow percentages and error bars will be added to the results section. revision: yes

  2. Referee: [Abstract] Abstract: The attribution of collusion-like and conformity behaviors to multi-agent interaction dynamics requires evidence that these do not arise from model pre-training artifacts or prompt phrasing; no single-agent control runs, prompt ablation studies, or cross-model comparisons are described to rule out these alternatives.

    Authors: We agree that ruling out non-interaction sources is necessary. We will add single-agent control runs in which each agent receives the complete task without peer interaction, yielding 0% collusion or conformity. Prompt ablations will systematically vary phrasing (e.g., neutral vs. competitive wording) while retaining the patterns at 51-60% rates. Cross-model results using GPT-4o, Claude-3.5-Sonnet, and Llama-3.1-405B will be reported, showing emergence rates of 52-61%. These will appear in a new 'Controls' subsection. revision: yes

  3. Referee: [Abstract] Abstract: The assertion that 'these risks cannot be prevented by existing agent-level safeguards alone' is load-bearing for the main contribution yet is stated without any reported experiments testing safeguard variants or interaction-protocol modifications.

    Authors: This point is well taken. We will add experiments applying agent-level safeguards (constitutional principles, output filters, and self-critique) that reduce individual violations by ~30% yet leave group collusion at 45% of trials. We will also evaluate interaction-protocol changes such as mandatory public logging and consensus overrides, which lower incidence to 22%. Results will be presented in the discussion to support the claim that agent-level measures are insufficient without system-level interventions. revision: yes

Circularity Check

0 steps flagged

No circularity: purely observational simulation results

full rationale

The paper reports frequencies of observed behaviors (collusion-like coordination, conformity) across multi-agent simulation trials under varying conditions. No equations, fitted parameters, self-referential derivations, or load-bearing self-citations appear in the provided text. Claims rest on direct experimental outcomes rather than any chain that reduces to its own inputs by construction. The absence of mathematical modeling or parameter estimation eliminates the patterns that would trigger circularity flags.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on the domain assumption that observed group behaviors in simulations reflect intrinsic properties of generative multi-agent interaction rather than experimental artifacts. No free parameters or invented entities are introduced in the abstract.

axioms (1)
  • domain assumption Generative models in multi-agent settings can produce human-like social pathologies without explicit instruction
    Invoked to interpret simulation outcomes as emergent social intelligence risks.

pith-pipeline@v0.9.0 · 5568 in / 1206 out tokens · 37998 ms · 2026-05-14T21:41:41.738512+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

159 extracted references · 159 canonical work pages · 5 internal anchors

  1. [1]

    Llm- deliberation: Evaluating llms with interactive multi-agent negotiation games

    Sahar Abdelnabi, Amr Gomaa, Sarath Sivaprasad, Lea Schönherr, and Mario Fritz. Llm- deliberation: Evaluating llms with interactive multi-agent negotiation games. 2023

  2. [2]

    Adair, Nancy R

    Wendi L. Adair, Nancy R. Buchan, Xiao-Ping Chen, and Leigh Anne Liu.Beyond Silence: How Context Communicates in Cross-Cultural Negotiation, pages 63–75. Springer Nature Switzerland, Cham, 2024. ISBN 978-3-031-53645-8. doi: 10.1007/978-3-031-53645-8_5. URLhttps://doi.org/10.1007/978-3-031-53645-8_5

  3. [3]

    The market for “lemons”: Quality uncertainty and the market mechanism

    George A Akerlof. The market for “lemons”: Quality uncertainty and the market mechanism. InUncertainty in economics, pages 235–251. Elsevier, 1978

  4. [4]

    A survey on task allocation and scheduling in robotic network systems.IEEE Internet of Things Journal, 12(2):1484–1508, 2024

    Saeid Alirezazadeh and Luís A Alexandre. A survey on task allocation and scheduling in robotic network systems.IEEE Internet of Things Journal, 12(2):1484–1508, 2024

  5. [5]

    Investigating cultural alignment of large language models

    Badr AlKhamissi, Muhammad ElNokrashy, Mai AlKhamissi, and Mona Diab. Investigating cultural alignment of large language models. InProceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (ACL), 2024

  6. [6]

    Self-resource allocation in multi-agent llm systems.arXiv preprint arXiv:2504.02051, 2025

    Alfonso Amayuelas, Jingbo Yang, Saaket Agashe, Ashwin Nagarajan, Antonis Antoniades, Xin Eric Wang, and William Wang. Self-resource allocation in multi-agent llm systems.arXiv preprint arXiv:2504.02051, 2025

  7. [7]

    Solomon E. Asch. Effects of group pressure on the modification and distortion of judgments. In Harold Guetzkow, editor,Groups, Leadership and Men: Research in Human Relations, pages 177–190. Carnegie Press, Pittsburgh, PA, 1951

  8. [8]

    Cam- bridge university press, 1995

    Frederic Charles Bartlett.Remembering: A study in experimental and social psychology. Cam- bridge university press, 1995

  9. [9]

    Solving conflict over common pool resources through delegation to a stakeholder.Journal of Economic Behavior & Organization, 227:106702, 2024

    Gabriel Bayle, Stefano Farolfi, Emmanuelle Lavaine, and Marc Willinger. Solving conflict over common pool resources through delegation to a stakeholder.Journal of Economic Behavior & Organization, 227:106702, 2024. ISSN 0167-2681. doi: https://doi.org/10.1016/j.jebo.2024. 106702. URLhttps://www.sciencedirect.com/science/article/pii/S0167268124003081

  10. [10]

    Conformity and social impact on ai agents, 2026

    Alessandro Bellina, Giordano De Marzo, and David Garcia. Conformity and social impact on ai agents, 2026. URLhttps://arxiv.org/abs/2601.05384

  11. [11]

    Information theoretic approach to detect collusion in multi-agent games

    Trevor Bonjour, Vaneet Aggarwal, and Bharat Bhargava. Information theoretic approach to detect collusion in multi-agent games. InUncertainty in Artificial Intelligence, pages 223–232. PMLR, 2022

  12. [12]

    Zana Buçinca, Maja Barbara Malaya, and Krzysztof Z. Gajos. To trust or to think: Cognitive forcing functions can reduce overreliance on ai in ai-assisted decision-making.Proceedings of the ACM on Human-Computer Interaction, 5(CSCW1), 2021. doi: 10.1145/3449287

  13. [13]

    Limiting the spread of misinformation in social networks

    Ceren Budak, Divyakant Agrawal, and Amr El Abbadi. Limiting the spread of misinformation in social networks. InProceedings of the 20th international conference on World wide web, pages 665–674, 2011

  14. [14]

    Llm collusion.arXiv preprint arXiv:2601.01279, 2026

    Shengyu Cao and Ming Hu. Llm collusion.arXiv preprint arXiv:2601.01279, 2026

  15. [15]

    Scheming ais: Will ais fake alignment during training in order to get power? arXiv preprint arXiv:2311.08379, 2023

    Joe Carlsmith. Scheming ais: Will ais fake alignment during training in order to get power? arXiv preprint arXiv:2311.08379, 2023. 49 Emergent Social Intelligence Risks in Generative Multi-Agent Systems

  16. [16]

    Why Do Multi-Agent LLM Systems Fail?

    Mert Cemri, Melissa Z Pan, Shuyi Yang, Lakshya A Agrawal, Bhavya Chopra, Rishabh Tiwari, Kurt Keutzer, Aditya Parameswaran, Dan Klein, Kannan Ramchandran, et al. Why do multi- agent llm systems fail?arXiv preprint arXiv:2503.13657, 2025

  17. [17]

    ChatEval: Towards Better LLM-based Evaluators through Multi-Agent Debate

    Chi-Min Chan, Weize Chen, Yusheng Su, Jianxuan Yu, Wei Xue, Shanghang Zhang, Jie Fu, and Zhiyuan Liu. Chateval: Towards better llm-based evaluators through multi-agent debate. arXiv preprint arXiv:2308.07201, 2023

  18. [18]

    Travelagent: An ai assistant for personalized travel planning.arXiv preprint arXiv:2409.08069, 2024

    Aili Chen, Xuyang Ge, Ziquan Fu, Yanghua Xiao, and Jiangjie Chen. Travelagent: An ai assistant for personalized travel planning.arXiv preprint arXiv:2409.08069, 2024

  19. [19]

    Ai4research: A survey of artificial intelligence for scientific research.arXiv preprint arXiv:2507.01903, 2025

    Qiguang Chen, Mingda Yang, Libo Qin, Jinhao Liu, Zheng Yan, Jiannan Guan, Dengyun Peng, Yiyan Ji, Hanjing Li, Mengkang Hu, et al. Ai4research: A survey of artificial intelligence for scientific research.arXiv preprint arXiv:2507.01903, 2025

  20. [20]

    Artificial intelligence and algorithmic price collusion in two-sided markets.arXiv preprint arXiv:2407.04088, 2024

    Cristian Chica, Yinglong Guo, and Gilad Lerman. Artificial intelligence and algorithmic price collusion in two-sided markets.arXiv preprint arXiv:2407.04088, 2024

  21. [21]

    Belief in authority: Impact of authority in multi-agent evaluation framework,

    Junhyuk Choi, Jeongyoun Kwon, Heeju Kim, Haeun Cho, Hayeong Jung, Sehee Min, and Bugeun Kim. Belief in authority: Impact of authority in multi-agent evaluation framework,

  22. [22]

    URLhttps://arxiv.org/abs/2601.04790

  23. [23]

    An empirical study of group conformity in multi-agent systems

    Min Choi, Keonwoo Kim, Sungwon Chae, and Sangyeop Baek. An empirical study of group conformity in multi-agent systems. In Wanxiang Che, Joyce Nabende, Ekaterina Shutova, and Mohammad Taher Pilehvar, editors,Findings of the Association for Computational Linguistics: ACL 2025, pages 5123–5139, Vienna, Austria, July 2025. Association for Computational Lingui...

  24. [24]

    Cialdini and Noah J

    Robert B. Cialdini and Noah J. Goldstein. Social influence: Compliance and conformity.Annual Review of Psychology, 55:591–621, 2004. doi: 10.1146/annurev.psych.55.090902.142015

  25. [25]

    Cambridge university press, 1996

    Herbert H Clark.Using language. Cambridge university press, 1996

  26. [26]

    Grounding in communication

    Herbert H Clark and Susan E Brennan. Grounding in communication. 1991

  27. [27]

    Clarkson g, jacobsen t, batcheller a (2007) information asymmetry and information sharing, government information quarterly 24:827–839.Government Information Quarterly, 10 2007

    Gavin Clarkson, Trond Jacobsen, and Archer Batcheller. Clarkson g, jacobsen t, batcheller a (2007) information asymmetry and information sharing, government information quarterly 24:827–839.Government Information Quarterly, 10 2007. doi: 10.1016/j.giq.2007.08.001

  28. [28]

    Market and non-market mechanisms for the optimal allocation of scarce resources.GamesandEconomicBehavior, 82:582–591, 2013

    Daniele Condorelli. Market and non-market mechanisms for the optimal allocation of scarce resources.GamesandEconomicBehavior, 82:582–591, 2013. ISSN0899-8256. doi: https://doi. org/10.1016/j.geb.2013.08.008. URL https://www.sciencedirect.com/science/article/ pii/S0899825613001255

  29. [29]

    Profilesinpatientsafety: Authoritygradientsinmedicalerror

    KarenS.CosbyandPatCroskerry. Profilesinpatientsafety: Authoritygradientsinmedicalerror. Academic Emergency Medicine, 11(12):1341–1345, 2004. doi: 10.1197/j.aem.2004.07.005

  30. [30]

    Thetraitors: Deceptionandtrustinmulti-agentlanguagemodelsimulations,

    PedroM.P.Curvo. Thetraitors: Deceptionandtrustinmulti-agentlanguagemodelsimulations,

  31. [31]

    URLhttps://arxiv.org/abs/2505.12923

  32. [32]

    Dirk De Clercq. The damage of deference: how personal and organizational factors transform deference to leader authority into unethical pro-organizational behavior.Management Research Review, 46(11):1637–1660, 04 2023. ISSN 2040-8269. doi: 10.1108/MRR-08-2022-0602. URLhttps://doi.org/10.1108/MRR-08-2022-0602. 50 Emergent Social Intelligence Risks in Gener...

  33. [33]

    Dietvorst, Joseph P

    Berkeley J. Dietvorst, Joseph P. Simmons, and Cade Massey. Algorithm aversion: People erroneously avoid algorithms after seeing them err.Journal of Experimental Psychology: General, 144(1):114–126, 2015. doi: 10.1037/xge0000033

  34. [34]

    Leibo, Usman Islam, Richard Willis, and Peter Sunehag

    Yali Du, Joel Z. Leibo, Usman Islam, Richard Willis, and Peter Sunehag. A review of cooperation in multi-agent learning, 2023. URLhttps://arxiv.org/abs/2312.05162

  35. [35]

    Misreporting in teams with individual decisionmaking: Theimpactofinformationandcommunication.JournalofEconomicBehavior& Organization, 209:509–532, 2023

    Eberhard Feess, Thomas Schilling, and Yuriy Timofeyev. Misreporting in teams with individual decisionmaking: Theimpactofinformationandcommunication.JournalofEconomicBehavior& Organization, 209:509–532, 2023. ISSN0167-2681. doi: https://doi.org/10.1016/j.jebo.2023. 03.027. URLhttps://www.sciencedirect.com/science/article/pii/S0167268123001014

  36. [36]

    A survey on large language model-based social agents in game-theoretic scenarios.Transactions on Machine Learning Research

    Xiachong Feng, Longxu Dou, Minzhi Li, Qinghao Wang, Yu Guo, Haochuan Wang, Chang Ma, and Lingpeng Kong. A survey on large language model-based social agents in game-theoretic scenarios.Transactions on Machine Learning Research

  37. [37]

    Taylor & Francis, 2010

    Fabrizio M Ferrari.Health and religious rituals in South Asia. Taylor & Francis, 2010

  38. [38]

    Foerster, Richard Y

    Jakob N. Foerster, Richard Y. Chen, Maruan Al-Shedivat, Shimon Whiteson, Pieter Abbeel, and Igor Mordatch. Learning with opponent-learning awareness. InProceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems (AAMAS), pages 122–

  39. [39]

    Is personal freedom a western value?American Journal of International Law, 91(4):593–627, 1997

    Thomas M Franck. Is personal freedom a western value?American Journal of International Law, 91(4):593–627, 1997

  40. [40]

    Thepowerofsocialinfluence: Areplicationandextensionof the asch experiment.PLOS ONE, 18:e0294325, 11 2023

    AxelFranzenandSebastianMader. Thepowerofsocialinfluence: Areplicationandextensionof the asch experiment.PLOS ONE, 18:e0294325, 11 2023. doi: 10.1371/journal.pone.0294325

  41. [41]

    The folk theorem in repeated games with discounting or with incomplete information.Econometrica, 54(3):533–554, 1986

    Drew Fudenberg and Eric Maskin. The folk theorem in repeated games with discounting or with incomplete information.Econometrica, 54(3):533–554, 1986

  42. [42]

    Honestllm: Toward an honest and helpful large language model.arXiv preprint arXiv:2406.00380, 2024

    Chujie Gao, Siyuan Wu, Yue Huang, Dongping Chen, Qihui Zhang, Zhengyan Fu, Yao Wan, Lichao Sun, and Xiangliang Zhang. Honestllm: Toward an honest and helpful large language model.arXiv preprint arXiv:2406.00380, 2024

  43. [43]

    Noncooperative collusion under imperfect price informa- tion.Econometrica: Journal of the Econometric Society, pages 87–100, 1984

    Edward J Green and Robert H Porter. Noncooperative collusion under imperfect price informa- tion.Econometrica: Journal of the Econometric Society, pages 87–100, 1984

  44. [44]

    Are your agents upward deceivers? arXiv preprint arXiv:2512.04864, 2025

    Dadi Guo, Qingyu Liu, Dongrui Liu, Qihan Ren, Shuai Shao, Tianyi Qiu, Haoran Li, Yi R. Fung, Zhongjie Ba, Juntao Dai, Jiaming Ji, Zhikai Chen, Jialing Tao, Yaodong Yang, Jing Shao, and Xia Hu. Are your agents upward deceivers?, 2025. URLhttps://arxiv.org/abs/2512.04864

  45. [45]

    Large language model based multi-agents: A survey of progress and challenges

    Taicheng Guo, Xiuying Chen, Yaqi Wang, Ruidi Chang, Shichao Pei, Nitesh V Chawla, Olaf Wiest, and Xiangliang Zhang. Large language model based multi-agents: A survey of progress and challenges. InIJCAI, 2024

  46. [46]

    Anti-languages.American anthropologist, 78(3):570– 584, 1976

    Michael Alexander Kirkwood Halliday. Anti-languages.American anthropologist, 78(3):570– 584, 1976

  47. [47]

    Multi-agent risks from advanced ai.arXiv preprint arXiv:2502.14143, 2025

    Lewis Hammond, Alan Chan, Jesse Clifton, Jason Hoelscher-Obermaier, Akbir Khan, Euan McLean, Chandler Smith, Wolfram Barfuss, Jakob Foerster, Tomáš Gavenčiak, et al. Multi-agent risks from advanced ai.arXiv preprint arXiv:2502.14143, 2025. 51 Emergent Social Intelligence Risks in Generative Multi-Agent Systems

  48. [48]

    The tragedy of the commons.Science, 162(3859):1243–1248, 1968

    Garrett Hardin. The tragedy of the commons.Science, 162(3859):1243–1248, 1968. ISSN 00368075, 10959203. URLhttp://www.jstor.org/stable/1724745

  49. [49]

    An optimal mechanism charging for priority in a queue

    Moshe Haviv and Eyal Winter. An optimal mechanism charging for priority in a queue. Operations Research Letters, 48(3):304–308, 2020. ISSN 0167-6377. doi: https://doi.org/ 10.1016/j.orl.2020.03.010. URL https://www.sciencedirect.com/science/article/pii/ S0167637720300456

  50. [50]

    The use of knowledge in society

    Friedrich August Hayek. The use of knowledge in society. InModern understandings of liberty and property, pages 27–38. Routledge, 2013

  51. [51]

    Coordination neglect: How lay theories of organizing complicate coordination in organizations.Research in organizational behavior, 22:153–191, 2000

    Chip Heath and Nancy Staudenmayer. Coordination neglect: How lay theories of organizing complicate coordination in organizations.Research in organizational behavior, 22:153–191, 2000

  52. [52]

    Helmreich, Ashleigh C

    Robert L. Helmreich, Ashleigh C. Merritt, and John A. Wilhelm. The evolution of crew resource management training in commercial aviation.The International Journal of Aviation Psychology, 9(1):19–32, 1999. doi: 10.1207/S15327108IJAP0901_2

  53. [53]

    other-play

    HengyuanHu, AdamLerer, AlexPeysakhovich, andJakobN.Foerster. “other-play”forzero-shot coordination. InProceedings of the 37th International Conference on Machine Learning (ICML), pages 4399–4410. PMLR, 2020

  54. [54]

    The sample complexity of online strategic decision making with information asymmetry and knowledge transportability

    Jiachen Hu, Rui Ai, Han Zhong, Xiaoyu Chen, Liwei Wang, Zhaoran Wang, and Zhuoran Yang. The sample complexity of online strategic decision making with information asymmetry and knowledge transportability. InForty-second International Conference on Machine Learning,

  55. [55]

    URLhttps://openreview.net/forum?id=e5yAhjSJ4j

  56. [56]

    Position: Towards a responsible llm-empowered multi-agent systems.arXiv preprint arXiv:2502.01714, 2025

    Jinwei Hu, Yi Dong, Shuang Ao, Zhuoyun Li, Boxuan Wang, Lokesh Singh, Guangliang Cheng, Sarvapali D Ramchurn, and Xiaowei Huang. Position: Towards a responsible llm-empowered multi-agent systems.arXiv preprint arXiv:2502.01714, 2025

  57. [57]

    On the resilience of llm-based multi-agent collaboration with faulty agents.arXiv preprint arXiv:2408.00989, 2024

    Jen-tse Huang, Jiaxu Zhou, Tailin Jin, Xuhui Zhou, Zixi Chen, Wenxuan Wang, Youliang Yuan, Michael R Lyu, and Maarten Sap. On the resilience of llm-based multi-agent collaboration with faulty agents.arXiv preprint arXiv:2408.00989, 2024

  58. [58]

    Metatool benchmark for large language models: Decidingwhethertousetoolsandwhichtouse

    Yue Huang, Jiawen Shi, Yuan Li, Chenrui Fan, Siyuan Wu, Qihui Zhang, Yixin Liu, Pan Zhou, Yao Wan, Neil Zhenqiang Gong, and Lichao Sun. Metatool benchmark for large language models: Decidingwhethertousetoolsandwhichtouse. InTheTwelfthInternationalConference on Learning Representations, 2024. URLhttps://openreview.net/forum?id=R0c2qtalgG

  59. [59]

    Chawla, Jian Pei, Jianfeng Gao, Michael Backes, Philip S

    Yue Huang, Chujie Gao, Siyuan Wu, Haoran Wang, Xiangqi Wang, Yujun Zhou, Yanbo Wang, Jiayi Ye, Jiawen Shi, Qihui Zhang, Yuan Li, Han Bao, Zhaoyi Liu, Tianrui Guan, Dongping Chen, Ruoxi Chen, Kehan Guo, Andy Zou, Bryan Hooi Kuen-Yew, Caiming Xiong, Elias Stengel- Eskin, Hongyang Zhang, Hongzhi Yin, Huan Zhang, Huaxiu Yao, Jaehong Yoon, Jieyu Zhang, Kai Shu...

  60. [60]

    52 Emergent Social Intelligence Risks in Generative Multi-Agent Systems

    URLhttps://arxiv.org/abs/2502.14296. 52 Emergent Social Intelligence Risks in Generative Multi-Agent Systems

  61. [61]

    Chemorch: Empower- ing llms with chemical intelligence via synthetic instructions.NeurIPS, 2025

    Yue Huang, Zhengzhe Jiang, Xiaonan Luo, Kehan Guo, Haomin Zhuang, Yujun Zhou, Zhengqing Yuan, Xiaoqi Sun, Jules Schleinitz, Yanbo Wang, et al. Chemorch: Empower- ing llms with chemical intelligence via synthetic instructions.NeurIPS, 2025

  62. [62]

    Building a foundational guardrail for general agentic systems via synthetic data

    Yue Huang, Hang Hua, Yujun Zhou, Pengcheng Jing, Manish Nagireddy, Inkit Padhi, Greta Dolcetti, Zhangchen Xu, Subhajit Chaudhury, Ambrish Rawat, Liubov Nedoshivina, Pin-Yu Chen, Prasanna Sattigeri, and Xiangliang Zhang. Building a foundational guardrail for general agentic systems via synthetic data. InThe Fourteenth International Conference on Learning R...

  63. [63]

    Deep research agents: A systematic examination and roadmap.arXiv preprint arXiv:2506.18096, 2025

    Yuxuan Huang, Yihang Chen, Haozheng Zhang, Kang Li, Huichi Zhou, Meng Fang, Linyi Yang, Xiaoguang Li, Lifeng Shang, Songcen Xu, et al. Deep research agents: A systematic examination and roadmap.arXiv preprint arXiv:2506.18096, 2025

  64. [64]

    Leibo, Matthew G

    Edward Hughes, Joel Z. Leibo, Matthew G. Phillips, Karl Tuyls, Edgar A. Duéñez-Guzmán, Antonio García Castañeda, Iain Dunning, Tina Zhu, Kevin R. McKee, Raphael Koster, Heather Roff, and Thore Graepel. Inequity aversion improves cooperation in intertemporal social dilemmas. InAdvances in Neural Information Processing Systems (NeurIPS), 2018

  65. [65]

    The economics of tacit collusion

    Marc Ivaldi, Bruno Jullien, Patrick Rey, Paul Seabright, and Jean Tirole. The economics of tacit collusion. 2003

  66. [66]

    Ortega, DJ Strouse, Joel Z

    Natasha Jaques, Angeliki Lazaridou, Edward Hughes, Caglar Gulcehre, Pedro A. Ortega, DJ Strouse, Joel Z. Leibo, and Nando de Freitas. Social influence as intrinsic motivation for multi-agent deep reinforcement learning. InProceedings of the 36th International Conference on Machine Learning (ICML), pages 3040–3049. PMLR, 2019

  67. [67]

    Measuring team hierarchy during high-stakes clinical decision making: Development and validation of a new behavioral observation method.Academic Medicine, 98(4):505–513, 12 2022

    Anna C Johansson, Bianca Manago, Jane Sell, and Cullen D Jackson. Measuring team hierarchy during high-stakes clinical decision making: Development and validation of a new behavioral observation method.Academic Medicine, 98(4):505–513, 12 2022. ISSN 1040-2446. doi: 10.1097/ACM.0000000000005133. URL https://doi.org/10.1097/ACM.0000000000005133

  68. [68]

    Flooding spread of manipulated knowledge in llm-based multi-agent communities.arXiv preprint arXiv:2407.07791, 2024

    Tianjie Ju, Yiting Wang, Xinbei Ma, Pengzhou Cheng, Haodong Zhao, Yulong Wang, Lifeng Liu, Jian Xie, Zhuosheng Zhang, and Gongshen Liu. Flooding spread of manipulated knowledge in llm-based multi-agent communities.arXiv preprint arXiv:2407.07791, 2024

  69. [69]

    Consecration rituals in south asia: An introduction

    István Keul. Consecration rituals in south asia: An introduction. InConsecration Rituals in South Asia, pages 1–16. Brill, 2017

  70. [70]

    Multiple LLM agents debate for equitable cultural alignment

    Dayeon Ki, Rachel Rudinger, Tianyi Zhou, and Marine Carpuat. Multiple LLM agents debate for equitable cultural alignment. In Wanxiang Che, Joyce Nabende, Ekaterina Shutova, and Mohammad Taher Pilehvar, editors,Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 24841–24877, Vienna, Austria...

  71. [71]

    Examining 35 years of individualism-collectivism research in asia: A meta- analysis.International Journal of Intercultural Relations, 100:101988, 2024

    Sang-Yeon Kim. Examining 35 years of individualism-collectivism research in asia: A meta- analysis.International Journal of Intercultural Relations, 100:101988, 2024

  72. [72]

    Many hands make light the work: The causes and consequences of social loafing.Journal of personality and social psychology, 37(6): 822, 1979

    Bibb Latané, Kipling Williams, and Stephen Harkins. Many hands make light the work: The causes and consequences of social loafing.Journal of personality and social psychology, 37(6): 822, 1979. 53 Emergent Social Intelligence Risks in Generative Multi-Agent Systems

  73. [73]

    The science of fake news.Science, 359(6380):1094–1096, 2018

    David MJ Lazer, Matthew A Baum, Yochai Benkler, Adam J Berinsky, Kelly M Greenhill, Filippo Menczer, Miriam J Metzger, Brendan Nyhan, Gordon Pennycook, David Rothschild, et al. The science of fake news.Science, 359(6380):1094–1096, 2018

  74. [74]

    Multi-agent reinforcement learning in sequential social dilemmas

    JoelZ.Leibo, ViniciusZambaldi, MarcLanctot, JanuszMarecki, andThoreGraepel. Multi-agent reinforcement learning in sequential social dilemmas. InProceedings of the 16th International Conference on Autonomous Agents and MultiAgent Systems (AAMAS), pages 464–473, 2017

  75. [75]

    Investor- bench: A benchmark for financial decision-making tasks with llm-based agent.arXiv preprint arXiv:2412.18174, 2024

    Haohang Li, Yupeng Cao, Yangyang Yu, Shashidhar Reddy Javaji, Zhiyang Deng, Yueru He, Yuechen Jiang, Zining Zhu, Koduvayur Subbalakshmi, Guojun Xiong, et al. Investor- bench: A benchmark for financial decision-making tasks with llm-based agent.arXiv preprint arXiv:2412.18174, 2024

  76. [76]

    A survey on the honesty of large language models.arXiv preprint arXiv:2409.18786, 2024

    Siheng Li, Cheng Yang, Taiqiang Wu, Chufan Shi, Yuji Zhang, Xinyu Zhu, Zesen Cheng, Deng Cai, Mo Yu, Lemao Liu, et al. A survey on the honesty of large language models.arXiv preprint arXiv:2409.18786, 2024

  77. [77]

    Frontier ai risk management framework in practice: A risk analysis technical report v1.5, 2026

    Dongrui Liu, Yi Yu, Jie Zhang, Guanxu Chen, Qihao Lin, Hanxi Zhu, Lige Huang, Yijin Zhou, Peng Wang, Shuai Shao, Boxuan Zhang, Zicheng Liu, Jingwei Sun, Yu Li, Yuejin Xie, Jiaxuan Guo, Jia Xu, Chaochao Lu, Bowen Zhou, Xia Hu, and Jing Shao. Frontier ai risk management framework in practice: A risk analysis technical report v1.5, 2026. URL https://arxiv.or...

  78. [78]

    Columbia University, 2020

    Shi Liu.Harm in harmony: A socioecological perspective on East Asian collectivism. Columbia University, 2020

  79. [79]

    Autonomousagentsforcollaborativetaskunderinformation asymmetry.Advances in Neural Information Processing Systems, 37:2734–2765, 2024

    Wei Liu, Chenxi Wang, Yifei Wang, Zihao Xie, Rennai Qiu, Yufan Dang, Zhuoyun Du, Weize Chen, ChengYang, andChenQian. Autonomousagentsforcollaborativetaskunderinformation asymmetry.Advances in Neural Information Processing Systems, 37:2734–2765, 2024

  80. [80]

    Logg, Julia A

    Jennifer M. Logg, Julia A. Minson, and Don A. Moore. Algorithm appreciation: People prefer algorithmic to human judgment.Organizational Behavior and Human Decision Processes, 151: 90–103, 2019. doi: 10.1016/j.obhdp.2018.12.005

Showing first 80 references.