pith. machine review for the scientific record. sign in

arxiv: 2605.13077 · v1 · submitted 2026-05-13 · 💻 cs.MA · cs.AI

Recognition: no theorem link

Counterfactual Reasoning for Causal Responsibility Attribution in Probabilistic Multi-Agent Systems

Chunyan Mu, Muhammad Najib

Pith reviewed 2026-05-14 02:10 UTC · model grok-4.3

classification 💻 cs.MA cs.AI
keywords responsibility attributioncounterfactual reasoningmulti-agent systemsShapley valuestochastic gamesNash equilibriumcausal responsibilityverification
0
0 comments X

The pith

Shapley values allocate responsibility fairly among agents in stochastic multi-agent games by quantifying retrospective counterfactual impact.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper models multi-agent systems as concurrent stochastic games in which agents select strategies that produce probabilistic outcomes. It defines retrospective counterfactual responsibility as the change in outcome probability that can be attributed to an individual agent's strategy when compared to what would have happened if that agent had chosen differently. Responsibility is then distributed using the Shapley value, which the authors prove satisfies fairness and consistency. The resulting framework supports verification of responsibility levels and the computation of Nash equilibria in which agents balance their assigned responsibility against expected rewards.

Core claim

In concurrent stochastic multi-player games, retrospective counterfactual responsibility quantifies an agent's accountability for an outcome under a given strategy profile by comparing the actual outcome probability to the probability that would have obtained if the agent had unilaterally deviated. Allocating this responsibility via the Shapley value produces a distribution that is fair, in that agents making identical marginal contributions receive identical shares, and consistent, in that the allocation remains stable when the set of agents changes.

What carries the argument

Retrospective counterfactual responsibility, which measures the marginal effect of an agent's strategy deviation on outcome probabilities, allocated via the Shapley value that averages each agent's contribution across all coalitions.

If this is right

  • Responsibility levels can be formally verified within the game model.
  • Agents can reach stable strategy profiles by trading off responsibility against expected reward at Nash equilibrium.
  • The allocation method applies uniformly to any probabilistic outcome without requiring case-by-case adjustments.
  • Fairness and consistency of the allocation follow directly from the properties of the Shapley value.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The approach could be used to design incentive mechanisms that penalize high-responsibility strategies in safety-critical systems.
  • It suggests a way to compare responsibility attributions across different solution concepts beyond Nash equilibrium.
  • The framework may help formalise accountability in joint AI decision-making where outcomes are probabilistic.

Load-bearing premise

That responsibility in probabilistic multi-agent systems is fully captured by backward counterfactual comparisons inside a concurrent stochastic game model, with no need for further domain-specific adjustments.

What would settle it

A concrete stochastic game in which the Shapley allocation of retrospective counterfactual responsibility produces shares that contradict clear intuitive judgments of accountability for the same outcome.

Figures

Figures reproduced from arXiv: 2605.13077 by Chunyan Mu, Muhammad Najib.

Figure 1
Figure 1. Figure 1: CSG model from the running example. to it that” an outcome occurs, whereas our framework evaluates responsibility through counterfactual dependence—what would happen were an agent to act differently—and seeks a fair quantitative attribution across agents. 2 Preliminaries 2.1 Concurrent stochastic game Definition 1. A concurrent stochastic multi-player game (CSG) is a tuple G = (Ag, S, s0 ,(Acti)i∈Ag, δ, Ap… view at source ↗
read the original abstract

Responsibility allocation -- determining the extent to which agents are accountable for outcomes -- is a fundamental challenge in the design and analysis of multi-agent systems. In this work, we model such systems as concurrent stochastic multi-player games and introduce a notion of retrospective (backward) counterfactual responsibility, which quantifies an agent's accountability for outcomes resulting from a given strategy profile. To allocate responsibility among agents, we utilise the Shapley value and formally show that this method satisfies key desirable properties, including fairness and consistency. Building on this foundation, we propose a formal framework that supports both verification and strategic reasoning in responsibility-aware multi-agent systems. Furthermore, by adopting Nash equilibrium as the solution concept, we demonstrate how to compute stable strategy profiles in which agents trade off responsibility against expected reward.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 1 minor

Summary. The paper models multi-agent systems as concurrent stochastic multi-player games, defines a retrospective counterfactual responsibility measure for a given strategy profile, applies the Shapley value to allocate responsibility, and claims to formally prove that this allocation satisfies fairness and consistency. It then outlines a verification and strategic-reasoning framework and shows how Nash equilibria can be computed that trade off responsibility against expected reward.

Significance. If the formal claims hold, the work supplies a principled, axiomatically grounded method for responsibility attribution in probabilistic MAS that could support accountability analysis in autonomous systems and verification tools. The reliance on the standard Shapley value (rather than ad-hoc weights) and the explicit use of Nash equilibrium for stable profiles are positive features.

major comments (2)
  1. [section introducing the responsibility measure and Shapley application] The definition of the characteristic function v(S) used for the Shapley value is not made explicit with respect to the behavior of agents outside coalition S. In a concurrent stochastic game, computing the counterfactual outcome for S requires a convention for the strategies of the remaining agents (original profile, equilibrium, or worst-case); without a fixed convention the marginal contributions are not uniquely determined, so the claimed fairness and consistency properties may hold only under additional implicit assumptions rather than in general.
  2. [proofs of properties] The abstract states that formal proofs of fairness and consistency are supplied, yet the manuscript provides no derivation details, small examples, or error analysis for the retrospective counterfactual responsibility measure. This absence makes it impossible to verify whether the properties survive the stochastic transition structure and the retrospective (backward) counterfactual construction.
minor comments (1)
  1. The abstract would be clearer if it briefly indicated the exact axioms or properties proven (e.g., efficiency, symmetry, dummy-player) rather than using the generic phrase 'key desirable properties'.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive review and for recognizing the potential significance of our framework for responsibility attribution in probabilistic multi-agent systems. We address each major comment below and will revise the manuscript to enhance clarity and accessibility.

read point-by-point responses
  1. Referee: [section introducing the responsibility measure and Shapley application] The definition of the characteristic function v(S) used for the Shapley value is not made explicit with respect to the behavior of agents outside coalition S. In a concurrent stochastic game, computing the counterfactual outcome for S requires a convention for the strategies of the remaining agents (original profile, equilibrium, or worst-case); without a fixed convention the marginal contributions are not uniquely determined, so the claimed fairness and consistency properties may hold only under additional implicit assumptions rather than in general.

    Authors: We thank the referee for highlighting this point. Our retrospective counterfactual responsibility measure defines v(S) by fixing the strategies of agents outside S to their actions in the given strategy profile. This convention is required by the retrospective (backward) nature of the definition, which evaluates deviations from the realized play rather than from an equilibrium or worst-case assumption. The formal definition in Section 3 states this explicitly, which ensures the marginal contributions are uniquely determined. We will add a dedicated clarifying paragraph and a small worked example in the revised manuscript to make the convention and its implications for stochastic transitions fully explicit. revision: yes

  2. Referee: [proofs of properties] The abstract states that formal proofs of fairness and consistency are supplied, yet the manuscript provides no derivation details, small examples, or error analysis for the retrospective counterfactual responsibility measure. This absence makes it impossible to verify whether the properties survive the stochastic transition structure and the retrospective (backward) counterfactual construction.

    Authors: The formal proofs of fairness and consistency appear in Appendix A, with derivations that account for the probabilistic transition function and the backward counterfactual construction. A small illustrative example demonstrating the properties is given in Section 4. To address the concern about accessibility, we will insert a high-level proof sketch into the main text (near the statement of the properties), expand the discussion of how the axioms hold under stochasticity, and add a brief error-analysis paragraph for the finite-horizon case in the revised version. revision: yes

Circularity Check

0 steps flagged

No circularity: standard Shapley value applied to independently defined responsibility measure

full rationale

The paper defines retrospective counterfactual responsibility directly from the concurrent stochastic game model and strategy profile. It then applies the standard Shapley value to this measure and proves the usual fairness/consistency axioms hold for the resulting allocation. This is a direct application of a known operator to a new set function v(S); the axioms follow from the definition of Shapley value rather than from any self-referential equation, fitted parameter, or self-citation chain. No step reduces the claimed result to its own inputs by construction.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 1 invented entities

The central claim rests on the domain assumption that multi-agent systems are accurately represented as concurrent stochastic multi-player games and on the standard mathematical properties of the Shapley value when applied to the newly defined responsibility measure.

axioms (2)
  • domain assumption Multi-agent systems can be modeled as concurrent stochastic multi-player games
    Explicit modeling choice stated in the abstract as the foundation for the responsibility definition.
  • standard math Shapley value satisfies fairness and consistency when applied to retrospective counterfactual responsibility
    Claimed formal result in the abstract; relies on established properties of the Shapley value.
invented entities (1)
  • retrospective counterfactual responsibility no independent evidence
    purpose: Quantifies an agent's accountability for outcomes resulting from a given strategy profile using backward counterfactuals
    New notion introduced to capture responsibility in the stochastic game setting.

pith-pipeline@v0.9.0 · 5419 in / 1323 out tokens · 54404 ms · 2026-05-14T02:10:25.733614+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

54 extracted references · 54 canonical work pages

  1. [1]

    In: AAMAS

    Abarca, A.I.R., Broersen, J.M.: A stit logic of responsibility. In: AAMAS. pp. 1717–1719 (2022)

  2. [2]

    In: Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems

    ˚Agotnes, T., van der Hoek, W., Wooldridge, M.: Normative system games. In: Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems. pp. 1–8 (2007)

  3. [3]

    MIT Press (2008)

    Baier, C., Katoen, J.: Principles of model checking. MIT Press (2008)

  4. [4]

    In: IJCAI

    Baier, C., Funke, F., Majumdar, R.: A game-theoretic account of responsibility allocation. In: IJCAI. pp. 1773–1779. ijcai.org (2021)

  5. [5]

    In: AAAI

    Baier, C., Funke, F., Majumdar, R.: Responsibility attribution in parameterized markovian models. In: AAAI. pp. 11734–11743. AAAI Press (2021) Counterfactual Reasoning for CR Attribution in Probabilistic MASs 21

  6. [6]

    Logic in high definition: Trends in logical semantics pp

    Baltag, A., Canavotto, I., Smets, S.: Causal agency and responsibility: a refinement of stit logic. Logic in high definition: Trends in logical semantics pp. 149–176 (2021)

  7. [7]

    In: Baier, C., Tinelli, C

    Basset, N., Kwiatkowska, M.Z., Topcu, U., Wiltsche, C.: Strategy synthesis for stochastic games with multiple long-run objectives. In: Baier, C., Tinelli, C. (eds.) TACAS. Lecture Notes in Computer Science, vol. 9035, pp. 256–271. Springer (2015)

  8. [8]

    Dagstuhl Reports 14, 75–91 (2024)

    Belle, V., Chockler, H., Vallor, S., Varshney, K.R., Vennekens, J., Beckers, S.: Trustworthiness and responsibility in ai-causality, learning, and verification (dagstuhl seminar 24121). Dagstuhl Reports 14, 75–91 (2024)

  9. [9]

    Studia Logica 51, 463–484 (1992)

    Belnap, N., Perloff, M.: The way of the agent. Studia Logica 51, 463–484 (1992)

  10. [10]

    Mind 121, 601–634 (2012)

    Braham, M., van Hees, M.: An anatomy of moral responsibility. Mind 121, 601–634 (2012)

  11. [11]

    International Journal of Game Theory 30, 309–319 (2002)

    van den Brink, R.: An axiomatization of the shapley value using a fairness property. International Journal of Game Theory 30, 309–319 (2002)

  12. [12]

    In: Proceedings of the twentieth annual ACM symposium on Theory of computing

    Canny, J.: Some algebraic and geometric computations in pspace. In: Proceedings of the twentieth annual ACM symposium on Theory of computing. pp. 460–467 (1988)

  13. [13]

    Reflections on artificial intelligence for humanity pp

    Chatila, R., Dignum, V., Fisher, M., Giannotti, F., Morik, K., Russell, S., Yeung, K.: Trustworthy ai. Reflections on artificial intelligence for humanity pp. 13–39 (2021)

  14. [14]

    Journal of Computer and System Sciences 78(2), 394–413 (2012)

    Chatterjee, K., Henzinger, T.A.: A survey of stochasticω-regular games. Journal of Computer and System Sciences 78(2), 394–413 (2012)

  15. [15]

    In: FSKD

    Chen, T., Lu, J.: Probabilistic alternating-time temporal logic and model checking algorithm. In: FSKD. pp. 35–39. IEEE Computer Society (2007)

  16. [16]

    Chockler, H., Halpern, J.Y.: Responsibility and blame: A structural-model ap- proach. J. Artif. Intell. Res. 22, 93–115 (2004)

  17. [17]

    Games and Economic Be- havior 1(2), 119–130 (1989)

    Chun, Y.: A new axiomatization of the shapley value. Games and Economic Be- havior 1(2), 119–130 (1989)

  18. [18]

    In: Logic, Rationality, and Interaction: Second International Workshop, LORI 2009, Chongqing, China, October 8-11, 2009

    Ciuni, R., Mastop, R.: Attributing distributed responsibility in stit logic. In: Logic, Rationality, and Interaction: Second International Workshop, LORI 2009, Chongqing, China, October 8-11, 2009. Proceedings 2. pp. 66–75. Springer (2009)

  19. [19]

    Cristau, J., David, C., Horn, F.: How do we remember the past in randomised strategies? Proceedings of GandALF 2010 25 (2010)

  20. [20]

    In: Proceedings of the Thirty-Fourth International Joint Conference on Artificial Intelligence, IJCAI (2025)

    De Giacomo, G., Lorini, E., Parker, T., Parretti, G.: Responsibility anticipation and attribution in ltlf. In: Proceedings of the Thirty-Fourth International Joint Conference on Artificial Intelligence, IJCAI (2025)

  21. [21]

    Logic Journal of the IGPL 18(1), 99–117 (2010)

    De Lima, T., Royakkers, L., Dignum, F.: A logic for reasoning about responsibility. Logic Journal of the IGPL 18(1), 99–117 (2010)

  22. [22]

    The oxford handbook of ethics of AI 4698, 215 (2020)

    Dignum, V.: Responsibility and artificial intelligence. The oxford handbook of ethics of AI 4698, 215 (2020)

  23. [23]

    In: Moral responsi- bility and alternative possibilities, pp

    Frankfurt, H.: Alternate possibilities and moral responsibility. In: Moral responsi- bility and alternative possibilities, pp. 17–25. Routledge (2018)

  24. [24]

    In: AAAI

    Friedenberg, M., Halpern, J.Y.: Blameworthiness in multi-agent settings. In: AAAI. pp. 525–532 (2019)

  25. [25]

    In: Proceedings of the annual meeting of the cognitive science society

    Gerstenberg, T., Ejova, A., Lagnado, D.: Blame the skilled. In: Proceedings of the annual meeting of the cognitive science society. vol. 33 (2011)

  26. [26]

    In: AAAI

    Halpern, J.Y., Kleiman-Weiner, M.: Towards formal definitions of blameworthi- ness, intention, and moral responsibility. In: AAAI. pp. 1853–1860 (2018)

  27. [27]

    part i: Causes

    Halpern, J.Y., Pearl, J.: Causes and explanations: A structural-model approach. part i: Causes. The British journal for the philosophy of science (2005) 22 Mu, Najib

  28. [28]

    Econometrica: Journal of the Econometric Society pp

    Hart, S., Mas-Colell, A.: Potential, value, and consistency. Econometrica: Journal of the Econometric Society pp. 589–614 (1989)

  29. [29]

    Journal of Artificial Intelligence Research 73, 173–208 (2022)

    Icarte, R.T., Klassen, T.Q., Valenzano, R., McIlraith, S.A.: Reward machines: Ex- ploiting reward function structure in reinforcement learning. Journal of Artificial Intelligence Research 73, 173–208 (2022)

  30. [30]

    ACM computing surveys (CSUR) 55(2), 1–38 (2022)

    Kaur, D., Uslu, S., Rittichier, K.J., Durresi, A.: Trustworthy artificial intelligence: a review. ACM computing surveys (CSUR) 55(2), 1–38 (2022)

  31. [31]

    In: International Symposium on Formal Methods

    Kobayashi, T., Bondu, M., Ishikawa, F.: Formal modelling of safety architecture for responsibility-aware autonomous vehicle via event-b refinement. In: International Symposium on Formal Methods. pp. 533–549. Springer (2023)

  32. [32]

    In: Quantitative Evaluation of Systems: 15th Interna- tional Conference, QEST 2018, Beijing, China, September 4-7, 2018, Proceedings

    Kwiatkowska, M., Norman, G., Parker, D., Santos, G.: Automated verification of concurrent stochastic games. In: Quantitative Evaluation of Systems: 15th Interna- tional Conference, QEST 2018, Beijing, China, September 4-7, 2018, Proceedings

  33. [33]

    pp. 223–239. Springer (2018)

  34. [34]

    In: International Symposium on Formal Methods

    Kwiatkowska, M., Norman, G., Parker, D., Santos, G.: Equilibria-based probabilis- tic model checking for concurrent stochastic games. In: International Symposium on Formal Methods. pp. 298–315. Springer (2019)

  35. [35]

    Formal Methods in System Design 58(1), 188–250 (2021)

    Kwiatkowska, M., Norman, G., Parker, D., Santos, G.: Automatic verification of concurrent stochastic systems. Formal Methods in System Design 58(1), 188–250 (2021)

  36. [36]

    John Wiley & Sons (2013)

    Lewis, D.: Counterfactuals. John Wiley & Sons (2013)

  37. [37]

    ACM Computing Surveys 55(9), 1–46 (2023)

    Li, B., Qi, P., Liu, B., Di, S., Liu, J., Pei, J., Yi, J., Zhou, B.: Trustworthy ai: From principles to practices. ACM Computing Surveys 55(9), 1–46 (2023)

  38. [38]

    In: Proceedings of the AAAI Conference on Artificial Intelligence

    Mu, C., Najib, M., Oren, N.: Responsibility-aware strategic reasoning in proba- bilistic multi-agent systems. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 39, pp. 23258–23266 (2025)

  39. [39]

    In: IJCAI

    Naumov, P., Tao, J.: Two forms of responsibility in strategic games. In: IJCAI. pp. 1989–1995. ijcai.org (2021)

  40. [40]

    In: Proceedings of the AAAI Conference on Artificial Intelligence

    Naumov, P., Tao, J.: Blameworthiness in strategic games. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 33, pp. 3011–3018 (2019)

  41. [41]

    The MIT Press (1994)

    Osborne, M.J., Rubinstein, A.: A course in game theory. The MIT Press (1994)

  42. [42]

    In: ECAI

    Parker, T., Grandi, U., Lorini, E.: Anticipating responsibility in multiagent plan- ning. In: ECAI. vol. 372, pp. 1859–1866 (2023)

  43. [43]

    Cambridge university press (2009)

    Pearl, J.: Causality. Cambridge university press (2009)

  44. [44]

    In: AAMAS

    Perelli, G.: Enforcing equilibria in multi-agent systems. In: AAMAS. vol. 1, pp. 188–196. International Foundation for Autonomous Agents and Multiagent Sys- tems (IFAAMAS) (2019)

  45. [45]

    In: Kuhn, H.W., Tucker, A.W

    Shapley, L.S.: A value for n-person games. In: Kuhn, H.W., Tucker, A.W. (eds.) Contributions to the Theory of Games II, pp. 307–317. Princeton University Press, Princeton (1953)

  46. [46]

    Journal of Philosophical Logic pp

    Shi, Q., Naumov, P.: Responsibility in multi-step decision schemes. Journal of Philosophical Logic pp. 1–39 (2025)

  47. [47]

    Cambridge University Press (2009)

    Shoham, Y., Leyton-Brown, K.: Multiagent Systems - Algorithmic, Game- Theoretic, and Logical Foundations. Cambridge University Press (2009)

  48. [48]

    In: 32nd EACSL Annual Conference on Computer Science Logic (CSL 2024)

    Stan, D., Najib, M., Lin, A.W., Abdulla, P.A.: Concurrent stochastic lossy channel games. In: 32nd EACSL Annual Conference on Computer Science Logic (CSL 2024). pp. 46–1. Schloss Dagstuhl-Leibniz-Zentrum f¨ ur Informatik (2024)

  49. [49]

    Routledge (2017)

    Widerker, D.: Moral responsibility and alternative possibilities: Essays on the im- portance of alternative possibilities. Routledge (2017)

  50. [50]

    Wooldridge, M.: Does game theory work? IEEE Intelligent Systems 27(6), 76–80 (2012) Counterfactual Reasoning for CR Attribution in Probabilistic MASs 23

  51. [51]

    Wright, R.W.: Causation in tort law. Calif. L. Rev. 73, 1735 (1985)

  52. [52]

    In: AAMAS

    Yazdanpanah, V., Dastani, M., Jamroga, W., Alechina, N., Logan, B.: Strategic responsibility under imperfect information. In: AAMAS. pp. 592–600 (2019)

  53. [53]

    Ai & Society 38(4), 1453–1464 (2023)

    Yazdanpanah, V., Gerding, E.H., Stein, S., Dastani, M., Jonker, C.M., Norman, T.J., Ramchurn, S.D.: Reasoning about responsibility in autonomous systems: chal- lenges and opportunities. Ai & Society 38(4), 1453–1464 (2023)

  54. [54]

    International Journal of Game Theory 14(2), 65–72 (1985)

    Young, H.P.: Monotonic solutions of cooperative games. International Journal of Game Theory 14(2), 65–72 (1985)