pith. machine review for the scientific record. sign in

arxiv: 2605.12536 · v1 · submitted 2026-05-03 · 🧬 q-bio.NC · cs.AI· cs.IT· math.IT

Recognition: unknown

Information as Maximum-Caliber Deviation: A bridge between Integrated Information Theory and the Free Energy Principle

Alexander Kearney

Authors on Pith no claims yet

Pith reviewed 2026-05-14 21:07 UTC · model grok-4.3

classification 🧬 q-bio.NC cs.AIcs.ITmath.IT
keywords integrated information theoryfree energy principlemaximum caliberprediction errorcause-effect repertoiresactive inferencevariational principlesconsciousness modeling
0
0 comments X

The pith

Information is the deviation of realized dynamics from a constrained maximum-caliber path ensemble, from which IIT 3.0's cause-effect repertoires emerge via variational principles.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper proposes defining information as the deviation ψ of actual dynamics from what a constrained maximum-caliber path ensemble would produce over a finite time horizon. Under this definition, the cause-effect repertoires and integrated information measures of IIT 3.0 arise directly from maximum-caliber variational principles, re-deriving the theory's phenomenological calculus from constrained entropy maximization. The same deviation is shown to equal prediction error in predictive coding models for Markov chains under the central limit theorem and for Ising models under large deviations theory. A sympathetic reader would care because the construction supplies a precise mathematical mapping between integrated information theory and the free energy principle's active inference framework, opening routes to extend both to new dynamical systems.

Core claim

The central claim is that information can be defined as the deviation ψ of realized dynamics from a constrained maximum-caliber path ensemble, from which each of the cause/effect repertoires central to IIT 3.0 emerge directly from MaxCal variational principles. This re-derives IIT's phenomenological calculus from constrained entropy-maximization, supplies a theoretical bridge to active inference which is mathematically dual under Langevin dynamics, and shows that ψ equals prediction error under the central limit theorem for Markov chains and large deviations theory for Ising models.

What carries the argument

The deviation ψ of realized dynamics from a constrained maximum-caliber path ensemble, which acts as the definition of information and generates IIT's cause-effect structures from variational entropy maximization.

If this is right

  • IIT 3.0's integrated information measures can be obtained from constrained entropy maximization alone.
  • The information measure ψ is equivalent to prediction error in predictive coding models for Markov chains and Ising models.
  • The framework supplies a principled route for extending IIT to dynamical regimes beyond its current scope.
  • It provides a rationale for studying convergence among FEP, IIT, and thermodynamic accounts of cognition such as fluctuation-dissipation violations.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The unification may predict that integrated information follows a hill-shaped trajectory during adaptation to sensory inputs in neural systems.
  • The approach could be tested by comparing ψ values computed from observed trajectories against empirical measures of information integration in biological preparations.
  • Consciousness-related quantities might be re-interpreted as measurable deviations from maximum-entropy path ensembles in physical systems.

Load-bearing premise

The proposed definition of information as maximum-caliber deviation is sufficient to recover the full set of IIT 3.0 cause-effect repertoires and measures without additional unstated constraints.

What would settle it

A direct computation on a small Markov chain or Ising model in which the cause-effect repertoires obtained from the maximum-caliber deviation do not match the repertoires produced by standard IIT 3.0 procedures.

Figures

Figures reproduced from arXiv: 2605.12536 by Alexander Kearney.

Figure 2.1
Figure 2.1. Figure 2.1: Our system X is partitioned into subsystems (V ,V⊥) at t = −1 and (Y ,Y⊥) at t = 0. We can also think of any part of our system as a “true insider” to its evolution, while thinking of any other part as extrinsic to the conscious part (assuming there is one) of X. Our task is to identify which of these perspectives matters most, and in what way, to the system X. To formalise this perspective, we define on… view at source ↗
Figure 2.2
Figure 2.2. Figure 2.2: Applying the cause function ζV ,Y : V −1 ⊥ = v⊥ becomes a background condition, Y 0 ⊥ is discarded, and the remaining subsystems are partitioned. 17 [PITH_FULL_IMAGE:figures/full_fig_p017_2_2.png] view at source ↗
Figure 2.3
Figure 2.3. Figure 2.3: Applying the effect function ζY ,V : Y 0 ⊥ = y⊥ becomes a background condition, V 1 ⊥ is discarded, and the remaining subsystems are partitioned. Definition 5 (Unconstrained Effect Repertoire). The unconstrained effect repertoire of a subsystem V with respect to subsystem Y (for a system X which has been observed as X0 = x 0 ), is the probability distribution of V 1 given that Y 0 is uniformly distribute… view at source ↗
Figure 4.1
Figure 4.1. Figure 4.1: Here we have a transition network which shows non-independent relationships between the inputs [PITH_FULL_IMAGE:figures/full_fig_p036_4_1.png] view at source ↗
Figure 4.2
Figure 4.2. Figure 4.2: Input nodes Y t ⊥ = (Xt 1 , Xt 4 ) have been fixed to background conditions y t ⊥, and output nodes Y t+1 ⊥ = (X t+1 1 , Xt+1 4 ) have been marginalized over, to select the subsidiary network GY over Y = (X2, X3). On the left we have the unconstrained case in which entropy across Y t ⊔ Y t+1 has been maximized. On the right hand side we have maximized entropy subject to the conditions Y t+1 = y t+1, retr… view at source ↗
Figure 4.3
Figure 4.3. Figure 4.3: In this case, Y = (X2, X3) and our partition is P = {Xt 2 ⊔ X t+1 2 , Xt 3 ⊔ X t+1 3 }. The subsidiary network G y t Z|Y conditions on Xt 3 = x t 3 and applies a MaxCal path ensemble over the network. In this case, Xt 2 should have a uniform distribution over its state space Ω2 while the value x t+1 2 of X t+1 2 should be fixed by the background conditions. For our network G y t Z⊥|Y we condition on Xt 2… view at source ↗
Figure 5.1
Figure 5.1. Figure 5.1: We understand our generative model as a graph with a series of biases and weights which combine to [PITH_FULL_IMAGE:figures/full_fig_p055_5_1.png] view at source ↗
Figure 6.1
Figure 6.1. Figure 6.1: Here, we have a graph G over the state space Ω = {1, 2, 3, 4, 5}. The adjacency matrix A represents connections between nodes. The degree d(x) represents the number of edges a node x belongs to, e.g. d(5) = 4. PGRW =   0 0 0 0 1 0 0 1 3 1 3 1 3 0 1 3 0 1 3 1 3 0 1 3 1 3 0 1 3 1 5 1 5 1 5 1 5 1 5   , πGRW = 1 15   1 3 3 3 5   Here, each x will have a conditional entropy value l… view at source ↗
Figure 6
Figure 6. Figure 6: figure 6.1 to illustrate, we may express the transition probabilities and stationary distribution of [PITH_FULL_IMAGE:figures/full_fig_p065_6.png] view at source ↗
Figure 6.2
Figure 6.2. Figure 6.2: Note that the colors are scaled independently across heatmaps. [PITH_FULL_IMAGE:figures/full_fig_p068_6_2.png] view at source ↗
Figure 6
Figure 6. Figure 6: displays heatmaps for [PITH_FULL_IMAGE:figures/full_fig_p068_6.png] view at source ↗
Figure 6
Figure 6. Figure 6: shows the skew and mean values of [PITH_FULL_IMAGE:figures/full_fig_p069_6.png] view at source ↗
Figure 7.1
Figure 7.1. Figure 7.1: Pipes and mazes constrain maximal path entropy by limiting movement in physical space. Water will [PITH_FULL_IMAGE:figures/full_fig_p074_7_1.png] view at source ↗
Figure 7
Figure 7. Figure 7: figure 7.1, is water flowing through a pipe. Left to its own devices, water would spread chaotically in all directions. [PITH_FULL_IMAGE:figures/full_fig_p074_7.png] view at source ↗
read the original abstract

The Free Energy Principle (FEP) is a leading framework for mathematically modeling self-organization and learning, while Integrated Information Theory (IIT) is a computational ontology of consciousness oriented around irreducible cause and effect. While conceptual unifications have been proposed and appear to be supported by empirical findings, the absence of a rigorous mathematical mapping places upper bounds on their precision and testability. This work proposes that information can be defined as the deviation $\psi$ of realized dynamics from a constrained maximum-caliber (MaxCal) path ensemble over a finite time horizon. Under this definition, each of the cause/effect repertoires central to IIT 3.0 emerge directly from MaxCal variational principles, allowing IIT's phenomenological calculus to be re-derived from constrained entropy-maximization (CMEP). This framework supplies a theoretical bridge to active inference, which is mathematically dual to CMEP under Langevin dynamics, and offers a principled route for extending IIT to new dynamical regimes. When the approach is applied under the Central Limit Theorem (CLT) for Markov chains and via large deviations theory (LDT) to Ising models, information $\psi$ is shown to be equivalent to prediction error under accompanying predictive coding models. This may hold relevance to the ``hill-shaped trajectory'' of $\Phi$ observed in neuronal cultures adapting to sensory inputs. Together, these results provide a physically and mathematically grounded rationale for studying the convergence of FEP, IIT, and thermodynamic frameworks of cognition such as recent work grounding consciousness in violations of the Fluctuation-Dissipation Theorem (FDT).

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 0 minor

Summary. The paper proposes defining information as the deviation ψ of realized dynamics from a constrained maximum-caliber (MaxCal) path ensemble over a finite time horizon. Under this definition, each of the cause/effect repertoires central to IIT 3.0 emerge directly from MaxCal variational principles, re-deriving IIT's phenomenological calculus from constrained entropy-maximization (CMEP). The framework bridges to active inference (dual to CMEP under Langevin dynamics) and shows ψ equivalent to prediction error under CLT for Markov chains and LDT for Ising models, with potential relevance to the hill-shaped trajectory of Φ in neuronal cultures.

Significance. If the claimed mappings hold without auxiliary constraints, the work supplies a physically grounded unification of IIT and FEP, grounding consciousness measures in thermodynamic path ensembles and offering a route to extend IIT beyond current regimes. The special-case equivalences to prediction error and the link to FDT violations are notable strengths if supported by explicit derivations.

major comments (2)
  1. [Abstract] Abstract: the central claim that IIT 3.0 repertoires 'emerge directly' from MaxCal variational principles is asserted without visible supporting equations, explicit mapping, or verification steps; this is load-bearing for the equivalence to prediction error and the bridge to FEP.
  2. [Abstract] Abstract: the definition of ψ is introduced as a proposal and then used to recover IIT quantities, but the equivalence to prediction error under CLT/LDT appears to depend on the specific form of the constraints on the path ensemble; without showing the general case is free of unstated auxiliary assumptions, the claimed generality risks circularity.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and detailed comments. We address each major point below and have revised the manuscript to improve the clarity and explicitness of the abstract and derivations where appropriate.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the central claim that IIT 3.0 repertoires 'emerge directly' from MaxCal variational principles is asserted without visible supporting equations, explicit mapping, or verification steps; this is load-bearing for the equivalence to prediction error and the bridge to FEP.

    Authors: We agree the abstract is concise and does not display the full equations. The manuscript derives the repertoires explicitly in Sections 3–4 by applying the MaxCal variational principle to the constrained path measure and showing that the resulting marginals recover the IIT cause-effect repertoires. To address the concern, we have revised the abstract to include a one-sentence outline of the variational step and added a forward reference to the relevant sections and equations. revision: partial

  2. Referee: [Abstract] Abstract: the definition of ψ is introduced as a proposal and then used to recover IIT quantities, but the equivalence to prediction error under CLT/LDT appears to depend on the specific form of the constraints on the path ensemble; without showing the general case is free of unstated auxiliary assumptions, the claimed generality risks circularity.

    Authors: The definition of ψ is the general deviation from the MaxCal ensemble under the observed constraints; the CLT and LDT equivalences follow from the standard statements of those theorems applied to the fluctuation statistics of the path measure, without further auxiliary constraints. We have added a clarifying paragraph in the revised discussion that states the assumptions explicitly and sketches the derivation steps to remove any appearance of circularity. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation proceeds from an explicit definitional proposal.

full rationale

The paper explicitly proposes a new definition of information as the deviation ψ from a constrained maximum-caliber path ensemble and then derives the IIT 3.0 cause/effect repertoires as consequences of the MaxCal variational principles applied to that definition. The claimed equivalence to prediction error is restricted to specific limiting cases (CLT for Markov chains and LDT for Ising models) rather than asserted as a general identity. No equations are presented in which an IIT quantity is shown to equal a fitted parameter or a self-referential constraint by construction, and no load-bearing self-citation chain is invoked to justify the central mapping. The argument is therefore self-contained as a theoretical re-expression rather than a tautological reduction of outputs to inputs.

Axiom & Free-Parameter Ledger

2 free parameters · 3 axioms · 1 invented entities

The central claim rests on a newly proposed definition of information together with standard mathematical tools; no numerical free parameters are explicitly fitted in the abstract, but the finite time horizon and choice of constraints function as definitional choices.

free parameters (2)
  • finite time horizon
    Chosen as part of the path-ensemble definition; affects the deviation measure ψ.
  • constraints on the path ensemble
    Specific constraints are required for the maximum-caliber construction but not enumerated in the abstract.
axioms (3)
  • standard math Variational principles of maximum caliber (CMEP)
    Invoked to derive cause/effect repertoires directly from constrained entropy maximization.
  • standard math Central Limit Theorem for Markov chains
    Applied to establish equivalence between ψ and prediction error.
  • standard math Large deviations theory for Ising models
    Used to show equivalence of ψ to prediction error in the Ising case.
invented entities (1)
  • information ψ no independent evidence
    purpose: Deviation of realized dynamics from constrained maximum-caliber path ensemble, serving as the bridge quantity between IIT and FEP
    Newly defined in the paper; no independent falsifiable handle outside the proposed framework is stated.

pith-pipeline@v0.9.0 · 5586 in / 1572 out tokens · 32946 ms · 2026-05-14T21:07:21.965052+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

230 extracted references · 194 canonical work pages · 17 internal anchors

  1. [1]

    (2004) An information integration theory of consciousness.BMC Neuroscience(2004)5(42)https://doi.org/10.1186/1471-2202-5-42

    TONONI, G. (2004) An information integration theory of consciousness.BMC Neuroscience(2004)5(42)https://doi.org/10.1186/1471-2202-5-42

  2. [2]

    (2008) Consciousness as integrated information: a provisional manifesto.The Biological Bulletin(2008)215(3): 216–242,https://doi.org/ 10.2307/25470707

    TONONI, G. (2008) Consciousness as integrated information: a provisional manifesto.The Biological Bulletin(2008)215(3): 216–242,https://doi.org/ 10.2307/25470707

  3. [3]

    and Gosseries, Olivia and Rosanova, Mario and Boly, M

    CASALI, A. G., GOSSERIES, O., ROSANOVA, M., BOLY, M., SARASSO, S., CASALI, K. R., CASAROTTO, S., BRUNO, M.-A., LAUREYS, S., TONONI, G., MASSIMINI, M. (2013) A Theoretically Based Index of Consciousness Independent of Sensory Processing and Behavior.Science Translational Medicine. 5(198)https://doi.org/10.1126/scitranslmed.3006294

  4. [4]

    OIZUMI, L.IS,ANDG

    M. OIZUMI, L.IS,ANDG. TONONI. (2014) From the phenomenology to the mechanisms of consciousness: integrated information theory 3.0.PLOS Compu- tational Biology10(5):e1003588, May 2014.https://doi.org/10.1371/journal.pcbi.1003588

  5. [5]

    BARRETT, A. B. (2014) An integration of integrated information theory with fundamental physics.Front. Psychol.5:63.https://doi.org/10.3389/fpsyg. 2014.00063

  6. [6]

    CERULLO, M. A. (2015). The Problem with Phi: A Critique of Integrated Information Theory.PLoS Computational Biology11(9), e1004286.https://doi. org/10.1371/journal.pcbi.1004286

  7. [7]

    TONONI, G. (2015). Integrated information theory.Scholarpedia,10(1), 4164.http://www.scholarpedia.org/article/Integrated_information_ theory

  8. [8]

    (2015) Integrated information-induced quantum collapse,Foundations of Physics45889-899https://doi.org/10.1007/ s10701-015-9905-6

    KREMNIZER, K., RANCHIN, A. (2015) Integrated information-induced quantum collapse,Foundations of Physics45889-899https://doi.org/10.1007/ s10701-015-9905-6

  9. [9]

    (2016) Improved Measures of Integrated Information.PLOS Computational Biology,12(11), e1005123.https://doi.org/10.1371/ journal.pcbi.1005123

    TEGMARK, M. (2016) Improved Measures of Integrated Information.PLOS Computational Biology,12(11), e1005123.https://doi.org/10.1371/ journal.pcbi.1005123

  10. [10]

    ZANARDI, P., TOMKA, M., VENUTI, L. C. (2018) Towards Quantum Integrated Information Theoryhttps://doi.org/10.48550/arXiv.1806.01421

  11. [11]

    MEDIANO, P. A. M., SETH, A. K., BARRETT, A. B. (2018) Measuring Integrated Information: Comparison of Candidate Measures in Theory and Simulation, https://doi.org/10.48550/arXiv.1806.09373

  12. [12]

    MEDIANO, P. A. M., ROSAS, F., CARHART-HARRIS, R. L., SETH, A. K., BARRETT, A. B. (2019) Beyond integrated information: A taxonomy of information dynamics phenomenahttps://doi.org/10.48550/arXiv.1909.02297

  13. [13]

    DOERIG, A., SCHURGER, A., HESS, K., HERZOG, M. H. (2019) The unfolding argument: Why IIT and other causal structure theories cannot explain consciousness,Consciousness and Cognition7249–59https://doi.org/10.1016/j.concog.2019.04.002

  14. [14]

    (2020) A general spectral decomposition of causal influences applied to integrated information.Journal of Neuroscience Methods3002020https://doi.org/10.1016/j.jneumeth.2019.108443

    COHEN, D., SASAI, S., TSUCHIYA, N., OIZUMI, M. (2020) A general spectral decomposition of causal influences applied to integrated information.Journal of Neuroscience Methods3002020https://doi.org/10.1016/j.jneumeth.2019.108443

  15. [15]

    (2021) The Mathematical Structure of Integrated Information Theory.Frontiers in Applied Mathematics and Statistics6https: //doi.org/10.3389/fams.2020.602973

    KLEINER, J., TULL, S. (2021) The Mathematical Structure of Integrated Information Theory.Frontiers in Applied Mathematics and Statistics6https: //doi.org/10.3389/fams.2020.602973

  16. [16]

    MEDIANO, P. A. M., FERNANDO, E. R., FARAH, J. C., SHANAHAN, M., BOR, D., BARRETT, A. B. (2022) Integrated information as a common signature of dynamical and information-processing complexity,Chaos32013115,https://doi.org/10.1063/5.0063384

  17. [17]

    NORTHOFF, G., ZILIO, F. (2022) From Shorter to Longer Timescales: Converging Integrated Information Theory (IIT) with the Temporo-Spatial Theory of Consciousness (TTC).Entropy (Basel).24(2): 270.https://doi.org/10.3390/e24020270

  18. [18]

    MEDIANO, P. A. M., ROSAS, F. E., BOR, D., SETH, A. K, BARRETT, A. B. (2022) The strength of weak integrated information theory.Trends in Cognitive Sciences26(8): 646–655https://doi.org/10.1016/j.tics.2022.04.008

  19. [19]

    (2023) Separating weak integrated information theory into inspired and aspirational approachesNeurosci Conscious.2023(1) https://doi.org/10.1093/nc/niad012

    LEUNG, A., TSUCHIYA, N. (2023) Separating weak integrated information theory into inspired and aspirational approachesNeurosci Conscious.2023(1) https://doi.org/10.1093/nc/niad012

  20. [20]

    (2023) Only what exists can cause: An intrinsic view of free will.https://doi.org/ 10.48550/arXiv.2206.02069 75 APREPRINT- 14THMAY, 2026

    TONONI, G., ALBANTAKIS, L., BOLY, M., CIRELLI, C., KOCH, C. (2023) Only what exists can cause: An intrinsic view of free will.https://doi.org/ 10.48550/arXiv.2206.02069 75 APREPRINT- 14THMAY, 2026

  21. [21]

    CEA, I., NEGRO, N., SIGNORELI, C. M. (2023) The Fundamental Tension in Integrated Information Theory 4.’s Realist Idealism.Entropy25(10), 1453 https://doi.org/10.3390/e25101453

  22. [22]

    (2023) Computing the Integrated Information of a Quantum Mechanism.Entropy25: 449https://doi

    ALBANTAKIS, L.; PRENTNER, R.; DURHAM, I. (2023) Computing the Integrated Information of a Quantum Mechanism.Entropy25: 449https://doi. org/10.3390/e25030449

  23. [23]

    ALBANTAKIS, L

    L. ALBANTAKIS, L. BARBOSA, G. FINDLAY, M. GRASSO, A. M. HAUN, W. MARSHALL, W. G. P. MAYNER, A. ZAEEMZADEH, M. BOLY, B. E. JUEL, S. SASAI, K. FUJII, I. DAVID, J. HENDREN, J. P. LANG,ANDG. TONONI. (2023) Integrated information theory (IIT) 4.0: Formulating the properties of phenomenal existence in physical terms.PLOS Comput. Biol.19(10):e1011465,https://doi...

  24. [24]

    (2025) Integrated Information Theory: A Consciousness-First Approach to What Exists.https://doi.org/10.48550/arXiv.2510

    TONONI, G., BOLY, M. (2025) Integrated Information Theory: A Consciousness-First Approach to What Exists.https://doi.org/10.48550/arXiv.2510. 25998

  25. [25]

    Adversarial testing of global neuronal workspace and integrated information theories of consciousness.Nature642, 133–142 (2025).https://doi.org/10.1038/s41586-025-08888-1

    COGITATECONSORTIUM ET AL. Adversarial testing of global neuronal workspace and integrated information theories of consciousness.Nature642, 133–142 (2025).https://doi.org/10.1038/s41586-025-08888-1

  26. [26]

    M., SCHNEIDER, S

    BAILEY, M. M., SCHNEIDER, S. L. (2026) When Wholes Resist Decomposition: A Spectral Measure of Epistemic EmergenceEntropy28(4): 380https: //doi.org/10.3390/e28040380 Predictive Coding, the Free Energy Principle, and Active Inference

  27. [27]

    RAO, P. N. R., BALLARD, D. H. (1999) Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects.Nat Neurosci.2(1): 79–87https://doi.org/10.1038/4580

  28. [28]

    (2003) Learning and inference in the brain.Neural Netw.16(9) 1325–1352https://doi.org/10.1016/j.neunet.2003.06.005

    FRISTON, K. (2003) Learning and inference in the brain.Neural Netw.16(9) 1325–1352https://doi.org/10.1016/j.neunet.2003.06.005

  29. [29]

    C., POUGET, A

    KNILL, D. C., POUGET, A. (2004) The Bayesian brain: the role of uncertainty in neural coding and computation,Trends in Neurosciences.27(12): 712–719 https://doi.org/10.1016/j.tins.2004.10.007

  30. [30]

    (2005) A theory of cortical responses.Philos Trans R Soc Lond B Biol Sci.360(1456), 815–836https://doi.org/10.1098/rstb.2005.1622

    FRISTON, K. (2005) A theory of cortical responses.Philos Trans R Soc Lond B Biol Sci.360(1456), 815–836https://doi.org/10.1098/rstb.2005.1622

  31. [31]

    Schneider, B

    FRISTON, K., KILNER, J., HARRISON, L. (2006) A free energy principle for the brain.J Physiol Paris.100(1-3): 70–87https://doi.org/10.1016/j. jphysparis.2006.10.001

  32. [32]

    Nature Reviews Neuroscience , year =

    FRISTON, K. (2010) The free-energy principle: a unified brain theory?.Nat Rev Neurosci11127–138https://doi.org/10.1038/nrn2787

  33. [33]

    (2011) Action understanding and active inferenceBiol Cybern.104(1-2): 137–160https://doi.org/10.1007/ s00422-011-0424-z

    FRISTON, K., MATTOUT, J., KILNER, J. (2011) Action understanding and active inferenceBiol Cybern.104(1-2): 137–160https://doi.org/10.1007/ s00422-011-0424-z

  34. [34]

    Journal of The Royal Society Interface , year =

    FRISTON, K. (2013) Life as we know it.J R Soc Interface10(86)https://doi.org/10.1098/rsif.2013.0475

  35. [35]

    (2016) Active inference and learning.Neuroscience & Biobehavioral Reviews68862–879https://doi.org/10.1016/j.neubiorev.2016.06.022

    FRISTON, K., FITZGERALD, T., RIGOLI, F., SCHWARTENBECK, P., O’DOHERTY, J., PEZZULO, G. (2016) Active inference and learning.Neuroscience & Biobehavioral Reviews68862–879https://doi.org/10.1016/j.neubiorev.2016.06.022

  36. [36]

    B., ADAMS, R

    MIRZA, M. B., ADAMS, R. A., MATHYS, C. D., FRISTON, K. J. (2016) Scene Construction, Visual Foraging, and Active Inference.Front. Comput. Neurosci. 10https://doi.org/10.3389/fncom.2016.00056

  37. [37]

    A., STEPHAN, K

    ADAMS, R. A., STEPHAN, K. E., BROWN, H. R., FRITH, C. D., FRISTON, K. J. (2016) The computational anatomy of psychosisFront. Psychiatry.4 https://doi.org/10.3389/fpsyt.2013.00047

  38. [38]

    (2016) Active inference and robot control: a case studyJ R Soc Interface13(122): 20160616 https://doi.org/10.1098/rsif.2016.0616

    PIO-LOPEZ, L., NIZARD, A., FRISTON, K., PEZZULO, G. (2016) Active inference and robot control: a case studyJ R Soc Interface13(122): 20160616 https://doi.org/10.1098/rsif.2016.0616

  39. [39]

    L., KIM, C

    BUCKLEY, C. L., KIM, C. S., MCGREGOR, S., SETH, A. K. (2017) The free energy principle for action and perception: A mathematical reviewJournal of Mathematical Psychology(2017)81: 55–79https://doi.org/10.1016/j.jmp.2017.09.004

  40. [40]

    D., PEZZULO, G., HOBSON, J

    FRISTON, K., LIN, M., FRITH, C. D., PEZZULO, G., HOBSON, J. A., ONDOBAKA, S. (2017) Active Inference, Curiosity and Insight.Neural Computation 29(10): 2633–2683https://doi.org/10.1162/neco_a_00999

  41. [41]

    Deep Active Inference as Variational Policy Gradients

    MILLIDGEB. (2019) Deep Active Inference as Variational Policy Gradients.arXiv:1907.03876https://doi.org/10.48550/arXiv.1907.03876

  42. [42]

    (2020) Markov blankets, information geometry and stochastic thermodynamics.Philos Trans A Math Phys Eng Sci378 (2164): 20190159https://doi.org/10.1098/rsta.2019.0159

    PARR, T., DACOSTA, L., FRISTON, K. (2020) Markov blankets, information geometry and stochastic thermodynamics.Philos Trans A Math Phys Eng Sci378 (2164): 20190159https://doi.org/10.1098/rsta.2019.0159

  43. [43]

    (2020) Reinforcement Learning through Active Inference.Bridging AI and Cognitive Science Workshop (BAICS), ICLR 2020https://baicsworkshop.github.io/pdf/BAICS_37.pdf

    TSCHANTZ, A., MILLIDGE, B., SETH, A., BUCKLEY, C. (2020) Reinforcement Learning through Active Inference.Bridging AI and Cognitive Science Workshop (BAICS), ICLR 2020https://baicsworkshop.github.io/pdf/BAICS_37.pdf

  44. [44]

    GOTTWALD, S., BRAUN, D. A. (2020) The two kinds of free energy and the Bayesian revolutionPLoS Comput Biol16(12): e1008420.https://doi.org/ 10.1371/journal.pcbi.1008420

  45. [45]

    (2020) Active inference on discrete state-spaces: A synthesis.Journal of Mathematical Psychology(2020)99https://doi.org/10.1016/j.jmp.2020.102447

    DACOSTA, L., PARR, T., SAJID, N., VESELIC, S., NEACSU, V., FRISTON, K. (2020) Active inference on discrete state-spaces: A synthesis.Journal of Mathematical Psychology(2020)99https://doi.org/10.1016/j.jmp.2020.102447

  46. [46]

    PEZZATO, C., FERRARI, R., CORBATO, C. H. (2020) A Novel Adaptive Controler for Robot Manipulators Based on Active InferenceIEEE Robotics and Automation Letters5(2): 2973–2980https://doi.org/10.1109/LRA.2020.2974451

  47. [47]

    MILLIDGE, B., SETH, A., BUCKLEY, C. L. (2021) Predictive Coding: a Theoretical and Experimental Review,https://doi.org/10.48550/arXiv.2107. 12979

  48. [48]

    A., KANAI, R

    BIEHL, M., POLLOCK, F. A., KANAI, R. (2021) A Technical Critique of Some Parts of the Free Energy Principle.Entropy (Basel)https://doi.org/10. 3390/e23030293 76 APREPRINT- 14THMAY, 2026

  49. [49]

    KAWAHARA, D., OZEKI, A., MIZUUCHI, I. (2022) A Curiosity Algorithm for Robots Based on the Free Energy Principle.2022 IEEE/SICE International Symposium on System Integration (SII), Narvik, Norway, 2022, pp 53-59doi:10.1109/SII52469.2022.9708819https://ieeexplore.ieee.org/document/ 9708819

  50. [50]

    (2022) How Active Inference Could Help Revolutionise RoboticsEntropy (Basel)24(3): 361https://doi.org/10.3390/e24030361

    DACOSTA, L., LANILLOS, P., SAJID, N., FRISTON, K., KHAN, S. (2022) How Active Inference Could Help Revolutionise RoboticsEntropy (Basel)24(3): 361https://doi.org/10.3390/e24030361

  51. [51]

    KAWAHARA, D., OZEKI, S., MIZUUCHI, I. (2022) A Curiosity Algorithm for Robots Based on the Free Energy Principle2022 IEEE/SICE Internatinoal Symposium on System Integration (SII)pp 53–59https://doi.org/10.1109/SII52469.2022.9708819

  52. [52]

    (2022) Unbiased Active Inference for Classical Control2022 International Conference on Intelligent Robots and Systemshttps://doi.org/10.48550/arXiv.2207.13409

    BAIOUMY, M., PEZZATO, C., FERRARI, R., HAWES, N. (2022) Unbiased Active Inference for Classical Control2022 International Conference on Intelligent Robots and Systemshttps://doi.org/10.48550/arXiv.2207.13409

  53. [53]

    SAKTHIVADIVEL, D. A. R. (2022) Towards a Geometry and Analysis for Bayesian Mechanics.https://doi.org/10.48550/arXiv.2204.11900

  54. [54]

    S., FRISTON, K

    BETTINGER, J. S., FRISTON, K. J. (2023) Conceptual foundations of physiological regulation incorporating the free energy principle and self-organized criticalityNeuroscience and Biobehavioral Reviews,155, Dec23, 105459,https://doi.org/10.1016/j.neubiorev.2023.105459

  55. [55]

    SAKTHIVADIVEL, D. A. R. (2023). A Worked Example of the Bayesian Mechanics of Classical Objects. In: Buckley, C. L.,et al.Active Inference.IWAI 2022. Communications in Computer and Information Science, vol 1721(Springer, Cham.)https://doi.org/10.1007/978-3-031-28719-0_21

  56. [56]

    RAMSTEAD, M. J. D., SAKITHIVADEL, D. A. R., HEINS, C., KOUDAHL, M., MILLIDGE, B., DACOSTA, L., KLEIN, B., FRISTON, K. J. (2023) On Bayesian Mechanics: A Physics of and by Beliefs.Interface Focus13(3): 20220029https://doi.org/10.1098/rsfs.2022.0029

  57. [57]

    FRISTON, K., DACOSTA, L., SAKTHIVADIVEL, D. A. R., HEINS, C., PAVLIOTIS, G. A., RAMSTEAD, M., PARR, T. (2023) Path integrals, particular kinds, and strange thingsPhysics of Life Reviews(2023)4732–62https://doi.org/10.1016/j.plrev.2023.08.016

  58. [58]

    CAUCHETEUX, C., GRAMFORT, A., KING, J. R. (2023) EVIDENCE OF A PREDICTIVE CODING HIERARCHY IN THE HUMAN BRAIN LISTENING TO SPEECH Nat Hum Behav7, 430–441c

  59. [59]

    (2024) Active inference as a theory of sentient behaviourBiological Psychology,186Feb2024 108741,https: //doi.org/10.1016/j.biopsycho.2023.108741

    PEZZULO, G., THOMAS, P., FRISTON, K. (2024) Active inference as a theory of sentient behaviourBiological Psychology,186Feb2024 108741,https: //doi.org/10.1016/j.biopsycho.2023.108741

  60. [60]

    Q., FIDERER, L

    PAZEM, J., KRUMM, M., VINING, A. Q., FIDERER, L. J., BRIEGEL, H. J. (2024) Free Energy Projective Simulation (FEPS): Active inference with inter- pretability.https://doi.org/10.48550/arXiv.2411.14991

  61. [61]

    (2024) The empirical status of predictive coding and active inference.Neuroscience & Biobehavioral Reviews157: 105473https://doi.org/10.1016/j.neubiorev.2023.105473

    HODSON, R., MEHTA, M., SMITH, R. (2024) The empirical status of predictive coding and active inference.Neuroscience & Biobehavioral Reviews157: 105473https://doi.org/10.1016/j.neubiorev.2023.105473

  62. [62]

    (2025) Addressing the Subsumption Thesis: A Formal Bridge between Microeconomics and Active Inferencehttps://doi.org/10.48550/ arXiv.2503.05048

    KUHN, N. (2025) Addressing the Subsumption Thesis: A Formal Bridge between Microeconomics and Active Inferencehttps://doi.org/10.48550/ arXiv.2503.05048

  63. [63]

    Noise Trading

    ADRA, S. (2025) The Free Energy Principle in Financial Markets: In PRaise of “Noise Trading”Review of Behavioral Economics12(2): 173–190https: //doi.org/10.1561/105.00000208

  64. [64]

    J., SLAGTER, H

    LAGEMAN, J., FAHRENFORT, J. J., SLAGTER, H. A. (2026) Prediction in action: toward an empirical science of active inference.PsyArXivhttps://doi. org/10.31234/osf.io/jg372_v1 IIT-FEP unification

  65. [65]

    (2014) Evolution of Integrated Causal Structures in Animats Exposed to Environments of Increasing ComplexityPLOS Comput Biol.https://doi.org/10.1371/journal.pcbi.1003966

    ALBANTAKISL., HINTZE, A., KOCH, C., ADAMI, C., TONONI, G. (2014) Evolution of Integrated Causal Structures in Animats Exposed to Environments of Increasing ComplexityPLOS Comput Biol.https://doi.org/10.1371/journal.pcbi.1003966

  66. [66]

    SAFRON, A. (2020) An Integrated World Modeling Theory (IWMT) of Consciousness: Combining Integrated Information and Global Neuronal Workspace Theories With the Free Energy Principle and Active Inference Framework; Toward Solving the Hard Problem and Characterizing Agentic Causation.Front. Artif. Intell.3https://doi.org/10.3389/frai.2020.00030

  67. [67]

    (2022) Integrated world modeling theory expanded: Implications for the future of consciousness,Frontiers in Computational Neuroscience,16 https://doi.org/10.3389/fncom.2022.642397

    SAFRONA. (2022) Integrated world modeling theory expanded: Implications for the future of consciousness,Frontiers in Computational Neuroscience,16 https://doi.org/10.3389/fncom.2022.642397

  68. [68]

    OLESEN, C.L., WAADE, P.T., ALBANTAKIS, L., MATHYS, C. (2023) Phi fluctuates with surprisal: An empirical pre-study for the synthesis of the free energy principle and integrated information theoryPLOS Computational Biologyhttps://doi.org/10.1371/journal.pcbi.1011346

  69. [69]

    (2025) Bridging integrated information theory and the free-energy principle in living neuronal networks.https://doi.org/10.48550/arXiv.2510.04084

    MAYAMA, T., SHIMIZU, S., TAKANO, Y., AKITA, D., TAKAHASHI, H. (2025) Bridging integrated information theory and the free-energy principle in living neuronal networks.https://doi.org/10.48550/arXiv.2510.04084

  70. [70]

    W., HAUN, A

    INTREPID CONSORTRIUM: CORCORAN, A. W., HAUN, A. M., DORMAN, R., TONONI, G., FRISTON, K. J., PENNARTZ, C. M. A. (2026) Integrated information and predictive processing theories of consciousness: An adversarial collaborative reviewhttps://doi.org/10.48550/arXiv.2509.00555 Consciousness

  71. [71]

    HELMHOLTZ, H., (PUBLISHED BY) PRATT, C. C. (1926)Helmholtz’s treatise on physiological optics, vols. i–iii Journal of Applied Psychology10(2)https: //psycnet.apa.org/doi/10.1037/h0068555

  72. [72]

    WHITEHEAD, A. N. (1929) Process and reality.Macmillan

  73. [73]

    (1974) What Is It Like to Be a Bat?The Philosophical Review83(4): 430–450https://doi.org/10.2307/2183914 77 APREPRINT- 14THMAY, 2026

    NAGEL, T. (1974) What Is It Like to Be a Bat?The Philosophical Review83(4): 430–450https://doi.org/10.2307/2183914 77 APREPRINT- 14THMAY, 2026

  74. [74]

    BAARS, B. J. (1988) A Cognitive Theory of Consciousness. (Cambridge University Press, New York, 1988)

  75. [75]

    CHALMERS, D. J. (1995) Facing up to the problem of consciousness.Journal of Consciousness Studies2(3):200–219, 1995.https://philpapers.org/rec/ CHAFUT

  76. [76]

    Cognition , year =

    DAHAENE, S., NACCACHE, L. (2001) Towards a cognitive neuroscience of consciousness: basic evidence and a workspace frameworkCognition79(1–2): 1–37 https://doi.org/10.1016/S0010-0277(00)00123-2

  77. [77]

    BAARS, B. J. (2005) Global workspace theory of consciousness: toward a cognitive neuroscience of human experienceProg Brain Res.150: 45–53https: //doi.org/10.1016/s0079-6123(05)50004-9

  78. [78]

    Consciousness in the universe: A review of the ‘Orch OR’ theory , journal =

    HAMEROFF, S., PENROSE, R. (2014) Consciousness in the universe: A review of the ‘Orch OR’ theoryPhysics of Life Reviews,11(1), 39-78.https: //doi.org/10.1016/j.plrev.2013.08.002

  79. [80]

    FRISTON, K. J. (2018) Of woodlice and men: A Bayesian account of cognition, life and consciousnessALIUS Interdisciplinary research group on the diversity of consciousnesshttps://doi.org/10.34700/h460-nz89

  80. [81]

    CHALMERS, D. J. (2018) The meta-problem of consciousness.Journal of Consciousness Studies,25(9-10), 6–61https://philpapers.org/archive/ chatmo-32.pdf

Showing first 80 references.