pith. machine review for the scientific record. sign in

arxiv: 2604.15679 · v1 · submitted 2026-04-17 · 💻 cs.LG · cs.AI· cs.CV

Recognition: unknown

Hierarchical Active Inference using Successor Representations

Authors on Pith no claims yet

Pith reviewed 2026-05-10 09:39 UTC · model grok-4.3

classification 💻 cs.LG cs.AIcs.CV
keywords active inferencesuccessor representationshierarchical planningreinforcement learningfree energy principleabstract statesabstract actionsnavigation tasks
0
0 comments X

The pith

Lower-level successor representations and active inference planning bootstrap higher-level abstract states and actions that enable efficient planning.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper proposes extending active inference, a framework based on the free energy principle for perception and action, by adding hierarchy and successor representations. Lower levels of the model use successor representations to identify abstract states and use their own planning to discover abstract actions. These abstractions then speed up planning at higher levels. The results are demonstrated on navigation, partially observable, and continuous control tasks where flat active inference struggles to scale. A sympathetic reader cares because this offers a concrete way to make brain-inspired planning models work on larger problems.

Core claim

The central claim is that a hierarchical model of the environment combined with successor representations allows lower-level successor representations to learn higher-level abstract states, lower-level active inference planning to bootstrap higher-level abstract actions, and the resulting abstractions to facilitate efficient planning. This is shown on a variant of the four rooms task, key-based navigation, partially observable planning, the Mountain Car problem, and PointMaze with continuous spaces. The work presents the first application of learned hierarchical state and action abstractions to active inference.

What carries the argument

The integration of successor representations inside a hierarchical active inference model, where lower-level representations define abstract states and lower-level policies define abstract actions for use at higher levels.

If this is right

  • Planning time decreases in environments with structure because decisions can be made over abstract states and actions rather than raw ones.
  • The same lower-level mechanisms can be reused to discover abstractions for new tasks without hand-designed hierarchies.
  • The approach extends active inference to continuous state and action spaces as well as partially observable settings.
  • Performance on standard reinforcement learning benchmarks improves once the bootstrapped abstractions are in place.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same bootstrapping process could be applied iteratively to create deeper hierarchies for even larger problems.
  • The method suggests a possible computational account for how the brain might use multi-scale successor-like representations during planning.
  • Hybrid systems could combine this active-inference hierarchy with other reinforcement learning algorithms that also learn abstractions.

Load-bearing premise

That lower-level successor representations and active inference planning can reliably produce higher-level abstract states and actions that generalize to improve planning.

What would settle it

Running the model on a new, larger environment and finding that the hierarchical version shows no reduction in planning steps or no gain in success rate compared with flat active inference would falsify the claim.

Figures

Figures reproduced from arXiv: 2604.15679 by Prashant Rangarajan, Rajesh P. N. Rao.

Figure 1
Figure 1. Figure 1: Hierarchical Active Inference using Successor Representations. (continued next page) 4 [PITH_FULL_IMAGE:figures/full_fig_p004_1.png] view at source ↗
Figure 1
Figure 1. Figure 1: (continued) (a) A 9 × 9 Gridworld with five rooms (separated by white walls with openings). The agent must navigate from the start state (top left) to the goal state (purple target symbol) to get a reward. In this example, one of the rooms has noisy observations. Active inference in this case would prefer an alternate (longer but less noisy) path to the goal. (b) By clustering of the successor matrix, the … view at source ↗
Figure 2
Figure 2. Figure 2: Generative model for a Discrete POMDP. The notation used here follows prior active inference literature (e.g., Smith et al., 2022). The EFE G depends on the prior C and generates the probabilities of different policies π. The policy π being a sequence of actions in turn affects the state transition matrix B. The observations are generated from the state using the observation model determined by the matrix … view at source ↗
Figure 3
Figure 3. Figure 3: Task 1: Serpentine MDP Gridworld Environment (a) A 9 × 9 Gridworld with a single start state (yellow), a single goal state (purple), and walls (solid white rectangles). (b) Successor representation of states with respect to the start state. The states closer to the start state have larger successor values and are represented by a lighter shade of brown. (c) Macro states learned by the model (4 in number). … view at source ↗
Figure 4
Figure 4. Figure 4: Macro Actions for Task 1. (a)-(c) represent macro actions that execute a policy at the lower level for transitioning from one macro state to another (here, macro actions for S2 → S0, S0 → S3, and S3 → S1 are shown). The white arrows indicate the most probable action from each micro state in the first macro state for reaching the bottleneck state of the second macro state. blue arrows in Figure 5a show the … view at source ↗
Figure 5
Figure 5. Figure 5: Planning using Hierarchical Active Inference for Task 1. (a) Higher-level transition function and higher-level plan. The arrows represent the possible transitions between the learned macro states using the learned macro actions. The blue arrows represent the optimal sequence of macro actions (obtained using active inference at the macro level) to navigate from macro state 2 (which contains the agent’s star… view at source ↗
Figure 6
Figure 6. Figure 6: Comparison of Planning Performance using Hierarchical, Flat Active Inference and RL baseline Models. (continued next page) 31 [PITH_FULL_IMAGE:figures/full_fig_p031_6.png] view at source ↗
Figure 6
Figure 6. Figure 6: (a) Total reward obtained by the agent as a function of number of training [PITH_FULL_IMAGE:figures/full_fig_p032_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Goal Change Results. (a) Performance when the goal is changed for the Q-learning agent and the hierarchical and flat agents based on successor representation (SR). The SR agents quickly re-plan, whereas Q-learning requires substantial retraining. (b) Four rooms environment illus￾trating the sequence of goals used during the experiment. 36 [PITH_FULL_IMAGE:figures/full_fig_p036_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Task 2: Gridworld with Key. (a) A 5×5 Gridworld with the agent’s start location (yellow), goal (purple), key location (orange), and walls (white). (b) Successor representation with respect to start micro state (brighter is higher successor value). The left half corresponds to the micro states of the agent prior to picking up the key, while the right half corresponds to after. (c) Macro states learned by th… view at source ↗
Figure 9
Figure 9. Figure 9: Macro Actions for Task 2. (a)-(d) show four of the macro actions, in terms of their corresponding lower level policy, for transitioning from one macro state to another. The macro actions shown are for the transitions S2 → S3, S3 → S0, S0 → S4 and S4 → S1 (Si represents macro state i in the plots). For each Si → Sj macro action, the white arrows indicate the most probable action for each micro state within … view at source ↗
Figure 10
Figure 10. Figure 10: Planning using Hierarchical Active Inference for Task 2. (a) Higher Level Planning: The arrows represent macro actions between pairs of ad￾jacent macro states. The blue arrows represent the macro action plan inferred by the agent through active inference at the macro level. This sequence of macro actions takes the agent from macro state 2 (containing the start location, no key) to 1 (containing the goal l… view at source ↗
Figure 11
Figure 11. Figure 11: Task 3: POMDP Gridworld Environment with five rooms. (a) A 9 × 9 Gridworld with a single agent (yellow), a single goal (purple), walls (white) and high noise room (gray) with a reward of 100 associated with the goal and overall observa￾tional noise parameter η = 0.1. (b) Successor representation of states with respect to initial state (brighter is closer). (c) Macro states learned by the model, which coin… view at source ↗
Figure 12
Figure 12. Figure 12: Macro Actions for Task 3. (a)-(c) represent macro actions that execute a policy at the lower level for transitioning from one macro state to another (here, macro actions for S0 → S1, S1 → S3, and S3 → S4 are shown). The white arrows indicate the most probable action from each micro state in the first macro state in order to eventually reach the bottleneck state of the second macro state. in the top right … view at source ↗
Figure 13
Figure 13. Figure 13: Higher Level Planning for Task 3. The arrows represent the potential macro actions between all pairs of adjacent macro states. The blue arrows represent the sequence of macro actions selected via higher-level active inference to get from macro state 0 (containing the start state) to macro state 4 (containing the goal state). Note that with active inference, the agent avoids the shorter path containing mac… view at source ↗
Figure 14
Figure 14. Figure 14: Active Inference optimizes the Tradeoff between Uncertainty and Re￾ward in Task 3. (a) Minimizing EFE using hierarchical active inference results in the agent naturally avoiding macro states with high noise (top right room), reaching the goal via the slightly longer path through two other rooms. (b) When the noise level in the top right room is decreased to the same noise level as the rest of the environm… view at source ↗
Figure 15
Figure 15. Figure 15: Tradeoff between Uncertainty and Reward as a Function of Noise in Task 3. The plot shows the probability that the agent, using active inference, chooses the shorter path to the goal in Task 3 as the noise in the top right room is increased by increasing the entropy of the observation model for that room. As the entropy increases, the agent increasingly prefers the longer path over the shorter one. Note th… view at source ↗
Figure 16
Figure 16. Figure 16: Mountain Car Problem and Value Function. (a) (Top panels) Three key stages depicted as snapshots in the MountainCar problem. We show the agent at its start state, the leftmost turning point where it begins accel￾erating to the right, and the final goal location at the top right hill. (Bottom panel). Trajectory of a successful run plotted in position-velocity state space, with the three stages shown in the… view at source ↗
Figure 17
Figure 17. Figure 17: Hierarchical Active Inference for Mountain Car. (a) State Space Clustering. The 100 micro states generated from a discretization of the continuous Mountain Car state space were clustered into 6 macro states using spec￾tral embedding. (b) Clustering Embeddings. Visualization of the macro states in the spectral embedding space (arbitrary units). (c) Successful Trajectory obtained using Hierarchical Active I… view at source ↗
Figure 18
Figure 18. Figure 18: PointMaze Variants. (a) UMaze (5 × 5), (b) Medium (8 × 8), and (c) Large (9 × 12). The green ball is the agent (point mass) and the red ball marks the goal location. Walls (brown) partition the continuous space into rooms connected by narrow corridors. Actions are continuous as well (see text for details). connected by narrow passages, the kind of bottleneck structure that spectral cluster￾ing of the succ… view at source ↗
Figure 19
Figure 19. Figure 19: UMaze and Macro State Clusters. (a) UMaze environment. (b) Four clus￾ters (colored regions) discovered by spectral clustering of the successor representation, overlaid on the discretized maze layout (dark areas denote walls). Arrows show the macro level policy directing the agent from each macro state toward the goal. The start (S) and goal (G) macro states are also marked on the figure. and flat agents c… view at source ↗
Figure 20
Figure 20. Figure 20: UMaze: Hierarchical vs. Flat Active Inference Performance Compari￾son. goal using ∼150 steps, largely because the macro level abstraction is robust to errors in the partially-learned successor representation. The flat agent eventually succeeds on UMaze given sufficient training episodes, but converges slower and with higher vari￾ance in both success rate and steps to goal. Multi-Goal Re-Planning The rewar… view at source ↗
Figure 21
Figure 21. Figure 21: UMaze Learning Curves. The plots show mean ± SEM over 20 seeds. Top: Success rate. Bottom: Micro level steps taken by the policy (capped at 5000). The hierarchical agent (blue) reaches 100% success by 200 episodes and reaches the goal in ∼150 steps; the flat agent (orange) lags behind in both metrics and settles with higher variance. On UMaze, both agents succeed because corridors are short enough for the… view at source ↗
Figure 22
Figure 22. Figure 22: PointMaze: Multi-Goal Re-Planning for Maze Variants. The agent learns the environment once (i.e., learns the successor representation), then navigates to a sequence of goals without re-learning. (a) UMaze with 5 goals. (b) Medium with 5 goals. (c) Large with 6 goals. 5 Conclusions and Future Work Active inference has emerged as a promising biologically-plausible approach to embod￾ied intelligence (Friston… view at source ↗
Figure 23
Figure 23. Figure 23: PointMaze: Hierarchy vs. Flat Performance Comparison. (a) UMaze. (b) Medium. (c) Large. Red crosses indicate failure to reach the goal. The hierarchical agent (blue) reaches all goals; the flat agent (orange) fails increasingly as the path to the goal gets longer. 59 [PITH_FULL_IMAGE:figures/full_fig_p059_23.png] view at source ↗
Figure 1
Figure 1. Figure 1: Results for Four Rooms MDP Problem. See text for details. 63 [PITH_FULL_IMAGE:figures/full_fig_p063_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Results for POMDP Version of Task 1. See text for details. 66 [PITH_FULL_IMAGE:figures/full_fig_p066_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Single-goal planning cost: Medium and Large. (a) Medium: both agents succeed, but the flat agent requires 170× more planning steps. (b) Large: the flat agent times out at 20,000 steps (the maximum) without reaching the goal, while the hierarchy succeeds in 6 steps. succeed, but the efficiency gap widens substantially compared to UMaze. On Large, the hierarchical agent succeeds in just 6 planning decisions,… view at source ↗
Figure 4
Figure 4. Figure 4: Trajectory for UMaze. Rendered frames (top) and corresponding trajectory segments (bottom). (a) to (b) to (c) correspond to macro actions (room-to-room navi￾gation). 73 [PITH_FULL_IMAGE:figures/full_fig_p073_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: Hierarchical trajectory on UMaze. The agent navigates from start (top-left) to goal (bottom-left), overlaid on the cluster map (faded regions). (a) Trajectory colored by current macro state membership. (b) Trajectory colored by macro action target (the cluster being navigated toward); gold indicates the goal phase within the final cluster. 74 [PITH_FULL_IMAGE:figures/full_fig_p074_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Macro State Clusters for the Three Maze Variants. (a) UMaze (4 clus￾ters). (b) Medium (8 clusters). (c) Large (12 clusters). Colored regions show clusters discovered by spectral clustering of the successor representation. Cluster boundaries align with narrow corridors connecting rooms. 75 [PITH_FULL_IMAGE:figures/full_fig_p075_6.png] view at source ↗
read the original abstract

Active inference, a neurally-inspired model for inferring actions based on the free energy principle (FEP), has been proposed as a unifying framework for understanding perception, action, and learning in the brain. Active inference has previously been used to model ecologically important tasks such as navigation and planning, but scaling it to solve complex large-scale problems in real-world environments has remained a challenge. Inspired by the existence of multi-scale hierarchical representations in the brain, we propose a model for planning of actions based on hierarchical active inference. Our approach combines a hierarchical model of the environment with successor representations for efficient planning. We present results demonstrating (1) how lower-level successor representations can be used to learn higher-level abstract states, (2) how planning based on active inference at the lower-level can be used to bootstrap and learn higher-level abstract actions, and (3) how these learned higher-level abstract states and actions can facilitate efficient planning. We illustrate the performance of the approach on several planning and reinforcement learning (RL) problems including a variant of the well-known four rooms task, a key-based navigation task, a partially observable planning problem, the Mountain Car problem, and PointMaze, a family of navigation tasks with continuous state and action spaces. Our results represent, to our knowledge, the first application of learned hierarchical state and action abstractions to active inference in FEP-based theories of brain function.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 3 minor

Summary. The manuscript proposes a hierarchical active inference framework that uses successor representations (SRs) to construct higher-level abstract states from lower-level SRs and to bootstrap higher-level abstract actions via lower-level active-inference policy selection. It claims three capabilities—learning abstract states, bootstrapping abstract actions, and improved planning efficiency—and demonstrates them on a four-rooms variant, key-based navigation, a POMDP, Mountain Car, and PointMaze tasks.

Significance. If the empirical claims hold under controlled evaluation, the work provides a concrete mechanism for scaling active inference to larger problems by importing SR-based abstraction from RL, while remaining grounded in the free-energy principle. This could strengthen links between FEP models and hierarchical RL and supply falsifiable predictions about multi-scale state-action representations.

major comments (2)
  1. [Results] Results sections (four-rooms, PointMaze, Mountain Car): performance gains are reported without explicit baselines (flat active inference, standard SR or DQN variants), error bars, or statistical tests; this leaves the three claimed demonstrations difficult to evaluate quantitatively.
  2. [Methods / Abstract action learning] Section describing abstract-action bootstrapping: the claim that lower-level planning reliably produces generalizable higher-level actions depends on specific choices of learning rates and discount factors (listed as free parameters); the manuscript should show sensitivity analysis or ablation to confirm the bootstrapping is robust rather than tuned to the reported environments.
minor comments (3)
  1. [Abstract] Abstract: states the three demonstrations but supplies no quantitative metrics or comparison points; adding one-sentence performance highlights would improve clarity.
  2. [Preliminaries] Notation: successor matrix and free-energy terms are introduced with varying symbols across equations; a single consistent notation table would reduce reader burden.
  3. [Figures] Figures: several trajectory plots lack axis labels or legend entries for the hierarchical vs. flat conditions; ensure all panels are self-contained.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the positive assessment, constructive comments, and recommendation for minor revision. We address the two major comments point by point below and will revise the manuscript accordingly to strengthen the quantitative evaluation and robustness analysis.

read point-by-point responses
  1. Referee: [Results] Results sections (four-rooms, PointMaze, Mountain Car): performance gains are reported without explicit baselines (flat active inference, standard SR or DQN variants), error bars, or statistical tests; this leaves the three claimed demonstrations difficult to evaluate quantitatively.

    Authors: We agree that explicit baselines, error bars, and statistical tests would improve the clarity and rigor of the empirical claims. In the revised manuscript we will add direct comparisons against flat active inference, standard successor-representation methods, and appropriate DQN or SR-based RL variants (where the task formulation permits). We will also report standard errors across independent runs and include statistical significance tests (e.g., paired t-tests or Wilcoxon rank-sum tests with appropriate corrections) for the performance differences on the four-rooms, PointMaze, and Mountain Car domains. These additions will make the three claimed capabilities quantitatively evaluable. revision: yes

  2. Referee: [Methods / Abstract action learning] Section describing abstract-action bootstrapping: the claim that lower-level planning reliably produces generalizable higher-level actions depends on specific choices of learning rates and discount factors (listed as free parameters); the manuscript should show sensitivity analysis or ablation to confirm the bootstrapping is robust rather than tuned to the reported environments.

    Authors: We acknowledge the importance of demonstrating that abstract-action bootstrapping does not hinge on narrowly tuned hyperparameters. In the revision we will add a sensitivity analysis that systematically varies the lower-level learning rate and discount factor over ranges centered on the values used in the reported experiments. We will also include targeted ablations that disable or alter the lower-level active-inference planning step to isolate its contribution to higher-level action learning. These results will be presented for the key-based navigation and POMDP tasks to confirm robustness across the environments where bootstrapping is demonstrated. revision: yes

Circularity Check

0 steps flagged

No significant circularity; derivation self-contained

full rationale

The manuscript constructs a hierarchical active inference model by combining established successor representations (SR) with active inference policy selection to learn abstract states and actions from lower-level SRs. The three claimed capabilities are shown via explicit algorithmic constructions and empirical results on standard benchmark tasks (four-rooms, key navigation, POMDP, Mountain Car, PointMaze). No equations reduce a prediction to a fitted parameter by construction, no load-bearing premise rests solely on self-citation, and the central claims remain independently verifiable through the reported performance gains and generalization tests. The derivation therefore does not collapse to its inputs.

Axiom & Free-Parameter Ledger

1 free parameters · 2 axioms · 0 invented entities

The model rests on standard assumptions from active inference and successor representation literature plus the novel hierarchical extension; specific free parameters such as learning rates are not detailed in the abstract but are expected in such frameworks.

free parameters (1)
  • learning rates and discount factors
    Typical hyperparameters in RL and active inference models that are tuned to enable the bootstrapping and planning demonstrations.
axioms (2)
  • domain assumption Successor representations can be used to learn higher-level abstract states from lower-level ones.
    Invoked in demonstration (1) of the abstract.
  • domain assumption Lower-level active inference planning can bootstrap higher-level abstract actions.
    Invoked in demonstration (2) of the abstract.

pith-pipeline@v0.9.0 · 5549 in / 1404 out tokens · 92699 ms · 2026-05-10T09:39:41.027064+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

6 extracted references · 4 canonical work pages · 1 internal anchor

  1. [1]

    Aitchison, L., & Lengyel, M. (2017). With or without you: Predictive coding and Bayesian inference in the brain.Current opinion in neurobiology,46, 219–227. Barreto, A., Dabney, W., Munos, R., Hunt, J. J., Schaul, T., van Hasselt, H. P., & Silver, D. (2017). Successor features for transfer in reinforcement learning.Advances in neural information processin...

  2. [2]

    Doya, K., Ishii, S., Pouget, A., & Rao, R. P. N. (Eds.). (2007).Bayesian brain: Proba- bilistic approaches to neural coding. MIT Press. Farama Foundation. (2023). Gymnasium-robotics: PointMaze environments [Accessed: 2024]. https://robotics.farama.org/envs/maze/point maze/ 76 FitzGerald, T. H. B., Dolan, R. J., & Friston, K. (2015). Dopamine, reward learn...

  3. [3]

    Neural Computation , volume=

    Friston, K. J., Salvatori, T., Isomura, T., Tschantz, A., Kiefer, A., Verbelen, T., Koudahl, M., Paul, A., Parr, T., Razi, A., Kagan, B. J., Buckley, C. L., & Ramstead, M. J. D. (2025). Active inference and intentional behavior.Neural Computa- tion,37(4), 666–700. https://doi.org/10.1162/neco a 01738 Gershman, S. J. (2018). The successor representation: I...

  4. [4]

    Pezzulo, G., Rigoli, F., & Friston, K. J. (2018). Hierarchical active inference: A theory of motivated control.Trends in cognitive sciences,22(4), 294–306. Proietti, R., Pezzulo, G., & Tessari, A. (2023). An active inference model of hierarchical action understanding, learning and imitation.Physics of Life Reviews,46, 92–

  5. [5]

    Rao, R. P. N. (2024). A sensory-motor theory of the neocortex.Nature Neuroscience, 27, 1221–1235. Rao, R. P. N., & Ballard, D. H. (1999). Predictive coding in the visual cortex: A func- tional interpretation of some extra-classical receptive-field effects.Nature neu- roscience,2(1), 79–87. 80 Rao, R. P. N., Gklezakos, D. C., & Sathish, V . (2024). Active ...

  6. [6]

    C., Muller, T

    https://doi.org/10.1038/s41467-024- 54257-3 Whittington, J. C., Muller, T. H., Mark, S., Chen, G., Barry, C., Burgess, N., & Behrens, T. E. (2020). The Tolman-Eichenbaum machine: Unifying space and relational memory through generalization in the hippocampal formation.Cell,183(5), 1249–1263. 82