Pathfinding with the Free Energy Principle

The Free Energy Principle, widely used in neuroscience and machine learning, provides a framework for decision-making under uncertainty. This simulation demonstrates how an agent can navigate a randomly generated maze using this principle. The agent forms beliefs about its environment, updates them based on observations, and selects actions that minimize expected free energy.


Maze and Terrain Generation

We use Prim's algorithm to generate a random maze. The maze is represented as a 2D grid where 1 denotes walls and 0 represents free paths.

The agent starts at and aims to reach , where is the maze size.

For terrain we represent the cells the same way, but just fill random cells on a grid.


Prior Preferences

The agent assigns a preference value to each position in the grid, decreasing with distance from the goal:

where is the goal preference weight and is the Euclidean distance to the goal.

Indeed we are encoding directions here for the agent to find the goal, but its the pathfinding using the FEP we are interested in.


Belief Representation

The agent maintains a belief matrix about the environment:

Observations update beliefs in a local neighborhood around the agent’s current position.


Expected Free Energy (EFE)

At each step, the agent calculates the expected free energy (EFE) for each possible move:

where:


Action Selection via Softmax

The agent selects its next action based on a softmax function over the negative EFE values:

where is an inverse temperature parameter controlling randomness.


Simulation and Visualization

The agent explores the maze, updating its beliefs and path history. The animation below illustrates the agent's movement towards the goal.


Conclusion

This simulation demonstrates how an agent can navigate an unknown environment by minimizing free energy. The method blends exploration and exploitation, making it a robust approach for pathfinding under uncertainty. Future extensions could include dynamic obstacles and multi-agent scenarios.