Modeling Minds: When AI Reveals How We Think

Author: Denis Avetisyan


A new framework treats artificial agents as models of human cognition, allowing researchers to statistically evaluate how underlying cognitive processes drive behavior.

Agentic Behavioral Modeling leverages Bayesian inference and reinforcement learning to analyze human behavior through the lens of computational agents.

Despite advances in computational cognitive science, formally bridging theoretical models of cognition with quantitative behavioral analysis remains a persistent challenge. This paper, ‘On Agentic Behavioral Modeling’, introduces a novel framework wherein artificial agents are treated as generative hypotheses about underlying cognitive mechanisms, evaluated directly through their capacity to explain observed human behavior. By formalizing task-agent-data systems as joint probability models, we demonstrate the utility of this approach on minimal laboratory paradigms, revealing connections between Bayesian inference and reinforcement learning-and offering a new, agent-centric interpretation of established psychometric functions. Could this agentic approach provide a unifying framework for cognitive behavioral science and unlock new insights into the computational principles governing human decision-making?


Emergent Order: Reconciling Mind, Task, and Action

For decades, cognitive science has faced a fundamental challenge: reconciling the complexities of what a person must do – the task demands – with who is doing it – encompassing an individual’s inherent characteristics, motivations, and limitations – and then aligning both with the actual behaviors observed. Traditional approaches frequently treat these elements as intertwined, leading to models that, while potentially descriptive, lack the ability to dissect why a particular behavior emerged in a specific context. This often results in a ‘black box’ effect, where a model can predict an outcome without revealing the underlying cognitive mechanisms. The difficulty lies in the fact that human behavior isn’t simply a response to stimuli, but a dynamic interplay between external pressures and internal states, a nuance often lost in attempts to create universally applicable cognitive frameworks.

Agentic Behavioral Modeling (ABM) presents a notable advancement in understanding complex behaviors by deliberately dissecting the interplay between task requirements, the characteristics of the agent performing the task, and the resulting observable actions. Unlike traditional approaches that often conflate these elements, ABM treats each as a distinct, manipulable component within a computational framework. This separation allows researchers to systematically investigate how specific agent properties – such as memory capacity, attentional biases, or motivational states – interact with task demands to produce particular behavioral outcomes. Consequently, ABM fosters the creation of models that are not only more aligned with the intricacies of real-world behavior, but also inherently more interpretable, as the influence of each component can be readily assessed and understood. The resulting clarity offers a pathway toward predicting behavior in novel situations and uncovering the underlying mechanisms driving it.

Agentic Behavioral Modeling represents a significant departure from traditional approaches that often prioritize simply describing behavior to instead focusing on the underlying computational processes that generate it. By explicitly defining agents with internal states, goals, and decision rules, ABM moves beyond correlational analyses and allows researchers to formulate testable hypotheses about the mechanisms driving observed actions. This shift enables the construction of models that aren’t merely predictive, but explanatory – revealing how an agent, whether biological or artificial, arrives at a particular behavior. Consequently, ABM opens doors to a deeper, mechanistic understanding of complex phenomena, facilitating insights into cognitive processes and offering the potential to extrapolate model behavior to novel situations with greater confidence.

The Principle of Least Surprise: A Foundation for Action

The Free Energy Principle (FEP) proposes that any self-organizing system, considered an agent, operates to minimize [latex]F = E_q(D) – K[q(D)][/latex], termed ‘free energy’. Here, [latex]E_q(D)[/latex] represents the expected surprise or epistemic error regarding sensory data [latex]D[/latex] under an approximate probability distribution [latex]q[/latex], and [latex]K[q(D)][/latex] is a measure of complexity, penalizing overly complex models. Minimization is achieved by updating the agent’s internal model – its beliefs about the causes of its sensations – to better predict incoming sensory input. Essentially, agents actively seek to reduce the difference between their predictions and actual experiences, thereby maintaining a stable internal state and minimizing surprise. This minimization is not necessarily conscious; it’s a fundamental principle governing the dynamics of any system attempting to maintain its integrity in a changing environment.

Perceptual Inference and Policy Selection are the mechanisms by which the Free Energy Principle is realized in an agent’s behavior. Perceptual Inference involves the agent constructing and maintaining an internal probabilistic model of its environment, allowing it to predict sensory inputs and minimize the difference between predicted and actual sensations. Simultaneously, Policy Selection determines which actions the agent will take, based on its internal model and goals; actions are chosen to alter sensory input in a way that confirms the agent’s predictions and minimizes expected free energy. These two processes are tightly coupled; accurate perception informs effective action, and actions, in turn, refine the internal model through Bayesian updating, creating a continuous cycle of prediction and response.

Active Inference operationalizes the Free Energy Principle by positing that an agent’s behavior is a cyclical process of generating predictions about incoming sensory data and subsequently selecting actions to minimize the difference between predicted and actual sensations – this difference being quantified as free energy. Specifically, an agent maintains a generative model that encodes beliefs about the causes of its sensations; predictions derived from this model are compared to observed sensory input. Discrepancies drive the selection of actions not to maximize reward, but to minimize surprise by changing the sensory input to better match the agent’s predictions. This process constitutes a continuous loop of prediction, action, and updated belief, effectively framing all behavior as an attempt to satisfy predicted sensory states and maintain homeostasis with the environment. [latex]Free\ Energy = D_{KL}(Q(x|o) || P(o|x))[/latex], where [latex]Q[/latex] is the approximate posterior, [latex]P[/latex] is the true posterior, and [latex]D_{KL}[/latex] is the Kullback-Leibler divergence.

Implementing the Free Energy Principle and Active Inference necessitates computational methods capable of representing and manipulating probabilistic beliefs about the world. Bayesian Inference provides a formal framework for updating these beliefs given new evidence, utilizing [latex]P(H|E) = \frac{P(E|H)P(H)}{P(E)}[/latex], where [latex]P(H|E)[/latex] represents the posterior probability of a hypothesis (H) given evidence (E). However, exact Bayesian Inference is often intractable for complex agents and environments. Variational Inference offers an approximate solution by formulating an optimization problem to find a probability distribution that closely matches the true posterior, allowing for efficient computation of beliefs and subsequent action selection. These methods provide the necessary precision to translate theoretical principles into functional agent behavior by quantifying uncertainty and enabling optimal inference and decision-making.

Formalizing Uncertainty: Modeling Beliefs and Actions

Partially Observable Markov Decision Processes (POMDPs) are a mathematical framework used to model decision-making in situations where the agent’s state is not fully known. Unlike standard Markov Decision Processes (MDPs), POMDPs incorporate a belief state – a probability distribution over possible states given the history of observations and actions. A POMDP is defined by a tuple [latex](S, A, O, T, R, \Omega, \gamma)[/latex], where S is the state space, A the action space, O the observation space, T the transition function, R the reward function, Ω the observation function, and γ the discount factor. The agent selects actions based on its current belief state, receives an observation, updates its belief using Bayes’ rule, and receives a reward. This process allows for the representation of uncertainty and the development of optimal policies for sequential decision problems where complete state information is unavailable.

Integrating Partially Observable Markov Decision Processes (POMDPs) into Agent-Based Models (ABMs) enables a detailed representation of agent cognition by explicitly modeling beliefs as probability distributions over possible states, given observations. This contrasts with traditional ABMs which often treat agent states as fully known. Within this framework, an agent’s actions are defined as policies – mappings from beliefs to action choices – and outcomes are determined by a probabilistic state transition function dependent on both the action and the true, but often unobservable, state of the environment. The POMDP component formalizes the process of belief updating based on new observations and the execution of chosen actions, creating a closed-loop system where an agent’s internal state of knowledge directly influences its behavior and subsequent learning.

Rescorla-Wagner Learning (RWL) is a model of classical conditioning that provides a computational mechanism for how agents learn predictive relationships between stimuli and events. The model posits that learning occurs when the difference between an expected reward and the actual reward received creates a prediction error [latex] \delta = \lambda (V – V’) [/latex], where λ is the learning rate, [latex] V [/latex] is the predicted value, and [latex] V’ [/latex] is the actual reward. This prediction error then updates the associative strength between the stimulus and the reward. Specifically, the associative strength [latex] V [/latex] is updated as [latex] \Delta V = \alpha \delta [/latex], where α is a learning parameter. By incorporating RWL into agent-based models, the agent’s beliefs about future events are directly tied to experienced outcomes, enabling the model to simulate learning and adaptation based on prediction errors and subsequent adjustments to associative strengths.

Formalizing adaptive behavior using frameworks like Partially Observable Markov Decision Processes (POMDPs) enables researchers to move beyond qualitative descriptions of behavior to quantitative, testable predictions. By explicitly defining agent beliefs, actions, and environmental dynamics within a computational model, specific hypotheses regarding cognitive mechanisms – such as learning rates, reward sensitivities, or belief update strategies – can be implemented and systematically varied. The resulting model behavior can then be compared against empirical data using statistical methods, allowing for rigorous evaluation of the proposed mechanisms and the quantification of model fit. This approach allows researchers to assess the explanatory power of different cognitive models and determine which mechanisms are most critical for driving observed adaptive behaviors in complex environments.

The Power of Prediction: Evaluating Models and Uncovering Principles

Rational Analysis offers a powerful framework for dissecting cognitive processes by evaluating models not on whether they achieve perfect accuracy, but on how optimally they perform given inherent limitations – be it computational power, time, or available information. This approach shifts the focus from simply describing behavior to understanding the underlying principles that shape it, positing that cognition often represents a pragmatic solution to complex problems. By explicitly incorporating constraints, researchers can determine if observed choices align with what would be expected from a perfectly rational agent operating under similar restrictions, or if deviations suggest systematic biases or adaptive strategies. The core tenet is that even seemingly imperfect behavior can be ‘rational’ when viewed through the lens of limited resources, offering a nuanced understanding of how the mind efficiently navigates a complex world.

Agent-based modeling (ABM) offers a powerful methodology for dissecting the underlying mechanisms of decision-making by contrasting observed behavioral patterns with the predictions of rational models. This approach allows researchers to determine whether individuals consistently make choices that maximize expected utility, given their informational constraints, or if systematic deviations from rationality are present. By simulating populations of autonomous agents, ABM can reveal if observed biases represent inherent limitations in cognitive processing, or if they stem from specific environmental pressures or task demands. Discrepancies between modeled rational behavior and empirical data then highlight potential areas for refining cognitive theories and understanding the constraints that shape human choice, moving beyond simply describing behavior to explaining why it occurs.

Protected Exceedance Probability (PEP) offers a statistically rigorous method for discerning the most likely cognitive model from a set of candidates. This approach moves beyond simple model comparison by calculating the probability that a given model is superior to all others, even accounting for potential model misspecification and parameter uncertainty. Simulations demonstrate the power of PEP; when the generative model accurately reflects the data-producing process, the resulting PEP values consistently approach 1.0, indicating strong evidence in favor of that model’s validity. Conversely, incorrect models yield substantially lower PEP values, providing a clear signal of their inadequacy and bolstering confidence in the selected explanation of observed behavior. This makes PEP a valuable tool for navigating the complexities of cognitive modeling and identifying the most plausible representation of underlying cognitive processes.

Analysis of participant data indicated that a Bayesian learning model, designated Model A1, emerged as the most plausible explanation for observed behavior in approximately 45% of individuals, as evidenced by consistently high Bayesian Information Criterion (BIC) values. This suggests that, for a substantial portion of the population, decisions can be effectively rationalized as inferences drawn from prior beliefs updated with new evidence. Supporting this finding, simulations demonstrated robust parameter recovery – a high degree of correlation between the values used to generate the simulated data and those estimated by the model – bolstering confidence in the model’s ability to accurately capture the underlying cognitive processes. These results collectively suggest that Bayesian principles offer a strong framework for understanding how individuals make decisions under uncertainty, and that the model reliably identifies those parameters driving the observed behavior.

The study’s embrace of Agentic Behavioral Modeling (ABM) as a method for dissecting cognitive processes aligns with a philosophy that prioritizes emergence over imposition. Rather than dictating how cognition should function, ABM proposes artificial agents as testable hypotheses, letting statistical analysis of behavioral data reveal underlying mechanisms. This echoes John Dewey’s sentiment: “Education is not preparation for life; education is life.” Just as Dewey advocated for learning through experience and adaptation, ABM allows cognitive models to evolve based on observed behavior, acknowledging that order arises from local rules and interactions, not from centralized control. The framework’s emphasis on probabilistic inference and model-based analysis reinforces the idea that understanding complex systems requires embracing uncertainty and allowing patterns to self-organize.

What’s Next?

The introduction of Agentic Behavioral Modeling shifts the emphasis from seeking a singular ‘model of the mind’ to evaluating a population of plausible agents. This isn’t progress toward a final answer, but a recognition that the question itself is flawed. The field often fixates on constructing internal representations, as if cognition were centrally planned. ABM acknowledges that complex behavior arises from the interaction of simpler, localized processes – small decisions by many participants produce global effects. The true challenge lies not in perfecting a single agent, but in understanding the dynamics of this agentic ecosystem.

Limitations remain, predictably. Current implementations rely heavily on task-specific modeling. Scaling this approach to more ecologically valid, unstructured environments will require a move away from precisely defined reward functions and toward agents capable of intrinsic motivation and open-ended exploration. Furthermore, the statistical comparison of agents often treats them as black boxes. Greater transparency-mechanistic interpretability-is needed to connect agentic behavior to underlying neural processes, though expecting a perfect mapping may be a category error.

The future likely resides in embracing this inherent messiness. Control is always an attempt to override natural order. Rather than striving for predictive dominance, the field should focus on leveraging agentic models as tools for generating novel hypotheses and exploring the space of possible behaviors. The goal isn’t to find the mind, but to create systems that exhibit interesting cognitive phenomena, and then to observe what emerges.


Original article: https://arxiv.org/pdf/2604.27894.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-05-02 22:54