Reasoning AI: Tailoring Explanations to Expert Minds

Author: Denis Avetisyan


A new approach leverages AI personas modeled on expert reasoning to deliver explanations that adapt to individual understanding and build trust in complex scientific domains.

The research introduces an adaptive explainability approach composed of two phases: the initial phase constructs agentic personas via embedding techniques, clustering algorithms, and large language model synthesis, while the subsequent phase leverages these personas to condition reward functions within a reinforcement learning framework.
The research introduces an adaptive explainability approach composed of two phases: the initial phase constructs agentic personas via embedding techniques, clustering algorithms, and large language model synthesis, while the subsequent phase leverages these personas to condition reward functions within a reinforcement learning framework.

This work introduces agentic personas, powered by knowledge graphs and reinforcement learning, to provide adaptive explainability for expert-facing AI systems.

Existing explainable AI (XAI) methods often fall short by assuming a uniform user, neglecting the diverse cognitive strategies of experts-a critical limitation in fields demanding deep understanding. This paper, ‘Agentic Personas for Adaptive Scientific Explanations with Knowledge Graphs’, introduces a reinforcement learning framework leveraging agentic personas-structured representations of expert reasoning-to generate explanations tailored to distinct epistemic stances within knowledge graphs. Results demonstrate that persona-driven explanations not only match state-of-the-art performance but also demonstrably align with expert preferences, significantly reducing the need for extensive human feedback. Could this approach unlock a new era of scalable, adaptive explainability for AI systems operating in complex, high-stakes scientific domains?


The Limits of Static Explanation: A Fundamental Flaw

Conventional explanation techniques within knowledge graphs frequently deliver uniform responses, irrespective of the individual seeking information or the specific context of their inquiry. This approach overlooks the inherent diversity in how people process information and construct understanding; a user with a strong statistical background, for instance, may benefit from an explanation emphasizing probabilities and confidence intervals, while another might prefer a narrative tracing a direct causal chain. The limitations of these static explanations become particularly apparent when dealing with complex problems where multiple lines of reasoning are possible, hindering effective decision-making and eroding trust in the system’s recommendations. Essentially, a single explanation, however logically sound, cannot adequately serve the varied needs and cognitive styles of a diverse user base, creating a crucial barrier to widespread adoption and practical utility.

Knowledge graph explanations, while increasingly prevalent, often falter when confronted with genuinely complex inquiries. Static, pre-defined reasoning paths struggle to accommodate the subtleties inherent in real-world problems, delivering explanations that lack the necessary granularity or context. This limitation is particularly acute in critical domains-such as medical diagnosis or financial risk assessment-where decisions carry significant consequences. An inability to provide nuanced, case-specific justifications erodes user trust, hindering effective comprehension and potentially leading to suboptimal or even harmful outcomes. Consequently, the demand for explanation methods that move beyond simple path enumeration and embrace the complexities of individual scenarios is paramount to fostering reliable and responsible AI systems.

While systems like MINERVA and PoLo represent significant advances in knowledge graph reasoning by providing explicit paths of inference, their explanations often fall short of true user utility. These foundational approaches typically deliver a single, pre-determined explanation, irrespective of the user’s prior knowledge, cognitive style, or specific information needs. This lack of adaptability means that a complex explanation, perfectly logical to the system, might overwhelm a novice user, while an expert could find the same explanation unnecessarily verbose or lacking crucial detail. Consequently, the delivered reasoning, though technically correct, may not resonate with – or be effectively utilized by – a diverse audience, highlighting a crucial need for explanations that can be tailored to individual cognitive profiles and epistemic stances.

Effective knowledge transfer hinges not simply on what is explained, but on how it aligns with an individual’s pre-existing beliefs and reasoning processes-their epistemic stance. Current explanation systems largely overlook this critical factor, presenting information uniformly regardless of whether a user favors holistic, pattern-based reasoning or a more analytical, step-by-step approach. This disconnect can lead to explanations being perceived as unhelpful, untrustworthy, or even misleading, particularly when dealing with complex or sensitive information. Addressing this gap requires systems capable of profiling user cognitive styles and tailoring explanations to resonate with their preferred methods of evidence evaluation and understanding construction, ultimately fostering greater comprehension and informed decision-making.

Participants demonstrated a significant preference for adaptive explanations over non-adaptive ones.
Participants demonstrated a significant preference for adaptive explanations over non-adaptive ones.

Agentic Personas: Modeling Expert Reasoning with Precision

Agentic personas are formalized models of expert reasoning developed through an Expert Evaluation Study focused on drug discovery processes. These personas are not intended to be simple user profiles, but rather structured representations of cognitive strategies employed by domain experts when analyzing data and forming conclusions. The development process involved capturing and analyzing expert feedback to identify distinct reasoning patterns, which were then codified into actionable models. These models aim to replicate the nuanced thought processes of experts, allowing for the simulation and application of expert-level reasoning within automated systems. The resulting personas serve as a basis for building AI agents capable of performing complex reasoning tasks in the drug discovery domain.

Agentic personas are implemented through the utilization of large language models (LLMs) to replicate the cognitive processes observed in expert reasoning. Specifically, LLMs are employed to model how experts assess evidence, identify relevant information, and construct hypotheses during problem-solving. This instantiation involves training the LLMs on datasets derived from expert evaluations, enabling them to capture the subtleties of expert interpretation, including weighting of evidence, consideration of alternative explanations, and the application of domain-specific knowledge. The resulting personas can then generate outputs that reflect these nuanced reasoning patterns, effectively mimicking the cognitive styles of the experts from whom they were derived.

The development of agentic personas utilizes LLM-based Narrative Synthesis to process and condense qualitative feedback from expert evaluations. This process begins with the generation of semantic embeddings for each expert response using Sentence-BERT, converting textual data into numerical vector representations. These embeddings are then subjected to clustering analysis employing K-Means, Agglomerative Clustering, and HDBSCAN algorithms to identify recurring themes and patterns in expert reasoning. The LLM then synthesizes these clustered themes into concise, representative narratives, effectively distilling the core insights from the larger dataset of expert feedback and forming the basis for persona definition.

Agentic personas function dynamically within the explanation generation process, moving beyond simple profile lookup. Each persona actively modulates the LLM’s output based on the identified cognitive style it represents. This is achieved through targeted prompting and parameter adjustments during inference, influencing factors such as the level of detail, the type of evidence emphasized, and the overall framing of the explanation. Consequently, the same underlying data can yield substantially different explanations, each optimized to resonate with a specific cognitive preference as determined by the expert evaluation study.

A t-SNE projection reveals two distinct clusters of expert response embeddings, corresponding to the distinct interaction styles of Elena (n=13, light blue) and Leo (n=2, dark blue).
A t-SNE projection reveals two distinct clusters of expert response embeddings, corresponding to the distinct interaction styles of Elena (n=13, light blue) and Leo (n=2, dark blue).

Reinforcement Learning: Dynamically Tailoring Explanations

Reinforcement Learning (RL) is utilized to dynamically tailor explanations generated from a knowledge graph by incorporating agentic personas. This process moves beyond static explanations by training an agent to select and frame explanatory pathways based on the desired persona. The RL agent learns to navigate the knowledge graph, identifying relevant nodes and edges, and then constructs an explanation that aligns with the characteristics and communication style of the assigned persona. This dynamic selection and framing allows the system to generate explanations that are not only factually accurate and relevant, but also specifically suited to the recipient, enhancing comprehension and trust.

The Reward Function is central to the system’s ability to generate persona-aligned explanations; it quantifies the quality of an explanation based on criteria derived from the assigned persona. This is achieved through Persona-Conditioned Reward, a mechanism that dynamically adjusts reward signals based on the specific characteristics and preferences defined by the persona. The function assesses factors such as the explanation’s comprehensiveness, clarity, and relevance, weighting these factors according to the persona’s profile – for example, a technical persona may prioritize precision, while a novice persona prioritizes simplicity. This allows the Reinforcement Learning agent to learn a policy that maximizes reward, effectively tailoring explanations to meet the needs and expectations associated with each persona.

The system employs a Knowledge Graph as the foundational source for explanatory information, enabling the identification of multiple potential paths to address a given query. To prioritize the selection of the most appropriate path, the system integrates with reasoning engines REx and RExLight. These engines evaluate candidate explanations based on fidelity – ensuring alignment with the facts contained within the Knowledge Graph – and relevance, assessing the degree to which the explanation directly addresses the user’s query. REx and RExLight operate by scoring potential explanations, allowing the system to select the highest-ranked path for subsequent narrative generation, thereby grounding explanations in verifiable knowledge and ensuring topical coherence.

Human-readable explanations are generated utilizing the GPT-4o-mini language model, which receives input derived from the knowledge graph and conditioned by the selected agentic persona. Evaluation by subject matter experts indicates a preference for these adaptive explanations in 63.3% to 76.0% of cases when compared to non-adaptive baseline explanations; this statistically significant result demonstrates a measurable improvement in explanation quality as perceived by experts and confirms the effectiveness of the persona-driven adaptive approach.

Utilizing persona-based feedback dramatically reduces training time for explanation policies, achieving up to a 100-fold improvement in efficiency.
Utilizing persona-based feedback dramatically reduces training time for explanation policies, achieving up to a 100-fold improvement in efficiency.

Implications for Trust and Decision Support: Towards Reliable AI

Current artificial intelligence systems often deliver explanations that, while technically accurate, fail to resonate with how an individual user approaches reasoning and understanding-their ‘epistemic stance’. This research demonstrates a shift towards personalized reasoning support by tailoring explanations to match a user’s existing beliefs and cognitive style. Rather than presenting a one-size-fits-all justification, the system adapts its communication to align with whether a user prefers, for example, causal, analogical, or rule-based reasoning. This alignment is crucial because explanations that acknowledge and build upon a user’s existing framework are demonstrably more effective at fostering trust and acceptance of AI-driven insights, moving beyond simply conveying information to genuinely supporting informed decision-making.

The capacity for adaptive explainability holds particular promise within high-stakes fields such as drug discovery, where the complexity of biological systems demands more than simple answers. Effective therapeutic development relies on a nuanced comprehension of why a particular molecule is predicted to be effective, not just that it is. By tailoring explanations to reflect a user’s existing knowledge and reasoning style, this approach facilitates a deeper, more critical engagement with AI-generated insights. This is crucial for researchers evaluating potential drug candidates, allowing them to assess the validity of predictions, identify potential biases, and ultimately make more informed decisions that could accelerate the development of life-saving treatments. The ability to move beyond opaque ‘black box’ predictions towards transparent, personalized reasoning support fosters trust and enables effective human-AI collaboration in this complex and vital domain.

Beyond simply delivering an output, advanced explainability aims to illuminate how an AI system arrived at a particular conclusion, fostering genuine comprehension rather than blind acceptance. This approach doesn’t prioritize merely confirming a ‘correct’ answer, but instead empowers individuals to dissect the reasoning process itself, enabling critical assessment of the information presented. By revealing the underlying logic – the chain of inferences, the data considered, and the assumptions made – users can move beyond surface-level validation to identify potential biases, limitations, or errors within the system’s reasoning. This deeper understanding is crucial not only for building trust, but also for informed decision-making, particularly in complex scenarios where simply knowing what is recommended is insufficient without grasping why.

Recent research demonstrates a substantial increase in the efficiency of training explainable AI systems through the innovative use of carefully constructed personas for feedback. Compared to traditional methods relying solely on full human feedback, this approach achieved training time reductions of up to 187x. Critically, the alignment between these persona-driven evaluations and those of human experts was remarkably strong, as evidenced by Pearson correlation coefficients ranging from 0.56 to 0.91. These findings suggest that personas can serve as a highly effective proxy for human judgment, accelerating development cycles and reducing the resource demands associated with building trustworthy and transparent AI solutions.

Persona-conditioned explanations adapt to provide more relevant insights into treatments like Fenofibrate for Coronary Artery Disease compared to non-adaptive explanations.
Persona-conditioned explanations adapt to provide more relevant insights into treatments like Fenofibrate for Coronary Artery Disease compared to non-adaptive explanations.

The pursuit of adaptive explainability, as detailed in this work, demands a rigorous foundation akin to mathematical proof. The framework’s reliance on agentic personas-LLM-derived representations of expert reasoning-highlights the need for logically complete and non-contradictory representations of knowledge. As Grace Hopper famously stated, “It’s easier to ask forgiveness than it is to get permission.” This resonates with the iterative nature of building complex systems; the agentic personas aren’t perfect from inception, but through reinforcement learning and conditioning on diverse epistemic stances, they refine their reasoning, mirroring the pragmatic approach Hopper advocated. The goal isn’t merely functional output, but provable correctness in explaining complex scientific concepts, especially within domains like drug discovery.

What’s Next?

The pursuit of adaptive explainability, as demonstrated by this work, inevitably encounters the limitations inherent in modeling expertise. While agentic personas offer a promising avenue for tailoring explanations, the true test lies not in simulating epistemic stances, but in formally verifying the correctness of the reasoning they embody. Current approaches rely heavily on the heuristic power of large language models; a concerning reliance given their acknowledged propensity for generating plausible, yet demonstrably false, statements. The field must move beyond behavioral validation-showing explanations appear helpful-toward rigorous mathematical proofs of their logical consistency.

A critical unresolved problem remains the integration of uncertainty. Scientific reasoning is rarely absolute; it is fundamentally probabilistic. Future work should explore methods for encoding and propagating uncertainty within these agentic frameworks, allowing explanations to reflect the degree of confidence in underlying inferences. Furthermore, the subjective nature of “trust” requires deeper investigation; a truly robust system will need to quantify the alignment between an explanation and a user’s prior knowledge – a task demanding a formal epistemology, not merely a statistical approximation.

In the chaos of data, only mathematical discipline endures. The challenge is not simply to explain scientific findings, but to ensure those explanations are themselves unassailable. The ultimate goal is not merely an interpretable AI, but a provably correct one; a system where every inference can be traced back to first principles, and every explanation is a logical consequence of established truths.


Original article: https://arxiv.org/pdf/2603.21846.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-03-24 18:48