Author: Denis Avetisyan
As AI agents gain autonomy, understanding their unique security vulnerabilities and building robust defenses is becoming critically important.

This review provides a comprehensive survey of the attack and defense landscape for agentic AI systems, outlining key risks, defense-in-depth strategies, and a taxonomy for future research.
While artificial intelligence promises unprecedented automation, the emergent complexity of agentic systems introduces novel security vulnerabilities distinct from traditional software. This paper, ‘The Attack and Defense Landscape of Agentic AI: A Comprehensive Survey’, presents the first systematic analysis of these risks, mapping the attack surface and outlining potential defense mechanisms for AI agents. Our work establishes a foundational framework for understanding agent security, revealing critical gaps in current approaches and highlighting open challenges in building robust, autonomous systems. As agentic AI becomes increasingly integrated into critical infrastructure, how can we proactively address these evolving security threats and ensure trustworthy operation?
Deconstructing Automation: The Rise of Agentic Systems
The integration of Large Language Models (LLMs) with conventional software marks a fundamental shift in the landscape of automation, moving beyond pre-programmed responses towards systems capable of independent action and decision-making. These āagenticā systems donāt simply react to instructions; they proactively formulate plans, leverage tools, and iteratively refine their approach to achieve specified goals. This represents a departure from earlier automation techniques, where tasks were rigidly defined and outcomes predictable. Instead, agentic systems exhibit a degree of autonomy, allowing them to address complex, ill-defined problems in dynamic environments – a capability previously reserved for human intelligence. The implications of this paradigm shift extend across numerous sectors, promising increased efficiency and innovation, but also demanding a re-evaluation of how systems are designed, secured, and governed.
Agentic AI systems, capable of autonomous action, fundamentally rely on three interconnected design dimensions to successfully complete tasks. The workflow defines the sequence of steps an agent takes, dictating how it breaks down a goal into manageable actions. Crucially, an agentās memory allows it to retain information across interactions, enabling learning and adaptation-without it, each task would begin anew. Finally, tool access grants the system the ability to interact with the external world, leveraging APIs and other resources to gather data or enact changes; a powerful agent is limited without the means to execute its plans. The interplay between these dimensions – a well-defined process, the capacity to remember past experiences, and the ability to utilize external resources – is what distinguishes these systems and unlocks their potential for complex problem-solving.
The escalating sophistication of agentic AI systems introduces a novel spectrum of security challenges beyond the scope of conventional defenses. Traditional cybersecurity protocols, designed to protect static code and well-defined network perimeters, are ill-equipped to handle the dynamic, autonomous behavior and emergent properties of these systems. Each added capability – such as tool utilization, memory recall, and workflow orchestration – creates new potential attack vectors, allowing malicious actors to exploit vulnerabilities in the systemās reasoning, planning, or execution. Moreover, the very nature of agentic systems – their ability to adapt and learn – means that security flaws can manifest unexpectedly and evolve rapidly, rendering static security measures obsolete. This necessitates a fundamental rethinking of security paradigms, moving towards proactive, adaptive defenses that can anticipate and mitigate risks within these complex, evolving AI architectures.
![This visualization details six potential attack vectors [latex]V_1-V_6[/latex] and seven associated security risks [latex]R_1-R_7[/latex] that could compromise AI agents.](https://arxiv.org/html/2603.11088v1/x2.png)
Unveiling the Attack Surface: Agentic Vulnerabilities
Agentic systems, due to their reliance on interpreting and executing instructions based on external inputs, are vulnerable to several injection attacks. Indirect Prompt Injection occurs when an agent processes data from an external source, such as a website, that contains malicious prompts altering its intended behavior. Path Traversal exploits insufficient input sanitization to access files and directories outside the intended working directory. OS Command Injection arises when user-supplied input is incorporated into system calls without proper validation, allowing attackers to execute arbitrary commands on the underlying operating system. These vulnerabilities stem from inadequate input validation and insufficient sandboxing of system calls, creating opportunities for malicious manipulation of agent functionality and potential system compromise.
Analysis of agentic systems, specifically AutoGPT, has revealed multiple concrete vulnerabilities documented through Common Vulnerabilities and Exposures (CVE) entries. Since 2023, at least five CVEs (including CVE-2023-37273, CVE-2023-37274, and CVE-2023-37275) have been identified and analyzed. These vulnerabilities stem from inadequate input validation and insufficient sandboxing of system calls, allowing attackers to manipulate agent behavior through crafted inputs. Detailed reports on these CVEs outline specific attack vectors, including methods for arbitrary code execution and unauthorized data access, alongside recommended mitigation strategies such as stricter input sanitization, enhanced access control mechanisms, and improved isolation of agent processes.
Agentic systems, due to insufficient input validation and access control mechanisms, are vulnerable to manipulation and unauthorized access. Specifically, a lack of robust input sanitization allows malicious actors to inject crafted prompts or commands that bypass intended system safeguards. This can lead to unintended agent actions, data exfiltration, or the execution of arbitrary system commands. Insufficient access control further exacerbates the issue by failing to restrict the agentās ability to interact with sensitive resources or execute privileged operations, enabling broader system compromise. These weaknesses collectively create a significant attack surface, allowing external actors to directly influence agent behavior and potentially gain control over the underlying system.

Fortifying the System: A Defense-in-Depth Strategy
A Defense in Depth strategy necessitates the concurrent implementation of both InputGuardrails and OutputGuardrails to comprehensively sanitize data flows. InputGuardrails operate on incoming data, validating and filtering potentially malicious or unintended prompts before they reach the agent. Conversely, OutputGuardrails function on the agentās generated output, ensuring that sensitive information is not disclosed and that responses adhere to predefined safety and compliance standards. This dual approach mitigates risks associated with both compromised inputs and unintended outputs, providing a more robust security posture than relying on either mechanism in isolation. Effective implementation requires defining clear validation criteria for inputs and sanitization rules for outputs, tailored to the specific application and data handled by the agent.
AccessControl mechanisms, when implemented with the principle of LeastPrivilege, function by granting agents only the minimum necessary permissions to perform assigned tasks. This restricts their ability to access or modify sensitive resources beyond defined parameters, thereby limiting the potential blast radius of a successful attack or compromised agent. Specifically, LeastPrivilege dictates that an agent should not have access to data or functionality it does not absolutely require, reducing the avenues for malicious activity or unintended data breaches. Implementation typically involves role-based access control (RBAC) or attribute-based access control (ABAC) systems, meticulously configured to enforce granular permissions and regularly audited to ensure ongoing adherence to the principle.
Current security implementations within AutoGPT largely concentrate on mitigating the effects of successful attacks through mechanisms like access control and output sanitization. However, a comprehensive analysis demonstrates these defenses are insufficient as they do not address vulnerabilities at earlier stages of operation. This study expands the scope of security considerations to include input guardrails, which validate incoming data, and information flow control, which restricts data propagation between agent components. By evaluating these multiple defense dimensions-input guardrails, output sanitization, information flow control, and access control-research indicates that fundamental vulnerabilities remain unaddressed, highlighting the need for a more holistic security strategy.

Beyond Reaction: Proactive Design and the Question of Trust
The efficacy of agentic systems hinges significantly on the reliability of information they process; therefore, designing with InputTrust – a rigorous evaluation of external data sources – is no longer optional, but foundational. These systems, capable of autonomous action, are inherently vulnerable if exposed to compromised or malicious data, potentially leading to flawed reasoning, incorrect outputs, and even harmful consequences. A comprehensive InputTrust framework necessitates not just verifying data authenticity, but also assessing the sourceās reputation, potential biases, and the likelihood of manipulation. By prioritizing trustworthy inputs, developers can mitigate risks across the confidentiality, integrity, and availability (CIA) triad, fostering more robust and dependable agentic technologies and building user confidence in their operation.
Responsible development of agentic systems hinges on a precise understanding of AccessSensitivity – the degree to which an agent can interact with sensitive data. This isnāt simply a binary āaccess grantedā or ādeniedā scenario; rather, itās a spectrum requiring granular control. Overly permissive access dramatically expands the potential attack surface, increasing the risk of data breaches and misuse, while excessively restrictive access can cripple an agentās functionality, rendering it ineffective. Developers must therefore carefully map data sensitivity levels to agent roles and permissions, employing techniques like data minimization – limiting access to only the data absolutely necessary for a task – and differential privacy to obscure individual data points while still enabling meaningful analysis. A thorough assessment of AccessSensitivity is not merely a technical challenge, but an ethical imperative, crucial for building trust and mitigating the potential harms associated with increasingly autonomous systems.
The research details a novel risk taxonomy designed to systematically categorize potential vulnerabilities in agentic systems. This framework moves beyond simple identification, offering a granular analysis of how various attack vectors – encompassing data poisoning, prompt injection, and denial-of-service, among others – intersect with core security concerns related to confidentiality, integrity, and availability – often represented by the CIA triad. Crucially, the study doesnāt simply list risks in isolation; it maps the complex interplay between them, revealing how exploitation of one vulnerability can cascade into others, amplifying potential damage. This interconnected understanding establishes a foundational basis for developing robust defense strategies, allowing developers to prioritize mitigation efforts based on the likelihood and potential impact of combined risks, ultimately fostering more secure and trustworthy agentic technologies.

The survey of agentic AI security reveals a fascinating paradox: the very autonomy designed to solve complex problems introduces novel attack vectors. This echoes Linus Torvaldsā sentiment: āMost good programmers do programming as a hobby, and then they get paid to do it.ā The exploration of prompt injection, access control vulnerabilities, and the need for defense-in-depth isnāt merely about patching flaws; itās about understanding how these systems can be broken, driven by the inherent curiosity to push boundaries. The paper demonstrates that robust security isn’t a destination, but a continual cycle of testing, reverse-engineering, and adaptation – a process mirroring the hackerās mindset, applied for constructive purposes. Itās a testament to the idea that true mastery comes from challenging assumptions and dissecting the underlying mechanisms, rather than blindly accepting pre-defined limitations.
What’s Next?
This survey reveals agentic AI security isn’t merely about patching vulnerabilities; itās about acknowledging that every instruction is a potential exploit. A bug, in this context, isnāt a mistake, but the system confessing its design sins – a transparent admission of the assumptions baked into its core. The current focus on prompt injection, while critical, addresses a symptom. The real challenge lies in constructing agents that fundamentally understand intent, not just parse syntax.
Future work must move beyond defense-in-depth – a reactive layering of safeguards – towards proactive resilience. A truly secure agent shouldnāt just resist malicious prompts; it should gracefully degrade under unexpected input, revealing its uncertainty rather than hallucinating a plausible response. The taxonomy of risks presented here is, inevitably, incomplete; adversarial creativity will always outpace categorization.
Ultimately, the pursuit of secure agentic systems is a forced march toward a deeper understanding of intelligence itself. If an agent can be fooled, it lacks true comprehension. The question isnāt whether these systems can be secured, but whether the very notion of ācontrolā is compatible with genuine autonomy. Perhaps the most fruitful research path involves accepting a degree of unpredictability, and designing systems that can recover-or at least, fail safely-when the inevitable cracks appear.
Original article: https://arxiv.org/pdf/2603.11088.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- CookieRun: Kingdom 5th Anniversary Finale update brings Episode 15, Sugar Swan Cookie, mini-game, Legendary costumes, and more
- Call the Midwife season 16 is confirmed ā but what happens next, after that end-of-an-era finale?
- Robots That React: Teaching Machines to Hear and Act
- PUBG Mobile collaborates with Apollo Automobil to bring its Hypercars this March 2026
- Taimanin SquadĀ coupon codes and how to use them (March 2026)
- Heeseung is leaving Enhypen to go solo. K-pop group will continue with six members
- Robots Learn by Example: Building Skills from Human Feedback
- How to get the new MLBB hero Marcel for free in Mobile Legends
- Alan Ritchsonās āWar Machineā Netflix Thriller Breaks Military Action Norms
- Peppa Pig will cheer on Daddy Pig at the London Marathon as he raises money for the National Deaf Childrenās Society after son Georgeās hearing loss
2026-03-13 10:14