Author: Denis Avetisyan
As AI agents take the reins of financial transactions, understanding and mitigating their unique security vulnerabilities is paramount.

This paper provides a systematic security assessment of fully autonomous large language model agents operating within agentic commerce, identifying cross-layer attack vectors and proposing a layered defense architecture.
While the promise of autonomous agents powered by large language models (LLMs) is rapidly expanding agentic commerce-encompassing negotiation, transactions, and asset management-existing security frameworks fail to adequately address the novel risks inherent in these systems. This paper, ‘SoK: Security of Autonomous LLM Agents in Agentic Commerce’, delivers a systematic analysis of this emerging threat landscape, identifying twelve cross-layer attack vectors stemming from vulnerabilities in reasoning, tooling, and underlying protocols. Our work reveals that securing agentic commerce demands coordinated controls spanning LLM safety, protocol design, identity management, and regulatory compliance, proposing a layered defense architecture to mitigate authorization gaps. What research and benchmarks are needed to build truly secure and trustworthy autonomous commerce ecosystems?
The Inevitable Bottleneck: Why Control Fails
Contemporary artificial intelligence, despite significant advancements, frequently necessitates continuous human intervention to ensure accuracy and prevent unintended consequences. This reliance on oversight creates a bottleneck, dramatically limiting the potential for widespread automation and true scalability. Current systems often excel at narrow, defined tasks, but struggle with adaptability and independent decision-making in dynamic, real-world scenarios; a human operator must validate outputs, correct errors, and manage unforeseen circumstances. This constant need for supervision not only increases operational costs but also prevents AI from reaching its full potential in applications demanding 24/7 operation or rapid response times, effectively confining its capabilities within the bounds of human availability and attention.
Large Language Model (LLM)-based agent systems represent a significant leap towards true automation by enabling artificial intelligence to move beyond simply responding to requests and begin acting on them independently. These systems leverage the reasoning and generative capabilities of LLMs not just for language processing, but to orchestrate complex tasks – from booking travel and managing finances to conducting research and interacting with various online services. Unlike traditional AI requiring constant human intervention, these agents can autonomously execute transactions, adapt to changing circumstances, and pursue defined goals without continuous oversight. This capability stems from their ability to break down objectives into sequential actions, utilize tools and APIs, and learn from both successes and failures, promising a future where AI handles routine and even complex processes with minimal human direction and unlocking scalability previously unattainable.
As autonomous agents proliferate, reliance on centralized control mechanisms becomes increasingly untenable and inefficient. The emerging paradigm demands sophisticated protocols governing agent-to-agent interactions, fostering a decentralized web of trust. These protocols must address critical challenges like verifying information authenticity, ensuring transactional integrity, and resolving conflicts without human intervention. Current research focuses on techniques such as cryptographic verification, reputation systems, and consensus algorithms – adapting blockchain principles and game theory to build reliable frameworks for multi-agent systems. Successfully navigating this shift requires a move away from top-down authority toward a system where agents can autonomously negotiate, collaborate, and establish trust based on verifiable data and pre-defined rules, ultimately unlocking the full potential of truly independent artificial intelligence.
The Architecture of Decentralized Trust
Virtuals Protocol and Agent Commerce Protocol (ACP) establish a decentralized framework for the development and economic participation of autonomous AI agents. Virtuals provides the foundational layer, enabling the creation of agent identities and secure communication channels. ACP builds upon this by defining standardized interfaces for agents to offer services, negotiate terms, and execute transactions. This infrastructure operates on a peer-to-peer network, eliminating the need for central intermediaries and allowing agents to directly interact with each other and with users. Key functionalities include agent discovery, service registration, and automated payment settlement, all facilitated through cryptographic verification and smart contract execution. The protocols are designed to support a wide range of agent-based services, from data analysis and content creation to automated trading and logistical operations.
Standardized payment protocols are critical for enabling scalable agent commerce. AP2 (Autonomous Payments Protocol 2) defines a framework for recurring and conditional payments between agents, ensuring automated and reliable transactions. The x402 protocol, built upon OIDC and FAPI, provides secure authorization flows for payment initiation, verifying agent identity and permissions. Meanwhile, MPP (Merchant Payment Protocol) focuses on streamlining the communication between agents and payment processors, reducing latency and transaction costs. These protocols collectively address security concerns through standardized encryption and authentication mechanisms, and enhance efficiency by automating payment flows without requiring manual intervention or centralized intermediaries.
Ethereum Request for Comments 8183 (ERC-8183) defines a standard for sending and receiving tokens on behalf of an agent, enabling autonomous operation without direct user intervention for each transaction. This differs from traditional account-based transfers by allowing agents to act as intermediaries, executing token operations based on predefined logic. Complementing this, ERC-8004 introduces a framework for agent-controlled token operations, specifically outlining how agents can manage and utilize tokens held in escrow or through delegation. These standards facilitate trustless transactions by leveraging smart contracts to enforce pre-agreed conditions and authorize token movements, eliminating the need for centralized intermediaries and increasing security within agent commerce systems. Both standards are designed to be compatible with existing Ethereum infrastructure and tooling, promoting interoperability and ease of integration.
The Illusion of Security: Layered Defenses in a Hostile System
Prompt injection represents a critical security vulnerability in LLM-based agent systems, stemming from the LLM’s inherent inability to consistently distinguish between legitimate instructions and malicious commands embedded within user-provided input. This allows an attacker to manipulate the agent’s behavior, potentially causing it to disregard its intended purpose, disclose confidential information, execute unauthorized actions, or generate harmful content. The risk is amplified as agents become more autonomous and interact with external tools and systems, as a compromised agent can leverage these integrations to inflict wider damage. Successful prompt injection attacks do not necessarily require access to the underlying model or system code; instead, they exploit the LLM’s natural language processing capabilities to override existing safeguards and redirect the agent’s operational flow.
A Layered Defense Architecture is essential for securing LLM-based agent systems against evolving threats. This architecture incorporates multiple security checkpoints across different system layers to reduce the risk of successful attacks. Complementing this is the implementation of robust Transaction Authorization systems leveraging Smart Contract technology; these systems enforce pre-defined rules and permissions for all agent transactions. Current research identifies 12 distinct cross-layer attack vectors – vulnerabilities that exploit weaknesses across multiple system layers simultaneously – that this combined approach is designed to address, including data injection, privilege escalation, and unauthorized access to resources. The systematization of these attack vectors allows for targeted implementation of security measures and continuous monitoring of system integrity.
Decentralized Identity (DID) and escrow mechanisms function as critical trust-building components within multi-agent systems. DIDs provide agents with verifiable, self-sovereign digital identities, eliminating reliance on centralized authorities and enabling secure authentication and authorization. Escrow services introduce a third-party holding of assets or commitments during transactions between agents, releasing them only upon fulfillment of pre-defined conditions. This mitigates the risk of fraud or non-performance by ensuring that neither agent can unilaterally benefit at the expense of the other; funds or commitments remain secure until contractual obligations are met, fostering reliable and predictable inter-agent relationships and reducing counterparty risk.
The Inevitable Cascade: Scaling Autonomy, Accepting Consequences
Agent-based systems, fueled by the capabilities of large language model (LLM)-based agents, represent a significant paradigm shift with the potential to reshape complex operational landscapes. These systems move beyond traditional automation by enabling autonomous entities to negotiate, collaborate, and problem-solve with a level of flexibility previously unattainable. In supply chain management, this translates to self-optimizing logistics, dynamic sourcing, and proactive disruption mitigation. Decentralized finance benefits through automated market making, personalized financial instruments, and enhanced security protocols. The core innovation lies in the LLM’s ability to understand nuanced requests, interpret complex data, and generate adaptive responses, allowing agents to operate with minimal human intervention and fostering a new era of efficiency and resilience across these critical sectors.
OpenClaw establishes a streamlined pathway for deploying autonomous agents into functional economic systems by integrating them with blockchain-based digital wallets. This framework allows agents to independently manage and transact with digital assets, facilitating decentralized commerce and complex supply chain interactions without centralized intermediaries. By leveraging blockchain’s security and transparency, OpenClaw ensures that all agent transactions are recorded and verifiable, building trust and accountability within the autonomous network. The practical implementation details focus on creating a secure and efficient bridge between agent actions and on-chain financial operations, allowing for scalable deployment across various applications – from automated trading and decentralized finance protocols to managing logistics and verifying data integrity in complex networks.
Within the Autonomous Collaboration Pipeline (ACP) framework, Evaluator Agents function as critical arbiters in agent-to-agent transactions, establishing a system of quality control and accountability previously absent in decentralized commerce. These specialized agents don’t simply execute tasks; they independently verify the outputs of other agents, assessing whether completed work meets predefined standards and contractual obligations. This verification process isn’t a centralized imposition, but rather a distributed consensus built upon the evaluations of multiple agents, creating a robust and tamper-proof audit trail. Consequently, the presence of Evaluator Agents dramatically reduces the risk associated with automated transactions, fostering trust and encouraging wider adoption of agent-based systems across diverse applications – from complex supply chain logistics to the rapidly evolving landscape of decentralized finance, ultimately building a more reliable and secure ecosystem for autonomous collaboration.
The study of autonomous LLM agents reveals a predictable pattern: systems, born of ambition, inevitably grapple with unforeseen consequences. Every dependency established, every smart contract deployed, is a promise made to the past, a commitment to a future that cannot be fully known. As Bertrand Russell observed, “The difficulty lies not so much in developing new ideas as in escaping from old ones.” This rings true as attempts to ‘control’ these agents through layered defenses often prove illusory, merely shifting vulnerabilities rather than eliminating them. The architecture isn’t a fortress, but an ecosystem-a complex web of interactions where failure isn’t an endpoint, but a catalyst for adaptation and self-correction. Everything built will one day start fixing itself, a cycle inherent in all complex systems.
The Looming Shadows
This analysis of autonomous agents in financial systems doesn’t reveal vulnerabilities so much as chart the inevitable avenues of their emergence. Each proposed defense-layered as it is-becomes, in time, merely another stratum for the next attack to burrow through. The very notion of “securing” such a system implies a static target, while the agents themselves, and the models that birth them, are in perpetual flux. The true cost won’t be the exploitation of prompt injection, but the slow realization that every automated decision carries the ghost of unforeseen consequence.
The focus on cross-layer attacks is astute, yet reveals a lingering faith in architectural boundaries. Systems aren’t built; they grow, and the tendrils of compromise will inevitably find the weakest capillary between layers – or simply bypass them entirely. Future work will not be about patching vulnerabilities, but about accepting the inherent ephemerality of control. The challenge isn’t to prevent failure, but to design for graceful degradation – a willingness to relinquish automation before it becomes calcified risk.
Ultimately, the field will shift from seeking perfect agents to understanding the ecology of imperfect ones. Monitoring won’t be about detecting malicious intent, but about tracking the subtle drift of behavior, the quiet erosion of trust. The question isn’t whether these agents will fail, but what form that failure will take, and how readily it will propagate through the interconnected web of automated finance.
Original article: https://arxiv.org/pdf/2604.15367.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- Quantum Agents: Scaling Reinforcement Learning with Distributed Quantum Computing
- Boruto: Two Blue Vortex Chapter 33 Preview — The Final Battle Vs Mamushi Begins
- All Skyblazer Armor Locations in Crimson Desert
- Every Melee and Ranged Weapon in Windrose
- How to Get the Sunset Reed Armor Set and Hollow Visage Sword in Crimson Desert
- One Piece Chapter 1180 Release Date And Where To Read
- New Avatar: The Last Airbender Movie Leaked Online
- USD RUB PREDICTION
- Zhuang Fangyi Build In Arknights Endfield
- All Shadow Armor Locations in Crimson Desert
2026-04-20 14:08