Meet Ledger Nano™ Gen5, the most playful signer yet

Discover now

Thought leadership | 02/10/2026

Agentic AI Is Loose. Your Security Model Is Not Ready

Ledger CTO, Charles Guillemet, explains the emergent risks of Agentic AI and how a a hardware-enforced architecture where Agents Propose and Humans Sign, keeps you in control of every action.

Before You Dive In:

  • Agentic AI is rapidly transitioning from research demos to production tools, but this speed has created a massive structural risk where autonomous agents act as “Trojan Horses” with elevated privileges on your local machine.
  • The combination of untrusted inputs and broad execution power granted to systems like OpenClaw can allow attackers to quietly exfiltrate sensitive data and manipulate on-chain transactions.
  • This is why Ledger advocates for the “Agents Propose, Humans Sign” model—a hardware-enforced architecture that ensures AI never accesses your private keys and humans remain the final authority for every action.

I’ve spent most of my career looking at new technologies at the moment they start to break. At Ledger Donjon, their job is to assume that every new system, no matter how promising, will eventually be used in ways its creators did not intend. That mindset comes from experience. Browsers, mobile apps, cloud infrastructure, smart contracts, every wave follows the same arc. Capability grows faster than security.  Adoption comes first, then comes learning. Attackers might initially fail, but they learn the fastest.

Agentic AI is following that pattern, except this time the acceleration is extreme.

Over the past months, we have seen agents move from research demos to production tools almost overnight. They schedule meetings, respond to emails, scrape the web, manage trading workflows, and execute code locally. The pitch is convenience. The trade-off is security and control. Once you allow an agent to run on your machine, you are no longer delegating tasks. You are delegating execution power inside your own security perimeter, pointing to some arbitrary external code that can leverage all your privileges.

That distinction matters more than most people realize.

OpenClaw, ClawdBot, Molbot: Things Are Moving Fast

One challenge with talking about agentic AI right now is that the landscape changes faster than the vocabulary.

OpenClaw is a framework and infrastructure that gives developers a way to build agents that crawl the web, load skills, and execute workflows. 

ClawdBots are the agents built on top of that framework. These are the tools people actually run on their machines. They combine web crawling, skill execution, and local system access.

Moltbot is simply the previous name for this same project rather than a separate tool. It is the exact same code and core idea. Local autonomy plus broad permissions.

Then you have co-work style agents, designed to collaborate with users across files, terminals, and internal tools. Different branding, same risk profile.

If this feels messy, that’s because it is. Frameworks, bots, skills, coworkers, all blurring together while capabilities expand week by week. These AI frameworks are built very quickly thanks to AI!

This Looks Familiar If You’ve Worked in Security Long Enough

Years ago, we saw the same pattern with browser extensions, on a much smaller scale. ActiveX and Flash died a death driven by critical security vulnerabilities. 

People installed helpers that promised productivity and convenience. Over time, those helpers gained more permissions, gaining access to pages, storage, and even credentials. Eventually, attackers realized that compromising one extension gave them leverage over millions of users.

Then came supply chain attacks. A trusted dependency gets poisoned, and this attack travels through the host, carrying malicious logic – infecting the whole system.

Agentic AI compresses all of this into a single system. An agent can ingest untrusted input, reason over it, execute commands locally, and communicate externally, all without a human in the loop.

From a security perspective, that combination should raise alarms. This is the recipe for disaster. You basically have a Trojan horse in your machine with elevated privileges.

The Incidents Were Predictable

When reports surfaced about agentic skills and cowork style agents being abused, no one at Ledger or the Donjon was surprised. The mechanics were straightforward. A useful-looking automation with convincing instructions, a local command executed with too much trust.

The outcomes were familiar. Files were accessed while credentials were exposed, as data was quietly exfiltrated – withdrawn without the user’s knowledge. Data might include passwords, sensitive pictures, and IP addresses. Anything that can be sold, used as a ransom, used for identity theft, or used for direct financial or reputational harm.

Public examples have started to pile up, across different agent stacks and use cases:

And there’s this Threat Intelligence report from Prompt Armor, explaining how Claude Cowork is vulnerable to file exfiltration attacks via indirect prompt injection.

Attackers are not doing anything novel here. This is social engineering with better tooling and more leverage.

Agentic AI Meets The Onchain Economy

This is now spilling directly on-chain.

Agents are starting to submit transactions, manage positions, and interact with protocols without continuous human involvement.

We are seeing this direction openly across the ecosystem. Circle is experimenting with agent-driven use cases through hackathons and developer programs.

Solana is enabling high-frequency, low-cost, onchain actions that suit autonomous systems; Base is pushing agent-friendly application patterns into production environments.

This is already in production.

From a security standpoint, that matters.

Once agents can initiate on-chain actions, security and control become the core question. Blockchains execute exactly what they receive. They do not know whether the intent originated from a human or from an automated system.

If an attacker can inject instructions into an agent’s context, through prompt injection, malicious skills, or poisoned inputs, the agent becomes the execution layer for the attacker. The permissions are already there. Attackers are always acting according to opportunity-cost principles. If they can directly steal money by attacking some systems, they will spend lots of time and effort attacking these. These models will be hacked again and again. People will lose money.

The risk is structural. Autonomous execution combined with irreversible settlement leaves very little margin for error.

The Pattern Has Been Documented Before

Years ago, Simon Willison described a failure mode that maps cleanly onto what we are seeing today.

  • Untrusted input.
  • Powerful execution.
  • Exfiltration.

Agentic systems hit all three conditions by default. They ingest external content continuously. They are designed to act on that content. They almost always have network access.

Once those conditions exist in the same system, exploitation becomes a question of timing. The failure happens at the architecture level.

If you want to go deeper, his writing on this remains one of the clearest references.

Where OpenClaw Fits Into This

As more incidents surfaced, researchers started to look closely at how OpenClaw-based skills were being abused in practice. What they found aligned perfectly with this pattern.

Malicious skills posed as legitimate automations. Users were encouraged to run commands locally. The agent executed them with full permissions. Sensitive data left the system.

A detailed breakdown of how this played out in the wild is available here.

Frameworks behave as expected. Boundaries are missing.

Autonomy Breaks How Humans Reason About Systems

Humans expect causality. You perform an action, and you observe a result, receiving an explainable outcome.

Agents do not behave this way. They chain decisions. They react to inputs you may never see. They operate asynchronously. When something goes wrong, users are often unable to reconstruct the sequence.

That loss of observability becomes a governance problem. Authority is automated. Accountability fades unless it is deliberately enforced.

Hardware Separation Alone is Not a Solution

I have seen many people respond to this risk by running agents on separate machines, Mac Minis, Raspberry Pis, and spare laptops. The assumption is that physical separation creates safety; it just depends on how you implement it. Isolation, and having dedicated hardware to do security stuff is a good security principle…but if you put all your credentials in another machine, it’s not a fail-safe solution.

Those machines are still connected to the internet. They still execute software. They still lack a trusted approval mechanism for sensitive actions. Without a way to verify and authorize what is happening, separation is cosmetic.

Security boundaries require enforcement.

Why We Insist on Secure Hardware-Backed Approval

At Ledger, we learned early that connected environments cannot be trusted by default. Screens can lie. Software can be manipulated, and inputs can be altered without the user knowing.

That is why high-value actions must be approved outside the environment where they are generated. When approval happens on a secure, isolated screen, controlled by hardware that never exposes keys to the host, the attack surface changes fundamentally.

This is the model behind Ledger’s Secure Element architecture and Clear Signing, which were designed specifically to prevent blind approvals in hostile environments:

An agent can propose an action. It cannot approve it. This should be the principle.

That separation between proposal and authorization is what keeps systems resilient under adversarial conditions.

We Have Seen This Cycle Too Many Times

Agentic AI will continue to evolve. That is inevitable. What is not inevitable is deploying it without boundaries, isolation, and explicit human approval for high-value actions.

Every major incident I have investigated over the years has shared one root cause. Trust was assumed where it should have been verified.

If your architecture cannot clearly separate what an agent suggests from what a human authorizes, then it is primed to fail under pressure.

At scale, systems like that do not degrade gracefully; they break loudly.

And sometimes, they break in ways that are unintentionally funny.

Last week, someone posted this on X:

Then there was this near miss:

Security usually fails quietly. Sometimes, it fails during standup. But not everyone is in a rush to give Agentic AI the keys to their life. 

Though I’m not endorsing the security of @jordanlyall‘s setup, his principles are sound. With one addition. If you don’t enforce security with a hardware wallet, it will get pwned.

Agents Propose, Humans Sign

You must never give an AI your private keys.  Instead, let the AI think, search, and propose. The actual signing authority, the power to move value, must remain exclusively with you. 

Your private key never enters the AI’s “brain.” It remains locked in the Secure Element. This isn’t a policy; it is a technical reality enforced by the hardware.

My colleague, Ian Rogers, Ledger’s CXO, jumped right in, building Agent Payments for the recent @circle hackathon. It’s not an official Ledger project, but the pattern is clear, and it scales:

  1. Agents Propose: The AI suggests the action.
  2. Humans Sign: You verify the intent via Clear Signing. 
  3. Ledger Enforces: Your signer ensures the key never leaves the device.

AI can certainly do the heavy lifting, but humans must always remain the ultimate authority. In our world, that authority doesn’t live in software or a cloud policy; it lives in the signer you hold in your hand. 


Authored by Ledger CTO, Charles Guillemet

Stay in touch

Announcements can be found in our blog. Press contact:
[email protected]

Subscribe to our
newsletter

New coins supported, blog updates and exclusive offers directly in your inbox


Your email address will only be used to send you our newsletter, as well as updates and offers. You can unsubscribe at any time using the link included in the newsletter. Learn more about how we manage your data and your rights.