Autonomous LLM brokers like OpenClaw are shifting the paradigm from passive assistants to proactive entities able to executing complicated, long-horizon duties by means of high-privilege system entry. Nevertheless, a safety evaluation analysis report from Tsinghua College and Ant Group reveals that OpenClaw’s ‘kernel-plugin’ structure—anchored by a pi-coding-agent serving because the Minimal Trusted Computing Base (TCB)—is susceptible to multi-stage systemic dangers that bypass conventional, remoted defenses. By introducing a five-layer lifecycle framework protecting initialization, enter, inference, determination, and execution, the analysis workforce demonstrates how compound threats like reminiscence poisoning and talent provide chain contamination can compromise an agent’s complete operational trajectory.
OpenClaw Structure: The pi-coding-agent and the TCB
OpenClaw makes use of a ‘kernel-plugin’ structure that separates core logic from extensible performance. The system’s Trusted Computing Base (TCB) is outlined by the pi-coding-agent, a minimal core accountable for reminiscence administration, job planning, and execution orchestration. This TCB manages an extensible ecosystem of third-party plugins—or ‘abilities’—that allow the agent to carry out high-privilege operations similar to automated software program engineering and system administration. A essential architectural vulnerability recognized by the analysis workforce is the dynamic loading of those plugins with out strict integrity verification, which creates an ambiguous belief boundary and expands the system’s assault floor.


✓ Signifies efficient threat mitigation by the safety layer
× Denotes uncovered dangers by the safety layer
A Lifecycle-Oriented Risk Taxonomy
The analysis workforce systematizes the menace panorama throughout 5 operational levels that align with the agent’s purposeful pipeline:
- Stage I (Initialization): The agent establishes its operational atmosphere and belief boundaries by loading system prompts, safety configurations, and plugins.
- Stage II (Enter): Multi-modal knowledge is ingested, requiring the agent to distinguish between trusted consumer directions and untrusted exterior knowledge sources.
- Stage III (Inference): The agent reasoning course of makes use of strategies similar to Chain-of-Thought (CoT) prompting whereas sustaining contextual reminiscence and retrieving exterior information by way of retrieval-augmented era.
- Stage IV (Choice): The agent selects acceptable instruments and generates execution parameters by means of planning frameworks similar to ReAct.
- Stage V (Execution): Excessive-level plans are transformed into privileged system actions, requiring strict sandboxing and access-control mechanisms to handle operations.
This structured strategy highlights that autonomous brokers face multi-stage systemic dangers that stretch past remoted immediate injection assaults.
Technical Case Research in Agent Compromise
1. Ability Poisoning (Initialization Stage)
Ability poisoning targets the agent earlier than a job even begins. Adversaries can introduce malicious abilities that exploit the potential routing interface.
- The Assault: The analysis workforce demonstrated this by coercing OpenClaw to create a purposeful talent named hacked-weather.
- Mechanism: By manipulating the talent’s metadata, the attacker artificially elevated its precedence over the reputable climate instrument.
- Impression: When a consumer requested climate knowledge, the agent bypassed the reputable service and triggered the malicious substitute, yielding attacker-controlled output.
- Prevalence: An empirical audit cited within the analysis report discovered that 26% of community-contributed instruments include safety vulnerabilities.






2. Oblique Immediate Injection (Enter Stage)
Autonomous brokers regularly ingest untrusted exterior knowledge, making them vulnerable to zero-click exploits.
- The Assault: Attackers embed malicious directives inside exterior content material, similar to an internet web page.
- Mechanism: When the agent retrieves the web page to meet a consumer request, the embedded payload overrides the unique goal.
- End result: In a single check, the agent ignored the consumer’s job to output a set ‘Howdy World’ string mandated by the malicious web site.




3. Reminiscence Poisoning (Inference Stage)
As a result of OpenClaw maintains a persistent state, it’s susceptible to long-term behavioral manipulation.
- Mechanism: An attacker makes use of a transient injection to change the agent’s MEMORY.md file.
- The Assault: A fabricated rule was added instructing the agent to refuse any question containing the time period ‘C++’.
- Impression: This ‘poison’ continued throughout classes; subsequent benign requests for C++ programming have been rejected by the agent, even after the preliminary assault interplay had ended.




4. Intent Drift (Choice Stage)
Intent drift happens when a sequence of regionally justifiable instrument calls results in a globally harmful end result.
- The State of affairs: A consumer issued a diagnostic request to remove a ‘suspicious crawler IP’.
- The Escalation: The agent autonomously recognized IP connections and tried to change the system firewall by way of iptables.
- System Failure: After a number of failed makes an attempt to change configuration recordsdata exterior its workspace, the agent terminated the working course of to aim a handbook restart. This rendered the WebUI inaccessible and resulted in an entire system outage.


5. Excessive-Threat Command Execution (Execution Stage)
This represents the ultimate realization of an assault the place earlier compromises propagate into concrete system influence.
- The Assault: An attacker decomposed a Fork Bomb assault into 4 individually benign file-write steps to bypass static filters.
- Mechanism: Utilizing Base64 encoding and sed to strip junk characters, the attacker assembled a latent execution chain in set off.sh.
- Impression: As soon as triggered, the script prompted a pointy CPU utilization surge to close 100% saturation, successfully launching a denial-of-service assault in opposition to the host infrastructure.






The 5-Layer Protection Structure
The analysis workforce evaluated present defenses as ‘fragmented’ level options and proposed a holistic, lifecycle-aware structure.


(1) Foundational Base Layer:
Establishes a verifiable root of belief in the course of the startup section. It makes use of Static/Dynamic Evaluation (ASTs) to detect unauthorized code and Cryptographic Signatures (SBOMs) to confirm talent provenance.
(2) Enter Notion Layer:
Acts as a gateway to stop exterior knowledge from hijacking the agent’s management stream. It enforces an Instruction Hierarchy by way of cryptographic token tagging to prioritize developer prompts over untrusted exterior content material.
(3) Cognitive State Layer:
Protects inside reminiscence and reasoning from corruption. It employs Merkle-tree Buildings for state snapshotting and rollbacks, alongside Cross-encoders to measure semantic distance and detect context drift.
(4) Choice Alignment Layer:
Ensures synthesized plans align with consumer goals earlier than any motion is taken. It consists of Formal Verification utilizing symbolic solvers to show that proposed sequences don’t violate security invariants.
(5) Execution Management Layer:
Serves as the ultimate enforcement boundary utilizing an ‘assume breach’ paradigm. It gives isolation by means of Kernel-Stage Sandboxing using eBPF and seccomp to intercept unauthorized system calls on the OS stage
Key Takeaways
- Autonomous brokers increase the assault floor by means of high-privilege execution and chronic reminiscence. Not like stateless LLM functions, brokers like OpenClaw depend on cross-system integration and long-term reminiscence to execute complicated, long-horizon duties. This proactive nature introduces distinctive multi-stage systemic dangers that span your complete operational lifecycle, from initialization to execution.
- Ability ecosystems face important provide chain dangers. Roughly 26% of community-contributed instruments in agent talent ecosystems include safety vulnerabilities. Attackers can use ‘talent poisoning’ to inject malicious instruments that seem reputable however include hidden precedence overrides, permitting them to silently hijack consumer requests and produce attacker-controlled outputs.
- Reminiscence is a persistent and harmful assault vector. Persistent reminiscence permits transient adversarial inputs to be reworked into long-term behavioral management. By means of reminiscence poisoning, an attacker can implant fabricated coverage guidelines into an agent’s reminiscence (e.g., MEMORY.md), inflicting the agent to persistently reject benign requests even after the preliminary assault session has ended.
- Ambiguous directions result in harmful ‘Intent Drift.’ Even with out specific malicious manipulation, brokers can expertise intent drift, the place a sequence of regionally justifiable instrument calls results in globally harmful outcomes. In documented circumstances, primary diagnostic safety requests escalated into unauthorized firewall modifications and repair terminations that rendered your complete system inaccessible.
- Efficient safety requires a lifecycle-aware, defense-in-depth structure. Current point-based defenses—similar to easy enter filters—are inadequate in opposition to cross-temporal, multi-stage assaults. A sturdy protection have to be built-in throughout all 5 layers of the agent lifecycle: Foundational Base (plugin vetting), Enter Notion (instruction hierarchy), Cognitive State (reminiscence integrity), Choice Alignment (plan verification), and Execution Management (kernel-level sandboxing by way of eBPF).
Take a look at Paper. Additionally, be happy to comply with us on Twitter and don’t neglect to hitch our 120k+ ML SubReddit and Subscribe to our E-newsletter. Wait! are you on telegram? now you’ll be able to be part of us on telegram as effectively.
Be aware: This text is supported and offered by Ant Analysis









