BT
Privacy ToolboxJournalProjectsResumeBookmarks
Feed
Privacy Toolbox
Journal
Projects
Resume
Bookmarks
Intel
NERF
The Vault
Threat Actors
Privacy Threats
Malware IoC
Dashboard
CVEs
Tags
Intel
NERFThe VaultThreat ActorsPrivacy ThreatsMalware IoCDashboardCVEsTags

Intel

  • Feed
  • Threat Actors
  • Privacy Threats
  • Dashboard
  • Privacy Toolbox
  • CVEs

Personal

  • Journal
  • Projects

Resources

  • Subscribe
  • Bookmarks
  • Developers
  • Tags
Cybersecurity News & Analysis
github
defconxt
•
© 2026
•
blacktemple.net
  1. Feed
  2. /Researchers Map Seven-Stage 'Promptware Kill Chain' for LLM-Based Malware

Researchers Map Seven-Stage 'Promptware Kill Chain' for LLM-Based Malware

mediumMalware & Threats|February 17, 20263 min read

Originally reported by Schneier on Security

#llm-security#prompt-injection#ai-malware#kill-chain#jailbreaking#lateral-movement
Share

TL;DR

Security researchers propose a structured framework mapping how AI prompt injection attacks evolve into sophisticated malware campaigns across seven distinct stages.

Why medium?

Academic research from Schneier et al. mapping a structured kill chain for LLM-based malware. Significant emerging threat research but no confirmed active exploitation of the described attack chain.

Beyond Simple Prompt Injection

The cybersecurity community's focus on "prompt injection" as a singular vulnerability fundamentally misrepresents the threat landscape surrounding large language models (LLMs). Security researchers Schneier, Brodt, Feldman, and Nassi have published new research demonstrating that attacks against AI systems have evolved into a distinct class of malware execution mechanisms they term "promptware."

The research proposes a structured seven-step kill chain model that mirrors traditional malware campaigns like Stuxnet and NotPetya, providing security practitioners with a framework to understand and defend against increasingly sophisticated AI-based attacks.

The Seven-Stage Kill Chain

Initial Access

Malicious payloads enter AI systems through direct prompt input or "indirect prompt injection," where adversaries embed instructions in content the LLM retrieves during inference, web pages, emails, shared documents, or even images and audio files in multimodal systems. The fundamental architectural flaw lies in LLMs processing all input as undifferentiated token sequences, eliminating the traditional boundary between trusted instructions and untrusted data.

Privilege Escalation (Jailbreaking)

Attackers circumvent safety training and policy guardrails through social engineering techniques, convincing models to adopt personas that ignore safety rules, or through sophisticated adversarial suffixes. This phase unlocks the model's full capabilities for malicious use.

Reconnaissance

The compromised LLM reveals information about connected services, assets, and capabilities, enabling autonomous progression through the kill chain. Unlike traditional malware reconnaissance, this occurs post-compromise and leverages the victim model's reasoning capabilities against itself.

Persistence

Promptware embeds itself into AI agents' long-term memory or poisons databases the agent relies upon. For example, malicious code can infect email archives, re-executing every time the AI summarizes past communications.

Command and Control (C2)

Established persistence enables dynamic command fetching during inference time, transforming static threats into controllable trojans whose behavior attackers can modify remotely.

Lateral Movement

Infected AI agents spread malware across connected systems, leveraging their access to emails, calendars, and enterprise platforms. Self-replicating attacks can trick email assistants into forwarding malicious payloads to all contacts, creating viral propagation patterns.

Actions on Objective

The final stage achieves tangible malicious outcomes including data exfiltration, financial fraud, or physical world impact. Documented examples include AI agents manipulated into selling cars for one dollar or transferring cryptocurrency to attacker wallets.

Real-World Demonstrations

The "Invitation Is All You Need" research demonstrated the kill chain by embedding malicious prompts in Google Calendar invitation titles. The attack achieved persistence through workspace memory, lateral movement by launching Zoom, and concluded with covert livestreaming of victims.

Similarly, "Here Comes the AI Worm" research showed end-to-end kill chain execution through email-based prompt injection, resulting in data exfiltration and viral propagation to new recipients.

Defense Strategy Implications

The research argues that prompt injection cannot be fixed in current LLM technology, requiring defense-in-depth strategies that assume initial access will occur. Effective defenses must focus on breaking the chain at subsequent stages through:

  • Limiting privilege escalation capabilities
  • Constraining reconnaissance activities
  • Preventing persistence mechanisms
  • Disrupting C2 communications
  • Restricting permitted agent actions

By reframing prompt injection as the initial stage of sophisticated malware campaigns rather than isolated vulnerabilities, security teams can shift from reactive patching to systematic risk management for AI-integrated systems.

Sources

  • The Promptware Kill Chain - Schneier on Security

Originally reported by Schneier on Security

Tags

#llm-security#prompt-injection#ai-malware#kill-chain#jailbreaking#lateral-movement

Tracked Companies

🇺🇸Google

Related Intelligence

  • Critical Infrastructure Under Siege: From Actively Exploited BeyondTrust RCE to Healthcare Ransomware Shutdowns

    criticalFeb 20, 2026
  • AI Agent Hackerbot-Claw Compromises Microsoft, DataDog, and CNCF GitHub Repositories

    highMar 9, 2026
  • Companies Deploy Hidden AI Prompt Injection to Bias Assistant Recommendations

    mediumMar 4, 2026

Related Knowledge

  • NERF Deep Training: Malware Analysis, Reverse Engineering, and Evasion Techniques

    offensive
  • Malware Analysis Deep Dive — NERF Training Module

    dfir
  • DFIR & Threat Hunting Deep Training — NERF Knowledge Base

    dfir

Explore

  • Dashboard
  • Privacy Threats
  • Threat Actors
← Back to the feed

Previous Article

← SANS ISC Stormcast Security Update - February 17, 2026

Next Article

Privacy, Anonymity, and OSINT: A Defender's Guide to Staying Safe Online→