DEV Community

Cover image for PROMPTFLUX and PROMPTSTEAL explained β€” AI Malware That Queries LLMs Mid-Attack (2026)
Mr Elite
Mr Elite

Posted on β€’ Originally published at securityelites.com

PROMPTFLUX and PROMPTSTEAL explained β€” AI Malware That Queries LLMs Mid-Attack (2026)

πŸ“° Originally published on Securityelites β€” AI Red Team Education β€” the canonical, fully-updated version of this article.

PROMPTFLUX and PROMPTSTEAL explained  β€” AI Malware That Queries LLMs Mid-Attack (2026)

Mandiant’s M-Trends 2026 report β€” released this week β€” named two malware families that represent a genuinely new category of threat in 2026: PROMPTFLUX and PROMPTSTEAL. These are not AI-assisted malware where humans use AI to write malicious code. They are malware families that actively query large language models during execution β€” using AI as part of their attack logic to evade detection and adapt in real time. My analysis of why this matters and what it changes for defenders.

What You’ll Learn

What PROMPTFLUX and PROMPTSTEAL are and how they differ from AI-generated malware
How querying an LLM mid-execution helps malware evade detection
Why traditional signature-based detection fails against this category
The defensive adaptations required to detect LLM-querying malware
What IBM calls β€œSlopoly” malware and the broader AI malware landscape

⏱️ 12 min read ### PROMPTFLUX β€” AI Malware Guide 2026 1. What PROMPTFLUX and PROMPTSTEAL Are 2. How LLM-Querying Malware Works 3. Why Signature Detection Fails 4. Slopoly β€” The AI Malware Ecosystem 5. Defensive Adaptations for AI Malware PROMPTFLUX represents the offensive convergence of the LLM capabilities I covered in What Is an LLM? with the adversarial ML techniques from Adversarial Machine Learning 2026. For the full AI malware picture including how AI is used to write malware, see Can AI Write Malware?

What PROMPTFLUX and PROMPTSTEAL Are

The key distinction I want to establish immediately: PROMPTFLUX is not malware written by AI. It is malware that uses AI during its execution. That’s a fundamentally different threat category. Traditional AI-generated malware (what IBM calls β€œSlopoly”) uses AI at the development stage β€” a human uses an LLM to help write malicious code, then deploys it. PROMPTFLUX and PROMPTSTEAL query LLMs during the attack itself, in real time, to make dynamic decisions about how to proceed.

PROMPTFLUX vs AI-GENERATED MALWARE β€” THE DISTINCTIONCopy

Traditional AI-generated malware (Slopoly)

Stage: development β€” human uses LLM to write malware code
Runtime: no LLM dependency β€” runs without AI after deployment
Detection: still detectable by behaviour-based AV (once behavioural pattern is known)

PROMPTFLUX / PROMPTSTEAL (LLM-querying malware)

Stage: runtime β€” malware queries LLM during execution to get instructions
Runtime: LLM is part of the attack logic β€” malware adapts based on AI responses
Detection: behaviour is dynamic and changes per-environment β†’ evades signature/behaviour profiles

Source

M-Trends 2026: β€œmalware families like PROMPTFLUX and PROMPTSTEAL actively query
large language models mid-execution to evade detection”
Released: March 2026, Mandiant/Google Threat Intelligence

How LLM-Querying Malware Works

My model for how LLM-querying malware evades detection by using AI during execution. The key insight is that the malware’s attack behaviour is not fixed at compile time β€” it’s generated at runtime by an external AI. This means every execution in a different environment can produce a different behaviour profile, which is precisely what defeats the detection approaches defenders currently rely on. The malware doesn’t have a fixed behaviour β€” it makes API calls to an LLM and uses the response to decide what to do next. This is adversarial use of the same flexibility that makes LLMs useful for legitimate software.

LLM-QUERYING MALWARE β€” EXECUTION MODELCopy

Execution flow (conceptual, based on M-Trends disclosure)

  1. Malware installs and gains initial foothold
  2. Reconnaissance phase: collects environment data (AV present, OS version, network config)
  3. LLM query: sends environment context to LLM API: β€œGiven [environment details], what evasion technique should I use to avoid detection by [AV product]?”
  4. AI response: returns specific evasion recommendation for that environment
  5. Malware implements the AI-recommended evasion and proceeds with attack

Why this breaks traditional detection

Signature-based: no fixed code pattern to match β€” LLM-generated evasion varies per environment
Behaviour-based: behaviour profile changes each run based on AI output
Sandbox analysis: sandbox environment β‰  target environment β†’ different AI response β†’ different behaviour

What PROMPTSTEAL specifically targets

PROMPTSTEAL: focused on extracting IP via β€œdistillation attacks” (M-Trends 2026)
Target: proprietary ML models β€” extracting specialised training data and logic
Method: systematic querying to reconstruct the proprietary model

securityelites.com

LLM-Querying Malware vs Traditional Malware β€” Detection Comparison

Detection Method
Traditional Malware
PROMPTFLUX-type
Signature matching
βœ… Detects known patterns
❌ No fixed signature
Behaviour baseline
βœ… Consistent behaviour
❌ Dynamic per environment
Sandbox analysis
βœ… Reproduces in sandbox
❌ Different AI response in sandbox
LLM API traffic monitoring
N/A
βœ… Detects LLM queries
Network egress analysis
βœ… C2 traffic patterns
⚠️ LLM API traffic looks legitimate

πŸ“Έ Detection method effectiveness against traditional vs LLM-querying malware. Three of the four standard detection approaches fail or degrade significantly against PROMPTFLUX-type malware. The only new effective detection method β€” LLM API traffic monitoring β€” requires defenders to build capability they didn’t previously need. My priority for any SOC upgrading their detection capability in 2026: add LLM API egress monitoring to the detection stack.

Why Signature Detection Fails

The adversarial ML and evasion concepts I covered in earlier guides β€” how AI classifiers can be fooled by carefully crafted inputs β€” come together in PROMPTFLUX in a way that makes the evasion more robust than any previous technique. Traditional malware evasion involves obfuscation β€” the code does the same thing but looks different. LLM-querying malware evasion involves adaptation β€” the code actually does something different based on the environment, and the AI determines what that different thing should be.


πŸ“– Read the complete guide on Securityelites β€” AI Red Team Education

This article continues with deeper technical detail, screenshots, code samples, and an interactive lab walk-through. Read the full article on Securityelites β€” AI Red Team Education β†’


This article was originally written and published by the Securityelites β€” AI Red Team Education team. For more cybersecurity tutorials, ethical hacking guides, and CTF walk-throughs, visit Securityelites β€” AI Red Team Education.

Top comments (0)