π° Originally published on Securityelites β AI Red Team Education β the canonical, fully-updated version of this article.
Mandiantβs M-Trends 2026 report β released this week β named two malware families that represent a genuinely new category of threat in 2026: PROMPTFLUX and PROMPTSTEAL. These are not AI-assisted malware where humans use AI to write malicious code. They are malware families that actively query large language models during execution β using AI as part of their attack logic to evade detection and adapt in real time. My analysis of why this matters and what it changes for defenders.
What Youβll Learn
What PROMPTFLUX and PROMPTSTEAL are and how they differ from AI-generated malware
How querying an LLM mid-execution helps malware evade detection
Why traditional signature-based detection fails against this category
The defensive adaptations required to detect LLM-querying malware
What IBM calls βSlopolyβ malware and the broader AI malware landscape
β±οΈ 12 min read ### PROMPTFLUX β AI Malware Guide 2026 1. What PROMPTFLUX and PROMPTSTEAL Are 2. How LLM-Querying Malware Works 3. Why Signature Detection Fails 4. Slopoly β The AI Malware Ecosystem 5. Defensive Adaptations for AI Malware PROMPTFLUX represents the offensive convergence of the LLM capabilities I covered in What Is an LLM? with the adversarial ML techniques from Adversarial Machine Learning 2026. For the full AI malware picture including how AI is used to write malware, see Can AI Write Malware?
What PROMPTFLUX and PROMPTSTEAL Are
The key distinction I want to establish immediately: PROMPTFLUX is not malware written by AI. It is malware that uses AI during its execution. Thatβs a fundamentally different threat category. Traditional AI-generated malware (what IBM calls βSlopolyβ) uses AI at the development stage β a human uses an LLM to help write malicious code, then deploys it. PROMPTFLUX and PROMPTSTEAL query LLMs during the attack itself, in real time, to make dynamic decisions about how to proceed.
PROMPTFLUX vs AI-GENERATED MALWARE β THE DISTINCTIONCopy
Traditional AI-generated malware (Slopoly)
Stage: development β human uses LLM to write malware code
Runtime: no LLM dependency β runs without AI after deployment
Detection: still detectable by behaviour-based AV (once behavioural pattern is known)
PROMPTFLUX / PROMPTSTEAL (LLM-querying malware)
Stage: runtime β malware queries LLM during execution to get instructions
Runtime: LLM is part of the attack logic β malware adapts based on AI responses
Detection: behaviour is dynamic and changes per-environment β evades signature/behaviour profiles
Source
M-Trends 2026: βmalware families like PROMPTFLUX and PROMPTSTEAL actively query
large language models mid-execution to evade detectionβ
Released: March 2026, Mandiant/Google Threat Intelligence
How LLM-Querying Malware Works
My model for how LLM-querying malware evades detection by using AI during execution. The key insight is that the malwareβs attack behaviour is not fixed at compile time β itβs generated at runtime by an external AI. This means every execution in a different environment can produce a different behaviour profile, which is precisely what defeats the detection approaches defenders currently rely on. The malware doesnβt have a fixed behaviour β it makes API calls to an LLM and uses the response to decide what to do next. This is adversarial use of the same flexibility that makes LLMs useful for legitimate software.
LLM-QUERYING MALWARE β EXECUTION MODELCopy
Execution flow (conceptual, based on M-Trends disclosure)
- Malware installs and gains initial foothold
- Reconnaissance phase: collects environment data (AV present, OS version, network config)
- LLM query: sends environment context to LLM API: βGiven [environment details], what evasion technique should I use to avoid detection by [AV product]?β
- AI response: returns specific evasion recommendation for that environment
- Malware implements the AI-recommended evasion and proceeds with attack
Why this breaks traditional detection
Signature-based: no fixed code pattern to match β LLM-generated evasion varies per environment
Behaviour-based: behaviour profile changes each run based on AI output
Sandbox analysis: sandbox environment β target environment β different AI response β different behaviour
What PROMPTSTEAL specifically targets
PROMPTSTEAL: focused on extracting IP via βdistillation attacksβ (M-Trends 2026)
Target: proprietary ML models β extracting specialised training data and logic
Method: systematic querying to reconstruct the proprietary model
securityelites.com
LLM-Querying Malware vs Traditional Malware β Detection Comparison
Detection Method
Traditional Malware
PROMPTFLUX-type
Signature matching
β
Detects known patterns
β No fixed signature
Behaviour baseline
β
Consistent behaviour
β Dynamic per environment
Sandbox analysis
β
Reproduces in sandbox
β Different AI response in sandbox
LLM API traffic monitoring
N/A
β
Detects LLM queries
Network egress analysis
β
C2 traffic patterns
β οΈ LLM API traffic looks legitimate
πΈ Detection method effectiveness against traditional vs LLM-querying malware. Three of the four standard detection approaches fail or degrade significantly against PROMPTFLUX-type malware. The only new effective detection method β LLM API traffic monitoring β requires defenders to build capability they didnβt previously need. My priority for any SOC upgrading their detection capability in 2026: add LLM API egress monitoring to the detection stack.
Why Signature Detection Fails
The adversarial ML and evasion concepts I covered in earlier guides β how AI classifiers can be fooled by carefully crafted inputs β come together in PROMPTFLUX in a way that makes the evasion more robust than any previous technique. Traditional malware evasion involves obfuscation β the code does the same thing but looks different. LLM-querying malware evasion involves adaptation β the code actually does something different based on the environment, and the AI determines what that different thing should be.
π Read the complete guide on Securityelites β AI Red Team Education
This article continues with deeper technical detail, screenshots, code samples, and an interactive lab walk-through. Read the full article on Securityelites β AI Red Team Education β
This article was originally written and published by the Securityelites β AI Red Team Education team. For more cybersecurity tutorials, ethical hacking guides, and CTF walk-throughs, visit Securityelites β AI Red Team Education.

Top comments (0)