Mitre Attack Framework 

Last Updated on February 19, 2026 by Arnav Sharma

For last three decades, cybersecurity has followed a familiar path. Attackers build something nasty, defenders figure out how to catch it, attackers tweak their approach, and the cycle continues. It was always a cat-and-mouse game, but the pace was manageable. Both sides were constrained by the same thing: human speed. Innovation happened over weeks, months, sometimes years.

That flow broke in 2025.

Google’s Threat Intelligence Group confirmed something many of us had been anticipating but hoping wouldn’t arrive quite so soon. They documented the first known instances of AI-augmented malware deployed in live operations. Not in a lab. Not in a proof-of-concept paper. In actual attacks against real targets.

We’re now dealing with malware that can rewrite itself mid-attack, generate novel evasion techniques on the fly, and make autonomous decisions about how to proceed based on the defenses it encounters. The era of adaptive, self-mutating code isn’t a warning about the future anymore. It’s an operational reality that security teams need to grapple with right now.

A Quick Walk Through Malware History

To appreciate why this moment is so significant, it helps to understand the road that got us here. Malware evolution didn’t happen overnight. Each generation built on the limitations of the last, and each forced defenders to rethink their approach.

The 1990s: Encryption and Early Polymorphism

The first polymorphic virus showed up in 1990, courtesy of Mark Washburn. The concept was elegant in its simplicity: encrypt the payload and swap out the decryption routine with every new copy. Suddenly, no two copies of the same virus looked alike to a scanner.

Viruses like Tequila in 1991, along with creations powered by Dark Avenger’s Mutation Engine, meant that even relatively unskilled authors could produce malware that slipped past signature-based detection. The security industry responded with emulation, essentially running suspicious code in a controlled sandbox to see what it actually did once decrypted.

The 2000s: Metamorphic Malware Gets Creative

Metamorphic malware raised the stakes considerably. Instead of just changing the encryption wrapper, these threats rewrote their entire codebase with each iteration. Techniques like instruction replacement (swapping operations for functionally equivalent ones), code permutation (reordering execution blocks connected by jump instructions), and dead-code insertion created variants that looked completely different to scanners while doing exactly the same malicious things under the hood.

Examples like W32/Simile and W32/Zmist demonstrated that you could eliminate the encrypted payload entirely and still evade detection. This forced the industry toward heuristic analysis, looking for suspicious behavioral patterns rather than matching specific byte sequences.

The 2010s: Going Fileless

Then malware decided it didn’t need to exist as a file at all. Living-off-the-land techniques leveraged legitimate system tools like PowerShell and WMI to execute attacks entirely in memory, never writing a traditional payload to disk. These fileless attacks were estimated to be up to ten times more successful than conventional malware because they blended seamlessly into normal system operations. If the tools being used are the same ones your IT team relies on daily, how do you tell the difference?

2024-2025: Enter AI

And then generative AI entered the picture. Everything accelerated.

The New Generation: AI as a Mutation Engine

Here’s what makes today’s adaptive malware fundamentally different from what came before. Traditional polymorphic malware cycled through a finite set of pre-programmed transformations. Think of it like a combination lock with a limited number of possible combinations. Eventually, defenders could map out the possibilities and detect them.

The new generation doesn’t have that limitation. It queries large language models at runtime to generate entirely novel code on demand. There’s no fixed set of transformations to catalogue. Every variant is genuinely new.

PROMPTFLUX: Self-Rewriting Code via Gemini

Discovered by Google’s Threat Intelligence Group in mid-2025, PROMPTFLUX is an experimental dropper that talks to Google’s Gemini API to request on-demand code obfuscation. One observed variant rewrote its entire source code every hour, generating structurally different but functionally identical versions of itself. Each iteration produced code that no scanner had ever seen before.

Think about what that means for signature-based detection. You can’t write a signature for something that doesn’t exist yet. Every hour, the malware essentially becomes a brand-new threat from the scanner’s perspective.

Google described PROMPTFLUX as a step toward more autonomous and adaptive malware, software that doesn’t just follow pre-programmed evasion routines but dynamically generates new ones based on the detection environment it encounters.

PROMPTSTEAL: Live LLM Queries in Real Operations

If PROMPTFLUX was the experimental phase, PROMPTSTEAL showed what operational deployment looks like. This malware, attributed to Russia-linked APT28 (Fancy Bear), was used against Ukrainian targets in June 2025. It marked Google’s first observation of malware querying a large language model during live operations.

PROMPTSTEAL used Hugging Face’s platform to interact with an open-source LLM, generating short Windows commands for reconnaissance and data exfiltration on the fly. It disguised itself as an image-generation tool while creating and executing commands in the background. The critical detail: none of the commands were hardcoded. Each was generated dynamically, which meant defenders couldn’t rely on known command patterns or static indicators of compromise.

I’ve worked with teams that build detection rules around specific command-line patterns. Against something like PROMPTSTEAL, those rules become almost worthless because the commands are different every time.

LameHug: On-Device AI for Autonomous Decision-Making

This one is probably the most concerning development of the three. LameHug embedded Alibaba Cloud’s Qwen 2.5-Coder-32B model, an open-source LLM, directly into its attack chain. Unlike PROMPTFLUX and PROMPTSTEAL, which relied on cloud-based API calls (which at least create a potential detection vector through unusual network traffic), LameHug used a local transformer model to semantically analyse system logs, interpret security events, and adapt its behaviour accordingly.

That’s the progression in a nutshell: from cloud-dependent AI malware to fully autonomous, on-device adaptive threats. No external API calls to flag. No unusual network connections to trigger alerts. The intelligence runs locally.

What Malware Can and Can’t Mutate

Before we spiral into doom and gloom, there’s a crucial concept that every defender needs to internalize. Malware can mutate its form, but it can’t avoid its needs.

No matter how cleverly malware rewrites itself, it still has to do certain things to accomplish its mission. It needs to execute instructions on a CPU. It needs to interact with the operating system through processes, files, registry entries, or services. If it wants to stick around, it needs persistence mechanisms. And almost always, it needs to communicate, whether that’s with a command-and-control server, through lateral movement, or via data exfiltration channels.

These are the invariants, the things that must happen regardless of what the code looks like. And invariants, not signatures, are where defenders need to focus their energy.

A useful mental model looks something like this:

  • Polymorphic code changes its runtime footprint, but the core objectives and OS interactions stay the same. Focus on behavior and memory-level telemetry.
  • Packing and runtime decryption hides the payload on disk, but the unpack/decrypt events and memory regions remain detectable. Lean on memory scanning and sandbox detonation.
  • Metamorphism rewrites code structure, but the semantics (what it actually does) persist. Semantic analysis and behavioral analytics are your friends here.
  • Behavioral polymorphism randomizes threading, timing, and task distribution, but end goals like credential access, persistence, and C2 communication remain constant. Correlate multi-event chains instead of relying on single-event detections.
  • Modular/adaptive configurations swap modules per victim, but loader behavior and C2 patterns tend to stay consistent. Network and endpoint correlation, combined with active threat hunting, is the play.

Even if the code turns to fog, the mission stays the same. That’s where your detection leverage lives.

Why This Breaks the Old Playbook

That said, the implications for existing defensive approaches are serious. Let’s not sugarcoat it.

Signature-Based Detection Is Effectively Dead for These Threats

This has been a slow death, and most of us have seen it coming for years. AI-adaptive malware delivers the final blow. When malware can generate a structurally unique variant every hour, or in some observed cases approximately every 15 seconds during execution, maintaining a signature database becomes an exercise in futility. No update pipeline can keep pace with that rate of change.

Signatures still have value for catching commodity malware and known threats. But against the adaptive stuff? They’re bringing a phonebook to a conversation that changes languages every minute.

Behavioural Heuristics Are Under Real Pressure

The traditional fallback when signatures fail has been behavioral analysis, detecting malware by what it does rather than what it looks like. That approach is still valuable, but it’s facing serious headwinds.

Adaptive malware that can modify its execution methods, switch between attack techniques based on the target’s security posture, and mimic legitimate application behavior is specifically designed to defeat behavioral heuristics. When the malware can observe the security stack it’s operating against and adjust accordingly, the detection problem becomes adversarial in a much deeper sense than we’ve dealt with before.

Research has even documented behavioral polymorphism, where malware splits tasks across multiple threads so that any single thread looks completely benign. Only when you correlate across all threads does the malicious pattern emerge. That’s a tough detection problem.

Indicators of Compromise Have a Very Short Shelf Life

IOCs (file hashes, IP addresses, domain names, specific command strings) have always been the currency of threat intelligence sharing. Against adaptive malware, IOCs go stale almost immediately. The hash changes every iteration. The commands are generated dynamically. The network indicators shift.

This doesn’t mean threat intelligence is dead. It means the level of intelligence needs to shift upward. From sharing IOCs to sharing TTPs (Tactics, Techniques, and Procedures), the higher-order patterns that remain consistent even when the implementation details are in constant flux. MITRE ATT&CK mapping of AI-enhanced techniques is going to become increasingly critical here.

The Barrier to Entry Is Dropping Fast

The Google report flagged something that should concern everyone in this space. This isn’t limited to nation-state actors anymore. Tools like WormGPT, FraudGPT, and Nytheon AI are available on dark web forums, offering malware generation capabilities that require minimal coding skills. Open-source models can be downloaded and run locally with safety guardrails removed, enabling anyone to create adaptive malware without relying on commercial AI APIs that might have built-in safeguards.

Underground marketplaces are already commercialising generative AI modules for phishing, deepfake production, and malware creation. The skill floor for producing sophisticated, evasive malware is dropping rapidly. And that economic shift matters enormously because it means the volume of adaptive threats is going to increase, not just the sophistication.

Attack Patterns Worth Watching

Based on current research and observed threat activity, several patterns are starting to crystallise that I think every security team should have on their radar.

  • Runtime metamorphism is code that restructures its execution flow, register usage, and API call sequences on each execution cycle, not just during propagation. This breaks both static signature matching and behavioral heuristics simultaneously because the behavioral fingerprint itself becomes non-deterministic. It’s a moving target in every dimension.
  • Environment-aware evasion is malware that profiles the sandbox, EDR tooling, or analysis framework it encounters and selectively activates mutation strategies tuned to bypass that specific stack. Samples have already been observed that behave completely benignly when they detect common analysis environments and only deploy their payload on production systems. I’ve seen this concept discussed in threat intel briefings, and it’s only going to get more sophisticated.
  • Objective drift is a fascinating and somewhat unsettling development. As malware becomes more autonomous, researchers are noting the potential for it to begin optimising for objectives beyond its original programming. A system designed for data theft might learn that persistence or lateral movement increases its lifespan and start optimising for those behaviors independently. This breaks the traditional link between attacker intent and malware action, which complicates attribution and response planning.
  • AI-powered botnets represent adaptive malware commanding entire networks that can adjust their attack strategies, evade containment, and reconfigure their infrastructure in real-time. According to the 2025 Imperva Bad Bot Report, automated traffic now exceeds 50% of web sessions, with malicious bots representing 37%. Layer AI-driven adaptation on top of that scale, and you’ve got a serious problem.

Defensive Architecture for the Adaptive Era

Alright, enough about the threats. Let’s talk about what actually works. Combating AI-driven threats requires a fundamental shift in defensive philosophy, from detection-first to verification-first, from reactive to predictive. Here’s what that looks like in practice.

Fight AI with AI

Behavioral analytics powered by machine learning needs to replace signature-based detection as the primary defensive layer. I’m talking about deploying User and Entity Behaviour Analytics (UEBA) that builds dynamic baselines for every user, device, and application and flags anomalies, those subtle deviations from established patterns that would be invisible to traditional tools.

The key advantage AI offers defenders is the ability to learn what “normal” looks like across a complex environment and detect novel threats without prior knowledge of their specific characteristics. You don’t need to know what the malware looks like. You need to know that something is behaving unlike everything else in the environment.

Microsoft’s Project Ire is a great example of where this is heading. It’s an autonomous agent designed to classify software by automating deep reverse engineering workflows and producing an auditable chain of evidence. Microsoft plans to integrate it into Defender as a Binary Analyser, with a longer-term vision of detecting novel malware in memory at scale. That’s the direction the industry needs to move: automating deeper reasoning rather than relying on shallow pattern matching.

Memory-Aware Detection and Forensics

Because packed and polymorphic malware frequently must reveal its real code in memory to execute, memory-centric approaches can catch what disk scanning misses entirely. This is one of the most reliable detection points against mutating malware. No matter how many times it rewrites itself on disk, it has to unpack and run somewhere.

Volexity has done excellent work developing memory forensics techniques aimed at detecting bypasses used to evade EDR inspection. The practical takeaway: make sure your EDR solution has strong memory scanning capabilities, and invest in memory forensics tooling for your incident response process.

Shift from Detection to Verification

Zero trust needs to extend beyond identity and access control into how we validate the integrity of security telemetry itself. If adaptive malware can falsify system logs and mimic legitimate processes, defenders must assume that some telemetry will be manipulated.

Use independent data sources to confirm integrity. Cross-reference system baselines against process-level attestations. If a single source tells you everything is fine but other sources tell a different story, trust the aggregated view, not the individual data point.

Shorten the Response Loop

Human-in-the-loop response processes are simply too slow against threats that adapt in real-time. Pre-authorise containment steps like session termination, host isolation, and key revocation so automation can act immediately while analysts investigate after the fact.

Here’s a number that should keep you up at night: the fastest recorded eCrime breakout time is now 51 seconds. From initial access to lateral movement in under a minute. Human reaction time cannot consistently match that. Automated response isn’t optional anymore. It’s a requirement.

Monitor for AI-Specific Indicators

We need new detection vectors tailored to AI-augmented attack chains. Look for unusual connections to AI APIs like OpenAI, Anthropic, or Hugging Face endpoints. Watch for unexpected LLM inference activity on endpoints. Flag abnormally high entropy in code segments, which can indicate dynamic obfuscation. Monitor for webhook-based exfiltration channels.

These aren’t traditional IOCs. They’re architectural indicators of AI-augmented attack chains, and they require updated detection logic in your SIEM and EDR platforms.

Consider Prevention-First Approaches for High-Risk Surfaces

For the highest-risk attack surfaces, Content Disarm and Reconstruction (CDR) technology deserves serious consideration. CDR assumes all incoming content is potentially dangerous and rebuilds files from the ground up, stripping out executable content or active code that could harbour zero-day threats. This bypasses the detection problem entirely for file-based attack vectors.

It’s not a silver bullet, and it won’t work for every use case. But for email attachments, file uploads, and document exchange workflows, CDR removes the guesswork.

Behavior Detection with Strong Correlation

Single-event behavioral rules are too easy to evade. The real power comes from correlation across multiple signals. Build detections that look for combinations: process tree anomalies paired with unusual privilege transitions, suspicious persistence chains combined with outbound beaconing, scripting engine execution linked to new scheduled tasks.

The more signals you correlate, the harder it becomes for adaptive malware to evade all of them simultaneously. No matter how clever the mutation engine, generating behavior that passes as normal across process trees, network patterns, privilege usage, and persistence mechanisms all at once is extremely difficult.

Don’t Forget the Basics

Even the most adaptive malware benefits from easy execution paths. The boring fundamentals still pay enormous dividends:

  • Aggressive patching and vulnerability management
  • Least privilege enforcement and removing local admin where possible
  • Application control and allowlisting where feasible
  • Script hardening for PowerShell, WSH, and Office macros
  • Network segmentation and egress control
  • Tested backups and restore drills for ransomware resilience

These measures reduce the attacker’s room to experiment with variants. A perfectly adaptive piece of malware still needs an entry point, and closing off easy paths forces it into noisier, more detectable approaches.

Invest in TTP-Level Threat Intelligence Sharing

Move beyond IOC-focused intelligence sharing to frameworks that capture the higher-order patterns of adaptive attacks. The techniques, procedures, and AI integration patterns that remain consistent even when the code changes constantly. MITRE ATT&CK mapping of AI-enhanced techniques will become the baseline for meaningful intelligence exchange in this new landscape.

Where This Is All Heading

Google’s researchers were careful to note that what they’ve observed so far represents early indicators, proofs-of-concept, and clear intent, not yet widespread deployment. But anyone who has watched how cybercrime evolves knows the pattern. Capability first appears in experiments, then trickles into mainstream malware kits, and eventually becomes standard practice. We’ve seen this movie before with ransomware-as-a-service, with living-off-the-land techniques, and with fileless malware. AI-adaptive malware will follow the same trajectory.

The convergence of generative AI with malware development fundamentally alters the economics of cyberattack. The cost and skill required to produce sophisticated, evasive malware is dropping rapidly, while the cost to defenders of detecting and responding to each unique variant is rising. That economic asymmetry, if left unaddressed, will reshape the threat landscape in ways that make the current environment look manageable by comparison.

But here’s the thing. This is an arms race, and AI gives defenders powerful tools too. Automated reverse engineering, semantic analysis, memory-based detection, behavioral correlation at machine speed, all of these are within reach. The question isn’t whether the technology exists to counter adaptive threats. It does. The question is whether organisations will invest in deploying it quickly enough.

The era of relying on static defences against static threats is ending. What comes next depends on how quickly the defensive community adapts. Not just our tools, but our architectures, our assumptions, and our operational tempo.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.