Google Threat Report Links AI-powered Malware to DPRK Crypto Theft

07-Nov-2025 Block News Media



In brief

  • Google identified five malware families that query LLMs to generate or hide malicious code.
  • A DPRK-linked group called UNC1069 used Gemini to probe wallet data and craft phishing scripts.
  • Google says it has disabled the accounts and tightened safeguards around model access.

Google has warned that several new malware families now use large language models during execution to modify or generate code, marking a new phase in how state-linked and criminal actors are deploying artificial intelligence in live operations.

In a report released this week, the Google Threat Intelligence Group said it has tracked at least five distinct strains of AI-enabled malware, some of which have already been used in ongoing and active attacks.

The newly-identified malware families “dynamically generate malicious scripts, obfuscate their own code to evade detection,” while also making use of AI models “to create malicious functions on demand,” instead of having those hard-coded into malware packages, the threat intelligence group stated.



Each variant leverages an external model such as Gemini or Qwen2.5-Coder during runtime to generate or obfuscate code, a method GTIG dubbed “just-in-time code creation.”

The technique represents a shift from traditional malware design, where malware logic is typically hard-coded into the binary.

By outsourcing parts of its functionality to an AI model, the malware can continuously make changes to harden itself against systems designed to deter it.

Two of the malware families, PROMPTFLUX and PROMPTSTEAL, demonstrate how attackers are integrating AI models directly into their operations.

GTIG’s technical brief describes how PROMPTFLUX runs a “Thinking Robot” process that calls Gemini’s API every hour to rewrite its own VBScript code, while PROMPTSTEAL, linked to Russia’s APT28 group, uses the Qwen model hosted on Hugging Face to generate Windows commands on demand.

The group also identified activity from a North Korean group known as UNC1069 (Masan) that misused Gemini.

Google’s research unit describes the group as “a North Korean threat actor known to conduct cryptocurrency theft campaigns leveraging social engineering,” with notable use of “language related to computer maintenance and credential harvesting.”

Per Google, the group’s queries to Gemini included instructions for locating wallet application data, generating scripts to access encrypted storage, and composing multilingual phishing content aimed at crypto exchange employees.

These activities, the report added, appeared to be part of a broader attempt to build code capable of stealing digital assets.

Google said it had already disabled the accounts tied to these activities and introduced new safeguards to limit model abuse, including refined prompt filters and tighter monitoring of API access.

The findings could point to a new attack surface where malware queries LLMs at runtime to locate wallet storage, generate bespoke exfiltration scripts, and craft highly credible phishing lures.

Decrypt has approached Google on how the new model could change approaches to threat modeling and attribution, but has yet to receive a response.

Generally Intelligent Newsletter

A weekly AI journey narrated by Gen, a generative AI model.



Source link
Also read: 7 Best Cryptos to Invest in 2025 at the Frontlines of Bull Run – One 1000X Presale Poised to Turn Small Bets Into Big Wins
About Author Lorem ipsum dolor sit amet, consectetur adipiscing elit. Nunc fermentum lectus eget interdum varius. Curabitur ut nibh vel velit cursus molestie. Cras sed sagittis erat. Nullam id ante hendrerit, lobortis justo ac, fermentum neque. Mauris egestas maximus tortor. Nunc non neque a quam sollicitudin facilisis. Maecenas posuere turpis arcu, vel tempor ipsum tincidunt ut.
WHAT'S YOUR OPINION?
Related News