AI

Researchers warn about the use of generative AI to evade malware detection

OpenAI itself has blocked more than 20 malicious networks that were seeking to exploit its platform

Researchers warn about the use of generative AI to evade malware detection

Pedro Domínguez

  • December 25, 2024
  • Updated: July 1, 2025 at 10:35 PM
Researchers warn about the use of generative AI to evade malware detection

A recent analysis by Unit 42 of Palo Alto Networks has revealed that language models like ChatGPT can be used to modify malware in JavaScript, making it more difficult to detect. Although these models do not generate malware from scratch, cybercriminals can ask them to rewrite or obfuscate existing malicious code. “These transformations are more natural, which complicates the identification of malware,” state cybersecurity experts. This approach could degrade malware classification systems by confusing them into labeling malicious code as harmless.

Despite the security restrictions implemented by providers like OpenAI, tools like WormGPT are being used to create more convincing phishing emails and new types of malware. In October 2024, OpenAI blocked more than 20 malicious networks seeking to exploit its platform. In tests, Unit 42 managed to generate 10,000 variants of malicious JavaScript, maintaining its functionality but decreasing its detection scores in models like Innocent Until Proven Guilty. Among the techniques used are changes in variable names, insertion of junk code, and complete script rewriting.

Additionally, machine learning algorithms can be tricked into classifying these variants as benign in 88% of cases, according to Unit 42. Even the most popular tools, such as VirusTotal, struggle to detect these codes. Researchers warn that these AI-based rewrites are harder to track than those generated by libraries like obfuscator.io. However, they suggest that these same techniques could improve detection models by generating more robust training data.

On another front, researchers from the North Carolina State University discovered an attack called TPUXtract, which allows stealing AI models run on Google‘s Edge TPUs through electromagnetic signals. However, although the technique is notable for its precision, it requires physical access to the device and specialized equipment, which fortunately limits its scope.

Additionally, Morphisec has demonstrated that the Exploit Prediction Scoring System (EPSS), used to assess vulnerabilities, can be manipulated with fake social media posts and empty repositories on GitHub. According to Ido Ikar, this technique allows “inflating indicators” and deceiving organizations by altering cyber risk management priorities.

NordVPN DOWNLOAD

Latest Articles

Loading next article