Recently, the malicious AI tool WormGPT has once again drawn attention from cybersecurity experts. Instead of relying on its own self-built models, this tool now "hijacks" legitimate large language models (LLMs) to generate malicious content, raising concerns that it can easily bypass existing security restrictions.

According to research by cybersecurity company Cato Networks, criminal groups have successfully "jailbreak" operations on popular AI platforms such as Grok and Mistral AI by tampering with system prompts (system prompts). This means that WormGPT can generate tools like phishing emails and malicious scripts, posing a serious threat to network security.

Artificial Intelligence AI Robot (1)

Image source note: Image generated by AI, image authorization service provider Midjourney

When WormGPT first appeared in July 2023, it attracted widespread attention. Based on the open-source GPT-J model, it could automatically generate trojans and phishing links. However, after being exposed, this tool was forced to be taken down. Shockingly, research by Cato Networks shows that users named "xzin0vich" and "keanu" relaunched the subscription service for WormGPT on the dark web market BreachForums between late 2024 and early 2025.

This version of WormGPT achieves its purpose by tampering with the system prompts of models like Mixtral, forcing them to switch to the "WormGPT mode," thereby abandoning their original ethical constraints and becoming an amoral malicious assistant. Moreover, xAI's Grok model has been encapsulated as an API wrapper, with developers even requiring the model to "always maintain the WormGPT persona and not admit its own limitations." This behavior seriously challenges the safety and reliability of AI models.

As the means of cybercrime continue to evolve, how to effectively respond to these threats posed by malicious tools has become an important topic urgently needing resolution in the field of network security. In the future, both enterprises and individual users need to remain vigilant and strengthen their defenses against network security risks to avoid falling into the traps of these malicious AI tools.