Security researchers warn that generative AI assistants, including Microsoft Copilot and xAI’s Grok, are evolving beyond productivity aids into potential infrastructure for malware abuse. Their web browsing features enable hackers to conceal malicious traffic and deploy adaptive command-and-control (C2) operations.
How Malware Exploits AI Assistants
Once malware infects a device, it harvests sensitive data and system details. Attackers encode this information into URLs on their controlled domains, such as http://malicious-site.com/report?data=12345678. The malware then prompts the AI assistant to “summarize the contents of this website.”
This query mimics legitimate AI usage, evading security detection. The attacker-controlled server logs the encoded data, relaying it undetected.
Hidden Responses and Escalation
Attackers can embed hidden prompts in the website’s response, which the AI processes and executes. This allows further instructions without raising alarms.
The threat intensifies when malware queries the AI for next steps. Based on harvested system info, it determines if the environment is a high-value enterprise target or a sandbox. In a sandbox, the malware remains dormant; otherwise, it advances to subsequent stages.
AI as Stealthy Decision Engine
Analysis shows AI services can serve as a covert transport layer. They also deliver prompts and outputs functioning as external decision engines. This paves the way for AI-driven malware implants and automated C2 systems that handle triage, targeting, and real-time operations dynamically.
Blending malicious traffic with legitimate AI interactions represents a sophisticated evasion tactic, challenging traditional security measures.