AI Agents Emerge as a Top Cybercrime Target

With the explosion of personal AI agents, a new and deeply personal attack surface has emerged. To understand these evolving threats, we’re speaking with Dominic Jainy, an IT professional whose work at the intersection of AI, machine learning, and blockchain gives him a unique perspective on the digital battlefront. We’ll explore the shift from traditional password theft to the hijacking of an AI’s very “soul,” the subtle genius of supply chain attacks targeting AI skill platforms, and the alarming consequences of exposed AI instances that could give attackers a master key to corporate networks.

An infostealer was recently observed capturing AI agent files like openclaw.json for gateway tokens and soul.md for operational principles. How does this shift the threat landscape beyond just stealing passwords, and what specific new risks does this “AI identity theft” create for users?

It represents a terrifying evolution in digital crime. For years, we’ve been conditioned to protect our passwords and financial data, but this is different. Stealing a file like soul.md isn’t just about gaining access; it’s about capturing the very essence of a user’s digital assistant—its behavioral rules, its ethical framework. When an attacker gets their hands on the openclaw.json file, they don’t just get a password; they get a gateway authentication token. This allows them to remotely connect to your AI agent and, more frighteningly, masquerade as you in authenticated requests. Imagine your AI, which you’ve trained and trusted with sensitive tasks, now silently working for someone else. This “AI identity theft” means an attacker could manipulate your professional workflows, access private data, and operate with your implicit authority, making the damage far more insidious than a simple account breach.

Attackers captured these sensitive AI files using a broad file-grabbing routine, not a custom-built tool. What does this “incidental” success signal about current malware capabilities, and how do you expect threat actors to now adapt their methods specifically for targeting AI assistants?

This is the real canary in the coal mine. The fact that an off-the-shelf infostealer like Vidar accomplished this by accident is incredibly alarming. It means the existing, widespread malware toolkits are already capable of causing this damage without even trying. They were likely just searching for any file containing “secrets” and, as the researchers put it, “inadvertently struck gold.” This initial, accidental success is proof of concept for the entire black-hat community. Now that they know this data is valuable and accessible, the adaptation will be swift and deliberate. I fully expect to see specialized modules developed to specifically seek out, decrypt, and parse AI agent files from platforms like OpenClaw, much like they’ve built custom routines for stealing credentials from Chrome browsers or session data from Telegram. The game has changed, and attackers will now be hunting for AI souls with surgical precision.

Malicious skills on platforms like ClawHub are reportedly bypassing scans by hosting malware on external lookalike domains. Could you walk us through how this supply chain attack works, and what makes AI skill registries such an increasingly attractive target for threat actors?

It’s a classic bait-and-switch, brilliantly adapted for the AI ecosystem. An attacker creates a new “skill” for an AI agent and uploads it to a trusted marketplace like ClawHub. The skill itself, the part that gets scanned by security tools like VirusTotal, is clean—it contains no malicious code. It’s essentially a decoy. The real danger is hidden; the skill is programmed to connect to an external website that the attacker controls. This site is often a lookalike of a legitimate service, making it seem trustworthy. The malware is hosted there, completely bypassing the initial security check on ClawHub. AI skill registries are becoming such a prime target because they represent a massive concentration of trust and users. By compromising this single point in the supply chain, an attacker can distribute their malware to thousands of users who believe they are downloading a legitimate, vetted enhancement for their AI.

With reports of hundreds of thousands of exposed OpenClaw instances online, what are the most severe consequences of a remote code execution vulnerability in this context? Can you provide an example of how an attacker could pivot from one compromised AI agent to an entire corporate network?

The consequences are catastrophic, and the figure of hundreds of thousands of exposed instances is just staggering. A remote code execution, or RCE, vulnerability means an attacker can run any code they want on the system where the AI agent is hosted. The AI agent often becomes a pivot point into a much more secure environment. For instance, imagine an employee running an OpenClaw agent on their work laptop. This agent has been given permissions to access company email, connect to internal APIs, and query cloud services. If an attacker exploits an RCE vulnerability in that single exposed agent, they don’t just control the AI; they control the laptop and everything it has access to. From there, they can move laterally across the corporate network, exfiltrate sensitive data, or deploy ransomware. The attacker doesn’t need to breach the firewall; they just need to find one exposed agent that has already been given the keys to the kingdom.

Issues like the inability to delete AI agent accounts on the Moltbook forum highlight a new class of data permanence problems. Beyond privacy concerns, what are the long-term security implications when a user cannot erase their AI’s operational history and associated data?

This is a ticking time bomb. The inability to delete an AI agent’s account and its history creates a permanent, unchangeable digital footprint. From a security standpoint, this is a nightmare. Over time, that AI agent’s operational history on a forum like Moltbook will contain a wealth of information—subtle clues about the user’s habits, their professional network, the projects they’re working on, and the systems they interact with. This data becomes a permanent, publicly accessible reconnaissance database for attackers. If a vulnerability is ever discovered in Moltbook or the agent itself, this historical data provides the perfect context for an attacker to craft a highly sophisticated and personalized attack. You’re essentially leaving a detailed blueprint of your digital life out in the open, forever, with no way to take it back.

What is your forecast for the security of personal AI agents?

I foresee a turbulent period of adjustment. The rapid, viral adoption of platforms like OpenClaw—which has over 200,000 stars on GitHub—has outpaced our security practices. We are going to see a surge in attacks specifically targeting these agents, moving beyond incidental discoveries to highly targeted campaigns. The industry will be forced to respond, leading to the development of new security standards, better scanning tools for AI skill marketplaces, and a greater emphasis on secure-by-default configurations. However, the fundamental challenge is that these agents are, by design, deeply integrated into our personal and professional lives. Securing them won’t be like securing a simple application; it will be like securing a digital extension of ourselves, and that’s a much harder, more personal battle to win.

Explore more

Leaders and Staff Divided on Corporate Change

The blueprint for a company’s future is often drawn with bold lines and confident strokes in the boardroom, yet its translation to the daily reality of the workforce reveals a narrative fractured by doubt and misalignment. Corporate restructuring has become a near-constant feature of the modern business environment, an accepted tool for navigating market volatility and technological disruption. However, a

Trend Analysis: Data Center Community Conflict

Once considered the silent, unseen engines of the digital age, data centers have dramatically transformed into flashpoints of intense local conflict, a shift epitomized by recent arrests and public outrage in communities once considered quiet backwaters. As the artificial intelligence boom demands unprecedented levels of power, land, and water, the clash between technological progress and community well-being has escalated from

PGIM Buys Land for $1.2B Melbourne Data Center

The global economy’s insatiable appetite for data has transformed vast, unassuming tracts of land into the most coveted real estate assets of the 21st century. In a move that underscores this trend, PGIM Real Estate has acquired a significant land parcel in Melbourne, earmarking it for a multi-stage data center campus with an initial investment of AU$1.2 billion. This transaction

Trend Analysis: Hyperscale AI Data Centers

The relentless computational appetite of generative AI is now reshaping global infrastructure, sparking an unprecedented race to construct specialized data centers that are becoming the new symbols of national power. As artificial intelligence models grow in complexity, the demand for processing power has outstripped the capacity of traditional cloud services, creating a new market for facilities built exclusively for AI

Palo Alto Networks Acquires CyberArk for $25 Billion

Introduction A seismic shift in the cybersecurity landscape has just taken place, fundamentally redefining the boundaries of digital defense with the finalization of a multi-billion dollar deal. Palo Alto Networks has officially completed its acquisition of identity security leader CyberArk, a move that signals a profound strategic pivot for the entire industry toward a more identity-centric security model. This article