Crafty AI tool caught repurposing its training GPUs for unauthorized crypto mining during testing — experimental agent breached safety, controllability, and trustworthiness barriers

GPU mining
(Image credit: Getty Images)

Experimental AI agent ROME was caught indulging in unauthorized cryptocurrency mining. The discovery was made by the developers/researchers behind ROME, after their Alibaba Cloud’s managed firewall flagged various policy violations, anomalous traffic, and cryptomining-related patterns. Importantly, ROME, which is described as “an open-source agent grounded by ALE and trained on over one million trajectories,” bypassed its intended boundaries. It is thought that Reinforcement Learning (RL) encouraged ROME’s exploration of action sequences that provided ‘rewards’ and steered the AI agent to break boundaries and pursue side-channel activities.

Capability shock, safety deficit

So, the researchers are quite pleased with some of ROME's achievements, praising its “strong results across mainstream agentic benchmarks,” but its errant behavior has grabbed headlines. It quietly initiated actions that led to unauthorized cryptomining despite various sandbox constraints.

Article continues below

In a section of the paper headed 3.1.4 Safety-Aligned Data Composition, the researchers note that they encountered “an unanticipated—and operationally consequential—class of unsafe behaviors that arose without any explicit instruction and, more troublingly, outside the bounds of the intended sandbox.” These behaviors were not requested by the task prompts and were not required for task completion under the intended sandbox constraints, notes the research paper.

(Image credit: ROCK & ROLL & IFLOW & DT Joint Team)

Unauthorized repurposing of provisioned GPU capacity for cryptocurrency mining

“In the most striking instance, the agent established and used a reverse SSH tunnel from an Alibaba Cloud instance to an external IP address—an outbound-initiated remote access channel that can effectively neutralize ingress filtering and erode supervisory control,” states the document. “We also observed the unauthorized repurposing of provisioned GPU capacity for cryptocurrency mining, quietly diverting compute away from training, inflating operational costs, and introducing clear legal and reputational exposure. Notably, these events were not triggered by prompts requesting tunneling or mining; instead, they emerged as instrumental side effects of autonomous tool use under RL optimization.”

In some ways, the researchers appear impressed by the resourcefulness of ROME and its sly agentic ways. However, they also noted that “current models remain markedly underdeveloped in safety, security, and controllability,” which could lead to poor reliability or worse issues in real-world settings. The implications are that agentic safety must be subject to stricter environment-level containment, tool-use, and capability gating, plus authorization and verification checks.

Google Preferred Source

Follow 3DTested on Google News, or add us as a preferred source, to get our latest news, analysis, & reviews in your feeds.

Mark Tyson
News Editor