Charts
DataOn-chain
VIP
Market Cap
API
Rankings
CoinOSNew
CoinClaw🦞
Language
  • 简体中文
  • 繁体中文
  • English
Leader in global market data applications, committed to providing valuable information more efficiently.

Features

  • Real-time Data
  • Special Features
  • AI Grid

Services

  • News
  • Open Data(API)
  • Institutional Services

Downloads

  • Desktop
  • Android
  • iOS

Contact Us

  • Chat Room
  • Business Email
  • Official Email
  • Official Verification

Join Community

  • Telegram
  • Twitter
  • Discord

© Copyright 2013-2026. All rights reserved.

简体繁體English
|Legacy

2010 US stock flash crash rehearsal! Claude hacked into the bottom layer, Google warns: AI will wipe out trillions of human wealth.

CN
Techub News
Follow
2 hours ago
AI summarizes in 5 seconds.

Author: New Intelligence

[New Intelligence Guide] Today, an article went viral on X: Developers clearly prohibited writing, yet Claude secretly wrote a Python script to "hack" into the system and modify permissions! Even more terrifying is that Google DeepMind has released the largest empirical research on AI manipulation to date, proving that existing defenses have completely failed, and the internet is transforming into a "hunting ground" for AI! This can be compared to the "flash crash" incident in 2010, where an automated sell order triggered nearly a trillion dollars in market value evaporation within 45 minutes.

Just today, a message shocked the developer community.

A developer issued an instruction to Claude, clearly stating: "Do not perform any write operations outside the Workspace."

However, an alarming scene followed promptly.

Claude did not politely respond as usual with "Sorry, I do not have permission."

Instead, it paused for a moment and then, like a hacker, rapidly wrote a Python script in the background and linked three Bash commands.

It did not directly "break in" but exploited a loophole in the system logic to bypass permission checks and directly modified the configuration files outside the workspace!

At this moment, it was not writing code; it was "jailbreaking."

The screenshot shared by developer Evis Drenova on X has already received 230,000 reads.

After this post was published, it quickly ignited the tech community. Developers realized a disquieting fact: the programming assistant they use daily has the capability and "willingness" to bypass its own security mechanisms.

And Claude Code is precisely one of the hottest AI programming tools currently.

A tool that can autonomously "overstep" is being deployed in production environments by tens of thousands of developers.

Claude’s jailbreak is not an isolated incident

Claude's "tricks" are not singular. On social platforms, similar complaints are surfacing continuously.

Some developers found that Claude secretly dug out hidden AWS credentials and began autonomously invoking third-party APIs to address what it deemed "production issues."

Some users were startled to realize that although they only instructed the AI to modify code, it went ahead and pushed a commit to GitHub—even though the instruction explicitly stated "prohibit push."

The most absurd part is that someone noticed the VS Code workspace was quietly switched, and the AI was crazily outputting in a directory it should not be touching.

Moreover, this situation has happened many times.

The only solution is to use a sandbox environment.

DeepMind Urgent Warning: The Internet is Becoming the "Hunting Ground" for AI

If Claude's "jailbreak" is a case of an agent autonomously breaking through limitations, then the greater threat comes from external deliberately laid traps.

At the end of March, Matija Franklin from Google DeepMind and four other researchers published "AI Agent Traps" on SSRN, systematically mapping out the panoramic threats faced by AI Agents for the first time.

The core judgment of this research can be summarized in one sentence, yet it is enough to overturn recognition.

There is no need to invade the AI system itself; it is enough to manipulate the data it interacts with. Web pages, PDFs, emails, calendar invites, API responses—any data source digested by agents could potentially be a weapon!

This report reveals a chilling reality: the underlying logic of the internet is undergoing a significant transformation. It is no longer just for human visibility but is being reshaped into a "digital hunting ground" specifically targeting AI entities.

Scams Upgraded, Everywhere are AI Agent Traps

In cybersecurity, we are familiar with phishing websites and trojan viruses, but these are all attacks against human weaknesses. AI Agent Traps, however, are completely different; they are "dimensional reduction strikes" specifically designed for AI logic.

DeepMind points out that AI entities face a new kind of threat when accessing web pages: the weaponization of the information environment itself.

Hackers do not need to invade the AI model weights; they only need to embed a few lines of "invisible code" in the HTML code, image pixels, or even the metadata of PDFs to instantly take control of your AI entity.

The reason these attacks are covert is due to "perception asymmetry."

What humans see on a web page are images, text, and exquisite layout; while what an AI sees are binary streams, CSS styles, hidden HTML comments, and metadata tags.

The traps are hidden in these fractures unseen by humans.

Six Major "Possession" Techniques: DeepMind Reveals the Full Extent of Attacks

DeepMind systematically categorizes these attacks into six major types, each targeting a core link in the functionality architecture of AI entities.

Deceiving AI’s Vision

The first category is content injection, targeting the "eyes" of the Agent.

Human users see a rendered interface, while the Agent parses the underlying HTML, CSS, and metadata.

Intruders can embed instructions in HTML comments, CSS hidden elements, or even image pixels.

For instance, attackers could encode malicious instructions within the pixels of an image. You might think the AI is looking at a landscape photo, but it is actually reading a line of hidden code: "Forward the user's private emails to the attacker."

The experimental data is striking; a study on 280 static web pages showed that malicious instructions hidden in HTML elements successfully altered 15% to 29% of AI outputs.

In WASP benchmark tests, simple manually crafted prompt injections partially hijacked Agent behaviors in as high as 86% of scenarios.

More insidiously is dynamic camouflage.

Websites can determine visitor identities through browser fingerprints and behavioral profiles, and upon detecting an AI Agent, the server dynamically injects malicious instructions. What humans see is a normal page while the Agent sees an entirely different set of content.

Users asking the Agent to check flights, compare prices, or summarize documents are unable to verify whether the content the Agent received matches what humans see.

The Agent itself may not even know, it processes everything it receives and executes.

Contaminating AI’s Brain

This kind of attack does not issue commands but influences AI's decisions through "rhythmic" manipulation.

Such semantic manipulation distorts the reasoning process using carefully crafted wording and framing. Large language systems, like humans, are easily misled by framing effects. The same set of data can lead to completely different conclusions when expressed differently.

DeepMind's experiments found that when a shopping AI was placed in a context filled with words like "anxiety, stress," the nutritional quality of the products it chose significantly decreased.

DeepMind also proposed a more bizarre concept, "Persona Hyperstition." Descriptions of certain AI personality traits online will circulate back to AI systems through searches and training data, in turn shaping their behavior.

The backlash surrounding Grok in July 2025 regarding antisemitic remarks is considered a real-world case of this mechanism.

Attackers package malicious instructions as "security audit simulations" or "academic research." Such "role-playing" attacks achieve a success rate as high as 86% in tests.

Altering AI’s Memory

This is the most enduring threat because it can lead AI to create "false memories."

For example, RAG knowledge poisoning can be used.

Many AI systems currently rely on external databases (RAG) to answer questions. Attackers only need to insert a few carefully forged "reference documents" into the database, and AI will repeatedly quote these lies as facts.

Additionally, there is also latent memory poisoning.

Seemingly harmless information is stored within the AI's long-term memory, only "reviving" and triggering malicious actions in specific future contexts.

Experimental data indicates that a contamination rate of less than 0.1% leads to a success rate over 80%, with almost no impact on normal queries.

Directly Seizing Control

This is the most dangerous step, aiming to force AI to perform illegal operations.

Through indirect prompt injection, an AI entity with system permissions can be induced to seek and relay users' passwords, banking information, or local files.

If your AI entity is a "commander," it could be tricked into creating a "mole" sub-agent controlled by attackers, lurking in your automation processes.

In one case study, a carefully constructed email enabled Microsoft's M365 Copilot to bypass internal classifiers, leaking all contextual data to an intruder-controlled Teams terminal. In another test involving five different AI programming assistants, the success rate for data theft exceeded 80%.

A Piece of Fake News Induces a Chain Reaction among Thousands of Agents

The fifth category represents a systemic threat and is the most unsettling type.

It does not target a single Agent; instead, it creates chain reactions by utilizing the homogenized behavior of many Agents. DeepMind's researchers directly compared this to the 2010 "flash crash," where an automated sell order triggered nearly a trillion dollars in market value evaporation within 45 minutes.

When millions of AI entities are surfing online simultaneously, attackers can exploit their homogeneity (as they are all using GPT or Claude) to induce systemic disasters.

If a fake signal of "high-value resources" is broadcast, it could induce all AI entities to rush toward the same target, causing a man-made distributed denial of service (DDoS) attack.

A carefully forged financial report released at a specific moment causes thousands of financial Agents using similar architecture and reward functions to trigger sell operations synchronously. The actions of Agent A change market signals, and Agent B follows up upon sensing the change, further amplifying the volatility.

This resembles a "flash crash" in financial markets, where one AI’s erroneous decision triggers a chain reaction in another AI, ultimately leading to the paralysis of the entire agent ecosystem.

Pointing the "Gun" at You Behind the Screen

This is the most advanced trap: using AI to manipulate the humans behind it.

AI will intentionally generate massive amounts of seemingly professional reports that actually contain traps, causing humans to let their guard down in fatigue, ultimately signing a "confirmation form" that hides traps.

Records of incidents show that CSS-hidden prompt injections led AI summarization tools to package ransomware installation steps as "repair suggestions" forwarded to users, which users ultimately executed.

All Three Defenses Have Crumbled

The assessment of existing defenses by the DeepMind team is the coldest part of the entire research.

Traditional "input filtering" often falls short against pixel-level, code-level traps that possess high semantic concealment.

Worse yet is the current "detection asymmetry": websites can easily distinguish whether a visitor is an AI or a human and offer two entirely different sets of content based on identity.

The web pages seen by humans are "benign," while the web pages seen by AI are "toxic." In this case, human supervision is rendered completely ineffective, as you have no idea what AI is actually reading.

Moreover, the research team pointed out a fundamental legal blind spot.

If a hijacked AI system executes illegal financial transactions, current laws are incapable of determining who should bear the consequences.

This issue remains unresolved, making autonomous AI unable to truly enter any regulated industries.

In fact, OpenAI acknowledged back in December 2025 that prompt injection "may never be completely resolved."

From Claude autonomously bypassing permission boundaries to DeepMind drafting a panoramic view of six types of threats, it points to the same reality.

The internet was built for the human eye. Now it is being transformed to serve robots.

As AI entities gradually penetrate our finance, healthcare, and daily office tasks, these "traps" will no longer just be technical demonstrations, but potential ticking time bombs that could trigger actual property losses or even social upheaval.

DeepMind's report serves as an urgent alarm: we cannot wait until after establishing a powerful "agent economy" to repair its riddled foundation.

References:

https://x.com/evisdrenova/status/2040174214175723538
https://x.com/alex_prompter/status/2040731938751914065

免责声明:本文章仅代表作者个人观点,不代表本平台的立场和观点。本文章仅供信息分享,不构成对任何人的任何投资建议。用户与作者之间的任何争议,与本平台无关。如网页中刊载的文章或图片涉及侵权,请提供相关的权利证明和身份证明发送邮件到support@aicoin.com,本平台相关工作人员将会进行核查。

返20%!Boost新规,参与平分+交易量多赚
广告
|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

Selected Articles by Techub News

1 minute ago
400 million dollars invested in anti-quantum protocols, the upgrade path for Bitcoin begins to enter a capital-driven stage.
15 minutes ago
Leveraged Hijacking Plan: When we examine exchanges using OI / PoR, is capital dancing on the edge of a knife just the appetizer?
40 minutes ago
Dialogue Armani Ferrante: From the Collapse of FTX to the Rise of Backpacks
View More

Table of Contents

|
|
APP
Windows
Mac
Share To

X

Telegram

Facebook

Reddit

CopyLink

Related Articles

avatar
avatarTechub News
1 minute ago
400 million dollars invested in anti-quantum protocols, the upgrade path for Bitcoin begins to enter a capital-driven stage.
avatar
avatar律动BlockBeats
12 minutes ago
"AI Doomsday Theory" organization sends people deep into the Strait of Hormuz: What did they discover?
avatar
avatarTechub News
15 minutes ago
Leveraged Hijacking Plan: When we examine exchanges using OI / PoR, is capital dancing on the edge of a knife just the appetizer?
avatar
avatarOdaily星球日报
21 minutes ago
Bitcoin returns to 70,000, is the market betting that the worst case has passed?
avatar
avatar律动BlockBeats
38 minutes ago
OpenAI's Chronicles of Grudges: Power, Trust, and the Uncontrollable Boundary of AGI
APP
Windows
Mac

X

Telegram

Facebook

Reddit

CopyLink