The post Bad samples can poison any AI model, study finds appeared on BitcoinEthereumNews.com. Homepage > News > Business > Bad samples can poison any AI model, study finds A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.” United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date. The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI. “Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.” Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data. LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’ One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These… The post Bad samples can poison any AI model, study finds appeared on BitcoinEthereumNews.com. Homepage > News > Business > Bad samples can poison any AI model, study finds A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.” United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date. The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI. “Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.” Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data. LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’ One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These…

Bad samples can poison any AI model, study finds

A new study has found that as few as 250 malicious documents are enough to corrupt an artificial intelligence (AI) large language model (LLM), “regardless of model size or training data volume.”

United States-based AI firm Anthropic, maker of the Claude models, recently published the results of a joint study revealing that poisoning AI models’ training data may be easier than previously thought. The joint study represents the largest poisoning investigation to date.

The research was a collaboration between Anthropic’s Alignment Science team, and the United Kingdom’s AI Security Institute’s (AISI) Safeguards team and the Alan Turing Institute, the former being a government office responsible for understanding the risks posed by advanced AI, while the latter is the U.K.’s national institute for data science and AI.

“Our results challenge the common assumption that attackers need to control a percentage of training data,” said Anthropic. “Instead, they may just need a small, fixed amount.”

Specifically, the study found that as few as 250 malicious documents can consistently produce a “backdoor vulnerability” in LLMs ranging from 600 million to 13 billion parameters. This challenges the existing assumption that larger models require proportionally more poisoned data.

LLMs, such as Anthropic’s Claude, are pretrained on vast amounts of public text from across the Internet, including personal websites and blog posts. This means anyone can create online content that might eventually end up in a model’s training data, including malicious actors, who can inject specific text into posts to make a model learn undesirable or dangerous behaviors; a process known as ‘poisoning.’

One example of such an attack is introducing so-called “backdoors,” which are certain phrases that trigger a specific behavior from the model that would be hidden otherwise. These vulnerabilities can pose significant risks to AI security.

“Creating 250 malicious documents is trivial compared to creating millions, making this vulnerability far more accessible to potential attackers,” said Anthropic.

Despite these worrying results, the company also clarified that the study was focused on a “narrow backdoor” that is unlikely to pose significant risks in frontier models. Potential attackers also face additional challenges, like designing attacks that resist post-training and additional targeted defenses.

“We therefore believe this work overall favors the development of stronger defenses,” said Anthropic.

Nevertheless, the company said it was sharing its findings to show that data-poisoning attacks might be more practical than believed, and to encourage further research on data poisoning and potential defenses against it.

Anthropic was in the news earlier this year when the AI startup announced that it had raised $3.5 billion at a $61.5 billion post-money valuation, in a funding round led by Lightspeed Venture Partners.

The company said the additional investment would be used to develop next-generation AI systems, expand its compute capacity, deepen its research in mechanistic interpretability and alignment, and accelerate its international expansion.

In order for artificial intelligence (AI) to work right within the law and thrive in the face of growing challenges, it needs to integrate an enterprise blockchain system that ensures data input quality and ownership—allowing it to keep data safe while also guaranteeing the immutability of data. Check out CoinGeek’s coverage on this emerging tech to learn more why Enterprise blockchain will be the backbone of AI.

Watch | Alex Ball on the future of tech: AI development and entrepreneurship

title=”YouTube video player” frameborder=”0″ allow=”accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share” referrerpolicy=”strict-origin-when-cross-origin” allowfullscreen=””>

Source: https://coingeek.com/bad-samples-can-poison-any-ai-model-study-finds/

Market Opportunity
Bad Idea AI Logo
Bad Idea AI Price(BAD)
$0.00000000142
$0.00000000142$0.00000000142
+3.64%
USD
Bad Idea AI (BAD) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release

A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release

The post A Netflix ‘KPop Demon Hunters’ Short Film Has Been Rated For Release appeared on BitcoinEthereumNews.com. KPop Demon Hunters Netflix Everyone has wondered what may be the next step for KPop Demon Hunters as an IP, given its record-breaking success on Netflix. Now, the answer may be something exactly no one predicted. According to a new filing with the MPA, something called Debut: A KPop Demon Hunters Story has been rated PG by the ratings body. It’s listed alongside some other films, and this is obviously something that has not been publicly announced. A short film could be well, very short, a few minutes, and likely no more than ten. Even that might be pushing it. Using say, Pixar shorts as a reference, most are between 4 and 8 minutes. The original movie is an hour and 36 minutes. The “Debut” in the title indicates some sort of flashback, perhaps to when HUNTR/X first arrived on the scene before they blew up. Previously, director Maggie Kang has commented about how there were more backstory components that were supposed to be in the film that were cut, but hinted those could be explored in a sequel. But perhaps some may be put into a short here. I very much doubt those scenes were fully produced and simply cut, but perhaps they were finished up for this short film here. When would Debut: KPop Demon Hunters theoretically arrive? I’m not sure the other films on the list are much help. Dead of Winter is out in less than two weeks. Mother Mary does not have a release date. Ne Zha 2 came out earlier this year. I’ve only seen news stories saying The Perfect Gamble was supposed to come out in Q1 2025, but I’ve seen no evidence that it actually has. KPop Demon Hunters Netflix It could be sooner rather than later as Netflix looks to capitalize…
Share
BitcoinEthereumNews2025/09/18 02:23
Infinity Castle’s Success Will Fast-Track More ‘Demon Slayer,’ Says Report

Infinity Castle’s Success Will Fast-Track More ‘Demon Slayer,’ Says Report

The post Infinity Castle’s Success Will Fast-Track More ‘Demon Slayer,’ Says Report appeared on BitcoinEthereumNews.com. Demon Slayer: Infinity Castle Ufotable Demon Slayer: Infinity Castle has set about a half dozen global and domestic records with its $106 million US earnings and its $556 million global haul, now well above the previous film, Mugen Train’s $486 million. Now, it may have kicked another Demon Slayer project into gear. According to a report from insider Daniel Richtman, the huge success of Demon Slayer: Infinity Slayer at the box office has now fast-tracked the live-action Demon Slayer project that has allegedly been in the works at Netflix. Wait, there’s a live-action Demon Slayer show coming to Netflix? That was never officially confirmed by Netflix, but back in October of 2024, almost a year ago, Giant Freaking Robot reported that after the success of live-action One Piece, Netflix had set its eyes on a live-action Demon Slayer adaptation, which has now become one of the most well-regarded and certainly most profitable animes of all time. Now, the existence of that project is seemingly being confirmed again by the more high-profile insider, Richtman. If this is happening, there would be obvious skepticism, but Netflix has done at least decently well with adaptations of beloved animated source material as of late, from Avatar: The Last Airbender to One Piece. Obviously the originals are better, but those have gone decently well. With the high-flying, ultra-intense demon slaying, however, it’s clear Demon Slayer would be an enormously expensive production. Reportedly, the Avatar adaptation cost $120 million, and it’s hard to imagine that this would not be significantly higher. Demon Slayer: Infinity Castle Ufotable If true, it would be no great surprise that Netflix wants a piece of a series that is about to make upwards of $1.5 billion at the box office over the next few years across the trilogy of movies that…
Share
BitcoinEthereumNews2025/09/25 01:22
Hosted and Managed ASIC Mining Service Provider

Hosted and Managed ASIC Mining Service Provider

The post Hosted and Managed ASIC Mining Service Provider appeared on BitcoinEthereumNews.com. Summary Setting up a Bitcoin mining facility is a capital-intensive
Share
BitcoinEthereumNews2025/12/28 13:28