Bitget App
Trade smarter
Buy cryptoMarketsTradeFuturesCopyBotsEarn

Anthropic launches $15,000 bug bounty program to strengthen AI safety

CryptopolitanCryptopolitan2024/08/07 16:00
By:By Chris Murithi

Share link:In this post: Anthropic launches $15K bug bounty for AI safety vulnerabilities. Ethical hackers invited to tackle AI ‘universal jailbreak’ attacks. Anthropic sets new transparency standards in AI safety.

Anthropic, an AI start-up supported by Amazon, has launched a bug bounty program and will pay up to $15,000 for every report that identifies critical weaknesses in its artificial intelligence systems. The initiative is one of the most extensive efforts by any company working with advanced language models to crowdsource security testing.

According to the company, the bounty targets “universal jailbreak” attacks, methods that can get around AI safety measures in areas such as bioweapons and cyber threats. Before making its next-generation safety mitigation system available to the public, Anthropic is planning to allow ethical hackers to test it to prevent potential misuse.

 

Starting as an invite-only initiative done in collaboration with HackerOne, Anthropic’s bug bounty program wants cybersecurity researchers’ skills in identifying and fixing vulnerabilities in its AI systems. The company plans to open it up more widely in the future, potentially offering a model of industry-wide cooperation on AI safety.

This comes as the UK’s Competition and Markets Authority (CMA) investigates Amazon’s $4bn investment into Anthropic over potential competition issues. Against this backdrop of increased regulatory scrutiny, focusing on safety could enhance Anthropic’s reputation and set it apart from rivals.

See also Musk sues OpenAI for prioritizing profit even as xAI expands

Anthropic sets new AI safety standards

While OpenAI and Google also have bug bounty programs, they mostly concentrate on traditional software vulnerabilities rather than those specific to artificial intelligence. Meta has been criticized for taking what some regard as a relatively closed approach toward research into ensuring the safe development of increasingly intelligent machines. By explicitly targeting such problems and inviting external examination of them, Anthropic sets a precedent for openness within the sector.

However, there are doubts over whether bug bounties alone can effectively address the full spectrum of concerns related to securing advanced machine learning systems. While valuable for identifying and patching particular flaws, they may not get to grips with broader challenges around AI alignment and long-term safety. A more holistic strategy involving extensive testing, improved interpretability, and potentially new governance structures could be needed to ensure that AI systems remain aligned with human values as they grow more powerful.

0

Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.

PoolX: Stake to earn
CEC, QTLX, GDV and other popular new coins are in hot progress!
Stake now!

You may also like

Texas court dismisses Consensys lawsuit against SEC

The Consensys suit was originally filed in April and included the allegation that the SEC had opened an investigation into Ethereum

Blockworks2024/09/20 06:54

Bitcoin is up 6% since the Fed rate cut. What’s next?

Despite short-term boost, Bybit executive warns investors of “potential challenges posed by economic uncertainty and market fluctuations”

Blockworks2024/09/20 06:10