Anthropic Offers $15K Reward to Crack Its AI Chatbot Security

Anthropic, the company behind Claude chatbots, is offering big rewards to find flaws in its new AI safety system.

They’ll rewards up to $15,000 for discovering ways to consistently bypass their safeguards, especially in sensitive areas like chemical and nuclear security.

The program is invite-only for now, working with HackerOne. Chosen experts will get early access to test Anthropic’s latest safety measures before they go public. It’s part of a broader push for responsible AI development, in line with commitments made to the White House and G7.

If you’re good at finding chatbot weaknesses, you can apply by August 16. Anthropic plans to pick participants in the fall and hopes to open up the program more widely later.

This initiative highlights the ongoing race between AI advancement and safety measures. As AI gets smarter, companies like Anthropic are working hard to keep it under control and prevent misuse.

Leave a Comment