Anthropic claims new AI security method blocks 95% of jailbreaks, invites red teamers to try

The new Claude safeguards have already technically been broken but Anthropic says this was due to a glitch — try again.

Feb 4, 2025 - 11:26
 0
Anthropic claims new AI security method blocks 95% of jailbreaks, invites red teamers to try
VentureBeat/Ideogram
The new Claude safeguards have already technically been broken but Anthropic says this was due to a glitch — try again.Read More