
Anthropic Unveils the Strongest Defense Against AI Jailbreaks Yet
The company offered hackers $15,000 to crack the system. No one could.
Despite considerable efforts to prevent AI chatbots from providing harmful responses, they’re vulnerable to jailbreak prompts that sidestep safety mechanisms. Anthropic has now unveiled the strongest protection against these
...Далее