Anthropic Developing Constitutional Classifiers to Safeguard AI Models From Jailbreak Attempts
Anthropic announced the development of a new system on Monday that can protect artificial intelligence (AI) models from jailbreaking attempts. Dubbed Constitutional Classifiers, it is a safeguarding technique that can detect when a jailbreaking attempt is made at the input level and prevent the AI from generating a harmful response as a result of it.
What's Your Reaction?