New AI Jailbreak Technique ‘Bad Likert Judge’ Increases Attack Success Rates by More Than 60%
Emerging Jailbreak Technique Poses New Threats to Language Models Cybersecurity research has recently unveiled a new jailbreak technique that undermines the safety mechanisms of large language models (LLMs), potentially enabling the generation of harmful or malicious content. This multi-turn attack strategy, termed “Bad Likert Judge,” has been revealed by researchers…