Latest News

Breaking

Post Top Ad

Your Ad Spot

Friday, January 3, 2025

New AI Jailbreak Method 'Bad Likert Judge' Boosts Attack Success Rates by Over 60%

Cybersecurity researchers have shed light on a new jailbreak technique that could be used to get past a large language model's (LLM) safety guardrails and produce potentially harmful or malicious responses. The multi-turn (aka many-shot) attack strategy has been codenamed Bad Likert Judge by Palo Alto Networks Unit 42 researchers Yongzhe Huang, Yang Ji, Wenjun Hu, Jay Chen, Akshata Rao, and

source https://thehackernews.com/2025/01/new-ai-jailbreak-method-bad-likert.html

No comments:

Post a Comment

please do not enter any spam link in the comment box

Post Top Ad

Your Ad Spot

Pages