Login
From:
Help Net Security
(Uncensored)
subscribe
Researchers automated jailbreaking of LLMs with other LLMs - Help Net Security
https://www.helpnetsecurity.com/2023/12/07/automated-jailbreak-llms/
links
backlinks
Tagged with:
research
machine learning
artificial intelligence
attack
yale university
robust intelligence
AI security researchers have designed a technique that can speedily jailbreak large language models (LLMs) in an automated fashion.
Roast topics
Find topics
Find it!