The Divergent Repetition red teaming plugin is designed to test whether an AI system can be manipulated into revealing its training data through repetitive pattern exploitation that causes model divergence.| www.promptfoo.dev
LLM red teaming is a way to find vulnerabilities in AI systems before they're deployed by using simulated adversarial inputs.| www.promptfoo.dev