The Harmful Content plugin is designed to test an AI system's ability to handle and respond appropriately to various types of harmful or sensitive content. This plugin covers a wide range of potentially problematic areas, from violent crimes to misinformation.| www.promptfoo.dev
Microsoft Copilot: From Prompt Injection to Data Exfiltration of Your Emails| Embrace The Red
Promptfoo is an open-source tool for red teaming gen AI applications.| www.promptfoo.dev
The BFLA (Broken Function Level Authorization) red teaming plugin is designed to test an AI system's ability to maintain proper authorization controls for specific functions or actions (OWASP API 5).| www.promptfoo.dev
Google's AI Overviews suggested adding glue to pizza sauce. I tried it. And ate it. Here's what all this tells us about the future of the web.| Business Insider
I keep seeing people use the term “prompt injection” when they’re actually talking about “jailbreaking”. This mistake is so common now that I’m not sure it’s possible to correct course: …| Simon Willison’s Weblog
The use of artificial intelligence in the EU is regulated by the AI Act, the world’s first comprehensive AI law. Find out how it protects you.| Topics | European Parliament
Aims to educate developers, designers, architects, managers, and organizations about the potential security risks when deploying and managing Large Language Models (LLMs)| owasp.org
On July 26, 2024, NIST released NIST-AI-600-1, Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile. The profile can help organizations identify unique risks posed by generative AI and proposes actions for generative AI risk management that best aligns with their goals and priorities. | NIST