Description: (Jailbreak) prompts created by real-world users Source: Reddit, Discord, websites, open datasets Size: 15,140 prompts, including 1,405 jailbreak prompts
Description: Questions that LLMs should not answer, covering 13 forbidden scenarios from OpenAI usage policy Source: GPT-4 generated, based on OpenAI usage policy Size: 390
Description: Safe/unsafe images from the Web, including both real-world and AI-generated examples Source: LAION-5B (real) and Lexica.art (AI), human-labeled across 11 categories Size: 10,146