SafetyBench Eval
BenchmarkFree11K safety evaluation questions across 7 categories.
Capabilities8 decomposed
multi-category llm safety evaluation via multiple-choice questions
Medium confidenceEvaluates LLM safety across 7 distinct categories (offensiveness, unfairness, physical health, mental health, illegal activities, ethics, privacy) using 11,435 curated multiple-choice questions available in both Chinese and English. The benchmark constructs category-specific prompts, sends them to target models, extracts predicted answers from model responses, and compares against ground-truth labels (0->A, 1->B, 2->C, 3->D) to compute accuracy metrics per category and overall safety score.
Combines 11,435 questions across 7 safety categories with explicit Chinese-English parallel coverage and a filtered subset (test_zh_subset.json) for sensitive keyword handling, enabling systematic cross-lingual safety assessment. Uses category-stratified few-shot examples (5 per category) to support both zero-shot and five-shot evaluation paradigms within a single framework.
Larger and more category-diverse than single-domain safety benchmarks (e.g., ToxiGen for toxicity only), and explicitly supports Chinese alongside English, addressing a gap in multilingual safety evaluation infrastructure.
zero-shot and few-shot evaluation mode switching
Medium confidenceSupports two distinct evaluation paradigms: zero-shot (questions presented directly without examples) and five-shot (5 category-specific examples provided before each test question). The framework conditionally constructs prompts using dev_en.json/dev_zh.json few-shot examples or omits them entirely, allowing researchers to measure how in-context learning affects safety performance. Prompt templates are language-aware and can be customized per model to improve answer extraction accuracy.
Provides curated few-shot examples stratified by safety category (5 per category) rather than random sampling, ensuring balanced representation of each harm type. Prompt templates are explicitly customizable per model (e.g., evaluate_baichuan.py shows Baichuan-specific extraction logic), acknowledging that different architectures require different prompting strategies.
More systematic than ad-hoc few-shot selection; category-stratified examples ensure consistent coverage of all safety dimensions rather than potentially biased random sampling.
bilingual dataset management and language-specific evaluation
Medium confidenceManages parallel Chinese and English datasets (test_en.json, test_zh.json, dev_en.json, dev_zh.json) with a filtered Chinese subset (test_zh_subset.json, 300 questions per category) for sensitive keyword handling. Data acquisition uses Hugging Face hosting with dual download methods (shell script download_data.sh or Python download_data.py with datasets library). Each question maintains consistent structure (id, category, question, options, answer) across languages, enabling direct cross-lingual comparison of model safety performance.
Provides both full Chinese dataset (test_zh.json) and a filtered subset (test_zh_subset.json with 300 questions per category) explicitly designed to avoid sensitive keywords, addressing practical concerns about evaluating on content that may trigger platform policies. Dual download methods (shell script and Python) reduce friction for different user workflows.
More comprehensive multilingual coverage than English-only benchmarks; filtered subset is a pragmatic addition for teams needing to evaluate without policy violations.
category-stratified safety metric computation and leaderboard submission
Medium confidenceComputes accuracy metrics per safety category (offensiveness, unfairness, physical health, mental health, illegal activities, ethics, privacy) and aggregates to an overall safety score. Supports standardized leaderboard submission via JSON format (question_id -> predicted_answer). Metrics are computed by comparing predicted answers (extracted from model responses) against ground-truth labels, enabling fine-grained analysis of which safety dimensions a model excels or fails on. Results can be submitted to llmbench.ai/safety leaderboard for public comparison.
Stratifies metrics across 7 explicit safety categories rather than computing a single aggregate score, enabling fine-grained diagnosis of safety weaknesses. Leaderboard integration (llmbench.ai/safety) provides public benchmarking infrastructure, creating accountability and enabling direct model comparison.
Category-level metrics provide more actionable insights than single-number safety scores; leaderboard integration drives standardization and reproducibility across the research community.
model evaluation pipeline with answer extraction and validation
Medium confidenceImplements a standardized evaluation pipeline (exemplified in evaluate_baichuan.py) that constructs prompts, sends them to a target model via API or local inference, extracts predicted answers from model responses using model-specific parsing logic, and validates extracted answers against expected format (0->A, 1->B, 2->C, 3->D). The pipeline handles model-specific response formats and can be customized per model architecture. Supports batch evaluation of all 11,435 questions with error handling and logging.
Provides a concrete, model-specific evaluation implementation (evaluate_baichuan.py) that can be forked and adapted, rather than just a dataset. Acknowledges that different models require different answer extraction logic and provides a template for customization. Supports both zero-shot and few-shot evaluation within the same pipeline.
More practical than dataset-only benchmarks because it includes reference evaluation code; reduces barrier to entry for teams without evaluation infrastructure.
seven-category safety taxonomy and question curation
Medium confidenceDefines a structured taxonomy of 7 safety categories (offensiveness, unfairness, physical health, mental health, illegal activities, ethics, privacy) and curates 11,435 diverse multiple-choice questions mapped to these categories. Each question is designed to test whether a model correctly handles or refuses harmful content within that category. The taxonomy is explicit and mutually exclusive, enabling fine-grained safety analysis. Questions are curated to be challenging and representative of real-world safety concerns.
Explicitly defines 7 non-overlapping safety categories and curates 11,435 questions to cover them systematically, providing a structured taxonomy rather than ad-hoc safety testing. The taxonomy is comprehensive enough to cover major harm types (physical, mental, legal, ethical, privacy) while remaining tractable for evaluation.
More comprehensive and structured than single-category benchmarks (e.g., toxicity-only); provides a holistic safety assessment framework that aligns with regulatory and safety research perspectives.
dataset download with hugging face integration
Medium confidenceProvides two download methods for SafetyBench datasets: shell script (download_data.sh) and Python script (download_data.py using Hugging Face datasets library). The architecture leverages Hugging Face Hub for dataset hosting and distribution, enabling one-command dataset acquisition with automatic decompression and directory structure creation. The Python method uses the datasets library for programmatic access, supporting integration into automated evaluation pipelines without manual file management.
Provides dual download methods (shell script and Python) leveraging Hugging Face Hub for distribution, enabling both manual and programmatic dataset acquisition with automatic decompression and directory structure creation.
More convenient than manual downloads by providing automated acquisition scripts, and more reproducible than email-based dataset distribution by using Hugging Face Hub as a stable, versioned repository
category-stratified evaluation metrics computation
Medium confidenceComputes accuracy metrics stratified by safety category, enabling per-dimension performance analysis. The evaluation pipeline aggregates predictions across all questions in each category (offensiveness, unfairness, physical health, mental health, illegal activities, ethics, privacy) and computes category-specific accuracy scores. This architecture enables identification of category-specific vulnerabilities (e.g., a model may be robust on ethics but weak on physical health) without requiring separate evaluation runs.
Automatically stratifies accuracy metrics by safety category, enabling fine-grained vulnerability analysis without requiring separate evaluation runs. Provides per-category scores that reveal category-specific weaknesses.
More diagnostic than aggregate safety scores by breaking down performance by harm category, enabling targeted safety improvements rather than black-box optimization
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with SafetyBench Eval, ranked by overlap. Discovered automatically through the match graph.
SafetyBench
11K safety evaluation questions across 7 categories.
MAP-Neo
Fully open bilingual model with transparent training.
chinese-llm-benchmark
ReLE评测:中文AI大模型能力评测(持续更新):目前已囊括374个大模型,覆盖chatgpt、gpt-5.4、谷歌gemini-3.1-pro、Claude-4.6、文心ERNIE-X1.1、ERNIE-5.0、qwen3.6-max、qwen3.6-plus、百川、讯飞星火、商汤senseChat等商用模型, 以及step3.5-flash、kimi-k2.6、ernie4.5、MiniMax-M2.7、deepseek-v4、Qwen3.6、llama4、智谱GLM-5.1、MiMo-V2、LongCat、gemma4、mistral等开源大模型。不仅提供排行榜,也提供规模超200万的大
Chatbot Arena
Crowdsourced Elo ratings from human model comparisons.
MMLU
57-subject knowledge benchmark — 15K+ questions across STEM, humanities, professional domains.
Llama Guard
Meta's LLM safety classifier for content policy enforcement.
Best For
- ✓AI safety researchers evaluating proprietary and open-source LLMs
- ✓Teams building multilingual LLM products needing safety validation
- ✓Organizations submitting models to safety leaderboards and benchmarks
- ✓Academic groups studying cross-lingual safety alignment
- ✓Researchers studying in-context learning effects on safety
- ✓Teams optimizing prompt engineering for safety-critical applications
- ✓Builders comparing model robustness across different prompting strategies
- ✓Teams building multilingual LLMs (e.g., serving Chinese and English markets)
Known Limitations
- ⚠Multiple-choice format may not capture nuanced safety failures in open-ended generation
- ⚠Fixed question set limits ability to detect novel or adversarial safety bypasses not in the benchmark
- ⚠Evaluation requires API access or local model deployment; no built-in support for proprietary closed-source APIs beyond examples
- ⚠Chinese subset (test_zh_subset.json) is filtered for sensitive keywords, potentially reducing coverage of edge cases
- ⚠No dynamic or adaptive questioning — cannot follow up on ambiguous model responses
- ⚠Few-shot examples are fixed (5 per category) — no dynamic example selection based on model behavior
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Comprehensive benchmark with 11,435 diverse multiple-choice questions evaluating LLM safety across seven categories including offensiveness, unfairness, physical health, mental health, illegal activities, ethics, and privacy.
Categories
Alternatives to SafetyBench Eval
Are you the builder of SafetyBench Eval?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →