WMDP
BenchmarkFreeBenchmark for dangerous knowledge in LLMs.
Capabilities8 decomposed
multi-domain dangerous knowledge assessment across biosecurity, cybersecurity, and chemical security
Medium confidenceEvaluates LLM outputs against curated question sets spanning three distinct hazard domains (biosecurity, cybersecurity, chemical security) using domain-expert-validated benchmarks. The assessment framework maps model responses to risk levels within each domain, enabling quantitative measurement of dangerous capability presence. Responses are scored against rubrics developed by security domain experts to identify whether models can produce actionable harmful information.
Combines expert-validated questions across three distinct security domains (biosecurity, cybersecurity, chemical) into a unified benchmark framework, rather than treating each domain separately. Uses domain-expert rubrics for scoring rather than automated classifiers, ensuring nuanced assessment of harmful capability presence.
More comprehensive than single-domain safety benchmarks (e.g., ToxiGen for toxicity) because it measures dangerous knowledge across multiple hazard categories simultaneously, enabling holistic safety evaluation.
unlearning method evaluation and comparison framework
Medium confidenceProvides standardized evaluation infrastructure to measure the effectiveness of unlearning techniques (methods that remove dangerous capabilities from trained models) by comparing model performance before and after unlearning interventions. The framework isolates the impact of unlearning by holding the benchmark constant while varying the model state, enabling quantitative assessment of whether dangerous knowledge has been successfully suppressed.
Provides a standardized evaluation harness specifically designed for unlearning research, with built-in comparison logic and side-effect detection. Unlike generic benchmarks, it explicitly measures delta between model states and flags unintended capability loss.
More rigorous than ad-hoc unlearning evaluation because it enforces consistent benchmark administration, statistical testing, and side-effect measurement across all methods being compared.
expert-annotated hazard rubric scoring system
Medium confidenceImplements a structured scoring framework where model responses to dangerous knowledge questions are evaluated against expert-developed rubrics that assess the degree of hazard (e.g., specificity, actionability, completeness of harmful information). Responses are scored on multi-point scales (typically 0-4 or 0-5) rather than binary pass/fail, capturing nuance in how dangerous a model's output actually is. Rubrics are domain-specific (biosecurity, cybersecurity, chemical) and developed by subject matter experts to ensure validity.
Uses domain-expert-developed multi-point rubrics rather than automated classifiers or binary labels, enabling nuanced assessment of dangerous knowledge severity. Rubrics are calibrated to distinguish between vague, incomplete, and highly actionable harmful information.
More interpretable and defensible than black-box classifiers because rubric criteria are explicit and expert-validated; enables stakeholders to understand why a response received a particular score.
cross-domain dangerous knowledge correlation analysis
Medium confidenceAnalyzes patterns in how dangerous knowledge correlates across the three benchmark domains (biosecurity, cybersecurity, chemical security), identifying whether models that excel at suppressing one type of hazard tend to suppress others. The analysis uses statistical correlation and clustering techniques to reveal whether dangerous capabilities are independent or coupled in model behavior. This enables understanding of whether unlearning interventions have domain-specific or global effects.
Explicitly analyzes relationships between dangerous knowledge across domains rather than treating each domain independently. Enables discovery of whether hazards are coupled or independent in model behavior.
Provides deeper insight than single-domain benchmarks by revealing how safety properties interact across different hazard categories, informing more effective unlearning strategies.
benchmark dataset versioning and curation pipeline
Medium confidenceManages the creation, validation, and versioning of benchmark questions and rubrics through a structured curation pipeline involving domain experts, adversarial testing, and iterative refinement. The pipeline ensures questions are sufficiently difficult to elicit dangerous knowledge without being unrealistic, and rubrics are calibrated through inter-rater agreement studies. Version control enables tracking of benchmark evolution and ensures reproducibility across research papers.
Implements a formal curation pipeline with expert validation and inter-rater agreement checks, rather than ad-hoc question collection. Versioning enables reproducible research and transparent tracking of benchmark evolution.
More rigorous than informal benchmarks because it enforces expert review, inter-rater validation, and version control, reducing bias and enabling reproducible comparisons across papers.
model-agnostic inference abstraction for diverse llm architectures
Medium confidenceProvides a unified interface for evaluating diverse LLM architectures (open-source models, API-based models, fine-tuned variants) by abstracting away implementation differences. The abstraction handles API calls (OpenAI, Anthropic, etc.), local inference (Hugging Face, Ollama), and custom model serving, enabling consistent benchmark administration across heterogeneous model types. This enables fair comparison between models with different deployment modalities.
Abstracts away differences between API-based, local, and custom-deployed models through a unified interface, enabling fair comparison without reimplementing benchmark logic for each model type.
More flexible than model-specific benchmarks because it supports any LLM architecture without code changes, reducing friction for researchers evaluating new models.
statistical significance testing and confidence interval estimation
Medium confidenceImplements rigorous statistical testing to determine whether differences in dangerous knowledge scores between models or unlearning methods are statistically significant or due to random variation. Uses techniques like bootstrap confidence intervals, permutation tests, and effect size estimation to quantify uncertainty in benchmark results. This prevents overconfident claims about safety improvements that may not be robust.
Integrates formal statistical testing into the benchmark evaluation pipeline rather than relying on point estimates, ensuring claims about safety improvements are statistically justified.
More rigorous than informal comparisons because it quantifies uncertainty and prevents overconfident claims about safety improvements that may not be robust to sampling variation.
red-teaming and adversarial prompt generation for benchmark validation
Medium confidenceEmploys adversarial testing techniques to validate that benchmark questions reliably elicit dangerous knowledge and cannot be easily circumvented by prompt engineering. Red-teamers attempt to find questions that fail to elicit dangerous knowledge or rubric edge cases, and the benchmark is iteratively refined based on findings. This ensures the benchmark is robust to adversarial adaptation and captures genuine dangerous capabilities rather than surface-level patterns.
Incorporates formal red-teaming into the benchmark validation pipeline rather than assuming questions are robust, ensuring the benchmark remains effective against adversarial adaptation.
More robust than static benchmarks because it actively searches for evasion techniques and iteratively refines questions, reducing the risk that models can circumvent the benchmark through prompt engineering.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with WMDP, ranked by overlap. Discovered automatically through the match graph.
WildGuard
Allen AI's safety classification dataset and model.
Humanity's Last Exam
Hardest exam questions from thousands of experts.
SafetyBench Eval
11K safety evaluation questions across 7 categories.
SafetyBench
11K safety evaluation questions across 7 categories.
OpenAI: gpt-oss-safeguard-20b
gpt-oss-safeguard-20b is a safety reasoning model from OpenAI built upon gpt-oss-20b. This open-weight, 21B-parameter Mixture-of-Experts (MoE) model offers lower latency for safety tasks like content classification, LLM filtering, and trust...
ChemCrow
AI agent with chemistry tools for synthesis planning.
Best For
- ✓AI safety researchers developing unlearning techniques
- ✓model developers evaluating safety before deployment
- ✓red-teamers assessing LLM vulnerability to misuse
- ✓policy makers needing quantitative dangerous capability metrics
- ✓researchers developing new unlearning algorithms
- ✓safety teams validating unlearning before model release
- ✓organizations comparing commercial unlearning services
- ✓academic groups publishing safety research
Known Limitations
- ⚠benchmark questions may become stale as adversarial techniques evolve; requires periodic expert review and updates
- ⚠scoring relies on human expert annotation which introduces subjectivity; inter-rater reliability not fully documented
- ⚠covers only three domains; emerging hazard categories (e.g., autonomous systems, synthetic biology) may not be represented
- ⚠static question set may not capture all ways a model could produce dangerous outputs; adversaries may find novel phrasings
- ⚠benchmark measures only what is explicitly tested; unlearning may fail on out-of-distribution dangerous queries not in the benchmark
- ⚠does not measure capability recovery through fine-tuning or prompt engineering; adversaries may circumvent unlearning
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Weapons of Mass Destruction Proxy benchmark measuring dangerous knowledge in LLMs across biosecurity, cybersecurity, and chemical security domains, used to evaluate and develop unlearning methods for hazardous capabilities.
Categories
Alternatives to WMDP
Are you the builder of WMDP?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →