MMMU
BenchmarkFreeExpert-level multimodal understanding across 30 subjects.
Capabilities8 decomposed
expert-level multimodal reasoning evaluation across 30 college subjects
Medium confidenceEvaluates AI models on 11,500 expert-level questions spanning 6 disciplines (Art & Design, Business, Science, Health & Medicine, Humanities & Social Science, Tech & Engineering) and 183 subfields, requiring simultaneous perception of heterogeneous visual modalities (charts, diagrams, chemical structures, music sheets, maps, tables) and application of college-level domain knowledge with deliberate multi-step reasoning. Questions are sourced from actual college exams, textbooks, and lectures to ensure authentic difficulty and real-world relevance.
MMMU is the only benchmark combining (1) 11,500 questions across 30 college subjects and 183 subfields, (2) 30 heterogeneous visual modality types (including domain-specific visuals like chemical structures and music sheets), and (3) explicit sourcing from authentic college exams/textbooks/lectures rather than synthetic or crowdsourced data. This scale and diversity of real-world academic content distinguishes it from narrower benchmarks like MMVP or ScienceQA which focus on single domains or simpler visual reasoning.
MMMU covers 6x more disciplines and 3x more subjects than domain-specific benchmarks (e.g., MedQA for medicine only), and includes heterogeneous visual types (chemical structures, music sheets) absent from general-purpose multimodal benchmarks like LVLM-eHub, making it the most comprehensive test of expert-level multimodal reasoning across academic domains.
discipline-specific performance stratification and diagnostic breakdown
Medium confidenceProvides granular performance metrics stratified across 6 core academic disciplines (Art & Design, Business, Science, Health & Medicine, Humanities & Social Science, Tech & Engineering) and 183 subfields, enabling identification of which knowledge domains and subject areas a model excels or struggles with. Leaderboard and evaluation infrastructure expose per-discipline accuracy, per-subject accuracy, and per-visual-modality accuracy to support targeted model improvement and domain-specific capability assessment.
MMMU's 183-subfield taxonomy enables fine-grained diagnostic analysis unavailable in coarser benchmarks. The explicit mapping of questions to both discipline and visual modality type allows simultaneous analysis of domain knowledge gaps and visual perception weaknesses, supporting root-cause analysis of model failures.
Unlike general multimodal benchmarks (LVLM-eHub, MMBench) that report only aggregate accuracy, MMMU's discipline-stratified breakdown enables targeted optimization for specific domains, making it actionable for domain-specific AI development rather than just comparative ranking.
heterogeneous visual modality evaluation with domain-specific visual types
Medium confidenceEvaluates multimodal model performance across 30 distinct visual modality types including domain-specific visuals (chemical structures, music sheets, mathematical diagrams) alongside common types (charts, tables, maps, photographs, illustrations). The benchmark explicitly tests whether models can perceive and reason over specialized visual representations used in professional and academic contexts, not just natural images or generic diagrams.
MMMU explicitly includes 30 heterogeneous visual modality types with emphasis on domain-specific visuals (chemical structures, music sheets, mathematical diagrams) rarely tested in general multimodal benchmarks. This design choice reflects real-world use cases where multimodal AI must handle specialized visual representations, not just natural images and generic charts.
Most multimodal benchmarks (MMBench, LLaVA-Bench) focus on natural images and simple charts; MMMU's inclusion of domain-specific visuals (chemistry, music, engineering) makes it the only benchmark validating multimodal AI for professional knowledge work requiring specialized visual literacy.
remote and local evaluation infrastructure with dual submission pathways
Medium confidenceProvides two evaluation pathways: (1) remote submission via EvalAI server (established 2023-12-04) with test set answers released for local verification (2026-02-12), and (2) local evaluation capability enabling offline batch evaluation of models on the full 11,500-question dataset. The dual infrastructure supports both cloud-based leaderboard submission and self-hosted evaluation for organizations with data privacy or latency constraints.
MMMU's dual evaluation infrastructure (remote EvalAI + local offline) is unusual for academic benchmarks, enabling both official leaderboard participation and privacy-preserving self-hosted evaluation. The 2026-02-12 release of test set answers for local verification suggests a hybrid model balancing leaderboard integrity with reproducibility.
Unlike benchmarks requiring cloud submission (e.g., GLUE, SuperGLUE), MMMU enables local evaluation for organizations with data privacy constraints, while still supporting official leaderboard ranking for research reproducibility.
multimodal perception and knowledge integration assessment
Medium confidenceExplicitly evaluates three integrated capabilities: (1) perception (understanding diverse visual modalities), (2) knowledge (domain-specific subject expertise), and (3) reasoning (deliberate multi-step reasoning over multimodal inputs). Questions are designed to require simultaneous visual understanding and domain knowledge application, preventing models from succeeding through either perception alone or knowledge lookup alone. This integration testing approach validates end-to-end multimodal reasoning rather than isolated sub-capabilities.
MMMU's explicit design to require simultaneous perception, knowledge, and reasoning (rather than testing each in isolation) reflects real-world expert tasks where these capabilities must be integrated. Questions cannot be solved by visual recognition alone or knowledge lookup alone, forcing genuine multimodal reasoning.
Most multimodal benchmarks (MMBench, LLaVA-Bench) test visual recognition or simple visual question-answering; MMMU's integration of expert-level domain knowledge with visual reasoning creates a more realistic assessment of multimodal AI readiness for professional applications.
mmmu-pro robust variant with enhanced evaluation rigor
Medium confidenceMMMU-Pro (introduced 2024-09-05) is a refined version of the base MMMU benchmark designed for more robust multimodal AI evaluation. The distinction from base MMMU is not fully documented in public materials, but the designation as 'robust' suggests improvements in question quality, answer verification, or evaluation methodology to reduce noise and improve benchmark reliability.
unknown — insufficient data. MMMU-Pro is mentioned as a 'robust version' but specific improvements over base MMMU are not documented in available materials.
unknown — insufficient data to compare MMMU-Pro against base MMMU or other robust benchmark variants.
human expert baseline and comparative performance analysis
Medium confidenceProvides human expert performance baseline on the full 11,500-question dataset, enabling assessment of whether AI models are approaching or exceeding human-level performance on expert-level multimodal reasoning tasks. The leaderboard (updated 2024-01-31) includes human expert scores, allowing direct comparison of AI model performance against domain expert accuracy.
MMMU's inclusion of human expert baseline (updated 2024-01-31) enables direct AI-vs-human comparison on expert-level tasks, a feature absent from many multimodal benchmarks. This design choice reflects the benchmark's focus on assessing AI readiness for professional knowledge work where human performance is the relevant reference point.
Unlike benchmarks with only AI baselines (GPT-4V, Claude), MMMU's human expert baseline enables assessment of whether AI is approaching human-level performance, critical for evaluating deployment readiness in professional domains.
college-level authentic sourcing from exams, textbooks, and lectures
Medium confidenceQuestions are explicitly sourced from authentic college-level materials (exams, textbooks, lectures) rather than synthetic generation or crowdsourcing, ensuring real-world difficulty, relevance, and alignment with actual academic standards. This sourcing approach guarantees that benchmark questions reflect genuine expert-level reasoning requirements rather than artificial or simplified tasks, and reduces risk of benchmark gaming through memorization of synthetic patterns.
MMMU's explicit commitment to sourcing questions from authentic college exams, textbooks, and lectures (rather than synthetic generation) ensures benchmark questions reflect genuine expert-level reasoning requirements. This design choice reduces benchmark gaming and improves correlation with real-world expert task performance.
Most multimodal benchmarks use crowdsourced or synthetically generated questions; MMMU's authentic sourcing from college materials ensures questions reflect real academic standards and reduces risk of AI systems gaming synthetic patterns without genuine reasoning capability.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with MMMU, ranked by overlap. Discovered automatically through the match graph.
MathVista
Visual mathematical reasoning benchmark.
11-777: MultiModal Machine Learning (Fall 2022) - Carnegie Mellon University

Tutorial on MultiModal Machine Learning (ICML 2023) - Carnegie Mellon University

Language Is Not All You Need: Aligning Perception with Language Models (Kosmos-1)
* ⭐ 03/2023: [PaLM-E: An Embodied Multimodal Language Model (PaLM-E)](https://arxiv.org/abs/2303.03378)
RealWorldQA
Real-world visual QA requiring spatial reasoning.
LLaVA 1.6
Open multimodal model for visual reasoning.
Best For
- ✓AI research teams developing multimodal large language models (LLMs) and vision-language models (VLMs)
- ✓Organizations evaluating commercial multimodal AI systems (GPT-4V, Claude, Gemini) for domain-specific applications
- ✓Academic institutions benchmarking student-built AI systems against established baselines
- ✓Enterprise teams assessing whether multimodal AI is ready for professional knowledge work (medicine, engineering, business analysis)
- ✓Model developers optimizing multimodal systems for specific professional domains (e.g., medical AI, engineering design tools)
- ✓Research teams analyzing failure modes and knowledge gaps in vision-language models
- ✓Enterprise teams selecting multimodal AI for domain-specific applications (e.g., medical diagnosis support, legal document analysis)
- ✓AI teams building domain-specific multimodal systems (chemistry analysis, music transcription, engineering design review)
Known Limitations
- ⚠Benchmark is static and college-level scoped — does not measure real-time interactive reasoning, multi-turn dialogue, or adversarial robustness
- ⚠Scoring methodology not explicitly documented in public materials — exact match vs. partial credit scoring formula unknown
- ⚠Train/dev/test split ratios and data contamination analysis not publicly disclosed, creating uncertainty about overlap with LLM training corpora
- ⚠No published analysis of demographic biases in question selection or subject representation balance across the 30 subjects
- ⚠Evaluation requires either remote submission to EvalAI server or local execution environment — no lightweight API-based evaluation option documented
- ⚠Performance ceiling not yet saturated (GPT-4V at 56% accuracy) but no analysis of whether benchmark has inherent ceiling effects at higher capability levels
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Massive Multi-discipline Multimodal Understanding benchmark with 11,500 expert-level questions across 30 subjects requiring college-level domain knowledge and deliberate reasoning over images, diagrams, and charts.
Categories
Alternatives to MMMU
Are you the builder of MMMU?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →