ARC-AGI
BenchmarkFreeAbstract reasoning benchmark with $1M prize for AGI.
Capabilities9 decomposed
interactive-agent-environment-rendering
Medium confidenceRenders abstract reasoning puzzle environments as interactive step-based simulations accessible via Python SDK or REST API, supporting dual render modes: 'terminal' for visual output and headless for high-speed evaluation at 2000+ FPS. Agents interact through GameAction command enums, receiving state updates after each step, enabling real-time agent-environment interaction loops without network latency in local mode.
Dual-mode rendering architecture (terminal + headless) with 2000+ FPS headless performance enables both interactive development and high-throughput benchmark evaluation without code changes, unlike static benchmark suites that require separate evaluation pipelines.
Faster than traditional visual puzzle benchmarks (which require image processing per task) because headless mode operates on abstract game state rather than pixel rendering, enabling 2K+ FPS evaluation vs. typical 10-100 FPS for vision-based benchmarks.
abstract-pattern-recognition-evaluation
Medium confidenceMeasures AI system performance on novel visual puzzles requiring fluid intelligence and abstract reasoning — specifically the ability to recognize patterns in limited examples and generalize to unseen puzzle variants. Tasks are designed to be 'easy for humans, hard for AI' by requiring exploration, perception-to-plan-to-action loops, memory, and goal acquisition without explicit task specifications, forcing genuine reasoning rather than pattern matching on known problem types.
Explicitly designed as an 'unbeaten benchmark' where no AI system has achieved human-level performance, using interactive agent environments rather than static puzzles to force genuine reasoning loops (exploration → perception → planning → action) and prevent shortcut solutions via memorization or pattern matching.
Measures reasoning robustness better than static benchmarks (MNIST, ImageNet) because novel puzzle variants prevent overfitting to known problem distributions, and interactive format forces agentic reasoning rather than single-pass classification.
scorecard-based-performance-tracking
Medium confidenceAggregates agent performance across multiple puzzle tasks into a unified scorecard data structure accessible via `arc.get_scorecard()` method, enabling comparative evaluation of different reasoning systems on the same benchmark. Scorecard system abstracts the underlying scoring formula (pass@k, accuracy, or custom metric) and provides structured output for leaderboard ranking and progress tracking.
Abstracts scoring complexity behind a single method call, enabling leaderboard-compatible evaluation without exposing underlying metric formula, reducing gaming of metrics while maintaining reproducibility across submissions.
Simpler than manual metric computation (typical in academic benchmarks) because scorecard automatically aggregates across all tasks, but less transparent than published formulas — trades interpretability for ease of use.
multi-version-benchmark-progression
Medium confidenceProvides three sequential benchmark versions (ARC-AGI, ARC-AGI-2, ARC-AGI-3) representing evolution from static visual puzzles to interactive agent environments, allowing researchers to track progress across versions and identify capability inflection points. Version progression reflects increasing complexity: from pattern recognition to agentic reasoning with memory and goal acquisition.
Intentionally evolves benchmark format (static → interactive) to match emerging AI capabilities rather than remaining static, enabling detection of capability phase transitions and preventing benchmark saturation that occurs with fixed task distributions.
More sensitive to reasoning capability emergence than single-version benchmarks because version progression forces systems to adapt to new interaction paradigms, preventing solutions that work only on static puzzle formats.
python-sdk-and-rest-api-dual-access
Medium confidenceProvides dual access patterns to benchmark evaluation: Python SDK (`arc_agi.Arcade()`) for local, low-latency evaluation and REST API for remote evaluation and leaderboard submission. SDK supports both authenticated (via ARC_API_KEY) and anonymous access, with authenticated keys enabling 'access to public games at release' and anonymous access providing reduced functionality. REST API enables integration into CI/CD pipelines and cloud-based evaluation infrastructure.
Dual-access architecture (local SDK + remote REST API) enables both rapid local iteration (2000+ FPS headless) and cloud-scale evaluation without code changes, with optional authentication for early access to new tasks — balancing developer velocity with controlled task release.
More flexible than API-only benchmarks (which require network round-trips) and more scalable than SDK-only approaches (which require local compute), enabling both rapid prototyping and distributed evaluation.
open-source-toolkit-with-agent-templates
Medium confidenceDistributes benchmark as open-source Python toolkit with reference agent implementations and templates, enabling researchers to build custom reasoning systems by extending provided base classes. Toolkit includes game environment abstraction, action enums, and scorecard computation, reducing boilerplate for agent development while maintaining compatibility with official leaderboard evaluation.
Open-source distribution with agent templates enables community-driven reasoning system development while maintaining official benchmark compatibility, preventing vendor lock-in and enabling reproducible research — unlike closed benchmarks that require proprietary evaluation infrastructure.
More accessible than academic benchmarks (which often lack reference implementations) and more flexible than commercial platforms (which restrict agent architecture choices), enabling rapid experimentation with novel reasoning approaches.
prize-incentivized-open-source-contribution
Medium confidenceStructures ARC Prize 2026 ($2,000,000 total) with explicit requirement that winning solutions be open-sourced, creating financial incentive for public release of novel reasoning techniques. Prize pool distributed across multiple tiers and submission windows via Kaggle partnership, enabling both individual researchers and teams to compete while ensuring breakthrough techniques become public knowledge.
Ties financial incentives ($2M) directly to open-source release requirement, creating alignment between individual researcher incentives and public knowledge advancement — unlike traditional academic publishing (which doesn't fund development) or commercial competitions (which restrict IP).
More effective at accelerating public AI research than academic grants (which don't incentivize open-source) or commercial competitions (which restrict IP), because it directly rewards both capability development and public release.
human-calibrated-benchmark-design
Medium confidenceBenchmark tasks are explicitly designed to be 'easy for humans, hard for AI' through human calibration methodology, ensuring evaluation measures genuine reasoning gaps rather than domain-specific knowledge or pattern matching. Tasks require exploration, perception-to-action loops, memory, and goal acquisition — capabilities that humans naturally possess but AI systems struggle with, creating a benchmark resistant to scaling-only approaches.
Explicitly designed to resist scaling-only approaches by measuring reasoning capabilities (exploration, memory, goal acquisition) that don't improve with more parameters or data, forcing genuine architectural innovation rather than training data expansion.
More revealing of fundamental capability gaps than scaling benchmarks (which improve with more compute) because it identifies reasoning limitations that scaling cannot overcome, enabling targeted architectural research.
agentic-reasoning-loop-evaluation
Medium confidenceEvaluates AI systems through multi-step agentic reasoning loops where agents must explore puzzle environments, perceive state changes, plan actions, and execute them iteratively — measuring not just final answers but the reasoning process itself. Interactive format forces agents to maintain memory across steps, acquire goals dynamically, and adapt strategies based on environmental feedback, preventing single-pass solutions.
Forces agentic reasoning loops (perception → planning → action → feedback) rather than single-pass classification, measuring agents' ability to maintain state, adapt strategy, and explore environments — capabilities essential for real-world AI systems but absent from static benchmarks.
More realistic than static benchmarks (which don't require adaptation) and more challenging than scripted environments (which have known solutions), because agents must discover effective reasoning strategies through interaction.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with ARC-AGI, ranked by overlap. Discovered automatically through the match graph.
agentscope
Build and run agents you can see, understand and trust.
Gridspace
Revolutionize call centers with AI-driven, real-time communication...
Composabl
Revolutionize industrial automation with intelligent, no-code AI...
MindPal
Build your AI Second Brain with a team of AI agents and multi-agent...
Simplifai
Automate complex business tasks with AI-driven efficiency and...
CXCortex
Revolutionizes business performance with real-time insights into customer interactions, personalized interactions, and automated task...
Best For
- ✓AI researchers benchmarking reasoning systems
- ✓teams building agentic AI systems that need fast iteration
- ✓developers prototyping novel reasoning architectures
- ✓AI capability researchers measuring progress toward AGI
- ✓teams building reasoning-focused AI systems (not domain-specific tools)
- ✓organizations seeking benchmarks resistant to scaling-only approaches
- ✓researchers submitting to ARC Prize competition
- ✓teams benchmarking multiple reasoning architectures
Known Limitations
- ⚠Headless mode requires explicit render_mode parameter; default terminal rendering adds latency overhead
- ⚠Action space (GameAction enum) specifics not documented — requires reverse-engineering from toolkit source
- ⚠No built-in visualization beyond terminal; programmatic rendering requires custom integration
- ⚠Interactive mode requires synchronous step-by-step execution; no batch evaluation API documented
- ⚠Benchmark explicitly shows 'scaling alone will not reach AGI' — may not reward current LLM scaling approaches
- ⚠Visual domain only — does not measure reasoning on language, mathematics, code, or real-world problems
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Abstraction and Reasoning Corpus benchmark designed to measure general intelligence in AI systems through novel visual puzzles requiring abstract pattern recognition, with a $1M prize for solutions matching human performance.
Categories
Alternatives to ARC-AGI
Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.
Compare →Amplication brings order to the chaos of large-scale software development by creating Golden Paths for developers - streamlined workflows that drive consistency, enable high-quality code practices, simplify onboarding, and accelerate standardized delivery across teams.
Compare →Are you the builder of ARC-AGI?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →