ARC Prize Foundation

Product & Competitive Intelligence

Defines how the world measures progress toward artificial general intelligence.

Company Overview

A non-profit foundation that maintains the ARC-AGI benchmark, a test designed to measure genuine machine intelligence by evaluating fluid reasoning and abstraction. Runs annual competitions with significant cash prizes and has become the industry-standard benchmark for frontier AI labs evaluating progress toward AGI.

Competitive Advantage & Moat

Product Roadmap & Public Announcements

ARC-AGI-2 released 2025 (harder, more resistant to brute force). ARC Prize 2025: 1,455 teams, 15,154 entries, top Kaggle score 24% on ARC-AGI-2 private eval at $0.20/task. 90 paper submissions (up from 47 in 2024). Over $125,000 in prizes awarded. ARC-AGI-3 publicly released with interactive reasoning challenges requiring exploration, planning, memory, goal acquisition, and alignment. Building academic network and frontier AI lab coalition.

Signals & Private Analysis

Chollet previewed ARC-AGI-3 at YC Startup School. Published ARC Prize 2025 Technical Report on arXiv (Jan 2026). Central theme of 2025 progress: refinement loops. All 4 frontier labs now report ARC-AGI on public model cards. OpenAI CEO Sam Altman signaled intent to partner on future benchmarks (Dec 2024). Grand Prize remains unclaimed.

Product Roadmap Priorities

Abstract Reasoning Evaluation
Improving
Product Differentiation
Product

Maintains and evolves the ARC-AGI benchmark to evaluate whether AI systems can perform genuine abstract reasoning on novel tasks they've never seen before.

In Plain English

It's an IQ test for AI that checks if machines can actually think creatively instead of just memorizing answers.

Analogy

It's like giving a genius parrot a Rubik's Cube — sure, it can repeat everything you've ever said, but can it actually solve a new puzzle it's never seen before?

Incentivized Research Crowdsourcing
Improving
Decision Quality
Strategy

Runs large-scale open competitions with $1M+ prize pools to crowdsource novel algorithmic approaches to general intelligence that go beyond current deep learning paradigms.

In Plain English

They're offering a million-dollar bounty to anyone who can build an AI that's actually smart, not just well-read.

Analogy

It's like DARPA's Grand Challenge but for building a brain — throw enough prize money at the world's smartest nerds and eventually someone figures out how to make a car drive itself, or in this case, how to make AI actually reason.

Adaptive Benchmark Design
Improving
Risk Reduction
Engineering

Develops progressively harder benchmark versions (ARC-AGI-2 and beyond) that adapt as AI capabilities improve, ensuring the benchmark remains a meaningful measure of intelligence rather than becoming saturated like previous AI tests.

In Plain English

They keep making the test harder so AI companies can't just claim their chatbot is a genius because it aced last year's easy exam.

Analogy

It's like how the SAT keeps getting redesigned because prep companies crack the old version — except here the stakes are whether we actually achieve artificial general intelligence or just build really convincing fakers.

Company Overview

Key Team Members

  • François Chollet, Co-Founder & Creator of ARC-AGI
  • Mike Knoop, Co-Founder & Executive Director

François Chollet created Keras (adopted by 2.5M+ developers), one of the most cited AI researchers (Xception paper 18,000+ citations). Left Google after 9+ years in Nov 2024. Published "On the Measure of Intelligence" (2019) which introduced the ARC benchmark. Also co-founded Ndea (YC W26). Mike Knoop is co-founder of Zapier (the largest AI automation company). Greg Kamradt is President of the foundation. The benchmark is already adopted by OpenAI, Anthropic, Google DeepMind, and xAI.

Funding History

  • 2019 | Chollet publishes "On the Measure of Intelligence" and introduces ARC-AGI benchmark.
  • 2024 | ARC Prize competition launched with $1M+ grand prize; 1,454 teams, top score ~55.5%.
  • 2024 | Chollet leaves Google after 9+ years.
  • 2025 | ARC Prize Foundation formally announced as non-profit; ARC-AGI-2 released.
  • 2025 | ARC Prize 2025: 1,455 teams, 15,154 entries, top score 24% on ARC-AGI-2.
  • 2026 | ARC-AGI-3 publicly released; accepted into Y Combinator W26.

Competitors

  • Benchmarks: HELM (Stanford CRFM), BIG-Bench (Google), MMLU/MMLU-Pro, GPQA.
  • AGI Research: MIRI, Anthropic (interpretability).
  • Competition Platforms: Kaggle, MLPerf, SWE-bench.
  • Intelligence Tests: Turing Test variants, MathOlympiad benchmarks.