APPLIED AI LABS
Interactive Data Experience · Dec 2022 — Jan 2026

36 Months
That Bent
the Curve

Explore how AI capability, cost, and adoption compressed what should have taken decades into a single, vertiginous sprint.

0 Years of capability gain
Cost reduction
0M Users in months

The 36 Months
That Changed Everything

Click on timeline or press play button and watch ten years of capability gain animate before you in seconds.

Dec 2022 → Jan 2026
Dec 2022 Jun 2023 Dec 2023 Jun 2024 Dec 2024 Jan 2026
2022
Press play or drag the scrubber to travel through AI's most compressed period of progress.
Adoption Speed
2 years
To reach hundreds of millions of users — vs 46 years for electricity to reach ¼ of US households.
Benchmark Saturation
90%+
Frontier models scoring above 90% on MMLU, MMLU-Pro, and GSM8K — tests designed for graduate-level humans.
SWE-Bench Leap
15×
More software engineering bugs solved in 2024 vs the year prior — from single digits to majority-solved.
Cost Drop
100×
Inference cost per million tokens fell from ~$20 to near-cents in roughly two years of continuous improvement.

When Machines
Crossed the Line

Watch AI benchmark curves sprint past the human expert ceiling — not over decades, but in months.

Language Understanding · MMLU-Pro
~90% Frontier models, early 2025
Coding Ability · SWE-Bench Verified
50%+ o3 family, majority solved
Reasoning · GSM8K Math
97%+ Frontier models, 2024–25
GUI Agents · Interface Sequences
24% Still early — large room to grow
Math Olympiad · AIME Competition
2022 — GPT-3 era
~2 / 15
problems solved
2025 — Frontier models
13–15 / 15
problems solved (90%+)
90%+ Near-perfect on high school competition math
Education · AI Adoption in K–12 Classrooms
60%
teachers use AI
29%
students use for math
21%
math teachers use AI for planning
50%
districts now train teachers on AI
Districts offering AI training doubled in one year
"It took speech recognition more than a decade to approach human-level performance. It took modern language models roughly 3–5 years to go from clumsy autocomplete to surpassing humans on broad exam suites. This isn't incremental improvement. This is benchmarks breaking under the speed of progress."

Everything Gets
10× Cheaper, Again

The cost of inference isn't following Moore's Law. It's lapping it.

Year: 2022
$20.00
per million tokens (GPT-4 class inference)
Drag to travel through time →
2022 2023 2024 2025+
Moore's Law
🤖
AI Inference
🚀
Peak Workloads
🏭
Steam Era
"Steam changed our muscles. Electricity changed the environments we occupied. The microchip changed our tools. AI is different: it is changing the thinking that runs everything else. For the first time, cognition itself is getting exponentially cheaper."

Same Intelligence,
Fraction of the Size

Instead of asking "how big can we build?" — what if the question is "how small can we go?"

THEN · 2022
540B parameters Data center required
PaLM-class model
540,000,000,000
Same capability.
Fraction of the size.
NOW · 2022
2022 baseline
540B Data center required
Equivalent performance
540,000,000,000
Drag the slider below. In 2022, matching expert-level performance on language benchmarks required a massive 540-billion-parameter model running in a data center. By 2025, a model 140× smaller achieves the same results — and fits on your phone.
Slide to see equivalent capability at different years →
2022 (540B) 2023 (70B) 2024 (7B) 2025 (3.8B)
Parameter Reduction
140×
Where It Runs
Phones
Capable AI assistants now run locally on consumer laptops and mobile devices — no data center needed.
Training Efficiency
50×
Faster improvement than Moore's Law in training costs, according to ARK Invest and Epoch AI analyses.
Historical Comparison
Decades → Years
Industrial engines improved gradually over generations. AI achieved orders-of-magnitude gains in 2–3 years.

The Fastest
Adoption Curve Ever

Electricity, smartphones, generative AI — three S-curves, each steeper than the last.

Electricity (1880–1930)
Smartphones (2007–2020)
Generative AI (2022–2026)
20–45%
2 hrs
Saved per day on repetitive tasks, reported by workers using AI tools
2 years
"Steam changed our muscles. Electricity changed the environments we occupied. The microchip changed our tools. AI is different: it is changing the thinking that runs everything else. Watch capability, cost, and adoption curves all bend — faster than any prior general-purpose technology documented in the economic or historical record."

Data Highlights
& Sources

The key numbers behind every claim in this experience — including a dedicated section on AI in mathematics and education — with direct links to the primary source for each.

Mathematics & Education
90%+
AI score on AIME 2025 — the elite high school math competition that qualifies students for the US Math Olympiad. Frontier models now solve 13–15 of 15 problems.
IntuitionLabs · AIME 2025 Benchmark Analysis
IMO Gold
AI achieved gold-medal standard at the 2025 International Mathematical Olympiad — problems specifically designed to require creative insight, not pattern matching.
IntuitionLabs · AI Reasoning at IMO 2025
<2%
AI success rate on FrontierMath — original, never-published problems created by 60+ mathematicians requiring hours to solve. This is the one wall AI hasn't crossed. Note: FrontierMath tests PhD research-level math — everything taught in K–12, including AP Calculus and competition prep, is well below this ceiling. For a high school math teacher, the more relevant number is the 90%+ AIME score above.
Epoch AI / ArXiv · FrontierMath Benchmark (2024)
60%
K-12 teachers used AI tools during the 2024–25 school year (Gallup/Walton). 32% use AI at least weekly. Preparing lessons is the #1 daily use at 20%.
YSU / Gallup–Walton Family Foundation Survey 2025
21%
Math teachers using AI for instructional planning — roughly half the rate of ELA and science teachers. Math teachers face unique challenges integrating AI into their practice.
Education Week / RAND Report · April 2025
Districts providing AI training to teachers doubled in one year — from ~25% in fall 2023 to ~50% in fall 2024. Low-poverty districts still lead higher-poverty ones.
RAND Corporation · American School District Panel 2025
29%
U.S. teens use ChatGPT specifically to solve math problems — the second most common student use after research (54%). Teen AI use for schoolwork doubled from 2023 to 2024.
Pew Research Center via YSU · 2024
Less bored
Students report being less bored in math class when teachers use ChatGPT to personalize lessons around student interests — with higher quality feedback and more relevant examples.
Wiley / School Science & Mathematics · April 2025
≈ 1-on-1
AI tutoring systems shown to match individualized human tutoring outcomes in K-12 math — identifying gaps, personalizing paths, and providing real-time feedback at scale.
Journal of Computer Assisted Learning · 2024
25%
Teachers who say AI tools do more harm than good in K-12 education. Only 6% say more benefit than harm. Math teachers cite concerns about errors, academic integrity, and loss of conceptual understanding.
Pew Research Center · May 2024