Karpathy Demonstrates Massive GPT-2 Training Cost Reduction
Andrej Karpathy reports that training GPT-2 originally in 2019 used 32 TPU v3 chips for 168 hours (~$43,000) to reach a 0.256525 CORE score. He says recent improvements merged into nanochat (from modded-nanogpt) now achieve a higher CORE score in 3.04 hours (~$73) on a single 8x H100 node, representing about a 600× cost reduction and an estimated 2.5× annual decline in cost over seven years.
Scoring Rationale
High practical impact and authoritative source, limited by single-source claim and brief, non-reproducible technical detail.
Practice with real Logistics & Shipping data
90 SQL & Python problems · 15 industry datasets
250 free problems · No credit card
See all Logistics & Shipping problemsStep-by-step roadmaps from zero to job-ready — curated courses, salary data, and the exact learning order that gets you hired.
Sources
- Read OriginalA quote from Andrej Karpathysimonwillison.net



