Live CompetitionEnds Apr 30, 2026

OpenAI Parameter Golf

Train a language model that fits in 16MB and trains in 10 minutes on 8×H100s. Lowest bits-per-byte wins. OpenAI put up $1M in compute credits. 1,500+ submissions. What's emerging is a masterclass in how to fit maximum knowledge into minimum parameters.

1.1147
Confirmed SOTA (BPB)
0.8265
Pending review (open PR)
1,500+
Submissions
1.2244
Baseline BPB

What's happening

Since launching on March 18, 2026, participants have compressed the naive baseline from 1.2244 to 1.1147 BPB — a 9% improvement — through innovations in three categories: quantization (ternary weights, GPTQ, int5/6 mixed precision), architecture (cross-sparse attention, depth recurrence, parallel residuals), and training tricks (test-time training, the Muon optimizer, sliding window evaluation).

Open PRs today claim scores as low as 0.83 BPB, suggesting the next wave of confirmed records will shatter existing marks. The techniques emerging here — fitting maximum knowledge into minimum parameters — are directly relevant to on-device AI and edge deployment.

Leaderboard (confirmed)

Top 10 as of April 10, 2026

#SubmissionBPB
1Self-Gen GPTQ + XSA-all1.1147
2LeakyReLU² + TTT + Muon1.1194
3EMA + GPTQ-lite1.1228
4Partial RoPE + LN Scale1.1248
5XSA4 + EMA + Int61.1271
6Efficient Partial XSA1.1307
7Int5-MLP + BigramHash1.1428
8SmearGate + BigramHash1.1458
9MLP3x + Int6 QAT1.1502
10Ternary U-Net 73.7M1.1570

Pending record claims

Open PRs with scores that would shatter the current SOTA if confirmed.

0.8265BPB

SLOT-24 + Pre-Quant AdamW TTT

ndokutovich

1.0600BPB

Recur345 + Par7 + Pre-Quant TTT

ndokutovich

1.0736BPB

Pre-quant TTT + Parallel Residuals

joshkmartinez

The rules

Artifact size
16 MB
Final compressed model must fit
Training budget
10 minutes
On 8×H100 GPUs, end-to-end
Metric
Bits-per-byte
On held-out text. Lower is better.
Prize pool
$1M credits
Awarded by OpenAI