Model card
GPT-4.1 mini.
OpenAIapitransformer
GPT-4.1 mini (2025-04-14), April 2025
§ 01 · Benchmarks
Every benchmark GPT-4.1 mini has a recorded score for.
| # | Benchmark | Area · Task | Metric | Value | Rank | Date | Source |
|---|---|---|---|---|---|---|---|
| 01 | HumanEval | Computer Code · Code Generation | pass@1 | 93.8% | #8 | — | source ↗ |
| 02 | HLE | Reasoning · Multi-step Reasoning | accuracy | 4.6% | #12 | — | source ↗ |
| 03 | GPQA | Reasoning · Multi-step Reasoning | accuracy | 66.4% | #21 | — | source ↗ |
| 04 | LiveCodeBench | Computer Code · Code Generation | pass@1 | 48.3% | #21 | — | source ↗ |
| 05 | SWE-bench Verified | Agentic AI · SWE-bench | resolve-rate | 23.6% | #79 | — | source ↗ |
Rank column shows this model’s position vs all other models scored on the same benchmark + metric (competitors after the slash). #1 in red means current SOTA. Sorted by rank, then newest result.
§ 02 · Strengths by area
Where GPT-4.1 mini actually performs.
§ 04 · Related models
Other OpenAI models scored on Codesota.
§ 05 · Sources & freshness
Where these numbers come from.
pricepertoken-leaderboard
3
results
openai-simple-evals
1
result
editorial
1
result
2 of 5 rows marked verified.