Physics-based continuous control benchmark. Evaluated on 15 DMControl tasks; metric is mean normalized score (0=random, 1000=expert) at 1M environment steps.
9 results indexed across 1 metric. Shaded row marks current SOTA; ties broken by submission date.
| # | Model | Org | Submitted | Paper / code | average-return |
|---|---|---|---|---|---|
| 01 | TD-MPC2 (317M params)OSS | UC San Diego | Mar 2026 | iclr-2024-paper | 960 |
| 02 | TD-MPC2 (19M params)OSS | UC San Diego | Mar 2026 | iclr-2024-paper | 953 |
| 03 | FOWMOSS | CMU | Mar 2026 | neurips-2024-paper | 945 |
| 04 | BROOSS | DeepMind / TU Warsaw | Mar 2026 | icml-2024-paper | 941 |
| 05 | TD-MPC2 (5M params)OSS | UC San Diego | Mar 2026 | iclr-2024-paper | 929 |
| 06 | DreamerV3OSS | Google DeepMind | Mar 2026 | arxiv-paper | 897 |
| 07 | TD-MPCOSS | UC San Diego | Mar 2026 | icml-2022-paper | 857 |
| 08 | DrQ-v2OSS | NYU / Google | Mar 2026 | icml-2022-paper | 799 |
| 09 | SAC (state-based)OSS | UC Berkeley | Mar 2026 | icml-2018-paper | 777 |
Each row below marks a model that broke the previous record on average-return. Intermediate submissions are kept in the leaderboard above; only SOTA-setting entries are re-listed here.
Higher scores win. Each subsequent entry improved upon the previous best.
Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.