Codesota · General · Coding Agents · MultiPL-ETasks/General/Coding Agents
Coding Agents · benchmark dataset · EN

MultiPL-E: A Scalable and Extensible Approach to Benchmarking Neural Code Generation.

MultiPL-E is a multi-programming-language benchmark for evaluating natural-language-to-code generation by large language models. It translates unit-test-driven Python benchmarks (OpenAI HumanEval and MBPP) into parallel problems in multiple programming languages, preserving prompts and test harnesses so models can be evaluated via execution-based metrics. The released dataset provides per-language configurations (e.g., humaneval-<lang>, mbpp-<lang>) containing prompts, tests, doctests, stop tokens and related metadata; the original project translated the Python benchmarks into 18 languages (and Hugging Face distributions expose many language-specific configs). Source code and dataset tooling are available from the NuPRL project (GitHub) and the authors published a paper describing the benchmark and methodology (arXiv:2208.08227 / IEEE TSE publication). License: MIT.

Paper Submit a result
§ 01 · Leaderboard

Best published scores.

1 result indexed across 1 metric. Shaded row marks current SOTA; ties broken by submission date.


Primary
Pass@1 · higher is better
Pass@1· primary
1 row
#ModelOrgSubmittedPaper / codePass@1
01Qwen2.5-PlusDec 2024Qwen2.5 Technical Report · code77
Fig 2 · Rows sorted by score within each metric. Shaded row marks SOTA. Dates reflect model or paper release where available, otherwise the date Codesota accessed the source.
§ 03 · Progress

1 steps
of state of the art.

Each row below marks a model that broke the previous record on Pass@1. Intermediate submissions are kept in the leaderboard above; only SOTA-setting entries are re-listed here.

Higher scores win. Each subsequent entry improved upon the previous best.

SOTA line · Pass@1
  1. Dec 19, 2024Qwen2.5-Plus77
Fig 3 · SOTA-setting models only. 1 entries span Dec 2024 Dec 2024.
§ 04 · Literature

1 paper
tied to this benchmark.

Every paper below corresponds to at least one row in the leaderboard above. Click through for the arXiv preprint and, when available, the reference implementation.

  • Qwen2.5 Technical Report
    Qwen:An YangBaosong YangBeichen ZhangBinyuan HuiBo ZhengBowen YuChengyuan LiDayiheng LiuFei HuangHaoran WeiHuan LinJian YangJianhong TuJianwei ZhangJianxin YangJiaxi YangJingren ZhouJunyang LinKai DangKeming LuKeqin BaoKexin YangLe YuMei LiMingfeng XuePei ZhangQin ZhuRui MenRunji LinTianHao LiTianyi TangTingyu XiaXingzhang RenXuancheng RenYang FanYang SuYichang ZhangYu WanYuqiong LiuZeyu CuiZhenru ZhangZihan Qiu
    Dec 2024·Qwen2.5-Plus
§ 06 · Contribute

Have a score that beats
this table?

Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.

Submit a result Read submission guide
What a submission needs
  • 01A public checkpoint or API endpoint
  • 02A reproduction script with frozen commit + seed
  • 03Declared evaluation environment (Python, deps)
  • 04One row per metric declared by this dataset
  • 05A contact so we can follow up on discrepancies