General Language Understanding Evaluation for masked language models
3 results indexed across 1 metric. Shaded row marks current SOTA; ties broken by submission date.
| # | Model | Org | Submitted | Paper / code | avg-score |
|---|---|---|---|---|---|
| 01 | DeBERTa-v3-largeOSS | Microsoft | Jan 2023 | DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Tra… | 91.37 |
| 02 | ALBERT-xxlarge-v2OSS | Feb 2020 | ALBERT: A Lite BERT for Self-supervised Learning of Lang… | 89.40 | |
| 03 | RoBERTa-largeOSS | Facebook AI | Jul 2019 | RoBERTa: A Robustly Optimized BERT Pretraining Approach | 88.50 |
Every paper below corresponds to at least one row in the leaderboard above. Click through for the arXiv preprint and, when available, the reference implementation.
Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.