In July 2025 Meta retired Papers with Code. A seven-year archive of leaderboards, submissions and citations went offline without notice, and the field was briefly reminded that benchmark infrastructure, however load-bearing it felt, had been sitting on a single corporate hosting decision the whole time. Codesota is the attempt to rebuild that infrastructure in a way that cannot be retired by a press release.
It is, plainly, a website and a JSON file. Each benchmark is a task-dataset-metric triple with a declared direction, a fixed split, a reproducibility package and a dated submission. Each row carries a verification tier — self-reported, community-reproduced, or Codesota-reproduced — and each score is stamped with the day it was run. The full standard is on the methodology page; this page is the editorial note behind it.
The philosophical commitment is modest: a registry, not a leaderboard. A leaderboard is a view of “who is on top right now”; a registry is the record that makes that view legible. When a model regresses between checkpoints, the preceding score stays visible so the regression itself is visible. When a score turns out to be wrong, the correction is visible too. Nothing is silently deleted.
The practical commitment is equally modest: no payment for listings, no sponsored rankings, no investor behind the curtain. If paid work ever exists — consulting, a commissioned evaluation — it is disclosed inline on the page where the result appears. The public registry is not for sale.
Codesota does not need to be large to be useful. It needs to be honest, dated, and available to anyone who wants the JSON.