Codesota · Computer Vision · Open-Vocabulary Object Detection · ODinW13 (subset of ODinW)Tasks/Computer Vision/Open-Vocabulary Object Detection
Open-Vocabulary Object Detection · benchmark dataset · EN

Object Detection in the Wild (ODinW) — subset: ODinW13.

Object Detection in the Wild (ODinW) is a benchmark/leaderboard (originating from the "Computer Vision in the Wild" community / EvalAI challenge) that aggregates multiple public object‑detection datasets to evaluate in-the-wild / zero-shot transfer performance of detectors. "ODinW13" refers to a specific subset of 13 datasets from the ODinW collection that is commonly reported as a single metric (average mAP across those 13 datasets) for measuring in-the-wild zero-shot detection. ODinW/ODinW13 is not a single stand-alone dataset with one canonical paper introducing it; instead it is a benchmark suite (used by many papers) and appears as an evaluation collection in numerous object-detection / open-vocabulary detection papers (for example: ScaleDet — arXiv:2306.04849 — which reports results on “13 datasets from Object Detection in the Wild (ODinW)”, and other open-vocabulary detection works that evaluate on ODinW). The ODinW benchmark is linked to the EvalAI challenge page for “Object Detection in the Wild” (Computer Vision in the Wild). Because ODinW13 is a reported subset/metric of that benchmark (not an independent dataset release), there is no single introducing arXiv paper or Hugging Face dataset page to link to; papers that use ODinW13 typically cite the benchmark or the CVinW/ELEVATER resources when reporting results.

Paper Submit a result
§ 01 · Leaderboard

Best published scores.

No results indexed yet — be the first to submit a score.

No benchmark results indexed yet
§ 06 · Contribute

Have a score that beats
this table?

Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.

Submit a result Read submission guide
What a submission needs
  • 01A public checkpoint or API endpoint
  • 02A reproduction script with frozen commit + seed
  • 03Declared evaluation environment (Python, deps)
  • 04One row per metric declared by this dataset
  • 05A contact so we can follow up on discrepancies