Codesota · General · Vision-Language Models · WISETasks/General/Vision-Language Models
Vision-Language Models · benchmark dataset · EN

WISE: A World Knowledge-Informed Semantic Evaluation.

WISE (World Knowledge-Informed Semantic Evaluation) is a benchmark and dataset for evaluating text-to-image (T2I) models on their ability to integrate world knowledge and complex semantic understanding into generated images. The benchmark contains 1,000 carefully crafted prompts organized across 25 sub-domains spanning cultural common sense, spatio-temporal reasoning, and natural science. The project introduces WiScore, a quantitative metric designed to assess knowledge–image alignment beyond traditional CLIP-based metrics. The repository includes prompt JSON files (structured prompts and explanations), evaluation code and scripts, example assets, and instructions to compute WiScore and run evaluations. Code and data are hosted in the public GitHub repository (https://github.com/PKU-YuanGroup/WISE). The accompanying paper is available at arXiv:2503.07265. (Also cited as [22] in the referencing paper.)

Paper Submit a result
§ 01 · Leaderboard

Best published scores.

No results indexed yet — be the first to submit a score.

No benchmark results indexed yet
§ 06 · Contribute

Have a score that beats
this table?

Submit a checkpoint and a reproduction script. We will run it, publish the score, and — if it takes the top — annotate the step on the progress chart with your name.

Submit a result Read submission guide
What a submission needs
  • 01A public checkpoint or API endpoint
  • 02A reproduction script with frozen commit + seed
  • 03Declared evaluation environment (Python, deps)
  • 04One row per metric declared by this dataset
  • 05A contact so we can follow up on discrepancies