LVIS is a large-scale, high-quality dataset for instance segmentation containing 160k-164k images and 2M instance annotations for over 1000 object categories. It focuses on long-tail object recognition, providing a larger and more detailed vocabulary than COCO. LVIS uses the same images as the COCO dataset but with different splits and annotations optimized for instance segmentation. The dataset includes common and rare object categories and provides standardized evaluation metrics like mean Average Precision (mAP) for instance segmentation.
MAP is the reported evaluation metric for LVIS (Instance Segmentation). Codesota tracks published model scores on this metric so readers can compare state-of-the-art results across sources and model families.
Higher is better
| Rank | Model | Trust | Score | Year | Source |
|---|---|---|---|---|---|
| 01 | Segment Anything Model (SAM) | paper | 44.7 | N/A | Source ↗ |