Codesota · Benchmark · COCO 2017 Instance SegmentationHome/Leaderboards/COCO 2017 Instance Segmentation
Unknown

COCO 2017 Instance Segmentation.

The Microsoft COCO 2017 Instance Segmentation dataset (COCO 2017) is a large-scale benchmark for object detection and instance segmentation. It provides images with per-instance segmentation annotations (polygon masks and RLE), bounding boxes, and category labels for commonly occurring object classes (the standard COCO set of 80 detection/segmentation categories). The 2017 split commonly used for benchmarking includes train2017 and val2017 (HF mirrors list ~118,287 training images and 5,000 validation images) and test splits; annotations are provided in COCO JSON format. COCO was introduced in Lin et al., "Microsoft COCO: Common Objects in Context" (arXiv:1405.0312 / ECCV 2014) and is widely used for evaluating instance segmentation, object detection, and related tasks.

Paper Leaderboard
§ 01 · SOTA history

Year over year.

Not enough data to show trend.
§ 02 · Leaderboard

Results by metric.

Only 1 model on this benchmark
Help build the community leaderboard — submit your model results.

MAP

MAP is the reported evaluation metric for COCO 2017 Instance Segmentation. Codesota tracks published model scores on this metric so readers can compare state-of-the-art results across sources and model families.

Higher is better

Trust tiers for MAPverifiedpapervendorcommunityunverified
RankModelTrustScoreYearSource
01Segment Anything Model (SAM)
dataset: COCO 2017 Instance Segmentation; task: 3
paper46.5N/ASource ↗
§ 04 · Submit a result

Add to the leaderboard.

← Back to Leaderboards