Natural Language Processingtoken-classification

Named Entity Recognition

Named entity recognition (NER) extracts structured mentions — people, organizations, locations, dates — from unstructured text, making it foundational to knowledge graphs, financial compliance, and clinical NLP. CoNLL-2003 English F1 scores have been above 93% since BERT, and current leaders like UniNER and GLiNER push past 95%, but these numbers mask the real difficulty: nested entities, emerging entity types, and cross-lingual transfer where performance drops 10-20 points. The shift from sequence labeling to generative NER (framing extraction as text generation) has opened the door for LLMs to compete, though latency-sensitive production systems still rely on encoder models like DeBERTa-v3 and SpanBERT.

1
Datasets
0
Results
f1
Canonical metric
Canonical Benchmark

CoNLL-2003

Reuters news stories annotated with 4 entity types: PER, ORG, LOC, MISC. The standard NER benchmark.

Primary metric: f1
View full leaderboard

Top 10

Leading models on CoNLL-2003.

No results yet. Be the first to contribute.

What were you looking for on Named Entity Recognition?

Didn't find the model, metric, or dataset you needed? Tell us in one line. We read every message and reply within 48 hours.

All datasets

1 dataset tracked for this task.

Related tasks

Other tasks in Natural Language Processing.

Reply within 48 hours · No newsletter

Didn't find what you came for?

Still looking for something on Named Entity Recognition? A missing model, a stale score, a benchmark we should cover — drop it here and we'll handle it.

Real humans read every message. We track what people are asking for and prioritize accordingly.