React Native Code Generation

Evaluating AI models on generating correct, production-quality React Native implementations. Covers animation, navigation, state management, lists, and platform APIs using real-world libraries (Reanimated, React Navigation, Zustand, FlashList).

1
Datasets
0
Results
requirement-satisfaction
Canonical metric
Canonical Benchmark

React Native Evals

A benchmark suite evaluating how AI coding models handle authentic React Native development tasks. 71 evals across 5 categories: animation (14), async-state management (14), lists (19), navigation (14), and React Native APIs (10). Each eval specifies explicit, judgeable requirements. Model outputs are scored on requirement satisfaction using LLM-based judging. Covers real libraries: Reanimated, React Navigation, Zustand, Jotai, React Query, FlatList, FlashList, LegendList.

Primary metric: requirement-satisfaction
View full leaderboard

Top 10

Leading models on React Native Evals.

No results yet. Be the first to contribute.

All datasets

1 dataset tracked for this task.