llm-leaderboard

Report Abuse

Basic Information

This repository is a community-maintained collection of structured data and benchmark results for large language models. It collects metadata for hundreds of LLMs including model sizes, context window limits, licensing, provider configurations and pricing, and measured performance such as throughput and latency. The project centralizes standardized benchmark scores and organizes all entries so contributors and consumers can compare models side by side. Data is stored under a clear directory layout with schemas for validation and a public contribution workflow. An interactive dashboard is provided to explore and visualize leaderboards and model comparisons. The repo is intended as a reference resource for anyone evaluating or documenting LLM capabilities and for community-driven updates to maintain accuracy.

Links

Categorization

App Details

Features
Comprehensive, structured data about hundreds of language models including parameters, context window sizes, license and organization details, provider-specific pricing and features, and performance metrics such as throughput and latency. Standardized benchmark definitions and per-model benchmark scores are included. A clear file layout places model metadata, providers, provider_models, benchmarks, model_benchmarks, organizations, and licenses under a data directory. JSON Schema files support validation of incoming contributions. The project includes contribution guidelines, a community review process for data quality, and an interactive leaderboard/dashboard for browsing and comparing models. Community-oriented tools and badges indicate openness to contributions and issue reporting.
Use Cases
The repo helps researchers, engineers, product managers and hobbyists compare LLMs by consolidating key technical and commercial attributes into a single, queryable dataset. It supports decision making about model selection by providing standardized benchmark results and provider pricing details so users can weigh performance versus cost. The organized data structure and validation schemas make it easier to reproduce or audit reported scores. Community contribution processes and source citation requirements increase transparency and help keep information current. The dashboard and leaderboard view speed up exploration and comparison without needing to aggregate disparate sources manually.

Please fill the required fields*