Benchmark LLMs in accuracy and translation across languages
Display benchmark results
GIFT-Eval: A Benchmark for General Time Series Forecasting
Upload ML model to Hugging Face Hub
View and submit machine learning model evaluations
Evaluate AI-generated results for accuracy
Browse and evaluate language models
Evaluate open LLMs in the languages of LATAM and Spain.
Explore and visualize diverse models
Find and download models from Hugging Face
Text-To-Speech (TTS) Evaluation using objective metrics.
Leaderboard of information retrieval models in French
Display and filter leaderboard models
The European Leaderboard is a benchmarking tool designed to evaluate and compare large language models (LLMs) in terms of accuracy and translation capabilities across multiple languages. It provides a comprehensive platform to assess model performance, enabling users to identify top-performing models for specific tasks and languages.
What is the main purpose of the European Leaderboard?
The primary purpose is to provide a standardized way to benchmark and compare LLMs across various European languages and tasks.
Which languages are supported by the European Leaderboard?
The tool supports a wide range of European languages, including English, French, German, Spanish, Italian, and many others. The exact list is updated regularly.
Can I benchmark my own model using European Leaderboard?
Yes, the platform allows users to submit and benchmark their own models, provided they meet the specified requirements.