Benchmark LLMs in accuracy and translation across languages
Rank machines based on LLaMA 7B v2 benchmark results
Convert Stable Diffusion checkpoint to Diffusers and open a PR
Display model benchmark results
Teach, test, evaluate language models with MTEB Arena
View and submit LLM evaluations
Browse and evaluate ML tasks in MLIP Arena
Convert Hugging Face model repo to Safetensors
Submit models for evaluation and view leaderboard
Upload ML model to Hugging Face Hub
Compare audio representation models using benchmark results
Compare and rank LLMs using benchmark scores
Search for model performance across languages and benchmarks
The European Leaderboard is a benchmarking tool designed to evaluate and compare large language models (LLMs) in terms of accuracy and translation capabilities across multiple languages. It provides a comprehensive platform to assess model performance, enabling users to identify top-performing models for specific tasks and languages.
What is the main purpose of the European Leaderboard?
The primary purpose is to provide a standardized way to benchmark and compare LLMs across various European languages and tasks.
Which languages are supported by the European Leaderboard?
The tool supports a wide range of European languages, including English, French, German, Spanish, Italian, and many others. The exact list is updated regularly.
Can I benchmark my own model using European Leaderboard?
Yes, the platform allows users to submit and benchmark their own models, provided they meet the specified requirements.