Open Persian LLM Leaderboard
GIFT-Eval: A Benchmark for General Time Series Forecasting
View and submit language model evaluations
Track, rank and evaluate open LLMs and chatbots
Search for model performance across languages and benchmarks
Retrain models for new data at edge devices
Measure over-refusal in LLMs using OR-Bench
Display and submit language model evaluations
View RL Benchmark Reports
Convert Hugging Face model repo to Safetensors
Submit deepfake detection models for evaluation
Persian Text Embedding Benchmark
Benchmark LLMs in accuracy and translation across languages
The Open Persian LLM Leaderboard is a comprehensive benchmarking platform designed to evaluate and compare the performance of Persian language models. It provides a transparent and standardized framework for assessing models across various tasks, enabling researchers and developers to identify top-performing models for specific use cases. The leaderboard is continuously updated to reflect the latest advancements in the field of Persian natural language processing.
What models are included in the Open Persian LLM Leaderboard?
The leaderboard includes a wide range of Persian language models, from state-of-the-art research models to open-source community models. The list is regularly updated as new models are released.
How often are the models updated?
Models are typically updated on a quarterly basis, but the leaderboard may be refreshed more frequently to include cutting-edge research advancements.
Why isn’t a specific model appearing on the leaderboard?
A model may not appear if it has not been submitted for evaluation or if it does not meet the leaderboard’s inclusion criteria. Users are encouraged to submit models for consideration.