Evaluate Persian LLMs on various tasks
Submit models for evaluation and view leaderboard
Search for model performance across languages and benchmarks
Evaluate code generation with diverse feedback types
Track, rank and evaluate open LLMs and chatbots
Browse and filter machine learning models by category and modality
Predict customer churn based on input details
View RL Benchmark Reports
Evaluate reward models for math reasoning
Pergel: A Unified Benchmark for Evaluating Turkish LLMs
Optimize and train foundation models using IBM's FMS
Track, rank and evaluate open LLMs and chatbots
Browse and filter ML model leaderboard data
The š¤ Persian LLM Leaderboard is a comprehensive resource for evaluating and comparing Persian language models (LLMs) across various tasks and metrics. It provides a centralized platform for researchers, developers, and users to assess the performance of different models and make informed decisions based on their needs. The leaderboard is designed to promote transparency and innovation in the field of Persian natural language processing.
⢠Model Performance Tracking: Detailed performance metrics for various Persian LLMs on tasks like text classification, summarization, and question answering.
⢠Task-Specific Benchmarking: Evaluation across a wide range of NLP tasks tailored to the Persian language.
⢠Comparative Analysis: Side-by-side comparison of models to identify strengths and weaknesses.
⢠Regular Updates: Continuous updates with new models, tasks, and metrics.
⢠Open Accessibility: Available to everyone, including researchers, developers, and enthusiasts.
⢠Documentation and Resources: Access to datasets, evaluation scripts, and best practices for benchmarking.
What models are included in the leaderboard?
The leaderboard includes a variety of Persian language models, ranging from smaller, efficient models to larger, state-of-the-art architectures. Models are added continuously as they are developed and benchmarked.
How are models rated or ranked?
Models are ranked based on their performance on specific tasks and metrics. The ranking is determined by evaluation results on standardized datasets and may vary depending on the task or metric being considered.
How often is the leaderboard updated?
The leaderboard is updated regularly to include new models, tasks, and metrics. Updates are typically announced on the official platform or through associated communication channels.