Generative Tasks Evaluation of Arabic LLMs
Parse and highlight entities in an email thread
Use title and abstract to predict future academic impact
Find the best matching text for a query
A benchmark for open-source multi-dialect Arabic ASR models
Calculate love compatibility using names
Load documents and answer questions from them
Analyze text using tuned lens and visualize predictions
Deduplicate HuggingFace datasets in seconds
Demo emotion detection
Detect harms and risks with Granite Guardian 3.1 8B
Generate answers by querying text in uploaded documents
Generate insights and visuals from text
AraGen Leaderboard is a platform designed for evaluating and comparing the performance of Arabic language models (LLMs) on generative tasks. It serves as a benchmarking tool to assess the capabilities of different models in understanding and generating Arabic text. AraGen Leaderboard is free to use and is particularly useful for researchers, developers, and users interested in Arabic NLP.
What is AraGen Leaderboard used for?
AraGen Leaderboard is used to evaluate and compare the performance of Arabic language models on generative tasks, helping users identify the best models for their needs.
How are models evaluated on AraGen Leaderboard?
Models are evaluated based on predefined metrics and custom tasks set by the user, ensuring a comprehensive assessment of their capabilities.
Can I use AraGen Leaderboard for languages other than Arabic?
No, AraGen Leaderboard is specifically designed for evaluating Arabic language models and does not support other languages.