Pergel: A Unified Benchmark for Evaluating Turkish LLMs
Convert PaddleOCR models to ONNX format
Browse and evaluate ML tasks in MLIP Arena
Create and upload a Hugging Face model card
Compare code model performance on benchmarks
Load AI models and prepare your space
Evaluate LLM over-refusal rates with OR-Bench
Convert Stable Diffusion checkpoint to Diffusers and open a PR
Predict customer churn based on input details
Run benchmarks on prediction models
Retrain models for new data at edge devices
Push a ML model to Hugging Face Hub
View NSQL Scores for Models
Cetvel is a unified benchmark tool designed for evaluating Turkish Large Language Models (LLMs). It provides a comprehensive framework to analyze and compare the performance of different models across a variety of tasks, making it an essential tool for researchers and developers working with Turkish NLP tasks.
• Comprehensive Task Coverage: Evaluate models on tasks such as translation, summarization, and question-answering specific to the Turkish language.
• Customizable Benchmarks: Create tailored benchmarking suites to focus on specific aspects of model performance.
• Cross-Model Comparisons: Compare multiple Turkish LLMs side-by-side to identify strengths and weaknesses.
• Detailed Reporting: Generate in-depth reports highlighting model accuracy, efficiency, and robustness.
• Integration with Popular LLMs: Supports integration with widely-used Turkish and multilingual LLMs.
What models are supported by Cetvel?
Cetvel supports a wide range of Turkish and multilingual LLMs, including but not limited to, models from leading NLP libraries.
Do I need NLP expertise to use Cetvel?
No, Cetvel is designed to be user-friendly. However, basic knowledge of NLP concepts may help in interpreting results.
Can I benchmark models in languages other than Turkish?
Cetvel is primarily optimized for Turkish, but it can be adapted for other languages with additional configuration.