Browse and compare language model leaderboards
Generate image descriptions
Turn your image and question into answers
Watch a video exploring AI, ethics, and Henrietta Lacks
Display a logo with a loading spinner
Convert screenshots to HTML code
Monitor floods in West Bengal in real-time
Ask questions about images
a tiny vision language model
Transcribe manga chapters with character names
finetuned florence2 model on VQA V2 dataset
PaliGemma2 LoRA finetuned on VQAv2
Rerun viewer with Gradio
Clembench is a tool designed to help users browse and compare language model leaderboards. It provides a platform for evaluating and analyzing the performance of different language models, particularly in the domain of Visual QA (Question Answering). Clembench enables users to explore benchmark results, compare model performance, and gain insights into the capabilities of various language models.
• Interactive Dashboard: Access a user-friendly interface to explore benchmark results. • Model Comparison: Compare performance metrics across multiple language models. • Real-Time Filtering: Narrow down results by metrics, datasets, or models. • Detailed Analytics: Dive into in-depth performance statistics for each model. • Benchmarking: Test and evaluate language models against standard benchmarks.
What types of models are supported on Clembench?
Clembench supports a wide range of language models, including but not limited to SOTA (State-of-the-Art) models in Visual QA.
How often are the leaderboards updated?
The leaderboards are regularly updated to reflect the latest advancements in language model research.
Can I export the comparison data?
Yes, Clembench allows users to export data and visualizations for further analysis or reporting.