View NSQL Scores for Models
Explore and submit models using the LLM Leaderboard
Launch web-based model application
Export Hugging Face models to ONNX
Text-To-Speech (TTS) Evaluation using objective metrics.
Browse and evaluate ML tasks in MLIP Arena
Measure execution times of BERT models using WebGPU and WASM
Create and upload a Hugging Face model card
Evaluate model predictions with TruLens
Calculate survival probability based on passenger details
Browse and submit LLM evaluations
Calculate VRAM requirements for LLM models
Run benchmarks on prediction models
DuckDB NSQL Leaderboard is a tool designed for model benchmarking, allowing users to view and compare NSQL scores for different models. It provides a centralized platform to evaluate and visualize performance metrics, making it easier to analyze and optimize model effectiveness.
• Real-time Tracking:Monitor NSQL scores as they are updated in real-time.
• Interactive Visualization:Explore performance metrics through graphs and charts.
• Filter and Sort:Easily filter and sort models based on criteria like score, date, or model type.
• Side-by-Side Comparison:Compare multiple models directly to identify strengths and weaknesses.
• Data Export:Export leaderboard data for further analysis or reporting.
What does NSQL stand for?
NSQL stands for Natural Language Query Benchmark, a measure of how well a model understands and processes natural language.
How are NSQL scores calculated?
NSQL scores are calculated by evaluating a model's ability to process and answer natural language queries, often based on standardized benchmarks.
Can I export the leaderboard data for my own analysis?
Yes, the DuckDB NSQL Leaderboard provides options to export data in formats like CSV or Excel for further analysis.