View NSQL Scores for Models
Generate leaderboard comparing DNA models
View and submit LLM benchmark evaluations
Visualize model performance on function calling tasks
Calculate VRAM requirements for LLM models
Explain GPU usage for model training
Browse and submit model evaluations in LLM benchmarks
Convert PyTorch models to waifu2x-ios format
Create demo spaces for models on Hugging Face
Browse and filter ML model leaderboard data
Evaluate RAG systems with visual analytics
Display and submit LLM benchmarks
View and submit LLM evaluations
DuckDB NSQL Leaderboard is a tool designed for model benchmarking, allowing users to view and compare NSQL scores for different models. It provides a centralized platform to evaluate and visualize performance metrics, making it easier to analyze and optimize model effectiveness.
• Real-time Tracking:Monitor NSQL scores as they are updated in real-time.
• Interactive Visualization:Explore performance metrics through graphs and charts.
• Filter and Sort:Easily filter and sort models based on criteria like score, date, or model type.
• Side-by-Side Comparison:Compare multiple models directly to identify strengths and weaknesses.
• Data Export:Export leaderboard data for further analysis or reporting.
What does NSQL stand for?
NSQL stands for Natural Language Query Benchmark, a measure of how well a model understands and processes natural language.
How are NSQL scores calculated?
NSQL scores are calculated by evaluating a model's ability to process and answer natural language queries, often based on standardized benchmarks.
Can I export the leaderboard data for my own analysis?
Yes, the DuckDB NSQL Leaderboard provides options to export data in formats like CSV or Excel for further analysis.