Submit deepfake detection models for evaluation
Evaluate code generation with diverse feedback types
Track, rank and evaluate open LLMs and chatbots
Optimize and train foundation models using IBM's FMS
Visualize model performance on function calling tasks
Find and download models from Hugging Face
Calculate survival probability based on passenger details
Measure BERT model performance using WASM and WebGPU
View and compare language model evaluations
Create and manage ML pipelines with ZenML Dashboard
Display leaderboard of language model evaluations
View RL Benchmark Reports
Compare model weights and visualize differences
The Deepfake Detection Arena Leaderboard is a platform designed for evaluating and comparing deepfake detection models. It provides a standardized environment where researchers and developers can submit their models for benchmarking against state-of-the-art algorithms. The leaderboard categorizes submissions under Model Benchmarking and focuses on identifying deepfake detection capabilities.
What models are eligible for submission?
Only deepfake detection models are eligible for submission. Ensure your model adheres to the platform's guidelines.
How are models evaluated on the leaderboard?
Models are evaluated based on accuracy, precision, and recall when detecting deepfake content.
Can I share my model's results publicly?
Yes, the platform allows users to share their model's performance metrics and insights with the community.