Display benchmark results
Display LLM benchmark leaderboard and info
View RL Benchmark Reports
Calculate VRAM requirements for LLM models
Quantize a model for faster inference
View and submit LLM benchmark evaluations
Explore and manage STM32 ML models with the STM32AI Model Zoo dashboard
Evaluate LLM over-refusal rates with OR-Bench
Calculate survival probability based on passenger details
Browse and filter ML model leaderboard data
Generate and view leaderboard for LLM evaluations
View and submit LLM evaluations
Submit models for evaluation and view leaderboard
Redteaming Resistance Leaderboard is a model benchmarking tool designed to evaluate and compare the performance of AI models in resisting adversarial attacks. It provides a comprehensive platform to display benchmark results, enabling researchers and developers to assess the robustness of their models against various threat scenarios. The leaderboard serves as a centralized resource for identifying top-performing models and tracking progress in adversarial defense.
What models are included in the leaderboard?
The leaderboard features a diverse range of AI models, including state-of-the-art architectures designed for adversarial defense.
How often are the results updated?
Results are updated in real-time to ensure the latest advancements in model resistance are reflected.
Can I contribute my own model to the leaderboard?
Yes, submissions are welcome. Please refer to the platform's documentation for guidelines on model submission and evaluation criteria.