Browse and filter AI model evaluation results
Generate images based on data
Display a welcome message on a webpage
Display document size plots
Evaluate LLMs using Kazakh MC tasks
Create detailed data reports
More advanced and challenging multi-task evaluation
Multilingual metrics for the LMSys Arena Leaderboard
Simulate causal effects and determine variable control
Browse and submit evaluation results for AI benchmarks
Calculate VRAM requirements for running large language models
Analyze Shark Tank India episodes
Browse and compare Indic language LLMs on a leaderboard
UnlearnDiffAtk Benchmark is a data visualization tool designed to help users evaluate and compare AI models through the lens of differentiable attacks. It provides an interactive platform to browse and filter results of model evaluations, enabling researchers and developers to understand model vulnerabilities and performance more effectively.
What is the UnlearnDiffAtk Benchmark used for?
The UnlearnDiffAtk Benchmark is used to evaluate and compare AI models based on their robustness against differentiable attacks. It helps identify vulnerabilities and understand model performance under various scenarios.
What types of attacks are supported by UnlearnDiffAtk Benchmark?
The benchmark supports a wide range of differentiable attacks, including gradient-based and black-box attacks. For a full list, refer to the documentation.
How do I install UnlearnDiffAtk Benchmark?
Installation instructions are provided in the documentation. Typically, it involves running a single command to set up the tool and its dependencies.