Browse and filter AI model evaluation results
More advanced and challenging multi-task evaluation
Migrate datasets from GitHub or Kaggle to Hugging Face Hub
This project is a GUI for the gpustack/gguf-parser-go
Browse and explore datasets from Hugging Face
A Leaderboard that demonstrates LMM reasoning capabilities
Explore and submit NER models
Explore and compare LLM models through interactive leaderboards and submissions
View and compare pass@k metrics for AI models
Transfer GitHub repositories to Hugging Face Spaces
Cluster data points using KMeans
Multilingual metrics for the LMSys Arena Leaderboard
Search for tagged characters in Animagine datasets
UnlearnDiffAtk Benchmark is a data visualization tool designed to help users evaluate and compare AI models through the lens of differentiable attacks. It provides an interactive platform to browse and filter results of model evaluations, enabling researchers and developers to understand model vulnerabilities and performance more effectively.
What is the UnlearnDiffAtk Benchmark used for?
The UnlearnDiffAtk Benchmark is used to evaluate and compare AI models based on their robustness against differentiable attacks. It helps identify vulnerabilities and understand model performance under various scenarios.
What types of attacks are supported by UnlearnDiffAtk Benchmark?
The benchmark supports a wide range of differentiable attacks, including gradient-based and black-box attacks. For a full list, refer to the documentation.
How do I install UnlearnDiffAtk Benchmark?
Installation instructions are provided in the documentation. Typically, it involves running a single command to set up the tool and its dependencies.