Multilingual Text Embedding Model Pruner
Teach, test, evaluate language models with MTEB Arena
View and submit LLM evaluations
Compare and rank LLMs using benchmark scores
GIFT-Eval: A Benchmark for General Time Series Forecasting
Measure BERT model performance using WASM and WebGPU
Open Persian LLM Leaderboard
Display leaderboard of language model evaluations
Evaluate open LLMs in the languages of LATAM and Spain.
Compare LLM performance across benchmarks
Submit models for evaluation and view leaderboard
View and compare language model evaluations
Evaluate adversarial robustness using generative models
MTEM Pruner is a Multilingual Text Embedding Model Pruner designed to simplify and optimize multilingual text embedding models. This tool allows users to prune multilingual models to focus on a single target language, making the model more efficient and specialized for specific use cases. By reducing the complexity of multilingual models, MTEM Pruner helps in improving inference speed, memory usage, and overall performance for monolingual applications.
pip install mtem-pruner
to install the MTEM Pruner package.model = load_multilingual_model("xlm-roberta-base")
pruned_model = mtem_pruner.prune(model, target_lang="en")
What models are supported by MTEM Pruner?
MTEM Pruner supports popular multilingual models such as Multilingual BERT, XLM-RoBERTa, and DistilMultilingualBERT. Support for additional models is continuously being added.
Does pruning affect the model's accuracy?
While pruning reduces the model size and complexity, it is designed to retain the most important features for the target language. In many cases, the accuracy for the specific language remains comparable or even improves due to the focus on relevant parameters.
Can I prune a model to support multiple languages?
MTEM Pruner is specifically designed for single-language pruning. However, you can run the pruning process multiple times for different languages if you need models for various languages.