Popular ~100-language instruction-tuned embedding model built on XLM-RoBERTa-large.
A strong 0.56B-parameter dense embedding model from intfloat (Microsoft Research). Treat the modality benchmarks above as the leading indicator of fit — composite scoring across modalities is still maturing.
Generated from this model’s benchmarks and ranking signals. Editor reviews refine it over time.
Access model weights, configuration files, and documentation.
See which devices can run this model and at what quality level.
A 560M-parameter multilingual embedding model from Microsoft Research's intfloat, initialized from XLM-RoBERTa-large and trained in two stages, weakly-supervised contrastive pre-training on 1B pairs followed by instruction-tuned fine-tuning using E5-Mistral synthetic data. Supports ~100 languages and uses natural-language task instructions to customize query embeddings.