Single 7B model unifying generation and embedding via Generative Representational Instruction Tuning.
A workable 7.2B-parameter dense embedding model from GritLM (Contextual AI). Treat the modality benchmarks above as the leading indicator of fit — composite scoring across modalities is still maturing.
Generated from this model’s benchmarks and ranking signals. Editor reviews refine it over time.
Access model weights, configuration files, and documentation.
See which devices can run this model and at what quality level.
GritLM-7B unifies text generation and representation in one Mistral-7B-based model via Generative Representational Instruction Tuning (GRIT), using bidirectional attention for embeddings and causal attention for generation, with task selected by the instruction. Set a new MTEB SOTA at release while matching instruction-tuned generative performance, and notably accelerates RAG by >60% on long documents because the same model serves both retrieval and generation.