Mixedbread

Embeddings

What are Embeddings?

Embeddings are numerical representations of data (words, phrases, images, etc.) mapped into continuous vector spaces. These vectors capture semantic and contextual relationships, enabling machine learning models to interpret the meaning and similarity between data points mathematically.

Generate your first embeddings

Try out the example to understand how to generate embeddings with our models.

Available Embedding Models

Choose the right model for your use case:

Model CardDescription
State-of-the-art English embedding model for semantic search and retrieval. Trained on 700M+ pairs, supports binary embeddings for fast, efficient storage.
German/English model fine-tuned on 30M+ German pairs. Supports binary quantization and Matryoshka learning for cost-effective, real-world use.
First 2D-Matryoshka embedding model for flexible speed, storage, and performance trade-offs. Remains competitive even at reduced sizes.

Last updated: July 14, 2025