voyage-3-large
Voyage AI has unveiled voyage-3-large, a cutting-edge general-purpose and multilingual embedding model that leads across eight evaluated domains, including law, finance, and code, outperforming OpenAI-v3-large and Cohere-v3-English by averages of 9.74% and 20.71%, respectively. Enabled by Matryoshka learning and quantization-aware training, it supports embeddings of 2048, 1024, 512, and 256 dimensions, along with multiple quantization options such as 32-bit floating point, signed and unsigned 8-bit integer, and binary precision, significantly reducing vector database costs with minimal impact on retrieval quality. Notably, voyage-3-large offers a 32K-token context length, surpassing OpenAI's 8K and Cohere's 512 tokens. Evaluations across 100 datasets in diverse domains demonstrate its superior performance, with flexible precision and dimensionality options enabling substantial storage savings without compromising quality.
Learn more
Codestral Embed
Codestral Embed is Mistral AI's first embedding model, specialized for code, optimized for high-performance code retrieval and semantic understanding. It significantly outperforms leading code embedders in the market today, such as Voyage Code 3, Cohere Embed v4.0, and OpenAI’s large embedding model. Codestral Embed can output embeddings with different dimensions and precisions; for instance, with a dimension of 256 and int8 precision, it still performs better than any model from competitors. The dimensions of the embeddings are ordered by relevance, allowing users to choose the first n dimensions for a smooth trade-off between quality and cost. It excels in retrieval use cases on real-world code data, particularly in benchmarks like SWE-Bench, which is based on real-world GitHub issues and corresponding fixes, and Text2Code (GitHub), relevant for providing context for code completion or editing.
Learn more
Mixedbread
Mixedbread is a fully-managed AI search engine that allows users to build production-ready AI search and Retrieval-Augmented Generation (RAG) applications. It offers a complete AI search stack, including vector stores, embedding and reranking models, and document parsing. Users can transform raw data into intelligent search experiences that power AI agents, chatbots, and knowledge systems without the complexity. It integrates with tools like Google Drive, SharePoint, Notion, and Slack. Its vector stores enable users to build production search engines in minutes, supporting over 100 languages. Mixedbread's embedding and reranking models have achieved over 50 million downloads and outperform OpenAI in semantic search and RAG tasks while remaining open-source and cost-effective. The document parser extracts text, tables, and layouts from PDFs, images, and complex documents, providing clean, AI-ready content without manual preprocessing.
Learn more
Azure AI Search
Deliver high-quality responses with a vector database built for advanced retrieval augmented generation (RAG) and modern search. Focus on exponential growth with an enterprise-ready vector database that comes with security, compliance, and responsible AI practices built in. Build better applications with sophisticated retrieval strategies backed by decades of research and customer validation. Quickly deploy your generative AI app with seamless platform and data integrations for data sources, AI models, and frameworks. Automatically upload data from a wide range of supported Azure and third-party sources. Streamline vector data processing with built-in extraction, chunking, enrichment, and vectorization, all in one flow. Support for multivector, hybrid, multilingual, and metadata filtering. Move beyond vector-only search with keyword match scoring, reranking, geospatial search, and autocomplete.
Learn more