Agentic methods and enterprise search depend upon sturdy knowledge retrieval that works effectively and precisely. Database supplier MongoDB thinks its latest embeddings fashions assist resolve falling retrieval high quality as extra AI methods go into manufacturing.
As agentic and RAG methods transfer into manufacturing, retrieval high quality is rising as a quiet failure level — one that may undermine accuracy, price, and person belief even when fashions themselves carry out effectively.
The corporate launched 4 new variations of its embeddings and reranking fashions. Voyage 4 will probably be accessible in 4 modes: voyage-4 embedding, voyage-4-large, voyage-4-lite, and voyage-4-nano.
MongoDB mentioned the voyage-4 embedding serves as its general-purpose mannequin; MongoDB considers Voyage-4-large its flagship mannequin. Voyage-4-lite focuses on duties requiring little latency and decrease prices, and voyage-4-nano is meant for extra native improvement and testing environments or for on-device knowledge retrieval.
Voyage-4-nano can also be MongoDB’s first open-weight mannequin. All fashions can be found through an API and on MongoDB’s Atlas platform.
The corporate mentioned the fashions outperform comparable fashions from Google and Cohere on the RTEB benchmark. Hugging Face’s RTEB benchmark places Voyage 4 as the highest embedding mannequin.
“Embedding fashions are a type of invisible decisions that may actually make or break AI experiences,” Frank Liu, product supervisor at MongoDB, mentioned in a briefing. “You get them incorrect, your search outcomes will really feel fairly random and shallow, however when you get them proper, your utility immediately feels prefer it understands your customers and your knowledge.”
He added that the aim of the Voyage 4 fashions is to enhance the retrieval of real-world knowledge, which regularly collapses as soon as agentic and RAG pipelines go into manufacturing.
MongoDB additionally launched a brand new multimodal embedding mannequin, voyage-multimodal-3.5, that may deal with paperwork that embrace textual content, pictures, and video. This mannequin vectorizes the info and extracts semantic that means from the tables, graphics, figures, and slides sometimes present in enterprise paperwork.
Enterprise’s embeddings issues
For enterprises, an agentic system is just nearly as good as its potential to reliably retrieve the precise data on the proper time. This requirement turns into tougher as workloads scale and context home windows fragment.
A number of mannequin suppliers goal that layer of agentic AI. Google’s Gemini Embedding mannequin topped the embedding leaderboards, and Cohere launched its Embed 4 multimodal mannequin, which processes paperwork greater than 200 pages lengthy. Mistral mentioned its coding-embedding mannequin, Codestral Embedding, outperforms Cohere, Google, and even MongoDB’s Voyage Code 3. MongoDB argues that benchmark efficiency alone doesn’t handle the operational complexity enterprises face in manufacturing.
MongoDB mentioned many purchasers have discovered that their knowledge stacks can’t deal with context-aware, retrieval-intensive workloads in manufacturing. The corporate mentioned it's seeing extra fragmentation with enterprises having to sew collectively completely different options to attach databases with a retrieval or reranking mannequin. To assist clients who don’t need fragmented options, the corporate is providing its fashions via a single knowledge platform, Atlas.
MongoDB’s wager is that retrieval can’t be handled as a unfastened assortment of best-of-breed elements anymore. For enterprise brokers to work reliably at scale, embeddings, reranking, and the info layer have to function as a tightly built-in system somewhat than a stitched-together stack.

