Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

Deepmind finds RAG limit with fixed-size embeddings

DeepMind shows classical BM25 still outperforms dense embeddings at scale.

byKerem Gülen
September 5, 2025
in Artificial Intelligence, Research
Home News Artificial Intelligence
Share on FacebookShare on TwitterShare on LinkedInShare on WhatsAppShare on e-mail

Google DeepMind has identified a fundamental architectural limitation within Retrieval-Augmented Generation (RAG) systems that rely on dense embeddings. This limitation reveals that fixed-size embeddings cannot represent all relevant document combinations as the database scales, impacting retrieval effectiveness.

The core issue lies in the representational capacity of fixed-size embeddings. A fixed dimension embedding cannot accurately represent all possible combinations of relevant documents when the database surpasses a certain size. This limitation is rooted in principles of communication complexity and sign-rank theory.

Theoretical capacity limits based on embedding size have been established. Embeddings of 512 dimensions reach their limit around 500,000 documents. Increasing the dimensions to 1024 extends the limit to approximately 4 million documents. A further increase to 4096 dimensions raises the ceiling to 250 million documents. These limits represent best-case estimates under free embedding optimization, where vectors are directly optimized against test labels. According to the Google DeepMind report, real-world language-constrained embeddings are anticipated to fail even sooner.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

To empirically demonstrate this limitation, Google DeepMind introduced the LIMIT benchmark, designed to stress-test embedders. The LIMIT benchmark includes two configurations: LIMIT full and LIMIT small. The LIMIT full configuration consists of 50,000 documents, where even strong embedders experience a collapse in performance, with recall@100 often falling below 20%. The LIMIT small configuration, comprising a mere 46 documents, still poses a challenge to models. Performance varies significantly, remaining far from reliable.

Specific results from testing the LIMIT small configuration include: Promptriever Llama3 8B achieved 54.3% recall@2 with 4096 dimensions. GritLM 7B obtained 38.4% recall@2, also with 4096 dimensions. E5-Mistral 7B reached 29.5% recall@2, utilizing 4096 dimensions. Gemini Embed achieved 33.7% recall@2 with 3072 dimensions. The research shows that even with only 46 documents, no embedder achieves full recall, emphasizing that the limitation stems from the single-vector embedding architecture itself, not solely from dataset size.

In contrast, BM25, a classical sparse lexical model, circumvents this limitation. Sparse models operate in effectively unbounded dimensional spaces, facilitating the capture of combinations that dense embeddings cannot effectively represent.

Current RAG implementations often assume that embeddings can scale indefinitely with increasing data volumes. Google DeepMind’s research demonstrates the incorrectness of this assumption, revealing that embedding size inherently constrains retrieval capacity. This constraint significantly impacts enterprise search engines managing millions of documents, agentic systems relying on complex logical queries, and instruction-following retrieval tasks where queries dynamically define relevance.

Existing benchmarks, such as MTEB, do not adequately capture these limitations because they test only a narrow subset of query-document combinations. The research team suggests that scalable retrieval requires moving beyond single-vector embeddings.

Alternatives to single-vector embeddings include Cross-Encoders, which achieve perfect recall on the LIMIT benchmark by directly scoring query-document pairs, albeit with high inference latency. Multi-Vector Models, such as ColBERT, offer more expressive retrieval by assigning multiple vectors per sequence, improving performance on LIMIT tasks. Sparse Models, including BM25, TF-IDF, and neural sparse retrievers, scale better in high-dimensional search but lack semantic generalization.

The key finding is that architectural innovation, rather than simply increasing embedder size, is essential. The research team’s analysis reveals that dense embeddings, despite their widespread use, are constrained by a mathematical limit. Dense embeddings cannot capture all possible relevance combinations once corpus sizes exceed limits tied to embedding dimensionality. This limitation is concretely demonstrated by the LIMIT benchmark, with recall@100 dropping below 20% on LIMIT full (50,000 documents) and even the best models maxing out at approximately 54% recall@2 on LIMIT small (46 documents). Classical techniques like BM25, or newer architectures such as multi-vector retrievers and cross-encoders, remain essential for building reliable retrieval engines at scale.


Featured image credit

Tags: DeepMindFeatured

Related Posts

Google replaces Assistant with Gemini on Android Auto in 2026

Google replaces Assistant with Gemini on Android Auto in 2026

November 28, 2025
Alibaba launches Quark AI glasses in two distinct price tiers

Alibaba launches Quark AI glasses in two distinct price tiers

November 28, 2025
High demand forces Google to tighten free access to Nano Banana Pro

High demand forces Google to tighten free access to Nano Banana Pro

November 28, 2025
Opera Neon update adds 1-minute Deep Research mode

Opera Neon update adds 1-minute Deep Research mode

November 27, 2025
Google Maps officially replaces Assistant with Gemini AI

Google Maps officially replaces Assistant with Gemini AI

November 27, 2025
OpenAI aims for 220 million paying subscribers by 2030 with new adult tier

OpenAI aims for 220 million paying subscribers by 2030 with new adult tier

November 27, 2025

LATEST NEWS

AWS introduces DNS failover feature to prevent future outages

Google replaces Assistant with Gemini on Android Auto in 2026

Amazon unveils Leo Ultra satellite terminal with 1 Gbps speeds

Asus issues critical warning RCE flaw hits AiCloud routers

BankBot YNRK is stealing crypto and bank data in total silence

Alibaba launches Quark AI glasses in two distinct price tiers

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.