Giuseppe Futia, PhD’s Post

🚀 How to Scale GraphRAG to Millions of Documents? 🧩 Traditional Graph-RAG approaches extract triples (subject, predicate, object) from documents, often via LLMs, to enable multi-hop reasoning over knowledge graphs. But this offline triple extraction step doesn’t scale well to web-scale corpora. 📘 A new paper by Huawei and The University of Edinburgh, “Millions of GeAR-s: Extending GraphRAG to Millions of Documents”, proposes an elegant alternative: 🔹 Pseudo-alignment instead of extraction Rather than generating new triples, the system performs an online pseudo-alignment between retrieved text chunks and existing Wikidata triples using a combination of dense and sparse retrieval. 🔹 Graph-guided multi-agent reasoning A multi-step (agentic) process decomposes complex queries into simpler sub-questions, traverses the KG to expand reasoning chains, and filters irrelevant passages before generating the final answer. 🔹 Scalable architecture By leveraging external KGs like Wikidata and avoiding heavy LLM extraction, the system scales to millions of passages while retaining graph-enhanced reasoning ability. The paper is available here: https://lnkd.in/enwgWDte

How reliable is the pseudo-alignment compared to full triple extraction—any loss in reasoning accuracy?

Like
Reply

To view or add a comment, sign in

Explore content categories