@jerryjliu0
A core retrieval idea that will lead to better results for your LLM QA system is decoupling embedding representations from raw text chunks (s/o @md_rumpf for inspiration). ✂️ There’s actually different ways to take advantage of this idea - and we’ll show how all of these are possible with @llama_index 👇 1️⃣ Embed a summary -> link to more documents associated with the text. ✅ This can help retrieve relevant documents at a high-level before retrieving chunks, vs. retrieving chunks directly (that might be in irrelevant documents). 2️⃣ Embed a sentence -> link to a window around the sentence. ✅ This allows for finer-grained retrieval of relevant context (embedding giant chunks leads to “lost in the middle” problems), but also ensures enough context for LLM synthesis. Guides 📗: 1️⃣ is possible with our recursive retriever, or our out of the box document summary index: Recursive Retriever: https://t.co/HmF2Dib6ho Document Summary Index: https://t.co/HjheQ8tV3N 2️⃣ is possible with our SentenceWindow parser + Metadata Sentence Window: https://t.co/3SN5Xt6vrT