11 September 2024 10:00 - 10:30
Next-gen context retrieval: Navigating through the RAG's challenges in generative AI
Large Language Models (LLMs) power cutting-edge Generative AI applications, yet their effectiveness is heavily dependent on the efficient retrieval of relevant context.
While Retrieval-Augmented Generation (RAG) has emerged as a powerful alternative to fine-tuning, real-world implementations reveal critical challenges: hallucinations, retrieval latency, lack of adaptability, and context drift.
In this talk, we dissect the limitations of both fine-tuning and traditional RAG-based methods, highlighting real-world bottlenecks in context retrieval. We introduce Adaptive Context Retrieval (ACR), a dynamic approach that refines retrieval mechanisms based on evolving user queries, document structures, and contextual dependencies.
By leveraging agentic workflows, hybrid retrieval techniques, and reinforcement learning, ACR promises to bridge the gap between static retrieval and real-time adaptability.