Advanced Self-Reflective RAG

Advanced Self-Reflective RAG

Overview As most LLMs are only trained periodically on a large amount of public data, they cannot access the latest information and/or private data. Retrieval-Augmented Generation (RAG) is a core paradigm for developing applications with LLMs, addressing this issue by connecting to external data sources. A basic RAG pipeline includes embedding user queries, retrieving relevant … Read more

The RAG vs Long-Context Debate: No Need to Fight

The RAG vs Long-Context Debate: No Need to Fight

Introduction Hello everyone, I am Liu Cong from NLP. As the context length supported by large models continues to increase, a debate has emerged online (many groups are discussing this topic, so I would like to share my thoughts) regarding RAG and Long-Context, which is really unnecessary… The main point is that the two are … Read more

Overview of Querying Process in LlamaIndex

Overview of Querying Process in LlamaIndex

Explanation Querying is the most important part of LLM applications. In LlamaIndex, once you have completed: data loading, building the index, and storing the index, you can proceed to the most crucial part of LLM applications: querying. A simple query is just a prompt call to the large language model: it can be a question … Read more

Pinecone + LangChain: Building an Efficient AI Retrieval System

Pinecone + LangChain: Building an Efficient AI Retrieval System

Pinecone + LangChain: Building an Efficient AI Retrieval System Recently, while experimenting with AI retrieval systems, I found that Pinecone and LangChain are a match made in heaven. Pinecone is a powerful vector database, and LangChain is a flexible framework. Combining the two allows you to easily build an efficient AI retrieval system. Today, I … Read more

LlamaIndex: A New Document Summary Index for QA Systems

LlamaIndex: A New Document Summary Index for QA Systems

In this blog post, we introduce a brand new data structure in LlamaIndex: the Document Summary Index. We describe how it helps provide better retrieval performance compared to traditional semantic search, along with an example. https://github.com/jerryjliu/llama_index 1 Background One of the core scenarios for large language models (LLM) is question answering on user data. To … Read more

Advanced RAG – Composable Retrieval with LlamaIndex

Advanced RAG - Composable Retrieval with LlamaIndex

LlamaIndex is a simple and flexible data framework that connects custom data sources with large language models. LlamaIndex provides comprehensive support for RAG. Advanced RAG (Retrieval-Augmented Generation) techniques can be modeled using a composable hierarchical abstraction. The retrieved text can be linked to the following elements: Retriever Text Pipeline Query Engine The retrieval of RAG … Read more