Open-Source AI Cookbook documentation

Multi-agent RAG System πŸ€–πŸ€πŸ€–

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Open In Colab

Multi-agent RAG System πŸ€–πŸ€πŸ€–

Authored by: Sergio Paniego

🚨 NOTE: This tutorial is advanced. You should have a solid understanding of the concepts discussed in the following cookbooks before diving in:

In this notebook, we will create a multi-agent RAG system, a system where multiple agents work together to retrieve and generate information, combining the strengths of retrieval-based systems and generative models.

What is a Multi-agent RAG System? πŸ€”

A Multi-agent Retrieval-Augmented Generation (RAG) system consists of multiple agents that collaborate to perform complex tasks. The retrieval agent retrieves relevant documents or information, while the generative agent synthesizes that information to generate meaningful outputs. There is a Manager Agent that orchestrates the system and selects the most appropriate agent for the task based on the user input.

The original idea for this recipe comes from this post. You may find more details about it there.

Below, you can find the architecture that we will build.

multiagent_rag_system.png

1. Install dependencies

First, let’s install the dependencies:

!pip install -q git+https://github.com/huggingface/transformers.git#egg=transformers[agents]
!pip install markdownify duckduckgo-search spaces gradio-tools langchain langchain-community langchain-huggingface faiss-cpu --upgrade -q

Let’s login in order to call the HF Inference API:

from huggingface_hub import notebook_login

notebook_login()

2. Let’s create our multi-agent RAG system

In this section, we will create each of the agents present in our RAG system.

We will have 3 agents managed by a central one (refer to the image for details):

  • πŸ•΅πŸ’¬ Web search agent: It will include the DuckDuckGoSearchTool tool and the VisitWebpageTool. As you can see, each agent may contain a list of tools.
  • πŸ•΅πŸ’¬ Retriever agent: It will include two tools for retrieving information from two different knowledge bases.
  • πŸ•΅πŸ’¬ Image generation agent: It will include a prompt generator tool in addition to the image generation tool.

πŸ’‘ In addition to these agents, the central/orchestrator agent will also have access to the code interpreter tool to execute code.

We will use Qwen/Qwen2.5-72B-Instruct as the LLM for each component, which will be accessed via the Inference API. Depending on the agent, a different LLM model may be used.

Note: The Inference API hosts models based on various criteria, and deployed models may be updated or replaced without prior notice. Learn more about it here.

from transformers.agents import HfApiEngine

llm_model = "Qwen/Qwen2.5-72B-Instruct"
llm_engine = HfApiEngine(llm_model)

Let’s dive into the details of each agent!

2.1 Web search agent πŸ”

The Web search agent will utilize the DuckDuckGoSearchTool to search the web and gather relevant information. This tool acts as a search engine, querying for results based on the specified keywords.

To make the search results actionable, we also need the agent to access the web pages retrieved by DuckDuckGo. That can be achieved by using the built-in VisitWebpageTool.

Let’s explore how to set it up and integrate it into our system!

The following code comes from the original Have several agents collaborate in a multi-agent hierarchy πŸ€–πŸ€πŸ€– recipe, so refer to it for more details.

2.1.1 Build our multi-tool web agent πŸ€–

Now that we’ve set up the basic search and webpage tools, let’s build our multi-tool web agent. This agent will combine several tools to perform more complex tasks, leveraging the capabilities of the ReactJsonAgent.

The ReactJsonAgent is particularly well-suited for web search tasks because its JSON action formulation requires only simple arguments and works seamlessly in sequential chains of single actions. This makes it an excellent choice for scenarios where we need to search the web for relevant information and retrieve detailed content from specific web pages. In contrast, CodeAgent action formulation is better suited for scenarios involving numerous or parallel tool calls.

By integrating multiple tools, we can ensure that our agent interacts with the web in a sophisticated and efficient manner.

Let’s dive into how to set this up and integrate it into our system!

from transformers.agents import ReactCodeAgent, ReactJsonAgent, ManagedAgent
from transformers.agents.search import DuckDuckGoSearchTool, VisitWebpageTool

web_agent = ReactJsonAgent(tools=[DuckDuckGoSearchTool(), VisitWebpageTool()], llm_engine=llm_engine)

Now that we have our first agent, let’s wrap it as a ManagedAgent so the central agent can use it.

managed_web_agent = ManagedAgent(
    agent=web_agent,
    name="search",
    description="Runs web searches for you. Give it your query as an argument.",
)

2.2 Retriever agent πŸ€–πŸ”

The second agent in our multi-agent system is the Retriever agent. This agent is responsible for gathering relevant information from different sources. To achieve this, it will utilize two tools that retrieve data from two separate knowledge bases.

We will reuse two data sources that were previously used in other RAG recipes, which will allow the retriever to efficiently gather information for further processing.

By leveraging these tools, the Retriever agent can access diverse datasets, ensuring a comprehensive collection of relevant information before passing it on to the next step in the system.

Let’s explore how to set up the retriever and integrate it into our multi-agent system!

2.2.1 HF docs retriever tool πŸ“š

The first retriever tool comes from the Agentic RAG: turbocharge your RAG with query reformulation and self-query! πŸš€ recipe.

For this retriever, we will use a dataset that contains a compilation of documentation pages for various huggingface packages, all stored as markdown files. This dataset serves as the knowledge base for the retriever agent to search and retrieve relevant documentation.

To make this dataset easily accessible for our agent, we will:

  1. Download the dataset: We will first fetch the markdown documentation.
  2. Embed the data: We will then convert the documentation into embeddings using a FAISS vector store for efficient similarity search.

By doing this, the retriever tool can quickly access the relevant pieces of documentation based on the search query, enabling the agent to provide accurate and detailed information.

Let’s go ahead and set up the tool to handle the documentation retrieval!

import datasets

knowledge_base = datasets.load_dataset("m-ric/huggingface_doc", split="train")
from tqdm import tqdm
from transformers import AutoTokenizer
from langchain.docstore.document import Document
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain.vectorstores import FAISS
from langchain_huggingface import HuggingFaceEmbeddings
from langchain_community.vectorstores.utils import DistanceStrategy

source_docs = [
    Document(page_content=doc["text"], metadata={"source": doc["source"].split("/")[1]}) for doc in knowledge_base
]

text_splitter = RecursiveCharacterTextSplitter.from_huggingface_tokenizer(
    AutoTokenizer.from_pretrained("thenlper/gte-small"),
    chunk_size=200,
    chunk_overlap=20,
    add_start_index=True,
    strip_whitespace=True,
    separators=["\n\n", "\n", ".", " ", ""],
)

# Split docs and keep only unique ones
print("Splitting documents...")
docs_processed = []
unique_texts = {}
for doc in tqdm(source_docs):
    new_docs = text_splitter.split_documents([doc])
    for new_doc in new_docs:
        if new_doc.page_content not in unique_texts:
            unique_texts[new_doc.page_content] = True
            docs_processed.append(new_doc)

print("Embedding documents...")
embedding_model = HuggingFaceEmbeddings(model_name="thenlper/gte-small")
huggingface_doc_vector_db = FAISS.from_documents(
    documents=docs_processed,
    embedding=embedding_model,
    distance_strategy=DistanceStrategy.COSINE,
)

Now that we have the documentation embedded in FAISS, let’s create the RetrieverTool. This tool will query the FAISS vector store to retrieve the most relevant documents based on the user’s query.

This will allow the retriever agent to access and provide relevant documentation when queried.

from transformers.agents import Tool
from langchain_core.vectorstores import VectorStore


class RetrieverTool(Tool):
    name = "retriever"
    description = "Using semantic similarity, retrieves some documents from the knowledge base that have the closest embeddings to the input query."
    inputs = {
        "query": {
            "type": "string",
            "description": "The query to perform. This should be semantically close to your target documents. Use the affirmative form rather than a question.",
        }
    }
    output_type = "string"

    def __init__(self, vectordb: VectorStore, **kwargs):
        super().__init__(**kwargs)
        self.vectordb = vectordb

    def forward(self, query: str) -> str:
        assert isinstance(query, str), "Your search query must be a string"

        docs = self.vectordb.similarity_search(
            query,
            k=7,
        )

        return "\nRetrieved documents:\n" + "".join(
            [f"===== Document {str(i)} =====\n" + doc.page_content for i, doc in enumerate(docs)]
        )
huggingface_doc_retriever_tool = RetrieverTool(huggingface_doc_vector_db)

2.2.2 PEFT issues retriever tool

For the second retriever, we will use the PEFT issues as data source as in the Simple RAG for GitHub issues using Hugging Face Zephyr and LangChain.

Again, the following code comes from that recipe so refer to it for more details!

from google.colab import userdata

GITHUB_ACCESS_TOKEN = userdata.get("GITHUB_PERSONAL_TOKEN")
from langchain.document_loaders import GitHubIssuesLoader

loader = GitHubIssuesLoader(repo="huggingface/peft", access_token=GITHUB_ACCESS_TOKEN, include_prs=False, state="all")
docs = loader.load()
splitter = RecursiveCharacterTextSplitter(chunk_size=512, chunk_overlap=30)
chunked_docs = splitter.split_documents(docs)
peft_issues_vector_db = FAISS.from_documents(chunked_docs, embedding=embedding_model)

Let’s now generate the second retriever tool using the same RetrieverTool.

peft_issues_retriever_tool = RetrieverTool(peft_issues_vector_db)

2.2.3 Build the Retriever agent

Now that we’ve created the two retriever tools, it’s time to build the Retriever agent. This agent will manage both tools and retrieve relevant information based on the user query.

We’ll use the ManagedAgent to integrate these tools and pass the agent to the central agent for coordination.

retriever_agent = ReactJsonAgent(
    tools=[huggingface_doc_retriever_tool, peft_issues_retriever_tool],
    llm_engine=llm_engine,
    max_iterations=4,
    verbose=2,
)
managed_retriever_agent = ManagedAgent(
    agent=retriever_agent,
    name="retriever",
    description="Retrieves documents from the knowledge base for you that are close to the input query. Give it your query as an argument. The knowledge base includes Hugging Face documentation and PEFT issues.",
)

2.3 Image generation agent 🎨

The third agent in our system is the Image generation agent. This agent will have two tools: one for refining the user query and another for generating the image based on the query. In this case, we will use the CodeAgent instead of a ReactAgent since the set of actions can be executed in one shot.

You can find more details about the image generation agent in the Agents, supercharged - Multi-agents, External tools, and more documentation.

Let’s dive into how these tools will work together to generate images based on user input!

from transformers import load_tool, CodeAgent

prompt_generator_tool = Tool.from_space(
    "sergiopaniego/Promptist", name="generator_tool", description="Optimizes user input into model-preferred prompts"
)
image_generation_tool = load_tool("m-ric/text-to-image", cache=False)
image_generation_agent = CodeAgent(tools=[prompt_generator_tool, image_generation_tool], llm_engine=llm_engine)

πŸ–Ό Again, we use ManagedAgent to tell the central agent that it can manage it. Additionally, we’ve included an additional_prompting parameter to ensure the agent returns the generated image instead of just a text description.

managed_image_generation_agent = ManagedAgent(
    agent=image_generation_agent,
    name="image_generation",
    description="Generates images from text prompts. Give it your prompt as an argument.",
    additional_prompting="\n\nYour final answer MUST BE only the generated image location.",
)

3. Let’s add the general agent manager to orchestrate the system

The central agent manager will coordinate tasks between the agents. It will:

  • Receive user input and decide which agent (Web search, Retriever, Image generation) handles it.
  • Delegate tasks to the appropriate agent based on the user’s query.
  • Collect and synthesize results from the agents.
  • Return the final output to the user.

We include all the agents we’ve developed as managed_agents and add any necessary imports for the code executor under additional_authorized_imports.

manager_agent = ReactCodeAgent(
    tools=[],
    llm_engine=llm_engine,
    managed_agents=[managed_web_agent, managed_retriever_agent, managed_image_generation_agent],
    additional_authorized_imports=["time", "datetime", "PIL"],
)

Now that everything is set up, let’s test the performance of the multi-agent RAG system!

To do so, we’ll provide some example queries and observe how the system delegates tasks between the agents, processes the information, and returns the final results.

This will help us understand the efficiency and effectiveness of our agents working together, and identify areas for optimization if necessary.

Let’s go ahead and run some tests!

3.1 Example trying to trigger the search agent

manager_agent.run("How many years ago was Stripe founded?")

3.2 Example trying to trigger the image generator agent

result = manager_agent.run(
    "Improve this prompt, then generate an image of it.", prompt="A rabbit wearing a space suit"
)
>>> from IPython.display import Image, display

>>> display(Image(filename=result))

3.3 Example trying to trigger the retriever agent for the HF docs knowledge base

manager_agent.run("How can I push a model to the Hub?")

3.4 Example trying to trigger the retriever agent for the PEFT issues knowledge base

manager_agent.run("How do you combine multiple adapters in peft?")

🏁 Final Thoughts

We have successfully built a multi-agent RAG system that integrates Web Search, Document Retrieval, and Image Generation agents, all orchestrated by a central agent manager. This architecture enables seamless task delegation, efficient processing, and the flexibility to handle a wide variety of user queries.

πŸ” Explore More

< > Update on GitHub