Skip to content

A tutorial on how to build Summary Brief from Evaluation Report - Offline+Open Source

Notifications You must be signed in to change notification settings

Edouard-Legoupil/rag_extraction

Repository files navigation

rag_extraction: Summarizing Evidence from Public Evaluation Reports

Retrieval-Augmented Generation (RAG) is an architecture that enhances the capabilities of a Large Language Model (LLM), by incorporating an information retrieval system. This system retrieves relevant data from external sources, here an evaluation report to ground the LLM’s responses.

Check the paper here

How RAG Works?

  • Generative AI Model: The LLM generates responses based on the input prompt.

  • Information Retrieval System: RAG adds an information retrieval component. When given a user query, it retrieves relevant information from external data sources.

  • Response Formulation: The top-ranked search results from the retrieval system are used as input to the LLM. The LLM then generates a response based on both the prompt and the retrieved information.

Benefits

  • Accuracy: By incorporating external facts, RAG improves the accuracy of generative AI models. It ensures that responses are grounded in real-world information.

  • Relevance: RAG allows the model to provide contextually relevant answers by pulling in data from relevant sources.

  • Trustworthiness: Users can trust the responses more when they are backed by factual evidence.

  • Customization: RAG can be constrained to use content from vectorized documents, images, and other data formats specific to the organization.

  • Control: Organizations have control over the grounding data used by the LLM, ensuring alignment with their content and requirements.

Steps

  • Index the Report: Use embedding model to index the report’s content.

  • User Query: When a user asks about specific aspects of the report, send their query to the retrieval system.

  • Retrieve Relevant Sections: The retrieval system fetches relevant sections from the report.

  • Generate Summary: The LLM combines the retrieved information with its own reasoning to generate a concise summary.

In summary, RAG bridges the gap between generative AI and factual information, making it a powerful tool for creating context-aware and trustworthy responses.

About

A tutorial on how to build Summary Brief from Evaluation Report - Offline+Open Source

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages