Skip to content

Latest commit

 

History

History
95 lines (63 loc) · 7.97 KB

README.md

File metadata and controls

95 lines (63 loc) · 7.97 KB
Aegis Athena mission badge

Aegis Athena: Project "Shielded Wisdom"

license

Welcome!

Aegis Athena is a little side-project for showcasing the capabilities of RAG. The user is thrown into an completely unknown environment (a space mission) and is given the task of solving a unknown problem in that environment. An LLM powered RAG pipeline is supposed to aid the user in solving that problem by providing answers to the user's questions concerning this specific domain.

Limitations

Although I achieved much better results, when using custom ReAct-like LLM agents, these pipelines require (only personal experience) very capable models (like GPT-4) to run satisfactorily. Thus, the main problem with this approach are the comparatively high operation costs.

This is why I used a simpler, less capable pipeline architecture.

Retrieval Augmented Generation (RAG) Pipeline

RAG pipeline

Data Preparation

  1. The documentation consists of multiple Markdown files. Firstly, these files are parsed using a custom Markdown reader (inherits the LlamaIndex BaseReader base class). This reader splits the documentation into Nodes, where each node represents a subsection with title and content. These Block Chunks represent the root nodes for the Auto-Merging Retriever.
  2. In the next step, Hierarchical Node Parsing is used to split these root nodes into individual sentences using a custom sentence parser. The resulting Sentence Chunks reference their parent node, so that they can be merged and replaced by the Auto-Merging Retriever. The sentence boundaries are determined using the NLTK PunktSentenceTokenizer.
  3. The root nodes are stored in a MongoDB instance, since there is no need to perform similarity search on them. However, the leaf nodes are being retrieved using similarity search (actually hybrid search, which also utilizes similarity search), which is why there are stored in a Weaviate instance, alongside their vector embedding representation. This two-level setup, follows the approach of retrieving based on a smaller context, which intuitively makes retrieval more precise and adding surrounding context, in order to give the LLM more information. This principle is used in several advanced RAG methods, like Sentence-Window Retrieval or in this case: Auto-Merging Retrieval.

Inference

  1. In order to get that chat-bot behavior, the LlamaIndex CondensePlusContextChatEngine is used. Every time, the user sends a message, given the chat history, the chat engine condenses the interaction into a Query. This Query is then used as input for the retrieval system.
  2. The first step of retrieval is to generate a hypothetical answer to the input query. This approach is called HyDE and it works under the assumption, that a hallucinated answer is often more similar to the actual answer document in the knowledge base, than the input query.
  3. In the next step, both the Query and the HyDE document are combined to create a vector embedding, to be used for similarity search.
  4. The combined document, as well as the vector embedding are then used to query the Weaviate instance and retrieve the top $k$ most similar documents. This step utilizes both similarity search (also called dense search), which is based on the distance between vector embeddings, and BM25F keyword based search (also called sparse search), which judges similarity based on matching keywords. This is thus called Hybrid Search.
  5. As some of the retrieved nodes might be part of the same Block Chunk, the Auto-Merging Retriever groups them together and determines, if they should be replaced/merged by their parent node. This is controlled using the simple_ratio_thresh hyper-parameter, which determines the ratio of child nodes, that need to be retrieved, in order for them to be replaced by their parent node.
  6. Although Hybrid Search already produces good results, they can be further improved using a Cross-Encoder based Reranker. The difference, between Bi-Encoder based retrieval (similarity search) and Cross-Encoder based retrieval is that, whereas the former approach creates vector embeddings for both the documents in the knowledge base as well as the input query and then computes similarity based on a distance metric (e.g.: cosine distance), the latter approach feeds both the input query and every potentially relevant document in the knowledge base into a classifier model, that outputs a relevance score between 0 and 1. Although this is computationally more expensive, we get a higher quality estimate of document relevance. This is why this step is preceeded by Cross-Encoder based retrieval, in order to narrow down the search space before using reranking to get the final context for the LLM.
  7. Lastly, the Chat History, the retrieved Context and a System Header are combined to form the final prompt for the LLM, which then produces the response shown to the user. The system prompt, provides the LLM with some basic information about the situation and persona description for the chat-bot.

Technologies

Showcase

The AI assistant provides the user with information about potential reasons for the OFFLINE status of the LN2 tank and possible solutions for identifying and resolving the underlyling problem.


The user retrieves status information of all systems.


The AI assistant provides the user with information about potential reasons for the MALFUNC status of the ECS system and possible solutions for identifying the underlyling problem.


The user retrieves detailed information of the ECS controller.

Contributing

Feel free to fork the project, create a feature branch, and send me a pull request!

License

This project is licensed under the MIT License.

Contact

You can reach out to me through Reddit. I'd love to chat about this project or any other interesting topics!


Buy Me A Coffee