Inspiration

We wanted to create a more customizable RAG pipeline. One of the common issues with RAG is the chunk breaking in the middle of important information. We wanted to separate the retrieval and generation, so that the user can see the retrieved chunks in the context of the original document. For the generation step, we wanted to allow the user to reposition or resize the chunk window as well.

What it does

The chunks are embedded with openAI API and then match the chunks to the original text to display. LanceDb is utilized to search for those chunks and retrieve them. User is prompted to put in an input that is a prompt for searching, then the program should compile the context and send it using OpenAI to generate a response.

How we built it

we used lancedb for searching for chunks as a vector database and index documents.

Challenges we ran into

our gui was being very stubborn

Accomplishments that we're proud of

We're proud of fixing the GUI to correctly call upon user input and utilise AI

What we learned

How to use LanceDb, call upon AI API through a GUI, and use TUI's

What's next for Golden Retriever

d

Built With

  • lancedb
  • lmstudio
  • python
  • simplepygui
Share this project:

Updates