Inspiration
We wanted to create a more customizable RAG pipeline. One of the common issues with RAG is the chunk breaking in the middle of important information. We wanted to separate the retrieval and generation, so that the user can see the retrieved chunks in the context of the original document. For the generation step, we wanted to allow the user to reposition or resize the chunk window as well.
What it does
The chunks are embedded with openAI API and then match the chunks to the original text to display. LanceDb is utilized to search for those chunks and retrieve them. User is prompted to put in an input that is a prompt for searching, then the program should compile the context and send it using OpenAI to generate a response.
How we built it
we used lancedb for searching for chunks as a vector database and index documents.
Challenges we ran into
our gui was being very stubborn
Accomplishments that we're proud of
We're proud of fixing the GUI to correctly call upon user input and utilise AI
What we learned
How to use LanceDb, call upon AI API through a GUI, and use TUI's
What's next for Golden Retriever
d
Built With
- lancedb
- lmstudio
- python
- simplepygui
Log in or sign up for Devpost to join the conversation.