🤭 Hello! In this project we will dive into our exciting work carried out in the Postgraduate discipline of Sequential Models, taught by professor Ivanovitch Silva, and offered by PPGEEC/UFRN.
This simple system allows users to engage in conversations about a YouTube video with itself, without the need to watch it firsthand. For this project, our group participants are Deborah Moreira, Miguel Euripedes and myself.
By using this app you can ask ChatGPT questions about things in a video. To minimize the cost per token, we use a Vector Store to perform a semantic search and return the most related documents.
The architecture is divided into two parts, data ingestion and information retrieval to feed ChatGPT (LLM).
We produced two explanatory materials about the architecture, a video about the architecture at a high level and a medium post with detailed content.
Simpler, 🎮 play on
Run notebook in notebooks/YoutubeRetrievalQA.ipynb
Install dependencies:
pip3 install -r requirements.txt
Run Gradio app:
python3 app.py
In short, in pipeline ingestion we have:
To retrieval, the first option is stateless app
The second option is a chain with a memory object