Dot is a standalone, open-source application designed for seamless interaction with documents and files using local LLMs and Retrieval Augmented Generation (RAG). It is inspired by solutions like Nvidia's Chat with RTX, providing a user-friendly interface for those without a programming background. Using the Phi-3 LLM by default, Dot ensures accessibility and simplicity right out of the box.
Screen.Recording.2024-05-20.at.20.27.52.2.mp4
Dot allows you to load multiple documents into an LLM and interact with them in a fully local environment. Supported document types include PDF, DOCX, PPTX, XLSX, and Markdown. Users can also engage with Big Dot for inquiries not directly related to their documents, similar to interacting with ChatGPT.
Built with Electron JS, Dot encapsulates a comprehensive Python environment that includes all necessary libraries. The application leverages libraries such as FAISS for creating local vector stores, Langchain, llama.cpp & Huggingface for setting up conversation chains, and additional tools for document management and interaction.
To use Dot:
- Visit the Dot website to download the application for Apple Silicon or Windows.
For developers:
- Clone the repository
$ https://github.com/alexpinel/Dot.git
- Install Node js and then run
npm install
inside the project repository, you can runnpm install --force
if you face any issues at this stage
Now, it is time to add a full python bundle to the app. The purpose of this is to create a distributable environment with all necessary libraries, if you only plan on using Dot from the console you might not need to follow this particular step but then make sure to replace the python path locations specified in src/index.js
. Creating the python bundle is covered in detail here: https://til.simonwillison.net/electron/python-inside-electron , the bundles can also be installed from here: https://github.com/indygreg/python-build-standalone/releases/tag/20240224
Having created the bundle, please rename it to 'python' and place it inside the llm
directory. It is now time to get all necessary libraries, keep in mind that running a simple pip install
will not work without specifying the actual path of the bundle so use this instead: path/to/python/.bin/or/.exe -m pip install
Required python libraries:
- pytorch link (CPU version recommended as it is lighter than GPU)
- langchain link
- FAISS link
- HuggingFace link
- llama.cpp link (Use CUDA implementation if you have an Nvidia GPU!)
- pypdf link
- docx2txt link
- Unstructured link (Use
pip install "unstructured[pptx, md, xlsx]
for the file formats)
Now python should be setup and running! However, there is still a few more steps left, now is the time to add the final magic to Dot! First, create a folder inside the llm
directory and name it mpnet
, there you will need to install sentence-transformers to use for the document embeddings, fetch all the files from the following link and place them inside the new folder: sentence-transformers/all-mpnet-base-v2
Finally, download the Mistral 7B LLM from the following link and place it inside the llm/scripts
directory alongside the python scripts used by Dot: TheBloke/Mistral-7B-Instruct-v0.2-GGUF
That's it! If you follow these steps you should be able to get it all running, please let me know if you are facing any issues :)
- Linux support
- Choice of LLM - Done!
- Image file support
- Enhanced document awareness beyond content
- Simplified file loading (select individual files, not just folders)
- Increased security measures for using local LLMs
- Support for additional document types
- Efficient file database management for quicker access to groups of files
Contributions are highly encouraged! As a student managing this project on the side, any help is greatly appreciated. Whether it's coding, documentation, or feature suggestions, please feel free to get involved!