An LLM driven recommendation system based on Radarr and Sonarr library or watch history information
-
Updated
Apr 14, 2025 - Vue
An LLM driven recommendation system based on Radarr and Sonarr library or watch history information
A framework for using local LLMs (Qwen2.5-coder 7B) that are fine-tuned using RL to generate, debug, and optimize code solutions through iterative refinement.
A fully customizable, super light-weight, cross-platform GenAI based Personal Assistant that can be run locally on your private hardware!
🤖 An Intelligent Chatbot: Powered by the locally hosted Ollama 3.2 LLM 🧠 and ChromaDB 🗂️, this chatbot offers semantic search 🔍, session-aware responses 🗨️, and an interactive Streamlit interface 🎨 for seamless user interaction. 🚀
🚀 A powerful Flutter-based AI chat application that lets you run LLMs directly on your mobile device or connect to local model servers. Features offline model execution, Ollama/LLMStudio integration, and a beautiful modern UI. Privacy-focused, cross-platform, and fully open source.
An AI-powered assistant to streamline knowledge management, member discovery, and content generation across Telegram and Twitter, while ensuring privacy with local LLM deployment.
Run large language models like Qwen and LLaMA locally on Android for offline, private, real-time question answering and chat — powered by ONNX Runtime.
Local Retrieval-Augmented Generation (RAG) pipeline using LangChain and ChromaDB to query PDF files with LLMs.
a vs code extension , to run local llms for code assistance
**Ask CLI** is a command-line tool for interacting with a local LLM (Large Language Model) server. It allows you to send queries and receive concise command-line responses.
AI-powered Bash terminal built with Python, Tkinter, tkterm, using local LLM through LMStudio for natural language command generation; features whitelist/blacklist management, intuitive interface.
AI-powered code and idea assistant for developers: local-first, doc-aware, and fully test-automated.
An AI-powered system for extracting cancer-related information from patient Electronic Health Record (EHR) notes
Privacy-First Local AI Chat for VS Code which is also Beautiful
Add a description, image, and links to the local-llm-integration topic page so that developers can more easily learn about it.
To associate your repository with the local-llm-integration topic, visit your repo's landing page and select "manage topics."