# 1. Run the menu
./menu.sh
# 2. Choose Option 1 (Quick Setup)
# 3. Choose Option 8 (Install a model - pick llama3.2:1b)
# 4. Choose Option 4 (Start Services)
# 5. Open browser: http://localhost:8501That's it! You now have a local AI chat interface.
You asked great questions about getting this local AI interface working simply and reliably. This guide answers them all.
Q1: Why isn't there a log file generated after install or health check?
- A: Fixed! All scripts now create logs in
./logs/directory with full paths and installation details.
Q2: Where are all the images and Docker running? Logs should mention paths.
- A: All paths are now logged:
- Logs:
./logs/ - Data:
./data/ - Models:
~/.ollama/models/ - Run
./menu.sh→ Option 9 to see all paths
- Logs:
Q3: If I restart or hibernate my laptop, do I need to start Docker again?
- A: Yes for manual mode, but you can auto-start services. See STARTUP_GUIDE.md for systemd setup.
Q4: Why don't we create a menu script?
- A: Done!
./menu.shgives you all options in one place.
Q5: Why are we still having problems?
- A: We've simplified everything. You DON'T need Docker for basic chat with local LLMs. Use native mode instead!
./menu.shOptions include:
- Quick Setup (Native - No Docker)
- Docker Setup (Full)
- Health Check
- Start Services (Native Mode)
- Stop All Services
- Restart Docker Containers
- View Logs
- Install/Pull Ollama Models
- System Information
- Run Troubleshooting
./run_local.shStarts everything automatically in native mode.
# Start Ollama
ollama serve &
# Install a model (first time only)
ollama pull llama3.2:1b
# Start the UI
streamlit run src/app.py- Python 3.8+ (you probably already have this)
- Ollama (we'll install it for you)
- A small model (1-3GB, we'll help you pick one)
- ❌ Docker (unless you want the full setup)
- ❌ GPU (CPU works fine for small models)
- ❌ Lots of RAM (8GB is enough for small models)
- ❌ OpenAI API key (this is local-only)
./menu.sh
# Choose Option 1: Quick SetupOr manually:
curl -fsSL https://ollama.ai/install.sh | sh./menu.sh
# Choose Option 8: Install/Pull Ollama Models
# Pick option 2: llama3.2:1b (recommended for testing)Model Recommendations:
| Model | Size | Speed | Use Case |
|---|---|---|---|
llama3.2:1b |
~1GB | Very Fast | Testing, quick responses |
llama3.2 (3B) |
~2GB | Fast | Balanced performance |
qwen2.5-coder:1.5b |
~1GB | Very Fast | Code-focused tasks |
phi3.5 |
~2GB | Fast | General purpose |
./menu.sh
# Choose Option 4: Start Services (Native Mode)Open your browser to: http://localhost:8501
You have 3 options:
./menu.sh
# Choose Option 4: Start Services./run_local.shSee the complete guide in STARTUP_GUIDE.md
Quick version:
# Create Ollama service
sudo tee /etc/systemd/system/ollama.service > /dev/null << 'EOF'
[Unit]
Description=Ollama Service
After=network-online.target
[Service]
Type=simple
User=$(whoami)
Environment="OLLAMA_HOST=0.0.0.0:11434"
ExecStart=/usr/local/bin/ollama serve
Restart=always
[Install]
WantedBy=multi-user.target
EOF
sudo systemctl enable ollama
sudo systemctl start ollamaRun this to see all paths:
./menu.sh
# Choose Option 9: System InformationOr check manually:
Your Project Directory: /workspace/
├── logs/ ← All logs here
│ ├── streamlit.log ← UI logs
│ ├── ollama.log ← Ollama logs
│ └── *.log ← Other logs
├── data/ ← Your data
│ ├── documents/ ← Uploaded documents
│ ├── conversations/ ← Saved chats
│ └── uploads/ ← File uploads
├── src/ ← App code
│ └── app.py ← Main application
├── menu.sh ← Interactive menu
├── run_local.sh ← Quick start script
└── .env ← Configuration
System Directories:
~/.ollama/models/ ← Downloaded models
/usr/local/bin/ollama ← Ollama binary
./menu.sh
# Choose Option 10: Run TroubleshootingThis creates a detailed report at logs/troubleshooting_*.log
# View live logs
tail -f logs/streamlit.log # UI logs
tail -f logs/ollama.log # Ollama logs
# Or use the menu
./menu.sh
# Choose Option 7: View Logs# Find what's using port 8501
lsof -i:8501
# Kill it
kill $(lsof -ti:8501)
# Or use menu
./menu.sh
# Choose Option 5: Stop All Services./menu.sh
# Choose Option 3: Health CheckThis will tell you exactly what's wrong and how to fix it.
No! Docker is optional. For basic local AI chat, native mode is simpler and faster.
Use Docker only if you:
- Want PostgreSQL database
- Need multiple users
- Want containerized deployment
For testing: llama3.2:1b (smallest, fastest)
For better quality: llama3.2 (3B version)
For coding: qwen2.5-coder:1.5b
In the UI:
- Go to sidebar
- Find "Model Settings"
- Select different model from dropdown
Or install new ones:
ollama pull model-nameYes! In the UI:
- Sidebar → Choose Provider → "OpenAI"
- Enter your API key
- Select OpenAI model
./data/conversations/
View them in the UI:
- Sidebar → Navigation → "🗂️ Conversations"
- Use smaller models:
llama3.2:1bis very fast - Native mode: Faster than Docker for local use
- SSD: Store models on SSD if possible
# Remove unused models
ollama rm model-name
# See all models and their sizes
ollama listOnce you're up and running:
- Explore Features: Check sidebar options
- Upload Documents: Use the Documents page to add files for RAG
- Try Different Agents: Switch between agents in sidebar
- Save Conversations: Use the save button to keep important chats
- Read Full Docs: See
docs/folder for advanced features
# Everything through menu
./menu.sh
# Quick start
./run_local.sh
# Manual control
ollama serve & # Start Ollama
streamlit run src/app.py # Start UI
ollama pull llama3.2:1b # Get a model
# Check status
curl http://localhost:11434/api/tags # Ollama
curl http://localhost:8501/_stcore/health # Streamlit
# View logs
tail -f logs/streamlit.log
tail -f logs/ollama.log
# Stop everything
./menu.sh # Option 5- Run troubleshooting:
./menu.sh→ Option 10 - Check logs:
./menu.sh→ Option 7 - System info:
./menu.sh→ Option 9 - Read docs:
docs/STARTUP_GUIDE.md - Health check:
./menu.sh→ Option 3
- Ollama installed and running
- At least one model downloaded
- Streamlit UI accessible at http://localhost:8501
- Can send a message and get a response
- Logs being created in
./logs/directory
If all checked, you're ready to go! 🎉
Focus: Simple local AI chat without complexity.
For advanced features (RAG, agents, MCP), see the full documentation in docs/.