Welcome to the Multi-Model AI Microservices Project! This project integrates three independent AI-powered services that can be deployed together seamlessly:
- Recommender Service – Suggests products based on user history.
- Fraud Detection Service – Detects suspicious transactions.
- Chatbot Service – Engages users in conversation.
Each service is built using FastAPI and is fully containerized with Docker for easy deployment. You can run everything locally using Docker Compose or deploy it to the cloud with Kubernetes.
Ensure Python is installed, then run:
pip install -r requirements.txtRun the following script to fetch the necessary AI models:
python scripts/download_models.pyThis ensures that each service has the correct model files before running.
Launch all three services at once:
docker compose up --buildThis will spin up the services on the following ports:
- Recommender Service → http://localhost:8001
- Fraud Detection Service → http://localhost:8002
- Chatbot Service → http://localhost:8003
You can test each service using cURL or Postman.
curl "http://localhost:8001/recommend?user_id=123"curl -X POST "http://localhost:8002/detect_fraud" \
-H "Content-Type: application/json" \
-d '{"amount": 1000, "location": "NY"}'curl -X POST "http://localhost:8003/chat" \
-H "Content-Type: application/json" \
-d '{"user_input": "Hello!"}'For cloud deployment, apply the Kubernetes configuration:
kubectl apply -f k8s/To run AI-Orchestrator efficiently, ensure your system meets the following hardware specifications:
- CPU: Multi-core processor (4+ cores, e.g., Intel i7/AMD Ryzen 7) for handling concurrent requests.
- GPU (Optional): Recommended for AI models requiring acceleration (e.g., deep learning). An NVIDIA GPU with CUDA support (8GB+ VRAM) is ideal.
- Minimum: 16GB RAM
- Recommended: 32GB RAM for handling multiple AI services efficiently.
- Minimum: 50GB free space per model
- Recommended: SSD (NVMe preferred) for faster data access and improved system responsiveness.
- Stable high-speed internet (especially if using cloud-based AI services or real-time processing).
- Linux (Ubuntu 20.04+ recommended) or Windows 10/11 with WSL2 for better compatibility with Docker and Kubernetes.
Ensure the following are installed:
- Docker (Latest version) for containerized deployments.
- Kubernetes (Optional, for orchestration at scale).
- Python 3.8+ for AI model execution and service scripts.
- GPU Drivers & CUDA (If using GPU for acceleration).
For production or large-scale deployments:
- Consider cloud-based solutions (AWS/GCP/Azure) for dynamic scaling.
- Use load balancing and Kubernetes clusters to optimize resource utilization.
- Ensure Docker Desktop is running before using
docker compose up. - Before downloading models, update
scripts/download_models.pywith the correct model URLs. - You can modify the Dockerfile to match your specific requirements.
