Skip to content

Operationalization of a machine learning microservice API using Docker and Kubernetes

Notifications You must be signed in to change notification settings

FavourDaniel/Udacity-Machine-Learning-Project

Repository files navigation

CircleCI

This is my 4th project which I built during my Udacity Cloud DevOps Nanodegree program

Project Overview

In this project, you will apply the skills you have acquired in this course to operationalize a Machine Learning Microservice API.

You are given a pre-trained, sklearn model that has been trained to predict housing prices in Boston according to several features, such as average rooms in a home and data about highway access, teacher-to-pupil ratios, and so on. You can read more about the data, which was initially taken from Kaggle, on the data source site. This project tests your ability to operationalize a Python flask app—in a provided file, app.py—that serves out predictions (inference) about housing prices through API calls. This project could be extended to any pre-trained machine learning model, such as those for image recognition and data labeling.

Project Tasks

Your project goal is to operationalize this working, machine learning microservice using kubernetes, which is an open-source system for automating the management of containerized applications. In this project you will:

  • Test your project code using linting
  • Complete a Dockerfile to containerize this application
  • Deploy your containerized application using Docker and make a prediction
  • Improve the log statements in the source code for this application
  • Configure Kubernetes and create a Kubernetes cluster
  • Deploy a container using Kubernetes and make a prediction
  • Upload a complete Github repo with CircleCI to indicate that your code has been tested

You can find a detailed project rubric, here.

The final implementation of the project will showcase your abilities to operationalize production microservices.


Setup the Environment

  • Create a virtualenv with Python 3.7 and activate it. Refer to this link for help on specifying the Python version in the virtualenv.
python3 -m pip install --user virtualenv
# You should have Python 3.7 available in your host. 
# Check the Python path using `which python3`
# Use a command similar to this one:
python3 -m virtualenv --python=<path-to-Python3.7> .devops
source .devops/bin/activate
  • Run make install to install the necessary dependencies

Running app.py

  1. Standalone: python app.py
  2. Run in Docker: ./run_docker.sh
  3. Run in Kubernetes: ./run_kubernetes.sh

Kubernetes Steps

  • Setup and Configure Docker locally
  • Setup and Configure Kubernetes locally
  • Create Flask app in Container
  • Run via kubectl =======

Cloud DevOps ND - C4- Microservices at Scale using AWS & Kubernetes - Supporting Material and Project Starter

This repository is associated with Cloud DevOps ND - Course 04 - Microservices at Scale using AWS & Kubernetes. In here, you'll find:

  1. Supporting material used in the video demonstration in the course
  2. Starting code for a project, in which you can containerize and deploy a machine learning srevice using Kubernetes.

A. Dependencies

A.1. Python

Download and install the python.

A.2. Docker Desktop

You would require you to install Docker Desktop to create containers for individual microservices. Refer the following links for instructions

A.3. Kubernetes

You would need to install any one tool for creating a Kubernetes cluster - KubeOne / Minikube / kubectl on top of Docker Desktop:

  1. Install and Set Up kubectl directly on top of Docker desktop - For Windows/macOS
  2. Install Minikube - For Linux/macOS

A.4. AWS account to access AWS Lambda

You'll need an AWS account to get started with AWS Lambda, which is a serverless computing platform on cloud.

A.5. An account with Circle CI

You may sign up on CircleCI.com with your GitHub credentials.


B. The Overarching Diagram

Overview


C. Tutorials

C.1. AWS Lambda & Serverless

  • Making Change: Create and deploy a serverless lambda function that responds to an input request; this example creates the correct amount of change to make up a value in US dollars.
  • Wikipedia Query: Deploy a lambda function that responds to an input, wikipedia page query; this example returns the first sentence of a specific wikipedia page upon being queried.

D. Project Instructions

To run any project code, you'll have to set up a virtual environment with the project dependencies. All of the following instructions are to be completed via a terminal/command line prompt.

E. Create and Activate an Environment

E.1. Git and version control

These instructions also assume you have git installed for working with Github from a terminal window, but if you do not, you can download that first from this Github installation page.

Now, you're ready to create your local environment!

  1. If you haven't already done so, clone the project repository, and navigate to the main project folder.
git clone https://github.com/udacity/DevOps_Microservices.git
cd DevOps_Microservices/project-ml-microservice-kubernetes
  1. Create (and activate) a new environment, named .devops with Python 3. If prompted to proceed with the install (Proceed [y]/n) type y.
python3 -m venv ~/.devops
source ~/.devops/bin/activate

At this point your command line should look something like: (.devops) <User>:project-ml-microservice-kubernetes<user>$. The (.devops) indicates that your environment has been activated, and you can proceed with further package installations.

  1. Installing dependencies via project Makefile. Many of the project dependencies are listed in the file requirements.txt; these can be installed using pip commands in the provided Makefile. While in your project directory, type the following command to install these dependencies.
make install

Now most of the .devops libraries are available to you. There are a couple of other libraries that we'll be using, which can be downloaded as specified, below.


E.2. Other Libraries

While you still have your .devops environment activated, you will still need to install:

  • Docker
  • Hadolint
  • Kubernetes (Minikube if you want to run Kubernetes locally)

E.3. Docker

You will need to use Docker to build and upload a containerized application. If you already have this installed and created a docker account, you may skip this step.

  1. You’ll need to create a free docker account, where you’ll choose a unique username and link your email to a docker account. Your username is your unique docker ID.

  2. To install the latest version of docker, choose the Community Edition (CE) for your operating system, on docker’s installation site. It is also recommended that you install the latest, stable release:

  3. After installation, you can verify that you’ve successfully installed docker by printing its version in your terminal: docker --version

E.4. Run Lint Checks

This project also must pass two lint checks; hadolint checks the Dockerfile for errors and pylint checks the app.py source code for errors.

  1. Install hadolint following the instructions, on hadolint's page:

For Mac:

brew install hadolint

For Windows:

scoop install hadolint
  1. In your terminal, type: make lint to run lint checks on the project code. If you haven’t changed any code, all requirements should be satisfied, and you should see a printed statement that rates your code (and prints out any additional comments):
------------------------------------
Your code has been rated at 10.00/10

That's about it! When working with kubernetes, you may need to install some other libraries, but these instructions will set you up with an environment that can build and deploy Docker containers.

4f1e50422012863102d588991b52319a94e1a53f

About

Operationalization of a machine learning microservice API using Docker and Kubernetes

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published