Skip to content

🕵️‍♂️ A custom model was created using TensorFlow 2 on a novel dataset. Dataset consisted of 2,400 images and had an accuracy of 85%.

Notifications You must be signed in to change notification settings

Purefekt/Custom-Object-Detection-with-TensorFlow-2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

81 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Custom Object Detection with TensorFlow-2

These are the steps I took to train an Object Detection model to detect my face. I used TensorFlow 2 Object Detection API on Windows 10. I used a virtual machine on Microsoft Azure. My VM was Standard DS2 v2, with 2 virtual CPUs and 7 GB of memory. This repository acts as a guide for others and also as a reference for myself to look back at.

In this project I will use the SSD MobileNet V2 FPNLite 640x640 pretrained model as it runs on low power devices like the raspberry pi and does not have too much of an accuracy hit. Others models with higher accuracy are better for projects where we use more powerful hardware, since I want to port this model to a raspberry pi 4, I stuck with this model. Other models

Softwares needed

Installing TensorFlow CPU

Open the anaconda command prompt. Create a virtual environment

conda create -n tensorflow pip python=3.8

Then activate the environment with

conda activate tensorflow

NOTE: The virtual environment has to be activated each time the anaconda terminal is closed.
Installing TensorFlow CPU

pip install tensorflow

Sanity Check

python
>>> import tensorflow as tf
>>> print(tf.__version__)
>>> exit()

A version number should be displayed.

Preparing the Workspace

Create a folder directly in C: and name it "TensorFlow". It can be created anywhere but the commands need to be changed accordingly.

cd C:\TensorFlow

Clone the TensorFlow models repository with

git clone https://github.com/tensorflow/models.git

This should clone all the files in a directory called models. After this stay inside C:\TensorFlow and download this repository into a .zip file. Then extract the three files, workspace, scripts and other scripts, directly in to the TensorFlow directory.

Installing the prequisites for the Object Detection API.

conda install -c anaconda protobuf
cd models\research
protoc object_detection\protos\*.proto --python_out=.

Close the terminal and open a new Anaconda prompt. Use the following command to activate it

conda activate tensorflow
pip install git+https://github.com/philferriere/cocoapi.git#subdirectory=PythonAPI
pip install cython
cd C:\TensorFlow\models\research
copy object_detection\packages\tf2\setup.py .
python -m pip install .

If there are no errors, run this command

python object_detection\builders\model_builder_tf2_test.py

Output should be similar to this

[       OK ] ModelBuilderTF2Test.test_create_ssd_models_from_config
[ RUN      ] ModelBuilderTF2Test.test_invalid_faster_rcnn_batchnorm_update
[       OK ] ModelBuilderTF2Test.test_invalid_faster_rcnn_batchnorm_update
[ RUN      ] ModelBuilderTF2Test.test_invalid_first_stage_nms_iou_threshold
[       OK ] ModelBuilderTF2Test.test_invalid_first_stage_nms_iou_threshold
[ RUN      ] ModelBuilderTF2Test.test_invalid_model_config_proto
[       OK ] ModelBuilderTF2Test.test_invalid_model_config_proto
[ RUN      ] ModelBuilderTF2Test.test_invalid_second_stage_batch_size
[       OK ] ModelBuilderTF2Test.test_invalid_second_stage_batch_size
[ RUN      ] ModelBuilderTF2Test.test_session
[  SKIPPED ] ModelBuilderTF2Test.test_session
[ RUN      ] ModelBuilderTF2Test.test_unknown_faster_rcnn_feature_extractor
[       OK ] ModelBuilderTF2Test.test_unknown_faster_rcnn_feature_extractor
[ RUN      ] ModelBuilderTF2Test.test_unknown_meta_architecture
[       OK ] ModelBuilderTF2Test.test_unknown_meta_architecture
[ RUN      ] ModelBuilderTF2Test.test_unknown_ssd_feature_extractor
[       OK ] ModelBuilderTF2Test.test_unknown_ssd_feature_extractor
----------------------------------------------------------------------
Ran 20 tests in 45.304s

OK (skipped=1)

Gathering and Labeling the Dataset [Inside Tensorflow/workspace/training_demo]

  • annotations: This is where we will store all our training data needed for our model, the CSV and RECORD files needed for the training pipeline. There is also a PBTXT File with the labels for my custom model.
  • exported-models: This is the folder where the exported inference graph is stored.
  • images: This folder consists of a test and train folder. Here the dataset images are stored with their label files (XML format).
  • models: In this folder we will store our training pipeline and checkpoint information from the training job as well as the CONFIG file needed for training.
  • pre-trained-models: Here we will store the pre-trained model that we will use as a starting checkpoint for training
  • The rest of the scripts are just used for training and exporting the model.
  • I clicked 1,400 images of myself (700 with glasses and 700 without glasses), in different outfits and different angles. I used the name_changer.py script to change the name of the images in an ordered manner to keep a track of the data.
  • Then I used the downscale_to_720p.py script to downscale all the 4k images to 720p. This converted all images from roughly 2-3 MB each to just 60-70 KB each.
  • Then I used mirror_image.py script to get a mirror image of all 1,400 (720p) images. This made my dataset twice as big. Now I had a total of 2,800 images.
  • Finally I used the RandomNames.bat to rename all images to random names and then used the name_changer.py script again on this folder. This renamed the randomised names from 1-2,800.
  • After performing these steps I had 2,800; 720p images in a random order with names from 1-2,800. I copied 20% (560) images to the test folder and 80% (2,240) images to the train folder.
  • To label the dataset I used LabelImg directly on the test and train images and labeled each image.

Generating Training Data

  • Now that the images and XML files are ready, I created the label_map. It is located in Tensorflow/workspace/training_demo/annotations/. Since this model only detects my face, the number of classes is just one, thus it looks like this.
item {
    id: 1
    name: 'Veer'
}
  • Now I generated the RECORD files for training.
pip install pandas

Navigate to the scripts\preprocessing directory

cd C:\TensorFlow\scripts\preprocessing

Running these 2 commands will generate the RECORD files.

python generate_tfrecord.py -x C:\Tensorflow\workspace\training_demo\images\train -l C:\Tensorflow\workspace\training_demo\annotations\label_map.pbtxt -o C:\Tensorflow\workspace\training_demo\annotations\train.record

python generate_tfrecord.py -x C:\Tensorflow\workspace\training_demo\images\test -l C:\Tensorflow\workspace\training_demo\annotations\label_map.pbtxt -o C:\Tensorflow\workspace\training_demo\annotations\test.record
  • Now under Tensorflow/workspace/training_demo/annotations/ there should be a test.record and train.record.

Configuring the Training Pipeline

  • I used the CONFIG File from one of the TensorFlow pre-trained models. There are plenty of models in the TensorFlow Model Zoo, but I used the SSD MobileNet V2 FPNLite 640x640[if download does not start after clicking on link, try right click -> open in a new tab]. After downloading I used 7zip to extract the contents and copied it to the Tensorflow/workspace/training_demo/pre-trained-models directory [make sure to copy the entire folder directly into the pre-trained models folder, dont copy the 3 files inside the extracted file individually]
  • To store the training pipeline I created a directory called my_ssd_mobilenet_v2_fpnlite in the Tensorflow/workspace/training_demo/models directory. Then copied the pipeline.config from the Tensorflow/workspace/training_demo/pre-trained-models.
  • I made the following changes to the pipeline.config file
  • Line 3. Change num_classes to num_classes: 1
  • Line 135. Change batch_size according to available memory (Higher values require more memory and vice-versa). I changed it to: batch_size: 5
  • Line 165. Change fine_tune_checkpoint to: fine_tune_checkpoint: "pre-trained-models/ssd_mobilenet_v2_fpnlite_640x640_coco17_tpu-8/checkpoint/ckpt-0"
  • Line 171. Change fine_tune_checkpoint_type to: fine_tune_checkpoint_type: "detection"
  • Line 175. Change label_map_path to: label_map_path: "annotations/label_map.pbtxt"
  • Line 177. Change input_path to: input_path: "annotations/train.record"
  • Line 185. Change label_map_path to: label_map_path: "annotations/label_map.pbtxt"
  • Line 189. Change input_path to: input_path: "annotations/test.record"

Training the Model

Finally I was ready to start the training process. After opening a new anaconda terminal

conda activate tensorflow
cd C:\TensorFlow\workspace\training_demo
python model_main_tf2.py --model_dir=models\my_ssd_mobilenet_v2_fpnlite --pipeline_config_path=models\my_ssd_mobilenet_v2_fpnlite\pipeline.config

After a few warnings, there should be the first 100 step summary

INFO:tensorflow:Step 100 per-step time 16.640s loss=0.454
I0810 11:56:12.520163 11172 model_lib_v2.py:644] Step 100 per-step time 16.640s loss=0.454

I ran the model for ~20,700 steps, which took around 95 hours. Since I could only get the cpu only VM I was not able to take advantage of TensorFlow GPU which would've brought my training time significantly down. I was trying to get a loss below 0.150, as this prevents underfitting and overfitting. Tensorflow logs the loss every 100 steps and Ctrl+C can be used to pause the training.

Note: The testing can be paused by simply pressing ctrl+C and then keeping the files intact and simply restarting the testing from the anaconda terminal. But once the inference graph is exported the testing cannot be resumed. Make sure to keep a copy of the entire "Tensorflow" folder before exporting the inference graph to resume the testing if necessary or to convert model to TensorFlowLite.

Monitoring Training with TensorBoard

I used TensorFlow's Tensorboard to monitor the training. It is a very powerful tool which lets the user monitor training and visualize training metrics. To start TensorBoard, open a new anaconda terminal

conda activate tensorflow
cd C:\TensorFlow\workspace\training_demo
tensorboard --logdir=models\my_ssd_mobilenet_v2_fpnlite

Output:

Serving TensorBoard on localhost; to expose to the network, use a proxy or pass --bind_all
TensorBoard 2.2.2 at http://localhost:6006/ (Press CTRL+C to quit)

http://localhost:6006/ in the browser will open TensorBoard.

We can see that the total loss is decreasing over time, but it starts to plateau, after this point there won't be much increase in accuracy and we will start getting diminishing returns, thus stopping at this point is a good idea.

Exporting the Inference Graph

To export the saved model I performed the following steps

conda activate tensorflow
cd C:\TensorFlow\workspace\training_demo
python .\exporter_main_v2.py --input_type image_tensor --pipeline_config_path .\models\my_ssd_mobilenet_v2_fpnlite\pipeline.config --trained_checkpoint_dir .\models\my_ssd_mobilenet_v2_fpnlite\ --output_directory .\exported-models\my_mobilenet_model

The finished model is stored in C:\TensorFlow\workspace\training_demo\exported-models\my_mobilenet_model\saved_model folder. There is a PB File called saved_model.pb. This is the inference graph! I also prefer to copy the label_map.pbtxt file in to this directory because it makes things a bit easier for testing. The label_map.pbtxt file is located in C:\TensorFlow\workspace\training_demo\annotations\label_map.pbtxt.

Testing out the Finished Model

There are several scripts which can be used to detect the model in different ways.

  • TF-image-od.py: This program uses the viz_utils module to visualize labels and bounding boxes. It performs object detection on a single image, and displays it with a cv2 window.
  • TF-image-object-counting.py: This program also performs inference on a single image. I have added my own labelling method with OpenCV which I prefer. It also counts the number of detections and displays it in the top left corner. The final image is, again, displayed with a cv2 window.
  • TF-video-od.py: This program is similar to the TF-image-od.py. However, it performs inference on each individual frame of a video and displays it via cv2 window.
  • TF-video-object-counting.py: This program is similar to TF-image-object-counting.py and has a similar labelling method with OpenCV. Takes a video for input, and also performs object detection on each frame, displaying the detection count in the top left corner.
  • TF-webcam-opencv.py: This program opens the webcam and detects the object in real time. Pressing Q will quit the webcam window.

The usage of each program looks like

usage: TF-image-od.py [-h] [--model MODEL] [--labels LABELS] [--image IMAGE] [--threshold THRESHOLD]

optional arguments:
  -h, --help            show this help message and exit
  --model MODEL         Folder that the Saved Model is Located In
  --labels LABELS       Where the Labelmap is Located
  --image IMAGE         Name of the single image to perform detection on
  --threshold THRESHOLD Minimum confidence threshold for displaying detected objects

Need to install opencv-python

pip install opencv-python

To test the model

conda activate tensorflow
cd C:\TensorFlow\workspace\training_demo

Then to run the script, just use

python TF-webcam-opencv.py

Link to my YouTube video!

This method can be used to create many different computer vision applications based on novel databases. .

About

🕵️‍♂️ A custom model was created using TensorFlow 2 on a novel dataset. Dataset consisted of 2,400 images and had an accuracy of 85%.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published