From 0e9f813a940efe5ec5b56d2bb41395dccc6e8b63 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ignacio=20Hern=C3=A1ndez=20Montilla?= Date: Fri, 10 Mar 2023 17:40:06 +0100 Subject: [PATCH] Minor updates --- Readme.md | 2 +- detect.py | 7 +++---- prepare_data.py | 2 ++ 3 files changed, 6 insertions(+), 5 deletions(-) diff --git a/Readme.md b/Readme.md index bfab297..f2667fa 100644 --- a/Readme.md +++ b/Readme.md @@ -15,7 +15,7 @@ For this experiment I'm using the following sources: - [Menpo2D dataset](https://github.com/jiankangdeng/MenpoBenchmark) - [AFW (Annotated Faces in the Wild)](https://ibug.doc.ic.ac.uk/resources/facial-point-annotations/) - Custom datasets: - - [Pexels](https://pexels.com): I downloaded 85 images from this website and annotated them using [CVAT](https://app.cvat.ai/). + - [Pexels](https://pexels.com): I downloaded 85 images from this website and annotated them using [CVAT](https://app.cvat.ai/). I am currently working on another batch of images. ## Results diff --git a/detect.py b/detect.py index 5e2b9d4..ffbbc54 100644 --- a/detect.py +++ b/detect.py @@ -12,11 +12,11 @@ model = YOLO(os.path.join(path_model, "best.pt")) # This will draw the detections - class_colors = spv.ColorPalette.from_hex(['#990099', '#ff66ff', '#00cc00', '#cc0000']) + class_colors = spv.ColorPalette.from_hex(['#ffff66', '#66ffcc', '#ff99ff', '#ffcc99']) box_annotator = spv.BoxAnnotator( thickness=2, - text_thickness=2, - text_scale=1, + text_thickness=1, + text_scale=0.5, color=class_colors ) @@ -45,4 +45,3 @@ if k == ord("q"): break - diff --git a/prepare_data.py b/prepare_data.py index 832c491..6dcee1a 100644 --- a/prepare_data.py +++ b/prepare_data.py @@ -154,6 +154,8 @@ def process_names(names, split, path_data, path_dest, skip): # PART 2: PROCESSING THE PEXELS IMAGES # ######################################## + # TODO: Add the images of the new batch + # Copying the images and the labels to the final folder pexels_labels = os.listdir(path_pexels_annotations) for l in pexels_labels: