FandosA / Face_Tracking_Real_Time_PyTorch

Implementation of a deep learning model based on the VGG16 neural network to track faces in a video or with a camera in real time.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Face Tracking In Real Time

Implementation of a deep learning model based on the VGG16 neural network to track faces on video or with a camera in real time. This proyect has been done following this tutorial, in which Nicholas Renotte implemented a model using the Tensorflow framework, and I took it to deploy and port it to PyTorch. To run the implementation, download this repository and run the setup.py script. It will create the necessary folders to prepare the dataset.

IMPORTANT: The model is designed to detect and track faces in images/videos with a single face, so, in your dataset, put images with only one face. Also, the dataset I used was made with images in landscape orientation (e.g. 1280x720, 640x360). Finally, make sure your images are in jpg format.

Dataset preparation

To prepare the dataset you can use your own images (as I did) or you can connect a camera to the PC and run the capture_images_camera.py script to capture images. This script will capture and store the desired number of images (by default 100). If you choose the first option just put the images in the data/images folder, otherwise try to take images from different points of view moving your head and face. The script is programmed so that images are taken every 0.5 seconds, and they will be stored automatically in the folder. To set the number of images to take, you can change the default value in the script or open a terminal and run the next command:

python capture_images_camera.py --images_to_take=<number of images>

Use images with and without faces on them. Once you have your images, run the resize_images.py script to make them be the same size. By default the output width and height are set to 640 and 360 respectively. In case you want to change them (not recommended) run the next command:

python resize_images.py --width=<width> --height=<height>

The resized images will be stored in the data/images_resized folder.

The next step is to label those images. For that, the tool Labelme will be used. Follow the tutorial mentioned above, starting at 20:05, to install the tool and label the images. Just note that when you click the Open Dir option in the tool, the input directory in our case is data/images_resized, and the output directory for the labels is data/labels.

If the labeling process was successful, running visualize_dataset.py should display the images and the bounding boxes around the faces in the images.

Data augmentation

To do the data augmentation, just run the data_augmentation.py script. For each image, using the library albumentations, one hundred images will be created taking subimages from the original image changing properties such as rotation, orientation, color, brightness... The resolution of these new images, i.e. their size, is set by default to 256x144. It can be changed by running the following command, but keep in mind that if this size is modified, the neural network and test files will also have to be modified, so it is not recommended. The number of subimages to take can be modified too (see the command).

python data_augmentation.py --width=<width> --height=<height> --num_subimages=<number of subimages>

The labels for those new images will be created automatically, setting the label to 1 or 0 depending on whether there is a face in the images or not, respectively, and converting the coordinates of the bounding boxes to the corresponding values. Images will be stored in the aug_data/images folder and labels will be stored in the aug_data/labels folder. To check that the data augmentation has been succesful, running visualize_augmented_dataset.py should display the new images with the bounding boxes around the faces.

Train the model

To train the model, the convolutional neural network VGG16 has been used. If you followed the steps correctly, it should be enough to run the train.py script to start the training. The hyperparameters can be modified both setting them in the script or in the command line, and they are:

  • dataset_path: path to the fodler that contains the images and the labels. In our case 'aug_data'.
  • train_split: percentage of the dataset to use for training.
  • log_dir: name of the folder that will be created to store the checkpoints and the loss values of the current training.
  • batch_size: batch size.
  • learning_rate: learning rate.
  • epochs: number of epochs.
  • device: device to run the training ('gpu' or 'cpu').

As an example, I uploaded the folder where the checkpoints of the model I trained were stored, as well as the loss and accuracy values along with their plots. Control points are saved every 10 epochs, and the best control point (the control point with the lowest loss validation value) is also saved.

IMPORTANT: the files inside the checkpoints/ folder are not actual checkpoints. They are empty files created to show the appearance of the folder that will be created when starting the training. On the other hand, the plot and the accuracy and loss values are the actual results that I got from training my model.

Test the model

Finally, you can test the model by running any of the test files: test_images.py, test_real_time_camera.py or test_real_time_video.py.

test_images.py

With this file it is possible to test the model with other images. The parameters to set here are the path to the folder containing the images we want to use to test the model, the name of the folder set in the train.py script to store the checkpoints, the checkpoint you want to use to test the model, the device to run the model ('gpu' or 'cpu') and the accuracy threshold to determine if there is a face in the test image or not and thus draw the bounding box if so. Images will be displayed continuously until finished.

test_real_time_camera.py

With this file it is possible to test the model with the camera. The parameters to set here are the name of the folder set in the train.py script to store the checkpoints, the checkpoint you want to use to test the model, the device to run the model ('gpu' or 'cpu') and the accuracy threshold to determine if there is a face in the test image or not and thus draw the bounding box if so. The images captured by the camera will be displayed in real time drawing the bounding box when a face is detected.

test_real_time_video.py

With this file it is possible to test the model with a video. The parameters to set here are the path to the folder containing the images we want to use to test the model, the path to the video, the name of the folder set in the train.py script to store the checkpoints, the checkpoint you want to use to test the model, the device to run the model ('gpu' or 'cpu') and the accuracy threshold to determine if there is a face in the test image or not and thus draw the bounding box if so. The video will be displayed in real time drawing the bounding box when a face is detected.

Results

In this video I show the results testing the model after training it with 32,200 images, 222 with a face and 100 without a face. 322 images that became 32,200 after data augmentation.

I hope you found this repository interesting and helpful in understanding how to track a face (or any other object, depending on the images you use to train the model) in images, video, or the camera.

Video: https://www.youtube.com/watch?v=m9huHfmKfKo&t=9s

About

Implementation of a deep learning model based on the VGG16 neural network to track faces in a video or with a camera in real time.


Languages

Language:Python 100.0%