prateekralhan / Hand-Sign-detection-using-TensorFlow-Object-Detection-API

Real-time Hand Sign detection using TensorFlow Object Detection API

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Real-time Hand Sign detection using TensorFlow Object Detection API πŸ‘ πŸ‘Ž πŸ‘ŒβœŒοΈπŸ‘ Project Status: Active

tensorflow-logo

An end-to-end implementation of training and leveraging our own custom object detection model using the tensorflow object detection API.

 

Installation and Usage:

Steps


Step 1. Clone this repository: https://github.com/prateekralhan/Hand-Sign-detection-using-TensorFlow-Object-Detection-API.git

Step 2. Create a new virtual environment
python -m venv env

Step 3. Activate your virtual environment
source tfod/bin/activate # Linux
.\tfod\Scripts\activate # Windows 

Step 4. Install dependencies and add virtual environment to the Python Kernel
python -m pip install --upgrade pip
pip install ipykernel
python -m ipykernel install --user --name=env

Step 5. Collect images using the Notebook 1. Image Collection.ipynb - ensure you change the kernel to the virtual environment as shown:

 
Step 6. Manually divide collected images into two folders train and test. So now all folders and annotations should be split between the two folders - train and test.


Step 7. Begin training process by opening 2. Training and Detection.ipynb, this notebook will walk you through installing Tensorflow Object Detection, making detections, saving and exporting your model.

Step 8. During this process the Notebook will install Tensorflow Object Detection. You should ideally receive a notification indicating that the API has installed successfully at Step 8 with the last line stating OK.
  If not, resolve installation errors by referring to the Error Guide.md in this folder.

Step 9. Once you get to step 6. Train the model, inside of the notebook, you may choose to train the model from within the notebook. I have noticed however that training inside of a separate terminal on a Windows machine you're able to display live loss metrics.  
Step 10. You can optionally evaluate your model inside of Tensorboard. Once the model has been trained and you have run the evaluation command under Step 7. Navigate to the evaluation folder for your trained model e.g.

 cd Tensorlfow/workspace/models/my_ssd_mobnet/eval

and open Tensorboard with the following command

tensorboard --logdir=. 

Tensorboard will be accessible through your browser and you will be able to see metrics including mAP - mean Average Precision, and Recall.

Training metrics:

 

Evaluation metrics:

   

Results

thumbsup thumbsdown thankyou livelong

 

About

Real-time Hand Sign detection using TensorFlow Object Detection API


Languages

Language:Jupyter Notebook 100.0%