utsavk28 / GesturalAI-Nerds

Sign Language Recognition using Machine and Deep Learning

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

GesturalAI

Full Demo

πŸ“Œ About

Sign Language Recognition using Machine and Deep Learning

🎯 Key Features

  • Classifies the American Sign Language to english letters in real time by taking video input using webcam

πŸ’» Development Process

Various Techniques/Models Used

  • I3D Transfer Learning : We worked on classifying most common words in English from sign language. For this, we used the WLASL dataset and fine tuned the I3D model which was pretrained on the Kinetics-400 dataset
  • CNN + RNN with self made dataset : We Created our own small dataset for 10 common words. Each word had 30 videos of 30 frames. Here we extracted the landmarks of pose, left hand and right hand using mediapipe holistic model. And used those landmarks of the frames for training the RNN model. This Model works accurately for the person who trained but the accuracy drops for the other persons
  • Transfer Learning with Resnet & Mobilnet : Trained on ASL dataset (consisting of sign languages of A-Z letters and some special characters like SPACE, DELETE and NOTHING). Predicts accurately on the training, validation and testing dataset but the accuracy drops in production
  • MediaPipe Feature Extraction + Machine Learning Algorithms : Using MediaPipe's Hand Model , features are extracted and engineered. Those are fed to machine learning model and are used to predict the labels

Datasets used

  • WLASL video dataset

    • WLASL is the largest video dataset for Word-Level American Sign Language (ASL) recognition
    • This dataset consists of around 20K videos for sign language representation of around 2000 commonly used words
  • American Sign Language Dataset for letters

    • Image data set for alphabets in the American Sign Language
    • Contains 87000 images of 200 X 200 pixels, divided into 29 classes (A - Z, SPACE, DELETE and NOTHING)
    • The test data set contains only 29 images, 1 image per labels.
  • American Sign Language Dataset for letters

    • RGB image dataset of American sign language alphabets.
    • This dataset contains 166k images of 200 X 200 pixels, divided into 28 classes (A - Z, SPACE and NOTHING)
    • The test data set contains 112 images , 4 images per labels.

Notebooks

Notebook's Folder Link

πŸ›  Project Setup

  1. Clone the repository using the git clone
 $ git clone https://github.com/utsavk28/Nerds.git
  1. Create a virtual environment
 $ virtualenv venv
 $ source venv/bin/activate
  1. Install the required packages
 $ pip install -r requirements.txt
  1. Run the app
 $ python main.py

Group members

ASL characters and their sign representations

ASL characters

πŸ“Έ Results

Demo link

About

Sign Language Recognition using Machine and Deep Learning


Languages

Language:Python 95.8%Language:CSS 2.2%Language:HTML 2.0%