Skip to content

Sign Language Recognition using Machine and Deep Learning

Notifications You must be signed in to change notification settings

utsavk28/GesturalAI-Nerds

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

59 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GesturalAI

Full Demo

📌 About

Sign Language Recognition using Machine and Deep Learning

🎯 Key Features

  • Classifies the American Sign Language to english letters in real time by taking video input using webcam

💻 Development Process

Various Techniques/Models Used

  • I3D Transfer Learning : We worked on classifying most common words in English from sign language. For this, we used the WLASL dataset and fine tuned the I3D model which was pretrained on the Kinetics-400 dataset
  • CNN + RNN with self made dataset : We Created our own small dataset for 10 common words. Each word had 30 videos of 30 frames. Here we extracted the landmarks of pose, left hand and right hand using mediapipe holistic model. And used those landmarks of the frames for training the RNN model. This Model works accurately for the person who trained but the accuracy drops for the other persons
  • Transfer Learning with Resnet & Mobilnet : Trained on ASL dataset (consisting of sign languages of A-Z letters and some special characters like SPACE, DELETE and NOTHING). Predicts accurately on the training, validation and testing dataset but the accuracy drops in production
  • MediaPipe Feature Extraction + Machine Learning Algorithms : Using MediaPipe's Hand Model , features are extracted and engineered. Those are fed to machine learning model and are used to predict the labels

Datasets used

  • WLASL video dataset

    • WLASL is the largest video dataset for Word-Level American Sign Language (ASL) recognition
    • This dataset consists of around 20K videos for sign language representation of around 2000 commonly used words
  • American Sign Language Dataset for letters

    • Image data set for alphabets in the American Sign Language
    • Contains 87000 images of 200 X 200 pixels, divided into 29 classes (A - Z, SPACE, DELETE and NOTHING)
    • The test data set contains only 29 images, 1 image per labels.
  • American Sign Language Dataset for letters

    • RGB image dataset of American sign language alphabets.
    • This dataset contains 166k images of 200 X 200 pixels, divided into 28 classes (A - Z, SPACE and NOTHING)
    • The test data set contains 112 images , 4 images per labels.

Notebooks

Notebook's Folder Link

🛠 Project Setup

  1. Clone the repository using the git clone
 $ git clone https://github.com/utsavk28/Nerds.git
  1. Create a virtual environment
 $ virtualenv venv
 $ source venv/bin/activate
  1. Install the required packages
 $ pip install -r requirements.txt
  1. Run the app
 $ python main.py

Group members

ASL characters and their sign representations

ASL characters

📸 Results

Demo link

About

Sign Language Recognition using Machine and Deep Learning

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published