There are 6 repositories under visual-speech-recognition topic.
Visual Speech Recognition for Multiple Languages
"LipNet: End-to-End Sentence-level Lipreading" in PyTorch
Python toolkit for Visual Speech Recognition
[ICASSP 2025] Official Pytorch implementation of "Large Language Models are Strong Audio-Visual Speech Recognition Learners".
Visual speech recognition with face inputs: code and models for F&G 2020 paper "Can We Read Speech Beyond the Lips? Rethinking RoI Selection for Deep Visual Speech Recognition"
Deep Visual Speech Recognition in arabic words
Official source code for the paper "Tailored Design of Audio-Visual Speech Recognition Models using Branchformers"
Visual Speech Recognition using deep learing methods
Implementation of "Combining Residual Networks with LSTMs for Lipreading" in Keras and Tensorflow2.0
EMOLIPS: TWO-LEVEL APPROACH FOR LIP-READING EMOTIONAL SPEECH
Speaker-Independent Speech Recognition using Visual Features
Visual Speech Recognition for Spanish
LipReadingITA: Keras implementation of the method described in the paper 'LipNet: End-to-End Sentence-level Lipreading'. Research project for University of Salerno.
In this repository, I try to use k2, icefall and Lhotse for lip reading. I will modify it for the lip reading task. Many different lip-reading datasets should be added. -_-
Deep Visual Speech Recognition in arabic words
Online Knowledge Distillation using LipNet and an Italian dataset. Master's Thesis Project.
Strong Gateway using Speech Processing ,3D Vision and Language processing . Deployed using Django