Visual Speech Recognition using deep learing methods
-
Updated
Apr 7, 2024 - Python
Visual Speech Recognition using deep learing methods
Online Knowledge Distillation using LipNet and an Italian dataset. Master's Thesis Project.
EMOLIPS: TWO-LEVEL APPROACH FOR LIP-READING EMOTIONAL SPEECH
In this repository, I try to use k2, icefall and Lhotse for lip reading. I will modify it for the lip reading task. Many different lip-reading datasets should be added. -_-
Implementation of "Combining Residual Networks with LSTMs for Lipreading" in Keras and Tensorflow2.0
Speaker-Independent Speech Recognition using Visual Features
Visual speech recognition with face inputs: code and models for F&G 2020 paper "Can We Read Speech Beyond the Lips? Rethinking RoI Selection for Deep Visual Speech Recognition"
Python toolkit for Visual Speech Recognition
"LipNet: End-to-End Sentence-level Lipreading" in PyTorch
Auto-AVSR: Lip-Reading Sentences Project
A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.
Visual Speech Recognition for Multiple Languages
Add a description, image, and links to the visual-speech-recognition topic page so that developers can more easily learn about it.
To associate your repository with the visual-speech-recognition topic, visit your repo's landing page and select "manage topics."