Visual Speech Recognition for Multiple Languages
-
Updated
Aug 17, 2023 - Python
Visual Speech Recognition for Multiple Languages
Auto-AVSR: Lip-Reading Sentences Project
A PyTorch implementation of the Deep Audio-Visual Speech Recognition paper.
"LipNet: End-to-End Sentence-level Lipreading" in PyTorch
Python toolkit for Visual Speech Recognition
Visual speech recognition with face inputs: code and models for F&G 2020 paper "Can We Read Speech Beyond the Lips? Rethinking RoI Selection for Deep Visual Speech Recognition"
Deep Visual Speech Recognition in arabic words
[ICASSP 2025] Official Pytorch implementation of "Large Language Models are Strong Audio-Visual Speech Recognition Learners".
Official source code for the paper "Tailored Design of Audio-Visual Speech Recognition Models using Branchformers"
Visual Speech Recognition using deep learing methods
Implementation of "Combining Residual Networks with LSTMs for Lipreading" in Keras and Tensorflow2.0
Speaker-Independent Speech Recognition using Visual Features
EMOLIPS: TWO-LEVEL APPROACH FOR LIP-READING EMOTIONAL SPEECH
Visual Speech Recognition for Spanish
Deep Visual Speech Recognition in arabic words
LipReadingITA: Keras implementation of the method described in the paper 'LipNet: End-to-End Sentence-level Lipreading'. Research project for University of Salerno.
A single-speaker reimplementation of the LipNet architecture for lip reading.
In this repository, I try to use k2, icefall and Lhotse for lip reading. I will modify it for the lip reading task. Many different lip-reading datasets should be added. -_-
Add a description, image, and links to the visual-speech-recognition topic page so that developers can more easily learn about it.
To associate your repository with the visual-speech-recognition topic, visit your repo's landing page and select "manage topics."