Skip to content

Determining the Audio-Video Synchronization between Mouth Motion and Speech in the Video.

License

Notifications You must be signed in to change notification settings

x0axz/Synthetic-Speech-Detection-in-Video

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Synthetic Speech Detection in Video using Deep Learning

Download lipsync_v4_73.mat and shape_predictor_68_face_landmarks.dat

This Project is forked from this Repository
Follow this Blog for step by step explanation.

Research paper refered throughout this project: https://www.robots.ox.ac.uk/~vgg/publications/2016/Chung16a/chung16a.pdf
They focused on determining the audio-video synchronization between mouth motion and speech in the video. They used audio-video synchronization for TV broadcasting. It's really a nice research paper, they developed a language-independent and speaker-independent solution to the lip-sync problem, without labeled data.

For the modeling and processing functions: https://github.com/voletiv/syncnet-in-keras
VidTIMIT dataset used in this project: http://conradsanderson.id.au/vidtimit

License

MIT

About

Determining the Audio-Video Synchronization between Mouth Motion and Speech in the Video.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published