Skip to content
/ Deep3D Public

Real-Time end-to-end 2D-to-3D Video Conversion, based on deep learning.

License

Notifications You must be signed in to change notification settings

HypoX64/Deep3D

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Deep3D

Real-Time end-to-end 2D-to-3D Video Conversion, based on deep learning.
Inspired by piiswrong/deep3d, we rebuild the network on pytorch and optimize it in time domain and faster inference speed. So, try it and enjoy your own 3D movies.


Left is input video and right is output video with parallax.

More examples:

Inference speed

Plan 360p (FPS) 720p (FPS) 1080p (FPS) 4k (FPS)
GPU (2080ti) 84 87 77 26
CPU (Xeon Platinum 8260) 27.7 14.1 7.2 2.0

Run Deep3D

Prerequisites

Dependencies

This code depends on opencv-python available via pip install.

pip install opencv-python

Clone this repo

git clone https://github.com/HypoX64/Deep3D
cd Deep3D

Get Pre-Trained Models

You can download pre_trained models from: [Google Drive] [百度云,提取码xxo0 ]
Note:

  • 360p can get the best result.
  • The published models are not inference optimized.
  • Models are still under training, 1080p and 4k models will be uploaded in the future.

Run it!

python inference.py --model ./export/deep3d_v1.0_640x360_cuda.pt --video ./medias/wood.mp4 --out ./result/wood.mp4 --inv
# some video need to reverse left and right views (--inv)

Acknowledgements

This code borrows heavily from [deep3d] [DeepMosaics]

About

Real-Time end-to-end 2D-to-3D Video Conversion, based on deep learning.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages