Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can we use this for realtime generation? #11

Closed
muhammadumair894 opened this issue May 27, 2024 · 2 comments
Closed

Can we use this for realtime generation? #11

muhammadumair894 opened this issue May 27, 2024 · 2 comments

Comments

@muhammadumair894
Copy link

Training and Inference Hardware.
Specifically, the GPU with 8G
VRAM can generate up to 3 minutes of video in one inference.

Hi there,

First of all, splendid work! I really appreciate the generalizability of this model and am looking forward to the code release.

I have a question regarding the paper's mention that a single 8GB GPU can produce a 3-minute video. Is it possible to make this process real-time or to stream the generated video so it appears as a real-time conversation?

@muhammadumair894
Copy link
Author

How much time one inference takes for 3 min generation?

@liutaocode
Copy link
Collaborator

Currently, motion generation is a long-term process (taking several seconds to several minutes). It cannot achieve real-time performance like vasa1. If the segments are too short, there will be a lack of smoothness between segments.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants