-
Notifications
You must be signed in to change notification settings - Fork 642
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to improve image fidelity? #390
Comments
You could try to use vqgan 16k or even the f=8 ( see the details on their
page)
…On Thu, Nov 25, 2021, 01:21 Soon-Yau Cheong ***@***.***> wrote:
[image: brown]
<https://user-images.githubusercontent.com/19167278/143328331-6fb1f1a7-1094-478d-a0c5-615192a9b3e9.png>
I have trained on mannequin dataset and the results look quite good.
However, the generated images are a bit blurry and fine details are loss.
Therefore, I wonder what changes do I need to do make them look better.
I currently use VQGAN pretrained on imagenet. I have also tried to train a
VAE from scratch (using default train_vae.py) but it is blurry. I tried
increasing the number of layers, number of tokens etc but didn't see
improvement and made it a bit more unstable. Any advise on what VAE
parameters to change?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<#390>, or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAR437XBIS5MZ6L4P3HSRQLUNV6RRANCNFSM5IXI2X6Q>
.
|
@soon-yau Hi, I am trying to train a model using vqgan_gumbel_f8_8192 as you do, but I get terrible results. What model parameters did you use to get this result, especially the number of layers and dimensions? Also, could you please provide a link to the "mannequin dataset"? Thanks |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
(Left:generated. Right: from dataset)
I have trained on mannequin dataset and the results look quite good. However, the generated images are a bit blurry and fine details are lost. Therefore, I wonder what changes do I need to do make them look crispier.
I currently use VQGAN pretrained on imagenet. I have also tried to train a VAE from scratch (using default train_vae.py) but it is blurry. I tried increasing the number of layers, number of tokens etc but didn't see improvement and made it a bit more unstable. Any advise on what VAE parameters to change?
The text was updated successfully, but these errors were encountered: