-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reconstruction results #20
Comments
Can you show me your code for reconstruction?I also meet this problem that reconstruction results of the ViT-VQGAN on ImageNet are very terrible. |
config = OmegaConf.load('configs/imagenet_vitvq_small.yaml') def preprocess(img):
original=Image.open('/home/marcelo/Downloads/212861459-e4113b34-622d-4602-afe4-f20e2d79425c.png') quant, _ = model.encode(image) |
Actually, I think the reason is the bad model checkpoint. Your script is
right. I measure the rFID, it is far away from VQGAN. I also train the
model on ImageNet, but it still works badly.
From: ***@***.***>
Date: Sat, Aug 12, 2023, 03:04
Subject: [External] Re: [thuanz123/enhancing-transformers] Reconstruction
results (Issue #20)
To: "thuanz123/enhancing-transformers"<
***@***.***>
Cc: ***@***.***>, "Comment"<
***@***.***>
The same as the one in the colab notebook
—
Reply to this email directly, view it on GitHub
<#20 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/A7ZMO7CSSW4EOBN4SOSLJVDXUZ63PANCNFSM6AAAAAA3L3JDYQ>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
So, after your training, you obtain a better model weights that improve the reconstruction? |
Hi, First of all thanks for you work.
Working with vit small, I see that results are far away from VQGAN, did you stop training when reached convergence? Do you think there is more room to improve the model performance/
Results with vit-small
input image
The text was updated successfully, but these errors were encountered: