Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

question about <mnist_minimal_example> #2

Closed
Johnson-yue opened this issue Jul 11, 2019 · 6 comments
Closed

question about <mnist_minimal_example> #2

Johnson-yue opened this issue Jul 11, 2019 · 6 comments

Comments

@Johnson-yue
Copy link
Contributor

Johnson-yue commented Jul 11, 2019

Hi, , thank you for your sharing , I am learning your method INN. and I can not understand the demo(mnist_minimal_example) loss ? what is the meaning?? I have ever seen before.
image

@Johnson-yue
Copy link
Contributor Author

Johnson-yue commented Jul 12, 2019

Hi, I know this nll loss compute is from Eq (6) in paper ,but where is L2 regularization?

And Question 2:
I run your experiments/mnist_minimal_example/train.py, and then run eval.py .Both run with default configure and plot those pictures...
Screenshot from 2019-07-12 14-16-18
I think the result is very bad . and how can I reimplement your experiments like your paper?

Validation loss:
-3.348240375518799

@ardizzone
Copy link
Member

ardizzone commented Jul 12, 2019

Hi, thank you for bringing this to our attention!

Concerning the L2 regularization: this is included in the pytorch optimizer, so we do not have to implement it by hand.

About the quality of results: You are right, they do not look good. I will investigate, and get back to you in a few days!

@ardizzone
Copy link
Member

ardizzone commented Jul 12, 2019

I could reproduce the issue on my end, and I also found the problem:
The amplitude for the noise augmentation was a factor of 10 off.
I just pushed commit e6bc8fd to fix it.
I also tweaked some other hyper-parameters, it now trains in half the time (15 minutes, 60 epochs).

Thank you again for bringing our attention to the issue!

If you want even better results, you can also increase the number of coupling blocks to 24, and train for 480 epochs (settings used in the paper, but takes several hours).

2019-07-12_16-45-00

I will close the issue, but feel free to reopen if there are further problems.

@Johnson-yue
Copy link
Contributor Author

@ardizzone Hi, I also test for style transfer experiments like Fig.(7), so I modify the eval.py into style_transfer.py and result like this:
source-vs-rec

I think in this configuration the cINN could not decompose into 10 essentially separate subnetworks.. because in this cause, some label is not generate image

@ardizzone
Copy link
Member

ardizzone commented Jul 16, 2019

Hi,

very interesting! For most of your samples it works fine, but as you said, some don't work at all.
This is just and issue of the training (and the model).

I just pushed the actual research code for the mnist task, used to produce the paper (some documentation and comments were still missing): 8ea0165

I also included a model checkpoint (see the readme in the experiments/mnist_cINN folder).

The model from the paper is slightly larger and trained for longer, you can test the style transfer by running eval.py in the experiments/mnist_cINN folder.

I am confident it can be made much smaller and more efficient using convolutional layers, instead of all fully connected, I may update in the future (or you can experiment yourself).

@Johnson-yue
Copy link
Contributor Author

@ardizzone Ok,thank you very much . when I test mnist experiments, I will check convolutional layers experiments like : colorization_minimal_experiment and colorization_cINN.

Thanks for your sharing again. and I will learn more for this.

ardizzone pushed a commit that referenced this issue Feb 10, 2021
Merge changes from VLL-HD
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants