Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question on experiences with Efficiency #10

Open
MarcCoru opened this issue Oct 20, 2020 · 2 comments
Open

Question on experiences with Efficiency #10

MarcCoru opened this issue Oct 20, 2020 · 2 comments

Comments

@MarcCoru
Copy link

Hi @lucidrains,
Thanks a lot for providing this implementation so quickly.
I have a question regarding your (or other's) experience on the efficiency of Lambdalayers.
I tried to implement a LambdaUNet where I changed the 3x3conv layers with lambdalayers and avg pooling.

The Conv-UNet has 17Mio parameters while the LambdaUNet only 3Mio. Still, inference and training take much longer in the LambdaUNet than in the ConvUNet (approx 1s ConvUnet vs 10s Lamndaunet). I also used a receptive field of r=23. I am not sure where this parameter originates from or what receptive field should be set. In the paper, the authors talk about "controlled experiments". I assume they chose the lambdalayer hyperparameter (in some way) similar to the conv parameters? It is not very clear from the paper (at least from my initial reading).

I was wondering if others share my experience on slower training and inference time when blindly changing conv layer with lambda layers. Maybe someone can share his expertise on how I can control my LambdaUnet to be comparable to a regular UNet to reproduce the performance and efficiency results from the paper.
Thanks again

@lucidrains
Copy link
Owner

@MarcCoru the paper hasn't even been reviewed yet, so I think we are all in uncharted territories. Let's just keep this open so people can add to the discussions

@yjxiong
Copy link

yjxiong commented Oct 20, 2020

Table 12 in the openreview version's appendix has provide references to the inference speed. Seems with more convolution layers replaced, the inference drastically slows down.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants