New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Calculating number of linear regions #10
Comments
Got a similar question here. Is this to reduce memory cost? |
Hi @YiteWang and @taoyang1122 Thank you for your interest in our work! Deep NNs typically have a very large number of linear regions. The larger the input dimension is (e.g. larger than 1x3x3), the more likely the input samples are separated into different linear regions. That means, if we use a larger input size and forward 3000 samples, we may end up with #Linear_Regions = 3000 (i.e. all input samples reside in different linear regions) for all NNs in the search space. To put it another way, reducing the input dimension makes the expressivities of different NNs more distinguishable. Hope that helps! |
@chenwydj Thanks very much! I am trying to reproduce the results on ImageNet. But I found the darts_evaluation to be very slow. It is going to take more than 4 days to train 350 epochs on 8 GPUs. Is this the same on your side? Thanks! |
@chenwydj I am having the similar issue as @taoyang1122 . Could you please let us know what hardware have you used to train found architecture on darts space on ImageNet with "batch_size = 768" on "8-gpu"? |
Hi @taoyang1122 and @maryanpetruk, We used V100 GPU to train ImageNet. It is true that training-from-scratch on ImageNet is slow: 4~5 days are very common. |
Hello, I'm trying to calculate number of linear regions but the linear region collector always returns number of dimensions (the same number for all networks).
Here are functions from |
Dear authors,
I am having a question for calculating number of linear regions. It seems that in TE-NAS, input images are augmented to be of size (1000,1,3,3):
lrc_model = Linear_Region_Collector(input_size=(1000, 1, 3, 3), sample_batch=3, dataset=xargs.dataset, data_path=xargs.data_path, seed=xargs.rand_seed)
Could you explain what the reason is behind this?
The text was updated successfully, but these errors were encountered: