New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question about the Experiment of Image Multi-label Classification #10
Comments
Could you please elaborate on the error details? |
@sacmehta Execute the commands of the above program. When I set s to 0.2, the program would be stuck there because of the small memory of my machine graphics card. ×#×--------------------------------------------------------------------------------------------------- In Table 1, when s = 0.2, FLOPs is 12M, and when s = 0.1, FLOPs is 6.5M. python train_classification.py --model dicenet In the experiment of image multi-label classification, execute the command of the above program. When I set s to 0.1, the program will report the error in the figure below. So, Could you tell me how to do it? Thank you very much. |
Uncomment this line and it should work. Note that we have not provided pretrained ImageNet weights at this configuration, so you might want to train on theImageNet first.
|
@sacmehta In the paper 《ELASTIC: Improving CNNs with Dynamic Scaling Policies》, the data set used is mscoco 2014, the size of training set/verification set is 82783/40504, |
In our paper, we used the same split as ELASTIC. |
Since COCO2017 is the latest version, we provided the default values to 2017 split. You can change them if you want. |
Code modified to version of data set 2014: `# -- coding: utf-8 -- """ class CocoDetection(datasets.coco.CocoDetection):
def main(args):
if name == 'main':
Dear Sir, |
You need to use lesser value of learning rate. Try 0.005 |
@sacmehta I use dicenet_s_0.2_imagenet_224x224.pth file to fine-tune on mscoco, using the following commands: The experimental process is as follows: The results of each epoch are as follows: Although there are improvements in each round, the effect is not very great. #------------------------------------------------------------------------------------------ I used clr(scheduler ) to carry out another group of experiments. The experimental results are as follows: In your training process, R_C and F_C are also slowly rising from a very small number (for example, in my experiment, 1.21/0.88)? On my desktop, GTX 970, batch size is 64, each round takes 40 minutes. If it takes 2.7 days to train 100 rounds, I don't know if the F_C value can exceed 71.08 (ELASTIC). Did my experiment go wrong? Can you give me some advice? Thank you very much for your patient reply ! |
With s=0.2, you cannot reach the value close to Elastic paper. You need to use the best Dicenet model. Also, batch size of 64 is too small. Try using something like 512 for best results. Since you are able to run the code and it is more of hyper-parameter tuning based on your machine setup which is beyond this repo, I am closing this issue. |
In the paper 《DiCENet: Dimension-wise Convolutions for Efficient Networks》, the network width scaling parameter s can be selected, but in the experiment of image multi-label classification, the experimental error of s=0.1 (when I run the corresponding program of s=0.2, the machine can't run, but the network parameter of s=0.1 is less than half) ... can you provide a program for s = 0.1?
The text was updated successfully, but these errors were encountered: