-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Docs]Some questions about the code reproduction results of LSKNet_T in DOTA-1.0 dataset #51
Comments
下面是日志文件 |
You may face the same situation as #43 . You can try to adjust the learning rate more or less to get better performance, but this may not compensate for the side effects of small batch size. |
非常感谢作者您的回复,我按照您的思路稍微调大了一点学习率,由于我们实验设备有限,只在多尺度数据集上训练了前三个epoch,并且在第三个epoch进行验证集精度对比, 3.使用1个GPU基础lr/4(0.00005),mAP=0.75420240328_182705.log 请问作者,我增大学习率,在第三个epoch精度有一定的上涨,那么增大学习率这个方向是对的吗,如果为了尽量复现原论文中的精度指标,我又需要做什么样的调整呢?我看到您日志文件中,预热为500个iter,这是否意味着我使用1个GPU需要预热4000个iteration呢? Thank you very much for your reply, the author. I have slightly increased the learning rate according to your idea. Due to limited experimental equipment, we only trained the first three epochs on a multi-scale dataset, and compared the accuracy of the validation set on the third epochs,
|
Adjusting the learning rate is an effective way, and you can also increase the number of training epochs appropriately. I have never changed the warmup, and I am not sure whether it will have a significant impact on performance. |
@gbdjxgp 您好,请问一下您对dotav1.0的分割,采用的是作者提供的代码吗?您的分割日志能发一下吗?我想参考一下,谢谢您这是我的日志 |
直接采用文档进行分割就行,我没找到分割文件 |
@gbdjxgp, Hi, Have you reproduce the work sucessfully?If successful, what adjustments have been made |
Hello, I did not conduct further experiments on multi-scale datasets. I only conducted experiments on single-scale datasets. It is recommended to adjust the learning rate of a single GPU to 0.00005, and the result of a single-scale dataset may be around 0.755 |
Branch
master branch https://mmrotate.readthedocs.io/en/latest/
📚 The doc issue
作者您好,我正在基于单个GPU复现您的代码,按照您文档中的说明,采用多尺度训练,仅读取预训练的主干,配置文件使用LSKNet_T,将syncBN改成BN,学习率从原来的0.0002改成0.0002/8,下面附件里是训练过程中的日志文件,从实验结果中看到精度差异较大,并不能达到您日志中的0.852,请问是我超参数设置的有问题吗,期待您的回复。
Hello author, I am replicating your code based on a single GPU. Following the instructions in your document, we are using multi-scale training to only read the pre trained backbone. The configuration file uses LSKNet-T, changing syncBN to BN, and changing the learning rate from 0.0002 to 0.0002/8. Attached is the log file from the training process. From the experimental results, we can see that there is a significant difference in accuracy, which cannot reach the 0.852 in your log. May I ask if there is an issue with my hyperparameter settings? Looking forward to your reply.
Suggest a potential alternative/fix
No response
The text was updated successfully, but these errors were encountered: