[Paper]
Zaizuo Tang1, Zongqi Yang1, Yubin Yang1,
1State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, China
Domain generalization methods can effectively enhance network performance on test samples with unknown distributions by isolating gradients between unstable and stable parameters. However, existing methods employ relatively coarse-grained partitioning of stable versus unstable parameters, leading to misclassified unstable parameters that degrade network feature processing capabilities. We first provide a theoretical analysis of gradient perturbations caused by unstable parameters. Based on this foundation, we propose Layer-Decomposition Training (LDT), which conducts fine-grained layer-wise partitioning guided by parameter instability levels, substantially improving parameter update stability. Furthermore, to address gradient amplitude disparities within stable layers and unstable layers respectively, we introduce a Dynamic Parameter Update (DPU) strategy that adaptively determines layer-specific update coefficients according to gradient variations, optimizing feature learning efficiency. Extensive experiments across diverse tasks (super-resolution, classification) and architectures (Transformer, Mamba, CNN) demonstrate LDT's superior generalization capability. Our code is available at https://github.com/ZaizuoTang/LDT.
-
Download the DRealSR dataset.
-
Download the pre-trained weights.
-
Start training:
python Train_pip.py
python basicsr/test.py -opt test_MambaIR_SR_x4.yml
This project is released under the Apache 2.0 license.
This code is based on BasicSR and MambaIR. Thanks for their awesome work.
If you have any questions, feel free to approach me at tangzz@smail.nju.edu.cn