Attention Guided Global Enhancement and Local Refinement Network for Semantic Segmentation [arXiv] [IEEE]
Jiangyun Li, Sen Zha, Chen Chen, Meng Ding, Tianxiang Zhang, and Hong Yu
In this paper, a novel AGLN is proposed to improve the encoder-decoder network for image segmentation. First, a Global Enhancement Method is designed to capture global semantic information from high-level features to complement the deficiency of global contexts in the upsampling process. Then, a Local Refinement Module is built to refine the noisy encoder features in both channel and spatial dimensions before the context fusion. After that, the proposed two methods are integrated into the Context Fusion Blocks, enabling the AGLN to generate semantically consistent segmentation masks on large-scale stuff and accurate boundaries on delicate objects.
AGLN achieves the state-of-the-art result (56.23% mean IOU) on the PASCAL Context dataset.
Please follow PyTorch-Encoding, and our work is in agln.py.
@article{li2022attention,
title={Attention guided global enhancement and local refinement network for semantic segmentation},
author={Li, Jiangyun and Zha, Sen and Chen, Chen and Ding, Meng and Zhang, Tianxiang and Yu, Hong},
journal={IEEE Transactions on Image Processing},
year={2022},
publisher={IEEE}
}
Thanks PyTorch-Encoding.