Skip to content
/ PRNet Public

The code of paper Efficient Camouflaged Object Detection Network Based on Global Localization Perception and Local Guidance Refinement

Notifications You must be signed in to change notification settings

hu-xh/PRNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 

Repository files navigation

PRNet

The paper was published by the IEEE Transactions on Circuits and Systems for Video Technology on January 2, 2024. The paper link is: - Link

Efficient Camouflaged Object Detection Network Based on Global Localization Perception and Local Guidance Refinement

PRNet

Camouflaged Object Detection (COD) is a challenging visual task due to its complex contour, diverse scales, and high similarity to the background. Existing COD methods encounter two predicaments: One is that they are prone to falling into local perception, resulting in inaccurate object localization; Another issue is the difficulty in achieving precise object segmentation due to a lack of detailed information. In addition, most COD methods typically require larger parameter amount and higher computational complexity in pursuit of better performance. To this end, we proposed an global localization perception and local guidance refinement network (PRNet), that simultaneously addresses performance and computational costs. Through aggregation and effective use of semantic and details information, PRNet achieves accurate localization and refined segmentation of camouflaged objects. Specifically, with the help of a designed Cascaded Attention Perceptron (CAP), we can effectively integrate and perceive multi-scale information to localize camouflaged objects. A Guided Refinement Decoder (GRD) is designed in a top-down manner helps to extract context information and aggregate details to refine camouflaged prediction results. Extensive experimental results demonstrate that our PRNet outperforms 10 state-of-the-art models on four challenging datasets across four evaluation metrics. Notably, PRNet exhibits a more compact structure, lower parameters, and reduced computational complexity. Source codes will be available at https://github.com/hu-xh/PRNet.

Network Architecture

fig1.png

Quantitative Evaluation Results and Qualitative Evaluation Results:

fig2.jpg fig3.jpg

Prerequisites

  • Python 3.6
  • Pytorch 1.10.2
  • Torchvision 0.11.3
  • Numpy 1.19.2

Pretrained Model

Download the following pth and put it into main folder

  • SMT-t with the fetch code:1234.

Datasets

Results

You can download the tested results map at - Baidu Pan link with the fetch code:1234.

You can download the results pth - Baidu Pan link with the fetch code:1234.

Contact

Feel free to send e-mails to me (1558239392@qq.com).

About

The code of paper Efficient Camouflaged Object Detection Network Based on Global Localization Perception and Local Guidance Refinement

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published