Skip to content

ShaohuaDong2021/DPLNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 

Repository files navigation

DPLNet (Efficient Multimodal Semantic Segmentation via Dual-Prompt Learning)

Welcome to the official code repository for Efficient Multimodal Semantic Segmentation via Dual-Prompt Learning. We're excited to share our work with you, please bear with us as we prepare the code and demo. Stay tuned for the reveal!

Motivation

Previous multimodal methods often need to fully fine-tune the entire network, which are training-costly due to massive parameter updates in the feature extraction and fusion, and thus increases the deployment burden of multimodal semantic segmentation. In this paper, we propose a novel and simple yet effective dual-prompt learning paradigm, dubbed DPLNet, for training-efficient multimodal semantic segmentation.

Editor

Framework

Overview architecture of the proposed DPLNet, which adapts a frozen pre-trained model using two specially designed prompting learning modules, MPG for multimodal prompt generation and MFA for multimodal feature adaption, with only a few learnable parameters to achieve multimodal semantic segmentation in a training-efficient way. Framework

Visualization

RGBD Semantic Segmentation Results

NYU-V2

Results

SUN-RGBD

Results

RGBT Semantic Segmentation Results

MFNet

Results

PST900

Results

RGB-D SOD Results

Results

RGB-T SOD Results

Results

RGB-T Video Semantic Segmentation Results

Results

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages