Skip to content

Latest commit

 

History

History
42 lines (27 loc) · 2.17 KB

README.md

File metadata and controls

42 lines (27 loc) · 2.17 KB

DPLNet (Efficient Multimodal Semantic Segmentation via Dual-Prompt Learning)

Welcome to the official code repository for Efficient Multimodal Semantic Segmentation via Dual-Prompt Learning. We're excited to share our work with you, please bear with us as we prepare the code and demo. Stay tuned for the reveal!

Motivation

Previous multimodal methods often need to fully fine-tune the entire network, which are training-costly due to massive parameter updates in the feature extraction and fusion, and thus increases the deployment burden of multimodal semantic segmentation. In this paper, we propose a novel and simple yet effective dual-prompt learning paradigm, dubbed DPLNet, for training-efficient multimodal semantic segmentation.

Editor

Framework

Overview architecture of the proposed DPLNet, which adapts a frozen pre-trained model using two specially designed prompting learning modules, MPG for multimodal prompt generation and MFA for multimodal feature adaption, with only a few learnable parameters to achieve multimodal semantic segmentation in a training-efficient way. Framework

Visualization

RGBD Semantic Segmentation Results

NYU-V2

Results

SUN-RGBD

Results

RGBT Semantic Segmentation Results

MFNet

Results

PST900

Results

RGB-D SOD Results

Results

RGB-T SOD Results

Results

RGB-T Video Semantic Segmentation Results

Results