Skip to content

PurdueDigitalTwin/MACP

Repository files navigation

MACP: Efficient Model Adaptation for Cooperative Perception

python BSD 3-Clause License


The official repository for the WACV 2024 paper MACP: Efficient Model Adaptation for Cooperative Perception. This work proposes a novel method to adapt a single-agent pretrained model to a V2V cooperative perception setting. It achieves state-of-the-art performance on both the V2V4Real and the OPV2V datasets.

Setup

Our project is based on MMDetection3D v1.1.0. Please refer to the official documentation to set up the environment.

Data Preparation

Download the V2V4Real and OPV2V datasets.

Once the data is downloaded, it's necessary organize the data in the following structure:

├── $REPO_ROOT
│   ├── data
│   │   ├── v2v4real
│   │   │   ├── train
│   │   │   │   ├── testoutput_CAV_data_2022-03-15-09-54-40_0 # data folder
│   │   │   ├── test
|   |   ├── openv2v
│   │   │   ├── train
│   │   │   │   ├── 2021_08_16_22_26_54 # data folder
│   │   │   ├── test
|   |   |   ├── validate
|   |   |   ├── test_culver_city

Then, run the script files scripts/create_v2v4real.sh and scripts/create_openv2v.sh to prepare the cached data.

Notes

  • The core code of our project is in the projects/Coperception folder.
  • The voxelization OP in the original implementation of BEVFusion is different from the implementation in MMCV. Please refer here to compile the OP on CUDA.

MACP Weights

If you are interested in including any other pretrained weights or details, please open an issue or contact us.

Model Backbone Checkpoint Config AP@50 AP@70 Log
MACP-V2V4Real BEVFusion-LiDAR Google Drive Google Drive 67.6 47.9 Google Drive
MACP-OPV2V BEVFusion-LiDAR Google Drive Google Drive 93.7 90.3 Google Drive

Training

We train our model on one NVIDIA RTX 4090 GPU with 24GB memory. The training command is as follows:

cd /path/to/repo
export PYTHONPATH=$PWD:$PYTHONPATH
python tools/train.py path/to/config

Evaluation

The evaluation command is as follows:

cd /path/to/repo
export PYTHONPATH=$PWD:$PYTHONPATH
python tools/test.py path/to/config path/to/checkpoint

Citation

If you find our work useful in your research, please consider citing:

@inproceedings{ma2024macp,
  title={MACP: Efficient Model Adaptation for Cooperative Perception},
  author={Ma, Yunsheng and Lu, Juanwu and Cui, Can and Zhao, Sicheng and Cao, Xu and Ye, Wenqian and Wang, Ziran},
  booktitle={Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision},
  pages={3373--3382},
  year={2024}
}

Acknowledgement

This project is based on code from several open-source projects. We would like to thank the authors for their great work: