Please cite the following paper if you intend to use this code for your research.
B. Salehi, G. Reus-Muns, D. Roy, Z. Wang, T. Jian, J. Dy, S. Ioannidis, and K. Chowdhury, “Deep Learning on Multimodal Sensor Data at the Wireless Edge for Vehicular Network,” IEEE Transactions on Vehicular Technology, vol. 71, no. 7, pp. 7639-7655, July 2022.
@article{salehi2022deep, title={Deep learning on multimodal sensor data at the wireless edge for vehicular network}, author={Salehi, Batool and Reus-Muns, Guillem and Roy, Debashri and Wang, Zifeng and Jian, Tong and Dy, Jennifer and Ioannidis, Stratis and Chowdhury, Kaushik}, journal={IEEE Transactions on Vehicular Technology}, volume={71}, number={7}, pages={7639--7655}, year={2022}, publisher={IEEE} }
We validate our approach on synthetic Raymobtime (https://pages.github.com/) and Real-world NEU dataset (https://genesys-lab.org/multimodal-fusion-nextg-v2x-communications). This repository is based on Raymobtime dataset. However, it can be applied to any other dataset (including NEU dataset), by changing the models to account for new input shapes.
We designed custom image feature extractors for Raymobtime dataset. To download the features please visit: (https://drive.google.com/drive/folders/1kRU8nmnvRj8DNU-VZYNqKnlZZNr84vu-?usp=sharing). The results in the paper are generated using the data mentioned in this link.
If you wish to generate the image features from scratch, please use the code available in create_image_feature directory.
The pre-trained model are available here (https://drive.google.com/drive/folders/1kRU8nmnvRj8DNU-VZYNqKnlZZNr84vu-?usp=sharing).
Please use the command below to run the framework. As an example, for testing on all three modalities, set the argument train_or_test as "test" and use the pre-trained models.
python main.py --id_gpu "gpu id to use" --data_folder "local path to dataset" --input coord img lidar --test_data_folder "local path to dataset" --model_folder "local path to model folders" --image_feature_to_use custom --train_or_test test