Skip to content
/ ODMT Public

[ACM'MM 2023 Oral] Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation

License

Notifications You must be signed in to change notification settings

xyliugo/ODMT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation (ACM MM 2023)

Wei Ji*, Xiangyan Liu*, An Zhang, Yinwei Wei, Yongxin Ni, Xiang Wang

arXiv | BibTeX

This repository is the official implementation for ACM MM 2023 paper (Oral) Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation. In this work:

  • To relieve the incompatibility issue between multi-modal features and existing sequential recommendation models, we introduce the ODMT framework, which comprises an ID-aware Multi-modal Transformer module for item representation;

  • To obtain robust predictions from multi-source input, we propose an online distillation training strategy in the prediction optimization stage, which marks the first instance of applying online distillation to a multi-modal recommendation task.

  • Comprehensive experiments on four diverse multi-modal recommendation datasets and three popular backbones for sequential recommendation validate the effectiveness and transferability of our proposed method, which is about 10% performance improvement compared with other baseline models.

Brief Introduction

The paper focuses on multi-modal recommendation systems, which integrate various types of information. While traditional collaborative filtering-based multi-modal recommendation systems have received significant attention, research on multi-modal sequential recommendation is still in its early stages. We investigate the importance of item representation learning and information fusion from heterogeneous data sources, and propose a new model-agnostic framework called "Online Distillation-enhanced Multi-modal Transformer (ODMT)" to enhance feature interaction and mutual learning among multi-source input (ID, text, and image) while improving recommendation accuracy. The framework includes an ID-aware Multi-modal Transformer module for information interaction and an online distillation training strategy to improve prediction robustness. Empirical experiments on video content and e-commerce recommendation datasets show that the proposed framework achieves approximately 10% performance improvement compared to baseline models.

Usage (Arts for example)

1. dataset preprocess

First, you need to download the corresponding data file from the Google Drive link provided and place it in the specified data folder. Then, you will need to run 1_lmdb_build.py and 2_lmdb_read.py to obtain the lmdb file.

2. run

python run_arts.py

You can freely change the range of hyperparameter values according to your needs.

News

[2023.11] Upload the Arts dataset!

[2023.11] We have released the code for our paper. However, the current code has not been thoroughly tested and may contain some unexpected issues. We will provide detailed explanations for the usage and dataset sections in the future.

[2023.10] Selected as an Oral at ACM MM 2023!

[2023.07] Accepted by ACM MM 2023!

Bibliography

If you find this repository helpful for your project, please consider citing our work:

@inproceedings{ji2023online,
  title={Online distillation-enhanced multi-modal transformer for sequential recommendation},
  author={Ji, Wei and Liu, Xiangyan and Zhang, An and Wei, Yinwei and Ni, Yongxin and Wang, Xiang},
  booktitle={Proceedings of the 31st ACM International Conference on Multimedia},
  pages={955--965},
  year={2023}
}

About

[ACM'MM 2023 Oral] Online Distillation-enhanced Multi-modal Transformer for Sequential Recommendation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages