ADMH-ER
Firstly, we try to generate all modal data based on pre-trained models.
- Text : we obtain the form of representations, i.e., {"name" : representation}
- Images : we obtain the form of representations, i.e., {"name" : representation}
- Videos : we obtain the form of representations, i.e., {"name" : representation}
Secondly, we generate the ids of entities and relations. entities : {entities' name : id} relations : {relations' name : id}
Thirdly, we generate the classification labels and linking labels. Note that, we create the groundtruth with the names. For each name, we have four files, i.e., "xxx.class", "xxx.json" and "xxx.link"
The raw datasets can be found below.
It is available at the following GitHub repository:
It can be accessed via:
Our processing Datasets for MMER can be found below.
The rough datasets can be available, and more details of them are coming. You can download them from:
- [MMER Datasets on BaiduPan](Link: https://pan.baidu.com/s/1kbUPWreOKUcrwaBXNDjJFA?pwd=heb9 password: heb9)
@article{Zhou2025ADMHERAD,
title={ADMH-ER: Adaptive Denoising Multi-Modal Hybrid for Entity Resolution},
author={Qian Zhou and Wei Chen and Li Zhang and An Liu and Jun Fang and Lei Zhao},
journal={IEEE Transactions on Knowledge and Data Engineering},
year={2025},
volume={37},
pages={1049-1063},
url={https://api.semanticscholar.org/CorpusID:275448400}
}