Code implementation of paper "SEMScene: Semantic-Consistency Enhanced Multi-Level Scene Graph Matching for Image-Text Retrieval" (ACM TOMM 2024).
-
Updated
Sep 20, 2024 - Python
Code implementation of paper "SEMScene: Semantic-Consistency Enhanced Multi-Level Scene Graph Matching for Image-Text Retrieval" (ACM TOMM 2024).
[IROS 2024] This repository contains the implementation of our paper: ModaLink: Unifying Modalities for Efficient Image-to-PointCloud Place Recognition
Noise of Web (NoW) is a challenging noisy correspondence learning (NCL) benchmark containing 100K image-text pairs for robust image-text matching/retrieval models.
[ACM MM 2024] Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives
Research project at AI·Robotics Institute, KIST
[KDD 2024] Improving the Consistency in Cross-Lingual Cross-Modal Retrieval with 1-to-K Contrastive Learning
[CVPR 2024] Do you remember? Dense Video Captioning with Cross-Modal Memory Retrieval
[TIP2024] The code of “Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching”
Official Pytorch implementation of "Improved Probabilistic Image-Text Representations" (ICLR 2024)
The Unified Code of Image-Text Retrieval for Further Exploration.
[TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”
[AAAI2021] The code of “Similarity Reasoning and Filtration for Image-Text Matching”
[IJCAI 2023] Text-Video Retrieval with Disentangled Conceptualization and Set-to-Set Alignment
[ICCV 2023] DiffusionRet: Generative Text-Video Retrieval with Diffusion Model
[CVPR 2023 Highlight] Video-Text as Game Players: Hierarchical Banzhaf Interaction for Cross-Modal Representation Learning
[NeurIPS 2022 Spotlight] Expectation-Maximization Contrastive Learning for Compact Video-and-Language Representations
Polysemous Visual-Semantic Embedding for Cross-Modal Retrieval (CVPR 2019)
Official Pytorch implementation of "Probabilistic Cross-Modal Embedding" (CVPR 2021)
Extended COCO Validation (ECCV) Caption dataset (ECCV 2022)
The unofficial implementation of paper, "Objects that Sound", from ECCV 2018.
Add a description, image, and links to the cross-modal-retrieval topic page so that developers can more easily learn about it.
To associate your repository with the cross-modal-retrieval topic, visit your repo's landing page and select "manage topics."