ViTAA: Visual-Textual Attributes Alignment in Person Search by Natural Language
-
Updated
Dec 22, 2021 - Python
ViTAA: Visual-Textual Attributes Alignment in Person Search by Natural Language
Flask Web App for ES-654 Machine Learning course project
Image-Text Matching Model Zoo
The Unified Code of Image-Text Retrieval for Further Exploration.
An attempt to transfer sentence to image style.
The code for the paper "GMMFormer: Gaussian-Mixture-Model Based Transformer for Efficient Partially Relevant Video Retrieval" (AAAI'24)
PyTorch code for cross-modal-retrieval on Flickr8k/30k using Bert and EfficientNet
Tensorflow implementation of UDIH
Code for the paper "Sentiment-Oriented Metric Learning for Text-to-Image Retrieval", ECIR'21
Implementation of "VSE++: Improving Visual-Semantic Embeddings with Hard Negatives" in Tensorflow.
This repository contains the code for the paper "Object-centric vs. Scene-centric Image-Text Cross-modal Retrieval: A Reproducibility Study" published at ECIR 2023.
PyTorch code for the paper "Complementarity is the king: A multi-modal and multi-grained hierarchical semantic enhancement network for cross-modal retrieval"
Deep Semisupervised Cross-modal Retrieval/Cross-view Recognition (IEEE TCYB 2022, PyTorch Code)
My master thesis: Siamese multi-hop attention for cross-modal retrieval.
Joint Versus Independent Multiview Hashing for Cross-View Retrieval[J] (IEEE TCYB 2021, PyTorch Code)
[TIP2024] The code of “Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching”
[ICASSP 2022] EEG - Music Cross Modal Learning
Add a description, image, and links to the cross-modal-retrieval topic page so that developers can more easily learn about it.
To associate your repository with the cross-modal-retrieval topic, visit your repo's landing page and select "manage topics."