Skip to content
/ mdfe Public

Framework for learning multi-domain image embeddings suitable for multi-domain image retrieval at instance-level

Notifications You must be signed in to change notification settings

morrisfl/mdfe

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MDFE - Multi-Domain Feature Extraction

model overview

Figure 1: Overview of the proposed multi-domain image embedding model. The model consists of a visual-semantic foundation model as backbone with an attached projection layer. The model was trained on a custom curated multi-domain training dataset (M4D-35k), using a margin-based softmax loss.

This repository is related to the research conducted for my master thesis entitled "Efficient and Discriminative Image Feature Extraction for Multi-Domain Image Retrieval". An executive summery of the thesis can be found here. For those interested in a comprehensive review of the entire work, please feel free to contact me to obtain a copy of the full thesis report.

Abstract

The prevalence of image capture devices has led to the growth of digital image collections, requiring advanced retrieval systems. Current methods are often limited by their domain specificity, struggle with out-of-domain images, and lack of generalization. This study addresses these limitations by focusing on multi-domain feature extraction. The goal entails in developing an efficient multi-domain image encoder for fine-grained retrieval while overcoming computational constraints. Therefore, a multi-domain training dataset, called M4D-35k, was curated, allowing for resource-efficient training. Dataset curation involved selecting from 15 datasets and optimizing their overall size in terms of samples and classes used. Additionally, the effectiveness of various visual-semantic foundation models and margin-based softmax loss were evaluated to assess their suitability for multi-domain feature extraction. Among the loss functions investigated, a proprietary approach was developed that refers to CLAM (class distribution aware additive angular margin loss). Even with computational limitations, a close to SOTA result was achieved on the Google Universal Image Embedding Challenge (GUIEC) evaluation dataset. Linear probing of the embedding model alone resulted in a mMP@5 score of 0.722. The total number of model parameters and the number of trainable parameters were reduced by 32% and 289 times, respectively. Despite the smaller model and without end-to-end fine-tuning, it trailed the GUIEC leaderboard by only 0.8%, surpassing 2nd place and closely behind 1st. It also outperformed the top-ranked method with similar computational prerequisites by 3.6%.

Results

GUIEC rank Method # total model params # trainable params mMP@5
1st place fine-tuning 661M 661M 0.730
2nd place fine-tuning 667M 667M 0.711
5th place linear probing 633M 1.1M 0.686
10th place linear probing 1,045M 22.0M 0.675
Own approach linear probing 431M 2.3M 0.722

Table: Comparison of the proposed approach with the top-ranked methods on the GUIEC evaluation dataset. It improves the total model parameters at inference by 32% compared to the leanest approach (5th place), reduces the number of trainable parameters by 289x compared to the fine-tuning approaches (1st and 2nd place), and achieves a performance close to SOTA, surpassing 2nd place and just behind 1st place.

Table of Contents

I. Setup

Here, we describe a step-by-step guide to setup and install dependencies on a UNIX-based system, such as Ubuntu, using conda as package manager. If conda is not available, alternative package managers such as venv can be used.

1. Create a virtual environment

conda create -n env_mdfe python=3.8
conda activate env_mdfe

2. Clone the repository

git clone git@github.com:morrisfl/mdfe.git

3. Install pytorch

Depending on your system and compute requirements, you may need to change the command below. See pytorch.org for more details. In order to submit the embedding models to the 2022 Google Universal Image Embedding Challenge, PyTorch 1.11.0 is required.

conda install pytorch==1.11.0 torchvision==0.12.0 cudatoolkit=11.3 -c pytorch

4. Install the repository with all dependencies

cd mdfe
python -m pip install .

If you want to make changes to the code, you can install the repository in editable mode:

python -m pip install -e .

5. Setup Google Drive access (optional)

In order to automatically upload checkpoints to Google Drive, you need to create a Google Drive API key. Setup instructions can be found here. If you don't want to upload checkpoints to Google Drive, please set the MODEL.cloud_upload parameter in the configuration file to False.

II. Data Preparation

In the process of fine-tuning/linear probing the embedding models, different datasets and dataset combinations can be used. The list of available datasets, and information about pre-processing, downloading and how to use them for training can be found here.

M4D-35k

The M4D-35k dataset is a custom curated multi-domain training dataset. It was created for resource-efficient training of multi-domain image embeddings. The curation process involved dataset selection and data sampling (optimize data size) by maximizing the performance on the GUIEC evaluation dataset. M4D-35k consists of 35k classes and 328k images sourced from four different datasets:

Domain Dataset # classes # images
Packaged goods Products-10k 9.5k 141.5k
Landmarks Google Landmarks v2 (subset) 10.0k 79.2k
Apparel & Accessories DeepFashion (Consumer to Shop) 14.3k 100.4k
Cars Stanford Cars (refined) 1.0k 7.3k
Multi-Domain M4D-35k 34.8k 328.4k

Notable, the Stanford Cars dataset was refined by enhancing the class granularity. Instead of classifying cars only by their model, the class labels were extended to the car color. More information about the refinement process can be found here.

The corresponding annotations of the M4D-35k dataset can be found in data/m4d-35k_train.csv. Make sure to download the corresponding datasets included in the M4D-35k dataset and place them in a <data_dir> of your choice. More information about the dataset and directory structure can be found here.

To use M4D-35k for training, add m4d_35k to the DATASET.names parameter in the configuration file in configs/.

III. Embedding Model

The architecture of the image embedding model is illustrated in Figure 1. The model consists of a visual-semantic foundation model as backbone with an attached projection layer. Different foundation models can be used, as shown in the table below.

Foundation Model Encoder architecture type model_name weights
OpenCLIP ViT clip see OpenCLIP see OpenCLIP
OpenCLIP ConvNeXt clip_convnext see OpenCLIP see OpenCLIP
CLIPA ViT clipav2 see OpenCLIP see OpenCLIP
EVA-CLIP ViT eva02 see timm -
MetaCLIP ViT meta-clip see OpenCLIP see OpenCLIP
SigLIP ViT siglip see timm -
DINOv2 ViT dinov2 see timm -
SAM ViT sam see timm -

In order to adjust the model architecture of the image embedding model, the following main parameters can be changed in the configuration file:

  • MODEL.embedding_dim: the dimension of the image embedding.
  • MODEL.BACKBONE.type: the type of the visual-semantic foundation model, supported types are those listed in the table above.
  • MODEL.BACKBONE.model_name: the name of the visual-semantic foundation model, specified by OpenCLIP or timm.
  • MODEL.BACKBONE.weights: the weights of the visual-semantic foundation model, only required for OpenCLIP models (corresponds to the pretrained parameter in OpenCLIP).
  • MODEL.NECK.type: the type to reduce the embedding dimension to the specified MODEL.embedding_dim, supported types are proj_layer and pooling.
  • MODEL.HEAD.name: the name of the margin-based softmax loss, supported names are ArcFace, DynM-ArcFace, AdaCos, LiArcFace, CurricularFace, and AdaFace.
  • MODEL.HEAD.k: the number of centers for the margin-based softmax loss.
  • MODEL.HEAD.s: the scaling factor for the margin-based softmax loss.
  • MODEL.HEAD.m: the margin for the margin-based softmax loss.

Further explanations of changeable parameters can be found in the default_cfg.py.

IV. Training

Training settings

The training settings can be changed in the configuration file found in configs/. The most important parameters are:

  • TRAIN.epoch_based: if True, the training is based on the number of epochs, otherwise on the number of iterations.
  • TRAIN.epochs: the number of epochs to train the model.
  • TRAIN.save_epoch: the frequency of saving the model checkpoint.
  • OPTIMIZER.name: the optimizer used for training, supported optimizers are Adam, AdamW and SGD.
  • OPTIMIZER.lr: the learning rate of the optimizer.
  • OPTIMIZER.weight_decay: the weight decay of the optimizer.
  • SCHEDULER.name: the learning rate scheduler used for training, supported schedulers are CosineAnnealingLR.
  • SCHEDULER.epoch_based: if True, the scheduler is based on the number of epochs, otherwise on the number of iterations.
  • SCEDULER.min_lr: the minimum learning rate of the scheduler.
  • SCHEDULER.warmup: the type of warmup to use, supported warmups are linear and exponential.
  • SCHEDULER.warmup_steps: the number of warmup steps. If the value is 1, the steps are equivalent to the number of iterations of one epoch.

Further explanations of changeable parameters can be found in the default_cfg.py.

Training run

To start the training, run the following command:

python tools/train.py configs/<config_file> <data_dir> \
    --output-dir results/ \
    --data_parallelism \
    --device cuda:0

The <config_file> corresponds to the configuration file in configs/ and <data_dir> to the directory where the datasets are stored. The --output-dir parameter specifies the directory where the training results are stored. The --data_parallelism parameter enables the use of multiple GPUs for training (available GPU IDs must be specified in the configuration file under TRAIN.gpu_ids). The --device parameter specifies the device to use for training.

Zero-shot model

To create a zero-shot model, which can be used for zero-shot evaluate, run the following command:

python tools/zero_shot_model.py configs/<config_file>

Within the <config_file>, the model architecture should be specified. The zero-shot model will be saved in the results/. If the MODEL.cloud_upload parameter in the configuration file is set to True, the zero-shot model will be uploaded to Google Drive.

V. Evaluation

This repository does not provide any evaluation scripts. However, the evaluation of the trained embedding models can be performed on the Google Universal Image Embedding Challenge hosted on Kaggle. The evaluation dataset consists of 5k query images and 200k index images across 11 different domains. The evaluation metric is the modified mean precision at 5 (mMP@5).

For evaluation, the trained embedding model has to be uploaded to Kaggle, where a scoring notebook performs feature extraction and metric computation on the evaluation dataset. In notebooks/ a template notebook is provided, which can be used to submit the trained embedding model to the challenge. Please note, this notebook downloads the embedding model from Google Drive. Therefore, the model_name and download url (shared link) have to be specified.