No description or website provided.
Switch branches/tags
Nothing to show
Clone or download
Latest commit b0b88d9 Dec 16, 2017
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
caffe Update README.md Dec 15, 2017
datasets first Dec 15, 2017
prototxt first Dec 15, 2017
.gitignore first Dec 15, 2017
README.md Update README.md Dec 16, 2017
framework.png Add files via upload Dec 15, 2017

README.md

GLAD

This repository provides the code for our ACM MM17 paper GLAD: Global-Local-Alignment Descriptor for Pedestrian Retrieval

Step.1 Pose Estimation

The first stage is to estimate the human keypoints. We used the deepercut model provided in DeeperCut. Especially, we utilize the single person pose estimation model provided by the authors.

Afer pose estimation, please detect the three parts according to our paper. Example image is as followes:

You can utilize any pose estimation methods to replace DeeperCut.

Step.2 Descriptor Learning

Make our caffe

We have modify the original caffe, please make our provided caffe before running our code.

Dataset

Download Market1501 Dataset. Then process these raw data as step.1.

ImageNet Pretrained model

Download GoogLeNet model pretrained on Imagenet.

Train our GLAD

  1. Modify the prototxt\train_val.prototxt. Please modify the dataset path in the file.
  2. End up training with 10,0000 iterations. More details, please see the prototxt\solver_stepsize_6400_2_step3_ver4_65.prototxt

Step.3 Test

  1. Extract fc6(and layer1/fc6, layer2/fc6, layer3/fc6) features.
  2. L1 normalization is needed.
  3. Adding weights for these four features according to our paper.

Our Model

  1. If you require our trained model, please contact Longhui Wei(weilh2568@gmail.com).
  2. If you have any questions about our code or paper, please contact Longhui Wei

Citation

Please cite our paper in your publications if it helps your research:

@inproceedings{wei2017glad,
  title={GLAD: Global-Local-Alignment Descriptor for Pedestrian Retrieval},
  author={Wei, Longhui and Zhang, Shiliang and Yao, Hantao and Gao, Wen and Tian, Qi},
  booktitle={ACM MM},
  year={2017}
}