Skip to content

chkim403/blstm-mtp

Repository files navigation

BLSTM-MTP

This repository contains the official Tensorflow implementation of Discriminative Appearance Modeling with Multi-track Pooling for Real-time Multi-object Tracking (CVPR 2021).

Dependencies

The code has been tested with:

  • Python 3.7.6
  • Tensorflow 1.15
  • CUDA 10.0
  • cuDNN 7.6.5
  • OpenCV 4.5.3

Download data

  1. Download the MOT17 Challenge dataset from this link. The zip file includes MOT Challenge public detections processed by Tracktor. We use this version of the public detections in our tracking demo below.
  2. If you already downloaded the dataset from the official MOT Challenge website before, please download the data from this link instead which doesn't include the image files.

Demo

  1. Set DATASET_DIR in config_tracker.py to your own directory where the dataset you download above is located.
  2. If you want to write the tracking output as images as well, set IS_VIS in config_tracker.py to True. Otherwise, leave it as it is.
  3. Download the model file from here and unzip the file. Use the location where the checkpoint file is located as model_path in the command below.
  4. Run the following command. Use your own paths for model_path and output_path. As for detector, you can use one of DPM, FRCNN, and SDP.
python run_tracker.py --model_path=YOUR_MODEL_FOLDER/model.ckpt --output_path=YOUR_OUTPUT_FOLDER  --detector=FRCNN --threshold=0.5 --network_type=appearance_motion_network
  1. This command will generate the tracking result that is shown in Table 6 of our paper. You can use these files to verify your output files.

Performance

When paired with Tracktor or CenterTrack, our method greatly improves the tracking performance in terms of IDF1 and IDS.

Method IDF1 MOTA IDS MT ML Frag FP FN
Tracktor++v2 55.1 56.3 1,987 21.1 35.3 3,763 8,866 235,449
Ours + Tracktor++v2 60.5 55.9 1,188 20.5 36.7 4,185 8,663 238,863

The data file that you download in the instructions above also includes MOT Challenge detections processed by CenterTrack (centertrack_prepr_det.txt). In order to use it as input to the tracker, you can simply change run_tracker.py in a way that it reads detections from centertrack_prepr_det.txt instead of tracktor_prepr_det.txt. The following is the result obtained by using the public detections processed by CenterTrack.

Method IDF1 MOTA IDS MT ML Frag FP FN
CTTrackPub 59.6 61.5 2,583 26.4 31.9 4,965 14,076 200,672
Ours + CTTrackPub 62.9 62.0 1,750 27.9 31.0 7,433 17,621 194,946

With NVIDIA TITAN Xp, the inference code runs at around 24 fps on the MOT17 Challenge test set (excluding time spent on I/O operations).

Training

The training code will be released soon in the future release. Stay tuned for more updates.

License

The code is released under the MIT License.

Contact

If you have any questions, please contact me at chkim@gatech.edu.

Citation

@InProceedings{Kim_2021_CVPR,
    author    = {Kim, Chanho and Fuxin, Li and Alotaibi, Mazen and Rehg, James M.},
    title     = {Discriminative Appearance Modeling With Multi-Track Pooling for Real-Time Multi-Object Tracking},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2021},
    pages     = {9553-9562}
}

About

Official Tensorflow Implementation of "Discriminative Appearance Modeling with Multi-track Pooling for Real-time Multi-object Tracking"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages