Skip to content

Latest commit

 

History

History
104 lines (67 loc) · 3.62 KB

README.md

File metadata and controls

104 lines (67 loc) · 3.62 KB

Real-time ReID Tracking w/. Lite but Strong Feature Extractor & GAN

Slides, Manuscript(Coming Soon...)

Introduction

From Sort to OCSort, we are becoming aware that deep feature extractor is crucial in both re-identification and multiple object tracking. The project integrates Yolo detection, GAN, deep feature extractor for re-identification and MOT. The baseline is a Yolov5(now is YoloV8!) based DeepSort(now is DeepOCSort & StrongSort) algorithm. Everything can be found here. The author has updated the repository with import of fast-reid package.
We yield even better results than ResNet50 with ResNet18-based model!!!

Dependency

Python >= 3.8 is recommended.

All code was developed with the latest version of PyTorch(2.0!) and Numpy. Please download and install the latest stable version with CUDA 11.x(ex. 11.8). All codes are tested with one single GPU. If you wish to accelerate training, you are advised to apply mixed-precision training. It is estimated to save training time as well as reduce memory cost. Distributed training is also available. If you wish to accelerate inference, you may refer to TorchScript, ONNX or TensorRT. A major reference is this.

You are suggested to install other dependencies with

pip --disable-pip-version-check install -r requirements.txt

With >= Python3.9, you can install CUML with

pip install cuml-cu11 --extra-index-url=https://pypi.nvidia.com

Optionally, using faiss is highly recommended.

conda install -c conda-forge faiss-gpu
conda install faiss-gpu=1.7.4

Datasets

The current datasets are a bit outdated. You are advised to use MOT17, MOT20 instead. Now we primarily use MOT16 evaluation as benchmark and Market1501/DukeMTMC/VeRi-776 to pre-train our re-id network.

Video Dataset: MOT16

Image Datasets: Market1501, DukeMTMC-reID, VeRi-776

Quick Start

External Quick Start Doc

Evaluation

Re-identification

External ReID Doc

Tracking

External Tracking Doc

Tracking Speed

[The following conclusion is outdated]

The baseline extractor in DeepSort-YoloV5 implementation is pure ResNet-18. The inference speed is 15 ~ 20 ms per frame [need to be re-assessed and OSNet 1.0 is heavy!!! ~100ms per frame] depending on the sparsity of pedestrians with 640 * 640 resolution with Tesla P100. It may be slower if bounding box is resized to (128, 256). The modified extractor is based on Dense skip connection in ResNet-18 with Squeeze and Excitation Network, only a minor increase on the number of learnable parameters. The tracking speed is 17 ms per frame under the same testing environment. The speed is acquired with time package after the synchronization of CUDA.

License

MIT License

Citation

If you find my work useful in your research, please consider citing:

@misc{SuperbTUM,
  author = {Mingzhe Hu},
  title = {real-time ReID Tracking},
  year = {2023},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/SuperbTUM/real-time-ReID-tracking}},
}