DEFT: Detection Embeddings for Tracking
DEFT: Detection Embeddings for Tracking,
Mohamed Chaabane, Peter Zhang, J. Ross Beveridge, Stephen O'Hara
arXiv technical report (arXiv 2102.02267)
@article{Chaabane2021deft,
title={DEFT: Detection Embeddings for Tracking},
author={Chaabane, Mohamed and Zhang, Peter and Beveridge, Ross and O'Hara, Stephen},
journal={arXiv preprint arXiv:2102.02267},
year={2021}
}
Contact: chaabane@colostate.edu. Any questions or discussion are welcome!
Most modern multiple object tracking (MOT) systems follow the tracking-by-detection paradigm, consisting of a detector followed by a method for associating detections into tracks. There is a long history in tracking of combining motion and appearance features to provide robustness to occlusions and other challenges, but typically this comes with the trade-off of a more complex and slower implementation. Recent successes on popular 2D tracking benchmarks indicate that top-scores can be achieved using a state-of-the-art detector and relatively simple associations relying on single-frame spatial offsets -- notably outperforming contemporary methods that leverage learned appearance features to help re-identify lost tracks. In this paper, we propose an efficient joint detection and tracking model named DEFT, or Detection Embeddings for Tracking. Our approach relies on an appearance-based object matching network jointly-learned with an underlying object detection network. An LSTM is also added to capture motion constraints. DEFT has comparable accuracy and speed to the top methods on 2D online tracking leaderboards while having significant advantages in robustness when applied to more challenging tracking data. DEFT raises the bar on the nuScenes monocular 3D tracking challenge, more than doubling the performance of the previous top method.
Dataset | AMOTA | MOTAR | MOTA |
---|---|---|---|
nuScenes | 17.7 | 48.4 | 15.6 |
The results are obtained on the nuScenes challenge evaluation server.
Dataset | MOTA | MOTP | IDF1 | IDS |
---|---|---|---|---|
MOT16 (Public) | 61.7 | 78.3 | 60.2 | 768 |
MOT16 (Private) | 68.03 | 78.71 | 66.39 | 925 |
MOT17 (Public) | 60.4 | 78.1 | 59.7 | 2581 |
MOT17 (Private) | 66.6 | 78.83 | 65.42 | 2823 |
The results are obtained on the MOT challenge evaluation server.
Dataset | MOTA | MOTP | MT | ML | IDS |
---|---|---|---|---|---|
KITTI | 88.95 | 84.55 | 84.77 | 1.85 | 343 |
The results are obtained on the KITTI challenge evaluation server.
- Clone this repo, and run the following commands.
- create a new conda environment and activate the environment.
# clone DEFT repository
cd DEFT
conda env create --file environment.yml
conda activate DEFT
- Install the dependencies.
pip install -r requirements.txt
- Install COCOAPI:
pip install -U 'git+https://github.com/cocodataset/cocoapi.git#subdirectory=PythonAPI'
- Compile DCNv2
cd src/lib/model/networks/
git clone https://github.com/CharlesShang/DCNv2
cd DCNv2
./make.sh
- Download the pretrained models and move them to src/models/ .
We use similar datasets preparation like in CenterTrack framework
- Download the dataset from nuScenes website. You only need to download the "Keyframe blobs", and only need the images data. You also need to download the maps and all metadata.
- Unzip, rename, and place the data as below. You will need to merge folders from different zip files.
${DEFT_ROOT}
|-- data
`-- |-- nuscenes
`-- |-- v1.0-trainval
| |-- samples
| | |-- CAM_BACK
| | | | -- xxx.jpg
| | |-- CAM_BACK_LEFT
| | |-- CAM_BACK_RIGHT
| | |-- CAM_FRONT
| | |-- CAM_FRONT_LEFT
| | |-- CAM_FRONT_RIGHT
|-- |-- maps
`-- |-- v1.0-trainval_meta
- Run the dataset preprocessing script.
cd src/tools/
convert_nuScenes.py
- Run the dataset preprocessing script.
cd src/tools/
sh get_mot_17.sh
- The output data structure should be:
${DEFT_ROOT}
|-- data
`-- |-- mot17
`-- |--- train
| |--- MOT17-02-FRCNN
| | |--- img1
| | |--- gt
| | | |--- gt.txt
| | |--- det
| | | |--- det.txt
| |--- ...
|--- test
| |--- MOT17-01-FRCNN
|---|--- ...
`---| annotations
|--- train.json
`--- test.json
- Download images, annotations, and calibration information from KITTI Tracking website and unzip. Place or symlink the data as below:
${DEFT_ROOT}
|-- data
`-- |-- kitti_tracking
`-- |-- data_tracking_image_2
| |-- training
| |-- |-- image_02
| |-- |-- |-- 0000
| |-- |-- |-- ...
|-- |-- testing
|-- label_02
| |-- 0000.txt
| |-- ...
`-- data_tracking_calib
- Run the dataset preprocessing script.
cd src/tools/
sh get_kitti_tracking.sh
- The resulting data structure should look like:
${DEFT_ROOT}
|-- data
`-- |-- kitti_tracking
`-- |-- data_tracking_image_2
| |-- training
| | |-- image_02
| | | |-- 0000
| | | |-- ...
|-- |-- testing
|-- label_02
| |-- 0000.txt
| |-- ...
|-- data_tracking_calib
`-- annotations
|-- tracking_train.json
|-- tracking_test.json
Please cite the corresponding References if you use the datasets.
@article{MOT16,
title = {{MOT}16: {A} Benchmark for Multi-Object Tracking},
shorttitle = {MOT16},
url = {http://arxiv.org/abs/1603.00831},
journal = {arXiv:1603.00831 [cs]},
author = {Milan, A. and Leal-Taix\'{e}, L. and Reid, I. and Roth, S. and Schindler, K.},
month = mar,
year = {2016},
note = {arXiv: 1603.00831},
keywords = {Computer Science - Computer Vision and Pattern Recognition}
}
@INPROCEEDINGS{Geiger2012CVPR,
author = {Andreas Geiger and Philip Lenz and Raquel Urtasun},
title = {Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite},
booktitle = {CVPR},
year = {2012}
}
@inproceedings{nuscenes2019,
title={{nuScenes}: A multimodal dataset for autonomous driving},
author={Holger Caesar and Varun Bankiti and Alex H. Lang and Sourabh Vora and Venice Erin Liong and Qiang Xu and Anush Krishnan and Yu Pan and Giancarlo Baldan and Oscar Beijbom},
booktitle={CVPR},
year={2020}
}
Scripts for training and evaluating DEFT on MOT, KITTI and nuScenes are available in the experiments folder. The outputs videos and results (same as submission format) will be on the folders $dataset_name$_videos and $dataset_name$_results.
A large portion of code is borrowed from xingyizhou/CenterTrack, shijieS/SST and Zhongdao/Towards-Realtime-MOT, many thanks to their wonderful work!