Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3569966.3570064acmotherconferencesArticle/Chapter ViewAbstractPublication PagescsseConference Proceedingsconference-collections
research-article

MOT-H: A Multi-Target Tracking Dataset Based on Horizontal View

Published: 20 December 2022 Publication History

Abstract

The computer vision field is quickly developing, including multiple object tracking, as the big data age approaches. The majority of the effort is focused on tracking methods while less attention is paid to the most important aspect, data. After an analysis of existing datasets, we find that they commonly ignore the breakpoint problem in tracking and have low image quality. Thus we present the dataset named Multiple Object Tracking on Horizontal view (MOT-H). MOT-H is meticulously annotated on crowded scenes from the horizontal view, with the primary goal of proving anti-jamming performance against complicated occlusion or even complete occlusion. The breakpoint issue is emphasized, which means the target object temporarily leaves the scene and returns after a while. The proposed MOT-H dataset has 10 sequences, 20,311 frames, and 337,440 annotation boxes in total, with all pictures having the resolution of 3840 × 2160 and being filmed at 30 frames per second (fps). We establish a fair benchmark for the future object tracking method development. The whole dataset can be found at: https://drive.google.com/drive/folders/1SCUJAdbqXQStyV-F2M9UyGfsuCaxR73a?usp=sharing.

References

[1]
Robert Collins, Xuhui Zhou, and Seng Keat Teh. 2005. An open source tracking testbed and evaluation web site. In IEEE International Workshop on Performance Evaluation of Tracking and Surveillance (IWPETS), Vol. 2. 35.
[2]
P. Dendorfer, H. Rezatofighi, A. Milan, J. Shi, D. Cremers, I. Reid, S. Roth, K. Schindler, and L. Leal-Taixé. 2019. CVPR19 Tracking and Detection Challenge: How crowded can it get?arXiv:1906.04567 [cs] (June 2019). http://arxiv.org/abs/1906.04567 arXiv:1906.04567.
[3]
P. Dendorfer, H. Rezatofighi, A. Milan, J. Shi, D. Cremers, I. Reid, S. Roth, K. Schindler, and L. Leal-Taixé. 2020. MOT20: A benchmark for multi object tracking in crowded scenes. arXiv:2003.09003[cs] (March 2020). http://arxiv.org/abs/1906.04567 arXiv:2003.09003.
[4]
James Ferryman and Ali Shahrokni. 2009. Pets2009: Dataset and challenge. In IEEE International Workshop on Performance Evaluation of Tracking and Surveillance (IWPETS). IEEE, 1–6.
[5]
Zheng Ge, Songtao Liu, Feng Wang, Zeming Li, and Jian Sun. 2021. YOLOX: Exceeding YOLO Series in 2021. arXiv preprint arXiv:2107.08430(2021).
[6]
Andreas Geiger, Philip Lenz, and Raquel Urtasun. 2012. Are we ready for autonomous driving? the kitti vision benchmark suite. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 3354–3361.
[7]
Ross Girshick. 2015. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV).
[8]
Erhan Gündoğdu and Abdullah Aydın Alatan. 2016. The Visual Object Tracking VOT2016 challenge results. In European Conference on Computer Vision Workshops (ECCVW).
[9]
Matej Kristan, Ales Leonardis, Jiri Matas, Michael Felsberg, Roman Pflugfelder, Luka ˇCehovin Zajc, Tomas Vojir, Gustav Hager, Alan Lukezic, Abdelrahman Eldesokey, 2017. The visual object tracking vot2017 challenge results. In Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW). 1949–1972.
[10]
L. Leal-Taixé, A. Milan, I. Reid, S. Roth, and K. Schindler. 2015. MOTChallenge 2015: Towards a Benchmark for Multi-Target Tracking. arXiv:1504.01942 [cs] (April 2015). http://arxiv.org/abs/1504.01942 arXiv:1504.01942.
[11]
A. Milan, L. Leal-Taixé, I. Reid, S. Roth, and K. Schindler. 2016. MOT16: A Benchmark for Multi-Object Tracking. arXiv:1603.00831 [cs] (March 2016). http://arxiv.org/abs/1603.00831 arXiv:1603.00831.
[12]
Matthias Muller, Adel Bibi, Silvio Giancola, Salman Alsubaihi, and Bernard Ghanem. 2018. TrackingNet: A Large-Scale Dataset and Benchmark for Object Tracking in the Wild. In European Conference on Computer Vision (ECCV).
[13]
Ergys Ristani, Francesco Solera, Roger Zou, Rita Cucchiara, and Carlo Tomasi. 2016. Performance measures and a data set for multi-target, multi-camera tracking. In European Conference on Computer Vision (ECCV). Springer, 17–35.
[14]
Rainer Stiefelhagen and John Garofolo. 2007. Multimodal Technologies for Perception of Humans: First International Evaluation Workshop on Classification of Events, Activities and Relationships, CLEAR 2006, Southampton, UK, April 6-7, 2006, Revised Selected Papers. Vol. 4122. Springer.
[15]
Longyin Wen, Dawei Du, Zhaowei Cai, Zhen Lei, Ming-Ching Chang, Honggang Qi, Jongwoo Lim, Ming-Hsuan Yang, and Siwei Lyu. 2020. UA-DETRAC: A new benchmark and protocol for multi-object detection and tracking. Computer Vision and Image Understanding (CVIU) 193 (2020), 102907.
[16]
Nicolai Wojke, Alex Bewley, and Dietrich Paulus. 2017. Simple Online and Realtime Tracking with a Deep Association Metric. In 2017 IEEE International Conference on Image Processing (ICIP). IEEE, 3645–3649. https://doi.org/10.1109/ICIP.2017.8296962
[17]
Yi Wu, Jongwoo Lim, and Ming-Hsuan Yang. 2013. Online Object Tracking: A Benchmark. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18]
Yifu Zhang, Peize Sun, Yi Jiang, Dongdong Yu, Zehuan Yuan, Ping Luo, Wenyu Liu, and Xinggang Wang. 2021. ByteTrack: Multi-Object Tracking by Associating Every Detection Box. arXiv preprint arXiv:2110.06864(2021).

Index Terms

  1. MOT-H: A Multi-Target Tracking Dataset Based on Horizontal View

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CSSE '22: Proceedings of the 5th International Conference on Computer Science and Software Engineering
    October 2022
    753 pages
    ISBN:9781450397780
    DOI:10.1145/3569966
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 20 December 2022

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Multiple object tracking
    2. benchmark.
    3. occlusion

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    CSSE 2022

    Acceptance Rates

    Overall Acceptance Rate 33 of 74 submissions, 45%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 34
      Total Downloads
    • Downloads (Last 12 months)13
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 26 Sep 2024

    Other Metrics

    Citations

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media