全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

基于深度学习的视频船舶目标追踪模型
Video Ship Target Tracking Model Based on Deep Learning

DOI: 10.12677/MOS.2024.131006, PP. 50-60

Keywords: 深度学习,船舶追踪,光流
Deep Learning
, Ship Tracking, Optical Flow

Full-Text   Cite this paper   Add to My Lib

Abstract:

船舶追踪是保证水上船舶交通安全的重要技术之一,船舶目标追踪模型可以对船舶的轨迹建模从而预测船只位置,监控船只交通安全,然而在船舶在交汇情境下,由于船舶被部分遮挡或处于复杂背景中,仅依赖视觉特征进行船舶追踪可能会导致追踪边界的不准确定位等问题。此外,当船只完全被遮挡并再次出现时,传统ReID方法往往不能为这些船只正确分配轨迹。为了解决这些问题,本文首先提出了一种融合光流场特征的单目标船舶追踪模型。该模型通过将视觉追踪特征与光流特征相结合,实现了更精确的目标边界定位。接着,通过结合Yolov5目标检测模型和经过改进的单目标追踪模型,实现了一种判别式多目标追踪模型,并采用更有效的匹配方法,一定程度上缓解了目标消失再出现id丢失的问题。实验表明,融合光流网络的目标追踪模型能更好地定位目标,而本文的联合多目标追踪框架在目标定位和ID分配方面均优于DeepSort、ByteTrack等追踪模型。
Ship tracking is one of the crucial technologies for ensuring the safety of maritime traffic. Ship tar-get tracking models can model the trajectory of ships to predict their positions and monitor mari-time traffic safety. However, in situations where ships intersect, issues such as partial ship obstruc-tions or complex backgrounds can lead to inaccurate tracking boundary positioning when relying solely on visual features. Additionally, traditional discriminative tracking models often struggle to correctly assign identities to ships when they reappear after being completely obscured. To address these challenges, this paper first introduces a single-object ship tracking model that integrates op-tical flow field features. This model combines visual tracking features with optical flow features, re-sulting in more precise target boundary localization. Subsequently, a joint multi-object tracking framework is implemented by combining the Yolov5 object detection model with an improved sin-gle-object tracking model. This framework also utilizes more effective matching methods, partially alleviating the issue of losing IDs when targets reappear. Experimental results demonstrate that the target tracking model incorporating optical flow performs better in target localization. Fur-thermore, the joint multi-object tracking framework presented in this paper outperforms tracking models like DeepSort and ByteTrack in both target localization and ID assignment.

References

[1]  Bewley, A., Ge, Z., Ott, L., et al. (2016) Simple Online and Real-Time Tracking. 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, 25-28 September 2016, 3464-3468.
https://doi.org/10.1109/ICIP.2016.7533003
[2]  Wojke, N., Bewley, A. and Paulus, D. (2017) Simple Online and Realtime Tracking with a Deep Association Metric. 2017 IEEE International Conference on Image Processing (ICIP), Beijing, 17-20 September 2017, 3645-3649.
https://doi.org/10.1109/ICIP.2017.8296962
[3]  Zhang, Y., Sun, P., Jiang, Y., et al. (2022) Bytetrack: Mul-ti-Object Tracking by Associating Every Detection Box. European Conference on Computer Vision, Tel Aviv, 23-27 October 2022, 1-21.
https://doi.org/10.1007/978-3-031-20047-2_1
[4]  Sun, P., Cao, J., Jiang, Y., et al. (2020) Transtrack: Multiple Object Tracking with Transformer.
https://doi.org/10.48550/arXiv.2012.15460
[5]  Meinhardt, T., Kirillov, A., Leal-Taixe, L., et al. (2022) Track-former: Multi-Object Tracking with Transformers. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, 18-24 June 2022, 8844-8854.
https://doi.org/10.1109/CVPR52688.2022.00864
[6]  Vaquero, L., Brea, V.M. and Mucientes, M. (2023) Re-al-Time Siamese Multiple Object Tracker with Enhanced Proposals. Pattern Recognition, 135, Article ID: 109141.
https://doi.org/10.1016/j.patcog.2022.109141
[7]  Cai, J., Xu, M., Li, W., et al. (2022) MeMOT: Multi-Object Tracking with Memory. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, 18-24 June 2022, 8090-8100.
https://doi.org/10.1109/CVPR52688.2022.00792
[8]  Bertinetto, L., Valmadre, J., Henriques, J.F., et al. (2016) Fully-Convolutional Siamese Networks for Object Tracking. Computer Vision ECCV 2016 Workshops, Amsterdam, 8-10 and 15-16 October 2016.
[9]  Bhat, G., Danelljan, M., Gool, L.V., et al. (2019) Learning Discriminative Model Prediction for Tracking. Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, 27 Octo-ber-2 November 2019, 6182-6191.
https://doi.org/10.1109/ICCV.2019.00628
[10]  Yan, B., Peng, H., Fu, J., et al. (2021) Learning Spatio-Temporal Transformer for Visual Tracking. Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, 10-17 October 2021, 10448-10457.
https://doi.org/10.1109/ICCV48922.2021.01028
[11]  Vaswani, A., Shazeer, N., Parmar, N., et al. (2017) Attention Is All You Need. Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, 4-9 December 2017, 6000-6010.
[12]  Chen, B., Li, P., Bai, L., et al. (2022) Backbone Is All Your Need: A Sim-plified Architecture for Visual Object Tracking. European Conference on Computer Vision, Tel Aviv, 23-27 October 2022, 375-392.
https://doi.org/10.1007/978-3-031-20047-2_22
[13]  Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al. (2020) An Image Is Worth 16×16 Words: Transformers for Image Recognition at Scale.
https://doi.org/10.48550/arXiv.2010.11929
[14]  Xu, H., Zhang, J., Cai, J., et al. (2022) Gmflow: Learning Optical Flow via Global Matching. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, 18-24 June 2022, 8121-8130.
https://doi.org/10.1109/CVPR52688.2022.00795
[15]  Redmon, J., Divvala, S., Girshick, R., et al. (2016) You Only Look Once: Unified, Real-Time Object Detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, 27-30 June 2016, 779-788.
https://doi.org/10.1109/CVPR.2016.91
[16]  Carion, N., Massa, F., Synnaeve, G., et al. (2020) End-to-End Object Detection with Transformers. European Conference on Computer Vision, Glasgow, 23-28 August 2020, 213-229.
https://doi.org/10.1007/978-3-030-58452-8_13
[17]  Liu, W., Anguelov, D., Erhan, D., et al. (2016) Ssd: Single Shot Multibox Detector. Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, 11-14 October 2016, 21-37.
https://doi.org/10.1007/978-3-319-46448-0_2
[18]  Lin, T.Y., Goyal, P., Girshick, R., et al. (2017) Focal Loss for Dense Object Detection. Proceedings of the IEEE International Conference on Computer Vision, Venice, 22-29 October 2017, 2980-2988.
https://doi.org/10.1109/ICCV.2017.324
[19]  Girshick, R., Donahue, J., Darrell, T., et al. (2014) Rich Feature Hi-erarchies for Accurate Object Detection and Semantic Segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, 23-28 June 2014, 580-587.
https://doi.org/10.1109/CVPR.2014.81
[20]  Girshick, R. (2015) Fast R-CNN. Proceedings of the IEEE Interna-tional Conference on Computer Vision, Santiago, 7-13 December 2015, 1440-1448.
https://doi.org/10.1109/ICCV.2015.169

Full-Text

comments powered by Disqus

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133

WeChat 1538708413