全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

融合注意力机制的电力巡检目标检测模型研究
Research on Target Detection Model for Electric Power Inspection Based on Fusion Attention Mechanism

DOI: 10.12677/airr.2024.132028, PP. 265-271

Keywords: 电力巡检,目标检测,Swin Transformer
Electric Power Inspection
, Object Detection, Swin Transformer

Full-Text   Cite this paper   Add to My Lib

Abstract:

针对传统电力电网设备巡检方法存在的巡检效率低和安全隐患多等问题,提出了一种基于人工智能技术的无人巡检方法。通过引入Swin Transformer模型,优化了目标检测算法,提高了巡检的精度和实时性。首先,分析了电力电网设备巡检的现状和传统方法的局限性,并对无人巡检系统的实际应用挑战和未来发展趋势进行了探讨,提出了以深度学习、计算机视觉为核心的技术框架。采用图像增强技术扩充了数据集,并手动标注获取了高质量数据集。将Faster RCNN与Swin Transformer结合,应用于自制数据集,实现了高效稳定的目标检测。与传统方法相比显著提升了巡检效率,降低了漏检率和误检率。本研究成果为电力行业的数字化转型和升级提供了理论和实践价值。
A unmanned inspection method based on artificial intelligence technology is proposed to address the problems of low inspection efficiency and multiple safety hazards in traditional inspection methods for power grid equipment. By introducing the Swin Transformer model, the object detection algorithm has been optimized, improving the accuracy and real-time performance of inspections. Firstly, the current situation of power grid equipment inspection and the limitations of traditional methods were analyzed, and the practical application challenges and future development trends of unmanned inspection systems were discussed. A technical framework centered on deep learning and computer vision was proposed. We expanded the dataset using image enhancement technology and manually annotated it to obtain high-quality datasets. The improved Faster RCNN was combined with Swin Transformer and applied to self-made datasets to achieve efficient and stable object detection. Compared with traditional methods, it significantly improves inspection efficiency, reduces missed detection rates and false detection rates. The results of this study provide theoretical and practical value for the digital transformation and upgrading of the power industry.

References

[1]  Lowe, D.G. (2004) Distinctive Image Features from Scale-Invariant Keypoints. International Journal of Computer Vision, 60, 91-110.
https://doi.org/10.1023/B:VISI.0000029664.99615.94
[2]  Dalal, N. and Triggs, B. (2005) Histograms of Oriented Gradients for Human Detection. 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), 1, 886-893.
[3]  Girshick, R., Donahue, J., Darrell, T. and Malik, J. (2014) Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, 23-28 June 2014, 580-587.
https://doi.org/10.1109/CVPR.2014.81
[4]  Girshick, R. (2015) Fast R-CNN. 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, 7-13 December 2015, 1440-1448.
https://doi.org/10.1109/ICCV.2015.169
[5]  Ren, S., He, K., Girshick, R. and Sun, J. (2017) Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39, 1137-1149.
https://doi.org/10.1109/TPAMI.2016.2577031
[6]  Redmon, J., Divvala, S., Girshick, R. and Farhadi, A. (2016) You Only Look Once: Unified, Real-Time Object Detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, 27-30 June 2016, 779-788.
https://doi.org/10.1109/CVPR.2016.91
[7]  Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y. and Berg, A.C. (2016) SSD: Single Shot MultiBox Detector. Proceedings of the European Conference on Computer Vision (ECCV), Springer, Cham., 21-37.
https://doi.org/10.1007/978-3-319-46448-0_2
[8]  Lin, T.-Y., Goyal, P., Girshick, R., He, K. and Dollár, P. (2017) Focal Loss for Dense Object Detection. 2017 IEEE International Conference on Computer Vision (ICCV), Venice, 22-29 October 2017, 2999-3007.
https://doi.org/10.1109/ICCV.2017.324
[9]  Vaswani, A., Shazeer, N., Parmar, N., et al. (2017) Attention Is All You Need. Neural Information Processing Systems (NIPS), Long Beach, CA, 4-9 December 2017, 5998-6008.
https://doi.org/10.48550/arXiv.1706.03762
[10]  Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al. (2021) An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. International Conference on Learning Representations (lCLR), Vienna, 3-7 May 2021.
https://doi.org/10.48550/arXiv.2010.11929
[11]  Liu, Z., et al. (2021) Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, 10-17 October 2021, 9992-10002.
https://doi.org/10.1109/ICCV48922.2021.00986
[12]  Lin, T.-Y., Dollár, P., Girshick, R., He, K., Hariharan, B. and Belongie, S. (2017) Feature Pyramid Networks for Object Detection. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, 21-26 July 2017, 936-944.
https://doi.org/10.1109/CVPR.2017.106

Full-Text

comments powered by Disqus

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133

WeChat 1538708413