图像引导和点云空间约束的公路洒落物检测定位方法

蔡怀宇,杨朝乾,崔子扬,等. 图像引导和点云空间约束的公路洒落物检测定位方法[J]. 光电工程,2024,51(3): 230317. doi: 10.12086/oee.2024.230317
引用本文: 蔡怀宇,杨朝乾,崔子扬,等. 图像引导和点云空间约束的公路洒落物检测定位方法[J]. 光电工程,2024,51(3): 230317. doi: 10.12086/oee.2024.230317
Cai H Y, Yang Z Q, Cui Z Y, et al. Image-guided and point cloud space-constrained method for detection and localization of abandoned objects on the road[J]. Opto-Electron Eng, 2024, 51(3): 230317. doi: 10.12086/oee.2024.230317
Citation: Cai H Y, Yang Z Q, Cui Z Y, et al. Image-guided and point cloud space-constrained method for detection and localization of abandoned objects on the road[J]. Opto-Electron Eng, 2024, 51(3): 230317. doi: 10.12086/oee.2024.230317

图像引导和点云空间约束的公路洒落物检测定位方法

详细信息
    作者简介:
    *通讯作者: 陈晓冬,xdchen@tju.edu.cn
  • 中图分类号: TP277

Image-guided and point cloud space-constrained method for detection and localization of abandoned objects on the road

More Information
  • 公路洒落物是影响交通安全的重要因素之一,为了解决中小尺度公路洒落物检测中的漏检、误检以及难以定位等问题,本文提出了一种图像引导和点云空间约束的公路洒落物检测定位方法。该方法使用改进的YOLOv7-OD网络处理图像数据获取二维目标预测框信息,将目标预测框投影到激光雷达坐标系下得到锥形感兴趣区域(region of interest, ROI)。在ROI区域内的点云空间约束下,联合点云聚类和点云生成算法获得不同尺度的洒落物在三维空间中的检测定位结果。实验表明:改进的YOLOv7-OD网络在中尺度目标上的召回率和平均精度分别为85.4%和82.0%,相比YOLOv7网络分别提升6.6和8.0个百分点;在小尺度目标上的召回率和平均精度分别为66.8%和57.3%,均提升5.3个百分点;洒落物定位方面,对于距离检测车辆30~40 m处的目标,深度定位误差为0.19 m,角度定位误差为0.082°,实现了多尺度公路洒落物的检测和定位。

  • Overview: Highways constitute a vital economic lifeline for a nation. With the continuous increase in highway mileage and traffic volume, the significance of daily maintenance work on the road has become more pronounced. The detection and localization of abandoned objects on the road are among the primary tasks in highway maintenance. Because if abandoned objects are not promptly cleared, they can easily lead to traffic congestion or even cause accidents. Detecting and locating abandoned objects on the road is a specific object detection task. In order to fully leverage the advantages of both image and point cloud data, solutions based on multisensor fusion have become a research hotspot. However, due to the sparse nature of the LiDAR point clouds, existing multisensor fusion methods usually encounter challenges such as missed detection, false alarms, and difficulties in localization when detecting small-to-medium-sized abandoned objects. To address the aforementioned issues, this paper proposes a method for detecting and locating abandoned objects on the road using image guidance and point cloud spatial constraints. Firstly, on the foundation of the YOLOv7, a small object detection layer has been added, and a channel attention mechanism has been introduced to enhance the network's ability to extract two-dimensional bounding boxes for small-to-medium-sized targets within the image. Subsequently, the predicted bounding boxes are projected onto the LiDAR coordinate system to generate a pyramidal region of interest (ROI). For larger targets, sufficient point cloud data allows for three-dimensional spatial position estimation through point cloud clustering within the ROI. For smaller targets, which have insufficient point cloud data for clustering within the ROI, spatial constraints from surrounding ground point cloud data are used. Using projection transformation relationships, point cloud data is generated to obtain spatial position information for the smaller targets, achieving the detection and localization of multiscale abandoned objects on the road in three-dimensional space. The experimental results show that the improved YOLOv7-OD network achieves recall and average precision rates of 85.4% and 82.0%, respectively, for medium-sized objects, representing improvements of 6.6% and 8% compared to the YOLOv7. The recall and average precision rates for small-sized objects are 66.8% and 57.3%, respectively, with an increase of 5.3%. In terms of localization, for abandoned objects located 30~40 m away from the detecting vehicle, the depth localization error is 0.19 m, and the angular localization error is 0.082°. The proposed algorithm can process 36 frames of data per second, effectively achieving real-time detection and localization of abandoned objects on the road.

  • 加载中
  • 图 1  所提方法的总体框架

    Figure 1.  The overall framework of the proposed method

    图 2  YOLOv7-OD网络结构

    Figure 2.  The network architecture of YOLOv7-OD

    图 3  SOD Layer网络结构

    Figure 3.  The network structure of SOD Layer

    图 4  SDK Attention模块

    Figure 4.  SDK Attention module

    图 5  根据图像目标检测框生成激光雷达坐标系下的ROI区域

    Figure 5.  Generation of ROI areas in the LiDAR coordinate system based on image object detection bounding boxes

    图 6  实验装置结构图

    Figure 6.  Structural diagram of the experimental device

    图 7  滤波结果图。(a)原始点云数据;(b)视场匹配获得有效点云数据;(c) CSF滤波后的非地面点云;(d) CSF滤波后的地面点云

    Figure 7.  Filtering results. (a) Original point cloud data; (b) Effective point cloud data obtained by field-of-view matching; (c) Non-ground point cloud after CSF filtering; (d) Ground point cloud after CSF filtering

    图 8  各评价指标在空间中的具体含义

    Figure 8.  The specific meanings of various evaluation metrics in space

    图 9  公路洒落物检测定位结果(场景一,截取部分区域)。(a)待测场景图像;(b) Method A的检测定位结果;(c)本文所提算法的检测定位结果

    Figure 9.  Detection and localization results for abandoned objects on the road (Scene one, selected partial area). (a) Image of the scene to be tested; (b) Detection and localization results by method A; (c) Detection and localization results by our method

    图 10  公路洒落物检测定位结果(场景二,截取部分区域)。(a)待测场景图像;(b) Method A的检测定位结果;(c)本文所提算法的检测定位结果

    Figure 10.  Detection and localization results for abandoned objects on the road (Scene two, selected partial area). (a) Image of the scene to be tested; (b) Detection and localization results by method A; (c) Detection and localization results by our method

    图 11  公路洒落物检测与定位实验结果。(a)场景一;(b)场景二

    Figure 11.  Experimental results for detecting and locating abandoned objects on the road. (a) Scene one; (b) Scene two

    表 1  WOD数据集上的消融实验

    Table 1.  Ablation experiments on the WOD dataset

    YOLOv7SOD
    Layer
    SDK
    Attention
    AP/% AP/%mAP0.5
    /%
    mAP0.5:0.95
    /%
    smallmediumlarge smallmediumlarge
    10.0035.8070.40 23.2047.2075.6057.1532.30
    11.9038.9066.9026.0049.7073.1059.2834.18
    11.2037.6066.8024.3048.7073.0058.1233.21
    12.0039.0067.7026.2050.8073.2059.4634.33
    下载: 导出CSV

    表 2  自制数据集上的消融实验

    Table 2.  Ablation experiments on custom dataset

    YOLOv7SOD
    Layer
    SDK
    Attention
    AP/% AP/%mAP0.5
    /%
    mAP0.5:0.95
    /%
    smallmediumlarge smallmediumlarge
    52.0074.0085.30 61.5078.8089.0094.2064.80
    55.0079.8092.7065.4084.0095.5094.3069.80
    54.1081.5093.4063.2085.3095.1093.8070.00
    57.3082.0092.0066.8085.4093.9095.3071.90
    下载: 导出CSV

    表 3  不同网络模型的其他评价指标

    Table 3.  Additional evaluation metrics for different network models

    YOLOv7SOD LayerSDK AttentionParams/MBGFLOPsFPS
    71.3103.382.2
    51.4108.273.2
    73.7118.375.4
    52123.365.8
    下载: 导出CSV

    表 4  YOLOv7-OD与其他目标检测算法的对比实验

    Table 4.  Comparative experiments of YOLOv7-OD with other object detection algorithms

    ModelParams
    /MB
    AP/% AP/%mAP0.5
    /%
    mAP0.5:0.95
    /%
    smallmediumlarge smallmediumlarge
    Faster-RCNN[2]79.03.70%19.0%41.7% 9.00%26.2%46.6%29.7%15.6%
    RetinaNet[7]61.73.00%29.8%64.7%10.6%40.4%71.4%43.9%24.7%
    YOLOX[22]1047.40%31.3%68.1%13.9%39.2%72.8%48.1%28.0%
    DETR[3]79.04.83%26.4%62.2%11.3%35.7%68.7%45.4%24.0%
    YOLOv3[5]1193.00%26.2%64.1%5.80%35.0%69.8%41.9%22.9%
    YOLOv5 [23]88.56.60%31.4%64.1%13.3%45.5%71.1%48.6%27.2%
    YOLOv6[6]72.47.50%37.5%71.7%16.7%46.4%78.0%51.9%31.0%
    YOLOv7[18]71.310.0%35.8%70.4%23.2%47.2%75.6%57.2%32.3%
    YOLOv8[24]83.68.30%38.9%71.3%17.7%46.7%76.5%53.0%32.0%
    YOLOv7-OD52.112.0%39.0%67.7%26.2%50.8%73.2%59.5%34.3%
    下载: 导出CSV

    表 5  两种点云定位方法的实验结果

    Table 5.  Experimental results of two point cloud localization methods

    MethodN (road objects)N (abandoned object)N (predicted)N (true)N (false)Precision/%Recall/%
    Method A920270150147398.0055.56
    Ours920270258250896.9095.56
    下载: 导出CSV

    表 6  不同距离下通过点云生成方法进行洒落物定位的误差

    Table 6.  Error in abandoned objects localization using point cloud generation method at different distances

    Distance/mMAE-error
    ΔD/mΔW/mΔθ/(°)
    0~200.1320.00990.195
    20~300.1560.01210.115
    30~400.1880.01620.0819
    Over 400.2230.02710.0541
    Total0.1810.02180.122
    下载: 导出CSV
  • [1]

    Liu W, Anguelov D, Erhan D, et al. SSD: single shot multibox detector[C]//14th European Conference on Computer Vision, Amsterdam, 2016: 21–37. https://doi.org/10.1007/978-3-319-46448-0_2.

    [2]

    Lin T Y, Goyal P, Girshick R, et al. Focal loss for dense object detection[C]//2017 IEEE International Conference on Computer Vision (ICCV), Venice, 2017: 2999–3007. https://doi.org/10.1109/ICCV.2017.324.

    [3]

    Carion N, Massa F, Synnaeve G, et al. End-to-end object detection with transformers[C]//16th European Conference on Computer Vision, Glasgow, 2020: 213–229. https://doi.org/10.1007/978-3-030-58452-8_13.

    [4]

    Bochkovskiy A, Wang C Y, Liao H Y M. YOLOv4: optimal speed and accuracy of object detection[Z]. arXiv: 2004.10934, 2020. https://doi.org/10.48550/arXiv.2004.10934.

    [5]

    Redmon J, Farhadi A. YOLOv3: an Incremental Improvement[Z]. arXiv: 1804.02767, 2018. https://doi.org/10.48550/arXiv.1804.02767.

    [6]

    Li C Y, Li L L, Jiang H L, et al. YOLOv6: a single-stage object detection framework for industrial applications[Z]. arXiv: 2209.02976, 2022. https://doi.org/10.48550/arXiv.2209.02976.

    [7]

    Ren S Q, He K M, Girshick R, et al. Faster R-CNN: towards real-time object detection with region proposal networks[J]. IEEE Trans Pattern Anal Mach Intell, 2017, 39(6): 1137−1149. doi: 10.1109/TPAMI.2016.2577031

    [8]

    He K M, Gkioxari G, Dollár P, et al. Mask R-CNN[C]//2017 IEEE International Conference on Computer Vision (ICCV), Venice, 2017: 2980–2988. https://doi.org/10.1109/ICCV.2017.322.

    [9]

    金瑶, 张锐, 尹东. 城市道路视频中小像素目标检测[J]. 光电工程, 2019, 46(9): 190053.

    Jin Y, Zhang R, Yin D. Object detection for small pixel in urban roads videos[J]. Opto-Electron Eng, 2019, 46(9): 190053.

    [10]

    Charles R Q, Hao S, Mo K C, et al. PointNet: deep learning on point sets for 3D classification and segmentation[C]//2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, 2017: 77–85. https://doi.org/10.1109/CVPR.2017.16.

    [11]

    Qi C R, Yi L, Su H, et al. PointNet++: deep hierarchical feature learning on point sets in a metric space[C]//Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, 2017: 5105–5114. https://doi.org/10.5555/3295222.3295263.

    [12]

    Qian G C, Li Y C, Peng H W, et al. PointNeXt: Revisiting pointnet++ with improved training and scaling strategies[C]//36th Conference on Neural Information Processing Systems, New Orleans, 2022: 23192–23204.

    [13]

    Qi C R, Liu W, Wu C X, et al. Frustum PointNets for 3D object detection from RGB-D data[C]//2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018: 918–927. https://doi.org/10.1109/CVPR.2018.00102.

    [14]

    Vora S, Lang A H, Helou B, et al. PointPainting: sequential fusion for 3D object detection[C]//2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 2020: 4603–4611. https://doi.org/10.1109/CVPR42600.2020.00466.

    [15]

    梁浩林, 蔡怀宇, 刘博翀, 等. 基于图像与点云融合的公路撒落物检测算法[J]. 激光与光电子学进展, 2023, 60(10): 1010001. doi: 10.3788/LOP213044

    Liang H L, Cai H Y, Liu B C, et al. Road falling objects detection algorithm based on image and point cloud fusion[J]. Laser Optoelectron Prog, 2023, 60(10): 1010001. doi: 10.3788/LOP213044

    [16]

    郑欣悦, 赖际舟, 吕品, 等. 基于红外视觉/激光雷达融合的目标识别与定位方法[J]. 导航定位与授时, 2021, 8(3): 34−41. doi: 10.19306/j.cnki.2095-8110.2021.03.005

    Zheng X Y, Lai J Z, Lv P, et al. Object detection and positioning method based on infrared vision/lidar fusion[J]. Navig Position Timing, 2021, 8(3): 34−41. doi: 10.19306/j.cnki.2095-8110.2021.03.005

    [17]

    范婉舒, 赖际舟, 吕品, 等. 基于改进DeepSORT的视觉/激光雷达目标跟踪与定位方法[J]. 导航定位与授时, 2022, 9(4): 77−84. doi: 10.19306/j.cnki.2095-8110.2022.04.009

    Fan W S, Lai J Z, Lv P, et al. Vision/lidar object tracking and localization method based on improved DeepSORT[J]. Navig Position Timing, 2022, 9(4): 77−84. doi: 10.19306/j.cnki.2095-8110.2022.04.009

    [18]

    Wang C Y, Bochkovskiy A, Liao H Y M. YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors[C]//2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, 2023: 7464–7475. https://doi.org/10.1109/CVPR52729.2023.00721.

    [19]

    Zhang W M, Qi J B, Wan P, et al. An easy-to-use airborne LiDAR data filtering method based on cloth simulation[J]. Remote Sens, 2016, 8(6): 501. doi: 10.3390/rs8060501

    [20]

    Huang J K, Grizzle J W. Improvements to target-based 3D LiDAR to camera calibration[J]. IEEE Access, 2020, 8: 134101−134110. doi: 10.1109/ACCESS.2020.3010734

    [21]

    Sun P, Kretzschmar H, Dotiwalla X, et al. Scalability in perception for autonomous driving: waymo open dataset[C]//2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, 2020: 2443–2451. https://doi.org/10.1109/CVPR42600.2020.00252.

    [22]

    Ge Z, Liu S T, Wang F, et al. YOLOX: exceeding YOLO series in 2021[Z]. arXiv: 2107.08430, 2021. https://doi.org/10.48550/arXiv.2107.08430.

    [23]

    Jocher G, Chaurasia A, Stoken A, et al. ultralytics/yolov5: v7.0 - YOLOv5 SOTA Realtime Instance Segmentation[EB/OL]. (2022-11-22). https://github.com/ultralytics/yolov5.

    [24]

    Jocher G, Chaurasia A, Qiu J. Ultralytics YOLO[EB/OL]. (2023-01-10). https://github.com/ultralytics/ultralytics.

  • 加载中

(12)

(6)

计量
  • 文章访问数: 
  • PDF下载数: 
  • 施引文献:  0
出版历程
收稿日期:  2023-12-27
修回日期:  2024-02-29
录用日期:  2024-03-01
刊出日期:  2024-04-05

目录

/

返回文章
返回