Journal of Jilin University(Engineering and Technology Edition) ›› 2024, Vol. 54 ›› Issue (12): 3589-3600.doi: 10.13229/j.cnki.jdxbgxb.20230105

Previous Articles     Next Articles

3D object detection algorithm fusing dense connectivity and Gaussian distance

Xin CHENG1,2(),Sheng-xian LIU1,Jing-mei ZHOU3(),Zhou ZHOU1,Xiang-mo ZHAO1   

  1. 1.School of Information Engineering,Chang'an University,Xi'an 710018,China
    2.Traffic Management Research Institute of the Ministry of Public Security,Wuxi 214151,China
    3.School of Electronics and Control Engineering,Chang'an University,Xi'an 710018,China
  • Received:2023-02-06 Online:2024-12-01 Published:2025-01-24
  • Contact: Jing-mei ZHOU E-mail:xincheng@chd.edu.cn;jmzhou@chd.edu.cn

Abstract:

To enhance the perception of small objects, based on the F-PointNet network,the FDG-PointNet 3D object detection model is proposed by combining dense connection and Gaussian distance features. Gaussian distance features is fused as additional attention features, and it effectively solves the low accuracy of instance segmentation in the F-PointNet network and enhances the noise filtering in the point cloud view cone. Based on the characteristics that dense connection can enhance feature extraction, the dense connection is used to improve PointNet++ network and enhance point cloud feature reuse. It alleviates low degree of feature extraction and gradient disappearance for small objects in the feature extraction process, and improves the accuracy of 3D object bounding box regression. The experimental results show that the proposed algorithm outperforms the benchmark method F-PointNet in three levels (easy, moderate, and hard) for the detection of car, pedestrian, and cyclist, which can achieve the average detection accuracy of 71.12%, 61.23%, and 55.71% for car, pedestrian, and cyclist at moderate level. It has the most significant improvement for pedestrian detection, and can increase 5.5% and 3.1% at easy and moderate levels, respectively. In summary,compared to F-PointNet algorithm, the proposed FDG-PointNet algorithm effectively solves the low accuracy of small objects detection and has strong applicability.

Key words: computer science and technology, 3D object detection, lidar, dense connectivity, gaussian distance

CLC Number: 

  • TP391.4

Fig.1

PointNet++ networks combined with densely connected structures"

Fig.2

Dense set abstraction module"

Fig.3

F-PointNet algorithm"

Fig.4

Masking situation"

Fig.5

FDG-PointNet network structure"

Table 1

Classification results on ModelNet 40 dataset"

方法输入总体精度OA/%
Subvolumevoxels89.2
PointNet11point cloud89.2
Kd-net26point cloud90.6
SO-net27point cloud90.9
PointNet++12point cloud90.7
本文point cloud91.3

Table 2

Part segmentation results on ShapeNet dataset"

方法meanaerobagcapcarchair

ear

phone

guitarknifelamplaptopmotormugpistolrocket

skate

board

table
PointNet[1183.783.478.782.574.989.673.091.585.980.895.365.293.081.257.972.880.6
SSCN2884.781.681.781.975.290.274.993.086.184.795.666.792.781.660.682.982.1
PointNet++1285.182.479.087.777.390.871.891.085.983.795.371.694.181.358.776.482.6
本文85.382.682.686.378.191.071.190.887.683.196.072.094.481.058.276.881.6

Table 3

Target detection accuracy of F-PointNet (v1) fused Gaussian distance features(BEV)"

模型汽车行人骑行人
EasyModerateHardEasyModerateHardEasyModerateHard
F-PointNet(v1)87.8282.4474.5070.3862.0756.7776.7257.7152.88
本文87.9082.2474.5371.6165.0757.3978.5558.4554.95

Table 4

Target detection accuracy of F-PointNet (v1) fused Gaussian distance features(3D)"

模型汽车行人骑行人
EasyModerateHardEasyModerateHardEasyModerateHard
F-PointNet(v1)83.2669.2862.5666.0457.0549.8470.5251.5050.95
本文83.8369.7363.1067.0857.9050.9071.1352.5949.40

Table 5

Impact of each part on 3D target detection"

密集连接高斯距离汽车行人骑行人
EasyModerateHardEasyModerateHardEasyModerateHard
83.7670.9263.6564.7558.1151.2373.1554.8951.62
81.1871.3463.8268.1559.8452.4775.0355.9952.74
83.8170.6463.7766.9959.1052.2576.8555.8151.67
84.3971.1264.2870.2161.2353.2775.3755.7151.71

Table 6

Comparison of different 3D target detection algorithms AP"

模型汽车行人骑行人
EasyModerateHardEasyModerateHardEasyModerateHard
MV3D(BV+FV)1871.1956.6055.30N/AN/AN/AN/AN/AN/A
MV3D(BV+FV+RGB)1871.2962.6856.56N/AN/AN/AN/AN/AN/A
RT3D[29]72.8561.6464.38N/AN/AN/AN/AN/AN/A
PointFusion2277.9263.0053.2733.3628.0423.3849.3429.4226.98
VoxelNet581.9765.4662.8557.8653.4248.8767.1747.6545.11
F-PointNet(v1)2083.2669.2862.5666.0457.0549.8470.5251.5050.95
F-PointNet(v2)2083.7670.9263.6564.7558.1151.2373.1554.8951.62
本文84.3971.1264.2870.2161.2353.2775.3755.7151.71

Table 7

Computate cost comparison experiment"

模型视锥体生成/ms

实例

分割/ms

边界框回归/ms总时间/ms

模型

大小/MB

F-PointNet(v2)60881916750
本文64902017451

Fig.6

Experimental results of FDG-PointNet model"

1 Cheng X, Zhou J M, Liu P Y, et al. 3D vehicle object tracking algorithm based on bounding box similarity measurement[J]. IEEE Transactions on Intelligent Transportation Systems, 2023, 24(12): 15844-15854.
2 Cheng X, Zhou J M, Song J C, et al. A highway traffic image enhancement algorithm based on improved GAN in complex weather conditions[J]. IEEE Transactions on Intelligent Transportation Systems, 2023, 24(8): 8716-8726.
3 Yurtsever E, Lambert J, Carballo A, et al. A survey of autonomous driving: common practices and emerging technologies[J]. IEEE Access, 2020, 8: 58443-58469.
4 Cui Y D, Chen R, Chu W B, et al. Deep learning for image and point cloud fusion in autonomous driving: a review[J]. IEEE Transactions on Intelligent Transportation Systems, 2022, 23(2): 722-739.
5 Zhou Y, Tuzel O. VoxelNet: end-to-end learning for point cloud based 3D object detection[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 4490-4499.
6 Lang A H, Vora S, Caesar H, et al. PointPillars: fast encoders for object detection from point clouds [C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, USA, 2019: 12697-12705.
7 Yan Y, Mao Y X, Li B. SECOND: sparsely embedded convolutional detection[J]. Sensors, 2018, 18(10): 3337.
8 Deng J J, Shi S S, Li P W, et al. Voxel R-CNN: towards high performance voxel-based 3D object detection[J]. Proceedings of the AAAI Conference on Artificial Intelligence, 2021, 35(2): 1201-1209.
9 Li B. 3D fully convolutional network for vehicle detection in point cloud[C]∥2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, Canada, 2017: 1513-1518.
10 程鑫, 王宏飞, 周经美, 等. 基于体素柱形的激光雷达点云车辆目标检测算法[J]. 中国公路学报, 2023, 36(3): 247-260.
Cheng Xin, Wang Hong-fei, Zhou Jing-mei, et al. Vehicle detection algorithm based on voxel pillars from lidar point clouds[J]. China Journal of Highway and Transport, 2023, 36(3): 247-260.
11 Qi C R, Su H, Mo K, et al. PointNet: deep learning on point sets for 3D classification and segmentation[C]∥2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, USA, 2017: 77-85.
12 Qi C R, Yi L, Su H, et al. PointNet++: deep hierarchical feature learning on point sets in a metric space[C]∥Advances in Neural Information Processing Systems, Long Beach, USA, 2017: 5105 - 5114 .
13 Shi S S, Wang X G, Li H S. PointRCNN: 3D object proposal generation and detection from point cloud[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, USA, 2019: 770-779.
14 Yang Z T, Sun Y A, Liu S, et al. 3DSSD: point-based 3D single stage object detector[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2020: 11040-11048.
15 Yang Z T, Sun Y A, Liu S, et al. STD: sparse-to-dense 3D object detector for point cloud[C]∥Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, USA, 2019: 1951-1960.
16 Shi W J, Rajkumar R. Point-GNN: graph neural network for 3D object detection in a point cloud[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle,USA, 2020: 1711-1719.
17 Yin T W, Zhou X Y, Krahenbuhl P. Center-based 3D object detection and tracking[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, USA, 2021: 11784-11793.
18 Chen X Z, Ma H M, Wan J, et al. Multi-view 3D object detection network for autonomous driving[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 2017: 1907-1915.
19 Ku J, Mozifian M, Lee J, et al. Joint 3D proposal generation and object detection from view aggregation[C]∥2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 2018: No.8594049.
20 Qi C R, Liu W, Wu C X, et al. Frustum PointNets for 3D object detection from RGB-D data[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 918-927.
21 Vora S, Lang A H, Helou B, et al. PointPainting: sequential fusion for 3D object detection[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, USA, 2020: 4604-4612.
22 Xu D F, Anguelov D, Jain A. PointFusion: deep sensor fusion for 3D bounding box estimation[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 244-253.
23 Wang Z X, Jia K. Frustum ConvNet: sliding frustums to aggregate local point-wise features for amodal 3D object detection[C]∥2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 2019: 1742-1749.
24 Huang G, Liu Z, Maaten L, et al. Densely connected convolutional networks[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, USA, 2017: 4700-4708.
25 Geiger A, Lenz P, Stiller C, et al. Vision meets robotics: the KITTI dataset[J]. The International Journal of Robotics Research, 2013, 32(11): 1231-1237.
26 Klokov R, Lempitsky V. Escape from cells: deep kd-networks for the recognition of 3D point cloud models[C]∥Proceedings of the IEEE International Conference on Computer Vision, Honolulu, USA, 2017: 863-872.
27 Li J X, Chen B M, Lee G H. SO-Net: self-organizing network for point cloud analysis[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 9397-9406.
28 Graham B, Engelcke M, Maaten L. 3D semantic segmentation with submanifold sparse convolutional networks[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, USA, 2018: 9224-9232.
[1] Bo TAO,Fu-wu YAN,Zhi-shuai YIN,Dong-mei WU. 3D object detection based on high⁃precision map enhancement [J]. Journal of Jilin University(Engineering and Technology Edition), 2023, 53(3): 802-809.
[2] Xue-mei LI,Chun-yang WANG,Xue-lian LIU,Da XIE. Time delay estimation of linear frequency-modulated continuous-wave lidar signals via SESTH [J]. Journal of Jilin University(Engineering and Technology Edition), 2022, 52(4): 950-958.
[3] Yi-cheng YU,Liang HU,Ling CHI,Jian-feng CHU. An anonymous authentication protocol for single⁃server architectures [J]. Journal of Jilin University(Engineering and Technology Edition), 2021, 51(2): 659-666.
[4] Chang-fu ZONG,Long WEN,Lei HE. Object detection based on Euclidean clustering algorithm with 3D laser scanner [J]. Journal of Jilin University(Engineering and Technology Edition), 2020, 50(1): 107-113.
[5] PAN Hai-yang, LIU Shun-an, YAO Yong-ming. Depth information-basd autonomous aerial refueling [J]. 吉林大学学报(工学版), 2014, 44(6): 1750-1756.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
[1] LI Shoutao, LI Yuanchun. Autonomous Mobile Robot Control Algorithm Based on Hierarchical Fuzzy Behaviors in Unknown Environments[J]. 吉林大学学报(工学版), 2005, 35(04): 391 -397 .
[2] Liu Qing-min,Wang Long-shan,Chen Xiang-wei,Li Guo-fa. Ball nut detection by machine vision[J]. 吉林大学学报(工学版), 2006, 36(04): 534 -538 .
[3] Li Hong-ying; Shi Wei-guang;Gan Shu-cai. Electromagnetic properties and microwave absorbing property
of Z type hexaferrite Ba3-xLaxCo2Fe24O41
[J]. 吉林大学学报(工学版), 2006, 36(06): 856 -0860 .
[4] Zhang Quan-fa,Li Ming-zhe,Sun Gang,Ge Xin . Comparison between flexible and rigid blank-holding in multi-point forming[J]. 吉林大学学报(工学版), 2007, 37(01): 25 -30 .
[5] Yang Shu-kai, Song Chuan-xue, An Xiao-juan, Cai Zhang-lin . Analyzing effects of suspension bushing elasticity
on vehicle yaw response character with virtual prototype method
[J]. 吉林大学学报(工学版), 2007, 37(05): 994 -0999 .
[6] . [J]. 吉林大学学报(工学版), 2007, 37(06): 1284 -1287 .
[7] Che Xiang-jiu,Liu Da-you,Wang Zheng-xuan . Construction of joining surface with G1 continuity for two NURBS surfaces[J]. 吉林大学学报(工学版), 2007, 37(04): 838 -841 .
[8] Liu Han-bing, Jiao Yu-ling, Liang Chun-yu,Qin Wei-jun . Effect of shape function on computing precision in meshless methods[J]. 吉林大学学报(工学版), 2007, 37(03): 715 -0720 .
[9] . [J]. 吉林大学学报(工学版), 2007, 37(04): 0 .
[10] Li Yue-ying,Liu Yong-bing,Chen Hua . Surface hardening and tribological properties of a cam materials[J]. 吉林大学学报(工学版), 2007, 37(05): 1064 -1068 .