• 中文核心
  • EI
  • 中国科技核心
  • Scopus
  • CSCD
  • 英国科学文摘

留言板

尊敬的读者、作者、审稿人, 关于本刊的投稿、审稿、编辑和出版的任何问题, 您可以本页添加留言。我们将尽快给您答复。谢谢您的支持!

姓名
邮箱
手机号码
标题
留言内容
验证码

基于结构频谱感知框架的配电网点云语义分割

唐友源 张辉 杜瑞 张恺宁 曹云康 别克扎提·巴合提 陈厚权 王耀南

唐友源, 张辉, 杜瑞, 张恺宁, 曹云康, 别克扎提·巴合提, 陈厚权, 王耀南. 基于结构频谱感知框架的配电网点云语义分割. 自动化学报, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c250540
引用本文: 唐友源, 张辉, 杜瑞, 张恺宁, 曹云康, 别克扎提·巴合提, 陈厚权, 王耀南. 基于结构频谱感知框架的配电网点云语义分割. 自动化学报, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c250540
Tang You-Yuan, Zhang Hui, Du Rui, Zhang Kai-Ning, Cao Yun-Kang, Biekezati Baheti, Chen Hou-Quan, Wang Yao-Nan. Semantic segmentation of distribution network point clouds based on a structure spectrum-aware framework. Acta Automatica Sinica, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c250540
Citation: Tang You-Yuan, Zhang Hui, Du Rui, Zhang Kai-Ning, Cao Yun-Kang, Biekezati Baheti, Chen Hou-Quan, Wang Yao-Nan. Semantic segmentation of distribution network point clouds based on a structure spectrum-aware framework. Acta Automatica Sinica, xxxx, xx(x): x−xx doi: 10.16383/j.aas.c250540

基于结构频谱感知框架的配电网点云语义分割

doi: 10.16383/j.aas.c250540 cstr: 32138.14.j.aas.c250540
基金项目: 国家自然科学基金重大项目(62595801), 湖南省十大技术攻关项目(2024GK1010), 湖南省自然科学基金重点项目(2025JJ30024), 国网湖南省电力有限公司科技项目(5216A522001Y、5216A5240003、5216AJ250008)资助
详细信息
    作者简介:

    唐友源:长沙理工大学人工智能学院硕士研究生. 2023年获得湘南学院学士学位. 主要研究方向为点云智能感知. E-mail: tangyouyuan@stu.csust.edu.cn

    张辉:湖南大学人工智能与机器人学院教授. 2007年获得湖南大学博士学位. 主要研究方向为数据分析, 图像处理和机器人控制. 本文通信作者. E-mail: zhanghui1983@hnu.edu.cn

    杜瑞:湖南大学人工智能与机器人学院博士研究生. 2020年获得湘潭大学硕士学位. 主要研究方向为数据分析, 图像处理. E-mail: durui@hnu.edu.cn

    张恺宁:湖南大学人工智能与机器人学院助理教授. 2025年获得武汉大学通信与信息系统博士学位, 主要研究方向为3D视觉、多模态感知. E-mail: carney@hnu.edu.cn

    曹云康:湖南大学人工智能与机器人学院助理教授. 2025年获得华中科技大学机械工程博士学位. 主要研究方向为复杂场景下的工业视觉检测与识别、工业基础模型及工业具身智能. E-mail: caoyunkang@hnu.edu.cn

    别克扎提·巴合提:湖南大学人工智能与机器人学院博士研究生. 2020年获得英国克兰菲尔德大学硕士学位. 主要研究方向为信号分析, 时间序列分析. E-mail: bbiekezati@hnu.edu.cn

    陈厚权:湖南大学人工智能与机器人学院博士研究生. 2022年获得中南林业科技大学计算机信息与工程学院计算机系学士学位. 主要研究方向为电力场景的多模态技术. E-mail: chq@hnu.edu.cn

    王耀南:中国工程院院士, 湖南大学人工智能与机器人学院教授. 1995年获得湖南大学博士学位. 主要研究方向为数据分析, 智能控制和图像处理. E-mail: yaonan@hnu.edu.cn

Semantic Segmentation of Distribution Network Point Clouds Based on a Structure Spectrum-Aware Framework

Funds: Major Program of the National Natural Science Foundation of China (62595801); Ten Technical Research Projects of Hunan Province (2024GK1010); Natural Science Foundation of Hunan Province (2025JJ30024); Science and Technology Project of State Grid Hunan Electric Power Co., Ltd. (5216A522001Y, 5216A5240003, and 5216AJ250008)
More Information
    Author Bio:

    TANG You-Yuan Master student candidate at the School of Artificial Intelligence, Changsha University of Science and Technology. He received his Bachelor's degree from Xiangnan University in 2023. His main research direction is point cloud intelligent perception

    ZHANG Hui Professor at the School of Artificial Intelligence and Robotics, Hunan University. He received his Ph.D. degree from Hunan University in 2007. His research interest covers data analysis, image processing, and robot control.Corresponding author of this paper

    DU Rui Ph.D. candidate at the School of Artificial Intelligence and Robotics, Hunan University. He received his master degree from Xiangtan University in 2020. His research interest covers data analysis and image processing

    ZHANG Kai-Ning Assistant Professor at the School of Artificial Intelligence and Robotics, Hunan University. He received his Ph.D. in Communication and Information Systems from Wuhan University in 2025. His research focuses on 3D vision and multimodal perception

    CAO Yun-Kang Assistant Professor at the School of Artificial Intelligence and Robotics, Hunan University. He received his Ph.D. degree in Mechanical Engineering from Huazhong University of Science and Technology in 2025. His research interest covers industrial vision detection and recognition in complex scenarios, industrial foundation models

    BIEKEZATI Baheti Ph. D. candidate at the School of Artificial Intelligence and Ro botics, Hunan University. He received his master degree from Cranfield University in 2020. His research in terest covers signal analysis and time series analysis

    CHEN Hou-Quan Ph. D. candidate at the School of Artificial Intelligence and Robotics, Hunan University. He received his bachelor degree from the School of Computer and Information Engineering, Central South University of Forestry and Technology in 2022. His research interest covers multimodal technologies for power-system applications

    WANG Yao-Nan Academician at Chinese Academy of Engineering, professor at the School of Artificial Intelligence and Robotics, Hunan University. He received his Ph.D. degree from Hunan University in 1995. His research in terest covers data analysis, intelligent control, and image processing

  • 摘要: 配电网点云语义分割对于实现无人化巡检与智能电网运维具有重要意义. 尽管已有方法在空间建模与结构增强方面取得了一定进展, 但在频谱特征挖掘与大规模点云处理效率上仍面临突出挑战. 为此, 本文提出一种结构频谱感知框架(Structure Spectrum-Aware Framework, SSAF), 以提升长距离配网场景下的点云表达能力. 在数据预处理阶段, 提出了一种结合结构引导的层级点云去噪策略与结构感知的样本划分方法, 在压缩冗余背景点云的同时, 有效保持杆塔、导线等关键目标的结构完整性与连续性. 在语义分割阶段, 构建空谱协同语义分割网络(Spatial-Spectral Collaborative Semantic Segmentation Network, SSCNet), 引入局部极坐标系以增强方向敏感特征的建模能力, 并设计基于注意力图的动态融合机制, 实现空间特征与频谱特征之间的自适应交互与信息增强. 实验结果表明, SSAF能在真实配电网场景点云数据集上实现更高的分割精度与推理效率, 在多个关键指标上显著优于现有主流方法, 验证了其在复杂场景下的实用性与工程潜力.
  • 图  1  结构频谱感知框架

    Fig.  1  The structure spectrum-aware framework

    图  2  数据预处理

    Fig.  2  data preprocessing

    图  3  空谱协同语义分割网络

    Fig.  3  Spectral-Spatial Collaborative Semantic Segmentation Network

    图  4  双向门控融合

    Fig.  4  Bidirectional Gated Fusion

    图  5  构建的配电网场景数据集可视化

    Fig.  5  Visualization of the constructed distribution network scene dataset

    图  6  配电网场景点云语义分割结果可视化

    Fig.  6  Visualization of cloud semantic segmentation results for distribution network site scenic spots

    图  7  不同处理阶段下分割精度与运行时间的对比

    Fig.  7  Comparison of segmentation accuracy and runtime at different processing stages

    图  8  不同配置下的特征激活热图

    Fig.  8  Feature activation heat maps under different configurations

    表  1  配电网场景点云语义分割实验(%)

    Table  1  Experiment on Semantic Segmentation of Distribution Network Site Cloud(%)

    方法 OA mAcc mIoU 背景 电力线 电力塔 Params (M) FLOPs (G)
    DPGCN[19] 98.26 71.29 78.92 98.20 63.10 75.46 3.60
    DGCNN[20] 98.89 78.88 83.36 99.03 62.13 88.92 1.30
    PointNet++[14] 98.40 86.04 84.75 99.62 67.48 87.14 1.00 7.20
    KPConv[31] 98.90 87.91 85.51 99.90 65.71 90.91 15.00
    PointNext-XL[32] 98.27 91.40 88.54 98.65 74.72 92.25 41.60 84.80
    PTv2[22] 98.36 92.63 88.42 99.28 80.04 85.93
    PointMetaBase-XL[33] 98.38 94.21 93.62 99.11 92.35 89.40 15.30 9.20
    DeLA[34] 96.87 95.34 93.65 96.89 92.69 91.36 7.00 14.00
    DeLA + X-3D[35] 98.59 96.81 94.29 96.36 94.30 92.22 18.10 9.80
    PTv3[23] 98.64 97.11 94.37 96.56 94.42 92.12
    PCM[36] 98.79 97.27 95.69 96.96 96.17 93.93 34.2
    SSCNet (Ours) 98.74 97.99 96.20 97.85 94.36 96.40 12.54 20.3
    下载: 导出CSV

    表  2  各场景滤波前后分类别点数统计(单位: 万点)

    Table  2  Category-wise Point Cloud Statistics Before and After Filtering (Unit: $ 10^4 $ points)

    场景原始点云滤波后点云背景点
    滤除率 (%)
    电力线点杆塔点背景点电力线点杆塔点背景点
    S1 86.9 16.4 4890.0 86.1 15.7 1415.0 71.1
    S2 67.9 1625.0 4042.0 65.7 1622.1 45.9 98.9
    S3 117.6 1090.4 3162.0 116.8 1087.5 397.0 87.5
    下载: 导出CSV

    表  3  消融实验结果(%)

    Table  3  The ablation experiment results(%)

    方法 PRST BGF OA mIoU mAcc
    1 98.12 94.13 91.74
    2 $ \checkmark$ 98.78 95.47 93.06
    3 $ \checkmark$ 98.57 95.86 96.84
    4 $ \checkmark$ $ \checkmark$ 98.74 96.20 97.99
    下载: 导出CSV

    表  4  主流方法在S3DIS数据集上的对比结果

    Table  4  Comparison of mainstream methods on the S3DIS dataset

    方法 参数量 (M) OA (%) mAcc (%) mIoU (%)
    PointNet++[14] 1.0 83.0 53.5
    DGCNN[20] 1.3 47.9
    KPConv[31] 15.0 72.8 67.1
    PointNext-XL[32] 41 91.0 77.2 71.1
    PTv1[37] 90.8 76.5 70.4
    PTv2[22] 11.3 91.6 78.0 72.7
    PointMetaBase-XL[33] 15.3 90.6 71.5
    DeLA[34] 7.0 92.2 80.0 74.1
    DeLA+X-3D[35] 8.0 92.2 80.1 74.3
    PTv3[23] 34.2 74.7
    PCM[36] 92.9 81.6 74.1
    SSCNet 10.2 92.3 82.1 75.1
    下载: 导出CSV
  • [1] Shen Y, Huang J, Wang J, Jiang J, Li J, Ferreira V. A review and future directions of techniques for extracting powerlines and pylons from LiDAR point clouds. International Journal of Applied Earth Observation and Geoinformation, 2024, 132: 104056 doi: 10.1016/j.jag.2024.104056
    [2] Jung J, Che E, Olsen M J, Shafer K C. Automated and efficient powerline extraction from laser scanning data using a voxel-based subsampling with hierarchical approach. ISPRS Journal of Photogrammetry and Remote Sensing, 2020, 163: 343−361 doi: 10.1016/j.isprsjprs.2020.03.018
    [3] Liu X, Miao X, Jiang H, Chen J, Wu M, Chen Z. Tower masking MIM: A self-supervised pretraining method for power line inspection. IEEE Transactions on Industrial Informatics, 2023, 20(1): 513−523
    [4] 王斐然, 韩庚, 郭昕阳, 石朝阳, 王金. 激光雷达数据下架空输电线路点云场景分割及净空入侵检测. 测绘通报, 2024(5): 133

    Wang F, Han G, Guo X, Shi Z, Wang J. Segmentation and clearance intrusion detection of overhead transmission line point cloud scene based on LiDAR data. Bulletin of Surveying and Mapping, 2024(5): 133
    [5] Kim H B, Sohn G, et al. 3D classification of power-line scene from airborne laser scanning data using random forests. International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 2010, 38(1999): 126−132
    [6] Lehtomäki M, Kukko A, Matikainen L, Hyyppä J, Kaartinen H, Jaakkola A. Power line mapping technique using all-terrain mobile laser scanning. Automation in Construction, 2019, 105: 102802 doi: 10.1016/j.autcon.2019.03.023
    [7] Fischler M A, Bolles R C. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM, 1981, 24(6): 381−395 doi: 10.1016/b978-0-08-051581-6.50070-2
    [8] Shen X, Qian C, Du Y, Yu X, Zhang R. An automatic extraction algorithm of high voltage transmission lines from airborne LIDAR point cloud data. Turkish Journal of Electrical Engineering and Computer Sciences, 2018, 26(4): 2043−2055 doi: 10.3906/elk-1801-23
    [9] Zhu S, et al. A deep-learning-based method for extracting an arbitrary number of individual power lines from uav-mounted laser scanning point clouds. Remote Sensing, 2024, 16(2): 393 doi: 10.3390/rs16020393
    [10] Maturana D, Scherer S. Voxnet: A 3d convolutional neural network for real-time object recognition//2015 IEEE/RSJ International Conference on Intelligent Robots and Systems. Hamburg, Germany: Institute of Electrical, 2015: 922-928.
    [11] 单铉洋, 孙战里, 曾志刚. RFNet: 用于三维点云分类的卷积神经网络. 自动化学报, 2023, 49(11): 2350−2359

    Shan X, Sun Z, Zeng Z. RFNet: Convolutional neural network for 3D point cloud classification. Acta Automatica Sinica, 2023, 49(11): 2350−2359
    [12] 鲁斌, 范晓明. 基于改进自适应k均值聚类的三维点云骨架提取的研究. 自动化学报, 2022, 48(8): 1994−2006 doi: 10.16383/j.aas.c200284

    Lu B, Fan X. Research on 3D point cloud skeleton extraction based on improved adaptive k-means clustering. Acta Automatica Sinica, 2022, 48(8): 1994−2006 doi: 10.16383/j.aas.c200284
    [13] Qi C R, Su H, Mo K, Guibas L J. Pointnet: Deep learning on point sets for 3d classification and segmentation//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, USA: IEEE, 2017: 652-660.
    [14] Qi C R, Yi L, Su H, Guibas L J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in Neural Information Processing Systems, 201730
    [15] Dong J, Chen H, Chen S, Zhao Y, Yang N. PSFE-Net: Semantic segmentation network for airborne LiDAR transmission corridor scenes inspection//2024 9th Asia Conference on Power and Electrical Engineering. Beijing, China: IEEE, 2024: 1538-1542.
    [16] Liu X, Shuang F, Li Y, Zhang L, Huang X, Qin J. SS-IPLE: Semantic segmentation of electric power corridor scene and individual power line extraction from UAV-based lidar point cloud. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2023, 16: 38−50 doi: 10.1109/JSTARS.2023.3289599
    [17] 黄郑, 顾徐, 王红星, 张星炜, 张欣. 基于改进PointNet++ 的输电杆塔点云语义分割模型. Electric Power, 2023, 56(3)

    Huang Z, Gu X, Wang H, Zhang X, Zhang X. An improved PointNet++ model for semantic segmentation of transmission tower point clouds. Electric Power, 2023, 56(3
    [18] Li W, Luo Z, Xiao Z, Chen Y, Wang C, Li J. A GCN-based method for extracting power lines and pylons from airborne LiDAR data. IEEE Transactions on Geoscience and Remote Sensing, 2021, 60: 1−14 doi: 10.1109/tgrs.2021.3076107
    [19] Li G, Muller M, Thabet A, Ghanem B. DeepGCNs: Can GCNs go as deep as CNNs?//Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul, South Korea: IEEE, 2019: 9267-9276.
    [20] Wang Y, Sun Y, Liu Z, Sarma S E, Bronstein M M, Solomon J M. Dynamic graph CNN for learning on point clouds. ACM Transactions on Graphics, 2019, 38(5): 1−12
    [21] 李建, 王健, 王雷, 李敏, 杨立克, 赵艺龙. 双重注意力机制的电力走廊点云语义分割. 测绘通报, 2023(4): 127

    Li J, Wang J, Wang L, Li M, Yang L, Zhao Y. Dual-attention mechanism for semantic segmentation of power corridor point clouds. Bulletin of Surveying and Mapping, 2023(4): 127
    [22] Wu X, Lao Y, Jiang L, Liu X, Zhao H. Point transformer v2: Grouped vector attention and partition-based pooling. Advances in Neural Information Processing Systems, 2022, 35: 33330−33342 doi: 10.52202/068431-2415
    [23] Wu X, et al. Point transformer v3: Simpler faster stronger//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2024: 4840-4851.
    [24] Bu L, Wang Y, Ma Q, Hou Z, Wang R, Bu F. Deep hierarchical learning on point clouds in feature space. Neurocomputing, 2025, 630: 129647 doi: 10.1016/j.neucom.2025.129647
    [25] Liu D, Hu W, Li X. Point cloud attacks in graph spectral domain: When 3d geometry meets graph signal processing. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023, 46(5): 3079−3095 doi: 10.1109/tpami.2023.3339130
    [26] Wen C, Long J, Yu B, Tao D. PointWavelet: Learning in Spectral Domain for 3-D Point Cloud Analysis. IEEE Transactions on Neural Networks and Learning Systems, 2025, 36(3): 4400−4412 doi: 10.1109/TNNLS.2024.3363244
    [27] Rizaldy A, Gloaguen R, Fassnacht F E, Ghamisi P. HyperPointFormer: Multimodal Fusion in 3-D Space With Dual-Branch Cross-Attention Transformers. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2025, 18: 21254−21274 doi: 10.1109/JSTARS.2025.3595648
    [28] Liang D, Feng T, Zhou X, Zhang Y, Zou Z, Bai X. Parameter-efficient fine-tuning in spectral domain for point cloud learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2025.
    [29] Yang Y, et al. RALoc: Enhancing Outdoor LiDAR Localization via Rotation Awareness//Proceedings of the IEEE/CVF International Conference on Computer Vision. 2025: 3304-3313.
    [30] Zhang W, et al. An easy-to-use airborne LiDAR data filtering method based on cloth simulation. Remote sensing, 2016, 8(6): 501 doi: 10.3390/rs8060501
    [31] Thomas H, Qi C R, Deschaud J E, Marcotegui B, Goulette F, Guibas L J. Kpconv: Flexible and deformable convolution for point clouds//Proceedings of the IEEE/CVF International Conference on Computer Vision. Seoul, South Korea: IEEE, 2019: 6411-6420.
    [32] Qian G, et al. Pointnext: Revisiting pointnet++ with improved training and scaling strategies. Advances in Neural Information Processing Systems, 2022, 35: 23192−23204 doi: 10.52202/068431-1685
    [33] Lin H, et al. Meta architecture for point cloud analysis//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada: IEEE, 2023: 17682-17691.
    [34] Yang W, et al. DeLA: An extremely faster network with decoupled local aggregation for large scale point cloud learning. International Journal of Applied Earth Observation and Geoinformation, 2024, 135: 104255 doi: 10.1016/j.jag.2024.104255
    [35] Sun S, Rao Y, Lu J, Yan H. X-3D: Explicit 3D structure modeling for point cloud recognition//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Seattle, USA: IEEE, 2024: 5074-5083.
    [36] Zhang T, et al. Point cloud mamba: Point cloud learning via state space model//Proceedings of the AAAI Conference on Artificial Intelligence: vol. 39: 10. 2025: 10121-10130.
    [37] Zhao H, Jiang L, Jia J, Torr P H S, Koltun V. Point transformer//Proceedings of the IEEE/CVF International Conference on Computer Vision. Montreal, Canada: IEEE, 2021: 16259-16268.
  • 加载中
计量
  • 文章访问数:  6
  • HTML全文浏览量:  3
  • 被引次数: 0
出版历程
  • 收稿日期:  2025-10-13
  • 录用日期:  2025-12-31
  • 网络出版日期:  2026-03-14

目录

    /

    返回文章
    返回