欢迎您访问《智慧农业(中英文)》官方网站! English

Smart Agriculture

• •    

基于图像和三维点云融合的休眠期高纺锤形苹果树剪枝点定位方法

刘龙1,2,3, 王宁1,2,3, 王嘉成1,2,3, 曹宇恒1,2,3, 张凯1,2,3, 康峰1,2,3, 王亚雄1,2,3()   

  1. 1. 北京林业大学 工学院,北京100083,中国
    2. 林木资源高效生产全国重点实验室,北京100083,中国
    3. 林业装备与自动化国家林业和草原局重点实验室,北京100083,中国
  • 收稿日期:2025-01-20 出版日期:2025-04-08
  • 基金项目:
    The sub-project of the National Key Research and Development Program of China(2018YFD0700603-2); Key Project of Beijing Forestry University's University-College Joint Fund(2024XY-G001); Ningxia Key Research and Development Program Project(2022BBF01002-03)
  • 作者简介:

    刘 龙,硕士,研究方向为林业装备自动化。E-mail:

  • 通信作者:
    王亚雄,博士,副教授,研究方向为农林装备自动化智能化技术、智慧果园可视化场景开发。E-mail:

Localization of Pruning Points of High Spindle Apple Trees in Dormant Period Based on Pictures and 3D Point Cloud

LIU Long1,2,3, WANG Ning1,2,3, WANG Jiacheng1,2,3, CAO Yuheng1,2,3, ZHANG Kai1,2,3, KANG Feng1,2,3, WANG Yaxiong1,2,3()   

  1. 1. School of Technology, Beijing Forestry University, Beijing 100083, China
    2. State Key Laboratory of Efficient Production of Forest Resources, Beijing 100083, China
    3. Key Laboratory of National Forestry and Grassland Administration on Forestry Equipment and Automation, Beijing 100083, China

摘要:

【目的/意义】 针对智能修剪机器人在复杂田间环境下对果树枝干识别精度不足及修剪点定位不准确的问题,提出一种基于图像和点云融合的深度学习方法,以实现休眠期高纺锤形苹果树剪枝点的自动识别与精准定位。 【方法】 首先,采用Realsense D435i相机采集苹果树RGB-D数据。其次,提出一种改进的U-Net模型,以VGG16(Visual Geometry Group 16)作为主干特征提取网络并在上采样阶段引入卷积块注意力模块CBAM(Convolutional Block Attention Module),实现对RGB图像中主干和一级枝的精确分割。然后,基于OpenCV的边缘检测与骨架提取算法,先提取一级枝连接点,再通过坐标平移在局部邻域内搜索潜在修剪点,并利用深度信息估算一级枝几何参数;同时,通过主干掩模与深度图融合,采用颜色筛选获取主干点云,并运用随机采样一致性算法进行圆柱拟合以估计主干直径。最后,基于智能修剪决策算法确定预测修剪点。 【结果和讨论】 改进的U-Net模型在枝干分割中的平均像素精度(Mean Pixel Accuracy, mPA)为95.52%,在背光和向光条件下表现出良好鲁棒性。相对于人工实测值,一级枝直径、间距和主干直径估计值的平均绝对误差分别为1.33、13.96和5.11 mm。此外,基于智能修剪决策系统识别修剪点的正确率为87.88%,单视角下平均处理时间约为4.2 s。 【结论】 本研究提出了一种高效且精准的苹果树剪枝点识别方法,为智能修剪机器人在现代农业中的应用提供了重要支持,进一步推动了农业生产向智能化和高效化方向发展。

关键词: 剪枝点识别, RGB-D, U-Net, 直径估计, 三维点云

Abstract:

[Objective] This study aims to solve the current problem of the intelligent pruning robot's insufficient recognition accuracy of fruit tree branches and inaccurate pruning point localization in complex field environments. To address this, a deep learning method based on the fusion of images and point clouds was proposed, enabling non-contact segmentation of dormant high-spindle apple tree branches and phenotypic parameter measurement. And complete the automatic identification and accurate localization of pruning points. [Methods] In this study, localized RGB-D data were gathered from apple trees using a Realsense D435i camera, a device capable of effective depth measurements within a range of 0.3~3.0 m. Data collection occurred between early and mid-January 2024, from 9:00 AM to 4:00 PM daily. During this period, the weather remained sunny, ensuring optimal conditions for high-quality data acquisition. To maintain consistency, the camera was mounted on a stand at a distance of 0.4~0.5 m from the main stem of the apple trees. After collecting the data, researchers manually labeled trunks and branches using Labelme software. They also employed the OpenCV library to enhance image data, which helped prevent overfitting during model training. To improve segmentation accuracy for tree trunks and branches in RGB images, the research team introduced an enhanced U-Net model. This model utilized VGG16 (Visual Geometry Group 16) as its backbone feature extraction network and incorporated the Convolutional Block Attention Module (CBAM) at the up-sampling stage. Based on the segmentation results, a multimodal data processing flow was established. Initially, the segmented branch mask maps were obtained from skeleton lines extracted using OpenCV's algorithm. The first-level branch connection points were identified based on their positions relative to the trunk. Subsequently, potential pruning points were searched for within local neighborhoods through coordinate translation. An edge detection algorithm was applied to locate the nearest edge pixels to these potential pruning points. By extending the diameter line of the branch pixel points on the images and combining this with depth information, the actual diameter of the branches could be estimated. Additionally, the branch spacing was calculated using the differences in vertical coordinates of potential pruning points in the pixel coordinate system, alongside depth information. Meanwhile, the trunk point cloud data were acquired by merging the trunk mask map with the depth map. Preprocessing of the point cloud enabled the estimation of the average trunk diameter in the local view through cylindrical fitting using the Randomized Sampling Consistency (RANSAC) algorithm. Finally, an intelligent pruning decision-making algorithm was developed through investigation of orchardists' pruning experience, analysis of relevant literature, and integration of phenotypic parameter acquisition methods, thus achieving accurate prediction of apple tree pruning points. [Results and Discussion] The improved U-Net model in this study achieved a mean pixel accuracy (mPA) of 95.52% for branch segmentation, representing a 2.74% improvement over the original architecture. Corresponding increases were observed in mean intersection over union (mIoU) and precision metrics. Comparative evaluations against DeepLabV3+, PSPNet, and the baseline U-Net were conducted under both backlight and front-light illumination conditions. The enhanced model demonstrated superior segmentation performance and robustness across all tested scenarios. Ablation experiments indicated that replacing the original feature extractor with VGG16 yielded a 1.52% mPA improvement, accompanied by simultaneous gains in mIoU and precision. The integration of the Convolutional Block Attention Module (CBAM) at the up sampling stage further augmented the model's capacity to resolve fine branch structures. Phenotypic parameter estimation using segmented branch masks combined with depth maps showed strong correlations with manual measurements. Specifically, the coefficient of determination (R2) values for primary branch diameter, branch spacing, and trunk diameter were 0.96, 0.95, and 0.91, respectively. The mean absolute errors (MAE) were recorded as 1.33, 13.96, and 5.11 mm, surpassing the accuracy of visual assessments by human pruning operators. The intelligent pruning decision system achieved an 87.88% correct identification rate for pruning points, with an average processing time of 4.2 s per viewpoint. These results validated the proposed methodology's practical feasibility and operational efficiency in real-world agricultural applications. [Conclusion] In summary, an efficient and accurate method was proposed for identifying pruning points on apple trees based on the fusion of image and point cloud data through deep learning. This comprehensive approach provides significant support for the application of intelligent pruning robots in modern agriculture, further advancing the shift towards smarter and more efficient agricultural production. The findings demonstrate that this method not only offers high feasibility but also exhibits outstanding efficiency and accuracy in practical applications, laying a solid foundation for future agricultural automation.

Key words: pruning point identification, RGB-D, U-Net, diameter estimation, 3D point clouds

中图分类号: