欢迎您访问《智慧农业(中英文)》官方网站! English

Smart Agriculture ›› 2025, Vol. 7 ›› Issue (3): 120-130.doi: 10.12133/j.smartag.SA202501022

• 信息处理与决策 • 上一篇    

基于改进U-Net模型的高纺锤形苹果树休眠期修剪点识别与定位方法

刘龙1,2,3, 王宁1,2,3, 王嘉成1,2,3, 曹宇恒1,2,3, 张凯1,2,3, 康峰1,2,3, 王亚雄1,2,3()   

  1. 1. 北京林业大学 工学院,北京 100083,中国
    2. 林木资源高效生产全国重点实验室,北京 100083,中国
    3. 林业装备与自动化国家林业和草原局重点实验室,北京 100083,中国
  • 收稿日期:2025-01-20 出版日期:2025-05-30
  • 基金项目:
    国家重点研发计划子课题(2018YFD0700603-2); 北京林业大学校院联合基金重点项目(2024XY-G001); 宁夏重点研发计划项目(2022BBF01002-03)
  • 作者简介:

    刘 龙,硕士,研究方向为林业装备自动化。E-mail:

  • 通信作者:
    王亚雄,博士,副教授,研究方向为农林装备自动化智能化技术、智慧果园可视化场景开发。E-mail:

Pruning Point Recognition and Localization for Spindle-Shaped Apple Trees During Dormant Season Using an Improved U-Net Model

LIU Long1,2,3, WANG Ning1,2,3, WANG Jiacheng1,2,3, CAO Yuheng1,2,3, ZHANG Kai1,2,3, KANG Feng1,2,3, WANG Yaxiong1,2,3()   

  1. 1. School of Technology, Beijing Forestry University, Beijing 100083, China
    2. State Key Laboratory of Efficient Production of Forest Resources, Beijing 100083, China
    3. Key Laboratory of National Forestry and Grassland Administration on Forestry Equipment and Automation, Beijing 100083, China
  • Received:2025-01-20 Online:2025-05-30
  • Foundation items:The Sub-Project of the National Key Research and Development Program of China(2018YFD0700603-2); Key Project of Beijing Forestry University's University-College Joint Fund(2024XY-G001); Ningxia Key Research and Development Program Project(2022BBF01002-03)
  • About author:

    LIU Long, E-mail:

  • Corresponding author:
    WANG Yaxiong, E-mail:

摘要:

【目的/意义】 针对智能修剪机器人在复杂田间环境下对果树枝干识别精度不足及修剪点定位不准确的问题,提出一种基于图像和点云融合的深度学习方法,以实现休眠期高纺锤形苹果树剪枝点的自动识别与精准定位。 【方法】 首先,采用Realsense D435i相机采集苹果树RGB-D数据。其次,提出一种改进的U-Net模型,以VGG16(Visual Geometry Group 16)作为主干特征提取网络并在上采样阶段引入卷积块注意力模块CBAM(Convolutional Block Attention Module),实现对RGB图像中主干和一级枝的精确分割。然后,基于OpenCV的边缘检测与骨架提取算法,先提取一级枝连接点,再通过坐标平移在局部邻域内搜索潜在修剪点,并利用深度信息估算一级枝几何参数;同时,通过主干掩模与深度图融合,采用颜色筛选获取主干点云,并运用随机采样一致性算法进行圆柱拟合以估计主干直径。最后,基于智能修剪决策算法确定预测修剪点。 【结果和讨论】 改进的U-Net模型在枝干分割中的平均像素精度(Mean Pixel Accuracy, mPA)为95.52%,在背光和向光条件下表现出良好鲁棒性。相对于人工实测值,一级枝直径、间距和主干直径估计值的平均绝对误差分别为1.33、13.96和5.11 mm。此外,基于智能修剪决策系统识别修剪点的正确率为87.88%,单视角下平均处理时间约为4.2 s。 【结论】 本研究提出了一种高效且精准的苹果树剪枝点识别方法,为智能修剪机器人在现代农业中的应用提供了重要支持,进一步推动了农业生产向智能化和高效化方向发展。

关键词: 剪枝点识别, RGB-D, U-Net, 直径估计, 三维点云, VGG16, 修剪机器人

Abstract:

[Objective] To address the current issues of intelligent pruning robots, such as insufficient recognition accuracy of fruit tree branches and inaccurate localization of pruning points in complex field environments, a deep learning method based on the fusion of images and point clouds was proposed in this research. The method enables non-contact segmentation of dormant high-spindle apple tree branches and measurement of phenotypic parameter, which also achieving automatic identification and precise localization of pruning points. [Methods] Localized RGB-D data were collected from apple trees using a Realsense D435i camera, a device capable of accurate depth measurements within the range of 0.3~3.0 m. Data acquisition took place between early and mid-January 2024, from 9:00 AM to 4:00 PM daily. To maintain uniformity, the camera was mounted on a stand at a distance of 0.4~0.5 m from the main stems of the apple trees. Following data collection, trunks and branches were manually annotated using Labelme software. The OpenCV library was also employed for data augmentation, which helped prevent overfitting during model training. To improve segmentation accuracy of tree trunks and branches in RGB images, an enhanced U-Net model was introduced. This model utilized VGG16 (Visual Geometry Group 16) as its backbone feature extraction network and incorporated the convolutional block attention module (CBAM) at the up-sampling stage. Based on the segmentation results, a multimodal data processing pipeline was established. First, segmented branch mask maps were derived from skeleton lines extracted using OpenCV's algorithm. The first-level branch connection points were identified based on their positions relative to the trunk. Subsequently, potential pruning points were then searched within local neighborhoods through coordinate translation. An edge detection algorithm was applied to locate the nearest edge pixels relative to these potential pruning points. By extending the diameter line of branch pixel points in the images and integrating with depth information, the actual diameter of the branches could be estimated. Additionally, branch spacing was calculated using vertical coordinates differences of potential pruning points in the pixel coordinate system, alongside depth information. Meanwhile, trunk point cloud data were acquired by merging the trunk mask maps with the depth maps. Preprocessing of the point cloud enabled the estimation of the average trunk diameter in the local view through cylindrical fitting using the randomized sampling consistency (RANSAC) algorithm. Finally, an intelligent pruning decision-making algorithm was developed by investigating of orchardists' pruning experience, analyzing relevant literature, and integrating phenotypic parameter acquisition methods, thus achieving accurate prediction of apple tree pruning points. [Results and Discussions] The improved U-Net model proposed achieved a mean pixel accuracy (mPA) of 95.52% for branch segmentation, representing a 2.74 percent point improvement over the original architecture. Corresponding increases were observed in mean intersection over union (mIoU) and precision metrics. Comparative evaluations against DeepLabV3+, PSPNet, and the baseline U-Net were conducted under both backlight and front-light illumination conditions. The improved model demonstrated superior segmentation performance and robustness across all tested scenarios. Ablation experiments indicated that replacing the original feature extractor with VGG16 yielded a 1.52 percent point mPA improvement, accompanied by simultaneous gains in mIoU and precision. The integration of the CBAM at the up sampling stage further augmented the model's capacity to resolve fine branch structures. Phenotypic parameter estimation using segmented branch masks combined with depth maps showed strong correlations with manual measurements. Specifically, the coefficient of determination (R2) values for primary branch diameter, branch spacing, and trunk diameter were 0.96, 0.95, and 0.91, respectively. The mean absolute errors (MAE) were recorded as 1.33, 13.96, and 5.11 mm, surpassing the accuracy of visual assessments by human pruning operators. The intelligent pruning decision system achieved an 87.88% correct identification rate for pruning points, with an average processing time of 4.2 s per viewpoint. These results confirm the practical feasibility and operational efficiency of the proposed method in real-world agricultural settings. [Conclusions] An efficient and accurate method for identifying pruning points on apple trees was proposed, which integrates image and point cloud data through deep learning. The results indicate that this method could provide significant support for the application of intelligent pruning robots in modern agriculture. It not only offers high feasibility but also exhibits outstanding efficiency and accuracy in practical applications, thus laying a solid foundation for the advancement of agricultural automation.

Key words: pruning point identification, RGB-D, U-Net, diameter estimation, 3D point clouds, VGG16, pruning robot

中图分类号: