Welcome to Smart Agriculture 中文

Most Down Articles

    Published in last 1 year | In last 2 years| In last 3 years| All| Most Downloaded in Recent Month | Most Downloaded in Recent Year|

    In last 2 years
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Crop Pest Target Detection Algorithm in Complex Scenes:YOLOv8-Extend
    ZHANG Ronghua, BAI Xue, FAN Jiangchuan
    Smart Agriculture    2024, 6 (2): 49-61.   DOI: 10.12133/j.smartag.SA202311007
    Abstract1897)   HTML194)    PDF(pc) (2287KB)(95218)       Save

    [Objective] It is of great significance to improve the efficiency and accuracy of crop pest detection in complex natural environments, and to change the current reliance on expert manual identification in the agricultural production process. Targeting the problems of small target size, mimicry with crops, low detection accuracy, and slow algorithm reasoning speed in crop pest detection, a complex scene crop pest target detection algorithm named YOLOv8-Entend was proposed in this research. [Methods] Firstly, the GSConv was introduecd to enhance the model's receptive field, allowing for global feature aggregation. This mechanism enables feature aggregation at both node and global levels simultaneously, obtaining local features from neighboring nodes through neighbor sampling and aggregation operations, enhancing the model's receptive field and semantic understanding ability. Additionally, some Convs were replaced with lightweight Ghost Convolutions and HorBlock was utilized to capture longer-term feature dependencies. The recursive gate convolution employed gating mechanisms to remember and transmit previous information, capturing long-term correlations. Furthermore, Concat was replaced with BiFPN for richer feature fusion. The bidirectional fusion of depth features from top to bottom and from bottom to top enhances the transmission of feature information acrossed different network layers. Utilizing the VoVGSCSP module, feature maps of different scales were connected to create longer feature map vectors, increasing model diversity and enhancing small object detection. The convolutional block attention module (CBAM) attention mechanism was introduced to strengthen features of field pests and reduce background weights caused by complexity. Next, the Wise IoU dynamic non-monotonic focusing mechanism was implemented to evaluate the quality of anchor boxes using "outlier" instead of IoU. This mechanism also included a gradient gain allocation strategy, which reduced the competitiveness of high-quality anchor frames and minimizes harmful gradients from low-quality examples. This approach allowed WIoU to concentrate on anchor boxes of average quality, improving the network model's generalization ability and overall performance. Subsequently, the improved YOLOv8-Extend model was compared with the original YOLOv8 model, YOLOv5, YOLOv8-GSCONV, YOLOv8-BiFPN, and YOLOv8-CBAM to validate the accuracy and precision of model detection. Finally, the model was deployed on edge devices for inference verification to confirm its effectiveness in practical application scenarios. [Results and Discussions] The results indicated that the improved YOLOv8-Extend model achieved notable improvements in accuracy, recall, mAP@0.5, and mAP@0.5:0.95 evaluation indices. Specifically, there were increases of 2.6%, 3.6%, 2.4% and 7.2%, respectively, showcasing superior detection performance. YOLOv8-Extend and YOLOv8 run respectively on the edge computing device JETSON ORIN NX 16 GB and were accelerated by TensorRT, mAP@0.5 improved by 4.6%, FPS reached 57.6, meeting real-time detection requirements. The YOLOv8-Extend model demonstrated better adaptability in complex agricultural scenarios and exhibited clear advantages in detecting small pests and pests sharing similar growth environments in practical data collection. The accuracy in detecting challenging data saw a notable increased of 11.9%. Through algorithm refinement, the model showcased improved capability in extracting and focusing on features in crop pest target detection, addressing issues such as small targets, similar background textures, and challenging feature extraction. [Conclusions] The YOLOv8-Extend model introduced in this study significantly boosts detection accuracy and recognition rates while upholding high operational efficiency. It is suitable for deployment on edge terminal computing devices to facilitate real-time detection of crop pests, offering technological advancements and methodologies for the advancement of cost-effective terminal-based automatic pest recognition systems. This research can serve as a valuable resource and aid in the intelligent detection of other small targets, as well as in optimizing model structures.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweighted Wheat Leaf Diseases and Pests Detection Model Based on Improved YOLOv8
    YANG Feng, YAO Xiaotong
    Smart Agriculture    2024, 6 (1): 147-157.   DOI: 10.12133/j.smartag.SA202309010
    Abstract1612)   HTML254)    PDF(pc) (1991KB)(63881)       Save

    Objective To effectively tackle the unique attributes of wheat leaf pests and diseases in their native environment, a high-caliber and efficient pest detection model named YOLOv8-SS (You Only Look Once Version 8-SS) was proposed. This innovative model is engineered to accurately identify pests, thereby providing a solid scientific foundation for their prevention and management strategies. Methods A total of 3 639 raw datasets of images of wheat leaf pests and diseases were collected from 6 different wheat pests and diseases in various farmlands in the Yuchong County area of Gansu Province, at different periods of time, using mobile phones. This collection demonstrated the team's proficiency and commitment to advancing agricultural research. The dataset was meticulously constructed using the LabelImg software to accurately label the images with targeted pest species. To guarantee the model's superior generalization capabilities, the dataset was strategically divided into a training set and a test set in an 8:2 ratio. The dataset includes thorough observations and recordings of the wheat leaf blade's appearance, texture, color, as well as other variables that could influence these characteristics. The compiled dataset proved to be an invaluable asset for both training and validation activities. Leveraging the YOLOv8 algorithm, an enhanced lightweight convolutional neural network, ShuffleNetv2, was selected as the basis network for feature extraction from images. This was accomplished by integrating a 3×3 Depthwise Convolution (DWConv) kernel, the h-swish activation function, and a Squeeze-and-Excitation Network (SENet) attention mechanism. These enhancements streamlined the model by diminishing the parameter count and computational demands, all while sustaining high detection precision. The deployment of these sophisticated methodologies exemplified the researchers' commitment and passion for innovation. The YOLOv8 model employs the SEnet attention mechanism module within both its Backbone and Neck components, significantly reducing computational load while bolstering accuracy. This method exemplifies the model's exceptional performance, distinguishing it from other models in the domain. By integrating a dedicated small target detection layer, the model's capabilities have been augmented, enabling more efficient and precise pest and disease detection. The introduction of a new detection feature map, sized 160×160 pixels, enables the network to concentrate on identifying small-targeted pests and diseases, thereby enhancing the accuracy of pest and disease recognition. Results and Discussion The YOLOv8-SS wheat leaf pests and diseases detection model has been significantly improved to accurately detect wheat leaf pests and diseases in their natural environment. By employing the refined ShuffleNet V2 within the DarkNet-53 framework, as opposed to the conventional YOLOv8, under identical experimental settings, the model exhibited a 4.53% increase in recognition accuracy and a 4.91% improvement in F1-Score, compared to the initial model. Furthermore, the incorporation of a dedicated small target detection layer led to a subsequent rise in accuracy and F1-Scores of 2.31% and 2.16%, respectively, despite a minimal upsurge in the number of parameters and computational requirements. The integration of the SEnet attention mechanism module into the YOLOv8 model resulted in a detection accuracy rate increase of 1.85% and an F1-Score enhancement of 2.72%. Furthermore, by swapping the original neural network architecture with an enhanced ShuffleNet V2 and appending a compact object detection sublayer (namely YOLOv8-SS), the resulting model exhibited a heightened recognition accuracy of 89.41% and an F1-Score of 88.12%. The YOLOv8-SS variant substantially outperformed the standard YOLOv8, showing a remarkable enhancement of 10.11% and 9.92% in accuracy, respectively. This outcome strikingly illustrates the YOLOv8-SS's prowess in balancing speed with precision. Moreover, it achieves convergence at a more rapid pace, requiring approximately 40 training epochs, to surpass other renowned models such as Faster R-CNN, MobileNetV2, SSD, YOLOv5, YOLOX, and the original YOLOv8 in accuracy. Specifically, the YOLOv8-SS boasted an average accuracy 23.01%, 15.13%, 11%, 25.21%, 27.52%, and 10.11% greater than that of the competing models, respectively. In a head-to-head trial involving a public dataset (LWDCD 2020) and a custom-built dataset, the LWDCD 2020 dataset yielded a striking accuracy of 91.30%, outperforming the custom-built dataset by a margin of 1.89% when utilizing the same network architecture, YOLOv8-SS. The AI Challenger 2018-6 and Plant-Village-5 datasets did not perform as robustly, achieving accuracy rates of 86.90% and 86.78% respectively. The YOLOv8-SS model has shown substantial improvements in both feature extraction and learning capabilities over the original YOLOv8, particularly excelling in natural environments with intricate, unstructured backdrops. Conclusion The YOLOv8-SS model is meticulously designed to deliver unmatched recognition accuracy while consuming a minimal amount of storage space. In contrast to conventional detection models, this groundbreaking model exhibits superior detection accuracy and speed, rendering it exceedingly valuable across various applications. This breakthrough serves as an invaluable resource for cutting-edge research on crop pest and disease detection within natural environments featuring complex, unstructured backgrounds. Our method is versatile and yields significantly enhanced detection performance, all while maintaining a lean model architecture. This renders it highly appropriate for real-world scenarios involving large-scale crop pest and disease detection.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Sensor: Research Progress, Challenges and Perspectives
    WANG Rujing
    Smart Agriculture    2024, 6 (1): 1-17.   DOI: 10.12133/j.smartag.SA202401017
    Abstract2187)   HTML441)    PDF(pc) (1179KB)(23271)       Save

    Significance Agricultural sensor is the key technology for developing modern agriculture. Agricultural sensor is a kind of detection device that can sense and convert physical signal, which is related to the agricultural environment, plants and animals, into an electrical signal. Agricultural sensors could be applied to monitor crops and livestock in different agricultural environments, including weather, water, atmosphere and soil. It is also an important driving force to promote the iterative upgrading of agricultural technology and change agricultural production methods. Progress The different agricultural sensors are categorized, the cutting-edge research trends of agricultural sensors are analyzed, and summarizes the current research status of agricultural sensors are summarized in different application scenarios. Moreover, a deep analysis and discussion of four major categories is conducted, which include agricultural environment sensors, animal and plant life information sensors, agricultural product quality and safety sensors, and agricultural machinery sensors. The process of research, development, the universality and limitations of the application of the four types of agricultural sensors are summarized. Agricultural environment sensors are mainly used for real-time monitoring of key parameters in agricultural production environments, such as the quality of water, gas, and soil. The soil sensors provide data support for precision irrigation, rational fertilization, and soil management by monitoring indicators such as soil humidity, pH, temperature, nutrients, microorganisms, pests and diseases, heavy metals and agricultural pollution, etc. Monitoring of dissolved oxygen, pH, nitrate content, and organophosphorus pesticides in irrigation and aquaculture water through water sensors ensures the rational use of water resources and water quality safety. The gas sensor monitors the atmospheric CO2, NH3, C2H2, CH4 concentration, and other information, which provides the appropriate environmental conditions for the growth of crops in greenhouses. The animal life information sensor can obtain the animal's growth, movement, physiological and biochemical status, which include movement trajectory, food intake, heart rate, body temperature, blood pressure, blood glucose, etc. The plant life information sensors monitor the plant's health and growth, such as volatile organic compounds of the leaves, surface temperature and humidity, phytohormones, and other parameters. Especially, the flexible wearable plant sensors provide a new way to measure plant physiological characteristics accurately and monitor the water status and physiological activities of plants non-destructively and continuously. These sensors are mainly used to detect various indicators in agricultural products, such as temperature and humidity, freshness, nutrients, and potentially hazardous substances (e.g., bacteria, pesticide residues, heavy metals, etc. Agricultural machinery sensors can achieve real-time monitoring and controlling of agricultural machinery to achieve real-time cultivation, planting, management, and harvesting, automated operation of agricultural machinery, and accurate application of pesticide, fertilizer. [Conclusions and Prospects In the challenges and prospects of agricultural sensors, the core bottlenecks of large-scale application of agricultural sensors at the present stage are analyzed in detail. These include low-cost, specialization, high stability, and adaptive intelligence of agricultural sensors. Furthermore, the concept of "ubiquitous sensing in agriculture" is proposed, which provides ideas and references for the research and development of agricultural sensor technology.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Shrimp Diseases Detection Method Based on Improved YOLOv8 and Multiple Features
    XU Ruifeng, WANG Yaohua, DING Wenyong, YU Junqi, YAN Maocang, CHEN Chen
    Smart Agriculture    2024, 6 (2): 62-71.   DOI: 10.12133/j.smartag.SA201311014
    Abstract1625)   HTML62)    PDF(pc) (1597KB)(19374)       Save

    [Objective] In recent years, there has been a steady increase in the occurrence and fatality rates of shrimp diseases, causing substantial impacts in shrimp aquaculture. These diseases are marked by their swift onset, high infectivity, complex control requirements, and elevated mortality rates. With the continuous growth of shrimp factory farming, traditional manual detection approaches are no longer able to keep pace with the current requirements. Hence, there is an urgent necessity for an automated solution to identify shrimp diseases. The main goal of this research is to create a cost-effective inspection method using computer vision that achieves a harmonious balance between cost efficiency and detection accuracy. The improved YOLOv8 (You Only Look Once) network and multiple features were employed to detect shrimp diseases. [Methods] To address the issue of surface foam interference, the improved YOLOv8 network was applied to detect and extract surface shrimps as the primary focus of the image. This target detection approach accurately recognizes objects of interest in the image, determining their category and location, with extraction results surpassing those of threshold segmentation. Taking into account the cost limitations of platform computing power in practical production settings, the network was optimized by reducing parameters and computations, thereby improving detection speed and deployment efficiency. Additionally, the Farnberck optical flow method and gray level co-occurrence matrix (GLCM) were employed to capture the movement and image texture features of shrimp video clips. A dataset was created using these extracted multiple feature parameters, and a Support Vector Machine (SVM) classifier was trained to categorize the multiple feature parameters in video clips, facilitating the detection of shrimp health. [Results and Discussions] The improved YOLOv8 in this study effectively enhanced detection accuracy without increasing the number of parameters and flops. According to the results of the ablation experiment, replacing the backbone network with FasterNet lightweight backbone network significantly reduces the number of parameters and computation, albeit at the cost of decreased accuracy. However, after integrating the efficient multi-scale attention (EMA) on the neck, the mAP0.5 increased by 0.3% compared to YOLOv8s, while mAP0.95 only decreased by 2.1%. Furthermore, the parameter count decreased by 45%, and FLOPs decreased by 42%. The improved YOLOv8 exhibits remarkable performance, ranking second only to YOLOv7 in terms of mAP0.5 and mAP0.95, with respective reductions of 0.4% and 0.6%. Additionally, it possesses a significantly reduced parameter count and FLOPS compared to YOLOv7, matching those of YOLOv5. Despite the YOLOv7-Tiny and YOLOv8-VanillaNet models boasting lower parameters and Flops, their accuracy lags behind that of the improved YOLOv8. The mAP0.5 and mAP0.95 of YOLOv7-Tiny and YOLOv8-VanillaNet are 22.4%, 36.2%, 2.3%, and 4.7% lower than that of the improved YOLOv8, respectively. Using a support vector machine (SVM) trained on a comprehensive dataset incorporating multiple feature, the classifier achieved an impressive accuracy rate of 97.625%. The 150 normal fragments and the 150 diseased fragments were randomly selected as test samples. The classifier exhibited a detection accuracy of 89% on this dataset of the 300 samples. This result indicates that the combination of features extracted using the Farnberck optical flow method and GLCM can effectively capture the distinguishing dynamics of movement speed and direction between infected and healthy shrimp. In this research, the majority of errors stem from the incorrect recognition of diseased segments as normal segments, accounting for 88.2% of the total error. These errors can be categorized into three main types: 1) The first type occurs when floating foam obstructs the water surface, resulting in a small number of shrimp being extracted from the image. 2) The second type is attributed to changes in water movement. In this study, nanotubes were used for oxygenation, leading to the generation of sprays on the water surface, which affected the movement of shrimp. 3) The third type of error is linked to video quality. When the video's pixel count is low, the difference in optical flow between diseased shrimp and normal shrimp becomes relatively small. Therefore, it is advisable to adjust the collection area based on the actual production environment and enhance video quality. [Conclusions] The multiple features introduced in this study effectively capture the movement of shrimp, and can be employed for disease detection. The improved YOLOv8 is particularly well-suited for platforms with limited computational resources and is feasible for deployment in actual production settings. However, the experiment was conducted in a factory farming environment, limiting the applicability of the method to other farming environments. Overall, this method only requires consumer-grade cameras as image acquisition equipment and has lower requirements on the detection platform, and can provide a theoretical basis and methodological support for the future application of aquatic disease detection methods.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    In Situ Identification Method of Maize Stalk Width Based on Binocular Vision and Improved YOLOv8
    ZUO Haoxuan, HUANG Qicheng, YANG Jiahao, MENG Fanjia, LI Sien, LI Li
    Smart Agriculture    2023, 5 (3): 86-95.   DOI: 10.12133/j.smartag.SA202309004
    Abstract698)   HTML88)    PDF(pc) (1659KB)(18545)       Save

    [Objective] The width of maize stalks is an important indicator affecting the lodging resistance of maize. The measurement of maize stalk width has many problems, such as cumbersome manual collection process and large errors in the accuracy of automatic equipment collection and recognition, and it is of great application value to study a method for in-situ detection and high-precision identification of maize stalk width. [Methods] The ZED2i binocular camera was used and fixed in the field to obtain real-time pictures from the left and right sides of maize stalks together. The picture acquisition system was based on the NVIDIA Jetson TX2 NX development board, which could achieve timed shooting of both sides view of the maize by setting up the program. A total of maize original images were collected and a dataset was established. In order to observe more features in the target area from the image and provide assistance to improve model training generalization ability, the original images were processed by five processing methods: image saturation, brightness, contrast, sharpness and horizontal flipping, and the dataset was expanded to 3500 images. YOLOv8 was used as the original model for identifying maize stalks from a complex background. The coordinate attention (CA) attention mechanism can bring huge gains to downstream tasks on the basis of lightweight networks, so that the attention block can capture long-distance relationships in one direction while retaining spatial information in the other direction, so that the position information can be saved in the generated attention map to focus on the area of interest and help the network locate the target better and more accurately. By adding the CA module multiple times, the CA module was fused with the C2f module in the original Backbone, and the Bottleneck in the original C2f module was replaced by the CA module, and the C2fCA network module was redesigned. Replacing the loss function Efficient IoU Loss(EIoU) splits the loss term of the aspect ratio into the difference between the predicted width and height and the width and height of the minimum outer frame, which accelerated the convergence of the prediction box, improved the regression accuracy of the prediction box, and further improved the recognition accuracy of maize stalks. The binocular camera was then calibrated so that the left and right cameras were on the same three-dimensional plane. Then the three-dimensional reconstruction of maize stalks, and the matching of left and right cameras recognition frames was realized through the algorithm, first determine whether the detection number of recognition frames in the two images was equal, if not, re-enter the binocular image. If they were equal, continue to judge the coordinate information of the left and right images, the width and height of the bounding box, and determine whether the difference was less than the given Ta. If greater than the given Ta, the image was re-imported; If it was less than the given Ta, the confidence level of the recognition frame of the image was determined whether it was less than the given Tb. If greater than the given Tb, the image is re-imported; If it is less than the given Tb, it indicates that the recognition frame is the same maize identified in the left and right images. If the above conditions were met, the corresponding point matching in the binocular image was completed. After the three-dimensional reconstruction of the binocular image, the three-dimensional coordinates (Ax, Ay, Az) and (Bx, By, Bz) in the upper left and upper right corners of the recognition box under the world coordinate system were obtained, and the distance between the two points was the width of the maize stalk. Finally, a comparative analysis was conducted among the improved YOLOv8 model, the original YOLOv8 model, faster region convolutional neural networks (Faster R-CNN), and single shot multiBox detector (SSD)to verify the recognition accuracy and recognition accuracy of the model. [Results and Discussions] The precision rate (P)、recall rate (R)、average accuracy mAP0.5、average accuracy mAP0.5:0.95 of the improved YOLOv8 model reached 96.8%、94.1%、96.6% and 77.0%. Compared with YOLOv7, increased by 1.3%、1.3%、1.0% and 11.6%, compared with YOLOv5, increased by 1.8%、2.1%、1.2% and 15.8%, compared with Faster R-CNN, increased by 31.1%、40.3%、46.2%、and 37.6%, and compared with SSD, increased by 20.6%、23.8%、20.9% and 20.1%, respectively. Respectively, and the linear regression coefficient of determination R2, root mean square error RMSE and mean absolute error MAE were 0.373, 0.265 cm and 0.244 cm, respectively. The method proposed in the research can meet the requirements of actual production for the measurement accuracy of maize stalk width. [Conclusions] In this study, the in-situ recognition method of maize stalk width based on the improved YOLOv8 model can realize the accurate in-situ identification of maize stalks, which solves the problems of time-consuming and laborious manual measurement and poor machine vision recognition accuracy, and provides a theoretical basis for practical production applications.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    The Key Issues and Evaluation Methods for Constructing Agricultural Pest and Disease Image Datasets: A Review
    GUAN Bolun, ZHANG Liping, ZHU Jingbo, LI Runmei, KONG Juanjuan, WANG Yan, DONG Wei
    Smart Agriculture    2023, 5 (3): 17-34.   DOI: 10.12133/j.smartag.SA202306012
    Abstract1396)   HTML179)    PDF(pc) (1576KB)(16101)       Save

    [Significance] The scientific dataset of agricultural pests and diseases is the foundation for monitoring and warning of agricultural pests and diseases. It is of great significance for the development of agricultural pest control, and is an important component of developing smart agriculture. The quality of the dataset affecting the effectiveness of image recognition algorithms, with the discovery of the importance of deep learning technology in intelligent monitoring of agricultural pests and diseases. The construction of high-quality agricultural pest and disease datasets is gradually attracting attention from scholars in this field. In the task of image recognition, on one hand, the recognition effect depends on the improvement strategy of the algorithm, and on the other hand, it depends on the quality of the dataset. The same recognition algorithm learns different features in different quality datasets, so its recognition performance also varies. In order to propose a dataset evaluation index to measure the quality of agricultural pest and disease datasets, this article analyzes the existing datasets and takes the challenges faced in constructing agricultural pest and disease image datasets as the starting point to review the construction of agricultural pest and disease datasets. [Progress] Firstly, disease and pest datasets are divided into two categories: private datasets and public datasets. Private datasets have the characteristics of high annotation quality, high image quality, and a large number of inter class samples that are not publicly available. Public datasets have the characteristics of multiple types, low image quality, and poor annotation quality. Secondly, the problems faced in the construction process of datasets are summarized, including imbalanced categories at the dataset level, difficulty in feature extraction at the dataset sample level, and difficulty in measuring the dataset size at the usage level. These include imbalanced inter class and intra class samples, selection bias, multi-scale targets, dense targets, uneven data distribution, uneven image quality, insufficient dataset size, and dataset availability. The main reasons for the problem are analyzed by two key aspects of image acquisition and annotation methods in dataset construction, and the improvement strategies and suggestions for the algorithm to address the above issues are summarized. The collection devices of the dataset can be divided into handheld devices, drone platforms, and fixed collection devices. The collection method of handheld devices is flexible and convenient, but it is inefficient and requires high photography skills. The drone platform acquisition method is suitable for data collection in contiguous areas, but the detailed features captured are not clear enough. The fixed device acquisition method has higher efficiency, but the shooting scene is often relatively fixed. The annotation of image data is divided into rectangular annotation and polygonal annotation. In image recognition and detection, rectangular annotation is generally used more frequently. It is difficult to label images that are difficult to separate the target and background. Improper annotation can lead to the introduction of more noise or incomplete algorithm feature extraction. In response to the problems in the above three aspects, the evaluation methods are summarized for data distribution consistency, dataset size, and image annotation quality at the end of the article. [Conclusions and Prospects] The future research and development suggestions for constructing high-quality agricultural pest and disease image datasets based are proposed on the actual needs of agricultural pest and disease image recognition:(1) Construct agricultural pest and disease datasets combined with practical usage scenarios. In order to enable the algorithm to extract richer target features, image data can be collected from multiple perspectives and environments to construct a dataset. According to actual needs, data categories can be scientifically and reasonably divided from the perspective of algorithm feature extraction, avoiding unreasonable inter class and intra class distances, and thus constructing a dataset that meets task requirements for classification and balanced feature distribution. (2) Balancing the relationship between datasets and algorithms. When improving algorithms, consider the more sufficient distribution of categories and features in the dataset, as well as the size of the dataset that matches the model, to improve algorithm accuracy, robustness, and practicality. It ensures that comparative experiments are conducted on algorithm improvement under the same evaluation standard dataset, and improved the pest and disease image recognition algorithm. Research the correlation between the scale of agricultural pest and disease image data and algorithm performance, study the relationship between data annotation methods and algorithms that are difficult to annotate pest and disease images, integrate recognition algorithms for fuzzy, dense, occluded targets, and propose evaluation indicators for agricultural pest and disease datasets. (3) Enhancing the use value of datasets. Datasets can not only be used for research on image recognition, but also for research on other business needs. The identification, collection, and annotation of target images is a challenging task in the construction process of pest and disease datasets. In the process of collecting image data, in addition to collecting images, attention can be paid to the collection of surrounding environmental information and host information. This method is used to construct a multimodal agricultural pest and disease dataset, fully leveraging the value of the dataset. In order to focus researchers on business innovation research, it is necessary to innovate the organizational form of data collection, develop a big data platform for agricultural diseases and pests, explore the correlation between multimodal data, improve the accessibility and convenience of data, and provide efficient services for application implementation and business innovation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Challenges of Oil Crop Yield Monitoring by Remote Sensing
    MA Yujing, WU Shangrong, YANG Peng, CAO Hong, TAN Jieyang, ZHAO Rongkun
    Smart Agriculture    2023, 5 (3): 1-16.   DOI: 10.12133/j.smartag.SA202303002
    Abstract966)   HTML187)    PDF(pc) (837KB)(10441)       Save

    [Significance] Oil crops play a significant role in the food supply, as well as the important source of edible vegetable oils and plant proteins. Real-time, dynamic and large-scale monitoring of oil crop growth is essential in guiding agricultural production, stabilizing markets, and maintaining health. Previous studies have made a considerable progress in the yield simulation of staple crops in regional scale based on remote sensing methods, but the yield simulation of oil crops in regional scale is still poor as its complexity of the plant traits and structural characteristics. Therefore, it is urgently needed to study regional oil crop yield estimation based on remote sensing technology. [Progress] This paper summarized the content of remote sensing technology in oil crop monitoring from three aspects: backgrounds, progressions, opportunities and challenges. Firstly, significances and advantages of using remote sensing technology to estimate the of oil crops have been expounded. It is pointed out that both parameter inversion and crop area monitoring were the vital components of yield estimation. Secondly, the current situation of oil crop monitoring was summarized based on remote sensing technology from three aspects of remote sensing parameter inversion, crop area monitoring and yield estimation. For parameter inversion, it is specified that optical remote sensors were used more than other sensors in oil crops inversion in previous studies. Then, advantages and disadvantages of the empirical model and physical model inversion methods were analyzed. In addition, advantages and disadvantages of optical and microwave data were further illustrated from the aspect of oil crops structure and traits characteristics. At last, optimal choice on the data and methods were given in oil crop parameter inversion. For crop area monitoring, this paper mainly elaborated from two parts of optical and microwave remote sensing data. Combined with the structure of oil crops and the characteristics of planting areas, the researches on area monitoring of oil crops based on different types of remote sensing data sources were reviewed, including the advantages and limitations of different data sources in area monitoring. Then, two yield estimation methods were introduced: remote sensing yield estimation and data assimilation yield estimation. The phenological period of oil crop yield estimation, remote sensing data source and modeling method were summarized. Next, data assimilation technology was introduced, and it was proposed that data assimilation technology has great potential in oil crop yield estimation, and the assimilation research of oil crops was expounded from the aspects of assimilation method and grid selection. All of them indicate that data assimilation technology could improve the accuracy of regional yield estimation of oil crops. Thirdly, this paper pointed out the opportunities of remote sensing technology in oil crop monitoring, put forward some problems and challenges in crop feature selection, spatial scale determination and remote sensing data source selection of oil crop yield, and forecasted the development trend of oil crop yield estimation research in the future. [Conclusions and Prospects] The paper puts forward the following suggestions for the three aspects: (1) Regarding crop feature selection, when estimating yields for oil crops such as rapeseed and soybeans, which have active photosynthesis in siliques or pods, relying solely on canopy leaf area index (LAI) as the assimilation state variable for crop yield estimation may result in significant underestimation of yields, thereby impacting the accuracy of regional crop yield simulation. Therefore, it is necessary to consider the crop plant characteristics and the agronomic mechanism of yield formation through siliques or pods when estimating yields for oil crops. (2) In determining the spatial scale, some oil crops are distributed in hilly and mountainous areas with mixed land cover. Using regularized yield simulation grids may result in the confusion of numerous background objects, introducing additional errors and affecting the assimilation accuracy of yield estimation. This poses a challenge to yield estimation research. Thus, it is necessary to choose appropriate methods to divide irregular unit grids and determine the optimal scale for yield estimation, thereby improving the accuracy of yield estimation. (3) In terms of remote sensing data selection, the monitoring of oil crops can be influenced by crop structure and meteorological conditions. Depending solely on spectral data monitoring may have a certain impact on yield estimation results. It is important to incorporate radar off-nadir remote sensing measurement techniques to perceive the response relationship between crop leaves and siliques or pods and remote sensing data parameters. This can bridge the gap between crop characteristics and remote sensing information for crop yield simulation. This paper can serve as a valuable reference and stimulus for further research on regional yield estimation and growth monitoring of oil crops. It supplements existing knowledge and provides insightful considerations for enhancing the accuracy and efficiency of oil crop production monitoring and management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Intelligent Identification of Crop Agronomic Traits and Morphological Structure Phenotypes: A Review
    ZHANG Jianhua, YAO Qiong, ZHOU Guomin, WU Wendi, XIU Xiaojie, WANG Jian
    Smart Agriculture    2024, 6 (2): 14-27.   DOI: 10.12133/j.smartag.SA202401015
    Abstract1847)   HTML119)    PDF(pc) (1376KB)(10164)       Save

    [Significance] The crop phenotype is the visible result of the complex interplay between crop genes and the environment. It reflects the physiological, ecological, and dynamic aspects of crop growth and development, serving as a critical component in the realm of advanced breeding techniques. By systematically analyzing crop phenotypes, researchers can gain valuable insights into gene function and identify genetic factors that influence important crop traits. This information can then be leveraged to effectively harness germplasm resources and develop breakthrough varieties. Utilizing data-driven, intelligent, dynamic, and non-invasive methods for measuring crop phenotypes allows researchers to accurately capture key growth traits and parameters, providing essential data for breeding and selecting superior crop varieties throughout the entire growth cycle. This article provides an overview of intelligent identification technologies for crop agronomic traits and morphological structural phenotypes. [Progress] Crop phenotype acquisition equipment serves as the essential foundation for acquiring, analyzing, measuring, and identifying crop phenotypes. This equipment enables detailed monitoring of crop growth status. The article presents an overview of the functions, performance, and applications of the leading high-throughput crop phenotyping platforms, as well as an analysis of the characteristics of various sensing and imaging devices used to obtain crop phenotypic information. The rapid advancement of high-throughput crop phenotyping platforms and sensory imaging equipment has facilitated the integration of cutting-edge imaging technology, spectroscopy technology, and deep learning algorithms. These technologies enable the automatic and high-throughput acquisition of yield, resistance, quality, and other relevant traits of large-scale crops, leading to the generation of extensive multi-dimensional, multi-scale, and multi-modal crop phenotypic data. This advancement supports the rapid progression of crop phenomics. The article also discusses the research progress of intelligent recognition technologies for agronomic traits such as crop plant height acquisition, crop organ detection, and counting, as well as crop ideotype recognition, crop morphological information measurement, and crop three-dimensional reconstruction for morphological structure intelligent recognition. Furthermore, this article outlines the main challenges faced in this field, including: difficulties in data collection in complex environments, high requirements for data scale, diversity, and preprocessing, the need to improve the lightweight nature and generalization ability of models, as well as the high cost of data collection equipment and the need to enhance practicality. [Conclusions and Prospects] Finally, this article puts forward the development directions of crop phenotype intelligent recognition technology, including: developing new and low cost intelligent field equipment for acquiring and analyzing crop phenotypes, enhancing the standardization and consistency of field crop phenotype acquisition, strengthening the generality of intelligent crop phenotype recognition models, researching crop phenotype recognition methods that involve multi-perspective, multimodal, multi-point continuous analysis, and spatiotemporal feature fusion, as well as improving model interpretability.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Prospects on Rapid Acquisition Technology of Farmland Soil Physical and Chemical Parameters
    QI Jiangtao, CHENG Panting, GAO Fangfang, GUO Li, ZHANG Ruirui
    Smart Agriculture    2024, 6 (3): 17-33.   DOI: 10.12133/j.smartag.SA202404003
    Abstract875)   HTML123)    PDF(pc) (1524KB)(8787)       Save

    [Significance] Soil stands as the fundamental pillar of agricultural production, with its quality being intrinsically linked to the efficiency and sustainability of farming practices. Historically, the intensive cultivation and soil erosion have led to a marked deterioration in some arable lands, characterized by a sharp decrease in soil organic matter, diminished fertility, and a decline in soil's structural integrity and ecological functions. In the strategic framework of safeguarding national food security and advancing the frontiers of smart and precision agriculture, the march towards agricultural modernization continues apace, intensifying the imperative for meticulous soil quality management. Consequently, there is an urgent need for the rrapid acquisition of soil's physical and chemical parameters. Interdisciplinary scholars have delved into soil monitoring research, achieving notable advancements that promise to revolutionize the way we understand and manage soil resource. [Progress] Utilizing the the Web of Science platform, a comprehensive literature search was conducted on the topic of "soil," further refined with supplementary keywords such as "electrochemistry", "spectroscopy", "electromagnetic", "ground-penetrating radar", and "satellite". The resulting literature was screened, synthesized, and imported into the CiteSpace visualization tool. A keyword emergence map was yielded, which delineates the trajectory of research in soil physical and chemical parameter detection technology. Analysis of the keyword emergence map reveals a paradigm shift in the acquisition of soil physical and chemical parameters, transitioning from conventional indoor chemical and spectrometry analyses to outdoor, real-time detection methods. Notably, soil sensors integrated into drones and satellites have garnered considerable interest. Additionally, emerging monitoring technologies, including biosensing and terahertz spectroscopy, have made their mark in recent years. Drawing from this analysis, the prevailing technologies for soil physical and chemical parameter information acquisition in agricultural fields have been categorized and summarized. These include: 1. Rapid Laboratory Testing Techniques: Primarily hinged on electrochemical and spectrometry analysis, these methods offer the dual benefits of time and resource efficiency alongside high precision; 2. Rapid Near-Ground Sensing Techniques: Leveraging electromagnetic induction, ground-penetrating radar, and various spectral sensors (multispectral, hyperspectral, and thermal infrared), these techniques are characterized by their high detection accuracy and swift operation. 3. Satellite Remote Sensing Techniques: Employing direct inversion, indirect inversion, and combined analysis methods, these approaches are prized for their efficiency and extensive coverage. 4. Innovative Rapid Acquisition Technologies: Stemming from interdisciplinary research, these include biosensing, environmental magnetism, terahertz spectroscopy, and gamma spectroscopy, each offering novel avenues for soil parameter detection. An in-depth examination and synthesis of the principles, applications, merits, and limitations of each technology have been provided. Moreover, a forward-looking perspective on the future trajectory of soil physical and chemical parameter acquisition technology has been offered, taking into account current research trends and hotspots. [Conclusions and Prospects] Current advancements in the technology for rapaid acquiring soil physical and chemical parameters in agricultural fields have been commendable, yet certain challenges persist. The development of near-ground monitoring sensors is constrained by cost, and their reliability, adaptability, and specialization require enhancement to effectively contend with the intricate and varied conditions of farmland environments. Additionally, remote sensing inversion techniques are confronted with existing limitations in data acquisition, processing, and application. To further develop the soil physical and chemical parameter acquisition technology and foster the evolution of smart agriculture, future research could beneficially delve into the following four areas: Designing portable, intelligent, and cost-effective near-ground soil information acquisition systems and equipment to facilitate rapid on-site soil information detection; Enhancing the performance of low-altitude soil information acquisition platforms and refine the methods for data interpretation to ensure more accurate insights; Integrating multifactorial considerations to construct robust satellite remote sensing inversion models, leveraging diverse and open cloud computing platforms for in-depth data analysis and mining; Engaging in thorough research on the fusion of multi-source data in the acquisition of soil physical and chemical parameter information, developing soil information sensing algorithms and models with strong generalizability and high reliability to achieve rapaid, precise, and intelligent acquisition of soil parameters.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospect of Multi-robot Collaborative SLAM in Complex Agricultural Scenarios
    MA Nan, CAO Shanshan, BAI Tao, KONG Fantao, SUN Wei
    Smart Agriculture    2024, 6 (6): 23-43.   DOI: 10.12133/j.smartag.SA202406005
    Abstract770)   HTML86)    PDF(pc) (2300KB)(8599)       Save

    [Significance] The rapid development of artificial intelligence and automation has greatly expanded the scope of agricultural automation, with applications such as precision farming using unmanned machinery, robotic grazing in outdoor environments, and automated harvesting by orchard-picking robots. Collaborative operations among multiple agricultural robots enhance production efficiency and reduce labor costs, driving the development of smart agriculture. Multi-robot simultaneous localization and mapping (SLAM) plays a pivotal role by ensuring accurate mapping and localization, which are essential for the effective management of unmanned farms. Compared to single-robot SLAM, multi-robot systems offer several advantages, including higher localization accuracy, larger sensing ranges, faster response times, and improved real-time performance. These capabilities are particularly valuable for completing complex tasks efficiently. However, deploying multi-robot SLAM in agricultural settings presents significant challenges. Dynamic environmental factors, such as crop growth, changing weather patterns, and livestock movement, increase system uncertainty. Additionally, agricultural terrains vary from open fields to irregular greenhouses, requiring robots to adjust their localization and path-planning strategies based on environmental conditions. Communication constraints, such as unstable signals or limited transmission range, further complicate coordination between robots. These combined challenges make it difficult to implement multi-robot SLAM effectively in agricultural environments. To unlock the full potential of multi-robot SLAM in agriculture, it is essential to develop optimized solutions that address the specific technical demands of these scenarios. [Progress] Existing review studies on multi-robot SLAM mainly focus on a general technological perspective, summarizing trends in the development of multi-robot SLAM, the advantages and limitations of algorithms, universally applicable conditions, and core issues of key technologies. However, there is a lack of analysis specifically addressing multi-robot SLAM under the characteristics of complex agricultural scenarios. This study focuses on the main features and applications of multi-robot SLAM in complex agricultural scenarios. The study analyzes the advantages and limitations of multi-robot SLAM, as well as its applicability and application scenarios in agriculture, focusing on four key components: multi-sensor data fusion, collaborative localization, collaborative map building, and loopback detection. From the perspective of collaborative operations in multi-robot SLAM, the study outlines the classification of SLAM frameworks, including three main collaborative types: centralized, distributed, and hybrid. Based on this, the study summarizes the advantages and limitations of mainstream multi-robot SLAM frameworks, along with typical scenarios in robotic agricultural operations where they are applicable. Additionally, it discusses key issues faced by multi-robot SLAM in complex agricultural scenarios, such as low accuracy in mapping and localization during multi-sensor fusion, restricted communication environments during multi-robot collaborative operations, and low accuracy in relative pose estimation between robots. [Conclusions and Prospects] To enhance the applicability and efficiency of multi-robot SLAM in complex agricultural scenarios, future research needs to focus on solving these critical technological issues. Firstly, the development of enhanced data fusion algorithms will facilitate improved integration of sensor information, leading to greater accuracy and robustness of the system. Secondly, the combination of deep learning and reinforcement learning techniques is expected to empower robots to better interpret environmental patterns, adapt to dynamic changes, and make more effective real-time decisions. Thirdly, large language models will enhance human-robot interaction by enabling natural language commands, improving collaborative operations. Finally, the integration of digital twin technology will support more intelligent path planning and decision-making processes, especially in unmanned farms and livestock management systems. The convergence of digital twin technology with SLAM is projected to yield innovative solutions for intelligent perception and is likely to play a transformative role in the realm of agricultural automation. This synergy is anticipated to revolutionize the approach to agricultural tasks, enhancing their efficiency and reducing the reliance on labor.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospects of Key Navigation Technologies for Facility Agricultural Robots
    HE Yong, HUANG Zhenyu, YANG Ningyuan, LI Xiyao, WANG Yuwei, FENG Xuping
    Smart Agriculture    2024, 6 (5): 1-19.   DOI: 10.12133/j.smartag.SA202404006
    Abstract1548)   HTML326)    PDF(pc) (2130KB)(7364)       Save

    [Significance] With the rapid development of robotics technology and the persistently rise of labor costs, the application of robots in facility agriculture is becoming increasingly widespread. These robots can enhance operational efficiency, reduce labor costs, and minimize human errors. However, the complexity and diversity of facility environments, including varying crop layouts and lighting conditions, impose higher demands on robot navigation. Therefore, achieving stable, accurate, and rapid navigation for robots has become a key issue. Advanced sensor technologies and algorithms have been proposed to enhance robots' adaptability and decision-making capabilities in dynamic environments. This not only elevates the automation level of agricultural production but also contributes to more intelligent agricultural management. [Progress] This paper reviews the key technologies of automatic navigation for facility agricultural robots. It details beacon localization, inertial positioning, simultaneous localization and mapping (SLAM) techniques, and sensor fusion methods used in autonomous localization and mapping. Depending on the type of sensors employed, SLAM technology could be subdivided into vision-based, laser-based and fusion systems. Fusion localization is further categorized into data-level, feature-level, and decision-level based on the types and stages of the fused information. The application of SLAM technology and fusion localization in facility agriculture has been increasingly common. Global path planning plays a crucial role in enhancing the operational efficiency and safety of facility aricultural robots. This paper discusses global path planning, classifying it into point-to-point local path planning and global traversal path planning. Furthermore, based on the number of optimization objectives, it was divided into single-objective path planning and multi-objective path planning. In regard to automatic obstacle avoidance technology for robots, the paper discusses sevelral commonly used obstacle avoidance control algorithms commonly used in facility agriculture, including artificial potential field, dynamic window approach and deep learning method. Among them, deep learning methods are often employed for perception and decision-making in obstacle avoidance scenarios. [Conclusions and Prospects] Currently, the challenges for facility agricultural robot navigation include complex scenarios with significant occlusions, cost constraints, low operational efficiency and the lack of standardized platforms and public datasets. These issues not only affect the practical application effectiveness of robots but also constrain the further advancement of the industry. To address these challenges, future research can focus on developing multi-sensor fusion technologies, applying and optimizing advanced algorithms, investigating and implementing multi-robot collaborative operations and establishing standardized and shared data platforms.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Orchard-Wide Visual Perception and Autonomous Operation of Fruit Picking Robots: A Review
    CHEN Mingyou, LUO Lufeng, LIU Wei, WEI Huiling, WANG Jinhai, LU Qinghua, LUO Shaoming
    Smart Agriculture    2024, 6 (5): 20-39.   DOI: 10.12133/j.smartag.SA202405022
    Abstract1103)   HTML179)    PDF(pc) (4030KB)(6452)       Save

    [Significance] Fruit-picking robot stands as a crucial solution for achieving intelligent fruit harvesting. Significant progress has been made in developing foundational methods for picking robots, such as fruit recognition, orchard navigation, path planning for picking, and robotic arm control, the practical implementation of a seamless picking system that integrates sensing, movement, and picking capabilities still encounters substantial technical hurdles. In contrast to current picking systems, the next generation of fruit-picking robots aims to replicate the autonomous skills exhibited by human fruit pickers. This involves effectively performing ongoing tasks of perception, movement, and picking without human intervention. To tackle this challenge, this review delves into the latest research methodologies and real-world applications in this field, critically assesses the strengths and limitations of existing methods and categorizes the essential components of continuous operation into three sub-modules: local target recognition, global mapping, and operation planning. [Progress] Initially, the review explores methods for recognizing nearby fruit and obstacle targets. These methods encompass four main approaches: low-level feature fusion, high-level feature learning, RGB-D information fusion, and multi-view information fusion, respectively. Each of these approaches incorporates advanced algorithms and sensor technologies for cluttered orchard environments. For example, low-level feature fusion utilizes basic attributes such as color, shapes and texture to distinguish fruits from backgrounds, while high-level feature learning employs more complex models like convolutional neural networks to interpret the contextual relationships within the data. RGB-D information fusion brings depth perception into the mix, allowing robots to gauge the distance to each fruit accurately. Multi-view information fusion tackles the issue of occlusions by combining data from multiple cameras and sensors around the robot, providing a more comprehensive view of the environment and enabling more reliable sensing. Subsequently, the review shifts focus to orchard mapping and scene comprehension on a broader scale. It points out that current mapping methods, while effective, still struggle with dynamic changes in the orchard, such as variations of fruits and light conditions. Improved adaptation techniques, possibly through machine learning models that can learn and adjust to different environmental conditions, are suggested as a way forward. Building upon the foundation of local and global perception, the review investigates strategies for planning and controlling autonomous behaviors. This includes not only the latest advancements in devising movement paths for robot mobility but also adaptive strategies that allow robots to react to unexpected obstacles or changes within the whole environment. Enhanced strategies for effective fruit picking using the Eye-in-Hand system involve the development of more dexterous robotic hands and improved algorithms for precisely predicting the optimal picking point of each fruit. The review also identifies a crucial need for further advancements in the dynamic behavior and autonomy of these technologies, emphasizing the importance of continuous learning and adaptive control systems to improve operational efficiency in diverse orchard environments. [Conclusions and Prospects] The review underscores the critical importance of coordinating perception, movement, and picking modules to facilitate the transition from a basic functional prototype to a practical machine. Moreover, it emphasizes the necessity of enhancing the robustness and stability of core algorithms governing perception, planning, and control, while ensuring their seamless coordination which is a central challenge that emerges. Additionally, the review raises unresolved questions regarding the application of picking robots and outlines future trends, include deeper integration of stereo vision and deep learning, enhanced global vision sampling, and the establishment of standardized evaluation criteria for overall operational performance. The paper can provide references for the eventual development of robust, autonomous, and commercially viable picking robots in the future.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Big Models in Agriculture: Key Technologies, Application and Future Directions
    GUO Wang, YANG Yusen, WU Huarui, ZHU Huaji, MIAO Yisheng, GU Jingqiu
    Smart Agriculture    2024, 6 (2): 1-13.   DOI: 10.12133/j.smartag.SA202403015
    Abstract3723)   HTML531)    PDF(pc) (1482KB)(6381)       Save

    [Significance] Big Models, or Foundation Models, have offered a new paradigm in smart agriculture. These models, built on the Transformer architecture, incorporate numerous parameters and have undergone extensive training, often showing excellent performance and adaptability, making them effective in addressing agricultural issues where data is limited. Integrating big models in agriculture promises to pave the way for a more comprehensive form of agricultural intelligence, capable of processing diverse inputs, making informed decisions, and potentially overseeing entire farming systems autonomously. [Progress] The fundamental concepts and core technologies of big models are initially elaborated from five aspects: the generation and core principles of the Transformer architecture, scaling laws of extending big models, large-scale self-supervised learning, the general capabilities and adaptions of big models, and the emerging capabilities of big models. Subsequently, the possible application scenarios of the big model in the agricultural field are analyzed in detail, the development status of big models is described based on three types of the models: Large language models (LLMs), large vision models (LVMs), and large multi-modal models (LMMs). The progress of applying big models in agriculture is discussed, and the achievements are presented. [Conclusions and Prospects] The challenges and key tasks of applying big models technology in agriculture are analyzed. Firstly, the current datasets used for agricultural big models are somewhat limited, and the process of constructing these datasets can be both expensive and potentially problematic in terms of copyright issues. There is a call for creating more extensive, more openly accessible datasets to facilitate future advancements. Secondly, the complexity of big models, due to their extensive parameter counts, poses significant challenges in terms of training and deployment. However, there is optimism that future methodological improvements will streamline these processes by optimizing memory and computational efficiency, thereby enhancing the performance of big models in agriculture. Thirdly, these advanced models demonstrate strong proficiency in analyzing image and text data, suggesting potential future applications in integrating real-time data from IoT devices and the Internet to make informed decisions, manage multi-modal data, and potentially operate machinery within autonomous agricultural systems. Finally, the dissemination and implementation of these big models in the public agricultural sphere are deemed crucial. The public availability of these models is expected to refine their capabilities through user feedback and alleviate the workload on humans by providing sophisticated and accurate agricultural advice, which could revolutionize agricultural practices.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Technology Knowledge Intelligent Question-Answering System Based on Large Language Model
    WANG Ting, WANG Na, CUI Yunpeng, LIU Juan
    Smart Agriculture    2023, 5 (4): 105-116.   DOI: 10.12133/j.smartag.SA202311005
    Abstract2822)   HTML391)    PDF(pc) (1475KB)(5811)       Save

    [Objective] The rural revitalization strategy presents novel requisites for the extension of agricultural technology. However, the conventional method encounters the issue of a contradiction between supply and demand. Therefore, there is a need for further innovation in the supply form of agricultural knowledge. Recent advancements in artificial intelligence technologies, such as deep learning and large-scale neural networks, particularly the advent of large language models (LLMs), render anthropomorphic and intelligent agricultural technology extension feasible. With the agricultural technology knowledge service of fruit and vegetable as the demand orientation, the intelligent agricultural technology question answering system was built in this research based on LLM, providing agricultural technology extension services, including guidance on new agricultural knowledge and question-and-answer sessions. This facilitates farmers in accessing high-quality agricultural knowledge at their convenience. [Methods] Through an analysis of the demands of strawberry farmers, the agricultural technology knowledge related to strawberry cultivation was categorized into six themes: basic production knowledge, variety screening, interplanting knowledge, pest diagnosis and control, disease diagnosis and control, and drug damage diagnosis and control. Considering the current situation of agricultural technology, two primary tasks were formulated: named entity recognition and question answering related to agricultural knowledge. A training corpus comprising entity type annotations and question-answer pairs was constructed using a combination of automatic machine annotation and manual annotation, ensuring a small yet high-quality sample. After comparing four existing Large Language Models (Baichuan2-13B-Chat, ChatGLM2-6B, Llama 2-13B-Chat, and ChatGPT), the model exhibiting the best performance was chosen as the base LLM to develop the intelligent question-answering system for agricultural technology knowledge. Utilizing a high-quality corpus, pre-training of a Large Language Model and the fine-tuning method, a deep neural network with semantic analysis, context association, and content generation capabilities was trained. This model served as a Large Language Model for named entity recognition and question answering of agricultural knowledge, adaptable to various downstream tasks. For the task of named entity recognition, the fine-tuning method of Lora was employed, fine-tuning only essential parameters to expedite model training and enhance performance. Regarding the question-answering task, the Prompt-tuning method was used to fine-tune the Large Language Model, where adjustments were made based on the generated content of the model, achieving iterative optimization. Model performance optimization was conducted from two perspectives: data and model design. In terms of data, redundant or unclear data was manually removed from the labeled corpus. In terms of the model, a strategy based on retrieval enhancement generation technology was employed to deepen the understanding of agricultural knowledge in the Large Language Model and maintain real-time synchronization of knowledge, alleviating the problem of LLM hallucination. Drawing upon the constructed Large Language Model, an intelligent question-answering system was developed for agricultural technology knowledge. This system demonstrates the capability to generate high-precision and unambiguous answers, while also supporting the functionalities of multi-round question answering and retrieval of information sources. [Results and Discussions] Accuracy rate and recall rate served as indicators to evaluate the named entity recognition task performance of the Large Language Models. The results indicated that the performance of Large Language Models was closely related to factors such as model structure, the scale of the labeled corpus, and the number of entity types. After fine-tuning, the ChatGLM Large Language Model demonstrated the highest accuracy and recall rate. With the same number of entity types, a higher number of annotated corpora resulted in a higher accuracy rate. Fine-tuning had different effects on different models, and overall, it improved the average accuracy of all models under different knowledge topics, with ChatGLM, Llama, and Baichuan values all surpassing 85%. The average recall rate saw limited increase, and in some cases, it was even lower than the values before fine-tuning. Assessing the question-answering task of Large Language Models using hallucination rate and semantic similarity as indicators, data optimization and retrieval enhancement generation techniques effectively reduced the hallucination rate by 10% to 40% and improved semantic similarity by more than 15%. These optimizations significantly enhanced the generated content of the models in terms of correctness, logic, and comprehensiveness. [Conclusion] The pre-trained Large Language Model of ChatGLM exhibited superior performance in named entity recognition and question answering tasks in the agricultural field. Fine-tuning pre-trained Large Language Models for downstream tasks and optimizing based on retrieval enhancement generation technology mitigated the problem of language hallucination, markedly improving model performance. Large Language Model technology has the potential to innovate agricultural technology knowledge service modes and optimize agricultural knowledge extension. This can effectively reduce the time cost for farmers to obtain high-quality and effective knowledge, guiding more farmers towards agricultural technology innovation and transformation. However, due to challenges such as unstable performance, further research is needed to explore optimization methods for Large Language Models and their application in specific scenarios.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Robots: Technology Progress, Challenges and Trends
    ZHAO Chunjiang, FAN Beibei, LI Jin, FENG Qingchun
    Smart Agriculture    2023, 5 (4): 1-15.   DOI: 10.12133/j.smartag.SA202312030
    Abstract3551)   HTML472)    PDF(pc) (2498KB)(4784)       Save

    [Significance] Autonomous and intelligent agricultural machinery, characterized by green intelligence, energy efficiency, and reduced emissions, as well as high intelligence and man-machine collaboration, will serve as the driving force behind global agricultural technology advancements and the transformation of production methods in the context of smart agriculture development. Agricultural robots, which utilize intelligent control and information technology, have the unique advantage of replacing manual labor. They occupy the strategic commanding heights and competitive focus of global agricultural equipment and are also one of the key development directions for accelerating the construction of China's agricultural power. World agricultural powers and China have incorporated the research, development, manufacturing, and promotion of agricultural robots into their national strategies, respectively strengthening the agricultural robot policy and planning layout based on their own agricultural development characteristics, thus driving the agricultural robot industry into a stable growth period. [Progress] This paper firstly delves into the concept and defining features of agricultural robots, alongside an exploration of the global agricultural robot development policy and strategic planning blueprint. Furthermore, sheds light on the growth and development of the global agricultural robotics industry; Then proceeds to analyze the industrial backdrop, cutting-edge advancements, developmental challenges, and crucial technology aspects of three representative agricultural robots, including farmland robots, orchard picking robots, and indoor vegetable production robots. Finally, summarizes the disparity between Chinese agricultural robots and their foreign counterparts in terms of advanced technologies. (1) An agricultural robot is a multi-degree-of-freedom autonomous operating equipment that possesses accurate perception, autonomous decision-making, intelligent control, and automatic execution capabilities specifically designed for agricultural environments. When combined with artificial intelligence, big data, cloud computing, and the Internet of Things, agricultural robots form an agricultural robot application system. This system has relatively mature applications in key processes such as field planting, fertilization, pest control, yield estimation, inspection, harvesting, grafting, pruning, inspection, harvesting, transportation, and livestock and poultry breeding feeding, inspection, disinfection, and milking. Globally, agricultural robots, represented by plant protection robots, have entered the industrial application phase and are gradually realizing commercialization with vast market potential. (2) Compared to traditional agricultural machinery and equipment, agricultural robots possess advantages in performing hazardous tasks, executing batch repetitive work, managing complex field operations, and livestock breeding. In contrast to industrial robots, agricultural robots face technical challenges in three aspects. Firstly, the complexity and unstructured nature of the operating environment. Secondly, the flexibility, mobility, and commoditization of the operation object. Thirdly, the high level of technology and investment required. (3) Given the increasing demand for unmanned and less manned operations in farmland production, China's agricultural robot research, development, and application have started late and progressed slowly. The existing agricultural operation equipment still has a significant gap from achieving precision operation, digital perception, intelligent management, and intelligent decision-making. The comprehensive performance of domestic products lags behind foreign advanced counterparts, indicating that there is still a long way to go for industrial development and application. Firstly, the current agricultural robots predominantly utilize single actuators and operate as single machines, with the development of multi-arm cooperative robots just emerging. Most of these robots primarily engage in rigid operations, exhibiting limited flexibility, adaptability, and functionality. Secondly, the perception of multi-source environments in agricultural settings, as well as the autonomous operation of agricultural robot equipment, relies heavily on human input. Thirdly, the progress of new teaching methods and technologies for human-computer natural interaction is rather slow. Lastly, the development of operational infrastructure is insufficient, resulting in a relatively low degree of "mechanization". [Conclusions and Prospects] The paper anticipates the opportunities that arise from the rapid growth of the agricultural robotics industry in response to the escalating global shortage of agricultural labor. It outlines the emerging trends in agricultural robot technology, including autonomous navigation, self-learning, real-time monitoring, and operation control. In the future, the path planning and navigation information perception of agricultural robot autonomy are expected to become more refined. Furthermore, improvements in autonomous learning and cross-scenario operation performance will be achieved. The development of real-time operation monitoring of agricultural robots through digital twinning will also progress. Additionally, cloud-based management and control of agricultural robots for comprehensive operations will experience significant growth. Steady advancements will be made in the innovation and integration of agricultural machinery and techniques.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Reconstruction of U.S. Regional-Scale Soybean SIF Based on MODIS Data and BP Neural Network
    YAO Jianen, LIU Haiqiu, YANG Man, FENG Jinying, CHEN Xiu, ZHANG Peipei
    Smart Agriculture    2024, 6 (5): 40-50.   DOI: 10.12133/j.smartag.SA202309006
    Abstract377)   HTML44)    PDF(pc) (1648KB)(4088)       Save

    [Objective] Sunlight-induced chlorophyll fluorescence (SIF) data obtained from satellites suffer from issues such as low spatial and temporal resolution, and discrete footprint because of the limitations imposed by satellite orbits. To address these problems, obtaining higher resolution SIF data, most reconstruction studies are based on low-resolution satellite SIF. Moreover, the spatial resolution of most SIF reconstruction products is still not enough to be directly used for the study of crop photosynthetic rate at the regional scale. Although some SIF products boast elevated resolutions, but these derive not from the original satellite SIF data reconstruct but instead evolve from secondary reconstructions based on preexisting SIF reconstruction products. Satellite OCO-2 (The Orbiting Carbon Obsevatory-2) equipped with a high-resolution spectrometer, OCO-2 SIF has higher spatial resolution (1.29×2.25 km) compared to other original SIF products, making it suitable in advancing the realm of high-resolution SIF data reconstruction, particularly within the context of regional-scale crop studies. [Methods] This research primarily exploration SIF reconstruct at the regional scale, mainly focused on the partial soybean planting regions nestled within the United States. The selection of MODIS raw data hinged on a meticulous consideration of environmental conditions, the distinctive physiological attributes of soybeans, and an exhaustive evaluation of factors intricately linked to OCO-2 SIF within these soybean planting regions. The primary tasks of this research encompassed reconstructing high resolution soybean SIF while concurrently executing a rigorous assessment of the reconstructed SIF's quality. During the dataset construction process, amalgamated SIF data from multiple soybean planting regions traversed by the OCO-2 satellite's footprint to retain as many of the available original SIF samples as possible. This approach provided the subsequent SIF reconstruction model with a rich source of SIF data. SIF data obtained beneath the satellite's trajectory were matched with various MODIS datasets, including enhanced vegetation index (EVI), fraction of photosynthetically active radiation (FPAR), and land surface temperature (LST), resulting in the creation of a multisource remote sensing dataset ultimately used for model training. Because of the multisource remote sensing dataset encompassed the most relevant explanatory variables within each SIF footprint coverage area concerning soybean physiological structure and environmental conditions. Through the activation functions in the BP neural network, it enhanced the understanding of the complex nonlinear relationships between the original SIF data and these MODIS products. Leveraging these inherent nonlinear relationships, compared and analyzed the effects of different combinations of explanatory variables on SIF reconstruction, mainly analyzing the three indicators of goodness of fit R2, root mean square error RMSE, and mean absolute error MAE, and then selecting the best SIF reconstruction model, generate a regional scale, spatially continuous, and high temporal resolution (500 m, 8 d) soybean SIF reconstruction dataset (BPSIF). [Results and Discussions] The research findings confirmed the strong performance of the SIF reconstruction model in predicting soybean SIF. After simultaneously incorporating EVI, FPAR, and LST as explanatory variables to model, achieved a goodness of fit with an R2 value of 0.84, this statistical metric validated the model's capability in predicting SIF data, it also reflected that the reconstructed 8 d time resolution of SIF data's reliability of applying to small-scale agricultural crop photosynthesis research with 500 m×500 m spatial scale. Based on this optimal model, generated the reconstructed SIF product (BPSIF). The Pearson correlation coefficient between the original OCO-2 SIF data and MODIS GPP stood were at a modest 0.53. In stark contrast, the correlation coefficient between BPSIF and MODIS Gross Primary Productivity (GPP) rosed significantly to 0.80. The increased correlation suggests that BPSIF could more accurately reflect the dynamic changes in GPP during the soybean growing season, making it more reliable compared to the original SIF data. Selected soybean planting areas in the United States with relatively single crop cultivation as the research area, based on high spatial resolution (1.29 km×2.25 km) OCO-2 SIF data, greatly reduced vegetation heterogeneity under a single SIF footprint. [Conclusions] The BPSIF proposed has significantly enhancing the regional and temporal continuity of OCO-2 SIF while preserving the time and spatial attributes contained in the original SIF dataset. Within the study area, BPSIF exhibits a significantly improved correlation with MODIS GPP compared to the original OCO-2 SIF. The proposed OCO-2 SIF data reconstruction method in this study holds the potential to provide a more reliable SIF dataset. This dataset has the potential to drive further understanding of soybean SIF at finer spatial and temporal scales, as well as find its relationship with soybean GPP.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    MSH-YOLOv8: Mushroom Small Object Detection Method with Scale Reconstruction and Fusion
    YE Dapeng, JING Jun, ZHANG Zhide, LI Huihuang, WU Haoyu, XIE Limin
    Smart Agriculture    2024, 6 (5): 139-152.   DOI: 10.12133/j.smartag.SA202404002
    Abstract650)   HTML82)    PDF(pc) (2660KB)(3808)       Save

    [Objective] Traditional object detection algorithms applied in the agricultural field, such as those used for crop growth monitoring and harvesting, often suffer from insufficient accuracy. This is particularly problematic for small crops like mushrooms, where recognition and detection are more challenging. The introduction of small object detection technology promises to address these issues, potentially enhancing the precision, efficiency, and economic benefits of agricultural production management. However, achieving high accuracy in small object detection has remained a significant challenge, especially when dealing with varying image sizes and target scales. Although the YOLO series models excel in speed and large object detection, they still have shortcomings in small object detection. To address the issue of maintaining high accuracy amid changes in image size and target scale, a novel detection model, Multi-Strategy Handling YOLOv8 (MSH-YOLOv8), was proposed. [Methods] The proposed MSH-YOLOv8 model builds upon YOLOv8 by incorporating several key enhancements aimed at improving sensitivity to small-scale targets and overall detection performance. Firstly, an additional detection head was added to increase the model's sensitivity to small objects. To address computational redundancy and improve feature extraction, the Swin Transformer detection structure was introduced into the input module of the head network, creating what was termed the "Swin Head (SH)". Moreover, the model integrated the C2f_Deformable convolutionv4 (C2f_DCNv4) structure, which included deformable convolutions, and the Swin Transformer encoder structure, termed "Swinstage", to reconstruct the YOLOv8 backbone network. This optimization enhanced feature propagation and extraction capabilities, increasing the network's ability to handle targets with significant scale variations. Additionally, the normalization-based attention module (NAM) was employed to improve performance without compromising detection speed or computational complexity. To further enhance training efficacy and convergence speed, the original loss function CIoU was replaced with wise-intersection over union (WIoU) Loss. Furthermore, experiments were conducted using mushrooms as the research subject on the open Fungi dataset. Approximately 200 images with resolution sizes around 600×800 were selected as the main research material, along with 50 images each with resolution sizes around 200×400 and 1 000×1 200 to ensure representativeness and generalization of image sizes. During the data augmentation phase, a generative adversarial network (GAN) was utilized for resolution reconstruction of low-resolution images, thereby preserving semantic quality as much as possible. In the post-processing phase, dynamic resolution training, multi-scale testing, soft non-maximum suppression (Soft-NMS), and weighted boxes fusion (WBF) were applied to enhance the model's small object detection capabilities under varying scales. [Results and Discussions] The improved MSH-YOLOv8 achieved an average precision at 50% (AP50) intersection over union of 98.49% and an AP@50-95 of 75.29%, with the small object detection metric APs reaching 39.73%. Compared to mainstream models like YOLOv8, these metrics showed improvements of 2.34%, 4.06% and 8.55%, respectively. When compared to the advanced TPH-YOLOv5 model, the improvements were 2.14%, 2.76% and 6.89%, respectively. The ensemble model, MSH-YOLOv8-ensemble, showed even more significant improvements, with AP50 and APs reaching 99.14% and 40.59%, respectively, an increase of 4.06% and 8.55% over YOLOv8. These results indicate the robustness and enhanced performance of the MSH-YOLOv8 model, particularly in detecting small objects under varying conditions. Further application of this methodology on the Alibaba Cloud Tianchi databases "Tomato Detection" and "Apple Detection" yielded MSH-YOLOv8-t and MSH-YOLOv8-a models (collectively referred to as MSH-YOLOv8). Visual comparison of detection results demonstrated that MSH-YOLOv8 significantly improved the recognition of dense and blurry small-scale tomatoes and apples. This indicated that the MSH-YOLOv8 method possesses strong cross-dataset generalization capability and effectively recognizes small-scale targets. In addition to quantitative improvements, qualitative assessments showed that the MSH-YOLOv8 model could handle complex scenarios involving occlusions, varying lighting conditions, and different growth stages of the crops. This demonstrates the practical applicability of the model in real-world agricultural settings, where such challenges are common. [Conclusions] The MSH-YOLOv8 improvement method proposed in this study effectively enhances the detection accuracy of small mushroom targets under varying image sizes and target scales. This approach leverages multiple strategies to optimize both the architecture and the training process, resulting in a robust model capable of high-precision small object detection. The methodology's application to other datasets, such as those for tomato and apple detection, further underscores its generalizability and potential for broader use in agricultural monitoring and management tasks.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Development Trend of Mountainous Tractor Leveling and Anti-Rollover System
    MU Xiaodong, YANG Fuzeng, DUAN Luojia, LIU Zhijie, SONG Zhuoying, LI Zonglin, GUAN Shouqing
    Smart Agriculture    2024, 6 (3): 1-16.   DOI: 10.12133/j.smartag.SA202312015
    Abstract967)   HTML111)    PDF(pc) (2448KB)(3743)       Save

    [Significance] The mechanization, automation and intelligentization of agricultural equipment are key factors to improve operation efficiency, free up labor force and promote the sustainable development of agriculture. It is also the hot spot of research and development of agricultural machinery industry in the future. In China, hills and mountains serves as vital production bases for agricultural products, accounting for about 70% of the country's land area. In addition, these regions face various environmental factors such as steep slopes, narrow road, small plots, complex terrain and landforms, as well as harsh working environment. Moreover, there is a lack of reliable agricultural machinery support across various production stages, along with a shortage of theoretical frameworks to guide the research and development of agricultural machinery tailored to hilly and mountainous locales. [Progress] This article focuses on the research advances of tractor leveling and anti-overturning systems in hilly and mountainous areas, including tractor body, cab and seat leveling technology, tractor rear suspension and implement leveling slope adaptive technology, and research progress on tractor anti-overturning protection devices and warning technology. The vehicle body leveling mechanism can be roughly divided into five types based on its different working modes: parallel four bar, center of gravity adjustable, hydraulic differential high, folding and twisting waist, and omnidirectional leveling. These mechanisms aim to address the issue of vehicle tilting and easy overturning when traversing or working on sloping or rugged roads. By keeping the vehicle body posture horizontal or adjusting the center of gravity within a stable range, the overall driving safety of the vehicle can be improved to ensure the accuracy of operation. Leveling the driver's cab and seats can mitigate the lateral bumps experienced by the driver during rough or sloping operations, reducing driver fatigue and minimizing strain on the lumbar and cervical spine, thereby enhancing driving comfort. The adaptive technology of tractor rear suspension and implement leveling on slopes can ensure that the tractor maintains consistent horizontal contact with the ground in hilly and mountainous areas, avoiding changes in the posture of the suspended implement with the swing of the body or the driving path, which may affect the operation effect. The tractor rollover protection device and warning technology have garnered significant attention in recent years. Prioritizing driver safety, rollover warning system can alert the driver in advance of the dangerous state of the tractor, automatically adjust the vehicle before rollover, or automatically open the rollover protection device when it is about to rollover, and timely send accident reports to emergency contacts, thereby ensuring the safety of the driver to the greatest extent possible. [Conclusions and Prospects] The future development directions of hill and mountain tractor leveling, anti-overturning early warning, unmanned, automatic technology were looked forward: Structure optimization, high sensitivity, good stability of mountain tractor leveling system research; Study on copying system of agricultural machinery with good slope adaptability; Research on anti-rollover early warning technology of environment perception and automatic interference; Research on precision navigation technology, intelligent monitoring technology and remote scheduling and management technology of agricultural machinery; Theoretical study on longitudinal stability of sloping land. This review could provide reference for the research and development of high reliability and high safety mountain tractor in line with the complex working environment in hill and mountain areas.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Lightweight Model for Detecting Small Targets of Litchi Pests Based on Improved YOLOv10n
    LI Zusheng, TANG Jishen, KUANG Yingchun
    Smart Agriculture    2025, 7 (2): 146-159.   DOI: 10.12133/j.smartag.SA202412003
    Abstract760)   HTML106)    PDF(pc) (2262KB)(3282)       Save

    Objective The accuracy of identifying litchi pests is crucial for implementing effective control strategies and promoting sustainable agricultural development. However, the current detection of litchi pests is characterized by a high percentage of small targets, which makes target detection models challenging in terms of accuracy and parameter count, thus limiting their application in real-world production environments. To improve the identification efficiency of litchi pests, a lightweight target detection model YOLO-LP (YOLO-Litchi Pests) based on YOLOv10n was proposed. The model aimed to enhance the detection accuracy of small litchi pest targets in multiple scenarios by optimizing the network structure and loss function, while also reducing the number of parameters and computational costs. Methods Two classes of litchi insect pests (Cocoon and Gall) images were collected as datasets for modeling in natural scenarios (sunny, cloudy, post-rain) and laboratory environments. The original data were expanded through random scaling, random panning, random brightness adjustments, random contrast variations, and Gaussian blurring to balance the category samples and enhance the robustness of the model, generating a richer dataset named the CG dataset (Cocoon and Gall dataset). The YOLO-LP model was constructed after the following three improvements. Specifically, the C2f module of the backbone network (Backbone) in YOLOv10n was optimized and the C2f_GLSA module was constructed using the global-to-local spatial aggregation (GLSA) module to focus on small targets and enhance the differentiation between the targets and the backgrounds, while simultaneously reducing the number of parameters and computation. A frequency-aware feature fusion module (FreqFusion) was introduced into the neck network (Neck) of YOLOv10n and a frequency-aware path aggregation network (FreqPANet) was designed to reduce the complexity of the model and address the problem of fuzzy and shifted target boundaries. The SCYLLA-IoU (SIoU) loss function replaced the Complete-IoU (CIoU) loss function from the baseline model to optimize the target localization accuracy and accelerate the convergence of the training process. Results and Discussions YOLO-LP achieved 90.9%, 62.2%, and 59.5% for AP50, AP50:95, and AP-Small50:95 in the CG dataset, respectively, and 1.9%, 1.0%, and 1.2% higher than the baseline model. The number of parameters and the computational costs were reduced by 13% and 17%, respectively. These results suggested that YOLO-LP had a high accuracy and lightweight design. Comparison experiments with different attention mechanisms validated the effectiveness of the GLSA module. After the GLSA module was added to the baseline model, AP50, AP50:95, and AP-Small50:95 achieved the highest performance in the CG dataset, reaching 90.4%, 62.0%, and 59.5%, respectively. Experiment results comparing different loss functions showed that the SIoU loss function provided better fitting and convergence speed in the CG dataset. Ablation test results revealed that the validity of each model improvement and the detection performance of any combination of the three improvements was significantly better than the baseline model in the YOLO-LP model. The performance of the models was optimal when all three improvements were applied simultaneously. Compared to several mainstream models, YOLO-LP exhibited the best overall performance, with a model size of only 5.1 MB, 1.97 million parameters (Params), and a computational volume of 5.4 GFLOPs. Compared to the baseline model, the detection of the YOLO-LP performance was significantly improved across four multiple scenarios. In the sunny day scenario, AP50, AP50:95, and AP-Small50:95 increased by 1.9%, 1.0 %, and 2.0 %, respectively. In the cloudy day scenario, AP50, AP50:95, and AP-Small50:95 increased by 2.5%, 1.3%, and 1.3%, respectively. In the post-rain scenario, AP50, AP50:95, and AP-Small50:95 increased by 2.0%, 2.4%, and 2.4%, respectively. In the laboratory scenario, only AP50 increased by 0.7% over the baseline model. These findings indicated that YOLO-LP achieved higher accuracy and robustness in multi-scenario small target detection of litchi pests. Conclusions The proposed YOLO-LP model could improve detection accuracy and effectively reduce the number of parameters and computational costs. It performed well in small target detection of litchi pests and demonstrated strong robustness across different scenarios. These improvements made the model more suitable for deployment on resource-constrained mobile and edge devices. The model provided a valuable technical reference for small target detection of litchi pests in various scenarios.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Three-Dimensional Dynamic Growth and Yield Simulation Model of Daylily Plants
    ZHANG Yue, LI Weijia, HAN Zhiping, ZHANG Kun, LIU Jiawen, HENKE Michael
    Smart Agriculture    2024, 6 (2): 140-153.   DOI: 10.12133/j.smartag.SA202310011
    Abstract446)   HTML39)    PDF(pc) (2579KB)(3084)       Save

    [Objective] The daylily, a perennial herb in the lily family, boasts a rich nutritional profile. Given its economic importance, enhancing its yield is a crucial objective. However, current research on daylily cultivation is limited, especially regarding three-dimensional dynamic growth simulation of daylily plants. In order to establish a technological foundation for improved cultivation management, growth dynamics prediction, and the development of plant variety types in daylily crops, this study introduces an innovative three-dimensional dynamic growth and yield simulation model for daylily plants. [Methods] The open-source GroIMP software platform was used to simulate and visualize three-dimensional scenes. With Datong daylily, the primary cultivated variety of daylily in the Datong area, as the research subject, a field experiment was conducted from March to September 2022, which covered the growth season of daylily. Through actual cultivation experiment measurements, morphological data and leaf photosynthetic physiological parameters of daylily leaves, flower stems, flower buds, and other organs were collected. The functional-structural plant model (FSPM) platform's three-dimensional modeling technology was employed to establish the Cloud Cover-based solar radiation models (CSRMs) and the Farquhar, von Camerer, and Berry model (FvCB model) suitable for daylily. Moreover, based on the source-sink relationship of daylily, the carbon allocation model of daylily photosynthetic products was developed. By using the β growth function, the growth simulation model of daylily organs was constructed, and the daily morphological data of daylily during the growth period were calculated, achieving the three-dimensional dynamic growth and yield simulation of daylily plants. Finally, the model was validated with measured data. [Results and Discussions] The coefficient of determination (R2) between the measured and simulated outdoor surface solar radiation was 0.87, accompanied by a Root Mean Squared Error (RMSE) of 28.52 W/m2. For the simulated model of each organ of the daylily plant, the R2 of the measured against the predicted values ranged from 0.896 to 0.984, with an RMSE varying between 1.4 and 17.7 cm. The R2 of the average flower bud yield simulation was 0.880, accompanied by an RMSE of 0.5 g. The overall F-value spanned from 82.244 to 1 168.533, while the Sig. value was consistently below the 0.05 significance level, suggesting a robust fit and statistical significance for the aforementioned models. Subsequently, a thorough examination of the light interaction, temperature influences, and photosynthetic attributes of daylily leaves throughout their growth cycle was carried out. The findings revealed that leaf nutrition growth played a pivotal role in the early phase of daylily's growth, followed by the contribution of leaf and flower stem nutrition in the middle stage, and finally the growth of daylily flower buds, which is the crucial period for yield formation, in the later stages. Analyzing the photosynthetic traits of daylily leaves comprehensively, it was observed that the photosynthetic rate was relatively low in the early spring as the new leaves were initially emerging and reached a plateau during the summer. Considering real-world climate conditions, the actual net photosynthetic rate was marginally lower than the rate verified under optimal conditions, with the simulated net assimilation rate typically ranging from 2 to 4 μmol CO2/(m2·s). [Conclusions] The three-dimensional dynamic growth model of daylily plants proposed in this study can faithfully articulate the growth laws and morphological traits of daylily plants across the three primary growth stages. This model not only illustrates the three-dimensional dynamic growth of daylily plants but also effectively mimics the yield data of daylily flower buds. The simulation outcomes concur with actual conditions, demonstrating a high level of reliability.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Identification Method of Wheat Grain Phenotype Based on Deep Learning of ImCascade R-CNN
    PAN Weiting, SUN Mengli, YUN Yan, LIU Ping
    Smart Agriculture    2023, 5 (3): 110-120.   DOI: 10.12133/j.smartag.SA202304006
    Abstract596)   HTML69)    PDF(pc) (1664KB)(3072)       Save

    [Objective] Wheat serves as the primary source of dietary carbohydrates for the human population, supplying 20% of the required caloric intake. Currently, the primary objective of wheat breeding is to develop wheat varieties that exhibit both high quality and high yield, ensuring an overall increase in wheat production. Additionally, the consideration of phenotype parameters, such as grain length and width, holds significant importance in the introduction, screening, and evaluation of germplasm resources. Notably, a noteworthy positive association has been observed between grain size, grain shape, and grain weight. Simultaneously, within the scope of wheat breeding, the occurrence of inadequate harvest and storage practices can readily result in damage to wheat grains, consequently leading to a direct reduction in both emergence rate and yield. In essence, the integrity of wheat grains directly influences the wheat breeding process. Nevertheless, distinguishing between intact and damaged grains remains challenging due to the minimal disparities in certain characteristics, thereby impeding the accurate identification of damaged wheat grains through manual means. Consequently, this study aims to address this issue by focusing on the detection of wheat kernel integrity and completing the attainment of grain phenotype parameters. [Methods] This study presented an enhanced approach for addressing the challenges of low detection accuracy, unclear segmentation of wheat grain contour, and missing detection. The proposed strategy involves utilizing the Cascade Mask R-CNN model and replacing the backbone network with ResNeXt to mitigate gradient dispersion and minimize the model's parameter count. Furthermore, the inclusion of Mish as an activation function enhanced the efficiency and versatility of the detection model. Additionally, a multilayer convolutional structure was introduced in the detector to thoroughly investigate the latent features of wheat grains. The Soft-NMS algorithm was employed to identify the candidate frame and achieve accurate segmentation of the wheat kernel adhesion region. Additionally, the ImCascade R-CNN model was developed. Simultaneously, to address the issue of low accuracy in obtaining grain contour parameters due to disordered grain arrangement, a grain contour-based algorithm for parameter acquisition was devised. Wheat grain could be approximated as an oval shape, and the grain edge contour could be obtained according to the mask, the distance between the farthest points could be iteratively obtained as the grain length, and the grain width could be obtained according to the area. Ultimately, a method for wheat kernel phenotype identification was put forth. The ImCascade R-CNN model was utilized to analyze wheat kernel images, extracting essential features and determining the integrity of the kernels through classification and boundary box regression branches. The mask generation branch was employed to generate a mask map for individual wheat grains, enabling segmentation of the grain contours. Subsequently, the number of grains in the image was determined, and the length and width parameters of the entire wheat grain were computed. [Results and Discussions] In the experiment on wheat kernel phenotype recognition, a comparison and improvement were conducted on the identification results of the Cascade Mask R-CNN model and the ImCascade R-CNN model across various modules. Additionally, the efficacy of the model modification scheme was verified. The comparison of results between the Cascade Mask R-CNN model and the ImCascade R-CNN model served to validate the proposed model's ability to significantly decrease the missed detection rate. The effectiveness and advantages of the ImCascade R-CNN model were verified by comparing its loss value, P-R value, and mAP_50 value with those of the Cascade Mask R-CNN model. In the context of wheat grain identification and segmentation, the detection results of the ImCascade R-CNN model were compared to those of the Cascade Mask R-CNN and Deeplabv3+ models. The comparison confirmed that the ImCascade R-CNN model exhibited superior performance in identifying and locating wheat grains, accurately segmenting wheat grain contours, and achieving an average accuracy of 90.2% in detecting wheat grain integrity. These findings serve as a foundation for obtaining kernel contour parameters. The grain length and grain width exhibited average error rates of 2.15% and 3.74%, respectively, while the standard error of the aspect ratio was 0.15. The statistical analysis and fitting of the grain length and width, as obtained through the proposed wheat grain shape identification method, yielded determination coefficients of 0.9351 and 0.8217, respectively. These coefficients demonstrated a strong agreement with the manually measured values, indicating that the method is capable of meeting the demands of wheat seed testing and providing precise data support for wheat breeding. [Conclusions] The findings of this study can be utilized for the rapid and precise detection of wheat grain integrity and the acquisition of comprehensive grain contour data. In contrast to current wheat kernel recognition technology, this research capitalizes on enhanced grain contour segmentation to furnish data support for the acquisition of wheat kernel contour parameters. Additionally, the refined contour parameter acquisition algorithm effectively mitigates the impact of disordered wheat kernel arrangement, resulting in more accurate parameter data compared to existing kernel appearance detectors available in the market, providing data support for wheat breeding and accelerating the cultivation of high-quality and high-yield wheat varieties.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Measurement of Mongolian Horse Body Based on Improved YOLOv8n-pose and 3D Point Cloud Analysis
    LI Minghuang, SU Lide, ZHANG Yong, ZONG Zheying, ZHANG Shun
    Smart Agriculture    2024, 6 (4): 91-102.   DOI: 10.12133/j.smartag.SA202312027
    Abstract1236)   HTML59)    PDF(pc) (2477KB)(3022)       Save

    [Objective] There exist a high genetic correlation among various morphological characteristics of Mongolian horses. Utilizing advanced technology to obtain body structure parameters related to athletic performance could provide data support for breeding institutions to develop scientific breeding plans and establish the groundwork for further improvement of Mongolian horse breeds. However, traditional manual measurement methods are time-consuming, labor-intensive, and may cause certain stress responses in horses. Therefore, ensuring precise and effective measurement of Mongolian horse body dimensions is crucial for formulating early breeding plans. [Method] Video images of 50 adult Mongolian horses in the suitable breeding stage at the Inner Mongolia Agricultural University Horse Breeding Technical Center was first collected. Fifty images per horse were captured to construct the training and validation sets, resulting in a total of 2 500 high-definition RGB images of Mongolian horses, with an equal ratio of images depicting horses in motion and at rest. To ensure the model's robustness and considering issues such as angles, lighting, and image blurring during actual image capture, a series of enhancement algorithms were applied to the original dataset, expanding the Mongolian horse image dataset to 4 000 images. The YOLOv8n-pose was employed as the foundational keypoint detection model. Through the design of the C2f_DCN module, deformable convolution (DCNV2) was integrated into the C2f module of the Backbone network to enhance the model's adaptability to different horse poses in real-world scenes. Besides, an SA attention module was added to the Neck network to improve the model's focus on critical features. The original loss function was replaced with SCYLLA-IoU (SIoU) to prioritize major image regions, and a cosine annealing method was employed to dynamically adjust the learning rate during model training. The improved model was named DSS-YOLO (DCNv2-SA-SIoU-YOLO) network model. Additionally, a test set comprising 30 RGB-D images of mature Mongolian horses was selected for constructing body dimension measurement tasks. DSS-YOLO was used for keypoint detection of body dimensions. The 2D keypoint coordinates from RGB images were fused with corresponding depth values from depth images to obtain 3D keypoint coordinates, and Mongolian horse's point cloud information was transformed. Point cloud processing and analysis were performed using pass-through filtering, random sample consensus (RANSAC) shape fitting, statistical outlier filtering, and principal component analysis (PCA) coordinate system correction. Finally, body height, body oblique length, croup height, chest circumference, and croup circumference were automatically computed based on keypoint spatial coordinates. [Results and Discussion] The proposed DSS-YOLO model exhibited parameter and computational costs of 3.48 M and 9.1 G, respectively, with an average accuracy mAP0.5:0.95 reaching 92.5%, and a dDSS of 7.2 pixels. Compared to Hourglass, HRNet, and SimCC, mAP0.5:0.95 increased by 3.6%, 2.8%, and 1.6%, respectively. By relying on keypoint coordinates for automatic calculation of body dimensions and suggesting the use of a mobile least squares curve fitting method to complete the horse's hip point cloud, experiments involving 30 Mongolian horses showed a mean average error (MAE) of 3.77 cm and mean relative error (MRE) of 2.29% in automatic measurements. [Conclusions] The results of this study showed that DSS-YOLO model combined with three-dimensional point cloud processing methods can achieve automatic measurement of Mongolian horse body dimensions with high accuracy. The proposed measurement method can also be extended to different breeds of horses, providing technical support for horse breeding plans and possessing practical application value.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Remote Sensing Identification Method of Cultivated Land at Hill County of Sichuan Basin Based on Deep Learning
    LI Hao, DU Yuqiu, XIAO Xingzhu, CHEN Yanxi
    Smart Agriculture    2024, 6 (3): 34-45.   DOI: 10.12133/j.smartag.SA202308002
    Abstract432)   HTML50)    PDF(pc) (2649KB)(3006)       Save

    [Objective] To fully utilize and protect farmland and lay a solid foundation for the sustainable use of land, it is particularly important to obtain real-time and precise information regarding farmland area, distribution, and other factors. Leveraging remote sensing technology to obtain farmland data can meet the requirements of large-scale coverage and timeliness. However, the current research and application of deep learning methods in remote sensing for cultivated land identification still requires further improvement in terms of depth and accuracy. The objective of this study is to investigate the potential application of deep learning methods in remote sensing for identifying cultivated land in the hilly areas of Southwest China, to provide insights for enhancing agricultural land utilization and regulation, and for harmonizing the relationship between cultivated land and the economy and ecology. [Methods] Santai county, Mianyang city, Sichuan province, China (30°42'34"~31°26'35"N, 104°43'04"~105°18'13"E) was selected as the study area. High-resolution imagery from two scenes captured by the Gaofen-6 (GF-6) satellite served as the primary image data source. Additionally, 30-meter resolution DEM data from the United States National Aeronautics and Space Administration (NASA) in 2020 was utilized. A land cover data product, SinoLC-1, was also incorporated for comparative evaluation of the accuracy of various extraction methods' results. Four deep learning models, namely Unet, PSPNet, DeeplabV3+, and Unet++, were utilized for remote sensing land identification research in cultivated areas. The study also involved analyzing the identification accuracy of cultivated land in high-resolution satellite images by combining the results of the random forest (RF) algorithm along with the deep learning models. A validation dataset was constructed by randomly generating 1 000 vector validation points within the research area. Concurrently, Google Earth satellite images with a resolution of 0.3 m were used for manual visual interpretation to determine the land cover type of the pixels where the validation points are located. The identification results of each model were compared using a confusion matrix to compute five accuracy evaluation metrics: Overall accuracy (OA), intersection over union (IoU), mean intersection over union (MIoU), F1-Score, and Kappa Coefficient to assess the cultivated land identification accuracy of different models and data products. [Results and Discussions] The deep learning models displayed significant advances in accuracy evaluation metrics, surpassing the performance of traditional machine learning approaches like RF and the latest land cover product, SinoLC-1 Landcover. Among the models assessed, the UNet++ model performed the best, its F1-Score, IoU, MIoU, OA, and Kappa coefficient values were 0.92, 85.93%, 81.93%, 90.60%, and 0.80, respectively. DeeplabV3+, UNet, and PSPNet methods followed suit. These performance metrics underscored the superior accuracy of the UNet++ model in precisely identifying and segmenting cultivated land, with a remarkable increase in accuracy of nearly 20% than machine learning methods and 50% for land cover products. Four typical areas of town, water body, forest land and contiguous cultivated land were selected to visually compare the results of cultivated land identification results. It could be observed that the deep learning models generally exhibited consistent distribution patterns with the satellite imageries, accurately delineating the boundaries of cultivated land and demonstrating overall satisfactory performance. However, due to the complex features in remote sensing images, the deep learning models still encountered certain challenges of omission and misclassification in extracting cultivated land. Among them, the UNet++ model showed the closest overall extraction results to the ground truth and exhibited advantages in terms of completeness of cultivated land extraction, discrimination between cultivated land and other land classes, and boundary extraction compared to other models. Using the UNet++ model with the highest recognition accuracy, two types of images constructed with different features—solely spectral features and spectral combined with terrain features—were utilized for cultivated land extraction. Based on the three metrics of IoU, OA, and Kappa, the model incorporating both spectral and terrain features showed improvements of 0.98%, 1.10%, and 0.01% compared to the model using only spectral features. This indicated that fusing spectral and terrain features can achieve information complementarity, further enhancing the identification effectiveness of cultivated land. [Conclusions] This study focuses on the practicality and reliability of automatic cultivated land extraction using four different deep learning models, based on high-resolution satellite imagery from the GF-6 in Santai county in China. Based on the cultivated land extraction results in Santai county and the differences in network structures among the four deep learning models, it was found that the UNet++ model, based on UNet, can effectively improve the accuracy of cultivated land extraction by introducing the mechanism of skip connections. Overall, this study demonstrates the effectiveness and practical value of deep learning methods in obtaining accurate farmland information from high-resolution remote sensing imagery.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Real-time Detection Algorithm of Expanded Feed Image on the Water Surface Based on Improved YOLOv11
    ZHOU Xiushan, WEN Luting, JIE Baifei, ZHENG Haifeng, WU Qiqi, LI Kene, LIANG Junneng, LI Yijian, WEN Jiayan, JIANG Linyuan
    Smart Agriculture    2024, 6 (6): 155-167.   DOI: 10.12133/j.smartag.SA202408014
    Abstract641)   HTML77)    PDF(pc) (1858KB)(2662)       Save

    [Objective] During the feeding process of fish populations in aquaculture, the video image characteristics of floating extruded feed on the water surface undergo continuous variations due to a myriad of environmental factors and fish behaviors. These variations pose significant challenges to the accurate detection of feed particles, which is crucial for effective feeding management. To address these challenges and enhance the detection of floating extruded feed particles on the water surface, ,thereby providing precise decision support for intelligent feeding in intensive aquaculture modes, the YOLOv11-AP2S model, an advanced detection model was proposed. [Methods] The YOLOv11-AP2S model enhanced the YOLOv11 algorithm by incorporating a series of improvements to its backbone network, neck, and head components. Specifically, an attention for fine-grained categorization (AFGC) mechanism was introduced after the 10th layer C2PSA of the backbone network. This mechanism aimed to boost the model's capability to capture fine-grained features, which were essential for accurately identifying feed particles in complex environments with low contrast and overlapping objects. Furthermore, the C3k2 module was replaced with the VoV-GSCSP module, which incorporated more sophisticated feature extraction and fusion mechanisms. This replacement further enhanced the network's ability to extract relevant features and improve detection accuracy. To improve the model's detection of small targets, a P2 layer was introduced. However, adding a P2 layer may increase computational complexity and resource consumption, so the overall performance and resource consumption of the model must be carefully balanced. To maintain the model's real-time performance while improving detection accuracy, a lightweight VoV-GSCSP module was utilized for feature fusion at the P2 layer. This approach enabled the YOLOv11-AP2S model to achieve high detection accuracy without sacrificing detection speed or model lightweights, making it suitable for real-time applications in aquaculture. [Results and Discussions] The ablation experimental results demonstrated the superiority of the YOLOv11-AP2S model over the original YOLOv11 network. Specifically, the YOLOv11-AP2S model achieved a precision ( P) and recall ( R) of 78.70%. The mean average precision (mAP50) at an intersection over union (IoU) threshold of 0.5 was as high as 80.00%, and the F1-Score had also reached 79.00%. These metrics represented significant improvements of 6.7%, 9.0%, 9.4% (for precision, as previously mentioned), and 8.0%, respectively, over the original YOLOv11 network. These improvements showed the effectiveness of the YOLOv11-AP2S model in detecting floating extruded feed particles in complex environments. When compared to other YOLO models, the YOLOv11-AP2S model exhibits clear advantages in detecting floating extruded feed images on a self-made dataset. Notably, under the same number of iterations, the YOLOv11-AP2S model achieved higher mAP50 values and lower losses, demonstrating its superiority in detection performance. This indicated that the YOLOv11-AP2S model strikes a good balance between learning speed and network performance, enabling it to efficiently and accurately detect images of floating extruded feed on the water surface. Furthermore, the YOLOv11-AP2S model's ability to handle complex detection scenarios, such as overlapping and adhesion of feed particles and occlusion by bubbles, was noteworthy. These capabilities were crucial for accurate detection in practical aquaculture environments, where such challenges were common and can significantly impair the performance of traditional detection systems. The improvements in detection accuracy and efficiency made the YOLOv11-AP2S model a valuable tool for intelligent feeding systems in aquaculture, as it could provide more reliable and timely information on fish feeding behavior. Additionally, the introduction of the P2 layer and the use of the lightweight VoV-GSCSP module for feature fusion at this layer contributed to the model's overall performance. These enhancements enabled the model to maintain high detection accuracy while keeping computational costs and resource consumption within manageable limits. This was particularly important for real-time applications in aquaculture, where both accuracy and efficiency were critical for effective feeding management. [Conclusions] The successful application of the YOLOv11-AP2S model in detecting floating extruded feed particles demonstrates its potential to intelligent feeding systems in aquaculture. By providing accurate and timely information on fish feeding behavior, the model can help optimize feeding strategies, reduce feed waste, and improve the overall efficiency and profitability of aquaculture operations. Furthermore, the model's ability to handle complex detection scenarios and maintain high detection accuracy while keeping computational costs within manageable limits makes it a practical and valuable tool for real-time applications in aquaculture. Therefore, the YOLOv11-AP2S model holds promise for wide application in intelligent aquaculture management, contributing to the sustainability and growth of the aquaculture industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospects of Key Technologies for Rice Smart Unmanned Farms
    YU Fenghua, XU Tongyu, GUO Zhonghui, BAI Juchi, XIANG Shuang, GUO Sien, JIN Zhongyu, LI Shilong, WANG Shikuan, LIU Meihan, HUI Yinxuan
    Smart Agriculture    2024, 6 (6): 1-22.   DOI: 10.12133/j.smartag.SA202410018
    Abstract989)   HTML163)    PDF(pc) (3047KB)(2626)       Save

    [Significance] Rice smart unmanned farm is the core component of smart agriculture, and it is a key path to realize the modernization of rice production and promote the high-quality development of agriculture. Leveraging advanced information technologies such as the Internet of Things (IoT) and artificial intelligence (AI), these farms enable deep integration of data-driven decision making and intelligent machines. This integration creates an unmanned production system that covers the entire process from planting and managing rice crops to harvesting, greatly improving the efficiency and precision of rice cultivation. [Progress] This paper systematically sorted out the key technologies of rice smart unmanned farms in the three main links of pre-production, production and post-production, and the key technologies of pre-production mainly include the construction of high-standard farmland, unmanned nursery, land leveling, and soil nutrient testing. The construction of high-standard farmland is the foundation of the physical environment of the smart unmanned farms of rice, which provides perfect operating environment for the operation of modernized smart farm machinery through the reasonable layout of the field roads, good drainage and irrigation systems, and the scientific planting structure. Agricultural machine operation provides a perfect operating environment. The technical level of unmanned nursery directly determines the quality of rice cultivation and harvesting in the later stage, and a variety of rice seeding machines and nursery plate setting machines have been put into use. Land leveling technology can improve the growing environment of rice and increase the land utilization rate, and the current land leveling technology through digital sensing and path planning technology, which improves the operational efficiency and reduces the production cost at the same time. Soil nutrient detection technology is mainly detected by electrochemical analysis and spectral analysis, but both methods have their advantages and disadvantages, how to integrate the two methods to achieve an all-round detection of soil nutrient content is the main direction of future research. The key technologies in production mainly include rice dry direct seeding, automated transplanting, precise variable fertilization, intelligent irrigation, field weed management, and disease diagnosis. Among them, the rice dry direct seeding technology requires the planter to have high precision and stability to ensure reasonable seeding depth and density. Automated rice transplanting technology mainly includes three ways: root washing seedling machine transplanting, blanket seedling machine transplanting, and potting blanket seedling machine transplanting; at present, the incidence of problems in the automated transplanting process should be further reduced, and the quality and efficiency of rice machine transplanting should be improved. Precision variable fertilization technology is mainly composed of three key technologies: information perception, prescription decision-making and precise operation, but there are still fewer cases of unmanned farms combining the three technologies, and in the future, the main research should be on the method of constructing the whole process operation system of variable fertilization. The smart irrigation system is based on the water demand of the whole life cycle of rice to realize adaptive irrigation control, and the current smart irrigation technology can automatically adjust the irrigation strategy through real-time monitoring of soil, climate and crop growth conditions to further improve irrigation efficiency and agricultural production benefits. The field weed management and disease diagnosis technology mainly recognizes rice weeds as well as diseases through deep learning and other methods, and combines them with precision application technology for prevention and intervention. Post-production key technologies mainly include rice yield estimation, unmanned harvesting, rice storage and processing quality testing. Rice yield estimation technology is mainly used to predict yield by combining multi-source data and algorithms, but there are still problems such as the difficulty of integrating multi-source data, which requires further research. In terms of unmanned aircraft harvesting technology, China's rice combine harvester market has tended to stabilize, and the safety of the harvester's autopilot should be further improved in the future. Rice storage and processing quality detection technology mainly utilizes spectral technology and machine vision technology to detect spectra and images, and future research can combine deep learning and multimodal fusion technology to improve the machine vision system's ability and adaptability to recognize the appearance characteristics of rice. [Conclusions and Prospects] This paper reviews the researches of the construction of intelligent unmanned rice farms at home and abroad in recent years, summarizes the main difficulties faced by the key technologies of unmanned farms in practical applications, analyzes the challenges encountered in the construction of smart unmanned farms, summarizes the roles and responsibilities of the government, enterprises, scientific research institutions, cooperatives and other subjects in promoting the construction of intelligent unmanned rice farms, and puts forward relevant suggestions. It provides certain support and development ideas for the construction of intelligent unmanned rice farms in China.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    The Path of Smart Agricultural Technology Innovation Leading Development of Agricultural New Quality Productivity
    CAO Bingxue, LI Hongfei, ZHAO Chunjiang, LI Jin
    Smart Agriculture    2024, 6 (4): 116-127.   DOI: 10.12133/j.smartag.SA202405004
    Abstract1322)   HTML224)    PDF(pc) (1102KB)(2532)       Save

    [Significance] Building the agricultural new quality productivity is of great significance. It is the advanced quality productivity which realizes the transformation, upgrading, and deep integration of substantive, penetrating, operational, and media factors, and has outstanding characteristics such as intelligence, greenness, integration, and organization. As a new technology revolution in the field of agriculture, smart agricultural technology transforms agricultural production mode by integrating agricultural biotechnology, agricultural information technology, and smart agricultural machinery and equipment, with information and knowledge as important core elements. The inherent characteristics of "high-tech, high-efficiency, high-quality, and sustainable" in agricultural new quality productivity are fully reflected in the practice of smart agricultural technology innovation. And it has become an important core and engine for promoting the agricultural new quality productivity. [Progress] Through literature review and theoretical analysis, this article conducts a systematic study on the practical foundation, internal logic, and problem challenges of smart agricultural technology innovation leading the development of agricultural new quality productivity. The conclusions show that: (1) At present, the global innovation capability of smart agriculture technology is constantly enhancing, and significant technology breakthroughs have been made in fields such as smart breeding, agricultural information perception, agricultural big data and artificial intelligence, smart agricultural machinery and equipment, providing practical foundation support for leading the development of agricultural new quality productivity. Among them, the smart breeding of 'Phenotype+Genotype+Environmental type' has entered the fast lane, the technology system for sensing agricultural sky, air, and land information is gradually maturing, the research and exploration on agricultural big data and intelligent decision-making technology continue to advance, and the creation of smart agricultural machinery and equipment for different fields has achieved fruitful results; (2) Smart agricultural technology innovation provides basic resources for the development of agricultural new quality productivity through empowering agricultural factor innovation, provides sustainable driving force for the development of agricultural new quality productivity through empowering agricultural technology innovation, provides practical paradigms for the development of agricultural new quality productivity through empowering agricultural scenario innovation, provides intellectual support for the development of agricultural new quality productivity through empowering agricultural entity innovation, and provides important guidelines for the development of agricultural new quality productivity through empowering agricultural value innovation; (3) Compared to the development requirements of agricultural new quality productivity in China and the advanced level of international smart agriculture technology, China's smart agriculture technology innovation is generally in the initial stage of multi-point breakthroughs, system integration, and commercial application. It still faces major challenges such as an incomplete policy system for technology innovation, key technologies with bottlenecks, blockages and breakpoints, difficulties in the transformation and implementation of technology achievements, and incomplete support systems for technology innovation. [Conclusions and Prospects] Regarding the issue of technology innovation in smart agriculture, this article proposes the 'Four Highs' path of smart agriculture technology innovation to fill the gaps in smart agriculture technology innovation and accelerate the formation of agricultural new quality productivity in China. The "Four Highs" path specifically includes the construction of high-energy smart agricultural technology innovation platforms, the breakthroughs in high-precision and cutting-edge smart agricultural technology products, the creation of high-level smart agricultural application scenarios, and the cultivation of high-level smart agricultural innovation talents. Finally, this article proposes four strategic suggestions such as deepening the understanding of smart agriculture technology innovation and agricultural new quality productivity, optimizing the supply of smart agriculture technology innovation policies, building a national smart agriculture innovation development pilot zone, and improving the smart agriculture technology innovation ecosystem.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Daylily Grading and Detection Model Based on Improved YOLOv10
    JIN Xuemeng, LIANG Xiyin, DENG Pengfei
    Smart Agriculture    2024, 6 (5): 108-118.   DOI: 10.12133/j.smartag.SA202407022
    Abstract507)   HTML68)    PDF(pc) (1532KB)(2498)       Save

    [Objective] In the agricultural production, accurately classifying dried daylily grades is a critical task with significant economic implications. However, current target detection models face challenges such as inadequate accuracy and excessive parameters when applied to dried daylily grading, limiting their practical application and widespread use in real-world settings. To address these issues, an innovative lightweight YOLOv10-AD network model was proposed. The model aims to enhance detection accuracy by optimizing the network structure and loss functions while reducing parameters and computational costs, making it more suitable for deployment in resource-constrained agricultural production environments. [Methods] The dried daylilies selected from the Qingyang region of Gansu province as the research subject. A large number of images of dried daylilies, categorized into three grades superior, medium, and inferior, were collected using mobile phones under varying lighting conditions and backgrounds. The images were carefully annotated and augmented to build a comprehensive dataset for dried daylily grade classification. YOLOv10 was chosen as the base network, and a newly designed backbone network called AKVanillaNet was introduced. AKVanillaNet combines AKConv (adaptive kernel convolution) with VanillaNet's deep learning and shallow inference mechanisms. The second convolutional layer in VanillaNet was replaced with AKConv, and AKConv was merged with standard convolution layers at the end of the training phase to optimize the model for capturing the unique shape characteristics of dried daylilies. This innovative design not only improved detection accuracy but also significantly reduced the number of parameters and computational costs. Additionally, the DysnakeConv module was integrated into the C2f structure, replacing the Bottleneck layer with a Bottleneck-DS layer to form the new C2f-DysnakeConv module. This module enhanced the model's sensitivity to the shapes and boundaries of targets, allowing the neural network to better capture the shape information of irregular objects like dried daylilies, further improving the model's feature extraction capability. The Powerful-IOU (PIOU) loss function was also employed, which introduced a target-size-adaptive penalty factor and a gradient adjustment function. This design guided the anchor box regression along a more direct path, helping the model better fit the data and improve overall performance. [Results and Discussions] The testing results on the dried daylily grade classification dataset demonstrated that the YOLOv10-AD model achieved a mean average precision (mAP) of 85.7%. The model's parameters, computational volume, and size were 2.45 M, 6.2 GFLOPs, and 5.0 M, respectively, with a frame rate of 156 FPS. Compared to the benchmark model, YOLOv10-AD improved mAP by 5.7% and FPS by 25.8%, while reducing the number of parameters, computational volume, and model size by 9.3%, 24.4%, and 9.1%, respectively. These results indicated that YOLOv10-AD not only improved detection accuracy but also reduced the model's complexity, making it easier to deploy in real-world production environments. Furthermore, YOLOv10-AD outperformed larger models in the same series, such as YOLOv10s and YOLOv10m. Specifically, the weight, parameters, and computational volume of YOLOv10-AD were only 31.6%, 30.5%, and 25.3% of those in YOLOv10s, and 15.7%, 14.8%, and 9.8% of YOLOv10m. Despite using fewer resources, YOLOv10-AD achieved a mAP increase of 2.4% over YOLOv10s and 1.9% over YOLOv10m. These findings confirm that YOLOv10-AD maintains high detection accuracy while requiring significantly fewer resources, making it more suitable for agricultural production environments where computational capacity may be limited. The study also examined the performance of YOLOv10-AD under different lighting conditions. The results showed that YOLOv10-AD achieved an average accuracy of 92.3% in brighter environments and 78.6% in darker environments. In comparison, the YOLOv10n model achieved 88.9% and 71.0% in the same conditions, representing improvements of 3.4% and 7.6%, respectively. These findings demonstrate that YOLOv10-AD has a distinct advantage in maintaining high accuracy and confidence in grading dried daylilies across varying lighting conditions. [Conclusions] The YOLOv10-AD network model proposed significantly reduces the number of parameters and computational costs without compromising detection accuracy. This model presents a valuable technical reference for intelligent classification of dried daylily grades in agricultural production environments, particularly where resources are constrained.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Advances, Problems and Challenges of Precise Estrus Perception and Intelligent Identification Technology for Cows
    ZHANG Zhiyong, CAO Shanshan, KONG Fantao, LIU Jifang, SUN Wei
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202305005
    Online available: 08 January 2025

    Wheat Lodging Types Detection Based on UAV Image Using Improved EfficientNetV2
    LONG Jianing, ZHANG Zhao, LIU Xiaohang, LI Yunxia, RUI Zhaoyu, YU Jiangfan, ZHANG Man, FLORES Paulo, HAN Zhexiong, HU Can, WANG Xufeng
    Smart Agriculture    2023, 5 (3): 62-74.   DOI: 10.12133/j.smartag.SA202308010
    Abstract598)   HTML52)    PDF(pc) (2022KB)(2447)       Save

    [Objective] Wheat, as one of the major global food crops, plays a key role in food production and food supply. Different influencing factors can lead to different types of wheat lodging, e.g., root lodging may be due to improper use of fertilizers. While stem lodging is mostly due to harsh environments, different types of wheat lodging can have different impacts on yield and quality. The aim of this study was to categorize the types of wheat lodging by unmanned aerial vehicle (UAV) image detection and to investigate the effect of UAV flight altitude on the classification performance. [Methods] Three UAV flight altitudes (15, 45, and 91 m) were set to acquire images of wheat test fields. The main research methods contained three parts: an automatic segmentation algorithm, wheat classification model selection, and an improved classification model based on EfficientNetV2-C. In the first part, the automatic segmentation algorithm was used to segment the UAV to acquire the wheat test field at three different heights and made it into the training dataset needed for the classification model. The main steps were first to preprocess the original wheat test field images acquired by the UAV through scaling, skew correction, and other methods to save computation time and improve segmentation accuracy. Subsequently, the pre-processed image information was analyzed, and the green part of the image was extracted using the super green algorithm, which was binarized and combined with the edge contour extraction algorithm to remove the redundant part of the image to extract the region of interest, so that the image was segmented for the first time. Finally, the idea of accumulating pixels to find sudden value added was used to find the segmentation coordinates of two different sizes of wheat test field in the image, and the region of interest of the wheat test field was segmented into a long rectangle and a short rectangle test field twice, so as to obtain the structural parameters of different sizes of wheat test field and then to generate the dataset of different heights. In the second part, four machine learning classification models of support vector machine (SVM), K nearest neighbor (KNN), decision tree (DT), and naive bayes (NB), and two deep learning classification models (ResNet101 and EfficientNetV2) were selected. Under the unimproved condition, six classification models were utilized to classify the images collected from three UAVs at different flight altitudes, respectively, and the optimal classification model was selected for improvement. In the third part, an improved model, EfficientNetV2-C, with EfficientNetV2 as the base model, was proposed to classify and recognized the lodging type of wheat in test field images. The main improvement points were attention mechanism improvement and loss function improvement. The attention mechanism was to replace the original model squeeze and excitation (SE) with coordinate attention (CA), which was able to embed the position information into the channel attention, aggregate the features along the width and height directions, respectively, during feature extraction, and capture the long-distance correlation in the width direction while retaining the long-distance correlation in the length direction, accurate location information, enhancing the feature extraction capability of the network in space. The loss function was replaced by class-balanced focal loss (CB-Focal Loss), which could assign different loss weights according to the number of valid samples in each class when targeting unbalanced datasets, effectively solving the impact of data imbalance on the classification accuracy of the model. [Results and Discussions] Four machine learning classification results: SVM average classification accuracy was 81.95%, DT average classification accuracy was 79.56%, KNN average classification accuracy was 59.32%, and NB average classification accuracy was 59.48%. The average classification accuracy of the two deep learning models, ResNet101 and EfficientNetV2, was 78.04%, and the average classification accuracy of ResNet101 was 81.61%. Comparing the above six classification models, the EfficientNetV2 classification model performed optimally at all heights. And the improved EfficientNetV2-C had an average accuracy of 90.59%, which was 8.98% higher compared to the average accuracy of EfficientNetV2. The SVM classification accuracies of UAVs at three flight altitudes of 15, 45, and 91 m were 81.33%, 83.57%, and 81.00%, respectively, in which the accuracy was the highest when the altitude was 45 m, and the classification results of the SVM model values were similar to each other, which indicated that the imbalance of the input data categories would not affect the model's classification effect, and the SVM classification model was able to solve the problem of high dimensionality of the data efficiently and had a good performance for small and medium-sized data sets. The SVM classification model could effectively solve the problem of the high dimensionality of data and had a better classification effect on small and medium-sized datasets. For the deep learning classification model, however, as the flight altitude increases from 15 to 91 m, the classification performance of the deep learning model decreased due to the loss of image feature information. Among them, the classification accuracy of ResNet101 decreased from 81.57% to 78.04%, the classification accuracy of EfficientNetV2 decreased from 84.40% to 81.61%, and the classification accuracy of EfficientNetV2-C decreased from 97.65% to 90.59%. The classification accuracy of EfficientNetV2-C at each of the three altitudes. The difference between the values of precision, recall, and F1-Score results of classification was small, which indicated that the improved model in this study could effectively solve the problems of unbalanced model classification results and poor classification effect caused by data imbalance. [Conclusions] The improved EfficientNetV2-C achieved high accuracy in wheat lodging type detection, which provides a new solution for wheat lodging early warning and crop management and is of great significance for improving wheat production efficiency and sustainable agricultural development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    GRA-WHO-TCN Combination Model for Forecasting Cold Chain Logistics Demand of Agricultural Products
    LIU Yan, JI Juncheng
    Smart Agriculture    2024, 6 (3): 148-158.   DOI: 10.12133/j.smartag.SA202310006
    Abstract521)   HTML39)    PDF(pc) (1280KB)(2409)       Save

    [Objective] As a critical component of agricultural product supply chain management, cold chain logistics demand prediction encounters challenges such as inadequate feature extraction, high nonlinearity of data, and the propensity for algorithms to become trapped in local optima during the digital transformation process. To address these issues and enhance the accuracy of demand prediction, achieve intelligent management of the agricultural product supply chain, a combined forecasting model that integrates grey relational analysis (GRA), the wild horse optimizer (WHO), and temporal convolutional networks (TCN) is proposed in this research. [Methods] Firstly, a cold chain logistics indicator system was established for the data of Zhejiang province, China, spanning the years 2000 to 2020. This system covered four key aspects: the economic scale of agricultural products, logistics transportation, digital technology, and agricultural product supply. Then, the GRA was applied to identify relevant indicators of cold chain logistics for agricultural products in Zhejiang province, with 17 indicators selected that had a correlation degree higher than 0.75. Sliding window technology, a problem-solving approach for data structures and algorithms, suitable for reducing the time complexity of data to a better level and improving the execution efficiency of algorithms, was used to partition the selected indicators. Secondly, the TCN model was employed to extract features of different scales by stacking multiple convolutional layers. Each layer utilized different-sized convolutional kernels to capture features within different time ranges. By utilizing the dilated convolutional module of TCN, temporal and spatial relationships within economic data were effectively mined, considering the temporal characteristics of socio-economic data and logistics information in the agricultural supply chain, and exploring the temporal and spatial features of economic data. Simultaneously, the WHO algorithm was applied to optimize five hyperparameters of the TCN model, including the number of TCN layers, the number of filters, residual blocks, Dense layers, and neurons within the Dense layer. Finally, the optimized GRA-WHO-TCN model was used to extract and analyze features from highly nonlinear multidimensional economic data, ultimately facilitating the prediction of cold chain logistics demand. [Results and Discussions] For comparative analysis of the superiority of the GRA-WHO-TCN model, the 17 selected indicators were input into long short-term memory (LSTM), TCN, WHO-LSTM, and WHO-TCN models. The parameters optimized by the WHO algorithm for the TCN model were set respectively: 2 TCN layer was, 2 residual blocks, 1 dense layer, 60 filters, and 16 neurons in the dense layer. The optimized GRA-WHO-TCN temporal model can effectively extract the temporal and spatial features of multidimensional data, fully explore the implicit relationships among indicator factors, and demonstrating good fitting effects. Compared to GRA-LSTM and GRA-TCN models, the GRA-TCN model exhibited superior performance, with a lower root mean square error of 37.34 and a higher correlation coefficient of 0.91, indicating the advantage of the TCN temporal model in handling complex nonlinear data. Furthermore, the GRA-WHO-LSTM and GRA-WHO-TCN models optimized by the WHO algorithm had improved prediction accuracy and stability compared to GRA-LSTM and GRA-TCN models, illustrating that the WHO algorithm effectively optimized model parameters to enhance the effectiveness of model fitting. When compared to the GRA-WHO-LSTM model, the GRA-WHO-TCN model displayed a lower root mean square error of 11.3 and an effective correlation coefficient of 0.95, predicting cold chain logistics demand quantities in Zhejiang province for the years 2016-2020 as 29.8, 30.46, 24.87, 26.45, and 27.99 million tons, with relative errors within 0.6%, achieving a high level of prediction accuracy. This achievement showcases a high level of prediction accuracy and underscores the utility of the GRA-WHO-TCN model in forecasting complex data scenarios. [Conclusions] The proposed GRA-WHO-TCN model demonstrated superior parameter optimization capabilities and predictive accuracy compared to the GRA-LSTM and GRA-TCN models. The predicted results align well with the development of cold chain logistics of agricultural products in Zhejiang province. This provides a scientific prediction foundation and practical reference value for the development of material flow and information flow in the agricultural supply chain under the digital economy context.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Apple Leaf Disease Detection Algorithm Based on Improved YOLOv8
    LUO Youlu, PAN Yonghao, XIA Shunxing, TAO Youzhi
    Smart Agriculture    2024, 6 (5): 128-138.   DOI: 10.12133/j.smartag.SA202406012
    Abstract658)   HTML97)    PDF(pc) (1702KB)(2246)       Save

    [Objective] As one of China's most important agricultural products, apples hold a significant position in cultivation area and yield. However, during the growth process, apples are prone to various diseases that not only affect the quality of the fruit but also significantly reduce the yield, impacting farmers' economic benefits and the stability of market supply. To reduce the incidence of apple diseases and increase fruit yield, developing efficient and fast apple leaf disease detection technology is of great significance. An improved YOLOv8 algorithm was proposed to identify the leaf diseases that occurred during the growth of apples. [Methods] YOLOv8n model was selected to detect various leaf diseases such as brown rot, rust, apple scab, and sooty blotch that apples might encounter during growth. SPD-Conv was introduced to replace the original convolutional layers to retain fine-grained information and reduce model parameters and computational costs, thereby improving the accuracy of disease detection. The multi-scale dilated attention (MSDA) attention mechanism was added at appropriate positions in the Neck layer to enhance the model's feature representation capability, which allowed the model to learn the receptive field dynamically and adaptively focus on the most representative regions and features in the image, thereby enhancing the ability to extract disease-related features. Finally, inspired by the RepVGG architecture, the original detection head was optimized to achieve a separation of detection and inference architecture, which not only accelerated the model's inference speed but also enhanced feature learning capability. Additionally, a dataset of apple leaf diseases containing the aforementioned diseases was constructed, and experiments were conducted. [Results and Discussions] Compared to the original model, the improved model showed significant improvements in various performance metrics. The mAP50 and mAP50:95 achieved 88.2% and 37.0% respectively, which were 2.7% and 1.3% higher than the original model. In terms of precision and recall, the improved model increased to 83.1% and 80.2%, respectively, representing an improvement of 0.9% and 1.1% over the original model. Additionally, the size of the improved model was only 7.8 MB, and the computational cost was reduced by 0.1 G FLOPs. The impact of the MSDA placement on model performance was analyzed by adding it at different positions in the Neck layer, and relevant experiments were designed to verify this. The experimental results showed that adding MSDA at the small target layer in the Neck layer achieved the best effect, not only improving model performance but also maintaining low computational cost and model size, providing important references for the optimization of the MSDA mechanism. To further verify the effectiveness of the improved model, various mainstream models such as YOLOv7-tiny, YOLOv9-c, RetinaNet, and Faster-RCNN were compared with the propoed model. The experimental results showed that the improved model outperformed these models by 1.4%, 1.3%, 7.8%, and 11.6% in mAP50, 2.8%, 0.2%, 3.4%, and 5.6% in mAP50:95. Moreover, the improved model showed significant advantages in terms of floating-point operations, model size, and parameter count, with a parameter count of only 3.7 MB, making it more suitable for deployment on hardware-constrained devices such as drones. In addition, to assess the model's generalization ability, a stratified sampling method was used, selecting 20% of the images from the dataset as the test set. The results showed that the improved model could maintain a high detection accuracy in complex and variable scenes, with mAP50 and mAP50:95 increasing by 1.7% and 1.2%, respectively, compared to the original model. Considering the differences in the number of samples for each disease in the dataset, a class balance experiment was also designed. Synthetic samples were generated using oversampling techniques to increase the number of minority-class samples. The experimental results showed that the class-balanced dataset significantly improved the model's detection performance, with overall accuracy increasing from 83.1% to 85.8%, recall from 80.2% to 83.6%, mAP50 from 88.2% to 88.9%, and mAP50:95 from 37.0% to 39.4%. The class-balanced dataset significantly enhanced the model's performance in detecting minority diseases, thereby improving the overall performance of the model. [Conclusions] The improved model demonstrated significant advantages in apple leaf disease detection. By introducing SPD-Conv and MSDA attention mechanisms, the model achieved noticeable improvements in both precision and recall while effectively reducing computational costs, leading to more efficient detection capabilities. The improved model could provide continuous health monitoring throughout the apple growth process and offer robust data support for farmers' scientific decision-making before fruit harvesting.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Path Tracking Control Algorithm of Tractor-Implement
    LIU Zhiyong, WEN Changkai, XIAO Yuejin, FU Weiqiang, WANG Hao, MENG Zhijun
    Smart Agriculture    2023, 5 (4): 58-67.   DOI: 10.12133/j.smartag.SA202308012
    Abstract400)   HTML25)    PDF(pc) (1386KB)(2161)       Save

    [Objective] The usual agricultural machinery navigation focuses on the tracking accuracy of the tractor, while the tracking effect of the trailed implement in the trailed agricultural vehicle is the core of the work quality. The connection mode of the tractor and the implement is non-rigid, and the implement can rotate around the hinge joint. In path tracking, this non-rigid structure, leads to the phenomenon of non-overlapping trajectories of the tractor and the implement, reduce the path tracking accuracy. In addition, problems such as large hysteresis and poor anti-interference ability are also very obvious. In order to solve the above problems, a tractor-implement path tracking control method based on variable structure sliding mode control was proposed, taking the tractor front wheel angle as the control variable and the trailed implement as the control target. [Methods] Firstly, the linear deviation model was established. Based on the structural relationship between the tractor and the trailed agricultural implements, the overall kinematics model of the vehicle was established by considering the four degrees of freedom of the vehicle: transverse, longitudinal, heading and articulation angle, ignoring the lateral force of the vehicle and the slip in the forward process. The geometric relationship between the vehicle and the reference path was integrated to establish the linear deviation model of vehicle-road based on the vehicle kinematic model and an approximate linearization method. Then, the control algorithm was designed. The switching function was designed considering three evaluation indexes: lateral deviation, course deviation and hinged angle deviation. The exponential reaching law was used as the reaching mode, the saturation function was used instead of the sign function to reduce the control variable jitter, and the convergence of the control law was verified by combining the Lyapunov function. The system was three-dimensional, in order to improve the dynamic response and steady-state characteristics of the system, the two conjugate dominant poles of the system were assigned within the required range, and the third point was kept away from the two dominant poles to reduce the interference on the system performance. The coefficient matrix of the switching function was solved based on the Ackermann formula, then the calculation formula of the tractor front wheel angle was obtained, and the whole control algorithm was designed. Finally, the path tracking control simulation experiment was carried out. The sliding mode controller was built in the MATLAB/Simulink environment, the controller was composed of the deviation calculation module and the control output calculation module. The tractor-implement model in Carsim software was selected with the front car as a tractor and the rear car as the single-axle implement, and tracking control simulation tests of different reference paths were conducted in the MATLAB/Carsim co-simulation environment. [Results and Discussions] Based on the co-simulation environment, the tracking simulation experiments of three reference paths were carried out. When tracking the double lane change path, the lateral deviation and heading deviation of the agricultural implement converged to 0 m and 0° after 8 s. When the reference heading changed, the lateral deviation and heading deviation were less than 0.1 m and less than 7°. When tracking the circular reference path, the lateral deviation of agricultural machinery tended to be stable after 7 s and was always less than 0.03 m, and the heading deviation of agricultural machinery tended to be stable after 7 s and remained at 0°. The simulation results of the double lane change path and the circular path showed that the controller could maintain good performance when tracking the constant curvature reference path. When tracking the reference path of the S-shaped curve, the tracking performance of the agricultural machinery on the section with constant curvature was the same as the previous two road conditions, and the maximum lateral deviation of the agricultural machinery at the curvature change was less than 0.05 m, the controller still maintained good tracking performance when tracking the variable curvature path. [Conclusions] The sliding mode variable structure controller designed in this study can effectively track the linear and circular reference paths, and still maintain a good tracking effect when tracking the variable curvature paths. Agricultural machinery can be on-line in a short time, which meets the requirements of speediness. In the tracking simulation test, the angle of the tractor front wheel and the articulated angle between the tractor and agricultural implement are kept in a small range, which meets the needs of actual production and reduces the possibility of safety accidents. In summary, the agricultural implement can effectively track the reference path and meet the requirements of precision, rapidity and safety. The model and method proposed in this study provide a reference for the automatic navigation of tractive agricultural implement. In future research, special attention will be paid to the tracking control effect of the control algorithm in the actual field operation and under the condition of large speed changes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospect of Quality Intelligent Control Technology in Facilities Environment of Characteristic Agricultural Products
    GUO Wei, WU Huarui, GUO Wang, GU Jingqiu, ZHU Huaji
    Smart Agriculture    2024, 6 (6): 44-62.   DOI: 10.12133/j.smartag.SA202411017
    Abstract511)   HTML64)    PDF(pc) (3153KB)(2145)       Save

    [Significance] In view of the lack of monitoring means of quality influence factors in the production process of characteristic agricultural products with in central and western regions of China, the weak ability of intelligent control, the unclear coupling relationship of quality control elements and the low degree of systematic application, the existing technologies described such as intelligent monitoring of facility environment, growth and nutrition intelligent control model, architecture of intelligent management and control platform and so on. Through the application of the Internet of Things, big data and the new generation of artificial intelligence technology, it provides technical support for the construction and application of intelligent process quality control system for the whole growth period of characteristic agricultural products. [Progress] The methods of environmental regulation and nutrition regulation are analyzed, including single parameters and combined control methods, such as light, temperature, humidity, CO2 concentration, fertilizer and water, etc. The multi-parameter coupling control method has the advantage of more comprehensive scene analysis. Based on the existing technology, a multi-factor coupling method of integrating growth state, agronomy, environment, input and agricultural work is put forward. This paper probes into the system architecture of the whole process service of quality control, the visual identification system of the growth process of agricultural products and the knowledge-driven agricultural technical service system, and introduces the technology of the team in the disease knowledge Q & A scene through multi-modal knowledge graph and large model technology. [Conclusions and Prospects] Based on the present situation of the production of characteristic facility agricultural products and the overall quality of farmers in the central and western regions of China, it is appropriate to transfer the whole technical system such as facility tomato, facility cucumber and so on. According to the varieties of characteristic agricultural products, cultivation models, quality control objectives to adapt to light, temperature, humidity and other parameters, as well as fertilizer, water, medicine and other input plans, a multi-factor coupling model suitable for a specific planting area is generated and long-term production verification and model correction are carried out. And popularize it in a wider area, making full use of the advantages of intelligent equipment and data elements will promote the realization of light simplification of production equipment, scene of intelligent technology, diversification of service models, on-line quality control, large-scale production of digital intelligence, and value of data elements, further cultivate facilities to produce new quality productivity.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grain Production Big Data Platform: Progress and Prospects
    YANG Guijun, ZHAO Chunjiang, YANG Xiaodong, YANG Hao, HU Haitang, LONG Huiling, QIU Zhengjun, LI Xian, JIANG Chongya, SUN Liang, CHEN Lei, ZHOU Qingbo, HAO Xingyao, GUO Wei, WANG Pei, GAO Meiling
    Smart Agriculture    2025, 7 (2): 1-12.   DOI: 10.12133/j.smartag.SA202409014
    Abstract534)   HTML81)    PDF(pc) (1811KB)(2105)       Save

    [Significance] The explosive development of agricultural big data has accelerated agricultural production into a new era of digitalization and intelligentialize. Agricultural big data is the core element to promote agricultural modernization and the foundation of intelligent agriculture. As a new productive forces, big data enhances the comprehensive intelligent management decision-making during the whole process of grain production. But it faces the problems such as the indistinct management mechanism of grain production big data resources, the lack of the full-chain decision-making algorithm system and big data platform for the whole process and full elements of grain production. [Progress] Grain production big data platform is a comprehensive service platform that uses modern information technologies such as big data, Internet of Things (IoT), remote sensing and cloud computing to provide intelligent decision-making support for the whole process of grain production based on intelligent algorithms for data collection, processing, analysis and monitoring related to grain production. In this paper, the progress and challenges in grain production big data, monitoring and decision-making algorithms are reviewed, as well as big data platforms in China and worldwide. With the development of the IoT and high-resolution multi-modal remote sensing technology, the massive agricultural big data generated by the "Space-Air-Ground" Integrated Agricultural Monitoring System, has laid an important foundation for smart agriculture and promoted the shift of smart agriculture from model-driven to data-driven. However, there are still some issues in field management decision-making, such as the requirements for high spatio-temporal resolution and timeliness of the information are difficult to meet, and the algorithm migration and localization methods based on big data need to be studied. In addition, the agricultural machinery operation and spatio-temporal scheduling algorithm based on remote sensing and IoT monitoring information to determine the appropriate operation time window and operation prescription, needs to be further developed, especially the cross-regional scheduling algorithm of agricultural machinery for summer harvest in China. Aiming to address the issues of non-bi-connected monitoring and decision-making algorithms in grain production, as well as the insufficient integration of agricultural machinery and information perception, a framework for the grain production big data intelligent platform based on digital twins is proposed. The platform leverages multi-source heterogeneous grain production big data and integrates a full-chain suit of standardized algorithms, including data acquisition, information extraction, knowledge map construction, intelligent decision-making, full-chain collaboration of agricultural machinery operations. It covers the typical application scenarios such as irrigation, fertilization, pests and disease management, emergency response to drought and flood disaster, all enabled by digital twins technology. [Conclusions and Prospects] The suggestions and trends for development of grain production big data platform are summarized in three aspects: (1) Creating an open, symbiotic grain production big data platform, with core characteristics such as open interface for crop and environmental sensors, maturity grading and a cloud-native packaging mechanism for core algorithms, highly efficient response to data and decision services; (2) Focusing on the typical application scenarios of grain production, take the exploration of technology integration and bi-directional connectivity as the base, and the intelligent service as the soul of the development path for the big data platform research; (3) The data-algorithm-service self-organizing regulation mechanism, the integration of decision-making information with the intelligent equipment operation, and the standardized, compatible and open service capabilities, can form the new quality productivity to ensure food safety, and green efficiency grain production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Fast Extracting Method for Strawberry Leaf Age and Canopy Width Based on Instance Segmentation Technology
    FAN Jiangchuan, WANG Yuanqiao, GOU Wenbo, CAI Shuangze, GUO Xinyu, ZHAO Chunjiang
    Smart Agriculture    2024, 6 (2): 95-106.   DOI: 10.12133/j.smartag.SA202310014
    Abstract877)   HTML48)    PDF(pc) (1903KB)(2098)       Save

    [Objective] There's a growing demand among plant cultivators and breeders for efficient methods to acquire plant phenotypic traits at high throughput, facilitating the establishment of mappings from phenotypes to genotypes. By integrating mobile phenotyping platforms with improved instance segmentation techniques, researchers have achieved a significant advancement in the automation and accuracy of phenotypic data extraction. Addressing the need for rapid extraction of leaf age and canopy width phenotypes in strawberry plants cultivated in controlled environments, this study introduces a novel high-throughput phenotyping extraction approach leveraging a mobile phenotyping platform and instance segmentation technology. [Methods] Data acquisition was conducted using a compact mobile phenotyping platform equipped with an array of sensors, including an RGB sensor, and edge control computers, capable of capturing overhead images of potted strawberry plants in greenhouses. Targeted adjustments to the network structure were made to develop an enhanced convolutional neural network (Mask R-CNN) model for processing strawberry plant image data and rapidly extracting plant phenotypic information. The model initially employed a split-attention networks (ResNeSt) backbone with a group attention module, replacing the original network to improve the precision and efficiency of image feature extraction. During training, the model adopted the Mosaic method, suitable for instance segmentation data augmentation, to expand the dataset of strawberry images. Additionally, it optimized the original cross-entropy classification loss function with a binary cross-entropy loss function to achieve better detection accuracy of plants and leaves. Based on this, the improved Mask R-CNN description involves post-processing of training results. It utilized the positional relationship between leaf and plant masks to statistically count the number of leaves. Additionally, it employed segmentation masks and image calibration against true values to calculate the canopy width of the plant. [Results and Discussions] This research conducted a thorough evaluation and comparison of the performance of an improved Mask R-CNN model, underpinned by the ResNeSt-101 backbone network. This model achieved a commendable mask accuracy of 80.1% and a detection box accuracy of 89.6%. It demonstrated the ability to efficiently estimate the age of strawberry leaves, demonstrating a high plant detection rate of 99.3% and a leaf count accuracy of 98.0%. This accuracy marked a significant improvement over the original Mask R-CNN model and meeting the precise needs for phenotypic data extraction. The method displayed notable accuracy in measuring the canopy widths of strawberry plants, with errors falling below 5% in about 98.1% of cases, highlighting its effectiveness in phenotypic dimension evaluation. Moreover, the model operated at a speed of 12.9 frames per second (FPS) on edge devices, effectively balancing accuracy and operational efficiency. This speed proved adequate for real-time applications, enabling rapid phenotypic data extraction even on devices with limited computational capabilitie. [Conclusions] This study successfully deployed a mobile phenotyping platform combined with instance segmentation techniques to analyze image data and extract various phenotypic indicators of strawberry plant. Notably, the method demonstrates remarkable robustness. The seamless fusion of mobile platforms and advanced image processing methods not only enhances efficiency but also ignifies a shift towards data-driven decision-making in agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Pig Back Transformer: Automatic 3D Pig Body Measurement Model
    WANG Yuxiao, SHI Yuanyuan, CHEN Zhaoda, WU Zhenfang, CAI Gengyuan, ZHANG Sumin, YIN Ling
    Smart Agriculture    2024, 6 (4): 76-90.   DOI: 10.12133/j.smartag.SA202401023
    Abstract1373)   HTML46)    PDF(pc) (2776KB)(2020)       Save

    [Objective] Nowadays most no contact body size measurement studies are based on point cloud segmentation method, they use a trained point cloud segmentation neural network to segment point cloud of pigs, then locate measurement points based on them. But point cloud segmentation neural network always need a larger graphics processing unit (GPU) memory, moreover, the result of the measurement key point still has room of improvement. This study aims to design a key point generating neural network to extract measurement key points from pig's point cloud. Reducing the GPU memory usage and improve the result of measurement points at the same time, improve both the efficiency and accuracy of the body size measurement. [Methods] A neural network model was proposed using improved Transformer attention mechanic called Pig Back Transformer for generating key points and back orientation points which were related to pig body dimensions. In the first part of the network, it was introduced an embedding structure for initial feature extraction and a Transformer encoder structure with edge attention which was a self-attention mechanic improved from Transformer's encoder. The embedding structure using two shared multilayer perceptron (MLP) and a distance embedding algorithm, it takes a set of points from the edge of pig back's point cloud as input and then extract information from the edge points set. In the encoder part, information about the offset distances between edge points and mass point which were their feature that extracted by the embedding structure mentioned before incorporated. Additionally, an extraction algorithm for back edge point was designed for extracting edge points to generate the input of the neural network model. In the second part of the network, it was proposed a Transformer encoder with improved self-attention called back attention. In the design of back attention, it also had an embedding structure before the encoder structure, this embedding structure extracted features from offset values, these offset values were calculated by the points which are none-edge and down sampled by farthest point sampling (FPS) to both the relative centroid point and model generated global key point from the first part that introduced before. Then these offset values were processed with max pooling with attention generated by the extracted features of the points' axis to extract more information that the original Transformer encoder couldn't extract with the same number of parameters. The output part of the model was designed to generate a set of offsets of the key points and points for back direction fitting, than add the set offset to the global key point to get points for pig body measurements. At last, it was introduced the methods for calculating body dimensions which were length, height, shoulder width, abdomen width, hip width, chest circumference and abdomen circumference using key points and back direction fitting points. [Results and Discussions] In the task of generating key points and points for back direction fitting, the improved Pig Back Transformer performed the best in the accuracy wise in the models tested with the same size of parameters, and the back orientation points generated by the model were evenly distributed which was a good preparation for a better body length calculation. A melting test for edge detection part with two attention mechanic and edge trim method both introduced above had being done, when the edge detection and the attention mechanic got cut off, the result had been highly impact, it made the model couldn't perform as well as before, when the edge trim method of preprocessing part had been cut off, there's a moderate impact on the trained model, but it made the loss of the model more inconsistence while training than before. When comparing the body measurement algorithm with human handy results, the relative error in length was 0.63%, which was an improvement compared to other models. On the other hand, the relative error of shoulder width, abdomen width and hip width had edged other models a little but there was no significant improvement so the performance of these measurement accuracy could be considered negligible, the relative error of chest circumference and abdomen circumference were a little bit behind by the other methods existed, it's because the calculate method of circumferences were not complicated enough to cover the edge case in the dataset which were those point cloud that have big holes in the bottom of abdomen and chest, it impacted the result a lot. [Conclusions] The improved Pig Back Transformer demonstrates higher accuracy in generating key points and is more resource-efficient, enabling the calculation of more accurate pig body measurements. And provides a new perspective for non-contact livestock body size measurements.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Hyperspectral Image-Based Method for Estimating Water and Chlorophyll Contents in Maize Leaves under Drought Stress
    WANG Jingyong, ZHANG Mingzhen, LING Huarong, WANG Ziting, GAI Jingyao
    Smart Agriculture    2023, 5 (3): 142-153.   DOI: 10.12133/j.smartag.SA202308018
    Abstract533)   HTML50)    PDF(pc) (2191KB)(1958)       Save

    [Objectives] Chlorophyll content and water content are key physiological indicators of crop growth, and their non-destructive detection is a key technology to realize the monitoring of crop growth status such as drought stress. This study took maize as an object to develop a hyperspectral-based approach for the rapid and non-destructive acquisition of the leaf chlorophyll content and water content for drought stress assessment. [Methods] Drought treatment experiments were carried out in a greenhouse of the College of Agriculture, Guangxi University. Maize plants were subjected to drought stress treatment at the seedling stage (four leaves). Four drought treatments were set up for normal water treatment [CK], mild drought [W1], moderate drought [W2], and severe drought [W3], respectively. Leaf samples were collected at the 3rd, 6th, and 9th days after drought treatments, and 288 leaf samples were collected in total, with the corresponding chlorophyll content and water content measured in a standard laboratory protocol. A pair of push-broom hyperspectral cameras were used to collect images of the 288 seedling maize leaf samples, and image processing techniques were used to extract the mean spectra of the leaf lamina part. The algorithm flow framework of "pre-processing - feature extraction - machine learning inversion" was adopted for processing the extracted spectral data. The effects of different pre-processing methods, feature wavelength extraction methods and machine learning regression models were analyzed systematically on the prediction performance of chlorophyll content and water content, respectively. Accordingly, the optimal chlorophyll content and water content inversion models were constructed. Firstly, 70% of the spectral data was randomly sampled and used as the training dataset for training the inversion model, whereas the remaining 30% was used as the testing dataset to evaluate the performance of the inversion model. Subsequently, the effects of different spectral pre-processing methods on the prediction performance of chlorophyll content and water content were compared. Different feature wavelengths were extracted from the optimal pre-processed spectra using different algorithms, then their capabilities in preserve the information useful for the inversion of leaf chlorophyll content and water content were compared. Finally, the performances of different machine learning regression model were compared, and the optimal inversion model was constructed and used to visualize the chlorophyll content and water content. Additionally, the construction of vegetation coefficients were explored for the inversion of chlorophyll content and water content and evaluated their inversion ability. The performance evaluation indexes used include determination coefficient and root mean squared error (RMSE). [Results and Discussions] With the aggravation of stress, the reflectivity of leaves in the wavelength range of 400~1700 nm gradually increased with the degree of drought stress. For the inversion of leaf chlorophyll content and water content, combining stepwise regression (SR) feature extraction with Stacking regression could obtain an optimal performance for chlorophyll content prediction, with an R2 of 0.878 and an RMSE of 0.317 mg/g. Compared with the full-band stacking model, SR-Stacking not only improved R2 by 2.9%, reduced RMSE by 0.0356mg/g, but also reduced the number of model input variables from 1301 to 9. Combining the successive projection algorithm (SPA) feature extraction with Stacking regression could obtain the optimal performance for water content prediction, with an R2 of 0.859 and RMSE of 3.75%. Compared with the full-band stacking model, SPA-Stacking not only increased R2 by 0.2%, reduced RMSE by 0.03%, but also reduced the number of model input variables from 1301 to 16. As the newly constructed vegetation coefficients, normalized difference vegetation index(NDVI) [(R410-R559)/(R410+R559)] and ratio index (RI) (R400/R1171) had the highest accuracy and were significantly higher than the traditional vegetation coefficients for chlorophyll content and water content inversion, respectively. Their R2 were 0.803 and 0.827, and their RMSE were 0.403 mg/g and 3.28%, respectively. The chlorophyll content and water content of leaves were visualized. The results showed that the physiological parameters of leaves could be visualized and the differences of physiological parameters in different regions of the same leaves can be found more intuitively and in detail. [Conclusions] The inversion models and vegetation indices constructed based on hyperspectral information can achieve accurate and non-destructive measurement of chlorophyll content and water content in maize leaves. This study can provide a theoretical basis and technical support for real-time monitoring of corn growth status. Through the leaf spectral information, according to the optimal model, the water content and chlorophyll content of each pixel of the hyperspectral image can be predicted, and the distribution of water content and chlorophyll content can be intuitively displayed by color. Because the field environment is more complex, transfer learning will be carried out in future work to improve its generalization ability in different environments subsequently and strive to develop an online monitoring system for field drought and nutrient stress.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Root Image Segmentation Method Based on Improved UNet and Transfer Learning
    TANG Hui, WANG Ming, YU Qiushi, ZHANG Jiaxi, LIU Liantao, WANG Nan
    Smart Agriculture    2023, 5 (3): 96-109.   DOI: 10.12133/j.smartag.SA202308003
    Abstract577)   HTML65)    PDF(pc) (2442KB)(1918)       Save

    [Objective] The root system is an important component of plant composition, and its growth and development are crucial for plants. Root image segmentation is an important method for obtaining root phenotype information and analyzing root growth patterns. Research on root image segmentation still faces difficulties, because of the noise and image quality limitations, the intricate and diverse soil environment, and the ineffectiveness of conventional techniques. This paper proposed a multi-scale feature extraction root segmentation algorithm that combined data augmentation and transfer learning to enhance the generalization and universality of the root image segmentation models in order to increase the speed, accuracy, and resilience of root image segmentation. [Methods] Firstly, the experimental datasets were divided into a single dataset and a mixed dataset. The single dataset acquisition was obtained from the experimental station of Hebei Agricultural University in Baoding city. Additionally, a self-made RhizoPot device was used to collect images with a resolution pixels of 10,200×14,039, resulting in a total of 600 images. In this experiment, 100 sheets were randomly selected to be manually labeled using Adobe Photoshop CC2020 and segmented into resolution pixels of 768×768, and divided into training, validation, and test sets according to 7:2:1. To increase the number of experimental samples, an open source multi-crop mixed dataset was obtained in the network as a supplement, and it was reclassified into training, validation, and testing sets. The model was trained using the data augmentation strategy, which involved performing data augmentation operations at a set probability of 0.3 during the image reading phase, and each method did not affect the other. When the probability was less than 0.3, changes would be made to the image. Specific data augmentation methods included changing image attributes, randomly cropping, rotating, and flipping those images. The UNet structure was improved by designing eight different multi-scale image feature extraction modules. The module structure mainly included two aspects: Image convolution and feature fusion. The convolution improvement included convolutional block attention module (CBAM), depthwise separable convolution (DP Conv), and convolution (Conv). In terms of feature fusion methods, improvements could be divided into concatenation and addition. Subsequently, ablation tests were conducted based on a single dataset, data augmentation, and random loading of model weights, and the optimal multi-scale feature extraction module was selected and compared with the original UNet. Similarly, a single dataset, data augmentation, and random loading of model weights were used to compare and validate the advantages of the improved model with the PSPNet, SegNet, and DeeplabV3Plus algorithms. The improved model used pre-trained weights from a single dataset to load and train the model based on mixed datasets and data augmentation, further improving the model's generalization ability and root segmentation ability. [Results and Discussions] The results of the ablation tests indicated that Conv_ 2+Add was the best improved algorithm. Compared to the original UNet, the mIoU, mRecall, and root F1 values of the model increased by 0.37%, 0.99%, and 0.56%, respectively. And, comparative experiments indicate Unet+Conv_2+Add model was superior to the PSPNet, SegNet, and DeeplabV3Plus models, with the best evaluation results. And the values of mIoU, mRecall, and the harmonic average of root F1 were 81.62%, 86.90%, and 77.97%, respectively. The actual segmented images obtained by the improved model were more finely processed at the root boundary compared to other models. However, for roots with deep color and low contrast with soil particles, the improved model could only achieve root recognition and the recognition was sparse, sacrificing a certain amount of information extraction ability. This study used the root phenotype evaluation software Rhizovision to analyze the root images of the Unet+Conv_2+Add improved model, PSPNet, SegNet, and DeeplabV3Plu, respectively, to obtain the values of the four root phenotypes (total root length, average diameter, surface area, and capacity), and the results showed that the average diameter and surface area indicator values of the improved model, Unet+Conv_2+Add had the smallest differences from the manually labeled indicator values and the SegNet indicator values for the two indicators. Total root length and volume were the closest to those of the manual labeling. The results of transfer learning experiments proved that compared with ordinary training, the transfer training of the improved model UNet+Conv_2+Add increased the IoU value of the root system by 1.25%. The Recall value of the root system was increased by 1.79%, and the harmonic average value of F1 was increased by 0.92%. Moreover, the overall convergence speed of the model was fast. Compared with regular training, the transfer training of the original UNet improved the root IoU by 0.29%, the root Recall by 0.83%, and the root F1 value by 0.21%, which indirectly confirmed the effectiveness of transfer learning. [Conclusions] The multi-scale feature extraction strategy proposed in this study can accurately and efficiently segment roots, and further improve the model's generalization ability using transfer learning methods, providing an important research foundation for crop root phenotype research.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Measurement Method of Beef Cattle Body Size Based on Multimodal Image Information and Improved Instance Segmentation Network
    WENG Zhi, FAN Qi, ZHENG Zhiqiang
    Smart Agriculture    2024, 6 (4): 64-75.   DOI: 10.12133/j.smartag.SA202310007
    Abstract1292)   HTML55)    PDF(pc) (3345KB)(1836)       Save

    [Objective] The body size parameter of cattle is a key indicator reflecting the physical development of cattle, and is also a key factor in the cattle selection and breeding process. In order to solve the demand of measuring body size of beef cattle in the complex environment of large-scale beef cattle ranch, an image acquisition device and an automatic measurement algorithm of body size were designed. [Methods] Firstly, the walking channel of the beef cattle was established, and when the beef cattle entered the restraining device through the channel, the RGB and depth maps of the image on the right side of the beef cattle were acquired using the Inter RealSense D455 camera. Secondly, in order to avoid the influence of the complex environmental background, an improved instance segmentation network based on Mask2former was proposed, adding CBAM module and CA module, respectively, to improve the model's ability to extract key features from different perspectives, extracting the foreground contour from the 2D image of the cattle, partitioning the contour, and comparing it with other segmentation algorithms, and using curvature calculation and other mathematical methods to find the required body size measurement points. Thirdly, in the processing of 3D data, in order to solve the problem that the pixel point to be measured in the 2D RGB image was null when it was projected to the corresponding pixel coordinates in the depth-valued image, resulting in the inability to calculate the 3D coordinates of the point, a series of processing was performed on the point cloud data, and a suitable point cloud filtering and point cloud segmentation algorithm was selected to effectively retain the point cloud data of the region of the cattle's body to be measured, and then the depth map was 16. Then the depth map was filled with nulls in the field to retain the integrity of the point cloud in the cattle body region, so that the required measurement points could be found and the 2D data could be returned. Finally, an extraction algorithm was designed to combine 2D and 3D data to project the extracted 2D pixel points into a 3D point cloud, and the camera parameters were used to calculate the world coordinates of the projected points, thus automatically calculating the body measurements of the beef cattle. [Results and Discussions] Firstly, in the part of instance segmentation, compared with the classical Mask R-CNN and the recent instance segmentation networks PointRend and Queryinst, the improved network could extract higher precision and smoother foreground images of cattles in terms of segmentation accuracy and segmentation effect, no matter it was for the case of occlusion or for the case of multiple cattles. Secondly, in three-dimensional data processing, the method proposed in the study could effectively extract the three-dimensional data of the target area. Thirdly, the measurement error of body size was analysed, among the four body size measurement parameters, the smallest average relative error was the height of the cross section, which was due to the more prominent position of the cross section, and the different standing positions of the cattle have less influence on the position of the cross section, and the largest average relative error was the pipe circumference, which was due to the influence of the greater overlap of the two front legs, and the higher requirements for the standing position. Finally, automatic body measurements were carried out on 137 beef cattle in the ranch, and the automatic measurements of the four body measurements parameters were compared with the manual measurements, and the results showed that the average relative errors of body height, cross section height, body slant length, and tube girth were 4.32%, 3.71%, 5.58% and 6.25%, respectively, which met the needs of the ranch. The shortcomings were that fewer body-size parameters were measured, and the error of measuring circumference-type body-size parameters was relatively large. Later studies could use a multi-view approach to increase the number of body rule parameters to be measured and improve the accuracy of the parameters in the circumference category. [Conclusions] The article designed an automatic measurement method based on two-dimensional and three-dimensional contactless body measurements of beef cattle. Moreover, the innovatively proposed method of measuring tube girth has higher accuracy and better implementation compared with the current research on body measurements in beef cattle. The relative average errors of the four body tape parameters meet the needs of pasture measurements and provide theoretical and practical guidance for the automatic measurement of body tape in beef cattle.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Path Planning and Motion Control Method for Sick and Dead Animal Transport Robots Integrating Improved A * Algorithm and Fuzzy PID
    XU Jishuang, JIAO Jun, LI Miao, LI Hualong, YANG Xuanjiang, LIU Xianwang, GUO Panpan, MA Zhirun
    Smart Agriculture    2023, 5 (4): 127-136.   DOI: 10.12133/j.smartag.SA202308001
    Abstract495)   HTML18)    PDF(pc) (1068KB)(1748)       Save

    [Objective] A key challenge for the harmless treatment center of sick and dead animal is to prevent secondary environmental pollution, especially during the process of transporting the animals from cold storage to intelligent treatment facilities. In order to solve this problem and achieve the intelligent equipment process of transporting sick and dead animal from storage cold storage to harmless treatment equipment in the harmless treatment center, it is necessary to conduct in-depth research on the key technical problems of path planning and autonomous walking of transport robots. [Methods] A * algorithm is mainly adopted for the robot path planning algorithm for indoor environments, but traditional A * algorithms have some problems, such as having many inflection points, poor smoothness, long calculation time, and many traversal nodes. In order to solve these problems, a path planning method for the harmless treatment of diseased and dead animal using transport robots based on the improved A algorithm was constructed, as well as a motion control method based on fuzzy proportional integral differential (PID). The Manhattan distance method was used to replace the heuristic function of the traditional A * algorithm, improving the efficiency of calculating the distance between the starting and ending points in the path planning process. Referring to the actual location of the harmless treatment site for sick and dead animal, vector cross product calculation was performed based on the vector from the starting point to the target point and the vector from the current position to the endpoint target. Additional values were added and dynamic adjustments were implemented, thereby changing the value of the heuristic function. In order to further improve the efficiency of path planning and reduce the search for nodes in the planning process, a method of adding function weights to the heuristic function was studied based on the actual situation on site, to change the weights according to different paths. When the current location node was relatively open, the search efficiency was improved by increasing the weight. When encountering situations such as corners, the weight was reduced to improve the credibility of the path. By improving the heuristic function, a driving path from the starting point to the endpoint was quickly obtained, but the resulting path was not smooth enough. Meanwhile, during the tracking process, the robot needs to accelerate and decelerate frequently to adapt to the path, resulting in energy loss. Therefore, according to the different inflection points and control points of the path, different orders of Bessel functions were introduced to smooth the planning process for the path, in order to achieve practical application results. By analyzing the kinematics of robot, the differential motion method of the track type was clarified. On this basis, a walking control algorithm for the robot based on fuzzy PID control was studied and proposed. Based on the actual operation status of the robot, the fuzzy rule conditions were recorded into a fuzzy control rule table, achieving online identification of the characteristic parameters of the robot and adjusting the angular velocity deviation of robot. When the robot controller received a fuzzy PID control signal, the angular velocity output from the control signal was converted into a motor rotation signal, which changed the motor speed on both sides of the robot to achieve differential control and adjust the steering of the robot. [Results and Discussions] Simulation experiments were conducted using the constructed environmental map obtained, verifying the effectiveness of the path planning method for the harmless treatment of sick and dead animal using the improved A algorithm. The comparative experiments between traditional A * algorithm and improved algorithm were conducted. The experimental results showed that the average traversal nodes of the improved A * algorithm decreased from 3 067 to 1 968, and the average time of the algorithm decreased from 20.34 s to 7.26 s. Through on-site experiments, the effectiveness and reliability of the algorithm were further verified. Different colors were used to identify the planned paths, and optimization comparison experiments were conducted on large angle inflection points, U-shaped inflection points, and continuous inflection points in the paths, verifying the optimization effect of the Bessel function on path smoothness. The experimental results showed that the path optimized by the Bessel function was smoother and more suitable for the walking of robot in practical scenarios. Fuzzy PID path tracking experiment results showed that the loading truck can stay close to the original route during both straight and turning driving, demonstrating the good effect of fuzzy PID on path tracking. Further experiments were conducted on the harmless treatment center to verify the effectiveness and practical application of the improved algorithm. Based on the path planning algorithm, the driving path of robot was quickly planned, and the fuzzy PID control algorithm was combined to accurately output the angular velocity, driving the robot to move. The transport robots quickly realized the planning of the transportation path, and during the driving process, could always be close to the established path, and the deviation error was maintained within a controllable range. [Conclusions] A path planning method for the harmless treatment of sick and dead animal using an transport robots based on an improved A * algorithm combined with a fuzzy PID motion control was proposed in this study. This method could effectively shorten the path planning time, reduce traversal nodes, and improve the efficiency and smoothness of path planning.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Prospect of Intelligent Monitoring Systems for the Physiological Indicators of Beef Cattle
    ZHANG Fan, ZHOU Mengting, XIONG Benhai, YANG Zhengang, LIU Minze, FENG Wenxiao, TANG Xiangfang
    Smart Agriculture    2024, 6 (4): 1-17.   DOI: 10.12133/j.smartag.SA202312001
    Abstract1240)   HTML81)    PDF(pc) (1471KB)(1718)       Save

    [Significance] The beef cattle industry plays a pivotal role in the development of China's agricultural economy and the enhancement of people's dietary structure. However, there exists a substantial disparity in feeding management practices and economic efficiency of beef cattle industry compared to developed countries. While the beef cattle industry in China is progressing towards intensive, modern, and large-scale development, it encounters challenges such as labor shortage and rising labor costs that seriously affect its healthy development. The determination of animal physiological indicators plays an important role in monitoring animal welfare and health status. Therefore, leveraging data collected from various sensors as well as technologies like machine learning, data mining, and modeling analysis enables automatic acquisition of meaningful information on beef cattle physiological indicators for intelligent management of beef cattle. In this paper, the intelligent monitoring technology of physiological indicators in beef cattle breeding process and its application value are systematically summarized, and the existing challenges and future prospects of intelligent beef cattle breeding process in China are prospected. [Progress] The methods of obtaining information on beef cattle physiological indicators include contact sensors worn on the body and non-contact sensors based on various image acquisitions. Monitoring the exercise behavior of beef cattle plays a crucial role in disease prevention, reproduction monitoring, and status assessment. The three-axis accelerometer sensor, which tracks the amount of time that beef cattle spend on lying, walking, and standing, is a widely used technique for tracking the movement behavior of beef cattle. Through machine vision analysis, individual recognition of beef cattle and identification of standing, lying down, and straddling movements can also be achieved, with the characteristics of non-contact, stress-free, low cost, and generating high data volume. Body temperature in beef cattle is associated with estrus, calving, and overall health. Sensors for monitoring body temperature include rumen temperature sensors and rectal temperature sensors, but there are issues with their inconvenience. Infrared temperature measurement technology can be utilized to detect beef cattle with abnormal temperatures by monitoring eye and ear root temperatures, although the accuracy of the results may be influenced by environmental temperature and monitoring distance, necessitating calibration. Heart rate and respiratory rate in beef cattle are linked to animal diseases, stress, and pest attacks. Monitoring heart rate can be accomplished through photoelectric volume pulse wave measurement and monitoring changes in arterial blood flow using infrared emitters and receivers. Respiratory rate monitoring can be achieved by identifying different nostril temperatures during inhalation and exhalation using thermal infrared imaging technology. The ruminating behavior of beef cattle is associated with health and feed nutrition. Currently, the primary tools used to detect rumination behavior are pressure sensors and three-axis accelerometer sensors positioned at various head positions. Rumen acidosis is a major disease in the rapid fattening process of beef cattle, however, due to limitations in battery life and electrode usage, real-time pH monitoring sensors placed in the rumen are still not widely utilized. Changes in animal physiology, growth, and health can result in alterations in specific components within body fluids. Therefore, monitoring body fluids or surrounding gases through biosensors can be employed to monitor the physiological status of beef cattle. By processing and analyzing the physiological information of beef cattle, indicators such as estrus, calving, feeding, drinking, health conditions, and stress levels can be monitored. This will contribute to the intelligent development of the beef cattle industry and enhance management efficiency. While there has been some progress made in developing technology for monitoring physiological indicators of beef cattle, there are still some challenges that need to be addressed. Contact sensors consume more energy which affects their lifespan. Various sensors are susceptible to environmental interference which affects measurement accuracy. Additionally, due to a wide variety of beef cattle breeds, it is difficult to establish a model database for monitoring physiological indicators under different feeding conditions, breeding stages, and breeds. Furthermore, the installation cost of various intelligent monitoring devices is relatively high, which also limits its utilization coverage. [Conclusion and Prospects] The application of intelligent monitoring technology for beef cattle physiological indicators is highly significance in enhancing the management level of beef cattle feeding. Intelligent monitoring systems and devices are utilized to acquire physiological behavior data, which are then analyzed using corresponding data models or classified through deep learning techniques to promptly monitor subtle changes in physiological indicators. This enables timely detection of sick, estrus, and calving cattle, facilitating prompt measures by production managers, reducing personnel workload, and improving efficiency. The future development of physiological indicators monitoring technologies in beef cattle primarily focuses on the following three aspects: (1) Enhancing the lifespan of contact sensors by reducing energy consumption, decreasing data transmission frequency, and improving battery life. (2) Integrating and analyzing various monitoring data from multiple perspectives to enhance the accuracy and utility value. (3) Strengthening research on non-contact, high-precision and automated analysis technologies to promote the precise and intelligent development within the beef cattle industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Imaging System for Plant Photosynthetic Phenotypes Incorporating Three-dimensional Structured Light and Chlorophyll Fluorescence
    SHU Hongwei, WANG Yuwei, RAO Yuan, ZHU Haojie, HOU Wenhui, WANG Tan
    Smart Agriculture    2024, 6 (1): 63-75.   DOI: 10.12133/j.smartag.SA202311018
    Abstract468)   HTML40)    PDF(pc) (1819KB)(1696)       Save

    Objective The investigation of plant photosynthetic phenotypes is essential for unlocking insights into plant physiological characteristics and dissecting morphological traits. However, traditional two-dimensional chlorophyll fluorescence imaging methods struggle to capture the complex three-dimensional spatial variations inherent in plant photosynthetic processes. To boost the efficacy of plant phenotyping and meet the increasingly demand for high-throughput analysis of photosynthetic phenotypes, the development and validation of a novel plant photosynthetic phenotype imaging system was explored, which uniquely combines three-dimensional structured light techniques with chlorophyll fluorescence technology. Methods The plant photosynthetic phenotype imaging system was composed of three primary parts: A tailored light source and projector, a camera, and a motorized filter wheel fitted with filters of various bandwidths, in addition to a terminal unit equipped with a development board and a touchscreen interface. The system was based on the principles and unique characteristics of chlorophyll fluorescence and structured light phase-shifted streak 3D reconstruction techniques. It utilized the custom-designed light source and projector, together with the camera's capability to choose specific wavelength bands, to its full potential. The system employed low-intensity white light within the 400–700 nm spectrum to elicit stable fluorescence, with blue light in the 440–450 nm range optimally triggering the fluorescence response. A projector was used to project dual-frequency, twelve-step phase-shifted stripes onto the plant, enabling the capture of both planar and stripe images, which were essential for the reconstruction of the plant's three-dimensional structure. An motorized filter wheel containing filters for red, green, blue, and near-infrared light, augmented by a filter less wheel for camera collaboration, facilitated the collection of images of plants at different wavelengths under varying lighting conditions. When illuminated with white light, filters corresponding to the red, green, and blue bands were applied to capture multiband images, resulting in color photographs that provides a comprehensive documentation of the plant's visual features. Upon exposure to blue light, the near-infrared filter was employed to capture near-infrared images, yielding data on chlorophyll fluorescence intensity. During the structured light streak projection, no filter was applied to obtain both planar and streak images of the plant, which were then employed in the 3D morphological reconstruction of the plant. The terminal, incorporating a development board and a touch screen, served as the control hub for the data acquisition and subsequent image processing within the plant photosynthetic phenotypic imaging system. It enabled the switching of light sources and the selection of camera bands through a combination of command and serial port control circuits. Following image acquisition, the data were transmitted back to the development board for analysis, processing, storage, and presentation. To validate the accuracy of 3D reconstruction and the reliability of photosynthetic efficiency assessments by the system, a prototype of the plant photosynthetic phenotypic imaging system was developed using 3D structured light and chlorophyll fluorescence technology, in accordance with the aforementioned methods, serving as an experimental validation platform. The accuracy of 3D reconstruction and the effectiveness of photosynthetic analysis capabilities of this imaging system were further confirmed through the analysis and processing of the experimental results, with comparative evaluations conducted against conventional 3D reconstruction methods and traditional chlorophyll fluorescence-based photosynthetic efficiency analyses. Results and Discussions The imaging system utilized for plant photosynthetic phenotypes incorporates a dual-frequency phase-shift algorithm to facilitate the reconstruction of three-dimensional (3D) plant phenotypes. Simultaneously, plant chlorophyll fluorescence images were employed to evaluate the plant's photosynthetic efficiency. This method enabled the analysis of the distribution of photosynthetic efficiency within a 3D space, offering a significant advancement over traditional plant photosynthetic imaging techniques. The 3D phenotype reconstructed using this method exhibits high precision, with an overall reconstruction accuracy of 96.69%. The total error was merely 3.31%, and the time required for 3D reconstruction was only 1.11 s. A comprehensive comparison of the 3D reconstruction approach presented with conventional methods had validated the accuracy of this technique, laying a robust foundation for the precise estimation of a plant's 3D photosynthetic efficiency. In the realm of photosynthetic efficiency analysis, the correlation coefficient between the photosynthetic efficiency values inferred from the chlorophyll fluorescence image analysis and those determined by conventional analysis exceeded 0.9. The experimental findings suggest a significant correlation between the photosynthetic efficiency values obtained using the proposed method and those from traditional methods, which could be characterized by a linear relationship, thereby providing a basis for more precise predictions of plant photosynthetic efficiency. Conclusions The method melds the 3D phenotype of plants with an analysis of photosynthetic efficiency, allowing for a more holistic assessment of the spatial heterogeneity in photosynthetic efficiency among plants by examining the pseudo-color images of chlorophyll fluorescence's spatial distribution. This approach elucidates the discrepancies in photosynthetic efficiency across various regions. The plant photosynthetic phenotype imaging system affords an intuitive and comprehensive view of the photosynthetic efficiency in plants under diverse stress conditions. Additionally, It provides technical support for the analysis of the spatial heterogeneity of high-throughput photosynthetic efficiency in plants.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Three-Dimensional Environment Perception Technology for Agricultural Wheeled Robots: A Review
    CHEN Ruiyun, TIAN Wenbin, BAO Haibo, LI Duan, XIE Xinhao, ZHENG Yongjun, TAN Yu
    Smart Agriculture    2023, 5 (4): 16-32.   DOI: 10.12133/j.smartag.SA202308006
    Abstract744)   HTML91)    PDF(pc) (1885KB)(1666)       Save

    [Significance] As the research focus of future agricultural machinery, agricultural wheeled robots are developing in the direction of intelligence and multi-functionality. Advanced environmental perception technologies serve as a crucial foundation and key components to promote intelligent operations of agricultural wheeled robots. However, considering the non-structured and complex environments in agricultural on-field operational processes, the environmental information obtained through conventional 2D perception technologies is limited. Therefore, 3D environmental perception technologies are highlighted as they can provide more dimensional information such as depth, among others, thereby directly enhancing the precision and efficiency of unmanned agricultural machinery operation. This paper aims to provide a detailed analysis and summary of 3D environmental perception technologies, investigate the issues in the development of agricultural environmental perception technologies, and clarify the future key development directions of 3D environmental perception technologies regarding agricultural machinery, especially the agricultural wheeled robot. [Progress] Firstly, an overview of the general status of wheeled robots was introduced, considering their dominant influence in environmental perception technologies. It was concluded that multi-wheel robots, especially four-wheel robots, were more suitable for the agricultural environment due to their favorable adaptability and robustness in various agricultural scenarios. In recent years, multi-wheel agricultural robots have gained widespread adoption and application globally. The further improvement of the universality, operation efficiency, and intelligence of agricultural wheeled robots is determined by the employed perception systems and control systems. Therefore, agricultural wheeled robots equipped with novel 3D environmental perception technologies can obtain high-dimensional environmental information, which is significant for improving the accuracy of decision-making and control. Moreover, it enables them to explore effective ways to address the challenges in intelligent environmental perception technology. Secondly, the recent development status of 3D environmental perception technologies in the agriculture field was briefly reviewed. Meanwhile, sensing equipment and the corresponding key technologies were also introduced. For the wheeled robots reported in the agriculture area, it was noted that the applied technologies of environmental perception, in terms of the primary employed sensor solutions, were divided into three categories: LiDAR, vision sensors, and multi-sensor fusion-based solutions. Multi-line LiDAR had better performance on many tasks when employing point cloud processing algorithms. Compared with LiDAR, depth cameras such as binocular cameras, TOF cameras, and structured light cameras have been comprehensively investigated for their application in agricultural robots. Depth camera-based perception systems have shown superiority in cost and providing abundant point cloud information. This study has investigated and summarized the latest research on 3D environmental perception technologies employed by wheeled robots in agricultural machinery. In the reported application scenarios of agricultural environmental perception, the state-of-the-art 3D environmental perception approaches have mainly focused on obstacle recognition, path recognition, and plant phenotyping. 3D environmental perception technologies have the potential to enhance the ability of agricultural robot systems to understand and adapt to the complex, unstructured agricultural environment. Furthermore, they can effectively address several challenges that traditional environmental perception technologies have struggled to overcome, such as partial sensor information loss, adverse weather conditions, and poor lighting conditions. Current research results have indicated that multi-sensor fusion-based 3D environmental perception systems outperform single-sensor-based systems. This superiority arises from the amalgamation of advantages from various sensors, which concurrently serve to mitigate individual shortcomings. [Conclusions and Prospects] The potential of 3D environmental perception technology for agricultural wheeled robots was discussed in light of the evolving demands of smart agriculture. Suggestions were made to improve sensor applicability, develop deep learning-based agricultural environmental perception technology, and explore intelligent high-speed online multi-sensor fusion strategies. Currently, the employed sensors in agricultural wheeled robots may not fully meet practical requirements, and the system's cost remains a barrier to widespread deployment of 3D environmental perception technologies in agriculture. Therefore, there is an urgent need to enhance the agricultural applicability of 3D sensors and reduce production costs. Deep learning methods were highlighted as a powerful tool for processing information obtained from 3D environmental perception sensors, improving response speed and accuracy. However, the limited datasets in the agriculture field remain a key issue that needs to be addressed. Additionally, multi-sensor fusion has been recognized for its potential to enhance perception performance in complex and changeable environments. As a result, it is clear that 3D environmental perception technology based on multi-sensor fusion is the future development direction of smart agriculture. To overcome challenges such as slow data processing speed, delayed processed data, and limited memory space for storing data, it is essential to investigate effective fusion schemes to achieve online multi-source information fusion with greater intelligence and speed.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agri-QA Net: Multimodal Fusion Large Language Model Architecture for Crop Knowledge Question-Answering System
    WU Huarui, ZHAO Chunjiang, LI Jingchen
    Smart Agriculture    2025, 7 (1): 1-10.   DOI: 10.12133/j.smartag.SA202411005
    Abstract1214)   HTML163)    PDF(pc) (1010KB)(1659)       Save

    [Objective] As agriculture increasingly relies on technological innovations to boost productivity and ensure sustainability, farmers need efficient and accurate tools to aid their decision-making processes. A key challenge in this context is the retrieval of specialized agricultural knowledge, which can be complex and diverse in nature. Traditional agricultural knowledge retrieval systems have often been limited by the modalities they utilize (e.g., text or images alone), which restricts their effectiveness in addressing the wide range of queries farmers face. To address this challenge, a specialized multimodal question-answering system tailored for cabbage cultivation was proposed. The system, named Agri-QA Net, integrates multimodal data to enhance the accuracy and applicability of agricultural knowledge retrieval. By incorporating diverse data modalities, Agri-QA Net aims to provide a holistic approach to agricultural knowledge retrieval, enabling farmers to interact with the system using multiple types of input, ranging from spoken queries to images of crop conditions. By doing so, it helps address the complexity of real-world agricultural environments and improves the accessibility of relevant information. [Methods] The architecture of Agri-QA Net was built upon the integration of multiple data modalities, including textual, auditory, and visual data. This multifaceted approach enables the system to develop a comprehensive understanding of agricultural knowledge, allowed the system to learn from a wide array of sources, enhancing its robustness and generalizability. The system incorporated state-of-the-art deep learning models, each designed to handle one specific type of data. Bidirectional Encoder Representations from Transformers (BERT)'s bidirectional attention mechanism allowed the model to understand the context of each word in a given sentence, significantly improving its ability to comprehend complex agricultural terminology and specialized concepts. The system also incorporated acoustic models for processing audio inputs. These models analyzed the spoken queries from farmers, allowing the system to understand natural language inputs even in noisy, non-ideal environments, which was a common challenge in real-world agricultural settings. Additionally, convolutional neural networks (CNNs) were employed to process images from various stages of cabbage growth. CNNs were highly effective in capturing spatial hierarchies in images, making them well-suited for tasks such as identifying pests, diseases, or growth abnormalities in cabbage crops. These features were subsequently fused in a Transformer-based fusion layer, which served as the core of the Agri-QA Net architecture. The fusion process ensured that each modality—text, audio, and image—contributes effectively to the final model's understanding of a given query. This allowed the system to provide more nuanced answers to complex agricultural questions, such as identifying specific crop diseases or determining the optimal irrigation schedules for cabbage crops. In addition to the fusion layer, cross-modal attention mechanisms and domain-adaptive techniques were incorporated to refine the model's ability to understand and apply specialized agricultural knowledge. The cross-modal attention mechanism facilitated dynamic interactions between the text, audio, and image data, ensuring that the model paid attention to the most relevant features from each modality. Domain-adaptive techniques further enhanced the system's performance by tailoring it to specific agricultural contexts, such as cabbage farming, pest control, or irrigation management. [Results and Discussions] The experimental evaluations demonstrated that Agri-QA Net outperforms traditional single-modal or simple multimodal models in agricultural knowledge tasks. With the support of multimodal inputs, the system achieved an accuracy rate of 89.5%, a precision rate of 87.9%, a recall rate of 91.3%, and an F1-Score of 89.6%, all of which are significantly higher than those of single-modality models. The integration of multimodal data significantly enhanced the system's capacity to understand complex agricultural queries, providing more precise and context-aware answers. The addition of cross-modal attention mechanisms enabled for more nuanced and dynamic interaction between the text, audio, and image data, which in turn improved the model's understanding of ambiguous or context-dependent queries, such as disease diagnosis or crop management. Furthermore, the domain-adaptive technique enabled the system to focus on specific agricultural terminology and concepts, thereby enhancing its performance in specialized tasks like cabbage cultivation and pest control. The case studies presented further validated the system's ability to assist farmers by providing actionable, domain-specific answers to questions, demonstrating its practical application in real-world agricultural scenarios. [Conclusions] The proposed Agri-QA Net framework is an effective solution for addressing agricultural knowledge questions, especially in the domain of cabbage cultivation. By integrating multimodal data and leveraging advanced deep learning techniques, the system demonstrates a high level of accuracy and adaptability. This study not only highlights the potential of multimodal fusion in agriculture but also paves the way for future developments in intelligent systems designed to support precision farming. Further work will focus on enhancing the model's performance by expanding the dataset to include more diverse agricultural scenarios, refining the handling of dialectical variations in audio inputs, and improving the system's ability to detect rare crop diseases. The ultimate goal is to contribute to the modernization of agricultural practices, offering farmers more reliable and effective tools to solve the challenges in crop management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Chilli-YOLO: An Intelligent Maturity Detection Algorithm for Field-Grown Chilli Based on Improved YOLOv10
    SI Chaoguo, LIU Mengchen, WU Huarui, MIAO Yisheng, ZHAO Chunjiang
    Smart Agriculture    2025, 7 (2): 160-171.   DOI: 10.12133/j.smartag.SA202411002
    Abstract327)   HTML58)    PDF(pc) (2545KB)(1624)       Save

    [Objective] In modern agriculture, the rapid and accurate detection of chillies at different maturity stages is a critical step for determining the optimal harvesting time and achieving intelligent sorting of field-grown chillies. However, existing target detection models face challenges in efficiency and accuracy when applied to the task of detecting chilli maturity, which limit their widespread use and effectiveness in practical applications. To address these challenges, a new algorithm, Chilli-YOLO, was proposed for achieving efficient and precise detection of chilli maturity in complex environments. [Methods] A comprehensive image dataset was collected, capturing chillis under diverse and realistic agricultural conditions, including varying lighting conditions, camera angles, and background complexities. These images were then meticulously categorized into four distinct maturity stages: Immature, transitional, mature, and dried. Data augmentation techniques were employed to expand the dataset and enhance the model's generalization capabilities. To develop an accurate and efficient chili maturity detection system, the YOLOv10s object detection network was chosen as the foundational architecture. The model's performance was further enhanced through strategic optimizations targeting the backbone network. Specifically, standard convolutional layers were replaced with Ghost convolutions. This technique generated more feature maps from fewer parameters, resulting in significant computational savings and improved processing speed without compromising feature extraction quality. Additionally, the C2f module was substituted with the more computationally efficient GhostConv module, further reducing redundancy and enhancing the model's overall efficiency. To improve the model's ability to discern subtle visual cues indicative of maturity, particularly in challenging scenarios involving occlusion, uneven lighting, or complex backgrounds, the partial self-attention (PSA) module within YOLOv10s was replaced with the second-order channel attention (SOCA) mechanism. SOCA leverages higher-order feature correlations to more effectively capture fine-grained characteristics of the chillis. This enabled the model to focus on relevant feature channels and effectively identify subtle maturity-related features, even when faced with significant visual noise and interference. Finally, to refine the precision of target localization and minimize bounding box errors, the extended intersection over union (XIoU) loss function was integrated into the model training process. XIoU enhances the traditional IoU loss by considering factors such as the aspect ratio difference and the normalized distance between the predicted and ground truth bounding boxes. By optimizing for these factors, the model achieved significantly improved localization accuracy, resulting in a more precise delineation of chillis in the images and contributing to the overall enhancement of the detection performance. The combined implementation of these improvements aimed to construct an effective approach to correctly classify the maturity level of chillis within the challenging and complex environment of a real-world farm. [Results and Discussion] The experimental results on the custom-built chilli maturity detection dataset showed that the Chilli-YOLO model performed excellently across multiple evaluation metrics. The model achieved an accuracy of 90.7%, a recall rate of 82.4%, and a mean average precision (mAP) of 88.9%. Additionally, the model's computational load, parameter count, model size, and inference time were 18.3 GFLOPs, 6.37 M, 12.6 M, and 7.3 ms, respectively. Compared to the baseline model, Chilli-YOLO improved accuracy by 2.6 percent point, recall by 2.8 percent point and mAP by 2.8 percent point. At the same time, the model's computational load decreased by 6.2 GFLOPs, the parameter count decreased by 1.67 M, model size reduced by 3.9 M. These results indicated that Chilli-YOLO strikes a good balance between accuracy and efficiency, making it capable of fast and precise detection of chilli maturity in complex agricultural environments. Moreover, compared to earlier versions of the YOLO model, Chilli-YOLO showed improvements in accuracy of 2.7, 4.8, and 5 percent point over YOLOv5s, YOLOv8n, and YOLOv9s, respectively. Recall rates were higher by 1.1, 0.3, and 2.3 percent point, and mAP increased by 1.2, 1.7, and 2.3 percent point, respectively. In terms of parameter count, model size, and inference time, Chilli-YOLO outperformed YOLOv5. This avoided the issue of YOLOv8n's lower accuracy, which was unable to meet the precise detection needs of complex outdoor environments. When compared to the traditional two-stage network Faster RCNN, Chilli-YOLO showed significant improvements across all evaluation metrics. Additionally, compared to the one-stage network SSD, Chilli-YOLO achieved substantial gains in accuracy, recall, and mAP, with increases of 16.6%, 12.1%, and 16.8%, respectively. Chilli-YOLO also demonstrated remarkable improvements in memory usage, model size, and inference time. These results highlighted the superior overall performance of the Chilli-YOLO model in terms of both memory consumption and detection accuracy, confirming its advantages for chilli maturity detection. [Conclusions] The proposed Chilli-YOLO model optimizes the network structure and loss functions, not only can significantly improve detection accuracy but also effectively reduce computational overhead, making it better suites for resource-constrained agricultural production environments. The research provides a reliable technical reference for intelligent harvesting of chillies in agricultural production environments, especially in resource-constrained settings.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Vegetable Crop Growth Modeling in Digital Twin Platform Based on Large Language Model Inference
    ZHAO Chunjiang, LI Jingchen, WU Huarui, YANG Yusen
    Smart Agriculture    2024, 6 (6): 63-71.   DOI: 10.12133/j.smartag.SA202410008
    Abstract1733)   HTML179)    PDF(pc) (1460KB)(1611)       Save

    [Objective] In the era of digital agriculture, real-time monitoring and predictive modeling of crop growth are paramount, especially in autonomous farming systems. Traditional crop growth models, often constrained by their reliance on static, rule-based methods, fail to capture the dynamic and multifactorial nature of vegetable crop growth. This research tried to address these challenges by leveraging the advanced reasoning capabilities of pre-trained large language models (LLMs) to simulate and predict vegetable crop growth with accuracy and reliability. Modeling the growth of vegetable crops within these platforms has historically been hindered by the complex interactions among biotic and abiotic factors. [Methods] The methodology was structured in several distinct phases. Initially, a comprehensive dataset was curated to include extensive information on vegetable crop growth cycles, environmental conditions, and management practices. This dataset incorporates continuous data streams such as soil moisture, nutrient levels, climate variables, pest occurrence, and historical growth records. By combining these data sources, the study ensured that the model was well-equipped to understand and infer the complex interdependencies inherent in crop growth processes. Then, advanced techniques was emploied for pre-training and fine-tuning LLMs to adapt them to the domain-specific requirements of vegetable crop modeling. A staged intelligent agent ensemble was designed to work within the digital twin platform, consisting of a central managerial agent and multiple stage-specific agents. The managerial agent was responsible for identifying transitions between distinct growth stages of the crops, while the stage-specific agents were tailored to handle the unique characteristics of each growth phase. This modular architecture enhanced the model's adaptability and precision, ensuring that each phase of growth received specialized attention and analysis. [Results and Discussions] The experimental validation of this method was conducted in a controlled agricultural setting at the Xiaotangshan Modern Agricultural Demonstration Park in Beijing. Cabbage (Zhonggan 21) was selected as the test crop due to its significance in agricultural production and the availability of comprehensive historical growth data. Over five years, the dataset collected included 4 300 detailed records, documenting parameters such as plant height, leaf count, soil conditions, irrigation schedules, fertilization practices, and pest management interventions. This dataset was used to train the LLM-based system and evaluate its performance using ten-fold cross-validation. The results of the experiments demonstrating the efficacy of the proposed system in addressing the complexities of vegetable crop growth modeling. The LLM-based model achieved 98% accuracy in predicting crop growth degrees and a 99.7% accuracy in identifying growth stages. These metrics significantly outperform traditional machine learning approaches, including long short-term memory (LSTM), XGBoost, and LightGBM models. The superior performance of the LLM-based system highlights its ability to reason over heterogeneous data inputs and make precise predictions, setting a new benchmark for crop modeling technologies. Beyond accuracy, the LLM-powered system also excels in its ability to simulate growth trajectories over extended periods, enabling farmers and agricultural managers to anticipate potential challenges and make proactive decisions. For example, by integrating real-time sensor data with historical patterns, the system can predict how changes in irrigation or fertilization practices will impact crop health and yield. This predictive capability is invaluable for optimizing resource allocation and mitigating risks associated with climate variability and pest outbreaks. [Conclusions] The study emphasizes the importance of high-quality data in achieving reliable and generalizable models. The comprehensive dataset used in this research not only captures the nuances of cabbage growth but also provides a blueprint for extending the model to other crops. In conclusion, this research demonstrates the transformative potential of combining large language models with digital twin technology for vegetable crop growth modeling. By addressing the limitations of traditional modeling approaches and harnessing the advanced reasoning capabilities of LLMs, the proposed system sets a new standard for precision agriculture. Several avenues also are proposed for future work, including expanding the dataset, refining the model architecture, and developing multi-crop and multi-region capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Localization Method for Agricultural Robots Based on Fusion of LiDAR and IMU
    LIU Yang, JI Jie, PAN Deng, ZHAO Lijun, LI Mingsheng
    Smart Agriculture    2024, 6 (3): 94-106.   DOI: 10.12133/j.smartag.SA202401009
    Abstract676)   HTML58)    PDF(pc) (3329KB)(1590)       Save

    [Objective] High-precision localization technology serves as the crucial foundation in enabling the autonomous navigation operations of intelligent agricultural robots. However, the traditional global navigation satellite system (GNSS) localization method faces numerous limitations, such as tree shadow, electromagnetic interference, and other factors in the agricultural environment brings challenges to the accuracy and reliability of localization technology. To address the deficiencies and achieve precise localization of agricultural robots independent of GNSS, a localization method was proposed based on the fusion of three-dimensional light detection and ranging (LiDAR) data and inertial measurement unit (IMU) information to enhance localization accuracy and reliability. [Methods] LiDAR was used to obtain point cloud data in the agricultural environment and realize self-localization via point cloud matching. By integrating real-time motion parameter measurements from the IMU with LiDAR data, a high-precision localization solution for agricultural robots was achieved through a specific fusion algorithm. Firstly, the LiDAR-obtained point cloud data was preprocessed and the depth map was used to save the data. This approach could reduce the dimensionality of the original LiDAR point cloud, and eliminate the disorder of the original LiDAR point cloud arrangement, facilitating traversal and clustering through graph search. Given the presence of numerous distinct crops like trees in the agricultural environment, an angle-based clustering method was adopted. Specific angle-based clustering criteria were set to group the point cloud data, leading to the segmentation of different clusters of points, and obvious crops in the agricultural environment was effectively perceived. Furthermore, to improve the accuracy and stability of positioning, an improved three-dimensional normal distribution transform (3D-NDT) localization algorithm was proposed. This algorithm operated by matching the LiDAR-scanned point cloud data in real time with the pre-existing down sampled point cloud map to achieve real-time localization. Considering that direct down sampling of LiDAR point clouds in the agricultural environment could result in the loss of crucial environmental data, a point cloud clustering operation was used in place of down sampling operation, thereby improving matching accuracy and positioning precision. Secondly, to address potential constraints and shortcomings of using a single sensor for robot localization, a multi-sensor information fusion strategy was deployed to improve the localization accuracy. Specifically, the extended Kalman filter algorithm (EKF) was chosen to fuse the localization data from LiDAR point cloud and the IMU odometer information. The IMU provided essential motion parameters such as acceleration and angular velocity of the agricultural robot, and by combining with the LiDAR-derived localization information, the localization of the agricultural robot could be more accurately estimated. This fusion approach maximized the advantages of different sensors, compensated for their individual limitations, and improved the overall localization accuracy of the agricultural robot. [Results and Discussions] A series of experimental results in the Gazebo simulation environment of the robot operating system (ROS) and real operation scenarios showed that the fusion localization method proposed had significant advantages. In the simulation environment, the average localization errors of the proposed multi-sensor data fusion localization method were 1.7 and 1.8 cm, respectively, while in the experimental scenario, these errors were 3.3 and 3.3 cm, respectively, which were significantly better than the traditional 3D-NDT localization algorithm. These findings showed that the localization method proposed in this study could achieve high-precision localization in the complex agricultural environment, and provide reliable localization assistance for the autonomous functioning of agricultural robots. [Conclusions] The proposed localization method based on the fusion of LiDAR data and IMU information provided a novel localization solution for the autonomous operation of agricultural robots in areas with limited GNSS reception. Through the comprehensive utilization of multi-sensor information and adopting advanced data processing and fusion algorithms, the localization accuracy of agricultural robots could be significantly improved, which could provide a new reference for the intelligence and automation of agricultural production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Knowledge Recommendation Model Integrating Time Perception and Context Filtering
    WANG Pengzhe, ZHU Huaji, MIAO Yisheng, LIU Chang, WU Huarui
    Smart Agriculture    2024, 6 (1): 123-134.   DOI: 10.12133/j.smartag.SA202312012
    Abstract899)   HTML35)    PDF(pc) (1503KB)(1585)       Save

    Objective Knowledge services in agricultural scenarios have the characteristics of long periodicity and prolonged activity time. Traditional recommendation models cannot effectively mine hidden information in agricultural scenarios, in order to improve the quality of agricultural knowledge recommendation services, agricultural contextual information based on agricultural time should be fully considered. To address these issues, a Time-aware and filter-enhanced sequential recommendation model for agricultural knowledge (TiFSA) was proposed, integrating temporal perception and enhanced filtering. Methods First, based on the temporal positional embedding, combining the temporal information of farmers' interactions with positional embedding based on time perception, it helped to learn project relevance based on agricultural season in agricultural contexts. A multi-head self-attention network recommendation algorithm based on time-awareness was proposed for the agricultural knowledge recommendation task, which extracted different interaction time information in the user interaction sequence and introduced it into the multi-head self-attention network to calculate the attention weight, which encoded the user's periodic interaction information based on the agricultural time, and also effectively captured the user's dynamic preference information over time. Then, through the temporal positional embedding, a filter filtering algorithm was introduced to adaptively attenuate the noise in farmers' situational data adaptively. The filtering algorithm was introduced to enhance the filtering module to effectively filter the noisy information in the agricultural dataset and alleviate the overfitting problem due to the poorly normalized and sparse agricultural dataset. By endowing the model with lower time complexity and adaptive noise attenuation capability. The applicability of this method in agricultural scenarios was improved. Next, a multi-head self attention network with temporal information was constructed to achieve unified modeling of time, projects, and features, and represent farmers' preferences of farmers over time in context, thereby providing reliable recommendation results for users. Finally, the AdamW optimizer was used to update and compute the model parameters. AdamW added L2 regularization and an appropriate penalty mechanism for larger weights, which could update all weights more smoothly and alleviate the problem of falling into local minima. Applied in the field of agricultural recommendation, it could further improve the training effect of the model. The experimental data came from user likes, comments, and corresponding time information in the "National Agricultural Knowledge Intelligent Service Cloud Platform", and the dataset ml-1m in the movie recommendation scenario was selected as an auxiliary validation of the performance of this model. Results and Discussions According to the user interaction sequence datasets in the "National Agricultural Knowledge Intelligent Service Cloud Platform", from the experimental results, it could be learned that TiFSA outperforms the other models on two different datasets, in which the enhancement was more obvious on the Agriculture dataset, where HR and NDCG were improved by 14.02% and 16.19%, respectively, compared to the suboptimal model, TiSASRec; while on the ml-1m dataset compared to the suboptimal model, SASRec, HR and NDCG were improved by 1.90% and 2.30%, respectively. In summary, the TiFSA model proposed in this paper has a large improvement compared with other models, which verifies verified the effectiveness of the TiFSA model and showed that the time interval information of farmer interaction and the filtering algorithm play an important role in the improvement of the model performance in the agricultural context. From the results of the ablation experiments, it could be seen that when the time-aware and enhanced filtering modules were removed, the values of the two metrics HR@10 and NDCG@10 were 0.293 6 and 0.203 9, respectively, and the recommended performance was poor. When only the time-aware module and only the augmentation filtering module were removed, the experimental results had different degrees of improvement compared to TiFSA-tf, and the TiFSA model proposed in this paper achieved the optimal performance in the two evaluation metrics. When only the multi-head self-attention network was utilized for recommendation, both recommendation metrics of the model were lower, indicating that the traditional sequence recommendation method that only considered the item number was not applicable to agricultural scenarios. When the augmented filtering module was introduced without the time-aware module, the model performance was improved, but still failed to achieve the ideal recommendation effect. When only the time-aware module was introduced without the augmented filtering module, there was a significant improvement in the model effect, which proved that the time-aware module was more applicable to agricultural scenarios and can effectively improve the model performance of the sequence recommendation task. When both time-aware and augmented filtering modules were introduced, the model performance was further improved, which on the one hand illustrated the dependence of the augmented filtering module on the time-aware module, and on the other hand verified the necessity of adopting the augmented filtering to the time-aware self-attention network model. Conclusions This research proposes an agricultural knowledge recommendation model that integrates time-awareness and augmented filtering, which introduces the user's interaction time interval into the embedded information, so that the model effectively learns the information of agricultural time in the agricultural scene, and the prediction of the user's interaction time and the object is more closely related to the actual scene; augmented filtering algorithms are used to attenuate the noise in the agricultural data. At the same time, the enhanced filtering algorithm is used to attenuate the noise in the agricultural data, and can be effectively integrated into the model for use, further improving the recommendation performance of the model. The experimental results show the effectiveness of the proposed TiFSA model on the agricultural dataset. The ablation experiments confirm the positive effect of time-awareness and enhanced filtering modules on the improvement of recommendation performance.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Remote Sensing Extraction Method of Terraced Fields Based on Improved DeepLab v3+
    ZHANG Jun, CHEN Yuyan, QIN Zhenyu, ZHANG Mengyao, ZHANG Jun
    Smart Agriculture    2024, 6 (3): 46-57.   DOI: 10.12133/j.smartag.SA202312028
    Abstract574)   HTML44)    PDF(pc) (2806KB)(1562)       Save

    [Objective] The accurate estimation of terraced field areas is crucial for addressing issues such as slope erosion control, water retention, soil conservation, and increasing food production. The use of high-resolution remote sensing imagery for terraced field information extraction holds significant importance in these aspects. However, as imaging sensor technologies continue to advance, traditional methods focusing on shallow features may no longer be sufficient for precise and efficient extraction in complex terrains and environments. Deep learning techniques offer a promising solution for accurately extracting terraced field areas from high-resolution remote sensing imagery. By utilizing these advanced algorithms, detailed terraced field characteristics with higher levels of automation can be better identified and analyzed. The aim of this research is to explore a proper deep learning algorithm for accurate terraced field area extraction in high-resolution remote sensing imagery. [Methods] Firstly, a terraced dataset was created using high-resolution remote sensing images captured by the Gaofen-6 satellite during fallow periods. The dataset construction process involved data preprocessing, sample annotation, sample cropping, and dataset partitioning with training set augmentation. To ensure a comprehensive representation of terraced field morphologies, 14 typical regions were selected as training areas based on the topographical distribution characteristics of Yuanyang county. To address misclassifications near image edges caused by limited contextual information, a sliding window approach with a size of 256 pixels and a stride of 192 pixels in each direction was utilized to vary the positions of terraced fields in the images. Additionally, geometric augmentation techniques were applied to both images and labels to enhance data diversity, resulting in a high-resolution terraced remote sensing dataset. Secondly, an improved DeepLab v3+ model was proposed. In the encoder section, a lightweight MobileNet v2 was utilized instead of Xception as the backbone network for the semantic segmentation model. Two shallow features from the 4th and 7th layers of the MobileNet v2 network were extracted to capture relevant information. To address the need for local details and global context simultaneously, the multi-scale feature fusion (MSFF) module was employed to replace the atrous spatial pyramid pooling (ASPP) module. The MSFF module utilized a series of dilated convolutions with increasing dilation rates to handle information loss. Furthermore, a coordinate attention mechanism was applied to both shallow and deep features to enhance the network's understanding of targets. This design aimed to lightweight the DeepLab v3+ model while maintaining segmentation accuracy, thus improving its efficiency for practical applications. [Results and Discussions] The research findings reveal the following key points: (1) The model trained using a combination of near-infrared, red, and green (NirRG) bands demonstrated the optimal overall performance, achieving precision, recall, F1-Score, and intersection over union (IoU) values of 90.11%, 90.22%, 90.17% and 82.10%, respectively. The classification results indicated higher accuracy and fewer discrepancies, with an error in reference area of only 12 hm2. (2) Spatial distribution patterns of terraced fields in Yuanyang county were identified through the deep learning model. The majority of terraced fields were found within the slope range of 8º to 25º, covering 84.97% of the total terraced area. Additionally, there was a noticeable concentration of terraced fields within the altitude range of 1 000 m to 2 000 m, accounting for 95.02% of the total terraced area. (3) A comparison with the original DeepLab v3+ network showed that the improved DeepLab v3+ model exhibited enhancements in terms of precision, recall, F1-Score, and IoU by 4.62%, 2.61%, 3.81% and 2.81%, respectively. Furthermore, the improved DeepLab v3+ outperformed UNet and the original DeepLab v3+ in terms of parameter count and floating-point operations. Its parameter count was only 28.6% of UNet and 19.5% of the original DeepLab v3+, while the floating-point operations were only 1/5 of UNet and DeepLab v3+. This not only improved computational efficiency but also made the enhanced model more suitable for resource-limited or computationally less powerful environments. The lightweighting of the DeepLab v3+ network led to improvements in accuracy and speed. However, the slection of the NirGB band combination during fallow periods significantly impacted the model's generalization ability. [Conclusions] The research findings highlights the significant contribution of the near-infrared (NIR) band in enhancing the model's ability to learn terraced field features. Comparing different band combinations, it was evident that the NirRG combination resulted in the highest overall recognition performance and precision metrics for terraced fields. In contrast to PSPNet, UNet, and the original DeepLab v3+, the proposed model showcased superior accuracy and performance on the terraced field dataset. Noteworthy improvements were observed in the total parameter count, floating-point operations, and the Epoch that led to optimal model performance, outperforming UNet and DeepLab v3+. This study underscores the heightened accuracy of deep learning in identifying terraced fields from high-resolution remote sensing imagery, providing valuable insights for enhanced monitoring and management of terraced landscapes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Severity Grading Model for Camellia Oleifera Anthracnose Infection Based on Improved YOLACT
    NIE Ganggang, RAO Honghui, LI Zefeng, LIU Muhua
    Smart Agriculture    2024, 6 (3): 138-147.   DOI: 10.12133/j.smartag.SA202402002
    Abstract417)   HTML34)    PDF(pc) (2130KB)(1553)       Save

    [Objective] Camellia oleifera is one of the four major woody oil plants in the world. Diseases is a significant factor leading to the decline in quality of Camellia oleifera and the financial loss of farmers. Among these diseases, anthracnose is a common and severe disease in Camellia oleifera forests, directly impacting yields and production rates. Accurate disease assessment can improve the prevention and control efficiency and safeguarding the farmers' profit. In this study, an improved You Only Look at CoefficienTs (YOLACT) based method was proposed to realize automatic and efficient grading of the severity of Camellia oleifera leaf anthracnose. [Methods] High-resolution images of Camellia oleifera anthracnose leaves were collected using a smartphone at the National Camellia oleifera Seed Base of Jiangxi Academy of Forestry, and finally 975 valid images were retained after a rigorous screening process. Five data enhancement means were applied, and a data set of 5 850 images was constructed finally, which was divided into training, validation, and test sets in a ratio of 7:2:1. For model selection, the Camellia-YOLACT model was proposed based on the YOLACT instance segmentation model, and by introducing improvements such as Swin-Transformer, weighted bi-directional feature pyramid network, and HardSwish activation function. The Swin Transformer was utilized for feature extraction in the backbone network part of YOLACT, leveraging the global receptive field and shift window properties of the self-attention mechanism in the Transformer architecture to enhance feature extraction capabilities. Additionally, a weighted bidirectional feature pyramid network was introduced to fuse feature information from different scales to improve the detection ability of the model for objects at different scales, thereby improving the detection accuracy. Furthermore, to increase the the model's robustness against the noise in the input data, the HardSwish activation function with stronger nonlinear capability was adopted to replace the ReLu activation function of the original model. Since images in natural environments usually have complex background and foreground information, the robustness of HardSwish helped the model better handling these situations and further improving the detection accuracy. With the above improvements, the Camellia-YOLACT model was constructed and experimentally validated by testing the Camellia oleifera anthracnose leaf image dataset. [Results and Discussions] A transfer learning approach was used for experimental validation on the Camellia oleifera anthracnose severity grading dataset, and the results of the ablation experiments showed that the mAP75 of Camellia-YOLACT proposed in this study was 86.8%, mAPall was 78.3%, mAR was 91.6% which were 5.7%, 2.5% and 7.9% higher than YOLACT model. In the comparison experiments, Camellia-YOLACT performed better than Segmenting Objects by Locations (SOLO) in terms of both accuracy and speed, and its detection speed was doubled compared to Mask R-CNN algorithm. Therefore, the Camellia-YOLACT algorithm was suitable in Camellia oleifera gardens for anthracnose real-time segmentation. In order to verify the outdoors detection performance of Camellia-YOLACT model, 36 groups of Camellia oleifera anthracnose grading experiments were conducted. Experimental results showed that the grading correctness of Camellia oleifera anthracnose injection severity reached 94.4%, and the average absolute error of K-value was 1.09%. Therefore, the Camellia-YOLACT model proposed in this study has a better performance on the grading of the severity of Camellia oleifera anthracnose. [Conclusions] The Camellia-YOLACT model proposed got high accuracy in leaf and anthracnose segmentation of Camellia oleifera, on the basis of which it can realize automatic grading of the severity of Camellia oleifera anthracnose. This research could provide technical support for the precise control of Camellia oleifera diseases.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Capacitive Soil Moisture Sensor Based on MoS2
    LI Lu, GE Yuqing, ZHAO Jianlong
    Smart Agriculture    2024, 6 (1): 28-35.   DOI: 10.12133/j.smartag.SA202309020
    Abstract499)   HTML38)    PDF(pc) (1373KB)(1435)       Save

    Objective The soil moisture content is a crucial factor that directly affected the growth and yield of crops. By using a soil measurement instrument to measure the soil's moisture content, lots of powerful data support for the development of agriculture can be provided. Furthermore, these data have guiding significance for the implementation of scientific irrigation and water-saving irrigation in farmland. In order to develop a reliable and efficient soil moisture sensor, a new capacitive soil moisture sensor based on microfabrication technology was proposed in this study. Capacitive moisture sensors have the advantages of low power consumption, good performance, long-term stability, and easy industrialization. Method The forked electrode array consists of multiple capacitors connected in parallel on the same plane. The ideal design parameters of 10 μm spacing and 75 pairs of forked electrodes were obtained by calculating the design of forked finger logarithms, forked finger spacing, forked finger width, forked finger length, and electrode thickness, and studying the influence of electrode parameters on capacitance sensitivity using COMSOL Multiphysics software. The size obtained an initial capacitance on the order of picofarads, and was not easily breakdown or failed. The sensor was constructed using microelectromechanical systems (MEMS) technology, where a 30 nm titanium adhesion layer was sputtered onto a glass substrate, followed by sputtering a 100 nm gold electrode to form a symmetrical structure of forked electrodes. Due to the strong adsorption capacity of water molecules of the MoS2 (molybdenum disulfide) layer, it exhibited high sensitivity to soil moisture and demonstrated excellent soil moisture sensing performance. The molybdenum disulfide was coated onto the completed electrodes as the humidity-sensitive material to create a humidity sensing layer. When the humidity changed, the dielectric constant of the electrode varied due to the moisture-absorbing characteristics of molybdenum disulfide, and the capacitance value of the device changed accordingly, thus enabling the measurement of soil moisture. Subsequently, the electrode was encapsulated with a polytetrafluoroethylene (PTFE) polymer film. The electrode encapsulated with the microporous film could be directly placed in the soil, which avoided direct contact between the soil/sand particles and the molybdenum disulfide on the device and allowed the humidity sensing unit to only capture the moisture in the soil for measuring humidity. This ensured the device's sensitivity to water moisture and improved its long-term stability. The method greatly reduced the size of the sensor, making it an ideal choice for on-site dynamic monitoring of soil moisture. Results and Discussions The surface morphology of molybdenum disulfide was characterized and analyzed using a Scanning Electron Microscope (SEM). It was observed that molybdenum disulfide nanomaterial exhibited a sheet-like two-dimensional structure, with smooth surfaces on the nanosheets. Some nanosheets displayed sharp edges or irregular shapes along the edges, and they were irregularly arranged with numerous gaps in between. The capacitive soil moisture sensor, which utilized molybdenum disulfide as the humidity-sensitive layer, exhibited excellent performance under varying levels of environmental humidity and soil moisture. At room temperature, a humidity generator was constructed using saturated salt solutions. Saturated solutions of lithium chloride, potassium acetate, magnesium chloride, copper chloride, sodium chloride, potassium chloride, and potassium sulfate were used to generate relative humidity levels of 11%, 23%, 33%, 66%, 75%, 84%, and 96%, respectively. The capacitance values of the sensor were measured at different humidity levels using an LCR meter (Agilent E4980A). The capacitance output of the sensor at a frequency of 200 Hz ranged from 12.13 pF to 187.42 nF as the relative humidity varied between 11% to 96%. The sensor exhibited high sensitivity and a wide humidity sensing range. Additionally, the frequency of the input voltage signal had a significant impact on the capacitance output of the sensor. As the testing frequency increased, the response of the sensor's system decreased. The humidity sensing performance of the sensor was tested in soil samples with moisture content of 8.66%, 13.91%, 22.02%, 31.11%, and 42.75%, respectively. As the moisture content in the soil increased from 8.66% to 42.75%, the capacitance output of the sensor at a frequency of 200 Hz increased from 119.51 nF to 377.98 nF, demonstrating a relatively high sensitivity. Similarly, as the frequency of the input voltage increased, the capacitance output of the sensor decreased. Additionally, the electrode exhibited good repeatability and the sensitivity of the sensor increased significantly as the testing frequency decreased. Conclusions The capacitive soil moisture sensor holds promise for effective and accurate monitoring of soil moisture levels, with its excellent performance, sensitivity, repeatability, and responsiveness to changes in humidity and soil moisture. The ultimate goal of this study is to achieve long-term monitoring of capacitance changes in capacitive soil moisture sensors, enabling monitoring of long-term changes in soil moisture. This will enable farmers to optimize irrigation systems, improve crop yields, and reduce water usage. In conclusion, the development of this innovative soil moisture sensor has the potential to promote agricultural modernization by providing accurate and reliable monitoring of soil moisture levels.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Image Segmentation Method Combined with VoVNetv2 and Shuffle Attention Mechanism for Fish Feeding in Aquaculture
    WANG Herong, CHEN Yingyi, CHAI Yingqian, XU Ling, YU Huihui
    Smart Agriculture    2023, 5 (4): 137-149.   DOI: 10.12133/j.smartag.SA202310003
    Abstract430)   HTML25)    PDF(pc) (2425KB)(1431)       Save

    [Objective] Intelligent feeding methods are significant for improving breeding efficiency and reducing water quality pollution in current aquaculture. Feeding image segmentation of fish schools is a critical step in extracting the distribution characteristics of fish schools and quantifying their feeding behavior for intelligent feeding method development. While, an applicable approach is lacking due to images challenges caused by blurred boundaries and similar individuals in practical aquaculture environment. In this study, a high-precision segmentation method was proposed for fish school feeding images and provides technical support for the quantitative analysis of fish school feeding behavior. [Methods] The novel proposed method for fish school feeding images segmentation combined VoVNetv2 with an attention mechanism named Shuffle Attention. Firstly, a fish feeding segmentation dataset was presented. The dataset was collected at the intensive aquaculture base of Laizhou Mingbo Company in Shandong province, with a focus on Oplegnathus punctatus as the research target. Cameras were used to capture videos of the fish school before, during, and after feeding. The images were annotated at the pixel level using Labelme software. According to the distribution characteristics of fish feeding and non-feeding stage, the data was classified into two semantic categories— non-occlusion and non-aggregation fish (fish1) and occlusion or aggregation fish (fish2). In the preprocessing stage, data cleaning and image augmentation were employed to further enhance the quality and diversity of the dataset. Initially, data cleaning rules were established based on the distribution of annotated areas within the dataset. Images with outlier annotations were removed, resulting in an improvement in the overall quality of the dataset. Subsequently, to prevent the risk of overfitting, five data augmentation techniques (random translation, random flip, brightness variation, random noise injection, random point addition) were applied for mixed augmentation on the dataset, contributing to an increased diversity of the dataset. Through data augmentation operations, the dataset was expanded to three times its original size. Eventually, the dataset was divided into a training dataset and testing dataset at a ratio of 8:2. Thus, the final dataset consisted of 1 612 training images and 404 testing images. In detail, there were a total of 116 328 instances of fish1 and 20 924 instances of fish2. Secondly, a fish feeding image segmentation method was proposed. Specifically, VoVNetv2 was used as the backbone network for the Mask R-CNN model to extract image features. VoVNetv2 is a backbone network with strong computational capabilities. Its unique feature aggregation structure enables effective fusion of features at different levels, extracting diverse feature representations. This facilitates better capturing of fish schools of different sizes and shapes in fish feeding images, achieving accurate identification and segmentation of targets within the images. To maximize feature mappings with limited resources, the experiment replaced the channel attention mechanism in the one-shot aggregation (OSA) module of VoVNetv2 with a more lightweight and efficient attention mechanism named shuffle attention. This improvement allowed the network to concentrate more on the location of fish in the image, thus reducing the impact of irrelevant information, such as noise, on the segmentation results. Finally, experiments were conducted on the fish segmentation dataset to test the performance of the proposed method. [Results and Discussions] The results showed that the average segmentation accuracy of the Mask R-CNN network reached 63.218% after data cleaning, representing an improvement of 7.018% compared to the original dataset. With both data cleaning and augmentation, the network achieved an average segmentation accuracy of 67.284%, indicating an enhancement of 11.084% over the original dataset. Furthermore, there was an improvement of 4.066% compared to the accuracy of the dataset after cleaning alone. These results demonstrated that data preprocessing had a positive effect on improving the accuracy of image segmentation. The ablation experiments on the backbone network revealed that replacing the ResNet50 backbone with VoVNetv2-39 in Mask R-CNN led to a 2.511% improvement in model accuracy. After improving VoVNetv2 through the Shuffle Attention mechanism, the accuracy of the model was further improved by 1.219%. Simultaneously, the parameters of the model decreased by 7.9%, achieving a balance between accuracy and lightweight design. Comparing with the classic segmentation networks SOLOv2, BlendMask and CondInst, the proposed model achieved the highest segmentation accuracy across various target scales. For the fish feeding segmentation dataset, the average segmentation accuracy of the proposed model surpassed BlendMask, CondInst, and SOLOv2 by 3.982%, 12.068%, and 18.258%, respectively. Although the proposed method demonstrated effective segmentation of fish feeding images, it still exhibited certain limitations, such as omissive detection, error segmentation, and false classification. [Conclusions] The proposed instance segmentation algorithm (SA_VoVNetv2_RCNN) effectively achieved accurate segmentation of fish feeding images. It can be utilized for counting the number and pixel quantities of two types of fish in fish feeding videos, facilitating quantitative analysis of fish feeding behavior. Therefore, this technique can provide technical support for the analysis of piscine feeding actions. In future research, these issues will be addressed to further enhance the accuracy of fish feeding image segmentation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Diagnosis of Grapevine Leafroll Disease Severity Infection via UAV Remote Sensing and Deep Learning
    LIU Yixue, SONG Yuyang, CUI Ping, FANG Yulin, SU Baofeng
    Smart Agriculture    2023, 5 (3): 49-61.   DOI: 10.12133/j.smartag.SA202308013
    Abstract846)   HTML125)    PDF(pc) (3044KB)(1403)       Save

    [Objective] Wine grapes are severely affected by leafroll disease, which affects their growth, and reduces the quality of the color, taste, and flavor of wine. Timely and accurate diagnosis of leafroll disease severity is crucial for preventing and controlling the disease, improving the wine grape fruit quality and wine-making potential. Unmanned aerial vehicle (UAV) remote sensing technology provides high-resolution images of wine grape vineyards, which can capture the features of grapevine canopies with different levels of leafroll disease severity. Deep learning networks extract complex and high-level features from UAV remote sensing images and perform fine-grained classification of leafroll disease infection severity. However, the diagnosis of leafroll disease severity is challenging due to the imbalanced data distribution of different infection levels and categories in UAV remote sensing images. [Method] A novel method for diagnosing leafroll disease severity was developed at a canopy scale using UAV remote sensing technology and deep learning. The main challenge of this task was the imbalanced data distribution of different infection levels and categories in UAV remote sensing images. To address this challenge, a method that combined deep learning fine-grained classification and generative adversarial networks (GANs) was proposed. In the first stage, the GANformer, a Transformer-based GAN model was used, to generate diverse and realistic virtual canopy images of grapevines with different levels of leafroll disease severity. To further analyze the image generation effect of GANformer. The t-distributed stochastic neighbor embedding (t-SNE) to visualize the learned features of real and simulated images. In the second stage, the CA-Swin Transformer, an improved image classification model based on the Swin Transformer and channel attention mechanism was used, to classify the patch images into different classes of leafroll disease infection severity. CA-Swin Transformer could also use a self-attention mechanism to capture the long-range dependencies of image patches and enhance the feature representation of the Swin Transformer model by adding a channel attention mechanism after each Transformer layer. The channel attention (CA) mechanism consisted of two fully connected layers and an activation function, which could extract correlations between different channels and amplify the informative features. The ArcFace loss function and instance normalization layer was also used to enhance the fine-grained feature extraction and downsampling ability for grapevine canopy images. The UAV images of wine grape vineyards were collected and processed into orthomosaic images. They labeled into three categories: healthy, moderate infection, and severe infection using the in-field survey data. A sliding window method was used to extract patch images and labels from orthomosaic images for training and testing. The performance of the improved method was compared with the baseline model using different loss functions and normalization methods. The distribution of leafroll disease severity was mapped in vineyards using the trained CA-Swin Transformer model. [Results and Discussions] The experimental results showed that the GANformer could generate high-quality virtual canopy images of grapevines with an FID score of 93.20. The images generated by GANformer were visually very similar to real images and could produce images with different levels of leafroll disease severity. The T-SNE visualization showed that the features of real and simulated images were well clustered and separated in two-dimensional space, indicating that GANformer learned meaningful and diverse features, which enriched the image dataset. Compared to CNN-based deep learning models, Transformer-based deep learning models had more advantages in diagnosing leafroll disease infection. Swin Transformer achieved an optimal accuracy of 83.97% on the enhanced dataset, which was higher than other models such as GoogLeNet, MobileNetV2, NasNet Mobile, ResNet18, ResNet50, CVT, and T2TViT. It was found that replacing the cross entropy loss function with the ArcFace loss function improved the classification accuracy by 1.50%, and applying instance normalization instead of layer normalization further improved the accuracy by 0.30%. Moreover, the proposed channel attention mechanism, named CA-Swin Transformer, enhanced the feature representation of the Swin Transformer model, achieved the highest classification accuracy on the test set, reaching 86.65%, which was 6.54% higher than using the Swin Transformer on the original test dataset. By creating a distribution map of leafroll disease severity in vineyards, it was found that there was a certain correlation between leafroll disease severity and grape rows. Areas with a larger number of severe leafroll diseases caused by Cabernet Sauvignon were more prone to have missing or weak plants. [Conclusions] A novel method for diagnosing grapevine leafroll disease severity at a canopy scale using UAV remote sensing technology and deep learning was proposed. This method can generate diverse and realistic virtual canopy images of grapevines with different levels of leafroll disease severity using GANformer, and classify them into different classes using CA-Swin Transformer. This method can also map the distribution of leafroll disease severity in vineyards using a sliding window method, and provides a new approach for crop disease monitoring based on UAV remote sensing technology.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    CSD-YOLOv8s: Dense Sheep Small Target Detection Model Based on UAV Images
    WENG Zhi, LIU Haixin, ZHENG Zhiqiang
    Smart Agriculture    2024, 6 (4): 42-52.   DOI: 10.12133/j.smartag.SA202401004
    Abstract1469)   HTML114)    PDF(pc) (1772KB)(1374)       Save

    [Objective] The monitoring of livestock grazing in natural pastures is a key aspect of the transformation and upgrading of large-scale breeding farms. In order to meet the demand for large-scale farms to achieve accurate real-time detection of a large number of sheep, a high-precision and easy-to-deploy small-target detection model: CSD-YOLOv8s was proposed to realize the real-time detection of small-targeted individual sheep under the high-altitude view of the unmanned aerial vehicle (UAV). [Methods] Firstly, a UAV was used to acquire video data of sheep in natural grassland pastures with different backgrounds and lighting conditions, and together with some public datasets downloaded formed the original image data. The sheep detection dataset was generated through data cleaning and labeling. Secondly, in order to solve the difficult problem of sheep detection caused by dense flocks and mutual occlusion, the SPPFCSPC module was constructed with cross-stage local connection based on the you only look once (YOLO)v8 model, which combined the original features with the output features of the fast spatial pyramid pooling network, fully retained the feature information at different stages of the model, and effectively solved the problem of small targets and serious occlusion of the sheep, and improved the detection performance of the model for small sheep targets. In the Neck part of the model, the convolutional block attention module (CBAM) convolutional attention module was introduced to enhance the feature information capture based on both spatial and channel aspects, suppressing the background information spatially and focusing on the sheep target in the channel, enhancing the network's anti-jamming ability from both channel and spatial dimensions, and improving the model's detection performance of multi-scale sheep under complex backgrounds and different illumination conditions. Finally, in order to improve the real-time and deploy ability of the model, the standard convolution of the Neck network was changed to a lightweight convolutional C2f_DS module with a changeable kernel, which was able to adaptively select the corresponding convolutional kernel for feature extraction according to the input features, and solved the problem of input scale change in the process of sheep detection in a more flexible way, and at the same time, the number of parameters of the model was reduced and the speed of the model was improved. [Results and Discussions] The improved CSD-YOLOv8s model exhibited excellent performance in the sheep detection task. Compared with YOLO, Faster R-CNN and other classical network models, the improved CSD-YOLOv8s model had higher detection accuracy and frames per second (FPS) of 87 f/s in the flock detection task with comparable detection speed and model size. Compared with the YOLOv8s model, Precision was improved from 93.0% to 95.2%, mAP was improved from 91.2% to 93.1%, and it had strong robustness to sheep targets with different degree of occlusion and different scales, which effectively solved the serious problems of missed and misdetection of sheep in the grassland pasture UAV-on-ground sheep detection task due to the small sheep targets, large background noise, and high degree of densification. misdetection serious problems. Validated by the PASCAL VOC 2007 open dataset, the CSD-YOLOv8s model proposed in this study improved the detection accuracy of 20 different objects, including transportation vehicles, animals, etc., especially in sheep detection, the detection accuracy was improved by 9.7%. [Conclusions] This study establishes a sheep dataset based on drone images and proposes a model called CSD-YOLOv8s for detecting grazing sheep in natural grasslands. The model addresses the serious issues of missed detections and false alarms in sheep detection under complex backgrounds and lighting conditions, enabling more accurate detection of grazing livestock in drone images. It achieves precise detection of targets with varying degrees of clustering and occlusion and possesses good real-time performance. This model provides an effective detection method for detecting sheep herds from the perspective of drones in natural pastures and offers technical support for large-scale livestock detection in breeding farms, with wide-ranging potential applications.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Tea Shoot Picking Point Recognition Model Based on Improved DeepLabV3+
    HU Chengxi, TAN Lixin, WANG Wenyin, SONG Min
    Smart Agriculture    2024, 6 (5): 119-127.   DOI: 10.12133/j.smartag.SA202403016
    Abstract423)   HTML44)    PDF(pc) (1379KB)(1361)       Save

    [Objective] The picking of famous and high-quality tea is a crucial link in the tea industry. Identifying and locating the tender buds of famous and high-quality tea for picking is an important component of the modern tea picking robot. Traditional neural network methods suffer from issues such as large model size, long training times, and difficulties in dealing with complex scenes. In this study, based on the actual scenario of the Xiqing Tea Garden in Hunan Province, proposes a novel deep learning algorithm was proposed to solve the precise segmentation challenge of famous and high-quality tea picking points. [Methods] The primary technical innovation resided in the amalgamation of a lightweight network architecture, MobilenetV2, with an attention mechanism known as efficient channel attention network (ECANet), alongside optimization modules including atrous spatial pyramid pooling (ASPP). Initially, MobilenetV2 was employed as the feature extractor, substituting traditional convolution operations with depth wise separable convolutions. This led to a notable reduction in the model's parameter count and expedited the model training process. Subsequently, the innovative fusion of ECANet and ASPP modules constituted the ECA_ASPP module, with the intention of bolstering the model's capacity for fusing multi-scale features, especially pertinent to the intricate recognition of tea shoots. This fusion strategy facilitated the model's capability to capture more nuanced features of delicate shoots, thereby augmenting segmentation accuracy. The specific implementation steps entailed the feeding of image inputs through the improved network, whereupon MobilenetV2 was utilized to extract both shallow and deep features. Deep features were then fused via the ECA_ASPP module for the purpose of multi-scale feature integration, reinforcing the model's resilience to intricate backgrounds and variations in tea shoot morphology. Conversely, shallow features proceeded directly to the decoding stage, undergoing channel reduction processing before being integrated with upsampled deep features. This divide-and-conquer strategy effectively harnessed the benefits of features at differing levels of abstraction and, furthermore, heightened the model's recognition performance through meticulous feature fusion. Ultimately, through a sequence of convolutional operations and upsampling procedures, a prediction map congruent in resolution with the original image was generated, enabling the precise demarcation of tea shoot harvesting points. [Results and Discussions] The experimental outcomes indicated that the enhanced DeepLabV3+ model had achieved an average Intersection over Union (IoU) of 93.71% and an average pixel accuracy of 97.25% on the dataset of tea shoots. Compared to the original model based on Xception, there was a substantial decrease in the parameter count from 54.714 million to a mere 5.818 million, effectively accomplishing a significant lightweight redesign of the model. Further comparisons with other prevalent semantic segmentation networks revealed that the improved model exhibited remarkable advantages concerning pivotal metrics such as the number of parameters, training duration, and average IoU, highlighting its efficacy and precision in the domain of tea shoot recognition. This considerable decreased in parameter numbers not only facilitated a more resource-economical deployment but also led to abbreviated training periods, rendering the model highly suitable for real-time implementations amidst tea garden ecosystems. The elevated mean IoU and pixel accuracy attested to the model's capacity for precise demarcation and identification of tea shoots, even amidst intricate and varied datasets, demonstrating resilience and adaptability in pragmatic contexts. [Conclusions] This study effectively implements an efficient and accurate tea shoot recognition method through targeted model improvements and optimizations, furnishing crucial technical support for the practical application of intelligent tea picking robots. The introduction of lightweight DeepLabV3+ not only substantially enhances recognition speed and segmentation accuracy, but also mitigates hardware requirements, thereby promoting the practical application of intelligent picking technology in the tea industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Collaborative Computing of Food Supply Chain Privacy Data Elements Based on Federated Learning
    XU Jiping, LI Hui, WANG Haoyu, ZHOU Yan, WANG Zhaoyang, YU Chongchong
    Smart Agriculture    2023, 5 (4): 79-91.   DOI: 10.12133/j.smartag.SA202309012
    Abstract457)   HTML30)    PDF(pc) (1719KB)(1359)       Save

    [Objective] The flow of private data elements plays a crucial role in the food supply chain, and the safe and efficient operation of the food supply chain can be ensured through the effective management and flow of private data elements. Through collaborative computing among the whole chain of the food supply chain, the production, transportation and storage processes of food can be better monitored and managed, so that possible quality and safety problems can be detected and solved in a timely manner, and the health and rights of consumers can be safeguarded. It can also be applied to the security risk assessment and early warning of the food supply chain. By analyzing big data, potential risk factors and abnormalities can be identified, and timely measures can be taken for early warning and intervention to reduce the possibility of quality and safety risks. This study combined the industrial Internet identification and resolution system with the federated learning algorithm, which can realize collaborative learning among multiple enterprises, and each enterprise can carry out collaborative training of the model without sharing the original data, which protects the privacy and security of the data while realizing the flow of the data, and it can also make use of the data resources distributed in different segments, which can realize more comprehensive and accurate collaborative calculations, and improve the safety and credibility of the industrial Internet system's security and credibility. [Methods] To address the problem of not being able to share and participate in collaborative computation among different subjects in the grain supply chain due to the privacy of data elements, this study first analyzed and summarized the characteristics of data elements in the whole link of grain supply chain, and proposed a grain supply chain data flow and collaborative computation architecture based on the combination of the industrial Internet mark resolution technology and the idea of federated learning, which was constructed in a layered and graded model to provide a good infrastructure for the decentralized between the participants. The data identification code for the flow of food supplied chain data elements and the task identification code for collaborative calculation of food supply chain, as well as the corresponding parameter data model, information data model and evaluation data model, were designed to support the interoperability of federated learning data. A single-link horizontal federation learning model with isomorphic data characteristics of different subjects and a cross-link vertical federation learning model with heterogeneous data characteristics were constructed, and the model parameters were quickly adjusted and calculated based on logistic regression algorithm, neural network algorithm and other algorithms, and the food supply chain security risk assessment scenario was taken as the object of the research, and the research was based on the open source FATE (Federated AI Technology) federation learning model. Enabler (Federated AI Technology) federated learning platform for testing and validation, and visualization of the results to provide effective support for the security management of the grain supply chain. [Results and Discussion] Compared with the traditional single-subject assessment calculation method, the accuracy of single-session isomorphic horizontal federation learning model assessment across subjects was improved by 6.7%, and the accuracy of heterogeneous vertical federation learning model assessment across sessions and subjects was improved by 8.3%. This result showed that the single-session isomorphic horizontal federated learning model assessment across subjects could make full use of the data information of each subject by merging and training the data of different subjects in the same session, thus improving the accuracy of security risk assessment. The heterogeneous vertical federated learning model assessment of cross-session and cross-subject further promotes the application scope of collaborative computing by jointly training data from different sessions and subjects, which made the results of safety risk assessment more comprehensive and accurate. The advantage of combining federated learning and logo resolution technology was that it could conduct model training without sharing the original data, which protected data privacy and security. At the same time, it could also realize the effective use of data resources and collaborative computation, improving the efficiency and accuracy of the assessment process. [Conclusions] The feasibility and effectiveness of this study in practical applications in the grain industry were confirmed by the test validation of the open-source FATE federated learning platform. This provides reliable technical support for the digital transformation of the grain industry and the security management of the grain supply chain, and helps to improve the intelligence level and competitiveness of the whole grain industry. Therefore, this study can provide a strong technical guarantee for realizing the safe, efficient and sustainable development of the grain supply chain.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Traceability Model of Plantation Agricultural Products Based on Blockchain and InterPlanetary File System
    CHEN Dandan, ZHANG Lijie, JIANG Shuangfeng, ZHANG En, ZHANG Jie, ZHAO Qing, ZHENG Guoqing, LI Guoqiang
    Smart Agriculture    2023, 5 (4): 68-78.   DOI: 10.12133/j.smartag.SA202307004
    Abstract494)   HTML37)    PDF(pc) (2011KB)(1345)       Save

    [Objective] The InterPlanetary File System (IPFS) is a peer-to-peer distributed file system, aiming to establish a global, open, and decentralized network for storage and sharing. Combining the IPFS and blockchain technology could alleviate the pressure on blockchain storage. The distinct features of the supply chain for agricultural products in the plantation industry, including extended production cycles, multiple, heterogeneous data sources, and relatively fragmented production, which can readily result in information gaps and opacity throughout the supply chain; in the traceability process of agricultural products, there are issues with sensitive data being prone to leakage and a lack of security, and the supply chain of plantation agricultural products is long, and the traceability data is often stored in multiple blocks, which requires frequent block tracing operations during tracing, resulting in low efficiency. Consequently, the aim of this study is to fully encapsulate the decentralized nature of blockchain, safeguard the privacy of sensitive data, and alleviate the storage strain of blockchain. [Method] A traceability model for plantation-based agricultural products was developed, leveraging the hyperledger fabric consortium chain and the IPFS. Based on data type, traceability data was categorized into structured and unstructured data. Given that blockchain ledgers were not optimized for direct storage of unstructured data, such as images and videos, to alleviate the storage strain on the blockchain, unstructured data was persisted in the IPFS, while structured data remains within the blockchain ledger. Based on data privacy categories, traceability data was categorized into public data and sensitive data. Public data was stored in the public ledger of hyperledger fabric, while sensitive data was stored in the private data collection of hyperledger fabric. This method allowed for efficient data access while maintaining data security, enhancing the efficiency of traceability. Hyperledger Fabric was the foundational platform for the development of the prototype system. The front-end website was based on the TCP/IP protocol stack. The website visualization was implemented through the React framework. Smart contracts were crafted using the Java programming language. The performance of the application layer interface was tested using the testing tool Postman. [Conclusions and Discussions] The blockchain-based plantation agricultural product traceability system was structured into a five-tiered architecture, starting from the top: the application layer, gateway layer, contract layer, consensus layer, and data storage layer. The primary service providers at the application layer were the enterprises and consumers involved in each stage of the traceability process. The gateway layer served as the middleware between users and the blockchain, primarily providing interface support for the front-end interface of the application layer. The contract layer mainly included smart contracts for planting, processing, warehousing, transportation, and sales. The consensus layer used the EtcdRaft consensus algorithm. The data storage layer was divided into the on-chain storage layer of the blockchain ledger and the off-chain storage layer of the IPFS cluster. In terms of data types, each piece of traceability data was categorized into structured data items and unstructured data items. Unstructured data was stored in the Interstellar File System cluster, and the returned content identifiers were integrated with the structured data items into the blockchain nodes within the traceability system. In the realm of data privacy, smart contracts were employed to segregate public and sensitive data, with public data directly integrating onto the blockchain, and sensitive data, adhering to predefined sharing policies, being stored in a private dataset designated by hyperledger fabric. In terms of user queries, consumers could retrieve product traceability information via a traceability system overseen by a reputable authority. The developed model website consisted of three parts: a login section, an agricultural product circulation information management and user data management section for enterprises in various links, and a traceability data query section for consumers. When using synchronous and asynchronous Application Program Interfaces, the average data on-chain latency was 2 138.9 and 37.6 ms, respectively, and the average data query latency was 12.3 ms. Blockchain, as the foundational data storage technology, enhances the credibility and transaction efficiency in agricultural product traceability. [Conclusions] This study designed and implemented a plantation agricultural product traceability model leveraging blockchain technology's private dataset and the IPFS cluster. This model ensured secure sharing and storage of traceability data, particularly sensitive data, across all stages. Compared to traditional centralized traceability models, it enhanced the reliability of the traceability data. Based on the evaluation through experimental systems, the traceability model proposed in this study effectively safeguarded the privacy of sensitive data in enterprises. Additionally, it offered high efficiency in data linking and querying. Applicable to the real-world traceability environment of plantation agricultural products, it showed potential for widespread application and promotion, offering fresh insights for designing blockchain traceability models in this sector. The model is still in its experimental phase and lacks applications across various types of crops in the farming industry. The subsequent step is to apply the model in real-world scenarios, continually enhance its efficiency, refine the model, advance the practical application of blockchain technology, and lay the foundation for agricultural modernization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Identification and Severity Classification of Typical Maize Foliar Diseases Based on Hyperspectral Data
    SHEN Yanyan, ZHAO Yutao, CHEN Gengshen, LYU Zhengang, ZHAO Feng, YANG Wanneng, MENG Ran
    Smart Agriculture    2024, 6 (2): 28-39.   DOI: 10.12133/j.smartag.SA202310016
    Abstract1357)   HTML83)    PDF(pc) (1519KB)(1345)       Save

    [Objective] In recent years, there has been a significant increase in the severity of leaf diseases in maize, with a noticeable trend of mixed occurrence. This poses a serious threat to the yield and quality of maize. However, there is a lack of studies that combine the identification of different types of leaf diseases and their severity classification, which cannot meet the needs of disease prevention and control under the mixed occurrence of different diseases and different severities in actual maize fields. [Methods] A method was proposed for identifying the types of typical leaf diseases in maize and classifying their severity using hyperspectral technology. Hyperspectral data of three leaf diseases of maize: northern corn leaf blight (NCLB), southern corn leaf blight (SCLB) and southern corn rust (SCR), were obtained through greenhouse pathogen inoculation and natural inoculation. The spectral data were preprocessed by spectral standardization, SG filtering, sensitive band extraction and vegetation index calculation, to explore the spectral characteristics of the three leaf diseases of maize. Then, the inverse frequency weighting method was utilized to balance the number of samples to reduce the overfitting phenomenon caused by sample imbalance. Relief-F and variable selection using random forests (VSURF) method were employed to optimize the sensitive spectral features, including band features and vegetation index features, to construct models for disease type identification based on the full stages of disease development (including all disease severities) and for individual disease severities using several representative machine learning approaches, demonstrating the effectiveness of the research method. Furthermore, the study individual occurrence severity classification models were also constructed for each single maize leaf disease, including the NCLB, SCLB and SCR severity classification models, respectively, aiming to achieve full-process recognition and disease severity classification for different leaf diseases. Overall accuracy (OA) and Macro F1 were used to evaluate the model accuracy in this study. Results and Discussion The research results showed significant spectrum changes of three kinds of maize leaf diseases primarily focusing on the visible (550-680 nm), red edge (740-760 nm), near-infrared (760-1 000 nm) and shortwave infrared (1 300-1 800 nm) bands. Disease-specific spectral features, optimized based on disease spectral response rules, effectively identified disease species and classify their severity. Moreover, vegetation index features were more effective in identifying disease-specific information than sensitive band features. This was primarily due to the noise and information redundancy present in the selected hyperspectral sensitive bands, whereas vegetation index could reduce the influence of background and atmospheric noise to a certain extent by integrating relevant spectral signals through band calculation, so as to achieve higher precision in the model. Among several machine learning algorithms, the support vector machine (SVM) method exhibited better robustness than random forest (RF) and decision tree (DT). In the full stage of disease development, the optimal overall accuracy (OA) of the disease classification model constructed by SVM based on vegetation index reached 77.51%, with a Macro F1 of 0.77, representing a 28.75% increase in OA and 0.30 higher of Macro F1 compared to the model based on sensitive bands. Additionally, the accuracy of the disease classification model with a single severity of the disease increased with the severity of the disease. The accuracy of disease classification during the early stage of disease development (OA=70.31%) closely approached that of the full disease development stage (OA=77.51%). Subsequently, in the moderate disease severity stage, the optimal accuracy of disease classification (OA=80.00%) surpassed the optimal accuracy of disease classification in the full disease development stage. Furthermore, the optimal accuracy of disease classification under severe severity reached 95.06%, with a Macro F1 of 0.94. This heightened accuracy during the severity stage can be attributed to significant changes in pigment content, water content and cell structure of the diseased leaves, intensifying the spectral response of each disease and enhancing the differentiation between different diseases. In disease severity classification model, the optimal accuracy of the three models for maize leaf disease severity all exceeded 70%. Among the three kinds of disease severity classification results, the NCLB severity classification model exhibited the best performance. The NCLB severity classification model, utilizing SVM based on the optimal vegetation index features, achieved an OA of 86.25%, with a Macro F1 of 0.85. In comparison, the accuracy of the SCLB severity classification model (OA=70.35%, Macro F1=0.70) and SCR severity classification model (OA=71.39%, Macro F1=0.69) were lower than that of NCLB. [Conclusions] The aforementioned results demonstrate the potential to effectively identify and classify the types and severity of common leaf diseases in maize using hyperspectral data. This lays the groundwork for research and provides a theoretical basis for large-scale crop disease monitoring, contributing to precision prevention and control as well as promoting green agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Using a Portable Visible-near Infrared Spectrometer and Machine Learning to Distinguish and Quantify Mold Contamination in Wheat
    JIA Wenshen, LYU Haolin, ZHANG Shang, QIN Yingdong, ZHOU Wei
    Smart Agriculture    2024, 6 (1): 89-100.   DOI: 10.12133/j.smartag.SA202311032
    Abstract432)   HTML24)    PDF(pc) (2050KB)(1344)       Save

    Objective Traditional methods for detecting mold are time-consuming, labor-intensive, and vulnerable to environmental influences, highlighting the need for a swift, precise, and dependable detection approach. Researchers have utilized visible-near infrared (NIR) spectroscopy for the non-destructive, rapid assessment of wheat moisture content, crude protein content, concealed pests, starch content, dry matter, weight, hardness, origin, and other attributes. However, most of these studies rely on research-grade Visible-NIR spectrometers typically found in laboratories. While these spectrometers offer superior detection accuracy and stability, their bulky size, lack of portability, and high cost hinder their widespread use and adoption across various agricultural product distribution channels. Methods A low-resolution Visible-NIR spectrometer (VNIAPD, with a resolution of 1.6 nm) was utilized to gather wheat data. The aim was to enhance the accuracy of moldy wheat detection by identifying suitable spectral data preprocessing methods using corresponding algorithms. A high-resolution Visible-NIR spectrometer (SINO2040, with a resolution of 0.19 nm) served as a control to validate the instrument and method's effectiveness. The Zhoumai (No. 22) wheat variety was adopted, with a total of 100 samples prepared. The spectra of fresh wheat were scanned and then placed in a constant temperature chamber at 35 °C to replicate the appropriate conditions for mold growth, thereby accelerating the reproduction of naturally occurring mold in the wheat. The degree of mold was categorized based on the cultivation time in the constant temperature chamber, with wheat classified as mildly, moderately, or severely moldy after 3, 6, and 9 days of cultivation, respectively. A total of 400 wheat spectral data points were collected, including 100 samples each of fresh wheat, wheat cultured for 3 days, wheat cultured for 6 days, and wheat cultured for 9 days. Preprocessing methods such as standard deviation normalization (SDN), standard normal variation (SNV), mean centrality (MC), first-order derivatives (1ST), Savitzky-Golay smoothing (SG), and multiple scattering correction (MSC) were applied to the spectral data. Outliers were identified and eliminated using the local outlier factor (LOF) method. Following this, the sequential projection algorithm (SPA) and Least absolute shrinkage and selection operator (LASSO) were used to extract characteristic wavelengths from the preprocessed spectra. Subsequently, six algorithms, including k-nearest neighbors (KNN), support vector machines (SVM), random forests (RF), Naïve-Bayes, back propagation neural networks (BPNN), and deep neural networks (DNN), were employed to model and analyze the feature wavelength spectra, differentiating moldy wheat and classifying the degree of mold. Evaluation criteria encompassed accuracy, modeling time, and model size to aid in selecting the most suitable model for specific application scenarios. Results and discussions Regarding accuracy, even when utilizing the computationally slower and more memory-demanding neural network models BPNN and DNN, both the VNIAPD and SINO2040 achieved a perfect 100% accuracy in the binary classification task of distinguishing between fresh and moldy wheat. They also maintained a faultless 100% accuracy in the ternary classification task that differentiates three varying levels of mold growth. Adopting faster and more memory-efficient shallow models such as KNN, SVM, RF, and Naïve-Bayes, the VNIAPD yielded a top test set accuracy of 97.72% when combined with RF for binary classification. Conversely, SINO2040 achieved 100% accuracy using Naïve-Bayes. In the ternary classification scenario, the VNIAPD hit the mark at 100% accuracy with both KNN and RF, while SINO2040 demonstrated 97.72% accuracy with KNN and SVM. Regarding modeling speed, the shallow machine learning algorithms, including KNN, SVM, RF, and Naïve-Bayes, exhibited quicker training times, with Naïve-Bayes being the swiftest at just 3 ms. In contrast, the neural network algorithms BPNN and DNN required more time for training, taking 3 293 and 18 614 ms, respectively. Regarding memory footprint, BPNN had the largest model size, occupying 4 028 kb, whereas SVM was the most memory-efficient, with a size of only 4 kb. Overall, the VNIAPD matched the SINO2040 in detection accuracy despite having lower optical parameters: A slightly lesser optical resolution of 1.6 nm compared to the SINO2040's 0.19 nm—and a lower cost, highlighting its efficiency and cost-effectiveness in the given context. Conclusions In this study, by comparing different preprocessing methods for spectral data, the optimal data optimization choices for corresponding algorithms were identified. As a result, the low-resolution spectrometer VNIAPD was able to achieve performance on par with the high-resolution spectrometer SINO2040 in detecting moldy wheat, providing a new option for low-cost, non-destructive detection of wheat mold and the degree of moldiness based on Visible-NIR spectroscopy.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automated Flax Seeds Testing Methods Based on Machine Vision
    MAO Yongwen, HAN Junying, LIU Chengzhong
    Smart Agriculture    2024, 6 (1): 135-146.   DOI: 10.12133/j.smartag.SA202309011
    Abstract532)   HTML57)    PDF(pc) (1671KB)(1290)       Save

    Objective Flax, characterized by its short growth cycle and strong adaptability, is one of the major cash crops in northern China. Due to its versatile uses and unique quality, it holds a significant position in China's oil and fiber crops. The quality of flax seeds directly affects the yield of the flax plant. Seed evaluation is a crucial step in the breeding process of flax. Common parameters used in the seed evaluation process of flax include circumference, area, length axis, and 1 000-seed weight. To ensure the high-quality production of flax crops, it is of great significance to understand the phenotypic characteristics of flax seeds, select different resources as parents based on breeding objectives, and adopt other methods for the breeding, cultivation, and evaluation of seed quality and traits of flax. Methods In response to the high error rates and low efficiency issues observed during the automated seed testing of flax seeds, the measurement methods were explored of flax seed contours based on machine vision research. The flax seed images were preprocessed, and the collected color images were converted to grayscale. A filtering and smoothing process was applied to obtain binary images. To address the issues of flax seed overlap and adhesion, a contour fitting image segmentation method based on fused corner features was proposed. This method incorporated adaptive threshold selection during edge detection of the image contour. Only multi-seed target areas that met certain criteria were subjected to image segmentation processing, while single-seed areas bypassed this step and were directly summarized for seed testing data. After obtaining the multi-seed adhesion target areas, the flax seeds underwent contour approximation, corner extraction, and contour fitting. Based on the provided image contour information, the image contour shape was approximated to another contour shape with fewer vertices, and the original contour curve was simplified to a more regular and compact line segment or polygon, minimizing computational complexity. All line shape characteristics in the image were marked as much as possible. Since the pixel intensity variations in different directions of image corners were significant, the second derivative matrix based on pixel grayscale values was used to detect image corners. Based on the contour approximation algorithm, contour corner detection was performed to obtain the coordinates of each corner. The resulting contour points and corners were used as outputs to further improve the accuracy and precision of subsequent contour fitting methods, resulting in a two-dimensional discrete point dataset of the image contour. Using the contour point dataset as an input, the geometric moments of the image contour were calculated, and the optimal solution for the ellipse parameters was obtained through numerical optimization based on the least squares method and the geometric features of the ellipse shape. Ultimately, the optimal contour was fitted to the given image, achieving the segmentation and counting of flax seed images. Meanwhile, each pixel in the digital image was a uniform small square in size and shape, so the circumference, area, and major and minor axes of the flax seeds could be represented by the total number of pixels occupied by the seeds in the image. The weight of a single seed could be calculated by dividing the total weight of the seeds by the total number of seeds detected by the contour, thereby obtaining the weight of the individual seed and converting it accordingly. Through the pixelization of the 1 yuan and 1 jiao coins from the fifth iteration of the 2019 Renminbi, a summary of the circumference, area, major axis, minor axis, and 1 000-seed weight of the flax seeds was achieved. Additionally, based on the aforementioned method, this study designed an automated real-time analysis system for flax seed testing data, realizing the automation of flax seed testing research. Experiments were conducted on images of flax seeds captured by an industrial camera. Results and Discussions The proposed automated seed identification method achieved an accuracy rate of 97.28% for statistically distinguishing different varieties of flax seeds. The average processing time for 100 seeds was 69.58 ms. Compared to the extreme erosion algorithm and the watershed algorithm based on distance transformation, the proposed method improved the average calculation accuracy by 19.6% over the extreme erosion algorithm and required a shorter average computation time than the direct use of the watershed algorithm. Considering the practical needs of automated seed identification, this method did not employ methods such as dilation or erosion for image morphology processing, thereby preserving the original features of the image to the greatest extent possible. Additionally, the flax seed automated seed identification data real-time analysis system could process image information in batches. By executing data summarization functions, it automatically generated corresponding data table folders, storing the corresponding image data summary tables. Conclusions The proposed method exhibits superior computational accuracy and processing speed, with shorter operation time and robustness. It is highly adaptable and able to accurately acquire the morphological feature parameters of flax seeds in bulk, ensuring measurement errors remain within 10%, which could provide technical support for future flax seed evaluation and related industrial development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Trajectory Tracking Method of Agricultural Machinery Multi-Robot Formation Operation Based on MPC Delay Compensator
    LUAN Shijie, SUN Yefeng, GONG Liang, ZHANG Kai
    Smart Agriculture    2024, 6 (3): 69-81.   DOI: 10.12133/j.smartag.SA202306013
    Abstract463)   HTML34)    PDF(pc) (1797KB)(1287)       Save

    [Objective] The technology of multi-machine convoy driving has emerged as a focal point in the field of agricultural mechanization. By organizing multiple agricultural machinery units into convoys, unified control and collaborative operations can be achieved. This not only enhances operational efficiency and reduces costs, but also minimizes human labor input, thereby maximizing the operational potential of agricultural machinery. In order to solve the problem of communication delay in cooperative control of multi-vehicle formation and its compensation strategy, the trajectory control method of multi-vehicle formation was proposed based on model predictive control (MPC) delay compensator. [Methods] The multi-vehicle formation cooperative control strategy was designed, which introduced the four-vehicle formation cooperative scenario in three lanes, and then introduced the design of the multi-vehicle formation cooperative control architecture, which was respectively enough to establish the kinematics and dynamics model and equations of the agricultural machine model, and laied down a sturdy foundation for solving the formation following problem later. The testing and optimization of automatic driving algorithms based on real vehicles need to invest too much time and economic costs, and were subject to the difficulties of laws and regulations, scene reproduction and safety, etc. Simulation platform testing could effectively solve the above question. For the agricultural automatic driving multi-machine formation scenarios, the joint simulation platform Carsim and Simulink were used to simulate and validate the formation driving control of agricultural machines. Based on the single-machine dynamics model of the agricultural machine, a delay compensation controller based on MPC was designed. Feedback correction first detected the actual output of the object and then corrected the model-based predicted output with the actual output and performed a new optimization. Based on the above model, the nonlinear system of kinematics and dynamics was linearized and discretized in order to ensure the real-time solution. The objective function was designed so that the agricultural machine tracks on the desired trajectory as much as possible. And because the operation range of the actuator was limited, the control increment and control volume were designed with corresponding constraints. Finally, the control increment constraints were solved based on the front wheel angle constraints, front wheel angle increments, and control volume constraints of the agricultural machine. [Results and Discussions] Carsim and MATLAB/Simulink could be effectively compatible, enabling joint simulation of software with external solvers. When the delay step size d=5 was applied with delay compensation, the MPC response was faster and smoother; the speed error curve responded faster and gradually stabilized to zero error without oscillations. Vehicle 1 effectively changed lanes in a short time and maintains the same lane as the lead vehicle. In the case of a longer delay step size d =10, controllers without delay compensation showed more significant performance degradation. Even under higher delay conditions, MPC with delay compensation applied could still quickly respond with speed error and longitudinal acceleration gradually stabilizing to zero error, avoiding oscillations. The trajectory of Vehicle 1 indicated that the effectiveness of the delay compensation mechanism decreased under extreme delay conditions. The simulation results validated the effectiveness of the proposed formation control algorithm, ensuring that multiple vehicles could successfully change lanes to form queues while maintaining specific distances and speeds. Furthermore, the communication delay compensation control algorithm enables vehicles with added delay to effectively complete formation tasks, achieving stable longitudinal and lateral control. This confirmed the feasibility of the model predictive controller with delay compensation proposed. [Conclusions] At present, most of the multi-machine formation coordination is based on simulation platform for verification, which has the advantages of safety, economy, speed and other aspects, however, there's still a certain gap between the idealized model in the simulation platform and the real machine experiment. Therefore, multi-machine formation operation of agricultural equipment still needs to be tested on real machines under sound laws and regulations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Rapid Detection Method for Wheat Seedling Leaf Number in Complex Field Scenarios Based on Improved YOLOv8
    HOU Yiting, RAO Yuan, SONG He, NIE Zhenjun, WANG Tan, HE Haoxu
    Smart Agriculture    2024, 6 (4): 128-137.   DOI: 10.12133/j.smartag.SA202403019
    Abstract607)   HTML104)    PDF(pc) (2913KB)(1255)       Save

    [Objective] The enumeration of wheat leaves is an essential indicator for evaluating the vegetative state of wheat and predicting its yield potential. Currently, the process of wheat leaf counting in field settings is predominantly manual, characterized by being both time-consuming and labor-intensive. Despite advancements, the efficiency and accuracy of existing automated detection and counting methodologies have yet to satisfy the stringent demands of practical agricultural applications. This study aims to develop a method for the rapid quantification of wheat leaves to refine the precision of wheat leaf tip detection. [Methods] To enhance the accuracy of wheat leaf detection, firstly, an image dataset of wheat leaves across various developmental stages—seedling, tillering, and overwintering—under two distinct lighting conditions and using visible light images sourced from both mobile devices and field camera equipmen, was constructed. Considering the robust feature extraction and multi-scale feature fusion capabilities of YOLOv8 network, the foundational architecture of the proposed model was based on the YOLOv8, to which a coordinate attention mechanism has been integrated. To expedite the model's convergence, the loss functions were optimized. Furthermore, a dedicated small object detection layer was introduced to refine the recognition of wheat leaf tips, which were typically difficult for conventional models to discern due to their small size and resemblance to background elements. This deep learning network was named as YOLOv8-CSD, tailored for the recognition of small targets such as wheat leaf tips, ascertains the leaf count by detecting the number of leaf tips present within the image. A comparative analysis was conducted on the YOLOv8-CSD model in comparison with the original YOLOv8 and six other prominent network architectures, including Faster R-CNN, Mask R-CNN, YOLOv7, and SSD, within a uniform training framework, to evaluate the model's effectiveness. In parallel, the performance of both the original and YOLOv8-CSD models was assessed under challenging conditions, such as the presence of weeds, occlusions, and fluctuating lighting, to emulate complex real-world scenarios. Ultimately, the YOLOv8-CSD model was deployed for wheat leaf number detection in intricate field conditions to confirm its practical applicability and generalization potential. [Results and Discussions] The research presented a methodology that achieved a recognition precision of 91.6% and an mAP0.5 of 85.1% for wheat leaf tips, indicative of its robust detection capabilities. This method exceled in adaptability within complex field environments, featuring an autonomous adjustment mechanism for different lighting conditions, which significantly enhanced the model's robustness. The minimal rate of missed detections in wheat seedlings' leaf counting underscored the method's suitability for wheat leaf tip recognition in intricate field scenarios, consequently elevating the precision of wheat leaf number detection. The sophisticated algorithm embedded within this model had demonstrated a heightened capacity to discern and focus on the unique features of wheat leaf tips during the detection process. This capability was essential for overcoming challenges such as small target sizes, similar background textures, and the intricacies of feature extraction. The model's consistent performance across diverse conditions, including scenarios with weeds, occlusions, and fluctuating lighting, further substantiated its robustness and its readiness for real-world application. [Conclusions] This research offers a valuable reference for accurately detecting wheat leaf numbers in intricate field conditions, as well as robust technical support for the comprehensive and high-quality assessment of wheat growth.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Image Segmentation Method of Chinese Yam Leaves in Complex Background Based on Improved ENet
    LU Bibo, LIANG Di, YANG Jie, SONG Aiqing, HUANGFU Shangwei
    Smart Agriculture    2024, 6 (6): 109-120.   DOI: 10.12133/j.smartag.SA202407007
    Abstract240)   HTML23)    PDF(pc) (2024KB)(1255)       Save

    [Objective] Crop leaf area is an important indicator reflecting light absorption efficiency and growth conditions. This paper established a diverse Chinese yam image dataset and proposesd a deep learning-based method for Chinese yam leaf image segmentation. This method can be used for real-time measurement of Chinese yam leaf area, addressing the inefficiency of traditional measurement techniques. This will provide more reliable data support for genetic breeding, growth and development research of Chinese yam, and promote the development and progress of the Chinese yam industry. [Methods] A lightweight segmentation network based on improved ENet was proposed. Firstly, based on ENet, the third stage was pruned to reduce redundant calculations in the model. This improved the computational efficiency and running speed, and provided a good basis for real-time applications. Secondly, PConv was used instead of the conventional convolution in the downsampling bottleneck structure and conventional bottleneck structure, the improved bottleneck structure was named P-Bottleneck. PConv applied conventional convolution to only a portion of the input channels and left the rest of the channels unchanged, which reduced memory accesses and redundant computations for more efficient spatial feature extraction. PConv was used to reduce the amount of model computation while increase the number of floating-point operations per second on the hardware device, resulting in lower latency. Additionally, the transposed convolution in the upsampling module was improved to bilinear interpolation to enhance model accuracy and reduce the number of parameters. Bilinear interpolation could process images smoother, making the processed images more realistic and clear. Finally, coordinate attention (CA) module was added to the encoder to introduce the attention mechanism, and the model was named CBPA-ENet. The CA mechanism not only focused on the channel information, but also keenly captured the orientation and position-sensitive information. The position information was embedded into the channel attention to globally encode the spatial information, capturing the channel information along one spatial direction while retaining the position information along the other spatial direction. The network could effectively enhance the attention to important regions in the image, and thus improve the quality and interpretability of segmentation results. [Results and Discussions] Trimming the third part resulted in a 28% decrease in FLOPs, a 41% decrease in parameters, and a 9 f/s increase in FPS. Improving the upsampling method to bilinear interpolation not only reduces the floating-point operation and parameters, but also slightly improves the segmentation accuracy of the model, increasing FPS by 4 f/s. Using P-Bottleneck instead of downsampling bottleneck structure and conventional bottleneck structure can reduce mIoU by only 0.04%, reduce FLOPs by 22%, reduce parameters by 16%, and increase FPS by 8 f/s. Adding CA mechanism to the encoder could only increase a small amount of FLOPs and parameters, improving the accuracy of the segmentation network. To verify the effectiveness of the improved segmentation algorithm, classic semantic segmentation networks of UNet, DeepLabV3+, PSPNet, and real-time semantic segmentation network LinkNet, DABNet were selected to train and validate. These six algorithms got quite high segmentation accuracy, among which UNet had the best mIoU and the mPA, but the model size was too large. The improved algorithm only accounts for 1% of the FLOPs and 0.41% of the parameters of UNet, and the mIoU and mPA were basically the same. Other classic semantic segmentation algorithms, such as DeepLabV3+, had similar accuracy to improved algorithms, but their large model size and slow inference speed were not conducive to embedded development. Although the real-time semantic segmentation algorithm LinkNet had a slightly higher mIoU, its FLOPs and parameters count were still far greater than the improved algorithm. Although the PSPNet model was relatively small, it was also much higher than the improved algorithm, and the mIoU and mPA were lower than the algorithm. The experimental results showed that the improved model achieved a mIoU of 98.61%. Compared with the original model, the number of parameters and FLOPs significantly decreased. Among them, the number of model parameters decreased by 51%, the FLOPs decreased by 49%, and the network operation speed increased by 38%. [Conclusions] The improved algorithm can accurately and quickly segment Chinese yam leaves, providing not only a more accurate means for determining Chinese yam phenotype data, but also a new method and approach for embedded research of Chinese yam. Using the model, the morphological feature data of Chinese yam leaves can be obtained more efficiently, providing a reliable foundation for further research and analysis.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight YOLOv8s-Based Strawberry Plug Seedling Grading Detection and Localization via Channel Pruning
    CHEN Junlin, ZHAO Peng, CAO Xianlin, NING Jifeng, YANG Shuqin
    Smart Agriculture    2024, 6 (6): 132-143.   DOI: 10.12133/j.smartag.SA202408001
    Abstract410)   HTML62)    PDF(pc) (3614KB)(1254)       Save

    [Objective] Plug tray seedling cultivation is a contemporary method known for its high germination rates, uniform seedling growth, shortened transplant recovery period, diminished pest and disease incidence, and enhanced labor efficiency. Despite these advantages, challenges such as missing or underdeveloped seedlings can arise due to seedling quality and environmental factors. To ensure uniformity and consistency of the seedlings, sorting is frequently necessary, and the adoption of automated seedling sorting technology can significantly reduce labor costs. Nevertheless, the overgrowth of seedlings within the plugs can effect the accuracy of detection algorithms. A method for grading and locating strawberry seedlings based on a lightweight YOLOv8s model was presented in this research to effectively mitigate the interference caused by overgrown seedlings. [Methods] The YOLOv8s model was selected as the baseline for detecting different categories of seedlings in the strawberry plug tray cultivation process, namely weak seedlings, normal seedlings, and plug holes. To improve the detection efficiency and reduce the model's computational cost, the layer-adaptive magnitude-based pruning(LAMP) score-based channel pruning algorithm was applied to compress the base YOLOv8s model. The pruning procedure involved using the dependency graph to derive the group matrices, followed by normalizing the group importance scores using the LAMP Score, and ultimately pruning the channels according to these processed scores. This pruning strategy effectively reduced the number of model parameters and the overall size of the model, thereby significantly enhancing its inference speed while maintaining the capability to accurately detect both seedlings and plug holes. Furthermore, a two-stage seedling-hole matching algorithm was introduced based on the pruned YOLOv8s model. In the first stage, seedling and plug hole bounding boxes were matched according to their the degree of overlap (Dp), resulting in an initial set of high-quality matches. This step helped minimize the number of potential matching holes for seedlings exhibiting overgrowth. Subsequently, before the second stage of matching, the remaining unmatched seedlings were ranked according to their potential matching hole scores (S), with higher scores indicating fewer potential matching holes. The seedlings were then prioritized during the second round of matching based on these scores, thus ensuring an accurate pairing of each seedling with its corresponding plug hole, even in cases where adjacent seedling leaves encroached into neighboring plug holes. [Results and Discussions] The pruning process inevitably resulted in the loss of some parameters that were originally beneficial for feature representation and model generalization. This led to a noticeable decline in model performance. However, through meticulous fine-tuning, the model's feature expression capabilities were restored, compensating for the information loss caused by pruning. Experimental results demonstrated that the fine-tuned model not only maintained high detection accuracy but also achieved significant reductions in FLOPs (86.3%) and parameter count (95.4%). The final model size was only 1.2 MB. Compared to the original YOLOv8s model, the pruned version showed improvements in several key performance metrics: precision increased by 0.4%, recall by 1.2%, mAP by 1%, and the F1-Score by 0.1%. The impact of the pruning rate on model performance was found to be non-linear. As the pruning rate increased, model performance dropped significantly after certain crucial channels were removed. However, further pruning led to a reallocation of the remaining channels' weights, which in some cases allowed the model to recover or even exceed its previous performance levels. Consequently, it was necessary to experiment extensively to identify the optimal pruning rate that balanced model accuracy and speed. The experiments indicated that when the pruning rate reached 85.7%, the mAP peaked at 96.4%. Beyond this point, performance began to decline, suggesting that this was the optimal pruning rate for achieving a balance between model efficiency and performance, resulting in a model size of 1.2 MB. To further validate the improved model's effectiveness, comparisons were conducted with different lightweight backbone networks, including MobileNetv3, ShuffleNetv2, EfficientViT, and FasterNet, while retaining the Neck and Head modules of the original YOLOv8s model. Results indicated that the modified model outperformed these alternatives, with mAP improvements of 1.3%, 1.8%, 1.5%, and 1.1%, respectively, and F1-Score increases of 1.5%, 1.8%, 1.1%, and 1%. Moreover, the pruned model showed substantial advantages in terms of floating-point operations, model size, and parameter count compared to these other lightweight networks. To verify the effectiveness of the proposed two-stage seedling-hole matching algorithm, tests were conducted using a variety of complex images from the test set. Results indicated that the proposed method achieved precise grading and localization of strawberry seedlings even under challenging overgrowth conditions. Specifically, the correct matching rate for normal seedlings reached 96.6%, for missing seedlings 84.5%, and for weak seedlings 82.9%, with an average matching accuracy of 88%, meeting the practical requirements of the strawberry plug tray cultivation process. [Conclusions] The pruned YOLOv8s model successfully maintained high detection accuracy while reducing computational costs and improving inference speed. The proposed two-stage seedling-hole matching algorithm effectively minimized the interference caused by overgrown seedlings, accurately locating and classifying seedlings of various growth stages within the plug tray. The research provides a robust and reliable technical solution for automated strawberry seedling sorting in practical plug tray cultivation scenarios, offering valuable insights and technical support for optimizing the efficiency and precision of automated seedling grading systems.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Disease Named Entity Recognition with Pointer Network Based on RoFormer Pre-trained Model
    WANG Tong, WANG Chunshan, LI Jiuxi, ZHU Huaji, MIAO Yisheng, WU Huarui
    Smart Agriculture    2024, 6 (2): 85-94.   DOI: 10.12133/j.smartag.SA202311021
    Abstract1589)   HTML37)    PDF(pc) (1219KB)(1189)       Save

    [Objective] With the development of agricultural informatization, a large amount of information about agricultural diseases exists in the form of text. However, due to problems such as nested entities and confusion of entity types, traditional named entities recognition (NER) methods often face challenges of low accuracy when processing agricultural disease text. To address this issue, this study proposes a new agricultural disease NER method called RoFormer-PointerNet, which combines the RoFormer pre-trained model with the PointerNet baseline model. The aim of this method is to improve the accuracy of entity recognition in agricultural disease text, providing more accurate data support for intelligent analysis, early warning, and prevention of agricultural diseases. [Methods] This method first utilized the RoFormer pre-trained model to perform deep vectorization processing on the input agricultural disease text. This step was a crucial foundation for the subsequent entity extraction task. As an advanced natural language processing model, the RoFormer pre-trained model's unique rotational position embedding approach endowed it with powerful capabilities in capturing textual positional information. In agricultural disease text, due to the diversity of terminology and the existence of polysemy, traditional entity recognition methods often faced challenges in confusing entity types. However, through its unique positional embedding mechanism, the RoFormer model was able to incorporate more positional information into the vector representation, effectively enriching the feature information of words. This characteristic enabled the model to more accurately distinguish between different entity types in subsequent entity extraction tasks, reducing the possibility of type confusion. After completing the vectorization representation of the text, this study further emploied a pointer network for entity extraction. The pointer network was an advanced sequence labeling approach that utilizes head and tail pointers to annotate entities within sentences. This labeling method was more flexible compared to traditional sequence labeling methods as it was not restricted by fixed entity structures, enabling the accurate extraction of all types of entities within sentences, including complex entities with nested relationships. In agricultural disease text, entity extraction often faced the challenge of nesting, such as when multiple different entity types are nested within a single disease symptom description. By introducing the pointer network, this study effectively addressed this issue of entity nesting, improving the accuracy and completeness of entity extraction. [Results and Discussions] To validate the performance of the RoFormer-PointerNet method, this study constructed an agricultural disease dataset, which comprised 2 867 annotated corpora and a total of 10 282 entities, including eight entity types such as disease names, crop names, disease characteristics, pathogens, infected areas, disease factors, prevention and control methods, and disease stages. In comparative experiments with other pre-trained models such as Word2Vec, BERT, and RoBERTa, RoFormer-PointerNet demonstrated superiority in model precision, recall, and F1-Score, achieving 87.49%, 85.76% and 86.62%, respectively. This result demonstrated the effectiveness of the RoFormer pre-trained model. Additionally, to verify the advantage of RoFormer-PointerNet in mitigating the issue of nested entities, this study compared it with the widely used bidirectional long short-term memory neural network (BiLSTM) and conditional random field (CRF) models combined with the RoFormer pre-trained model as decoding methods. RoFormer-PointerNet outperformed the RoFormer-BiLSTM, RoFormer-CRF, and RoFormer-BiLSTM-CRF models by 4.8%, 5.67% and 3.87%, respectively. The experimental results indicated that RoFormer-PointerNet significantly outperforms other models in entity recognition performance, confirming the effectiveness of the pointer network model in addressing nested entity issues. To validate the superiority of the RoFormer-PointerNet method in agricultural disease NER, a comparative experiment was conducted with eight mainstream NER models such as BiLSTM-CRF, BERT-BiLSTM-CRF, and W2NER. The experimental results showed that the RoFormer-PointerNet method achieved precision, recall, and F1-Score of 87.49%, 85.76% and 86.62%, respectively in the agricultural disease dataset, reaching the optimal level among similar methods. This result further verified the superior performance of the RoFormer-PointerNet method in agricultural disease NER tasks. [Conclusions] The agricultural disease NER method RoFormer-PointerNet, proposed in this study and based on the RoFormer pre-trained model, demonstrates significant advantages in addressing issues such as nested entities and type confusion during the entity extraction process. This method effectively identifies entities in Chinese agricultural disease texts, enhancing the accuracy of entity recognition and providing robust data support for intelligent analysis, early warning, and prevention of agricultural diseases. This research outcome holds significant importance for promoting the development of agricultural informatization and intelligence.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grape Recognition and Localization Method Based on 3C-YOLOv8n and Depth Camera
    LIU Chang, SUN Yu, YANG Jing, WANG Fengchao, CHEN Jin
    Smart Agriculture    2024, 6 (6): 121-131.   DOI: 10.12133/j.smartag.SA202407008
    Abstract361)   HTML38)    PDF(pc) (2020KB)(1146)       Save

    [Objective] Grape picking is a key link in increasing production. However, in this process, a large amount of manpower and material resources are required, which makes the picking process complex and slow. To enhance harvesting efficiency and achieve automated grape harvesting, an improved YOLOv8n object detection model named 3C-YOLOv8n was proposed, which integrates the RealSense D415 depth camera for grape recognition and localization. [Methods] The propoesed 3C-YOLOv8n incorporated a convolutional block attention module (CBAM) between the first C2f module and the third Conv module in the backbone network. Additionally, a channel attention (CA) module was added at the end of the backbone structure, resulting in a new 2C-C2f backbone network architecture. This design enabled the model to sequentially infer attention maps across two independent dimensions (channel and spatial), optimize features by considering relationships between channels and positional information. The network structure was both flexible and lightweight. Furthermore, the Content-aware ReAssembly of Features up sampling operator was implemented to support instance-specific kernels (such as deconvolution) for feature reconstruction with neighboring pixels, replacing the nearest neighbor interpolation operator in the YOLOv8n neck network. This enhancement increased the receptive field and guided the reconstruction process based on input features while maintaining low parameter and computational complexity, thereby forming the 3C-YOLOv8n model. The pyrealsense2 library was utilized to obtain pixel position information from the target area using the Intel RealSense D415 camera. During this process, the depth camera was used to capture images, and target detection algorithms were employed to pinpoint the location of grapes. The camera's depth sensor facilitated the acquisition of the three-dimensional point cloud of grapes, allowing for the calculation of the distance from the pixel point to the camera and the subsequent determination of the three-dimensional coordinates of the center of the target's bounding box in the camera coordinate system, thus achieving grape recognition and localization. [Results and Discussions] Comparative and ablation experiments were conducted. it was observed that the 3C-YOLOv8n model achieved a mean average precision (mAP) of 94.3% at an intersection ratio of 0.5 (IOU=0.5), surpassing the YOLOv8n model by 1%. The accuracy (P) and recall (R) rates were recorded at 91.6% and 86.4%, respectively, reflecting increases of 0.1% and 0.7%. The F1-Score also improved by 0.4%, demonstrating that the improved network model met the experimental accuracy and recall requirements. In terms of loss, the 3C-YOLOv8n algorithm exhibited superior performance, with a rapid decrease in loss values and minimal fluctuations, ultimately leading to a minimized loss value. This indicated that the improved algorithm quickly reached a convergence state, enhancing both model accuracy and convergence speed. The ablation experiments revealed that the original YOLOv8n model yielded a mAP of 93.3%. The integration of the CBAM and CA attention mechanisms into the YOLOv8n backbone resulted in mAP values of 93.5% each. The addition of the Content-aware ReAssembly of Features up sampling operator to the neck network of YOLOv8n produced a 0.5% increase in mAP, culminating in a value of 93.8%. The combination of the three improvement strategies yielded mAP increases of 0.3, 0.7, and 0.8%, respectively, compared to the YOLOv8n model. Overall, the 3C-YOLOv8n model demonstrated the best detection performance, achieving the highest mAP of 94.3%. The ablation results confirmed the positive impact of the proposed improvement strategies on the experimental outcomes. Compared to other mainstream YOLO series algorithms, all evaluation metrics showed enhancements, with the lowest missed detection and false detection rates among all tested algorithms, underscoring its practical advantages in detection tasks. [Conclusions] By effectively addressing the inefficiencies of manual labor, 3C-YOLOv8n network model not only enhances the precision of grape recognition and localization but also significantly optimizes overall harvesting efficiency. Its superior performance in evaluation metrics such as precision, recall, mAP, and F1-Score, alongside the lowest recorded loss values among YOLO series algorithms, indicates a remarkable advancement in model convergence and operational effectiveness. Furthermore, the model's high accuracy in grape target recognition not only lays the groundwork for automated harvesting systems but also enables the implementation of complementary intelligent operations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Visible/NIR Spectral Inversion of Malondialdehyde Content in JUNCAO Based on Deep Convolutional Gengrative Adversarial Network
    YE Dapeng, CHEN Chen, LI Huilin, LEI Yingxiao, WENG Haiyong, QU Fangfang
    Smart Agriculture    2023, 5 (3): 132-141.   DOI: 10.12133/j.smartag.SA202307011
    Abstract359)   HTML26)    PDF(pc) (1784KB)(1143)       Save

    [Objective] JUNCAO, a perennial herbaceous plant that can be used as medium for cultivating edible and medicinal fungi. It has important value for promotion, but the problem of overwintering needs to be overcome when planting in the temperate zone. Low-temperature stress can adversely impact the growth of JUNCAO plants. Malondialdehyde (MDA) is a degradation product of polyunsaturated fatty acid peroxides, which can serve as a useful diagnostic indicator for studying plant growth dynamics. Because the more severe the damage caused by low temperature stress on plants, the higher their MDA content. Therefore, the detection of MDA content can provide instruct for low-temperature stress diagnosis and JUNCAO plants breeding. With the development of optical sensors and machine learning technologies, visible/near-infrared spectroscopy technology combined with algorithmic models has great potential in rapid, non-destructive and high-throughput inversion of MDA content and evaluation of JUNCAO growth dynamics. [Methods] In this research, six varieties of JUNCAO plants were selected as experimental subjects. They were divided into a control group planted at ambient temperature (28°C) and a stress group planted at low temperature (4°C). The hyperspectral reflectances of JUNCAO seedling leaves during the seedling stage were collected using an ASD spectroradiomete and a near-infrared spectrometer, and then the leaf physiological indicators were measured to obtain leaf MDA content. Machine learning methods were used to establish the MDA content inversion models based on the collected spectral reflectance data. To enhance the prediction accuracy of the model, an improved one-dimensional deep convolutional generative adversarial network (DCAGN ) was proposed to increase the sample size of the training set. Firstly, the original samples were divided into a training set (96 samples) and a prediction set (48 samples) using the Kennard stone (KS) algorithm at a ratio of 2:1. Secondly, the 96 training set samples were generated through the DCGAN model, resulting in a total of 384 pseudo samples that were 4 times larger than the training set. The pseudo samples were randomly shuffled and sequentially added to the training set to form an enhanced modeling set. Finally, the MDA quantitative detection models were established based on random forest (RF), partial least squares regression (PLSR), and convolutional neural network (CNN) algorithms. By comparing the prediction accuracies of the three models after increasing the sample size of the training set, the best MDA regression detection model of JUNCAO was obtained. [Results and Discussions] (1) The MDA content of the six varieties of JUNCAO plants ranged from 12.1988 to 36.7918 nmol/g. Notably, the MDA content of JUNCAO under low-temperature stress was remarkably increased compared to the control group with significant differences (P<0.05). Moreover, the visible/near-infrared spectral reflectance in the stressed group also exhibited an increasing trend compared to the control group. (2) Samples generated by the DCAGN model conformed to the distribution patterns of the original samples. The spectral curves of the generated samples retained the shape and trends of the original data. The corresponding MDA contented of generated samples consistently falling within the range of the original samples, with the average and standard deviation only decreased by 0.6650 and 0.9743 nmol/g, respectively. (3) Prior to the inclusion of generated samples, the detection performance of the three models differed significantly, with a correlation coefficient (R2) of 0.6967 for RF model, that of 0.6729 for CNN model, and that of 0.5298 for the PLSR model. After the introduction of generated samples, as the number of samples increased, all three models exhibited an initial increase followed by a decrease in R2 on the prediction set, while the root mean square error of prediction (RMSEP) first decreased and then increased. (4) The prediction results of the three regression models indicated that augmenting the sample size by using DCGAN could effectively enhance the prediction performance of models. Particularly, utilizing DCGAN in combination with the RF model achieved the optimal MDA content detection performance, with the R2 of 0.7922 and the RMSEP of 2.1937. [Conclusions] Under low temperature stress, the MDA content and spectral reflectance of the six varieties of JUNCAO leaves significantly increased compared to the control group, which might due to the damage of leaf pigments and tissue structure, and the decrease in leaf water content. Augmenting the sample size using DCGAN effectively enhanced the reliability and detection accuracy of the models. This improvement was evident across different regression models, illustrating the robust generalization capabilities of this DCGAN deep learning network. Specifically, the combination of DCGAN and RF model achieved optimal MDA content detection performance, as expanding to a sufficient sample dataset contributed to improve the modeling accuracy and stability. This research provides valuable insights for JUNCAO plants breeding and the diagnosis of low-temperature stress based on spectral technology and machine learning methods, offering a scientific basis for achieving high, stable, and efficient utilization of JUNCAO plants.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Cow Hoof Slippage Detecting Method Based on Enhanced DeepLabCut Model
    NIAN Yue, ZHAO Kaixuan, JI Jiangtao
    Smart Agriculture    2024, 6 (5): 153-163.   DOI: 10.12133/j.smartag.SA202406014
    Abstract867)   HTML27)    PDF(pc) (1765KB)(1111)       Save

    [Objective] The phenomenon of hoof slipping occurs during the walking process of cows, which indicates the deterioration of the farming environment and a decline in the cows' locomotor function. Slippery grounds can lead to injuries in cows, resulting in unnecessary economic losses for farmers. To achieve automatically recognizing and detecting slippery hoof postures during walking, the study focuses on the localization and analysis of key body points of cows based on deep learning methods. Motion curves of the key body points were analyzed, and features were extracted. The effectiveness of the extracted features was verified using a decision tree classification algorithm, with the aim of achieving automatic detection of slippery hoof postures in cows. [Method] An improved localization method for the key body points of cows, specifically the head and four hooves, was proposed based on the DeepLabCut model. Ten networks, including ResNet series, MobileNet-V2 series, and EfficientNet series, were selected to respectively replace the backbone network structure of DeepLabCut for model training. The root mean square error(RMSE), model size, FPS, and other indicators were chosen, and after comprehensive consideration, the optimal backbone network structure was selected as the pre-improved network. A network structure that fused the convolutional block attention module (CBAM) attention mechanism with ResNet-50 was proposed. A lightweight attention module, CBAM, was introduced to improve the ResNet-50 network structure. To enhance the model's generalization ability and robustness, the CBAM attention mechanism was embedded into the first convolution layer and the last convolution layer of the ResNet-50 network structure. Videos of cows with slippery hooves walking in profile were predicted for key body points using the improved DeepLabCut model, and the obtained key point coordinates were used to plot the motion curves of the cows' key body points. Based on the motion curves of the cows' key body points, the feature parameter Feature1 for detecting slippery hooves was extracted, which represented the local peak values of the derivative of the motion curves of the cows' four hooves. The feature parameter Feature2 for predicting slippery hoof distances was extracted, specifically the minimum local peak points of the derivative curve of the hooves, along with the local minimum points to the left and right of these peaks. The effectiveness of the extracted Feature1 feature parameters was verified using a decision tree classification model. Slippery hoof feature parameters Feature1 for each hoof were extracted, and the standard deviation of Feature1 was calculated for each hoof. Ultimately, a set of four standard deviations for each cow was extracted as input parameters for the classification model. The classification performance was evaluated using four common objective metrics, including accuracy, precision, recall, and F1-Score. The prediction accuracy for slippery hoof distances was assessed using RMSE as the evaluation metric. [Results and Discussion] After all ten models reached convergence, the loss values ranked from smallest to largest were found in the EfficientNet series, ResNet series, and MobileNet-V2 series, respectively. Among them, ResNet-50 exhibited the best localization accuracy in both the training set and validation set, with RMSE values of only 2.69 pixels and 3.31 pixels, respectively. The MobileNet series had the fastest inference speed, reaching 48 f/s, while the inference speeds of the ResNet series and MobileNet series were comparable, with ResNet series performing slightly better than MobileNet series. Considering the above factors, ResNet-50 was ultimately selected as the backbone network for further improvements on DeepLabCut. Compared to the original ResNet-50 network, the ResNet-50 network improved by integrating the CBAM module showed a significant enhancement in localization accuracy. The accuracy of the improved network increased by 3.7% in the training set and by 9.7% in the validation set. The RMSE between the predicted body key points and manually labeled points was only 2.99 pixels, with localization results for the right hind hoof, right front hoof, left hind hoof, left front hoof, and head improved by 12.1%, 44.9%, 0.04%, 48.2%, and 39.7%, respectively. To validate the advancement of the improved model, a comparison was made with the mainstream key point localization model, YOLOv8s-pose, which showed that the RMSE was reduced by 1.06 pixels compared to YOLOv8s-pose. This indicated that the ResNet-50 network integrated with the CBAM attention mechanism possessed superior localization accuracy. In the verification of the cow slippery hoof detection classification model, a 10-fold cross-validation was conducted to evaluate the performance of the cow slippery hoof classification model, resulting in average values of accuracy, precision, recall, and F1-Score at 90.42%, 0.943, 0.949, and 0.941, respectively. The error in the calculated slippery hoof distance of the cows, using the slippery hoof distance feature parameter Feature2, compared to the manually calibrated slippery hoof distance was found to be 1.363 pixels. [Conclusion] The ResNet-50 network model improved by integrating the CBAM module showed a high accuracy in the localization of key body points of cows. The cow slippery hoof judgment model and the cow slippery hoof distance prediction model, based on the extracted feature parameters for slippery hoof judgment and slippery hoof distance detection, both exhibited small errors when compared to manual detection results. This indicated that the proposed enhanced deeplabcut model obtained good accuracy and could provide technical support for the automatic detection of slippery hooves in cows.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Detection Method of Effective Tillering of Rice in Field Based on Lightweight Ghost-YOLOv8 and Smart Phone
    CUI Jiale, ZENG Xiangfeng, REN Zhengwei, SUN Jian, TANG Chen, YANG Wanneng, SONG Peng
    Smart Agriculture    2024, 6 (5): 98-107.   DOI: 10.12133/j.smartag.SA202407012
    Abstract481)   HTML84)    PDF(pc) (2128KB)(1094)       Save

    [Objective] The number of effective tillers per plant is one of the important agronomic traits affecting rice yield. In order to solve the problems of high cost and low accuracy of effective tiller detection caused by dense tillers, mutual occlusion and ineffective tillers in rice, a method for dividing effective tillers and ineffective tillers in rice was proposed. Combined with the deep learning model, a high-throughput and low-cost mobile phone App for effective tiller detection in rice was developed to solve the practical problems of effective tiller investigation in rice under field conditions. [Methods] The investigations of rice tillering showed that the number of effective tillers of rice was often higher than that of ineffective tillers. Based on the difference in growth height between effective and ineffective tillers of rice, a new method for distinguishing effective tillers from ineffective tillers was proposed. A fixed height position of rice plants was selected to divide effective tillers from ineffective tillers, and rice was harvested at this position. After harvesting, cross-sectional images of rice tillering stems were taken using a mobile phone, and the stems were detected and counted by the YOLOv8 model. Only the cross-section of the stem was identified during detection, while the cross-section of the panicle was not identified. The number of effective tillers of rice was determined by the number of detected stems. In order to meet the needs of field work, a mobile phone App for effective tiller detection of rice was developed for real-time detection. GhostNet was used to lighten the YOLOv8 model. Ghost Bottle-Neck was integrated into C2f to replace the original BottleNeck to form C2f-Ghost module, and then the ordinary convolution in the network was replaced by Ghost convolution to reduce the complexity of the model. Based on the lightweight Ghost-YOLOv8 model, a mobile App for effective tiller detection of rice was designed and constructed using the Android Studio development platform and intranet penetration counting. [Results and Discussions] The results of field experiments showed that there were differences in the growth height of effective tillers and ineffective tillers of rice. The range of 52 % to 55 % of the total plant height of rice plants was selected for harvesting, and the number of stems was counted as the number of effective tillers per plant. The range was used as the division standard of effective tillers and ineffective tillers of rice. The accuracy and recall rate of effective tillers counting exceeded 99%, indicating that the standard was accurate and comprehensive in guiding effective tillers counting. Using the GhostNet lightweight YOLOv8 model, the parameter quantity of the lightweight Ghost-YOLOv8 model was reduced by 43%, the FPS was increased by 3.9, the accuracy rate was 0.988, the recall rate was 0.980, and the mAP was 0.994. The model still maintains excellent performance while light weighting. Based on the lightweight Ghost-YOLOv8 model, a mobile phone App for detecting effective tillers of rice was developed. The App was tested on 100 cross-sectional images of rice stems collected under the classification criteria established in this study. Compared with the results of manual counting of effective tillers per plant, the accuracy of the App's prediction results was 99.61%, the recall rate was 98.76%, and the coefficient of determination was 0.985 9, indicating the reliability of the App and the established standards in detecting effective tillers of rice. [Conclusions] Through the lightweight Ghost-YOLOv8 model, the number of stems in the cross-sectional images of stems collected under the standard was detected to obtain the effective tiller number of rice. An Android-side rice effective tillering detection App was developed, which can meet the field investigation of rice effective tillering, help breeders to collect data efficiently, and provide a basis for field prediction of rice yield. Further research could supplement the cross-sectional image dataset of multiple rice stems to enable simultaneous measurement of effective tillers across multiple rice plants and improve work efficiency. Further optimization and enhancement of the App's functionality is necessary to provide more tiller-related traits, such as tiller angle.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Precision Target Spraying System Integrated with Remote Deep Learning Recognition Model for Cabbage Plant Centers
    ZHANG Hui, HU Jun, SHI Hang, LIU Changxi, WU Miao
    Smart Agriculture    2024, 6 (6): 85-95.   DOI: 10.12133/j.smartag.SA202406013
    Abstract316)   HTML25)    PDF(pc) (1923KB)(1056)       Save

    [Objective] Spraying calcium can effectively prevent the occurrence of dry burning heart disease in Chinese cabbage. Accurately targeting spraying calcium can more effectively improve the utilization rate of calcium. Since the sprayer needs to move rapidly in the field, this can lead to over-application or under-application of the pesticide. This study aims to develop a targeted spray control system based on deep learning technology, explore the relationship between the advance speed, spray volume, and coverage of the sprayer, thereby addressing the uneven application issues caused by different nebulizer speeds by studying the real scenario of calcium administration to Chinese cabbage hearts. [Methods] The targeted spraying control system incorporates advanced sensors and computing equipment that were capable of obtaining real-time data regarding the location of crops and the surrounding environmental conditions. This data allowed for dynamic adjustments to be made to the spraying system, ensuring that pesticides were delivered with high precision. To further enhance the system's real-time performance and accuracy, the YOLOv8 object detection model was improved. A Ghost-Backbone lightweight network structure was introduced, integrating remote sensing technologies along with the sprayer's forward speed and the frequency of spray responses. This innovative combination resulted in the creation of a YOLOv8-Ghost-Backbone lightweight model specifically tailored for agricultural applications. The model operated on the Jetson Xavier NX controller, which was a high-performance, low-power computing platform designed for edge computing. The system was allowed to process complex tasks in real time directly in the field. The targeted spraying system was composed of two essential components: A pressure regulation unit and a targeted control unit. The pressure regulation unit was responsible for adjusting the pressure within the spraying system to ensure that the output remains stable under various operational conditions. Meanwhile, the targeted control unit played a crucial role in precisely controlling the direction, volume, and coverage of the spray to ensure that the pesticide was applied effectively to the intended areas of the plants. To rigorously evaluate the performance of the system, a series of intermittent spray tests were conducted. During these tests, the forward speed of the sprayer was gradually increased, allowing to assess how well the system responded to changes in speed. Throughout the testing phase, the response frequency of the electromagnetic valve was measured to calculate the corresponding spray volume for each nozzle. [Results and Conclusions] The experimental results indicated that the overall performance of the targeted spraying system was outstanding, particularly under conditions of high-speed operation. By meticulously recording the response times of the three primary components of the system, the valuable data were gathered. The average time required for image processing was determined to be 29.50 ms, while the transmission of decision signals took an average of 6.40 ms. The actual spraying process itself required 88.83 ms to complete. A thorough analysis of these times revealed that the total response time of the spraying system lagged by approximately 124.73 ms when compared to the electrical signal inputs. Despite the inherent delays, the system was able to maintain a high level of spraying accuracy by compensating for the response lag of the electromagnetic valve. Specifically, when tested at a speed of 7.2 km/h, the difference between the actual spray volume delivered and the required spray volume, after accounting for compensation, was found to be a mere 0.01 L/min. This minimal difference indicates that the system met the standard operational requirements for effective pesticide application, thereby demonstrating its precision and reliability in practical settings. [Conclusions] In conclusion, this study developed and validated a deep learning-based targeted spraying control system that exhibited excellent performance regarding both spraying accuracy and response speed. The system serves as a significant technical reference for future endeavors in agricultural automation. Moreover, the research provides insights into how to maintain consistent spraying effectiveness and optimize pesticide utilization efficiency by dynamically adjusting the spraying system as the operating speed varies. The findings of this research will offer valuable experiences and guidance for the implementation of agricultural robots in the precise application of pesticides, with a particular emphasis on parameter selection and system optimization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Prediction and Mapping of Soil Total Nitrogen Using GF-5 Image Based on Machine Learning Optimization Modeling
    LIU Liqi, WEI Guangyuan, ZHOU Ping
    Smart Agriculture    2024, 6 (5): 61-73.   DOI: 10.12133/j.smartag.SA202405011
    Abstract325)   HTML46)    PDF(pc) (3325KB)(1032)       Save

    [Objective] Nitrogen in soil is an absolutely crucial element for plant growth. Insufficient nitrogen supply can severely affect crop yield and quality, while excessive use of nitrogen fertilizers can lead to significant environmental issues such as water eutrophication and groundwater pollution. Therefore, large-scale, rapid detection of soil nitrogen content and precise fertilization are of great importance for smart agriculture. In this study, the hyperspectral data from the GF-5 satellite was emploied, and the various machine learning algorithms introduced to establish a prediction model for soil total nitrogen (TN) content and a distribution map of soil TN content was generated in the study area, aiming to provide scientific evidence for intelligent monitoring in smart agriculture. [Method] The study area was the Jian Sanjiang Reclamation Area in Fujin city, Heilongjiang province. Fieldwork involved the careful collection of 171 soil samples, obtaining soil spectral data, chemical analysis data of soil TN content, and the GF-5 hyperspectral data. Among these samples, 140 were randomly selected as the modeling sample set for calibration, and the remaining 31 samples were used as the test sample set. Three machine learning algorithms were introduced: Partial least squares regression (PLSR), backpropagation neural network (BPNN), and support vector machine (SVM) driven by a polynomial kernel function (Poly). Three distinct soil TN inversion models were constructed using these algorithms. To optimize model performance, ten-fold cross-validation was employed to determine the optimal parameters for each model. Additionally, multiple scatter correction (MSC) was applied to obtain band characteristic values, thus enhancing the model's prediction capability. Model performance was evaluated using three indicators: Coefficient of determination (R²), root mean square error (RMSE), and relative prediction deviation (RPD), to assess the prediction accuracy of different models. [Results and Discussions] MSC-Poly-SVM model exhibited the best prediction performance on the test sample set, with an R² of 0.863, an RMSE of 0.203, and an RPD of 2.147. This model was used to perform soil TN content inversion mapping using GF-5 satellite hyperspectral data. In accordance with the stringent requirements of land quality geochemical evaluation, the GF-5 hyperspectral land organic nitrogen parameter distribution map was drawn based on the "Determination of Land Quality Geochemical Evaluation". The results revealed that 86.1% of the land in the Jian Sanjiang study area had a total nitrogen content of more than 2.0 g/kg, primarily concentrated in first and second-grade plots, while third and fourth-grade plots accounted for only 11.83% of the total area. The study area exhibited sufficient soil nitrogen reserves, with high TN background values mainly concentrated along the riverbanks in the central part, distributed in a northeast-east direction. Specifically, in terms of soil spectral preprocessing, the median filtering method performed best in terms of smoothness and maintaining spectral characteristics. The spectra extracted from GF-5 imagery were generally quite similar to ground-measured spectral data, despite some noise, which had a minimal overall impact. [Conclusions] This study demonstrates the clear feasibility of using GF-5 satellite hyperspectral remote sensing data and machine learning algorithm for large-scale quantitative detection and visualization analysis of soil TN content. The soil TN content distribution map generated based on GF-5 hyperspectral remote sensing data is detailed and consistent with results from other methods, providing technical support for future large-scale quantitative detection of soil nutrient status and rational fertilization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Ecological Risk Assessment of Cultivated Land Based on Landscape Pattern: A Case Study of Tongnan District, Chongqing
    ZHANG Xingshan, YANG Heng, MA Wenqiu, YANG Minli, WANG Haiyi, YOU Yong, HUI Yunting, GONG Zeqi, WANG Tianyi
    Smart Agriculture    2024, 6 (3): 58-68.   DOI: 10.12133/j.smartag.SA202306008
    Abstract430)   HTML28)    PDF(pc) (1636KB)(1025)       Save

    [Objective] Farmland consolidation for agricultural mechanization in hilly and mountainous areas can alter the landscape pattern, elevation, slope and microgeomorphology of cultivated land. It is of great significance to assess the ecological risk of cultivated land to provide data reference for the subsequent farmland consolidation for agricultural mechanization. This study aims to assess the ecological risk of cultivated land before and after farmland consolidation for agricultural mechanization in hilly and mountainous areas, and to explore the relationship between cultivated land ecological risk and cultivated land slope. [Methods] Twenty counties in Tongnan district of Chongqing city was selected as the assessment units. Based on the land use data in 2010 and 2020 as two periods, ArcGIS 10.8 and Excel software were used to calculate landscape pattern indices. The weights for each index were determined by entropy weight method, and an ecological risk assessment model was constructed, which was used to reveal the temporal and spatial change characteristics of ecological risk. Based on the principle of mathematical statistics, the correlation analysis between cultivated land ecological risk and cultivated land slope was carried out, which aimed to explore the relationship between cultivated land ecological risk and cultivated land slope. [Results and Discussions] Comparing to 2010, patch density (PD), division (D), fractal dimension (FD), and edge density (ED) of cultivated land all decreased in 2020, while meant Patch Size (MPS) increased, indicating an increase in the contiguity of cultivated land. The mean shape index (MSI) of cultivated land increased, indicating that the shape of cultivated land tended to be complicated. The landscape disturbance index (U) decreased from 0.97 to 0.94, indicating that the overall resistance to disturbances in cultivated land has increased. The landscape vulnerability index (V) increased from 2.96 to 3.20, indicating that the structure of cultivated land become more fragile. The ecological risk value of cultivated land decreased from 3.10 to 3.01, indicating the farmland consolidation for agricultural mechanization effectively improved the landscape pattern of cultivated land and enhanced the safety of the agricultural ecosystem. During the two periods, the ecological risk areas were primarily composed of low-risk and relatively low-risk zones. The area of low-risk zones increased by 6.44%, mainly expanding towards the northern part, while the area of relatively low-risk zones increased by 6.17%, primarily spreading towards the central-eastern and southeastern part. The area of moderate-risk zones increased by 24.4%, mainly extending towards the western and northwestern part, while the area of relatively high-risk zones decreased by 60.70%, with some new additions spreading towards the northeastern part. The area of high-risk zones increased by 16.30%, with some new additions extending towards the northwest part. Overall, the ecological safety zones of cultivated relatively increased. The cultivated land slope was primarily concentrated in the range of 2° to 25°. On the one hand, when the cultivated land slope was less than 15°, the proportion of the slope area was negatively correlated with the ecological risk value. On the other hand, when the slope was above 15°, the proportion of the slope area was positively correlated with the ecological risk value. In 2010, there was a highly significant correlation between the proportion of slope area and ecological risk value for cultivated land slope within the ranges of 5° to 8°, 15° to 25°, and above 25°, with corresponding correlation coefficients of 0.592, 0.609, and 0.849, respectively. In 2020, there was a highly significant correlation between the proportion of slope area and ecological risk value for cultivated land slope within the ranges of 2° to 5°, 5° to 8°, 15° to 25°, and above 25°, with corresponding correlation coefficients of 0.534, 0.667, 0.729, and 0.839, respectively. [Conclusions] The assessment of cultivated land ecological risk in Tongnan district of Chongqing city before and after the farmland consolidation for agricultural mechanization, as well as the analysis of the correlation between ecological risk and cultivated land slope, demonstrate that the farmland consolidation for agricultural mechanization can reduce cultivated land ecological risk, and the proportion of cultivated land slope can be an important basis for precision guidance in the farmland consolidation for agricultural mechanization. Considering the occurrence of moderate sheet erosion from a slope of 5° and intense erosion from a slope of 10° to 15°, and taking into account the reduction of ecological risk value and the actual topographic conditions, the subsequent farmland consolidation for agricultural mechanization in Tongnan district should focus on areas with cultivated land slope ranging from 5° to 8° and 15° to 25°.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    ReluformerN: Lightweight High-Low Frequency Enhanced for Hyperspectral Agricultural Lancover Classification
    LIU Yi, ZHANG Yanjun
    Smart Agriculture    2024, 6 (5): 74-87.   DOI: 10.12133/j.smartag.SA202406008
    Abstract350)   HTML23)    PDF(pc) (3072KB)(970)       Save

    [Objective] In order to intelligently monitor the distribution of agricultural land cover types, high-spectral cameras are usually mounted on drones to collect high-spectral data, followed by classification of the high-spectral data to automatically draw crop distribution maps. Different crops have similar shapes, and the same crop has significant differences in different growth stages, so the network model for agricultural land cover classification requires a high degree of accuracy. However, network models with high classification accuracy are often complex and cannot be deployed on hardware systems. In view of this problem, a lightweight high-low frequency enhanced Reluformer network (ReluformerN) was proposed in this research. [Methods] Firstly, an adaptive octave convolution was proposed, which utilized the softmax function to automatically adjust the spectral dimensions of high-frequency features and low-frequency features, effectively alleviating the influence of manually setting the spectral dimensions and benefiting the subsequent extraction of spatial and spectral domain features of hyperspectral images. Secondly, a Reluformer was proposed to extract global features, taking advantage of the fact that low-frequency information could capture global features. Reluformer replaced the softmax function with a function of quadratic computational complexity, and through theoretical and graphical analysised, Relu function, LeakRelu function, and Gelu function were compared, it was found that the ReLU function and the softmax function both had non-negativity, which could be used for feature relevance analysis. Meanwhile, the ReLU function has a linearization feature, which is more suitable for self-relevance analysis. Therefore, the ReLU self-attention mechanism was proposed, which used the ReLU function to perform feature self-attention analysis. In order to extract deep global features, multi-scale feature fusion was used, and the ReLU self-attention mechanism was used as the core to construct the multi-head ReLU self-attention mechanism. Similar to the transformer architecture, the Reluformer structure was built by combining multi-head ReLU self-attention mechanism, feedforward layers, and normalization layers. With Reluformer as the core, the Reluformer network (ReluformerN) was proposed. This network considered frequency from the perspective of high-frequency information, taking into account the local features of image high-frequency information, and used deep separable convolution to design a lightweight network for fine-grained feature extraction of high-frequency information. It proposed Reluformer to extract global features for low-frequency information, which represented the global features of the image. ReluformerN was experimented on three public high-spectral data sets (Indian Pines, WHU-Hi-LongKou and Salinas) for crop variety fine classification, and was compared with five popular classification networks (2D-CNN, HybirdSN, ViT, CTN and LSGA-VIT). [Results and Discussion] ReluformerN performed best in overall accuracy (OA), average accuracy (AA), and other accuracy evaluation indicators. In the evaluation indicators of model parameters, model computation (FLOPs), and model complexity, ReluformerN had the smallest number of parameters and was less than 0.3 M, and the lowest computation. In the visualization comparison, the classification effect diagram of the model using ReluformerN had clearer image edges and more complete morphological structures, with fewer classification errors. The validity of the adaptive octave convolution was verified by comparing it with the traditional eightfold convolution. The classification accuracy of the adaptive octave convolution was 0.1% higher than that of the traditional octave convolution. When the artificial parameters were set to different values, the maximum and minimum classification accuracies of the traditional octave convolution were about 0.3% apart, while those of the adaptive octave convolution were only 0.05%. This showed that the adaptive octave convolution not only had the highest classification accuracy, but was also less sensitive to the artificial parameter setting, effectively overcoming the influence of the artificial parameter setting on the classification result. To validated the Reluformer module, it was compared with transformer, LeakRelufromer, and Linformer in terms of accuracy evaluation metrics such as OA and AA. The Reluformer achieved the highest classification accuracy and the lowest model parameter count among these models. This indicated that Reluformer not only effectively extracted global features but also reduced computational complexity. Finally, the effectiveness of the high-frequency and low-frequency branch networks was verified. The effectiveness of the high-frequency and low-frequency feature extraction branches was verified, and the characteristics of the feature distribution after high-frequency feature extraction, after high-low frequency feature extraction, and after the classifier were displayed using a 2D t-sne, compared with the original feature distribution. It was found that after high-frequency feature extraction, similar features were generally clustered together, but the spacing between different features was small, and there were also some features with overlapping situations. After low-frequency feature extraction, it was obvious that similar features were clustered more tightly. After high-low frequency feature fusion, and after the classifier, it was obvious that similar features were clustered, and different types of features were clearly separated, indicating that high-low frequency feature extraction enhanced the classification effect. [Conclusion] This network achieves a good balance between crop variety classification accuracy and model complexity, and is expected to be deployed on hardware systems with limited resources in the future to achieve real-time classification functions.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Differential Privacy-enhanced Blockchain-Based Quality Control Model for Rice
    WU Guodong, HU Quanxing, LIU Xu, QIN Hui, GAO Bowen
    Smart Agriculture    2024, 6 (4): 149-159.   DOI: 10.12133/j.smartag.SA202311027
    Abstract309)   HTML20)    PDF(pc) (1858KB)(969)       Save

    [Objective] Rice plays a crucial role in daily diet. The rice industry involves numerous links, from paddy planting to the consumer's table, and the integrity of the quality control data chain directly affects the credibility of rice quality control and traceability information. The process of rice traceability also faces security issues, such as the leakage of privacy information, which need immediate solutions. Additionally, the previous practice of uploading all information onto the blockchain leads to high storage costs and low system efficiency. To address these problems, this study proposed a differential privacy-enhanced blockchain-based quality control model for rice, providing new ideas and solutions to optimize the traditional quality regulation and traceability system. [Methods] By exploring technologies of blockchain, interplanetary file system (IPFS), and incorporating differential privacy techniques, a blockchain-based quality control model for rice with differential privacy enhancement was constructed. Firstly, the data transmission process was designed to cover the whole industry chain of rice, including cultivation, acquisition, processing, warehousing, and sales. Each module stored the relevant data and a unique number from the previous link, forming a reliable information chain and ensuring the continuity of the data chain for quality control. Secondly, to address the issue of large data volume and low efficiency of blockchain storage, the key quality control data of each link in the rice industry chain was stored in the IPFS. Subsequently, the hash value of the stored data was returned and recorded on the blockchain. Lastly, to enhance the traceability of the quality control model information, the sensitive information in the key quality control data related to the cultivation process was presented to users after undergoing differential privacy processing. Individual data was obfuscated to increase the credibility of the quality control information while also protecting the privacy of farmers' cultivation practices. Based on this model, a differential privacy-enhanced blockchain-based quality control system for rice was designed. [Results and Discussions] The architecture of the differential privacy-enhanced blockchain-based quality control system for rice consisted of the physical layer, transport layer, storage layer, service layer, and application layer. The physical layer included sensor devices and network infrastructure, ensuring data collection from all links of the industry chain. The transport layer handled data transmission and communication, securely uploading collected data to the cloud. The storage layer utilized a combination of traditional databases, IPFS, and blockchain to efficiently store and manage key data on and off the blockchain. The traditional database was used for the management and querying of structured data. IPFS stored the key quality control data in the whole industry chain, while blockchain was employed to store the hash values returned by IPFS. This integrated storage method improved system efficiency, ensured the continuity, reliability, and traceability of quality control data, and provided consumers with reliable information. The service layer was primarily responsible for handling business logic and providing functional services. The implementation of functions in the application layer relied heavily on the design of a series of interfaces within the service layer. Positioned at the top of the system architecture, the application layer was responsible for providing user-centric functionality and interfaces. This encompassed a range of applications such as web applications and mobile applications, aiming to present data and facilitate interactive features to fulfill the requirements of both consumers and businesses. Based on the conducted tests, the average time required for storing data in a single link of the whole industry chain within the system was 1.125 s. The average time consumed for information traceability query was recorded as 0.691 s. Compared to conventional rice quality regulation and traceability systems, the proposed system demonstrated a reduction of 6.64% in the storage time of single-link data and a decrease of 16.44% in the time required to perform information traceability query. [Conclusions] This study proposes a differential privacy-enhanced blockchain-based quality control model for rice. The model ensures the continuity of the quality control data chain by integrating the various links of the whole industry chain of rice. By combining blockchain with IPFS storage, the model addresses the challenges of large data volume and low efficiency of blockchain storage in traditional systems. Furthermore, the model incorporates differential privacy protection to enhance traceability while safeguarding the privacy of individual farmers. This study can provide reference for the design and improvement of rice quality regulation and traceability systems.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Parametric Reconstruction Method of Wheat Leaf Curved Surface Based on Three-Dimensional Point Cloud
    ZHU Shunyao, QU Hongjun, XIA Qian, GUO Wei, GUO Ya
    Smart Agriculture    2025, 7 (1): 85-96.   DOI: 10.12133/j.smartag.SA202410004
    Abstract376)   HTML48)    PDF(pc) (2475KB)(930)       Save

    [Objective] Plant leaf shape is an important part of plant architectural model. Establishment of a three-dimensional structural model of leaves may assist simulating and analyzing plant growth. However, existing leaf modeling approaches lack interpretability, invertibility, and operability, which limit the estimation of model parameters, the simulation of leaf shape, the analysis and interpretation of leaf physiology and growth state, and model reusage. Aiming at the interoperability between three-dimensional structure representation and mathematical model parameters, this study paid attention to three aspects in wheat leaf shape parametric reconstruction: (1) parameter-driven model structure, (2) model parameter inversion, and (3) parameter dynamic mapping during growth. Based on this, a set of parameter-driven and point cloud inversion model for wheat leaf interoperability was proposed in this study. [Methods] A parametric surface model of a wheat leaf with seven characteristic parameters by using parametric modeling technology was built, and the forward parametric construction of the wheat leaf structure was realized. Three parameters, maximum leaf width, leaf length, and leaf shape factor, were used to describe the basic shape of the blade on the leaf plane. On this basis, two parameters, namely the angle between stems and leaves and the curvature degree, were introduced to describe the bending characteristics of the main vein of the blade in the three-dimensional space. Two parameters, namely the twist angle around the axis and the twist deviation angle around the axis, were introduced to represent the twisted structure of the leaf blade along the vein. The reverse parameter estimation module was built according to the surface model. The point cloud was divided by the uniform segmentation method along the Y-axis, and the veins were fit by a least squares regression method. Then, the point cloud was re-segmented according to the fit vein curve. Subsequently, the rotation angle was precisely determined through the segment-wise transform estimation method, with all parameters being optimally fit using the RANSAC regression algorithm. To validate the reliability of the proposed methodology, a set of sample parameters was randomly generated, from which corresponding sample point clouds were synthesized. These sample point clouds were then subjected to estimation using the described method. Then error analyzing was carried out on the estimation results. Three-dimensional imaging technology was used to collect the point clouds of Zhengmai 136, Yangmai 34, and Yanmai 1 samples. After noise reduction and coordinate registration, the model parameters were inverted and estimated, and the reconstructed point clouds were produced using the parametric model. The reconstruction error was validated by calculating the dissimilarity, represented by the Chamfer Distance, between the reconstructed point cloud and the measured point cloud. [Results and Discussions] The model could effectively reconstruct wheat leaves, and the average deviation of point cloud based parametric reconstruction results was about 1.2 mm, which had a high precision. Parametric modeling technology based on prior knowledge and point cloud fitting technology based on posterior data was integrated in this study to construct a digital twin model of specific species at the 3D structural level. Although some of the detailed characteristics of the leaves were moderately simplified, the geometric shape of the leaves could be highly restored with only a few parameters. This method was not only simple, direct and efficient, but also had more explicit geometric meaning of the obtained parameters, and was both editable and interpretable. In addition, the practice of using only tools such as rulers to measure individual characteristic parameters of plant organs in traditional research was abandoned in this study. High-precision point cloud acquisition technology was adopted to obtain three-dimensional data of wheat leaves, and pre-processing work such as point cloud registration, segmentation, and annotation was completed, laying a data foundation for subsequent research. [Conclusions] There is interoperability between the reconstructed model and the point cloud, and the parameters of the model can be flexibly adjusted to generate leaf clusters with similar shapes. The inversion parameters have high interpretability and can be used for consistent and continuous estimation of point cloud time series. This research is of great value to the simulation analysis and digital twinning of wheat leaves.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Low-Cost Chlorophyll Fluorescence Imaging System Applied in Plant Physiology Status Detection
    YANG Zhenyu, TANG Hao, GE Wei, XIA Qian, TONG Dezhi, FU Lijiang, GUO Ya
    Smart Agriculture    2023, 5 (3): 154-165.   DOI: 10.12133/j.smartag.SA202306006
    Abstract531)   HTML42)    PDF(pc) (1735KB)(917)       Save

    [Objective] Chlorophyll fluorescence (ChlF) emission from photosystem II (PSII) is closely coupled with photochemical reactions. As an efficient and non-destructive means of obtaining plant photosynthesis efficiency and physiological state information, the collection of fluorescence signals is often used in many fields such as plant physiological research, smart agricultural information sensing, etc. Chlorophyll fluorescence imaging systems, which is the experimental device for collecting the fluorescence signal, have difficulties in application due to their high price and complex structure. In order to solve the issues, this paper investigates and constructs a low-cost chlorophyll fluorescence imaging system based on a micro complementary metal oxide semiconductor (CMOS) camera and a smartphone, and carries out experimental verifications and applications on it. [Method] The chlorophyll fluorescence imaging system is mainly composed of three parts: excitation light, CMOS camera and its control circuit, and a upper computer based on a smartphone. The light source of the excitation light group is based on the principle and characteristics of chlorophyll fluorescence, and uses a blue light source of 460 nm band to achieve the best fluorescence excitation effect. In terms of structure, the principle of integrating sphere was borrowed, the bowl-shaped light source structure was adopted, and the design of the LED surface light source was used to meet the requirements of chlorophyll fluorescence signal measurement for the uniformity of the excitation light field. For the adjustment of light source intensity, the control scheme of pulse width modulation was adopted, which could realize sequential control of different intensities of excitation light. Through the simulation analysis of the light field, the light intensity and distribution characteristics of the light field were stuidied, and the calibration of the excitation light group was completed according to the simulation results. The OV5640 micro CMOS camera was used to collect fluorescence images. Combined with the imaging principle of the CMOS camera, the fluorescence imaging intensity of the CMOS camera was calculated, and its ability to collect chlorophyll fluorescence was analyzed and discussed. The control circuit of the CMOS camera uses an STM32 microcontroller as the microcontroller unit, and completes the data communication between the synchronous light group control circuit and the smartphone through the RS232 to TTL serial communication module and the full-speed universal serial bus, respectively. The smartphone upper computer software is the operating software of the chlorophyll fluorescence imaging system user terminal and the overall control program for fluorescence image acquisition. The overall workflow could be summarized as the user sets the relevant excitation light parameters and camera shooting instructions in the upper computer as needed, sends the instructions to the control circuit through the universal serial bus and serial port, and completes the control of excitation light and CMOS camera image acquisition. After the chlorophyll fluorescence image collection was completed, the data would be sent back to the smart phone or server for analysis, processing, storage, and display. In order to verify the design of the proposed scheme, a prototype of the chlorophyll fluorescence imaging system based on this scheme was made for experimental verification. Firstly, the uniformity of the light field was measured on the excitation light to test the actual performance of the excitation light designed in this article. On this basis, a chlorophyll fluorescence imaging experiment under continuous light excitation and modulated pulse light protocols was completed. Through the analysis and processing of the experimental results and comparison with mainstream chlorophyll fluorometers, the fluorescence imaging capabilities and low-cost advantages of this chlorophyll fluorometer were further verified. [Results and Discussions] The maximum excitation light intensity of the chlorophyll fluorescence imaging system designed in this article was 6250 µmol/(m2·s). Through the simulation analysis of the light field and the calculation and analysis of the fluorescence imaging intensity of the CMOS camera, the feasibility of collecting chlorophyll fluorescence images by the OV5640 micro CMOS camera was demonstrated, which provided a basis for the specific design and implementation of the fluorometer. In terms of hardware circuits, it made full use of the software and hardware advantages of smartphones, and only consisted of the control circuits of the excitation light and CMOS camera and the corresponding communication modules to complete the fluorescence image collection work, simplifying the circuit structure and reducing hardware costs to the greatest extent. The final fluorescence instrument achieved a collection resolution of 5 million pixels, a spectral range of 400~1000 nm, and a stable acquisition frequency of up to 42 f/s. Experimental results showed that the measured data was consistent with theoretical analysis and simulation, which could meet the requirements of fluorescence detection. The instrument was capable of collecting images of chlorophyll fluorescence under continuous light excitation or the protocol of modulated pulsed light. The acquired chlorophyll fluorescence images could reflect the two-dimensional heterogeneity of leaves and could effectively distinguish the photosynthetic characteristics of different leaves. Typical chlorophyll fluorescence parameter images of Fv/Fm, Rfd, etc. were in line with expectations. Compared with the existing chlorophyll fluorescence imaging system, the chlorophyll fluorescence imaging system designed in this article has obvious cost advantages while realizing the rapid detection function of chlorophyll fluorescence. [Conclusions] The instrument is with a simple structure and low cost, and has good application value for the detection of plant physiology and environmental changes. The system is useful for developing other fluorescence instruments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Navigation and Spraying Robot in Sheep Farm
    FAN Mingshuo, ZHOU Ping, LI Miao, LI Hualong, LIU Xianwang, MA Zhirun
    Smart Agriculture    2024, 6 (4): 103-115.   DOI: 10.12133/j.smartag.SA202312016
    Abstract1069)   HTML33)    PDF(pc) (2160KB)(916)       Save

    [Objective] Manual disinfection in large-scale sheep farm is laborious, time-consuming, and often results in incomplete coverage and inadequate disinfection. With the rapid development of the application of artificial intelligence and automation technology, the automatic navigation and spraying robot for livestock and poultry breeding, has become a research hotspot. To maintain shed hygiene and ensure sheep health, an automatic navigation and spraying robot was proposed for sheep sheds. [Methods] The automatic navigation and spraying robot was designed with a focus on three aspects: hardware, semantic segmentation model, and control algorithm. In terms of hardware, it consisted of a tracked chassis, cameras, and a collapsible spraying device. For the semantic segmentation model, enhancements were made to the lightweight semantic segmentation model ENet, including the addition of residual structures to prevent network degradation and the incorporation of a squeeze-and-excitation network (SENet) attention mechanism in the initialization module. This helped to capture global features when feature map resolution was high, addressing precision issues. The original 6-layer ENet network was reduced to 5 layers to balance the encoder and decoder. Drawing inspiration from dilated spatial pyramid pooling, a context convolution module (CCM) was introduced to improve scene understanding. A criss-cross attention (CCA) mechanism was adapted to acquire context global features of different scales without cascading, reducing information loss. This led to the development of a double attention enet (DAENet) semantic segmentation model was proposed to achieve real-time and accurate segmentation of sheep shed surfaces. Regarding control algorithms, a method was devised to address the robot's difficulty in controlling its direction at junctions. Lane recognition and lane center point identification algorithms were proposed to identify and mark navigation points during the robot's movement outside the sheep shed by simulating real roads. Two cameras were employed, and a camera switching algorithm was developed to enable seamless switching between them while also controlling the spraying device. Additionally, a novel offset and velocity calculation algorithm was proposed to control the speeds of the robot's left and right tracks, enabling control over the robot's movement, stopping, and turning. [Results and Discussions] The DAENet model achieved a mean intersection over union (mIoU) of 0.945 3 in image segmentation tasks, meeting the required segmentation accuracy. During testing of the camera switching algorithm, it was observed that the time taken for the complete transition from camera to spraying device action does not exceed 15 seconds when road conditions changed. Testing of the center point and offset calculation algorithm revealed that, when processing multiple frames of video streams, the algorithm averages 0.04 to 0.055 per frame, achieving frame rates of 20 to 24 frames per second, meeting real-time operational requirements. In field experiments conducted in sheep farm, the robot successfully completed automatic navigation and spraying tasks in two sheds without colliding with roadside troughs. The deviation from the road and lane centerlines did not exceed 0.3 meters. Operating at a travel speed of 0.2 m/s, the liquid in the medicine tank was adequate to complete the spraying tasks for two sheds. Additionally, the time taken for the complete transition from camera to spraying device action did not exceed 15 when road conditions changed. The robot maintained an average frame rate of 22.4 frames per second during operation, meeting the experimental requirements for accurate and real-time information processing. Observation indicated that the spraying coverage rate of the robot exceeds 90%, meeting the experimental coverage requirements. [Conclusions] The proposed automatic navigation and spraying robot, based on the DAENet semantic segmentation model and center point recognition algorithm, combined with hardware design and control algorithms, achieves comprehensive disinfection within sheep sheds while ensuring safety and real-time operation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progresses of Crop Growth Monitoring Based on Synthetic Aperture Radar Data
    HONG Yujiao, ZHANG Shuo, LI Li
    Smart Agriculture    2024, 6 (1): 46-62.   DOI: 10.12133/j.smartag.SA202308019
    Abstract538)   HTML80)    PDF(pc) (1147KB)(914)       Save

    Significance Crop production is related to national food security, economic development and social stability, so timely information on the growth of major crops is of great significance for strengthening the crop production management and ensuring food security. The traditional crop growth monitoring mainly judges the growth of crops by manually observing the shape, color and other appearance characteristics of crops through the external industry, which has better reliability and authenticity, but it will consume a lot of manpower, is inefficient and difficult to carry out monitoring of a large area. With the development of space technology, satellite remote sensing technology provides an opportunity for large area crop growth monitoring. However, the acquisition of optical remote sensing data is often limited by the weather during the peak crop growth season when rain and heat coincide. Synthetic aperture radar (SAR) compensates well for the shortcomings of optical remote sensing, and has a wide demand and great potential for application in crop growth monitoring. However, the current research on crop growth monitoring using SAR data is still relatively small and lacks systematic sorting and summarization. In this paper, the research progress of SAR inversion of crop growth parameters were summarized through comprehensive analysis of existing literature, clarify the main technical methods and application of SAR monitoring of crop growth, and explore the existing problems and look forward to its future research direction. Progress] The current research status of SAR crop growth monitoring were reviewed, the application of SAR technology had gone through several development stages: from the early single-polarization, single-band stage, gradually evolving to the mid-term multi-polarization, multi-band stage, and then to the stage of joint application of tight polarization and optical remote sensing. Then, the research progress and milestone achievements of crop growth monitoring based on SAR data were summarized in three aspects, namely, crop growth SAR remote sensing monitoring indexes, crop growth SAR remote sensing monitoring data and crop growth SAR remote sensing monitoring methods. First, the key parameters of crop growth were summarized, and the crop growth monitoring indexes were divided into morphological indicators, physiological and biochemical indicators, yield indicators and stress indicators. Secondly, the core principle of SAR monitoring of crop growth parameters was introduced, which was based on the interaction between SAR signals and vegetation, and then the specific scattering model and inversion algorithm were used to estimate the crop growth parameters. Then, a detailed summary and analysis of the radar indicators mainly applied to crop growth monitoring were also presented. Finally, SAR remote sensing methods for crop growth monitoring, including mechanistic modeling, empirical modeling, semi-empirical modeling, direct monitoring, and assimilation monitoring of crop growth models, were described, and their applicability and applications in growth monitoring were analyzed. Conclusions and Prospects Four challenges exist in SAR crop growth monitoring are proposed: 1) Compared with the methods of crop growth monitoring using optical remote sensing data, the methods of crop growth monitoring using SAR data are obviously relatively small. The reason may be that SAR remote sensing itself has some inherent shortcomings; 2) Insufficient mining of microwave scattering characteristics, at present, a large number of studies have applied the backward scattering intensity and polarization characteristics to crop growth monitoring, but few have applied the phase information to crop growth monitoring, especially the application study of polarization decomposition parameters to growth monitoring. The research on the application of polarization decomposition parameter to crop growth monitoring is still to be deepened; 3) Compared with the optical vegetation index, the radar vegetation index applied to crop growth monitoring is relatively less; 4 ) Crop growth monitoring based on SAR scattered intensity is mainly based on an empirical model, which is difficult to be extended to different regions and types of crops, and the existence of this limitation prevents the SAR scattering intensity-based technology from effectively realizing its potential in crop growth monitoring. Finally, future research should focus on mining microwave scattering features, utilizing SAR polarization decomposition parameters, developing and optimizing radar vegetation indices, and deepening scattering models for crop growth monitoring.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Individual Tree Skeleton Extraction and Crown Prediction Method of Winter Kiwifruit Trees
    LI Zhengkai, YU Jiahui, PAN Shijia, JIA Zefeng, NIU Zijie
    Smart Agriculture    2023, 5 (4): 92-104.   DOI: 10.12133/j.smartag.SA202308015
    Abstract381)   HTML18)    PDF(pc) (2529KB)(907)       Save

    [Objective] The proliferation of kiwifruit trees severely overlaps, resulting in a complex canopy structure, rendering it impossible to extract their skeletons or predict their canopies using conventional methods. The objective of this research is to propose a crown segmentation method that integrates skeleton information by optimizing image processing algorithms and developing a new scheme for fusing winter and summer information. In cases where fruit trees are densely distributed, achieving accurate segmentation of fruit tree canopies in orchard drone images can efficiently and cost-effectively obtain canopy information, providing a foundation for determining summer kiwifruit growth size, spatial distribution, and other data. Furthermore, it facilitates the automation and intelligent development of orchard management. [Methods] The 4- to 8-year-old kiwifruit trees were chosen and remote sensing images of winter and summer via unmanned aerial vehicles were obtain as the primary analysis visuals. To tackle the challenge of branch extraction in winter remote sensing images, a convolutional attention mechanism was integrated into the PSP-Net network, along with a joint attention loss function. This was designed to boost the network's focus on branches, enhance the recognition and targeting capabilities of the target area, and ultimately improve the accuracy of semantic segmentation for fruit tree branches.For the generation of the skeleton, digital image processing technology was employed for screening. The discrete information of tree branches was transformed into the skeleton data of a single fruit tree using growth seed points. Subsequently, the semantic segmentation results were optimized through mathematical morphology calculations, enabling smooth connection of the branches. In response to the issue of single tree canopy segmentation in summer, the growth characteristics of kiwifruit trees were taken into account, utilizing the outward expansion of branches growing from the trunk.The growth of tree branches was simulated by using morphological expansion to predict the summer canopy. The canopy prediction results were analyzed under different operators and parameters, and the appropriate expansion operators along with their corresponding operation lengths were selected. The skeleton of a single tree was extracted from summer images. By combining deep learning with mathematical morphology methods through the above steps, the optimized single tree skeleton was used as a prior condition to achieve canopy segmentation. [Results and Discussions] In comparison to traditional methods, the accuracy of extracting kiwifruit tree canopy information images at each stage of the process has been significantly enhanced. The enhanced PSP Net was evaluated using three primary regression metrics: pixel accuracy (PA), mean intersection over union ratio (MIoU), and weighted F1 Score (WF1). The PA, MIoU and WF1 of the improved PSP-Net were 95.84%, 95.76% and 95.69% respectively, which were increased by 12.30%, 22.22% and 17.96% compared with U-Net, and 21.39% , 21.51% and 18.12% compared with traditional PSP-Net, respectively. By implementing this approach, the skeleton extraction function for a single fruit tree was realized, with the predicted PA of the canopy surpassing 95%, an MIoU value of 95.76%, and a WF1 of canopy segmentation approximately at 94.07%.The average segmentation precision of the approach surpassed 95%, noticeably surpassing the original skeleton's 81.5%. The average conformity between the predicted skeleton and the actual summer skeleton stand at 87%, showcasing the method's strong prediction performance. Compared with the original skeleton, the PA, MIoU and WF1 of the optimized skeleton increased by 13.2%, 10.9% and 18.4%, respectively. The continuity of the predicted skeleton had been optimized, resulting in a significant improvement of the canopy segmentation index. The solution effectively addresses the issue of semantic segmentation fracture, and a single tree canopy segmentation scheme that incorporates skeleton information could effectively tackle the problem of single fruit tree canopy segmentation in complex field environments. This provided a novel technical solution for efficient and low-cost orchard fine management. [Conclusions] A method for extracting individual kiwifruit plant skeletons and predicting canopies based on skeleton information was proposed. This demonstrates the enormous potential of drone remote sensing images for fine orchard management from the perspectives of method innovation, data collection, and problem solving. Compared with manual statistics, the overall efficiency and accuracy of kiwifruit skeleton extraction and crown prediction have significantly improved, effectively solving the problem of case segmentation in the crown segmentation process.The issue of semantic segmentation fragmentation has been effectively addressed, resulting in the development of a single tree canopy segmentation method that incorporates skeleton information. This approach can effectively tackle the challenges of single fruit tree canopy segmentation in complex field environments, thereby offering a novel technical solution for efficient and cost-effective orchard fine management. While the research is primarily centered on kiwifruit trees, the methodology possesses strong universality. With appropriate modifications, it can be utilized to monitor canopy changes in other fruit trees, thereby showcasing vast application potential.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Identification Method of Kale Leaf Ball Based on Improved UperNet
    ZHU Yiping, WU Huarui, GUO Wang, WU Xiaoyan
    Smart Agriculture    2024, 6 (3): 128-137.   DOI: 10.12133/j.smartag.SA202401020
    Abstract464)   HTML114)    PDF(pc) (1568KB)(904)       Save

    [Objective] Kale is an important bulk vegetable crop worldwide, its main growth characteristics are outer leaves and leaf bulbs. The traits of leaf bulb kale are crucial for adjusting water and fertilizer parameters in the field to achieve maximum yield. However, various factors such as soil quality, light exposure, leaf overlap, and shading can affect the growth of in practical field conditions. The similarity in color and texture between leaf bulbs and outer leaves complicates the segmentation process for existing recognition models. In this paper, the segmentation of kale outer leaves and leaf bulbs in complex field background was proposed, using pixel values to determine leaf bulb size for intelligent field management. A semantic segmentation algorithm, UperNet-ESA was proposed to efficiently and accurately segment nodular kale outer leaf and leaf bulb in field scenes using the morphological features of the leaf bulbs and outer leaves of nodular kale to realize the intelligent management of nodular kale in the field. [Methods] The UperNet-ESA semantic segmentation algorithm, which uses the unified perceptual parsing network (UperNet) as an efficient semantic segmentation framework, is more suitable for extracting crop features in complex environments by integrating semantic information across different scales. The backbone network was improved using ConvNeXt, which is responsible for feature extraction in the model. The similarity between kale leaf bulbs and outer leaves, along with issues of leaf overlap affecting accurate target contour localization, posed challenges for the baseline network, leading to low accuracy. ConvNeXt effectively combines the strengths of convolutional neural networks (CNN) and Transformers, using design principles from Swin Transformer and building upon ResNet50 to create a highly effective network structure. The simplicity of the ConvNeXt design not only enhances segmentation accuracy with minimal model complexity, but also positions it as a top performer among CNN architectures. In this study, the ConvNeXt-B version was chosen based on considerations of computational complexity and the background characteristics of the knotweed kale image dataset. To enhance the model's perceptual acuity, block ratios for each stage were set at 3:3:27:3, with corresponding channel numbers of 128, 256, 512 and 1 024, respectively. Given the visual similarity between kale leaf bulbs and outer leaves, a high-efficiency channel attention mechanism was integrated into the backbone network to improve feature extraction in the leaf bulb region. By incorporating attention weights into feature mapping through residual inversion, attention parameters were cyclically trained within each block, resulting in feature maps with attentional weights. This iterative process facilitated the repeated training of attentional parameters and enhanced the capture of global feature information. To address challenges arising from direct pixel addition between up-sampling and local features, potentially leading to misaligned context in feature maps and erroneous classifications at kale leaf boundaries, a feature alignment module and feature selection module were introduced into the feature pyramid network to refine target boundary information extraction and enhance model segmentation accuracy. [Results and Discussions] The UperNet-ESA semantic segmentation model outperforms the current mainstream UNet model, PSPNet model, DeepLabV3+ model in terms of segmentation accuracy, where mIoU and mPA reached 92.45% and 94.32%, respectively, and the inference speed of up to 16.6 frames per second (fps). The mPA values were better than that of the UNet model, PSPNet model, ResNet-50 based, MobilenetV2, and DeepLabV3+ model with Xception as the backbone, showing improvements of 11.52%, 13.56%, 8.68%, 4.31%, and 6.21%, respectively. Similarly, the mIoU exhibited improvements of 12.21%, 13.04%, 10.65%, 3.26% and 7.11% compared to the mIoU of the UNet-based model, PSPNet model, and DeepLabV3+ model based on the ResNet-50, MobilenetV2, and Xception backbones, respectively. This performance enhancement can be attributed to the introduction of the ECA module and the improvement made to the feature pyramid network in this model, which strengthen the judgement of the target features at each stage to obtain effective global contextual information. In addition, although the PSPNet model had the fastest inference speed, the overall accuracy was too low to for developing kale semantic segmentation models. On the contrary, the proposed model exhibited superior inference speed compared to all other network models. [Conclusions] The experimental results showed that the UperNet-ESA semantic segmentation model proposed in this study outperforms the original network in terms of performance. The improved model achieves the best accuracy-speed balance compared to the current mainstream semantic segmentation networks. In the upcoming research, the current model will be further optimized and enhanced, while the kale dataset will be expanded to include a wider range of samples of nodulated kale leaf bulbs. This expansion is intended to provide a more robust and comprehensive theoretical foundation for intelligent kale field management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Adaptive Time Horizon MPC Path Tracking Control Method for Mowing Robot
    HE Qing, JI Jie, FENG Wei, ZHAO Lijun, ZHANG Bohan
    Smart Agriculture    2024, 6 (3): 82-93.   DOI: 10.12133/j.smartag.SA202401010
    Abstract441)   HTML32)    PDF(pc) (2278KB)(897)       Save

    [Objective] The traditional predictive control approach usually employs a fixed time horizon and often overlooks the impact of changes in curvature and road bends. This oversight leads to subpar tracking performance and inadequate adaptability of robots for navigating curves and paths. Although extending the time horizon of the standard fixed time horizon model predictive control (MPC) can improve curve path tracking accuracy, it comes with high computational costs, making it impractical in situations with restricted computing resources. Consequently, an adaptive time horizon MPC controller was developed to meet the requirements of complex tasks such as autonomous mowing. [Methods] Initially, it was crucial to establish a kinematic model for the mowing robot, which required employing Taylor linearization and Euler method discretization techniques to ensure accurate path tracking. The prediction equation for the error model was derived after conducting a comprehensive analysis of the robot's kinematics model employed in mowing. Second, the size of the previewing area was determined by utilizing the speed data and reference path information gathered from the mowing robot. The region located a certain distance ahead of the robot's current position, was identified to as the preview region, enabling a more accurate prediction of the robot's future traveling conditions. Calculations for both the curve factor and curve change factor were carried out within this preview region. The curvature factor represented the initial curvature of the path, while the curvature change factor indicated the extent of curvature variation in this region. These two variables were then fed into a fuzzy controller, which adjusted the prediction time horizon of the MPC. The integration enabled the mowing robot to promptly adjust to changes in the path's curvature, thereby improving its accuracy in tracking the desired trajectory. Additionally, a novel technique for triggering MPC execution was developed to reduce computational load and improve real-time performance. This approach ensured that MPC activation occurred only when needed, rather than at every time step, resulting in reduced computational expenses especially during periods of smooth robot motion where unnecessary computation overhead could be minimized. By meeting kinematic and dynamic constraints, the optimization algorithm successfully identified an optimal control sequence, ultimately enhancing stability and reliability of the control system. Consequently, these set of control algorithms facilitated precise path tracking while considering both kinematic and dynamic limitations in complex environments. [Results and Discussion] The adaptive time-horizon MPC controller effectively limited the maximum absolute heading error and maximum absolute lateral error to within 0.13 rad and 11 cm, respectively, surpassing the performance of the MPC controller in the control group. Moreover, compared to both the first and fourth groups, the adaptive time-horizon MPC controller achieved a remarkable reduction of 75.39% and 57.83% in mean values for lateral error and heading error, respectively (38.38% and 31.84%, respectively). Additionally, it demonstrated superior tracking accuracy as evidenced by its significantly smaller absolute standard deviation of lateral error (0.025 6 m) and course error (0.025 5 rad), outperforming all four fixed time-horizon MPC controllers tested in the study. Furthermore, this adaptive approach ensured precise tracking and control capabilities for the mowing robot while maintaining a remarkably low average solution time of only 0.004 9 s, notably faster than that observed with other control data sets-reducing computational load by approximately 10.9 ms compared to maximum time-horizon MPC. [Conclusions] The experimental results demonstrated that the adaptive time-horizon MPC tracking approach effectively addressed the trade-off between control accuracy and computational complexity encountered in fixed time-horizon MPC. By dynamically adjusting the time horizon length the and performing MPC calculations based on individual events, this approach can more effectively handle scenarios with restricted computational resources, ensuring superior control precision and stability. Furthermore, it achieves a balance between control precision and real-time performance in curve route tracking for mowing robots, offering a more practical and reliable solution for their practical application.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Market Monitoring and Early Warning: An Integrated Forecasting Approach Based on Deep Learning
    XU Shiwei, LI Qianchuan, LUAN Rupeng, ZHUANG Jiayu, LIU Jiajia, XIONG Lu
    Smart Agriculture    2025, 7 (1): 57-69.   DOI: 10.12133/j.smartag.SA202411004
    Abstract380)   HTML29)    PDF(pc) (1936KB)(889)       Save

    [Significance] The fluctuations in the supply, consumption, and prices of agricultural products directly affect market monitoring and early warning systems. With the ongoing transformation of China's agricultural production methods and market system, advancements in data acquisition technologies have led to an explosive growth in agricultural data. However, the complexity of the data, the narrow applicability of existing models, and their limited adaptability still present significant challenges in monitoring and forecasting the interlinked dynamics of multiple agricultural products. The efficient and accurate forecasting of agricultural market trends is critical for timely policy interventions and disaster management, particularly in a country with a rapidly changing agricultural landscape like China. Consequently, there is a pressing need to develop deep learning models that are tailored to the unique characteristics of Chinese agricultural data. These models should enhance the monitoring and early warning capabilities of agricultural markets, thus enabling precise decision-making and effective emergency responses. [Methods] An integrated forecasting methodology was proposed based on deep learning techniques, leveraging multi-dimensional agricultural data resources from China. The research introduced several models tailored to different aspects of agricultural market forecasting. For production prediction, a generative adversarial network and residual network collaborative model (GAN-ResNet) was employed. For consumption forecasting, a variational autoencoder and ridge regression (VAE-Ridge) model was used, while price prediction was handled by an Adaptive-Transformer model. A key feature of the study was the adoption of an "offline computing and visualization separation" strategy within the Chinese agricultural monitoring and early warning system (CAMES). This strategy ensures that model training and inference are performed offline, with the results transmitted to the front-end system for visualization using lightweight tools such as ECharts. This approach balances computational complexity with the need for real-time early warnings, allowing for more efficient resource allocation and faster response times. The corn, tomato, and live pig market data used in this study covered production, consumption and price data from 1980 to 2023, providing comprehensive data support for model training. [Results and Discussions] The deep learning models proposed in this study significantly enhanced the forecasting accuracy for various agricultural products. For instance, the GAN-ResNet model, when used to predict maize yield at the county level, achieved a mean absolute percentage error (MAPE) of 6.58%. The VAE-Ridge model, applied to pig consumption forecasting, achieved a MAPE of 6.28%, while the Adaptive-Transformer model, used for tomato price prediction, results in a MAPE of 2.25%. These results highlighted the effectiveness of deep learning models in handling complex, nonlinear relationships inherent in agricultural data. Additionally, the models demonstrate notable robustness and adaptability when confronted with challenges such as sparse data, seasonal market fluctuations, and heterogeneous data sources. The GAN-ResNet model excels in capturing the nonlinear fluctuations in production data, particularly in response to external factors such as climate conditions. Its capacity to integrate data from diverse sources—including weather data and historical yield data—made it highly effective for production forecasting, especially in regions with varying climatic conditions. The VAE-Ridge model addressed the issue of data sparsity, particularly in the context of consumption data, and provided valuable insights into the underlying relationships between market demand, macroeconomic factors, and seasonal fluctuations. Finally, the Adaptive-Transformer model stand out in price prediction, with its ability to capture both short-term price fluctuations and long-term price trends, even under extreme market conditions. [Conclusions] This study presents a comprehensive deep learning-based forecasting approach for agricultural market monitoring and early warning. The integration of multiple models for production, consumption, and price prediction provides a systematic, effective, and scalable tool for supporting agricultural decision-making. The proposed models demonstrate excellent performance in handling the nonlinearities and seasonal fluctuations characteristic of agricultural markets. Furthermore, the models' ability to process and integrate heterogeneous data sources enhances their predictive power and makes them highly suitable for application in real-world agricultural monitoring systems. Future research will focus on optimizing model parameters, enhancing model adaptability, and expanding the system to incorporate additional agricultural products and more complex market conditions. These improvements will help increase the stability and practical applicability of the system, thus further enhancing its potential for real-time market monitoring and early warning capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Tomato Leaf Disease and Pest Detection Method Based on Improved YOLOv10n
    WU Liuai, XU Xueke
    Smart Agriculture    2025, 7 (1): 146-155.   DOI: 10.12133/j.smartag.SA202410023
    Abstract555)   HTML97)    PDF(pc) (1834KB)(865)       Save

    [Objective] To address the challenges in detecting tomato leaf diseases and pests, such as complex environments, small goals, low precision, redundant parameters, and high computational complexity, a novel lightweight, high-precision, real-time detection model was proposed called YOLOv10n-YS. This model aims to accurately identify diseases and pests, thereby providing a solid scientific basis for their prevention and management strategies. Methods] The dataset was collected using mobile phones to capture images from multiple angles under natural conditions, ensuring complete and clear leaf images. It included various weather conditions and covered nine types: Early blight, leaf mold, mosaic virus, septoria, spider mites damage, yellow leaf curl virus, late blight, leaf miner disease, and healthy leaves, with all images having a resolution of 640×640 pixels. In the proposed YOLOv10n-YS model, firstly, the C2f in the backbone network was replaced with C2f_RepViTBlock, thereby reducing the computational load and parameter volume and achieving a lightweight design. Secondly, through the introduction of a sliced operation SimAM attention mechanism, the Conv_SWS module was formed, which enhanced the extraction of small target features. Additionally, the DySample lightweight dynamic up sampling module was used to replace the up sampling module in the neck network, concentrating sampling points on target areas and ignoring backgrounds, thereby effectively identifying defects. Finally, the efficient channel attention (ECA) was improved by performing average pooling and max pooling on the input layer to aggregate features and then adding them together, which further enhanced global perspective information and features of different scales. The improved module, known as efficient channel attention with cross-channel interaction (EMCA) attention, was introduced, and the pyramid spatial attention (PSA) in the backbone network was replaced with the EMCA attention mechanism, thereby enhancing the feature extraction capability of the backbone network. [Results and Discussions] After introducing the C2f_RepViTBlock, the model's parameter volume and computational load were reduced by 12.3% and 9.7%, respectively, with mAP@0.5 and F1-Score each increased by 0.2% and 0.3%. Following the addition of the Conv_SWS and the replacement of the original convolution, mAP@0.5 and F1-Score were increased by 1.2% and 2%, respectively, indicating that the Conv_SWS module significantly enhanced the model's ability to extract small target features. After the introduction of DySample, mAP@0.5 and F1-Score were increased by 1.8% and 2.6%, respectively, but with a slight increase in parameter volume and computational load. Finally, the addition of the EMCA attention mechanism further enhanced the feature extraction capability of the backbone network. Through these four improvements, the YOLOv10n-YS model was formed. Compared with the YOLOv10n algorithm, YOLOv10n-YS reduced parameter volume and computational load by 13.8% and 8.5%, respectively, with both mAP@0.5 and F1-Score increased. These improvements not only reduced algorithm complexity but also enhanced detection accuracy, making it more suitable for industrial real-time detection. The detection accuracy of tomato diseases and pests using the YOLOv10n-YS algorithm was significantly better than that of comparative algorithms, and it had the lowest model parameter volume and computational load. The visualization results of detection by different models showed that the YOLOv10n-YS network could provide technical support for the detection and identification of tomato leaf diseases and pests. To verify the performance and robustness of the YOLOv10n-YS algorithm, comparative experiments were conducted on the public Plant-Village-9 dataset with different algorithms. The results showed that the average detection accuracy of YOLOv10n-YS on the Plant-Village dataset reached 91.1%, significantly higher than other algorithms. [Conclusions] The YOLOv10n-YS algorithm is not only characterized by occupying a small amount of space but also by possessing high recognition accuracy. On the tomato leaf dataset, excellent performance was demonstrated by this algorithm, thereby verifying its broad applicability and showcasing its potential to play an important role in large-scale crop pest and disease detection applications. Deploying the model on drone platforms and utilizing multispectral imaging technology can achieve real-time detection and precise localization of pests and diseases in complex field environments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    An Rapeseed Unmanned Seeding System Based on Cloud-Terminal High Precision Maps
    LU Bang, DONG Wanjing, DING Youchun, SUN Yang, LI Haopeng, ZHANG Chaoyu
    Smart Agriculture    2023, 5 (4): 33-44.   DOI: 10.12133/j.smartag.SA202310004
    Abstract409)   HTML50)    PDF(pc) (2408KB)(828)       Save

    [Objective] Unmanned seeding of rapeseed is an important link to construct unmanned rapeseed farm. Aiming at solving the problems of cumbersome manual collection of small and medium-sized field boundary information in the south, the low efficiency of turnaround operation of autonomous tractor, and leaving a large leakage area at the turnaround point, this study proposes to build an unmanned rapeseed seeding operation system based on cloud-terminal high-precision maps, and to improve the efficiency of the turnaround operation and the coverage of the operation. [Methods] The system was mainly divided into two parts: the unmanned seeding control cloud platform for oilseed rape is mainly composed of a path planning module, an operation monitoring module and a real-time control module; the navigation and control platform for rapeseed live broadcasting units is mainly composed of a Case TM1404 tractor, an intelligent seeding and fertilizing machine, an angle sensor, a high-precision Beidou positioning system, an electric steering wheel, a navigation control terminal and an on-board controller terminal. The process of constructing the high-precision map was as follows: determining the operating field, laying the ground control points; collecting the positional data of the ground control points and the orthophoto data from the unmanned aerial vehicle (UAV); processing the image data and constructing the complete map; slicing the map, correcting the deviation and transmitting it to the webpage. The field boundary information was obtained through the high-precision map. The equal spacing reduction algorithm and scanning line filling algorithm was adopted, and the spiral seeding operation path outside the shuttle row was automatically generated. According to the tractor geometry and kinematics model and the size of the distance between the tractor position and the field boundary, the specific parameters of the one-back and two-cut turning model were calculated, and based on the agronomic requirements of rapeseed sowing operation, the one-back-two-cut turn operation control strategy was designed to realize the rapeseed direct seeding unit's sowing operation for the omitted operation area of the field edges and corners. The test included map accuracy test, operation area simulation test and unmanned seeding operation field test. For the map accuracy test, the test field at the edge of Lake Yezhi of Huazhong Agricultural Universit was selected as the test site, where high-precision maps were constructed, and the image and position (POS) data collected by the UAV were processed, synthesized, and sliced, and then corrected for leveling according to the actual coordinates of the correction point and the coordinates of the image. Three rectangular fields of different sizes were selected for the operation area simulation test to compare the operation area and coverage rate of the three operation modes: set row, shuttle row, and shuttle row outer spiral. The Case TM1404 tractor equipped with an intelligent seeding and fertilizer application integrated machine was used as the test platform for the unmanned seeding operation test, and data such as tracking error and operation speed were recorded in real time by software algorithms. The data such as tracking error and operation speed were recorded in real-time. After the flowering of rapeseed, a series of color images of the operation fields were obtained by aerial photography using a drone during the flowering period of rapeseed, and the color images of the operation fields were spliced together, and then the seedling and non-seedling areas were mapped using map surveying and mapping software. [Results and Discussions] The results of the map accuracy test showed that the maximum error of the high-precision map ground verification point was 3.23 cm, and the results of the operation area simulation test showed that the full-coverage path of the helix outside the shuttle row reduced the leakage rate by 18.58%-26.01% compared with that of the shuttle row and the set of row path. The results of unmanned seeding operation field test showed that the average speed of unmanned seeding operation was 1.46 m/s, the maximum lateral deviation was 7.94 cm, and the maximum average absolute deviation was 1.85 cm. The test results in field showed that, the measured field area was 1 018.61 m2, and the total area of the non-growing oilseed rape area was 69.63 m2, with an operating area of 948.98 m2, and an operating coverage rate of 93.16%. [Conclusions] The effectiveness and feasibility of the constructed unmanned seeding operation system for rapeseed were demonstrated. This study can provide technical reference for unmanned seeding operation of rapeseed in small and medium-sized fields in the south. In the future, the unmanned seeding operation mode of rapeseed will be explored in irregular field conditions to further improve the applicability of the system.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Supply Chains for Agricultural Products: Key Technologies, Research Progress and Future Direction
    HAN Jiawei, YANG Xinting
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501006
    Online available: 07 May 2025

    Dynamic Prediction Method for Carbon Emissions of Cold Chain Distribution Vehicle under Multi-Source Information Fusion
    YANG Lin, LIU Shuangyin, XU Longqin, HE Min, SHENG Qingfeng, HAN Jiawei
    Smart Agriculture    2024, 6 (4): 138-148.   DOI: 10.12133/j.smartag.SA202403020
    Abstract378)   HTML17)    PDF(pc) (2240KB)(811)       Save

    [Objective] The dynamic prediction of carbon emission from cold chain distribution is an important basis for the accurate assessment of carbon emission and its green credit grade. Facing the fact that the carbon emission of vehicles is affected by multiple factors, such as road condition information, driving characteristics, refrigeration parameters, etc., a dynamic prediction model of carbon emission was proposed from refrigerated vehicles that integrates multi-source information. [Methods] The backbone feature extraction network, neck feature fusion network and loss function of YOLOv8s was firstly improved. The full-dimensional dynamic convolution was introduced into the backbone feature extraction network, and the multidimensional attention mechanism was introduced to capture the contextual key information to improve the model feature extraction capability. A progressive feature pyramid network was introduced into the feature extraction network, which reduced the loss of key information by fusing features layer by layer and improved the feature fusion efficiency. The road condition information recognition model based on improved YOLOv8s was constructed to characterize the road condition information in terms of the number of road vehicles and the percentage of pixel area. Pearson's correlation coefficient was used to compare and analyze the correlation between carbon emissions of refrigerated vehicles and different influencing factors, and to verify the necessity and criticality of the selection of input parameters of the carbon emission prediction model. Then the iTransformer temporal prediction model was improved, and the external attention mechanism was introduced to enhance the feature extraction ability of iTransformer model and reduce the computational complexity. The dynamic prediction model of carbon emission of refrigerated vehicles based on the improved iTransformer was constructed by taking the road condition information, driving characteristics (speed, acceleration), cargo weight, and refrigeration parameters (temperature, power) as inputs. Finally, the model was compared and analyzed with other models to verify the robustness of the road condition information and the prediction accuracy of the vehicle carbon emission dynamic prediction model, respectively. [Results and Discussions] The results of correlation analysis showed that the vehicle driving parameters were the main factor affecting the intensity of vehicle carbon emissions, with a correlation of 0.841. The second factor was cargo weight, with a correlation of 0.807, which had a strong positive correlation. Compared with the vehicle refrigeration parameters, the road condition information had a stronger correlation between vehicle carbon emissions, the correlation between refrigeration parameters and the vehicle carbon emissions impact factor were above 0.67. In order to further ensure the accuracy of the vehicle carbon emissions prediction model, The paper was selected as the input parameters for the carbon emissions prediction model. The improved YOLOv8s road information recognition model achieved 98.1%, 95.5%, and 98.4% in precision, recall, and average recognition accuracy, which were 1.2%, 3.7%, and 0.2% higher than YOLOv8s, respectively, with the number of parameters and the amount of computation being reduced by 12.5% and 31.4%, and the speed of detection being increased by 5.4%. This was due to the cross-dimensional feature learning through full-dimensional dynamic convolution, which fully captured the key information and improved the feature extraction capability of the model, and through the progressive feature pyramid network after fusing the information between different classes through gradual step-by-step fusion, which fully retained the important feature information and improved the recognition accuracy of the model. The predictive performance of the improved iTransformer carbon emission prediction model was better than other time series prediction models, and its prediction curve was closest to the real carbon emission curve with the best fitting effect. The introduction of the external attention mechanism significantly improved the prediction accuracy, and its MSE, MAE, RMSE and R2 were 0.026 1 %VOL, 0.079 1 %VOL, 0.161 5 %VOL and 0.940 0, respectively, which were 0.4%, 15.3%, 8.7% and 1.3% lower, respectively, when compared with iTransformer. As the degree of road congestion increased, the prediction accuracy of the constructed carbon emission prediction model increased. [Conclusions] The carbon emission prediction model for cold chain distribution under multi-source information fusion proposed in this study can realize accurate prediction of carbon emission from refrigerated vehicles, provide theoretical basis for rationally formulating carbon emission reduction strategies and promoting the development of low-carbon cold chain distribution.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Large Language Model Based on Precise Knowledge Retrieval and Knowledge Collaborative Generation
    JIANG Jingchi, YAN Lian, LIU Jie
    Smart Agriculture    2025, 7 (1): 20-32.   DOI: 10.12133/j.smartag.SA202410025
    Abstract922)   HTML56)    PDF(pc) (2081KB)(811)       Save

    [Objective] The rapid advancement of large language models (LLMs) has positioned them as a promising novel research paradigm in smart agriculture, leveraging their robust cognitive understanding and content generative capabilities. However, due to the lack of domain-specific agricultural knowledge, general LLMs often exhibit factual errors or incomplete information when addressing specialized queries, which is particularly prominent in agricultural applications. Therefore, enhancing the adaptability and response quality of LLMs in agricultural applications has become an important research direction. [Methods] To improve the adaptability and precision of LLMs in the agricultural applications, an innovative approach named the knowledge graph-guided agricultural LLM (KGLLM) was proposed. This method integrated information entropy for effective knowledge filtering and applied explicit constraints on content generation during the decoding phase by utilizing semantic information derived from an agricultural knowledge graph. The process began by identifying and linking key entities from input questions to the agricultural knowledge graph, which facilitated the formation of knowledge inference paths and the development of question-answering rationales. A critical aspect of this approach was ensuring the validity and reliability of the external knowledge incorporated into the model. This was achieved by evaluating the entropy difference in the model's outputs before and after the introduction of each piece of knowledge. Knowledge that didn't enhance the certainty of the answers was systematically filtered out. The knowledge paths that pass this entropy evaluation were used to adjust the token prediction probabilities, prioritizing outputs that were closely aligned with the structured knowledge. This allowed the knowledge graph to exert explicit guidance over the LLM's outputs, ensuring higher accuracy and relevance in agricultural applications. [Results and Discussions] The proposed knowledge graph-guided technique was implemented on five mainstream general-purpose LLMs, including open-source models such as Baichuan, ChatGLM, and Qwen. These models were compared with state-of-the-art knowledge graph-augmented generation methods to evaluate the effectiveness of the proposed approach. The results demonstrate that the proposed knowledge graph-guided approach significantly improved several key performance metrics of fluency, accuracy, factual correctness, and domain relevance. Compared to GPT-4o, the proposed method achieved notable improvements by an average of 2.592 3 in Mean BLEU, 2.815 1 in ROUGE, and 9.84% in BertScore. These improvements collectively signify that the proposed approach effectively leverages agricultural domain knowledge to refine the outputs of general-purpose LLMs, making them more suitable for agricultural applications. Ablation experiments further validated that the knowledge-guided agricultural LLM not only filtered out redundant knowledge but also effectively adjusts token prediction distributions during the decoding phase. This enhanced the adaptability of general-purpose LLMs in agriculture contexts and significantly improves the interpretability of their responses. The knowledge filtering and knowledge graph-guided model decoding method proposed in this study, which was based on information entropy, effectively identifies and selects knowledge that carried more informational content through the comparison of information entropy.Compared to existing technologies in the agricultural field, this method significantly reduced the likelihood of "hallucination" phenomena during the generation process. Furthermore, the guidance of the knowledge graph ensured that the model's generated responses were closely related to professional agricultural knowledge, thereby avoiding vague and inaccurate responses generated from general knowledge. For instance, in the application of pest and disease control, the model could accurately identify the types of crop diseases and corresponding control measures based on the guided knowledge path, thereby providing more reliable decision support. [Conclusions] This study provides a valuable reference for the construction of future agricultural large language models, indicating that the knowledge graphs guided mehtod has the potential to enhance the domain adaptability and answer quality of models. Future research can further explore the application of similar knowledge-guided strategies in other vertical fields to enhance the adaptability and practicality of LLMs across various professional domains.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    AgNPs Prepared by Lemon Juice Reduction Method for SERS Rapid Detection of Pesticide Residues in Fruits and Vegetables
    DONG Shanshan, ZHANG Fengqiu, XIA Qi, LI Jialin, LIU Chao, LIU Shaowei, CHEN Xiangyu, WANG Rujing, HUANG Qing
    Smart Agriculture    2024, 6 (1): 101-110.   DOI: 10.12133/j.smartag.SA202311010
    Abstract365)   HTML19)    PDF(pc) (2305KB)(774)       Save

    Objective The use of pesticides is one of the root causes of food safety problems. Pesticide exposure and pesticide residues can not only lead to environmental pollution issues but also seriously affect human health. In order to meet the rapid and sensitive detection needs of pesticide residues in agricultural products, a method based on lemon juice reduction to prepare silver nanoparticles (AgNPs) is reported in this research. Methods First, fresh lemon juice was filtered through filter paper and diluted to a 2% lemon juice aqueous solution. Then, a certain concentration of AgNO3 solution, 50 mm NaOH solution were prepared and stored at room temperature. Then, 10 mL ddH2O, 2 mL NaOH, 2 mL 2% lemon juice, and 5 mL AgNO3 solution were mixed. When the solution turned to a clear yellow color, the solution was centrifuged to obtain AgNPs. The morphology and structure of AgNPs were observed by transmission electron microscopy (TEM). In order to verify the successful synthesis of the nanoparticles and the distribution characteristics of the nanoparticles, ultraviolet spectroscopy was used for measurement and analysis, and 4-ATP was used as a SERS probe to preliminarily verify the SERS enhancement performance of AgNPs. Furthermore, the content of the main reducing components in lemon juice, namely ascorbic acid, glucose, and fructose was analyzed. The content of ascorbic acid in lemon juice was determined by high-performance liquid chromatography, and the content of glucose and fructose in lemon juice was determined by UV-visible spectrophotometry. To verify the stability and uniformity of the SERS signal of the nanoparticles, 4-ATP was used as an surface enhancement of raman scattering (SERS) probe for detection analysis. The stability of the SERS performance of the colloidal substrate within 41 days and the SERS performance at temperatures ranging from 0-80 °C were analyzed. Using 4-ATP as the SERS probe, the experimental conditions were optimized for the preparation of AgNPs by the lemon juice method, including pH and AgNO3 concentration. To validate the practical usability of the nanoparticles, the solutions of paraquat and carbendazim and the detection limits of pesticide residues on different fruits and vegetables were detected by SERS. Results and discussions The method for preparing AgNPs has the advantages of simple operation, green and easy synthesis. The particle morphology and size of the prepared AgNPs were basically uniform, with a size of about 20 nm. The ultraviolet-visible spectrum of AgNPs solution showed that the absorption peak was about 400 nm and the peak shape was narrow, indicating that the colloidal solution had good homogeneity. The detection limit of 4-ATP as the SERS probe was 10-14 M, indicating that the nanoparticle had a good SERS. In addition, the content of ascorbic acid, the main reducing ingredient, in lemon juice measured by high-performance liquid chromatography (HPLC) was 395.76 μg/mL. The contents of glucose and fructose, which were the main reducing components in lemon juice, were 5.95 and 5.90 mg/mL, respectively. Furthermore, the characterization and analysis results of the AgNPs prepared by the mixed reducing solution prepared according to the concentration data of each component showed that the AgNPs obtained were also uniform in morphology and size, with a diameter of about 20 nm, but the SERS enhancement performance was not as good as that of the AgNPs reduced by lemon juice. The SERS signal uniformity of the AgNPs reduced by lemon juice analyzed results showed that the peak intensity of the SERS spectral of 4-ATP at different sites at the same concentration was not significantly changed for 15 times, and its standard deviation RSD=5.03%, which was much lower than the intersubstrate RSD value (<16%) of the qualified new SERS active substrate for quantitative analysis. The temporal stability and temperature stability of AgNPs analyzed results showed that the nanoparticles still had SERS enhanced performance after 41 days of storage, and had SERS enhanced performance stability over a wide temperature range (0~80 °C). In addition, the optimization results of experimental conditions showed that the optimal pH for the preparation of AgNPs was around 7.5, and the optimal range of AgNO3 concentration used was 1.76×10-4~3.33×10-4 mol/L. Finally, using AgNPs prepared by lemon juice reduction method for pesticide residue SERS detection on the surface of fruits and vegetables, the detection limits for paraquat and carbendazim in solution were as low as 10-14 and 10-10 M, respectively, and the concentration of pesticides showed a good linear relationship with Raman spectral intensity. The lowest detection limits for paraquat and carbendazim residues on different fruits and vegetables were as low as 3.90 ng/kg and 0.22 µg/kg, respectively. Conclusions This work provides a green and convenient method for preparing SERS materials for rapid detection of pesticide residues on fruits and vegetables. This method has practical value for universal operation. The prepared AgNPs can be used for trace pesticide residue detection, providing a pathway for rapid and sensitive detection of pesticide residues in agricultural products.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Intelligent Detection and Alarm System for Ferrous Metal Foreign Objects in Silage Machines
    ZHANG Qing, LI Yang, YOU Yong, WANG Decheng, HUI Yunting
    Smart Agriculture    2024, 6 (1): 111-122.   DOI: 10.12133/j.smartag.SA202306010
    Abstract373)   HTML20)    PDF(pc) (2026KB)(757)       Save

    Objective During the operation of the silage machine, the inclusion of ferrous metal foreign objects such as stray iron wires can inflict severe damage to the machine's critical components and livestock organs. To safeguard against that, a metal detection system with superior performance was developed in this research to enable precise and efficient identification of metal foreign bodies during field operations, ensuring the integrity of the silage process and the well-being of the animals. Methods The ferrous metal detection principle of silage machine was firstly analyzed. The detection coil is the probe of the metal detection system. After being connected in parallel with a capacitor, it is connected to the detection module. The detection coil received the alternating signal generated by the detection module to generate an alternating magnetic field. After the metal object entered the magnetic field, it affects the equivalent resistance and equivalent inductance of the detection coil. The detection module detected the change of the equivalent resistance and equivalent inductance, and then transmited the signal to the control module through the serial peripheral interface (SPI). The control module filtered the signal and transmited it to the display terminal through the serial port. The display terminal could set the threshold. When the data exceeded the threshold, the system performed sound and light alarm and other processing. Hardware part of the metal detection system of silage machine were firstly design. The calculation of the planar spiral coil and the cylindrical coil was carried out and the planar spiral coil was selected as the research object. By using the nondominated sorting genetic algorithm-Ⅱ (NSGA-II) combined with the method of finite element simulation analysis, the wire diameter, inner diameter, outer diameter, layer number and frequency of the coil were determined, and the calculation of the bent coil and the unbent coil and the array coil was carried out. The hardware system was integrated. The software system for the metal detection system was also designed, utilizing an STM32 microcontroller as the control module and LabView for writing the primary program on the upper computer. The system continuously displayed the read data and time-equivalent impedance graph in real-time, allowing for the setting of upper and lower alarm thresholds. When a metal foreign object was detected, the warning light turned red and an alarm sound was emitted, causing the feed roll to stop. To simulate the scenario of metal detection during the operation of a silage machine, a test bench was set up to validate the performance of the metal detection system. Results and Discussions The test results of the metal detection function showed that for a metal wire with a diameter of 0.6 mm and a length of 20 mm, as the inner diameter of the detection coil increased, the maximum alarm distance increased first and then decreased. The maximum alarm distance occured when the inner diameter was 35 mm, which was consistent with the optimization result. The maximum alarm distance was the largest when the detection coil was two layers, and there was no data readout when it was three layers. Therefore, the optimal thickness of the detection coil for this metal detection system was two layers. When the detection distance was greater than 80 mm, the alarm rate began to decrease, and the detection effect was weakened. When the detection distance was within 70 mm, the metal detection system could achieve a 100% alarm rate. The test results of the system response time showed that the average system response time was 0.105 0 s, which was less than the safe transportation time of 0.202 0 s. The system can give an alarm before the metal foreign object reaches the cutter, so the system is safe and effective. Conclusion In this study, a metal detection system for silage machines was designed. A set of optimization methods for metal detection coils was proposed, and the corresponding metal detection software and hardware systems were developed, and the functions of the metal detection system were verified through experiments, which could provide strong technical support for the safe operation of silage machines.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Price Game Model and Competitive Strategy of Agricultural Products Retail Market in the Context of Blockchain
    XUE Bing, SUN Chuanheng, LIU Shuangyin, LUO Na, LI Jinhui
    Smart Agriculture    2024, 6 (4): 160-173.   DOI: 10.12133/j.smartag.SA202309027
    Abstract375)   HTML28)    PDF(pc) (1265KB)(748)       Save

    [Objective] In the retail market for agricultural products, consumers are increasingly concerned about the safety and health aspects of those products. Traceability of blockchain has emerged as a crucial solution to address these concerns. Essentially, a blockchain functions as a dynamic, distributed, and shared database. When implemented in the agricultural supply chain, it not only improves product transparency to attract more consumers but also raises concerns about consumer privacy disclosure. The level of consumer apprehension regarding privacy will directly influence their choice to purchase agricultural products traced through blockchain-traced. Moreover, retailers' choices to sell blockchain-traced produce are influenced by consumer privacy concerns. By analyzing the impact of blockchain technology on the competitive strategies, pricing, and decision-making among agricultural retailers, they can develop market competition strategies that suit their market conditions to bolster their competitiveness and optimize the agricultural supply chain to maximize overall benefits. [Methods] Based on Nash equilibrium and Stackelberg game theory, a market competition model was developed to analyze the interactions between existing and new agricultural product retailers. The competitive strategies adopted by agricultural product retailers were analyzed under four different options of whether two agricultural retailers sell blockchain agricultural products. It delved into product utility, optimal pricing, demand, and profitability for each retailer under these different scenarios. How consumer privacy concerns impact pricing and profits of two agricultural product retailers and the optimal response strategy choice of another retailer when the competitor made the decision choice first were also analyzed. This analysis aimed to guide agricultural product retailers in making strategic choices that would safeguard their profits and market positions. To address the cooperative game problem of agricultural product retailers in market competition, ensure that retailers could better cooperate in the game, blockchain smart contract technology was used. By encoding the process and outcomes of the Stackelberg game into smart contracts, retailers could input their specific variables and receive tailored strategy recommendations. Uploading game results onto the blockchain network ensured transparency and encouraged cooperative behavior among retailers. By using the characteristics of blockchain, the game results were uploaded to the blockchain network to regulate the cooperative behavior, to ensure the maximization of the overall interests of the supply chain. [Results and Discussions] The research highlighted the significant improvement in agricultural product quality transparency through blockchain traceability technology. However, concerns regarding consumer privacy arising from this traceability could directly impact the pricing, profitability and retailers' decisions to provide blockchain-traceable items. Furthermore, an analysis of the strategic balance between two agricultural product retailers revealed that in situations of low and high product information transparency, both retailers were inclined to simultaneously offer sell traceable products. In such a scenario, blockchain traceability technology enhanced the utility and profitability of retail agricultural products, leading consumers to prefer purchase these traceable products from retailers. In cases where privacy concerns and agricultural product information transparency were both moderate, the initial retailer was more likely to opt for blockchain-based traceable products. This was because consumers had higher trust in the initial retailer, enabling them to bear a higher cost associated with privacy concerns. Conversely, new retailers failed to gain a competitive advantage and eventually exit the market. When consumer privacy concerns exceeded a certain threshold, both competing agricultural retailers discovered that offering blockchain-based traceable products led to a decline in their profits. [Conclusions] When it comes to agricultural product quality and safety, incorporating blockchain technology in traceability significantly improves the transparency of quality-related information for agricultural products. However, it is important to recognize that the application of blockchain for agricultural product traceability is not universally suitable for all agricultural retailers. Retailers must evaluate their unique circumstances and make the most suitable decisions to enhance the effectiveness of agricultural products, drive sales demand, and increase profits. Within the competitive landscape of the agricultural product retail market, nurturing a positive collaborative relationship is essential to maximize mutual benefits and optimize the overall profitability of the agricultural product supply chain.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Multi-Focal Green Plant Image Fusion Method Based on Stationary Wavelet Transform and Parameter-Adaptation Dual Channel Pulse-Coupled Neural Network
    LI Jiahao, QU Hongjun, GAO Mingzhe, TONG Dezhi, GUO Ya
    Smart Agriculture    2023, 5 (3): 121-131.   DOI: 10.12133/j.smartag.SA202308005
    Abstract420)   HTML32)    PDF(pc) (1435KB)(745)       Save

    [Objective] To construct the 3D point cloud model of green plants a large number of clear images are needed. Due to the limitation of the depth of field of the lens, part of the image would be out of focus when the green plant image with a large depth of field is collected, resulting in problems such as edge blurring and texture detail loss, which greatly affects the accuracy of the 3D point cloud model. However, the existing processing algorithms are difficult to take into account both processing quality and processing speed, and the actual effect is not ideal. The purpose of this research is to improve the quality of the fused image while taking into account the processing speed. [Methods] A plant image fusion method based on non-subsampled shearlet transform (NSST) based parameter-adaptive dual channel pulse-coupled neural network (PADC-PCNN) and stationary wavelet transform (SWT) was proposed. Firstly, the RGB image of the plant was separated into three color channels, and the G channel with many features such as texture details was decomposed by NSST in four decomposition layers and 16 directions, which was divided into one group of low frequency subbands and 64 groups of high frequency subbands. The low frequency subband used the gradient energy fusion rule, and the high frequency subband used the PADC-PCNN fusion rule. In addition, the weighting of the eight-neighborhood modified Laplacian operator was used as the link strength of the high-frequency fusion part, which enhanced the fusion effect of the detailed features. At the same time, for the R and B channels with more contour information and background information, a SWT with fast speed and translation invariance was used to suppress the pseudo-Gibbs effect. Through the high-precision and high-stability multi-focal length plant image acquisition system, 480 images of 8 experimental groups were collected. The 8 groups of data were divided into an indoor light group, natural light group, strong light group, distant view group, close view group, overlooking group, red group, and yellow group. Meanwhile, to study the application range of the algorithm, the focus length of the collected clear plant image was used as the reference (18 mm), and the image acquisition was adjusted four times before and after the step of 1.5 mm, forming the multi-focus experimental group. Subjective evaluation and objective evaluation were carried out for each experimental group to verify the performance of the algorithm. Subjective evaluation was analyzed through human eye observation, detail comparison, and other forms, mainly based on the human visual effect. The image fusion effect of the algorithm was evaluated using four commonly used objective indicators, including average gradient (AG), spatial frequency (SF), entropy (EN), and standard deviation (SD). [Results and Discussions] The proposed PADC-PCNN-SWT algorithm and other five algorithms of common fast guided filtering algorithm (FGF), random walk algorithm (RW), non-subsampled shearlet transform based PCNN (NSST-PCNN) algorithm, SWT algorithm and non-subsampled shearlet transform based parameter-adaptive dual-channel pulse-coupled neural network (NSST-PADC) and were compared. In the objective evaluation data except for the red group and the yellow group, each index of the PADC-PCNN-SWT algorithm was second only to the NSST-PADC algorithm, but the processing speed was 200.0% higher than that of the NSST-PADC algorithm on average. At the same time, compared with the FDF, RW, NSST-PCNN, and SWT algorithms, the PADC-PCN -SWT algorithm improved the clarity index by 5.6%, 8.1%, 6.1%, and 17.6%, respectively, and improved the spatial frequency index by 2.9%, 4.8%, 7.1%, and 15.9%, respectively. However, the difference between the two indicators of information entropy and standard deviation was less than 1%, and the influence was ignored. In the yellow group and the red group, the fusion quality of the non-green part of the algorithm based on PADC-PCNN-SWT was seriously degraded. Compared with other algorithms, the sharpness index of the algorithm based on PADC-PCNN-SWT decreased by an average of 1.1%, and the spatial frequency decreased by an average of 5.1%. However, the indicators of the green part of the fused image were basically consistent with the previous several groups of experiments, and the fusion effect was good. Therefore, the algorithm based on PADC-PCNN-SWT only had a good fusion effect on green plants. Finally, by comparing the quality of four groups of fused images with different focal length ranges, the results showed that the algorithm based on PADC-PCNN-SWT had a better contour and color restoration effect for out-of-focus images in the range of 15-21 mm, and the focusing range based on PADC-PCNN-SWT was about 6 mm. [Conclusions] The multi-focal length image fusion algorithm based on PADC-PCNN-SWT achieved better detail fusion performance and higher image fusion efficiency while ensuring fusion quality, providing high-quality data, and saving a lot of time for building 3D point cloud model of green plants.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Regional Farming Pig Counting System Based on Improved Instance Segmentation Algorithm
    ZHANG Yanqi, ZHOU Shuo, ZHANG Ning, CHAI Xiujuan, SUN Tan
    Smart Agriculture    2024, 6 (4): 53-63.   DOI: 10.12133/j.smartag.SA202310001
    Abstract1308)   HTML76)    PDF(pc) (2077KB)(729)       Save

    [Objective] Currently, pig farming facilities mainly rely on manual counting for tracking slaughtered and stored pigs. This is not only time-consuming and labor-intensive, but also prone to counting errors due to pig movement and potential cheating. As breeding operations expand, the periodic live asset inventories put significant strain on human, material and financial resources. Although methods based on electronic ear tags can assist in pig counting, these ear tags are easy to break and fall off in group housing environments. Most of the existing methods for counting pigs based on computer vision require capturing images from a top-down perspective, necessitating the installation of cameras above each hogpen or even the use of drones, resulting in high installation and maintenance costs. To address the above challenges faced in the group pig counting task, a high-efficiency and low-cost pig counting method was proposed based on improved instance segmentation algorithm and WeChat public platform. [Methods] Firstly, a smartphone was used to collect pig image data in the area from a human view perspective, and each pig's outline in the image was annotated to establish a pig count dataset. The training set contains 606 images and the test set contains 65 images. Secondly, an efficient global attention module was proposed by improving convolutional block attention module (CBAM). The efficient global attention module first performed a dimension permutation operation on the input feature map to obtain the interaction between its channels and spatial dimensions. The permuted features were aggregated using global average pooling (GAP). One-dimensional convolution replaced the fully connected operation in CBAM, eliminating dimensionality reduction and significantly reducing the model's parameter number. This module was integrated into the YOLOv8 single-stage instance segmentation network to build the pig counting model YOLOv8x-Ours. By adding an efficient global attention module into each C2f layer of the YOLOv8 backbone network, the dimensional dependencies and feature information in the image could be extracted more effectively, thereby achieving high-accuracy pig counting. Lastly, with a focus on user experience and outreach, a pig counting WeChat mini program was developed based on the WeChat public platform and Django Web framework. The counting model was deployed to count pigs using images captured by smartphones. [Results and Discussions] Compared with existing methods of Mask R-CNN, YOLACT(Real-time Instance Segmentation), PolarMask, SOLO and YOLOv5x, the proposed pig counting model YOLOv8x-Ours exhibited superior performance in terms of accuracy and stability. Notably, YOLOv8x-Ours achieved the highest accuracy in counting, with errors of less than 2 and 3 pigs on the test set. Specifically, 93.8% of the total test images had counting errors of less than 3 pigs. Compared with the two-stage instance segmentation algorithm Mask R-CNN and the YOLOv8x model that applies the CBAM attention mechanism, YOLOv8x-Ours showed performance improvements of 7.6% and 3%, respectively. And due to the single-stage design and anchor-free architecture of the YOLOv8 model, the processing speed of a single image was only 64 ms, 1/8 of Mask R-CNN. By embedding the model into the WeChat mini program platform, pig counting was conducted using smartphone images. In cases where the model incorrectly detected pigs, users were given the option to click on the erroneous location in the result image to adjust the statistical outcomes, thereby enhancing the accuracy of pig counting. [Conclusions] The feasibility of deep learning technology in the task of pig counting was demonstrated. The proposed method eliminates the need for installing hardware equipment in the breeding area of the pig farm, enabling pig counting to be carried out effortlessly using just a smartphone. Users can promptly spot any errors in the counting results through image segmentation visualization and easily rectify any inaccuracies. This collaborative human-machine model not only reduces the need for extensive manpower but also guarantees the precision and user-friendliness of the counting outcomes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Key Technologies and Construction model for Unmanned Smart Farms: Taking the "1.5-Ton Grain per Mu" Unmanned Farm as An Example
    LIU lining, ZHANG Hongqi, ZHANG Ziwen, ZHANG Zhenghui, WANG Jiayu, LI Xuanxuan, ZHU Ke, LIU Pingzeng
    Smart Agriculture    2025, 7 (1): 70-84.   DOI: 10.12133/j.smartag.SA202410033
    Abstract429)   HTML52)    PDF(pc) (2674KB)(717)       Save

    [Objective] As a key model of smart agriculture, the unmanned smart farm aims to develop a highly intelligent and automated system for high grain yields. This research uses the "1.5-Ton grain per Mu" farm in Dezhou city, Shandong province, as the experimental site, targeting core challenges in large-scale smart agriculture and exploring construction and service models for such farms. [Methods] The "1.5-Ton grain per Mu" unmanned smart farm comprehensively utilized information technologies such as the internet of things (IoT) and big data to achieve full-chain integration and services for information perception, transmission, mining, and application. The overall construction architecture consisted of the perception layer, transmission layer, processing layer, and application layer. This architecture enabled precise perception, secure transmission, analysis and processing, and application services for farm data. A perception system for the unmanned smart farm of wheat was developed, which included a digital perception network and crop phenotypic analysis. The former achieved precise perception, efficient transmission, and precise measurement and control of data information within the farm through perception nodes, self-organizing networks, and edge computing core processing nodes. Phenotypic analysis utilized methods such as deep learning to extract phenotypic characteristics at different growth stages, such as the phenological classification of wheat and wheat ear length. An intelligent controlled system had been developed. The system consisted of an intelligent agricultural machinery system, a field irrigation system, and an aerial pesticided application system. The intelligent agricultural machinery system was composed of three parts: the basic layer, decision-making layer, and application service layer. They were responsible for obtaining real-time status information of agricultural machinery, formulating management decisions for agricultural machinery, and executing operational commands, respectively. Additionally, appropriate agricultural machinery models and configuration references were provided. A refined irrigation scheme was designed based on the water requirements and soil conditions at different developmental stages of wheat. And, an irrigation control algorithm based on fuzzy PID was proposed. Finally, relying on technologies such as multi-source data fusion, distributed computing, and geographic information system (GIS), an intelligent management and control platform for the entire agricultural production process was established. [Results and Discussions] The digital perception network enabled precise sensing and networked transmission of environmental information within the farm. The data communication quality of the sensor network remained above 85%, effectively ensuring data transmission quality. The average relative error in extracting wheat spike length information based on deep learning algorithms was 1.24%. Through the coordinated operation of intelligent control system, the farm achieved lean and unmanned production management, enabling intelligent control throughout the entire production chain, which significantly reduced labor costs and improved the precision and efficiency of farm management. The irrigation model not only saved 20% of irrigation water but also increased the yield of "Jinan 17" and "Jimai 44" by 10.18% and 7%, respectively. Pesticide application through spraying drones reduced pesticide usage by 55%. The big data platform provided users with production guidance services such as meteorological disaster prediction, optimal sowing time, environmental prediction, and water and fertilizer management through intelligent scientific decision support, intelligent agricultural machinery operation, and producted quality and safety traceability modules, helping farmers manage their farms scientifically. [Conclusions] The study achieved comprehensive collection of environmental information within the farm, precise phenotypic analysis, and intelligent control of agricultural machinery, irrigation equipment, and other equipment. Additionally, it realized digital services for agricultural management through a big data platform. The development path of the "1.5-Ton grain per Mu" unmanned smart farm can provid references for the construction of smart agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Detection Method of Dairy Cow Lameness from Top-view Based on the Fusion of Spatiotemporal Stream Features
    DAI Xin, WANG Junhao, ZHANG Yi, WANG Xinjie, LI Yanxing, DAI Baisheng, SHEN Weizheng
    Smart Agriculture    2024, 6 (4): 18-28.   DOI: 10.12133/j.smartag.SA202405025
    Abstract936)   HTML33)    PDF(pc) (1828KB)(713)       Save

    [Objective] The detection of lameness in dairy cows is an important issue that needs to be solved urgently in the process of large-scale dairy farming. Timely detection and effective intervention can reduce the culling rate of young dairy cows, which has important practical significance for increasing the milk production of dairy cows and improving the economic benefits of pastures. Due to the low efficiency and low degree of automation of traditional manual detection and contact sensor detection, the mainstream cow lameness detection method is mainly based on computer vision. The detection perspective of existing computer vision-based cow lameness detection methods is mainly side view, but the side view perspective has limitations that are difficult to eliminate. In the actual detection process, there are problems such as cows blocking each other and difficulty in deployment. The cow lameness detection method from the top view will not be difficult to use on the farm due to occlusion problems. The aim is to solve the occlusion problem under the side view. [Methods] In order to fully explore the movement undulations of the trunk of the cow and the movement information in the time dimension during the walking process of the cow, a cow lameness detection method was proposed from a top view based on fused spatiotemporal flow features. By analyzing the height changes of the lame cow in the depth video stream during movement, a spatial stream feature image sequence was constructed. By analyzing the instantaneous speed of the lame cow's body moving forward and swaying left and right when walking, optical flow was used to capture the instantaneous speed of the cow's movement, and a time flow characteristic image sequence was constructed. The spatial flow and time flow features were combined to construct a fused spatiotemporal flow feature image sequence. Different from traditional image classification tasks, the image sequence of cows walking includes features in both time and space dimensions. There would be a certain distinction between lame cows and non-lame cows due to their related postures and walking speeds when walking, so using video information analysis was feasible to characterize lameness as a behavior. The video action classification network could effectively model the spatiotemporal information in the input image sequence and output the corresponding category in the predicted result. The attention module Convolutional Block Attention Module (CBAM) was used to improve the PP-TSMv2 video action classification network and build the Cow-TSM cow lameness detection model. The CBAM module could perform channel weighting on different modes of cows, while paying attention to the weights between pixels to improve the model's feature extraction capabilities. Finally, cow lameness experiments were conducted on different modalities, different attention mechanisms, different video action classification networks and comparison of existing methods. The data was used for cow lameness included a total of 180 video streams of cows walking. Each video was decomposed into 100‒400 frames. The ratio of the number of video segments of lame cows and normal cows was 1:1. For the feature extraction of cow lameness from the top view, RGB images had less extractable information, so this work mainly used depth video streams. [Results and Discussions] In this study, a total of 180 segments of cow image sequence data were acquired and processed, including 90 lame cows and 90 non-lame cows with a 1:1 ratio of video segments, and the prediction accuracy of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features reaches 88.7%, the model size was 22 M, and the offline inference time was 0.046 s. The prediction accuracy of the common mainstream video action classification models TSM, PP-TSM, SlowFast and TimesFormer models on the data set of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features reached 66.7%, 84.8%, 87.1% and 85.7%, respectively. The comprehensive performance of the improved Cow-TSM model in this paper was the most. At the same time, the recognition accuracy of the fused spatiotemporal flow feature image was improved by 12% and 4.1%, respectively, compared with the temporal mode and spatial mode, which proved the effectiveness of spatiotemporal flow fusion in this method. By conducting ablation experiments on different attention mechanisms of SE, SK, CA and CBAM, it was proved that the CBAM attention mechanism used has the best effect on the data of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features. The channel attention in CBAM had a better effect on fused spatiotemporal flow data, and the spatial attention could also focus on the key spatial information in cow images. Finally, comparisons were made with existing lameness detection methods, including different methods from side view and top view. Compared with existing methods in the side-view perspective, the prediction accuracy of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features was slightly lower, because the side-view perspective had more effective cow lameness characteristics. Compared with the method from the top view, a novel fused spatiotemporal flow feature detection method with better performance and practicability was proposed. [Conclusions] This method can avoid the occlusion problem of detecting lame cows from the side view, and at the same time improves the prediction accuracy of the detection method from the top view. It is of great significance for reducing the incidence of lameness in cows and improving the economic benefits of the pasture, and meets the needs of large-scale construction of the pasture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Obstacle Avoidance Control Method of Electric Skid-Steering Chassis Based on Fuzzy Logic Control
    LI Lei, SHE Xiaoming, TANG Xinglong, ZHANG Tao, DONG Jiwei, GU Yuchuan, ZHOU Xiaohui, FENG Wei, YANG Qinghui
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202408003
    Online available: 27 December 2024

    Electrochemical Immunosensor for in Situ Detection of Brassinolide
    WEI Qian, GAO Yuanyuan, LI Aixue
    Smart Agriculture    2024, 6 (1): 76-88.   DOI: 10.12133/j.smartag.SA202311001
    Abstract944)   HTML20)    PDF(pc) (1903KB)(682)       Save

    Objective Brassinolide is an important endogenous plant hormone. In this work, an electrochemical immunosensor for in situ detection of brassinolide was constructed using screen-printed electrode (SPE). Methods Au nanoparticles (AuNPs) were firstly electrodeposited on the surface of SPE electrode by electrochemical workstation, and then CuCl2 nanowires (CuCl2 NWs) were added to the electrode, which can not only increase the conductivity of the electrode, but also Cu2+ can be used as a REDOX probe for the sensor. Finally, Mxene and polydopamine nanocomposite (Mxene@PDA) were selected as the modification materials for SPE electrodes because Mxene has the advantages of large surface area and good electrical conductivity, which can further amplify Cu2+ signals. However, Mxene is easily oxidized and unstable in air. Polydopamine (PDA) contains a large number of catechol and amino groups, which are coated on the surface of Mxene after self-polymerization by dopamine, cutting off the path of oxygen penetration, making Mxene difficult to be oxidized. Mxene@PDA can also be used as a coupling agent to fix more antibodies on the electrode surface, improving the overall biocompatibility, and improve the overall biocompatibility. Results and Discussions The sensor has a wide linear detection range: 0.1 pg/mL to 1 mg/mL, and the detection limit was 0.015 pg/ml (S/N=3). In addition, the content of endogenous brassinolide in wheat was detected by SPE electrodes in vitro and the recovery rate was 98.13% to 104.74%.While verifying the accuracy of the sensor, it also demonstrated its superior stability and sensitivity. Besides, the sensor also showed excellent application potential in the in situ detection of brassinosteroids from wheat leaves. Compared with other brassinolide detection methods, the immunosensor developed in this study has better analytical performance. Conclusions An electrochemical immunosensor for in situ detection of brassinolide was developed for the first time, providing a good electrochemical platform for in situ determination of brassinolide in plant leaves, which has great application potential in precision agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    The Development Logic, Influencing Factors and Realization Path for Low-Carbon Agricultural Mechanization
    YANG Yinsheng, WEI Xin
    Smart Agriculture    2023, 5 (4): 150-159.   DOI: 10.12133/j.smartag.SA202304008
    Abstract440)   HTML35)    PDF(pc) (870KB)(677)       Save

    [Significance] With the escalating global climate change and ecological pollution issues, the "dual carbon" target of Carbon Peak and Carbon Neutrality has been incorporated into various sectors of China's social development. To ensure the green and sustainable development of agriculture, it is imperative to minimize energy consumption and reduce pollution emissions at every stage of agricultural mechanization, meet the diversified needs of agricultural machinery and equipment in the era of intelligent information, and develop low-carbon agricultural mechanization. The development of low-carbon agricultural mechanization is not only an important part of the transformation and upgrading of agricultural mechanization in China but also an objective requirement for the sustainable development of agriculture under the "dual carbon" target. Progress] The connotation and objectives of low-carbon agricultural mechanization are clarified and the development logic of low-carbon agricultural mechanization from three dimensions: theoretical, practical, and systematic are expounded. The "triple-win" of life, production, and ecology is proposed, it is an important criterion for judging the functional realization of low-carbon agricultural mechanization system from a theoretical perspective. The necessity and urgency of low-carbon agricultural mechanization development from a practical perspective is revealed. The "human-machine-environment" system of low-carbon agricultural mechanization development is analyzed and the principles and feasibility of coordinated development of low-carbon agricultural mechanization based on a systemic perspective is explained. Furthermore, the deep-rooted reasons affecting the development of low-carbon agricultural mechanization from six aspects are analyzed: factor conditions, demand conditions, related and supporting industries, production entities, government, and opportunities. Conclusion and Prospects] Four approaches are proposed for the realization of low-carbon agricultural mechanization development: (1) Encouraging enterprises to implement agricultural machinery ecological design and green manufacturing throughout the life cycle through key and core technology research, government policies, and financial support; (2) Guiding agricultural entities to implement clean production operations in agricultural mechanization, including but not limited to innovative models of intensive agricultural land, exploration and promotion of new models of clean production in agricultural mechanization, and the construction of a carbon emission measurement system for agricultural low-carbonization; (3) Strengthening the guidance and implementation of the concept of socialized services for low-carbon agricultural machinery by government departments, constructing and improving a "8S" system of agricultural machinery operation services mainly consisting of Sale, Spare part, Service, Survey, Show, School, Service, and Scrap, to achieve the long-term development of dematerialized agricultural machinery socialized services and green shared operation system; (4) Starting from concept guidance, policy promotion, and financial support, comprehensively advancing the process of low-carbon disposal and green remanufacturing of retired and waste agricultural machinery by government departments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Spectroscopic Detection of Rice Leaf Blast Infection at Different Leaf Positions at The Early Stages With Solar-Induced Chlorophyll Fluorescence
    CHENG Yuxin, XUE Bowen, KONG Yuanyuan, YAO Dongliang, TIAN Long, WANG Xue, YAO Xia, ZHU Yan, CAO Weixing, CHENG Tao
    Smart Agriculture    2023, 5 (3): 35-48.   DOI: 10.12133/j.smartag.SA202309008
    Abstract645)   HTML47)    PDF(pc) (5433KB)(675)       Save

    [Objective] Rice blast is considered as the most destructive disease that threatens global rice production and causes severe economic losses worldwide. The detection of rice blast in an early manner plays an important role in resistance breeding and plant protection. At present, most studies on rice blast detection have been devoted to its symptomatic stage, while none of previous studies have used solar-induced chlorophyll fluorescence (SIF) to monitor rice leaf blast (RLB) at early stages. This research was conducted to investigate the early identification of RLB infected leaves based on solar-induced chlorophyll fluorescence at different leaf positions. [Methods] Greenhouse experiments and field trials were conducted separately in Nanjing and Nantong in July and August, 2021, in order to record SIF data of the top 1th to 4th leaves of rice plants at jointing and heading stages with an Analytical Spectral Devices (ASD) spectrometer coupled with a FluoWat leaf clip and a halogen lamp. At the same time, the disease severity levels of the measured samples were manually collected according to the GB/T 15790-2009 standard. After the continuous wavelet transform (CWT) of SIF spectra, separability assessment and feature selection were applied to SIF spectra. Wavelet features sensitive to RLB were extracted, and the sensitive features and their identification accuracy of infected leaves for different leaf positions were compared. Finally, RLB identification models were constructed based on linear discriminant analysis (LDA). [Results and Discussion] The results showed that the upward and downward SIF in the far-red region of infected leaves at each leaf position were significantly higher than those of healthy leaves. This may be due to the infection of the fungal pathogen Magnaporthe oryzae, which may have destroyed the chloroplast structure, and ultimately inhibited the primary reaction of photosynthesis. In addition, both the upward and downward SIF in the red region and the far-red region increased with the decrease of leaf position. The sensitive wavelet features varied by leaf position, while most of them were distributed in the steep slope of the SIF spectrum and wavelet scales 3, 4 and 5. The sensitive features of the top 1th leaf were mainly located at 665-680 nm, 755-790 nm and 815-830 nm. For the top 2th leaf, the sensitive features were mainly found at 665-680 nm and 815-830 nm. For the top 3th one, most of the sensitive features lay at 690 nm, 755-790 nm and 815-830 nm, and the sensitive bands around 690 nm were observed. The sensitive features of the top 4th leaf were primarily located at 665-680 nm, 725 nm and 815-830 nm, and the sensitive bands around 725 nm were observed. The wavelet features of the common sensitive region (665-680 nm), not only had physiological significance, but also coincided with the chlorophyll absorption peak that allowed for reasonable spectral interpretation. There were differences in the accuracy of RLB identification models at different leaf positions. Based on the upward and downward SIF, the overall accuracies of the top 1th leaf were separately 70% and 71%, which was higher than other leaf positions. As a result, the top 1th leaf was an ideal indicator leaf to diagnose RLB in the field. The classification accuracy of SIF wavelet features were higher than the original SIF bands. Based on CWT and feature selection, the overall accuracy of the upward and downward optimal features of the top 1th to 4th leaves reached 70.13%、63.70%、64.63%、64.53% and 70.90%、63.12%、62.00%、64.02%, respectively. All of them were higher than the canopy monitoring feature F760, whose overall accuracy was 69.79%, 61.31%, 54.41%, 61.33% and 69.99%, 58.79%, 54.62%, 60.92%, respectively. This may be caused by the differences in physiological states of the top four leaves. In addition to RLB infection, the SIF data of some top 3th and top 4th leaves may also be affected by leaf senescence, while the SIF data of top 1th leaf, the latest unfolding leaf of rice plants was less affected by other physical and chemical parameters. This may explain why the top 1th leaf responded to RLB earlier than other leaves. The results also showed that the common sensitive features of the four leaf positions were also concentrated on the steep slope of the SIF spectrum, with better classification performance around 675 and 815 nm. The classification accuracy of the optimal common features, ↑WF832,3 and ↓WF809,3, reached 69.45%, 62.19%, 60.35%, 63.00% and 69.98%, 62.78%, 60.51%, 61.30% for the top 1th to top 4th leaf positions, respectively. The optimal common features, ↑WF832,3 and ↓WF809,3, were both located in wavelet scale 3 and 800-840nm, which may be related to the destruction of the cell structure in response to Magnaporthe oryzae infection. [Conclusions] In this study, the SIF spectral response to RLB was revealed, and the identification models of the top 1th leaf were found to be most precise among the top four leaves. In addition, the common wavelet features sensitive to RLB, ↑WF832,3 and ↓WF809,3, were extracted with the identification accuracy of 70%. The results proved the potential of CWT and SIF for RLB detection, which can provide important reference and technical support for the early, rapid and non-destructive diagnosis of RLB in the field.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Identification Method of Wheat Field Lodging Area Based on Deep Learning Semantic Segmentation and Transfer Learning
    ZHANG Gan, YAN Haifeng, HU Gensheng, ZHANG Dongyan, CHENG Tao, PAN Zhenggao, XU Haifeng, SHEN Shuhao, ZHU Keyu
    Smart Agriculture    2023, 5 (3): 75-85.   DOI: 10.12133/j.smartag.SA202309013
    Abstract502)   HTML57)    PDF(pc) (2219KB)(662)       Save

    [Objective] Lodging constitutes a severe crop-related catastrophe, resulting in a reduction in photosynthesis intensity, diminished nutrient absorption efficiency, diminished crop yield, and compromised crop quality. The utilization of unmanned aerial vehicles (UAV) to acquire agricultural remote sensing imagery, despite providing high-resolution details and clear indications of crop lodging, encounters limitations related to the size of the study area and the duration of the specific growth stages of the plants. This limitation hinders the acquisition of an adequate quantity of low-altitude remote sensing images of wheat fields, thereby detrimentally affecting the performance of the monitoring model. The aim of this study is to explore a method for precise segmentation of lodging areas in limited crop growth periods and research areas. [Methods] Compared to the images captured at lower flight altitudes, the images taken by UAVs at higher altitudes cover a larger area. Consequently, for the same area, the number of images taken by UAVs at higher altitudes is fewer than those taken at lower altitudes. However, the training of deep learning models requires huge amount supply of images. To make up the issue of insufficient quantity of high-altitude UAV-acquired images for the training of the lodging area monitoring model, a transfer learning strategy was proposed. In order to verify the effectiveness of the transfer learning strategy, based on the Swin-Transformer framework, the control model, hybrid training model and transfer learning training model were obtained by training UAV images in 4 years (2019, 2020, 2021, 2023)and 3 study areas(Shucheng, Guohe, Baihe) under 2 flight altitudes (40 and 80 m). To test the model's performance, a comparative experimental approach was adopted to assess the accuracy of the three models for segmenting 80 m altitude images. The assessment relied on five metrics: intersection of union (IoU), accuracy, precision, recall, and F1-score. [Results and Discussions] The transfer learning model shows the highest accuracy in lodging area detection. Specifically, the mean IoU, accuracy, precision, recall, and F1-score achieved 85.37%, 94.98%, 91.30%, 92.52% and 91.84%, respectively. Notably, the accuracy of lodging area detection for images acquired at a 40 m altitude surpassed that of images captured at an 80 m altitude when employing a training dataset composed solely of images obtained at the 40 m altitude. However, when adopting mixed training and transfer learning strategies and augmenting the training dataset with images acquired at an 80 m altitude, the accuracy of lodging area detection for 80 m altitude images improved, inspite of the expense of reduced accuracy for 40 m altitude images. The performance of the mixed training model and the transfer learning model in lodging area detection for both 40 and 80 m altitude images exhibited close correspondence. In a cross-study area comparison of the mean values of model evaluation indices, lodging area detection accuracy was slightly higher for images obtained in Baihu area compared to Shucheng area, while accuracy for images acquired in Shucheng surpassed that of Guohe. These variations could be attributed to the diverse wheat varieties cultivated in Guohe area through drill seeding. The high planting density of wheat in Guohe resulted in substantial lodging areas, accounting for 64.99% during the late mature period. The prevalence of semi-lodging wheat further exacerbated the issue, potentially leading to misidentification of non-lodging areas. Consequently, this led to a reduction in the recall rate (mean recall for Guohe images was 89.77%, which was 4.88% and 3.57% lower than that for Baihu and Shucheng, respectively) and IoU (mean IoU for Guohe images was 80.38%, which was 8.80% and 3.94% lower than that for Baihu and Shucheng, respectively). Additionally, the accuracy, precision, and F1-score for Guohe were also lower compared to Baihu and Shucheng. [Conclusions] This study inspected the efficacy of a strategy aimed at reducing the challenges associated with the insufficient number of high-altitude images for semantic segmentation model training. By pre-training the semantic segmentation model with low-altitude images and subsequently employing high-altitude images for transfer learning, improvements of 1.08% to 3.19% were achieved in mean IoU, accuracy, precision, recall, and F1-score, alongside a notable mean weighted frame rate enhancement of 555.23 fps/m2. The approach proposed in this study holds promise for improving lodging monitoring accuracy and the speed of image segmentation. In practical applications, it is feasible to leverage a substantial quantity of 40 m altitude UAV images collected from diverse study areas including various wheat varieties for pre-training purposes. Subsequently, a limited set of 80 m altitude images acquired in specific study areas can be employed for transfer learning, facilitating the development of a targeted lodging detection model. Future research will explore the utilization of UAV images captured at even higher flight altitudes for further enhancing lodging area detection efficiency.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Artificial Intelligence-Driven High-Quality Development of New-Quality Productivity in Animal Husbandry: Restraining Factors, Generation Logic and Promotion Paths
    LIU Jifang, ZHOU Xiangyang, LI Min, HAN Shuqing, GUO Leifeng, CHI Liang, YANG Lu, WU Jianzhai
    Smart Agriculture    2025, 7 (1): 165-177.   DOI: 10.12133/j.smartag.SA202407010
    Abstract1353)   HTML29)    PDF(pc) (1692KB)(654)       Save

    [Significance] Developing new-quality productivity is of great significance for promoting high-quality development of animal husbandry. However, there is currently limited research on new-quality productivity in animal husbandry, and there is a lack of in-depth analysis on its connotation, characteristics, constraints, and promotion path. [Progress] This article conducts a systematic study on the high-quality development of animal husbandry productivity driven by artificial intelligence. The new-quality productivity of animal husbandry is led by cutting-edge technological innovations such as biotechnology, information technology, and green technology, with digitalization, greening, and ecologicalization as the direction of industrial upgrading. Its basic connotation is manifested as higher quality workers, more advanced labor materials, and a wider range of labor objects. Compared with traditional productivity, the new-quality productivity of animal husbandry is an advanced productivity guided by technological innovation, new development concepts, and centered on the improvement of total factor productivity. It has significant characteristics of high production efficiency, good industrial benefits, and strong sustainable development capabilities. China's new-quality productivity in animal husbandry has a good foundation for development, but it also faces constraints such as insufficient innovation in animal husbandry breeding technology, weak core competitiveness, low mechanization rate of animal husbandry, weak independent research and development capabilities of intelligent equipment, urgent demand for "machine replacement", shortcomings in the quantity and quality of animal husbandry talents, low degree of scale of animal husbandry, and limited level of intelligent management. Artificial intelligence in animal husbandry can be widely used in environmental control, precision feeding, health monitoring and disease prevention and control, supply chain optimization and other fields. Artificial intelligence, through revolutionary breakthroughs in animal husbandry technology represented by digital technology, innovative allocation of productivity factors in animal husbandry linked by data elements, and innovative allocation of productivity factors in animal husbandry adapted to the digital economy, has given birth to new-quality productivity in animal husbandry and empowered the high-quality development of animal husbandry. [Conclusions and Prospects] This article proposes a path to promote the development of new-quality productivity in animal husbandry by improving the institutional mechanism of artificial intelligence to promote the development of modern animal husbandry industry, strengthening the application of artificial intelligence in animal husbandry technology innovation and promotion, and improving the management level of artificial intelligence in the entire industry chain of animal husbandry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0