Welcome to Smart Agriculture 中文

Top Read Articles

    Published in last 1 year |  In last 2 years |  In last 3 years |  All
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Pig Back Transformer: Automatic 3D Pig Body Measurement Model
    WANG Yuxiao, SHI Yuanyuan, CHEN Zhaoda, WU Zhenfang, CAI Gengyuan, ZHANG Sumin, YIN Ling
    Smart Agriculture    2024, 6 (4): 76-90.   DOI: 10.12133/j.smartag.SA202401023
    Abstract1278)   HTML41)    PDF(pc) (2776KB)(1174)       Save

    [Objective] Nowadays most no contact body size measurement studies are based on point cloud segmentation method, they use a trained point cloud segmentation neural network to segment point cloud of pigs, then locate measurement points based on them. But point cloud segmentation neural network always need a larger graphics processing unit (GPU) memory, moreover, the result of the measurement key point still has room of improvement. This study aims to design a key point generating neural network to extract measurement key points from pig's point cloud. Reducing the GPU memory usage and improve the result of measurement points at the same time, improve both the efficiency and accuracy of the body size measurement. [Methods] A neural network model was proposed using improved Transformer attention mechanic called Pig Back Transformer for generating key points and back orientation points which were related to pig body dimensions. In the first part of the network, it was introduced an embedding structure for initial feature extraction and a Transformer encoder structure with edge attention which was a self-attention mechanic improved from Transformer's encoder. The embedding structure using two shared multilayer perceptron (MLP) and a distance embedding algorithm, it takes a set of points from the edge of pig back's point cloud as input and then extract information from the edge points set. In the encoder part, information about the offset distances between edge points and mass point which were their feature that extracted by the embedding structure mentioned before incorporated. Additionally, an extraction algorithm for back edge point was designed for extracting edge points to generate the input of the neural network model. In the second part of the network, it was proposed a Transformer encoder with improved self-attention called back attention. In the design of back attention, it also had an embedding structure before the encoder structure, this embedding structure extracted features from offset values, these offset values were calculated by the points which are none-edge and down sampled by farthest point sampling (FPS) to both the relative centroid point and model generated global key point from the first part that introduced before. Then these offset values were processed with max pooling with attention generated by the extracted features of the points' axis to extract more information that the original Transformer encoder couldn't extract with the same number of parameters. The output part of the model was designed to generate a set of offsets of the key points and points for back direction fitting, than add the set offset to the global key point to get points for pig body measurements. At last, it was introduced the methods for calculating body dimensions which were length, height, shoulder width, abdomen width, hip width, chest circumference and abdomen circumference using key points and back direction fitting points. [Results and Discussions] In the task of generating key points and points for back direction fitting, the improved Pig Back Transformer performed the best in the accuracy wise in the models tested with the same size of parameters, and the back orientation points generated by the model were evenly distributed which was a good preparation for a better body length calculation. A melting test for edge detection part with two attention mechanic and edge trim method both introduced above had being done, when the edge detection and the attention mechanic got cut off, the result had been highly impact, it made the model couldn't perform as well as before, when the edge trim method of preprocessing part had been cut off, there's a moderate impact on the trained model, but it made the loss of the model more inconsistence while training than before. When comparing the body measurement algorithm with human handy results, the relative error in length was 0.63%, which was an improvement compared to other models. On the other hand, the relative error of shoulder width, abdomen width and hip width had edged other models a little but there was no significant improvement so the performance of these measurement accuracy could be considered negligible, the relative error of chest circumference and abdomen circumference were a little bit behind by the other methods existed, it's because the calculate method of circumferences were not complicated enough to cover the edge case in the dataset which were those point cloud that have big holes in the bottom of abdomen and chest, it impacted the result a lot. [Conclusions] The improved Pig Back Transformer demonstrates higher accuracy in generating key points and is more resource-efficient, enabling the calculation of more accurate pig body measurements. And provides a new perspective for non-contact livestock body size measurements.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    CSD-YOLOv8s: Dense Sheep Small Target Detection Model Based on UAV Images
    WENG Zhi, LIU Haixin, ZHENG Zhiqiang
    Smart Agriculture    2024, 6 (4): 42-52.   DOI: 10.12133/j.smartag.SA202401004
    Abstract1269)   HTML92)    PDF(pc) (1772KB)(538)       Save

    [Objective] The monitoring of livestock grazing in natural pastures is a key aspect of the transformation and upgrading of large-scale breeding farms. In order to meet the demand for large-scale farms to achieve accurate real-time detection of a large number of sheep, a high-precision and easy-to-deploy small-target detection model: CSD-YOLOv8s was proposed to realize the real-time detection of small-targeted individual sheep under the high-altitude view of the unmanned aerial vehicle (UAV). [Methods] Firstly, a UAV was used to acquire video data of sheep in natural grassland pastures with different backgrounds and lighting conditions, and together with some public datasets downloaded formed the original image data. The sheep detection dataset was generated through data cleaning and labeling. Secondly, in order to solve the difficult problem of sheep detection caused by dense flocks and mutual occlusion, the SPPFCSPC module was constructed with cross-stage local connection based on the you only look once (YOLO)v8 model, which combined the original features with the output features of the fast spatial pyramid pooling network, fully retained the feature information at different stages of the model, and effectively solved the problem of small targets and serious occlusion of the sheep, and improved the detection performance of the model for small sheep targets. In the Neck part of the model, the convolutional block attention module (CBAM) convolutional attention module was introduced to enhance the feature information capture based on both spatial and channel aspects, suppressing the background information spatially and focusing on the sheep target in the channel, enhancing the network's anti-jamming ability from both channel and spatial dimensions, and improving the model's detection performance of multi-scale sheep under complex backgrounds and different illumination conditions. Finally, in order to improve the real-time and deploy ability of the model, the standard convolution of the Neck network was changed to a lightweight convolutional C2f_DS module with a changeable kernel, which was able to adaptively select the corresponding convolutional kernel for feature extraction according to the input features, and solved the problem of input scale change in the process of sheep detection in a more flexible way, and at the same time, the number of parameters of the model was reduced and the speed of the model was improved. [Results and Discussions] The improved CSD-YOLOv8s model exhibited excellent performance in the sheep detection task. Compared with YOLO, Faster R-CNN and other classical network models, the improved CSD-YOLOv8s model had higher detection accuracy and frames per second (FPS) of 87 f/s in the flock detection task with comparable detection speed and model size. Compared with the YOLOv8s model, Precision was improved from 93.0% to 95.2%, mAP was improved from 91.2% to 93.1%, and it had strong robustness to sheep targets with different degree of occlusion and different scales, which effectively solved the serious problems of missed and misdetection of sheep in the grassland pasture UAV-on-ground sheep detection task due to the small sheep targets, large background noise, and high degree of densification. misdetection serious problems. Validated by the PASCAL VOC 2007 open dataset, the CSD-YOLOv8s model proposed in this study improved the detection accuracy of 20 different objects, including transportation vehicles, animals, etc., especially in sheep detection, the detection accuracy was improved by 9.7%. [Conclusions] This study establishes a sheep dataset based on drone images and proposes a model called CSD-YOLOv8s for detecting grazing sheep in natural grasslands. The model addresses the serious issues of missed detections and false alarms in sheep detection under complex backgrounds and lighting conditions, enabling more accurate detection of grazing livestock in drone images. It achieves precise detection of targets with varying degrees of clustering and occlusion and possesses good real-time performance. This model provides an effective detection method for detecting sheep herds from the perspective of drones in natural pastures and offers technical support for large-scale livestock detection in breeding farms, with wide-ranging potential applications.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospects of Key Navigation Technologies for Facility Agricultural Robots
    HE Yong, HUANG Zhenyu, YANG Ningyuan, LI Xiyao, WANG Yuwei, FENG Xuping
    Smart Agriculture    2024, 6 (5): 1-19.   DOI: 10.12133/j.smartag.SA202404006
    Abstract1193)   HTML294)    PDF(pc) (2130KB)(3995)       Save

    [Significance] With the rapid development of robotics technology and the persistently rise of labor costs, the application of robots in facility agriculture is becoming increasingly widespread. These robots can enhance operational efficiency, reduce labor costs, and minimize human errors. However, the complexity and diversity of facility environments, including varying crop layouts and lighting conditions, impose higher demands on robot navigation. Therefore, achieving stable, accurate, and rapid navigation for robots has become a key issue. Advanced sensor technologies and algorithms have been proposed to enhance robots' adaptability and decision-making capabilities in dynamic environments. This not only elevates the automation level of agricultural production but also contributes to more intelligent agricultural management. [Progress] This paper reviews the key technologies of automatic navigation for facility agricultural robots. It details beacon localization, inertial positioning, simultaneous localization and mapping (SLAM) techniques, and sensor fusion methods used in autonomous localization and mapping. Depending on the type of sensors employed, SLAM technology could be subdivided into vision-based, laser-based and fusion systems. Fusion localization is further categorized into data-level, feature-level, and decision-level based on the types and stages of the fused information. The application of SLAM technology and fusion localization in facility agriculture has been increasingly common. Global path planning plays a crucial role in enhancing the operational efficiency and safety of facility aricultural robots. This paper discusses global path planning, classifying it into point-to-point local path planning and global traversal path planning. Furthermore, based on the number of optimization objectives, it was divided into single-objective path planning and multi-objective path planning. In regard to automatic obstacle avoidance technology for robots, the paper discusses sevelral commonly used obstacle avoidance control algorithms commonly used in facility agriculture, including artificial potential field, dynamic window approach and deep learning method. Among them, deep learning methods are often employed for perception and decision-making in obstacle avoidance scenarios. [Conclusions and Prospects] Currently, the challenges for facility agricultural robot navigation include complex scenarios with significant occlusions, cost constraints, low operational efficiency and the lack of standardized platforms and public datasets. These issues not only affect the practical application effectiveness of robots but also constrain the further advancement of the industry. To address these challenges, future research can focus on developing multi-sensor fusion technologies, applying and optimizing advanced algorithms, investigating and implementing multi-robot collaborative operations and establishing standardized and shared data platforms.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Regional Farming Pig Counting System Based on Improved Instance Segmentation Algorithm
    ZHANG Yanqi, ZHOU Shuo, ZHANG Ning, CHAI Xiujuan, SUN Tan
    Smart Agriculture    2024, 6 (4): 53-63.   DOI: 10.12133/j.smartag.SA202310001
    Abstract1179)   HTML64)    PDF(pc) (2077KB)(592)       Save

    [Objective] Currently, pig farming facilities mainly rely on manual counting for tracking slaughtered and stored pigs. This is not only time-consuming and labor-intensive, but also prone to counting errors due to pig movement and potential cheating. As breeding operations expand, the periodic live asset inventories put significant strain on human, material and financial resources. Although methods based on electronic ear tags can assist in pig counting, these ear tags are easy to break and fall off in group housing environments. Most of the existing methods for counting pigs based on computer vision require capturing images from a top-down perspective, necessitating the installation of cameras above each hogpen or even the use of drones, resulting in high installation and maintenance costs. To address the above challenges faced in the group pig counting task, a high-efficiency and low-cost pig counting method was proposed based on improved instance segmentation algorithm and WeChat public platform. [Methods] Firstly, a smartphone was used to collect pig image data in the area from a human view perspective, and each pig's outline in the image was annotated to establish a pig count dataset. The training set contains 606 images and the test set contains 65 images. Secondly, an efficient global attention module was proposed by improving convolutional block attention module (CBAM). The efficient global attention module first performed a dimension permutation operation on the input feature map to obtain the interaction between its channels and spatial dimensions. The permuted features were aggregated using global average pooling (GAP). One-dimensional convolution replaced the fully connected operation in CBAM, eliminating dimensionality reduction and significantly reducing the model's parameter number. This module was integrated into the YOLOv8 single-stage instance segmentation network to build the pig counting model YOLOv8x-Ours. By adding an efficient global attention module into each C2f layer of the YOLOv8 backbone network, the dimensional dependencies and feature information in the image could be extracted more effectively, thereby achieving high-accuracy pig counting. Lastly, with a focus on user experience and outreach, a pig counting WeChat mini program was developed based on the WeChat public platform and Django Web framework. The counting model was deployed to count pigs using images captured by smartphones. [Results and Discussions] Compared with existing methods of Mask R-CNN, YOLACT(Real-time Instance Segmentation), PolarMask, SOLO and YOLOv5x, the proposed pig counting model YOLOv8x-Ours exhibited superior performance in terms of accuracy and stability. Notably, YOLOv8x-Ours achieved the highest accuracy in counting, with errors of less than 2 and 3 pigs on the test set. Specifically, 93.8% of the total test images had counting errors of less than 3 pigs. Compared with the two-stage instance segmentation algorithm Mask R-CNN and the YOLOv8x model that applies the CBAM attention mechanism, YOLOv8x-Ours showed performance improvements of 7.6% and 3%, respectively. And due to the single-stage design and anchor-free architecture of the YOLOv8 model, the processing speed of a single image was only 64 ms, 1/8 of Mask R-CNN. By embedding the model into the WeChat mini program platform, pig counting was conducted using smartphone images. In cases where the model incorrectly detected pigs, users were given the option to click on the erroneous location in the result image to adjust the statistical outcomes, thereby enhancing the accuracy of pig counting. [Conclusions] The feasibility of deep learning technology in the task of pig counting was demonstrated. The proposed method eliminates the need for installing hardware equipment in the breeding area of the pig farm, enabling pig counting to be carried out effortlessly using just a smartphone. Users can promptly spot any errors in the counting results through image segmentation visualization and easily rectify any inaccuracies. This collaborative human-machine model not only reduces the need for extensive manpower but also guarantees the precision and user-friendliness of the counting outcomes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Prospect of Intelligent Monitoring Systems for the Physiological Indicators of Beef Cattle
    ZHANG Fan, ZHOU Mengting, XIONG Benhai, YANG Zhengang, LIU Minze, FENG Wenxiao, TANG Xiangfang
    Smart Agriculture    2024, 6 (4): 1-17.   DOI: 10.12133/j.smartag.SA202312001
    Abstract1172)   HTML80)    PDF(pc) (1471KB)(1434)       Save

    [Significance] The beef cattle industry plays a pivotal role in the development of China's agricultural economy and the enhancement of people's dietary structure. However, there exists a substantial disparity in feeding management practices and economic efficiency of beef cattle industry compared to developed countries. While the beef cattle industry in China is progressing towards intensive, modern, and large-scale development, it encounters challenges such as labor shortage and rising labor costs that seriously affect its healthy development. The determination of animal physiological indicators plays an important role in monitoring animal welfare and health status. Therefore, leveraging data collected from various sensors as well as technologies like machine learning, data mining, and modeling analysis enables automatic acquisition of meaningful information on beef cattle physiological indicators for intelligent management of beef cattle. In this paper, the intelligent monitoring technology of physiological indicators in beef cattle breeding process and its application value are systematically summarized, and the existing challenges and future prospects of intelligent beef cattle breeding process in China are prospected. [Progress] The methods of obtaining information on beef cattle physiological indicators include contact sensors worn on the body and non-contact sensors based on various image acquisitions. Monitoring the exercise behavior of beef cattle plays a crucial role in disease prevention, reproduction monitoring, and status assessment. The three-axis accelerometer sensor, which tracks the amount of time that beef cattle spend on lying, walking, and standing, is a widely used technique for tracking the movement behavior of beef cattle. Through machine vision analysis, individual recognition of beef cattle and identification of standing, lying down, and straddling movements can also be achieved, with the characteristics of non-contact, stress-free, low cost, and generating high data volume. Body temperature in beef cattle is associated with estrus, calving, and overall health. Sensors for monitoring body temperature include rumen temperature sensors and rectal temperature sensors, but there are issues with their inconvenience. Infrared temperature measurement technology can be utilized to detect beef cattle with abnormal temperatures by monitoring eye and ear root temperatures, although the accuracy of the results may be influenced by environmental temperature and monitoring distance, necessitating calibration. Heart rate and respiratory rate in beef cattle are linked to animal diseases, stress, and pest attacks. Monitoring heart rate can be accomplished through photoelectric volume pulse wave measurement and monitoring changes in arterial blood flow using infrared emitters and receivers. Respiratory rate monitoring can be achieved by identifying different nostril temperatures during inhalation and exhalation using thermal infrared imaging technology. The ruminating behavior of beef cattle is associated with health and feed nutrition. Currently, the primary tools used to detect rumination behavior are pressure sensors and three-axis accelerometer sensors positioned at various head positions. Rumen acidosis is a major disease in the rapid fattening process of beef cattle, however, due to limitations in battery life and electrode usage, real-time pH monitoring sensors placed in the rumen are still not widely utilized. Changes in animal physiology, growth, and health can result in alterations in specific components within body fluids. Therefore, monitoring body fluids or surrounding gases through biosensors can be employed to monitor the physiological status of beef cattle. By processing and analyzing the physiological information of beef cattle, indicators such as estrus, calving, feeding, drinking, health conditions, and stress levels can be monitored. This will contribute to the intelligent development of the beef cattle industry and enhance management efficiency. While there has been some progress made in developing technology for monitoring physiological indicators of beef cattle, there are still some challenges that need to be addressed. Contact sensors consume more energy which affects their lifespan. Various sensors are susceptible to environmental interference which affects measurement accuracy. Additionally, due to a wide variety of beef cattle breeds, it is difficult to establish a model database for monitoring physiological indicators under different feeding conditions, breeding stages, and breeds. Furthermore, the installation cost of various intelligent monitoring devices is relatively high, which also limits its utilization coverage. [Conclusion and Prospects] The application of intelligent monitoring technology for beef cattle physiological indicators is highly significance in enhancing the management level of beef cattle feeding. Intelligent monitoring systems and devices are utilized to acquire physiological behavior data, which are then analyzed using corresponding data models or classified through deep learning techniques to promptly monitor subtle changes in physiological indicators. This enables timely detection of sick, estrus, and calving cattle, facilitating prompt measures by production managers, reducing personnel workload, and improving efficiency. The future development of physiological indicators monitoring technologies in beef cattle primarily focuses on the following three aspects: (1) Enhancing the lifespan of contact sensors by reducing energy consumption, decreasing data transmission frequency, and improving battery life. (2) Integrating and analyzing various monitoring data from multiple perspectives to enhance the accuracy and utility value. (3) Strengthening research on non-contact, high-precision and automated analysis technologies to promote the precise and intelligent development within the beef cattle industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Measurement Method of Beef Cattle Body Size Based on Multimodal Image Information and Improved Instance Segmentation Network
    WENG Zhi, FAN Qi, ZHENG Zhiqiang
    Smart Agriculture    2024, 6 (4): 64-75.   DOI: 10.12133/j.smartag.SA202310007
    Abstract1171)   HTML40)    PDF(pc) (3345KB)(1278)       Save

    [Objective] The body size parameter of cattle is a key indicator reflecting the physical development of cattle, and is also a key factor in the cattle selection and breeding process. In order to solve the demand of measuring body size of beef cattle in the complex environment of large-scale beef cattle ranch, an image acquisition device and an automatic measurement algorithm of body size were designed. [Methods] Firstly, the walking channel of the beef cattle was established, and when the beef cattle entered the restraining device through the channel, the RGB and depth maps of the image on the right side of the beef cattle were acquired using the Inter RealSense D455 camera. Secondly, in order to avoid the influence of the complex environmental background, an improved instance segmentation network based on Mask2former was proposed, adding CBAM module and CA module, respectively, to improve the model's ability to extract key features from different perspectives, extracting the foreground contour from the 2D image of the cattle, partitioning the contour, and comparing it with other segmentation algorithms, and using curvature calculation and other mathematical methods to find the required body size measurement points. Thirdly, in the processing of 3D data, in order to solve the problem that the pixel point to be measured in the 2D RGB image was null when it was projected to the corresponding pixel coordinates in the depth-valued image, resulting in the inability to calculate the 3D coordinates of the point, a series of processing was performed on the point cloud data, and a suitable point cloud filtering and point cloud segmentation algorithm was selected to effectively retain the point cloud data of the region of the cattle's body to be measured, and then the depth map was 16. Then the depth map was filled with nulls in the field to retain the integrity of the point cloud in the cattle body region, so that the required measurement points could be found and the 2D data could be returned. Finally, an extraction algorithm was designed to combine 2D and 3D data to project the extracted 2D pixel points into a 3D point cloud, and the camera parameters were used to calculate the world coordinates of the projected points, thus automatically calculating the body measurements of the beef cattle. [Results and Discussions] Firstly, in the part of instance segmentation, compared with the classical Mask R-CNN and the recent instance segmentation networks PointRend and Queryinst, the improved network could extract higher precision and smoother foreground images of cattles in terms of segmentation accuracy and segmentation effect, no matter it was for the case of occlusion or for the case of multiple cattles. Secondly, in three-dimensional data processing, the method proposed in the study could effectively extract the three-dimensional data of the target area. Thirdly, the measurement error of body size was analysed, among the four body size measurement parameters, the smallest average relative error was the height of the cross section, which was due to the more prominent position of the cross section, and the different standing positions of the cattle have less influence on the position of the cross section, and the largest average relative error was the pipe circumference, which was due to the influence of the greater overlap of the two front legs, and the higher requirements for the standing position. Finally, automatic body measurements were carried out on 137 beef cattle in the ranch, and the automatic measurements of the four body measurements parameters were compared with the manual measurements, and the results showed that the average relative errors of body height, cross section height, body slant length, and tube girth were 4.32%, 3.71%, 5.58% and 6.25%, respectively, which met the needs of the ranch. The shortcomings were that fewer body-size parameters were measured, and the error of measuring circumference-type body-size parameters was relatively large. Later studies could use a multi-view approach to increase the number of body rule parameters to be measured and improve the accuracy of the parameters in the circumference category. [Conclusions] The article designed an automatic measurement method based on two-dimensional and three-dimensional contactless body measurements of beef cattle. Moreover, the innovatively proposed method of measuring tube girth has higher accuracy and better implementation compared with the current research on body measurements in beef cattle. The relative average errors of the four body tape parameters meet the needs of pasture measurements and provide theoretical and practical guidance for the automatic measurement of body tape in beef cattle.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Measurement of Mongolian Horse Body Based on Improved YOLOv8n-pose and 3D Point Cloud Analysis
    LI Minghuang, SU Lide, ZHANG Yong, ZONG Zheying, ZHANG Shun
    Smart Agriculture    2024, 6 (4): 91-102.   DOI: 10.12133/j.smartag.SA202312027
    Abstract1141)   HTML55)    PDF(pc) (2477KB)(1631)       Save

    [Objective] There exist a high genetic correlation among various morphological characteristics of Mongolian horses. Utilizing advanced technology to obtain body structure parameters related to athletic performance could provide data support for breeding institutions to develop scientific breeding plans and establish the groundwork for further improvement of Mongolian horse breeds. However, traditional manual measurement methods are time-consuming, labor-intensive, and may cause certain stress responses in horses. Therefore, ensuring precise and effective measurement of Mongolian horse body dimensions is crucial for formulating early breeding plans. [Method] Video images of 50 adult Mongolian horses in the suitable breeding stage at the Inner Mongolia Agricultural University Horse Breeding Technical Center was first collected. Fifty images per horse were captured to construct the training and validation sets, resulting in a total of 2 500 high-definition RGB images of Mongolian horses, with an equal ratio of images depicting horses in motion and at rest. To ensure the model's robustness and considering issues such as angles, lighting, and image blurring during actual image capture, a series of enhancement algorithms were applied to the original dataset, expanding the Mongolian horse image dataset to 4 000 images. The YOLOv8n-pose was employed as the foundational keypoint detection model. Through the design of the C2f_DCN module, deformable convolution (DCNV2) was integrated into the C2f module of the Backbone network to enhance the model's adaptability to different horse poses in real-world scenes. Besides, an SA attention module was added to the Neck network to improve the model's focus on critical features. The original loss function was replaced with SCYLLA-IoU (SIoU) to prioritize major image regions, and a cosine annealing method was employed to dynamically adjust the learning rate during model training. The improved model was named DSS-YOLO (DCNv2-SA-SIoU-YOLO) network model. Additionally, a test set comprising 30 RGB-D images of mature Mongolian horses was selected for constructing body dimension measurement tasks. DSS-YOLO was used for keypoint detection of body dimensions. The 2D keypoint coordinates from RGB images were fused with corresponding depth values from depth images to obtain 3D keypoint coordinates, and Mongolian horse's point cloud information was transformed. Point cloud processing and analysis were performed using pass-through filtering, random sample consensus (RANSAC) shape fitting, statistical outlier filtering, and principal component analysis (PCA) coordinate system correction. Finally, body height, body oblique length, croup height, chest circumference, and croup circumference were automatically computed based on keypoint spatial coordinates. [Results and Discussion] The proposed DSS-YOLO model exhibited parameter and computational costs of 3.48 M and 9.1 G, respectively, with an average accuracy mAP0.5:0.95 reaching 92.5%, and a dDSS of 7.2 pixels. Compared to Hourglass, HRNet, and SimCC, mAP0.5:0.95 increased by 3.6%, 2.8%, and 1.6%, respectively. By relying on keypoint coordinates for automatic calculation of body dimensions and suggesting the use of a mobile least squares curve fitting method to complete the horse's hip point cloud, experiments involving 30 Mongolian horses showed a mean average error (MAE) of 3.77 cm and mean relative error (MRE) of 2.29% in automatic measurements. [Conclusions] The results of this study showed that DSS-YOLO model combined with three-dimensional point cloud processing methods can achieve automatic measurement of Mongolian horse body dimensions with high accuracy. The proposed measurement method can also be extended to different breeds of horses, providing technical support for horse breeding plans and possessing practical application value.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    The Path of Smart Agricultural Technology Innovation Leading Development of Agricultural New Quality Productivity
    CAO Bingxue, LI Hongfei, ZHAO Chunjiang, LI Jin
    Smart Agriculture    2024, 6 (4): 116-127.   DOI: 10.12133/j.smartag.SA202405004
    Abstract1121)   HTML205)    PDF(pc) (1102KB)(2355)       Save

    [Significance] Building the agricultural new quality productivity is of great significance. It is the advanced quality productivity which realizes the transformation, upgrading, and deep integration of substantive, penetrating, operational, and media factors, and has outstanding characteristics such as intelligence, greenness, integration, and organization. As a new technology revolution in the field of agriculture, smart agricultural technology transforms agricultural production mode by integrating agricultural biotechnology, agricultural information technology, and smart agricultural machinery and equipment, with information and knowledge as important core elements. The inherent characteristics of "high-tech, high-efficiency, high-quality, and sustainable" in agricultural new quality productivity are fully reflected in the practice of smart agricultural technology innovation. And it has become an important core and engine for promoting the agricultural new quality productivity. [Progress] Through literature review and theoretical analysis, this article conducts a systematic study on the practical foundation, internal logic, and problem challenges of smart agricultural technology innovation leading the development of agricultural new quality productivity. The conclusions show that: (1) At present, the global innovation capability of smart agriculture technology is constantly enhancing, and significant technology breakthroughs have been made in fields such as smart breeding, agricultural information perception, agricultural big data and artificial intelligence, smart agricultural machinery and equipment, providing practical foundation support for leading the development of agricultural new quality productivity. Among them, the smart breeding of 'Phenotype+Genotype+Environmental type' has entered the fast lane, the technology system for sensing agricultural sky, air, and land information is gradually maturing, the research and exploration on agricultural big data and intelligent decision-making technology continue to advance, and the creation of smart agricultural machinery and equipment for different fields has achieved fruitful results; (2) Smart agricultural technology innovation provides basic resources for the development of agricultural new quality productivity through empowering agricultural factor innovation, provides sustainable driving force for the development of agricultural new quality productivity through empowering agricultural technology innovation, provides practical paradigms for the development of agricultural new quality productivity through empowering agricultural scenario innovation, provides intellectual support for the development of agricultural new quality productivity through empowering agricultural entity innovation, and provides important guidelines for the development of agricultural new quality productivity through empowering agricultural value innovation; (3) Compared to the development requirements of agricultural new quality productivity in China and the advanced level of international smart agriculture technology, China's smart agriculture technology innovation is generally in the initial stage of multi-point breakthroughs, system integration, and commercial application. It still faces major challenges such as an incomplete policy system for technology innovation, key technologies with bottlenecks, blockages and breakpoints, difficulties in the transformation and implementation of technology achievements, and incomplete support systems for technology innovation. [Conclusions and Prospects] Regarding the issue of technology innovation in smart agriculture, this article proposes the 'Four Highs' path of smart agriculture technology innovation to fill the gaps in smart agriculture technology innovation and accelerate the formation of agricultural new quality productivity in China. The "Four Highs" path specifically includes the construction of high-energy smart agricultural technology innovation platforms, the breakthroughs in high-precision and cutting-edge smart agricultural technology products, the creation of high-level smart agricultural application scenarios, and the cultivation of high-level smart agricultural innovation talents. Finally, this article proposes four strategic suggestions such as deepening the understanding of smart agriculture technology innovation and agricultural new quality productivity, optimizing the supply of smart agriculture technology innovation policies, building a national smart agriculture innovation development pilot zone, and improving the smart agriculture technology innovation ecosystem.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Artificial Intelligence-Driven High-Quality Development of New-Quality Productivity in Animal Husbandry: Restraining Factors, Generation Logic and Promotion Paths
    LIU Jifang, ZHOU Xiangyang, LI Min, HAN Shuqing, GUO Leifeng, CHI Liang, YANG Lu, WU Jianzhai
    Smart Agriculture    2025, 7 (1): 165-177.   DOI: 10.12133/j.smartag.SA202407010
    Abstract1059)   HTML15)    PDF(pc) (1692KB)(193)       Save

    [Significance] Developing new-quality productivity is of great significance for promoting high-quality development of animal husbandry. However, there is currently limited research on new-quality productivity in animal husbandry, and there is a lack of in-depth analysis on its connotation, characteristics, constraints, and promotion path. [Progress] This article conducts a systematic study on the high-quality development of animal husbandry productivity driven by artificial intelligence. The new-quality productivity of animal husbandry is led by cutting-edge technological innovations such as biotechnology, information technology, and green technology, with digitalization, greening, and ecologicalization as the direction of industrial upgrading. Its basic connotation is manifested as higher quality workers, more advanced labor materials, and a wider range of labor objects. Compared with traditional productivity, the new-quality productivity of animal husbandry is an advanced productivity guided by technological innovation, new development concepts, and centered on the improvement of total factor productivity. It has significant characteristics of high production efficiency, good industrial benefits, and strong sustainable development capabilities. China's new-quality productivity in animal husbandry has a good foundation for development, but it also faces constraints such as insufficient innovation in animal husbandry breeding technology, weak core competitiveness, low mechanization rate of animal husbandry, weak independent research and development capabilities of intelligent equipment, urgent demand for "machine replacement", shortcomings in the quantity and quality of animal husbandry talents, low degree of scale of animal husbandry, and limited level of intelligent management. Artificial intelligence in animal husbandry can be widely used in environmental control, precision feeding, health monitoring and disease prevention and control, supply chain optimization and other fields. Artificial intelligence, through revolutionary breakthroughs in animal husbandry technology represented by digital technology, innovative allocation of productivity factors in animal husbandry linked by data elements, and innovative allocation of productivity factors in animal husbandry adapted to the digital economy, has given birth to new-quality productivity in animal husbandry and empowered the high-quality development of animal husbandry. [Conclusions and Prospects] This article proposes a path to promote the development of new-quality productivity in animal husbandry by improving the institutional mechanism of artificial intelligence to promote the development of modern animal husbandry industry, strengthening the application of artificial intelligence in animal husbandry technology innovation and promotion, and improving the management level of artificial intelligence in the entire industry chain of animal husbandry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Navigation and Spraying Robot in Sheep Farm
    FAN Mingshuo, ZHOU Ping, LI Miao, LI Hualong, LIU Xianwang, MA Zhirun
    Smart Agriculture    2024, 6 (4): 103-115.   DOI: 10.12133/j.smartag.SA202312016
    Abstract960)   HTML30)    PDF(pc) (2160KB)(661)       Save

    [Objective] Manual disinfection in large-scale sheep farm is laborious, time-consuming, and often results in incomplete coverage and inadequate disinfection. With the rapid development of the application of artificial intelligence and automation technology, the automatic navigation and spraying robot for livestock and poultry breeding, has become a research hotspot. To maintain shed hygiene and ensure sheep health, an automatic navigation and spraying robot was proposed for sheep sheds. [Methods] The automatic navigation and spraying robot was designed with a focus on three aspects: hardware, semantic segmentation model, and control algorithm. In terms of hardware, it consisted of a tracked chassis, cameras, and a collapsible spraying device. For the semantic segmentation model, enhancements were made to the lightweight semantic segmentation model ENet, including the addition of residual structures to prevent network degradation and the incorporation of a squeeze-and-excitation network (SENet) attention mechanism in the initialization module. This helped to capture global features when feature map resolution was high, addressing precision issues. The original 6-layer ENet network was reduced to 5 layers to balance the encoder and decoder. Drawing inspiration from dilated spatial pyramid pooling, a context convolution module (CCM) was introduced to improve scene understanding. A criss-cross attention (CCA) mechanism was adapted to acquire context global features of different scales without cascading, reducing information loss. This led to the development of a double attention enet (DAENet) semantic segmentation model was proposed to achieve real-time and accurate segmentation of sheep shed surfaces. Regarding control algorithms, a method was devised to address the robot's difficulty in controlling its direction at junctions. Lane recognition and lane center point identification algorithms were proposed to identify and mark navigation points during the robot's movement outside the sheep shed by simulating real roads. Two cameras were employed, and a camera switching algorithm was developed to enable seamless switching between them while also controlling the spraying device. Additionally, a novel offset and velocity calculation algorithm was proposed to control the speeds of the robot's left and right tracks, enabling control over the robot's movement, stopping, and turning. [Results and Discussions] The DAENet model achieved a mean intersection over union (mIoU) of 0.945 3 in image segmentation tasks, meeting the required segmentation accuracy. During testing of the camera switching algorithm, it was observed that the time taken for the complete transition from camera to spraying device action does not exceed 15 seconds when road conditions changed. Testing of the center point and offset calculation algorithm revealed that, when processing multiple frames of video streams, the algorithm averages 0.04 to 0.055 per frame, achieving frame rates of 20 to 24 frames per second, meeting real-time operational requirements. In field experiments conducted in sheep farm, the robot successfully completed automatic navigation and spraying tasks in two sheds without colliding with roadside troughs. The deviation from the road and lane centerlines did not exceed 0.3 meters. Operating at a travel speed of 0.2 m/s, the liquid in the medicine tank was adequate to complete the spraying tasks for two sheds. Additionally, the time taken for the complete transition from camera to spraying device action did not exceed 15 when road conditions changed. The robot maintained an average frame rate of 22.4 frames per second during operation, meeting the experimental requirements for accurate and real-time information processing. Observation indicated that the spraying coverage rate of the robot exceeds 90%, meeting the experimental coverage requirements. [Conclusions] The proposed automatic navigation and spraying robot, based on the DAENet semantic segmentation model and center point recognition algorithm, combined with hardware design and control algorithms, achieves comprehensive disinfection within sheep sheds while ensuring safety and real-time operation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Real-Time Monitoring Method for Cow Rumination Behavior Based on Edge Computing and Improved MobileNet v3
    ZHANG Yu, LI Xiangting, SUN Yalin, XUE Aidi, ZHANG Yi, JIANG Hailong, SHEN Weizheng
    Smart Agriculture    2024, 6 (4): 29-41.   DOI: 10.12133/j.smartag.SA202405023
    Abstract945)   HTML45)    PDF(pc) (1694KB)(307)       Save

    [Objective] Real-time monitoring of cow ruminant behavior is of paramount importance for promptly obtaining relevant information about cow health and predicting cow diseases. Currently, various strategies have been proposed for monitoring cow ruminant behavior, including video surveillance, sound recognition, and sensor monitoring methods. However, the application of edge device gives rise to the issue of inadequate real-time performance. To reduce the volume of data transmission and cloud computing workload while achieving real-time monitoring of dairy cow rumination behavior, a real-time monitoring method was proposed for cow ruminant behavior based on edge computing. [Methods] Autonomously designed edge devices were utilized to collect and process six-axis acceleration signals from cows in real-time. Based on these six-axis data, two distinct strategies, federated edge intelligence and split edge intelligence, were investigated for the real-time recognition of cow ruminant behavior. Focused on the real-time recognition method for cow ruminant behavior leveraging federated edge intelligence, the CA-MobileNet v3 network was proposed by enhancing the MobileNet v3 network with a collaborative attention mechanism. Additionally, a federated edge intelligence model was designed utilizing the CA-MobileNet v3 network and the FedAvg federated aggregation algorithm. In the study on split edge intelligence, a split edge intelligence model named MobileNet-LSTM was designed by integrating the MobileNet v3 network with a fusion collaborative attention mechanism and the Bi-LSTM network. [Results and Discussions] Through comparative experiments with MobileNet v3 and MobileNet-LSTM, the federated edge intelligence model based on CA-MobileNet v3 achieved an average Precision rate, Recall rate, F1-Score, Specificity, and Accuracy of 97.1%, 97.9%, 97.5%, 98.3%, and 98.2%, respectively, yielding the best recognition performance. [Conclusions] It is provided a real-time and effective method for monitoring cow ruminant behavior, and the proposed federated edge intelligence model can be applied in practical settings.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Vegetable Crop Growth Modeling in Digital Twin Platform Based on Large Language Model Inference
    ZHAO Chunjiang, LI Jingchen, WU Huarui, YANG Yusen
    Smart Agriculture    2024, 6 (6): 63-71.   DOI: 10.12133/j.smartag.SA202410008
    Abstract871)   HTML150)    PDF(pc) (1460KB)(880)       Save

    [Objective] In the era of digital agriculture, real-time monitoring and predictive modeling of crop growth are paramount, especially in autonomous farming systems. Traditional crop growth models, often constrained by their reliance on static, rule-based methods, fail to capture the dynamic and multifactorial nature of vegetable crop growth. This research tried to address these challenges by leveraging the advanced reasoning capabilities of pre-trained large language models (LLMs) to simulate and predict vegetable crop growth with accuracy and reliability. Modeling the growth of vegetable crops within these platforms has historically been hindered by the complex interactions among biotic and abiotic factors. [Methods] The methodology was structured in several distinct phases. Initially, a comprehensive dataset was curated to include extensive information on vegetable crop growth cycles, environmental conditions, and management practices. This dataset incorporates continuous data streams such as soil moisture, nutrient levels, climate variables, pest occurrence, and historical growth records. By combining these data sources, the study ensured that the model was well-equipped to understand and infer the complex interdependencies inherent in crop growth processes. Then, advanced techniques was emploied for pre-training and fine-tuning LLMs to adapt them to the domain-specific requirements of vegetable crop modeling. A staged intelligent agent ensemble was designed to work within the digital twin platform, consisting of a central managerial agent and multiple stage-specific agents. The managerial agent was responsible for identifying transitions between distinct growth stages of the crops, while the stage-specific agents were tailored to handle the unique characteristics of each growth phase. This modular architecture enhanced the model's adaptability and precision, ensuring that each phase of growth received specialized attention and analysis. [Results and Discussions] The experimental validation of this method was conducted in a controlled agricultural setting at the Xiaotangshan Modern Agricultural Demonstration Park in Beijing. Cabbage (Zhonggan 21) was selected as the test crop due to its significance in agricultural production and the availability of comprehensive historical growth data. Over five years, the dataset collected included 4 300 detailed records, documenting parameters such as plant height, leaf count, soil conditions, irrigation schedules, fertilization practices, and pest management interventions. This dataset was used to train the LLM-based system and evaluate its performance using ten-fold cross-validation. The results of the experiments demonstrating the efficacy of the proposed system in addressing the complexities of vegetable crop growth modeling. The LLM-based model achieved 98% accuracy in predicting crop growth degrees and a 99.7% accuracy in identifying growth stages. These metrics significantly outperform traditional machine learning approaches, including long short-term memory (LSTM), XGBoost, and LightGBM models. The superior performance of the LLM-based system highlights its ability to reason over heterogeneous data inputs and make precise predictions, setting a new benchmark for crop modeling technologies. Beyond accuracy, the LLM-powered system also excels in its ability to simulate growth trajectories over extended periods, enabling farmers and agricultural managers to anticipate potential challenges and make proactive decisions. For example, by integrating real-time sensor data with historical patterns, the system can predict how changes in irrigation or fertilization practices will impact crop health and yield. This predictive capability is invaluable for optimizing resource allocation and mitigating risks associated with climate variability and pest outbreaks. [Conclusions] The study emphasizes the importance of high-quality data in achieving reliable and generalizable models. The comprehensive dataset used in this research not only captures the nuances of cabbage growth but also provides a blueprint for extending the model to other crops. In conclusion, this research demonstrates the transformative potential of combining large language models with digital twin technology for vegetable crop growth modeling. By addressing the limitations of traditional modeling approaches and harnessing the advanced reasoning capabilities of LLMs, the proposed system sets a new standard for precision agriculture. Several avenues also are proposed for future work, including expanding the dataset, refining the model architecture, and developing multi-crop and multi-region capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Automatic Detection Method of Dairy Cow Lameness from Top-view Based on the Fusion of Spatiotemporal Stream Features
    DAI Xin, WANG Junhao, ZHANG Yi, WANG Xinjie, LI Yanxing, DAI Baisheng, SHEN Weizheng
    Smart Agriculture    2024, 6 (4): 18-28.   DOI: 10.12133/j.smartag.SA202405025
    Abstract865)   HTML31)    PDF(pc) (1828KB)(536)       Save

    [Objective] The detection of lameness in dairy cows is an important issue that needs to be solved urgently in the process of large-scale dairy farming. Timely detection and effective intervention can reduce the culling rate of young dairy cows, which has important practical significance for increasing the milk production of dairy cows and improving the economic benefits of pastures. Due to the low efficiency and low degree of automation of traditional manual detection and contact sensor detection, the mainstream cow lameness detection method is mainly based on computer vision. The detection perspective of existing computer vision-based cow lameness detection methods is mainly side view, but the side view perspective has limitations that are difficult to eliminate. In the actual detection process, there are problems such as cows blocking each other and difficulty in deployment. The cow lameness detection method from the top view will not be difficult to use on the farm due to occlusion problems. The aim is to solve the occlusion problem under the side view. [Methods] In order to fully explore the movement undulations of the trunk of the cow and the movement information in the time dimension during the walking process of the cow, a cow lameness detection method was proposed from a top view based on fused spatiotemporal flow features. By analyzing the height changes of the lame cow in the depth video stream during movement, a spatial stream feature image sequence was constructed. By analyzing the instantaneous speed of the lame cow's body moving forward and swaying left and right when walking, optical flow was used to capture the instantaneous speed of the cow's movement, and a time flow characteristic image sequence was constructed. The spatial flow and time flow features were combined to construct a fused spatiotemporal flow feature image sequence. Different from traditional image classification tasks, the image sequence of cows walking includes features in both time and space dimensions. There would be a certain distinction between lame cows and non-lame cows due to their related postures and walking speeds when walking, so using video information analysis was feasible to characterize lameness as a behavior. The video action classification network could effectively model the spatiotemporal information in the input image sequence and output the corresponding category in the predicted result. The attention module Convolutional Block Attention Module (CBAM) was used to improve the PP-TSMv2 video action classification network and build the Cow-TSM cow lameness detection model. The CBAM module could perform channel weighting on different modes of cows, while paying attention to the weights between pixels to improve the model's feature extraction capabilities. Finally, cow lameness experiments were conducted on different modalities, different attention mechanisms, different video action classification networks and comparison of existing methods. The data was used for cow lameness included a total of 180 video streams of cows walking. Each video was decomposed into 100‒400 frames. The ratio of the number of video segments of lame cows and normal cows was 1:1. For the feature extraction of cow lameness from the top view, RGB images had less extractable information, so this work mainly used depth video streams. [Results and Discussions] In this study, a total of 180 segments of cow image sequence data were acquired and processed, including 90 lame cows and 90 non-lame cows with a 1:1 ratio of video segments, and the prediction accuracy of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features reaches 88.7%, the model size was 22 M, and the offline inference time was 0.046 s. The prediction accuracy of the common mainstream video action classification models TSM, PP-TSM, SlowFast and TimesFormer models on the data set of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features reached 66.7%, 84.8%, 87.1% and 85.7%, respectively. The comprehensive performance of the improved Cow-TSM model in this paper was the most. At the same time, the recognition accuracy of the fused spatiotemporal flow feature image was improved by 12% and 4.1%, respectively, compared with the temporal mode and spatial mode, which proved the effectiveness of spatiotemporal flow fusion in this method. By conducting ablation experiments on different attention mechanisms of SE, SK, CA and CBAM, it was proved that the CBAM attention mechanism used has the best effect on the data of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features. The channel attention in CBAM had a better effect on fused spatiotemporal flow data, and the spatial attention could also focus on the key spatial information in cow images. Finally, comparisons were made with existing lameness detection methods, including different methods from side view and top view. Compared with existing methods in the side-view perspective, the prediction accuracy of automatic detection method for dairy cow lameness based on fusion of spatiotemporal stream features was slightly lower, because the side-view perspective had more effective cow lameness characteristics. Compared with the method from the top view, a novel fused spatiotemporal flow feature detection method with better performance and practicability was proposed. [Conclusions] This method can avoid the occlusion problem of detecting lame cows from the side view, and at the same time improves the prediction accuracy of the detection method from the top view. It is of great significance for reducing the incidence of lameness in cows and improving the economic benefits of the pasture, and meets the needs of large-scale construction of the pasture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Orchard-Wide Visual Perception and Autonomous Operation of Fruit Picking Robots: A Review
    CHEN Mingyou, LUO Lufeng, LIU Wei, WEI Huiling, WANG Jinhai, LU Qinghua, LUO Shaoming
    Smart Agriculture    2024, 6 (5): 20-39.   DOI: 10.12133/j.smartag.SA202405022
    Abstract818)   HTML147)    PDF(pc) (4030KB)(3930)       Save

    [Significance] Fruit-picking robot stands as a crucial solution for achieving intelligent fruit harvesting. Significant progress has been made in developing foundational methods for picking robots, such as fruit recognition, orchard navigation, path planning for picking, and robotic arm control, the practical implementation of a seamless picking system that integrates sensing, movement, and picking capabilities still encounters substantial technical hurdles. In contrast to current picking systems, the next generation of fruit-picking robots aims to replicate the autonomous skills exhibited by human fruit pickers. This involves effectively performing ongoing tasks of perception, movement, and picking without human intervention. To tackle this challenge, this review delves into the latest research methodologies and real-world applications in this field, critically assesses the strengths and limitations of existing methods and categorizes the essential components of continuous operation into three sub-modules: local target recognition, global mapping, and operation planning. [Progress] Initially, the review explores methods for recognizing nearby fruit and obstacle targets. These methods encompass four main approaches: low-level feature fusion, high-level feature learning, RGB-D information fusion, and multi-view information fusion, respectively. Each of these approaches incorporates advanced algorithms and sensor technologies for cluttered orchard environments. For example, low-level feature fusion utilizes basic attributes such as color, shapes and texture to distinguish fruits from backgrounds, while high-level feature learning employs more complex models like convolutional neural networks to interpret the contextual relationships within the data. RGB-D information fusion brings depth perception into the mix, allowing robots to gauge the distance to each fruit accurately. Multi-view information fusion tackles the issue of occlusions by combining data from multiple cameras and sensors around the robot, providing a more comprehensive view of the environment and enabling more reliable sensing. Subsequently, the review shifts focus to orchard mapping and scene comprehension on a broader scale. It points out that current mapping methods, while effective, still struggle with dynamic changes in the orchard, such as variations of fruits and light conditions. Improved adaptation techniques, possibly through machine learning models that can learn and adjust to different environmental conditions, are suggested as a way forward. Building upon the foundation of local and global perception, the review investigates strategies for planning and controlling autonomous behaviors. This includes not only the latest advancements in devising movement paths for robot mobility but also adaptive strategies that allow robots to react to unexpected obstacles or changes within the whole environment. Enhanced strategies for effective fruit picking using the Eye-in-Hand system involve the development of more dexterous robotic hands and improved algorithms for precisely predicting the optimal picking point of each fruit. The review also identifies a crucial need for further advancements in the dynamic behavior and autonomy of these technologies, emphasizing the importance of continuous learning and adaptive control systems to improve operational efficiency in diverse orchard environments. [Conclusions and Prospects] The review underscores the critical importance of coordinating perception, movement, and picking modules to facilitate the transition from a basic functional prototype to a practical machine. Moreover, it emphasizes the necessity of enhancing the robustness and stability of core algorithms governing perception, planning, and control, while ensuring their seamless coordination which is a central challenge that emerges. Additionally, the review raises unresolved questions regarding the application of picking robots and outlines future trends, include deeper integration of stereo vision and deep learning, enhanced global vision sampling, and the establishment of standardized evaluation criteria for overall operational performance. The paper can provide references for the eventual development of robust, autonomous, and commercially viable picking robots in the future.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Cow Hoof Slippage Detecting Method Based on Enhanced DeepLabCut Model
    NIAN Yue, ZHAO Kaixuan, JI Jiangtao
    Smart Agriculture    2024, 6 (5): 153-163.   DOI: 10.12133/j.smartag.SA202406014
    Abstract808)   HTML26)    PDF(pc) (1765KB)(644)       Save

    [Objective] The phenomenon of hoof slipping occurs during the walking process of cows, which indicates the deterioration of the farming environment and a decline in the cows' locomotor function. Slippery grounds can lead to injuries in cows, resulting in unnecessary economic losses for farmers. To achieve automatically recognizing and detecting slippery hoof postures during walking, the study focuses on the localization and analysis of key body points of cows based on deep learning methods. Motion curves of the key body points were analyzed, and features were extracted. The effectiveness of the extracted features was verified using a decision tree classification algorithm, with the aim of achieving automatic detection of slippery hoof postures in cows. [Method] An improved localization method for the key body points of cows, specifically the head and four hooves, was proposed based on the DeepLabCut model. Ten networks, including ResNet series, MobileNet-V2 series, and EfficientNet series, were selected to respectively replace the backbone network structure of DeepLabCut for model training. The root mean square error(RMSE), model size, FPS, and other indicators were chosen, and after comprehensive consideration, the optimal backbone network structure was selected as the pre-improved network. A network structure that fused the convolutional block attention module (CBAM) attention mechanism with ResNet-50 was proposed. A lightweight attention module, CBAM, was introduced to improve the ResNet-50 network structure. To enhance the model's generalization ability and robustness, the CBAM attention mechanism was embedded into the first convolution layer and the last convolution layer of the ResNet-50 network structure. Videos of cows with slippery hooves walking in profile were predicted for key body points using the improved DeepLabCut model, and the obtained key point coordinates were used to plot the motion curves of the cows' key body points. Based on the motion curves of the cows' key body points, the feature parameter Feature1 for detecting slippery hooves was extracted, which represented the local peak values of the derivative of the motion curves of the cows' four hooves. The feature parameter Feature2 for predicting slippery hoof distances was extracted, specifically the minimum local peak points of the derivative curve of the hooves, along with the local minimum points to the left and right of these peaks. The effectiveness of the extracted Feature1 feature parameters was verified using a decision tree classification model. Slippery hoof feature parameters Feature1 for each hoof were extracted, and the standard deviation of Feature1 was calculated for each hoof. Ultimately, a set of four standard deviations for each cow was extracted as input parameters for the classification model. The classification performance was evaluated using four common objective metrics, including accuracy, precision, recall, and F1-Score. The prediction accuracy for slippery hoof distances was assessed using RMSE as the evaluation metric. [Results and Discussion] After all ten models reached convergence, the loss values ranked from smallest to largest were found in the EfficientNet series, ResNet series, and MobileNet-V2 series, respectively. Among them, ResNet-50 exhibited the best localization accuracy in both the training set and validation set, with RMSE values of only 2.69 pixels and 3.31 pixels, respectively. The MobileNet series had the fastest inference speed, reaching 48 f/s, while the inference speeds of the ResNet series and MobileNet series were comparable, with ResNet series performing slightly better than MobileNet series. Considering the above factors, ResNet-50 was ultimately selected as the backbone network for further improvements on DeepLabCut. Compared to the original ResNet-50 network, the ResNet-50 network improved by integrating the CBAM module showed a significant enhancement in localization accuracy. The accuracy of the improved network increased by 3.7% in the training set and by 9.7% in the validation set. The RMSE between the predicted body key points and manually labeled points was only 2.99 pixels, with localization results for the right hind hoof, right front hoof, left hind hoof, left front hoof, and head improved by 12.1%, 44.9%, 0.04%, 48.2%, and 39.7%, respectively. To validate the advancement of the improved model, a comparison was made with the mainstream key point localization model, YOLOv8s-pose, which showed that the RMSE was reduced by 1.06 pixels compared to YOLOv8s-pose. This indicated that the ResNet-50 network integrated with the CBAM attention mechanism possessed superior localization accuracy. In the verification of the cow slippery hoof detection classification model, a 10-fold cross-validation was conducted to evaluate the performance of the cow slippery hoof classification model, resulting in average values of accuracy, precision, recall, and F1-Score at 90.42%, 0.943, 0.949, and 0.941, respectively. The error in the calculated slippery hoof distance of the cows, using the slippery hoof distance feature parameter Feature2, compared to the manually calibrated slippery hoof distance was found to be 1.363 pixels. [Conclusion] The ResNet-50 network model improved by integrating the CBAM module showed a high accuracy in the localization of key body points of cows. The cow slippery hoof judgment model and the cow slippery hoof distance prediction model, based on the extracted feature parameters for slippery hoof judgment and slippery hoof distance detection, both exhibited small errors when compared to manual detection results. This indicated that the proposed enhanced deeplabcut model obtained good accuracy and could provide technical support for the automatic detection of slippery hooves in cows.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Development Trend of Mountainous Tractor Leveling and Anti-Rollover System
    MU Xiaodong, YANG Fuzeng, DUAN Luojia, LIU Zhijie, SONG Zhuoying, LI Zonglin, GUAN Shouqing
    Smart Agriculture    2024, 6 (3): 1-16.   DOI: 10.12133/j.smartag.SA202312015
    Abstract795)   HTML100)    PDF(pc) (2448KB)(3655)       Save

    [Significance] The mechanization, automation and intelligentization of agricultural equipment are key factors to improve operation efficiency, free up labor force and promote the sustainable development of agriculture. It is also the hot spot of research and development of agricultural machinery industry in the future. In China, hills and mountains serves as vital production bases for agricultural products, accounting for about 70% of the country's land area. In addition, these regions face various environmental factors such as steep slopes, narrow road, small plots, complex terrain and landforms, as well as harsh working environment. Moreover, there is a lack of reliable agricultural machinery support across various production stages, along with a shortage of theoretical frameworks to guide the research and development of agricultural machinery tailored to hilly and mountainous locales. [Progress] This article focuses on the research advances of tractor leveling and anti-overturning systems in hilly and mountainous areas, including tractor body, cab and seat leveling technology, tractor rear suspension and implement leveling slope adaptive technology, and research progress on tractor anti-overturning protection devices and warning technology. The vehicle body leveling mechanism can be roughly divided into five types based on its different working modes: parallel four bar, center of gravity adjustable, hydraulic differential high, folding and twisting waist, and omnidirectional leveling. These mechanisms aim to address the issue of vehicle tilting and easy overturning when traversing or working on sloping or rugged roads. By keeping the vehicle body posture horizontal or adjusting the center of gravity within a stable range, the overall driving safety of the vehicle can be improved to ensure the accuracy of operation. Leveling the driver's cab and seats can mitigate the lateral bumps experienced by the driver during rough or sloping operations, reducing driver fatigue and minimizing strain on the lumbar and cervical spine, thereby enhancing driving comfort. The adaptive technology of tractor rear suspension and implement leveling on slopes can ensure that the tractor maintains consistent horizontal contact with the ground in hilly and mountainous areas, avoiding changes in the posture of the suspended implement with the swing of the body or the driving path, which may affect the operation effect. The tractor rollover protection device and warning technology have garnered significant attention in recent years. Prioritizing driver safety, rollover warning system can alert the driver in advance of the dangerous state of the tractor, automatically adjust the vehicle before rollover, or automatically open the rollover protection device when it is about to rollover, and timely send accident reports to emergency contacts, thereby ensuring the safety of the driver to the greatest extent possible. [Conclusions and Prospects] The future development directions of hill and mountain tractor leveling, anti-overturning early warning, unmanned, automatic technology were looked forward: Structure optimization, high sensitivity, good stability of mountain tractor leveling system research; Study on copying system of agricultural machinery with good slope adaptability; Research on anti-rollover early warning technology of environment perception and automatic interference; Research on precision navigation technology, intelligent monitoring technology and remote scheduling and management technology of agricultural machinery; Theoretical study on longitudinal stability of sloping land. This review could provide reference for the research and development of high reliability and high safety mountain tractor in line with the complex working environment in hill and mountain areas.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Method for Calculating Semantic Similarity of Short Agricultural Texts Based on Transfer Learning
    JIN Ning, GUO Yufeng, HAN Xiaodong, MIAO Yisheng, WU Huarui
    Smart Agriculture    2025, 7 (1): 33-43.   DOI: 10.12133/j.smartag.SA202410026
    Abstract751)   HTML10)    PDF(pc) (1239KB)(139)       Save

    [Objective] Intelligent services of agricultural knowledge have emerged as a current hot research domain, serving as a significant support for the construction of smart agriculture. The platform "China Agricultural Technology Extension" provides users with efficient and convenient agricultural information consultation services via mobile terminals, and has accumulated a vast amount of Q&A data. These data are characterized by a huge volume of information, rapid update and iteration, and a high degree of redundancy, resulting in the platform encountering issues such as frequent repetitive questions, low timeliness of problem responses, and inaccurate information retrieval. There is an urgent requirement for a high-quality text semantic similarity calculation approach to confront these challenges and effectively enhance the information service efficiency and intelligent level of the platform. In view of the problems of incomplete feature extraction and lack of short agro-text annotation data sets in existing text semantic similarity calculation models, a semantic similarity calculation model for short agro-text, namely CWPT-SBERT, based on transfer learning and BERT pre-training model, was proposed. [Methods] CWPT-SBERT was based on Siamese architecture with identical left and right sides and shared parameters, which had the advantages of low structural complexity and high training efficiency. This network architecture effectively reduced the consumption of computational resources by sharing parameters and ensures that input texts were compared in the same feature space. CWPT-SBERT consisted of four main parts: Semantic enhancement layer, embedding layer, pooling layer, and similarity measurement layer. The CWPT method based on the word segmentation unit was proposed in the semantic enhancement layer to further divide Chinese characters into more fine-grained sub-units maximizing the semantic features in short Chinese text and effectively enhancing the model's understanding of complex Chinese vocabulary and character structures. In the embedding layer, a transfer learning strategy was used to extract features from agricultural short texts based on SBERT. It captured the semantic features of Chinese text in the general domain, and then generated a more suitable semantic feature vector representation after fine-tuning. Transfer learning methods to train models on large-scale general-purposed domain annotation datasets solve the problem of limited short agro-text annotation datasets and high semantic sparsity. The pooling layer used the average pooling strategy to map the high-dimensional semantic vector of Chinese short text to a low-dimensional vector space. The similarity measurement layer used the cosine similarity calculation method to measure the similarity between the semantic feature vector representations of the two output short texts, and the computed similarity degree was finally input into the loss function to guide model training, optimize model parameters, and improve the accuracy of similarity calculation. [Results and Discussions] For the task of calculating semantic similarity in agricultural short texts, on a dataset containing 19 968 pairs of short ago-texts, the CWPT-SBERT model achieved an accuracy rate of 97.18% and 96.93%, a recall rate of 97.14%, and an F1-Score value of 97.04%, which are higher than 12 models such as TextCNN_Attention, MaLSTM and SBERT. By analyzing the Pearson and Spearman coefficients of CWPT-SBERT, SBERT, SALBERT and SRoBERTa trained on short agro-text datasets, it could be observed that the initial training value of the CWPT-SBERT model was significantly higher than that of the comparison models and was close to the highest value of the comparison models. Moreover, it exhibited a smooth growth trend during the training process, indicating that CWPT-SBERT had strong correlation, robustness, and generalization ability from the initial state. During the training process, it could not only learn the features in the training data but also effectively apply these features to new domain data. Additionally, for ALBERT, RoBERTa and BERT models, fine-tuning training was conducted on short agro-text datasets, and optimization was performed by utilizing the morphological structure features to enrich text semantic feature expression. Through ablation experiments, it was evident that both optimization strategies could effectively enhance the performance of the models. By analyzing the attention weight heatmap of Chinese character morphological structure, the importance of Chinese character radicals in representing Chinese character attributes was highlighted, enhancing the semantic representation of Chinese characters in vector space. There was also complex correlation within the morphological structure of Chinese characters. [Conclusions] CWPT-SBERT uses transfer learning methods to solve the problem of limited short agro-text annotation datasets and high semantic sparsity. By leveraging the Chinese-oriented word segmentation method CWPT to break down Chinese characters, the semantic representation of word vectors is enhanced, and the semantic feature expression of short texts is enriched. CWPT-SBERT model has high accuracy of semantic similarity on small-scale short agro-text and obvious performance advantages, which provides an effective technical reference for semantic intelligence matching.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Advances and Prospects on Rapid Acquisition Technology of Farmland Soil Physical and Chemical Parameters
    QI Jiangtao, CHENG Panting, GAO Fangfang, GUO Li, ZHANG Ruirui
    Smart Agriculture    2024, 6 (3): 17-33.   DOI: 10.12133/j.smartag.SA202404003
    Abstract748)   HTML114)    PDF(pc) (1524KB)(5188)       Save

    [Significance] Soil stands as the fundamental pillar of agricultural production, with its quality being intrinsically linked to the efficiency and sustainability of farming practices. Historically, the intensive cultivation and soil erosion have led to a marked deterioration in some arable lands, characterized by a sharp decrease in soil organic matter, diminished fertility, and a decline in soil's structural integrity and ecological functions. In the strategic framework of safeguarding national food security and advancing the frontiers of smart and precision agriculture, the march towards agricultural modernization continues apace, intensifying the imperative for meticulous soil quality management. Consequently, there is an urgent need for the rrapid acquisition of soil's physical and chemical parameters. Interdisciplinary scholars have delved into soil monitoring research, achieving notable advancements that promise to revolutionize the way we understand and manage soil resource. [Progress] Utilizing the the Web of Science platform, a comprehensive literature search was conducted on the topic of "soil," further refined with supplementary keywords such as "electrochemistry", "spectroscopy", "electromagnetic", "ground-penetrating radar", and "satellite". The resulting literature was screened, synthesized, and imported into the CiteSpace visualization tool. A keyword emergence map was yielded, which delineates the trajectory of research in soil physical and chemical parameter detection technology. Analysis of the keyword emergence map reveals a paradigm shift in the acquisition of soil physical and chemical parameters, transitioning from conventional indoor chemical and spectrometry analyses to outdoor, real-time detection methods. Notably, soil sensors integrated into drones and satellites have garnered considerable interest. Additionally, emerging monitoring technologies, including biosensing and terahertz spectroscopy, have made their mark in recent years. Drawing from this analysis, the prevailing technologies for soil physical and chemical parameter information acquisition in agricultural fields have been categorized and summarized. These include: 1. Rapid Laboratory Testing Techniques: Primarily hinged on electrochemical and spectrometry analysis, these methods offer the dual benefits of time and resource efficiency alongside high precision; 2. Rapid Near-Ground Sensing Techniques: Leveraging electromagnetic induction, ground-penetrating radar, and various spectral sensors (multispectral, hyperspectral, and thermal infrared), these techniques are characterized by their high detection accuracy and swift operation. 3. Satellite Remote Sensing Techniques: Employing direct inversion, indirect inversion, and combined analysis methods, these approaches are prized for their efficiency and extensive coverage. 4. Innovative Rapid Acquisition Technologies: Stemming from interdisciplinary research, these include biosensing, environmental magnetism, terahertz spectroscopy, and gamma spectroscopy, each offering novel avenues for soil parameter detection. An in-depth examination and synthesis of the principles, applications, merits, and limitations of each technology have been provided. Moreover, a forward-looking perspective on the future trajectory of soil physical and chemical parameter acquisition technology has been offered, taking into account current research trends and hotspots. [Conclusions and Prospects] Current advancements in the technology for rapaid acquiring soil physical and chemical parameters in agricultural fields have been commendable, yet certain challenges persist. The development of near-ground monitoring sensors is constrained by cost, and their reliability, adaptability, and specialization require enhancement to effectively contend with the intricate and varied conditions of farmland environments. Additionally, remote sensing inversion techniques are confronted with existing limitations in data acquisition, processing, and application. To further develop the soil physical and chemical parameter acquisition technology and foster the evolution of smart agriculture, future research could beneficially delve into the following four areas: Designing portable, intelligent, and cost-effective near-ground soil information acquisition systems and equipment to facilitate rapid on-site soil information detection; Enhancing the performance of low-altitude soil information acquisition platforms and refine the methods for data interpretation to ensure more accurate insights; Integrating multifactorial considerations to construct robust satellite remote sensing inversion models, leveraging diverse and open cloud computing platforms for in-depth data analysis and mining; Engaging in thorough research on the fusion of multi-source data in the acquisition of soil physical and chemical parameter information, developing soil information sensing algorithms and models with strong generalizability and high reliability to achieve rapaid, precise, and intelligent acquisition of soil parameters.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospects of Key Technologies for Rice Smart Unmanned Farms
    YU Fenghua, XU Tongyu, GUO Zhonghui, BAI Juchi, XIANG Shuang, GUO Sien, JIN Zhongyu, LI Shilong, WANG Shikuan, LIU Meihan, HUI Yinxuan
    Smart Agriculture    2024, 6 (6): 1-22.   DOI: 10.12133/j.smartag.SA202410018
    Abstract703)   HTML124)    PDF(pc) (3047KB)(1634)       Save

    [Significance] Rice smart unmanned farm is the core component of smart agriculture, and it is a key path to realize the modernization of rice production and promote the high-quality development of agriculture. Leveraging advanced information technologies such as the Internet of Things (IoT) and artificial intelligence (AI), these farms enable deep integration of data-driven decision making and intelligent machines. This integration creates an unmanned production system that covers the entire process from planting and managing rice crops to harvesting, greatly improving the efficiency and precision of rice cultivation. [Progress] This paper systematically sorted out the key technologies of rice smart unmanned farms in the three main links of pre-production, production and post-production, and the key technologies of pre-production mainly include the construction of high-standard farmland, unmanned nursery, land leveling, and soil nutrient testing. The construction of high-standard farmland is the foundation of the physical environment of the smart unmanned farms of rice, which provides perfect operating environment for the operation of modernized smart farm machinery through the reasonable layout of the field roads, good drainage and irrigation systems, and the scientific planting structure. Agricultural machine operation provides a perfect operating environment. The technical level of unmanned nursery directly determines the quality of rice cultivation and harvesting in the later stage, and a variety of rice seeding machines and nursery plate setting machines have been put into use. Land leveling technology can improve the growing environment of rice and increase the land utilization rate, and the current land leveling technology through digital sensing and path planning technology, which improves the operational efficiency and reduces the production cost at the same time. Soil nutrient detection technology is mainly detected by electrochemical analysis and spectral analysis, but both methods have their advantages and disadvantages, how to integrate the two methods to achieve an all-round detection of soil nutrient content is the main direction of future research. The key technologies in production mainly include rice dry direct seeding, automated transplanting, precise variable fertilization, intelligent irrigation, field weed management, and disease diagnosis. Among them, the rice dry direct seeding technology requires the planter to have high precision and stability to ensure reasonable seeding depth and density. Automated rice transplanting technology mainly includes three ways: root washing seedling machine transplanting, blanket seedling machine transplanting, and potting blanket seedling machine transplanting; at present, the incidence of problems in the automated transplanting process should be further reduced, and the quality and efficiency of rice machine transplanting should be improved. Precision variable fertilization technology is mainly composed of three key technologies: information perception, prescription decision-making and precise operation, but there are still fewer cases of unmanned farms combining the three technologies, and in the future, the main research should be on the method of constructing the whole process operation system of variable fertilization. The smart irrigation system is based on the water demand of the whole life cycle of rice to realize adaptive irrigation control, and the current smart irrigation technology can automatically adjust the irrigation strategy through real-time monitoring of soil, climate and crop growth conditions to further improve irrigation efficiency and agricultural production benefits. The field weed management and disease diagnosis technology mainly recognizes rice weeds as well as diseases through deep learning and other methods, and combines them with precision application technology for prevention and intervention. Post-production key technologies mainly include rice yield estimation, unmanned harvesting, rice storage and processing quality testing. Rice yield estimation technology is mainly used to predict yield by combining multi-source data and algorithms, but there are still problems such as the difficulty of integrating multi-source data, which requires further research. In terms of unmanned aircraft harvesting technology, China's rice combine harvester market has tended to stabilize, and the safety of the harvester's autopilot should be further improved in the future. Rice storage and processing quality detection technology mainly utilizes spectral technology and machine vision technology to detect spectra and images, and future research can combine deep learning and multimodal fusion technology to improve the machine vision system's ability and adaptability to recognize the appearance characteristics of rice. [Conclusions and Prospects] This paper reviews the researches of the construction of intelligent unmanned rice farms at home and abroad in recent years, summarizes the main difficulties faced by the key technologies of unmanned farms in practical applications, analyzes the challenges encountered in the construction of smart unmanned farms, summarizes the roles and responsibilities of the government, enterprises, scientific research institutions, cooperatives and other subjects in promoting the construction of intelligent unmanned rice farms, and puts forward relevant suggestions. It provides certain support and development ideas for the construction of intelligent unmanned rice farms in China.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Chinese Kiwifruit Text Named Entity Recognition Method Based on Dual-Dimensional Information and Pruning
    QI Zijun, NIU Dangdang, WU Huarui, ZHANG Lilin, WANG Lunfeng, ZHANG Hongming
    Smart Agriculture    2025, 7 (1): 44-56.   DOI: 10.12133/j.smartag.SA202410022
    Abstract612)   HTML5)    PDF(pc) (1225KB)(104)       Save

    [Objective] Chinese kiwifruit texts exhibit unique dual-dimensional characteristics. The cross-paragraph dependency is complex semantic structure, whitch makes it challenging to capture the full contextual relationships of entities within a single paragraph, necessitating models capable of robust cross-paragraph semantic extraction to comprehend entity linkages at a global level. However, most existing models rely heavily on local contextual information and struggle to process long-distance dependencies, thereby reducing recognition accuracy. Furthermore, Chinese kiwifruit texts often contain highly nested entities. This nesting and combination increase the complexity of grammatical and semantic relationships, making entity recognition more difficult. To address these challenges, a novel named entity recognition (NER) method, KIWI-Coord-Prune(kiwifruit-CoordKIWINER-PruneBi-LSTM) was proposed in this research, which incorporated dual-dimensional information processing and pruning techniques to improve recognition accuracy. [Methods] The proposed KIWI-Coord-Prune model consisted of a character embedding layer, a CoordKIWINER layer, a PruneBi-LSTM layer, a self-attention mechanism, and a CRF decoding layer, enabling effective entity recognition after processing input character vectors. The CoordKIWINER and PruneBi-LSTM modules were specifically designed to handle the dual-dimensional features in Chinese kiwifruit texts. The CoordKIWINER module applied adaptive average pooling in two directions on the input feature maps and utilized convolution operations to separate the extracted features into vertical and horizontal branches. The horizontal and vertical features were then independently extracted using the Criss-Cross Attention (CCNet) mechanism and Coordinate Attention (CoordAtt) mechanism, respectively. This module significantly enhanced the model's ability to capture cross-paragraph relationships and nested entity structures, thereby generating enriched character vectors containing more contextual information, which improved the overall representation capability and robustness of the model. The PruneBi-LSTM module was built upon the enhanced dual-dimensional vector representations and introduced a pruning strategy into Bi-LSTM to effectively reduce redundant parameters associated with background descriptions and irrelevant terms. This pruning mechanism not only enhanced computational efficiency while maintaining the dynamic sequence modeling capability of Bi-LSTM but also improved inference speed. Additionally, a dynamic feature extraction strategy was employed to reduce the computational complexity of vector sequences and further strengthen the learning capacity for key features, leading to improved recognition of complex entities in kiwifruit texts. Furthermore, the pruned weight matrices become sparser, significantly reducing memory consumption. This made the model more efficient in handling large-scale agricultural text-processing tasks, minimizing redundant information while achieving higher inference and training efficiency with fewer computational resources. [Results and Discussions] Experiments were conducted on the self-built KIWIPRO dataset and four public datasets: People's Daily, ClueNER, Boson, and ResumeNER. The proposed model was compared with five advanced NER models: LSTM, Bi-LSTM, LR-CNN, Softlexicon-LSTM, and KIWINER. The experimental results showed that KIWI-Coord-Prune achieved F1-Scores of 89.55%, 91.02%, 83.50%, 83.49%, and 95.81%, respectively, outperforming all baseline models. Furthermore, controlled variable experiments were conducted to compare and ablate the CoordKIWINER and PruneBi-LSTM modules across the five datasets, confirming their effectiveness and necessity. Additionally, the impact of different design choices was explored for the CoordKIWINER module, including direct fusion, optimized attention mechanism fusion, and network structure adjustment residual optimization. The experimental results demonstrated that the optimized attention mechanism fusion method yielded the best performance, which was ultimately adopted in the final model. These findings highlight the significance of properly designing attention mechanisms to extract dual-dimensional features for NER tasks. Compared to existing methods, the KIWI-Coord-Prune model effectively addressed the issue of underutilized dual-dimensional information in Chinese kiwifruit texts. It significantly improved entity recognition performance for both overall text structures and individual entity categories. Furthermore, the model exhibited a degree of generalization capability, making it applicable to downstream tasks such as knowledge graph construction and question-answering systems. [Conclusions] This study presents an novel NER approach for Chinese kiwifruit texts, which integrating dual-dimensional information extraction and pruning techniques to overcome challenges related to cross-paragraph dependencies and nested entity structures. The findings offer valuable insights for researchers working on domain-specific NER and contribute to the advancement of agriculture-focused natural language processing applications. However, two key limitations remain: 1) The balance between domain-specific optimization and cross-domain generalization requires further investigation, as the model's adaptability to non-agricultural texts has yet to be empirically validated; 2) the multilingual applicability of the model is currently limited, necessitating further expansion to accommodate multilingual scenarios. Future research should focus on two key directions: 1) Enhancing domain robustness and cross-lingual adaptability by incorporating diverse textual datasets and leveraging pre-trained multilingual models to improve generalization, and 2) Validating the model's performance in multilingual environments through transfer learning while refining linguistic adaptation strategies to further optimize recognition accuracy.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Localization Method for Agricultural Robots Based on Fusion of LiDAR and IMU
    LIU Yang, JI Jie, PAN Deng, ZHAO Lijun, LI Mingsheng
    Smart Agriculture    2024, 6 (3): 94-106.   DOI: 10.12133/j.smartag.SA202401009
    Abstract587)   HTML52)    PDF(pc) (3329KB)(1071)       Save

    [Objective] High-precision localization technology serves as the crucial foundation in enabling the autonomous navigation operations of intelligent agricultural robots. However, the traditional global navigation satellite system (GNSS) localization method faces numerous limitations, such as tree shadow, electromagnetic interference, and other factors in the agricultural environment brings challenges to the accuracy and reliability of localization technology. To address the deficiencies and achieve precise localization of agricultural robots independent of GNSS, a localization method was proposed based on the fusion of three-dimensional light detection and ranging (LiDAR) data and inertial measurement unit (IMU) information to enhance localization accuracy and reliability. [Methods] LiDAR was used to obtain point cloud data in the agricultural environment and realize self-localization via point cloud matching. By integrating real-time motion parameter measurements from the IMU with LiDAR data, a high-precision localization solution for agricultural robots was achieved through a specific fusion algorithm. Firstly, the LiDAR-obtained point cloud data was preprocessed and the depth map was used to save the data. This approach could reduce the dimensionality of the original LiDAR point cloud, and eliminate the disorder of the original LiDAR point cloud arrangement, facilitating traversal and clustering through graph search. Given the presence of numerous distinct crops like trees in the agricultural environment, an angle-based clustering method was adopted. Specific angle-based clustering criteria were set to group the point cloud data, leading to the segmentation of different clusters of points, and obvious crops in the agricultural environment was effectively perceived. Furthermore, to improve the accuracy and stability of positioning, an improved three-dimensional normal distribution transform (3D-NDT) localization algorithm was proposed. This algorithm operated by matching the LiDAR-scanned point cloud data in real time with the pre-existing down sampled point cloud map to achieve real-time localization. Considering that direct down sampling of LiDAR point clouds in the agricultural environment could result in the loss of crucial environmental data, a point cloud clustering operation was used in place of down sampling operation, thereby improving matching accuracy and positioning precision. Secondly, to address potential constraints and shortcomings of using a single sensor for robot localization, a multi-sensor information fusion strategy was deployed to improve the localization accuracy. Specifically, the extended Kalman filter algorithm (EKF) was chosen to fuse the localization data from LiDAR point cloud and the IMU odometer information. The IMU provided essential motion parameters such as acceleration and angular velocity of the agricultural robot, and by combining with the LiDAR-derived localization information, the localization of the agricultural robot could be more accurately estimated. This fusion approach maximized the advantages of different sensors, compensated for their individual limitations, and improved the overall localization accuracy of the agricultural robot. [Results and Discussions] A series of experimental results in the Gazebo simulation environment of the robot operating system (ROS) and real operation scenarios showed that the fusion localization method proposed had significant advantages. In the simulation environment, the average localization errors of the proposed multi-sensor data fusion localization method were 1.7 and 1.8 cm, respectively, while in the experimental scenario, these errors were 3.3 and 3.3 cm, respectively, which were significantly better than the traditional 3D-NDT localization algorithm. These findings showed that the localization method proposed in this study could achieve high-precision localization in the complex agricultural environment, and provide reliable localization assistance for the autonomous functioning of agricultural robots. [Conclusions] The proposed localization method based on the fusion of LiDAR data and IMU information provided a novel localization solution for the autonomous operation of agricultural robots in areas with limited GNSS reception. Through the comprehensive utilization of multi-sensor information and adopting advanced data processing and fusion algorithms, the localization accuracy of agricultural robots could be significantly improved, which could provide a new reference for the intelligence and automation of agricultural production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospect of Multi-robot Collaborative SLAM in Complex Agricultural Scenarios
    MA Nan, CAO Shanshan, BAI Tao, KONG Fantao, SUN Wei
    Smart Agriculture    2024, 6 (6): 23-43.   DOI: 10.12133/j.smartag.SA202406005
    Abstract553)   HTML73)    PDF(pc) (2300KB)(4493)       Save

    [Significance] The rapid development of artificial intelligence and automation has greatly expanded the scope of agricultural automation, with applications such as precision farming using unmanned machinery, robotic grazing in outdoor environments, and automated harvesting by orchard-picking robots. Collaborative operations among multiple agricultural robots enhance production efficiency and reduce labor costs, driving the development of smart agriculture. Multi-robot simultaneous localization and mapping (SLAM) plays a pivotal role by ensuring accurate mapping and localization, which are essential for the effective management of unmanned farms. Compared to single-robot SLAM, multi-robot systems offer several advantages, including higher localization accuracy, larger sensing ranges, faster response times, and improved real-time performance. These capabilities are particularly valuable for completing complex tasks efficiently. However, deploying multi-robot SLAM in agricultural settings presents significant challenges. Dynamic environmental factors, such as crop growth, changing weather patterns, and livestock movement, increase system uncertainty. Additionally, agricultural terrains vary from open fields to irregular greenhouses, requiring robots to adjust their localization and path-planning strategies based on environmental conditions. Communication constraints, such as unstable signals or limited transmission range, further complicate coordination between robots. These combined challenges make it difficult to implement multi-robot SLAM effectively in agricultural environments. To unlock the full potential of multi-robot SLAM in agriculture, it is essential to develop optimized solutions that address the specific technical demands of these scenarios. [Progress] Existing review studies on multi-robot SLAM mainly focus on a general technological perspective, summarizing trends in the development of multi-robot SLAM, the advantages and limitations of algorithms, universally applicable conditions, and core issues of key technologies. However, there is a lack of analysis specifically addressing multi-robot SLAM under the characteristics of complex agricultural scenarios. This study focuses on the main features and applications of multi-robot SLAM in complex agricultural scenarios. The study analyzes the advantages and limitations of multi-robot SLAM, as well as its applicability and application scenarios in agriculture, focusing on four key components: multi-sensor data fusion, collaborative localization, collaborative map building, and loopback detection. From the perspective of collaborative operations in multi-robot SLAM, the study outlines the classification of SLAM frameworks, including three main collaborative types: centralized, distributed, and hybrid. Based on this, the study summarizes the advantages and limitations of mainstream multi-robot SLAM frameworks, along with typical scenarios in robotic agricultural operations where they are applicable. Additionally, it discusses key issues faced by multi-robot SLAM in complex agricultural scenarios, such as low accuracy in mapping and localization during multi-sensor fusion, restricted communication environments during multi-robot collaborative operations, and low accuracy in relative pose estimation between robots. [Conclusions and Prospects] To enhance the applicability and efficiency of multi-robot SLAM in complex agricultural scenarios, future research needs to focus on solving these critical technological issues. Firstly, the development of enhanced data fusion algorithms will facilitate improved integration of sensor information, leading to greater accuracy and robustness of the system. Secondly, the combination of deep learning and reinforcement learning techniques is expected to empower robots to better interpret environmental patterns, adapt to dynamic changes, and make more effective real-time decisions. Thirdly, large language models will enhance human-robot interaction by enabling natural language commands, improving collaborative operations. Finally, the integration of digital twin technology will support more intelligent path planning and decision-making processes, especially in unmanned farms and livestock management systems. The convergence of digital twin technology with SLAM is projected to yield innovative solutions for intelligent perception and is likely to play a transformative role in the realm of agricultural automation. This synergy is anticipated to revolutionize the approach to agricultural tasks, enhancing their efficiency and reducing the reliance on labor.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Rapid Detection Method for Wheat Seedling Leaf Number in Complex Field Scenarios Based on Improved YOLOv8
    HOU Yiting, RAO Yuan, SONG He, NIE Zhenjun, WANG Tan, HE Haoxu
    Smart Agriculture    2024, 6 (4): 128-137.   DOI: 10.12133/j.smartag.SA202403019
    Abstract512)   HTML97)    PDF(pc) (2913KB)(835)       Save

    [Objective] The enumeration of wheat leaves is an essential indicator for evaluating the vegetative state of wheat and predicting its yield potential. Currently, the process of wheat leaf counting in field settings is predominantly manual, characterized by being both time-consuming and labor-intensive. Despite advancements, the efficiency and accuracy of existing automated detection and counting methodologies have yet to satisfy the stringent demands of practical agricultural applications. This study aims to develop a method for the rapid quantification of wheat leaves to refine the precision of wheat leaf tip detection. [Methods] To enhance the accuracy of wheat leaf detection, firstly, an image dataset of wheat leaves across various developmental stages—seedling, tillering, and overwintering—under two distinct lighting conditions and using visible light images sourced from both mobile devices and field camera equipmen, was constructed. Considering the robust feature extraction and multi-scale feature fusion capabilities of YOLOv8 network, the foundational architecture of the proposed model was based on the YOLOv8, to which a coordinate attention mechanism has been integrated. To expedite the model's convergence, the loss functions were optimized. Furthermore, a dedicated small object detection layer was introduced to refine the recognition of wheat leaf tips, which were typically difficult for conventional models to discern due to their small size and resemblance to background elements. This deep learning network was named as YOLOv8-CSD, tailored for the recognition of small targets such as wheat leaf tips, ascertains the leaf count by detecting the number of leaf tips present within the image. A comparative analysis was conducted on the YOLOv8-CSD model in comparison with the original YOLOv8 and six other prominent network architectures, including Faster R-CNN, Mask R-CNN, YOLOv7, and SSD, within a uniform training framework, to evaluate the model's effectiveness. In parallel, the performance of both the original and YOLOv8-CSD models was assessed under challenging conditions, such as the presence of weeds, occlusions, and fluctuating lighting, to emulate complex real-world scenarios. Ultimately, the YOLOv8-CSD model was deployed for wheat leaf number detection in intricate field conditions to confirm its practical applicability and generalization potential. [Results and Discussions] The research presented a methodology that achieved a recognition precision of 91.6% and an mAP0.5 of 85.1% for wheat leaf tips, indicative of its robust detection capabilities. This method exceled in adaptability within complex field environments, featuring an autonomous adjustment mechanism for different lighting conditions, which significantly enhanced the model's robustness. The minimal rate of missed detections in wheat seedlings' leaf counting underscored the method's suitability for wheat leaf tip recognition in intricate field scenarios, consequently elevating the precision of wheat leaf number detection. The sophisticated algorithm embedded within this model had demonstrated a heightened capacity to discern and focus on the unique features of wheat leaf tips during the detection process. This capability was essential for overcoming challenges such as small target sizes, similar background textures, and the intricacies of feature extraction. The model's consistent performance across diverse conditions, including scenarios with weeds, occlusions, and fluctuating lighting, further substantiated its robustness and its readiness for real-world application. [Conclusions] This research offers a valuable reference for accurately detecting wheat leaf numbers in intricate field conditions, as well as robust technical support for the comprehensive and high-quality assessment of wheat growth.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    MSH-YOLOv8: Mushroom Small Object Detection Method with Scale Reconstruction and Fusion
    YE Dapeng, JING Jun, ZHANG Zhide, LI Huihuang, WU Haoyu, XIE Limin
    Smart Agriculture    2024, 6 (5): 139-152.   DOI: 10.12133/j.smartag.SA202404002
    Abstract504)   HTML71)    PDF(pc) (2660KB)(2294)       Save

    [Objective] Traditional object detection algorithms applied in the agricultural field, such as those used for crop growth monitoring and harvesting, often suffer from insufficient accuracy. This is particularly problematic for small crops like mushrooms, where recognition and detection are more challenging. The introduction of small object detection technology promises to address these issues, potentially enhancing the precision, efficiency, and economic benefits of agricultural production management. However, achieving high accuracy in small object detection has remained a significant challenge, especially when dealing with varying image sizes and target scales. Although the YOLO series models excel in speed and large object detection, they still have shortcomings in small object detection. To address the issue of maintaining high accuracy amid changes in image size and target scale, a novel detection model, Multi-Strategy Handling YOLOv8 (MSH-YOLOv8), was proposed. [Methods] The proposed MSH-YOLOv8 model builds upon YOLOv8 by incorporating several key enhancements aimed at improving sensitivity to small-scale targets and overall detection performance. Firstly, an additional detection head was added to increase the model's sensitivity to small objects. To address computational redundancy and improve feature extraction, the Swin Transformer detection structure was introduced into the input module of the head network, creating what was termed the "Swin Head (SH)". Moreover, the model integrated the C2f_Deformable convolutionv4 (C2f_DCNv4) structure, which included deformable convolutions, and the Swin Transformer encoder structure, termed "Swinstage", to reconstruct the YOLOv8 backbone network. This optimization enhanced feature propagation and extraction capabilities, increasing the network's ability to handle targets with significant scale variations. Additionally, the normalization-based attention module (NAM) was employed to improve performance without compromising detection speed or computational complexity. To further enhance training efficacy and convergence speed, the original loss function CIoU was replaced with wise-intersection over union (WIoU) Loss. Furthermore, experiments were conducted using mushrooms as the research subject on the open Fungi dataset. Approximately 200 images with resolution sizes around 600×800 were selected as the main research material, along with 50 images each with resolution sizes around 200×400 and 1 000×1 200 to ensure representativeness and generalization of image sizes. During the data augmentation phase, a generative adversarial network (GAN) was utilized for resolution reconstruction of low-resolution images, thereby preserving semantic quality as much as possible. In the post-processing phase, dynamic resolution training, multi-scale testing, soft non-maximum suppression (Soft-NMS), and weighted boxes fusion (WBF) were applied to enhance the model's small object detection capabilities under varying scales. [Results and Discussions] The improved MSH-YOLOv8 achieved an average precision at 50% (AP50) intersection over union of 98.49% and an AP@50-95 of 75.29%, with the small object detection metric APs reaching 39.73%. Compared to mainstream models like YOLOv8, these metrics showed improvements of 2.34%, 4.06% and 8.55%, respectively. When compared to the advanced TPH-YOLOv5 model, the improvements were 2.14%, 2.76% and 6.89%, respectively. The ensemble model, MSH-YOLOv8-ensemble, showed even more significant improvements, with AP50 and APs reaching 99.14% and 40.59%, respectively, an increase of 4.06% and 8.55% over YOLOv8. These results indicate the robustness and enhanced performance of the MSH-YOLOv8 model, particularly in detecting small objects under varying conditions. Further application of this methodology on the Alibaba Cloud Tianchi databases "Tomato Detection" and "Apple Detection" yielded MSH-YOLOv8-t and MSH-YOLOv8-a models (collectively referred to as MSH-YOLOv8). Visual comparison of detection results demonstrated that MSH-YOLOv8 significantly improved the recognition of dense and blurry small-scale tomatoes and apples. This indicated that the MSH-YOLOv8 method possesses strong cross-dataset generalization capability and effectively recognizes small-scale targets. In addition to quantitative improvements, qualitative assessments showed that the MSH-YOLOv8 model could handle complex scenarios involving occlusions, varying lighting conditions, and different growth stages of the crops. This demonstrates the practical applicability of the model in real-world agricultural settings, where such challenges are common. [Conclusions] The MSH-YOLOv8 improvement method proposed in this study effectively enhances the detection accuracy of small mushroom targets under varying image sizes and target scales. This approach leverages multiple strategies to optimize both the architecture and the training process, resulting in a robust model capable of high-precision small object detection. The methodology's application to other datasets, such as those for tomato and apple detection, further underscores its generalizability and potential for broader use in agricultural monitoring and management tasks.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Apple Leaf Disease Detection Algorithm Based on Improved YOLOv8
    LUO Youlu, PAN Yonghao, XIA Shunxing, TAO Youzhi
    Smart Agriculture    2024, 6 (5): 128-138.   DOI: 10.12133/j.smartag.SA202406012
    Abstract491)   HTML80)    PDF(pc) (1702KB)(1056)       Save

    [Objective] As one of China's most important agricultural products, apples hold a significant position in cultivation area and yield. However, during the growth process, apples are prone to various diseases that not only affect the quality of the fruit but also significantly reduce the yield, impacting farmers' economic benefits and the stability of market supply. To reduce the incidence of apple diseases and increase fruit yield, developing efficient and fast apple leaf disease detection technology is of great significance. An improved YOLOv8 algorithm was proposed to identify the leaf diseases that occurred during the growth of apples. [Methods] YOLOv8n model was selected to detect various leaf diseases such as brown rot, rust, apple scab, and sooty blotch that apples might encounter during growth. SPD-Conv was introduced to replace the original convolutional layers to retain fine-grained information and reduce model parameters and computational costs, thereby improving the accuracy of disease detection. The multi-scale dilated attention (MSDA) attention mechanism was added at appropriate positions in the Neck layer to enhance the model's feature representation capability, which allowed the model to learn the receptive field dynamically and adaptively focus on the most representative regions and features in the image, thereby enhancing the ability to extract disease-related features. Finally, inspired by the RepVGG architecture, the original detection head was optimized to achieve a separation of detection and inference architecture, which not only accelerated the model's inference speed but also enhanced feature learning capability. Additionally, a dataset of apple leaf diseases containing the aforementioned diseases was constructed, and experiments were conducted. [Results and Discussions] Compared to the original model, the improved model showed significant improvements in various performance metrics. The mAP50 and mAP50:95 achieved 88.2% and 37.0% respectively, which were 2.7% and 1.3% higher than the original model. In terms of precision and recall, the improved model increased to 83.1% and 80.2%, respectively, representing an improvement of 0.9% and 1.1% over the original model. Additionally, the size of the improved model was only 7.8 MB, and the computational cost was reduced by 0.1 G FLOPs. The impact of the MSDA placement on model performance was analyzed by adding it at different positions in the Neck layer, and relevant experiments were designed to verify this. The experimental results showed that adding MSDA at the small target layer in the Neck layer achieved the best effect, not only improving model performance but also maintaining low computational cost and model size, providing important references for the optimization of the MSDA mechanism. To further verify the effectiveness of the improved model, various mainstream models such as YOLOv7-tiny, YOLOv9-c, RetinaNet, and Faster-RCNN were compared with the propoed model. The experimental results showed that the improved model outperformed these models by 1.4%, 1.3%, 7.8%, and 11.6% in mAP50, 2.8%, 0.2%, 3.4%, and 5.6% in mAP50:95. Moreover, the improved model showed significant advantages in terms of floating-point operations, model size, and parameter count, with a parameter count of only 3.7 MB, making it more suitable for deployment on hardware-constrained devices such as drones. In addition, to assess the model's generalization ability, a stratified sampling method was used, selecting 20% of the images from the dataset as the test set. The results showed that the improved model could maintain a high detection accuracy in complex and variable scenes, with mAP50 and mAP50:95 increasing by 1.7% and 1.2%, respectively, compared to the original model. Considering the differences in the number of samples for each disease in the dataset, a class balance experiment was also designed. Synthetic samples were generated using oversampling techniques to increase the number of minority-class samples. The experimental results showed that the class-balanced dataset significantly improved the model's detection performance, with overall accuracy increasing from 83.1% to 85.8%, recall from 80.2% to 83.6%, mAP50 from 88.2% to 88.9%, and mAP50:95 from 37.0% to 39.4%. The class-balanced dataset significantly enhanced the model's performance in detecting minority diseases, thereby improving the overall performance of the model. [Conclusions] The improved model demonstrated significant advantages in apple leaf disease detection. By introducing SPD-Conv and MSDA attention mechanisms, the model achieved noticeable improvements in both precision and recall while effectively reducing computational costs, leading to more efficient detection capabilities. The improved model could provide continuous health monitoring throughout the apple growth process and offer robust data support for farmers' scientific decision-making before fruit harvesting.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Remote Sensing Extraction Method of Terraced Fields Based on Improved DeepLab v3+
    ZHANG Jun, CHEN Yuyan, QIN Zhenyu, ZHANG Mengyao, ZHANG Jun
    Smart Agriculture    2024, 6 (3): 46-57.   DOI: 10.12133/j.smartag.SA202312028
    Abstract490)   HTML42)    PDF(pc) (2806KB)(1055)       Save

    [Objective] The accurate estimation of terraced field areas is crucial for addressing issues such as slope erosion control, water retention, soil conservation, and increasing food production. The use of high-resolution remote sensing imagery for terraced field information extraction holds significant importance in these aspects. However, as imaging sensor technologies continue to advance, traditional methods focusing on shallow features may no longer be sufficient for precise and efficient extraction in complex terrains and environments. Deep learning techniques offer a promising solution for accurately extracting terraced field areas from high-resolution remote sensing imagery. By utilizing these advanced algorithms, detailed terraced field characteristics with higher levels of automation can be better identified and analyzed. The aim of this research is to explore a proper deep learning algorithm for accurate terraced field area extraction in high-resolution remote sensing imagery. [Methods] Firstly, a terraced dataset was created using high-resolution remote sensing images captured by the Gaofen-6 satellite during fallow periods. The dataset construction process involved data preprocessing, sample annotation, sample cropping, and dataset partitioning with training set augmentation. To ensure a comprehensive representation of terraced field morphologies, 14 typical regions were selected as training areas based on the topographical distribution characteristics of Yuanyang county. To address misclassifications near image edges caused by limited contextual information, a sliding window approach with a size of 256 pixels and a stride of 192 pixels in each direction was utilized to vary the positions of terraced fields in the images. Additionally, geometric augmentation techniques were applied to both images and labels to enhance data diversity, resulting in a high-resolution terraced remote sensing dataset. Secondly, an improved DeepLab v3+ model was proposed. In the encoder section, a lightweight MobileNet v2 was utilized instead of Xception as the backbone network for the semantic segmentation model. Two shallow features from the 4th and 7th layers of the MobileNet v2 network were extracted to capture relevant information. To address the need for local details and global context simultaneously, the multi-scale feature fusion (MSFF) module was employed to replace the atrous spatial pyramid pooling (ASPP) module. The MSFF module utilized a series of dilated convolutions with increasing dilation rates to handle information loss. Furthermore, a coordinate attention mechanism was applied to both shallow and deep features to enhance the network's understanding of targets. This design aimed to lightweight the DeepLab v3+ model while maintaining segmentation accuracy, thus improving its efficiency for practical applications. [Results and Discussions] The research findings reveal the following key points: (1) The model trained using a combination of near-infrared, red, and green (NirRG) bands demonstrated the optimal overall performance, achieving precision, recall, F1-Score, and intersection over union (IoU) values of 90.11%, 90.22%, 90.17% and 82.10%, respectively. The classification results indicated higher accuracy and fewer discrepancies, with an error in reference area of only 12 hm2. (2) Spatial distribution patterns of terraced fields in Yuanyang county were identified through the deep learning model. The majority of terraced fields were found within the slope range of 8º to 25º, covering 84.97% of the total terraced area. Additionally, there was a noticeable concentration of terraced fields within the altitude range of 1 000 m to 2 000 m, accounting for 95.02% of the total terraced area. (3) A comparison with the original DeepLab v3+ network showed that the improved DeepLab v3+ model exhibited enhancements in terms of precision, recall, F1-Score, and IoU by 4.62%, 2.61%, 3.81% and 2.81%, respectively. Furthermore, the improved DeepLab v3+ outperformed UNet and the original DeepLab v3+ in terms of parameter count and floating-point operations. Its parameter count was only 28.6% of UNet and 19.5% of the original DeepLab v3+, while the floating-point operations were only 1/5 of UNet and DeepLab v3+. This not only improved computational efficiency but also made the enhanced model more suitable for resource-limited or computationally less powerful environments. The lightweighting of the DeepLab v3+ network led to improvements in accuracy and speed. However, the slection of the NirGB band combination during fallow periods significantly impacted the model's generalization ability. [Conclusions] The research findings highlights the significant contribution of the near-infrared (NIR) band in enhancing the model's ability to learn terraced field features. Comparing different band combinations, it was evident that the NirRG combination resulted in the highest overall recognition performance and precision metrics for terraced fields. In contrast to PSPNet, UNet, and the original DeepLab v3+, the proposed model showcased superior accuracy and performance on the terraced field dataset. Noteworthy improvements were observed in the total parameter count, floating-point operations, and the Epoch that led to optimal model performance, outperforming UNet and DeepLab v3+. This study underscores the heightened accuracy of deep learning in identifying terraced fields from high-resolution remote sensing imagery, providing valuable insights for enhanced monitoring and management of terraced landscapes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Real-time Detection Algorithm of Expanded Feed Image on the Water Surface Based on Improved YOLOv11
    ZHOU Xiushan, WEN Luting, JIE Baifei, ZHENG Haifeng, WU Qiqi, LI Kene, LIANG Junneng, LI Yijian, WEN Jiayan, JIANG Linyuan
    Smart Agriculture    2024, 6 (6): 155-167.   DOI: 10.12133/j.smartag.SA202408014
    Abstract473)   HTML62)    PDF(pc) (1858KB)(935)       Save

    [Objective] During the feeding process of fish populations in aquaculture, the video image characteristics of floating extruded feed on the water surface undergo continuous variations due to a myriad of environmental factors and fish behaviors. These variations pose significant challenges to the accurate detection of feed particles, which is crucial for effective feeding management. To address these challenges and enhance the detection of floating extruded feed particles on the water surface, ,thereby providing precise decision support for intelligent feeding in intensive aquaculture modes, the YOLOv11-AP2S model, an advanced detection model was proposed. [Methods] The YOLOv11-AP2S model enhanced the YOLOv11 algorithm by incorporating a series of improvements to its backbone network, neck, and head components. Specifically, an attention for fine-grained categorization (AFGC) mechanism was introduced after the 10th layer C2PSA of the backbone network. This mechanism aimed to boost the model's capability to capture fine-grained features, which were essential for accurately identifying feed particles in complex environments with low contrast and overlapping objects. Furthermore, the C3k2 module was replaced with the VoV-GSCSP module, which incorporated more sophisticated feature extraction and fusion mechanisms. This replacement further enhanced the network's ability to extract relevant features and improve detection accuracy. To improve the model's detection of small targets, a P2 layer was introduced. However, adding a P2 layer may increase computational complexity and resource consumption, so the overall performance and resource consumption of the model must be carefully balanced. To maintain the model's real-time performance while improving detection accuracy, a lightweight VoV-GSCSP module was utilized for feature fusion at the P2 layer. This approach enabled the YOLOv11-AP2S model to achieve high detection accuracy without sacrificing detection speed or model lightweights, making it suitable for real-time applications in aquaculture. [Results and Discussions] The ablation experimental results demonstrated the superiority of the YOLOv11-AP2S model over the original YOLOv11 network. Specifically, the YOLOv11-AP2S model achieved a precision ( P) and recall ( R) of 78.70%. The mean average precision (mAP50) at an intersection over union (IoU) threshold of 0.5 was as high as 80.00%, and the F1-Score had also reached 79.00%. These metrics represented significant improvements of 6.7%, 9.0%, 9.4% (for precision, as previously mentioned), and 8.0%, respectively, over the original YOLOv11 network. These improvements showed the effectiveness of the YOLOv11-AP2S model in detecting floating extruded feed particles in complex environments. When compared to other YOLO models, the YOLOv11-AP2S model exhibits clear advantages in detecting floating extruded feed images on a self-made dataset. Notably, under the same number of iterations, the YOLOv11-AP2S model achieved higher mAP50 values and lower losses, demonstrating its superiority in detection performance. This indicated that the YOLOv11-AP2S model strikes a good balance between learning speed and network performance, enabling it to efficiently and accurately detect images of floating extruded feed on the water surface. Furthermore, the YOLOv11-AP2S model's ability to handle complex detection scenarios, such as overlapping and adhesion of feed particles and occlusion by bubbles, was noteworthy. These capabilities were crucial for accurate detection in practical aquaculture environments, where such challenges were common and can significantly impair the performance of traditional detection systems. The improvements in detection accuracy and efficiency made the YOLOv11-AP2S model a valuable tool for intelligent feeding systems in aquaculture, as it could provide more reliable and timely information on fish feeding behavior. Additionally, the introduction of the P2 layer and the use of the lightweight VoV-GSCSP module for feature fusion at this layer contributed to the model's overall performance. These enhancements enabled the model to maintain high detection accuracy while keeping computational costs and resource consumption within manageable limits. This was particularly important for real-time applications in aquaculture, where both accuracy and efficiency were critical for effective feeding management. [Conclusions] The successful application of the YOLOv11-AP2S model in detecting floating extruded feed particles demonstrates its potential to intelligent feeding systems in aquaculture. By providing accurate and timely information on fish feeding behavior, the model can help optimize feeding strategies, reduce feed waste, and improve the overall efficiency and profitability of aquaculture operations. Furthermore, the model's ability to handle complex detection scenarios and maintain high detection accuracy while keeping computational costs within manageable limits makes it a practical and valuable tool for real-time applications in aquaculture. Therefore, the YOLOv11-AP2S model holds promise for wide application in intelligent aquaculture management, contributing to the sustainability and growth of the aquaculture industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    GRA-WHO-TCN Combination Model for Forecasting Cold Chain Logistics Demand of Agricultural Products
    LIU Yan, JI Juncheng
    Smart Agriculture    2024, 6 (3): 148-158.   DOI: 10.12133/j.smartag.SA202310006
    Abstract437)   HTML37)    PDF(pc) (1280KB)(1589)       Save

    [Objective] As a critical component of agricultural product supply chain management, cold chain logistics demand prediction encounters challenges such as inadequate feature extraction, high nonlinearity of data, and the propensity for algorithms to become trapped in local optima during the digital transformation process. To address these issues and enhance the accuracy of demand prediction, achieve intelligent management of the agricultural product supply chain, a combined forecasting model that integrates grey relational analysis (GRA), the wild horse optimizer (WHO), and temporal convolutional networks (TCN) is proposed in this research. [Methods] Firstly, a cold chain logistics indicator system was established for the data of Zhejiang province, China, spanning the years 2000 to 2020. This system covered four key aspects: the economic scale of agricultural products, logistics transportation, digital technology, and agricultural product supply. Then, the GRA was applied to identify relevant indicators of cold chain logistics for agricultural products in Zhejiang province, with 17 indicators selected that had a correlation degree higher than 0.75. Sliding window technology, a problem-solving approach for data structures and algorithms, suitable for reducing the time complexity of data to a better level and improving the execution efficiency of algorithms, was used to partition the selected indicators. Secondly, the TCN model was employed to extract features of different scales by stacking multiple convolutional layers. Each layer utilized different-sized convolutional kernels to capture features within different time ranges. By utilizing the dilated convolutional module of TCN, temporal and spatial relationships within economic data were effectively mined, considering the temporal characteristics of socio-economic data and logistics information in the agricultural supply chain, and exploring the temporal and spatial features of economic data. Simultaneously, the WHO algorithm was applied to optimize five hyperparameters of the TCN model, including the number of TCN layers, the number of filters, residual blocks, Dense layers, and neurons within the Dense layer. Finally, the optimized GRA-WHO-TCN model was used to extract and analyze features from highly nonlinear multidimensional economic data, ultimately facilitating the prediction of cold chain logistics demand. [Results and Discussions] For comparative analysis of the superiority of the GRA-WHO-TCN model, the 17 selected indicators were input into long short-term memory (LSTM), TCN, WHO-LSTM, and WHO-TCN models. The parameters optimized by the WHO algorithm for the TCN model were set respectively: 2 TCN layer was, 2 residual blocks, 1 dense layer, 60 filters, and 16 neurons in the dense layer. The optimized GRA-WHO-TCN temporal model can effectively extract the temporal and spatial features of multidimensional data, fully explore the implicit relationships among indicator factors, and demonstrating good fitting effects. Compared to GRA-LSTM and GRA-TCN models, the GRA-TCN model exhibited superior performance, with a lower root mean square error of 37.34 and a higher correlation coefficient of 0.91, indicating the advantage of the TCN temporal model in handling complex nonlinear data. Furthermore, the GRA-WHO-LSTM and GRA-WHO-TCN models optimized by the WHO algorithm had improved prediction accuracy and stability compared to GRA-LSTM and GRA-TCN models, illustrating that the WHO algorithm effectively optimized model parameters to enhance the effectiveness of model fitting. When compared to the GRA-WHO-LSTM model, the GRA-WHO-TCN model displayed a lower root mean square error of 11.3 and an effective correlation coefficient of 0.95, predicting cold chain logistics demand quantities in Zhejiang province for the years 2016-2020 as 29.8, 30.46, 24.87, 26.45, and 27.99 million tons, with relative errors within 0.6%, achieving a high level of prediction accuracy. This achievement showcases a high level of prediction accuracy and underscores the utility of the GRA-WHO-TCN model in forecasting complex data scenarios. [Conclusions] The proposed GRA-WHO-TCN model demonstrated superior parameter optimization capabilities and predictive accuracy compared to the GRA-LSTM and GRA-TCN models. The predicted results align well with the development of cold chain logistics of agricultural products in Zhejiang province. This provides a scientific prediction foundation and practical reference value for the development of material flow and information flow in the agricultural supply chain under the digital economy context.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Detection Method of Effective Tillering of Rice in Field Based on Lightweight Ghost-YOLOv8 and Smart Phone
    CUI Jiale, ZENG Xiangfeng, REN Zhengwei, SUN Jian, TANG Chen, YANG Wanneng, SONG Peng
    Smart Agriculture    2024, 6 (5): 98-107.   DOI: 10.12133/j.smartag.SA202407012
    Abstract408)   HTML77)    PDF(pc) (2128KB)(657)       Save

    [Objective] The number of effective tillers per plant is one of the important agronomic traits affecting rice yield. In order to solve the problems of high cost and low accuracy of effective tiller detection caused by dense tillers, mutual occlusion and ineffective tillers in rice, a method for dividing effective tillers and ineffective tillers in rice was proposed. Combined with the deep learning model, a high-throughput and low-cost mobile phone App for effective tiller detection in rice was developed to solve the practical problems of effective tiller investigation in rice under field conditions. [Methods] The investigations of rice tillering showed that the number of effective tillers of rice was often higher than that of ineffective tillers. Based on the difference in growth height between effective and ineffective tillers of rice, a new method for distinguishing effective tillers from ineffective tillers was proposed. A fixed height position of rice plants was selected to divide effective tillers from ineffective tillers, and rice was harvested at this position. After harvesting, cross-sectional images of rice tillering stems were taken using a mobile phone, and the stems were detected and counted by the YOLOv8 model. Only the cross-section of the stem was identified during detection, while the cross-section of the panicle was not identified. The number of effective tillers of rice was determined by the number of detected stems. In order to meet the needs of field work, a mobile phone App for effective tiller detection of rice was developed for real-time detection. GhostNet was used to lighten the YOLOv8 model. Ghost Bottle-Neck was integrated into C2f to replace the original BottleNeck to form C2f-Ghost module, and then the ordinary convolution in the network was replaced by Ghost convolution to reduce the complexity of the model. Based on the lightweight Ghost-YOLOv8 model, a mobile App for effective tiller detection of rice was designed and constructed using the Android Studio development platform and intranet penetration counting. [Results and Discussions] The results of field experiments showed that there were differences in the growth height of effective tillers and ineffective tillers of rice. The range of 52 % to 55 % of the total plant height of rice plants was selected for harvesting, and the number of stems was counted as the number of effective tillers per plant. The range was used as the division standard of effective tillers and ineffective tillers of rice. The accuracy and recall rate of effective tillers counting exceeded 99%, indicating that the standard was accurate and comprehensive in guiding effective tillers counting. Using the GhostNet lightweight YOLOv8 model, the parameter quantity of the lightweight Ghost-YOLOv8 model was reduced by 43%, the FPS was increased by 3.9, the accuracy rate was 0.988, the recall rate was 0.980, and the mAP was 0.994. The model still maintains excellent performance while light weighting. Based on the lightweight Ghost-YOLOv8 model, a mobile phone App for detecting effective tillers of rice was developed. The App was tested on 100 cross-sectional images of rice stems collected under the classification criteria established in this study. Compared with the results of manual counting of effective tillers per plant, the accuracy of the App's prediction results was 99.61%, the recall rate was 98.76%, and the coefficient of determination was 0.985 9, indicating the reliability of the App and the established standards in detecting effective tillers of rice. [Conclusions] Through the lightweight Ghost-YOLOv8 model, the number of stems in the cross-sectional images of stems collected under the standard was detected to obtain the effective tiller number of rice. An Android-side rice effective tillering detection App was developed, which can meet the field investigation of rice effective tillering, help breeders to collect data efficiently, and provide a basis for field prediction of rice yield. Further research could supplement the cross-sectional image dataset of multiple rice stems to enable simultaneous measurement of effective tillers across multiple rice plants and improve work efficiency. Further optimization and enhancement of the App's functionality is necessary to provide more tiller-related traits, such as tiller angle.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Daylily Grading and Detection Model Based on Improved YOLOv10
    JIN Xuemeng, LIANG Xiyin, DENG Pengfei
    Smart Agriculture    2024, 6 (5): 108-118.   DOI: 10.12133/j.smartag.SA202407022
    Abstract398)   HTML57)    PDF(pc) (1532KB)(1274)       Save

    [Objective] In the agricultural production, accurately classifying dried daylily grades is a critical task with significant economic implications. However, current target detection models face challenges such as inadequate accuracy and excessive parameters when applied to dried daylily grading, limiting their practical application and widespread use in real-world settings. To address these issues, an innovative lightweight YOLOv10-AD network model was proposed. The model aims to enhance detection accuracy by optimizing the network structure and loss functions while reducing parameters and computational costs, making it more suitable for deployment in resource-constrained agricultural production environments. [Methods] The dried daylilies selected from the Qingyang region of Gansu province as the research subject. A large number of images of dried daylilies, categorized into three grades superior, medium, and inferior, were collected using mobile phones under varying lighting conditions and backgrounds. The images were carefully annotated and augmented to build a comprehensive dataset for dried daylily grade classification. YOLOv10 was chosen as the base network, and a newly designed backbone network called AKVanillaNet was introduced. AKVanillaNet combines AKConv (adaptive kernel convolution) with VanillaNet's deep learning and shallow inference mechanisms. The second convolutional layer in VanillaNet was replaced with AKConv, and AKConv was merged with standard convolution layers at the end of the training phase to optimize the model for capturing the unique shape characteristics of dried daylilies. This innovative design not only improved detection accuracy but also significantly reduced the number of parameters and computational costs. Additionally, the DysnakeConv module was integrated into the C2f structure, replacing the Bottleneck layer with a Bottleneck-DS layer to form the new C2f-DysnakeConv module. This module enhanced the model's sensitivity to the shapes and boundaries of targets, allowing the neural network to better capture the shape information of irregular objects like dried daylilies, further improving the model's feature extraction capability. The Powerful-IOU (PIOU) loss function was also employed, which introduced a target-size-adaptive penalty factor and a gradient adjustment function. This design guided the anchor box regression along a more direct path, helping the model better fit the data and improve overall performance. [Results and Discussions] The testing results on the dried daylily grade classification dataset demonstrated that the YOLOv10-AD model achieved a mean average precision (mAP) of 85.7%. The model's parameters, computational volume, and size were 2.45 M, 6.2 GFLOPs, and 5.0 M, respectively, with a frame rate of 156 FPS. Compared to the benchmark model, YOLOv10-AD improved mAP by 5.7% and FPS by 25.8%, while reducing the number of parameters, computational volume, and model size by 9.3%, 24.4%, and 9.1%, respectively. These results indicated that YOLOv10-AD not only improved detection accuracy but also reduced the model's complexity, making it easier to deploy in real-world production environments. Furthermore, YOLOv10-AD outperformed larger models in the same series, such as YOLOv10s and YOLOv10m. Specifically, the weight, parameters, and computational volume of YOLOv10-AD were only 31.6%, 30.5%, and 25.3% of those in YOLOv10s, and 15.7%, 14.8%, and 9.8% of YOLOv10m. Despite using fewer resources, YOLOv10-AD achieved a mAP increase of 2.4% over YOLOv10s and 1.9% over YOLOv10m. These findings confirm that YOLOv10-AD maintains high detection accuracy while requiring significantly fewer resources, making it more suitable for agricultural production environments where computational capacity may be limited. The study also examined the performance of YOLOv10-AD under different lighting conditions. The results showed that YOLOv10-AD achieved an average accuracy of 92.3% in brighter environments and 78.6% in darker environments. In comparison, the YOLOv10n model achieved 88.9% and 71.0% in the same conditions, representing improvements of 3.4% and 7.6%, respectively. These findings demonstrate that YOLOv10-AD has a distinct advantage in maintaining high accuracy and confidence in grading dried daylilies across varying lighting conditions. [Conclusions] The YOLOv10-AD network model proposed significantly reduces the number of parameters and computational costs without compromising detection accuracy. This model presents a valuable technical reference for intelligent classification of dried daylily grades in agricultural production environments, particularly where resources are constrained.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Identification Method of Kale Leaf Ball Based on Improved UperNet
    ZHU Yiping, WU Huarui, GUO Wang, WU Xiaoyan
    Smart Agriculture    2024, 6 (3): 128-137.   DOI: 10.12133/j.smartag.SA202401020
    Abstract392)   HTML112)    PDF(pc) (1568KB)(656)       Save

    [Objective] Kale is an important bulk vegetable crop worldwide, its main growth characteristics are outer leaves and leaf bulbs. The traits of leaf bulb kale are crucial for adjusting water and fertilizer parameters in the field to achieve maximum yield. However, various factors such as soil quality, light exposure, leaf overlap, and shading can affect the growth of in practical field conditions. The similarity in color and texture between leaf bulbs and outer leaves complicates the segmentation process for existing recognition models. In this paper, the segmentation of kale outer leaves and leaf bulbs in complex field background was proposed, using pixel values to determine leaf bulb size for intelligent field management. A semantic segmentation algorithm, UperNet-ESA was proposed to efficiently and accurately segment nodular kale outer leaf and leaf bulb in field scenes using the morphological features of the leaf bulbs and outer leaves of nodular kale to realize the intelligent management of nodular kale in the field. [Methods] The UperNet-ESA semantic segmentation algorithm, which uses the unified perceptual parsing network (UperNet) as an efficient semantic segmentation framework, is more suitable for extracting crop features in complex environments by integrating semantic information across different scales. The backbone network was improved using ConvNeXt, which is responsible for feature extraction in the model. The similarity between kale leaf bulbs and outer leaves, along with issues of leaf overlap affecting accurate target contour localization, posed challenges for the baseline network, leading to low accuracy. ConvNeXt effectively combines the strengths of convolutional neural networks (CNN) and Transformers, using design principles from Swin Transformer and building upon ResNet50 to create a highly effective network structure. The simplicity of the ConvNeXt design not only enhances segmentation accuracy with minimal model complexity, but also positions it as a top performer among CNN architectures. In this study, the ConvNeXt-B version was chosen based on considerations of computational complexity and the background characteristics of the knotweed kale image dataset. To enhance the model's perceptual acuity, block ratios for each stage were set at 3:3:27:3, with corresponding channel numbers of 128, 256, 512 and 1 024, respectively. Given the visual similarity between kale leaf bulbs and outer leaves, a high-efficiency channel attention mechanism was integrated into the backbone network to improve feature extraction in the leaf bulb region. By incorporating attention weights into feature mapping through residual inversion, attention parameters were cyclically trained within each block, resulting in feature maps with attentional weights. This iterative process facilitated the repeated training of attentional parameters and enhanced the capture of global feature information. To address challenges arising from direct pixel addition between up-sampling and local features, potentially leading to misaligned context in feature maps and erroneous classifications at kale leaf boundaries, a feature alignment module and feature selection module were introduced into the feature pyramid network to refine target boundary information extraction and enhance model segmentation accuracy. [Results and Discussions] The UperNet-ESA semantic segmentation model outperforms the current mainstream UNet model, PSPNet model, DeepLabV3+ model in terms of segmentation accuracy, where mIoU and mPA reached 92.45% and 94.32%, respectively, and the inference speed of up to 16.6 frames per second (fps). The mPA values were better than that of the UNet model, PSPNet model, ResNet-50 based, MobilenetV2, and DeepLabV3+ model with Xception as the backbone, showing improvements of 11.52%, 13.56%, 8.68%, 4.31%, and 6.21%, respectively. Similarly, the mIoU exhibited improvements of 12.21%, 13.04%, 10.65%, 3.26% and 7.11% compared to the mIoU of the UNet-based model, PSPNet model, and DeepLabV3+ model based on the ResNet-50, MobilenetV2, and Xception backbones, respectively. This performance enhancement can be attributed to the introduction of the ECA module and the improvement made to the feature pyramid network in this model, which strengthen the judgement of the target features at each stage to obtain effective global contextual information. In addition, although the PSPNet model had the fastest inference speed, the overall accuracy was too low to for developing kale semantic segmentation models. On the contrary, the proposed model exhibited superior inference speed compared to all other network models. [Conclusions] The experimental results showed that the UperNet-ESA semantic segmentation model proposed in this study outperforms the original network in terms of performance. The improved model achieves the best accuracy-speed balance compared to the current mainstream semantic segmentation networks. In the upcoming research, the current model will be further optimized and enhanced, while the kale dataset will be expanded to include a wider range of samples of nodulated kale leaf bulbs. This expansion is intended to provide a more robust and comprehensive theoretical foundation for intelligent kale field management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Lightweight Model for Detecting Small Targets of Litchi Pests Based on Improved YOLOv10n
    LI Zusheng, TANG Jishen, KUANG Yingchun
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202412003
    Online available: 24 January 2025

    Seedling Stage Corn Line Detection Method Based on Improved YOLOv8
    LI Hongbo, TIAN Xin, RUAN Zhiwen, LIU Shaowen, REN Weiqi, SU Zhongbin, GAO Rui, KONG Qingming
    Smart Agriculture    2024, 6 (6): 72-84.   DOI: 10.12133/j.smartag.SA202408008
    Abstract381)   HTML80)    PDF(pc) (3458KB)(236)       Save

    [Objective] Crop line extraction is critical for improving the efficiency of autonomous agricultural machines in the field. However, traditional detection methods struggle to maintain high accuracy and efficiency under challenging conditions, such as strong light exposure and weed interference. The aims are to develop an effective crop line extraction method by combining YOLOv8-G, Affinity Propagation, and the Least Squares method to enhance detection accuracy and performance in complex field environments. [Methods] The proposed method employs machine vision techniques to address common field challenges. YOLOv8-G, an improved object detection algorithm that combines YOLOv8 and GhostNetV2 for lightweight, high-speed performance, was used to detect the central points of crops. These points were then clustered using the Affinity Propagation algorithm, followed by the application of the Least Squares method to extract the crop lines. Comparative tests were conducted to evaluate multiple backbone networks within the YOLOv8 framework, and ablation studies were performed to validate the enhancements made in YOLOv8-G. [Results and Discussions] The performance of the proposed method was compared with classical object detection and clustering algorithms. The YOLOv8-G algorithm achieved average precision (AP) values of 98.22%, 98.15%, and 97.32% for corn detection at 7, 14, and 21 days after emergence, respectively. Additionally, the crop line extraction accuracy across all stages was 96.52%. These results demonstrate the model's ability to maintain high detection accuracy despite challenging conditions in the field. [Conclusions] The proposed crop line extraction method effectively addresses field challenges such as lighting and weed interference, enabling rapid and accurate crop identification. This approach supports the automatic navigation of agricultural machinery, offering significant improvements in the precision and efficiency of field operations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Trajectory Tracking Method of Agricultural Machinery Multi-Robot Formation Operation Based on MPC Delay Compensator
    LUAN Shijie, SUN Yefeng, GONG Liang, ZHANG Kai
    Smart Agriculture    2024, 6 (3): 69-81.   DOI: 10.12133/j.smartag.SA202306013
    Abstract381)   HTML31)    PDF(pc) (1797KB)(871)       Save

    [Objective] The technology of multi-machine convoy driving has emerged as a focal point in the field of agricultural mechanization. By organizing multiple agricultural machinery units into convoys, unified control and collaborative operations can be achieved. This not only enhances operational efficiency and reduces costs, but also minimizes human labor input, thereby maximizing the operational potential of agricultural machinery. In order to solve the problem of communication delay in cooperative control of multi-vehicle formation and its compensation strategy, the trajectory control method of multi-vehicle formation was proposed based on model predictive control (MPC) delay compensator. [Methods] The multi-vehicle formation cooperative control strategy was designed, which introduced the four-vehicle formation cooperative scenario in three lanes, and then introduced the design of the multi-vehicle formation cooperative control architecture, which was respectively enough to establish the kinematics and dynamics model and equations of the agricultural machine model, and laied down a sturdy foundation for solving the formation following problem later. The testing and optimization of automatic driving algorithms based on real vehicles need to invest too much time and economic costs, and were subject to the difficulties of laws and regulations, scene reproduction and safety, etc. Simulation platform testing could effectively solve the above question. For the agricultural automatic driving multi-machine formation scenarios, the joint simulation platform Carsim and Simulink were used to simulate and validate the formation driving control of agricultural machines. Based on the single-machine dynamics model of the agricultural machine, a delay compensation controller based on MPC was designed. Feedback correction first detected the actual output of the object and then corrected the model-based predicted output with the actual output and performed a new optimization. Based on the above model, the nonlinear system of kinematics and dynamics was linearized and discretized in order to ensure the real-time solution. The objective function was designed so that the agricultural machine tracks on the desired trajectory as much as possible. And because the operation range of the actuator was limited, the control increment and control volume were designed with corresponding constraints. Finally, the control increment constraints were solved based on the front wheel angle constraints, front wheel angle increments, and control volume constraints of the agricultural machine. [Results and Discussions] Carsim and MATLAB/Simulink could be effectively compatible, enabling joint simulation of software with external solvers. When the delay step size d=5 was applied with delay compensation, the MPC response was faster and smoother; the speed error curve responded faster and gradually stabilized to zero error without oscillations. Vehicle 1 effectively changed lanes in a short time and maintains the same lane as the lead vehicle. In the case of a longer delay step size d =10, controllers without delay compensation showed more significant performance degradation. Even under higher delay conditions, MPC with delay compensation applied could still quickly respond with speed error and longitudinal acceleration gradually stabilizing to zero error, avoiding oscillations. The trajectory of Vehicle 1 indicated that the effectiveness of the delay compensation mechanism decreased under extreme delay conditions. The simulation results validated the effectiveness of the proposed formation control algorithm, ensuring that multiple vehicles could successfully change lanes to form queues while maintaining specific distances and speeds. Furthermore, the communication delay compensation control algorithm enables vehicles with added delay to effectively complete formation tasks, achieving stable longitudinal and lateral control. This confirmed the feasibility of the model predictive controller with delay compensation proposed. [Conclusions] At present, most of the multi-machine formation coordination is based on simulation platform for verification, which has the advantages of safety, economy, speed and other aspects, however, there's still a certain gap between the idealized model in the simulation platform and the real machine experiment. Therefore, multi-machine formation operation of agricultural equipment still needs to be tested on real machines under sound laws and regulations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Remote Sensing Identification Method of Cultivated Land at Hill County of Sichuan Basin Based on Deep Learning
    LI Hao, DU Yuqiu, XIAO Xingzhu, CHEN Yanxi
    Smart Agriculture    2024, 6 (3): 34-45.   DOI: 10.12133/j.smartag.SA202308002
    Abstract369)   HTML48)    PDF(pc) (2649KB)(2441)       Save

    [Objective] To fully utilize and protect farmland and lay a solid foundation for the sustainable use of land, it is particularly important to obtain real-time and precise information regarding farmland area, distribution, and other factors. Leveraging remote sensing technology to obtain farmland data can meet the requirements of large-scale coverage and timeliness. However, the current research and application of deep learning methods in remote sensing for cultivated land identification still requires further improvement in terms of depth and accuracy. The objective of this study is to investigate the potential application of deep learning methods in remote sensing for identifying cultivated land in the hilly areas of Southwest China, to provide insights for enhancing agricultural land utilization and regulation, and for harmonizing the relationship between cultivated land and the economy and ecology. [Methods] Santai county, Mianyang city, Sichuan province, China (30°42'34"~31°26'35"N, 104°43'04"~105°18'13"E) was selected as the study area. High-resolution imagery from two scenes captured by the Gaofen-6 (GF-6) satellite served as the primary image data source. Additionally, 30-meter resolution DEM data from the United States National Aeronautics and Space Administration (NASA) in 2020 was utilized. A land cover data product, SinoLC-1, was also incorporated for comparative evaluation of the accuracy of various extraction methods' results. Four deep learning models, namely Unet, PSPNet, DeeplabV3+, and Unet++, were utilized for remote sensing land identification research in cultivated areas. The study also involved analyzing the identification accuracy of cultivated land in high-resolution satellite images by combining the results of the random forest (RF) algorithm along with the deep learning models. A validation dataset was constructed by randomly generating 1 000 vector validation points within the research area. Concurrently, Google Earth satellite images with a resolution of 0.3 m were used for manual visual interpretation to determine the land cover type of the pixels where the validation points are located. The identification results of each model were compared using a confusion matrix to compute five accuracy evaluation metrics: Overall accuracy (OA), intersection over union (IoU), mean intersection over union (MIoU), F1-Score, and Kappa Coefficient to assess the cultivated land identification accuracy of different models and data products. [Results and Discussions] The deep learning models displayed significant advances in accuracy evaluation metrics, surpassing the performance of traditional machine learning approaches like RF and the latest land cover product, SinoLC-1 Landcover. Among the models assessed, the UNet++ model performed the best, its F1-Score, IoU, MIoU, OA, and Kappa coefficient values were 0.92, 85.93%, 81.93%, 90.60%, and 0.80, respectively. DeeplabV3+, UNet, and PSPNet methods followed suit. These performance metrics underscored the superior accuracy of the UNet++ model in precisely identifying and segmenting cultivated land, with a remarkable increase in accuracy of nearly 20% than machine learning methods and 50% for land cover products. Four typical areas of town, water body, forest land and contiguous cultivated land were selected to visually compare the results of cultivated land identification results. It could be observed that the deep learning models generally exhibited consistent distribution patterns with the satellite imageries, accurately delineating the boundaries of cultivated land and demonstrating overall satisfactory performance. However, due to the complex features in remote sensing images, the deep learning models still encountered certain challenges of omission and misclassification in extracting cultivated land. Among them, the UNet++ model showed the closest overall extraction results to the ground truth and exhibited advantages in terms of completeness of cultivated land extraction, discrimination between cultivated land and other land classes, and boundary extraction compared to other models. Using the UNet++ model with the highest recognition accuracy, two types of images constructed with different features—solely spectral features and spectral combined with terrain features—were utilized for cultivated land extraction. Based on the three metrics of IoU, OA, and Kappa, the model incorporating both spectral and terrain features showed improvements of 0.98%, 1.10%, and 0.01% compared to the model using only spectral features. This indicated that fusing spectral and terrain features can achieve information complementarity, further enhancing the identification effectiveness of cultivated land. [Conclusions] This study focuses on the practicality and reliability of automatic cultivated land extraction using four different deep learning models, based on high-resolution satellite imagery from the GF-6 in Santai county in China. Based on the cultivated land extraction results in Santai county and the differences in network structures among the four deep learning models, it was found that the UNet++ model, based on UNet, can effectively improve the accuracy of cultivated land extraction by introducing the mechanism of skip connections. Overall, this study demonstrates the effectiveness and practical value of deep learning methods in obtaining accurate farmland information from high-resolution remote sensing imagery.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Ecological Risk Assessment of Cultivated Land Based on Landscape Pattern: A Case Study of Tongnan District, Chongqing
    ZHANG Xingshan, YANG Heng, MA Wenqiu, YANG Minli, WANG Haiyi, YOU Yong, HUI Yunting, GONG Zeqi, WANG Tianyi
    Smart Agriculture    2024, 6 (3): 58-68.   DOI: 10.12133/j.smartag.SA202306008
    Abstract369)   HTML24)    PDF(pc) (1636KB)(611)       Save

    [Objective] Farmland consolidation for agricultural mechanization in hilly and mountainous areas can alter the landscape pattern, elevation, slope and microgeomorphology of cultivated land. It is of great significance to assess the ecological risk of cultivated land to provide data reference for the subsequent farmland consolidation for agricultural mechanization. This study aims to assess the ecological risk of cultivated land before and after farmland consolidation for agricultural mechanization in hilly and mountainous areas, and to explore the relationship between cultivated land ecological risk and cultivated land slope. [Methods] Twenty counties in Tongnan district of Chongqing city was selected as the assessment units. Based on the land use data in 2010 and 2020 as two periods, ArcGIS 10.8 and Excel software were used to calculate landscape pattern indices. The weights for each index were determined by entropy weight method, and an ecological risk assessment model was constructed, which was used to reveal the temporal and spatial change characteristics of ecological risk. Based on the principle of mathematical statistics, the correlation analysis between cultivated land ecological risk and cultivated land slope was carried out, which aimed to explore the relationship between cultivated land ecological risk and cultivated land slope. [Results and Discussions] Comparing to 2010, patch density (PD), division (D), fractal dimension (FD), and edge density (ED) of cultivated land all decreased in 2020, while meant Patch Size (MPS) increased, indicating an increase in the contiguity of cultivated land. The mean shape index (MSI) of cultivated land increased, indicating that the shape of cultivated land tended to be complicated. The landscape disturbance index (U) decreased from 0.97 to 0.94, indicating that the overall resistance to disturbances in cultivated land has increased. The landscape vulnerability index (V) increased from 2.96 to 3.20, indicating that the structure of cultivated land become more fragile. The ecological risk value of cultivated land decreased from 3.10 to 3.01, indicating the farmland consolidation for agricultural mechanization effectively improved the landscape pattern of cultivated land and enhanced the safety of the agricultural ecosystem. During the two periods, the ecological risk areas were primarily composed of low-risk and relatively low-risk zones. The area of low-risk zones increased by 6.44%, mainly expanding towards the northern part, while the area of relatively low-risk zones increased by 6.17%, primarily spreading towards the central-eastern and southeastern part. The area of moderate-risk zones increased by 24.4%, mainly extending towards the western and northwestern part, while the area of relatively high-risk zones decreased by 60.70%, with some new additions spreading towards the northeastern part. The area of high-risk zones increased by 16.30%, with some new additions extending towards the northwest part. Overall, the ecological safety zones of cultivated relatively increased. The cultivated land slope was primarily concentrated in the range of 2° to 25°. On the one hand, when the cultivated land slope was less than 15°, the proportion of the slope area was negatively correlated with the ecological risk value. On the other hand, when the slope was above 15°, the proportion of the slope area was positively correlated with the ecological risk value. In 2010, there was a highly significant correlation between the proportion of slope area and ecological risk value for cultivated land slope within the ranges of 5° to 8°, 15° to 25°, and above 25°, with corresponding correlation coefficients of 0.592, 0.609, and 0.849, respectively. In 2020, there was a highly significant correlation between the proportion of slope area and ecological risk value for cultivated land slope within the ranges of 2° to 5°, 5° to 8°, 15° to 25°, and above 25°, with corresponding correlation coefficients of 0.534, 0.667, 0.729, and 0.839, respectively. [Conclusions] The assessment of cultivated land ecological risk in Tongnan district of Chongqing city before and after the farmland consolidation for agricultural mechanization, as well as the analysis of the correlation between ecological risk and cultivated land slope, demonstrate that the farmland consolidation for agricultural mechanization can reduce cultivated land ecological risk, and the proportion of cultivated land slope can be an important basis for precision guidance in the farmland consolidation for agricultural mechanization. Considering the occurrence of moderate sheet erosion from a slope of 5° and intense erosion from a slope of 10° to 15°, and taking into account the reduction of ecological risk value and the actual topographic conditions, the subsequent farmland consolidation for agricultural mechanization in Tongnan district should focus on areas with cultivated land slope ranging from 5° to 8° and 15° to 25°.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Adaptive Time Horizon MPC Path Tracking Control Method for Mowing Robot
    HE Qing, JI Jie, FENG Wei, ZHAO Lijun, ZHANG Bohan
    Smart Agriculture    2024, 6 (3): 82-93.   DOI: 10.12133/j.smartag.SA202401010
    Abstract363)   HTML25)    PDF(pc) (2278KB)(794)       Save

    [Objective] The traditional predictive control approach usually employs a fixed time horizon and often overlooks the impact of changes in curvature and road bends. This oversight leads to subpar tracking performance and inadequate adaptability of robots for navigating curves and paths. Although extending the time horizon of the standard fixed time horizon model predictive control (MPC) can improve curve path tracking accuracy, it comes with high computational costs, making it impractical in situations with restricted computing resources. Consequently, an adaptive time horizon MPC controller was developed to meet the requirements of complex tasks such as autonomous mowing. [Methods] Initially, it was crucial to establish a kinematic model for the mowing robot, which required employing Taylor linearization and Euler method discretization techniques to ensure accurate path tracking. The prediction equation for the error model was derived after conducting a comprehensive analysis of the robot's kinematics model employed in mowing. Second, the size of the previewing area was determined by utilizing the speed data and reference path information gathered from the mowing robot. The region located a certain distance ahead of the robot's current position, was identified to as the preview region, enabling a more accurate prediction of the robot's future traveling conditions. Calculations for both the curve factor and curve change factor were carried out within this preview region. The curvature factor represented the initial curvature of the path, while the curvature change factor indicated the extent of curvature variation in this region. These two variables were then fed into a fuzzy controller, which adjusted the prediction time horizon of the MPC. The integration enabled the mowing robot to promptly adjust to changes in the path's curvature, thereby improving its accuracy in tracking the desired trajectory. Additionally, a novel technique for triggering MPC execution was developed to reduce computational load and improve real-time performance. This approach ensured that MPC activation occurred only when needed, rather than at every time step, resulting in reduced computational expenses especially during periods of smooth robot motion where unnecessary computation overhead could be minimized. By meeting kinematic and dynamic constraints, the optimization algorithm successfully identified an optimal control sequence, ultimately enhancing stability and reliability of the control system. Consequently, these set of control algorithms facilitated precise path tracking while considering both kinematic and dynamic limitations in complex environments. [Results and Discussion] The adaptive time-horizon MPC controller effectively limited the maximum absolute heading error and maximum absolute lateral error to within 0.13 rad and 11 cm, respectively, surpassing the performance of the MPC controller in the control group. Moreover, compared to both the first and fourth groups, the adaptive time-horizon MPC controller achieved a remarkable reduction of 75.39% and 57.83% in mean values for lateral error and heading error, respectively (38.38% and 31.84%, respectively). Additionally, it demonstrated superior tracking accuracy as evidenced by its significantly smaller absolute standard deviation of lateral error (0.025 6 m) and course error (0.025 5 rad), outperforming all four fixed time-horizon MPC controllers tested in the study. Furthermore, this adaptive approach ensured precise tracking and control capabilities for the mowing robot while maintaining a remarkably low average solution time of only 0.004 9 s, notably faster than that observed with other control data sets-reducing computational load by approximately 10.9 ms compared to maximum time-horizon MPC. [Conclusions] The experimental results demonstrated that the adaptive time-horizon MPC tracking approach effectively addressed the trade-off between control accuracy and computational complexity encountered in fixed time-horizon MPC. By dynamically adjusting the time horizon length the and performing MPC calculations based on individual events, this approach can more effectively handle scenarios with restricted computational resources, ensuring superior control precision and stability. Furthermore, it achieves a balance between control precision and real-time performance in curve route tracking for mowing robots, offering a more practical and reliable solution for their practical application.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Severity Grading Model for Camellia Oleifera Anthracnose Infection Based on Improved YOLACT
    NIE Ganggang, RAO Honghui, LI Zefeng, LIU Muhua
    Smart Agriculture    2024, 6 (3): 138-147.   DOI: 10.12133/j.smartag.SA202402002
    Abstract352)   HTML30)    PDF(pc) (2130KB)(1038)       Save

    [Objective] Camellia oleifera is one of the four major woody oil plants in the world. Diseases is a significant factor leading to the decline in quality of Camellia oleifera and the financial loss of farmers. Among these diseases, anthracnose is a common and severe disease in Camellia oleifera forests, directly impacting yields and production rates. Accurate disease assessment can improve the prevention and control efficiency and safeguarding the farmers' profit. In this study, an improved You Only Look at CoefficienTs (YOLACT) based method was proposed to realize automatic and efficient grading of the severity of Camellia oleifera leaf anthracnose. [Methods] High-resolution images of Camellia oleifera anthracnose leaves were collected using a smartphone at the National Camellia oleifera Seed Base of Jiangxi Academy of Forestry, and finally 975 valid images were retained after a rigorous screening process. Five data enhancement means were applied, and a data set of 5 850 images was constructed finally, which was divided into training, validation, and test sets in a ratio of 7:2:1. For model selection, the Camellia-YOLACT model was proposed based on the YOLACT instance segmentation model, and by introducing improvements such as Swin-Transformer, weighted bi-directional feature pyramid network, and HardSwish activation function. The Swin Transformer was utilized for feature extraction in the backbone network part of YOLACT, leveraging the global receptive field and shift window properties of the self-attention mechanism in the Transformer architecture to enhance feature extraction capabilities. Additionally, a weighted bidirectional feature pyramid network was introduced to fuse feature information from different scales to improve the detection ability of the model for objects at different scales, thereby improving the detection accuracy. Furthermore, to increase the the model's robustness against the noise in the input data, the HardSwish activation function with stronger nonlinear capability was adopted to replace the ReLu activation function of the original model. Since images in natural environments usually have complex background and foreground information, the robustness of HardSwish helped the model better handling these situations and further improving the detection accuracy. With the above improvements, the Camellia-YOLACT model was constructed and experimentally validated by testing the Camellia oleifera anthracnose leaf image dataset. [Results and Discussions] A transfer learning approach was used for experimental validation on the Camellia oleifera anthracnose severity grading dataset, and the results of the ablation experiments showed that the mAP75 of Camellia-YOLACT proposed in this study was 86.8%, mAPall was 78.3%, mAR was 91.6% which were 5.7%, 2.5% and 7.9% higher than YOLACT model. In the comparison experiments, Camellia-YOLACT performed better than Segmenting Objects by Locations (SOLO) in terms of both accuracy and speed, and its detection speed was doubled compared to Mask R-CNN algorithm. Therefore, the Camellia-YOLACT algorithm was suitable in Camellia oleifera gardens for anthracnose real-time segmentation. In order to verify the outdoors detection performance of Camellia-YOLACT model, 36 groups of Camellia oleifera anthracnose grading experiments were conducted. Experimental results showed that the grading correctness of Camellia oleifera anthracnose injection severity reached 94.4%, and the average absolute error of K-value was 1.09%. Therefore, the Camellia-YOLACT model proposed in this study has a better performance on the grading of the severity of Camellia oleifera anthracnose. [Conclusions] The Camellia-YOLACT model proposed got high accuracy in leaf and anthracnose segmentation of Camellia oleifera, on the basis of which it can realize automatic grading of the severity of Camellia oleifera anthracnose. This research could provide technical support for the precise control of Camellia oleifera diseases.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospect of Quality Intelligent Control Technology in Facilities Environment of Characteristic Agricultural Products
    GUO Wei, WU Huarui, GUO Wang, GU Jingqiu, ZHU Huaji
    Smart Agriculture    2024, 6 (6): 44-62.   DOI: 10.12133/j.smartag.SA202411017
    Abstract351)   HTML51)    PDF(pc) (3153KB)(973)       Save

    [Significance] In view of the lack of monitoring means of quality influence factors in the production process of characteristic agricultural products with in central and western regions of China, the weak ability of intelligent control, the unclear coupling relationship of quality control elements and the low degree of systematic application, the existing technologies described such as intelligent monitoring of facility environment, growth and nutrition intelligent control model, architecture of intelligent management and control platform and so on. Through the application of the Internet of Things, big data and the new generation of artificial intelligence technology, it provides technical support for the construction and application of intelligent process quality control system for the whole growth period of characteristic agricultural products. [Progress] The methods of environmental regulation and nutrition regulation are analyzed, including single parameters and combined control methods, such as light, temperature, humidity, CO2 concentration, fertilizer and water, etc. The multi-parameter coupling control method has the advantage of more comprehensive scene analysis. Based on the existing technology, a multi-factor coupling method of integrating growth state, agronomy, environment, input and agricultural work is put forward. This paper probes into the system architecture of the whole process service of quality control, the visual identification system of the growth process of agricultural products and the knowledge-driven agricultural technical service system, and introduces the technology of the team in the disease knowledge Q & A scene through multi-modal knowledge graph and large model technology. [Conclusions and Prospects] Based on the present situation of the production of characteristic facility agricultural products and the overall quality of farmers in the central and western regions of China, it is appropriate to transfer the whole technical system such as facility tomato, facility cucumber and so on. According to the varieties of characteristic agricultural products, cultivation models, quality control objectives to adapt to light, temperature, humidity and other parameters, as well as fertilizer, water, medicine and other input plans, a multi-factor coupling model suitable for a specific planting area is generated and long-term production verification and model correction are carried out. And popularize it in a wider area, making full use of the advantages of intelligent equipment and data elements will promote the realization of light simplification of production equipment, scene of intelligent technology, diversification of service models, on-line quality control, large-scale production of digital intelligence, and value of data elements, further cultivate facilities to produce new quality productivity.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Recognition Method of Facility Cucumber Farming Behaviours Based on Improved SlowFast Model
    HE Feng, WU Huarui, SHI Yangming, ZHU Huaji
    Smart Agriculture    2024, 6 (3): 118-127.   DOI: 10.12133/j.smartag.SA202402001
    Abstract348)   HTML104)    PDF(pc) (1737KB)(450)       Save

    [Objective] The identification of agricultural activities plays a crucial role for greenhouse vegetables production, particularly in the precise management of cucumber cultivation. By monitoring and analyzing the timing and procedures of agricultural operations, effective guidance can be provided for agricultural production, leading to increased crop yield and quality. However, in practical applications, the recognition of agricultural activities in cucumber cultivation faces significant challenges. The complex and ever-changing growing environment of cucumbers, including dense foliage and internal facility structures that may obstruct visibility, poses difficulties in recognizing agricultural activities. Additionally, agricultural tasks involve various stages such as planting, irrigation, fertilization, and pruning, each with specific operational intricacies and skill requirements. This requires the recognition system to accurately capture the characteristics of various complex movements to ensure the accuracy and reliability of the entire recognition process. To address the complex challenges, an innovative algorithm: SlowFast-SMC-ECA (SlowFast-Spatio-Temporal Excitation, Channel Excitation, Motion Excitation-Efficient Channel Attention) was proposed for the recognition of agricultural activity behaviors in cucumber cultivation within facilities. [Methods] This algorithm represents a significant enhancement to the traditional SlowFast model, with the goal of more accurately capturing hand motion features and crucial dynamic information in agricultural activities. The fundamental concept of the SlowFast model involved processing video streams through two distinct pathways: the Slow Pathway concentrated on capturing spatial detail information, while the Fast Pathway emphasized capturing temporal changes in rapid movements. To further improve information exchange between the Slow and Fast pathways, lateral connections were incorporated at each stage. Building upon this foundation, the study introduced innovative enhancements to both pathways, improving the overall performance of the model. In the Fast Pathway, a multi-path residual network (SMC) concept was introduced, incorporating convolutional layers between different channels to strengthen temporal interconnectivity. This design enabled the algorithm to sensitively detect subtle temporal variations in rapid movements, thereby enhancing the recognition capability for swift agricultural actions. Meanwhile, in the Slow Pathway, the traditional residual block was replaced with the ECA-Res structure, integrating an effective channel attention mechanism (ECA) to improve the model's capacity to capture channel information. The adaptive adjustment of channel weights by the ECA-Res structure enriched feature expression and differentiation, enhancing the model's understanding and grasp of key spatial information in agricultural activities. Furthermore, to address the challenge of class imbalance in practical scenarios, a balanced loss function (Smoothing Loss) was developed. By introducing regularization coefficients, this loss function could automatically adjust the weights of different categories during training, effectively mitigating the impact of class imbalance and ensuring improved recognition performance across all categories. [Results and Discussions] The experimental results significantly demonstrated the outstanding performance of the improved SlowFast-SMC-ECA model on a specially constructed agricultural activity dataset. Specifically, the model achieved an average recognition accuracy of 80.47%, representing an improvement of approximately 3.5% compared to the original SlowFast model. This achievement highlighted the effectiveness of the proposed improvements. Further ablation studies revealed that replacing traditional residual blocks with the multi-path residual network (SMC) and ECA-Res structures in the second and third stages of the SlowFast model leads to superior results. This highlighted that the improvements made to the Fast Pathway and Slow Pathway played a crucial role in enhancing the model's ability to capture details of agricultural activities. Additional ablation studies also confirmed the significant impact of these two improvements on improving the accuracy of agricultural activity recognition. Compared to existing algorithms, the improved SlowFast-SMC-ECA model exhibited a clear advantage in prediction accuracy. This not only validated the potential application of the proposed model in agricultural activity recognition but also provided strong technical support for the advancement of precision agriculture technology. In conclusion, through careful refinement and optimization of the SlowFast model, it was successfully enhanced the model's recognition capabilities in complex agricultural scenarios, contributing valuable technological advancements to precision management in greenhouse cucumber cultivation. [Conclusions] By introducing advanced recognition technologies and intelligent algorithms, this study enhances the accuracy and efficiency of monitoring agricultural activities, assists farmers and agricultural experts in managing and guiding the operational processes within planting facilities more efficiently. Moreover, the research outcomes are of immense value in improving the traceability system for agricultural product quality and safety, ensuring the reliability and transparency of agricultural product quality.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Tea Shoot Picking Point Recognition Model Based on Improved DeepLabV3+
    HU Chengxi, TAN Lixin, WANG Wenyin, SONG Min
    Smart Agriculture    2024, 6 (5): 119-127.   DOI: 10.12133/j.smartag.SA202403016
    Abstract337)   HTML40)    PDF(pc) (1379KB)(720)       Save

    [Objective] The picking of famous and high-quality tea is a crucial link in the tea industry. Identifying and locating the tender buds of famous and high-quality tea for picking is an important component of the modern tea picking robot. Traditional neural network methods suffer from issues such as large model size, long training times, and difficulties in dealing with complex scenes. In this study, based on the actual scenario of the Xiqing Tea Garden in Hunan Province, proposes a novel deep learning algorithm was proposed to solve the precise segmentation challenge of famous and high-quality tea picking points. [Methods] The primary technical innovation resided in the amalgamation of a lightweight network architecture, MobilenetV2, with an attention mechanism known as efficient channel attention network (ECANet), alongside optimization modules including atrous spatial pyramid pooling (ASPP). Initially, MobilenetV2 was employed as the feature extractor, substituting traditional convolution operations with depth wise separable convolutions. This led to a notable reduction in the model's parameter count and expedited the model training process. Subsequently, the innovative fusion of ECANet and ASPP modules constituted the ECA_ASPP module, with the intention of bolstering the model's capacity for fusing multi-scale features, especially pertinent to the intricate recognition of tea shoots. This fusion strategy facilitated the model's capability to capture more nuanced features of delicate shoots, thereby augmenting segmentation accuracy. The specific implementation steps entailed the feeding of image inputs through the improved network, whereupon MobilenetV2 was utilized to extract both shallow and deep features. Deep features were then fused via the ECA_ASPP module for the purpose of multi-scale feature integration, reinforcing the model's resilience to intricate backgrounds and variations in tea shoot morphology. Conversely, shallow features proceeded directly to the decoding stage, undergoing channel reduction processing before being integrated with upsampled deep features. This divide-and-conquer strategy effectively harnessed the benefits of features at differing levels of abstraction and, furthermore, heightened the model's recognition performance through meticulous feature fusion. Ultimately, through a sequence of convolutional operations and upsampling procedures, a prediction map congruent in resolution with the original image was generated, enabling the precise demarcation of tea shoot harvesting points. [Results and Discussions] The experimental outcomes indicated that the enhanced DeepLabV3+ model had achieved an average Intersection over Union (IoU) of 93.71% and an average pixel accuracy of 97.25% on the dataset of tea shoots. Compared to the original model based on Xception, there was a substantial decrease in the parameter count from 54.714 million to a mere 5.818 million, effectively accomplishing a significant lightweight redesign of the model. Further comparisons with other prevalent semantic segmentation networks revealed that the improved model exhibited remarkable advantages concerning pivotal metrics such as the number of parameters, training duration, and average IoU, highlighting its efficacy and precision in the domain of tea shoot recognition. This considerable decreased in parameter numbers not only facilitated a more resource-economical deployment but also led to abbreviated training periods, rendering the model highly suitable for real-time implementations amidst tea garden ecosystems. The elevated mean IoU and pixel accuracy attested to the model's capacity for precise demarcation and identification of tea shoots, even amidst intricate and varied datasets, demonstrating resilience and adaptability in pragmatic contexts. [Conclusions] This study effectively implements an efficient and accurate tea shoot recognition method through targeted model improvements and optimizations, furnishing crucial technical support for the practical application of intelligent tea picking robots. The introduction of lightweight DeepLabV3+ not only substantially enhances recognition speed and segmentation accuracy, but also mitigates hardware requirements, thereby promoting the practical application of intelligent picking technology in the tea industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agri-QA Net: Multimodal Fusion Large Language Model Architecture for Crop Knowledge Question-Answering System
    WU Huarui, ZHAO Chunjiang, LI Jingchen
    Smart Agriculture    2025, 7 (1): 1-10.   DOI: 10.12133/j.smartag.SA202411005
    Abstract334)   HTML74)    PDF(pc) (1010KB)(288)       Save

    [Objective] As agriculture increasingly relies on technological innovations to boost productivity and ensure sustainability, farmers need efficient and accurate tools to aid their decision-making processes. A key challenge in this context is the retrieval of specialized agricultural knowledge, which can be complex and diverse in nature. Traditional agricultural knowledge retrieval systems have often been limited by the modalities they utilize (e.g., text or images alone), which restricts their effectiveness in addressing the wide range of queries farmers face. To address this challenge, a specialized multimodal question-answering system tailored for cabbage cultivation was proposed. The system, named Agri-QA Net, integrates multimodal data to enhance the accuracy and applicability of agricultural knowledge retrieval. By incorporating diverse data modalities, Agri-QA Net aims to provide a holistic approach to agricultural knowledge retrieval, enabling farmers to interact with the system using multiple types of input, ranging from spoken queries to images of crop conditions. By doing so, it helps address the complexity of real-world agricultural environments and improves the accessibility of relevant information. [Methods] The architecture of Agri-QA Net was built upon the integration of multiple data modalities, including textual, auditory, and visual data. This multifaceted approach enables the system to develop a comprehensive understanding of agricultural knowledge, allowed the system to learn from a wide array of sources, enhancing its robustness and generalizability. The system incorporated state-of-the-art deep learning models, each designed to handle one specific type of data. Bidirectional Encoder Representations from Transformers (BERT)'s bidirectional attention mechanism allowed the model to understand the context of each word in a given sentence, significantly improving its ability to comprehend complex agricultural terminology and specialized concepts. The system also incorporated acoustic models for processing audio inputs. These models analyzed the spoken queries from farmers, allowing the system to understand natural language inputs even in noisy, non-ideal environments, which was a common challenge in real-world agricultural settings. Additionally, convolutional neural networks (CNNs) were employed to process images from various stages of cabbage growth. CNNs were highly effective in capturing spatial hierarchies in images, making them well-suited for tasks such as identifying pests, diseases, or growth abnormalities in cabbage crops. These features were subsequently fused in a Transformer-based fusion layer, which served as the core of the Agri-QA Net architecture. The fusion process ensured that each modality—text, audio, and image—contributes effectively to the final model's understanding of a given query. This allowed the system to provide more nuanced answers to complex agricultural questions, such as identifying specific crop diseases or determining the optimal irrigation schedules for cabbage crops. In addition to the fusion layer, cross-modal attention mechanisms and domain-adaptive techniques were incorporated to refine the model's ability to understand and apply specialized agricultural knowledge. The cross-modal attention mechanism facilitated dynamic interactions between the text, audio, and image data, ensuring that the model paid attention to the most relevant features from each modality. Domain-adaptive techniques further enhanced the system's performance by tailoring it to specific agricultural contexts, such as cabbage farming, pest control, or irrigation management. [Results and Discussions] The experimental evaluations demonstrated that Agri-QA Net outperforms traditional single-modal or simple multimodal models in agricultural knowledge tasks. With the support of multimodal inputs, the system achieved an accuracy rate of 89.5%, a precision rate of 87.9%, a recall rate of 91.3%, and an F1-Score of 89.6%, all of which are significantly higher than those of single-modality models. The integration of multimodal data significantly enhanced the system's capacity to understand complex agricultural queries, providing more precise and context-aware answers. The addition of cross-modal attention mechanisms enabled for more nuanced and dynamic interaction between the text, audio, and image data, which in turn improved the model's understanding of ambiguous or context-dependent queries, such as disease diagnosis or crop management. Furthermore, the domain-adaptive technique enabled the system to focus on specific agricultural terminology and concepts, thereby enhancing its performance in specialized tasks like cabbage cultivation and pest control. The case studies presented further validated the system's ability to assist farmers by providing actionable, domain-specific answers to questions, demonstrating its practical application in real-world agricultural scenarios. [Conclusions] The proposed Agri-QA Net framework is an effective solution for addressing agricultural knowledge questions, especially in the domain of cabbage cultivation. By integrating multimodal data and leveraging advanced deep learning techniques, the system demonstrates a high level of accuracy and adaptability. This study not only highlights the potential of multimodal fusion in agriculture but also paves the way for future developments in intelligent systems designed to support precision farming. Further work will focus on enhancing the model's performance by expanding the dataset to include more diverse agricultural scenarios, refining the handling of dialectical variations in audio inputs, and improving the system's ability to detect rare crop diseases. The ultimate goal is to contribute to the modernization of agricultural practices, offering farmers more reliable and effective tools to solve the challenges in crop management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Intelligent Decision-Making Method for Personalized Vegetable Crop Water and Fertilizer Management Based on Large Language Models
    WU Huarui, LI Jingchen, YANG Yusen
    Smart Agriculture    2025, 7 (1): 11-19.   DOI: 10.12133/j.smartag.SA202410007
    Abstract329)   HTML35)    PDF(pc) (1181KB)(244)       Save

    [Objective] The current crop management faces the challenges of difficulty in capturing personalized needs and the lack of flexibility in the decision-making process. To address the limitations of conventional precision agriculture systems, optimize key aspects of agricultural production, including crop yield, labor efficiency, and water and fertilizer use, while ensure sustainability and adaptability to diverse farming conditions, in this research, an intelligent decision-making method was presents for personalized vegetable crop water and fertilizer management using large language model (LLM) by integrating user-specific preferences into decision-making processes through natural language interactions. [Methods] The method employed artificial intelligence techniques, combining natural language processing (NLP) and reinforcement learning (RL). Initially, LLM engaged users through structured dialogues to identify their unique preferences related to crop production goals, such as maximizing yield, reducing resource consumption, or balancing multiple objectives. These preferences were then modeled as quantifiable parameters and incorporated into a multi-objective optimization framework. To realize this framework, proximal policy optimization (PPO) was applied within a reinforcement learning environment to develop dynamic water and fertilizer management strategies. Training was conducted in the gym-DSSAT simulation platform, a system designed for agricultural decision support. The RL model iteratively learned optimal strategies by interacting with the simulation environment, adjusting to diverse conditions and balancing conflicting objectives effectively. To refine the estimation of user preferences, the study introduced a two-phase process comprising prompt engineering to guide user responses and adversarial fine-tuning for enhanced accuracy. These refinements ensured that user inputs were reliably transformed into structured decision-making criteria. Customized reward functions were developed for RL training to address specific agricultural goals. The reward functions account for crop yield, resource efficiency, and labor optimization, aligning with the identified user priorities. Through iterative training and simulation, the system dynamically adapted its decision-making strategies to varying environmental and operational conditions. [Results and Discussions] The experimental evaluation highlighted the system's capability to effectively personalize crop management strategies. Using simulations, the method demonstrated significant improvements over traditional approaches. The LLM-based model accurately captured user-specific preferences through structured natural language interactions, achieving reliable preference modeling and integration into the decision-making process. The system's adaptability was evident in its ability to respond dynamically to changes in user priorities and environmental conditions. For example, in scenarios emphasizing resource conservation, water and fertilizer use were significantly reduced without compromising crop health. Conversely, when users prioritized yield, the system optimized irrigation and fertilization schedules to enhance productivity. These results showcased the method's flexibility and its potential to balance competing objectives in complex agricultural settings. Additionally, the integration of user preferences into RL-based strategy development enabled the generation of tailored management plans. These plans aligned with diverse user goals, including maximizing productivity, minimizing resource consumption, and achieving sustainable farming practices. The system's multi-objective optimization capabilities allowed it to navigate trade-offs effectively, providing actionable insights for decision-making. The experimental validation also demonstrated the robustness of the PPO algorithm in training the RL model. The system's strategies were refined iteratively, resulting in consistent performance improvements across various scenarios. By leveraging LLM to capture nuanced user preferences and combining them with RL for adaptive decision-making, the method bridges the gap between generic precision agriculture solutions and personalized farming needs. [Conclusions] This study established a novel framework for intelligent decision-making in agriculture, integrating LLM with reinforcement learning to address personalized crop management challenges. By accurately capturing user-specific preferences and dynamically adapting to environmental and operational variables, the method offers a transformative approach to optimizing agricultural productivity and sustainability. Future work will focus on expanding the system's applicability to a wider range of crops and environmental contexts, enhancing the interpretability of its decision-making processes, and facilitating integration with real-world agricultural systems. These advancements aim to further refine the precision and impact of intelligent agricultural decision-making systems, supporting sustainable and efficient farming practices globally.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Dense Nursery Stock Detecting and Counting Based on UAV Aerial Images and Improved LSC-CNN
    PENG Xiaodan, CHEN Fengjun, ZHU Xueyan, CAI Jiawei, GU Mengmeng
    Smart Agriculture    2024, 6 (5): 88-97.   DOI: 10.12133/j.smartag.SA202404011
    Abstract323)   HTML43)    PDF(pc) (2507KB)(372)       Save

    [Objective] The number, location, and crown spread of nursery stock are important foundations data for their scientific management. Traditional approach of conducting nursery stock inventories through on-site individual plant surveys is labor-intensive and time-consuming. Low-cost and convenient unmanned aerial vehicles (UAVs) for on-site collection of nursery stock data are beginning to be utilized, and the statistical analysis of nursery stock information through technical means such as image processing achieved. During the data collection process, as the flight altitude of the UAV increases, the number of trees in a single image also increases. Although the anchor box can cover more information about the trees, the cost of annotation is enormous in the case of a large number of densely populated tree images. To tackle the challenges of tree adhesion and scale variance in images captured by UAVs over nursery stock, and to reduce the annotation costs, using point-labeled data as supervisory signals, an improved dense detection and counting model was proposed to accurately obtain the location, size, and quantity of the targets. [Method] To enhance the diversity of nursery stock samples, the spruce dataset, the Yosemite, and the KCL-London publicly available tree datasets were selected to construct a dense nursery stock dataset. A total of 1 520 nursery stock images were acquired and divided into training and testing sets at a ratio of 7:3. To enhance the model's adaptability to tree data of different scales and variations in lighting, data augmentation methods such as adjusting the contrast and resizing the images were applied to the images in the training set. After enhancement, the training set consists of 3 192 images, and the testing set contains 456 images. Considering the large number of trees contained in each image, to reduce the cost of annotation, the method of selecting the center point of the trees was used for labeling. The LSC-CNN model was selected as the base model. This model can detect the quantity, location, and size of trees through point-supervised training, thereby obtaining more information about the trees. The LSC-CNN model was made improved to address issues of missed detections and false positives that occurred during the testing process. Firstly, to address the issue of missed detections caused by severe adhesion of densely packed trees, the last convolutional layer of the feature extraction network was replaced with dilated convolution. This change enlarges the receptive field of the convolutional kernel on the input while preserving the detailed features of the trees. So the model is better able to capture a broader range of contextual information, thereby enhancing the model's understanding of the overall scene. Secondly, the convolutional block attention module (CBAM) attention mechanism was introduced at the beginning of each scale branch. This allowed the model to focus on the key features of trees at different scales and spatial locations, thereby improving the model's sensitivity to multi-scale information. Finally, the model was trained using label smooth cross-entropy loss function and grid winner-takes-all strategy, emphasizing regions with highest losses to boost tree feature recognition. [Results and Discussions] The mean counting accuracy (MCA), mean absolute error (MAE), and root mean square error (RMSE) were adopted as evaluation metrics. Ablation studies and comparative experiments were designed to demonstrate the performance of the improved LSC-CNN model. The ablation experiment proved that the improved LSC-CNN model could effectively resolve the issues of missed detections and false positives in the LSC-CNN model, which were caused by the density and large-scale variations present in the nursery stock dataset. IntegrateNet, PSGCNet, CANet, CSRNet, CLTR and LSC-CNN models were chosen as comparative models. The improved LSC-CNN model achieved MCA, MAE, and RMSE of 91.23%, 14.24, and 22.22, respectively, got an increase in MCA by 6.67%, 2.33%, 6.81%, 5.31%, 2.09% and 2.34%, respectively; a reduction in MAE by 21.19, 11.54, 18.92, 13.28, 11.30 and 10.26, respectively; and a decrease in RMSE by 28.22, 28.63, 26.63, 14.18, 24.38 and 12.15, respectively, compared to the IntegrateNet, PSGCNet, CANet, CSRNet, CLTR and LSC-CNN models. These results indicate that the improved LSC-CNN model achieves high counting accuracy and exhibits strong generalization ability. [Conclusions] The improved LSC-CNN model integrated the advantages of point supervision learning from density estimation methods and the generation of target bounding boxes from detection methods.These improvements demonstrate the enhanced performance of the improved LSC-CNN model in terms of accuracy, precision, and reliability in detecting and counting trees. This study could hold practical reference value for the statistical work of other types of nursery stock.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Reconstruction of U.S. Regional-Scale Soybean SIF Based on MODIS Data and BP Neural Network
    YAO Jianen, LIU Haiqiu, YANG Man, FENG Jinying, CHEN Xiu, ZHANG Peipei
    Smart Agriculture    2024, 6 (5): 40-50.   DOI: 10.12133/j.smartag.SA202309006
    Abstract319)   HTML44)    PDF(pc) (1648KB)(1970)       Save

    [Objective] Sunlight-induced chlorophyll fluorescence (SIF) data obtained from satellites suffer from issues such as low spatial and temporal resolution, and discrete footprint because of the limitations imposed by satellite orbits. To address these problems, obtaining higher resolution SIF data, most reconstruction studies are based on low-resolution satellite SIF. Moreover, the spatial resolution of most SIF reconstruction products is still not enough to be directly used for the study of crop photosynthetic rate at the regional scale. Although some SIF products boast elevated resolutions, but these derive not from the original satellite SIF data reconstruct but instead evolve from secondary reconstructions based on preexisting SIF reconstruction products. Satellite OCO-2 (The Orbiting Carbon Obsevatory-2) equipped with a high-resolution spectrometer, OCO-2 SIF has higher spatial resolution (1.29×2.25 km) compared to other original SIF products, making it suitable in advancing the realm of high-resolution SIF data reconstruction, particularly within the context of regional-scale crop studies. [Methods] This research primarily exploration SIF reconstruct at the regional scale, mainly focused on the partial soybean planting regions nestled within the United States. The selection of MODIS raw data hinged on a meticulous consideration of environmental conditions, the distinctive physiological attributes of soybeans, and an exhaustive evaluation of factors intricately linked to OCO-2 SIF within these soybean planting regions. The primary tasks of this research encompassed reconstructing high resolution soybean SIF while concurrently executing a rigorous assessment of the reconstructed SIF's quality. During the dataset construction process, amalgamated SIF data from multiple soybean planting regions traversed by the OCO-2 satellite's footprint to retain as many of the available original SIF samples as possible. This approach provided the subsequent SIF reconstruction model with a rich source of SIF data. SIF data obtained beneath the satellite's trajectory were matched with various MODIS datasets, including enhanced vegetation index (EVI), fraction of photosynthetically active radiation (FPAR), and land surface temperature (LST), resulting in the creation of a multisource remote sensing dataset ultimately used for model training. Because of the multisource remote sensing dataset encompassed the most relevant explanatory variables within each SIF footprint coverage area concerning soybean physiological structure and environmental conditions. Through the activation functions in the BP neural network, it enhanced the understanding of the complex nonlinear relationships between the original SIF data and these MODIS products. Leveraging these inherent nonlinear relationships, compared and analyzed the effects of different combinations of explanatory variables on SIF reconstruction, mainly analyzing the three indicators of goodness of fit R2, root mean square error RMSE, and mean absolute error MAE, and then selecting the best SIF reconstruction model, generate a regional scale, spatially continuous, and high temporal resolution (500 m, 8 d) soybean SIF reconstruction dataset (BPSIF). [Results and Discussions] The research findings confirmed the strong performance of the SIF reconstruction model in predicting soybean SIF. After simultaneously incorporating EVI, FPAR, and LST as explanatory variables to model, achieved a goodness of fit with an R2 value of 0.84, this statistical metric validated the model's capability in predicting SIF data, it also reflected that the reconstructed 8 d time resolution of SIF data's reliability of applying to small-scale agricultural crop photosynthesis research with 500 m×500 m spatial scale. Based on this optimal model, generated the reconstructed SIF product (BPSIF). The Pearson correlation coefficient between the original OCO-2 SIF data and MODIS GPP stood were at a modest 0.53. In stark contrast, the correlation coefficient between BPSIF and MODIS Gross Primary Productivity (GPP) rosed significantly to 0.80. The increased correlation suggests that BPSIF could more accurately reflect the dynamic changes in GPP during the soybean growing season, making it more reliable compared to the original SIF data. Selected soybean planting areas in the United States with relatively single crop cultivation as the research area, based on high spatial resolution (1.29 km×2.25 km) OCO-2 SIF data, greatly reduced vegetation heterogeneity under a single SIF footprint. [Conclusions] The BPSIF proposed has significantly enhancing the regional and temporal continuity of OCO-2 SIF while preserving the time and spatial attributes contained in the original SIF dataset. Within the study area, BPSIF exhibits a significantly improved correlation with MODIS GPP compared to the original OCO-2 SIF. The proposed OCO-2 SIF data reconstruction method in this study holds the potential to provide a more reliable SIF dataset. This dataset has the potential to drive further understanding of soybean SIF at finer spatial and temporal scales, as well as find its relationship with soybean GPP.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Price Game Model and Competitive Strategy of Agricultural Products Retail Market in the Context of Blockchain
    XUE Bing, SUN Chuanheng, LIU Shuangyin, LUO Na, LI Jinhui
    Smart Agriculture    2024, 6 (4): 160-173.   DOI: 10.12133/j.smartag.SA202309027
    Abstract312)   HTML21)    PDF(pc) (1265KB)(582)       Save

    [Objective] In the retail market for agricultural products, consumers are increasingly concerned about the safety and health aspects of those products. Traceability of blockchain has emerged as a crucial solution to address these concerns. Essentially, a blockchain functions as a dynamic, distributed, and shared database. When implemented in the agricultural supply chain, it not only improves product transparency to attract more consumers but also raises concerns about consumer privacy disclosure. The level of consumer apprehension regarding privacy will directly influence their choice to purchase agricultural products traced through blockchain-traced. Moreover, retailers' choices to sell blockchain-traced produce are influenced by consumer privacy concerns. By analyzing the impact of blockchain technology on the competitive strategies, pricing, and decision-making among agricultural retailers, they can develop market competition strategies that suit their market conditions to bolster their competitiveness and optimize the agricultural supply chain to maximize overall benefits. [Methods] Based on Nash equilibrium and Stackelberg game theory, a market competition model was developed to analyze the interactions between existing and new agricultural product retailers. The competitive strategies adopted by agricultural product retailers were analyzed under four different options of whether two agricultural retailers sell blockchain agricultural products. It delved into product utility, optimal pricing, demand, and profitability for each retailer under these different scenarios. How consumer privacy concerns impact pricing and profits of two agricultural product retailers and the optimal response strategy choice of another retailer when the competitor made the decision choice first were also analyzed. This analysis aimed to guide agricultural product retailers in making strategic choices that would safeguard their profits and market positions. To address the cooperative game problem of agricultural product retailers in market competition, ensure that retailers could better cooperate in the game, blockchain smart contract technology was used. By encoding the process and outcomes of the Stackelberg game into smart contracts, retailers could input their specific variables and receive tailored strategy recommendations. Uploading game results onto the blockchain network ensured transparency and encouraged cooperative behavior among retailers. By using the characteristics of blockchain, the game results were uploaded to the blockchain network to regulate the cooperative behavior, to ensure the maximization of the overall interests of the supply chain. [Results and Discussions] The research highlighted the significant improvement in agricultural product quality transparency through blockchain traceability technology. However, concerns regarding consumer privacy arising from this traceability could directly impact the pricing, profitability and retailers' decisions to provide blockchain-traceable items. Furthermore, an analysis of the strategic balance between two agricultural product retailers revealed that in situations of low and high product information transparency, both retailers were inclined to simultaneously offer sell traceable products. In such a scenario, blockchain traceability technology enhanced the utility and profitability of retail agricultural products, leading consumers to prefer purchase these traceable products from retailers. In cases where privacy concerns and agricultural product information transparency were both moderate, the initial retailer was more likely to opt for blockchain-based traceable products. This was because consumers had higher trust in the initial retailer, enabling them to bear a higher cost associated with privacy concerns. Conversely, new retailers failed to gain a competitive advantage and eventually exit the market. When consumer privacy concerns exceeded a certain threshold, both competing agricultural retailers discovered that offering blockchain-based traceable products led to a decline in their profits. [Conclusions] When it comes to agricultural product quality and safety, incorporating blockchain technology in traceability significantly improves the transparency of quality-related information for agricultural products. However, it is important to recognize that the application of blockchain for agricultural product traceability is not universally suitable for all agricultural retailers. Retailers must evaluate their unique circumstances and make the most suitable decisions to enhance the effectiveness of agricultural products, drive sales demand, and increase profits. Within the competitive landscape of the agricultural product retail market, nurturing a positive collaborative relationship is essential to maximize mutual benefits and optimize the overall profitability of the agricultural product supply chain.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Dynamic Prediction Method for Carbon Emissions of Cold Chain Distribution Vehicle under Multi-Source Information Fusion
    YANG Lin, LIU Shuangyin, XU Longqin, HE Min, SHENG Qingfeng, HAN Jiawei
    Smart Agriculture    2024, 6 (4): 138-148.   DOI: 10.12133/j.smartag.SA202403020
    Abstract304)   HTML15)    PDF(pc) (2240KB)(691)       Save

    [Objective] The dynamic prediction of carbon emission from cold chain distribution is an important basis for the accurate assessment of carbon emission and its green credit grade. Facing the fact that the carbon emission of vehicles is affected by multiple factors, such as road condition information, driving characteristics, refrigeration parameters, etc., a dynamic prediction model of carbon emission was proposed from refrigerated vehicles that integrates multi-source information. [Methods] The backbone feature extraction network, neck feature fusion network and loss function of YOLOv8s was firstly improved. The full-dimensional dynamic convolution was introduced into the backbone feature extraction network, and the multidimensional attention mechanism was introduced to capture the contextual key information to improve the model feature extraction capability. A progressive feature pyramid network was introduced into the feature extraction network, which reduced the loss of key information by fusing features layer by layer and improved the feature fusion efficiency. The road condition information recognition model based on improved YOLOv8s was constructed to characterize the road condition information in terms of the number of road vehicles and the percentage of pixel area. Pearson's correlation coefficient was used to compare and analyze the correlation between carbon emissions of refrigerated vehicles and different influencing factors, and to verify the necessity and criticality of the selection of input parameters of the carbon emission prediction model. Then the iTransformer temporal prediction model was improved, and the external attention mechanism was introduced to enhance the feature extraction ability of iTransformer model and reduce the computational complexity. The dynamic prediction model of carbon emission of refrigerated vehicles based on the improved iTransformer was constructed by taking the road condition information, driving characteristics (speed, acceleration), cargo weight, and refrigeration parameters (temperature, power) as inputs. Finally, the model was compared and analyzed with other models to verify the robustness of the road condition information and the prediction accuracy of the vehicle carbon emission dynamic prediction model, respectively. [Results and Discussions] The results of correlation analysis showed that the vehicle driving parameters were the main factor affecting the intensity of vehicle carbon emissions, with a correlation of 0.841. The second factor was cargo weight, with a correlation of 0.807, which had a strong positive correlation. Compared with the vehicle refrigeration parameters, the road condition information had a stronger correlation between vehicle carbon emissions, the correlation between refrigeration parameters and the vehicle carbon emissions impact factor were above 0.67. In order to further ensure the accuracy of the vehicle carbon emissions prediction model, The paper was selected as the input parameters for the carbon emissions prediction model. The improved YOLOv8s road information recognition model achieved 98.1%, 95.5%, and 98.4% in precision, recall, and average recognition accuracy, which were 1.2%, 3.7%, and 0.2% higher than YOLOv8s, respectively, with the number of parameters and the amount of computation being reduced by 12.5% and 31.4%, and the speed of detection being increased by 5.4%. This was due to the cross-dimensional feature learning through full-dimensional dynamic convolution, which fully captured the key information and improved the feature extraction capability of the model, and through the progressive feature pyramid network after fusing the information between different classes through gradual step-by-step fusion, which fully retained the important feature information and improved the recognition accuracy of the model. The predictive performance of the improved iTransformer carbon emission prediction model was better than other time series prediction models, and its prediction curve was closest to the real carbon emission curve with the best fitting effect. The introduction of the external attention mechanism significantly improved the prediction accuracy, and its MSE, MAE, RMSE and R2 were 0.026 1 %VOL, 0.079 1 %VOL, 0.161 5 %VOL and 0.940 0, respectively, which were 0.4%, 15.3%, 8.7% and 1.3% lower, respectively, when compared with iTransformer. As the degree of road congestion increased, the prediction accuracy of the constructed carbon emission prediction model increased. [Conclusions] The carbon emission prediction model for cold chain distribution under multi-source information fusion proposed in this study can realize accurate prediction of carbon emission from refrigerated vehicles, provide theoretical basis for rationally formulating carbon emission reduction strategies and promoting the development of low-carbon cold chain distribution.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight YOLOv8s-Based Strawberry Plug Seedling Grading Detection and Localization via Channel Pruning
    CHEN Junlin, ZHAO Peng, CAO Xianlin, NING Jifeng, YANG Shuqin
    Smart Agriculture    2024, 6 (6): 132-143.   DOI: 10.12133/j.smartag.SA202408001
    Abstract301)   HTML53)    PDF(pc) (3614KB)(646)       Save

    [Objective] Plug tray seedling cultivation is a contemporary method known for its high germination rates, uniform seedling growth, shortened transplant recovery period, diminished pest and disease incidence, and enhanced labor efficiency. Despite these advantages, challenges such as missing or underdeveloped seedlings can arise due to seedling quality and environmental factors. To ensure uniformity and consistency of the seedlings, sorting is frequently necessary, and the adoption of automated seedling sorting technology can significantly reduce labor costs. Nevertheless, the overgrowth of seedlings within the plugs can effect the accuracy of detection algorithms. A method for grading and locating strawberry seedlings based on a lightweight YOLOv8s model was presented in this research to effectively mitigate the interference caused by overgrown seedlings. [Methods] The YOLOv8s model was selected as the baseline for detecting different categories of seedlings in the strawberry plug tray cultivation process, namely weak seedlings, normal seedlings, and plug holes. To improve the detection efficiency and reduce the model's computational cost, the layer-adaptive magnitude-based pruning(LAMP) score-based channel pruning algorithm was applied to compress the base YOLOv8s model. The pruning procedure involved using the dependency graph to derive the group matrices, followed by normalizing the group importance scores using the LAMP Score, and ultimately pruning the channels according to these processed scores. This pruning strategy effectively reduced the number of model parameters and the overall size of the model, thereby significantly enhancing its inference speed while maintaining the capability to accurately detect both seedlings and plug holes. Furthermore, a two-stage seedling-hole matching algorithm was introduced based on the pruned YOLOv8s model. In the first stage, seedling and plug hole bounding boxes were matched according to their the degree of overlap (Dp), resulting in an initial set of high-quality matches. This step helped minimize the number of potential matching holes for seedlings exhibiting overgrowth. Subsequently, before the second stage of matching, the remaining unmatched seedlings were ranked according to their potential matching hole scores (S), with higher scores indicating fewer potential matching holes. The seedlings were then prioritized during the second round of matching based on these scores, thus ensuring an accurate pairing of each seedling with its corresponding plug hole, even in cases where adjacent seedling leaves encroached into neighboring plug holes. [Results and Discussions] The pruning process inevitably resulted in the loss of some parameters that were originally beneficial for feature representation and model generalization. This led to a noticeable decline in model performance. However, through meticulous fine-tuning, the model's feature expression capabilities were restored, compensating for the information loss caused by pruning. Experimental results demonstrated that the fine-tuned model not only maintained high detection accuracy but also achieved significant reductions in FLOPs (86.3%) and parameter count (95.4%). The final model size was only 1.2 MB. Compared to the original YOLOv8s model, the pruned version showed improvements in several key performance metrics: precision increased by 0.4%, recall by 1.2%, mAP by 1%, and the F1-Score by 0.1%. The impact of the pruning rate on model performance was found to be non-linear. As the pruning rate increased, model performance dropped significantly after certain crucial channels were removed. However, further pruning led to a reallocation of the remaining channels' weights, which in some cases allowed the model to recover or even exceed its previous performance levels. Consequently, it was necessary to experiment extensively to identify the optimal pruning rate that balanced model accuracy and speed. The experiments indicated that when the pruning rate reached 85.7%, the mAP peaked at 96.4%. Beyond this point, performance began to decline, suggesting that this was the optimal pruning rate for achieving a balance between model efficiency and performance, resulting in a model size of 1.2 MB. To further validate the improved model's effectiveness, comparisons were conducted with different lightweight backbone networks, including MobileNetv3, ShuffleNetv2, EfficientViT, and FasterNet, while retaining the Neck and Head modules of the original YOLOv8s model. Results indicated that the modified model outperformed these alternatives, with mAP improvements of 1.3%, 1.8%, 1.5%, and 1.1%, respectively, and F1-Score increases of 1.5%, 1.8%, 1.1%, and 1%. Moreover, the pruned model showed substantial advantages in terms of floating-point operations, model size, and parameter count compared to these other lightweight networks. To verify the effectiveness of the proposed two-stage seedling-hole matching algorithm, tests were conducted using a variety of complex images from the test set. Results indicated that the proposed method achieved precise grading and localization of strawberry seedlings even under challenging overgrowth conditions. Specifically, the correct matching rate for normal seedlings reached 96.6%, for missing seedlings 84.5%, and for weak seedlings 82.9%, with an average matching accuracy of 88%, meeting the practical requirements of the strawberry plug tray cultivation process. [Conclusions] The pruned YOLOv8s model successfully maintained high detection accuracy while reducing computational costs and improving inference speed. The proposed two-stage seedling-hole matching algorithm effectively minimized the interference caused by overgrown seedlings, accurately locating and classifying seedlings of various growth stages within the plug tray. The research provides a robust and reliable technical solution for automated strawberry seedling sorting in practical plug tray cultivation scenarios, offering valuable insights and technical support for optimizing the efficiency and precision of automated seedling grading systems.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Design and Test of Dust Removal Seeding Rate Monitoring System for Rapeseed Seeders
    LI Qiang, YU Qiuli, LI Haopeng, XU Chunbao, DING Youchun
    Smart Agriculture    2024, 6 (3): 107-117.   DOI: 10.12133/j.smartag.SA202401011
    Abstract300)   HTML11)    PDF(pc) (1747KB)(354)       Save

    [Objective] The pneumatic rapeseed seeder easily inhales the dust generated during live broadcast operations into the seeding pipe, and then releases it along with the rapeseed. Consequently, when monitoring the rapeseed flow, dust interference can affect the flow, detection sensitivity and accuracy. This research aims to develop a dust removal rapeseed seeding rate monitoring system suitable for air-assisted pneumatic rapeseed seeders to improve the transparency and intelligence of the sowing process. [Methods] The monitoring system comprised a dust removal rapeseed seed flow detection device and a sowing monitoring terminal, which could be adjusted for different seeders widths by altering the number of detection devices. The rapeseed stream sensing structure operated on the principle of photoelectric induction. A delicate light layer was generated using an LED light source and a narrow slit structure. The convex lens condenses the light and directed it onto the sensing area of the silicon photovoltaic cell. When the rapeseed seeds pass through the sensing light layer, the silicon photovoltaic cell produced a voltage change signal. The signal converts it into a pulse signal that can be recognized by the microcontroller. A dust removal mechanism was designed by analyzing dust sources in the seeding system during normal field operation of the air-assisted rapeseed seeding machine and understanding the impact mechanism of the dust detection device on the accuracy of rapeseed flow monitoring. This mechanism employed a transparent plate to protect the photoelectric induction device in a relatively enclosed space and used a stepper motor screw mechanism to generate friction between the transparent plate and the dust removal cloth for effective dust removal. The appropriate size of the dust shield was determined by comparing its movement stroke with other structural dimensions of the detection device. The relationship between the silicon photocells voltage and detection accuracy was established through experiments at seeding frequencies of 10‒40 Hz. To ensure that the real-time detection accuracy was not less than 90%, the dust removal control threshold was set at 82% of the initial voltage value. In order to prevent congestion and data loss during data transmission and improve the scalability and compatibility of the monitoring system, data transmission between the detection device and the monitoring terminal was implemented based on the CAN2.0A communication protocol. The structural framework and monitoring terminal functions of the rapeseed sowing monitoring system were outlined. Software functions of the detection device were designed to meet the dust removal, communication, and rapeseed flow detection needs. The program execution process of the detection device was explained. In order to provide data support for the dust flow rate that should be controlled at various seeding frequencies during the bench test, experiments were conducted in the field to obtain theoretical data. [Results and Discussions] The comparison bench test of the detection device indicates that with the average seeding frequency ranging from 12.4 to 36.3 Hz and the average dust flow rate ranging from 252 to 386 mg/s, the detection accuracy after two dust removal cycles without a dustproof and dust removal detection device was not higher than 80.2%. The dust detection device with dust removal got an accuracy rate of not less than 90.2%, and the average detection accuracy rate within a single dust removal cycle was not less than 93.6%. The seeding amount monitoring bench test showed that when the average seeding frequency was no higher than 37.6 Hz, the seeding rate monitoring accuracy was not less than 92.2%. Furthermore, the field sowing experiment results demonstrated that at a normal operating speed (2.8‒4.6 km/h) of the rapeseed direct seeder, with a field sowing frequency of 14.8‒31.1 Hz, the accuracy of sowing quantity monitoring was not less than 93.1%. [Conclusions] The rapeseed sowing quality monitoring system provides effective support for precise detection even when operating in dusty conditions with the pneumatic rapeseed direct seeder. In the future, by integrating positioning data, sowing information, and fertilization monitoring data through CAN bus technology, a comprehensive field sowing and fertilization status map can be created to further enhance the system's capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Tomato Leaf Disease and Pest Detection Method Based on Improved YOLOv10n
    WU Liuai, XU Xueke
    Smart Agriculture    2025, 7 (1): 146-155.   DOI: 10.12133/j.smartag.SA202410023
    Abstract285)   HTML62)    PDF(pc) (1834KB)(204)       Save

    [Objective] To address the challenges in detecting tomato leaf diseases and pests, such as complex environments, small goals, low precision, redundant parameters, and high computational complexity, a novel lightweight, high-precision, real-time detection model was proposed called YOLOv10n-YS. This model aims to accurately identify diseases and pests, thereby providing a solid scientific basis for their prevention and management strategies. Methods] The dataset was collected using mobile phones to capture images from multiple angles under natural conditions, ensuring complete and clear leaf images. It included various weather conditions and covered nine types: Early blight, leaf mold, mosaic virus, septoria, spider mites damage, yellow leaf curl virus, late blight, leaf miner disease, and healthy leaves, with all images having a resolution of 640×640 pixels. In the proposed YOLOv10n-YS model, firstly, the C2f in the backbone network was replaced with C2f_RepViTBlock, thereby reducing the computational load and parameter volume and achieving a lightweight design. Secondly, through the introduction of a sliced operation SimAM attention mechanism, the Conv_SWS module was formed, which enhanced the extraction of small target features. Additionally, the DySample lightweight dynamic up sampling module was used to replace the up sampling module in the neck network, concentrating sampling points on target areas and ignoring backgrounds, thereby effectively identifying defects. Finally, the efficient channel attention (ECA) was improved by performing average pooling and max pooling on the input layer to aggregate features and then adding them together, which further enhanced global perspective information and features of different scales. The improved module, known as efficient channel attention with cross-channel interaction (EMCA) attention, was introduced, and the pyramid spatial attention (PSA) in the backbone network was replaced with the EMCA attention mechanism, thereby enhancing the feature extraction capability of the backbone network. [Results and Discussions] After introducing the C2f_RepViTBlock, the model's parameter volume and computational load were reduced by 12.3% and 9.7%, respectively, with mAP@0.5 and F1-Score each increased by 0.2% and 0.3%. Following the addition of the Conv_SWS and the replacement of the original convolution, mAP@0.5 and F1-Score were increased by 1.2% and 2%, respectively, indicating that the Conv_SWS module significantly enhanced the model's ability to extract small target features. After the introduction of DySample, mAP@0.5 and F1-Score were increased by 1.8% and 2.6%, respectively, but with a slight increase in parameter volume and computational load. Finally, the addition of the EMCA attention mechanism further enhanced the feature extraction capability of the backbone network. Through these four improvements, the YOLOv10n-YS model was formed. Compared with the YOLOv10n algorithm, YOLOv10n-YS reduced parameter volume and computational load by 13.8% and 8.5%, respectively, with both mAP@0.5 and F1-Score increased. These improvements not only reduced algorithm complexity but also enhanced detection accuracy, making it more suitable for industrial real-time detection. The detection accuracy of tomato diseases and pests using the YOLOv10n-YS algorithm was significantly better than that of comparative algorithms, and it had the lowest model parameter volume and computational load. The visualization results of detection by different models showed that the YOLOv10n-YS network could provide technical support for the detection and identification of tomato leaf diseases and pests. To verify the performance and robustness of the YOLOv10n-YS algorithm, comparative experiments were conducted on the public Plant-Village-9 dataset with different algorithms. The results showed that the average detection accuracy of YOLOv10n-YS on the Plant-Village dataset reached 91.1%, significantly higher than other algorithms. [Conclusions] The YOLOv10n-YS algorithm is not only characterized by occupying a small amount of space but also by possessing high recognition accuracy. On the tomato leaf dataset, excellent performance was demonstrated by this algorithm, thereby verifying its broad applicability and showcasing its potential to play an important role in large-scale crop pest and disease detection applications. Deploying the model on drone platforms and utilizing multispectral imaging technology can achieve real-time detection and precise localization of pests and diseases in complex field environments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Detection Method of Apple Alternaria Leaf Spot Based on Deep-Semi-NMF
    FU Zhuojun, HU Zheng, DENG Yangjun, LONG Chenfeng, ZHU Xinghui
    Smart Agriculture    2024, 6 (6): 144-154.   DOI: 10.12133/j.smartag.SA202409001
    Abstract278)   HTML28)    PDF(pc) (1901KB)(137)       Save

    [Objective] Apple Alternaria leaf spot can easily lead to premature defoliation of apple tree leaves, thereby affecting the quality and yield of apples. Consequently, accurately detecting of the disease has become a critical issue in the precise prevention and control of apple tree diseases. Due to factors such as backlighting, traditional image segmentation-based methods for detecting disease spots struggle to accurately identify the boundaries of diseased areas against complex backgrounds. There is an urgent need to develop new methods for detecting apple Alternaria leaf spot, which can assist in the precise prevention and control of apple tree diseases. [Methods] A novel detection method named Deep Semi-Non-negative Matrix Factorization-based Mahalanobis Distance Anomaly Detection (DSNMFMAD) was proposed, which combines Deep Semi-Non-negative Matrix Factorization (DSNMF) with Mahalanobis distance for robust anomaly detection in complex image backgrounds. The proposed method began by utilizing DSNMF to extract low-rank background components and sparse anomaly features from the apple Alternaria leaf spot images. This enabled effective separation of the background and anomalies, mitigating interference from complex background noise while preserving the non-negativity constraints inherent in the data. Subsequently, Mahalanobis distance was employed, based on the Singular Value Decomposition (SVD) feature subspace, to construct a lesion detector. The detector identified lesions by calculating the anomaly degree of each pixel in the anomalous regions. The apple tree leaf disease dataset used was provided by PaddlePaddle AI-Studio. Each image in the dataset has a resolution of 512×512 pixels, in RGB color format, and was in JPEG format. The dataset was captured in both laboratory and natural environments. Under laboratory conditions, 190 images of apple leaves with spot-induced leaf drop were used, while 237 images were collected under natural conditions. Furthermore, the dataset was augmented with geometric transformations and random changes in brightness, contrast, and hue, resulting in 1 145 images under laboratory conditions and 1 419 images under natural conditions. These images reflect various real-world scenarios, capturing apple leaves at different stages of maturity, in diverse lighting conditions, angles, and noise environments. This diversed dataset ensured that the proposed method could be tested under a wide range of practical conditions, providing a comprehensive evaluation of its effectiveness in detecting apple Alternaria leaf spot. [Results and Discussions] DSNMFMAD demonstrated outstanding performance under both laboratory and natural conditions. A comparative analysis was conducted with several other detection methods, including GRX (Reed-Xiaoli detector), LRX (Local Reed-Xiaoli detector), CRD (Collaborative-Representation-Based Detector), LSMAD (LRaSMD-Based Mahalanobis Distance Detector), and the deep learning model Unet. The results demonstrated that DSNMFMAD exhibited superior performance in the laboratory environment. The results demonstrated that DSNMFMAD attained a recognition accuracy of 99.8% and a detection speed of 0.087 2 s/image. The accuracy of DSNMFMAD was found to exceed that of GRX, LRX, CRD, LSMAD, and Unet by 0.2%, 37.9%, 10.3%, 0.4%, and 24.5%, respectively. Additionally, the DSNMFMAD exhibited a substantially superior detection speed in comparison to LRX, CRD, LSMAD, and Unet, with an improvement of 8.864, 107.185, 0.309, and 1.565 s, respectively. In a natural environment, where a dataset of 1 419 images of apple Alternaria leaf spot was analysed, DSNMFMAD demonstrated an 87.8% recognition accuracy, with an average detection speed of 0.091 0 s per image. In this case, its accuracy outperformed that of GRX, LRX, CRD, LSMAD, and Unet by 2.5%, 32.7%, 5%, 14.8%, and 3.5%, respectively. Furthermore, the detection speed was faster than that of LRX, CRD, LSMAD, and Unet by 2.898, 132.017, 0.224, and 1.825 s, respectively. [Conclusions] The DSNMFMAD proposed in this study was capable of effectively extracting anomalous parts of an image through DSNMF and accurately detecting the location of apple Alternaria leaf spot using a constructed lesion detector. This method achieved higher detection accuracy compared to the benchmark methods, even under complex background conditions, demonstrating excellent performance in lesion detection. This advancement could provide a valuable technical reference for the detection and prevention of apple Alternaria leaf spot.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Rice Leaf Disease Image Enhancement Based on Improved CycleGAN
    YAN Congkuan, ZHU Dequan, MENG Fankai, YANG Yuqing, TANG Qixing, ZHANG Aifang, LIAO Juan
    Smart Agriculture    2024, 6 (6): 96-108.   DOI: 10.12133/j.smartag.SA202407019
    Abstract273)   HTML31)    PDF(pc) (1744KB)(205)       Save

    Objective Rice diseases significantly impact both the yield and quality of rice production. Automatic recognition of rice diseases using computer vision is crucial for ensuring high yields, quality, and efficiency. However, rice disease image recognition faces challenges such as limited availability of datasets, insufficient sample sizes, and imbalanced sample distributions across different disease categories. To address these challenges, a data augmentation method for rice leaf disease images was proposed based on an improved CycleGAN model in this reseach which aimed to expand disease image datasets by generating disease features, thereby alleviating the burden of collecting real disease data and providing more comprehensive and diverse data to support automatic rice disease recognition. Methods The proposed approach built upon the CycleGAN framework, with a key modification being the integration of a convolutional block attention module (CBAM) into the generator's residual module. This enhancement strengthened the network's ability to extract both local key features and global contextual information pertaining to rice disease-affected areas. The model increased its sensitivity to small-scale disease targets and subtle variations between healthy and diseased domains. This design effectively mitigated the potential loss of critical feature information during the image generation process, ensuring higher fidelity in the resulting images. Additionally, skip connections were introduced between the residual modules and the CBAM. These connections facilitate improved information flow between different layers of the network, addressing common issues such as gradient vanishing during the training of deep networks. Furthermore, a perception similarity loss function, designed to align with the human visual system, was incorporated into the overall loss function. This addition enabled the deep learning model to more accurately measure perceptual differences between the generated images and real images, thereby guiding the network towards producing higher-quality samples. This adjustment also helped to reduce visual artifacts and excessive smoothing, while concurrently improving the stability of the model during the training process. To comprehensively evaluate the quality of the rice disease images generated by the proposed model and to assess its impact on disease recognition performance, both subjective and objective evaluation metrics were utilized. These included user perception evaluation (UPE), structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), and the performance of disease recognition within object detection frameworks. Comparative experiments were conducted across multiple GAN models, enabling a thorough assessment of the proposed model's performance in generating rice disease images. Additionally, different attention mechanisms, including efficient channel attention (ECA), coordinate attention (CA), and CBAM, were individually embedded into the generator's residual module. These variations allowed for a detailed comparison of the effects of different attention mechanisms on network performance and the visual quality of the generated images. Ablation studies were further performed to validate the effectiveness of the CBAM residual module and the perception similarity loss function in the network's overall architecture. Based on the generated rice disease samples, transfer learning experiments were conducted using various object detection models. By comparing the performance of these models before and after transfer learning, the effectiveness of the generated disease image data in enhancing the performance of object detection models was empirically verified. Results and Discussions The rice disease images generated by the improved CycleGAN model surpassed those produced by other GAN variants in terms of image detail clarity and the prominence of disease-specific features. In terms of objective quality metrics, the proposed model exhibited a 3.15% improvement in SSIM and an 8.19% enhancement in PSNR compared to the original CycleGAN model, underscoring its significant advantage in structural similarity and signal-to-noise ratio. The comparative experiments involving different attention mechanisms and ablation studies revealed that embedding the CBAM into the generator effectively increased the network's focus on critical disease-related features, resulting in more realistic and clearly defined disease-affected regions in the generated images. Furthermore, the introduction of the perception similarity loss function substantially enhanced the network's ability to perceive and represent disease-related information, thereby improving the visual fidelity and realism of the generated images. Additionally, transfer learning applied to object detection models such as YOLOv5s, YOLOv7-tiny, and YOLOv8s led to significant improvements in disease detection performance on the augmented dataset. Notably, the detection accuracy of the YOLOv5s model increased from 79.7% to 93.8%, representing a considerable enhancement in both generalization ability and robustness. This improvement also effectively reduced the rates of false positives and false negatives, resulting in more stable and reliable performance in rice disease detection tasks. Conclusions The rice leaf disease image generation method based on the improved CycleGAN model, as proposed in this study, effectively transforms images of healthy leaves into those depicting disease symptoms. By addressing the challenge of insufficient disease samples, this method significantly improves the disease recognition capabilities of object detection models. Therefore, it holds considerable application potential in the domain of leaf disease image augmentation and offers a promising new direction for expanding datasets of disease images for other crops.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Prediction and Mapping of Soil Total Nitrogen Using GF-5 Image Based on Machine Learning Optimization Modeling
    LIU Liqi, WEI Guangyuan, ZHOU Ping
    Smart Agriculture    2024, 6 (5): 61-73.   DOI: 10.12133/j.smartag.SA202405011
    Abstract265)   HTML39)    PDF(pc) (3325KB)(572)       Save

    [Objective] Nitrogen in soil is an absolutely crucial element for plant growth. Insufficient nitrogen supply can severely affect crop yield and quality, while excessive use of nitrogen fertilizers can lead to significant environmental issues such as water eutrophication and groundwater pollution. Therefore, large-scale, rapid detection of soil nitrogen content and precise fertilization are of great importance for smart agriculture. In this study, the hyperspectral data from the GF-5 satellite was emploied, and the various machine learning algorithms introduced to establish a prediction model for soil total nitrogen (TN) content and a distribution map of soil TN content was generated in the study area, aiming to provide scientific evidence for intelligent monitoring in smart agriculture. [Method] The study area was the Jian Sanjiang Reclamation Area in Fujin city, Heilongjiang province. Fieldwork involved the careful collection of 171 soil samples, obtaining soil spectral data, chemical analysis data of soil TN content, and the GF-5 hyperspectral data. Among these samples, 140 were randomly selected as the modeling sample set for calibration, and the remaining 31 samples were used as the test sample set. Three machine learning algorithms were introduced: Partial least squares regression (PLSR), backpropagation neural network (BPNN), and support vector machine (SVM) driven by a polynomial kernel function (Poly). Three distinct soil TN inversion models were constructed using these algorithms. To optimize model performance, ten-fold cross-validation was employed to determine the optimal parameters for each model. Additionally, multiple scatter correction (MSC) was applied to obtain band characteristic values, thus enhancing the model's prediction capability. Model performance was evaluated using three indicators: Coefficient of determination (R²), root mean square error (RMSE), and relative prediction deviation (RPD), to assess the prediction accuracy of different models. [Results and Discussions] MSC-Poly-SVM model exhibited the best prediction performance on the test sample set, with an R² of 0.863, an RMSE of 0.203, and an RPD of 2.147. This model was used to perform soil TN content inversion mapping using GF-5 satellite hyperspectral data. In accordance with the stringent requirements of land quality geochemical evaluation, the GF-5 hyperspectral land organic nitrogen parameter distribution map was drawn based on the "Determination of Land Quality Geochemical Evaluation". The results revealed that 86.1% of the land in the Jian Sanjiang study area had a total nitrogen content of more than 2.0 g/kg, primarily concentrated in first and second-grade plots, while third and fourth-grade plots accounted for only 11.83% of the total area. The study area exhibited sufficient soil nitrogen reserves, with high TN background values mainly concentrated along the riverbanks in the central part, distributed in a northeast-east direction. Specifically, in terms of soil spectral preprocessing, the median filtering method performed best in terms of smoothness and maintaining spectral characteristics. The spectra extracted from GF-5 imagery were generally quite similar to ground-measured spectral data, despite some noise, which had a minimal overall impact. [Conclusions] This study demonstrates the clear feasibility of using GF-5 satellite hyperspectral remote sensing data and machine learning algorithm for large-scale quantitative detection and visualization analysis of soil TN content. The soil TN content distribution map generated based on GF-5 hyperspectral remote sensing data is detailed and consistent with results from other methods, providing technical support for future large-scale quantitative detection of soil nutrient status and rational fertilization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Suitable Sowing Date Method of Winter Wheat at the County Level Based on ECMWF Long-Term Reanalysis Data
    LIU Ruixuan, ZHANG Fangzhao, ZHANG Jibo, LI Zhenhai, YANG Juntao
    Smart Agriculture    2024, 6 (5): 51-60.   DOI: 10.12133/j.smartag.SA202309019
    Abstract264)   HTML20)    PDF(pc) (1500KB)(321)       Save

    [Objective] Acurately determining the suitable sowing date for winter wheat is of great significance for improving wheat yield and ensuring national food security. Traditional visual interpretation method is not only time-consuming and labor-intensive, but also covers a relatively small area. Remote sensing monitoring, belongs to post-event monitoring, exhibits a time lag. The aim of this research is to use the temperature threshold method and accumulated thermal time requirements for wheat leaves appearance method to analyze the suitable sowing date for winter wheat in county-level towns under the influence of long-term sequence of climate warming. [Methods] The research area were various townships in Qihe county, Shandong province. Based on European centre for medium-range weather forecasts (ECMWF) reanalysis data from 1997 to 2022, 16 meteorological data grid points in Qihe county were selected. Firstly, the bilinear interpolation method was used to interpolate the temperature data of grid points into the approximate center points of each township in Qihe county, and the daily average temperatures for each township were obtained. Then, temperature threshold method was used to determine the final dates of stable passage through 18, 16, 14 and 0 ℃. Key sowing date indicators such as suitable sowing temperature for different wheat varieties, growing degree days (GDD)≥0 ℃ from different sowing dates to before overwintering, and daily average temperature over the years were used for statistical analysis of the suitable sowing date for winter wheat. Secondly, the accumulated thermal time requirements for wheat leaves appearance method was used to calculate the appropriate date of GDD for strong seedlings before winter by moving forward from the stable date of dropping to 0 ℃. Accumulating the daily average temperatures above 0 ℃ to the date when the GDD above 0 ℃ was required for the formation of strong seedlings of wheat, a range of ±3 days around this calculated date was considered the theoretical suitable sowing date. Finally, combined with actual production practices, the appropriate sowing date of winter wheat in various townships of Qihe county was determined under the trend of climate warming. [Results and Discussions] The results showed that, from November 1997 to early December 2022, winter and annual average temperatures in Qihe county had all shown an upward trend, and there was indeed a clear trend of climate warming in various townships of Qihe county. Judging from the daily average temperature over the years, the temperature fluctuation range in November was the largest in a year, with a maximum standard deviation was 2.61 ℃. This suggested a higher likelihood of extreme weather conditions in November. Therefore, it was necessary to take corresponding measures to prevent and reduce disasters in advance to avoid affecting the growth and development of wheat. In extreme weather conditions, it was limited to determine the sowing date only by temperature or GDD. In cold winter years, it was too one-sided to consider only from the perspective of GDD. It was necessary to expand the range of GDD required for winter wheat before overwintering based on temperature changes to ensure the normal growth and development of winter wheat. The suitable sowing date for semi winter wheat obtained by temperature threshold method was from October 4th to October 16th, and the suitable sowing date for winter wheat was from September 27th to October 4th. Taking into account the GDD required for the formation of strong seedlings before winter, the suitable sowing date for winter wheat was from October 3rd to October 13th, and the suitable sowing date for semi winter wheat was from October 15th to October 24th, which was consisted with the suitable sowing date for winter wheat determined by the accumulated thermal time requirements for wheat leaves appearance method. Considering the winter wheat varieties planted in Qihe county, the optimal sowing date for winter wheat in Qihe county was from October 3rd to October 16th, and the optimal sowing date was from October 5th to October 13th. With the gradual warming of the climate, the suitable sowing date for wheat in various townships of Qihe county in 2022 was later than that in 2002. However, the sowing date for winter wheat was still influenced by factors such as soil moisture, topography, and seeding quality. The suitable sowing date for a specific year still needed to be adjusted to local conditions and flexibly sown based on the specific situation of that year. [Conclusions] The experimental results proved the feasibility of the temperature threshold method and accumulated thermal time requirements for wheat leaves appearance method in determining the suitable sowing date for winter wheat. The temperature trend can be used to identify cold or warm winters, and the sowing date can be adjusted in a timely manner to enhance wheat yield and reduce the impact of excessively high or low temperatures on winter wheat. The research results can not only provide decision-making reference for winter wheat yield assessment in Qihe county, but also provide an important theoretical basis for scientifically arrangement of agricultural production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grape Recognition and Localization Method Based on 3C-YOLOv8n and Depth Camera
    LIU Chang, SUN Yu, YANG Jing, WANG Fengchao, CHEN Jin
    Smart Agriculture    2024, 6 (6): 121-131.   DOI: 10.12133/j.smartag.SA202407008
    Abstract262)   HTML34)    PDF(pc) (2020KB)(583)       Save

    [Objective] Grape picking is a key link in increasing production. However, in this process, a large amount of manpower and material resources are required, which makes the picking process complex and slow. To enhance harvesting efficiency and achieve automated grape harvesting, an improved YOLOv8n object detection model named 3C-YOLOv8n was proposed, which integrates the RealSense D415 depth camera for grape recognition and localization. [Methods] The propoesed 3C-YOLOv8n incorporated a convolutional block attention module (CBAM) between the first C2f module and the third Conv module in the backbone network. Additionally, a channel attention (CA) module was added at the end of the backbone structure, resulting in a new 2C-C2f backbone network architecture. This design enabled the model to sequentially infer attention maps across two independent dimensions (channel and spatial), optimize features by considering relationships between channels and positional information. The network structure was both flexible and lightweight. Furthermore, the Content-aware ReAssembly of Features up sampling operator was implemented to support instance-specific kernels (such as deconvolution) for feature reconstruction with neighboring pixels, replacing the nearest neighbor interpolation operator in the YOLOv8n neck network. This enhancement increased the receptive field and guided the reconstruction process based on input features while maintaining low parameter and computational complexity, thereby forming the 3C-YOLOv8n model. The pyrealsense2 library was utilized to obtain pixel position information from the target area using the Intel RealSense D415 camera. During this process, the depth camera was used to capture images, and target detection algorithms were employed to pinpoint the location of grapes. The camera's depth sensor facilitated the acquisition of the three-dimensional point cloud of grapes, allowing for the calculation of the distance from the pixel point to the camera and the subsequent determination of the three-dimensional coordinates of the center of the target's bounding box in the camera coordinate system, thus achieving grape recognition and localization. [Results and Discussions] Comparative and ablation experiments were conducted. it was observed that the 3C-YOLOv8n model achieved a mean average precision (mAP) of 94.3% at an intersection ratio of 0.5 (IOU=0.5), surpassing the YOLOv8n model by 1%. The accuracy (P) and recall (R) rates were recorded at 91.6% and 86.4%, respectively, reflecting increases of 0.1% and 0.7%. The F1-Score also improved by 0.4%, demonstrating that the improved network model met the experimental accuracy and recall requirements. In terms of loss, the 3C-YOLOv8n algorithm exhibited superior performance, with a rapid decrease in loss values and minimal fluctuations, ultimately leading to a minimized loss value. This indicated that the improved algorithm quickly reached a convergence state, enhancing both model accuracy and convergence speed. The ablation experiments revealed that the original YOLOv8n model yielded a mAP of 93.3%. The integration of the CBAM and CA attention mechanisms into the YOLOv8n backbone resulted in mAP values of 93.5% each. The addition of the Content-aware ReAssembly of Features up sampling operator to the neck network of YOLOv8n produced a 0.5% increase in mAP, culminating in a value of 93.8%. The combination of the three improvement strategies yielded mAP increases of 0.3, 0.7, and 0.8%, respectively, compared to the YOLOv8n model. Overall, the 3C-YOLOv8n model demonstrated the best detection performance, achieving the highest mAP of 94.3%. The ablation results confirmed the positive impact of the proposed improvement strategies on the experimental outcomes. Compared to other mainstream YOLO series algorithms, all evaluation metrics showed enhancements, with the lowest missed detection and false detection rates among all tested algorithms, underscoring its practical advantages in detection tasks. [Conclusions] By effectively addressing the inefficiencies of manual labor, 3C-YOLOv8n network model not only enhances the precision of grape recognition and localization but also significantly optimizes overall harvesting efficiency. Its superior performance in evaluation metrics such as precision, recall, mAP, and F1-Score, alongside the lowest recorded loss values among YOLO series algorithms, indicates a remarkable advancement in model convergence and operational effectiveness. Furthermore, the model's high accuracy in grape target recognition not only lays the groundwork for automated harvesting systems but also enables the implementation of complementary intelligent operations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    ReluformerN: Lightweight High-Low Frequency Enhanced for Hyperspectral Agricultural Lancover Classification
    LIU Yi, ZHANG Yanjun
    Smart Agriculture    2024, 6 (5): 74-87.   DOI: 10.12133/j.smartag.SA202406008
    Abstract245)   HTML21)    PDF(pc) (3072KB)(639)       Save

    [Objective] In order to intelligently monitor the distribution of agricultural land cover types, high-spectral cameras are usually mounted on drones to collect high-spectral data, followed by classification of the high-spectral data to automatically draw crop distribution maps. Different crops have similar shapes, and the same crop has significant differences in different growth stages, so the network model for agricultural land cover classification requires a high degree of accuracy. However, network models with high classification accuracy are often complex and cannot be deployed on hardware systems. In view of this problem, a lightweight high-low frequency enhanced Reluformer network (ReluformerN) was proposed in this research. [Methods] Firstly, an adaptive octave convolution was proposed, which utilized the softmax function to automatically adjust the spectral dimensions of high-frequency features and low-frequency features, effectively alleviating the influence of manually setting the spectral dimensions and benefiting the subsequent extraction of spatial and spectral domain features of hyperspectral images. Secondly, a Reluformer was proposed to extract global features, taking advantage of the fact that low-frequency information could capture global features. Reluformer replaced the softmax function with a function of quadratic computational complexity, and through theoretical and graphical analysised, Relu function, LeakRelu function, and Gelu function were compared, it was found that the ReLU function and the softmax function both had non-negativity, which could be used for feature relevance analysis. Meanwhile, the ReLU function has a linearization feature, which is more suitable for self-relevance analysis. Therefore, the ReLU self-attention mechanism was proposed, which used the ReLU function to perform feature self-attention analysis. In order to extract deep global features, multi-scale feature fusion was used, and the ReLU self-attention mechanism was used as the core to construct the multi-head ReLU self-attention mechanism. Similar to the transformer architecture, the Reluformer structure was built by combining multi-head ReLU self-attention mechanism, feedforward layers, and normalization layers. With Reluformer as the core, the Reluformer network (ReluformerN) was proposed. This network considered frequency from the perspective of high-frequency information, taking into account the local features of image high-frequency information, and used deep separable convolution to design a lightweight network for fine-grained feature extraction of high-frequency information. It proposed Reluformer to extract global features for low-frequency information, which represented the global features of the image. ReluformerN was experimented on three public high-spectral data sets (Indian Pines, WHU-Hi-LongKou and Salinas) for crop variety fine classification, and was compared with five popular classification networks (2D-CNN, HybirdSN, ViT, CTN and LSGA-VIT). [Results and Discussion] ReluformerN performed best in overall accuracy (OA), average accuracy (AA), and other accuracy evaluation indicators. In the evaluation indicators of model parameters, model computation (FLOPs), and model complexity, ReluformerN had the smallest number of parameters and was less than 0.3 M, and the lowest computation. In the visualization comparison, the classification effect diagram of the model using ReluformerN had clearer image edges and more complete morphological structures, with fewer classification errors. The validity of the adaptive octave convolution was verified by comparing it with the traditional eightfold convolution. The classification accuracy of the adaptive octave convolution was 0.1% higher than that of the traditional octave convolution. When the artificial parameters were set to different values, the maximum and minimum classification accuracies of the traditional octave convolution were about 0.3% apart, while those of the adaptive octave convolution were only 0.05%. This showed that the adaptive octave convolution not only had the highest classification accuracy, but was also less sensitive to the artificial parameter setting, effectively overcoming the influence of the artificial parameter setting on the classification result. To validated the Reluformer module, it was compared with transformer, LeakRelufromer, and Linformer in terms of accuracy evaluation metrics such as OA and AA. The Reluformer achieved the highest classification accuracy and the lowest model parameter count among these models. This indicated that Reluformer not only effectively extracted global features but also reduced computational complexity. Finally, the effectiveness of the high-frequency and low-frequency branch networks was verified. The effectiveness of the high-frequency and low-frequency feature extraction branches was verified, and the characteristics of the feature distribution after high-frequency feature extraction, after high-low frequency feature extraction, and after the classifier were displayed using a 2D t-sne, compared with the original feature distribution. It was found that after high-frequency feature extraction, similar features were generally clustered together, but the spacing between different features was small, and there were also some features with overlapping situations. After low-frequency feature extraction, it was obvious that similar features were clustered more tightly. After high-low frequency feature fusion, and after the classifier, it was obvious that similar features were clustered, and different types of features were clearly separated, indicating that high-low frequency feature extraction enhanced the classification effect. [Conclusion] This network achieves a good balance between crop variety classification accuracy and model complexity, and is expected to be deployed on hardware systems with limited resources in the future to achieve real-time classification functions.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research on the Spatio-temporal Characteristics and Driving Factors of Smart Farm Development in the Yangtze River Economic Belt
    GAO Qun, WANG Hongyang, CHEN Shiyao
    Smart Agriculture    2024, 6 (6): 168-179.   DOI: 10.12133/j.smartag.SA202404005
    Abstract242)   HTML29)    PDF(pc) (1099KB)(343)       Save

    [Objective] In order to summarize exemplary cases of high-quality development in regional smart agriculture and contribute strategies for the sustainable advancement of the national smart agriculture cause, the spatiotemporal characteristics and key driving factors of smart farms in the Yangtze River Economic Belt were studied. [Methods] Based on data from 11 provinces (municipalities) spanning the years 2014 to 2023, a comprehensive analysis was conducted on the spatio-temporal differentiation characteristics of smart farms in the Yangtze River Economic Belt using methods such as kernel density analysis, spatial auto-correlation analysis, and standard deviation ellipse. Including the overall spatial clustering characteristics, high-value or low-value clustering phenomena, centroid characteristics, and dynamic change trends. Subsequently, the geographic detector was employed to identify the key factors driving the spatio-temporal differentiation of smart farms and to discern the interactions between different factors. The analysis was conducted across seven dimensions: special fiscal support, industry dependence, human capital, urbanization, agricultural mechanization, internet infrastructure, and technological innovation. [Results and Discussions] Firstly, in terms of temporal characteristics, the number of smart farms in the Yangtze River Economic Belt steadily increased over the past decade. The year 2016 marked a significant turning point, after which the growth rate of smart farms had accelerated noticeably. The development of the upper, middle, and lower reaches exhibited both commonalities and disparities. Specifically, the lower sub-regions got a higher overall development level of smart farms, with a fluctuating upward growth rate; the middle sub-regions were at a moderate level, showing a fluctuating upward growth rate and relatively even provincial distribution; the upper sub-regions got a low development level, with a stable and slow growth rate, and an unbalanced provincial distribution. Secondly, in terms of spatial distribution, smart farms in the Yangtze River Economic Belt exhibited a dispersed agglomeration pattern. The results of global auto-correlation indicated that smart farms in the Yangtze River Economic Belt tended to be randomly distributed. The results of local auto-correlation showed that the predominant patterns of agglomeration were H-L and L-H types, with the distribution across provinces being somewhat complex; H-H type agglomeration areas were mainly concentrated in Sichuan, Hubei, and Anhui; L-L type agglomeration areas were primarily in Yunnan and Guizhou. The standard deviation ellipse results revealed that the mean center of smart farms in the Yangtze River Economic Belt had shifted from Anqing city in Anhui province in 2014 to Jingzhou city in Hubei province in 2023, with the spatial distribution showing an overall trend of shifting southwestward and a slow expansion toward the northeast and south. Finally, in terms of key driving factors, technological innovation was the primary critical factor driving the formation of the spatio-temporal distribution pattern of smart farms in the Yangtze River Economic Belt, with a factor explanatory degree of 0.311 1. Moreover, after interacting with other indicators, it continued to play a crucial role in the spatio-temporal distribution of smart farms, which aligned with the practical logic of smart farm development. Urbanization and agricultural mechanization levels were the second and third largest key factors, with factor explanatory degrees of 0.292 2 and 0.251 4, respectively. The key driving factors for the spatio-temporal differentiation of smart farms in the upper, middle, and lower sub-regions exhibited both commonalities and differences. Specifically, the top two key factors driver identification in the upper region were technological innovation (0.841 9) and special fiscal support (0.782 3). In the middle region, they were technological innovation (0.619 0) and human capital (0.600 1), while in the lower region, they were urbanization (0.727 6) and technological innovation (0.425 4). The identification of key driving factors and the detection of their interactive effects further confirmed that the spatio-temporal distribution characteristics of smart farms in the Yangtze River Economic Belt were the result of the comprehensive action of multiple factors. [Conclusions] The development of smart farms in the Yangtze River Economic Belt is showing a positive momentum, with both the total number of smart farms and the number of sub-regions experiencing stable growth. The development speed and level of smart farms in the sub-regions exhibit a differentiated characteristic of "lower reaches > middle reaches > upper reaches". At the same time, the overall distribution of smart farms in the Yangtze River Economic Belt is relatively balanced, with the degree of sub-regional distribution balance being "middle reaches (Hubei province, Hunan province, Jiangxi province are balanced) > lower reaches (dominated by Anhui) > upper reaches (Sichuan stands out)". The coverage of smart farm site selection continues to expand, forming a "northeast-southwest" horizontal diffusion pattern. In addition, the spatio-temporal characteristics of smart farms in the Yangtze River Economic Belt are the result of the comprehensive action of multiple factors, with the explanatory power of factors ranked from high to low as follows: Technological innovation > urbanization > agricultural mechanization > human capital > internet infrastructure > industry dependence > special fiscal support. Moreover, the influence of each factor is further strengthened after interaction. Based on these conclusions, suggestions are proposed to promote the high-quality development of smart farms in the Yangtze River Economic Belt. This study not only provides a theoretical basis and reference for the construction of smart farms in the Yangtze River Economic Belt and other regions, but also helps to grasp the current status and future trends of smart farm development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Differential Privacy-enhanced Blockchain-Based Quality Control Model for Rice
    WU Guodong, HU Quanxing, LIU Xu, QIN Hui, GAO Bowen
    Smart Agriculture    2024, 6 (4): 149-159.   DOI: 10.12133/j.smartag.SA202311027
    Abstract242)   HTML17)    PDF(pc) (1858KB)(591)       Save

    [Objective] Rice plays a crucial role in daily diet. The rice industry involves numerous links, from paddy planting to the consumer's table, and the integrity of the quality control data chain directly affects the credibility of rice quality control and traceability information. The process of rice traceability also faces security issues, such as the leakage of privacy information, which need immediate solutions. Additionally, the previous practice of uploading all information onto the blockchain leads to high storage costs and low system efficiency. To address these problems, this study proposed a differential privacy-enhanced blockchain-based quality control model for rice, providing new ideas and solutions to optimize the traditional quality regulation and traceability system. [Methods] By exploring technologies of blockchain, interplanetary file system (IPFS), and incorporating differential privacy techniques, a blockchain-based quality control model for rice with differential privacy enhancement was constructed. Firstly, the data transmission process was designed to cover the whole industry chain of rice, including cultivation, acquisition, processing, warehousing, and sales. Each module stored the relevant data and a unique number from the previous link, forming a reliable information chain and ensuring the continuity of the data chain for quality control. Secondly, to address the issue of large data volume and low efficiency of blockchain storage, the key quality control data of each link in the rice industry chain was stored in the IPFS. Subsequently, the hash value of the stored data was returned and recorded on the blockchain. Lastly, to enhance the traceability of the quality control model information, the sensitive information in the key quality control data related to the cultivation process was presented to users after undergoing differential privacy processing. Individual data was obfuscated to increase the credibility of the quality control information while also protecting the privacy of farmers' cultivation practices. Based on this model, a differential privacy-enhanced blockchain-based quality control system for rice was designed. [Results and Discussions] The architecture of the differential privacy-enhanced blockchain-based quality control system for rice consisted of the physical layer, transport layer, storage layer, service layer, and application layer. The physical layer included sensor devices and network infrastructure, ensuring data collection from all links of the industry chain. The transport layer handled data transmission and communication, securely uploading collected data to the cloud. The storage layer utilized a combination of traditional databases, IPFS, and blockchain to efficiently store and manage key data on and off the blockchain. The traditional database was used for the management and querying of structured data. IPFS stored the key quality control data in the whole industry chain, while blockchain was employed to store the hash values returned by IPFS. This integrated storage method improved system efficiency, ensured the continuity, reliability, and traceability of quality control data, and provided consumers with reliable information. The service layer was primarily responsible for handling business logic and providing functional services. The implementation of functions in the application layer relied heavily on the design of a series of interfaces within the service layer. Positioned at the top of the system architecture, the application layer was responsible for providing user-centric functionality and interfaces. This encompassed a range of applications such as web applications and mobile applications, aiming to present data and facilitate interactive features to fulfill the requirements of both consumers and businesses. Based on the conducted tests, the average time required for storing data in a single link of the whole industry chain within the system was 1.125 s. The average time consumed for information traceability query was recorded as 0.691 s. Compared to conventional rice quality regulation and traceability systems, the proposed system demonstrated a reduction of 6.64% in the storage time of single-link data and a decrease of 16.44% in the time required to perform information traceability query. [Conclusions] This study proposes a differential privacy-enhanced blockchain-based quality control model for rice. The model ensures the continuity of the quality control data chain by integrating the various links of the whole industry chain of rice. By combining blockchain with IPFS storage, the model addresses the challenges of large data volume and low efficiency of blockchain storage in traditional systems. Furthermore, the model incorporates differential privacy protection to enhance traceability while safeguarding the privacy of individual farmers. This study can provide reference for the design and improvement of rice quality regulation and traceability systems.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Agriculture    2024, 6 (6): 0-0.  
    Abstract238)            Save
    Related Articles | Metrics | Comments0
    Precision Target Spraying System Integrated with Remote Deep Learning Recognition Model for Cabbage Plant Centers
    ZHANG Hui, HU Jun, SHI Hang, LIU Changxi, WU Miao
    Smart Agriculture    2024, 6 (6): 85-95.   DOI: 10.12133/j.smartag.SA202406013
    Abstract234)   HTML24)    PDF(pc) (1923KB)(764)       Save

    [Objective] Spraying calcium can effectively prevent the occurrence of dry burning heart disease in Chinese cabbage. Accurately targeting spraying calcium can more effectively improve the utilization rate of calcium. Since the sprayer needs to move rapidly in the field, this can lead to over-application or under-application of the pesticide. This study aims to develop a targeted spray control system based on deep learning technology, explore the relationship between the advance speed, spray volume, and coverage of the sprayer, thereby addressing the uneven application issues caused by different nebulizer speeds by studying the real scenario of calcium administration to Chinese cabbage hearts. [Methods] The targeted spraying control system incorporates advanced sensors and computing equipment that were capable of obtaining real-time data regarding the location of crops and the surrounding environmental conditions. This data allowed for dynamic adjustments to be made to the spraying system, ensuring that pesticides were delivered with high precision. To further enhance the system's real-time performance and accuracy, the YOLOv8 object detection model was improved. A Ghost-Backbone lightweight network structure was introduced, integrating remote sensing technologies along with the sprayer's forward speed and the frequency of spray responses. This innovative combination resulted in the creation of a YOLOv8-Ghost-Backbone lightweight model specifically tailored for agricultural applications. The model operated on the Jetson Xavier NX controller, which was a high-performance, low-power computing platform designed for edge computing. The system was allowed to process complex tasks in real time directly in the field. The targeted spraying system was composed of two essential components: A pressure regulation unit and a targeted control unit. The pressure regulation unit was responsible for adjusting the pressure within the spraying system to ensure that the output remains stable under various operational conditions. Meanwhile, the targeted control unit played a crucial role in precisely controlling the direction, volume, and coverage of the spray to ensure that the pesticide was applied effectively to the intended areas of the plants. To rigorously evaluate the performance of the system, a series of intermittent spray tests were conducted. During these tests, the forward speed of the sprayer was gradually increased, allowing to assess how well the system responded to changes in speed. Throughout the testing phase, the response frequency of the electromagnetic valve was measured to calculate the corresponding spray volume for each nozzle. [Results and Conclusions] The experimental results indicated that the overall performance of the targeted spraying system was outstanding, particularly under conditions of high-speed operation. By meticulously recording the response times of the three primary components of the system, the valuable data were gathered. The average time required for image processing was determined to be 29.50 ms, while the transmission of decision signals took an average of 6.40 ms. The actual spraying process itself required 88.83 ms to complete. A thorough analysis of these times revealed that the total response time of the spraying system lagged by approximately 124.73 ms when compared to the electrical signal inputs. Despite the inherent delays, the system was able to maintain a high level of spraying accuracy by compensating for the response lag of the electromagnetic valve. Specifically, when tested at a speed of 7.2 km/h, the difference between the actual spray volume delivered and the required spray volume, after accounting for compensation, was found to be a mere 0.01 L/min. This minimal difference indicates that the system met the standard operational requirements for effective pesticide application, thereby demonstrating its precision and reliability in practical settings. [Conclusions] In conclusion, this study developed and validated a deep learning-based targeted spraying control system that exhibited excellent performance regarding both spraying accuracy and response speed. The system serves as a significant technical reference for future endeavors in agricultural automation. Moreover, the research provides insights into how to maintain consistent spraying effectiveness and optimize pesticide utilization efficiency by dynamically adjusting the spraying system as the operating speed varies. The findings of this research will offer valuable experiences and guidance for the implementation of agricultural robots in the precise application of pesticides, with a particular emphasis on parameter selection and system optimization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Parametric Reconstruction Method of Wheat Leaf Curved Surface Based on Three-Dimensional Point Cloud
    ZHU Shunyao, QU Hongjun, XIA Qian, GUO Wei, GUO Ya
    Smart Agriculture    2025, 7 (1): 85-96.   DOI: 10.12133/j.smartag.SA202410004
    Abstract223)   HTML30)    PDF(pc) (2475KB)(411)       Save

    [Objective] Plant leaf shape is an important part of plant architectural model. Establishment of a three-dimensional structural model of leaves may assist simulating and analyzing plant growth. However, existing leaf modeling approaches lack interpretability, invertibility, and operability, which limit the estimation of model parameters, the simulation of leaf shape, the analysis and interpretation of leaf physiology and growth state, and model reusage. Aiming at the interoperability between three-dimensional structure representation and mathematical model parameters, this study paid attention to three aspects in wheat leaf shape parametric reconstruction: (1) parameter-driven model structure, (2) model parameter inversion, and (3) parameter dynamic mapping during growth. Based on this, a set of parameter-driven and point cloud inversion model for wheat leaf interoperability was proposed in this study. [Methods] A parametric surface model of a wheat leaf with seven characteristic parameters by using parametric modeling technology was built, and the forward parametric construction of the wheat leaf structure was realized. Three parameters, maximum leaf width, leaf length, and leaf shape factor, were used to describe the basic shape of the blade on the leaf plane. On this basis, two parameters, namely the angle between stems and leaves and the curvature degree, were introduced to describe the bending characteristics of the main vein of the blade in the three-dimensional space. Two parameters, namely the twist angle around the axis and the twist deviation angle around the axis, were introduced to represent the twisted structure of the leaf blade along the vein. The reverse parameter estimation module was built according to the surface model. The point cloud was divided by the uniform segmentation method along the Y-axis, and the veins were fit by a least squares regression method. Then, the point cloud was re-segmented according to the fit vein curve. Subsequently, the rotation angle was precisely determined through the segment-wise transform estimation method, with all parameters being optimally fit using the RANSAC regression algorithm. To validate the reliability of the proposed methodology, a set of sample parameters was randomly generated, from which corresponding sample point clouds were synthesized. These sample point clouds were then subjected to estimation using the described method. Then error analyzing was carried out on the estimation results. Three-dimensional imaging technology was used to collect the point clouds of Zhengmai 136, Yangmai 34, and Yanmai 1 samples. After noise reduction and coordinate registration, the model parameters were inverted and estimated, and the reconstructed point clouds were produced using the parametric model. The reconstruction error was validated by calculating the dissimilarity, represented by the Chamfer Distance, between the reconstructed point cloud and the measured point cloud. [Results and Discussions] The model could effectively reconstruct wheat leaves, and the average deviation of point cloud based parametric reconstruction results was about 1.2 mm, which had a high precision. Parametric modeling technology based on prior knowledge and point cloud fitting technology based on posterior data was integrated in this study to construct a digital twin model of specific species at the 3D structural level. Although some of the detailed characteristics of the leaves were moderately simplified, the geometric shape of the leaves could be highly restored with only a few parameters. This method was not only simple, direct and efficient, but also had more explicit geometric meaning of the obtained parameters, and was both editable and interpretable. In addition, the practice of using only tools such as rulers to measure individual characteristic parameters of plant organs in traditional research was abandoned in this study. High-precision point cloud acquisition technology was adopted to obtain three-dimensional data of wheat leaves, and pre-processing work such as point cloud registration, segmentation, and annotation was completed, laying a data foundation for subsequent research. [Conclusions] There is interoperability between the reconstructed model and the point cloud, and the parameters of the model can be flexibly adjusted to generate leaf clusters with similar shapes. The inversion parameters have high interpretability and can be used for consistent and continuous estimation of point cloud time series. This research is of great value to the simulation analysis and digital twinning of wheat leaves.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Key Technologies and Construction model for Unmanned Smart Farms: Taking the "1.5-Ton Grain per Mu" Unmanned Farm as An Example
    LIU lining, ZHANG Hongqi, ZHANG Ziwen, ZHANG Zhenghui, WANG Jiayu, LI Xuanxuan, ZHU Ke, LIU Pingzeng
    Smart Agriculture    2025, 7 (1): 70-84.   DOI: 10.12133/j.smartag.SA202410033
    Abstract208)   HTML27)    PDF(pc) (2674KB)(175)       Save

    [Objective] As a key model of smart agriculture, the unmanned smart farm aims to develop a highly intelligent and automated system for high grain yields. This research uses the "1.5-Ton grain per Mu" farm in Dezhou city, Shandong province, as the experimental site, targeting core challenges in large-scale smart agriculture and exploring construction and service models for such farms. [Methods] The "1.5-Ton grain per Mu" unmanned smart farm comprehensively utilized information technologies such as the internet of things (IoT) and big data to achieve full-chain integration and services for information perception, transmission, mining, and application. The overall construction architecture consisted of the perception layer, transmission layer, processing layer, and application layer. This architecture enabled precise perception, secure transmission, analysis and processing, and application services for farm data. A perception system for the unmanned smart farm of wheat was developed, which included a digital perception network and crop phenotypic analysis. The former achieved precise perception, efficient transmission, and precise measurement and control of data information within the farm through perception nodes, self-organizing networks, and edge computing core processing nodes. Phenotypic analysis utilized methods such as deep learning to extract phenotypic characteristics at different growth stages, such as the phenological classification of wheat and wheat ear length. An intelligent controlled system had been developed. The system consisted of an intelligent agricultural machinery system, a field irrigation system, and an aerial pesticided application system. The intelligent agricultural machinery system was composed of three parts: the basic layer, decision-making layer, and application service layer. They were responsible for obtaining real-time status information of agricultural machinery, formulating management decisions for agricultural machinery, and executing operational commands, respectively. Additionally, appropriate agricultural machinery models and configuration references were provided. A refined irrigation scheme was designed based on the water requirements and soil conditions at different developmental stages of wheat. And, an irrigation control algorithm based on fuzzy PID was proposed. Finally, relying on technologies such as multi-source data fusion, distributed computing, and geographic information system (GIS), an intelligent management and control platform for the entire agricultural production process was established. [Results and Discussions] The digital perception network enabled precise sensing and networked transmission of environmental information within the farm. The data communication quality of the sensor network remained above 85%, effectively ensuring data transmission quality. The average relative error in extracting wheat spike length information based on deep learning algorithms was 1.24%. Through the coordinated operation of intelligent control system, the farm achieved lean and unmanned production management, enabling intelligent control throughout the entire production chain, which significantly reduced labor costs and improved the precision and efficiency of farm management. The irrigation model not only saved 20% of irrigation water but also increased the yield of "Jinan 17" and "Jimai 44" by 10.18% and 7%, respectively. Pesticide application through spraying drones reduced pesticide usage by 55%. The big data platform provided users with production guidance services such as meteorological disaster prediction, optimal sowing time, environmental prediction, and water and fertilizer management through intelligent scientific decision support, intelligent agricultural machinery operation, and producted quality and safety traceability modules, helping farmers manage their farms scientifically. [Conclusions] The study achieved comprehensive collection of environmental information within the farm, precise phenotypic analysis, and intelligent control of agricultural machinery, irrigation equipment, and other equipment. Additionally, it realized digital services for agricultural management through a big data platform. The development path of the "1.5-Ton grain per Mu" unmanned smart farm can provid references for the construction of smart agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Accurate Detection of Tree Planting Locations in Inner Mongolia for The Three North Project Based on YOLOv10-MHSA
    XIE Jiyuan, ZHANG Dongyan, NIU Zhen, CHENG Tao, YUAN Feng, LIU Yaling
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202410010
    Online available: 24 January 2025

    Agricultural Large Language Model Based on Precise Knowledge Retrieval and Knowledge Collaborative Generation
    JIANG Jingchi, YAN Lian, LIU Jie
    Smart Agriculture    2025, 7 (1): 20-32.   DOI: 10.12133/j.smartag.SA202410025
    Abstract200)   HTML24)    PDF(pc) (2081KB)(213)       Save

    [Objective] The rapid advancement of large language models (LLMs) has positioned them as a promising novel research paradigm in smart agriculture, leveraging their robust cognitive understanding and content generative capabilities. However, due to the lack of domain-specific agricultural knowledge, general LLMs often exhibit factual errors or incomplete information when addressing specialized queries, which is particularly prominent in agricultural applications. Therefore, enhancing the adaptability and response quality of LLMs in agricultural applications has become an important research direction. [Methods] To improve the adaptability and precision of LLMs in the agricultural applications, an innovative approach named the knowledge graph-guided agricultural LLM (KGLLM) was proposed. This method integrated information entropy for effective knowledge filtering and applied explicit constraints on content generation during the decoding phase by utilizing semantic information derived from an agricultural knowledge graph. The process began by identifying and linking key entities from input questions to the agricultural knowledge graph, which facilitated the formation of knowledge inference paths and the development of question-answering rationales. A critical aspect of this approach was ensuring the validity and reliability of the external knowledge incorporated into the model. This was achieved by evaluating the entropy difference in the model's outputs before and after the introduction of each piece of knowledge. Knowledge that didn't enhance the certainty of the answers was systematically filtered out. The knowledge paths that pass this entropy evaluation were used to adjust the token prediction probabilities, prioritizing outputs that were closely aligned with the structured knowledge. This allowed the knowledge graph to exert explicit guidance over the LLM's outputs, ensuring higher accuracy and relevance in agricultural applications. [Results and Discussions] The proposed knowledge graph-guided technique was implemented on five mainstream general-purpose LLMs, including open-source models such as Baichuan, ChatGLM, and Qwen. These models were compared with state-of-the-art knowledge graph-augmented generation methods to evaluate the effectiveness of the proposed approach. The results demonstrate that the proposed knowledge graph-guided approach significantly improved several key performance metrics of fluency, accuracy, factual correctness, and domain relevance. Compared to GPT-4o, the proposed method achieved notable improvements by an average of 2.592 3 in Mean BLEU, 2.815 1 in ROUGE, and 9.84% in BertScore. These improvements collectively signify that the proposed approach effectively leverages agricultural domain knowledge to refine the outputs of general-purpose LLMs, making them more suitable for agricultural applications. Ablation experiments further validated that the knowledge-guided agricultural LLM not only filtered out redundant knowledge but also effectively adjusts token prediction distributions during the decoding phase. This enhanced the adaptability of general-purpose LLMs in agriculture contexts and significantly improves the interpretability of their responses. The knowledge filtering and knowledge graph-guided model decoding method proposed in this study, which was based on information entropy, effectively identifies and selects knowledge that carried more informational content through the comparison of information entropy.Compared to existing technologies in the agricultural field, this method significantly reduced the likelihood of "hallucination" phenomena during the generation process. Furthermore, the guidance of the knowledge graph ensured that the model's generated responses were closely related to professional agricultural knowledge, thereby avoiding vague and inaccurate responses generated from general knowledge. For instance, in the application of pest and disease control, the model could accurately identify the types of crop diseases and corresponding control measures based on the guided knowledge path, thereby providing more reliable decision support. [Conclusions] This study provides a valuable reference for the construction of future agricultural large language models, indicating that the knowledge graphs guided mehtod has the potential to enhance the domain adaptability and answer quality of models. Future research can further explore the application of similar knowledge-guided strategies in other vertical fields to enhance the adaptability and practicality of LLMs across various professional domains.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Market Monitoring and Early Warning: An Integrated Forecasting Approach Based on Deep Learning
    XU Shiwei, LI Qianchuan, LUAN Rupeng, ZHUANG Jiayu, LIU Jiajia, XIONG Lu
    Smart Agriculture    2025, 7 (1): 57-69.   DOI: 10.12133/j.smartag.SA202411004
    Abstract198)   HTML16)    PDF(pc) (1936KB)(105)       Save

    [Significance] The fluctuations in the supply, consumption, and prices of agricultural products directly affect market monitoring and early warning systems. With the ongoing transformation of China's agricultural production methods and market system, advancements in data acquisition technologies have led to an explosive growth in agricultural data. However, the complexity of the data, the narrow applicability of existing models, and their limited adaptability still present significant challenges in monitoring and forecasting the interlinked dynamics of multiple agricultural products. The efficient and accurate forecasting of agricultural market trends is critical for timely policy interventions and disaster management, particularly in a country with a rapidly changing agricultural landscape like China. Consequently, there is a pressing need to develop deep learning models that are tailored to the unique characteristics of Chinese agricultural data. These models should enhance the monitoring and early warning capabilities of agricultural markets, thus enabling precise decision-making and effective emergency responses. [Methods] An integrated forecasting methodology was proposed based on deep learning techniques, leveraging multi-dimensional agricultural data resources from China. The research introduced several models tailored to different aspects of agricultural market forecasting. For production prediction, a generative adversarial network and residual network collaborative model (GAN-ResNet) was employed. For consumption forecasting, a variational autoencoder and ridge regression (VAE-Ridge) model was used, while price prediction was handled by an Adaptive-Transformer model. A key feature of the study was the adoption of an "offline computing and visualization separation" strategy within the Chinese agricultural monitoring and early warning system (CAMES). This strategy ensures that model training and inference are performed offline, with the results transmitted to the front-end system for visualization using lightweight tools such as ECharts. This approach balances computational complexity with the need for real-time early warnings, allowing for more efficient resource allocation and faster response times. The corn, tomato, and live pig market data used in this study covered production, consumption and price data from 1980 to 2023, providing comprehensive data support for model training. [Results and Discussions] The deep learning models proposed in this study significantly enhanced the forecasting accuracy for various agricultural products. For instance, the GAN-ResNet model, when used to predict maize yield at the county level, achieved a mean absolute percentage error (MAPE) of 6.58%. The VAE-Ridge model, applied to pig consumption forecasting, achieved a MAPE of 6.28%, while the Adaptive-Transformer model, used for tomato price prediction, results in a MAPE of 2.25%. These results highlighted the effectiveness of deep learning models in handling complex, nonlinear relationships inherent in agricultural data. Additionally, the models demonstrate notable robustness and adaptability when confronted with challenges such as sparse data, seasonal market fluctuations, and heterogeneous data sources. The GAN-ResNet model excels in capturing the nonlinear fluctuations in production data, particularly in response to external factors such as climate conditions. Its capacity to integrate data from diverse sources—including weather data and historical yield data—made it highly effective for production forecasting, especially in regions with varying climatic conditions. The VAE-Ridge model addressed the issue of data sparsity, particularly in the context of consumption data, and provided valuable insights into the underlying relationships between market demand, macroeconomic factors, and seasonal fluctuations. Finally, the Adaptive-Transformer model stand out in price prediction, with its ability to capture both short-term price fluctuations and long-term price trends, even under extreme market conditions. [Conclusions] This study presents a comprehensive deep learning-based forecasting approach for agricultural market monitoring and early warning. The integration of multiple models for production, consumption, and price prediction provides a systematic, effective, and scalable tool for supporting agricultural decision-making. The proposed models demonstrate excellent performance in handling the nonlinearities and seasonal fluctuations characteristic of agricultural markets. Furthermore, the models' ability to process and integrate heterogeneous data sources enhances their predictive power and makes them highly suitable for application in real-world agricultural monitoring systems. Future research will focus on optimizing model parameters, enhancing model adaptability, and expanding the system to incorporate additional agricultural products and more complex market conditions. These improvements will help increase the stability and practical applicability of the system, thus further enhancing its potential for real-time market monitoring and early warning capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grain Production Big Data Platform: Progress and Prospects
    YANG Guijun, ZHAO Chunjiang, YANG Xiaodong, YANG Hao, HU Haitang, LONG Huiling, QIU Zhengjun, LI Xian, JIANG Chongya, SUN Liang, CHEN Lei, ZHOU Qingbo, HAO Xingyao, GUO Wei, WANG Pei, GAO Meiling
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202409014
    Online available: 21 January 2025

    Image Segmentation Method of Chinese Yam Leaves in Complex Background Based on Improved ENet
    LU Bibo, LIANG Di, YANG Jie, SONG Aiqing, HUANGFU Shangwei
    Smart Agriculture    2024, 6 (6): 109-120.   DOI: 10.12133/j.smartag.SA202407007
    Abstract181)   HTML21)    PDF(pc) (2024KB)(677)       Save

    [Objective] Crop leaf area is an important indicator reflecting light absorption efficiency and growth conditions. This paper established a diverse Chinese yam image dataset and proposesd a deep learning-based method for Chinese yam leaf image segmentation. This method can be used for real-time measurement of Chinese yam leaf area, addressing the inefficiency of traditional measurement techniques. This will provide more reliable data support for genetic breeding, growth and development research of Chinese yam, and promote the development and progress of the Chinese yam industry. [Methods] A lightweight segmentation network based on improved ENet was proposed. Firstly, based on ENet, the third stage was pruned to reduce redundant calculations in the model. This improved the computational efficiency and running speed, and provided a good basis for real-time applications. Secondly, PConv was used instead of the conventional convolution in the downsampling bottleneck structure and conventional bottleneck structure, the improved bottleneck structure was named P-Bottleneck. PConv applied conventional convolution to only a portion of the input channels and left the rest of the channels unchanged, which reduced memory accesses and redundant computations for more efficient spatial feature extraction. PConv was used to reduce the amount of model computation while increase the number of floating-point operations per second on the hardware device, resulting in lower latency. Additionally, the transposed convolution in the upsampling module was improved to bilinear interpolation to enhance model accuracy and reduce the number of parameters. Bilinear interpolation could process images smoother, making the processed images more realistic and clear. Finally, coordinate attention (CA) module was added to the encoder to introduce the attention mechanism, and the model was named CBPA-ENet. The CA mechanism not only focused on the channel information, but also keenly captured the orientation and position-sensitive information. The position information was embedded into the channel attention to globally encode the spatial information, capturing the channel information along one spatial direction while retaining the position information along the other spatial direction. The network could effectively enhance the attention to important regions in the image, and thus improve the quality and interpretability of segmentation results. [Results and Discussions] Trimming the third part resulted in a 28% decrease in FLOPs, a 41% decrease in parameters, and a 9 f/s increase in FPS. Improving the upsampling method to bilinear interpolation not only reduces the floating-point operation and parameters, but also slightly improves the segmentation accuracy of the model, increasing FPS by 4 f/s. Using P-Bottleneck instead of downsampling bottleneck structure and conventional bottleneck structure can reduce mIoU by only 0.04%, reduce FLOPs by 22%, reduce parameters by 16%, and increase FPS by 8 f/s. Adding CA mechanism to the encoder could only increase a small amount of FLOPs and parameters, improving the accuracy of the segmentation network. To verify the effectiveness of the improved segmentation algorithm, classic semantic segmentation networks of UNet, DeepLabV3+, PSPNet, and real-time semantic segmentation network LinkNet, DABNet were selected to train and validate. These six algorithms got quite high segmentation accuracy, among which UNet had the best mIoU and the mPA, but the model size was too large. The improved algorithm only accounts for 1% of the FLOPs and 0.41% of the parameters of UNet, and the mIoU and mPA were basically the same. Other classic semantic segmentation algorithms, such as DeepLabV3+, had similar accuracy to improved algorithms, but their large model size and slow inference speed were not conducive to embedded development. Although the real-time semantic segmentation algorithm LinkNet had a slightly higher mIoU, its FLOPs and parameters count were still far greater than the improved algorithm. Although the PSPNet model was relatively small, it was also much higher than the improved algorithm, and the mIoU and mPA were lower than the algorithm. The experimental results showed that the improved model achieved a mIoU of 98.61%. Compared with the original model, the number of parameters and FLOPs significantly decreased. Among them, the number of model parameters decreased by 51%, the FLOPs decreased by 49%, and the network operation speed increased by 38%. [Conclusions] The improved algorithm can accurately and quickly segment Chinese yam leaves, providing not only a more accurate means for determining Chinese yam phenotype data, but also a new method and approach for embedded research of Chinese yam. Using the model, the morphological feature data of Chinese yam leaves can be obtained more efficiently, providing a reliable foundation for further research and analysis.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Detection and Recognition Model for Small Target Pests on Sticky Traps in Multi-Source Scenarios
    YANG Xinting, HU Huan, CHEN Xiao, LI Wenzheng, ZHOU Zijie, LI Wenyong
    Smart Agriculture    2025, 7 (1): 111-123.   DOI: 10.12133/j.smartag.SA202410019
    Abstract171)   HTML17)    PDF(pc) (2514KB)(123)       Save

    [Objective] In crop cultivation and production, pests have gradually become one of the main issues affecting agricultural yield. Traditional models often focus on achieving high accuracy, however, to facilitate model application, lightweighting is necessary. The targets in yellow sticky trap images are often very small with low pixel resolution, so modifications in network structure, loss functions, and lightweight convolutions need to adapt to the detection of small-object pests. Ensuring a balance between model lightweighting and small-object pest detection is particularly important. To improve the detection accuracy of small target pests on sticky trap images from multi-source scenarios, a lightweight detection model named MobileNetV4+VN-YOLOv5s was proposed in this research to detect two main small target pests in agricultural production, whiteflies and thrips. [Methods] In the backbone layer of MobileNetV4+VN-YOLOv5s, an EM block constructed with the MobileNetV4 backbone network was introduced for detecting small, high-density, and overlapping targets, making it suitable for deployment on mobile devices. Additionally, the Neck layer of MobileNetV4+VN-YOLOv5s incorporates the GSConv and VoV-GSCSP modules to replace regular convolutional modules with lightweight design, effectively reducing the parameter size of the model while improving detection accuracy. Lastly, a normalized wasserstein distance (NWD)loss function was introduced into the framework to enhance the sensitivity for low-resolution small target pests. Extensive experiments including state-of-the-art comparison, ablation evaluation, performance analysis on image splitting, pest density and multi-source data were conducted. [Results and Discussions] Through ablation tests, it was concluded that the EM module and the VoV-GSCSP convolution module had significant effects in reducing the model parameter size and frame rate, the NWD loss function significantly improved the mean average precision (mAP) of the model. By comparing tests with different loss functions, the NWD loss function improves the mAP by 6.1, 10.8 and 8.2 percentage compared to the DIoU, GIoU and EIoU loss functions, respectively, so the addition of the NWD loss function achieved good results. Comparative performance tests were detected wiht different light weighting models, the experimental results showed that the mAP of the proposed MobileNetV4+VN-YOLOv5s model in three scenarios (Indoor, Outdoor, Indoor&Outdoor) was 82.5%, 70.8%, and 74.7%, respectively. Particularly, the MobileNetV4+VN-YOLOv5s model had a parameter size of only 4.2 M, 58% of the YOLOv5s model, the frame rate was 153.2 fps, an increase of 6.0 fps compared to the YOLOv5s model. Moreover, the precision and mean average precision reach 79.7% and 82.5%, which were 5.6 and 8.4 percentage points higher than the YOLOv5s model, respectively. Comparative tests were conducted in the upper scenarios based on four splitting ratios: 1×1, 2×2, 5×5, and 10×10. The most superior was the result by using 5×5 ratio in indoor scenario, and the mAP of this case reached 82.5%. The mAP of the indoor scenario was the highest in the low-density case, reaching 83.8%, and the model trained based on the dataset from indoor condition achieves the best performance. Comparative tests under different densities of pest data resulted in a decreasing trend in mAP from low to high densities for the MobileNetV4+VN-YOLOv5s model in the three scenarios. Based on the comparison of the experimental results of different test sets in different scenarios, all three models achieved the best detection accuracy on the IN dataset. Specifically, the IN-model had the highest mAP at 82.5%, followed by the IO-model. At the same time, the detection performance showed the same trend across all three test datasets: The IN model performed the best, followed by the IO-model, and the OUT-model performed the lowest. By comparing the tests with different YOLO improvement models, it was concluded that MobileNetV4+VN-YOLOv5s had the highest mAP, EVN-YOLOv8s was the second highest, and EVN-YOLOv11s was the lowest. Besides, after deploying the model to the Raspberry Pi 4B motherboard, it was concluded that the detection results of the YOLOv5s model had more misdetections and omissions than those of the MobileNetV4+VN-YOLOv5s model, and the time of the model was shortened by about 33% compared to that of the YOLOv5s model, which demonstrated that the model had a good prospect of being deployed in the application. [Conclusions] The MobileNetV4+VN-YOLOv5s model proposed in this study achieved a balance between lightweight design and accuracy. It can be deployed on embedded devices, facilitating practical applications. The model can provide a reference for detecting small target pests in sticky trap images under various multi-source scenarios.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Tomato Growth Height Prediction Method by Phenotypic Feature Extraction Using Multi-modal Data
    GONG Yu, WANG Ling, ZHAO Rongqiang, YOU Haibo, ZHOU Mo, LIU Jie
    Smart Agriculture    2025, 7 (1): 97-110.   DOI: 10.12133/j.smartag.SA202410032
    Abstract164)   HTML30)    PDF(pc) (1307KB)(79)       Save

    [Objective] Accurate prediction of tomato growth height is crucial for optimizing production environments in smart farming. However, current prediction methods predominantly rely on empirical, mechanistic, or learning-based models that utilize either images data or environmental data. These methods fail to fully leverage multi-modal data to capture the diverse aspects of plant growth comprehensively. [Methods] To address this limitation, a two-stage phenotypic feature extraction (PFE) model based on deep learning algorithm of recurrent neural network (RNN) and long short-term memory (LSTM) was developed. The model integrated environment and plant information to provide a holistic understanding of the growth process, emploied phenotypic and temporal feature extractors to comprehensively capture both types of features, enabled a deeper understanding of the interaction between tomato plants and their environment, ultimately leading to highly accurate predictions of growth height. [Results and Discussions] The experimental results showed the model's effectiveness: When predicting the next two days based on the past five days, the PFE-based RNN and LSTM models achieved mean absolute percentage error (MAPE) of 0.81% and 0.40%, respectively, which were significantly lower than the 8.00% MAPE of the large language model (LLM) and 6.72% MAPE of the Transformer-based model. In longer-term predictions, the 10-day prediction for 4 days ahead and the 30-day prediction for 12 days ahead, the PFE-RNN model continued to outperform the other two baseline models, with MAPE of 2.66% and 14.05%, respectively. [Conclusions] The proposed method, which leverages phenotypic-temporal collaboration, shows great potential for intelligent, data-driven management of tomato cultivation, making it a promising approach for enhancing the efficiency and precision of smart tomato planting management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Domain Generalization Method of Strawberry Disease Recognition Based on Instance Whitening and Restitution
    HU Xiaobo, XU Taosheng, WANG Chengjun, ZHU Hongbo, GAN Lei
    Smart Agriculture    2025, 7 (1): 124-135.   DOI: 10.12133/j.smartag.SA202411016
    Abstract153)   HTML13)    PDF(pc) (1496KB)(200)       Save

    [Objective] Strawberry disease recognition models based on deep neural networks generally assume that the training (source domain) and the test (target domain) datasets are identically and independently distributed. However, in practical applications, due to the influence of illumination, background and strawberry variety, the target domain often exhibits significant domain shift from the source domain. The domain shift result in accuracy decline of the models in target domain. To address this problem, a domain generalization method based on instant whitening and restitution (IWR) was proposed to improve the generalization performance of strawberry disease identification models in this research. [Methods] Samples from different source often exhibit great domain shift due to variations in strawberry varieties, regional climate, and photography methods. Therefore, a dataset was constructed for domain generalization research on strawberry disease using two distinct approaches. The first dataset was acquired using a Nikon D810 camera at multiple strawberry farms in Changfeng county, Anhui province, with a fixed sampling schedule and fixed camera distance. In contrast, the second dataset was an open-source collection, primarily comprising images captured using smartphones in multiple strawberry greenhouses in Korea, with varied and random shooting distances and angles. The IWR module mitigated style variations (e.g., illumination, color) through instance whitening, where features were normalized to reduce domain discrepancies between the datasets. However, such operation was task-ignorant and inevitable removed some task-relevant information, which may be harmful to classification performance of the models. To remedy this, the removed task-relevant features were attempted to recover. Specifically, two modules were designed to extract task-relevant and task-irrelevant feature from the filtered style features, respectively. A dual restitution loss was utilized to constraint the modules' feature correlation between the task and a mutual loss was used to ensure the independence of the features. In addition, a separation optimization strategy was adopted to further enhance the feature separation effect of the two modules. [Results and Discussions] The F1-Score was adopted as evaluation metrics. A series of ablations studies and comparative experiments were conducted to demonstrate the effectiveness of the proposed IWR. The ablation experiments proved that the IWR could effectively eliminate the style variations between different datasets and separate task-relevant feature from the filtered style features, which could simultaneously enhance model generalization and discrimination capabilities. The recognition accuracy increased when IWR pluged to AlexNet, GoogLeNet, ResNet-18, ResNet-50, MobileNetV2 and MobileNetV3. It demonstrated that the proposed IWR was an effective way to improve the generalization of the models. Compared with other domain generalization methods such as IBNNet, SW and SNR, the generalization performance of the proposed algorithm on test datasets could be improved by 2.63%, 2.35% and 1.14%, respectively. To better understand how IWR works, the intermediate feature maps of ResNet-50 without and with IWR were compared. The visualization result showed that the model with IWR was more robust when the image style changed. These results indicated that the proposed IWR achieves high classification accuracy and boosts the generalization performance of the models. [Conclusions] An instance whitening and restitution module was presented, which aimed to learn generalizable and discriminative feature representations for effective domain generalization. The IWR was a plug-and-play module, it could be inserted into existing convolutional networks for strawberry disease recognition. Style normalization and restitution (SNR) reduced the style information through instance whitening operation and then restitutes the task-relevant discriminative features caused by instance whitening. The introduced dual restitution loss and mutual loss further facilitate the separation of task-relevant and task-irrelevant feature. The schemes powered by IWR achieves the state-of-the-art performance on strawberry disease identification.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research on Agricultural Drought Prediction Based on GCN-BiGRU-STMHSA
    QUAN Jialu, CHEN Wenbai, WANG Yiqun, CHENG Jiajing, LIU Yilong
    Smart Agriculture    2025, 7 (1): 156-164.   DOI: 10.12133/j.smartag.SA202410027
    Abstract152)   HTML15)    PDF(pc) (1086KB)(84)       Save

    [Objective] Agricultural drought has a negative impact on the development of agricultural production and even poses a threat to food security. To reduce disaster losses and ensure stable crop yields, accurately predicting and classifying agricultural drought severity based on the standardized soil moisture index (SSMI) is of significant importance. [Methods] An agricultural drought prediction model, GCN-BiGRU-STMHSA was proposed, which integrated a graph convolutional network (GCN), a bidirectional gated recurrent unit (BiGRU), and a multi-head self-attention (MHSA) mechanism, based on remote sensing data. In terms of model design, the proposed method first employed GCN to fully capture the spatial correlations among different meteorological stations. By utilizing GCN, a spatial graph structure based on meteorological stations was constructed, enabling the extraction and modeling of spatial dependencies between stations. Additionally, a spatial multi-head self-attention mechanism (S-MHSA) was introduced to further enhance the model's ability to capture spatial features. For temporal modeling, BiGRU was utilized as the time-series feature extraction module. BiGRU considers both forward and backward dependencies in time-series data, enabling a more comprehensive understanding of the temporal dynamics of agricultural drought. Meanwhile, a temporal multi-head self-attention mechanism (T-MHSA) was incorporated to enhance the model's capability to learn long-term temporal dependencies and improve prediction stability across different time scales. Finally, the model employed a fully connected layer to perform regression prediction of the SSMI. Based on the classification criteria for agricultural drought severity levels, the predicted SSMI values were mapped to the corresponding drought severity categories, achieving precise agricultural drought classification. To validate the effectiveness of the proposed model, the global land data assimilation system (GLDAS_2.1) dataset and conducted modeling and experiments was utilized on five representative meteorological stations in the North China Plain (Xinyang, Gushi, Fuyang, Huoqiu, and Dingyuan). Additionally, the proposed model was compared with multiple deep learning models, including GRU, LSTM, and Transformer, to comprehensively evaluate its performance in agricultural drought prediction tasks. The experimental design covered different forecasting horizons to analyze the model's generalization capability in both short-term and long-term predictions, thereby providing a more reliable early warning system for agricultural drought. [Results and Discussions] Experimental results demonstrated that the proposed GCN-BiGRU-STMHSA model outperforms baseline models in both SSMI prediction and agricultural drought classification tasks. Specifically, across the five study stations, the model achieved significantly lower mean absolute error (MAE) and root mean squared error (RMSE), while attaining higher coefficient of determination ( R²), classification accuracy (ACC), and F1-Score ( F1). Notably, at the Gushi station, the model exhibited the best performance in predicting SSMI 10 days ahead, achieving an MAE of 0.053, a RMSE of 0.071, a R² of 0.880, an ACC of 0.925, and a F1 of 0.924. Additionally, the model's generalization capability was investigated under different forecasting horizons (7, 14, 21, and 28 days). Results indicated that the model achieved the highest accuracy in short-term predictions (7 days). Although errors increase slightly as the prediction horizon extends, the model maintained high classification accuracy even for long-term predictions (up to 28 days). This highlighted the model's robustness and effectiveness in agricultural drought prediction over varying time scales. [Conclusions] The proposed model achieves superior accuracy and generalization capability in agricultural drought prediction and classification. By effectively integrating spatial graph modeling, temporal sequence feature extraction, and self-attention mechanisms, the model outperforms conventional deep learning approaches in both short-term and long-term forecasting tasks. Its strong performance provides accurate drought early warnings, assisting agricultural management authorities in formulating efficient water resource management strategies and optimizing irrigation plans. This contributes to safeguarding agricultural production and mitigating the potential adverse effects of agricultural drought.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Recognition of Sugarcane Leaf Diseases in Complex Backgrounds Based on Deep Network Ensembles
    MA Weiwei, CHEN Yue, WANG Yongmei
    Smart Agriculture    2025, 7 (1): 136-145.   DOI: 10.12133/j.smartag.SA202411026
    Abstract148)   HTML24)    PDF(pc) (1385KB)(80)       Save

    [Objective] Sugarcane is an important cash crop, and its health status affects crop yields. However, under natural environmental conditions, the identification of sugarcane leaf diseases is a challenging problem. There are various issues such as disease spots on sugarcane leaves being blocked and interference from lighting, which make it extremely difficult to comprehensively obtain disease information, thus significantly increasing the difficulty of disease identification. Early image recognition algorithms cannot accurately extract disease features and are prone to misjudgment and missed judgment in practical applications. To solve the problem of identifying sugarcane leaf diseases under natural conditions and break through the limitations of traditional methods, a novel identification model, XEffDa was proposed in this research. [Methods] The XEffDa model proposed implemented a series of improvement measures based on the ensemble learning framework, aiming to significantly improve the accuracy of classifying and identifying sugarcane leaf diseases. Firstly, the images in the sugarcane leaf disease dataset under natural conditions were pre-processed. Real-time data augmentation techniques were used to expand the scale of the dataset. Meanwhile, HSV image segmentation and edge-processing techniques were adopted to effectively remove redundant backgrounds and interference factors in the images. Considering that sugarcane leaf disease images were fine-grained images, in order to fully extract the semantic information of the images, the transfer learning strategy was employed. The pre-trained models of EfficientNetB0, Xception, and DenseNet201 were loaded respectively, and with the help of the pre-trained weight parameters based on the ImageNet dataset, the top layers of the models were frozen. The performance of the validation set was monitored through the Bayesian optimization method, and the parameters of the top-layer structure were replaced, thus achieving a good balance between optimizing the number of model parameters and the overall performance. In the top-layer structure, the improved ElasticNet regularization and Dropout layer were integrated. These two mechanisms cooperated with each other to double-suppress overfitting and significantly enhance the generalization ability of the model. During the training process, the MSprop optimizer was selected and combined with the sparse categorical cross - entropy loss function to better adapt to the multi-classification problem of sugarcane disease identification. After each model completed training independently, an exponential weight-allocation strategy was used to organically integrate the prediction features of each model and accurately map them to the final disease categories. To comprehensively evaluate the model performance, the accuracy indicator was continuously monitored, and an early-stopping mechanism was introduced to avoid overfitting and further strengthen the generalization ability of the model. Through the implementation of this series of refined optimization and integration strategies, the XEffDa model for sugarcane leaf diseases was finally successfully constructed. [Results and Discussions] The results of the confusion matrix showed that the XEffDa model performed very evenly across various disease categories, and all indicators achieved excellent results. Especially in the identification of red rot disease, its F1-Score was as high as 99.09%. This result was not only higher than that of other single models (such as EfficientNetB0 and Xception) but also superior to the combination of EfficientNetB0 and other deep networks (such as DenseNet121 and DenseNet201). This indicated that the XEffDa model significantly improved the ability to extract and classify features of complex pathological images by integrating the advantages of different network architectures. The comparison experiments of different models showed that the recognition accuracy of the XEffDa model reached 97.62%. Compared with the single models of EfficientNetB0 and Xception, as well as the combined models of EfficientNetB0 and other deep networks, the recognition accuracy increased by 9.96, 6.04, 8.09, 4.19, and 1.78 percentage points, respectively. The fusion experiments further showed that the accuracy, precision, recall, and F1-Score of the network improved by ElasticNet regularization increased by 3.76, 3.76, 3.67, and 3.72 percentage points respectively compared with the backbone network. The results of the maximum-probability scatter plot showed that the proportion of the maximum prediction probability value not lower than 0.5 was as high as 99.4%. [Conclusions] The XEffDa model demonstrated stronger robustness and stability. In the identification task of small sugarcane leaf disease datasets, it showed good generalization ability. This model can provide a powerful reference for the accurate prevention and control of sugarcane crop leaf diseases in practical scenarios, and it has positive significance for promoting the intelligent and precise management of sugarcane production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Advances, Problems and Challenges of Precise Estrus Perception and Intelligent Identification Technology for Cows
    ZHANG Zhiyong, CAO Shanshan, KONG Fantao, LIU Jifang, SUN Wei
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202305005
    Online available: 08 January 2025

    Research on Defogging Remote Sensing Images Based on A Hybrid Attention-Based Generative Adversarial Network
    MA Liu, MAO Kebiao, GUO Zhonghua
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202410011
    Online available: 24 January 2025

    Smart Agriculture    2025, 7 (1): 0-1.  
    Abstract99)            Save
    Related Articles | Metrics | Comments0
    Obstacle Avoidance Control Method of Electric Skid-Steering Chassis Based on Fuzzy Logic Control
    LI Lei, SHE Xiaoming, TANG Xinglong, ZHANG Tao, DONG Jiwei, GU Yuchuan, ZHOU Xiaohui, FENG Wei, YANG Qinghui
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202408003
    Online available: 27 December 2024

    Chilli-YOLO: An Intelligent Maturity Detection Algorithm for Field-Grown Chilli Based on Improved YOLOv10
    SI Chaoguo, LIU Mengchen, WU Huarui, MIAO Yisheng, ZHAO Chunjiang
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202411002
    Online available: 24 March 2025

    Improvement of HLM modeling for Winter Wheat Yield Estimation Under Drought Conditions
    ZHAO Peiqin, LIU Changbin, ZHENG Jie, MENG Yang, MEI Xin, TAO Ting, ZHAO Qian, MEI Guangyuan, YANG Xiaodong
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202408009
    Online available: 24 January 2025

    Bi-Intentional Modeling and Knowledge Graph Diffusion for Rice Variety Selection and Breeding Recommendation
    QIAO Lei, CHEN Lei, YUAN Yuan
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202412025
    Online available: 03 March 2025

    Analysis of the Spatial and Temporal Patterns of Grain Production and Their Influencing Factors in Sichuan Province
    ZHENG Ling, MA Qianran, JIANG Tao, LIU Xiaojing, MOU Jiahui, WANG Canhui, LAN Yu
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202411013
    Online available: 18 March 2025

    Smart Agriculture    2024, 6 (5): 0-0.  
    Abstract33)            Save
    Related Articles | Metrics | Comments0
    Localization of Pruning Points of High Spindle Apple Trees in Dormant Period Based on Pictures and 3D Point Cloud
    LIU Long, WANG Ning, WANG Jiacheng, CAO Yuheng, ZHANG Kai, KANG Feng, WANG Yaxiong
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501022
    Online available: 08 April 2025

    Smart Agriculture    2024, 6 (4): 0-0.  
    Abstract24)            Save
    Related Articles | Metrics | Comments0
    Knowledge Graph Driven Grain Big Data Applications:Overview and Perspective
    YANG Chenxue, LI Xian, ZHOU Qingbo
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501004
    Online available: 25 April 2025

    Spatiotemporal Pattern and Multi-Scenario Simulation of Land Use Conflicts: A Case Study of Shandong Section of The Yellow River Basin
    DONG Guanglong, YIN Haiyang, YAO Rongyan, YUAN Chenzhao, QU Chengchuang, TIAN Yuan, JIA Min
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202409007
    Online available: 14 April 2025

    Smart Agriculture    2024, 6 (3): 0-0.  
    Abstract17)            Save
    Related Articles | Metrics | Comments0
    Research Progress on Remote Sensing Monitoring and Intelligent Decision-Making Algorithms for Rice Production
    ZHAO Bingting, HUA Chuanhai, YE Chenyang, XIONG Yuchun, QIAN Tao, CHENG Tao, YAO Xia, ZHENG Hengbiao, ZHU Yan, CAO Weixing, JIANG Chongya
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501002
    Online available: 29 April 2025

    Design and Test of Rotating Envelope Comb Stripping Tobacco Picking Mechanism
    WANG Xiaohan, RAN Yunliang, GE Chao, GUO Ting, LIU Yihao, CHEN Du, WANG Shumao
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501020
    Online available: 24 April 2025

    Key Technologies and Prospects of Laser Weeding Robots
    YU Zhongyi, WANG Hongyu, HE Xiongkui, ZHAO Lei, WANG Yuanyuan, SUN Hai
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202410031
    Online available: 29 April 2025

    Agricultural Drought Monitoring in Arid Irrigated Areas Based on TVDI Combined with ICEEMDAN-ARIMA Model
    WEI Yuxin, LI Qiao, TAO Hongfei, LU Chunlei, LUO Xu, MAHEMUJIANG Aihemaiti, JIANG Youwei
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202502005
    Online available: 30 April 2025