Welcome to Smart Agriculture 中文

Top Read Articles

    Published in last 1 year |  In last 2 years |  In last 3 years |  All
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Vegetable Crop Growth Modeling in Digital Twin Platform Based on Large Language Model Inference
    ZHAO Chunjiang, LI Jingchen, WU Huarui, YANG Yusen
    Smart Agriculture    2024, 6 (6): 63-71.   DOI: 10.12133/j.smartag.SA202410008
    Abstract2184)   HTML187)    PDF(pc) (1460KB)(1698)       Save

    [Objective] In the era of digital agriculture, real-time monitoring and predictive modeling of crop growth are paramount, especially in autonomous farming systems. Traditional crop growth models, often constrained by their reliance on static, rule-based methods, fail to capture the dynamic and multifactorial nature of vegetable crop growth. This research tried to address these challenges by leveraging the advanced reasoning capabilities of pre-trained large language models (LLMs) to simulate and predict vegetable crop growth with accuracy and reliability. Modeling the growth of vegetable crops within these platforms has historically been hindered by the complex interactions among biotic and abiotic factors. [Methods] The methodology was structured in several distinct phases. Initially, a comprehensive dataset was curated to include extensive information on vegetable crop growth cycles, environmental conditions, and management practices. This dataset incorporates continuous data streams such as soil moisture, nutrient levels, climate variables, pest occurrence, and historical growth records. By combining these data sources, the study ensured that the model was well-equipped to understand and infer the complex interdependencies inherent in crop growth processes. Then, advanced techniques was emploied for pre-training and fine-tuning LLMs to adapt them to the domain-specific requirements of vegetable crop modeling. A staged intelligent agent ensemble was designed to work within the digital twin platform, consisting of a central managerial agent and multiple stage-specific agents. The managerial agent was responsible for identifying transitions between distinct growth stages of the crops, while the stage-specific agents were tailored to handle the unique characteristics of each growth phase. This modular architecture enhanced the model's adaptability and precision, ensuring that each phase of growth received specialized attention and analysis. [Results and Discussions] The experimental validation of this method was conducted in a controlled agricultural setting at the Xiaotangshan Modern Agricultural Demonstration Park in Beijing. Cabbage (Zhonggan 21) was selected as the test crop due to its significance in agricultural production and the availability of comprehensive historical growth data. Over five years, the dataset collected included 4 300 detailed records, documenting parameters such as plant height, leaf count, soil conditions, irrigation schedules, fertilization practices, and pest management interventions. This dataset was used to train the LLM-based system and evaluate its performance using ten-fold cross-validation. The results of the experiments demonstrating the efficacy of the proposed system in addressing the complexities of vegetable crop growth modeling. The LLM-based model achieved 98% accuracy in predicting crop growth degrees and a 99.7% accuracy in identifying growth stages. These metrics significantly outperform traditional machine learning approaches, including long short-term memory (LSTM), XGBoost, and LightGBM models. The superior performance of the LLM-based system highlights its ability to reason over heterogeneous data inputs and make precise predictions, setting a new benchmark for crop modeling technologies. Beyond accuracy, the LLM-powered system also excels in its ability to simulate growth trajectories over extended periods, enabling farmers and agricultural managers to anticipate potential challenges and make proactive decisions. For example, by integrating real-time sensor data with historical patterns, the system can predict how changes in irrigation or fertilization practices will impact crop health and yield. This predictive capability is invaluable for optimizing resource allocation and mitigating risks associated with climate variability and pest outbreaks. [Conclusions] The study emphasizes the importance of high-quality data in achieving reliable and generalizable models. The comprehensive dataset used in this research not only captures the nuances of cabbage growth but also provides a blueprint for extending the model to other crops. In conclusion, this research demonstrates the transformative potential of combining large language models with digital twin technology for vegetable crop growth modeling. By addressing the limitations of traditional modeling approaches and harnessing the advanced reasoning capabilities of LLMs, the proposed system sets a new standard for precision agriculture. Several avenues also are proposed for future work, including expanding the dataset, refining the model architecture, and developing multi-crop and multi-region capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Artificial Intelligence-Driven High-Quality Development of New-Quality Productivity in Animal Husbandry: Restraining Factors, Generation Logic and Promotion Paths
    LIU Jifang, ZHOU Xiangyang, LI Min, HAN Shuqing, GUO Leifeng, CHI Liang, YANG Lu, WU Jianzhai
    Smart Agriculture    2025, 7 (1): 165-177.   DOI: 10.12133/j.smartag.SA202407010
    Abstract1635)   HTML38)    PDF(pc) (1692KB)(733)       Save

    [Significance] Developing new-quality productivity is of great significance for promoting high-quality development of animal husbandry. However, there is currently limited research on new-quality productivity in animal husbandry, and there is a lack of in-depth analysis on its connotation, characteristics, constraints, and promotion path. [Progress] This article conducts a systematic study on the high-quality development of animal husbandry productivity driven by artificial intelligence. The new-quality productivity of animal husbandry is led by cutting-edge technological innovations such as biotechnology, information technology, and green technology, with digitalization, greening, and ecologicalization as the direction of industrial upgrading. Its basic connotation is manifested as higher quality workers, more advanced labor materials, and a wider range of labor objects. Compared with traditional productivity, the new-quality productivity of animal husbandry is an advanced productivity guided by technological innovation, new development concepts, and centered on the improvement of total factor productivity. It has significant characteristics of high production efficiency, good industrial benefits, and strong sustainable development capabilities. China's new-quality productivity in animal husbandry has a good foundation for development, but it also faces constraints such as insufficient innovation in animal husbandry breeding technology, weak core competitiveness, low mechanization rate of animal husbandry, weak independent research and development capabilities of intelligent equipment, urgent demand for "machine replacement", shortcomings in the quantity and quality of animal husbandry talents, low degree of scale of animal husbandry, and limited level of intelligent management. Artificial intelligence in animal husbandry can be widely used in environmental control, precision feeding, health monitoring and disease prevention and control, supply chain optimization and other fields. Artificial intelligence, through revolutionary breakthroughs in animal husbandry technology represented by digital technology, innovative allocation of productivity factors in animal husbandry linked by data elements, and innovative allocation of productivity factors in animal husbandry adapted to the digital economy, has given birth to new-quality productivity in animal husbandry and empowered the high-quality development of animal husbandry. [Conclusions and Prospects] This article proposes a path to promote the development of new-quality productivity in animal husbandry by improving the institutional mechanism of artificial intelligence to promote the development of modern animal husbandry industry, strengthening the application of artificial intelligence in animal husbandry technology innovation and promotion, and improving the management level of artificial intelligence in the entire industry chain of animal husbandry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Key Technologies and Prospects of Laser Weeding Robots
    YU Zhongyi, WANG Hongyu, HE Xiongkui, ZHAO Lei, WANG Yuanyuan, SUN Hai
    Smart Agriculture    2025, 7 (2): 132-145.   DOI: 10.12133/j.smartag.SA202410031
    Abstract1627)   HTML53)    PDF(pc) (6774KB)(687)       Save

    [Significance] Grass damage in farmland seriously restricts the quality and yield of crop planting and production, and promotes the occurrence of pests and diseases. Weed control is a necessary measure for high yield and high quality of crops. Currently, there are five main weed control methods: Manual, biological, thermal, mechanical, and chemical weed control. Traditional chemical weed control methods are gradually limited due to soil pollution and ecological balance disruption. Intelligent laser weeding technology, with the characteristics of environmental protection, high efficiency, flexibility, and automation, as an emerging and promising ecological and environmental protection new object control method for field weeds, has become the core direction to replace chemical weeding in recent years. The laser weeding robot is the carrier of laser weeding technology, an important manifestation of the development of modern agriculture towards intelligence and precision, and has great application and promotion value. [Progress] Laser weeding is currently a research hotspot to develop and study key technologies and equipment for smart agriculture, and has achieved a series of significant results, greatly promoting the promotion and application of intelligent laser weeding robots in the field. Laser weed control technology achieves precise weed control through thermal, photochemical, and photodynamic effects. In this article, the research background of laser weeding was introduced, its key technologies, operation system and equipment were discussed in details, covering aspects such as operating principles, system architecture, seedling, weed recognition and localization, robot navigation and path planning, as well as actuator control technologies. Then, based on the current research status of laser weeding robots, the existing problems and development trends of intelligent laser weeding robots were prospected. [Conclusion and Prospect] Based on the different field grass conditions in different regions, a large number of indoor and outdoor experiments on laser weed control should be carried out in the future to further verify the technical effectiveness and feasibility of laser field weed control, providing support for the research and application of laser weed control equipment technology. Despite facing challenges such as high costs and poor environmental adaptability, with the integration of technologies such as artificial intelligence and the Internet of Things, as well as policy support, laser weeding is expected to become an important support for sustainable agricultural development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agri-QA Net: Multimodal Fusion Large Language Model Architecture for Crop Knowledge Question-Answering System
    WU Huarui, ZHAO Chunjiang, LI Jingchen
    Smart Agriculture    2025, 7 (1): 1-10.   DOI: 10.12133/j.smartag.SA202411005
    Abstract1558)   HTML197)    PDF(pc) (1010KB)(1805)       Save

    [Objective] As agriculture increasingly relies on technological innovations to boost productivity and ensure sustainability, farmers need efficient and accurate tools to aid their decision-making processes. A key challenge in this context is the retrieval of specialized agricultural knowledge, which can be complex and diverse in nature. Traditional agricultural knowledge retrieval systems have often been limited by the modalities they utilize (e.g., text or images alone), which restricts their effectiveness in addressing the wide range of queries farmers face. To address this challenge, a specialized multimodal question-answering system tailored for cabbage cultivation was proposed. The system, named Agri-QA Net, integrates multimodal data to enhance the accuracy and applicability of agricultural knowledge retrieval. By incorporating diverse data modalities, Agri-QA Net aims to provide a holistic approach to agricultural knowledge retrieval, enabling farmers to interact with the system using multiple types of input, ranging from spoken queries to images of crop conditions. By doing so, it helps address the complexity of real-world agricultural environments and improves the accessibility of relevant information. [Methods] The architecture of Agri-QA Net was built upon the integration of multiple data modalities, including textual, auditory, and visual data. This multifaceted approach enables the system to develop a comprehensive understanding of agricultural knowledge, allowed the system to learn from a wide array of sources, enhancing its robustness and generalizability. The system incorporated state-of-the-art deep learning models, each designed to handle one specific type of data. Bidirectional Encoder Representations from Transformers (BERT)'s bidirectional attention mechanism allowed the model to understand the context of each word in a given sentence, significantly improving its ability to comprehend complex agricultural terminology and specialized concepts. The system also incorporated acoustic models for processing audio inputs. These models analyzed the spoken queries from farmers, allowing the system to understand natural language inputs even in noisy, non-ideal environments, which was a common challenge in real-world agricultural settings. Additionally, convolutional neural networks (CNNs) were employed to process images from various stages of cabbage growth. CNNs were highly effective in capturing spatial hierarchies in images, making them well-suited for tasks such as identifying pests, diseases, or growth abnormalities in cabbage crops. These features were subsequently fused in a Transformer-based fusion layer, which served as the core of the Agri-QA Net architecture. The fusion process ensured that each modality—text, audio, and image—contributes effectively to the final model's understanding of a given query. This allowed the system to provide more nuanced answers to complex agricultural questions, such as identifying specific crop diseases or determining the optimal irrigation schedules for cabbage crops. In addition to the fusion layer, cross-modal attention mechanisms and domain-adaptive techniques were incorporated to refine the model's ability to understand and apply specialized agricultural knowledge. The cross-modal attention mechanism facilitated dynamic interactions between the text, audio, and image data, ensuring that the model paid attention to the most relevant features from each modality. Domain-adaptive techniques further enhanced the system's performance by tailoring it to specific agricultural contexts, such as cabbage farming, pest control, or irrigation management. [Results and Discussions] The experimental evaluations demonstrated that Agri-QA Net outperforms traditional single-modal or simple multimodal models in agricultural knowledge tasks. With the support of multimodal inputs, the system achieved an accuracy rate of 89.5%, a precision rate of 87.9%, a recall rate of 91.3%, and an F1-Score of 89.6%, all of which are significantly higher than those of single-modality models. The integration of multimodal data significantly enhanced the system's capacity to understand complex agricultural queries, providing more precise and context-aware answers. The addition of cross-modal attention mechanisms enabled for more nuanced and dynamic interaction between the text, audio, and image data, which in turn improved the model's understanding of ambiguous or context-dependent queries, such as disease diagnosis or crop management. Furthermore, the domain-adaptive technique enabled the system to focus on specific agricultural terminology and concepts, thereby enhancing its performance in specialized tasks like cabbage cultivation and pest control. The case studies presented further validated the system's ability to assist farmers by providing actionable, domain-specific answers to questions, demonstrating its practical application in real-world agricultural scenarios. [Conclusions] The proposed Agri-QA Net framework is an effective solution for addressing agricultural knowledge questions, especially in the domain of cabbage cultivation. By integrating multimodal data and leveraging advanced deep learning techniques, the system demonstrates a high level of accuracy and adaptability. This study not only highlights the potential of multimodal fusion in agriculture but also paves the way for future developments in intelligent systems designed to support precision farming. Further work will focus on enhancing the model's performance by expanding the dataset to include more diverse agricultural scenarios, refining the handling of dialectical variations in audio inputs, and improving the system's ability to detect rare crop diseases. The ultimate goal is to contribute to the modernization of agricultural practices, offering farmers more reliable and effective tools to solve the challenges in crop management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Tomato Growth Height Prediction Method by Phenotypic Feature Extraction Using Multi-modal Data
    GONG Yu, WANG Ling, ZHAO Rongqiang, YOU Haibo, ZHOU Mo, LIU Jie
    Smart Agriculture    2025, 7 (1): 97-110.   DOI: 10.12133/j.smartag.SA202410032
    Abstract1361)   HTML74)    PDF(pc) (1307KB)(242)       Save

    [Objective] Accurate prediction of tomato growth height is crucial for optimizing production environments in smart farming. However, current prediction methods predominantly rely on empirical, mechanistic, or learning-based models that utilize either images data or environmental data. These methods fail to fully leverage multi-modal data to capture the diverse aspects of plant growth comprehensively. [Methods] To address this limitation, a two-stage phenotypic feature extraction (PFE) model based on deep learning algorithm of recurrent neural network (RNN) and long short-term memory (LSTM) was developed. The model integrated environment and plant information to provide a holistic understanding of the growth process, emploied phenotypic and temporal feature extractors to comprehensively capture both types of features, enabled a deeper understanding of the interaction between tomato plants and their environment, ultimately leading to highly accurate predictions of growth height. [Results and Discussions] The experimental results showed the model's effectiveness: When predicting the next two days based on the past five days, the PFE-based RNN and LSTM models achieved mean absolute percentage error (MAPE) of 0.81% and 0.40%, respectively, which were significantly lower than the 8.00% MAPE of the large language model (LLM) and 6.72% MAPE of the Transformer-based model. In longer-term predictions, the 10-day prediction for 4 days ahead and the 30-day prediction for 12 days ahead, the PFE-RNN model continued to outperform the other two baseline models, with MAPE of 2.66% and 14.05%, respectively. [Conclusions] The proposed method, which leverages phenotypic-temporal collaboration, shows great potential for intelligent, data-driven management of tomato cultivation, making it a promising approach for enhancing the efficiency and precision of smart tomato planting management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Seedling Stage Corn Line Detection Method Based on Improved YOLOv8
    LI Hongbo, TIAN Xin, RUAN Zhiwen, LIU Shaowen, REN Weiqi, SU Zhongbin, GAO Rui, KONG Qingming
    Smart Agriculture    2024, 6 (6): 72-84.   DOI: 10.12133/j.smartag.SA202408008
    Abstract1260)   HTML104)    PDF(pc) (3458KB)(323)       Save

    [Objective] Crop line extraction is critical for improving the efficiency of autonomous agricultural machines in the field. However, traditional detection methods struggle to maintain high accuracy and efficiency under challenging conditions, such as strong light exposure and weed interference. The aims are to develop an effective crop line extraction method by combining YOLOv8-G, Affinity Propagation, and the Least Squares method to enhance detection accuracy and performance in complex field environments. [Methods] The proposed method employs machine vision techniques to address common field challenges. YOLOv8-G, an improved object detection algorithm that combines YOLOv8 and GhostNetV2 for lightweight, high-speed performance, was used to detect the central points of crops. These points were then clustered using the Affinity Propagation algorithm, followed by the application of the Least Squares method to extract the crop lines. Comparative tests were conducted to evaluate multiple backbone networks within the YOLOv8 framework, and ablation studies were performed to validate the enhancements made in YOLOv8-G. [Results and Discussions] The performance of the proposed method was compared with classical object detection and clustering algorithms. The YOLOv8-G algorithm achieved average precision (AP) values of 98.22%, 98.15%, and 97.32% for corn detection at 7, 14, and 21 days after emergence, respectively. Additionally, the crop line extraction accuracy across all stages was 96.52%. These results demonstrate the model's ability to maintain high detection accuracy despite challenging conditions in the field. [Conclusions] The proposed crop line extraction method effectively addresses field challenges such as lighting and weed interference, enabling rapid and accurate crop identification. This approach supports the automatic navigation of agricultural machinery, offering significant improvements in the precision and efficiency of field operations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospects of Key Technologies for Rice Smart Unmanned Farms
    YU Fenghua, XU Tongyu, GUO Zhonghui, BAI Juchi, XIANG Shuang, GUO Sien, JIN Zhongyu, LI Shilong, WANG Shikuan, LIU Meihan, HUI Yinxuan
    Smart Agriculture    2024, 6 (6): 1-22.   DOI: 10.12133/j.smartag.SA202410018
    Abstract1250)   HTML175)    PDF(pc) (3047KB)(2742)       Save

    [Significance] Rice smart unmanned farm is the core component of smart agriculture, and it is a key path to realize the modernization of rice production and promote the high-quality development of agriculture. Leveraging advanced information technologies such as the Internet of Things (IoT) and artificial intelligence (AI), these farms enable deep integration of data-driven decision making and intelligent machines. This integration creates an unmanned production system that covers the entire process from planting and managing rice crops to harvesting, greatly improving the efficiency and precision of rice cultivation. [Progress] This paper systematically sorted out the key technologies of rice smart unmanned farms in the three main links of pre-production, production and post-production, and the key technologies of pre-production mainly include the construction of high-standard farmland, unmanned nursery, land leveling, and soil nutrient testing. The construction of high-standard farmland is the foundation of the physical environment of the smart unmanned farms of rice, which provides perfect operating environment for the operation of modernized smart farm machinery through the reasonable layout of the field roads, good drainage and irrigation systems, and the scientific planting structure. Agricultural machine operation provides a perfect operating environment. The technical level of unmanned nursery directly determines the quality of rice cultivation and harvesting in the later stage, and a variety of rice seeding machines and nursery plate setting machines have been put into use. Land leveling technology can improve the growing environment of rice and increase the land utilization rate, and the current land leveling technology through digital sensing and path planning technology, which improves the operational efficiency and reduces the production cost at the same time. Soil nutrient detection technology is mainly detected by electrochemical analysis and spectral analysis, but both methods have their advantages and disadvantages, how to integrate the two methods to achieve an all-round detection of soil nutrient content is the main direction of future research. The key technologies in production mainly include rice dry direct seeding, automated transplanting, precise variable fertilization, intelligent irrigation, field weed management, and disease diagnosis. Among them, the rice dry direct seeding technology requires the planter to have high precision and stability to ensure reasonable seeding depth and density. Automated rice transplanting technology mainly includes three ways: root washing seedling machine transplanting, blanket seedling machine transplanting, and potting blanket seedling machine transplanting; at present, the incidence of problems in the automated transplanting process should be further reduced, and the quality and efficiency of rice machine transplanting should be improved. Precision variable fertilization technology is mainly composed of three key technologies: information perception, prescription decision-making and precise operation, but there are still fewer cases of unmanned farms combining the three technologies, and in the future, the main research should be on the method of constructing the whole process operation system of variable fertilization. The smart irrigation system is based on the water demand of the whole life cycle of rice to realize adaptive irrigation control, and the current smart irrigation technology can automatically adjust the irrigation strategy through real-time monitoring of soil, climate and crop growth conditions to further improve irrigation efficiency and agricultural production benefits. The field weed management and disease diagnosis technology mainly recognizes rice weeds as well as diseases through deep learning and other methods, and combines them with precision application technology for prevention and intervention. Post-production key technologies mainly include rice yield estimation, unmanned harvesting, rice storage and processing quality testing. Rice yield estimation technology is mainly used to predict yield by combining multi-source data and algorithms, but there are still problems such as the difficulty of integrating multi-source data, which requires further research. In terms of unmanned aircraft harvesting technology, China's rice combine harvester market has tended to stabilize, and the safety of the harvester's autopilot should be further improved in the future. Rice storage and processing quality detection technology mainly utilizes spectral technology and machine vision technology to detect spectra and images, and future research can combine deep learning and multimodal fusion technology to improve the machine vision system's ability and adaptability to recognize the appearance characteristics of rice. [Conclusions and Prospects] This paper reviews the researches of the construction of intelligent unmanned rice farms at home and abroad in recent years, summarizes the main difficulties faced by the key technologies of unmanned farms in practical applications, analyzes the challenges encountered in the construction of smart unmanned farms, summarizes the roles and responsibilities of the government, enterprises, scientific research institutions, cooperatives and other subjects in promoting the construction of intelligent unmanned rice farms, and puts forward relevant suggestions. It provides certain support and development ideas for the construction of intelligent unmanned rice farms in China.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Intelligent Decision-Making Method for Personalized Vegetable Crop Water and Fertilizer Management Based on Large Language Models
    WU Huarui, LI Jingchen, YANG Yusen
    Smart Agriculture    2025, 7 (1): 11-19.   DOI: 10.12133/j.smartag.SA202410007
    Abstract1249)   HTML98)    PDF(pc) (1181KB)(534)       Save

    [Objective] The current crop management faces the challenges of difficulty in capturing personalized needs and the lack of flexibility in the decision-making process. To address the limitations of conventional precision agriculture systems, optimize key aspects of agricultural production, including crop yield, labor efficiency, and water and fertilizer use, while ensure sustainability and adaptability to diverse farming conditions, in this research, an intelligent decision-making method was presents for personalized vegetable crop water and fertilizer management using large language model (LLM) by integrating user-specific preferences into decision-making processes through natural language interactions. [Methods] The method employed artificial intelligence techniques, combining natural language processing (NLP) and reinforcement learning (RL). Initially, LLM engaged users through structured dialogues to identify their unique preferences related to crop production goals, such as maximizing yield, reducing resource consumption, or balancing multiple objectives. These preferences were then modeled as quantifiable parameters and incorporated into a multi-objective optimization framework. To realize this framework, proximal policy optimization (PPO) was applied within a reinforcement learning environment to develop dynamic water and fertilizer management strategies. Training was conducted in the gym-DSSAT simulation platform, a system designed for agricultural decision support. The RL model iteratively learned optimal strategies by interacting with the simulation environment, adjusting to diverse conditions and balancing conflicting objectives effectively. To refine the estimation of user preferences, the study introduced a two-phase process comprising prompt engineering to guide user responses and adversarial fine-tuning for enhanced accuracy. These refinements ensured that user inputs were reliably transformed into structured decision-making criteria. Customized reward functions were developed for RL training to address specific agricultural goals. The reward functions account for crop yield, resource efficiency, and labor optimization, aligning with the identified user priorities. Through iterative training and simulation, the system dynamically adapted its decision-making strategies to varying environmental and operational conditions. [Results and Discussions] The experimental evaluation highlighted the system's capability to effectively personalize crop management strategies. Using simulations, the method demonstrated significant improvements over traditional approaches. The LLM-based model accurately captured user-specific preferences through structured natural language interactions, achieving reliable preference modeling and integration into the decision-making process. The system's adaptability was evident in its ability to respond dynamically to changes in user priorities and environmental conditions. For example, in scenarios emphasizing resource conservation, water and fertilizer use were significantly reduced without compromising crop health. Conversely, when users prioritized yield, the system optimized irrigation and fertilization schedules to enhance productivity. These results showcased the method's flexibility and its potential to balance competing objectives in complex agricultural settings. Additionally, the integration of user preferences into RL-based strategy development enabled the generation of tailored management plans. These plans aligned with diverse user goals, including maximizing productivity, minimizing resource consumption, and achieving sustainable farming practices. The system's multi-objective optimization capabilities allowed it to navigate trade-offs effectively, providing actionable insights for decision-making. The experimental validation also demonstrated the robustness of the PPO algorithm in training the RL model. The system's strategies were refined iteratively, resulting in consistent performance improvements across various scenarios. By leveraging LLM to capture nuanced user preferences and combining them with RL for adaptive decision-making, the method bridges the gap between generic precision agriculture solutions and personalized farming needs. [Conclusions] This study established a novel framework for intelligent decision-making in agriculture, integrating LLM with reinforcement learning to address personalized crop management challenges. By accurately capturing user-specific preferences and dynamically adapting to environmental and operational variables, the method offers a transformative approach to optimizing agricultural productivity and sustainability. Future work will focus on expanding the system's applicability to a wider range of crops and environmental contexts, enhancing the interpretability of its decision-making processes, and facilitating integration with real-world agricultural systems. These advancements aim to further refine the precision and impact of intelligent agricultural decision-making systems, supporting sustainable and efficient farming practices globally.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Large Language Model Based on Precise Knowledge Retrieval and Knowledge Collaborative Generation
    JIANG Jingchi, YAN Lian, LIU Jie
    Smart Agriculture    2025, 7 (1): 20-32.   DOI: 10.12133/j.smartag.SA202410025
    Abstract1127)   HTML70)    PDF(pc) (2081KB)(895)       Save

    [Objective] The rapid advancement of large language models (LLMs) has positioned them as a promising novel research paradigm in smart agriculture, leveraging their robust cognitive understanding and content generative capabilities. However, due to the lack of domain-specific agricultural knowledge, general LLMs often exhibit factual errors or incomplete information when addressing specialized queries, which is particularly prominent in agricultural applications. Therefore, enhancing the adaptability and response quality of LLMs in agricultural applications has become an important research direction. [Methods] To improve the adaptability and precision of LLMs in the agricultural applications, an innovative approach named the knowledge graph-guided agricultural LLM (KGLLM) was proposed. This method integrated information entropy for effective knowledge filtering and applied explicit constraints on content generation during the decoding phase by utilizing semantic information derived from an agricultural knowledge graph. The process began by identifying and linking key entities from input questions to the agricultural knowledge graph, which facilitated the formation of knowledge inference paths and the development of question-answering rationales. A critical aspect of this approach was ensuring the validity and reliability of the external knowledge incorporated into the model. This was achieved by evaluating the entropy difference in the model's outputs before and after the introduction of each piece of knowledge. Knowledge that didn't enhance the certainty of the answers was systematically filtered out. The knowledge paths that pass this entropy evaluation were used to adjust the token prediction probabilities, prioritizing outputs that were closely aligned with the structured knowledge. This allowed the knowledge graph to exert explicit guidance over the LLM's outputs, ensuring higher accuracy and relevance in agricultural applications. [Results and Discussions] The proposed knowledge graph-guided technique was implemented on five mainstream general-purpose LLMs, including open-source models such as Baichuan, ChatGLM, and Qwen. These models were compared with state-of-the-art knowledge graph-augmented generation methods to evaluate the effectiveness of the proposed approach. The results demonstrate that the proposed knowledge graph-guided approach significantly improved several key performance metrics of fluency, accuracy, factual correctness, and domain relevance. Compared to GPT-4o, the proposed method achieved notable improvements by an average of 2.592 3 in Mean BLEU, 2.815 1 in ROUGE, and 9.84% in BertScore. These improvements collectively signify that the proposed approach effectively leverages agricultural domain knowledge to refine the outputs of general-purpose LLMs, making them more suitable for agricultural applications. Ablation experiments further validated that the knowledge-guided agricultural LLM not only filtered out redundant knowledge but also effectively adjusts token prediction distributions during the decoding phase. This enhanced the adaptability of general-purpose LLMs in agriculture contexts and significantly improves the interpretability of their responses. The knowledge filtering and knowledge graph-guided model decoding method proposed in this study, which was based on information entropy, effectively identifies and selects knowledge that carried more informational content through the comparison of information entropy.Compared to existing technologies in the agricultural field, this method significantly reduced the likelihood of "hallucination" phenomena during the generation process. Furthermore, the guidance of the knowledge graph ensured that the model's generated responses were closely related to professional agricultural knowledge, thereby avoiding vague and inaccurate responses generated from general knowledge. For instance, in the application of pest and disease control, the model could accurately identify the types of crop diseases and corresponding control measures based on the guided knowledge path, thereby providing more reliable decision support. [Conclusions] This study provides a valuable reference for the construction of future agricultural large language models, indicating that the knowledge graphs guided mehtod has the potential to enhance the domain adaptability and answer quality of models. Future research can further explore the application of similar knowledge-guided strategies in other vertical fields to enhance the adaptability and practicality of LLMs across various professional domains.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospect of Multi-robot Collaborative SLAM in Complex Agricultural Scenarios
    MA Nan, CAO Shanshan, BAI Tao, KONG Fantao, SUN Wei
    Smart Agriculture    2024, 6 (6): 23-43.   DOI: 10.12133/j.smartag.SA202406005
    Abstract1088)   HTML96)    PDF(pc) (2300KB)(8854)       Save

    [Significance] The rapid development of artificial intelligence and automation has greatly expanded the scope of agricultural automation, with applications such as precision farming using unmanned machinery, robotic grazing in outdoor environments, and automated harvesting by orchard-picking robots. Collaborative operations among multiple agricultural robots enhance production efficiency and reduce labor costs, driving the development of smart agriculture. Multi-robot simultaneous localization and mapping (SLAM) plays a pivotal role by ensuring accurate mapping and localization, which are essential for the effective management of unmanned farms. Compared to single-robot SLAM, multi-robot systems offer several advantages, including higher localization accuracy, larger sensing ranges, faster response times, and improved real-time performance. These capabilities are particularly valuable for completing complex tasks efficiently. However, deploying multi-robot SLAM in agricultural settings presents significant challenges. Dynamic environmental factors, such as crop growth, changing weather patterns, and livestock movement, increase system uncertainty. Additionally, agricultural terrains vary from open fields to irregular greenhouses, requiring robots to adjust their localization and path-planning strategies based on environmental conditions. Communication constraints, such as unstable signals or limited transmission range, further complicate coordination between robots. These combined challenges make it difficult to implement multi-robot SLAM effectively in agricultural environments. To unlock the full potential of multi-robot SLAM in agriculture, it is essential to develop optimized solutions that address the specific technical demands of these scenarios. [Progress] Existing review studies on multi-robot SLAM mainly focus on a general technological perspective, summarizing trends in the development of multi-robot SLAM, the advantages and limitations of algorithms, universally applicable conditions, and core issues of key technologies. However, there is a lack of analysis specifically addressing multi-robot SLAM under the characteristics of complex agricultural scenarios. This study focuses on the main features and applications of multi-robot SLAM in complex agricultural scenarios. The study analyzes the advantages and limitations of multi-robot SLAM, as well as its applicability and application scenarios in agriculture, focusing on four key components: multi-sensor data fusion, collaborative localization, collaborative map building, and loopback detection. From the perspective of collaborative operations in multi-robot SLAM, the study outlines the classification of SLAM frameworks, including three main collaborative types: centralized, distributed, and hybrid. Based on this, the study summarizes the advantages and limitations of mainstream multi-robot SLAM frameworks, along with typical scenarios in robotic agricultural operations where they are applicable. Additionally, it discusses key issues faced by multi-robot SLAM in complex agricultural scenarios, such as low accuracy in mapping and localization during multi-sensor fusion, restricted communication environments during multi-robot collaborative operations, and low accuracy in relative pose estimation between robots. [Conclusions and Prospects] To enhance the applicability and efficiency of multi-robot SLAM in complex agricultural scenarios, future research needs to focus on solving these critical technological issues. Firstly, the development of enhanced data fusion algorithms will facilitate improved integration of sensor information, leading to greater accuracy and robustness of the system. Secondly, the combination of deep learning and reinforcement learning techniques is expected to empower robots to better interpret environmental patterns, adapt to dynamic changes, and make more effective real-time decisions. Thirdly, large language models will enhance human-robot interaction by enabling natural language commands, improving collaborative operations. Finally, the integration of digital twin technology will support more intelligent path planning and decision-making processes, especially in unmanned farms and livestock management systems. The convergence of digital twin technology with SLAM is projected to yield innovative solutions for intelligent perception and is likely to play a transformative role in the realm of agricultural automation. This synergy is anticipated to revolutionize the approach to agricultural tasks, enhancing their efficiency and reducing the reliance on labor.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    A Lightweight Model for Detecting Small Targets of Litchi Pests Based on Improved YOLOv10n
    LI Zusheng, TANG Jishen, KUANG Yingchun
    Smart Agriculture    2025, 7 (2): 146-159.   DOI: 10.12133/j.smartag.SA202412003
    Abstract1084)   HTML127)    PDF(pc) (2262KB)(3419)       Save

    [Objective] The accuracy of identifying litchi pests is crucial for implementing effective control strategies and promoting sustainable agricultural development. However, the current detection of litchi pests is characterized by a high percentage of small targets, which makes target detection models challenging in terms of accuracy and parameter count, thus limiting their application in real-world production environments. To improve the identification efficiency of litchi pests, a lightweight target detection model YOLO-LP (YOLO-Litchi Pests) based on YOLOv10n was proposed. The model aimed to enhance the detection accuracy of small litchi pest targets in multiple scenarios by optimizing the network structure and loss function, while also reducing the number of parameters and computational costs. [Methods] Two classes of litchi insect pests (Cocoon and Gall) images were collected as datasets for modeling in natural scenarios (sunny, cloudy, post-rain) and laboratory environments. The original data were expanded through random scaling, random panning, random brightness adjustments, random contrast variations, and Gaussian blurring to balance the category samples and enhance the robustness of the model, generating a richer dataset named the CG dataset (Cocoon and Gall dataset). The YOLO-LP model was constructed after the following three improvements. Specifically, the C2f module of the backbone network (Backbone) in YOLOv10n was optimized and the C2f_GLSA module was constructed using the global-to-local spatial aggregation (GLSA) module to focus on small targets and enhance the differentiation between the targets and the backgrounds, while simultaneously reducing the number of parameters and computation. A frequency-aware feature fusion module (FreqFusion) was introduced into the neck network (Neck) of YOLOv10n and a frequency-aware path aggregation network (FreqPANet) was designed to reduce the complexity of the model and address the problem of fuzzy and shifted target boundaries. The SCYLLA-IoU (SIoU) loss function replaced the Complete-IoU (CIoU) loss function from the baseline model to optimize the target localization accuracy and accelerate the convergence of the training process. [Results and Discussions] YOLO-LP achieved 90.9%, 62.2%, and 59.5% for AP50, AP50:95, and AP-Small50:95 in the CG dataset, respectively, and 1.9%, 1.0%, and 1.2% higher than the baseline model. The number of parameters and the computational costs were reduced by 13% and 17%, respectively. These results suggested that YOLO-LP had a high accuracy and lightweight design. Comparison experiments with different attention mechanisms validated the effectiveness of the GLSA module. After the GLSA module was added to the baseline model, AP50, AP50:95, and AP-Small50:95 achieved the highest performance in the CG dataset, reaching 90.4%, 62.0%, and 59.5%, respectively. Experiment results comparing different loss functions showed that the SIoU loss function provided better fitting and convergence speed in the CG dataset. Ablation test results revealed that the validity of each model improvement and the detection performance of any combination of the three improvements was significantly better than the baseline model in the YOLO-LP model. The performance of the models was optimal when all three improvements were applied simultaneously. Compared to several mainstream models, YOLO-LP exhibited the best overall performance, with a model size of only 5.1 MB, 1.97 million parameters (Params), and a computational volume of 5.4 GFLOPs. Compared to the baseline model, the detection of the YOLO-LP performance was significantly improved across four multiple scenarios. In the sunny day scenario, AP50, AP50:95, and AP-Small50:95 increased by 1.9%, 1.0 %, and 2.0 %, respectively. In the cloudy day scenario, AP50, AP50:95, and AP-Small50:95 increased by 2.5%, 1.3%, and 1.3%, respectively. In the post-rain scenario, AP50, AP50:95, and AP-Small50:95 increased by 2.0%, 2.4%, and 2.4%, respectively. In the laboratory scenario, only AP50 increased by 0.7% over the baseline model. These findings indicated that YOLO-LP achieved higher accuracy and robustness in multi-scenario small target detection of litchi pests. [Conclusions] The proposed YOLO-LP model could improve detection accuracy and effectively reduce the number of parameters and computational costs. It performed well in small target detection of litchi pests and demonstrated strong robustness across different scenarios. These improvements made the model more suitable for deployment on resource-constrained mobile and edge devices. The model provided a valuable technical reference for small target detection of litchi pests in various scenarios.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Method for Calculating Semantic Similarity of Short Agricultural Texts Based on Transfer Learning
    JIN Ning, GUO Yufeng, HAN Xiaodong, MIAO Yisheng, WU Huarui
    Smart Agriculture    2025, 7 (1): 33-43.   DOI: 10.12133/j.smartag.SA202410026
    Abstract946)   HTML15)    PDF(pc) (1239KB)(427)       Save

    [Objective] Intelligent services of agricultural knowledge have emerged as a current hot research domain, serving as a significant support for the construction of smart agriculture. The platform "China Agricultural Technology Extension" provides users with efficient and convenient agricultural information consultation services via mobile terminals, and has accumulated a vast amount of Q&A data. These data are characterized by a huge volume of information, rapid update and iteration, and a high degree of redundancy, resulting in the platform encountering issues such as frequent repetitive questions, low timeliness of problem responses, and inaccurate information retrieval. There is an urgent requirement for a high-quality text semantic similarity calculation approach to confront these challenges and effectively enhance the information service efficiency and intelligent level of the platform. In view of the problems of incomplete feature extraction and lack of short agro-text annotation data sets in existing text semantic similarity calculation models, a semantic similarity calculation model for short agro-text, namely CWPT-SBERT, based on transfer learning and BERT pre-training model, was proposed. [Methods] CWPT-SBERT was based on Siamese architecture with identical left and right sides and shared parameters, which had the advantages of low structural complexity and high training efficiency. This network architecture effectively reduced the consumption of computational resources by sharing parameters and ensures that input texts were compared in the same feature space. CWPT-SBERT consisted of four main parts: Semantic enhancement layer, embedding layer, pooling layer, and similarity measurement layer. The CWPT method based on the word segmentation unit was proposed in the semantic enhancement layer to further divide Chinese characters into more fine-grained sub-units maximizing the semantic features in short Chinese text and effectively enhancing the model's understanding of complex Chinese vocabulary and character structures. In the embedding layer, a transfer learning strategy was used to extract features from agricultural short texts based on SBERT. It captured the semantic features of Chinese text in the general domain, and then generated a more suitable semantic feature vector representation after fine-tuning. Transfer learning methods to train models on large-scale general-purposed domain annotation datasets solve the problem of limited short agro-text annotation datasets and high semantic sparsity. The pooling layer used the average pooling strategy to map the high-dimensional semantic vector of Chinese short text to a low-dimensional vector space. The similarity measurement layer used the cosine similarity calculation method to measure the similarity between the semantic feature vector representations of the two output short texts, and the computed similarity degree was finally input into the loss function to guide model training, optimize model parameters, and improve the accuracy of similarity calculation. [Results and Discussions] For the task of calculating semantic similarity in agricultural short texts, on a dataset containing 19 968 pairs of short ago-texts, the CWPT-SBERT model achieved an accuracy rate of 97.18% and 96.93%, a recall rate of 97.14%, and an F1-Score value of 97.04%, which are higher than 12 models such as TextCNN_Attention, MaLSTM and SBERT. By analyzing the Pearson and Spearman coefficients of CWPT-SBERT, SBERT, SALBERT and SRoBERTa trained on short agro-text datasets, it could be observed that the initial training value of the CWPT-SBERT model was significantly higher than that of the comparison models and was close to the highest value of the comparison models. Moreover, it exhibited a smooth growth trend during the training process, indicating that CWPT-SBERT had strong correlation, robustness, and generalization ability from the initial state. During the training process, it could not only learn the features in the training data but also effectively apply these features to new domain data. Additionally, for ALBERT, RoBERTa and BERT models, fine-tuning training was conducted on short agro-text datasets, and optimization was performed by utilizing the morphological structure features to enrich text semantic feature expression. Through ablation experiments, it was evident that both optimization strategies could effectively enhance the performance of the models. By analyzing the attention weight heatmap of Chinese character morphological structure, the importance of Chinese character radicals in representing Chinese character attributes was highlighted, enhancing the semantic representation of Chinese characters in vector space. There was also complex correlation within the morphological structure of Chinese characters. [Conclusions] CWPT-SBERT uses transfer learning methods to solve the problem of limited short agro-text annotation datasets and high semantic sparsity. By leveraging the Chinese-oriented word segmentation method CWPT to break down Chinese characters, the semantic representation of word vectors is enhanced, and the semantic feature expression of short texts is enriched. CWPT-SBERT model has high accuracy of semantic similarity on small-scale short agro-text and obvious performance advantages, which provides an effective technical reference for semantic intelligence matching.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grain Production Big Data Platform: Progress and Prospects
    YANG Guijun, ZHAO Chunjiang, YANG Xiaodong, YANG Hao, HU Haitang, LONG Huiling, QIU Zhengjun, LI Xian, JIANG Chongya, SUN Liang, CHEN Lei, ZHOU Qingbo, HAO Xingyao, GUO Wei, WANG Pei, GAO Meiling
    Smart Agriculture    2025, 7 (2): 1-12.   DOI: 10.12133/j.smartag.SA202409014
    Abstract846)   HTML118)    PDF(pc) (1811KB)(2255)       Save

    [Significance] The explosive development of agricultural big data has accelerated agricultural production into a new era of digitalization and intelligentialize. Agricultural big data is the core element to promote agricultural modernization and the foundation of intelligent agriculture. As a new productive forces, big data enhances the comprehensive intelligent management decision-making during the whole process of grain production. But it faces the problems such as the indistinct management mechanism of grain production big data resources, the lack of the full-chain decision-making algorithm system and big data platform for the whole process and full elements of grain production. [Progress] Grain production big data platform is a comprehensive service platform that uses modern information technologies such as big data, Internet of Things (IoT), remote sensing and cloud computing to provide intelligent decision-making support for the whole process of grain production based on intelligent algorithms for data collection, processing, analysis and monitoring related to grain production. In this paper, the progress and challenges in grain production big data, monitoring and decision-making algorithms are reviewed, as well as big data platforms in China and worldwide. With the development of the IoT and high-resolution multi-modal remote sensing technology, the massive agricultural big data generated by the "Space-Air-Ground" Integrated Agricultural Monitoring System, has laid an important foundation for smart agriculture and promoted the shift of smart agriculture from model-driven to data-driven. However, there are still some issues in field management decision-making, such as the requirements for high spatio-temporal resolution and timeliness of the information are difficult to meet, and the algorithm migration and localization methods based on big data need to be studied. In addition, the agricultural machinery operation and spatio-temporal scheduling algorithm based on remote sensing and IoT monitoring information to determine the appropriate operation time window and operation prescription, needs to be further developed, especially the cross-regional scheduling algorithm of agricultural machinery for summer harvest in China. Aiming to address the issues of non-bi-connected monitoring and decision-making algorithms in grain production, as well as the insufficient integration of agricultural machinery and information perception, a framework for the grain production big data intelligent platform based on digital twins is proposed. The platform leverages multi-source heterogeneous grain production big data and integrates a full-chain suit of standardized algorithms, including data acquisition, information extraction, knowledge map construction, intelligent decision-making, full-chain collaboration of agricultural machinery operations. It covers the typical application scenarios such as irrigation, fertilization, pests and disease management, emergency response to drought and flood disaster, all enabled by digital twins technology. [Conclusions and Prospects] The suggestions and trends for development of grain production big data platform are summarized in three aspects: (1) Creating an open, symbiotic grain production big data platform, with core characteristics such as open interface for crop and environmental sensors, maturity grading and a cloud-native packaging mechanism for core algorithms, highly efficient response to data and decision services; (2) Focusing on the typical application scenarios of grain production, take the exploration of technology integration and bi-directional connectivity as the base, and the intelligent service as the soul of the development path for the big data platform research; (3) The data-algorithm-service self-organizing regulation mechanism, the integration of decision-making information with the intelligent equipment operation, and the standardized, compatible and open service capabilities, can form the new quality productivity to ensure food safety, and green efficiency grain production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Chinese Kiwifruit Text Named Entity Recognition Method Based on Dual-Dimensional Information and Pruning
    QI Zijun, NIU Dangdang, WU Huarui, ZHANG Lilin, WANG Lunfeng, ZHANG Hongming
    Smart Agriculture    2025, 7 (1): 44-56.   DOI: 10.12133/j.smartag.SA202410022
    Abstract794)   HTML17)    PDF(pc) (1225KB)(614)       Save

    [Objective] Chinese kiwifruit texts exhibit unique dual-dimensional characteristics. The cross-paragraph dependency is complex semantic structure, whitch makes it challenging to capture the full contextual relationships of entities within a single paragraph, necessitating models capable of robust cross-paragraph semantic extraction to comprehend entity linkages at a global level. However, most existing models rely heavily on local contextual information and struggle to process long-distance dependencies, thereby reducing recognition accuracy. Furthermore, Chinese kiwifruit texts often contain highly nested entities. This nesting and combination increase the complexity of grammatical and semantic relationships, making entity recognition more difficult. To address these challenges, a novel named entity recognition (NER) method, KIWI-Coord-Prune(kiwifruit-CoordKIWINER-PruneBi-LSTM) was proposed in this research, which incorporated dual-dimensional information processing and pruning techniques to improve recognition accuracy. [Methods] The proposed KIWI-Coord-Prune model consisted of a character embedding layer, a CoordKIWINER layer, a PruneBi-LSTM layer, a self-attention mechanism, and a CRF decoding layer, enabling effective entity recognition after processing input character vectors. The CoordKIWINER and PruneBi-LSTM modules were specifically designed to handle the dual-dimensional features in Chinese kiwifruit texts. The CoordKIWINER module applied adaptive average pooling in two directions on the input feature maps and utilized convolution operations to separate the extracted features into vertical and horizontal branches. The horizontal and vertical features were then independently extracted using the Criss-Cross Attention (CCNet) mechanism and Coordinate Attention (CoordAtt) mechanism, respectively. This module significantly enhanced the model's ability to capture cross-paragraph relationships and nested entity structures, thereby generating enriched character vectors containing more contextual information, which improved the overall representation capability and robustness of the model. The PruneBi-LSTM module was built upon the enhanced dual-dimensional vector representations and introduced a pruning strategy into Bi-LSTM to effectively reduce redundant parameters associated with background descriptions and irrelevant terms. This pruning mechanism not only enhanced computational efficiency while maintaining the dynamic sequence modeling capability of Bi-LSTM but also improved inference speed. Additionally, a dynamic feature extraction strategy was employed to reduce the computational complexity of vector sequences and further strengthen the learning capacity for key features, leading to improved recognition of complex entities in kiwifruit texts. Furthermore, the pruned weight matrices become sparser, significantly reducing memory consumption. This made the model more efficient in handling large-scale agricultural text-processing tasks, minimizing redundant information while achieving higher inference and training efficiency with fewer computational resources. [Results and Discussions] Experiments were conducted on the self-built KIWIPRO dataset and four public datasets: People's Daily, ClueNER, Boson, and ResumeNER. The proposed model was compared with five advanced NER models: LSTM, Bi-LSTM, LR-CNN, Softlexicon-LSTM, and KIWINER. The experimental results showed that KIWI-Coord-Prune achieved F1-Scores of 89.55%, 91.02%, 83.50%, 83.49%, and 95.81%, respectively, outperforming all baseline models. Furthermore, controlled variable experiments were conducted to compare and ablate the CoordKIWINER and PruneBi-LSTM modules across the five datasets, confirming their effectiveness and necessity. Additionally, the impact of different design choices was explored for the CoordKIWINER module, including direct fusion, optimized attention mechanism fusion, and network structure adjustment residual optimization. The experimental results demonstrated that the optimized attention mechanism fusion method yielded the best performance, which was ultimately adopted in the final model. These findings highlight the significance of properly designing attention mechanisms to extract dual-dimensional features for NER tasks. Compared to existing methods, the KIWI-Coord-Prune model effectively addressed the issue of underutilized dual-dimensional information in Chinese kiwifruit texts. It significantly improved entity recognition performance for both overall text structures and individual entity categories. Furthermore, the model exhibited a degree of generalization capability, making it applicable to downstream tasks such as knowledge graph construction and question-answering systems. [Conclusions] This study presents an novel NER approach for Chinese kiwifruit texts, which integrating dual-dimensional information extraction and pruning techniques to overcome challenges related to cross-paragraph dependencies and nested entity structures. The findings offer valuable insights for researchers working on domain-specific NER and contribute to the advancement of agriculture-focused natural language processing applications. However, two key limitations remain: 1) The balance between domain-specific optimization and cross-domain generalization requires further investigation, as the model's adaptability to non-agricultural texts has yet to be empirically validated; 2) the multilingual applicability of the model is currently limited, necessitating further expansion to accommodate multilingual scenarios. Future research should focus on two key directions: 1) Enhancing domain robustness and cross-lingual adaptability by incorporating diverse textual datasets and leveraging pre-trained multilingual models to improve generalization, and 2) Validating the model's performance in multilingual environments through transfer learning while refining linguistic adaptation strategies to further optimize recognition accuracy.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Real-time Detection Algorithm of Expanded Feed Image on the Water Surface Based on Improved YOLOv11
    ZHOU Xiushan, WEN Luting, JIE Baifei, ZHENG Haifeng, WU Qiqi, LI Kene, LIANG Junneng, LI Yijian, WEN Jiayan, JIANG Linyuan
    Smart Agriculture    2024, 6 (6): 155-167.   DOI: 10.12133/j.smartag.SA202408014
    Abstract793)   HTML82)    PDF(pc) (1858KB)(2749)       Save

    [Objective] During the feeding process of fish populations in aquaculture, the video image characteristics of floating extruded feed on the water surface undergo continuous variations due to a myriad of environmental factors and fish behaviors. These variations pose significant challenges to the accurate detection of feed particles, which is crucial for effective feeding management. To address these challenges and enhance the detection of floating extruded feed particles on the water surface, ,thereby providing precise decision support for intelligent feeding in intensive aquaculture modes, the YOLOv11-AP2S model, an advanced detection model was proposed. [Methods] The YOLOv11-AP2S model enhanced the YOLOv11 algorithm by incorporating a series of improvements to its backbone network, neck, and head components. Specifically, an attention for fine-grained categorization (AFGC) mechanism was introduced after the 10th layer C2PSA of the backbone network. This mechanism aimed to boost the model's capability to capture fine-grained features, which were essential for accurately identifying feed particles in complex environments with low contrast and overlapping objects. Furthermore, the C3k2 module was replaced with the VoV-GSCSP module, which incorporated more sophisticated feature extraction and fusion mechanisms. This replacement further enhanced the network's ability to extract relevant features and improve detection accuracy. To improve the model's detection of small targets, a P2 layer was introduced. However, adding a P2 layer may increase computational complexity and resource consumption, so the overall performance and resource consumption of the model must be carefully balanced. To maintain the model's real-time performance while improving detection accuracy, a lightweight VoV-GSCSP module was utilized for feature fusion at the P2 layer. This approach enabled the YOLOv11-AP2S model to achieve high detection accuracy without sacrificing detection speed or model lightweights, making it suitable for real-time applications in aquaculture. [Results and Discussions] The ablation experimental results demonstrated the superiority of the YOLOv11-AP2S model over the original YOLOv11 network. Specifically, the YOLOv11-AP2S model achieved a precision ( P) and recall ( R) of 78.70%. The mean average precision (mAP50) at an intersection over union (IoU) threshold of 0.5 was as high as 80.00%, and the F1-Score had also reached 79.00%. These metrics represented significant improvements of 6.7%, 9.0%, 9.4% (for precision, as previously mentioned), and 8.0%, respectively, over the original YOLOv11 network. These improvements showed the effectiveness of the YOLOv11-AP2S model in detecting floating extruded feed particles in complex environments. When compared to other YOLO models, the YOLOv11-AP2S model exhibits clear advantages in detecting floating extruded feed images on a self-made dataset. Notably, under the same number of iterations, the YOLOv11-AP2S model achieved higher mAP50 values and lower losses, demonstrating its superiority in detection performance. This indicated that the YOLOv11-AP2S model strikes a good balance between learning speed and network performance, enabling it to efficiently and accurately detect images of floating extruded feed on the water surface. Furthermore, the YOLOv11-AP2S model's ability to handle complex detection scenarios, such as overlapping and adhesion of feed particles and occlusion by bubbles, was noteworthy. These capabilities were crucial for accurate detection in practical aquaculture environments, where such challenges were common and can significantly impair the performance of traditional detection systems. The improvements in detection accuracy and efficiency made the YOLOv11-AP2S model a valuable tool for intelligent feeding systems in aquaculture, as it could provide more reliable and timely information on fish feeding behavior. Additionally, the introduction of the P2 layer and the use of the lightweight VoV-GSCSP module for feature fusion at this layer contributed to the model's overall performance. These enhancements enabled the model to maintain high detection accuracy while keeping computational costs and resource consumption within manageable limits. This was particularly important for real-time applications in aquaculture, where both accuracy and efficiency were critical for effective feeding management. [Conclusions] The successful application of the YOLOv11-AP2S model in detecting floating extruded feed particles demonstrates its potential to intelligent feeding systems in aquaculture. By providing accurate and timely information on fish feeding behavior, the model can help optimize feeding strategies, reduce feed waste, and improve the overall efficiency and profitability of aquaculture operations. Furthermore, the model's ability to handle complex detection scenarios and maintain high detection accuracy while keeping computational costs within manageable limits makes it a practical and valuable tool for real-time applications in aquaculture. Therefore, the YOLOv11-AP2S model holds promise for wide application in intelligent aquaculture management, contributing to the sustainability and growth of the aquaculture industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Supply Chains for Agricultural Products: Key Technologies, Research Progress and Future Direction
    HAN Jiawei, YANG Xinting
    Smart Agriculture    2025, 7 (3): 1-16.   DOI: 10.12133/j.smartag.SA202501006
    Abstract765)   HTML88)    PDF(pc) (1373KB)(1017)       Save

    [Significance] The smart transformation of agricultural product supply chains is an essential solution to the challenges faced by traditional supply chains, such as information asymmetry, high logistics costs, and difficulties in quality traceability. This transformation also serves as a vital pathway to modernize agriculture and enhance industrial competitiveness. By integrating technologies such as the Internet of Things (IoT), big data, and artificial intelligence (AI), smart supply chains facilitate precise production and processing, efficient logistics distribution, and transparent quality supervision. As a result, they improve circulation efficiency, ensure product safety, increase farmers' incomes, and promote sustainable agricultural development. Furthermore, in light of global shifts in agricultural trade, this transformation bolsters the international competitiveness of China's agricultural products and propels the agricultural industrial chain toward higher value-added segments. This paper systematically examines the conceptual framework, technological applications, and future trends of smart supply chains, aiming to provide a theoretical foundation for industry practices and insights for policymaking and technological innovation. [Progress] In the production phase, IoT and remote sensing technologies enable real-time monitoring of crop growth conditions, including soil moisture, temperature, and pest infestation, facilitating precision irrigation, fertilization, and pest management. Big data analysis, coupled with AI algorithms, helps in predicting crop yields, optimizing resource allocation, and minimizing waste. Additionally, AI-driven smart pest control systems can dynamically adjust pesticide application, reducing chemical usage and environmental impact. The processing stage leverages advanced technologies for efficient sorting, grading, cleaning, and packaging. Computer vision and hyperspectral imaging technologies enhance the sorting efficiency and quality inspection of agricultural products, ensuring only high-quality products proceed to the next stage. Novel cleaning techniques, such as ultrasonic and nanobubble cleaning, effectively remove surface contaminants and reduce microbial loads without compromising product quality. Moreover, AI-integrated systems optimize processing lines, reduce downtime and enhance overall throughput. Warehousing employs IoT sensors to monitor environmental conditions like temperature, humidity, and gas concentrations, ensure optimal storage conditions for diverse agricultural products. AI algorithms predict inventory demand, optimize stock levels to minimize waste and maximize freshness. Robotics and automation in warehouses improve picking, packing, and palletizing efficiency, reduce labor costs and enhance accuracy. The transportation sector focuses on cold chain innovations to maintain product quality during transit. IoT-enabled temperature-controlled containers and AI-driven scheduling systems ensure timely and efficient delivery. Additionally, the integration of blockchain technology provides immutable records of product handling and conditions, enhances transparency and trust. The adoption of new energy vehicles, such as electric and hydrogen-powered trucks, further reduces carbon footprints and operating costs. In the distribution and sales stages, big data analytics optimize delivery routes, reducing transportation time and costs. AI-powered demand forecasting enables precise inventory management, minimizes stockouts and excess inventories. Moreover, AI and machine learning algorithms personalize marketing efforts, improve customer engagement and satisfaction. Blockchain technology ensures product authenticity and traceability, enhances consumer trust. [Conclusions and Prospects] As technological advancements and societal demands continue to evolve, the smart transformation of agricultural product supply chains has become increasingly urgent. Future development should prioritize unmanned operations to alleviate labor shortages and enhance product quality and safety. Establishing information-sharing platforms and implementing refined management practices are crucial for optimizing resource allocation, improving operational efficiency, and enhancing international competitiveness. Additionally, aligning with the "dual-carbon" strategy by promoting clean energy adoption, optimizing transportation methods, and advocating for sustainable packaging will drive the supply chain toward greater sustainability. However, the application of emerging technologies in agricultural supply chains faces challenges such as data governance, technical adaptability, and standardization. Addressing these issues requires policy guidance, technological innovation, and cross-disciplinary collaboration. By overcoming these challenges, the comprehensive intelligent upgrade of agricultural product supply chains can be achieved, ultimately contribute to the modernization and sustainable development of the agricultural sector.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Advances, Problems and Challenges of Precise Estrus Perception and Intelligent Identification Technology for Cows
    ZHANG Zhiyong, CAO Shanshan, KONG Fantao, LIU Jifang, SUN Wei
    Smart Agriculture    2025, 7 (3): 48-68.   DOI: 10.12133/j.smartag.SA202305005
    Abstract729)   HTML34)    PDF(pc) (2463KB)(2600)       Save

    [Significance] Estrus monitoring and identification in cows is a crucial aspect of breeding management in beef and dairy cattle farming. Innovations in precise sensing and intelligent identification methods and technologies for estrus in cows are essential not only for scientific breeding, precise management, and smart breeding on a population level but also play a key supportive role in health management, productivity enhancement, and animal welfare improvement at the individual level. The aims are to provide a reference for scientific management and the study of modern production technologies in the beef and dairy cattle industry, as well as theoretical methodologies for the research and development of key technologies in precision livestock farming. [Progress] Based on describing the typical characteristics of normal and abnormal estrus in cows, this paper systematically categorizes and summarizes the recent research progress, development trends, and methodological approaches in estrus monitoring and identification technologies, focusing on the monitoring and diagnosis of key physiological signs and behavioral characteristics during the estrus period. Firstly, the paper outlines the digital monitoring technologies for three critical physiological parameters, body temperature, rumination, and activity levels, and their applications in cow estrus monitoring and identification. It analyzes the intrinsic reasons for performance bottlenecks in estrus monitoring models based on body temperature, compares the reliability issues faced by activity-based estrus monitoring, and addresses the difficulties in balancing model generalization and robustness design. Secondly, the paper examines the estrus sensing and identification technologies based on three typical behaviors: feeding, vocalization, and sexual desire. It highlights the latest applications of new artificial intelligence technologies, such as computer vision and deep learning, in estrus monitoring and points out the critical role of these technologies in improving the accuracy and timeliness of monitoring. Finally, the paper focuses on multi-factor fusion technologies for estrus perception and identification, summarizing how different researchers combine various physiological and behavioral parameters using diverse monitoring devices and algorithms to enhance accuracy in estrus monitoring. It emphasizes that multi-factor fusion methods can improve detection rates and the precision of identification results, being more reliable and applicable than single-factor methods. The importance and potential of multi-modal information fusion in enhancing monitoring accuracy and adaptability are underlined. The current shortcomings of multi-factor information fusion methods are analyzed, such as the potential impact on animal welfare from parameter acquisition methods, the singularity of model algorithms used for representing multi-factor information fusion, and inadequacies in research on multi-factor feature extraction models and estrus identification decision algorithms. [Conclusions and Prospects] From the perspectives of system practicality, stability, environmental adaptability, cost-effectiveness, and ease of operation, several key issues are discussed that need to be addressed in the further research of precise sensing and intelligent identification technologies for cow estrus within the context of high-quality development in digital livestock farming. These include improving monitoring accuracy under weak estrus conditions, overcoming technical challenges of audio extraction and voiceprint construction amidst complex background noise, enhancing the adaptability of computer vision monitoring technologies, and establishing comprehensive monitoring and identification models through multi-modal information fusion. It specifically discusses the numerous challenges posed by these issues to current technological research and explains that future research needs to focus not only on improving the timeliness and accuracy of monitoring technologies but also on balancing system cost-effectiveness and ease of use to achieve a transition from the concept of smart farming to its practical implementation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Tomato Leaf Disease and Pest Detection Method Based on Improved YOLOv10n
    WU Liuai, XU Xueke
    Smart Agriculture    2025, 7 (1): 146-155.   DOI: 10.12133/j.smartag.SA202410023
    Abstract721)   HTML103)    PDF(pc) (1834KB)(920)       Save

    [Objective] To address the challenges in detecting tomato leaf diseases and pests, such as complex environments, small goals, low precision, redundant parameters, and high computational complexity, a novel lightweight, high-precision, real-time detection model was proposed called YOLOv10n-YS. This model aims to accurately identify diseases and pests, thereby providing a solid scientific basis for their prevention and management strategies. Methods] The dataset was collected using mobile phones to capture images from multiple angles under natural conditions, ensuring complete and clear leaf images. It included various weather conditions and covered nine types: Early blight, leaf mold, mosaic virus, septoria, spider mites damage, yellow leaf curl virus, late blight, leaf miner disease, and healthy leaves, with all images having a resolution of 640×640 pixels. In the proposed YOLOv10n-YS model, firstly, the C2f in the backbone network was replaced with C2f_RepViTBlock, thereby reducing the computational load and parameter volume and achieving a lightweight design. Secondly, through the introduction of a sliced operation SimAM attention mechanism, the Conv_SWS module was formed, which enhanced the extraction of small target features. Additionally, the DySample lightweight dynamic up sampling module was used to replace the up sampling module in the neck network, concentrating sampling points on target areas and ignoring backgrounds, thereby effectively identifying defects. Finally, the efficient channel attention (ECA) was improved by performing average pooling and max pooling on the input layer to aggregate features and then adding them together, which further enhanced global perspective information and features of different scales. The improved module, known as efficient channel attention with cross-channel interaction (EMCA) attention, was introduced, and the pyramid spatial attention (PSA) in the backbone network was replaced with the EMCA attention mechanism, thereby enhancing the feature extraction capability of the backbone network. [Results and Discussions] After introducing the C2f_RepViTBlock, the model's parameter volume and computational load were reduced by 12.3% and 9.7%, respectively, with mAP@0.5 and F1-Score each increased by 0.2% and 0.3%. Following the addition of the Conv_SWS and the replacement of the original convolution, mAP@0.5 and F1-Score were increased by 1.2% and 2%, respectively, indicating that the Conv_SWS module significantly enhanced the model's ability to extract small target features. After the introduction of DySample, mAP@0.5 and F1-Score were increased by 1.8% and 2.6%, respectively, but with a slight increase in parameter volume and computational load. Finally, the addition of the EMCA attention mechanism further enhanced the feature extraction capability of the backbone network. Through these four improvements, the YOLOv10n-YS model was formed. Compared with the YOLOv10n algorithm, YOLOv10n-YS reduced parameter volume and computational load by 13.8% and 8.5%, respectively, with both mAP@0.5 and F1-Score increased. These improvements not only reduced algorithm complexity but also enhanced detection accuracy, making it more suitable for industrial real-time detection. The detection accuracy of tomato diseases and pests using the YOLOv10n-YS algorithm was significantly better than that of comparative algorithms, and it had the lowest model parameter volume and computational load. The visualization results of detection by different models showed that the YOLOv10n-YS network could provide technical support for the detection and identification of tomato leaf diseases and pests. To verify the performance and robustness of the YOLOv10n-YS algorithm, comparative experiments were conducted on the public Plant-Village-9 dataset with different algorithms. The results showed that the average detection accuracy of YOLOv10n-YS on the Plant-Village dataset reached 91.1%, significantly higher than other algorithms. [Conclusions] The YOLOv10n-YS algorithm is not only characterized by occupying a small amount of space but also by possessing high recognition accuracy. On the tomato leaf dataset, excellent performance was demonstrated by this algorithm, thereby verifying its broad applicability and showcasing its potential to play an important role in large-scale crop pest and disease detection applications. Deploying the model on drone platforms and utilizing multispectral imaging technology can achieve real-time detection and precise localization of pests and diseases in complex field environments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Status and Prospect of Quality Intelligent Control Technology in Facilities Environment of Characteristic Agricultural Products
    GUO Wei, WU Huarui, GUO Wang, GU Jingqiu, ZHU Huaji
    Smart Agriculture    2024, 6 (6): 44-62.   DOI: 10.12133/j.smartag.SA202411017
    Abstract689)   HTML69)    PDF(pc) (3153KB)(2182)       Save

    [Significance] In view of the lack of monitoring means of quality influence factors in the production process of characteristic agricultural products with in central and western regions of China, the weak ability of intelligent control, the unclear coupling relationship of quality control elements and the low degree of systematic application, the existing technologies described such as intelligent monitoring of facility environment, growth and nutrition intelligent control model, architecture of intelligent management and control platform and so on. Through the application of the Internet of Things, big data and the new generation of artificial intelligence technology, it provides technical support for the construction and application of intelligent process quality control system for the whole growth period of characteristic agricultural products. [Progress] The methods of environmental regulation and nutrition regulation are analyzed, including single parameters and combined control methods, such as light, temperature, humidity, CO2 concentration, fertilizer and water, etc. The multi-parameter coupling control method has the advantage of more comprehensive scene analysis. Based on the existing technology, a multi-factor coupling method of integrating growth state, agronomy, environment, input and agricultural work is put forward. This paper probes into the system architecture of the whole process service of quality control, the visual identification system of the growth process of agricultural products and the knowledge-driven agricultural technical service system, and introduces the technology of the team in the disease knowledge Q & A scene through multi-modal knowledge graph and large model technology. [Conclusions and Prospects] Based on the present situation of the production of characteristic facility agricultural products and the overall quality of farmers in the central and western regions of China, it is appropriate to transfer the whole technical system such as facility tomato, facility cucumber and so on. According to the varieties of characteristic agricultural products, cultivation models, quality control objectives to adapt to light, temperature, humidity and other parameters, as well as fertilizer, water, medicine and other input plans, a multi-factor coupling model suitable for a specific planting area is generated and long-term production verification and model correction are carried out. And popularize it in a wider area, making full use of the advantages of intelligent equipment and data elements will promote the realization of light simplification of production equipment, scene of intelligent technology, diversification of service models, on-line quality control, large-scale production of digital intelligence, and value of data elements, further cultivate facilities to produce new quality productivity.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Detection Method of Apple Alternaria Leaf Spot Based on Deep-Semi-NMF
    FU Zhuojun, HU Zheng, DENG Yangjun, LONG Chenfeng, ZHU Xinghui
    Smart Agriculture    2024, 6 (6): 144-154.   DOI: 10.12133/j.smartag.SA202409001
    Abstract657)   HTML32)    PDF(pc) (1901KB)(270)       Save

    [Objective] Apple Alternaria leaf spot can easily lead to premature defoliation of apple tree leaves, thereby affecting the quality and yield of apples. Consequently, accurately detecting of the disease has become a critical issue in the precise prevention and control of apple tree diseases. Due to factors such as backlighting, traditional image segmentation-based methods for detecting disease spots struggle to accurately identify the boundaries of diseased areas against complex backgrounds. There is an urgent need to develop new methods for detecting apple Alternaria leaf spot, which can assist in the precise prevention and control of apple tree diseases. [Methods] A novel detection method named Deep Semi-Non-negative Matrix Factorization-based Mahalanobis Distance Anomaly Detection (DSNMFMAD) was proposed, which combines Deep Semi-Non-negative Matrix Factorization (DSNMF) with Mahalanobis distance for robust anomaly detection in complex image backgrounds. The proposed method began by utilizing DSNMF to extract low-rank background components and sparse anomaly features from the apple Alternaria leaf spot images. This enabled effective separation of the background and anomalies, mitigating interference from complex background noise while preserving the non-negativity constraints inherent in the data. Subsequently, Mahalanobis distance was employed, based on the Singular Value Decomposition (SVD) feature subspace, to construct a lesion detector. The detector identified lesions by calculating the anomaly degree of each pixel in the anomalous regions. The apple tree leaf disease dataset used was provided by PaddlePaddle AI-Studio. Each image in the dataset has a resolution of 512×512 pixels, in RGB color format, and was in JPEG format. The dataset was captured in both laboratory and natural environments. Under laboratory conditions, 190 images of apple leaves with spot-induced leaf drop were used, while 237 images were collected under natural conditions. Furthermore, the dataset was augmented with geometric transformations and random changes in brightness, contrast, and hue, resulting in 1 145 images under laboratory conditions and 1 419 images under natural conditions. These images reflect various real-world scenarios, capturing apple leaves at different stages of maturity, in diverse lighting conditions, angles, and noise environments. This diversed dataset ensured that the proposed method could be tested under a wide range of practical conditions, providing a comprehensive evaluation of its effectiveness in detecting apple Alternaria leaf spot. [Results and Discussions] DSNMFMAD demonstrated outstanding performance under both laboratory and natural conditions. A comparative analysis was conducted with several other detection methods, including GRX (Reed-Xiaoli detector), LRX (Local Reed-Xiaoli detector), CRD (Collaborative-Representation-Based Detector), LSMAD (LRaSMD-Based Mahalanobis Distance Detector), and the deep learning model Unet. The results demonstrated that DSNMFMAD exhibited superior performance in the laboratory environment. The results demonstrated that DSNMFMAD attained a recognition accuracy of 99.8% and a detection speed of 0.087 2 s/image. The accuracy of DSNMFMAD was found to exceed that of GRX, LRX, CRD, LSMAD, and Unet by 0.2%, 37.9%, 10.3%, 0.4%, and 24.5%, respectively. Additionally, the DSNMFMAD exhibited a substantially superior detection speed in comparison to LRX, CRD, LSMAD, and Unet, with an improvement of 8.864, 107.185, 0.309, and 1.565 s, respectively. In a natural environment, where a dataset of 1 419 images of apple Alternaria leaf spot was analysed, DSNMFMAD demonstrated an 87.8% recognition accuracy, with an average detection speed of 0.091 0 s per image. In this case, its accuracy outperformed that of GRX, LRX, CRD, LSMAD, and Unet by 2.5%, 32.7%, 5%, 14.8%, and 3.5%, respectively. Furthermore, the detection speed was faster than that of LRX, CRD, LSMAD, and Unet by 2.898, 132.017, 0.224, and 1.825 s, respectively. [Conclusions] The DSNMFMAD proposed in this study was capable of effectively extracting anomalous parts of an image through DSNMF and accurately detecting the location of apple Alternaria leaf spot using a constructed lesion detector. This method achieved higher detection accuracy compared to the benchmark methods, even under complex background conditions, demonstrating excellent performance in lesion detection. This advancement could provide a valuable technical reference for the detection and prevention of apple Alternaria leaf spot.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Graph Neural Networks for Knowledge Graph Construction: Research Progress, Agricultural Development Potential, and Future Directions
    YUAN Huan, FAN Beilei, YANG Chenxue, LI Xian
    Smart Agriculture    2025, 7 (2): 41-56.   DOI: 10.12133/j.smartag.SA202501007
    Abstract620)   HTML48)    PDF(pc) (3048KB)(617)       Save

    [Significance] Graph neural networks (GNN) have emerged as a powerful tool in the realm of data analysis, particularly in knowledge graph construction. By capitalizing on the interaction and message passing among nodes in a graph, GNN can capture intricate relationships, making them widely applicable in various tasks, including knowledge representation, extraction, fusion, and inference. In the context of agricultural knowledge graph (AKG) development and knowledge service application, however, the agricultural domain presents unique challenges. These challenges encompass data with high multi-source heterogeneity, dynamic spatio-temporal changes in knowledge, complex relationships, and stringent requirements for interpretability. Given its strengths in graph structure data modeling, GNNs hold great promise in addressing these difficulties. For instance, in agricultural data, information from weather sensors, soil monitoring devices, and historical crop yield records varies significantly in format and type, and the ability of GNNs to handle such heterogeneity becomes crucial. [Progress] Firstly, this paper provides a comprehensive overview of the representation methods and fundamental concepts of GNNs was presented. The main structures, basic principles, characteristics, and application directions of five typical GNN models were discussed, including recursive graph neural networks (RGNN), convolutional graph neural networks (CGNN), graph auto-encoder networks (GAE), graph attention networks (GAT), and spatio-temporal graph neural networks(STGNN). Each of these models has distinct advantages in graph feature extraction, which are leveraged for tasks such as dynamic updates, knowledge completion, and complex relationship modeling in knowledge graphs. For example, STGNNs are particularly adept at handling the time-series and spatial data prevalent in agriculture, enabling more accurate prediction of crop growth patterns. Secondly, how GNN utilize graph structure information and message passing mechanisms to address issues in knowledge extraction related to multi-source heterogeneous data fusion and knowledge representation was elucidated. It can enhance the capabilities of entity recognition disambiguation and multi-modal data entity recognition. For example, when dealing with both textual descriptions of agricultural pests and corresponding image data, GNNs can effectively integrate these different modalities to accurately identify the pests. It also addresses the tasks of modeling complex dependencies and long-distance relationships or multi-modal relation extraction, achieving precise extraction of complex, missing information, or multi-modal events. Furthermore, GNNs possess unique characteristics, such as incorporating node or subgraph topology information, learning deep hidden associations between entities and relationships, generating low-dimensional representations encoding structure and semantics, and learning or fusing iterative non-linear neighborhood feature relationships on the graph structure, make it highly suitable for tasks like entity prediction, relation prediction, denoising, and anomaly information inference. These applications significantly enhance the construction quality of knowledge graphs. In an agricultural setting, this means more reliable predictions of disease outbreaks based on the relationships between environmental factors and crop health. Finally, in-depth analyses of typical cases of intelligent applications based on GNNs in agricultural knowledge question answering, recommendation systems, yield prediction, and pest monitoring and early warning are conducted. The potential of GNNs for constructing temporal agricultural knowledge models is explored, and its ability to adapt to the changing nature of agricultural data over time is highlighted. [Conclusions and Prospects] Research on constructing AKGs using GNNs is in its early stages. Future work should focus on key technologies like deep multi-source heterogeneous data fusion, knowledge graph evolution, scenario-based complex reasoning, and improving interpretability and generalization. GNN-based AKGs are expected to take on professional roles such as virtual field doctors and agricultural experts. Applications in pest control and planting decisions will be more precise, and intelligent tools like smart agricultural inputs and encyclopedia retrieval systems will be more comprehensive. By representing and predicting entities and relationships in agriculture, GNN-based AKGs can offer efficient knowledge services and intelligent solutions for sustainable agricultural development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Key Technologies and Construction model for Unmanned Smart Farms: Taking the "1.5-Ton Grain per Mu" Unmanned Farm as An Example
    LIU lining, ZHANG Hongqi, ZHANG Ziwen, ZHANG Zhenghui, WANG Jiayu, LI Xuanxuan, ZHU Ke, LIU Pingzeng
    Smart Agriculture    2025, 7 (1): 70-84.   DOI: 10.12133/j.smartag.SA202410033
    Abstract606)   HTML59)    PDF(pc) (2674KB)(800)       Save

    [Objective] As a key model of smart agriculture, the unmanned smart farm aims to develop a highly intelligent and automated system for high grain yields. This research uses the "1.5-Ton grain per Mu" farm in Dezhou city, Shandong province, as the experimental site, targeting core challenges in large-scale smart agriculture and exploring construction and service models for such farms. [Methods] The "1.5-Ton grain per Mu" unmanned smart farm comprehensively utilized information technologies such as the internet of things (IoT) and big data to achieve full-chain integration and services for information perception, transmission, mining, and application. The overall construction architecture consisted of the perception layer, transmission layer, processing layer, and application layer. This architecture enabled precise perception, secure transmission, analysis and processing, and application services for farm data. A perception system for the unmanned smart farm of wheat was developed, which included a digital perception network and crop phenotypic analysis. The former achieved precise perception, efficient transmission, and precise measurement and control of data information within the farm through perception nodes, self-organizing networks, and edge computing core processing nodes. Phenotypic analysis utilized methods such as deep learning to extract phenotypic characteristics at different growth stages, such as the phenological classification of wheat and wheat ear length. An intelligent controlled system had been developed. The system consisted of an intelligent agricultural machinery system, a field irrigation system, and an aerial pesticided application system. The intelligent agricultural machinery system was composed of three parts: the basic layer, decision-making layer, and application service layer. They were responsible for obtaining real-time status information of agricultural machinery, formulating management decisions for agricultural machinery, and executing operational commands, respectively. Additionally, appropriate agricultural machinery models and configuration references were provided. A refined irrigation scheme was designed based on the water requirements and soil conditions at different developmental stages of wheat. And, an irrigation control algorithm based on fuzzy PID was proposed. Finally, relying on technologies such as multi-source data fusion, distributed computing, and geographic information system (GIS), an intelligent management and control platform for the entire agricultural production process was established. [Results and Discussions] The digital perception network enabled precise sensing and networked transmission of environmental information within the farm. The data communication quality of the sensor network remained above 85%, effectively ensuring data transmission quality. The average relative error in extracting wheat spike length information based on deep learning algorithms was 1.24%. Through the coordinated operation of intelligent control system, the farm achieved lean and unmanned production management, enabling intelligent control throughout the entire production chain, which significantly reduced labor costs and improved the precision and efficiency of farm management. The irrigation model not only saved 20% of irrigation water but also increased the yield of "Jinan 17" and "Jimai 44" by 10.18% and 7%, respectively. Pesticide application through spraying drones reduced pesticide usage by 55%. The big data platform provided users with production guidance services such as meteorological disaster prediction, optimal sowing time, environmental prediction, and water and fertilizer management through intelligent scientific decision support, intelligent agricultural machinery operation, and producted quality and safety traceability modules, helping farmers manage their farms scientifically. [Conclusions] The study achieved comprehensive collection of environmental information within the farm, precise phenotypic analysis, and intelligent control of agricultural machinery, irrigation equipment, and other equipment. Additionally, it realized digital services for agricultural management through a big data platform. The development path of the "1.5-Ton grain per Mu" unmanned smart farm can provid references for the construction of smart agriculture.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress on Remote Sensing Monitoring and Intelligent Decision-Making Algorithms for Rice Production
    ZHAO Bingting, HUA Chuanhai, YE Chenyang, XIONG Yuchun, QIAN Tao, CHENG Tao, YAO Xia, ZHENG Hengbiao, ZHU Yan, CAO Weixing, JIANG Chongya
    Smart Agriculture    2025, 7 (2): 57-72.   DOI: 10.12133/j.smartag.SA202501002
    Abstract602)   HTML49)    PDF(pc) (1365KB)(449)       Save

    [Significance] Rice is a staple food crop worldwide, and ccurate monitoring of its growth is crucial for global food security. Remote sensing serves as a powerful tool in modern agriculture. By integrating remote sensing with intelligent decision-making algorithms, farmers can achieve more precise and sustainable rice cultivation. To provide actionable insights and guidance for researchers in this field, this review examines the latest advancements in remote sensing and smart algorithms for rice farming, while addressing current challenges and future trends. [Progress] Currently, remote sensing-based monitoring systems for rice production have been comprehensively implemented across the entire production cycle. For planting distribution identification, optical remote sensing and synthetic aperture radar (SAR) technologies complement each other to enhance accuracy through data fusion. Regarding growth period monitoring, a robust technical framework has been established, incorporating the empirical threshold method, shape model approach, and machine learning classification techniques. Dynamic evaluation of growth status is enabled by constructing correlation models between remote sensing features and biophysical parameters. Disaster monitoring systems provide rapid responses to various natural disasters. Yield and quality predictions integrate crop models, remote sensing data, and machine learning algorithms. Intelligent decision-making algorithms are deeply embedded in all stages of rice production. For instance, during planting planning, the integration of geographic information systems (GIS) and multi-criteria evaluation methods facilitates regional suitability assessments and farm-level quantitative designs. In topdressing management, nitrogen-based intelligent algorithms have significantly improved fertilization precision. Irrigation optimization achieves water conservation and emission reduction by synthesizing soil moisture and meteorological data. Finally, precise pesticide application prescriptions are generated using remote sensing and unmanned aerial vehicle (UAV) technologies. [Conclusions and Prospects] Despite significant progress, current research faces persistent challenges, including difficulties in multi-source data fusion, complexities in acquiring prior knowledge, insufficient model standardization, and barriers to large-scale technology implementation. Future efforts should prioritize the following six directions: (1) Technological innovation: Advance collaborative analysis of multi-source remote sensing data, design optimized data fusion algorithms, and construct an integrated air-space-ground monitoring network; (2) Intelligent algorithms: Explore cutting-edge techniques such as generative adversarial networks (GANs) and federated learning to enhance model adaptability across diverse environments; (3) Research scale: Establish a global rice growth monitoring system and develop multi-factor coupling models to assess climate change impacts; (4) Technology dissemination: Strengthen demonstration projects, reduce equipment costs, and cultivate interdisciplinary professionals; (5) Standards and protocols: Promote internationally unified standards for monitoring and decision - making frameworks; (6) System integration: Leverage technologies such as digital twins and blockchain to develop smart agriculture platforms for end-to-end intelligent management. Through multi-dimensional innovation, these advancements will significantly elevate the intelligence of rice production, offering robust support for global food security and sustainable agricultural development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Chilli-YOLO: An Intelligent Maturity Detection Algorithm for Field-Grown Chilli Based on Improved YOLOv10
    SI Chaoguo, LIU Mengchen, WU Huarui, MIAO Yisheng, ZHAO Chunjiang
    Smart Agriculture    2025, 7 (2): 160-171.   DOI: 10.12133/j.smartag.SA202411002
    Abstract539)   HTML80)    PDF(pc) (2545KB)(1740)       Save

    [Objective] In modern agriculture, the rapid and accurate detection of chillies at different maturity stages is a critical step for determining the optimal harvesting time and achieving intelligent sorting of field-grown chillies. However, existing target detection models face challenges in efficiency and accuracy when applied to the task of detecting chilli maturity, which limit their widespread use and effectiveness in practical applications. To address these challenges, a new algorithm, Chilli-YOLO, was proposed for achieving efficient and precise detection of chilli maturity in complex environments. [Methods] A comprehensive image dataset was collected, capturing chillis under diverse and realistic agricultural conditions, including varying lighting conditions, camera angles, and background complexities. These images were then meticulously categorized into four distinct maturity stages: Immature, transitional, mature, and dried. Data augmentation techniques were employed to expand the dataset and enhance the model's generalization capabilities. To develop an accurate and efficient chili maturity detection system, the YOLOv10s object detection network was chosen as the foundational architecture. The model's performance was further enhanced through strategic optimizations targeting the backbone network. Specifically, standard convolutional layers were replaced with Ghost convolutions. This technique generated more feature maps from fewer parameters, resulting in significant computational savings and improved processing speed without compromising feature extraction quality. Additionally, the C2f module was substituted with the more computationally efficient GhostConv module, further reducing redundancy and enhancing the model's overall efficiency. To improve the model's ability to discern subtle visual cues indicative of maturity, particularly in challenging scenarios involving occlusion, uneven lighting, or complex backgrounds, the partial self-attention (PSA) module within YOLOv10s was replaced with the second-order channel attention (SOCA) mechanism. SOCA leverages higher-order feature correlations to more effectively capture fine-grained characteristics of the chillis. This enabled the model to focus on relevant feature channels and effectively identify subtle maturity-related features, even when faced with significant visual noise and interference. Finally, to refine the precision of target localization and minimize bounding box errors, the extended intersection over union (XIoU) loss function was integrated into the model training process. XIoU enhances the traditional IoU loss by considering factors such as the aspect ratio difference and the normalized distance between the predicted and ground truth bounding boxes. By optimizing for these factors, the model achieved significantly improved localization accuracy, resulting in a more precise delineation of chillis in the images and contributing to the overall enhancement of the detection performance. The combined implementation of these improvements aimed to construct an effective approach to correctly classify the maturity level of chillis within the challenging and complex environment of a real-world farm. [Results and Discussion] The experimental results on the custom-built chilli maturity detection dataset showed that the Chilli-YOLO model performed excellently across multiple evaluation metrics. The model achieved an accuracy of 90.7%, a recall rate of 82.4%, and a mean average precision (mAP) of 88.9%. Additionally, the model's computational load, parameter count, model size, and inference time were 18.3 GFLOPs, 6.37 M, 12.6 M, and 7.3 ms, respectively. Compared to the baseline model, Chilli-YOLO improved accuracy by 2.6 percent point, recall by 2.8 percent point and mAP by 2.8 percent point. At the same time, the model's computational load decreased by 6.2 GFLOPs, the parameter count decreased by 1.67 M, model size reduced by 3.9 M. These results indicated that Chilli-YOLO strikes a good balance between accuracy and efficiency, making it capable of fast and precise detection of chilli maturity in complex agricultural environments. Moreover, compared to earlier versions of the YOLO model, Chilli-YOLO showed improvements in accuracy of 2.7, 4.8, and 5 percent point over YOLOv5s, YOLOv8n, and YOLOv9s, respectively. Recall rates were higher by 1.1, 0.3, and 2.3 percent point, and mAP increased by 1.2, 1.7, and 2.3 percent point, respectively. In terms of parameter count, model size, and inference time, Chilli-YOLO outperformed YOLOv5. This avoided the issue of YOLOv8n's lower accuracy, which was unable to meet the precise detection needs of complex outdoor environments. When compared to the traditional two-stage network Faster RCNN, Chilli-YOLO showed significant improvements across all evaluation metrics. Additionally, compared to the one-stage network SSD, Chilli-YOLO achieved substantial gains in accuracy, recall, and mAP, with increases of 16.6%, 12.1%, and 16.8%, respectively. Chilli-YOLO also demonstrated remarkable improvements in memory usage, model size, and inference time. These results highlighted the superior overall performance of the Chilli-YOLO model in terms of both memory consumption and detection accuracy, confirming its advantages for chilli maturity detection. [Conclusions] The proposed Chilli-YOLO model optimizes the network structure and loss functions, not only can significantly improve detection accuracy but also effectively reduce computational overhead, making it better suites for resource-constrained agricultural production environments. The research provides a reliable technical reference for intelligent harvesting of chillies in agricultural production environments, especially in resource-constrained settings.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Market Monitoring and Early Warning: An Integrated Forecasting Approach Based on Deep Learning
    XU Shiwei, LI Qianchuan, LUAN Rupeng, ZHUANG Jiayu, LIU Jiajia, XIONG Lu
    Smart Agriculture    2025, 7 (1): 57-69.   DOI: 10.12133/j.smartag.SA202411004
    Abstract535)   HTML36)    PDF(pc) (1936KB)(981)       Save

    [Significance] The fluctuations in the supply, consumption, and prices of agricultural products directly affect market monitoring and early warning systems. With the ongoing transformation of China's agricultural production methods and market system, advancements in data acquisition technologies have led to an explosive growth in agricultural data. However, the complexity of the data, the narrow applicability of existing models, and their limited adaptability still present significant challenges in monitoring and forecasting the interlinked dynamics of multiple agricultural products. The efficient and accurate forecasting of agricultural market trends is critical for timely policy interventions and disaster management, particularly in a country with a rapidly changing agricultural landscape like China. Consequently, there is a pressing need to develop deep learning models that are tailored to the unique characteristics of Chinese agricultural data. These models should enhance the monitoring and early warning capabilities of agricultural markets, thus enabling precise decision-making and effective emergency responses. [Methods] An integrated forecasting methodology was proposed based on deep learning techniques, leveraging multi-dimensional agricultural data resources from China. The research introduced several models tailored to different aspects of agricultural market forecasting. For production prediction, a generative adversarial network and residual network collaborative model (GAN-ResNet) was employed. For consumption forecasting, a variational autoencoder and ridge regression (VAE-Ridge) model was used, while price prediction was handled by an Adaptive-Transformer model. A key feature of the study was the adoption of an "offline computing and visualization separation" strategy within the Chinese agricultural monitoring and early warning system (CAMES). This strategy ensures that model training and inference are performed offline, with the results transmitted to the front-end system for visualization using lightweight tools such as ECharts. This approach balances computational complexity with the need for real-time early warnings, allowing for more efficient resource allocation and faster response times. The corn, tomato, and live pig market data used in this study covered production, consumption and price data from 1980 to 2023, providing comprehensive data support for model training. [Results and Discussions] The deep learning models proposed in this study significantly enhanced the forecasting accuracy for various agricultural products. For instance, the GAN-ResNet model, when used to predict maize yield at the county level, achieved a mean absolute percentage error (MAPE) of 6.58%. The VAE-Ridge model, applied to pig consumption forecasting, achieved a MAPE of 6.28%, while the Adaptive-Transformer model, used for tomato price prediction, results in a MAPE of 2.25%. These results highlighted the effectiveness of deep learning models in handling complex, nonlinear relationships inherent in agricultural data. Additionally, the models demonstrate notable robustness and adaptability when confronted with challenges such as sparse data, seasonal market fluctuations, and heterogeneous data sources. The GAN-ResNet model excels in capturing the nonlinear fluctuations in production data, particularly in response to external factors such as climate conditions. Its capacity to integrate data from diverse sources—including weather data and historical yield data—made it highly effective for production forecasting, especially in regions with varying climatic conditions. The VAE-Ridge model addressed the issue of data sparsity, particularly in the context of consumption data, and provided valuable insights into the underlying relationships between market demand, macroeconomic factors, and seasonal fluctuations. Finally, the Adaptive-Transformer model stand out in price prediction, with its ability to capture both short-term price fluctuations and long-term price trends, even under extreme market conditions. [Conclusions] This study presents a comprehensive deep learning-based forecasting approach for agricultural market monitoring and early warning. The integration of multiple models for production, consumption, and price prediction provides a systematic, effective, and scalable tool for supporting agricultural decision-making. The proposed models demonstrate excellent performance in handling the nonlinearities and seasonal fluctuations characteristic of agricultural markets. Furthermore, the models' ability to process and integrate heterogeneous data sources enhances their predictive power and makes them highly suitable for application in real-world agricultural monitoring systems. Future research will focus on optimizing model parameters, enhancing model adaptability, and expanding the system to incorporate additional agricultural products and more complex market conditions. These improvements will help increase the stability and practical applicability of the system, thus further enhancing its potential for real-time market monitoring and early warning capabilities.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight YOLOv8s-Based Strawberry Plug Seedling Grading Detection and Localization via Channel Pruning
    CHEN Junlin, ZHAO Peng, CAO Xianlin, NING Jifeng, YANG Shuqin
    Smart Agriculture    2024, 6 (6): 132-143.   DOI: 10.12133/j.smartag.SA202408001
    Abstract531)   HTML67)    PDF(pc) (3614KB)(1276)       Save

    [Objective] Plug tray seedling cultivation is a contemporary method known for its high germination rates, uniform seedling growth, shortened transplant recovery period, diminished pest and disease incidence, and enhanced labor efficiency. Despite these advantages, challenges such as missing or underdeveloped seedlings can arise due to seedling quality and environmental factors. To ensure uniformity and consistency of the seedlings, sorting is frequently necessary, and the adoption of automated seedling sorting technology can significantly reduce labor costs. Nevertheless, the overgrowth of seedlings within the plugs can effect the accuracy of detection algorithms. A method for grading and locating strawberry seedlings based on a lightweight YOLOv8s model was presented in this research to effectively mitigate the interference caused by overgrown seedlings. [Methods] The YOLOv8s model was selected as the baseline for detecting different categories of seedlings in the strawberry plug tray cultivation process, namely weak seedlings, normal seedlings, and plug holes. To improve the detection efficiency and reduce the model's computational cost, the layer-adaptive magnitude-based pruning(LAMP) score-based channel pruning algorithm was applied to compress the base YOLOv8s model. The pruning procedure involved using the dependency graph to derive the group matrices, followed by normalizing the group importance scores using the LAMP Score, and ultimately pruning the channels according to these processed scores. This pruning strategy effectively reduced the number of model parameters and the overall size of the model, thereby significantly enhancing its inference speed while maintaining the capability to accurately detect both seedlings and plug holes. Furthermore, a two-stage seedling-hole matching algorithm was introduced based on the pruned YOLOv8s model. In the first stage, seedling and plug hole bounding boxes were matched according to their the degree of overlap (Dp), resulting in an initial set of high-quality matches. This step helped minimize the number of potential matching holes for seedlings exhibiting overgrowth. Subsequently, before the second stage of matching, the remaining unmatched seedlings were ranked according to their potential matching hole scores (S), with higher scores indicating fewer potential matching holes. The seedlings were then prioritized during the second round of matching based on these scores, thus ensuring an accurate pairing of each seedling with its corresponding plug hole, even in cases where adjacent seedling leaves encroached into neighboring plug holes. [Results and Discussions] The pruning process inevitably resulted in the loss of some parameters that were originally beneficial for feature representation and model generalization. This led to a noticeable decline in model performance. However, through meticulous fine-tuning, the model's feature expression capabilities were restored, compensating for the information loss caused by pruning. Experimental results demonstrated that the fine-tuned model not only maintained high detection accuracy but also achieved significant reductions in FLOPs (86.3%) and parameter count (95.4%). The final model size was only 1.2 MB. Compared to the original YOLOv8s model, the pruned version showed improvements in several key performance metrics: precision increased by 0.4%, recall by 1.2%, mAP by 1%, and the F1-Score by 0.1%. The impact of the pruning rate on model performance was found to be non-linear. As the pruning rate increased, model performance dropped significantly after certain crucial channels were removed. However, further pruning led to a reallocation of the remaining channels' weights, which in some cases allowed the model to recover or even exceed its previous performance levels. Consequently, it was necessary to experiment extensively to identify the optimal pruning rate that balanced model accuracy and speed. The experiments indicated that when the pruning rate reached 85.7%, the mAP peaked at 96.4%. Beyond this point, performance began to decline, suggesting that this was the optimal pruning rate for achieving a balance between model efficiency and performance, resulting in a model size of 1.2 MB. To further validate the improved model's effectiveness, comparisons were conducted with different lightweight backbone networks, including MobileNetv3, ShuffleNetv2, EfficientViT, and FasterNet, while retaining the Neck and Head modules of the original YOLOv8s model. Results indicated that the modified model outperformed these alternatives, with mAP improvements of 1.3%, 1.8%, 1.5%, and 1.1%, respectively, and F1-Score increases of 1.5%, 1.8%, 1.1%, and 1%. Moreover, the pruned model showed substantial advantages in terms of floating-point operations, model size, and parameter count compared to these other lightweight networks. To verify the effectiveness of the proposed two-stage seedling-hole matching algorithm, tests were conducted using a variety of complex images from the test set. Results indicated that the proposed method achieved precise grading and localization of strawberry seedlings even under challenging overgrowth conditions. Specifically, the correct matching rate for normal seedlings reached 96.6%, for missing seedlings 84.5%, and for weak seedlings 82.9%, with an average matching accuracy of 88%, meeting the practical requirements of the strawberry plug tray cultivation process. [Conclusions] The pruned YOLOv8s model successfully maintained high detection accuracy while reducing computational costs and improving inference speed. The proposed two-stage seedling-hole matching algorithm effectively minimized the interference caused by overgrown seedlings, accurately locating and classifying seedlings of various growth stages within the plug tray. The research provides a robust and reliable technical solution for automated strawberry seedling sorting in practical plug tray cultivation scenarios, offering valuable insights and technical support for optimizing the efficiency and precision of automated seedling grading systems.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Parametric Reconstruction Method of Wheat Leaf Curved Surface Based on Three-Dimensional Point Cloud
    ZHU Shunyao, QU Hongjun, XIA Qian, GUO Wei, GUO Ya
    Smart Agriculture    2025, 7 (1): 85-96.   DOI: 10.12133/j.smartag.SA202410004
    Abstract502)   HTML51)    PDF(pc) (2475KB)(957)       Save

    [Objective] Plant leaf shape is an important part of plant architectural model. Establishment of a three-dimensional structural model of leaves may assist simulating and analyzing plant growth. However, existing leaf modeling approaches lack interpretability, invertibility, and operability, which limit the estimation of model parameters, the simulation of leaf shape, the analysis and interpretation of leaf physiology and growth state, and model reusage. Aiming at the interoperability between three-dimensional structure representation and mathematical model parameters, this study paid attention to three aspects in wheat leaf shape parametric reconstruction: (1) parameter-driven model structure, (2) model parameter inversion, and (3) parameter dynamic mapping during growth. Based on this, a set of parameter-driven and point cloud inversion model for wheat leaf interoperability was proposed in this study. [Methods] A parametric surface model of a wheat leaf with seven characteristic parameters by using parametric modeling technology was built, and the forward parametric construction of the wheat leaf structure was realized. Three parameters, maximum leaf width, leaf length, and leaf shape factor, were used to describe the basic shape of the blade on the leaf plane. On this basis, two parameters, namely the angle between stems and leaves and the curvature degree, were introduced to describe the bending characteristics of the main vein of the blade in the three-dimensional space. Two parameters, namely the twist angle around the axis and the twist deviation angle around the axis, were introduced to represent the twisted structure of the leaf blade along the vein. The reverse parameter estimation module was built according to the surface model. The point cloud was divided by the uniform segmentation method along the Y-axis, and the veins were fit by a least squares regression method. Then, the point cloud was re-segmented according to the fit vein curve. Subsequently, the rotation angle was precisely determined through the segment-wise transform estimation method, with all parameters being optimally fit using the RANSAC regression algorithm. To validate the reliability of the proposed methodology, a set of sample parameters was randomly generated, from which corresponding sample point clouds were synthesized. These sample point clouds were then subjected to estimation using the described method. Then error analyzing was carried out on the estimation results. Three-dimensional imaging technology was used to collect the point clouds of Zhengmai 136, Yangmai 34, and Yanmai 1 samples. After noise reduction and coordinate registration, the model parameters were inverted and estimated, and the reconstructed point clouds were produced using the parametric model. The reconstruction error was validated by calculating the dissimilarity, represented by the Chamfer Distance, between the reconstructed point cloud and the measured point cloud. [Results and Discussions] The model could effectively reconstruct wheat leaves, and the average deviation of point cloud based parametric reconstruction results was about 1.2 mm, which had a high precision. Parametric modeling technology based on prior knowledge and point cloud fitting technology based on posterior data was integrated in this study to construct a digital twin model of specific species at the 3D structural level. Although some of the detailed characteristics of the leaves were moderately simplified, the geometric shape of the leaves could be highly restored with only a few parameters. This method was not only simple, direct and efficient, but also had more explicit geometric meaning of the obtained parameters, and was both editable and interpretable. In addition, the practice of using only tools such as rulers to measure individual characteristic parameters of plant organs in traditional research was abandoned in this study. High-precision point cloud acquisition technology was adopted to obtain three-dimensional data of wheat leaves, and pre-processing work such as point cloud registration, segmentation, and annotation was completed, laying a data foundation for subsequent research. [Conclusions] There is interoperability between the reconstructed model and the point cloud, and the parameters of the model can be flexibly adjusted to generate leaf clusters with similar shapes. The inversion parameters have high interpretability and can be used for consistent and continuous estimation of point cloud time series. This research is of great value to the simulation analysis and digital twinning of wheat leaves.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Defogging Remote Sensing Images Method Based on a Hybrid Attention-Based Generative Adversarial Network
    MA Liu, MAO Kebiao, GUO Zhonghua
    Smart Agriculture    2025, 7 (2): 172-182.   DOI: 10.12133/j.smartag.SA202410011
    Abstract493)   HTML31)    PDF(pc) (1999KB)(298)       Save

    [Objective] Remote sensing images have become an important data source in fields such as surface observation, environmental monitoring, and natural disaster prediction. However, the acquisition of remote sensing images is often affected by weather phenomena such as fog and clouds, which reduces the image quality and poses challenges to subsequent analysis and processing tasks. In recent years, the introduction of attention mechanisms has enabled models to better capture and utilize important features in images, thereby significantly improving defogging performance. However, traditional channel attention mechanisms usually rely on global average pooling to summarize feature information. Although this method simplifies the complexity of calculations, it is not satisfactory when dealing with images with significant local changes and sensitivity to outliers. In addition, remote sensing images usually cover a wide area, and the diverse terrain makes the fog pattern more complex. Therefore, to address this issue, a hybrid attention-based generative adversarial network hybrid attention-based generative adversarial network (HAB-GAN) was proposed in this research, which integrates an efficient channel attention (ECA) module and a spatial attention block (SAB). [Method] By merging feature extraction from both channel and spatial dimensions, the model effectively enhanced its ability to identify and recover hazy areas in remote sensing images. In HAB-GAN, the ECA module captured local cross-channel interactions, addressing the shortcomings of traditional global averaged pooling in terms of insufficient sensitivity to local detail information. The ECA module used a global average pooling strategy without dimensionality reduction, automatically adapting to the characteristics of each channel without introducing extra parameters, thereby enhancing the inter-channel dependencies. ECA emploied a one-dimensional convolution operation, which used a learnable kernel size to adaptively determine the range of channel interactions. This design effectively avoided the over-smoothing of global features common in traditional pooling layers, allowing the model to more precisely extract local detailed while maintaining low computational complexity. The SAB module introduced a weighted mechanism on the spatial dimension by constructing a spatial attention map to enhance the model's ability to identify hazy areas in the image. This module extracted feature maps through convolution operations and applies attention weighting in both horizontal and vertical directions, highlighting regions with severe haze, allowing the model to better capture spatial information in the image, thereby enhancing dehazing performance. The generator of HAB-GAN combined residual network structures with hybrid attention modules. It first extracted initial features from input images through convolutional layers and then passed these features through several residual blocks. The residual blocks effectively mitigated the vanishing gradient problem in deep neural networks and maintain feature consistency and continuity by passing input features directly to deeper network layers through skip connections. Each residual block incorporated ECA and SAB modules, enabling precise feature learning through weighted processing in both channel and spatial dimensions. After extracting effective features, the generator generated dehazed images through convolution operations. The discriminator adopted a standard convolutional neural network architecture, focusing on extracting local detail features from the images generated by the generator. It consisted of multiple convolutional layers, batch normalization layers, and Leaky ReLU activation functions. By extracting local features layer by layer and down-sampling, the discriminator progressively reduced the spatial resolution of the images, evaluating their realism at both global and local levels. The generator and discriminator were jointly optimized through adversarial training, where the generator aimed to produce increasingly realistic dehazed images, and the discriminator continually improved its ability to distinguish between real and generated images, thereby enhancing the learning effectiveness and image quality of the generator. [Results and Discussions] To validate the effectiveness of HAB-GAN, experiments were conducted on the remote sensing image scene classification 45 (RESISC45) dataset. The experimental results demonstrated that compared to existing dehazing models, HAB-GAN excels in key evaluation metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). Specifically, compared to SpA GAN, HAB-GAN improved PSNR by 2.642 5 dB and SSIM by 0.012 2; Compared to HyA-GAN, PSNR improved by 1.138 dB and SSIM by 0.001 9. Additionally, to assess the generalization capability of HAB-GAN, further experiments were conducted on the RICE2 dataset to verify its performance in cloud removal tasks. The results showed that HAB-GAN also performs exceptionally well in cloud removal tasks, with PSNR improving by 3.593 2 dB and SSIM improving by 0.040 2. Compared to HyA-GAN, PSNR and SSIM increased by 1.854 dB and 0.012 4, respectively. To further explored the impact of different modules on the model's performance, ablation experiments were designed, gradually removing the ECA module, the SAB module, and the entire hybrid attention module. The experimental results showed that removing the ECA module reduced PSNR by 2.642 5 dB and SSIM by 0.012 2; Removing the SAB module reduced PSNR by 2.955 dB and SSIM by 0.008 7, and removing the entire hybrid attention module reduced PSNR and SSIM by 3.866 1 dB and 0.033 4, respectively. [Conclusions] The proposed HAB-GAN model not only performs excellently in dehazing and beclouding tasks but also significantly enhances the clarity and detail recovery of dehazed images through the synergistic effect of the ECA module and the SAB module. Additionally, its strong performance across different remote sensing datasets further validates its effectiveness and generalization ability, showcasing broad application potential particularly in fields such as agriculture, environmental monitoring, and disaster prediction, where high-quality remote sensing data is crucial. HAB-GAN is poised to become a valuable tool for improving data reliability and supporting more accurate decision-making and analysis.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Accurate Detection of Tree Planting Locations in Inner Mongolia for The Three North Project Based on YOLOv10-MHSA
    XIE Jiyuan, ZHANG Dongyan, NIU Zhen, CHENG Tao, YUAN Feng, LIU Yaling
    Smart Agriculture    2025, 7 (3): 108-119.   DOI: 10.12133/j.smartag.SA202410010
    Abstract484)   HTML74)    PDF(pc) (3031KB)(327)       Save

    [Objective] The traditional manual field investigation method of the tree planting locations is not only inefficient but also error-prone, and the low-altitude unmanned aerial vehicle (UAV) has become the best choice to solve these problems. To solve the problem of accuracy and efficiency in the detection of tree planting locations (tree pits) in Inner Mongolia of China's Three North Project, an accurate recognition and detection model of tree planting locations based on YOLOv10-MHSA was proposed. [Methods] A long-endurance, multi-purpose vertical take-off and landing (VTOL) fixed-wing UAV was used to collect images of tree planting locations. Equipped with a 26-megapixel camera with high spatial resolution, the UAV was well-suited for high-precision field mapping. Aerial photography was conducted between 11:00 and 12:00 on August 1, 2024. Flight parameters were set as follows: Altitude of 150 m (yielding a ground resolution of approximately 2.56 cm), course overlap rate of 75%, side overlap rate of 65%, and flight speed of 20 m/s. To prevent overfitting during network training, the original data set was enhanced. To improve the quality and efficiency of model training, different attention mechanisms and optimizing loss functions were introduced. Specifically, a more effective EIOU loss function was introduced, comprising three components: IOU loss, distance loss, and azimuth loss. This function directly minimizes the width and height discrepancies between the target frame and anchor, leading to faster convergence and more accurate positioning. Additionally, the Focal-EIOU loss function was adopted to address sample imbalance in bounding box regression tasks, further improving the model's convergence speed and positioning precision. [Results and Discussions] After the introduction of the multi-head self-attention mechanism (MHSA), the model achieved improvements of 1.4% and 1.7% in the evaluation metrics AP@0.5 and AP@0.5:0.95, respectively, and the accuracy and recall rate were also improved. This indicates that MHSA effectively aids the model in extracting the feature information of the target and improving the detection accuracy in complex background. Although the processing speed of the model decreases slightly after adding the attention mechanism, it could still meet the requirements of real-time detection. The experiment compared four loss functions: CIOU, SIOU, EIOU and Focal-EIOU. The results showed that the Focal-EIOU loss function yielded significant increases in precision and recall. This demonstrated that the Focal-EIOU loss function could accelerate the convergence speed of the model and improve the positioning accuracy when dealing with the sample imbalance problem in small target detection. Finally, an improved model, YOLOv10-MHSA, was proposed, incorporating MHSA attention mechanism, small target detection layer and Focal-EIOU loss function. The results of ablation experiments showed that AP@0.5 and AP@0.5:0.95 were increased by 2.2% and 0.9%, respectively, after adding only small target detection layer on the basis of YOLOv10n, and the accuracy and recall rate were also significantly improved. When the MHSA and Focal-EIOU loss functions were further added, the model detection effect was significantly improved. Compared with the baseline model YOLOv10n, the AP@0.5, AP@0.5:0.95, P-value and R-value were improved by 6.6%, 10.0%, 4.1% and 5.1%, respectively. Although the FPS was reduced, the detection performance of the improved model was significantly better than that of the original model in various complex scenes, especially for small target detection in densely distributed and occluded scenes. [Conclusions] By introducing MHSA and the optimized loss function (Focal-EIOU) into YOLOv10n model, the research significantly improved the accuracy and efficiency of tree planting location detection in the Three North Project in Inner Mongolia. The experimental results show that MHSA can enhance the ability of the model to extract local and global information of the target in complex background, and effectively reduce the phenomenon of missed detection and false detection. The Focal-EIOU loss function accelerates the convergence speed of the model and improves the positioning accuracy by optimizing the sample imbalance problem in the bounding box regression task. Although the model processing speed has decreased, the method proposed still meets the real-time detection requirements, provides strong technical support for the scientific afforestation of the Three North Project.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Knowledge Graph Driven Grain Big Data Applications: Overview and Perspective
    YANG Chenxue, LI Xian, ZHOU Qingbo
    Smart Agriculture    2025, 7 (2): 26-40.   DOI: 10.12133/j.smartag.SA202501004
    Abstract469)   HTML55)    PDF(pc) (1435KB)(629)       Save

    [Significance] Grain production spans multiple stages and involves numerous heterogeneous factors, including agronomic inputs, natural resources, environmental conditions, and socio-economic variables. However, the associated data generated throughout the entire production process, ranging from cultivation planning to harvest evaluation, remains highly fragmented, unstructured, and semantically diverse. This complexity data, combined with the lack of integrated core algorithms to support decision-making, has severely limited the potential of big data to drive innovation in grain production. Knowledge graph technology, by offering structured and semantically-rich representations of complex data, enables the integration of multi-source and heterogeneous data, enhances semantic mining and reasoning capabilities, and provides intelligent, knowledge-driven support for sustainable grain production, thereby addressing these challenges effectively. [Progress] This paper systematically reviewed the current research and application progress of knowledge graphs in the grain production big data. A comprehensive knowledge graph driven framework was proposed based on a hybrid paradigm combining data-driven modeling and domain knowledge guidance to support the entire grain production lifecycle and addressed three primary dimensions of data complexity: Structural diversity, relational heterogeneity, and semantic ambiguity. The key techniques of constructing multimodal knowledge map and temporal reasoning for grain production were described. First, an agricultural ontology system for grain production was designed, incorporating domain-specific concepts, hierarchical relationships, and attribute constraints. This ontology provided the semantic foundation for knowledge modeling and alignment. Second, multimodal named entity recognition (NER) techniques were employed to extract entities such as crops, varieties, weather conditions, operations, and equipment from structured and unstructured data sources, including satellite imagery, agronomic reports, Internet of Things sensor data, and historical statistics. Advanced deep learning models, such as bidirectional encoder representations from transformers (BERT) and vision-language transformers, were used to enhance recognition accuracy across text and image modalities. Third, the system implemented multimodal entity linking and disambiguation, which connected identical or semantically similar entities across different data sources by leveraging graph embeddings, semantic similarity measures, and rule-based matching. Finally, temporal reasoning modules were constructed using temporal knowledge graphs and logical rules to support dynamic inference over time-sensitive knowledge, such as crop growth stages, climate variations, and policy interventions. The proposed knowledge graph driven system enabled the development of intelligent applications across multiple stages of grain production. In the pre-production stage, knowledge graphs supported decision-making in resource allocation, crop variety selection, and planting schedule optimization based on past data patterns and predictive inference. During the in-production stage, the system facilitated precision operations, such as real-time fertilization and irrigation by reasoning over current field status, real-time sensor inputs, and historical trends. In the post-production stage, it enabled yield assessment and economic evaluation through integration of production outcomes, environmental factors, and policy constraints. [Conclusions and Prospects] Knowledge graph technologies offer a scalable and semantically-enhanced approach for unlocking the full potential of grain production big data. By integrating heterogeneous data sources, representing domain knowledge explicitly, and supporting intelligent reasoning, knowledge graphs can provide visualization, explainability, and decision support across various spatial scales, including national, provincial, county-level, and large-scale farm contexts. These technologies are of great scientific and practical significance in supporting China's national food security strategy and advancing the goals of storing grain in the land and storing grain in technology. Future directions include the construction of cross-domain agricultural knowledge fusion systems, dynamic ontology evolution mechanisms, and federated knowledge graph platforms for multi-region data collaboration under data privacy constraints.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Extracting Method of the Cultivation Aera of Rice Based on Sentinel-1/2 and Google Earth Engine (GEE): A Case Study of the Hangjiahu Plain
    E Hailin, ZHOU Decheng, LI Kun
    Smart Agriculture    2025, 7 (2): 81-94.   DOI: 10.12133/j.smartag.SA202502003
    Abstract469)   HTML36)    PDF(pc) (2961KB)(293)       Save

    [Objective] Accurate monitoring of rice planting areas is vital for ensuring national food security, evaluating greenhouse gas emissions, optimizing water resource allocation, and maintaining agricultural ecosystems. In recent years, the integration of remote sensing technologies—particularly the fusion of optical and synthetic aperture radar (SAR) data—has significantly enhanced the capacity to monitor crop distribution, even under challenging weather conditions. However, many current studies still rely heavily on phenological features captured at specific key stages, such as the transplanting phase, while overlooking the complete temporal dynamics of vegetation and water-related indices throughout the entire rice growth cycle. There is an urgent need for a method that fully leverages the time-series characteristics of remote sensing indices to enable accurate, scalable, and timely rice mapping. [Methods] Focusing on the Hangjiahu Plain, a typical rice-growing region in eastern China, a novel approach—dynamic NDVI-SDWI Fusion method for rice mapping (DNSF-Rice) was proposed in this research to accurately extract rice planting areas by synergistically integrating Sentinel-1 SAR and Sentinel-2 optical imagery on the google earth engine (GEE) platform. The methodological framework included the following three steps: First, using Sentinel-2 imagery, a time series of the normalized difference vegetation index (NDVI) was constructed. By analyzing its temporal dynamics across key rice growth stages, potential rice planting areas were identified through a threshold-based classification method; Second, a time series of the Sentinel-1 dual-polarized water index (SDWI) was generated to analyze its dynamic changes throughout the rice growth cycle. A thresholding algorithm was then applied to extract rice field distribution based on microwave data, considering the significant irrigation involved in rice cultivation; Finally, the spatial intersection of the NDVI-derived and SDWI-derived results was intersected to generate the final rice planting map. This step ensures that only pixels exhibiting both vegetation growth and irrigation signals were classified as rice. The classification datasets spanned five consecutive years from 2019 to 2023, with a spatial resolution of 10 m. [Results and Discussions] The proposed method demonstrated high accuracy and robust performance in mapping rice planting areas. Over the study period, the method achieved an overall accuracy of over 96% and an F1-Score exceeding 0.96, outperforming several benchmark products in terms of spatial consistency and precision. The integration of NDVI and SDWI time-series features enabled effective identification of rice fields, even under the challenging conditions of frequent cloud cover and variable precipitation typical in the study area. Interannual analysis revealed a consistent increase in rice planting areas across the Hangjiahu Plain from 2019 to 2023. The remote sensing-based rice area estimates were in strong agreement with official agricultural statistics, further validating the reliability of the proposed method. The fusion of optical and SAR data proved to be a valuable strategy, effectively compensating for the limitations inherent in single-source imagery, especially during the cloudy and rainy seasons when optical imagery alone was often insufficient. Furthermore, the use of GEE facilitated the rapid processing of large-scale time-series data, supporting the operational scalability required for regional rice monitoring. This study emphasized the critical importance of capturing the full temporal dynamics of both vegetation and water signals throughout the entire rice growth cycle, rather than relying solely on fixed phenological stages. [Conclusions] By leveraging the complementary advantages of optical and SAR imagery and utilizing the complete time-series behavior of NDVI and SDWI indices, the proposed approach successfully mapped rice planting areas across a complex monsoon climate region over a five-year period. The method has been proven to be stable, reproducible, and adaptable for large-scale agricultural monitoring applications.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Extraction Method of Maize Plant Skeleton and Phenotypic Parameters Based on Improved YOLOv11-Pose
    NIU Ziang, QIU Zhengjun
    Smart Agriculture    2025, 7 (2): 95-105.   DOI: 10.12133/j.smartag.SA202501001
    Abstract456)   HTML55)    PDF(pc) (2260KB)(540)       Save

    [Objective] Accurate extraction of maize plant skeletons and phenotypic parameters is fundamental for acquisition of plant growth data, morphological analysis, and agricultural management. However, leaf occlusion and complex backgrounds in dense planting environments pose significant challenges to skeleton and parameters extraction. A maize plant skeleton and phenotypic parameters extraction method suitable for dense field environments was proposed in this research to enhance the extraction precision and efficiency, and provide technical support for maize growth data acquisition. [Methods] An improved YOLOv11-Pose multi-object keypoint detection network was introduced, a top-down detection framework was adopted to detect maize plant keypoints and reconstruct skeletons. A uniform sampling algorithm was used to design a keypoint representation method tailored for maize skeletons and optimize task adaptability. Additionally, a single-head self-attention mechanism and a convolutional block attention module were incorporated to guide the model's focus on occluded regions and connected parts, thereby improve its adaptability to complex scenarios. [Results and Discussion] In dense field maize environments, experimental results showed that when the number of uniformly sampled keypoints was set to 10, the Fréchet distance reached its minimum value of 79.008, effectively preserving the original skeleton's morphological features while avoiding the negative impact of redundant points. Under this configuration, the improved YOLOv11-Pose model achieved a bounding box detection precision of 0.717. The keypoint detection mAP50 and mAP50-95 improved by 10.9% and 23.8%, respectively, compared to the original model, with an inference time of 52.7 ms per image. The results demonstrated the model's superior performance and low computational cost in complex field environments, particularly in keypoint detection tasks with enhanced accuracy and robustness. The study further combined the results of skeleton extraction and spatial geometric information to achieve a plant height measurement mean average error (MAE) of 2.435 cm, the detection error of leaf age was less than one growth period, and the measurement error of leaf length was 3.482%, verifying the effectiveness and practicability of the proposed method in the application of phenotypic parameter measurement. [Conclusion] The proposed improved YOLOv11-Pose model can efficiently and accurately extract maize plant skeletons, meeting the demands of ground-based maize growth data acquisition. The research could provide technical support for phenotypic data acquisition in grain production and precision agricultural management.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grape Recognition and Localization Method Based on 3C-YOLOv8n and Depth Camera
    LIU Chang, SUN Yu, YANG Jing, WANG Fengchao, CHEN Jin
    Smart Agriculture    2024, 6 (6): 121-131.   DOI: 10.12133/j.smartag.SA202407008
    Abstract444)   HTML41)    PDF(pc) (2020KB)(1182)       Save

    [Objective] Grape picking is a key link in increasing production. However, in this process, a large amount of manpower and material resources are required, which makes the picking process complex and slow. To enhance harvesting efficiency and achieve automated grape harvesting, an improved YOLOv8n object detection model named 3C-YOLOv8n was proposed, which integrates the RealSense D415 depth camera for grape recognition and localization. [Methods] The propoesed 3C-YOLOv8n incorporated a convolutional block attention module (CBAM) between the first C2f module and the third Conv module in the backbone network. Additionally, a channel attention (CA) module was added at the end of the backbone structure, resulting in a new 2C-C2f backbone network architecture. This design enabled the model to sequentially infer attention maps across two independent dimensions (channel and spatial), optimize features by considering relationships between channels and positional information. The network structure was both flexible and lightweight. Furthermore, the Content-aware ReAssembly of Features up sampling operator was implemented to support instance-specific kernels (such as deconvolution) for feature reconstruction with neighboring pixels, replacing the nearest neighbor interpolation operator in the YOLOv8n neck network. This enhancement increased the receptive field and guided the reconstruction process based on input features while maintaining low parameter and computational complexity, thereby forming the 3C-YOLOv8n model. The pyrealsense2 library was utilized to obtain pixel position information from the target area using the Intel RealSense D415 camera. During this process, the depth camera was used to capture images, and target detection algorithms were employed to pinpoint the location of grapes. The camera's depth sensor facilitated the acquisition of the three-dimensional point cloud of grapes, allowing for the calculation of the distance from the pixel point to the camera and the subsequent determination of the three-dimensional coordinates of the center of the target's bounding box in the camera coordinate system, thus achieving grape recognition and localization. [Results and Discussions] Comparative and ablation experiments were conducted. it was observed that the 3C-YOLOv8n model achieved a mean average precision (mAP) of 94.3% at an intersection ratio of 0.5 (IOU=0.5), surpassing the YOLOv8n model by 1%. The accuracy (P) and recall (R) rates were recorded at 91.6% and 86.4%, respectively, reflecting increases of 0.1% and 0.7%. The F1-Score also improved by 0.4%, demonstrating that the improved network model met the experimental accuracy and recall requirements. In terms of loss, the 3C-YOLOv8n algorithm exhibited superior performance, with a rapid decrease in loss values and minimal fluctuations, ultimately leading to a minimized loss value. This indicated that the improved algorithm quickly reached a convergence state, enhancing both model accuracy and convergence speed. The ablation experiments revealed that the original YOLOv8n model yielded a mAP of 93.3%. The integration of the CBAM and CA attention mechanisms into the YOLOv8n backbone resulted in mAP values of 93.5% each. The addition of the Content-aware ReAssembly of Features up sampling operator to the neck network of YOLOv8n produced a 0.5% increase in mAP, culminating in a value of 93.8%. The combination of the three improvement strategies yielded mAP increases of 0.3, 0.7, and 0.8%, respectively, compared to the YOLOv8n model. Overall, the 3C-YOLOv8n model demonstrated the best detection performance, achieving the highest mAP of 94.3%. The ablation results confirmed the positive impact of the proposed improvement strategies on the experimental outcomes. Compared to other mainstream YOLO series algorithms, all evaluation metrics showed enhancements, with the lowest missed detection and false detection rates among all tested algorithms, underscoring its practical advantages in detection tasks. [Conclusions] By effectively addressing the inefficiencies of manual labor, 3C-YOLOv8n network model not only enhances the precision of grape recognition and localization but also significantly optimizes overall harvesting efficiency. Its superior performance in evaluation metrics such as precision, recall, mAP, and F1-Score, alongside the lowest recorded loss values among YOLO series algorithms, indicates a remarkable advancement in model convergence and operational effectiveness. Furthermore, the model's high accuracy in grape target recognition not only lays the groundwork for automated harvesting systems but also enables the implementation of complementary intelligent operations.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Rice Leaf Disease Image Enhancement Based on Improved CycleGAN
    YAN Congkuan, ZHU Dequan, MENG Fankai, YANG Yuqing, TANG Qixing, ZHANG Aifang, LIAO Juan
    Smart Agriculture    2024, 6 (6): 96-108.   DOI: 10.12133/j.smartag.SA202407019
    Abstract442)   HTML44)    PDF(pc) (1744KB)(302)       Save

    Objective Rice diseases significantly impact both the yield and quality of rice production. Automatic recognition of rice diseases using computer vision is crucial for ensuring high yields, quality, and efficiency. However, rice disease image recognition faces challenges such as limited availability of datasets, insufficient sample sizes, and imbalanced sample distributions across different disease categories. To address these challenges, a data augmentation method for rice leaf disease images was proposed based on an improved CycleGAN model in this reseach which aimed to expand disease image datasets by generating disease features, thereby alleviating the burden of collecting real disease data and providing more comprehensive and diverse data to support automatic rice disease recognition. Methods The proposed approach built upon the CycleGAN framework, with a key modification being the integration of a convolutional block attention module (CBAM) into the generator's residual module. This enhancement strengthened the network's ability to extract both local key features and global contextual information pertaining to rice disease-affected areas. The model increased its sensitivity to small-scale disease targets and subtle variations between healthy and diseased domains. This design effectively mitigated the potential loss of critical feature information during the image generation process, ensuring higher fidelity in the resulting images. Additionally, skip connections were introduced between the residual modules and the CBAM. These connections facilitate improved information flow between different layers of the network, addressing common issues such as gradient vanishing during the training of deep networks. Furthermore, a perception similarity loss function, designed to align with the human visual system, was incorporated into the overall loss function. This addition enabled the deep learning model to more accurately measure perceptual differences between the generated images and real images, thereby guiding the network towards producing higher-quality samples. This adjustment also helped to reduce visual artifacts and excessive smoothing, while concurrently improving the stability of the model during the training process. To comprehensively evaluate the quality of the rice disease images generated by the proposed model and to assess its impact on disease recognition performance, both subjective and objective evaluation metrics were utilized. These included user perception evaluation (UPE), structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), and the performance of disease recognition within object detection frameworks. Comparative experiments were conducted across multiple GAN models, enabling a thorough assessment of the proposed model's performance in generating rice disease images. Additionally, different attention mechanisms, including efficient channel attention (ECA), coordinate attention (CA), and CBAM, were individually embedded into the generator's residual module. These variations allowed for a detailed comparison of the effects of different attention mechanisms on network performance and the visual quality of the generated images. Ablation studies were further performed to validate the effectiveness of the CBAM residual module and the perception similarity loss function in the network's overall architecture. Based on the generated rice disease samples, transfer learning experiments were conducted using various object detection models. By comparing the performance of these models before and after transfer learning, the effectiveness of the generated disease image data in enhancing the performance of object detection models was empirically verified. Results and Discussions The rice disease images generated by the improved CycleGAN model surpassed those produced by other GAN variants in terms of image detail clarity and the prominence of disease-specific features. In terms of objective quality metrics, the proposed model exhibited a 3.15% improvement in SSIM and an 8.19% enhancement in PSNR compared to the original CycleGAN model, underscoring its significant advantage in structural similarity and signal-to-noise ratio. The comparative experiments involving different attention mechanisms and ablation studies revealed that embedding the CBAM into the generator effectively increased the network's focus on critical disease-related features, resulting in more realistic and clearly defined disease-affected regions in the generated images. Furthermore, the introduction of the perception similarity loss function substantially enhanced the network's ability to perceive and represent disease-related information, thereby improving the visual fidelity and realism of the generated images. Additionally, transfer learning applied to object detection models such as YOLOv5s, YOLOv7-tiny, and YOLOv8s led to significant improvements in disease detection performance on the augmented dataset. Notably, the detection accuracy of the YOLOv5s model increased from 79.7% to 93.8%, representing a considerable enhancement in both generalization ability and robustness. This improvement also effectively reduced the rates of false positives and false negatives, resulting in more stable and reliable performance in rice disease detection tasks. Conclusions The rice leaf disease image generation method based on the improved CycleGAN model, as proposed in this study, effectively transforms images of healthy leaves into those depicting disease symptoms. By addressing the challenge of insufficient disease samples, this method significantly improves the disease recognition capabilities of object detection models. Therefore, it holds considerable application potential in the domain of leaf disease image augmentation and offers a promising new direction for expanding datasets of disease images for other crops.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Detection and Recognition Model for Small Target Pests on Sticky Traps in Multi-Source Scenarios
    YANG Xinting, HU Huan, CHEN Xiao, LI Wenzheng, ZHOU Zijie, LI Wenyong
    Smart Agriculture    2025, 7 (1): 111-123.   DOI: 10.12133/j.smartag.SA202410019
    Abstract439)   HTML39)    PDF(pc) (2514KB)(463)       Save

    [Objective] In crop cultivation and production, pests have gradually become one of the main issues affecting agricultural yield. Traditional models often focus on achieving high accuracy, however, to facilitate model application, lightweighting is necessary. The targets in yellow sticky trap images are often very small with low pixel resolution, so modifications in network structure, loss functions, and lightweight convolutions need to adapt to the detection of small-object pests. Ensuring a balance between model lightweighting and small-object pest detection is particularly important. To improve the detection accuracy of small target pests on sticky trap images from multi-source scenarios, a lightweight detection model named MobileNetV4+VN-YOLOv5s was proposed in this research to detect two main small target pests in agricultural production, whiteflies and thrips. [Methods] In the backbone layer of MobileNetV4+VN-YOLOv5s, an EM block constructed with the MobileNetV4 backbone network was introduced for detecting small, high-density, and overlapping targets, making it suitable for deployment on mobile devices. Additionally, the Neck layer of MobileNetV4+VN-YOLOv5s incorporates the GSConv and VoV-GSCSP modules to replace regular convolutional modules with lightweight design, effectively reducing the parameter size of the model while improving detection accuracy. Lastly, a normalized wasserstein distance (NWD)loss function was introduced into the framework to enhance the sensitivity for low-resolution small target pests. Extensive experiments including state-of-the-art comparison, ablation evaluation, performance analysis on image splitting, pest density and multi-source data were conducted. [Results and Discussions] Through ablation tests, it was concluded that the EM module and the VoV-GSCSP convolution module had significant effects in reducing the model parameter size and frame rate, the NWD loss function significantly improved the mean average precision (mAP) of the model. By comparing tests with different loss functions, the NWD loss function improves the mAP by 6.1, 10.8 and 8.2 percentage compared to the DIoU, GIoU and EIoU loss functions, respectively, so the addition of the NWD loss function achieved good results. Comparative performance tests were detected wiht different light weighting models, the experimental results showed that the mAP of the proposed MobileNetV4+VN-YOLOv5s model in three scenarios (Indoor, Outdoor, Indoor&Outdoor) was 82.5%, 70.8%, and 74.7%, respectively. Particularly, the MobileNetV4+VN-YOLOv5s model had a parameter size of only 4.2 M, 58% of the YOLOv5s model, the frame rate was 153.2 fps, an increase of 6.0 fps compared to the YOLOv5s model. Moreover, the precision and mean average precision reach 79.7% and 82.5%, which were 5.6 and 8.4 percentage points higher than the YOLOv5s model, respectively. Comparative tests were conducted in the upper scenarios based on four splitting ratios: 1×1, 2×2, 5×5, and 10×10. The most superior was the result by using 5×5 ratio in indoor scenario, and the mAP of this case reached 82.5%. The mAP of the indoor scenario was the highest in the low-density case, reaching 83.8%, and the model trained based on the dataset from indoor condition achieves the best performance. Comparative tests under different densities of pest data resulted in a decreasing trend in mAP from low to high densities for the MobileNetV4+VN-YOLOv5s model in the three scenarios. Based on the comparison of the experimental results of different test sets in different scenarios, all three models achieved the best detection accuracy on the IN dataset. Specifically, the IN-model had the highest mAP at 82.5%, followed by the IO-model. At the same time, the detection performance showed the same trend across all three test datasets: The IN model performed the best, followed by the IO-model, and the OUT-model performed the lowest. By comparing the tests with different YOLO improvement models, it was concluded that MobileNetV4+VN-YOLOv5s had the highest mAP, EVN-YOLOv8s was the second highest, and EVN-YOLOv11s was the lowest. Besides, after deploying the model to the Raspberry Pi 4B motherboard, it was concluded that the detection results of the YOLOv5s model had more misdetections and omissions than those of the MobileNetV4+VN-YOLOv5s model, and the time of the model was shortened by about 33% compared to that of the YOLOv5s model, which demonstrated that the model had a good prospect of being deployed in the application. [Conclusions] The MobileNetV4+VN-YOLOv5s model proposed in this study achieved a balance between lightweight design and accuracy. It can be deployed on embedded devices, facilitating practical applications. The model can provide a reference for detecting small target pests in sticky trap images under various multi-source scenarios.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research on the Spatio-temporal Characteristics and Driving Factors of Smart Farm Development in the Yangtze River Economic Belt
    GAO Qun, WANG Hongyang, CHEN Shiyao
    Smart Agriculture    2024, 6 (6): 168-179.   DOI: 10.12133/j.smartag.SA202404005
    Abstract426)   HTML32)    PDF(pc) (1099KB)(482)       Save

    [Objective] In order to summarize exemplary cases of high-quality development in regional smart agriculture and contribute strategies for the sustainable advancement of the national smart agriculture cause, the spatiotemporal characteristics and key driving factors of smart farms in the Yangtze River Economic Belt were studied. [Methods] Based on data from 11 provinces (municipalities) spanning the years 2014 to 2023, a comprehensive analysis was conducted on the spatio-temporal differentiation characteristics of smart farms in the Yangtze River Economic Belt using methods such as kernel density analysis, spatial auto-correlation analysis, and standard deviation ellipse. Including the overall spatial clustering characteristics, high-value or low-value clustering phenomena, centroid characteristics, and dynamic change trends. Subsequently, the geographic detector was employed to identify the key factors driving the spatio-temporal differentiation of smart farms and to discern the interactions between different factors. The analysis was conducted across seven dimensions: special fiscal support, industry dependence, human capital, urbanization, agricultural mechanization, internet infrastructure, and technological innovation. [Results and Discussions] Firstly, in terms of temporal characteristics, the number of smart farms in the Yangtze River Economic Belt steadily increased over the past decade. The year 2016 marked a significant turning point, after which the growth rate of smart farms had accelerated noticeably. The development of the upper, middle, and lower reaches exhibited both commonalities and disparities. Specifically, the lower sub-regions got a higher overall development level of smart farms, with a fluctuating upward growth rate; the middle sub-regions were at a moderate level, showing a fluctuating upward growth rate and relatively even provincial distribution; the upper sub-regions got a low development level, with a stable and slow growth rate, and an unbalanced provincial distribution. Secondly, in terms of spatial distribution, smart farms in the Yangtze River Economic Belt exhibited a dispersed agglomeration pattern. The results of global auto-correlation indicated that smart farms in the Yangtze River Economic Belt tended to be randomly distributed. The results of local auto-correlation showed that the predominant patterns of agglomeration were H-L and L-H types, with the distribution across provinces being somewhat complex; H-H type agglomeration areas were mainly concentrated in Sichuan, Hubei, and Anhui; L-L type agglomeration areas were primarily in Yunnan and Guizhou. The standard deviation ellipse results revealed that the mean center of smart farms in the Yangtze River Economic Belt had shifted from Anqing city in Anhui province in 2014 to Jingzhou city in Hubei province in 2023, with the spatial distribution showing an overall trend of shifting southwestward and a slow expansion toward the northeast and south. Finally, in terms of key driving factors, technological innovation was the primary critical factor driving the formation of the spatio-temporal distribution pattern of smart farms in the Yangtze River Economic Belt, with a factor explanatory degree of 0.311 1. Moreover, after interacting with other indicators, it continued to play a crucial role in the spatio-temporal distribution of smart farms, which aligned with the practical logic of smart farm development. Urbanization and agricultural mechanization levels were the second and third largest key factors, with factor explanatory degrees of 0.292 2 and 0.251 4, respectively. The key driving factors for the spatio-temporal differentiation of smart farms in the upper, middle, and lower sub-regions exhibited both commonalities and differences. Specifically, the top two key factors driver identification in the upper region were technological innovation (0.841 9) and special fiscal support (0.782 3). In the middle region, they were technological innovation (0.619 0) and human capital (0.600 1), while in the lower region, they were urbanization (0.727 6) and technological innovation (0.425 4). The identification of key driving factors and the detection of their interactive effects further confirmed that the spatio-temporal distribution characteristics of smart farms in the Yangtze River Economic Belt were the result of the comprehensive action of multiple factors. [Conclusions] The development of smart farms in the Yangtze River Economic Belt is showing a positive momentum, with both the total number of smart farms and the number of sub-regions experiencing stable growth. The development speed and level of smart farms in the sub-regions exhibit a differentiated characteristic of "lower reaches > middle reaches > upper reaches". At the same time, the overall distribution of smart farms in the Yangtze River Economic Belt is relatively balanced, with the degree of sub-regional distribution balance being "middle reaches (Hubei province, Hunan province, Jiangxi province are balanced) > lower reaches (dominated by Anhui) > upper reaches (Sichuan stands out)". The coverage of smart farm site selection continues to expand, forming a "northeast-southwest" horizontal diffusion pattern. In addition, the spatio-temporal characteristics of smart farms in the Yangtze River Economic Belt are the result of the comprehensive action of multiple factors, with the explanatory power of factors ranked from high to low as follows: Technological innovation > urbanization > agricultural mechanization > human capital > internet infrastructure > industry dependence > special fiscal support. Moreover, the influence of each factor is further strengthened after interaction. Based on these conclusions, suggestions are proposed to promote the high-quality development of smart farms in the Yangtze River Economic Belt. This study not only provides a theoretical basis and reference for the construction of smart farms in the Yangtze River Economic Belt and other regions, but also helps to grasp the current status and future trends of smart farm development.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Precision Target Spraying System Integrated with Remote Deep Learning Recognition Model for Cabbage Plant Centers
    ZHANG Hui, HU Jun, SHI Hang, LIU Changxi, WU Miao
    Smart Agriculture    2024, 6 (6): 85-95.   DOI: 10.12133/j.smartag.SA202406013
    Abstract408)   HTML26)    PDF(pc) (1923KB)(1073)       Save

    [Objective] Spraying calcium can effectively prevent the occurrence of dry burning heart disease in Chinese cabbage. Accurately targeting spraying calcium can more effectively improve the utilization rate of calcium. Since the sprayer needs to move rapidly in the field, this can lead to over-application or under-application of the pesticide. This study aims to develop a targeted spray control system based on deep learning technology, explore the relationship between the advance speed, spray volume, and coverage of the sprayer, thereby addressing the uneven application issues caused by different nebulizer speeds by studying the real scenario of calcium administration to Chinese cabbage hearts. [Methods] The targeted spraying control system incorporates advanced sensors and computing equipment that were capable of obtaining real-time data regarding the location of crops and the surrounding environmental conditions. This data allowed for dynamic adjustments to be made to the spraying system, ensuring that pesticides were delivered with high precision. To further enhance the system's real-time performance and accuracy, the YOLOv8 object detection model was improved. A Ghost-Backbone lightweight network structure was introduced, integrating remote sensing technologies along with the sprayer's forward speed and the frequency of spray responses. This innovative combination resulted in the creation of a YOLOv8-Ghost-Backbone lightweight model specifically tailored for agricultural applications. The model operated on the Jetson Xavier NX controller, which was a high-performance, low-power computing platform designed for edge computing. The system was allowed to process complex tasks in real time directly in the field. The targeted spraying system was composed of two essential components: A pressure regulation unit and a targeted control unit. The pressure regulation unit was responsible for adjusting the pressure within the spraying system to ensure that the output remains stable under various operational conditions. Meanwhile, the targeted control unit played a crucial role in precisely controlling the direction, volume, and coverage of the spray to ensure that the pesticide was applied effectively to the intended areas of the plants. To rigorously evaluate the performance of the system, a series of intermittent spray tests were conducted. During these tests, the forward speed of the sprayer was gradually increased, allowing to assess how well the system responded to changes in speed. Throughout the testing phase, the response frequency of the electromagnetic valve was measured to calculate the corresponding spray volume for each nozzle. [Results and Conclusions] The experimental results indicated that the overall performance of the targeted spraying system was outstanding, particularly under conditions of high-speed operation. By meticulously recording the response times of the three primary components of the system, the valuable data were gathered. The average time required for image processing was determined to be 29.50 ms, while the transmission of decision signals took an average of 6.40 ms. The actual spraying process itself required 88.83 ms to complete. A thorough analysis of these times revealed that the total response time of the spraying system lagged by approximately 124.73 ms when compared to the electrical signal inputs. Despite the inherent delays, the system was able to maintain a high level of spraying accuracy by compensating for the response lag of the electromagnetic valve. Specifically, when tested at a speed of 7.2 km/h, the difference between the actual spray volume delivered and the required spray volume, after accounting for compensation, was found to be a mere 0.01 L/min. This minimal difference indicates that the system met the standard operational requirements for effective pesticide application, thereby demonstrating its precision and reliability in practical settings. [Conclusions] In conclusion, this study developed and validated a deep learning-based targeted spraying control system that exhibited excellent performance regarding both spraying accuracy and response speed. The system serves as a significant technical reference for future endeavors in agricultural automation. Moreover, the research provides insights into how to maintain consistent spraying effectiveness and optimize pesticide utilization efficiency by dynamically adjusting the spraying system as the operating speed varies. The findings of this research will offer valuable experiences and guidance for the implementation of agricultural robots in the precise application of pesticides, with a particular emphasis on parameter selection and system optimization.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Drought Monitoring in Arid Irrigated Areas Based on TVDI Combined with ICEEMDAN-ARIMA Model
    WEI Yuxin, LI Qiao, TAO Hongfei, LU Chunlei, LUO Xu, MAHEMUJIANG Aihemaiti, JIANG Youwei
    Smart Agriculture    2025, 7 (2): 117-131.   DOI: 10.12133/j.smartag.SA202502005
    Abstract404)   HTML17)    PDF(pc) (3498KB)(160)       Save

    [Objective] Drought, one of the most frequent natural disasters globally, is characterized by its extensive impact area, prolonged duration, and significant harm. Large scale irrigation areas, as important pillars of China's agricultural economy, often have their benefits severely restricted by drought disasters. Therefore, quickly and accurately grasping the regional drought situation is of great significance. It can not only effectively improve the utilization efficiency of water resources and reduce agricultural production losses but also promote the sustainable development of regional agriculture. [Methods] The Santun river irrigation area in Xinjiang, an arid - zone irrigation area, was taken as an research object. Based on Landsat TM/ETM+/OLI_TIRS series data, the temperature vegetation drought index (TVDI) and the vegetation temperature condition index (VTCI) were calculated. Using in situ the soil water content of the 0-10 cm soil layer in the study area measured by the Smart Soil Moisture Monitor, an applicability analysis of the drought monitoring effects of TVDI and VTCI was carried out to select the remote sensing monitoring index suitable for drought research in the study area. Based on the selected drought monitoring index, methods such as linear trend analysis and Theil - Sen + Mann - Kendall trend test were used to explore the temporal and spatial distribution characteristics and change trends of drought in the study area from 2005 to 2022. Meanwhile, with the help of machine learning algorithms, an ICEEMDAN - ARIMA combined model was constructed to predict the drought situation in the study area in spring, summer, and autumn of 2023. The prediction performance of the ICEEMDAN - ARIMA combined model was evaluated using root mean square error (RMSE), mean absolute error (MAE), and coefficient of determination (R2). [Results and Discussions] The research results show that there were varying degrees of linear correlations between the two drought indices, TVDI and VTCI, inverted from remote sensing data, and the soil water content of the 0-10 cm surface soil layer in the Santun river irrigation area of Xinjiang. The coefficient of determination between TVDI and the measured soil water content was greater than 0.51 in all periods, with an overall fitting coefficient of 0.57, and the slopes of the fitting equations were all negative, indicating a significant negative correlation. In contrast, the highest coefficient of determination of VTCI was only 0.33, and its overall monitoring effect was significantly weaker than that of TVDI. In terms of temporal and spatial distribution, the drought situation in the study area showed a slow-increasing trend from 2005 to 2022. The growth rate of TVDI was 0.01/10 a, and it had strong spatial heterogeneity, specifically manifested as the spatial distribution characteristic that the southern and southwestern regions of the irrigation area were drier than the northern and northeastern regions. The results of the drought trend analysis indicated that from 2005 to 2022, the distribution of Sen change rate data in the study area conforms to the normal distribution (P < 0.01), and the Sen slopes of more than 72.83% of the regions were greater than zero. At the same time, according to the classification criteria of the Sen + Mann - Kendall trend test, six types of drought change trends were divided. The area proportions of the extremely significant mitigation, significant mitigation, slight mitigation, extremely significant drying, significant drying, and slight drying categories were 0.73%, 1.78%, 24.31%, 5.33%, 9.43%, and 58.42%, respectively. The area proportions of the slight drying and slight mitigation categories were the largest, accounting for a total of 82.73% of the total area of the study area. The ICEEMDAN - ARIMA combined model constructed with the help of machine learning algorithms achieved good results in predicting the drought situation in the study area in 2023. The average value of R2 reached 0.962, demonstrating high robustness and good prediction performance. [Conclusions] The research results systematically characterizes the characteristics of agricultural drought changes in the Santun river irrigation area of Xinjiang over a long - time series, and reveals that the ICEEMDAN - ARIMA combined model has good prediction accuracy in agricultural drought prediction research. This study can provide important references for the construction of drought early warning and forecasting systems, water resource management, and the sustainable development of agriculture in arid-zone irrigation areas.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Analysis of the Spatial Temporal Evolution Pattern and Influencing Factors of Grain Production in Sichuan Province of China
    ZHENG Ling, MA Qianran, JIANG Tao, LIU Xiaojing, MOU Jiahui, WANG Canhui, LAN Yu
    Smart Agriculture    2025, 7 (2): 13-25.   DOI: 10.12133/j.smartag.SA202411013
    Abstract404)   HTML34)    PDF(pc) (3213KB)(169)       Save

    [Objective] Sichuan province, recognized as a strategic core region for China's food security, exhibits spatiotemporal dynamics in grain production that have significant implications for regional resource allocation and national food strategies. Previous studies have primarily focused on spatial dimension of grain production, treating temporal and spatial characteristics separately. This approach neglected the non-stationary effects of time and failed to integrate the evolution patterns of time and space in a coherent manner. Consequently, the interrelationship between the temporal evolution and spatial distribution of grain production was not fully elucidated. A spatiotemporal integrated analysis framework along with a method for extracting spatiotemporal features was proposed in this study. The objective is to elucidate the spatial pattern temporal variations in Sichuan province, thereby providing a scientific basis for regional food security management. [Methods] The study was based on county-level panel data of Sichuan province spanning the years 2000 to 2019. Multiple spatiotemporal analysis techniques were employed to comprehensively examine the evolution of grain production and to identify its driving mechanisms. Initially, standard deviation ellipse analysis and the centroid migration trajectory model were applied to assess the spatial distribution of major grain-producing areas and their temporal migration trends. This analysis enabled the identification of spatial agglomeration patterns and the direction of change in grain production. Subsequently, a three-dimensional spatiotemporal framework was constructed based on the space-time cube model. This framework integrated both temporal and spatial information. Hotspot analysis and the local Moran's I statistic were then utilized to systematically identify the distribution of cold and hot spots as well as spatial clustering patterns in county-level grain output. This approach revealed the spatiotemporal hotspots, clustering characteristics, and the evolving trends of grain production over time. Finally, a spatiotemporal geographically weighted regression model was employed to quantitatively assess the influence of various factors on grain production. These factors included natural elements (such as topography, climate, and soil properties), agricultural factors (such as the total sown area, mechanization level, and irrigation conditions), economic factors (such as per capita gross domestic product and rural per capita disposable income), and human factors (such as rural population and nighttime light intensity). The analysis elucidated the spatial heterogeneity and evolution of the principal driving forces affecting grain production in the province. [Results and Discussions] A high-yield core area was established on the eastern Sichuan plain, with its spatial distribution exhibiting a pronounced northeast-southwest orientation. The production centroid consistently remained near Lezhi County, although it experienced significant shifts during the periods 2000-2001 and 2009-2010. In contrast, the grain production levels in the western Sichuan plateau and the central hilly regions were relatively low. Over the past two decades, the province demonstrated seven distinct patterns in the distribution of cold and hot spots and three clustering patterns in grain production. Specifically, grain output on the Chengdu Plain continuously increased, the decline in production on the western plateau decelerated, and production in the central region consistently decreased. Approximately 64.77% of the province exhibited potential for increased production, particularly in the western region, where improvements in natural conditions and the gradual enhancement of agricultural infrastructure contributed to significant yield growth potential. Conversely, roughly 16.93% of the areas, characterized by complex topography and limited resources, faced potential yield reductions due to resource scarcity and restrictive cultivation conditions. The analysis further revealed that agricultural factors served as the dominant determinants influencing the spatiotemporal characteristics of grain production. In this regard, the total sown area and the area of cultivated land acted as positive contributors. Natural factors, including slope, soil pH, and annual sunshine duration, exerted negative effects. Although human and economic factors had relatively minor influences, indicators such as population density and nighttime light intensity also played a moderating role in regional grain production. The maintenance of agricultural land area proved crucial in safeguarding and enhancing grain yields, while improvements in natural resource conditions further bolstered production capacity. These findings underscored the inherent spatiotemporal disparities in grain production within Sichuan province and revealed the impact of agricultural resource allocation, environmental conditions, and policy support on the heterogeneity of spatial production changes. [Conclusions] The proposed spatiotemporal integrated analysis framework provided a novel perspective for elucidating the dynamic evolution and driving mechanisms of grain production in Sichuan province. The findings demonstrated that the grain production pattern exhibited complex characteristics, including regional concentration, dynamic spatiotemporal evolution, and the interplay of multiple factors. Based on these results, future policies should emphasize the construction of high-standard farmland, the promotion of precision agriculture technologies, and the rational adjustment of agricultural resource allocation. Such measures are intended to enhance agricultural production efficiency and to improve the regional eco-agricultural system. Ultimately, these recommendations aim to furnish both theoretical support and practical guidance for the establishment of a stable and efficient grain production system and for advancing the development of Sichuan province as a key granary.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Recognition of Sugarcane Leaf Diseases in Complex Backgrounds Based on Deep Network Ensembles
    MA Weiwei, CHEN Yue, WANG Yongmei
    Smart Agriculture    2025, 7 (1): 136-145.   DOI: 10.12133/j.smartag.SA202411026
    Abstract396)   HTML45)    PDF(pc) (1385KB)(232)       Save

    [Objective] Sugarcane is an important cash crop, and its health status affects crop yields. However, under natural environmental conditions, the identification of sugarcane leaf diseases is a challenging problem. There are various issues such as disease spots on sugarcane leaves being blocked and interference from lighting, which make it extremely difficult to comprehensively obtain disease information, thus significantly increasing the difficulty of disease identification. Early image recognition algorithms cannot accurately extract disease features and are prone to misjudgment and missed judgment in practical applications. To solve the problem of identifying sugarcane leaf diseases under natural conditions and break through the limitations of traditional methods, a novel identification model, XEffDa was proposed in this research. [Methods] The XEffDa model proposed implemented a series of improvement measures based on the ensemble learning framework, aiming to significantly improve the accuracy of classifying and identifying sugarcane leaf diseases. Firstly, the images in the sugarcane leaf disease dataset under natural conditions were pre-processed. Real-time data augmentation techniques were used to expand the scale of the dataset. Meanwhile, HSV image segmentation and edge-processing techniques were adopted to effectively remove redundant backgrounds and interference factors in the images. Considering that sugarcane leaf disease images were fine-grained images, in order to fully extract the semantic information of the images, the transfer learning strategy was employed. The pre-trained models of EfficientNetB0, Xception, and DenseNet201 were loaded respectively, and with the help of the pre-trained weight parameters based on the ImageNet dataset, the top layers of the models were frozen. The performance of the validation set was monitored through the Bayesian optimization method, and the parameters of the top-layer structure were replaced, thus achieving a good balance between optimizing the number of model parameters and the overall performance. In the top-layer structure, the improved ElasticNet regularization and Dropout layer were integrated. These two mechanisms cooperated with each other to double-suppress overfitting and significantly enhance the generalization ability of the model. During the training process, the MSprop optimizer was selected and combined with the sparse categorical cross - entropy loss function to better adapt to the multi-classification problem of sugarcane disease identification. After each model completed training independently, an exponential weight-allocation strategy was used to organically integrate the prediction features of each model and accurately map them to the final disease categories. To comprehensively evaluate the model performance, the accuracy indicator was continuously monitored, and an early-stopping mechanism was introduced to avoid overfitting and further strengthen the generalization ability of the model. Through the implementation of this series of refined optimization and integration strategies, the XEffDa model for sugarcane leaf diseases was finally successfully constructed. [Results and Discussions] The results of the confusion matrix showed that the XEffDa model performed very evenly across various disease categories, and all indicators achieved excellent results. Especially in the identification of red rot disease, its F1-Score was as high as 99.09%. This result was not only higher than that of other single models (such as EfficientNetB0 and Xception) but also superior to the combination of EfficientNetB0 and other deep networks (such as DenseNet121 and DenseNet201). This indicated that the XEffDa model significantly improved the ability to extract and classify features of complex pathological images by integrating the advantages of different network architectures. The comparison experiments of different models showed that the recognition accuracy of the XEffDa model reached 97.62%. Compared with the single models of EfficientNetB0 and Xception, as well as the combined models of EfficientNetB0 and other deep networks, the recognition accuracy increased by 9.96, 6.04, 8.09, 4.19, and 1.78 percentage points, respectively. The fusion experiments further showed that the accuracy, precision, recall, and F1-Score of the network improved by ElasticNet regularization increased by 3.76, 3.76, 3.67, and 3.72 percentage points respectively compared with the backbone network. The results of the maximum-probability scatter plot showed that the proportion of the maximum prediction probability value not lower than 0.5 was as high as 99.4%. [Conclusions] The XEffDa model demonstrated stronger robustness and stability. In the identification task of small sugarcane leaf disease datasets, it showed good generalization ability. This model can provide a powerful reference for the accurate prevention and control of sugarcane crop leaf diseases in practical scenarios, and it has positive significance for promoting the intelligent and precise management of sugarcane production.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Improvement of HLM Modeling for Winter Wheat Yield Estimation Under Drought Conditions
    ZHAO Peiqin, LIU Changbin, ZHENG Jie, MENG Yang, MEI Xin, TAO Ting, ZHAO Qian, MEI Guangyuan, YANG Xiaodong
    Smart Agriculture    2025, 7 (2): 106-116.   DOI: 10.12133/j.smartag.SA202408009
    Abstract394)   HTML18)    PDF(pc) (1608KB)(346)       Save

    [Objective] Winter wheat yield is crucial for national food security and the standard of living of the population. Existing crop yield prediction models often show low accuracy under disaster-prone climatic conditions. This study proposed an improved hierarchical linear model (IHLM) based on a drought weather index reduction rate, aiming to enhance the accuracy of crop yield estimation under drought conditions. [Methods] HLM was constructed using the maximum enhanced vegetation index-2 (EVI2max), meteorological data (precipitation, radiation, and temperature from March to May), and observed winter wheat yield data from 160 agricultural survey stations in Shandong province (2018-2021). To validate the model's accuracy, 70% of the data from Shandong province was randomly selected for model construction, and the remaining data was used to validate the accuracy of the yield model. HLM considered the variation in meteorological factors as a key obstacle affecting crop growth and improved the model by calculating the relative meteorological factors. The calculation of relative meteorological factors helped reduce the impact of inter-annual differences in meteorological data. The accuracy of the HLM model was compared with that of the random forest (RF), Support Vector Regression (SVR), and Extreme Gradient Boosting (XGBoost) models. The HLM model provided more intuitive interpretation, especially suitable for processing hierarchical data, which helped capture the variability of winter wheat yield data under drought conditions. Therefore, a drought weather index reduction rate model from the agricultural insurance industry was introduced to further optimize the HLM model, resulting in the construction of the IHLM model. The IHLM model was designed to improve crop yield prediction accuracy under drought conditions. Since the precipitation differences between Henan and Shandong provinces were small, to test the transferability of the IHLM model, Henan province sample data was processed in the same way as in Shandong, and the IHLM model was applied to Henan province to evaluate its performance under different geographical conditions. [Results and Discussions] The accuracy of the HLM model, improved based on relative meteorological factors (rMF), was higher than that of RF, SVR, and XGBoost. The validation accuracy showed a Pearson correlation coefficient (r) of 0.76, a root mean squared error (RMSE) of 0.60 t/hm2, and a normalized RMSE (nRMSE) of 11.21%. In the drought conditions dataset, the model was further improved by incorporating the relationship between the winter wheat drought weather index and the reduction rate of winter wheat yield. After the improvement, the RMSE decreased by 0.48 t/hm2, and the nRMSE decreased by 28.64 percentage points, significantly enhancing the accuracy of the IHLM model under drought conditions. The IHLM model also demonstrated good applicability when transferred to Henan province. [Conclusions] The IHLM model developed in this study improved the accuracy and stability of crop yield predictions, especially under drought conditions. Compared to RF, SVR, and XGBoost models, the IHLM model was more suitable for predicting winter wheat yield. This research can be widely applied in the agricultural insurance field, playing a significant role in the design of agricultural insurance products, rate setting, and risk management. It enables more accurate predictions of winter wheat yield under drought conditions, with results that are closer to actual outcomes.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research on Agricultural Drought Prediction Based on GCN-BiGRU-STMHSA
    QUAN Jialu, CHEN Wenbai, WANG Yiqun, CHENG Jiajing, LIU Yilong
    Smart Agriculture    2025, 7 (1): 156-164.   DOI: 10.12133/j.smartag.SA202410027
    Abstract389)   HTML32)    PDF(pc) (1086KB)(669)       Save

    [Objective] Agricultural drought has a negative impact on the development of agricultural production and even poses a threat to food security. To reduce disaster losses and ensure stable crop yields, accurately predicting and classifying agricultural drought severity based on the standardized soil moisture index (SSMI) is of significant importance. [Methods] An agricultural drought prediction model, GCN-BiGRU-STMHSA was proposed, which integrated a graph convolutional network (GCN), a bidirectional gated recurrent unit (BiGRU), and a multi-head self-attention (MHSA) mechanism, based on remote sensing data. In terms of model design, the proposed method first employed GCN to fully capture the spatial correlations among different meteorological stations. By utilizing GCN, a spatial graph structure based on meteorological stations was constructed, enabling the extraction and modeling of spatial dependencies between stations. Additionally, a spatial multi-head self-attention mechanism (S-MHSA) was introduced to further enhance the model's ability to capture spatial features. For temporal modeling, BiGRU was utilized as the time-series feature extraction module. BiGRU considers both forward and backward dependencies in time-series data, enabling a more comprehensive understanding of the temporal dynamics of agricultural drought. Meanwhile, a temporal multi-head self-attention mechanism (T-MHSA) was incorporated to enhance the model's capability to learn long-term temporal dependencies and improve prediction stability across different time scales. Finally, the model employed a fully connected layer to perform regression prediction of the SSMI. Based on the classification criteria for agricultural drought severity levels, the predicted SSMI values were mapped to the corresponding drought severity categories, achieving precise agricultural drought classification. To validate the effectiveness of the proposed model, the global land data assimilation system (GLDAS_2.1) dataset and conducted modeling and experiments was utilized on five representative meteorological stations in the North China Plain (Xinyang, Gushi, Fuyang, Huoqiu, and Dingyuan). Additionally, the proposed model was compared with multiple deep learning models, including GRU, LSTM, and Transformer, to comprehensively evaluate its performance in agricultural drought prediction tasks. The experimental design covered different forecasting horizons to analyze the model's generalization capability in both short-term and long-term predictions, thereby providing a more reliable early warning system for agricultural drought. [Results and Discussions] Experimental results demonstrated that the proposed GCN-BiGRU-STMHSA model outperforms baseline models in both SSMI prediction and agricultural drought classification tasks. Specifically, across the five study stations, the model achieved significantly lower mean absolute error (MAE) and root mean squared error (RMSE), while attaining higher coefficient of determination ( R²), classification accuracy (ACC), and F1-Score ( F1). Notably, at the Gushi station, the model exhibited the best performance in predicting SSMI 10 days ahead, achieving an MAE of 0.053, a RMSE of 0.071, a R² of 0.880, an ACC of 0.925, and a F1 of 0.924. Additionally, the model's generalization capability was investigated under different forecasting horizons (7, 14, 21, and 28 days). Results indicated that the model achieved the highest accuracy in short-term predictions (7 days). Although errors increase slightly as the prediction horizon extends, the model maintained high classification accuracy even for long-term predictions (up to 28 days). This highlighted the model's robustness and effectiveness in agricultural drought prediction over varying time scales. [Conclusions] The proposed model achieves superior accuracy and generalization capability in agricultural drought prediction and classification. By effectively integrating spatial graph modeling, temporal sequence feature extraction, and self-attention mechanisms, the model outperforms conventional deep learning approaches in both short-term and long-term forecasting tasks. Its strong performance provides accurate drought early warnings, assisting agricultural management authorities in formulating efficient water resource management strategies and optimizing irrigation plans. This contributes to safeguarding agricultural production and mitigating the potential adverse effects of agricultural drought.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Lightweight Cattle Facial Recognition Method Based on Improved YOLOv11
    HAN Yu, QI Kangkang, ZHENG Jiye, LI Jinai, JIANG Fugui, ZHANG Xianglun, YOU Wei, ZHANG Xia
    Smart Agriculture    2025, 7 (3): 173-184.   DOI: 10.12133/j.smartag.SA202502010
    Abstract387)   HTML37)    PDF(pc) (2163KB)(107)       Save

    [Objective] Beef cattle breeding constitutes a pivotal component of modern animal husbandry, where accurate individual identification serves as the cornerstone for advancing automated technologies, including intelligent weight measurement, body condition scoring, body conformation assessment, and behavior monitoring. However, practical breeding environments are riddled with challenges: soiled conditions, cluttered backgrounds, and constant animal movement collectively result in high variability in cattle facial data features. Furthermore, inconsistent lighting and diverse shooting angles often obscure key features, increasing the risk of misjudgment during detection. To tackle these issues, this study introduces an improved model, YOLO-PCW, which enhances detection performance while maintaining a lightweight structure, effectively addressing the complexities of precise cattle face recognition in challenging breeding settings. [Methods] The research leveraged the Cow Fusion dataset (CFD), a comprehensive collection of real-world cattle face images captured under variable lighting conditions, from multiple angles, and against complex backgrounds, for model training and validation. Concurrently, a Custom Cow Monitor Dataset (CMD) was created from video footage obtained through the a breeding farm's monitoring system, providing a robust basis for evaluating the model's generalization capabilities. The YOLOv11 architecture served as the foundational framework for implementing the following performance improvements. The partial convolution (PConv) was seamlessly integrated into the C3K2 module within the YOLOv11 head network. Utilizing the sparse convolutional properties of PConv on the feature maps, the convolutional structure was meticulously optimized, reducing computational redundancy and memory access while preserving the model's accuracy, rendering it highly suitable for real-time applications. Additionally, the convolutional block attention module (CBAM) was incorporated to enhance feature map processing through adaptive channel-wise and spatial attentions. This refinement enabled precise extraction of target regions by mitigating background interference, allowing the model to focus on critical anatomical features such as the eyes, mouth, and nose. Furthermore, the weighted intersection over union (WIoU) loss function was adopted to replace the CIoU, optimizing the weighted strategy for bounding box regression errors. This innovation reduced the adverse effects of large or outlier gradients in extreme samples, enabling the model to prioritize average-quality samples for refinement. The resulting improvement in key region localization accuracy bolstered the model's generalization capability and overall performance, establishing a state-of-the-art cattle face recognition framework. [Results and Discussion] The YOLO-PCW model achieved a remarkable accuracy rate (P) of 96.4%, recall rate (R) of 96.7%, and mean average precision (mAP) of 98.7%. With 2.3 million parameters and a computational load of 5.6 GFLOPs, it not only improved accuracy, recall, and mAP by 3.6, 5, and 4.4 percentage point respectively, but also achieved a significant reduction in floating-point computational load and parameter size, down to 88.9% and 88.5% of the original model's, respectively. Ablation studies revealed that the CBAM module enhanced precision from 92.8% to 95.2%. The WIoU loss function optimized target positioning accuracy, achieving a precision of 93.8%. The PConv module substantially reduced computational load from 6.3 GFLOPs to 5.5 GFLOPs, thereby lightening the model's computational burden significantly. The synergistic collaboration of these components provided robust support for enhancing the performance of the cattle face recognition model. Comparative experiments demonstrated that under identical conditions, the YOLO-PCW model outperformed algorithms such as Faster-RCNN, SSD, YOLOv5, YOLOv7-tiny, and YOLOv8 under identical conditions, exhibiting the most outstanding performance. It effectively balanced recognition accuracy with computational efficiency, achieving optimal utilization of computational resources. [Conclusions] The improved YOLO-PCW model, featuring a lightweight architecture and optimized attention mechanism, could successfully improve detection accuracy while simplifies deployment. It can achieve precise cattle face recognition in real-world breeding environments, providing an efficient and practical solution for individual identification in applications such as animal welfare breeding, intelligent ranch management, smart ranch construction, and animal health monitoring.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Bi-Intentional Modeling and Knowledge Graph Diffusion for Rice Variety Selection and Breeding Recommendation
    QIAO Lei, CHEN Lei, YUAN Yuan
    Smart Agriculture    2025, 7 (2): 73-80.   DOI: 10.12133/j.smartag.SA202412025
    Abstract370)   HTML18)    PDF(pc) (1070KB)(657)       Save

    [Objective] Selection of rice varieties requires consideration of several factors, such as yield, fertility, disease resistance and resistance to downfall. In order to meet the user's rice variety selection needs, help users quickly access to the rice varieties they need, improve efficiency, and further promote the informatization and intelligence of rice breeding work, the bi-intentional modeling and knowledge graph diffusion model, an advanced method was proposed. [Methods] The research work was mainly carried out at two levels: Data and methodology. At the data level, considering the current lack of relevant data support for rice variety selection and breeding recommendation, a certain amount of recommendation dataset was constructed. The rice variety selection recommendation dataset consisted of two parts: Interaction data and knowledge graph. For the interaction data, the rice varieties that had been planted in the region were collected on a region-by-region basis, and then a batch of users was simulated and generated from the region. The corresponding rice varieties were assigned to the generated users according to the random sampling method to construct the user-item interaction data. For the knowledge graph, detailed text descriptions of rice varieties were first collected, and then information was extracted from them to construct data in ternary format from multiple varietal characteristics, such as selection unit, varietal category, disease resistance, and cold tolerance. At the methodological level, a model of bi-intentional modeling and knowledge graph diffusion (BMKGD) was proposed. The intent factor in the interaction behavior and the denoising process of the knowledge graph were both taken into account by the BMKGD model. Intentions were usually considered from two perspectives: individual independence and conformity. A dual intent space was chosen to be built by the model to represent both perspectives. For the problem of noisy data in the knowledge graph, denoising was carried out by combining the idea of the diffusion model. Random noise was introduced to destroy the original structure when the knowledge graph was initialized, and the original structure was restored through iterative learning. The denoising was completed in this process. After that, cross-view contrastive learning was carried out in both views. [Results and Discussions] The method proposed achieved optimal performance in the rice variety selection dataset, with recall and normalized discounted cumulative gain (NDCG) values improved by 2.9% and 3.7% compared to the suboptimal model. The performance improvement validated the effectiveness of the method to some extent, indicating that the BMKGD model was more suitable for rice variety recommendation. The Recall value of the BMKGD model on the rice variety selection dataset was 0.327 6, meeting the basic requirements of the recommendation system. The analysis revealed that the collaborative signals in the interaction data played a major role, while the quality of the constructed knowledge graph still had some room for improvement. The module variants with key components removed all exhibited a decrease in performance compared to the original model, which validated the effectiveness of the modules. The performance degradation of the model variants with each component removed varied, indicating that different components played different roles. The performance drop of the model variant with the cross-view contrastive learning module removed was small, indicating that there was some room for improvement in the module to fully utilize the collaborative relationship between the two views. Conclusions The BMKGD model proposed in this paper achieves good performance on the rice variety selection dataset and accomplishes the recommendation task well. It shows that the model can be used to support the rice variety selection and breeding work and help users to select suitable rice varieties.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Agricultural Big Data Governance: Key Technologies, Applications Analysis and Future Directions
    GUO Wei, WU Huarui, ZHU Huaji, WANG Feifei
    Smart Agriculture    2025, 7 (3): 17-34.   DOI: 10.12133/j.smartag.SA202503020
    Abstract365)   HTML71)    PDF(pc) (4090KB)(609)       Save

    [Significance] To provide a reference for advancing high-quality agricultural production driven by data, this paper focuses on the issues of inconsistent acquisition standards, incomplete data collection, and ambiguous governance mechanisms in China's agricultural production data, examines existing governance models for agricultural production big data, and clarifies the technical pathways for realizing the value of data elements through the integrated and innovative application of key big data governance technologies and tools in practical scenarios. [Progress] From the perspective of agricultural production big data governance, this paper explores 17 types of big data governance technologies and tools across six core processes: Data acquisition and processing, data storage and exchange, data management, data analysis, large models, and data security guarantee. It conducts in-depth research on the application methods of big data governance technologies in agricultural production, revealing that: Remote sensing, unmanned aerial vehicle(UAV), Internet of Things (IoT), and terminal data acquisition and processing systems are already reatively mature; data storage and exchange system are developing rapidly, data management technologies remain in the initial stage; data analysis technologies have been widely applied; large model technology systems have taken initially shape; and data security assurance systems are gradually being into parctice. The above technologies are effectively applied in scenarios through tools and middleware such as data matching, computing power matching, network adaptation, model matching, scenario matching, and business configuration. This paper also analyzes the data governance throughout the entire agricultural production chain, including pre-production, in-production, and post-production, stages, as well as service cases involving different types of agricultural parks, research institutes and universities, production entities, and farmers. It demonstrates that sound data governance can provide sufficient planning and input analysis prior to production, helping planting entities in making rational plans. In production, it can provide data-driven guidance for key scenarios such as agricultural machinery operations and agricultural technical services, thereby fully supporting decision-making in the production process; and based on massive data, it can achieve reliable results in yield assessment and production benefit evaluation. Additionally, the paper introduces governance experience from national-level industrial parks, provincial-level agricultural science and technology parks, and some single-product entities, and investigates domestic and international technologies, practices, and tools related to agricultural production big data governance, indicating that there is a need to break through the business chains and service model of agricultural production across regions, themes, and scenarios. [Conclusions and Prospects] This paper presents insights into the future development directions of agricultural production big data governance, encompassing the promotion of standard formulation and implementation for agricultural production big data governance, the establishment of a universal resource pool for such governance, the expansion of diversified application scenarios, adaptation to the new paradigm of large-model- and massive-data-driven agricultural production big data governance, and the enhancement of security and privacy protection for agricultural production big data.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Domain Generalization Method of Strawberry Disease Recognition Based on Instance Whitening and Restitution
    HU Xiaobo, XU Taosheng, WANG Chengjun, ZHU Hongbo, GAN Lei
    Smart Agriculture    2025, 7 (1): 124-135.   DOI: 10.12133/j.smartag.SA202411016
    Abstract350)   HTML25)    PDF(pc) (1496KB)(276)       Save

    [Objective] Strawberry disease recognition models based on deep neural networks generally assume that the training (source domain) and the test (target domain) datasets are identically and independently distributed. However, in practical applications, due to the influence of illumination, background and strawberry variety, the target domain often exhibits significant domain shift from the source domain. The domain shift result in accuracy decline of the models in target domain. To address this problem, a domain generalization method based on instant whitening and restitution (IWR) was proposed to improve the generalization performance of strawberry disease identification models in this research. [Methods] Samples from different source often exhibit great domain shift due to variations in strawberry varieties, regional climate, and photography methods. Therefore, a dataset was constructed for domain generalization research on strawberry disease using two distinct approaches. The first dataset was acquired using a Nikon D810 camera at multiple strawberry farms in Changfeng county, Anhui province, with a fixed sampling schedule and fixed camera distance. In contrast, the second dataset was an open-source collection, primarily comprising images captured using smartphones in multiple strawberry greenhouses in Korea, with varied and random shooting distances and angles. The IWR module mitigated style variations (e.g., illumination, color) through instance whitening, where features were normalized to reduce domain discrepancies between the datasets. However, such operation was task-ignorant and inevitable removed some task-relevant information, which may be harmful to classification performance of the models. To remedy this, the removed task-relevant features were attempted to recover. Specifically, two modules were designed to extract task-relevant and task-irrelevant feature from the filtered style features, respectively. A dual restitution loss was utilized to constraint the modules' feature correlation between the task and a mutual loss was used to ensure the independence of the features. In addition, a separation optimization strategy was adopted to further enhance the feature separation effect of the two modules. [Results and Discussions] The F1-Score was adopted as evaluation metrics. A series of ablations studies and comparative experiments were conducted to demonstrate the effectiveness of the proposed IWR. The ablation experiments proved that the IWR could effectively eliminate the style variations between different datasets and separate task-relevant feature from the filtered style features, which could simultaneously enhance model generalization and discrimination capabilities. The recognition accuracy increased when IWR pluged to AlexNet, GoogLeNet, ResNet-18, ResNet-50, MobileNetV2 and MobileNetV3. It demonstrated that the proposed IWR was an effective way to improve the generalization of the models. Compared with other domain generalization methods such as IBNNet, SW and SNR, the generalization performance of the proposed algorithm on test datasets could be improved by 2.63%, 2.35% and 1.14%, respectively. To better understand how IWR works, the intermediate feature maps of ResNet-50 without and with IWR were compared. The visualization result showed that the model with IWR was more robust when the image style changed. These results indicated that the proposed IWR achieves high classification accuracy and boosts the generalization performance of the models. [Conclusions] An instance whitening and restitution module was presented, which aimed to learn generalizable and discriminative feature representations for effective domain generalization. The IWR was a plug-and-play module, it could be inserted into existing convolutional networks for strawberry disease recognition. Style normalization and restitution (SNR) reduced the style information through instance whitening operation and then restitutes the task-relevant discriminative features caused by instance whitening. The introduced dual restitution loss and mutual loss further facilitate the separation of task-relevant and task-irrelevant feature. The schemes powered by IWR achieves the state-of-the-art performance on strawberry disease identification.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Spatiotemporal Pattern and Multi-Scenario Simulation of Land Use Conflicts: A Case Study of Shandong Section of the Yellow River Basin
    DONG Guanglong, YIN Haiyang, YAO Rongyan, YUAN Chenzhao, QU Chengchuang, TIAN Yuan, JIA Min
    Smart Agriculture    2025, 7 (2): 183-195.   DOI: 10.12133/j.smartag.SA202409007
    Abstract341)   HTML12)    PDF(pc) (3337KB)(560)       Save

    [Objective] The frequent occurrence of land use conflicts, such as the occupation of arable land by urban construction land expansion, non-grain use of arable land, and the shrinking of ecological space, poses multiple pressures on the Shandong section of the Yellow River in terms of economic development, arable land protection, and ecological conservation. Accurately identifying and predicting future trends of land-use conflicts in the Shandong section of the Yellow River under various scenarios will provide a reference for the governance of land use conflicts, rational land resource utilization, and optimization of the national land spatial pattern in this region. [Methods] The data used mainly includs land use data, elevation data, basic geographic information data, meteorological data, protected area data, and socio-economic data. Drawing from the concept of ecological risk assessment, an "External Pressure + Vulnerability-Stability" model was constructed. Indicators such as area-weighted average patch fractal dimension, landscape value of land use types, and patch density were used to quantify and characterize land use conflicts in the Shandong section of the Yellow River from 2000 to 2020. Subsequently, the CA-Markov model was employed to establish cellular automata transition rules, with a 10-year simulation period using a default 5×5 cellular filter matrix, projecting 2030 land use conflict patterns under natural development, cultivated land protection, and ecological conservation scenarios. [Results and Discussions] From 2000 to 2020, significant changes in land use were observed in the Shandong section of the Yellow River, mainly characterized by rapid expansion of urban construction land and a reduction in grassland and arable land. Urban construction land increased by 4 346 km2, with its proportion rising from 13.50% in 2000 to 18.67% in 2020. During the study period, the level of land use conflict showed a mitigating trend, with the average land use conflict index decreasing from 0.567 in 2000 to 0.522 in 2020. Medium conflict has been the dominant type of land use conflict in the Shandong section of the Yellow River, followed by low conflict, while high conflict accounted for the smallest proportion. This indicates that land use conflicts in the region were generally controllable. The spatial pattern of land use conflicts in the Shandong section of the Yellow River remained relatively stable. Low conflicts were mainly distributed in areas with high concentration of arable land and water bodies, as well as in urban built-up areas. Medium conflicts were most widespread, especially in the transitional zones between arable land and rural settlements, and between arable land and forest land. The proportion of high conflict decreased from 19.34% in 2000 to 8.61% in 2020, mainly clustering in the transitional zones between urban construction land and other land types, the land type interlacing belt in the Central Shandong Hills, and along the Yellow River. The multi-scenario land use simulation results for 2030 showed significant differences in land use changes under different scenarios. Under the natural development scenario, the level of land use conflict was expected to deteriorate, with the most severe conflict situation. While both arable land protection and ecological conservation scenarios demonstrate partial conflict mitigation, the expansion of arable land occurs at the expense of ecological spaces, potentially compromising regional ecological security. In contrast, under the ecological conservation scenario, by prioritizing ecological protection and highlighting the protection of the ecological environment, the expansion of urban construction land and the reclamation of arable land, which cause ecological damage, were effectively curbed. Notably, this scenario exhibited the lowest proportion of high conflicts and demonstrated superior conflict mitigation effectiveness. [Conclusions] Land use conflicts in the Shandong section of the Yellow River have been somewhat mitigated, with the main form of conflict being the rapid expansion of urban construction land encroaching on arable land and ecological land. The ecological conservation scenario effectively balances the relationship between arable land protection, ecological security, and urbanization development, and is an optimal strategy for alleviating land use conflicts in the Shandong section of the Yellow River.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Image Segmentation Method of Chinese Yam Leaves in Complex Background Based on Improved ENet
    LU Bibo, LIANG Di, YANG Jie, SONG Aiqing, HUANGFU Shangwei
    Smart Agriculture    2024, 6 (6): 109-120.   DOI: 10.12133/j.smartag.SA202407007
    Abstract336)   HTML26)    PDF(pc) (2024KB)(1298)       Save

    [Objective] Crop leaf area is an important indicator reflecting light absorption efficiency and growth conditions. This paper established a diverse Chinese yam image dataset and proposesd a deep learning-based method for Chinese yam leaf image segmentation. This method can be used for real-time measurement of Chinese yam leaf area, addressing the inefficiency of traditional measurement techniques. This will provide more reliable data support for genetic breeding, growth and development research of Chinese yam, and promote the development and progress of the Chinese yam industry. [Methods] A lightweight segmentation network based on improved ENet was proposed. Firstly, based on ENet, the third stage was pruned to reduce redundant calculations in the model. This improved the computational efficiency and running speed, and provided a good basis for real-time applications. Secondly, PConv was used instead of the conventional convolution in the downsampling bottleneck structure and conventional bottleneck structure, the improved bottleneck structure was named P-Bottleneck. PConv applied conventional convolution to only a portion of the input channels and left the rest of the channels unchanged, which reduced memory accesses and redundant computations for more efficient spatial feature extraction. PConv was used to reduce the amount of model computation while increase the number of floating-point operations per second on the hardware device, resulting in lower latency. Additionally, the transposed convolution in the upsampling module was improved to bilinear interpolation to enhance model accuracy and reduce the number of parameters. Bilinear interpolation could process images smoother, making the processed images more realistic and clear. Finally, coordinate attention (CA) module was added to the encoder to introduce the attention mechanism, and the model was named CBPA-ENet. The CA mechanism not only focused on the channel information, but also keenly captured the orientation and position-sensitive information. The position information was embedded into the channel attention to globally encode the spatial information, capturing the channel information along one spatial direction while retaining the position information along the other spatial direction. The network could effectively enhance the attention to important regions in the image, and thus improve the quality and interpretability of segmentation results. [Results and Discussions] Trimming the third part resulted in a 28% decrease in FLOPs, a 41% decrease in parameters, and a 9 f/s increase in FPS. Improving the upsampling method to bilinear interpolation not only reduces the floating-point operation and parameters, but also slightly improves the segmentation accuracy of the model, increasing FPS by 4 f/s. Using P-Bottleneck instead of downsampling bottleneck structure and conventional bottleneck structure can reduce mIoU by only 0.04%, reduce FLOPs by 22%, reduce parameters by 16%, and increase FPS by 8 f/s. Adding CA mechanism to the encoder could only increase a small amount of FLOPs and parameters, improving the accuracy of the segmentation network. To verify the effectiveness of the improved segmentation algorithm, classic semantic segmentation networks of UNet, DeepLabV3+, PSPNet, and real-time semantic segmentation network LinkNet, DABNet were selected to train and validate. These six algorithms got quite high segmentation accuracy, among which UNet had the best mIoU and the mPA, but the model size was too large. The improved algorithm only accounts for 1% of the FLOPs and 0.41% of the parameters of UNet, and the mIoU and mPA were basically the same. Other classic semantic segmentation algorithms, such as DeepLabV3+, had similar accuracy to improved algorithms, but their large model size and slow inference speed were not conducive to embedded development. Although the real-time semantic segmentation algorithm LinkNet had a slightly higher mIoU, its FLOPs and parameters count were still far greater than the improved algorithm. Although the PSPNet model was relatively small, it was also much higher than the improved algorithm, and the mIoU and mPA were lower than the algorithm. The experimental results showed that the improved model achieved a mIoU of 98.61%. Compared with the original model, the number of parameters and FLOPs significantly decreased. Among them, the number of model parameters decreased by 51%, the FLOPs decreased by 49%, and the network operation speed increased by 38%. [Conclusions] The improved algorithm can accurately and quickly segment Chinese yam leaves, providing not only a more accurate means for determining Chinese yam phenotype data, but also a new method and approach for embedded research of Chinese yam. Using the model, the morphological feature data of Chinese yam leaves can be obtained more efficiently, providing a reliable foundation for further research and analysis.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Agriculture    2024, 6 (6): 0-0.  
    Abstract331)            Save
    Related Articles | Metrics | Comments0
    Artificial Intelligence for Agricultural Science (AI4AS): Key Elements, Challenges and Pathways
    ZHAO Ruixue, YANG Xiao, ZHANG Dandan, LI Jiao, HUANG Yongwen, XIAN Guojian, KOU Yuantao, SUN Tan
    Smart Agriculture    2025, 7 (3): 35-47.   DOI: 10.12133/j.smartag.SA202502019
    Abstract303)   HTML28)    PDF(pc) (1295KB)(61)       Save

    [Significance] Artificial intelligence for science (AI4S), as an emerging paradigm that deeply integrates artificial intelligence(AI) with scientific research, has triggered profound transformations in research methodologies. By accelerating scientific discovery through AI technologies, it is driving a shift in scientific research from traditional approaches reliant on experience and intuition to methodologies co-driven by data and AI. This transition has spurred innovative breakthroughs across numerous scientific domains and presents new opportunities for the transformation of agricultural research. With its powerful capabilities in data processing, intelligent analysis, and pattern recognition, AI can transcend the cognitive limitations of researchers in the field and is gradually emerging as an indispensable tool in modern agricultural scientific research, injecting new impetus into the intelligent, efficient, and collaborative development of agricultural scientific research. [Progress] This paper systematically reviews the current advancements in AI4S and its implications for agricultural research. It reveals that AI4S has triggered a global race among countries around the world vying for the commanding heights of a new round of scientific and technological strategies. Developed nations in Europe and America, for instance, have laid out the frontier areas in AI4S and rolled out relevant policies. Meanwhile, some top universities and research institutions are accelerating related research, and tech giants are actively cultivating related industries to advance the application and deployment of AI technologies in scientific research. In recent years, AI4S has achieved remarkable development, showing great potential across multiple disciplines and finding widespread application in data mining, model construction, and result prediction. In the field of agricultural scientific research, AI4S has played an important role in accelerating multi-disciplinary integration, promoting the improvement of the scientific research efficiency, facilitating the breakthrough of complex problems, driving the transformation of the scientific research paradigm, and upgrading scientific research infrastructure. The continuous progress of information technology and synthetic biology has made the interdisciplinary integration of agriculture and multiple disciplines increasingly closer. The deep integration of AI and agricultural scientific research not only improves the application level of AI in the agricultural field but also drives the transformation of traditional agricultural scientific research models towards intelligence, data-driven, and collaborative directions, providing new possibilities for agricultural scientific and technological innovation. The new agricultural digital infrastructure is characterized by intelligent data collection, edge computing power deployment, high-throughput network transmission, and distributed storage architecture, aiming to break through the bottlenecks of traditional agricultural scientific research facilities in terms of real-time performance, collaboration, and scalability. Taking emerging disciplines such as Agrinformatics and climate-focused Agriculture-Forestry-AI (AgFoAI) as examples, they focus on using AI technology to analyze agricultural data, construct crop growth models, and climate change models, etc., to promote the development and innovation of agricultural scientific research. [Conclusions and Prospects] With its robust capabilities in data processing, intelligent analysis, and pattern recognition, AI is increasingly becoming an indispensable tool in modern agricultural scientific research. To address emerging demands, core domains, and research processes in agricultural research, the concept of agricultural intelligent research is proposed, characterized by human-machine collaboration and interdisciplinary integration. This paradigm employs advanced data analytics, pattern recognition, and predictive modeling to perform in-depth mining and precise interpretation of multidimensional, full-lifecycle, large-scale agricultural datasets. By comprehensively unraveling the intrinsic complexities and latent patterns of research subjects, it autonomously generates novel, scientifically grounded, and high-value research insights, thereby driving agricultural research toward greater intelligence, precision, and efficiency. The framework's core components encompass big science infrastructure (supporting large-scale collaborative research), big data resources (integrating heterogeneous agricultural datasets), advanced AI model algorithms (enabling complex simulations and predictions), and collaborative platforms (facilitating cross-disciplinary and cross-institutional synergy). Finally, in response to challenges related to data resources, model capabilities, research ecosystems, and talent development, actionable pathways and concrete recommendations are outlined from the perspectives of top-level strategic planning, critical technical ecosystems, collaborative innovation ecosystems, disciplinary system construction, and interdisciplinary talent cultivation, aiming to establish a new AI4S-oriented agricultural research framework.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Remote Sensing for Rice Growth Stages Monitoring: Research Progress, Bottleneck Problems and Technical Optimization Paths
    LI Ruijie, WANG Aidong, WU Huaxing, LI Ziqiu, FENG Xiangqian, HONG Weiyuan, TANG Xuejun, QIN Jinhua, WANG Danying, CHU Guang, ZHANG Yunbo, CHEN Song
    Smart Agriculture    2025, 7 (3): 89-107.   DOI: 10.12133/j.smartag.SA202412019
    Abstract300)   HTML47)    PDF(pc) (2568KB)(280)       Save

    [Significance] ​The efficient and precise identification of rice growth stages through remote sensing technology holds critical significance for varietal breeding optimization and production management enhancement. Remote sensing, characterized by high spatial-temporal resolution and automated monitoring capabilities, provides transformative solutions for large-scale dynamic phenology monitoring, offering essential technical support to address climate change impacts and food security challenges in complex agroecosystems where precise monitoring of growth stage transitions enables yield prediction and stress-resilient cultivation management.​ [Progress] In recent years, the technical system for monitoring rice growth stages has achieved systematic breakthroughs in the perception layer, decision-making layer, and execution layer, forming a technological ecosystem covering the entire chain of "data acquisition-feature analysis-intelligent decision-making-precise operation". At the perception layer, a "space-air-ground" three-dimensional monitoring network has been constructed: High-altitude satellites (Sentinel-2, Landsat) realize regional-scale phenological dynamic tracking through wide-spectrum multi-temporal observations; low-altitude unmanned aerial vehicle (UAV) equipped with hyperspectral and light detection and ranging (LiDAR) sensors analyze the heterogeneity of canopy three-dimensional structure; near-ground sensor networks real-timely capture leaf-scale photosynthetic efficiency and nitrogen metabolism parameters. Radiometric calibration and temporal interpolation algorithms eliminate the spatio-temporal heterogeneity of multi-source data, forming continuous and stable monitoring capabilities. Innovations in technical methods show three integration trends: Firstly, multimodal data collaboration mechanisms break through the physical characteristic barriers between optical and radar data; secondly, deep integration of mechanistic models and data-driven approaches embeds the scattering by arbitrarily inclined leaves by arbitrary inclined leaves (PROSPECT + SAIL, PROSAIL) radiative transfer model into the long short-term memory (LSTM) network architecture; thirdly, cross-scale feature analysis technology breaks through by constructing organ-population association models based on dynamic attention mechanisms, realizing multi-granularity mapping between panicle texture features and canopy leaf area index (LAI) fluctuations. The current technical system has completed three-dimensional leaps: From discrete manual observations to full-cycle continuous perception, with monitoring frequency upgraded from weekly to hourly; from empirical threshold-based judgment to mechanism-data hybrid-driven, the cross-regional generalization ability of the model can be significantly improved; from independent link operations to full-chain collaboration of "perception-decision-execution", constructing a digital management closed-loop covering rice sowing to harvest, providing core technical support for smart farm construction. [Conclusions and Prospects] Current technologies face three-tiered challenges in data heterogeneity, feature limitations and algorithmic constraints. Future research should focus on three aspects: 1) Multi-source data assimilation systems to reconcile spatiotemporal heterogeneity through UAV-assisted satellite calibration and GAN-based cloud-contaminated data reconstruction; 2) Cross-scale physiological-spectral models integrating 3D canopy architecture with adaptive soil-adjusted indices to overcome spectral saturation; 3) Mechanism-data hybrid paradigms embedding thermal-time models into LSTM networks for environmental adaptation, developing lightweight CNNs with multi-scale attention for occlusion-resistant panicle detection, and implementing transfer learning for cross-regional model generalization. The convergence of multi-source remote sensing, intelligent algorithms, and physiological mechanisms will establish a full-cycle dynamic monitoring system based on agricultural big data.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Electrochemical Sensors for Plant Active Small Molecule Detection: A review
    ZHANG Le, LI Aixue, CHEN Liping
    Smart Agriculture    2025, 7 (3): 69-88.   DOI: 10.12133/j.smartag.SA202502023
    Abstract276)   HTML30)    PDF(pc) (6524KB)(976)       Save

    [Significance] Plant active small molecules play an indispensable role in plants. They form the basis of the core physiological mechanisms that regulate the plant growth and development and enhance resilience to environmental stress. Achieving highly precise quantitative analysis of these active small molecules is therefore vital for promoting precise management practice in the agricultural and accelerating the development of smart agriculture. Currently, various technologies exist for the detection and analysis of these small molecules in plants. Among them, electrochemical sensing platforms have attracted extensive attention due to their significant advantages, including high sensitivity, excellent selectivity, and low cost. These advantages enable them to effectively detect trace levels of various active small molecules in plant samples. They also have the potential for real-time and in-situ detection. [Progress] Based on a comprehensive review of relevant academic literature, this article systematically summarizes the current research progress and status of electrochemical sensors applied in detecting plant active small molecule. Based on this, the article further analyzes the core sensing mechanisms of different electrochemical sensors types, signal amplification technologies for enhancing detection performance, and their huge potential in practical applications. Furthermore, this paper explores a notable development direction in this field: Sensor technology is evolving from the traditional in vitro detection mode to more challenging in vivo detection and in-situ real-time monitoring methods. Meanwhile, the article particularly emphasizes and elaborates in detail the indispensable and significant role of nanomaterials in key links such as constructing high-performance sensing interfaces and significantly enhancing detection sensitivity and selectivity. Finally, it prospectively discusses the innovative integration of electrochemical sensors with cutting-edge flexible electronic technology and powerful artificial intelligence (AI)-based data analysis, along with their potential for broad application. [Conclusions and Prospects] This article comprehensively identifies and summarizes the core technical challenges that electrochemical sensors currently face in detecting plant active small molecule. In terms of environmental detection, due to the influence of the complex matrix within plants, the response signal of the sensor is prone to drift, and its stability and sensitivity show a decline. Regarding electrolytes, the external application of liquid electrolytes dilutes the target molecules concentration in plant samples, lowering the detection accuracy. Furthermore, the transition from principle development to mature productization and industrialization of electrochemical sensors is relatively lengthy, and there are few types of sensors available for the detection of plant physiological indicators: Limiting their application in actual agricultural production. On this basis, the article prospectively analyzes the key directions of future research. First, continuously improving sensor performance indicators such as sensitivity, selectivity and reliability. Second, exploring and optimizing electrolyte material systems with stronger adaptability to significantly improve detection accuracy and long-term stability. Third, promoting deeper integration and innovation of sensor technology with advanced micro-nano electronic technology and powerful AI algorithms. The core objective of this review is to provide a theoretical guidance framework for in-depth research and systematic performance optimization of electrochemical sensing technology for plant active small molecules, as well as practical guidance for the actual application of related sensors in complex plant substrate environments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Pruning Point Recognition and Localization for Spindle-Shaped Apple Trees During Dormant Season Using an Improved U-Net Model
    LIU Long, WANG Ning, WANG Jiacheng, CAO Yuheng, ZHANG Kai, KANG Feng, WANG Yaxiong
    Smart Agriculture    2025, 7 (3): 120-130.   DOI: 10.12133/j.smartag.SA202501022
    Abstract274)   HTML31)    PDF(pc) (2725KB)(127)       Save

    [Objective] To address the current issues of intelligent pruning robots, such as insufficient recognition accuracy of fruit tree branches and inaccurate localization of pruning points in complex field environments, a deep learning method based on the fusion of images and point clouds was proposed in this research. The method enables non-contact segmentation of dormant high-spindle apple tree branches and measurement of phenotypic parameter, which also achieving automatic identification and precise localization of pruning points. [Methods] Localized RGB-D data were collected from apple trees using a Realsense D435i camera, a device capable of accurate depth measurements within the range of 0.3~3.0 m. Data acquisition took place between early and mid-January 2024, from 9:00 AM to 4:00 PM daily. To maintain uniformity, the camera was mounted on a stand at a distance of 0.4~0.5 m from the main stems of the apple trees. Following data collection, trunks and branches were manually annotated using Labelme software. The OpenCV library was also employed for data augmentation, which helped prevent overfitting during model training. To improve segmentation accuracy of tree trunks and branches in RGB images, an enhanced U-Net model was introduced. This model utilized VGG16 (Visual Geometry Group 16) as its backbone feature extraction network and incorporated the convolutional block attention module (CBAM) at the up-sampling stage. Based on the segmentation results, a multimodal data processing pipeline was established. First, segmented branch mask maps were derived from skeleton lines extracted using OpenCV's algorithm. The first-level branch connection points were identified based on their positions relative to the trunk. Subsequently, potential pruning points were then searched within local neighborhoods through coordinate translation. An edge detection algorithm was applied to locate the nearest edge pixels relative to these potential pruning points. By extending the diameter line of branch pixel points in the images and integrating with depth information, the actual diameter of the branches could be estimated. Additionally, branch spacing was calculated using vertical coordinates differences of potential pruning points in the pixel coordinate system, alongside depth information. Meanwhile, trunk point cloud data were acquired by merging the trunk mask maps with the depth maps. Preprocessing of the point cloud enabled the estimation of the average trunk diameter in the local view through cylindrical fitting using the randomized sampling consistency (RANSAC) algorithm. Finally, an intelligent pruning decision-making algorithm was developed by investigating of orchardists' pruning experience, analyzing relevant literature, and integrating phenotypic parameter acquisition methods, thus achieving accurate prediction of apple tree pruning points. [Results and Discussions] The improved U-Net model proposed achieved a mean pixel accuracy (mPA) of 95.52% for branch segmentation, representing a 2.74 percent point improvement over the original architecture. Corresponding increases were observed in mean intersection over union (mIoU) and precision metrics. Comparative evaluations against DeepLabV3+, PSPNet, and the baseline U-Net were conducted under both backlight and front-light illumination conditions. The improved model demonstrated superior segmentation performance and robustness across all tested scenarios. Ablation experiments indicated that replacing the original feature extractor with VGG16 yielded a 1.52 percent point mPA improvement, accompanied by simultaneous gains in mIoU and precision. The integration of the CBAM at the up sampling stage further augmented the model's capacity to resolve fine branch structures. Phenotypic parameter estimation using segmented branch masks combined with depth maps showed strong correlations with manual measurements. Specifically, the coefficient of determination (R2) values for primary branch diameter, branch spacing, and trunk diameter were 0.96, 0.95, and 0.91, respectively. The mean absolute errors (MAE) were recorded as 1.33, 13.96, and 5.11 mm, surpassing the accuracy of visual assessments by human pruning operators. The intelligent pruning decision system achieved an 87.88% correct identification rate for pruning points, with an average processing time of 4.2 s per viewpoint. These results confirm the practical feasibility and operational efficiency of the proposed method in real-world agricultural settings. [Conclusions] An efficient and accurate method for identifying pruning points on apple trees was proposed, which integrates image and point cloud data through deep learning. The results indicate that this method could provide significant support for the application of intelligent pruning robots in modern agriculture. It not only offers high feasibility but also exhibits outstanding efficiency and accuracy in practical applications, thus laying a solid foundation for the advancement of agricultural automation.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Agriculture    2025, 7 (1): 0-1.  
    Abstract266)            Save
    Related Articles | Metrics | Comments0
    High-Precision Fish Pose Estimation Method Based on Improved HRNet
    PENG Qiujun, LI Weiran, LIU Yeqiang, LI Zhenbo
    Smart Agriculture    2025, 7 (3): 160-172.   DOI: 10.12133/j.smartag.SA202502001
    Abstract254)   HTML27)    PDF(pc) (2577KB)(54)       Save

    [Objective] Fish pose estimation (FPE) provides fish physiological information, facilitating health monitoring in aquaculture. It aids decision-making in areas such as fish behavior recognition. When fish are injured or deficient, they often display abnormal behaviors and noticeable changes in the positioning of their body parts. Moreover, the unpredictable posture and orientation of fish during swimming, combined with the rapid swimming speed of fish, restrict the current scope of research in FPE. In this research, a FPE model named HPFPE is presented to capture the swimming posture of fish and accurately detect their key points. [Methods] On the one hand, this model incorporated the CBAM module into the HRNet framework. The attention module enhanced accuracy without adding computational complexity, while effectively capturing a broader range of contextual information. On the other hand, the model incorporated dilated convolution to increase the receptive field, allowing it to capture more spatial context. [Results and Discussions] Experiments showed that compared with the baseline method, the average precision (AP) of HPFPE based on different backbones and input sizes on the oplegnathus punctatus datasets had increased by 0.62, 1.35, 1.76, and 1.28 percent point, respectively, while the average recall (AR) had also increased by 0.85, 1.50, 1.40, and 1.00, respectively. Additionally, HPFPE outperformed other mainstream methods, including DeepPose, CPM, SCNet, and Lite-HRNet. Furthermore, when compared to other methods using the ornamental fish data, HPFPE achieved the highest AP and AR values of 52.96%, and 59.50%, respectively. [Conclusions] The proposed HPFPE can accurately estimate fish posture and assess their swimming patterns, serving as a valuable reference for applications such as fish behavior recognition.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Multi Environmental Factor Optimization Strategies for Venlo-type Greenhouses Based on CFD
    NIE Pengcheng, CHEN Yufei, HUANG Lu, LI Xuehan
    Smart Agriculture    2025, 7 (3): 199-209.   DOI: 10.12133/j.smartag.SA202502002
    Abstract234)   HTML15)    PDF(pc) (2374KB)(479)       Save

    [Objective] In the context of modern agricultural practices, regulating the indoor microclimate of Venlo-type greenhouses during the summer months through mechanical ventilation remains a significant challenge. This is primarily due to the nonlinear dynamics and strong coupling characteristics inherent in greenhouse environmental systems, where variables such as temperature, humidity, and CO₂ concentration interact in complex, interdependent ways. Traditional control strategies, which often rely heavily on empirical knowledge and manual intervention, are insufficient to cope with such dynamic conditions. These approaches tend to result in imprecise control, substantial time delays in response to environmental fluctuations, and unnecessarily high energy consumption during ventilation operations. Therefore, this study combines computational fluid dynamics (CFD) model with a multi-objective particle swarm optimization (MOPSO) algorithm to establish a joint optimization framework, aiming to address the issues of significant time delays and excessive operational energy consumption caused by vague environmental control strategies in this scenario. [Methods] To build a reliable simulation and optimization framework, environmental parameters such as temperature, wind speed, and CO₂ concentration were continuously collected via a network of environmental monitoring sensors installed at various positions inside the greenhouse. These real-time data served as validation benchmarks for the CFD model and supported the verification of mesh independence. In the CFD model construction, the internal structure of the Venlo-type greenhouse was precisely reconstructed, and appropriate boundary conditions were set based on empirical data. The airflow dynamics and thermal field were simulated using a finite volume-based solver. Four grid resolutions were evaluated for grid independence by comparing the variations in output metrics. The controllable parameters in the model included fan outlet wind speed and cooling pad condensation temperature. These parameters were systematically varied within predefined ranges. To evaluate the greenhouse environmental quality and energy consumption under different control conditions, three custom-defined objective functions were proposed: temperature suitability, CO2 uniformity, and fan operating energy consumption. The MOPSO algorithm was then applied to conduct iterative optimization over the defined parameter space. At each step, the objective functions were recalculated based on CFD outputs, and Pareto-optimal solutions were identified using non-dominated sorting. After iterative optimization using the algorithm, the conflicting objectives of environmental deviation and energy consumption were balanced, leading to the optimal range for the greenhouse environmental control strategy in this scenario. [Results and Discussions] The experimental results showed that the environmental field simulation accuracy of the CFD model was high, with an average relative error of 5.7%. In the grid independence test, three grid types, coarse, medium, and fine, were selected. The variations in the grid divisions were 1.7% and 0.6%, respectively. After considering both computational accuracy and efficiency, the medium grid division standard was adopted for subsequent simulations. The optimization strategy proposed in this study allows for closed-loop evaluation of the environment. The algorithm set the population size to 100 particles, and within the specified range of fan outlet wind speed and cooling pad condensation temperature, each particle iterates 5 times for optimization. The position updated in each iteration was used to calculate the values of the three objective evaluation functions, followed by non-dominated comparison and adaptation of the solutions, until the optimization was complete. In the Pareto surface fitted by the output results, the fan outlet wind speed ranges from 2.8 to 5.4 m/s, and the inlet temperature ranges from 295.3 to 299.7 K. Since the evaluation functions under the environmental control strategy were all in an ideal, non-dominated state, two sets of boundary control conditions were randomly selected for simulation: operating Condition A [296 K, 3.5 m/s] and operating Condition B [299 K, 5 m/s]. Post-processing contour plots showed that both operating conditions achieve good environmental optimization uniformity. The approximate ranges for each parameter were: temperature from 300.3 to 303.9 K, wind speed from 0.7 to 2.3 m/s, and CO2 concentration from 2.43 × 10-5 to 3.56 × 10-5 kmol/m3. Based on environmental uniformity optimization, operating Condition A focused on adjusting the suitable temperature for crops by lowering the cooling pad condensation temperature, but there was a relative stagnation of CO2. Operating Condition B, by increasing the fan outlet wind speed, focused on regulating CO2 flow and diffusion, but the gradient change of airflow near the two side walls was relatively abrupt. [Conclusions] This study complements the research on the systematic adjustment of greenhouse environmental parameters, while its closed-loop iterative features also enhance the simulation efficiency. The simulation results show that by arbitrarily combining the optimal solution set within the theoretical range of the strategy output, optimization of the targeted objectives can be achieved by appropriately discarding other secondary objectives, providing a reference for regulating the uniformity and economy of mechanical ventilation in greenhouses. Subsequent research can further quantify the coupling effects and weight settings of each objective function to improve the overall optimization of the functions.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Obstacle Avoidance Control Method of Electric Skid-Steering Chassis Based on Fuzzy Logic Control
    LI Lei, SHE Xiaoming, TANG Xinglong, ZHANG Tao, DONG Jiwei, GU Yuchuan, ZHOU Xiaohui, FENG Wei, YANG Qinghui
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202408003
    Online available: 27 December 2024

    Current Status and Development Trend of the Low-Altitude Economy Industry in Orchards
    WANG Xuechang, XU Wenbo, ZHENG Yongjun, YANG Shenghui, LIU Xingxing, SU Daobilige, WANG Zimeng
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202506008
    Online available: 24 July 2025

    Design and Test of Rotary Envelope Combing-Type Tobacco Leaf Harvesting Mechanism
    WANG Xiaohan, RAN Yunliang, GE Chao, GUO Ting, LIU Yihao, CHEN Du, WANG Shumao
    Smart Agriculture    2025, 7 (3): 210-223.   DOI: 10.12133/j.smartag.SA202501020
    Abstract195)   HTML11)    PDF(pc) (2231KB)(41)       Save

    [Objective] China is a big tobacco producer in the world, where tobacco significantly contributes to the national economy. Among all production stages, the leaf harvesting process requires the most labor. Currently, tobacco leaf harvesting in China remains predominantly manual, characterized by low mechanization, high labor demand, a limited harvesting window, and high labor intensity. With the advancement of agricultural modernization, mechanized tobacco leaf harvesting has become increasingly essential. However, existing tobacco harvesters are oversized and cause substantial leaf damage, making them unsuitable for China's conditions. To address this, a rotary envelope combing-type harvesting mechanism is proposed to minimize leaf damage and loss during harvesting. [Methods] A tobacco plant model was developed based on morphological characteristics and assessed the mechanical properties of tobacco leaves using a digital push-pull force gauge to measure tensile and bending characteristics. Force measurements for leaf separation in various directions revealed minimal force requirements when separating leaves from top to bottom. Based on these findings, a rotary envelope comb-type harvesting mechanism was designed, featuring both a transmission mechanism and a picking wheel. During operation, the picking wheel rotates around the tobacco stem, employing inertial combing from top to bottom for efficient leaf separation. Analysis of interactions between the picking mechanism and tobacco leaves identified combing speed as the parameter with greatest impact on picking efficiency. The mechanism's structural parameters affecting the picking wheel's movement trajectory were examined, and an improved particle swarm optimization algorithm was applied using MATLAB to refine these parameters. Additionally, Abaqus finite element simulation software was utilized to optimize the wheel structure's mechanical combing process. Dynamic simulation tests using Adams software modeled the mechanism's process of enveloping the tobacco stem and separating leaves, validating suction efficiency and determining optimal envelope range and speed parameters at various traveling speeds. To evaluate the picking effect and effectiveness of the tobacco leaf picking mechanism designed in this study, a field experiment was conducted in Sanxiang town, Yiyang county, Henan province. The performance of the harvesting mechanism was analyzed based on two critical evaluation criteria: the rate of tobacco leaf damage and the leakage rate. [Results and Discussions] By optimizing the mechanism's structural parameters using MATLAB, horizontal movement was reduced by 50.66%, and the movement trajectory was aligned vertically with the tobacco leaves, significantly reducing the risk of collision during the picking process. Finite element analysis identified the diameter of the picking rod as the key structural parameter influencing picking performance. Following extensive simulations, the optimal picking rod diameter was determined to be 15 mm, offering an ideal balance between structural strength and functional performance. The optimal envelope circle diameter for the mechanism was established at 70 mm. Aluminum alloy was selected as the material for the picking rod due to its lightweight nature, high strength-to-weight ratio, and excellent corrosion resistance. Dynamics analysis further revealed that the combing speed should not exceed 2.5 m/s to minimize leaf damage. The ideal rotational speed range for the picking mechanism was determined to be between 120 and 210 r/min, balancing operational efficiency with leaf preservation. These findings provide crucial guidance for refining the design and enhancing the practical performance of the picking mechanism. Field tests confirmed that the mechanism significantly improved operational performance, achieving a leakage rate below 7% and a damage rate below 10%, meeting the requirements for efficient tobacco picking. It was observed that excessive leaf leakage primarily occurred when leaves were steeply inclined, which hindered effective stem envelopment by the harvesting mechanism. Consequently, the mechanism proved particularly effective for picking centrally positioned leaves, while drooping leaves resulted in higher leakage and damage rates. The primary cause of leaf damage was found to be mechanical contact between the harvesting mechanism and the leaves during operation. Notably, while increasing striking speed reduced leakage, it simultaneously led to a higher damage rate. Compared to the existing harvesting mechanism, this newly developed mechanism is more compact and supports layered leaf picking, making it especially well-suited for integration into small- and medium-sized harvesting machinery. [Conclusions] This study presents an effective and practical solution for tobacco leaf harvesting mechanization, specifically addressing the critical challenges of leaf damage and leakage. The proposed solution not only improves harvesting quality but also features a significantly simplified mechanical structure. By combining innovative technology with optimized design, this approach minimizes impact on delicate leaves, reduces leakage, and ensures higher yields with minimal human intervention. Analysis and testing demonstrate this mechanized solution's potential to significantly reduce production losses, offering both economic and operational benefits for the tobacco industry.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Grading Asparagus officinalis L. Using Improved YOLOv11
    YANG Qilang, YU Lu, LIANG Jiaping
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202501024
    Online available: 03 June 2025

    Dynamic Prediction Model of Crop Canopy Temperature Based on VMD-LSTM
    WANG Yuxi, HUANG Lyuwen, DUAN Xiaolin
    Smart Agriculture    2025, 7 (3): 143-159.   DOI: 10.12133/j.smartag.SA202502015
    Abstract178)   HTML14)    PDF(pc) (4491KB)(43)       Save

    [Objective] Accurate prediction of crop canopy temperature is essential for comprehensively assessing crop growth status and guiding agricultural production. This study focuses on kiwifruit and grapes to address the challenges in accurately predicting crop canopy temperature. [Methods] A dynamic prediction model for crop canopy temperature was developed based on Long Short-Term Memory (LSTM), Variational Mode Decomposition (VMD), and the Rime Ice Morphology-based Optimization Algorithm (RIME) optimization algorithm, named RIME-VMD-RIME-LSTM (RIME2-VMD-LSTM). Firstly, crop canopy temperature data were collected by an inspection robot suspended on a cableway. Secondly, through the performance of multiple pre-test experiments, VMD-LSTM was selected as the base model. To reduce cross-interference between different frequency components of VMD, the K-means clustering algorithm was applied to cluster the sample entropy of each component, reconstructing them into new components. Finally, the RIME optimization algorithm was utilized to optimize the parameters of VMD and LSTM, enhancing the model's prediction accuracy. [Results and Discussions] The experimental results demonstrated that the proposed model achieved lower Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) (0.360 1 and 0.254 3 °C, respectively) in modeling different noise environments than the comparator model. Furthermore, the R2 value reached a maximum of 0.994 7. [Conclusions] This model provides a feasible method for dynamically predicting crop canopy temperature and offers data support for assessing crop growth status in agricultural parks.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Vegetable Price Prediction Based on Optimized Neural Network Time Series Models
    HOU Ying, SUN Tan, CUI Yunpeng, WANG Xiaodong, ZHAO Anping, WANG Ting, WANG Zengfei, YANG Weijia, GU Gang
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202410037
    Online available: 22 May 2025

    The Bee Pollination Recognition Model Based On The Lightweight YOLOv10n-CHL
    CHANG Jian, WANG Bingbing, YIN Long, LI Yanqing, LI Zhaoxin, LI Zhuang
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202502033
    Online available: 06 June 2025

    The Lightweight Bee Pollination Recognition Model Based On YOLOv10n-CHL
    CHANG Jian, WANG Bingbing, YIN Long, LI Yanqing, LI Zhaoxin, LI Zhuang
    Smart Agriculture    2025, 7 (3): 185-198.   DOI: 10.12133/j.smartag.SA202503033
    Abstract148)   HTML14)    PDF(pc) (2411KB)(26)       Save

    [Objective] Bee pollination is pivotal to plant reproduction and crop yield, making its identification and monitoring highly significant for agricultural production. However, practical detection of bee pollination poses various challenges, including the small size of bee targets, their low pixel occupancy in images, and the complexity of floral backgrounds. Aimed to scientifically evaluate pollination efficiency, accurately detect the pollination status of flowers, and provide reliable data to guide flower and fruit thinning in orchards, ultimately supports the scientific management of bee colonies and enhances agricultural efficiency, a lightweight recognition model that can effectively overcome the above obstacles was proposed, thereby advancing the practical application of bee pollination detection technology in smart agriculture. [Methods] A specialized bee pollination dataset was constructed comprising three flower types: strawberry, blueberry, and chrysanthemum. High-resolution cameras were used to record videos of the pollination process, which were then subjected to frame sampling to extract representative images. These initial images underwent manual screening to ensure quality and relevance. To address challenges such as limited data diversity and class imbalance, a comprehensive data augmentation strategy was employed. Techniques including rotation, flipping, brightness adjustment, and mosaic augmentation were applied, significantly expanding the dataset's size and variability. The enhanced dataset was subsequently split into training and validation sets at an 8:2 ratio to ensure robust model evaluation. The base detection model was built upon an improved YOLOv10n architecture. The conventional C2f module in the backbone was replaced with a novel cross stage partial network_multi-scale edge information enhance (CSP_MSEE) module, which synergizes the cross-stage partial connections from cross stage partial network (CSPNet) with a multi-scale edge enhancement strategy. This design greatly improved feature extraction, particularly in scenarios involving fine-grained structures and small-scale targets like bees. For the neck, a hybrid-scale feature pyramid network (HS-FPN) was implemented, incorporating a channel attention (CA) mechanism and a dimension matching (DM) module to refine and align multi-scale features. These features were further integrated through a selective feature fusion (SFF) module, enabling the effective combination of low-level texture details and high-level semantic representations. The detection head was replaced with the lightweight shared detail enhanced convolutional detection head (LSDECD), an enhanced version of the Lightweight shared convolutional detection head (LSCD) detection head. It incorporated detail enhancement convolution (DEConv) from DEA-Net to improve the extraction of fine-grained bee features. Additionally, the standard convolution_groupnorm (Conv_GN) layers were replaced with detail enhancement convolution_ groupnorm (DEConv_GN), significantly reducing model parameters and enhancing the model's sensitivity to subtle bee behaviors. This lightweight yet accurate model design made it highly suitable for real-time deployment on resource-constrained edge devices in agricultural environments. [Results and Discussions] Experimental results on the three bee pollination datasets: strawberry, blueberry, and chrysanthemum, demonstrated the effectiveness of the proposed improvements over the baseline YOLOv10n model. The enhanced model achieved significant reductions in computational overhead, lowering the computational complexity by 3.1 GFLOPs and the number of parameters by 1.3 M. The computational cost of the improved model reached 5.1 GFLOPS, and the number of parameters was 1.3 M. These reductions contribute to improved efficiency, making the model more suitable for deployment on edge devices with limited processing capabilities, such as mobile platforms or embedded systems used in agricultural monitoring. In terms of detection performance, the improved model showed consistent gains across all three datasets. Specifically, the recall rates reached 82.6% for strawberry flowers, 84.0% for blueberry flowers, and 84.8% for chrysanthemum flowers. Corresponding mAP50 (Mean Average Precision at IoU threshold of 0.5) scores were 89.3%, 89.5%, and 88.0%, respectively. Compared to the original YOLOv10n model, these results marked respective improvements of 2.1% in recall and 1.7% in mAP50 on the strawberry dataset, 2.0% and 2.6% on the blueberry dataset, and 2.1% and 2.2% on the chrysanthemum dataset. [Conclusions] The proposed YOLOv10n-CHL lightweight bee pollination detection model, through coordinated enhancements at multiple architectural levels, achieved notable improvements in both detection accuracy and computational efficiency across multiple bee pollination datasets. The model significantly improved the detection performance for small objects while substantially reducing computational overhead, facilitating its deployment on edge computing platforms such as drones and embedded systems. This research could provide a solid technical foundation for the precise monitoring of bee pollination behavior and the advancement of smart agriculture. Nevertheless, the model's adaptability to extreme lighting and complex weather conditions remains an area for improvement. Future work will focus on enhancing the model's robustness in these scenarios to support its broader application in real-world agricultural environments.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Smart Agriculture    2025, 7 (2): 0-0.  
    Abstract133)            Save
    Related Articles | Metrics | Comments0
    Embodied Intelligent Agricultural Robots: Key Technologies, Application Analysis, Challenges and Prospects
    WEI Peigang, CAO Shanshan, LIU Jifang, LIU Zhenhu, SUN Wei, KONG Fantao
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505008
    Online available: 30 June 2025

    U-Net Greenhouse Sweet Cherry Image Segmentation Method Integrating PDE Plant Temporal Image Contrastive Learning and GCN Skip Connections
    HU Lingyan, GUO Ruiya, GUO Zhanjun, XU Guohui, GAI Rongli, WANG Zumin, ZHANG Yumeng, JU Bowen, NIE Xiaoyu
    Smart Agriculture    2025, 7 (3): 131-142.   DOI: 10.12133/j.smartag.SA202502008
    Abstract123)   HTML15)    PDF(pc) (2118KB)(23)       Save

    [Objective] Within the field of plant phenotyping feature extraction, the accurate delineation of small targets boundaries and the adequate recovery of spatial details during upsampling operations have long been recognized as significant obstacles hindering progress. To address these limitations, an improved U-Net architecture designed for greenhouse sweet cherry image segmentation. [Methods] Taking temporal phenotypic images of sweet cherries as the research subject, the U-Net segmentation model was employed to delineate the specific organ regions of the plant. This architecture was referred to as the U-Net integrating self-supervised contrastive learning method for plant time-series images with priori distance embedding (PDE) pre-training and graph convolutional networks (GCN ) skip connection for greenhouse sweet cherry image segmentation. To accelerate model convergence, the pre-trained weights derived from the PDE plant temporal image contrastive learning method were transferred to. Concurrently, the incorporation of a GCN local feature fusion layer was incorporated as a skip connection to optimize feature fusion, thereby providing robust technical support for image segmentation task. The PDE plant temporal image contrastive learning method pre-training required the construction of image pairs corresponding to different phenological periods. A classification distance loss function, which incorporated prior knowledge, was employed to construct an Encoder with adjusted parameters. Pre-trained weights obtained from the PDE plant temporal image contrastive learning method were effectively transferred and and applied to the semantic segmentation task, enabling the network to accurately learn semantic information and detailed textures of various sweet cherry organs. The Encoder module performs multi-scale feature extraction by convolutional and pooling layers. This process enabled the hierarchical processing of the semantic information embedded in the input image to construct representations that progress transitions from low-level texture features to high-level semantic features. This allows consistent extraction of semantic features from images across various scales and abstraction of underlying information, enhancing feature discriminability and optimizing modeling of complex targets. The Decoder module was employed to conduct up sampling operations, which facilitated the integration of features from diverse scales and the restoration of the original image resolution. This enabled the results to effectively reconstruct spatial details and significantly improve the efficiency of model optimization. At the interface between the Encoder and Decoder modules, a GCN layer designed for local feature fusion was strategically integrated as a skip connection, enabling the network to better capture and learn the local features in multi-scale images. [Results and Discussions] Utilizing a set of evaluation metrics including accuracy, precision, recall, and F1-Score, an in-depth and rigorous assessment of the model's performance capabilities was conducted. The research findings revealed that the improved U-Net model achieved superior performance in semantic segmentation of sweet cherry images, with an accuracy of up to 0.955 0. Ablation experiments results further revealed that the proposed method attained a precision of 0.932 8, a recall of 0.927 4, and an F1-Score of 0.912 8. The accuracy of improved U-Net is higher by 0.069 9, 0.028 8, and 0.042 compared to the original U-Net, U-Net with PDE plant temporal image contrastive learning method, and U-Net with GCN skip connections, respectively. Meanwhile the F1-Score is 0.078 3, 0.033 8, and 0.043 8 higher respectively. In comparative experiments against DeepLabV3, Swin Transformer and Segment Anything Model segmentation methods, the proposed model surpassed the above models by 0.022 2, 0.027 6 and 0.042 2 in accuracy; 0.063 7, 0.147 1 and 0.107 7 in precision; 0.035 2, 0.065 4 and 0.050 8 in recall; and 0.076 8, 0.127 5 and 0.103 4 in F1-Score. [Conclusions] The incorporation of the PDE plant temporal image contrastive learning method and the GCN techniques was utilized to develop an advanced U-Net architecture that is specifically designed and optimized for the analysis of sweat cherry plant phenotyping. The results demonstrate that the proposed method is capable of effectively addressing the issues of boundary blurring and detail loss associated with small targets in complex orchard scenarios. It enables the precise segmentation of the primary organs and background regions in sweet cherry images, thereby enhancing the segmentation accuracy of the original model. This improvement provides a solid foundation for subsequent crop modeling research and holds significant practical importance for the advancement of agricultural intelligence.

    Table and Figures | Reference | Related Articles | Metrics | Comments0
    Research Progress and Prospects of Multi-Agent Large Language Models in Agricultural Applications
    ZHAO Yingping, LIANG Jinming, CHEN Beizhang, DENG Xiaoling, ZHANG Yi, XIONG Zheng, PAN Ming, MENG Xiangbao
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202503026
    Online available: 24 July 2025

    Estimation of Corn Aboveground Biomass Based on CNN-LSTM-SA
    WANG Yi, XUE Rong, HAN Wenting, SHAO Guomin, HOU Yanqiao, CUI Xitong
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202412004
    Online available: 27 June 2025

    Intelligent Inspection Path Planning Algorithm for Large-Scale Cattle Farms
    CHEN Ruotong, LIU Jifang, ZHANG Zhiyong, MA Nan, WEI Peigang, WANG Yi
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202504004
    Online available: 12 June 2025

    A Study on Corn Borer Detection Using Low-Altitude Close-Range UAV Imagery
    ZHAO Jun, NIE Zhigang, LI Guang, LIU Jiayu
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505006
    Online available: 09 July 2025

    LightTassel-YOLO: A Real-Time Detection Method for Maize Tassels Based on UAV Remote Sensing
    CAO Yuying, LIU Yinchuan, GAO Xinyue, JIA Yinjiang, DONG Shoutian
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505021
    Online available: 01 August 2025

    Multi-objective Planting Planning Method Based on Connected Components and Genetic Algorithm: A Case Study of Fujin City
    XU Menghua, WANG Xiujuan, LENG Pei, ZHANG Mengmeng, WANG Haoyu, HUA Jing, KANG Mengzhen
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202504012
    Online available: 27 June 2025

    Beef Cattle Object Detection Method Based on Improved YOLOv12
    LIU Yiheng, LIU Libo
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202503018
    Online available: 23 July 2025

    A Transfer Learning-Based Multimodal Model for Grape Detection and Counting
    XU Wenwen, YU Kejian, DAI Zexu, WU Yunzhi
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202504005
    Online available: 16 June 2025

    Forecasting Method for China's Soybean Demand Based on Improved Temporal Fusion Transformers
    LIU Jiajia, QIN Xiaojing, LI Qianchuan, XU Shiwei, ZHAO Jichun, WANG Yigang, XIONG Lu, LIANG Xiaohe
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505017
    Online available: 28 July 2025

    Smart Agriculture    2025, 7 (3): 0-1.  
    Abstract60)            Save
    Related Articles | Metrics | Comments0
    Rapid Tea Identification and Polyphenol Detection in Fresh Tea Leaves Using Visible/Shortwave and Longwave Near-Infrared Spectroscopy
    XU Jinchai, LI Xiaoli, WENG Haiyong, HE Yong, ZHU Xuesong, LIU Hongfei, HUANG Zhenxiong, YE Dapeng
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505034
    Online available: 22 July 2025

    Current Status and Trends of Application Scenarios and Industrial Development in the Agricultural Low-Altitude Economy
    HE Yong, DAI Fushuang, ZHU Jiangpeng, HE Liwen, WANG Yueying
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202507014
    Online available: 26 August 2025

    Application of Photoacoustic Spectroscopy in Quality Assessment of Agricultural and Forestry Products
    XIE Weijun, CHEN Keying, QIAO Mengmeng, WU Bin, GUO Qing, ZHAO Maocheng
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505026
    Online available: 08 July 2025

    Semi-Supervised Deep Convolutional Generative Adversarial Network for Imbalanced Hyperspectral Viability Detection of Naturally Aged Soybean Germplasm
    LI Fei, WANG Ziqiang, WU Jing, XIN Xia, LI Chunmei, XU Hubo
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505013
    Online available: 13 August 2025

    Chinese Tea Pest and Disease Named Entity Recognition Method Based on Improved Boundary Offset Prediction Network
    XIE Yuxin, WEI Jiangshu, ZHANG Yao, LI Fang
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505007
    Online available: 01 August 2025

    Detection Method of Tea Geometrid Larvae in Canopy Environments Based on YOLO and Diffusion Models
    LUO xuelun, GOUDA Mostafa, SONG xinbei, HU yan, ZHANG wenkai, HE yong, ZHANG jin, LI Xiaoli
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505023
    Online available: 13 August 2025

    Advances in the Application of Multi-source Data Fusion Technology in Non-Destructive Detection of Apple
    GUO Qi, FAN Yixuan, YAN Xinhuan, LIU Xuemei, CAO Ning, WANG Zhen, PAN Shaoxiang, TAN Mengnan, ZHENG Xiaodong, SONG Ye
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505036
    Online available: 13 August 2025

    Research on The Acoustic-Vibration Detection Method for The Apple Moldy Core Disease Based on D-S Evidence Theory
    LIU Jie, ZHAO Kang, ZHAO Qinjun, SONG Ye
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505032
    Online available: 13 August 2025

    Spectral Technology in Vegetable Production Detection: Research Progress, Challenges and Suggestions
    BAI Juekun, LIU Yachao, DONG Daming, YUE Xiaolong, DU Xiuke
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202504027
    Online available: 14 August 2025

    An Embedded Fluorescence Imaging Detection System for Fruit and Vegetable Quality Deterioration Based on Improved YOLOv8
    GAO Chenhong, ZHU Qibing, HUANG Min
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505038
    Online available: 04 September 2025

    Non-destructive Detection of Apple Water Core Disease Based on Hyperspectral and X-ray CT Imaging
    YU Xinyuan, WANG Zhenjie, YOU Sicong, TU Kang, LAN Weijie, PENG Jing, ZHU Lixia, CHEN Tao, PAN Leiqing
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202507022
    Online available: 04 September 2025

    Rapid and Non-Destructive Analysis of Hawthorn Moisture Content Based on Hyperspectral Imaging Technology
    BAI Ruibin, WANG Hui, WANG Hongpeng, HONG Jiashun, ZHOU Junhui, YANG Jian
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505033
    Online available: 14 August 2025

    Small Target Detection Method of Maize Leaf Disease Based on DCC-YOLOv10n
    DANG Shanshan, QIAO Shicheng, BAI Mingyu, ZHANG Mingyue, ZHAO Chenyu, PAN Chunyu, WANG Guochen
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202504017
    Online available: 18 August 2025

    Remote Sensing Approaches for Cropland Abandonment Perception in Southern Hilly and Mountainous Areas of China: A Review
    LONG Yuqiao, SUN Jing, WEN Yanru, WANG Chuya, DONG Xiuchun, HUANG Ping, WU Wenbin, CHEN Jin, DING Mingzhong
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505022
    Online available: 28 August 2025

    Detection Method for Log-Cultivated Shiitake Mushrooms Based on Improved RT-DETR
    WANG Fengyun, WANG Xuanyu, AN Lei, FENG Wenjie
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202506034
    Online available: 11 September 2025

    Monte Carlo Simulation of Light Propagation in Orah Mandarin Tissues and Optimization of Spectral Detection in Diffuse Reflection Mode
    OUYANG Aiguo, WANG Yang, HOU Youfei, WANG Guantian, LIU Yande
    Smart Agriculture    DOI: 10.12133/j.smartag.SA202505029
    Online available: 26 August 2025