1. Introduction
Cotton is an important economic crop and raw material for the textile industry and plays a crucial role in the national economy and people’s livelihoods [
1]. However, cotton often suffers from various diseases and pests during its growth, with pest infestation being one of the most prominent factors affecting the high yield and quality of cotton in China, directly impacting the economic income of cotton farmers [
2]. The cotton aphid is an important pest during the seedling stage that is widely distributed in cotton areas throughout China and easily spreads once it occurs. Cotton aphids insert their piercing-sucking mouthparts into the underside or tender parts of cotton leaves, sucking sap. The infested leaves curl upward, forming an arc, with aphid excretions visible on the leaf surface, often promoting mold growth. Infected cotton plants exhibit stunted growth, relatively small leaves, and reduced leaf numbers [
3].
Xinjiang, owing to its unique geographical location, is the largest cotton planting and processing base in China [
1]. Owing to its large and concentrated cotton planting scale, which is conducive to large-scale and continuous agricultural machinery operations, it possesses characteristics conducive to the development of mechanical intelligence. Rapid and accurate identification of the severity of cotton aphid damage in the field, followed by direct implementation of further control measures based on the identified results, such as precise variable pesticide spraying, has positive effects on improving efficiency, preventing further spread of pests, and thus enhancing the quality and yield of cotton. Currently, the discovery of cotton aphid damage relies primarily on manual observation, which is not only time-consuming and laborious but also prone to inaccuracies in human judgment, leading to the further spread of pest problems and highlighting serious drawbacks in large-scale cotton pest control and management.
With the continuous advancement of computer vision technology and deep learning methods, the speed and accuracy of traditional image analysis techniques have become insufficient to meet the demands of modern agricultural production. The utilization of deep learning methods for pest and disease detection has gradually become mainstream [
4]. The model learns from a large number of data features to efficiently identify crop pests and diseases without much human involvement. For example, Alves et al. [
5] introduced a field image cotton pest classification method based on deep residual networks. They constructed a dataset containing the 15 most common cotton pests and harmless categories and proposed a specific residual deep learning model, ResNet34*, which automatically classifies major pests from given images. A performance evaluation against other convolutional neural networks revealed that the proposed model achieved the highest accuracy, with an f score of 0.98. Bao et al. [
6] developed a lightweight CA_DenseNet_BC_40 model based on DenseNet for classifying leaf damage caused by cotton aphids. They introduced coordinated attention mechanisms into each submodule of the dense block to better extract features of cotton aphid-induced symptoms, achieving a high classification accuracy of 97.3%. With the rapid development of drone technology, there has been an increasing amount of academic research using drone-based remote sensing technology for crop pest and disease detection. This approach allows for precise and rapid monitoring of pests and diseases [
7]. Li et al. [
8] utilized high-resolution multispectral images captured by drones. Using data analysis methods such as partial least squares regression (PLSR) and neural network models, they developed a Verticillium wilt monitoring model for cotton based on drone remote sensing images. On this basis, they integrated the use of agricultural drones to generate prescription maps for spraying against Verticillium wilt, thereby achieving an integrated approach from disease detection to pest control.
Object detection is an important direction where computer vision and deep learning intersect [
9], which not only classifies objects in images but also accurately identifies their positions. It has extensive applications in industrial manufacturing and smart agriculture, with representative algorithms including the one-stage YOLO [
10] series and the two-stage Faster R-CNN [
11]. Among them, the YOLO series models, known for their rapid and accurate performance, are widely used in agricultural detection. Some scholars [
12] have combined drone multispectral imaging with object detection algorithms to perform seedling detection and counting in cotton crops. By utilizing multispectral images collected by drones, models suitable for counting throughout the cotton seedling stage were established via three deep learning algorithms: YOLOv5, YOLOv7, and CenterNet. The results indicated that YOLOv7 performed better in terms of detection and counting. Li et al. [
13] integrated three new modules, SCFEM, DSCFEM, and SPPM, into the YOLOv5 model to extract image features effectively. They applied this model to the detection of ten types of jute diseases and pests, achieving an average accuracy of 96.63%, demonstrating the broad prospects of machine learning and deep learning in precise agricultural disease prevention. Xu et al. [
14] collected images of cotton seedling aphid infections via smartphones and compiled them into an image dataset. They constructed three different object detection models, Faster R-CNN, YOLOv5, and SSD, with YOLOv5 achieving the highest mAP (95.7%) and FPS (61.73) and successfully deploying it on a mobile Android phone. Many researchers are dedicated to developing lightweight models that can be more easily deployed on mobile devices, thereby facilitating the effective application of pest and disease detection models in real-world scenarios. Xu et al. [
15] developed a lightweight SSV2-YOLO model based on YOLOv5s. By reconstructing the backbone network via Stem and ShuffleNet V2 and adjusting the neck network width, it was used for sugarcane aphid detection. The feature level, data augmentation method, and loss function were further reconstructed to improve the detection performance for small, dense, and overlapping targets. The improved method is extremely lightweight and can be used for real-time detection on mobile devices. Lin et al. [
16] developed a system for automatically detecting peanut leaf diseases. The system is based on the YOLOv8n model and combines the advanced FasterNeXt and DSConv modules and the generalized intersection over union loss function. It aims to improve the lightweight nature of the model so that it can adapt to edge computing devices and realize real-time detection of peanut leaf diseases.
In summary, deep learning has demonstrated good performance in crop pest and disease detection because of its powerful feature extraction and autonomous learning capabilities [
17]. Many scholars have achieved high real-time detection performance on standard platforms. However, the large size of these computing platforms makes them unsuitable for the detection of pests and diseases in field farm environments. Furthermore, most studies in this field focus on detecting a single image at a time, which has limited detection coverage and low efficiency. This approach is inadequate for the management and control of pests and diseases in large-scale cotton fields in Xinjiang. We aim to use edge computing devices with low power consumption and small sizes for real-time detection of pests and diseases. Real-time detection of video data can display the occurrence of pests and diseases in a larger range and improve detection efficiency. Moreover, the deployment of pest and disease detection models in edge computing scenarios can be combined with pesticide application tractors and drones. By placing the computing unit on the edge, large-area pest and disease detection can be achieved without the need for long-distance data transmission, greatly reducing labor and improving efficiency, thereby achieving precise pesticide spraying, which is more valuable for practical research. However, the computing resources of commonly used edge devices remain significantly limited [
18]. Existing general models have numerous network parameters, complex calculations, and high memory consumption, making real-time detection challenging and unable to improve the detection efficiency for pests and diseases in practical applications [
19]. Therefore, developing a cotton aphid damage detection model tailored for actual cotton field scenarios, characterized by low computational complexity, effective detection performance, and ease of deployment, is essential.
In response to the aforementioned challenges, this paper proposes an efficient cotton aphid damage detection model designed for real-time deployment on drones or tractors in cotton fields. We build upon the YOLOv8n model by integrating a module inspired by the Slim-Neck design, which reduces computational overhead to enhance detection speed at the edge. Additionally, an attention mechanism is incorporated to ensure high detection accuracy. The resulting model is named GVC-YOLO. The GVC-YOLO model is optimized and accelerated via TensorRT, deployed on the edge computing device Jetson Xavier NX, and connected to cameras to obtain real-time field videos for timely inference and detection of cotton aphid damage. This research can be combined with IoT agricultural machinery to achieve integrated intelligent pest and disease recognition and precise variable pesticide spraying, providing a convenient and effective new method for large-scale pest and disease management in the current cotton industry.
2. Materials and Methods
2.1. Data Acquisition
The image data for this experiment were collected at the Korla Experimental Station of the Institute of Plant Protection, Chinese Academy of Agricultural Sciences (41°44′59″N, 85°48′30″E). This region has a typical temperate continental climate, is full of light throughout the year, has a large temperature difference between day and night, has low precipitation, and is suitable for high-quality cotton growth. The cotton varieties used were ‘Zhongmiansuo49’ and ‘Xinluzhong66’, which were sown in mid-to-late April in mulch cultivation mode with drip irrigation under the film. The areas in the cotton field where the degree of cotton aphid infestation was severe were selected for image collection. During the experiment, no pesticides were applied to inhibit the growth of the cotton aphids. Cotton aphid infestations occurred naturally without any artificial intervention.
The data were collected from late June to mid–July 2019 and from early June to early July 2022. This period is the most serious period of cotton aphid infestation, and it is also a critical period for pesticide control. All image data were obtained under natural conditions to enrich the diversity of the image dataset, including varying light intensities, cotton leaf obstructions, overlaps, small targets, and complex backgrounds. Data acquisition was conducted with HUAWEI Nova, iPhone 8 Plus, and Mi Note 3 smartphones. The acquisition method was overhead vertical shooting. The researchers stood next to the cotton plants with a smartphone in hand and vertically shot images of the cotton seedlings from a vantage of 1.2–1.5 m. The smartphone was parallel to the ground, and the camera was 90 degrees to the ground. The image resolution was 4030 × 3024 pixels. Additionally, smartphones were used to record videos via parallel movement to obtain video data, with a resolution of 1920 × 1080 pixels and a frame rate of 30 FPS.
2.2. Data Preprocessing
The collected raw images were manually screened to remove ineffective images, such as those that were blurry or unclear. For video files, key frames were extracted. Owing to the large size of the collected raw image data and redundant information, such as the soil contained in the images, which could slow down the training speed of deep learning models, the images were uniformly cropped to a size of 3000 × 3000 pixels to remove irrelevant information. Additionally, high-definition smartphones are used for data collection, resulting in large original image resolutions, whereas lower-resolution industrial cameras are commonly used in practical applications. Models trained on high-resolution images may not perform optimally in real-world scenarios. Therefore, the image resolutions were further uniformly reduced to 1024 × 1024 pixels.
Using the annotation tool LabelImg, we conducted graded annotations of cotton aphid damage to cotton leaf canopy regions and stored the annotated data in the PASCAL VOC format. The grading of cotton aphid damage in this study followed both expert experience and national grading standards (GB/T 15799-2011) [
20], as outlined in
Table 1. Examples of acquired images for this experiment are shown in
Figure 1. Deep learning requires learning and extracting features from a large amount of labeled data. To enhance the model’s generalizability and robustness, optimize its learning ability regarding the data, and avoid overfitting, various data augmentation methods, including horizontal flipping, vertical flipping, introducing Gaussian noise, and adjusting the brightness of the original images, were applied to the 3051 annotated images, resulting in a total of four augmentation methods [
21]. The original dataset was expanded through these methods, resulting in a total of 16,950 images. These images were randomly divided into training, validation, and testing sets at a ratio of 8:1:1. The basic information of the cotton aphid damage dataset is presented in
Table 2.
2.3. YOLOv8 Object Detection Model
YOLOv8 is a state-of-the-art (SOTA) model in the field of object detection that integrates and improves upon previous generations of YOLO models. As a one-stage object detection network, it offers unparalleled performance in terms of flexibility and accuracy; its high performance and low latency enable it to accomplish various computer vision tasks quickly. To meet the practical needs of different specific scenarios, YOLOv8 provides a total of five models at different scales: N, S, M, L, and X. The depth of the network and the number of parameters increase progressively, resulting in improved accuracy of detection results, albeit with a corresponding decrease in detection speed [
22].
The network structure of YOLOv8 consists of four main parts: the input, backbone, neck, and head. The backbone network comprises standard convolution modules and C2f and SPPF modules. This section still adopts the CSPDarknet structure from YOLOv5, replacing the C3 module with a more gradient-rich and lightweight C2f module. This step of the process effectively extracts features from input images at different levels to capture crucial information. The neck section uses a PAN-FPN structure, combining the advantages of PANet and FPN to fuse multiscale features from different levels more effectively [
23]. The head section abandons the anchor-based approach used in previous generations of YOLO structures, foregoing the use of predefined anchor boxes. Instead, it adopts the anchor-free method to directly learn and predict the position and size information of targets from image features without relying on predefined reference boxes. During output, the Decoupled-Head approach separates the localization and classification tasks of object detection into two independent branches. The first branch performs regression tasks to output prediction boxes, whereas the other branch handles classification tasks to output the probabilities of the target classes. This decoupled design avoids mutual interference between tasks, thus preventing inaccurate detection results.
The classification loss of YOLOv8 employs binary cross-entropy loss (BCEL) to measure the model’s accuracy in classifying targets, as shown in Equation (1), where yi represents the label and σ(xi) represents the model’s predicted result. The bounding box regression loss adopts an improved form of the positional loss function, CIoU loss, and DFL. Distribution focal loss (DFL) attempts to learn and optimize the position of bounding boxes in a more direct manner; this allows the network to focus more quickly on the distribution of neighboring regions near the target position. In Equation (2), y represents the ground truth label, yi and yi+1 are two adjacent quality labels to y, and Si and Si+1 are the sigmoid outputs corresponding to the quality labels. The weighted sum of the three losses mentioned above constitutes the final loss function.
2.4. GVC-YOLO Model Construction
In this study, to achieve real-time and accurate detection of cotton aphid damage, the YOLOv8n network, which is the fastest in speed and most suitable for deployment on edge devices, was chosen as the base network. Lightweight reconstructions were made for this network. The standard convolutions in the backbone network and neck section of the original structure were replaced with lightweight GSConv to reduce the computational load of the model. Additionally, the VoVGSCSP structure was introduced in the neck section to further reduce the model’s weight while also improving the fusion of features related to cotton aphid damage. Furthermore, a CA mechanism was added after the C2f module in the backbone network to increase the model’s ability to extract features of cotton aphid damage. The SPPF was replaced with the more computationally efficient SimSPPF network.
Figure 2 illustrates the structure of the improved lightweight GVC-YOLO network model.
The convolution operations can efficiently conduct feature extraction and processing through local perception and parameter-sharing mechanisms; they are an indispensable part of computer vision and deep learning. However, standard convolution accumulates many parameters when dealing with high-resolution or high-channel data, increasing computational costs and memory usage.
The lightweight GSConv module [
24] combines depthwise separable convolution and standard convolution, aiming to retain as many hidden connections between channels as possible while reducing model complexity and maintaining good accuracy. The structure of the GSConv module is illustrated in
Figure 3. Suppose that the input channel is C
1, and the output channel is C
2. First, the number of channels is changed to C
2/2 after a standard convolution module; then, a depthwise separable convolution module is performed, and the number of channels remains unchanged. Then, the above two convoluted feature graphs are connected, and nonlinear shuffle operations are carried out to enhance the learning ability of the features. The mathematical expression of GSConv is depicted in Equation (3). The computational cost of GSConv is approximately 50% that of standard convolution, while its learning capability is comparable. This advantage is more evident for lightweight detectors.
In the feature fusion stage, referring to the idea of ResNet [
25], we continue to introduce the GS bottleneck [
24] on the basis of GSConv. Then, based on the Slim-Neck design concept, we adopt the one-shot aggregation method to design the cross-stage partial network (CSPNet) module VoVGSCSP [
24]. The structure is shown in
Figure 4. The calculations are illustrated in Equations (4) and (5). This feature fusion module effectively balances model accuracy and speed. Replacing the C2f module in the neck part with the VoVGSCSP module enables the model to maintain sufficient accuracy and a high feature reuse rate while reducing the complexity of the computation and network structure. Models based on the above improvements are easier to deploy in edge computing devices.
The attention mechanism [
26] was initially proposed in the field of natural language processing (NLP) and has since been widely applied in computer vision. This mechanism dynamically optimizes the allocation of importance to local regions of images, enabling models to focus more on key information relevant to the current task while reducing attention to less relevant areas, thereby enhancing the effectiveness of target detection.
Coordinate attention (CA) [
27] is an attention mechanism used to enhance deep learning models’ understanding of the spatial structure of input data. It encodes positional information in both the horizontal and vertical directions into channel attention, allowing the model to focus on large-scale spatial relationships without introducing excessive computational overhead. As illustrated in
Figure 5, the CA mechanism first performs global average pooling operations on the input feature map in both the width and height directions. The resulting two feature maps are subsequently stacked to form a new feature layer with a shape of [C, 1, H + W]. Subsequently, convolutional dimension reduction, batch normalization, and nonlinear activation functions are applied to enrich the feature representation. Then, the features of the feature map in the width and height directions are separated, and the number of channels is adjusted via convolution. The attention weights g
h and g
w in two directions are obtained via the sigmoid activation function. Finally, the multiplicative weighted calculation is performed with the original feature map, and the final feature map with the attention weight is output. The multiplication weighting formula is shown in Equation (6).
The CA mechanism, by embedding positional information, effectively enhances the model’s ability to localize cotton aphid damage to leaves in complex scenarios, thereby improving the accuracy of cotton aphid damage detection. Moreover, the CA mechanism is highly flexible, avoiding significant computational overhead and not increasing the complexity of the model.
Simplified spatial pyramid pooling-fast (SimSPPF) [
28] is an efficient network proposed in YOLOv6, which is an improved version of SPPF. Its structure, as shown in
Figure 6, serially processes three 5 × 5 scale max pooling operations and optimizes the SiLU activation function in SPPF to ReLU, thereby accelerating the convergence speed of the network and improving operational efficiency.
This network does not change the size or channels of the input feature map; its main function is to extract and fuse high-level features. During the fusion process, multiple max pooling operations are applied, greatly enriching the expressive power of the obtained feature maps. This allows for the extraction of semantic features from as many layers as possible, effectively avoiding the loss of local information in cotton aphid damage images and improving the efficiency and accuracy of feature extraction.
2.5. Experimental Platform
The experiments in this paper were conducted on a server platform with Ubuntu 20.04 LTS as the operating system. The hardware and software configuration were as follows: an Xeon(R) Gold 5218 CPU @ 2.30 GHz processor, 252 GB of RAM, and an NVIDIA A100 GPU with 40 GB of VRAM. The deep learning framework used was PyTorch 1.10 with CUDA 11.3, and GPU acceleration was performed via cuDNN 8.0.5. Multiple comparative experiments and ablation studies were conducted. The hyperparameters were kept consistent for each experiment to ensure the validity and reliability of the experiments. Detailed information is provided in
Table 3.
2.6. Edge Computing Equipment
Traditional desktop PCs are not suitable for mounting on mobile mechanical equipment in the field due to issues such as their large size and high power consumption. Moreover, traditional cloud computing paradigms have drawbacks such as unstable video data transmission and slow speed [
29]. Edge computing methods can effectively address these issues to achieve real-time identification of cotton aphid damage in the field. The Nvidia Jetson Xavier NX embedded edge computing device was selected for model deployment, as shown in
Figure 7, with the device parameters detailed in
Table 4.
The Nvidia Jetson Xavier NX is compact, powerful, and equipped with rich I/O interfaces. It supports a range of mainstream AI frameworks and algorithms, making it easy for developers to integrate models into the device for tasks such as image recognition, segmentation, and object detection in various computer vision applications [
30]. The device’s software environment consists of the Ubuntu 18.04 operating system, Python 3.6, CUDA 10.2, and TensorRT 8.2 for accelerating PyTorch model inference. OpenCV 4.1.1 is used with support for GPU acceleration provided by CUDA.
2.7. Evaluation Indicators
In object detection algorithms, the performance of network models is evaluated via several metrics: precision (P), recall (R), and mean average precision (mAP). Additionally, when lightweight models are considered, parameters, floating-point operations (FLOPs), frames per second (FPS), and the size of the optimal model weight file are also important indicators for evaluating model performance [
31]. The metrics
[email protected] and
[email protected]:0.95 represent the mAP values when the IoU threshold is 0.5 and when the IoU ranges from 0.5 to 0.95 with a step size of 0.05, respectively. A higher mAP indicates a higher overall accuracy of the model. FPS refers to the number of frames processed per second, which is primarily used to measure the detection speed of the model at the edge. A higher FPS value indicates that the model can identify and locate targets more quickly. The formula for each indicator is shown below:
where TP represents positive samples correctly predicted as positive samples, FP represents negative samples incorrectly predicted as positive samples, FN represents positive samples incorrectly predicted as negative samples, N represents the number of categories, Np represents the number of images, and Ti represents the detection time.
4. Discussion
The utilization of intelligent methods for pest and disease detection in farmlands is a critical component of achieving smart agriculture. Currently, there is extensive research on pest and disease detection or monitoring. To monitor pests and diseases over large areas, researchers have employed satellite remote sensing technology to establish monitoring models capable of long-term damage prediction before the onset of infestations [
44]. With the rapid development of drone technology, methods for monitoring farmland pests and diseases via drones equipped with multispectral and hyperspectral cameras have been extensively studied by scholars both domestically and internationally [
45]. Drone remote sensing has unique advantages for monitoring pests and diseases.
The aforementioned studies are both macroscopic, providing qualitative monitoring of pest and disease occurrence over large areas of farmland. This paper aims to utilize deep learning-based object detection algorithms to perform quantitative assessments of pest and disease infestations in farmland by annotating RGB images or videos captured by smartphones or drones. Advanced object detection models were trained on many labeled images, demonstrating high-precision recognition of pest and disease images. Currently, most related research is carried out in static scenes with traditional computing methods. Video detection at the edge can better improve the efficiency of pest and disease detection, which is more challenging and has more research value. Owing to the limited computational resources of edge mobile devices, running object detection models on such devices poses significant challenges, especially for real-time identification, which demands more computational resources. We are also working on improving the algorithm to develop a lightweight model. The complex background of farmland scenes and the occlusion of cotton leaves pose challenges for recognition. We aim to increase the detection performance by adding attention mechanisms and improving the backbone network.
On the basis of the above research, we developed the CVC-YOLO model for high-precision identification of cotton aphid damage in the field. The model is characterized by high accuracy and speed, enabling precise and rapid identification of leaf damage caused by cotton aphids. The size of the GVC-YOLO model is only 5.4 MB. In terms of detection accuracy, the
[email protected]:0.95 reached 90.3%. Finally, we deployed the proposed GVC-YOLO model on the edge computing device Jetson Xavier NX, achieving fast and accurate detection of aphid-damaged leaves at the edge.
We compared the GVC-YOLO model with both classic and state-of-the-art methods in the field. Two-stage methods, such as Faster R-CNN, are not suitable for deployment on edge mobile devices because of their slow detection speeds. In contrast, the YOLO series models feature a simple structure, making them easy to deploy across various platforms while demonstrating strong adaptability. By optimizing the network structure, we further reduce the model’s computational requirements, thereby increasing its detection speed on edge devices and enabling real-time performance. The deployed edge computing device is small and low-power and can be connected to cameras mounted on drones and spraying tractors, enabling real-time detection at the edge. By analyzing real-time detection results and integrating them with an automatic spraying control system, a precise pest control technology system can be established. This provides technical support for precise pesticide application, reducing pesticide waste and lowering agricultural costs.
Despite our efforts to conduct thorough and comprehensive research, our work has several shortcomings. First, although our improvements to the original model greatly reduced the complexity of the model and increased the speed of detection, the improvement in detection accuracy was not significant. This suggests that continued efforts are needed to further improve the accuracy of pest and disease detection. In addition, to fully demonstrate the benefits of the lightweight model, it needs to be tested and deployed on edge device platforms with more limited computational resources, such as the Jetson nano. This will also be the focus of research in our future work.
5. Conclusions
This study proposes a lightweight object detection model called GVC-YOLO to address prominent issues such as the inability to run high-complexity models efficiently in real time on edge devices with limited computational resources. On the basis of the YOLOv8n model, the model’s volume is significantly reduced by introducing the lightweight modules GSConv and VoVGSCSP, thereby reducing model complexity and accelerating detection speed. Additionally, the CA mechanism and SimSPPF network are incorporated to enhance the model’s detection capability for cotton aphid damage features and improve detection accuracy. To further validate the proposed model’s effectiveness, the lightweight GVC-YOLO model is optimized and deployed to the edge computing device Jetson Xavier NX via TensorRT and tested for detection analysis on video data captured in actual cotton field scenarios.
The main conclusions are as follows: the GVC-YOLO model achieves
[email protected] and
[email protected]:0.95 values of 97.9% and 90.3%, respectively, demonstrating high accuracy in recognizing cotton aphid damage and meeting practical detection requirements. Moreover, the parameters and FLOPs are 2.5 M and 6.8 G, respectively, with the optimal model weight size being only 5.4 MB. Compared with YOLOv8n, these figures represent reductions of 16.7%, 17.1%, and 14.3%, respectively, making it the most lightweight and efficient model of its kind for the detection of cotton aphid-damaged leaves. When deployed in actual cotton field scenarios, the GVC-YOLO model achieves a detection speed of 48 FPS with FP16 precision, meeting the real-time detection requirements at the edge.
The lightweight detection method proposed in this study can also be transferred and applied to real-time detection and analysis of various other crop pests and diseases, which is highly important for the mechanization and intelligent development of crop pest control. Edge computing devices can be installed on ground sprayers or low-altitude agricultural drones, and real-time pest information detected by the model can be displayed directly. Farmers can make decisions on the basis of detection results, control pesticide spraying in real time, and realize an integrated process from pest detection to prevention and control. This significantly reduces manual labor and achieves more precise and efficient development in the agricultural industry.