Figures
Abstract
This study proposes the S-YOLOv10-ASI algorithm to improve the accuracy of tea identification and harvesting by robots, integrating a slice-assisted super-reasoning technique. The algorithm improves the partial structure of the YOLOv10 network through space-to-depth convolution. The Progressive Feature Pyramid Network minimizes information loss during multi-stage transmission, enhances the saliency of key layers, resolves conflicts between objects, and improves the fusion of non-adjacent layers. Intersection over Union (IoU) is used to optimize the loss function calculation. The slice-assisted super-reasoning algorithm is integrated to improve the recognition ability of YOLOv10 network for long-distance and small-target tea. The experimental results demonstrate that when compared to YOLOv10, S-YOLOv10-ASI shows significant improvements across various metrics. Specifically, Bounding Box Regression Loss decreases by over 30% in the training set, while Classification Loss and Bounding Box Regression Loss drop by more than 60% in the validation set. Additionally, Distribution Focal Loss reduces by approximately 10%. Furthermore, Precision, Recall, and mAP have all increased by 7.1%, 6.69%, and 6.78% respectively. Moreover, the AP values for single bud, one bud and one leaf, and one bud and two leaves have seen improvements of 6.10%, 7.99%, and 8.28% respectively. The improved model effectively addresses challenges such as long-distance detection, small targets, and low resolution. It also offers high precision and recall, laying the foundation for the development of an Anji White Tea picking robot.
Citation: Yang C, Yuan W, Zhao Q, Wang Z, Song B, Dong X, et al. (2025) Identification of fresh leaves of Anji White Tea: S-YOLOv10-ASI algorithm fusing asymptotic feature pyra-mid network. PLoS One 20(7): e0325527. https://doi.org/10.1371/journal.pone.0325527
Editor: Yile Chen, Macau University of Science and Technology, MACAO
Received: January 4, 2025; Accepted: May 14, 2025; Published: July 2, 2025
Copyright: © 2025 Yang et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data for this study are publicly available from the IEEE DATAPort repository (http://doi.org/10.21227/eefx-a805).
Funding: This research was funded by the Major special projects of Yunnan Provincial Science and Tech-nology Department in 2023(202302AE090020), Key Projects of Basic Research Program of Yunnan Province in 2023 (202201AS070325) and Yunnan Provincial Science and Technology Department 2023 Rural Revitalization Science and Technology Special Project(202304Bl090013). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: No authors have competing interests.
Instruction
Anji White Tea [1,2], renowned both in China and internationally, is celebrated for its distinctive flavor and exceptional quality. The region’s unique natural conditions create an ideal environment for cultivation, but also present challenges for the mechanization and automation of tea harvesting. Traditional manual picking is not only labor-intensive and inefficient, but also easy to vary in quality due to manual operation. Moreover, recent years have seen an increasing problem with the aging of tea farmers and a shortage of labor, which has severely hampered the development of tea production. Therefore, applying mechanization and intelligent technologies to enhance the efficiency and quality of tea production has become crucial for improving tea picking processes.
With the advancement of agricultural science and technology, mechanized and intelligent systems have made significant strides, including developments such as tomato-picking robots [3,4] and intelligent irrigation systems. However, in tea production, especially in the tea picking process, the application of mechanization and intelligence is still in its infancy. Compared to robots for tomato picking, apple picking, and others [5,6], tea picking robots [7] face more stringent requirements for recognition and detection algorithms due to the dense growth of tea leaves, complex environments, and significant variations in lighting.
The advancement of deep learning technology [8,9] has significantly enhanced capabilities in target detection and image recognition tasks. YOLO (You Only Look Once) series network, as a representative of it, has become a research hotspot because of its efficient real-time processing ability and good detection accuracy. In 2023, our research team proposed the ShuffleNetv2-YOLOv5-Lite-E algorithm for Yunnan large-leaf tea tree picking [10]. The YOLOv5 network was deeply improved by network structure re-placement and channel pruning. A bud and two leaves recognition rate of 94.52% was achieved on edge devices. Lixue Zhu et al. [11] improved the YOLOv5 network through the ECANet (efficient channel attention network) module and the BiFPN (bidirectional feature pyramid network). The accuracy of tea bud detection reached 94.4%, and the recall rate reached 90.38%. Aiming at the problem of low detection accuracy caused by the complex background of tea, Shuang Xie et al. [12] by integrating deformable convolution, attention mechanism and improved spatial pyramid pooling, proposed an improved Tea-YOLOv8s algorithm. And the accuracy of tea bud detection reached 88.27%. Nevertheless, these methods fail to overcome real-world challenges including long-range detection, small targets, and low-resolution conditions.
To address these challenges, this study proposes an enhanced YOLOv10 architecture incorporating Space-to-Depth Convolution, which effectively preserves fine-grained details in long-distance and low-resolution imagery. The asymptotic feature pyramid network is used to optimize the YOLOv10 network framework, reduce the information loss of the network in multi-stage transmission, enhance the saliency of the key layer, alleviate the contradictory information of different objects, and optimize the fusion of non-adjacent layers. The Inner-IoU [13] optimization loss function calculation is used to improve the convergence speed of the model and enhance the universality of the model. The slice-assisted super-reasoning algorithm is integrated to improve the recognition ability of YOLOv10 network for long-distance and small-target tea. This study aims to develop a high-precision, robust visual recognition model specifically optimized for Anji white tea harvesting robots, with particular emphasis on accuracy and generalization capability across varying field conditions. Thereby laying the groundwork for the creation of specialized robots for picking Anji white tea.
Materials and methods
Image collection and dataset preparation
The datasets for this study were collected on-site in Anji County, Huzhou City, Zhejiang Province. The acquisition device uses Canon EOS R5 camera with RF24–105 mm lens to ensure high resolution and clarity of the original image. To improve the model’s detection accuracy in real-world conditions, this study simulated the distance between the tea-picking robot and the tea tree. The shooting distance ranged from 20 to 100 centimeters, while the distance between the robot and the ground varied from 100 to 160 centimeters. The ISO is set to 200, the aperture is f/4.0, and the shutter speed is 1/300. In order to ensure that the model can accurately identify tea leaves under different illumination, different slopes and different angles, and has stronger universality in practical applications, a multi-period and multi-angle image acquisition method is adopted. The resolution of the image is 5152 pixels (horizontal) × 3864 pixels (vertical), and the storage format is jpg. In this study, the tea posture is divided into side angle and positive angle according to the different position of tea. The image acquisition device is located in the middle of the robot, and the height can be freely adjusted. When the tea leaves grow on both sides of the field of view, they are considered to be in a side view. The robot, equipped with an occlusal end effector, must adjust the effector’s angle to approach the base of the tea leaves for harvesting. When the tea grows at the top of the field of view, the fresh leaves of the tea are considered to be a positive angle of view, and the occlusal end-effector can directly complete the picking, as shown in Fig 1.
To enhance the model’s generalization capability in complex and low-light conditions [14,15] and better capture tea features from different distances and angles, as shown in Fig 2, the collected data was added to improve the performance and robustness of the final detection model under real-world operating conditions. To improve the model’s recognition performance under diverse lighting conditions in the tea garden, the brightness of the images was randomly adjusted within a range of 0.5 to 1.5 times. Additionally, to enhance the model’s performance in fuzzy conditions, the image contrast was randomly varied from 0.4 to 1.8 times. The random range is mainly set based on the actual environmental variations in the tea garden. Furthermore, to strengthen the model’s ability to extract tea features at different distances and fields of view, the images were randomly scaled along both the x-axis and y-axis.
In this study, a total of 2265 original samples were collected, and 15855 images were enhanced. After screening, a total of 15000 images were selected. After annotating the image with Make Sense, it is randomly divided into training set, verification set and test set, with the number of 9000, 3000 and 3000 respectively. The data set is based on image brightness, image contrast and scaling, and the division results are shown in Table 1.
YOLOv10 network improvements
YOLOv10 is the latest version of the YOLO series [16,17]. It is a deep learning network for target detection. The network demonstrates notable efficiency in real-time processing while maintaining robust detection accuracy. Nevertheless, the YOLOv10 network exhibits two notable limitations: (1) significant detail degradation when processing low-resolution images, (2) suboptimal recognition performance for small tea targets at long distances. To address these issues, this study employs Space-to-Depth Convolution to enhance specific components of the YOLOv10 network, reducing the loss of detail in long-distance and low-resolution images. The Asymptotic Feature Pyramid Network (AFPN) is used to optimize the YOLOv10 framework. It reduces information loss during multi-stage transmission, improves the saliency of key layers, alleviates conflicting information between objects, and enhances the fusion of non-adjacent layers. The Inner-IoU optimization loss function calculation is used to improve the convergence speed of the model and enhance the universality of the model. The slice-assisted super-reasoning algorithm is integrated to improve the recognition ability of YOLOv10 network for long-distance and small-target tea. The improved YOLOv10 network is shown in Fig 3. The enhancements to the Space-to-Depth Convolution primarily target the layers subsequent to the second through fifth convolutional layers within the backbone. Regarding the advancements in the Asymptotic Feature Pyramid Network, the primary focus lies on the neck section. The Slicing Aided Hyper Inference Algorithm is positioned between the input and the backbone during the model’s detection phase.
Space-to-depth convolution enhancement
During deployment, tea-picking robots often capture blurred images [18–20] due to various factors, including high movement speed, mechanical vibrations, variable lighting conditions, and suboptimal focal length adjustments. The YOLOv10 network tends to lose significant detail when processing low-resolution images or distant tea leaves. To address these limitations, this study employs Space-to-Depth Convolution to enhance specific components of the YOLOv10 architecture. As shown in Fig 4, the Space-to-Depth Convolution is mainly composed of a SPD (Space-to-Depth) layer and a non-striding convolution layer. The SPD divides the input feature map into multiple sub-feature maps according to a certain stride, and stitches these sub-feature maps on the channel dimension, thereby reducing the space size but retaining all information. Non-strided convolution is applied to reduce the number of channels while retaining discriminative feature information.
As shown in Equations (1–3), where X represents the original feature map, and the size is .
represents the sub-feature map, which is generated by the proportional division of the original feature map, so each sub-feature map needs to down sample the original feature map. For example, when scale = 2, four subgraphs
,
, and
can be obtained, and the size is
. As shown in Equations 4 and 5,
represents the sub-feature map after linking along the channel dimension.
represents the transformed
, and a non-stepped convolutional layer with a
filter is added during the conversion process. The non-stepped convolutional layer is mainly used to preserve all discriminant feature information.
Asymptotic feature pyramid network improvement
During field operations, variations in the distance between the data acquisition device and tea leaves, along with inherent differences in leaf sizes, can lead to feature loss in the YOLOv10 network. This adversely affects the fusion of non-adjacent layers. To address these issues, this study introduces an enhanced Asymptotic Feature Pyramid Network (AFPN) to optimize the YOLOv10 architecture. This improvement enables effective fusion between non-adjacent layers by progressively incorporating higher-level features. In the refined YOLOv10 network, the initial stage focuses on integrating low-level features, which are rich in detailed information.
The high-level feature fusion occurs in the final stage, as illustrated in Fig 5. In this schematic, blue arrows represent convolution operations, yellow arrows indicate adaptive spatial fusion, purple arrows denote upsampling, and green arrows indicate downsampling. During this fusion process, the integration of low-level features with high-level features facilitates semantic information fusion [21,22]. In the process of asymptotic feature pyramid network, adaptive spatial fusion [23,24] is mainly used for multi-layer feature fusion. As shown in Equations (6) and (7), ASF (Adaptive Spatial Fusion) allocates spatial weights to enhance the saliency of key layers and reduce conflicting information between objects. represents the result feature vector,
,
and
represent the spatial weights of the three levels at the l level, respectively, and the adaptive spatial fusion of multi-level features is obtained.
represents the feature vector from n-level to l-level, and the position is
. In this way, the adaptive spatial fusion operation can effectively fuse feature vectors from different levels at each spatial position
to enhance the effect of feature extraction.
Inner-IoU improvement
While the YOLOv10 network’s loss function performs well in bounding box regression, it has limitations in adaptive adjustment and weak generalization when applied to different tea quality grades. In order to improve the detection performance of the YOLOv10 network and accelerate the bounding box regression, while enhancing the generalization ability of the model, this study further introduces Inner-IoU to optimize the YOLOv10 network, as shown in Fig 6. The Target Box and Anchor Box are represented in blue, while the Inner Target Box and Inner Anchor Box are shown in green. Inner-IoU evaluates the overlap between the predicted box and the ground truth by calculating the ratio of the green region to the yellow region. The center points of Inner Target Box and Target Box are c and , respectively. The center points of Anchor Box and Inner Anchor Box are c and
, respectively. The width and height are w and h, respectively. The improved YOLOv10 network uses a small auxiliary bounding box to accelerate convergence for samples with high IoU, and a large auxiliary bounding box to compute the loss for samples with low IoU.
To further improve the model’s generalization capability, this study introduces a scaling factor (ratio [0.5, 1.5]) to regulate auxiliary bounding box dimensions and loss computation, as detailed in Equations (8–15). When the ratio is less than 1, the auxiliary bounding box becomes proportionally smaller than the ground truth box. Although the effective range for regression is smaller than that of IoU, the gradient’s absolute value is higher than IoU, which accelerates the model’s convergence. Which is highly beneficial for accelerating the model’s convergence speed. When the ratio is greater than 1, the auxiliary bounding box can enhance the regression effect by expanding the effective range of regression.
Slicing aided hyper inference algorithm integration
Experimental results indicate that detection performance significantly declines beyond certain distance thresholds. For single buds and the “one bud, one leaf” configurations, both accuracy and recall rates decrease sharply when targets exceed 40 cm from the tea-picking robot. In contrast, the “one bud, two leaves” configuration shows similar decline beyond 60 cm. In the detection process for tea garden picking robots, accurately identifying and positioning long-distance and small targets can be challenging when relying solely on enhancements to the YOLOv10 network. To address this, this study combines the improved YOLOv10 network with the Slicing-aided Hyper Inference algorithm [25,26].This combination aims to enhance the visual recognition model’s ability to accurately detect and identify both long-distance and small targets. The Slicing Aided Hyper Inference Algorithm is integrated at the image input stage. As shown in Fig 7, the algorithm first divides the input image or video into multiple slices, with the number of slices customizable. It then performs target detection on each slice individually. The results are then merged back into the original image. During the slicing phase, the data is split into smaller chunks to reduce the computational and memory requirements for a single inference. In the hyper-inference phase, the algorithm processes each slice in parallel, significantly speeding up the overall operation. Finally, in the result fusion phase, the inference results from all slices are combined to ensure the accuracy of the final output. For overlapping detection boxes, the algorithm automatically merges them during the stitching process to enhance the accuracy of the recognition results. Additionally, by incorporating the Slicing-Aided Hyper Inference method, the model optimizes memory usage through image decomposition, significantly reducing hardware resource demands.
Comparison experimental design
To validate the detection performance of the enhanced YOLOv10 network, which integrates Space-to-Depth Convolution, Asymptotic Feature Pyramid Network, and Inner-IoU, this study designed a comprehensive experimental framework. The framework includes four comparative test groups focused on Anji white tea’s single buds, “one bud, one leaf,” and “one bud, two leaves” configurations. The experiments involved training and testing models using the improved YOLOv10 network, YOLOv10, RT-DETR, CornerNet, and SSD networks on the dataset. Model training was conducted on a Windows 10 system with a GPU. The host machine featured an i5 12600KF processor (10 cores, 16 threads), a Colorful B760M-T WIFI V20 motherboard, a 512GB Colorful CN600 M.2 NVMe SSD, and a Colorful RTX4060TI ULTRA W OC 16GB graphics card. The network development was carried out using Python 3.9 and Pycharm 2023. The number of worker threads for data loading was set to 8, the batch size was set to 64, and the input image size was uniformly and automatically adjusted to 640 × 640 pixels. The training process was set for 1000 epochs.
As shown in Equations (15) and (17), P represents Precision, R represents Recall, and F1 function represents the harmonic mean of Precision and Recall, where FP represents the number of recognition errors, and FN represents the number of undetected. mAP represents the average accuracy, and FLOPs represents the number of floating-point operations. Precision serves as a key metric for assessing a model’s predictive accuracy regarding positive class identification. The higher the Precision, the less the model misclassifies negative classes as positive classes. Recall is generally used to evaluate the model ‘s ability to cover positive samples. The higher the Recall, the lower the false negative rate of the model. F1 is the value after considering the accuracy and recall rate. When there is a big difference between the accuracy and recall rate, F1 can better measure the comprehensive performance of the model.
Results and analysis
Ablation study
To assess the contribution of each component in the enhanced YOLOv10 network, an ablation study was performed. Various versions of the network were tested by removing one component at a time and measuring the resulting performance differences. The test results are presented in Table 2.
The results indicate that the Asymptotic Feature Pyramid Network enhances the YOLOv10 network by improving Precision by 1.75%, Recall by 2.58%, and mAP by 2.45%. Additionally, the Space-to-Depth Convolution boosts Precision by 3.63%, Recall by 0.75%, and mAP by 3.18%. The Inner-IoU technique increases Precision by 2.32%, Recall by 0.45%, and mAP by 2.29%. After the overall improvement, the Precision, Re-call and mAP of the YOLOv10-ASI network are increased by 7.1%, 6.69% and 6.78% respectively compared with the original YOLOv10 network. Compared to the 94.52% mAP value of our team’s previous ShuffleNetv2-YOLOv5-Lite-E network, it shows an improvement of 2.4%. In order to further comprehensively evaluate the performance of the model, this study introduces the Gradient-weighted Class Activation Mapping heat map to visualize the model and judge the area of concern of the model. The GradCAM (Gradient-weighted Class Activation Mapping) heat map related to the ablation experiment is shown in Fig 8.
Model analysis
In machine learning and deep learning, Loss Function [27,28] is mainly used to measure the difference between the model detection value and the true value. The smaller the value, the closer the model detection result is to the target value. In the YOLOv10 network, the Loss Function primarily consists of Classification Loss, Bounding Box Regression Loss, and Distribution Focal Loss. Classification Loss evaluates the model’s accuracy in the classification task, while Bounding Box Regression Loss assesses the positional discrepancy between the predicted bounding box and the actual one. Distribution Focal Loss measures the divergence between the predicted distribution and the true distribution. As shown in Fig 9, the loss function of the YOLOv10-ASI network is significantly slowed down around 500 rounds, which is about 200 rounds earlier than the original YOLOv10 network. After 800 training rounds, the loss function of the YOLOv10-ASI network stabilized, achieving this milestone approximately 100 rounds earlier than the original YOLOv10 network. Specifically, Classification Loss, Bounding Box Regression Loss, and Distribution Focal Loss were ultimately maintained below 0.4, 0.4, and 0.9, respectively. Classification Loss and Distribution Focal Loss showed minimal variation in the training set compared to the original YOLOv10, while Bounding Box Regression Loss declined by over 30%. In the validation set, Classification Loss and Bounding Box Regression Loss each decreased by more than 60%, and Distribution Focal Loss experienced a reduction of around 10%.
Precision [29], Recall [30,31], and F1 [32,33], are commonly used to evaluate machine learning models and measure the performance of classification models. As shown in Fig 10, the YOLOv10-ASI network demonstrates a notable enhancement in Precision, Recall, and F1-Score compared to the original YOLOv10 network. Specifically, the improved YOLOv10 network boosts accuracy by 7.1 percentage points, recall by 6.69 percentage points, and the F1 by 6.78 percentage points.
Model comparison experiment
To comprehensively evaluate the performance advantages of the proposed YOLOv10-ASI network, this study conducted model training and testing using the same dataset for the YOLOv10 network, YOLOv10, RT-DETR, CornerNet, and SSD networks. The test results are presented in Table 3. The findings indicate that the improved YOLOv10-ASI outperformed YOLOv10, RT-DETR, CornerNet, and SSD in single bud recognition, with AP values increasing by 6.10%, 8.47%, 23.70%, and 19.97%, respectively. For the recognition of one bud and one leaf, the AP values rose by 7.99%, 11.71%, 25.48%, and 19.85%, respectively. When recognizing one bud and two leaves, the AP values improved by 8.28%, 10.06%, 23.65%, and 20.76%, respectively. Overall, the mAP increased by 7.44%, 10.06%, 24.28%, and 20.19%, respectively.
Additionally, in order to network combined with Slicing Aided Hyper Inference algorithm in abnormal conditions such as long distance, small target, low illumination and blur, this study performs additional external verification experiments. The external validation data were collected from the tea garden in Anji County, Huzhou City, Zhejiang Province, and some of the external validation results are shown in Fig 11. Experimental results demonstrate that both YOLOv10-ASI and the original YOLOv10 achieve the highest detection confidence among compared models, with YOLOv10-ASI exhibiting an average confidence score 5% higher than its baseline counterpart. In addition, for the long-distance and small target objects that cannot be detected by the YOLOv10-ASI network, the S-YOLOv10-ASI network after fusing the Slicing Aided Hyper Inference algorithm shows better detection results.
Discussion
In recent years, the aging rural population and the migration of young people to urban areas have led to a growing scarcity of labor resources for tea picking. Additionally, the tea picking process requires a high level of selectivity for the buds, and manual picking [7,34] can be impacted by factors such as fatigue or inexperience, which can compromise the quality of the tea. With the advancement of agricultural automation and intelligence, the development and application of tea picking robots have become particularly important. Aiming at the challenges of long distance, small target and low resolution encountered by traditional tea picking robots, this study uses Space-to-Depth Convolution to improve some structures of YOLOv10 network to reduce the loss of detail information of long distance and low resolution images.
The proposed framework incorporates an Asymptotic Feature Pyramid Network to address several limitations in YOLOv10: information attenuation across network stages, insufficient key feature emphasis, object feature conflicts, and suboptimal distant-layer feature fusion. At the same time, the Inner-IoU optimization loss function is used to improve the convergence speed of the model and enhance the universality of the model. Combined with the Slicing Aided Hyper Inference algorithm [35,36], the performance in the detection of long-distance and small targets is improved. The visual recognition model for the Anji white tea picking robot developed in this study offers robust technical support for the intelligent harvesting of Anji white tea, demonstrating significant application potential and value for promotion.
Furthermore, the proposed models and methodologies exhibit substantial potential for broad applications. Beyond facilitating the harvesting of diverse tea varieties, these techniques can be readily adapted for intelligent crop harvesting across agricultural domains. This research holds considerable theoretical and practical value in advancing the intelligent modernization of agricultural production [37,38] and achieving sustainable agricultural development.
Conclusion
In this study, the partial structure of YOLOv10 network is improved by Space-to-Depth Convolution, which reduces the loss of detail information of long-distance and low-resolution images. The YOLOv10 network framework is optimized by the asymptotic feature pyramid network, which enhances the saliency of the key layer, alleviates the contradictory information of different objects, and optimizes the fusion of non-adjacent layers. Through the calculation of Inner-IoU optimization loss function, the convergence speed of the model is improved, and the universality of the model is further enhanced. After fusing the slice-assisted hyper-inference algorithm, the S-YOLOv10-ASI network shows better detection results for long-distance and small target objects that cannot be detected by the YOLOv10-ASI network. The experimental results show that:
- (1) From the perspective of loss function trends, the Classification Loss, Bounding Box Regression Loss, and Distribution Focal Loss of the S-YOLOv10-ASI network eventually stabilized at values below 0.4, 0.4, and 0.9, respectively. While the Classification Loss and Distribution Focal Loss showed no significant changes in the training set compared to the original YOLOv10 network, the Bounding Box Regression Loss decreased by more than 30%. In the validation set, the Classification Loss and Bounding Box Regression Loss decreased by more than 60%, while the Distribution Focal Loss decreased by approximately 10%. This indicates a noticeable improvement in model stability. Additionally, the loss function of the YOLOv10-ASI network begins to stabilize significantly around 500 iterations, which is approximately 200 iterations earlier than in the original YOLOv10 network.
- (2) Model Parameters and Detection Performance: Compared to YOLOv10, RT-DETR, CornerNet, and SSD, the improved S-YOLOv10-ASI network showed AP improvements of 6.10%, 8.47%, 23.70%, and 19.97% for single bud recognition, respectively. For one bud with one leaf recognition, AP values improved by 7.99%, 11.71%, 25.48%, and 19.85%, respectively. For one bud with two leaves recognition, AP values improved by 8.28%, 10.06%, 23.65%, and 20.76%, respectively. The final mAP im-proved by 7.44%, 10.06%, 24.28%, and 20.19%, respectively. Additionally, Precision increased by 8.37%, 12.13%, 17.36%, and 13.10%, while Recall increased by 7.49%, 11.87%, 30.20%, and 27.38%. The S-YOLOv10-ASI network also demonstrated superior detection performance for long-distance and small-target objects that traditional networks could not detect.
References
- 1. Zeng C, Lin H, Liu Z, Liu Z. Analysis of young shoots of “Anji Baicha” (Camellia sinensis) at three developmental stages using nontargeted LC-MS-based metabolomics. J Food Sci. 2019;84(7):1746–57. pmid:31206686
- 2. Yan S-M, Hu Z-F, Wu C-X, Jin L, Chen G, Zeng X-Y, et al. Electronic tongue combined with chemometrics to provenance discrimination for a Green Tea (Anji-White Tea). J Food Quality. 2017;2017:1–6.
- 3. Gao G, Shuai C, Wang S, Ding T. Using improved YOLO V5s to recognize tomatoes in a continuous working environment. SIViP. 2024;18(5):4019–28.
- 4. Fukada K, Hara K, Cai J, Teruya D, Shimizu I, Kuriyama T, et al. An automatic tomato growth analysis system using YOLO transfer learning. Appl Sci. 2023;13(12):6880.
- 5. Liu J, Zhao G, Liu S, Liu Y, Yang H, Sun J, et al. New progress in intelligent picking: online detection of apple maturity and fruit diameter based on machine vision[J]. Agronomy. 2024;14(4).
- 6. Wang J, Su Y, Yao J, Liu M, Du Y, Wu X, et al. Apple rapid recognition and processing method based on an improved version of YOLOv5. Ecol Informatics. 2023;77:102196.
- 7. Wu L, Liu H, Ye C, Wu Y. Development of a premium tea-picking robot incorporating deep learning and computer vision for leaf detection. Appl Sci. 2024;14(13):5748.
- 8. Khan Z, Liu H, Shen Y, Zeng X. Deep learning improved YOLOv8 algorithm: Real-time precise instance segmentation of crown region orchard canopies in natural environment. Computers Electron Agric. 2024;224:109168.
- 9. Orka NA, Haque E, Uddin MN, Ahamed T. Nutrispace: a novel color space to enhance deep learning based early detection of cucurbits nutritional deficiency. Computers Electron Agric. 2024;225:109296.
- 10. Zhang S, Yang H, Yang C, Yuan W, Li X, Wang X, et al. Edge device detection of tea leaves with one bud and two leaves based on shuffleNetv2-YOLOv5-Lite-E. Agronomy. 2023;13(2):577.
- 11. Zhu L, Zhang Z, Lin G, Chen P, Li X, Zhang S. Detection and localization of tea bud based on improved YOLOv5s and 3D point cloud processing. Agronomy. 2023;13:2412.
- 12. Xie S, Sun H. Tea-YOLOv8s: a tea bud detection model based on deep learning and computer vision. Sensors (Basel). 2023;23(14):6576. pmid:37514870
- 13. Zhao B, Zhang Q, Liu Y. Detection method for rice seedling planting conditions based on image processing and an improved YOLOv8n model. Appl Sci. 2024;14(6).
- 14. Zhang T, Pan P, Zhang J. Steel surface defect detection algorithm based on improved YOLOv8n. Appl Sci. 2024;14(12):5325.
- 15. Huynh NB, Groendahl RA, Tomic O. Deep learning with uncertainty estimation for automatic tumor segmentation in PET/CT of head and neck cancers: impact of model complexity, image processing and augmentation. Biomed Physics Eng Express. 2024.
- 16. Li J, Gao T, Mi Q. Phasic parallel-network policy: a deep reinforcement learning framework based on action correlation. Computing. 2024:1–18.
- 17. Li J, Feng Y, Shao Y, Liu F. IDP-YOLOV9: improvement of object detection model in severe weather scenarios from drone perspective. Appl Sci. 2024;14(12):5277.
- 18. Wang J, Wang J. A lightweight YOLOv8 based on attention mechanism for mango pest and disease detection. J Real-Time Image Proc. 2024;21(4).
- 19. Li Y, Lin S, Ma J. A recognition algorithm applied to apple orchard picking robot. J Phys. 2024;2824(1):012001.
- 20. Fan P, Zheng C, Sun J, Chen D, Lang G, Li Y. Enhanced real-time target detection for picking robots using lightweight centernet in complex orchard environments. Agriculture. 2024;14(7):1059.
- 21. Qiu Y, Yang M. A deep vision sensing-based smart control method for apple-picking robots under the context of agricultural E-commerce. J Circuit Syst Comp. 2024;33(11).
- 22. Wu H, Wang Y, Zhou Y, Liu X, Zhai X, Sun K, et al. Tire defect detection based on low and high-level feature fusion. Meas Sci Technol. 2024;35(9):095401.
- 23. Kechen S, Han W, Ying Z. Lightweight multi-level feature difference fusion network for RGB-D-T salient object detection. J King Saud University - Computer Information Sci. 2023;35(8).
- 24. Chao Z, Xu W, Liu R, Cho H, Jia F. Surgical action detection based on path aggregation adaptive spatial network. Multimed Tools Appl. 2023;82(17):26971–86.
- 25. Zhou W, Shi Y, Huang X. Multi-view scene classification based on feature integration and evidence decision fusion. Remote Sensing. 2024;16(5):738.
- 26. Zhorif NN, Anandyto RK, Rusyadi AU, Irwansyah E. Implementation of Slicing Aided Hyper Inference (SAHI) in YOLOv8 to counting oil palm trees using high-resolution aerial imagery data. IJACSA. 2024;15(7).
- 27. Xinzhu Z, Guoxiang S, Naimin X. A method of modern standardized apple orchard flowering monitoring based on S-YOLO. Agriculture. 2023;13(2):380.
- 28. Kolli SC, Seelamanthula S, V R K V. Privacy enhanced course recommendations through deep learning in Federated Learning environments. Int J Information Technol. 2024:1–7.
- 29. Suo Y, He Z, Liu Y. Deep learning CS-ResNet-101 model for diabetic retinopathy classification. Biomed Signal Processing Control. 2024;97:106661.
- 30. Zhao Y, Lv W, Zhang Y, Tang M, Wang H. Enhanced data preprocessing with novel window function in Raman spectroscopy: Leveraging feature selection and machine learning for raspberry origin identification. Spectrochim Acta A Mol Biomol Spectrosc. 2024;323:124913. pmid:39126867
- 31. Guo W, Wang L, Zhu L, Ye Y, Zhang Z, Yang B, et al. Accelerated discovery and formation mechanism of high-entropy carbide ceramics using machine learning based on low-cost descriptors. J Alloys Compounds. 2024;1004:175929.
- 32. Hady AAD, Abd El-Hafeez T. Utilizing machine learning to analyze trunk movement patterns in women with postpartum low back pain. Sci Rep. 2024;14(1):18726. pmid:39134567
- 33. Dai T, Bao M, Zhang M, Wang Z, Tang J, Liu Z. A risk prediction model based on machine learning algorithm for parastomal hernia after permanent colostomy. BMC Med Inform Decis Mak. 2024;24(1):224. pmid:39118122
- 34. Kun L, Zhengmin W, Chengmao C. Biomechanical characterization of bionic mechanical harvesting of tea buds. Agriculture. 2022;12(9):1361.
- 35. Fan D, Yang Y, Feng S, Dai W, Liang B, Xiong J. SIPNet & SAHI: multiscale sunspot extraction for high-resolution full solar images. Appl Sci. 2023;14(1):7.
- 36. Wang X, He N, Hong C, Wang Q, Chen M. Improved YOLOX-X based UAV aerial photography object detection algorithm. Image Vision Computing. 2023;135:104697.
- 37. Qiu Y, Wang F, Zhang Z, Shi K, Song Y, Lu J, et al. Quantitative softness and texture bimodal haptic sensors for robotic clinical feature identification and intelligent picking. Sci Adv. 2024;10(30):eadp0348. pmid:39047112
- 38. Li C, Wu H, Zhang T. Lightweight network of multi-stage strawberry detection based on improved YOLOv7-Tiny. Agriculture. 2024;14(7):1132.