Figures
Abstract
With the accelerated development of the technological power of society, aerial images of drones gradually penetrated various industries. Due to the variable speed of drones, the captured images are shadowed, blurred, and obscured. Second, drones fly at varying altitudes, leading to changing target scales and making it difficult to detect and identify small targets. In order to solve the above problems, an improved ASG-YOLOv5 model is proposed in this paper. Firstly, this research proposes a dynamic contextual attention module, which uses feature scores to dynamically assign feature weights and output feature information through channel dimensions to improve the model’s attention to small target feature information and increase the network’s ability to extract contextual information; secondly, this research designs a spatial gating filtering multi-directional weighted fusion module, which uses spatial filtering and weighted bidirectional fusion in the multi-scale fusion stage to improve the characterization of weak targets, reduce the interference of redundant information, and better adapt to the detection of weak targets in images under unmanned aerial vehicle remote sensing aerial photography; meanwhile, using Normalized Wasserstein Distance and CIoU regression loss function, the similarity metric value of the regression frame is obtained by modeling the Gaussian distribution of the regression frame, which increases the smoothing of the positional difference of the small targets and solves the problem that the positional deviation of the small targets is very sensitive, so that the model’s detection accuracy of the small targets is effectively improved. This paper trains and tests the model on the VisDrone2021 and AI-TOD datasets. This study used the NWPU-RESISC dataset for visual detection validation. The experimental results show that ASG-YOLOv5 has a better detection effect in unmanned aerial vehicle remote sensing aerial images, and the frames per second (FPS) reaches 86, which meets the requirement of real-time small target detection, and it can be better adapted to the detection of the weak and small targets in the aerial image dataset, and ASG-YOLOv5 outperforms many existing target detection methods, and its detection accuracy reaches 21.1% mAP value. The mAP values are improved by 2.9% and 1.4%, respectively, compared with the YOLOv5 model. The project is available at https://github.com/woaini-shw/asg-yolov5.git.
Citation: Shi H, Yang W, Chen D, Wang M (2024) ASG-YOLOv5: Improved YOLOv5 unmanned aerial vehicle remote sensing aerial images scenario for small object detection based on attention and spatial gating. PLoS ONE 19(6): e0298698. https://doi.org/10.1371/journal.pone.0298698
Editor: Rajmohan Rajendran, VIT-AP Campus, INDIA
Received: August 3, 2023; Accepted: January 29, 2024; Published: June 3, 2024
Copyright: © 2024 Shi et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the manuscript and its Supporting information files.
Funding: Funding provided by: [1] The National Natural Science Foundation of China: "Research on Early Discovery and Situation Awareness of Multilingual Network Public Opinion Events"; Project No.: 202204120017; [2] Special research and development task of the autonomous region: "Research and development of key technologies for edge computing data security in digital monitoring scenarios"; Project No.: 2022B01008-2; [3] Major science and technology projects of the autonomous region: "Research on common key technologies of blockchain"; Project No.: 2020A02001-1; [4] Optimization of low resolution equipment defect recognition algorithm based on image enhancement; Project No.: SGXJXTOOJFJS2200076; [5] Research Project of "Tianshan Talents": "Research and Application of Multilingual and Multimodal Information Content Security"; Project No.: 202304120002; [6] Research on Small Target Detection Methods for UAV Scenarios; Project No.: XJ2023G097.
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
With the rapid development of AI technology and deep learning, small target detection has progressed rapidly [1, 2], and models in the direction of small target detection have emerged. Small targets of different scales in images under aerial photography of unmanned aerial vehicles (UAV) have problems such as difficult recognition and multi-scale detection, category ambiguity, etc. In order to solve these problems, a small target detection model is introduced. UAV remote sensing aerial photography is widely used in road monitoring targets, photography, life images, agricultural development, and other kinds of applications; for example, UAVs remote sensing shoot pedestrians and vehicles on the road, and among the vehicles, there are large vehicles and small vehicles that need to be identified and detected in real-time. This research uses a small target detection model under deep learning to identify and analyze them through the real-time images captured by the UAV. The drone flight height produces pictures of target scales and the kinds of targets at different scales. There are no particularly weak targets in general images [3]. At the same time, there are different kinds of weak targets in aerial images, and the model completes the work of artificial intelligence through computational analysis. The detection of small targets in UAV remote sensing aerial images has the following three main problems: 1) UAV remote sensing aerial images need to be detected in real-time, and the images taken at high altitudes will have particularly weak targets that are easy to detect when there is a lack of targets to be detected and wrong detection targets; 2) there are a large number of different kinds of targets in the UAV remote sensing aerial photography images, which need to be divided into different kinds and detect different types of weak targets, some of which tend to overlap and be covered up by some unimportant or important kinds, which will cause particular problems for detection and identification; 3) the UAV remote sensing aerial photography images will appear at night and during the day, as well as various environmental factors such as lighting, making it difficult to identify and detect weak targets as well. The above three problems can be seen through the image comparison in Fig 1.
a) the problem of target occlusion in the images taken by the UAV; b) the problem of target recognition arising from the images taken by the UAV in different environments; c) the factors such as blurring and exposure of the images taken by the UAV at different speeds of motion affect target detection; d) the problems of different scales and many dense types of targets arising from images taken by UAVs flying at different altitudes. Fig 1 is attributed to the NWPU-RESISC database and are available from the NWPU-RESISC database (url(s) https://tensorflow.google.cn/datasets/catalog/resisc45).
To address the above problems, this paper proposes the ASG-YOLOv5 model, designs the dynamic contextual attention module to solve the influence of dense scenes and blurred exposure scenes on small target detection and improve the model’s attention to small targets, and excessive space gating bidirectional filtering modules are designed to solve the impact of target occlusion and excessive redundant target interference on small target detection; The Normalized Wasserstein Distance [4] and CIoU [5] regression loss function are used to measure the similarity between the regression frames better and optimize the smoothing of the small target localization difference, which improves the accuracy and effectiveness of the model for small target detection. The model was extensively experimented on VisDrone2021 dataset [6] and AI-TOD dataset [7] and achieved better results than other models. In VisDrone2021-DET-val dataset and AI-TOD-test dataset, the mAP values reached 21.1% and 26.1%, respectively. A visual comparison was performed in the NWPU-RESISC dataset [8] and better detection results were obtained. The ASG-YOLOv5 model is more effective in detecting weak targets under UAV remote sensing aerial photography, and it can recognize more categories of weak targets and images of different scales.
The contribution of this paper is specified as follows.
- A dynamic contextual attention module (DCA) is proposed to filter redundant information using information bottlenecks, dynamically assign weights by target feature scores to enhance the extraction of global information, and output feature context information through channel dimensions to enable the network to focus better and extract useful target features.
- A spatially gated filtering multi-directional weighted fusion module (SGM) is proposed to enhance the characterization of small target categories by suppressing large target categories through spatially gated units, introducing global attentional up-sampling to increase the small target feature information in the underlying feature layer, and performing a weighted fusion of feature layers at different scales to enhance the characterization of small target categories.
- This research optimizes the regression loss function of the model by introducing Normalized Wasserstein Distance loss, which is combined with the CIoU loss function to increase the smoothness of the positional difference of the target frames, increase the similarity between the small target categories, and ultimately improve the model’s detection accuracy and effectiveness for small targets in UAV scenarios.
- On the unmanned aerial vehicles remote sensing datasets VisDrone2021-DET-val and AI-TOD-test, the ASG-YOLOv5 model designed in this paper achieves 21.1% mAP and 26.1% mAP values, respectively, which significantly improves the small target detection capability by 2.9% and 1.4% mAP values, respectively, compared with the baseline YOLOv5 model. A visual comparison was performed in the NWPU-RESISC dataset and better detection results were obtained. At the same time, the frames per second (FPS) reach 86, which meets the requirement of real-time small target detection in UAV remote sensing scenarios and has better accuracy and speed advantages than most UAV remote sensing target detection models.
The remaining chapters of this thesis are organized as follows: Section 2 describes the related work used to design the model used in this paper and its details. Section 3 presents the overall structure of the ASG-YOLOv5 model, the main modules designed, and the loss function. Section 4 presents the details of the experiments done by the model on the two UAV datasets, the comparison experiments, and the ablation experiments. Finally, Section 5 summarizes the results of the methodology study presented in this paper, methodological limitations, and the outlook of the next research.
2 Related work
As small target detection methods have evolved, their models’ accuracy in detecting small targets has gradually improved. However, the UAV remote sensing shooting scene has many target objects, and detection is still difficult. Hence, this paper introduces a redesigned model structure for small target detection that incorporates multi-scale feature fusion and data enhancement techniques for input images. These methods have improved the accuracy and effectiveness of small target detection in UAV scenes.
Small Target Detection. The initial violent approach to target detection is a sliding window in different directions, using classification to identify targets. Windows of different sizes and aspect ratios detect different target types at different observation distances. The existing models for small target detection can be divided into one-stage and two-stage models, and in 2014, Ross Girshick proposed R-CNN [9], the initial two-stage target detection model. R-CNN is a model for object detection based on a deep neural network, which has attracted the world’s attention for its excellent performance at that time. The R-CNN is a model for object detection based on deep neural networks, and it has been noted for its excellent performance at that time. It still uses selective search to select 2000 suggestion boxes; due to the different sizes of suggestion boxes, the obtained feature boxes need to be transformed to the same size by the ROI Pooling layer. Fast R-CNN [10] classifies and predicts the position size of boxes output by the convolutional neural network. The network finally uses SVD instead of a fully connected layer to improve the computational speed. In 2015, Kaiming He, Ross Girshick, et al. proposed the famous Faster R-CNN [11] algorithm, a breakthrough in proposing RPN (Region Proposal Network) network, using an anchor mechanism to link region generation with the convolutional network, abandoning the selective search. The algorithm has improved the detection speed to 17FPS (Frames Per Second) and achieved 70.4% detection results on the VOC2012 test set.
The one-stage models started with Wei Liu’s presentation of SSD at ECCV 2016 [12] and YOLO [13] presented by Joseph Redmon at the University of Washington in 2016. YOLO has significantly improved small target detection from the original YOLOv1 to YOLOv5 [14]. The main contribution of YOLO is the development of a real-time high-performance target detection framework for one-stage detection, which predicts the location and class of targets in an image by simply feeding the image into the network at once. In contrast to YOLOV4 [15], proposed by Glenn Jocher in 2020, YOLOv5 enhances each batch of training data by passing it through a data loader. The data loader performs three types of data enhancement: scaling, color space adjustment, and mosaic enhancement. YOLOv5 and YOLOV4 use CSPDarknet as the backbone to extract rich target feature information from the input image. CSPNet solves the problem of duplicating gradient information for network optimization in the backbone of other large convolutional neural network frameworks. This is done by integrating the gradient changes into the feature map from beginning to end, reducing the number of parameters and FLOPS values of the model, ensuring both inference speed and accuracy, and reducing the model size [15]. For the aerial pictures taken by UAVs, the small target category is more difficult to identify. Hence, this paper chooses to use the faster YOLOv5 model as the base model, adding an attention mechanism and gating unit to filter and detect small targets in the UAV view, which can detect weak targets in more real-time and more accurately. Eventually, there was a significant improvement in accuracy and mAP for the improved YOLOv5 model in several experiments.
Feature fusion and enhancement at different scales. Multi-scale takes sample information of signals with different granularity and effectively improves the detection of weak targets by fusion and enhancement of multi-scale features. In the field of target detection, the SSD (Single Shot MultiBox Detector) [12] network proposed by Wei Liu in 2016 is the initial network for target detection direction using different scale feature fusion. Later, the FPN proposed in 2017 [16] led to a significant improvement in the fusion of scale features in the small target detection direction, and there have been researchers who have continuously improved the FPN since then. The PAFPN [17] structure used in the subsequent YOLOv5 [14] better transfers the target information from different feature layers, from top to bottom and from bottom to top. The subsequent BiFPN [18] structure saves the connection transfer between some of these feature layers and introduces weights to learn the importance between the different feature layers of the input. The subsequent AF-FPN [19] structure, proposed in 2021, uses the adaptive attention module (AAM) to reduce the absence of contextual information in the high-level feature maps and the feature enhancement module (FEM) to enhance the representational power of the pyramid structure. The network mentioned above structures are relatively effective in fusing and enhancing features at different scales but need more simultaneous fusion and enhancement of feature information, and the communication between feature layers is complicated, which can easily disrupt the model’s focus on small target detection.
In this paper, this research improves FPN by adding a new spatial gated filtering multi-directional weighted fusion module to provide a better-weighted fusion of features at different scales. At the same time, by filtering out the redundant information of other feature layers and increasing the small-target pixel information of the underlying features, the model pays more attention to the information of the weak and small targets to enhance the characterization ability of the small-target features.
Data Enhancement. The aerial images taken by drones as the dataset used for experiments will have the problems of a small number of small targets in the images and a small percentage of tiny target labeled areas, resulting in uneven distribution of small targets in the images, and therefore poor generalization ability in terms of location. In the actual test images, however, small targets are likely to appear in areas that did not appear in the training set [20] and are difficult to detect. The images contain a large number of background categories. At the same time, data enhancement enables the dataset to be extended to increase the effect of the model in detecting small targets, which essentially solves some of the problems contained in the original dataset and improves the robustness of the model to different data. So, this research chose to perform data augmentation on the dataset. Now, there are many processing methods; researchers use Copy-Pasting [21] (Copy-Pasting strategy), meaning that the small target is pasted to any location in the image. A new annotation is generated, and the pasted small target can be randomly transformed (scaled, folded, rotated, etc.). This process ensures that the Context in which the small target is located is appropriate [20]. In this way, by increasing the number of small targets in each image, the number of matched anchors increases, which improves the contribution of small targets to the loss calculation in the training phase. Some researchers use Cutouts [22] to randomly select a square region of fixed size and then use all-0 padding. Meanwhile, Mosaic [15] improved on CutMix [23] by using four images for splicing while randomly selecting a region as the region of interest and splicing it, and passing the newly synthesized image into the neural network for learning as a way to improve the robustness and generalization of the model.
In this paper, this research uses Mosaic, Cutout, Copy-Past, and traditional data enhancement methods to simultaneously perform data enhancement on the dataset. Target detection dataset once the image data is enhanced (target box is changed), the labels need to be modified accordingly. These three methods are chosen because they can be easily embedded in the training framework of this research, help increase the model’s detection effect in scenarios with small targets and lack of target background, and are an excellent way to solve the unbalanced data set with small numbers and a proportion of samples.
3 Method
In this section, this research provided a thorough description of the compositional architecture of the ASG-YOLOv5 method designed in this paper. It elaborated on the details of the custom-designed Dynamic Contextual Attention Module (DCA), Spatial Gating Filtering Multi-directional Weighted Fusion Module (SGM), and NWD-CIoU Loss regression loss function. This research effectively improved the overall model performance for detecting small targets in UAV scenarios by leveraging the modules designed in this paper and network structure.
3.1 ASG-YOLOv5 model
In this paper, based on the design of Dynamic Contextual Attention Module (DCA) and Spatial Gating Filtering Multi-directional Weighted Fusion Module (SGM), this research incorporates the Dynamic Contextual Attention Module at the end of the backbone network, and at the end of the Spatial Gating Filtering Multi-directional Weighted Fusion Module (SGM) module is introduced in PAFPN [17]. First, the input feature maps are extracted from different scales by the DCA module and SPPF module at the backbone and end of the network to improve the network’s focus on useful target features and output the P2, P3, P4, and P5 feature layers. The detailed structure of spatial pyramid pooling—fast (SPPF) module is shown in Fig 2. The feature layers are up-sampled to obtain C5, C4, C3, and C2 feature layers. The Pi(i = 2,3,4) and Pi+1 feature layers extracted from the backbone network are filtered and localized to the small target category through the fusion of SGM module using spatial gating unit (SG) and BiFPN-Concat [18] in order to filter out the redundant feature-target information. Finally, the filtered feature layers are fused with Ci(i = 2,3,4) to obtain the final output feature layer Fi(i = 2,3,4). Finally, the output four-layer feature layer is used as the head for predicting weak and small targets. The overall structure and details of the relevant modules of the ASG-YOLOv5 model proposed in this paper are shown in Fig 3.
The CBS module denotes the convolution, the BN layer, and the SiLU activation function.
(a) ASG-YOLOv5 model structure. (b) C3 module. (c) Consists of a C3 layer, BiFPN fusion, upsampling operation, and a convolutional layer. (d) Consists of a convolutional layer, BiFPN fusion, and a C3 layer. Adding DCA module at the end with CSPDarknet53 of base model YOLOv5 as the backbone; introducing a new spatially gated filtered multi-directionally weighted fusion module, SGM, with PAFPN similar to base model YOLOv5 as the neck module.
3.2 Dynamic Contextual Attention module (DCA)
Attention mechanisms are widely used in the field of vision, such as Vision Transformer (VIT) [24] and Swin-Transformer [25, 26], whose methods aim to make the network pay better attention to useful feature target information. There is very little small target information in the UAV remote sensing shooting dataset. The network has a low ability to pay attention, and the existing attention assigns weights from the channel and spatial dimensions to improve the overall attentional effect of the model. The existing attention mechanism cannot fully utilize the connection to global information in the feature layer, and the computational volume and parameters are huge, making the overall inference speed of the model slow. Based on the above problems, this research proposes a new dynamic contextual attention module—DCA, whose structure is shown in Fig 4, which firstly utilizes feature similarity scores to assign feature information weights dynamically, integrates global contextual feature information, in which this research adds an information bottleneck structure, reduces redundant information in the output features, and retains the most relevant global information of the input and output features, which reduces the impact of the introduced redundant information. The module reduces the introduction of additional parameters and network inference speed. Moreover, this research uses the separable depth convolution to fuse the local feature information to get the fused contextual feature information. As shown by Eq 1.
(1)
where x denotes the input feature, τ and φ denote the learnable scalar, respectively, S denotes the target feature information score, and α and β denote the mean and standard deviation of the target feature information score, respectively.
Input feature x goes through a global contextual information extraction structure containing feature similarity score and information bottleneck structure to get global information; the global information is fused with the original input feature information, and the weights of different channels are recalibrated to adjust the channel dependency, and the obtained feature weights of different channels are multiplicatively weighted and fused with the input features.
Inspired by the Squeeze-and-Excitation Networks (SENet) [27, 28] network model shown in Fig 5, this research improves the generalization ability of the module by using channel attention to compute the weights of the fused contextual feature information on the channel dimension. The channel attention extracts the fused global information and input feature information using depth separable convolution (DSConv), which has a relatively low number of parameters and computational cost, and after the normalization layer and ReLU activation function, performs the weight normalization operation on the extracted fused features and multiply fuses the normalized weights with the features of different location information, as shown in Eq 2.
The input feature x is residual-connected, and a part of it is dimensionally compressed using global average pooling. It goes through two fully connected layers to predict each channel and get the importance of different channels. Then, a normalization operation is performed using the sigmoid activation function, and then feature fusion is performed with the other part of the input features.
This research adds the DCA module to the end of the backbone network of YOLOv5 [14] to ensure the inference time is manageable and the number of model parameters is manageable. In the feature extraction stage, the DCA module makes the backbone network more effective in extracting the feature information of the target by fully extracting the fusion context information and performing inter-channel computation on the fused information, which improves the ability of the network to focus on useful target features.
3.3 Spatial Gating Filtering Multi-directional Weighted Fusion Module (SGM)
In the UAV remote sensing aerial images, weak and small targets are mixed with large targets, which are difficult to recognize. The input image of the backbone network will get different feature layers, and different feature layers pay different attention to the details of small targets; the P5 feature layer, as a deep layer, exists more target semantic information and pays more attention to the details of the features of the target to be detected, and the P2 feature layer, as a shallow layer, has particularly low pixels of target information and pays more attention to the overall features of the target to be detected. At the same time, each feature layer contains different size targets, and the fusion of feature layers of different scales will produce redundant feature target information. In order to solve the above problems, this research designs a spatial gated filtering multi-way weighted fusion module (SGM), which filters out the interference of large targets and redundant information through spatial gated filtering units, enhances the small target pixel information of the underlying feature layer by global attentional up-sampling, and fuses the two neighboring feature layers in a weighted manner, so that the model can better extract the useful target feature information and focus on detecting the small target features. The main structural details are shown in Fig 6.
Simupsample is Content Aware Reassembly of Features (CARAFE), the CBS module consists of Conv layer, BN layer, Silu layer, SG is a spatial gating unit, GAU denotes Global Attention UpSample module, and the BiFPN-Concat module is the weighted feature fusion module.
The input Pi+1 (i = 2, 3, 4) feature layer is first subjected to a lightweight, universal upsampling operation: Content Aware Reassembly of Features (CARAFE) [29]. The upsampling kernel prediction module and feature recombination module are used respectively, introducing only a small amount of parameters and computing costs while obtaining a larger Receptive field to aggregate the context information of features. The feature information of this layer is then transformed and extracted through the CBS module, which consists of a Conv layer, a BN (batch normalization) layer, and an activation function Silu layer. Next, after the spatial gating unit SG, before multiplying the input features pixel by pixel, the ordinary convolution is replaced by using the dilated convolution [30, 31], which captures a larger range of contextual information by adjusting the dilation rate while keeping the convolution kernel size unchanged. This enables the dilated convolution to capture a broader range of semantic associations and better understand the global structure and contextual information in the image. Also by adjusting the dilation rate, a larger sensory field can be obtained without introducing additional parameters, thus reducing the number of parameters and computational complexity of the model to some extent. The use of multiple dilated convolutional layers to enrich the receptive field, as well as the final normalization operation using the ReLU activation function, can dynamically adjust the weights of the target features in the feature map in order to better capture the spatial information of the target object. Then, a point-by-point multiplication operation is performed with the Pi (i = 2,3,4) feature layer to increase the representation of small target features. The function obtained from the previous module is then subtracted point-by-point with the Pi(i = 2,3,4) feature layer to eliminate large target features and filter small target features. The output feature layer is then fused with the shallow feature layer using the Global Attention Upsample (GAU) [32] module to increase the small target pixel information of the underlying feature layer. The detailed structure diagram is shown in Fig 7. There is more small target semantic information in the deep feature layer; this research extracted the global context feature information from the new feature layer obtained after point-by-point subtraction by using 1×1 convolutional layer and Softmax function normalization operation and performing 3×3 convolutional processing with channel dimension on the shallow Pi (i = 2,3,4) feature layer. The global context information features of the deep layer are weighted and fused with the shallow features in the channel dimension. Finally, the deep feature layer is up-sampled and fused with the weighted shallow features to output the new feature layer information. The small target semantic information of the shallow feature layer is fused into the shallow feature layer to enhance the small target feature information so that the model can locate the small target information and valuable information more accurately when extracting features. Finally, the new feature layer is output with Ci(i = 2,3,4) feature layer by BiFPN-Concat [18]; it contains the connection of multi-directional different scale feature layers and the fusion of fast normalization. BiFPN-Concat adds weight coefficients on top of the concat and normalizes the weights to fuse different numbers of branching feature layers separately to fuse better feature layers of different scales and better focus on small targets.
Pi denotes the shallow feature layer, and Pi+1 denotes the deep feature layer.
3.4 Normalized Wasserstein Distance- Complete IoU Loss (NWD-CIoU)
Since small targets contain only a few pixel sizes, and at the same time, small target features lack appearance information, it is difficult for state-of-the-art target detection models to localize and identify small targets accurately. In the dataset captured in the UAV scenario, the small target category accounts for most of the data, making the existing model less accurate and less effective for target category detection. The existing IoU metrics in YOLOv5 are extremely sensitive to the positional difference of the small targets, which also leads to the poor accuracy of the small target detection. Therefore, in order to improve the model’s detection accuracy and effect for small targets, this research introduced Normalized Wasserstein Distance (NWD) [4], which is used as an evaluation method for small target detection through Wasserstein distance, and combined with CIoU [5] to optimize the existing loss function. Firstly, this research calculated the center distance, overlap rate, and aspect ratio between the target and the anchor frame in order to make the anchor frame regression loss more stable. The specific formula is expressed as follows.
(3)
(4)
(5)
where ρ2(b, bgt) denotes the Euclidean distance between the centroids of the predicted and real frames, respectively, c denotes the diagonal distance of the minimum closure region, α denotes the equilibrium coefficient, ν denotes the aspect ratio of the anchor frame, w and h denotes the length and width of the anchor frame.
Meanwhile, this research modeled the 2D Gaussian distribution of the BBox box to more fully show the weights of different positional feature information, calculate the distance metric for the positional information of the two target boxes, and finally normalize to get the similarity metric value between 0–1. The formulas are expressed as shown in Eqs 6 and 7.
(6)
Where
and
the Gaussian distribution modeling the two target frames is denoted as (Na, Nb),
denotes the difference measure between the target frames.
(7)
Where C denotes the value of the constant required to be associated with the data in the dataset.
This research did equal weight coefficients of the obtained metrics with the CIoU regression loss and added them as the final regression loss function value. As a result, this research proposed the NWD-CIoU Loss, whose formulaic expression is shown in Eq 8.
(8)
where α, β denote the equilibrium coefficients of the regulation loss function.
This research optimized the loss function by Normalized Wasserstein Distance, which ensures the smoothness of the model for the positional differences and, at the same time, increases the similarity between the two target frames. Finally, the detection accuracy and effect of the model for small target features are effectively improved.
4 Results and discussion
This section demonstrated the effectiveness of the ASG-YOLOv5 method proposed in this paper by performing a comprehensive series of comparisons on the VisDrone2021 and AI-TOD datasets, and ablation experiments on the VisDrone2021-DET-val. First, this research described the VisDrone2021 and AI-TOD datasets used for the experiments, the parameters used for model training and testing, and the evaluation metrics used for validation. Subsequently, this research compared the validation results and visualization plots of the current state-of-the-art target detection methods on the VisDrone2021-DET-val and AI-TOD-test datasets. Finally, this research performed ablation studies for each of the proposed modules.
4.1 Experimental details
In this paper, this research used NVIDIA A40 GPUs (48G of video memory) for training and testing; the model code is written under Python 3.6 to implement the ASG-YOLOv5 model for target detection in UAV remote sensing scenarios based on the Pytorch 1.10 deep learning framework. The ASG-YOLOv5 model designed in this paper has many of the same parts as YOLOv5s, such as the backbone and neck, so the model also uses the weights of YOLOv5s. This research used 640×640 pixels for the input image size, set the batch size to 64, and set the epochs to 300. The training loss of the optimizer decreases faster than the SGD [33]. The initial learning rate lr0 is set to 0.005, and the periodic learning rate lrf is set to 0.1.
4.2 Experimental dataset
In this experiment, this research used the training and validation sets from the VisDrone2021 dataset [6] for training and testing, which contains ten categories, 6,471 images for the training set, and 1,610 images captured by different UAVs for the validation set, which cover a variety of aspects, such as different cities and their environments, people, and scene densities, etc. This research also used the AI-TOD dataset [7] for the training and test sets, which contain eight categories, 11214 aerial images from the training set and 14018 aerial images from the test set covering weak target objects such as cars, airplanes, pedestrians, etc. The pixels of the image are 12.8 pixels, much smaller than those of other aviation scene images. Also, this study uses the NWPU-RESISC dataset [8] for target detection visualization comparison. After the experiments, this research can see that the model designed in this paper better detects weak targets in UAV remote sensing aerial images and can better focus on and identify weak targets. During the experiment, this research used data enhancement to expand the data set to improve the robustness of the network. First, this research used traditional data enhancement to scale and rotate the image to crop and change the color gamut for adjustment. This research then used Cutout [22] randomly selected regions for cropping, Mosaic [15] method to perform a four-image puzzle, and the Copy-Paste [21] method to paste and copy the images to fuse them. After data enhancement, this research finally got good test results.
4.3 Evaluation metrics
This research chose mAP (average of all 10 IoU thresholds, ranging from [0.5: 0.95] for all categories of AP calculations), AP50, and AP75 as the judging results. mAP (APval) was used to measure the overall accuracy rate.
AP50 calculates the average accuracy of all images in each category when IoU is set to 0.5 (accuracy at different recall points is averaged) and then averages over all categories. The expression can be written as
(9)
where r denotes recall and p denotes accuracy. mAP is the result of taking the average accuracy value between 0.5 and 0.95 for IoU and then taking the average value, and the expression can be written as
(10)
where K denotes the category.
Moreover, this research uses APvt, APt, APS, APL, and APM as the detection accuracy values of the model for minute, tiny, small, medium, and large targets; APS denotes average precision (AP) measurements for target frames with pixel areas less than 322; APM denotes AP measurements for target frames with pixel areas between 322 and 962; APL denotes AP measurements for target frames with pixel areas greater than 962. Since the AI-TOD dataset has minimal, tiny, and small targets, this research includes APvt and APt as the detection accuracy values for minute and tiny targets in evaluating the model using the AI-TOD dataset. APvt denotes the target detection accuracy value for pixels between 2 and 8, and APt denotes the target detection accuracy value for pixels between 8 and 16. The size and detection speed of the model are compared using the number of parameters and frames per second (FPS) metrics.
4.4 Experimental comparison with other models
4.4.1. Experiments with the VisDrone2021 dataset.
This research has done experimental validation of several existing better small target detection models on the VisDrone2021-DET-val dataset [6], as shown in Table 1. It can be seen that the ASG-YOLOv5 model designed in this paper is higher than the mAP value of the YOLOv5 model by 2.9%. The accuracy for small target detection is higher than that of the YOLOv5 model by 6.3% of the APS value. The accuracy of the YOLOv5 model is higher by 4.7% and 6.6% for medium and large targets. The ASG-YOLOv5 model has a 1.4% higher APval value than the YOLOv7 model and a 5.3% higher APS value than the YOLOv7 model for detection accuracy of small targets. At the same time, this research adds the designed modules into YOLOv7 and YOLOv8 [34]. As shown in Table 1, the detection accuracy of the model designed in this paper is higher than that of ASG-YOLOv7 by 1.3% mAP, although the detection accuracy is slightly higher than that of ASG-YOLOv8, as shown in Table 2, the computation, parameter counts, and speeds of the model designed in this paper are better than those of ASG-YOLOv8, thus proving that the model designed in this paper well balances the speed and detection accuracy and has a large performance advantage. As shown in Table 2, the computational and parametric quantities of the model designed in this paper are relatively low. The FPS value can reach 86, which meets the real-time demand of deploying UAVs for small-target detection. In contrast, the accuracy of the model designed in this paper is much higher than that of the YOLOv7-tiny model, which reflects the balance of small-target detection accuracy and speed of the model designed in UAV scenarios, demonstrating the performance advantage.
In order to realize the real-time performance of the small target detection model in the UAV scenario, this research designed a lightweight module, but there is also a certain number of parameters, which makes the model as a whole increase a certain amount of computational complexity; at the same time, the accuracy of the small target detection in two UAV scenarios has been significantly improved, and shows the advantage of the detection accuracy in the comparison of existing small target detection models, so this research has a better balance between detection real-time speed and detection accuracy. Although the added modules have a certain number of parameters, the accuracy is well improved, which reflects the advantage of the overall performance of the model designed in this paper.
Figs 8–10 display the confusion matrices for the YOLOv5, YOLOv7 models, and ASG-YOLOv5 model, respectively, detecting ten different categories. After being trained, comparing these matrices reveals that the ASG-YOLOv5 model can accurately detect all ten categories in UAV images with higher precision than the YOLOv5 and YOLOv7 models. The ASG-YOLOv5 model scores lower for missed and falsely detected objects than the YOLOv5 and YOLOv7 models, especially for small objects that occur less frequently in certain categories. Notably, the confusion matrix summarizes the object classification performance of the trained models for four different object classes.
It contains ten categories of the VisDrone2021 dataset. In the matrix, "background FP" signifies the instances where the model missed detecting non-background category target objects. In contrast, "background FN" indicates the occurrences where the model falsely detected category target objects that were not present.
Thus, comparing the experimental results, it can be seen that the model designed in this paper has better results on the VisDrone2021 dataset, with a significant improvement in the mAP value, and has a well-balanced performance between the model’s running speed and the number of parameters, which is more adaptable to the detection of small targets under the aerial images of UAVs, and is superior to the majority of the small target detection models.
4.4.2 Experiments with the AI-TOD dataset.
This research experimentally validates several existing better small target detection models on the AI-TOD dataset, as shown in Table 3. It can be seen that the ASG-YOLOv5 model designed in this paper has 1.4% higher mAP than the YOLOv5 model. The minimal target detection accuracy is 1.9% higher than the APvt value of the YOLOv5 model, the weak target detection accuracy is 1.7% higher than the APt value of the YOLOv5 model, and the small target detection accuracy is 2.9% higher than the APS value of the YOLOv5 model. For medium-sized targets, the accuracy of the YOLOv5 model is 4.6% higher. Adding the modules designed in this paper to YOLOv7 and YOLOv8, as shown in Table 3, ASG-YOLOv7 and ASG-YOLOv8 are less effective in detecting UAV scenarios compared to the ASG-YOLOv5 model, which is 1.2% mAP higher than the ASG-YOLOv8 model. This indicates that due to the differences in the structure of these three networks, the YOLOv5 network can effectively integrate and utilize the attention and spatial gating mechanisms that this research has designed, but it is not able to adapt or utilize these mechanisms well in the structure of YOLOv7 and YOLOv8; joining to the YOLOv5 network its model can take advantage of its feature expression ability due to the way of feature extraction, the number of layers, and the sensory field, and other factors result in the inability to utilize or adapt this module in YOLOv7 and YOLOv8 fully.
4.4.3 Experiments with the NWPU-RESISC dataset.
As can be seen from the results of the comparison of the detection detail visualization of the NWPU-RESISC dataset in Fig 11, as well as the detail portions of the red and blue boxes, the ASG-YOLOv5 model in Fig 11a has high confidence in detecting a variety of weak target classes and is able to detect them accurately. In contrast, the YOLOv5 model in Fig 11b suffers from omission problems in detecting target categories. Also, the confidence of target detection is significantly lower than that of the ASG-YOLOv5 model designed in this paper. This verifies that the model designed in this paper is more effective in detecting target objects from tiny to medium target objects in UAV remote sensing captured images with higher detection accuracy.
a. denotes the graph of the visualization effect of the ASG-YOLOv5 model for the UAV remote sensing captured images; b. denotes the graph of the visualization effect of the YOLOv5 model for the UAV remote sensing shooting picture visualization effect diagram. Fig 11 is attributed to the NWPU-RESISC database and are available from the NWPU-RESISC database (url(s) https://tensorflow.google.cn/datasets/catalog/resisc45).
Overall, it can be seen from the visualization of the NWPU-RESISC dataset detection comparison details in Fig 11 and its details in the red and green boxes that this study is effective in detecting the model in specific scenarios and conditions such as dense targets, light changes, tiny targets, target occlusion, and target motion blurring. Meanwhile, from the comparison of the details in the two figures, it can be seen that the ASG-YOLOv5 model designed in this paper, on the other hand, can be better applied to specific scenes and conditions such as dense targets, light changes (day and night), tiny targets, target occlusion, target motion blurring, etc., and has a better detection ability for capturing tiny targets compared with the YOLOv5 model.
4.5 Ablation experiments
On VisDrone2021-DET-val, this research trained different components to analyze different effects, as shown in Table 4.
- A. First, the study adds the DCA module to the end of the backbone network of YOLOv5 to compare with the original YOLOv5 model on VisDrone2021-DET-val. As shown in Table 4, it can be seen that the mAP value is improved by 1.5% after adding the DCA module, and from the APS value, it can be seen that the addition of the DCA module improves the model’s focus on small targets, and the detection accuracy is improved by 3.4%. Meanwhile, the study compared the effectiveness and accuracy of Squeeze-and-Excitation Networks and the DCA module designed in this paper for target detection of the YOLOv5 model in UAV filming scenarios, as shown in Table 5, which shows that the mAP value is improved by 1.5% after the addition of the DCA module designed in this paper, which is a good proof of the effectiveness of the DCA module designed in this paper in improving the model’s ability to focus on and detect useful target information. Meanwhile, the DCA module usually needs a large amount of training data to learn effective contextual relationships; however, in domains or tasks such as industry, agriculture, etc., data may be very scarce, limiting the module’s performance. Meanwhile, the DCA module, although lightweight, also needs to compute a large number of attentional weights, which will have a certain number of parameters.
- B. Secondly, the study adds the SGM module in the Neck layer of YOLOv5 with the original YOLOv5 model on VisDrone2021-DET-val for comparison. As shown in Table 4, it can be seen that after adding the SGM module, the features extracted by the backbone network are fed into the neck structure, which is filtered by the SGM module on the feature space of the deep feature layer and then combined with the shallow features to introduce the global attentional up-sampling and the weighted fusion of the feature layers of different scales, which enhances the characterization of the useful target categories, reduces the interference of the redundant feature information, and increases the backbone useful feature information in the extracted features of the network. Eventually, the detection accuracy of the model for useful feature targets is improved so that the mAP value of the model is increased by about 1.1%, and it can be seen from the APS value that the model filters out redundant targets and improves the detection accuracy of small targets by 2.6%. The module uses a lightweight structure, and the overall number of parameters of the model has less improvement after the addition of the neck structure. At the same time, the module may be insufficient for eliminating redundant information and feature fusion strategy for small target detection in other scenarios, and the number of parameters of the module still needs to be reduced so that the model can have a faster detection speed in other scenarios.
- C. Finally, the study replaced the CIOU regression loss function and added the NWD-CIoU regression loss function. According to Eq 8, the study compared the model detection accuracy of the NWD-CIoU regression loss function on the UAV dataset taking different values of the balance coefficient, as shown in Table 6, the NWD-CIoU loss designed in this paper based on the ASG-YOLOv5 model improves the mAP value of CIOU Loss by 0.3% over CIOU Loss, and the value of the small-target detection accuracy improves by 0.3%; at the same time, balance coefficients taking the values of 0.7NWD and 0.3IoU have the best model detection accuracy. As shown in Figs 12 and 13, the study compared the stability of the model loss function curves under different values of the balance coefficient in the two UAV datasets. When the balance coefficient takes the values of 0.7NWD and 0.3IoU, the loss function curves of the model are the smoothest and the fastest convergence, proving that the loss function’s stability is better in this case.
5 Conclusion
In this paper, for the problem of small target detection in unmanned aerial vehicle remote sensing aerial images, the study proposes the ASG-YOLOv5 model by taking YOLOv5 as the base model, adding some cutting-edge target detection techniques and improving the backbone network and neck structure of YOLOv5. The study designs a dynamic contextual attention module, which is added to the end of the backbone network to dynamically assign target weights and fuse global and local contextual information to improve the model’s attention to small target feature information in the extracted features. In the multi-scale fusion stage, this research designed spatial gated filtering multi-directional weighted fusion module to reduce the interference of redundant features in the features extracted from the backbone network and to increase the information of small target pixels in the underlying feature layer so that the model can better extract useful target features; the introduction of the Normalized Wasserstein Distance improves the regression loss function, measures the similarity of Gaussian distributions between regression frames, and increases the ability to measure the similarity between mutually inclusive bounding boxes and the smoothing of small target differences. On the VisDrone2021 and AI-TOD datasets, the model designed in this paper has been subjected to several comparative experiments with other newer small target detection models, achieving better detection accuracy and significant results in applying attention and spatial gating mechanisms. The combination with different state-of-the-art techniques makes the improved YOLOv5 model more suitable for detecting small targets in UAV remote sensing aerial photography images. It also solves the problems of tediousness, scale uncertainty, and environmental changes when there are multiple object classes in different scale images in UAV remote sensing aerial photography and better prevents the phenomenon of missed detection. By optimizing and adjusting for UAV scenarios, an effective solution is provided for this UAV small target detection field, which promotes the development of small target detection technology and improves the effectiveness of UAV missions. However, the method in this paper still has some limitations; although the lightweight module is added, the overall parameter count of the model still needs to be increased, which makes the model slow down in the application scenarios. Meanwhile, the model designed in this paper and existing models still have low detection accuracy for small targets in UAV scenarios. This study will further investigate the limitations of the improved model and apply the model to scenarios such as image segmentation.
References
- 1.
Haoyang Zhang, Ying Wang, Feras Dayoub, and Niko Sunderhauf. Varifocalnet: An iou-aware dense object detector. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8514–8523,2021.
- 2.
Yukang Chen, Peizhen Zhang, Zeming Li, Yanwei Li, Xiangyu Zhang, Gaofeng Meng, et al: Feedback-driven data provider for object detection. arXiv e-prints, pp. arXiv-2004,2020.
- 3.
Zhu, Xingkui, et al. "TPH-YOLOv5: Improved YOLOv5 based on transformer prediction head for object detection on drone-captured scenarios." Proceedings of the IEEE/CVF international conference on computer vision. 2021.
- 4.
Wang, Jinwang, et al. "A normalized Gaussian Wasserstein distance for tiny object detection." arXiv preprint arXiv:2110.13389, 2021.
- 5.
Zhaohui Zheng, Ping Wang, Dongwei Ren, Wei Liu, Rongguang Ye, Qinghua Hu, et al. Enhancing Geometric Factors in Model Learning and Inference for Object Detection and Instance Segmentation. arXiv preprint arXiv:2005.03572,2020.
- 6.
Cao, Yaru, et al. "VisDrone-DET2021: The vision meets drone object detection challenge results." Proceedings of the IEEE/CVF International conference on computer vision. 2021.
- 7.
Wang, Jinwang, et al. "Tiny object detection in aerial images." 2020 25th international conference on pattern recognition (ICPR). IEEE, 2021.
- 8.
Cheng G, Han J, Lu X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proceedings of the IEEE. 2017;105:1865–1883.
- 9.
R. Girshick, J. Donahue, T. Darrell, and J. Malik. rich feature hierarchies for accurate object detection and semantic segmentation. in IEEE Conference on Computer Vision and Pattern Recognition (CVPR),2014.
- 10.
Ross Girshick. Fast R-CNN. In Proceedings of the IEEE international conference on computer vision, pages 1440–1448,2015.
- 11.
Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems (NIPS), pages 91–99,2015.
- 12.
Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, et al. SSD: Single shot multibox detector. In European conference on computer vision, pages 21–37. Springer,2016.
- 13.
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi. You Only Look Once: Unified, Real-Time Object Detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, 27–30 June 2016,779–788.
- 14.
Glenn Jocher, Alex Stoken, Jirka Borovec, NanoCode012, Ayush Chaurasia, TaoXie, et al. ultralytics/YOLOv5:v5.0—YOLOv5-P6 1280 models, AWS, Supervise.ly, and YouTube integrations, Apr. 2021.
- 15.
Alexey Bochkovskiy, Chien-Yao Wang, and Hong-Yuan Mark Liao. YOLOV4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934, 2020.
- 16.
Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125.
- 17.
Liu, Shu, et al. Path aggregation network for instance segmentation. Proceedings of the IEEE conference on computer vision and pattern recognition.2018.
- 18.
Mingxing Tan, Ruoming Pang, and Quoc V Le, Efficientdet: Scalable and efficient object detection. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10781–10790,2020.
- 19.
Wang, J.; Chen, Y.; Gao, M.; Dong, Z. Improved YOLOv5 network for real-time multi-scale traffic sign detection. arXiv 2021, arXiv:2112.08782.
- 20.
Mate Kisantal, Zbigniew Wojna, Jakub Murawski, Jacek Naruniec, and Kyunghyun Cho. Augmentation for small object detection. arXiv preprint arXiv: 1902.07296,2019.
- 21.
Golnaz Ghiasi, Yin Cui, Aravind Srinivas, Rui Qian, Tsung Yi Lin, Ekin D. Cubuk, et al. Simple copy-paste is a strong data augmentation method for instance segmentation. arXiv preprint arXiv:2012.07177,2020.
- 22.
Terrance DeVries and Graham W Taylor. Improved regularization of convolutional neural networks with CutOut. arXiv preprint arXiv:1708.04552,2017.
- 23.
Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6023–6032,2019.
- 24.
A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, et al. An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929,2020.
- 25.
Liu, Ze, et al. "Swin transformer: Hierarchical vision transformer using shifted windows." Proceedings of the IEEE/CVF international conference on computer vision. 2021.
- 26.
Zhang, Cui, et al. "SwinSUNet: Pure transformer network for remote sensing image change detection." IEEE Transactions on Geoscience and Remote Sensing 60: 1–13, 2022.
- 27.
Hu, Jie, Li Shen, and Gang Sun. "Squeeze-and-excitation networks." Proceedings of the IEEE conference on computer vision and pattern recognition. 2018.
- 28.
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, et al. Attention is all you need. In Advances in Neural Information Processing Systems, pages 5998–6008,2017.
- 29.
Wang, Jiaqi, et al. "Carafe: Content-aware reassembly of features." Proceedings of the IEEE/CVF international conference on computer vision. 2019.
- 30.
Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, Thomas S. Huang. Free-Form Image Inpainting With Gated Convolution. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019, pp. 4471–4480.
- 31.
Yu, Fisher,and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions. arXiv preprint arXiv:1511.07122, 2015.
- 32.
Li, Hanchao, et al. "Pyramid attention network for semantic segmentation." arXiv preprint arXiv:1805.10180, 2018.
- 33.
Sebastian U Stich, Jean-Baptiste Cordonnier, and Martin Jaggi. Sparsified SGD with memory. in Advances in Neural Information Processing Systems, pages 4452–4463,2018.
- 34.
Glenn Jocher, Ayush Chaurasia, and Jing Qiu. YOLO by Ultralytics, 1 2023.
- 35.
J. Pang, K. Chen, J. Shi, H. Feng, W. Ouyang, and D. Lin, "Libra CNN: Towards balanced learning for object detection,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2019, pp. 821–830.
- 36.
S. Zhang, L. Wen, X. Bian, Z. Lei, and S. Z. Li. Single-shot refinement neural network for object detection. in CVPR, pages 4203–4212, 2018.
- 37.
Lin, Tsung-Yi, et al. "Focal loss for dense object detection." Proceedings of the IEEE international conference on computer vision. 2017.
- 38.
J. Redmon, and A. Farhadi, “YOLOv3: an incremental improvement,” arXiv:1804.02767. 2018.
- 39.
Z. Li, C. Peng, G. Yu, X. Zhang, Y. Deng, and J. Sun. Detnet: a backbone network for object detection. coRR, abs/1804.06215,2018.
- 40.
Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: Fully convolutional one-stage object detection. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9627–9636.
- 41.
Zhang, Shifeng, et al. "Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection." Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2020.
- 42.
Y. Li, Y. Chen, N. Wang, and Z. Zhang. 2019. Scale-aware trident networks for object detection. In Proceedings of the IEEE International Conference on Computer Vision. 6054–6063.
- 43. Wen L., et al. "VisDrone-DET2020: The Vision Meets Drone Object Detection in Image Challenge Results." Lecture Notes in Computer Science 12538 (2020): 692–712.
- 44.
Wang, Chien-Yao, Alexey Bochkovskiy, and Hong-Yuan Mark Liao. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv preprint arXiv:2207.02696, 2022.
- 45.
Duan, Kaiwen, et al. "Centernet: Keypoint triplets for object detection." Proceedings of the IEEE/CVF international conference on computer vision. 2019.
- 46.
Cai, Zhaowei, and Nuno Vasconcelos. "Cascade r-cnn: Delving into high quality object detection." Proceedings of the IEEE conference on computer vision and pattern recognition. 2018.
- 47.
Xu, Chang, et al. "RFLA: Gaussian receptive field based label assignment for tiny object detection." European conference on computer vision. Cham: Springer Nature Switzerland, 2022.