Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

OS-DETR: End-to-end brain tumor detection framework based on orthogonal channel shuffle networks

  • Kaixin Deng ,

    Contributed equally to this work with: Kaixin Deng, Quan Wen, Fan Yang, Hang Ouyang

    Roles Conceptualization, Investigation, Software, Writing – original draft, Writing – review & editing

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Quan Wen ,

    Contributed equally to this work with: Kaixin Deng, Quan Wen, Fan Yang, Hang Ouyang

    Roles Funding acquisition, Validation, Writing – review & editing

    wenq@cdut.edu.cn

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Fan Yang ,

    Contributed equally to this work with: Kaixin Deng, Quan Wen, Fan Yang, Hang Ouyang

    Roles Data curation, Software, Validation

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Hang Ouyang ,

    Contributed equally to this work with: Kaixin Deng, Quan Wen, Fan Yang, Hang Ouyang

    Roles Formal analysis, Investigation, Software, Visualization

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Zhuohang Shi,

    Roles Formal analysis, Validation

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Shiyu Shuai,

    Roles Project administration, Software, Validation

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

  • Zhaowang Wu

    Roles Validation, Visualization

    Affiliation College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, China

Abstract

OrthoNets use the Gram-Schmidt process to achieve orthogonality among filters but do not impose constraints on the internal orthogonality of individual filters. To reduce the risk of overfitting, especially in scenarios with limited data such as medical image, this study explores an enhanced network that ensures the internal orthogonality within individual filters, named the Orthogonal Channel Shuffle Network ( OSNet). This network is integrated into the Detection Transformer (DETR) framework for brain tumor detection, resulting in the OS-DETR. To further optimize model performance, this study also incorporates deformable attention mechanisms and an Intersection over Union strategy that emphasizes the internal region influence of bounding boxes and the corner distance disparity. Experimental results on the Br35H brain tumor dataset demonstrate the significant advantages of OS-DETR over mainstream object detection frameworks. Specifically, OS-DETR achieves a Precision of 95.0%, Recall of 94.2%, mAP@50 of 95.7%, and mAP@50:95 of 74.2%. The code implementation and experimental results are available at https://github.com/dkx2077/OS-DETR.git.

Introduction

In recent years, significant advancements in medical technology have greatly contributed to brain tumor research and treatment, deepening our understanding of their molecular and structural characteristics. Among these technologies, gene sequencing plays a pivotal role by uncovering the molecular biology underlying brain tumors and enabling the development of personalized treatments [1]. High-quality imaging diagnostics, including skull CT [2] and MRI [3], are essential for the accurate detection and quantitative evaluation of brain tumors. These imaging tools provide comprehensive lesion visualization, empowering physicians to make informed clinical decisions and improve patient outcomes. Despite these advancements, there remains substantial potential for further innovation, particularly in applying object detection techniques to enhance diagnostic efficiency.

Object detection, a fundamental area in computer vision, is widely applied to identify and localize object instances within predefined categories. Traditional object detection often relies on the careful design and extraction of manual features [4, 5]. However, the development of deep learning, particularly convolutional neural networks (CNN) [6, 7], has revolutionized this field, enabling more robust and accurate detection. Modern deep learning-based object detection approaches are typically categorized into one-stage and two-stage methods [8]. In one-stage methods, the most influential framework is You Only Look Once (YOLO). With its efficient feature extraction capabilities, YOLO has set a benchmark for real-time object detection tasks and is widely applied across various fields [9, 10]. These applications highlight its adaptability in addressing complex challenges. The introduction of DETR [11] marks a milestone in object detection, framing it as a set prediction problem. Leveraging the Transformer’s sequence transformation capabilities [12], DETR directly converts image sequences into set sequences. This eliminates the need for region proposal networks and anchor mechanisms, enabling an end-to-end detection framework.

However, within medical image, the majority of studies have focused on classification [13] and segmentation [14, 15], with relatively few studies on brain tumor object detection. Currently, most existing approaches are the YOLO-based framework [1619] in the medical image, while research exploring end-to-end object detection framework remains limited. DETR has demonstrated highly competitive performance on the large-scale COCO [20] dataset. However, Research [21] shows a significant decline in its performance when applied to small-scale datasets, indicating a strong dependency on data quantity. To address this issue, this study introduces a simple and efficient feature extraction network as the backbone of the detection framework. This network reduces the risk of overfitting and exhibits notable advantages, particularly in the field of medical imaging, where data availability is often limited. Specifically, this study explores an enhanced model, the OS-DETR, as depicted in Fig 1. Similar to other DETR-based frameworks, OS-DETR consists of a backbone network, an improved efficient encoder, and a decoder.

thumbnail
Fig 1. The architecture of the OS-DETR.

Compared to RT-DETR, this study introduces several improvements to the network architecture. First, OS-DETR incorporates the OSNet into the backbone network to enhance feature extraction capabilities. Second, OS-DETR optimizes the encoder part to strengthen the fusion of cross-scale information.

https://doi.org/10.1371/journal.pone.0320757.g001

The primary contributions in this study are as follows:

  1. This study introduces the OSNet, which utilizes the Special Orthogonal Group process to initialize filters, ensuring orthogonality within their internal spatial structure.
  2. In the encoder component, OS-DETR incorporates the deformable attention mechanism to replace the multi-head self-attention mechanism commonly used in other DETR-based framework.
  3. OS-DETR incorporates ICAD-IoU, which integrates the influence of internal regions with corner distance discrepancies.

The structure of this paper is organized as follows: Related Work section provides a comprehensive review of mainstream object detection algorithms and analyzes the literature focusing on algorithms designed for brain tumor detection, systematically outlining the latest progress and findings in the field. Materials and Methods section details the research methods employed in this study, including the implementation specifics, key techniques, and innovative approaches. Experiments and Results section presents the experimental setup and environment, reports the results of comparative and ablation studies, and conducts an analysis of the findings. Finally, Discussion and Conclusion section summarizes the primary content and conclusions of the study while proposing future research directions and potential improvements.

Related work

You Only Look Once (YOLO)

Real-time object detection technology is necessary in various fields. The YOLO framework stands out among object detection algorithms for its ability to accurately complete detection tasks in a single pass, positioning it as a mainstream option. Originally proposed by Joseph Redmon [22], YOLO differs from prior methods in that it does not require sliding windows, the execution of numerous classifiers per image, or a two-stage approach involving the detection of potential object-containing areas followed by classification. Instead, YOLO utilizes a straightforward regression output for prediction, unlike Fast R-CNN [23], which employs dual independent outputs for classification probability and bounding box coordinate regression.

The YOLO series consistently aims to balance speed with accuracy for real-time performance. However, the initial model compromised some accuracy, especially with small objects or overlapping boxes, in pursuit of high-speed detection. As YOLO evolved, striking a balance between speed and accuracy became a focal point. With the integration of anchor boxes and transfer layers in YOLOv2 [24], the accuracy of object localization improved. YOLOv3 [25] adopted a multi-scale feature extraction architecture, enhancing adaptability to objects of varying scales. As the framework progressed, achieving a nuanced balance between speed and accuracy emerged as a key challenge. YOLOv4 [26] and YOLOv5 introduced novel network backbones, enhanced data augmentation methods, and refined training strategies, significantly boosting accuracy while retaining real-time performance.

YOLOv7 [27] introduces the Extended Efficient Layer Aggregation Network (E-ELAN), a strategy that enables deep models to learn and converge more efficiently by managing the shortest and longest gradient paths. E-ELAN enhances the network’s learning capability without disrupting the original gradient paths through shuffling and merging bases to combine features from different groups. This approach is particularly effective for models with infinite stacks of computational blocks. The architecture of YOLOv7 is cascade-based, and when standard scaling techniques like depth scaling are applied, a transition layer is created. YOLOv8 represents an advanced model for object detection and semantic segmentation. It introduces a redesigned Cross Stage Partial Layer (CSPLayer) [28], referred to as the Cross Stage Partial Bottleneck with two Convolutions. This model employs a decoupled anchorless head to address likelihood, classification, and regression tasks. Additionally, it incorporates Complete Intersection over Union [29], Distribution Focal Loss [30], and Binary Cross-Entropy loss functions.

YOLOv9 [31] overcomes the challenges of information bottlenecks and invertible function limitations in deep networks by introducing two novel techniques: Programmable Gradient Information (PGI) and the Generalized Efficient Layer Aggregation Network (GELAN). These advancements achieve higher parameter efficiency while relying solely on traditional convolutional operators. YOLOv10 [32] introduces consistent dual allocation, enabling end-to-end training without the need for Non-Maximal Suppression (NMS). The model optimizes each component of YOLO using a comprehensive efficiency-accuracy-driven design strategy. YOLO11 [33] enhances feature extraction capabilities and improves object detection accuracy through its optimized backbone network and neck structure. Additionally, the refined architectural design and streamlined training processes increase processing speed while maintaining a strong balance between accuracy and performance. On the COCO dataset, YOLO11m achieves a higher mean Average Precision compared to YOLOv8m, reducing the number of parameters by 22% and significantly boosting computational efficiency.

Detection Transformer (DETR)

In recent times, the Transformer has seen widespread application in computer vision’s object classification domain, as evidenced by the Vision Transformer (ViT) and others [3437]. DETR marks the first use of the classic Encoder-Decoder structure of Transformer in object detection. DETR’s backbone network employs a convolutional network, while its Encoder and Decoder are based on the Transformer framework. Its output layer is a Multi-Layer Perceptron (MLP), and the model utilizes a loss function derived from bipartite graph matching, forming a graph by aligning the ground truth with predicted bounding boxes. DETR comprises four key modules: the backbone, encoder, decoder, and prediction head.

Nevertheless, the initial DETR model faced two major challenges: a slow convergence rate, largely due to the uniform self-attention across each position, requiring more time to identify sparse yet crucial spots; and subpar small object performance, attributed to the absence of a multi-scale structure akin to Feature Pyramid Network (FPN) and the self-attention’s computational complexity being square proportional to the feature map’s area. Addressing these issues, Deformable DETR [38] introduced Deformable Attention to tackle the problem inherent in standard Transformer’s Attention, thereby hastening model convergence, reducing algorithmic complexity, and incorporating multi-scale features to enhance small object detection efficiency.

However, the DETR series models’ high computational demands constrain their practical application effectiveness and hinder full utilization of post-processing-free benefits, like NMS. To tackle these challenges, Wenyu Lv and colleagues introduced the first real-time, end-to-end object detection framework, Real-time Detection Transformer (RT-DETR) [39]. RT-DETR features an efficient hybrid encoder adept at handling multi-scale features by segregating intra-scale interaction from cross-scale fusion. It also introduces IoU-aware query selection, further enhancing performance by supplying the decoder with superior initial object queries. Additionally, RT-DETR’s capability to adjust inference speed through varying decoder layers, without necessitating retraining, broadens its applicability in diverse real-time contexts. Notably, RT-DETR was the first DETR-based framework that surpassed contemporary YOLO-based framework of comparable scale in both detection speed and accuracy.

Applications of object detection in brain tumor

RCS-YOLO [17] is an innovative design that achieves an effective balance between feature cascading and computational efficiency. It integrates Reparameterized Convolution based on Channel Shuffle (RCS) and One-Shot Aggregation of RCS (RCS-OSA) methods, which significantly enhance information extraction while minimizing time consumption. The RCS method is inspired by ShuffleNet [40, 41] and leverages channel shuffle to deeply fuse input features across different channels. This approach optimizes the model’s performance in both training and inference phases.

BGF-YOLO [16] is an advanced architecture based on YOLOv8, specifically designed for brain tumor detection. It integrates three key enhancements: Bi-level Routing Attention (BRA) [42], Generalized Feature Pyramid Network (GFPN), and an additional fourth detecting head. BRA employs a dynamic sparse attention mechanism that selectively focuses on salient regions, effectively reducing feature redundancy and improving detection precision. GFPN enables multilevel feature fusion by incorporating dense connections and skip pathways, ensuring robust handling of both high-level semantic and low-level spatial information. The inclusion of the fourth detecting head extends the detection capacity to finer scales, addressing variability in tumor sizes and improving accuracy for larger objects.

Materials and methods

Feature extraction network for OS-DETR model

Orthogonal Channel Shuffle Network (OSNet).

The concept of channel attention was initially introduced by Squeeze-and-Excitation Networks (SENet) [43]. SENet employs Global Average Pooling (GAP) to reduce the spatial dimension of each feature channel to a single scalar. However, a drawback of SENet is its sole reliance on GAP for channel compression. Addressing this limitation, Frequency Channel Attention Networks (FcaNet) [44] posited that GAP might neglect vital low-frequency information. FcaNet emphasizes the importance of frequency selection in Discrete Cosine Transforms (DCTs) for enhanced information provision. The success of FcaNet is largely due to the orthogonality in DCT compression mapping. Building upon this, Hadi Salman et al. developed a channel attention mechanism that utilizes random orthogonal filters to compress spatial information of each feature [45], termed Orthogonal Channel Attention. OrthoNets demonstrated notable results on the ImageNet dataset and achieved superior performance on the Birds and Places365 datasets.

In OrthoNets, the Gram-Schmidt process flattens randomly generated filters into vectors and ensures pairwise orthogonality among these vectors, achieving overall orthogonality between filters. However, this approach does not explicitly constrain the internal orthogonality of individual filters, meaning the orthogonality between rows or columns within each filter is not guaranteed. Constraining internal orthogonality within individual filters can effectively reduce the risk of overfitting, particularly in the medical image field where data is often limited. To address this, this study introduces the OSNet that adopts the special orthogonal group method to initialize filters, ensuring orthogonality within the spatial structure of each filter. This approach guarantees the independence of features within each filter, such as orthogonality among row or column vectors, without imposing orthogonality constraints between different filters, which are independently sampled. The structure of the OSNet Block is shown in Fig 2.

thumbnail
Fig 2. The structure of the OSNet Block.

The proposed method consists of three stages. Stage 0: Orthogonal matrices are sampled from the Special Orthogonal Group (SO) to construct orthogonal filters; Stage 1: These orthogonal filters are applied to the input feature maps, followed by dimensionality reduction along the spatial dimensions. This process generates channel attention weights, which are further refined through a fully connected layer and subsequently used to reweight the channels of the original feature map; Stage 2: The channels of the output feature map are then rearranged.

https://doi.org/10.1371/journal.pone.0320757.g002

The OSNet Block is designed based on the ResNet architecture and comprises two primary branches: a shortcut branch and a main branch. The shortcut branch adjusts spatial dimensions and channel numbers to match the main branch output. The main branch consists of two convolutional layers. The first layer adjusts the spatial dimensions of the input feature map X, while the second layer extracts features and outputs the feature map F. For the feature map F with dimensions (b, c, h, w), orthogonal matrices are randomly sampled from the Special Orthogonal Group , since only square matrices can be ortho gonal matrices, where n = h = w. For each channel i (out of a total of c channels), an orthogonal matrix is sampled and used as the filter weights. These filters maintain orthogonality within the spatial dimensions, ensuring diversity and independence in the feature extraction process.

All filters are concatenated along the channel dimension, resulting in a combined filter with dimensions (c, h, w). Element-wise multiplication is performed between the feature map F and the corresponding filter weights for each channel i, preserving the original dimensions (b, c, h, w) of the feature map F. This operation effectively applies orthogonal constraints to the feature map F. A summation operation is then applied across the spatial dimensions (height and width). This operation compresses the spatial information of each channel into a tensor f with dimensions (b, c, 1, 1). This feature transformation projects the input feature map F onto the orthogonal subspace represented by the concatenated orthogonal filter . The tensor value of each channel in f represents its projection coefficient in the orthogonal subspace. The orthogonal filter , satisfying orthogonal constraints, enables this projection to extract components of feature map F along different orthogonal directions. This process reduces redundancy and correlation between features, enhancing the independence and diversity of feature representations.

The orthogonal constraint promotes feature decoupling and diverse representations, allowing the network to better capture the essential structures and characteristics of input data. The resulting tensors f undergo a subnetwork comprising two fully connected layers with a ReLU activation and a Sigmoid activation function, producing channel attention weights with dimensions (b,c), representing importance weights for each channel. The attention weights are applied channel-wise to the original feature map F, reweighting the importance of each channel. This attention mechanism enables the network to dynamically focus on relevant features, enhancing its ability to capture critical information across different channels.

The attention weights are reshaped into the dimensions (b, c, 1, 1) to align with the structure required for element-wise multiplication. These weights are then applied to each channel of the original feature map F, producing the reweighted feature map . This operation highlights key feature channels while reducing the influence of less important ones, thereby enhancing the network’s ability to focus on essential information within the input data. The reweighted feature map is then combined with the residual connection and activated using the ReLU function, resulting in the final feature map . This step merges the attention-learned feature map with the original input features through the residual connection , preserving input information while enhancing the feature representation capacity. Subsequently, a Channel Shuffle operation is applied to the merged feature map . This channel rearrangement promotes information exchange and feature fusion between different channel groups, thereby improving feature complementarity and diversity. Finally, a non-linear activation function, such as ReLU, is applied to the merged feature map . This step increases the network’s expressive power, allowing it to capture more complex patterns and relationships within the data.

In OS-DETR, OSNet Blocks are stacked, enabling the network to progressively refine feature representations and capture hierarchical patterns. Compared to the Gram-Schmidt method employed in the original OrthoNets, this approach constrains the orthogonality within individual filters, effectively reducing the risk of model overfitting to some extent.

Orthogonality and its implications for neural networks.

The Special Orthogonal Group is a significant matrix group comprising orthogonal matrices over real numbers with determinants equal to 1. Formally, the n-dimensional Special Orthogonal Group SO(n) is defined as:

(1)

Where R is an real matrix, is the transpose of R, I is the identity matrix, and is the determinant of R. Matrices in SO(n) are orthogonal, satisfying RTR = I, ensuring their column (or row) vectors are mutually orthogonal and of unit length. This orthogonality guarantees that the column (or row) space forms an orthonormal basis. By leveraging SO(n)’s parameterized representation, manifold sampling, symmetry, and dimensional flexibility, this method obtain a set of orthogonal bases with strong orthogonality, good diversity, and adaptability to different task requirements. Utilizing orthogonal matrices generated from SO(n) as convolutional filters enhances feature representation quality. This approach ensures minimal redundancy and maximum independence between feature channels, enabling the network to capture richer and more diverse feature patterns. It also provides a regularization effect, mitigating overfitting and improving model generalization.

Consider a multi-layer network:

(2)

Where represents a 1-Lipschitz activation function, such as ReLU, and denotes the weight matrix or the unfolded convolutional kernel of the l-th layer. The function represents the output of the deep neural network, with x as the input. Applying orthogonal constraints to some or all , such that:

(3)

Where denotes the spectral norm of the matrix, ensures the following property:

(4)

Here, the Lipschitz constant of the entire network is . Under ideal conditions, where all layer weight matrices are strictly orthogonal, these layers do not amplify input perturbations. This helps maintain control over the overall Lipschitz constant of the network. Even when strict orthogonality cannot be achieved for all layers in practice, applying orthogonal constraints to critical layers and keeping their spectral norms within a small range (e.g., ) effectively tightens the network’s overall Lipschitz constant. This constraint limits the network’s effective capacity, reducing the risk of overfitting.

Improved efficient encoder for OS-DETR

In the encoder section of other multi-scale models, the extended input sequence length significantly impacts the computational efficiency, impeding the real-time capabilities of the DETR model. To address this challenge, this study introduced an improved efficient encoder, which structure is similar to the RT-DETR model. This encoder comprises two key components: the Attention-based Intra-scale Feature Interaction (AIFI) module and the CNN-based Cross-scale Feature-fusion Module (CCFM). The structure of this encoder is illustrated in Fig 3. This approach optimizes the processing of multi-scale features, enhancing both computational efficiency and model performance.

thumbnail
Fig 3. The structure of improved efficient encoder for OS-DETR Model.

This encoder comprises an AIFI module and a CCFM module. The features extracted by the feature extraction network need to be dimensionally varied before they can be fed into the AIFI and CCFM modules.

https://doi.org/10.1371/journal.pone.0320757.g003

Unlike previous DETR models, such as Deformable-DETR [38], which typically convert feature maps from multiple scales into a lengthy vector, this study adopts the approach used in RT-DETR. It applies the AIFI module for within-scale feature interaction exclusively to the S5 feature plane extracted by the feature extractor. This strategy leverages the deeper, higher-level, and more semantically rich characteristics of S5 features compared to the shallower S3 and S4 features, making them crucial for the Transformer model. These features play a vital role in differentiating various object characteristics, as shallower features often lack rich semantic content.

In the AIFI module, the two-dimensional S5 features undergo initial flattening into a vector. They are then processed through deformable attention and Feed-Forward Network (FFN), before being reshaped back into two dimensions for subsequent cross-scale feature fusion using the CCFM module. The CCFM module comprises multiple fusion blocks with convolutional layers inserted into the fusion pathway. These blocks merge adjacent features into new ones, incorporating 3 RepBlocks. The outputs from both pathways are combined through element-wise addition.

This process can be described with the following mathematical expression:

(5)(6)

The Reshape operation restores the feature’s shape to match that of S5, reversing the Flatten operation. Deformable attention [36] (DAT) is employed to address a limitation of the conventional multi-head self-attention mechanism used in models like RT-DETR. In the standard approach, the attention pattern remains fixed across all positions, with each attention head focusing on a weighted combination of all positions in the input sequence. These patterns do not change during training, which limits the model’s ability to capture complex relationships and dependencies within the data. Fig 4 illustrates the DAT structure. The deformable attention mechanism concentrates only on a select portion of key areas in the image, thus maintaining high performance while substantially reducing computational load. DAT dynamically chooses sampling points instead of uniformly processing the entire image, enabling the model to focus more on areas most relevant to the current task [46]. By learning a set of offsets for each input image, DAT allows all queries to interact with a common set of keys and values, shifting the key and value to crucial positions. DAT enhances the representation ability of sparse attention and also has linear spatial complexity.

Given an input feature map , a uniform grid of points p is generated as references, which . Specifically, the grid size is downsampled from the input feature map size by a factor r, where HG = H/r and WG = W/r. The values of the reference points are linearly spaced 2D coordinates ranging from (0,0) to . These coordinates are then normalized to the range based on the grid dimensions , where (–1,–1) represents the top-left corner and corresponds to the bottom-right corner. DAT first linearly maps the input feature map x to the query:

(7)

Next, initialize a set of uniform grids as the reference points p for the sampling points, and the offset is generated by the sub-network :

(8)

Where s is the predefined control offset amplitude. Then, add the offset to the reference point p to get the coordinates of the sampling points, and use bilinear interpolation to sample the original feature map x:

(9)

Where represents the bilinear sampling operation, the first operand A is the feature map, and the second operand B is the coordinates of the sampling points. The features are then sampled at the locations of the deformed points, serving as keys and values for the subsequent projection matrices. The deformed keys and values are obtained through the following expression:

(10)

DAT employs a “Deformable relative position bias” method for position encoding, introducing spatial information into the visual attention mechanism. This approach encodes the relative position bias when considering the relative positions between queries and keys in the attention mechanism. Traditional attention mechanisms base the interaction between queries and keys solely on content similarity, neglecting their spatial relationships. To address this limitation, relative position bias is introduced to capture spatial information between queries and keys.

For a feature map of shape H W, the relative coordinate displacements range from [–H,H] and [–W,W] in the two dimensions. In Swin Transformer [47], the authors construct a relative position bias table , indexing this table by relative displacements in both directions to obtain the relative position bias B. However, in deformable attention mechanisms, key positions are continuous rather than discrete. To accommodate this, the authors propose a number of improvements:

  1. Continuous positions: Deformable attention allows keys to have continuous positions, not limited to discrete grid locations.
  2. Normalized range: Relative displacements are normalized to the range, enhancing robustness across different feature map scales.
  3. Interpolation: Interpolation in the parameterized bias table covers all possible offset values, not just discrete relative displacements.
  4. Parameterized bias table: A learnable parameterized bias table is used instead of fixed position encoding, allowing adaptive learning of position biases.

This method provides a flexible and adaptive approach to incorporating spatial information into the attention mechanism, particularly suited for deformable attention scenarios. The mathematical expression for this approach is as follows:

(11)

Where is a smaller-sized bias matrix, is used to calculate the relative position of the query and the deformed key, and finally the output of each head in the multi-head attention is calculated as follows:

(12)

Where d is the dimension of the query and key, which is the same as in the regular Transformer.

Decoder for OS-DETR model

The Decoder’s input comprises three essential components. The first part includes the content information of bounding boxes, which encapsulates the category data. The second part utilizes the coordinate information of bounding boxes, represented as 4D data (x,y,w,h), denoting the position and dimensions of each anchor. The (x,y) values indicate location, while (w,h) represent size and shape. The third input is the memory, which consists of multi-scale feature maps extracted from the image. The initial values for both the content and coordinate information of bounding boxes are derived from two sources. The primary source is a set of 300 queries selected using the IOU-aware query-selection method. Unlike the original DETR, where queries lack prior information leading to slower convergence, RT-DETR’s queries are chosen from the memory based on category confidence scores. The secondary source involves adding noise to ground truth data. The denoising approach aligns closely with the DN-DETR [48] model. It introduces noise to ground truth categories, which then serve as part of the Decoder input. Similarly, object coordinates undergo noise addition, contributing to the anchor coordinates in the Decoder input. The Decoder employs Multi-Scale Deformable-Attention for its attention mechanism. Its output encompasses both category predictions and bounding box coordinates. This design enables model to perform end-to-end learning directly from image pixels to bounding boxes and category predictions, eliminating the need for additional post-processing steps such as NMS.

Intersection over Union (IoU) for OS-DETR

In object detection tasks, IoU [49] serves as a crucial metric for evaluating the precision of bounding box regression. IoU-based loss functions, such as GIoU, provide gradient signals for model training and accelerate convergence. DETR-based framework often employ IoU-related strategies to guide the generation of high-quality feature representations. For instance, the original DETR model utilizes GIoU and its associated IoU-aware query selection method. This strategy prioritizes matching predicted boxes with higher IoU to ground truth boxes during the training process’s matching algorithm. This matching results in a greater update magnitude for these high-quality predicted boxes during backpropagation.

However, standard IoU or GIoU may encounter challenges in accurately reflecting subtle discrepancies between predicted and ground truth boxes, particularly when the two boxes exhibit similar shapes but possess slight differences in size or position. To further enhance IoU’s discriminative capability, several improvement methods have been proposed. Inner-IoU [50] emphasizes the influence of the inner region of bounding boxes on the loss. It scales both the ground truth box and the predicted box by the same ratio, using their respective center points as references. It then computes the IoU using the scaled boxes, thus giving greater attention to the overlap of inner regions. By reinforcing the internal regions, this approach can accelerate bounding box regression and improve the model’s generalization performance.

Another significant improvement is MPDIoU [51], which addresses the challenge of distinguishing between predicted and ground truth boxes that share the same aspect ratio but differ in size. MPDIoU enhances discrimination by minimizing the distance between the top-left and bottom-right corner points of the predicted and ground truth boxes. It combines this distance information with IoU. MPDIoU forms the final metric by a weighted combination of the normalized corner point distance and IoU. MPDIoU demonstrates improved discrimination and optimization in bounding box regression tasks, particularly in distinguishing cases with similar positions but different sizes, or cases with close center points but insufficient overlap.

On this basis, OS-DETR integrates the strengths of Inner-IoU and MPDIoU, introducing the Interior-Aware Corner Distance IoU (IACD-IoU). This method simultaneously emphasizes the internal region influence of bounding boxes and the corner distance disparity. During IoU computation, a scaling factor is employed to highlight the overlap within the interior regions, followed by an additional penalty based on the distances between the top-left and bottom-right corners of the predicted and ground truth boxes. Specifically, the calculation of IACD-IoU can be seen as deriving an internally focused IoU, then subtracting a penalty term reflecting the corner distance, thus balancing fine-grained internal differences and global structural deviations, as illustrated in the Fig 5.

The method utilizes a scaling factor, denoted as ratio, set to 0.7 in this study, to compute the coordinates of the scaled ground truth and anchor boxes. The coordinates of the top-left and bottom-right corners of the ground truth box are represented as and , respectively, while those of the anchor box are denoted as and . The center of the ground truth box and its internal box is represented as , with its width and height expressed as and . The scaled coordinates of the ground truth box are computed as follows:

(13)(14)

Similarly, the center, width, and height of the anchor box are denoted as , , and . The scaled coordinates of the anchor box are given by:

(15)(16)

The intersection and union of the scaled ground truth and anchor boxes are then calculated as follows:

(17)(18)

This results in the interior-focused IoU, denoted as :

(19)

After computing , a penalty term is subtracted to account for the corner distance. The Euclidean distances between the corners of the ground truth and anchor boxes, and , are defined as:

(20)(21)

With w and h representing the width and height of the input image, the final form of IACD-IoU is given by:

(22)

Experiments and results

Dataset

This study employs the 2020 brain tumor detection dataset (Br35H) [52] for experiments. For a more objective evaluation, this study utilized the dataset provided by the RCS-YOLO project, comprising 701 images across train and val folders. Specifically, 500 images from the train folder were designated as the training set, and the remaining 201 images in the val folder served as the test set. Fig 6 shows the distribution details of the bounding boxes. The area of most bounding boxes is less than 20% of the total image size. This study further conducts a analysis of the performance of different orthogonality methods, using the Brain Tumor Dataset [53] as the experimental basis. The Brain Tumor Dataset is a simple dataset designed for brain tumor classification, comprising data samples from four categories: glioma, meningioma, no tumor (healthy brain), and pituitary.

Experimental environment

This experiments were conducted on a system running Ubuntu 20.04, using Python 3.8 as the programming language and PyTorch 1.11.0 as the deep learning framework. The hardware setup consisted of an NVIDIA RTX 3090 GPU with 24 GB of memory.

Evaluation metrics

This study employed a comprehensive and rigorous evaluation metric system to analyze the object detection performance of the model. At the same time, this study systematically evaluated the strengths and limitations of the proposed model. Specifically, the selected evaluation metrics encompass multiple aspects, including the model’s speed, complexity, and accuracy. Among them, FPS (frames per second) was used to measure the model’s real-time performance, and its calculation formula is as follows:

(23)

Here, , , represent the preprocessing time, inference time, and postprocessing time per frame, respectively. Each of these metrics is measured in milliseconds (ms). By calculating the total processing time across all stages and taking its reciprocal, the processing efficiency of the model in practical tasks can be determined. The number of parameters (measured in millions) and the GFLOPS (floating-point operations) indicate the computational complexity of the model. Precision and recall are used to evaluate the accuracy and completeness of the detection results, respectively. Their calculation formulas are as follows:

(24)

Here, TP (True Positive) represents the number of correctly detected targets, FP (False Positive) refers to the number of incorrectly detected targets, and FN (False Negative) indicates the number of targets that were not detected. mAP@50 and mAP@50:95 denote the mean Average Precision at an IoU of 0.5 and the mean of a series of AP values calculated at IoU thresholds ranging from 0.5 to 0.95 with a step size of 0.05, respectively. These metrics comprehensively evaluate the model’s performance across different detection thresholds. IoU (Intersection over Union) is defined as the ratio of the intersection area to the union area of the predicted bounding box and the ground truth bounding box. The calculation of AP involves two steps: first, obtaining a series of Precision and Recall values under varying confidence thresholds to plot the Precision-Recall (P-R) curve; second, computing the area under the P-R curve to derive the AP. The mAP is obtained by calculating the arithmetic mean of the AP values across all categories, providing a comprehensive assessment of the model’s overall performance in multi-class detection tasks, ensuring the evaluation’s objectivity and thoroughness.

Experimental parameter setting

The model underwent a training process spanning 100 epochs, utilizing a batch size of 4 and an image resolution of 640 pixels. The initial learning rate was set to 0.0001, while the final OneCycleLR learning rate reached 1.0. Early stopping patience was disabled to allow for comprehensive training. The training procedure employed 4 worker processes to optimize computational efficiency. Data augmentation techniques were strategically applied, with horizontal flipping implemented at a probability of 0.5. However, vertical flipping, Mosaic augmentation, and Mixup augmentation were not utilized in this particular training setup.

Training results of OS-DETR

Fig 7 illustrates the performance metrics of OS-DETR under varying training epochs, including Precision, Recall, mAP@50, and mAP@50:95. Each subfigure presents a solid line representing the average experimental results and a semi-transparent shaded region indicating the standard deviation range calculated from multiple trials. The graphs reveal that all four metrics increase significantly during the initial training phase (the first 30 epochs), with the improvement rate slowing down after 40–60 epochs and eventually stabilizing. This trend suggests that the model undergoes rapid learning in the early stages, followed by a fine-tuning phase. Analyzing the error range shows substantial fluctuations in the early training phase, indicating sensitivity to initialization conditions and the order of training data. As training progresses, the error range gradually narrows, demonstrating enhanced stability in later stages. While YOLO-based framework typically require approximately 200 epochs to achieve convergence, the OS-DETR converges within 100 epochs, showcasing a notably faster convergence rate.

thumbnail
Fig 7. OS-DETR performance metrics across training epochs.

https://doi.org/10.1371/journal.pone.0320757.g007

Comparative experiments

Comparative experiments of orthogonality methods.

This section compares the performance differences between two orthogonalization methods: orthogonality among filters and orthogonality within filters. Specifically, experiments were conducted on the simple brain tumor classification dataset [53]. To minimize the influence of other modules, two classification models with simple structures were constructed based on standard convolutional neural networks. The only difference between the two models lies in the method of filter orthogonality employed.

To evaluate the proposed OSNet, which reduce the risk of overfitting, especially in scenarios with limited data such as medical image, the two models were trained using training set proportions of 100%, 50%, 20%, and 10%. The training results are illustrated in Fig 8.

thumbnail
Fig 8. The comparison of different orthogonality methods at different training set proportions.

https://doi.org/10.1371/journal.pone.0320757.g008

As shown in the figure, when the training set proportions were 100% and 50%, the Accuracy and Loss curves for both orthogonality methods exhibited similar trends, with some instances of overlap. However, as the training set size decreased to 20% and 10%, the model which using orthogonality within filters method demonstrated closer alignment of the Loss curves between the training and validation sets. Furthermore, its Accuracy curves exhibited greater consistency.

These results indicate that, in scenarios involving small-scale medical image datasets, the orthogonality within filters method effectively reduce the risk of overfitting. This validates the effectiveness of the proposed approach.

Comparative experiments on Br35h dataset.

This study conducted comparative experiments on the Br35H dataset with other mainstream object detection algorithms (such as the YOLO series and DETR series models) and specialized brain tumor detection algorithms (such as BGF-YOLO, RCS-YOLO, and LSM-YOLO). All experiments were conducted on the same hardware to ensure consistency. To address inconsistencies in dataset preprocessing among different algorithms, the experiments utilized the dataset provided by the RCS-YOLO project for standardization. The experimental results are presented in Table 1.

thumbnail
Table 1. Comparison with YOLO-based models and other detectors.

https://doi.org/10.1371/journal.pone.0320757.t001

The results of the comparative experiments demonstrate that the proposed OS-DETR achieves significantly better performance than most mainstream YOLO and DETR series models in terms of precision, recall, mAP@50, and mAP@50:95 metrics. Even when compared with models specifically designed for brain tumor detection, OS-DETR exhibits superior performance. Specifically, OS-DETR outperforms RCS-YOLO, RepVGG-GELAN, and LSM-YOLO in the mAP@50 metric. Furthermore, OS-DETR achieves the best results in the mAP@50:95 metric, further confirming its outstanding performance.

Benefiting from the concise and efficient structure of OSNet, OS-DETR has fewer parameters and lower computational complexity than other DETR-based frameworks. Additionally, it surpasses all these DETR-based frameworks in terms of FPS, which highlights its high efficiency. Notably, some moedls suffer from extended post-processing time, causing the FPS of their smaller parameter versions to show limited improvement over larger parameter versions. Leveraging its end-to-end design, OS-DETR also achieves better FPS compared to YOLOv8m, YOLOv10b, YOLOv11m, and RT-DETR models with similar parameter scales.

Visualization of comparative experiments.

The comparison of the prediction results of the OS-DETR model on the validation dataset with the Ground Truth labels is shown in Fig 9. This figure presents a visual comparison between the model’s predictions and the ground truth annotations on several representative images from the validation dataset.

thumbnail
Fig 9. The comparison of OS-DETR predictions and the ground truth labels.

https://doi.org/10.1371/journal.pone.0320757.g009

By comparing the prediction results of the OS-DETR model on the validation dataset with the ground truth labels, the results demonstrate that the OS-DETR model achieves outstanding accuracy in detecting brain tumors.

This study further compared the detection performance of OS-DETR and RT-DETR on challenging samples. As illustrated in Fig 10, in scenarios with blurred boundaries, such as cases (a) and (c), OS-DETR demonstrates superior capability in distinguishing targets from the background. In more complex scenarios, represented by cases (b) and (d) in Fig 10, RT-DETR primarily focuses on local target features, whereas OS-DETR exhibits a broader global perspective. This enables OS-DETR to generate more precise prediction boxes with higher alignment to the ground truth labels.

thumbnail
Fig 10. The comparison of OS-DETR and RT-DETR on challenging samples.

https://doi.org/10.1371/journal.pone.0320757.g010

To investigate the regions of interest at different feature scales in the target detection task, this study applied the Grad-CAM++ method to visualize the multi-scale features of RT-DETR and OS-DETR. Specifically, two key layers from the backbone network were selected for analysis: layer 4, representing the first feature extraction stage, and layer 7, corresponding to the final feature extraction stage. Grad-CAM++ was used to compute activations and gradients during forward and backward propagation for these layers. The resulting activation heatmaps were normalized and overlaid on the original images to provide a more intuitive representation of the network’s focus areas.

Fig 11 illustrates the differences in the network’s focus at various layers for the same image. Features extracted at layer 4, after the initial Convolution Block and OSNet Block, primarily capture edges and local textures. In contrast, features from layer 7, processed through the final Convolution Block and OSNet Block, encapsulate higher-level semantic information, capturing more complete target representations.

thumbnail
Fig 11. The comparison of heatmaps Between RT-DETR and OS-DETR.

https://doi.org/10.1371/journal.pone.0320757.g011

To further analyze the effectiveness of multi-scale feature fusion in the detection head and neck, representative output layers were selected, and their fused outputs were calculated. By examining the generated heatmaps, the model’s focus across different feature layers and its capacity to perceive target regions were intuitively analyzed. This further demonstrates that RT-DETR primarily focuses on local target features, while OS-DETR exhibits a broader global perspective.

Ablation experiment

To comprehensively evaluate the impact of key components in OS-DETR, a series of systematic ablation experiments were conducted, with the results detailed in Table 2. The results demonstrate that the integration of OSNet shows significant performance improvements across all evaluation metrics. Specifically, mAP@50 increased by 0.5%, while mAP@50:95 achieved a 1% improvement. When the enhanced efficient encoder with deformable attention mechanisms was incorporated independently, mAP@50 improved by 0.4%, and mAP@50:95 showed a 1.4% increase compared to the baseline model.

Furthermore, when OSNet and the efficient encoder were integrated simultaneously, the model maintained its excellent performance. Notably, the introduction of ICAD-IoU further optimized the model, achieving the best overall performance. Compared to the original model, mAP@50 improved by 1.5%, and mAP@50:95 increased by 1.4%, reflecting a substantial enhancement in detection accuracy.

The combined effects of OSNet, deformable attention mechanisms, and ICAD-IoU significantly strengthened the target detection capabilities of the OS-DETR model. The incremental improvements achieved by each component highlight their complementarity and synergy, collectively driving continuous advancements in model performance.

Discussion and conclusion

This study introduces an improved detection framework for brain tumor detection tasks in medical imaging, named OS-DETR (Orthogonal Channel Shuffle Detection Transformer). The framework incorporates OSNet as the backbone network and integrates deformable attention mechanisms along with an enhanced ICAD-IoU strategy within the DETR framework. OS-DETR demonstrates significant advantages in brain tumor detection tasks. Experimental results on the Br35H brain tumor dataset show that OS-DETR achieves superior performance across key metrics, with a Precision of 95.0%, Recall of 94.2%, mAP@50 of 95.7%, and mAP@50:95 of 74.2%, outperforming most mainstream object detection frameworks. This study highlights the potential of incorporating orthogonal characteristics and efficient encoding methods into deep learning models, offering a novel approach for small-sample object detection in medical imaging.

Although the proposed OS-DETR framework demonstrates significant advantages in brain tumor detection tasks, several areas require further research and optimization:

  1. Although OS-DETR performs well on small-scale datasets, its computational cost still has room for further optimization. Future work could incorporate techniques such as network pruning, quantization, and distillation to design more lightweight models, addressing the constraints on computational resources in practical medical scenarios.
  2. The current experiments are primarily conducted on the Br35H dataset. Future research could extend the application of OS-DETR to a broader range of medical imaging datasets, such as those involving different types of tumors or other diseases, to evaluate its generalization ability.
  3. OS-DETR explores filter initialization based on the Special Orthogonal Group method. Future studies could investigate alternative methods for generating orthogonal matrices to further enhance model performance.

References

  1. 1. Sahm F, Schrimpf D, Jones DTW, Meyer J, Kratz A, Reuss D, et al. Next-generation sequencing in routine brain tumor diagnostics enables an integrated diagnosis and identifies actionable targets. Acta Neuropathol. 2016;131(6):903–10. pmid:26671409
  2. 2. Jain R. Perfusion CT imaging of brain tumors: an overview. AJNR Am J Neuroradiol. 2011;32(9):1570–7. pmid:21051510
  3. 3. Abd-Ellah MK, Awad AI, Khalaf AAM, Hamed HFA. A review on brain tumor diagnosis from MRI images: practical implications, key achievements, and lessons learned. Magn Reson Imaging. 2019;61:300–18. pmid:31173851
  4. 4. Lienhart R, Maydt J. An extended set of haar-like features for rapid object detection. In: Proceedings of the international conference on image processing. vol. 1. IEEE; 2002. p. I.
  5. 5. Wang Y. An analysis of the Viola-Jones face detection algorithm. Image Processing On Line. 2014;4:128–48.
  6. 6. LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–324.
  7. 7. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Adv Neural Inf Process Syst. 2012;25.
  8. 8. Xiao Y, Tian Z, Yu J, Zhang Y, Liu S, Du S, et al. A review of object detection based on deep learning. Multim Tools Appl. 2020;79:23729–91.
  9. 9. Aishwarya A, Wadhwa P, Owais O, Vashisht V. A waste management technique to detect and separate non-biodegradable waste using machine learning and YOLO algorithm. In: 2021 11th International Conference on Cloud Computing, Data Science and Engineering (Confluence). IEEE; 2021. p. 443–7.
  10. 10. Tan L, Huangfu T, Wu L, Chen W. Comparison of RetinaNet, SSD, and YOLO v3 for real-time pill identification. BMC Med Inf Decis Making. 2021;21:1–11.
  11. 11. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S. End-to-end object detection with transformers. In: European Conference on Computer Vision. Springer; 2020. p. 213–29.
  12. 12. Vaswani A. Attention is all you need. Adv Neural Inf Process Syst. 2017;30:1–10.
  13. 13. Abiwinanda N, Hanif M, Hesaputra ST, Handayani A, Mengko TR. Brain tumor classification using convolutional neural network. In: World Congress on Medical Physics and Biomedical Engineering 2018: June 3-8, 2018, Prague, Czech Republic. vol. 1. Springer; 2019. p. 183–9.
  14. 14. Chen W, Liu B, Peng S, Sun J, Qiao X. S3D-UNet: separable 3D U-Net for brain tumor segmentation. In: Brainlesion: glioma, multiple sclerosis, stroke and traumatic brain injuries: 4th international workshop, BrainLes 2018, held in conjunction with MICCAI 2018, Granada, Spain, September 16, 2018, Revised Selected Papers, Part II 4. Springer; 2019. p. 358–68.
  15. 15. Deng K, Lei J, Li X, Shuai S, Lin M, Li S. An improved lightweight segmentation neural network for dermoscopic lesion images based on knowledge distillation. In: 2023 International Conference on Image Processing, Computer Vision and Machine Learning (ICICML). IEEE; 2023. p. 268–71.
  16. 16. Kang M, Ting CM, Ting FF, Phan RCW. Bgf-yolo: enhanced yolov8 with multiscale attentional feature fusion for brain tumor detection. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2024. p. 35–45.
  17. 17. Kang M, Ting CM, Ting FF, Phan RCW. RCS-YOLO: a fast and high-accuracy object detector for brain tumor detection. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer; 2023. p. 600–10.
  18. 18. Balakrishnan T, Sengar S. RepVGG-GELAN: Enhanced GELAN with VGG-STYLE ConvNets for brain tumour detection. arXiv preprint 2024. https://arxiv.org/abs/2405.03541
  19. 19. Yu Z, Guan Q, Yang J, Yang Z, Zhou Q, Chen Y. LSM-YOLO: a compact and effective ROI detector for medical detection. arXiv preprint 2024. https://arxiv.org/abs/2408.14087
  20. 20. Lin TY, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al. Microsoft coco: common objects in context. In: Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13. Springer; 2014. p. 740–55.
  21. 21. Wang W, Zhang J, Cao Y, Shen Y, Tao D. Towards data-efficient detection transformers. In: European conference on computer vision. Springer; 2022. p. 88–105.
  22. 22. Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2016. p. 779–88.
  23. 23. Girshick R. Fast r-cnn. In: Proceedings of the IEEE international conference on computer vision; 2015. p. 1440–8.
  24. 24. Redmon J, Farhadi A. YOLO9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2017. p. 7263–71.
  25. 25. Farhadi A, Redmon J. Yolov3: an incremental improvement. Computer vision and pattern recognition. 2018;1804:1–6.
  26. 26. Bochkovskiy A, Wang C, Liao H. Yolov4: Optimal speed and accuracy of object detection. arXiv preprint 2020. https://arxiv.org/abs/2004.10934
  27. 27. Wang C, Bochkovskiy A, Liao H. YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023. p. 7464–75.
  28. 28. Wang C, Liao H, Wu Y, Chen P, Hsieh J, Yeh I. CSPNet: A new backbone that can enhance learning capability of CNN. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 2020. p. 390–1.
  29. 29. Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D. Distance-IoU loss: faster and better learning for bounding box regression. In: Proceedings of the AAAI Conference on Artificial Intelligence. 2020;34:12993–3000.
  30. 30. Li X, Wang W, Wu L, Chen S, Hu X, Li J. Generalized focal loss: learning qualified and distributed bounding boxes for dense object detection. Adv Neural Inf Process Syst. 2020;33:21002–12.
  31. 31. Wang C, Yeh I, Liao H. Yolov9: learning what you want to learn using programmable gradient information. In: European conference on computer vision. Springer; 2025. p. 1–21.
  32. 32. Wang A, Chen H, Liu L, Chen K, Lin Z, Han J, et al. Yolov10: Real-time end-to-end object detection. arXiv preprint 2024. https://arxiv.org/abs/2405.14458
  33. 33. Jocher G, Qiu J. Ultralytics YOLO11; 2024. Available from: https://github.com/ultralytics/ultralytics
  34. 34. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T. An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint. 2020. https://arxiv.org/abs/2010.11929
  35. 35. Zhuohang S. Dynamic convolution-based image dehazing network. Multim Tools Appl. 2024;83(16):49039–56.
  36. 36. Xia Z, Pan X, Song S, Li LE, Huang G. Vision transformer with deformable attention. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022. p. 4794–803.
  37. 37. Wu Z, Wen Q, Yang F, Deng K. An innovative personalized recommendation approach based on deep learning and user review content. IEEE Access. 2024.
  38. 38. Zhu X, Su W, Lu L, Li B, Wang X, Dai J. Deformable detr: deformable transformers for end-to-end object detection. arXiv preprint. 2020. https://arxiv.org/abs/2010.04159
  39. 39. Zhao Y, Lv W, Xu S, Wei J, Wang G, Dang Q. Detrs beat yolos on real-time object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024. p. 16965–74.
  40. 40. Zhang X, Zhou X, Lin M, Sun J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018. p. 6848–56.
  41. 41. Ma N, Zhang X, Zheng HT, Sun J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. Proceedings of the European conference on computer vision (ECCV). 2018. p. 116–31.
  42. 42. Zhu L, Wang X, Ke Z, Zhang W, Lau R. Biformer: vision transformer with bi-level routing attention. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023. p. 10323–33.
  43. 43. Hu J, Shen L, Sun G. Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018. p. 7132–41.
  44. 44. Qin Z, Zhang P, Wu F, Li X. Fcanet: frequency channel attention networks. Proceedings of the IEEE/CVF International Conference on Computer Vision. p. 783–92.
  45. 45. Salman H, Parks C, Swan M, Gauch J. Orthonets: orthogonal channel attention networks. In: 2023 IEEE International Conference on Big Data (BigData). IEEE; 2023. p. 829–37.
  46. 46. Yue Y, Li N, Zhang G, Zhu Z, Liu X, Song S, et al. Automatic segmentation of esophageal gross tumor volume in 18F-FDG PET/CT images via GloD-LoATUNet. Comput Methods Programs Biomed. 2023;229:107266. pmid:36470035
  47. 47. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z. Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021. p. 10012–22.
  48. 48. Li F, Zhang H, Liu S, Guo J, Ni LM, Zhang L. Dn-detr: accelerate detr training by introducing query denoising. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022. p. 13619–27.
  49. 49. Yu J, Jiang Y, Wang Z, Cao Z, Huang T. Unitbox: an advanced object detection network. In: Proceedings of the 24th ACM International Conference on Multimedia. 2016. p. 516–20.
  50. 50. Zhang H, Xu C, Zhang S. Inner-IoU: more effective intersection over union loss with auxiliary bounding box. arXiv preprint 2023. https://arxiv.org/abs/2311.02877
  51. 51. Siliang M, Yong X. MPDIoU: a loss for efficient and accurate bounding box regression. arXiv preprint 2023. https://arxiv.org/abs/2307.07662
  52. 52. Hamada A. Br35H: brain tumor detection; 2020. Available from: https://www.kaggle.com/datasets/ahmedhamada0/brain-tumor-detection
  53. 53. Backert T. Brain tumor dataset; 2024. Available from: https://www.kaggle.com/datasets/tombackert/brain-tumor-mri-data