Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

An improved Deeplabv3+ semantic segmentation algorithm with multiple loss constraints

  • Yunyan Wang,

    Roles Data curation, Formal analysis, Investigation

    Affiliations School of Electrical and Electronic Engineering, Hubei University of Technology, Wuhan, China, Hubei University of Technology Cooperative Innovation Center of Hubei Province for Efficient Use of Solar Energy, Wuhan, China

  • Chongyang Wang ,

    Roles Conceptualization, Writing – original draft

    18838960004@163.com

    Affiliation School of Electrical and Electronic Engineering, Hubei University of Technology, Wuhan, China

  • Huaxuan Wu,

    Roles Writing – review & editing

    Affiliation School of Electrical and Electronic Engineering, Hubei University of Technology, Wuhan, China

  • Peng Chen

    Roles Investigation

    Affiliation School of Electrical and Electronic Engineering, Hubei University of Technology, Wuhan, China

Abstract

Aiming at the problems of low segmentation accuracy and inaccurate object boundary segmentation in current semantic segmentation algorithms, a semantic segmentation algorithm using multiple loss function constraints and multi-level cascading residual structure is proposed. The multi-layer cascaded residual unit was used to increase the range of the network layer receptive field. A parallel network was constructed to extract different depth feature information, then different depth feature information and encoder output features are fused to obtain multiple outputs feature which build multiple losses with the label, thereby constraining the model to optimize the network. The proposed network was evaluated on Cityscapes and CamVid datasets. The experimental results show that the mean Intersection over Union ratio (MIoU) of the proposed algorithm is 3.07% and 3.59% higher than the original Deeplabv3+ algorithm, respectively.

1 Introduction

Image segmentation, as an important branch in the field of computer vision, can achieve accurate positioning and identification of object locations while reducing costs, and is widely used in unmanned driving, medical image processing, industrial robots and other fields [1, 2]. In 2015, the Fully Convolutional Neural Network (FCN) was proposed, which made deep learning formally used in the field of semantic segmentation [3]. FCN used VGG Convolutional Neural Network as the backbone Network. After removing all the full connection layers, a deconvolution layer is added at the end of the network to upsample the input image, and the dense pixel prediction results are obtained. Compared with the traditional segmentation algorithm [4], FCN has a great improvement in the segmentation accuracy, but it has poor segmentation effect in the segmentation of small target objects [5]. The problem of FCN is that it only relies on high-level feature information to classify pixels, and lacks the use of low-level feature images with rich detailed information, which leads to rough final segmentation results of the network. Therefore, how to rationally use the underlying feature information to enhance the semantic segmentation effect has become a hot research issue. Researchers have proposed some semantic segmentation models using codec structure, such as U-Net [6], Multi-Path Refinement Networks (Refinenet) and SegNet [7, 8]. The above segmentation model uses the codec structure to supplement the details lost by the model’s lower sampling in the process of up-sampling to improve the segmentation performance of the model [9]. Zhao et al. proposed the Pyramid Scene Parsing Network (PSPNet) semantic segmentation model, which can integrates multi-scale context information through the pyramid scene analysis network to obtain excellent segmentation results [10, 11]. In addition, using the attention mechanism to strengthen the network’s learning and attention to important features is also a way to improve the semantic segmentation effect. Danet uses the dual attention mechanism of space and channel to capture the semantic interdependence in space and channel dimensions, which significantly improves the segmentation effect [12, 13]. In addition, Chen et al. proposed the Deeplab series semantic segmentation model, in which V1 proposed to replace the pooling layer with void convolution to avoid the loss of detailed information and obtain a larger receptive field range. In addition, Conditional Random Field (CRF) was added at the end of the model to optimize the boundary segmentation [14]. In DeepLabV2, a void space pyramid module was proposed to obtain global semantic information at multiple scales [15]. Then, DeepLabV3 and DeepLabV3+ have improved the void space pyramid module again [16, 17] and replaced Inception as a new backbone network, which has become the most effective semantic segmentation algorithm in the current semantic segmentation field [18, 19]. Recently, Fu et al. [13] proposed a dual-focus network for scene segmentation, which can capture rich context dependencies based on self-focus mechanism. Specifically, they attached two types of focus modules to the expanded FCN, which modeled semantic interdependence in FDI. The location notice module selectively aggregates the features of each location by a weighted sum of the features of all locations. Huo et al. [20] proposed the concept of bar pooling and used horizontal average pooling and vertical average pooling to model global information in an attempt to approach the long-term dependency problem in semantic segmentation.

Through the analysis of the above semantic segmentation algorithms, it can be seen that the mainstream semantic segmentation algorithms improve the accuracy of segmentation algorithms by using the underlying feature information, extracting multi-scale context information and attention mechanism. [2123] Therefore, how to find appropriate underlying feature information to help restore images in the process of up-sampling, obtain more semantic information of scale and build appropriate attention mechanism module to enhance the dependency between pixels has become a hot research issue.

Starting from capturing multi-scale semantic information and reusing underlying features, this paper improves the basic residual unit and loss function of the original backbone network model. First, use the multi-level residual structure to obtain more scale semantic information. Secondly, construct a multi-loss function constraint model to use the underlying feature information in the optimization process. Finally, multiple datasets are used to verify the proposed algorithm.

This paper will introduce our proposed method from the five chapters. The first chapter introduces the latest research technology and research status. Chapter II shows the structure and function of ResNet101 and DeepLabV3+. Chapter III presents our method and introduces the experimental process in detail. Then Chapter IV is the experiment of the proposed method, including comparison with other algorithms and some visualization and segmentation results. Finally, Chapter V concludes the contributions and deficiencies of this article.

2 Related work

2.1 RESNET-101

For traditional convolutional neural networks, it was previously generally believed that the deeper the network was, the stronger the ability of nonlinear expression would be. When a shallow neural network achieves good results in a certain kind of task, more characteristic information can be obtained theoretically by adding a network layer at the back end. However, in fact, with the deepening of the network, the problem of network degradation will occur. Theoretically speaking, if the identity mapping is maintained in the subsequent network layers, that is, when it does not work, the effect of the deep network should be equal to that of the shallow network, and there should be no degradation problem. The reasonable explanation is that the deep network layer is difficult to maintain the identity mapping, resulting in the loss of learning objectives. Therefore, how to make the deep network layer keep the identity mapping is the key factor to solve the degradation problem. The residual structure is shown in Fig 1.

thumbnail
Fig 1. The structure diagram of residual.

Residual network uses jump connection to construct residual structure, which solves the degradation problem to some extent.

https://doi.org/10.1371/journal.pone.0261582.g001

It is assumed that the input of the residual unit is i, and the output after passing through three layers of convolutional layer is F(i). Its definition is shown in Eq (1).

(1)

In the equation, w1,w2,w3 is the weight parameter of the three-layer convolution layer, and σ is the activation function of ReLU. The actual output H(i) of the residual element is obtained after adding F(i) to the input i, as shown in Eq (2).

(2)

Generally speaking, the one-layer network can be regarded as y = H(i), and the output of each residual unit can be expressed as H(i) = F(i)+i. In the unit mapping, y = i is the observed value and H(i) corresponds to the predicted value, so F(i) is called the residual. The function of the residual unit is to skillfully convert the learning target into the residual part. As for the neural network, it is obvious that fitting F(i) = 0 is much easier than fitting H(i) = i, so the depth of the network layer of the residual network can reach far exceeds that of other network structures.

Benefit from the fact that the residual structure can solve the degradation problem caused by the deep network, the depth of the ResNet-101 network can reach 101 layers, and its structure is shown in Table 1. The ResNeT-101 network consists of five parts, which are: Conv1, Conv2_x, Conv3_x, Conv4_x and Conv5_x, where Conv1 is a convolutional layer with step size of 2 and scale of 7×7×64, while Conv2_x-Conv5_x is stacked by residual structure, which contains 3, 4, 23 and 3 residual blocks respectively. Each residual block contains 3 network layers, plus the full connection layer between Conv1 and the end of the network, so there are 101 network layers in total.

2.2 Deeplabv3+

The structure of Deeplabv3 + [18] is shown in Fig 2.

thumbnail
Fig 2. The structure diagram of DeppLabv3+.

The Deeplabv3 + adopts residual network as the backbone network, and uses encoding and decoding structure to improve semantic segmentation effect [18].

https://doi.org/10.1371/journal.pone.0261582.g002

Deeplabv3+ encoding structure is composed of ResNet-101 network and void space pyramid module, in which ResNet-101 is mainly divided into 5 parts. The first part is a convolutional layer of 7×7×64, and the other 4 parts contain 3, 4, 23, and 3 blocks respectively. As shown in Fig 3(A), each Block is composed of three convolutional layers plus shortcut connections. The void space pyramid pooling module is connected at the end of the ResNet-101 network to obtain the semantic information of multi-scale images. The decoding structure performs 4-fold bilinear up-sampling on the feature map output by the encoder network, and then merges it with the low-level feature information at the channel level to restore the spatial information lost during the downsampling process.

3 The proposed method

3.1 Res2net residual cell structure

We use Res2Net structure as the basic unit to construct the backbone network of semantic segmentation framework. As shown in Fig 3, the Res2Net structure on the right constructs a multi-layer similar residue-cascading structure in a single residual unit, which can obtain a larger receptive field range and more scale representation capability compared with the traditional residual unit structure on the left.

The specific operation is as follows:

  1. The feature graph output by 1×1 convolution kernel is evenly divided into S feature blocks i1~is in channel dimension;
  2. Starting from i2, the feature block performs concat operation on the number of channels after the next feature block goes through 3×3 convolution. Therefore, the output feature block starting from j2 contains all the feature information on its left, and with the increase of S, the receptive field range of the output feature graph is increased indirectly.
  3. j1~js is spliced and fused, and the number of channels is reduced by 1×1 convolution to get the output feature.

As can be seen from Fig 4, the reason for the indirect increase of receptive field in feature figure j1~js is as follows:

  1. For the feature block is-1, the feature graph js-1 is obtained after 3×3 convolution. It is assumed that the receptive field of the pixel on the feature block js-1 is 3×3.
  2. Since the 3×3 convolution is filled, the size of the feature block js-1 is equal to that of is, is+ is obtained after the fusion of the two, and js is obtained after the 3×3 convolution of is+.
  3. Analyzing js it can be seen from the figure that the receptive field range of the center point of js corresponds to the 3×3 Central area of the is+. The 3×3 region of is+ center contains the feature information in the 5×5 region of is-1 feature graph. Therefore, for js, its receptive field was indirectly expanded to 5×5 area. In addition, the receptive fields of different channel feature maps are different in size, which improves the ability to obtain the features of objects at different scales, so as to obtain more scale semantic features.

3.2 Optimization of multi-loss constrained models

In order to make the features of the shallow network gain the semantic segmentation results effectively, this paper establishes several loss functions to constrain the optimization direction of the model. As shown in Fig 5, the feature map Out finally output by the residual network is up-sampled twice and then channel fusion is performed with the feature map Feature map2 output by Block2, and then up-sampling is performed 8 times and the label image is constructed with a middle layer loss. Then use the feature map Feature map3 and the out feature map output by the residual network Block3 to perform channel fusion, and after 16 times up-sampling, build a high-level loss with the label image. Each loss part contains the feature information of different network depths, which helps the model use shallow feature information as much as possible to correct the model parameters in the process of backpropagation.

thumbnail
Fig 5. Optimization structure diagram of multi-loss constraint.

https://doi.org/10.1371/journal.pone.0261582.g005

The weight adaptive method is adopted to assign different weights to different loss parts, and the total loss function Lossg is obtained as shown in Eq 3.

(3)

Where Loss is the original loss of the network, Lossm is the middle-level loss, and Lossh is the high-level loss. The detailed mathematical expressions of the loss function are shown in Eqs 4 and 5.

(4)(5)

3.3 Improve the Deeplabv3+ algorithm process

Algorithm flow chart shown in Fig 6, we firstly using operations such as standardization, random cutting to deal with data sets. Secondly, the main network was constructed by using multi-level coupling residual structure, and the receptive fields of residual blocks were increased to extract features with multi-scale information. then use of empty space module further extract multiscale pyramid semantic information. Finally, feature images of extracted multi-scale semantic information and network features of different depths were fused to construct multiple loss functions to constrain-model optimization direction, and the prediction results were obtained by up-sampling.

thumbnail
Fig 6. Flow chart of semantic segmentation of urban road images.

https://doi.org/10.1371/journal.pone.0261582.g006

The original algorithm is improved in two directions, namely, obtaining multi-scale semantic information and reusing the underlying features, by using the improved strategies such as multi-level linkage residual structure and multi-loss function constraint, which improves the segmentation accuracy of the model and reduces the occurrence of false segmentation and discontinuous segmentation. The specific algorithm flow is shown as follows.

Step1: Preprocess the training data set by using random cropping, scale transformation, and standard normalization, increase the diversity of training samples, improve the generalization ability of the model, and convert the size of the input image to 513 or 321 according to the type of data.

Step2: Build the backbone network of the model, replace the residual units in the traditional residual network through the multi-level linkage residual structure, so as to obtain more scale feature information, and adjust the number of pooling layers according to the output step size.

Step3: Build the void space pyramid module to process the feature map output by the backbone network, and further obtain the multi-scale semantic information.

Step4: Feature fusion is carried out between the network output feature graphs of different depths in the backbone network and the final network output. After bilinear up-sampling of different degrees, multiple loss functions are constructed with the original label images to constrain the model optimization and realize the reuse of the underlying features.

4 Experiment

4.1 Datasets and evaluation criteria

In order to verify the effectiveness of the proposed algorithm, experimental data were used from the Cityscapes [24] dataset and CamVid [25, 26] dataset, Cityscapes dataset is divided into 19 categories in semantic segmentation task, including 3475 urban road images with a resolution of 2048*1024, including 2975 training sets and 500 test sets. The CamVid dataset contains 12 semantic categories (including unlabeled categories). There are 701 images in this dataset, including 367 training images, 101 validation images and 233 test images, with a resolution of 480*320 for each image.

In order to evaluate the effectiveness of the proposed algorithm, the average pixel accuracy and average crosswise ratio index were used to evaluate the experimental results. Mean pixel accuracy (mPA) is calculated as shown in Eq 6, which means to calculate the mean value of the proportion of correctly classified pixel points in each class.

(6)

The Intersection over Union (IOU) was calculated to calculate the degree of overlap between the real and predicted values. Its definition is shown in Eq 7. The Mean Intersection over Union (IOU) index was obtained by taking the Intersection over Union index of each class and averaging it. mIoU index intuitively reflects the quality of the segmentation effect.

(7)

4.2 Experiments settings

The experimental environment is Ubuntu16.04 system, the CPU is Intel-I7-7700, the memory is 16G, the GPU is GeForce GTX 1070, the video memory is 8G, and the deep learning framework is Pytorch 1.4.0. In order to ensure the authenticity of the improved network, the original network and the improved network are set the same hyperparameters, and the control variable method is used to prove the effectiveness of the improved algorithm. Due to the different characteristics of the two sets of data, some parameters of the two sets of data are different as follows: batch_size of Cityscapes dataset is set to 4, and the total number of iterations is 3e4; The CamVid dataset batch size is set to 6, and the total number of iterations is 2e4.

4.3 Experimental results and analysis

The Cityscapes dataset contains 19 categories of objects. The overall segmentation accuracy is shown in Table 2, from which it can be seen that the original Deeplabv3 + algorithm can achieve 94.11%, 78.66% and 69.08% for the segmentation of PA, mPA and mIoU index of the Cityscapes dataset.

thumbnail
Table 2. Overall segmentation results of Cityscapes (unit: %).

https://doi.org/10.1371/journal.pone.0261582.t002

After the construction of the multi-loss function constraint model optimization (Improvement Ⅰ), PA, mPA and mIoU are increased by 0.1%, 2.23% and 2% respectively. while after the replacement of the original residual structure by the multi-level linkage residual structure (Improvement Ⅱ), PA, mPA and mIoU indexes are increased by 0.32%, 3.25% and 2.77% respectively.

Table 3 shows the detailed segmentation results of the experiment on the Cityscapes dataset. It can be seen that, for the 19 categories contained in the dataset, the segmentation accuracy of the proposed algorithm in 18 specific categories is higher than that of the original algorithm, especially for the two categories of bus and Fence. The IoU ratio increases by 9.24% and 6%, respectively.

thumbnail
Table 3. Segmentation results of Cityscapes dataset experiment (unit: %).

https://doi.org/10.1371/journal.pone.0261582.t003

Table 4 shows the overall segmentation results of this algorithm and its ablation experiment on the CamVid dataset. It can be seen that the pixel accuracy, average pixel accuracy, and mean Intersection over Union index of this algorithm can reach 86.17%, 67.90%, and 55.95%, respectively. Compared with the original algorithm that uses a single improvement strategy, its three segmentation indexes have been improved, especially compared to the original Deeplabv3Plus algorithm, the PA, mPA, and mIoU indexes increased by 0.83%, 5.52%, and 3.59%, respectively.

thumbnail
Table 4. Overall segmentation results of CamVid dataset (unit: %).

https://doi.org/10.1371/journal.pone.0261582.t004

The detailed segmentation accuracy of each category of CamVid is shown in Table 5. It can be seen that among the 12 semantic categories contained in CamVid, the segmentation accuracy of this article is better than the original algorithm in 11 categories, especially in the two categories of Bicyclist and Fence, which are Increased by 10.03% and 10.86% respectively.

thumbnail
Table 5. Segmentation results of CamVid dataset experiment (unit: %).

https://doi.org/10.1371/journal.pone.0261582.t005

Fig 7 is a partial segmentation result diagram of the cityscape, where (a)- (e) are the original label image and the segmentation results corresponding to the 4 sets of ablation experiments.

thumbnail
Fig 7. Partial segmentation results of Cityscape dataset.

Comparing the first column of images, it can be seen from the labeled area of the yellow box in the figure that the truck head in the original algorithm segmentation result is mis-segmented, and the segmentation results of the Improved Ⅰ and Improved Ⅱ algorithms reduce the mis-segmented area. After applying two innovative methods at the same time, the segmentation area is minimized. In the second column of image labeling area, the original, Improved Ⅰ, and Improved Ⅱ algorithm segmentation results have the phenomenon of mis-segmentation of cars into trucks, and the proposed algorithm is more correct to segment the contours and categories of cars. In the third column of segmentation results, the bicycle contours in the original algorithm and the Improved Ⅰ are relatively fuzzy. In addition, there is a problem of discontinuous segmentation of street lights, and the proposed algorithm can get a clear bicycle contour shape. In the fourth column of images, it can be seen that part of the truck area in the original algorithm was mistakenly divided into cars and sky categories. The Improved Ⅰ and the Improved Ⅱ algorithm reduced the mis- segmented area, and the proposed algorithm segmented the truck area completely and correctly.

https://doi.org/10.1371/journal.pone.0261582.g007

Fig 8 shows the segmentation results of the algorithm and its ablation comparison test on some CamVid datasets.

thumbnail
Fig 8. Results of partial segmentation of CamVid dataset.

From the image segmentation results in column (a), it can be seen that the Deeplabv3plus algorithm, the Improved Ⅰ and the Improved Ⅱ have different degrees of under-segmentation on the street light poles, and the complete street light pole area cannot be segmented. Compared with the above three methods, the proposed algorithm can segment a more complete street light pole area and get more precise details. In view of the images in column (b) and their segmentation results, Deeplabv3plus, Improved Ⅰ, and Improved Ⅱ algorithms have different degrees of mis-segmentation in the segmentation of the motor vehicle lane area, and part of the motor vehicle lane area is mistakenly divided into sidewalks. In practical applications, the misclassification of these two semantic categories is a fatal error. The algorithm of this paper, which combines two improved strategies, correctly divides the motor vehicle lanes and sidewalks, and solves the mis-segmentation phenomenon. For the images in column (c) and their segmentation results, the sidewalk on the left and the street lights on the right have different degrees of mis-segmentation, especially in the segmentation results of Deeplabv3plus and the improved algorithm. Although the Improved Ⅱ segmented the sidewalk area relatively correctly, the streetlight area marked on the right side was seriously under-segmented. The segmentation result of the algorithm in this paper is still relatively rough compared to the label, but compared with the above segmentation result, the error segmentation phenomenon is reduced. In the image (d) and a series of segmentation results, Deeplabv3plus and the Improved Ⅰ algorithm are not accurate in the segmentation of the marked sidewalk area, while the Improved Ⅱ algorithm and our algorithm obtain more accurate segmentation results.

https://doi.org/10.1371/journal.pone.0261582.g008

5 Conclusion

In this paper, an improved Deeplabv3 + algorithm combined with multi-loss constraint model optimization is proposed to solve the problem that the traditional Deeplabv3 + image semantic segmentation algorithm can not reuse the multi-scale feature information and the underlying feature sufficiently. The specific contributions of this paper are summarized as follows:

  1. This paper proposes a solution to replace ResNet101 with Res2Net. Compared with the traditional residual unit structure, it can obtain a larger receptive field range and more scale representation ability to solve the problem of poor basic feature extraction ability of the backbone network;
  2. This paper innovatively constructs a new loss function composed of multiple loss functions for the network to make the network converge more effectively;
  3. This paper combines Res2Net, the new loss function formed by multiple loss functions and the pyramid space module together, and proposes a new more efficient network.

The improvement of the algorithm depends on the multi-layer class residual structure to acquire better features of representation, and the multi-loss function is used to limit the direction of model optimization in the process of back propagation, which causes two problems: first, model training consumes more computing resources and reduces real-time performance. Second, the construction of multiple loss function depends on the quality of artificial design, so it is difficult to ensure the effectiveness of multiple loss function.

References

  1. 1. Ye L, Duan T, Zhu J. Neural network-based semantic segmentation model for robot perception of driverless vision. IET Cyber-systems and Robotics. 2020 2020-01-01;2(4):190–6.
  2. 2. Peng Huang, Qi Zheng, Chao Liang. Review of image segmentation methods [J/OL]. Journal of Wuhan University (Natural Science Edition). 2020;66(6):519–31.
  3. 3. J. L, E. S, T. D, ^editors. Fully convolutional networks for semantic segmentation. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2015 0007-12-20. Pub Place; Year Published.
  4. 4. J. K, J. KL, K. ML, ^editors. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2016 2016-01-01. Pub Place; Year Published.
  5. 5. W. S, R. W. Fully Convolutional Networks for Semantic Segmentation of Very High Resolution Remotely Sensed Images Combined With DSM. IEEE GEOSCI REMOTE S. 2018 2018-01-01;15(3):474–8.
  6. 6. Ronneberger O., Fischer P., and Brox T., “U-Net: Convolutional networks for biomedical image segmentation,” 2015, arXiv:1505.04597. [Online]. Available: http://arxiv.org/abs/1505.04597.
  7. 7. V. B, A. K, R. C. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE T PATTERN ANAL. 2017 2017-01-01;39(12):2481–95. pmid:28060704
  8. 8. G. L, A. M, C. S, I. R, ^editors. RefineNet: Multi-path Refinement Networks for High-Resolution Semantic Segmentation. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017 2017-01-01. Pub Place; Year Published.
  9. 9. Lifu Chen, Yanzhi Liu, Peng Zhang, Zhihui Yuan, Xuemin Xing. A road extraction algorithm for high-resolution SAR image with multiple features based on multi-path Refinenet [J]. Computer science. 2020 2020-03-15;47(03):156–61.
  10. 10. H. Z, J. S, X. Q, X. W, J. J, ^editors. Pyramid Scene Parsing Network. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017 2017-01-01. Pub Place; Year Published.
  11. 11. Dan Gao, Jianying Chen, Ying Xie. A-PSPNet: A PSPNet Image Semantic Segmentation Model with Attention Mechanism [J]. Journal of China Academy of Electronics, 2020, 15(06): 518–523.
  12. 12. J. F, J. L, H. T, Y. L, Y. B, Z. F, et al., ^editors. Dual Attention Network for Scene Segmentation. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2019 2019-01-01. Pub Place; Year Published.
  13. 13. J. F, J. L, J. J, Y. L, Y. B, H. L. Scene Segmentation With Dual Relation-Aware Attention Network. IEEE T NEUR NET LEAR. 2021 2021-01-01;32(6):2547–60. pmid:32745005
  14. 14. Chen L, Papandreou G, Kokkinos I, Murphy K, Yuille AL, ^editors. Semantic image segmentation with deep convolutional nets and fully connected CRFs; 2015 2015-01-01; San Diego, CA, United states. Pub Place: International Conference on Learning Representations, ICLR; Year Published. https://doi.org/10.1109/TPAMI.2017.2699184 pmid:28463186
  15. 15. Chen L, Papandreou G, Kokkinos I, Murphy K, Yuille AL. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE T PATTERN ANAL. 2018 2018-01-01;40(4):834–48.
  16. 16. Chen L, Papandreou G, Schroff F, Adam H. Rethinking Atrous Convolution for Semantic Image Segmentation. 2017 2017-01-01.
  17. 17. C. L, L. C, F. S, H. A, W. H, A. LY, et al., ^editors. Auto-DeepLab: Hierarchical Neural Architecture Search for Semantic Image Segmentation. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2019 2019-01-01. Pub Place; Year Published.
  18. 18. Chen L.-C., Zhu Y., Papandreou G., Schroff F., and Adam H., “Encoder-decoder with atrous separable convolution for semantic image segmentation,” 2018, arXiv:1802.02611. [Online]. Available: http://arxiv.org/abs/1802.02611.
  19. 19. Roy Choudhury A, Vanguri R, Jambawalikar SR, Kumar P. Segmentation of Brain Tumors Using DeepLabv3. Cham: Springer International Publishing; 2019. p. 154–67.
  20. 20. Q. H, L. Z, M. MC, J. F, ^editors. Strip Pooling: Rethinking Spatial Pooling for Scene Parsing. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2020 2020-01-01. Pub Place; Year Published.
  21. 21. Wang G, Zhai Q. Feature fusion network based on strip pooling. SCI REP-UK. 2021;11(1). pmid:34711889
  22. 22. Q. Z, H. W. A Novel Data-based Stochastic Distribution Control for Non-Gaussian Stochastic Systems. IEEE T AUTOMAT CONTR. 2021 2021-01-01:1.
  23. 23. Liu Y, Zhang Q, Yue H. Stochastic distribution tracking control for stochastic non-linear systems via probability density function vectorisation. T I MEAS CONTROL. 2021;43(14):3149–57.
  24. 24. CORDTS M , OMRAN M, RAMOS , et al. The cityscapes dataset for semantic urban scene understanding[C]. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016:3213–3223.
  25. 25. BROSTOW G J, SHOTTON J, FAUQUEUR J, et al. Segmentation and recognition using structure from motion point clouds[C]. European conference on computer vision, 2008:44–57.
  26. 26. BROSTOW G J, FAUQUEUR J, CIPOLLA R. Semantic object classes in video: A high-definition ground truth database[J]. Pattern Recognition Letters, 2009, 30(2):88–97.