Qr code
News Official network 中文
Chen Guanzhou


Main positions:助理研究员
Gender:Male
Status:Employed
School/Department:测绘遥感信息工程国家重点实验室
  • Discipline: Photogrammetry and Remote Sensing
  • Click: times

    Open Time:..

    The Last Update Time:..

    Current position: Home >> Scientific Research >> Paper Publications

    SDFCNv2: An Improved FCN Framework for Remote Sensing Images Semantic Segmentation

    Hits : Praise

    DOI number:10.3390/rs13234902

    Journal:REMOTE SENSING

    Key Words:fully convolutional networks (FCNs),convolutional neural networks (CNNs),deep learning,semantic segmentation,remote sensing,SDFCN

    Abstract:Semantic segmentation is a fundamental task in remote sensing image analysis (RSIA). Fully convolutional networks (FCNs) have achieved state-of-the-art performance in the task of semantic segmentation of natural scene images. However, due to distinctive differences between natural scene images and remotely-sensed (RS) images, FCN-based semantic segmentation methods from the field of computer vision cannot achieve promising performances on RS images without modifications. In previous work, we proposed an RS image semantic segmentation framework SDFCNv1, combined with a majority voting postprocessing method. Nevertheless, it still has some drawbacks, such as small receptive field and large number of parameters. In this paper, we propose an improved semantic segmentation framework SDFCNv2 based on SDFCNv1, to conduct optimal semantic segmentation on RS images. We first construct a novel FCN model with hybrid basic convolutional (HBC) blocks and spatial-channel-fusion squeeze-and-excitation (SCFSE) modules, which occupies a larger receptive field and fewer network model parameters. We also put forward a data augmentation method based on spectral-specific stochastic-gamma-transform-based (SSSGT-based) during the model training process to improve generalizability of our model. Besides, we design a mask-weighted voting decision fusion postprocessing algorithm for image segmentation on overlarge RS images. We conducted several comparative experiments on two public datasets and a real surveying and mapping dataset. Extensive experimental results demonstrate that compared with the SDFCNv1 framework, our SDFCNv2 framework can increase the mIoU metric by up to 5.22% while only using about half of parameters.

    Co-author:Xiaoliang Tan,Beibei Guo,Kun Zhu,Puyun Liao,Tong Wang,Qing Wang

    Indexed by:Journal paper

    Correspondence Author:Xiaodong Zhang

    Document Type:J

    Volume:13

    Issue:23

    Translation or Not:no

    Date of Publication:2021-11-03

    Included Journals:SCI