A Prior-Guided Dual Branch Multi-Feature Fusion Network for Building Segmentation in Remote Sensing Images
Auteur(s): |
Yingbin Wu
Peng Zhao Fubo Wang Mingquan Zhou Shengling Geng Dan Zhang |
---|---|
Médium: | article de revue |
Langue(s): | anglais |
Publié dans: | Buildings, 2 juillet 2024, n. 7, v. 14 |
Page(s): | 2006 |
DOI: | 10.3390/buildings14072006 |
Abstrait: |
The domain of remote sensing image processing has witnessed remarkable advancements in recent years, with deep convolutional neural networks (CNNs) establishing themselves as a prominent approach for building segmentation. Despite the progress, traditional CNNs, which rely on convolution and pooling for feature extraction during the encoding phase, often fail to precisely delineate global pixel interactions, potentially leading to the loss of vital semantic details. Moreover, conventional CNN-based segmentation models frequently neglect the nuanced semantic differences between shallow and deep features during the decoding phase, which can result in subpar feature integration through rudimentary addition or concatenation techniques. Additionally, the unique boundary characteristics of buildings in remote sensing images, which offer a rich vein of prior information, have not been fully harnessed by traditional CNNs. This paper introduces an innovative approach to building segmentation in remote sensing images through a prior-guided dual branch multi-feature fusion network (PDBMFN). The network is composed of a prior-guided branch network (PBN) in the encoding process, a parallel dilated convolution module (PDCM) designed to incorporate prior information, and a multi-feature aggregation module (MAM) in the decoding process. The PBN leverages prior region and edge information derived from superpixels and edge maps to enhance edge detection accuracy during the encoding phase. The PDCM integrates features from both branches and applies dilated convolution across various scales to expand the receptive field and capture a more comprehensive semantic context. During the decoding phase, the MAM utilizes deep semantic information to direct the fusion of features, thereby optimizing segmentation efficacy. Through a sequence of aggregations, the MAM gradually merges deep and shallow semantic information, culminating in a more enriched and holistic feature representation. Extensive experiments are conducted across diverse datasets, such as WHU, Inria Aerial, and Massachusetts, revealing that PDBMFN outperforms other sophisticated methods in terms of segmentation accuracy. In the key segmentation metrics, including mIoU, precision, recall, and F1 score, PDBMFN shows a marked superiority over contemporary techniques. The ablation studies further substantiate the performance improvements conferred by the PBN’s prior information guidance and the efficacy of the PDCM and MAM modules. |
Copyright: | © 2024 by the authors; licensee MDPI, Basel, Switzerland. |
License: | Cette oeuvre a été publiée sous la license Creative Commons Attribution 4.0 (CC-BY 4.0). Il est autorisé de partager et adapter l'oeuvre tant que l'auteur est crédité et la license est indiquée (avec le lien ci-dessus). Vous devez aussi indiquer si des changements on été fait vis-à-vis de l'original. |
3.83 MB
- Informations
sur cette fiche - Reference-ID
10795238 - Publié(e) le:
01.09.2024 - Modifié(e) le:
01.09.2024