Abstract

This study presents a hybrid architecture tailored for semantic segmentation challenges, mainly targeting the water area extraction for flood detection and monitoring. The model integrates an efficient transformer-based encoder, utilizing an efficient multi-head self-attention module for capturing hierarchical feature maps through a ‘downsample-upsample’ strategy. The proposed decoder architecture comprises one feature refinement head block and three CNN-based dual-branch context blocks. The convolutional block attention module is employed within the feature refinement head block to refine feature representation. The depth-wise separable atrous spatial pyramid pooling module is central to this architecture, facilitating efficient multi-scale contextual information capture. Compared to the state-of-the-art models, our model and the PSPNet model obtained the highest precision, recall, and F1-scores of above 80%, and mIoU surpassing 70%. The proposed method outperformed PSPNet in recall, F1-score, mIoU, and pixel accuracy, albeit with a slight deficit in precision. In terms of scale and efficiency, compared to the PSPNet model, our model has lower complexity and slightly higher inference speed, highlighting its effectiveness and efficiency in the water area segmentation for flood detection. The source code is available at https://github.com/manhhv87/mmsegmentation.git.

Details

Title
Enhanced deep learning-based water area segmentation for flood detection and monitoring
Author
Pham, Thang M 1 ; Do, Nam 1 ; Bui, Hanh T 2 ; Hoang, Manh V 1   VIAFID ORCID Logo 

 Faculty of Engineering Mechanics and Automation, University of Engineering and Technology, Vietnam National University , Hanoi 10000, Vietnam 
 Faculty of Fundamental Science, Phenikaa University , Yen Nghia, Ha-Dong district, Hanoi 10000, Vietnam 
First page
045025
Publication year
2024
Publication date
Dec 2024
Publisher
IOP Publishing
e-ISSN
26322153
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3121582496
Copyright
© 2024 The Author(s). Published by IOP Publishing Ltd. This work is published under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.