Abstract

Salient object detection is vital for non-specific class subject segmentation in computer vision applications. However, accurately segmenting foreground subjects with complex backgrounds and intricate boundaries remains a challenge for existing methods. To address these limitations, our study proposes SU2GE-Net, which introduces several novel improvements. We replace the traditional CNN-based backbone with the transformer-based Swin-TransformerV2, known for its effectiveness in capturing long-range dependencies and rich contextual information. To tackle under and over-attention phenomena, we introduce Gated Channel Transformation (GCT). Furthermore, we adopted an edge-based loss (Edge Loss) for network training to capture spatial-wise structural details. Additionally, we propose Training-only Augmentation Loss (TTA Loss) to enhance spatial stability using augmented data. Our method is evaluated using six common datasets, achieving an impressive Fβ score of 0.883 on DUTS-TE. Compared with other models, SU2GE-Net demonstrates excellent performance in various segmentation scenarios.

Details

Title
SU2GE-Net: a saliency-based approach for non-specific class foreground segmentation
Author
Lei, Xiaochun 1 ; Cai, Xiang 2 ; Lu, Linjun 2 ; Cui, Zihang 2 ; Jiang, Zetao 1 

 Guilin University of Electronic Technology, School of Computer Science and Information Security, GuiLin, China (GRID:grid.440723.6) (ISNI:0000 0001 0807 124X); Guilin University of Electronic Technology, Guangxi Key Laboratory of Image and Graphic Intelligent Processing, Guilin, China (GRID:grid.440723.6) (ISNI:0000 0001 0807 124X) 
 Guilin University of Electronic Technology, School of Computer Science and Information Security, GuiLin, China (GRID:grid.440723.6) (ISNI:0000 0001 0807 124X) 
Pages
13263
Publication year
2023
Publication date
2023
Publisher
Nature Publishing Group
e-ISSN
20452322
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2850929310
Copyright
© Springer Nature Limited 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.