Content area

Abstract

Human Activity Recognition (HAR) plays a crucial role in identifying and digitizing human behaviors. Among various approaches, sound-based HAR offers distinct advantages, such as overcoming visual limitations and enabling recognition in diverse environments. This study introduces an innovative application of sound segmentation with SegNet, originally designed for image segmentation, to sound-based HAR. Traditionally, labeling sound data has been challenging due to its limited scope, often restricted to specific events or time frames. To address this issue, a novel labeling approach was developed, allowing detailed annotations across the entire temporal and frequency domains. This method facilitates the use of SegNet, which requires pixel-level labeling for accurate segmentation, leading to more granular and explainable activity recognition. A dataset comprising six distinct human activities—speech, groaning, screaming, coughing, toilet and snoring—was constructed to enable comprehensive evaluation. The trained neural network, utilizing this annotated dataset, achieved F1 scores ranging from 0.68 to 0.95. The model’s practical applicability was further validated through recognition tests conducted in a professional office environment. This study presents a novel framework for quantifying daily human activities through sound segmentation, contributing to advancements in intelligent system technology.

Details

1009240
Title
Granular and explainable human activity recognition through sound segmentation and deep learning
Author
Kim, Jisoo 1   VIAFID ORCID Logo  ; Yoo, Byounghyun 2   VIAFID ORCID Logo 

 Department of Artificial Intelligence, Jeju National University , 102 Jejudaehak-ro, Jeju-si, Jeju Special Self-Governing Province, 63243 , South Korea 
 Intelligence and Interaction Research Center, Korea Institute of Science and Technology , 5 Hwarangro14-gil, Seongbuk-gu, Seoul 02792 , South Korea 
Volume
12
Issue
8
Pages
252-269
Publication year
2025
Publication date
Aug 2025
Publisher
Oxford University Press
Place of publication
Oxford
Country of publication
United Kingdom
ISSN
22885048
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2025-07-28
Milestone dates
2025-03-18 (Received); 2025-07-21 (Accepted); 2025-07-17 (Rev-recd); 2025-08-22 (Corrected)
Publication history
 
 
   First posting date
28 Jul 2025
ProQuest document ID
3242038496
Document URL
https://www.proquest.com/scholarly-journals/granular-explainable-human-activity-recognition/docview/3242038496/se-2?accountid=208611
Copyright
© The Author(s) 2025. Published by Oxford University Press on behalf of the Society for Computational Design and Engineering. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-08-27
Database
3 databases
  • Coronavirus Research Database
  • ProQuest One Academic
  • ProQuest One Academic