Content area

Abstract

Action recognition in videos has become crucial in computer vision because of its diverse applications, such as multimedia indexing and surveillance in public environments. The incorporation of attention mechanisms into deep learning has gained considerable attention. This approach aims to emulate the human visual processing system by enabling models to focus on pertinent aspects of a scene and derive significant insights. This study introduces an advanced soft attention mechanism designed to enhance the CNN-sLSTM architecture for recognizing human actions in videos. We used the VGG19 convolutional neural network to extract spatial features from the video frames, whereas the sLSTM network models the temporal relationships between frames. The performance of our model was assessed using two widely used datasets, HMDB-51 and UCF-101, with precision as the key evaluation metric. Our results indicate substantial improvements, achieving accuracy scores of 53.12% (base approach) and 67.18% (with attention) for HMDB-51 and 83.98% (base approach) and 94.15% (with attention) for UCF-101. These results underscore the effectiveness of the proposed soft attention mechanism in improving the performance of video action recognition models.

Details

1009240
Title
Improving Human Action Recognition in Videos with CNN– sLSTM and Soft Attention Mechanism
Author
Khaled, Merit 1 ; Mohammed, Beladgham 1 

 Laboratory of TIT, Department of Electrical Engineering, Tahri Mohammed University of Bechar, Algeria 
Publication title
Volume
21
Issue
1
Pages
122-137
Number of pages
17
Publication year
2025
Publication date
2025
Publisher
Engineering and Scientific Research Groups
Place of publication
Paris
Country of publication
France
e-ISSN
11125209
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
ProQuest document ID
3224658062
Document URL
https://www.proquest.com/scholarly-journals/improving-human-action-recognition-videos-with/docview/3224658062/se-2?accountid=208611
Copyright
© 2025. This work is published under https://creativecommons.org/licenses/by/4.0/legalcode (the"License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-11-14
Database
ProQuest One Academic