Full text

Turn on search term navigation

© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.

Abstract

Advancements in music emotion prediction are driving AI-driven algorithmic composition, enabling the generation of complex melodies. However, bridging neural and auditory domains remains challenging due to the semantic gap between brain-derived low-level features and high-level musical concepts, making alignment computationally demanding. This study proposes a deep learning framework for generating MIDI sequences aligned with labeled emotion predictions through supervised feature extraction from neural and auditory domains. EEGNet is employed to process neural data, while an autoencoder-based piano algorithm handles auditory data. To address modality heterogeneity, Centered Kernel Alignment is incorporated to enhance the separation of emotional states. Furthermore, regression between feature domains is applied to reduce intra-subject variability in extracted Electroencephalography (EEG) patterns, followed by the clustering of latent auditory representations into denser partitions to improve MIDI reconstruction quality. Using musical metrics, evaluation on real-world data shows that the proposed approach improves emotion classification (namely, between arousal and valence) and the system’s ability to produce MIDI sequences that better preserve temporal alignment, tonal consistency, and structural integrity. Subject-specific analysis reveals that subjects with stronger imagery paradigms produced higher-quality MIDI outputs, as their neural patterns aligned more closely with the training data. In contrast, subjects with weaker performance exhibited auditory data that were less consistent.

Details

Title
EEG-Based Music Emotion Prediction Using Supervised Feature Extraction for MIDI Generation
Author
Gomez-Morales, Oscar 1   VIAFID ORCID Logo  ; Perez-Nastar, Hernan 2 ; Andrés Marino Álvarez-Meza 2   VIAFID ORCID Logo  ; Torres-Cardona, Héctor 3   VIAFID ORCID Logo  ; Castellanos-Dominguez, Germán 2   VIAFID ORCID Logo 

 Faculty of Systems and Telecommunications, Universidad Estatal Península de Santa Elena, Avda. La Libertad, Santa Elena 7047, Ecuador; [email protected] 
 Signal Processing and Recognition Group, Universidad Nacional de Colombia, Manizales 170003, Colombia; [email protected] (A.M.Á.-M.); [email protected] (G.C.-D.) 
 Transmedia Research Center, Universidad de Caldas, Manizales 170003, Colombia; [email protected] 
First page
1471
Publication year
2025
Publication date
2025
Publisher
MDPI AG
e-ISSN
14248220
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
3176350326
Copyright
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.