You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2025. This work is published under http://creativecommons.org/licenses/by/4.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Background
Voice biomarkers can effectively indicate early Mild Cognitive Impairment (MCI)1‐10. This study evaluates how each speech type performs in screening through active tasks. We also examine which combinations of these tasks best detect MCI.
Method
We designed three tasks to that end: Scripted Reading (Task 1) and Picture‐Based Question and Answer (Task 2) for structured speech, and Spontaneous Speech‐Based Storytelling (Task 3) for semi‐structured speech (Figure 1). We collected 129 speech samples from 21 participants. Using Recursive Feature Elimination with Cross‐Validation (RFECV), we selected 32 key features from over 1,700 acoustic and linguistic ones for classification (Figure 2). We framed the evaluation process as a combinatorial problem where y=f(Task1 Task 2 Task 3). Here, y indicates whether a person is at risk of MCI, and f represents the predictive model that we trained with the speech samples.
Result
The decoding analysis revealed that the combination of Task_1, 2 and 3 achieved the highest AUC performance (AUC 0.963; 100% ratio). Relative to this maximum performance, the combination of Task_2 + Task_3 achieved 0.869 (90.2%), followed by Task_3 0.822 (85.4%), and Task_1 + Task_3 0.817 (84.8%; Figure 3A). Among these, using all tasks resulted in the best classification
Results
an AUC of 0.963, specificity of 0.633, and sensitivity of 0.977 (Figure 3B). The feature importance analysis revealed that 1st quantile regression coefficient of MFCC[14] and 50% upper level time of MFCC delta[9] were the most significant features contributing to the classification model (Figure 3C). Lastly, the acoustic features derived from Task_3, pcm_zcr_sma_de_lpc4 (representing the fourth coefficient of linear predictive coding; LPC) and pcm_zcr_sma_de_lpgain (reflecting the signal‐to‐noise ratio based on energy distribution), showed clear differences between normal and patient groups (Figure 3D).
Conclusion
Among various task combinations, the combination of Task_1+Task_2+Task_3 consistently achieved the best results, with Task 3 being included in all high‐performance combinations. Feature importance analysis and target variable distribution further emphasized the greater contribution of acoustic features compared to linguistic features in classification performance.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
1 HAII Inc., Seoul, Korea, Republic of (South), HCI Lab, Yonsei University, Seoul, Korea, Republic of (South)
2 HAII Inc., Seoul, Korea, Republic of (South)
3 Graduate School of AI and Informatics, Sangmyung University, Seoul, Korea, Republic of (South), Department of Human‐Centred AI, Sangmyung University, Seoul, Korea, Republic of (South)
4 Graduate School of AI and Informatics, Sangmyung University, Seoul, Korea, Republic of (South), Department of Human‐Centred AI, Sangmyung University, Seoul, Korea, Republic of (South), Institute for Advanced Intelligence Study, Daejeon, Korea, Republic of (South)
5 Ewha Medical Research Institute, Ewha Womans University, Seoul, Korea, Republic of (South), Neurology, Ewha Womans University Mokdong Hospital, Ewha Womans University College of Medicine, Seoul, Korea, Republic of (South)





