Content area

Abstract

This research explores how Voice User Interface (VUI) and Graphical User Interface (GUI) can be combined into a multimodal user interface to improve usability and user experience, specifically for mental health professionals working with patients who have Neurodevelopmental Disorders (NDD). These patients often face significant challenges in expressing their emotions or thoughts, making it difficult for therapists to fully understand their mental state. To address this, the study focuses on creating an intuitive and effective interface that allows professionals to dynamically generate, modify, approve, and save AI-generated illustrations representing patients’ emotions and thoughts.

Using a Human-Centered Design (HCD) approach, the study involved iterative prototype development and evaluation. By combining qualitative and quantitative data collection through observation, surveys, and semi-structured interviews, feedback was gathered from participants using standardized tools like the System Usability Scale (SUS) and the User Experience Questionnaire (UEQ). Participants included psychologists, therapists, and medical students, providing a comprehensive perspective on the system’s usability and user experience.

The findings highlight the importance of integrating VUI and GUI to overcome the limitations of relying solely on one modality. The multimodal interface significantly improved accessibility and usability, offering a more inclusive solution for therapeutic communication. This thesis contributes to the field by providing actionable design recommendations for multimodal systems, with the potential to inform future research and development in mental health applications and beyond.

Details

Title
INVESTIGATING THE EFFICIENCY OF MULTIMODAL INTERACTIVE SYSTEMS Integrating Voice and Graphical User Interfaces for Supporting Mental Health Professionals
Author
Ruzindana, Fidele
Publication year
2025
Publisher
ProQuest Dissertations & Theses
ISBN
9798265410955
Source type
Dissertation or Thesis
Language of publication
English
ProQuest document ID
3275672610
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.