It appears you don't have support to open PDFs in this web browser. To view this file, Open with your PDF reader
Abstract
Facial gestures are defined as the movements of facial muscles that occur in areas such as the eyebrows, cheeks, and mouth, and are utilized by humans to convey a multitude of meanings in different situations. These meanings can range from expressing emotions through facial expressions, performing speech-related actions with the mouth and teeth, to conveying specific linguistic constructs. Consequently, detecting facial gestures has emerged as a popular research topic that has gained increasing interest from academia and industry. In addition, acoustic sensing has won significant popularity in hand tracking, gesture recognition, and silent speech recognition due to its inexpensive, widespread availability and compact size. With previous research highlighting the potentials of ubiquitous acoustic sensing and the growing usage of smart speakers and earphones, hence we investigate the feasibility of adopting acoustic sensing for recognizing various facial gestures in three critical application domains: detecting emotional facial expressions, recognizing silent speech, and identifying facial information for sign language translation. Accurately recognizing emotional facial expressions can be a valuable means of understanding audience feedback and engagement with entertainment content. In light of this, the dissertation presents SonicFace, the first-ever system that utilizes a smart speaker to identify various subtle emotional facial expressions. To ensure privacy protection during speech interactions and also enhance resistance to ambient noise, the dissertation introduces EarCommand, a new earphone-based, hands-free interaction technique that employs silent speech commands recognition by leveraging the connection between ear canal deformation and articulator movements. As sign language is the primary communication mode for individuals with hearing impairments, the dissertation proposes TransASL, an innovative smart glasses based system that simultaneously translates hand gestures and facial expressions into natural language by utilizing two pairs of microphone and speaker mounted on the earphones. One pair captures front-facing signals for hand gestures, while the other pair records inner-facing signals for recognizing facial expressions. All these scientific explorations in this dissertation are expected to lay a foundation for facial gesture recognition in emotional facial expression, silent speech recognition based on mouth motion and facial information in ASL recognition with a convenient and ubiquitous method.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer