Sensitive Talking Heads

Thomas S. Huang, Mark A. Hasegawa-Johnson, Stephen M. Chu, Zhihong Zeng, Hao Tang

Research output: Contribution to journalArticlepeer-review

Abstract

A description with regards to the experiments done in emotive spoken language user interfaces is given. It has been found out that when the use of multimodal, synthesizing, and recognizing information has been optimized in both the audio and video modalities, there has been an improvement when it comes to recognition accuracy and synthesis quality. Specific topics being covered include: the speech and emotion recognition by humans; the automatic audiovisual speech and emotion recognition; the audiovisual speech synthesis; the emotive prosody; and finally the emotionally nuanced audiovisual speech.

Original languageEnglish (US)
Pages (from-to)67-72
Number of pages6
JournalIEEE Signal Processing Magazine
Volume26
Issue number4
DOIs
StatePublished - 2009

ASJC Scopus subject areas

  • Signal Processing
  • Electrical and Electronic Engineering
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Sensitive Talking Heads'. Together they form a unique fingerprint.

Cite this