Minimal representation of speech signals for generation of emotion speech and human-robot interaction

Heyoung Lee, Z. Zenn Bien

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this paper minimal representation of voiced speech based on decomposition into AM-FM components is proposed for generation of emotion speech. For the decomposition, firstly time-frequency boundaries of AM-FM components are estimated and secondary each AM-FM component is extracted by using the variable bandwidth filter [17] adaptive to the estimated time-frequency boundaries. Finally, two parameters, that is, instantaneous frequency and instantaneous amplitude of each AM-FM component are estimated. The set composed of instantaneous amplitudes and instantaneous frequencies is the minimal representation of voiced speech signals. The minimal representation is optimal feature set since the set describes effectively the biomechanical characteristics of the vocal codes and the vocal track. Raw speech signals are modified by changing the parameters for generation of emotion speech.

Original languageEnglish
Title of host publication16th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN
Pages137-140
Number of pages4
DOIs
StatePublished - 2007
Event16th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN - Jeju, Korea, Republic of
Duration: 26 Aug 200729 Aug 2007

Publication series

NameProceedings - IEEE International Workshop on Robot and Human Interactive Communication

Conference

Conference16th IEEE International Conference on Robot and Human Interactive Communication, RO-MAN
Country/TerritoryKorea, Republic of
CityJeju
Period26/08/0729/08/07

Fingerprint

Dive into the research topics of 'Minimal representation of speech signals for generation of emotion speech and human-robot interaction'. Together they form a unique fingerprint.

Cite this