Minimal Representation of Speech Signals for Generation of Emotion Speech and Human-Robot Interaction

In this paper minimal representation of voiced speech based on decomposition into AM-FM components is proposed for generation of emotion speech. For the decomposition, firstly time-frequency boundaries of AM-FM components are estimated and secondary each AM-FM component is extracted by using the var...

Full description

Saved in:
Bibliographic Details
Published in:RO-MAN 2007 - The 16th IEEE International Symposium on Robot and Human Interactive Communication pp. 137 - 140
Main Authors: Heyoung Lee, Bien, Z.Z.
Format: Conference Proceeding
Language:English
Published: IEEE 01-08-2007
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In this paper minimal representation of voiced speech based on decomposition into AM-FM components is proposed for generation of emotion speech. For the decomposition, firstly time-frequency boundaries of AM-FM components are estimated and secondary each AM-FM component is extracted by using the variable bandwidth filter adaptive to the estimated time-frequency boundaries. Finally, two parameters, that is, instantaneous frequency and instantaneous amplitude of each AM-FM component are estimated. The set composed of instantaneous amplitudes and instantaneous frequencies is the minimal representation of voiced speech signals. The minimal representation is optimal feature set since the set describes effectively the biomechanical characteristics of the vocal codes and the vocal track. Raw speech signals are modified by changing the parameters for generation of emotion speech.
ISBN:9781424416349
1424416345
ISSN:1944-9445
1944-9437
DOI:10.1109/ROMAN.2007.4415068