go back

Robust intonation pattern classification in human robot interaction

Martin Heckmann and Kazuhiro Nakadai, "Robust intonation pattern classification in human robot interaction", Proc. INTERSPEECH, 2011.

Abstract

We present a system for the classification of intonation patterns in human robot interaction. The system distinguishes questions from other types of utterances and can deal with additional reverberations, background noise, as well as music interfering with the speech signal. The main building blocks of our system are a multi channel source separation, robust fundamental frequency extraction and tracking, segmentation of the speech signal, and classification of the fundamental frequency pattern of the last speech segment. We evaluate the system with Japanese sentences which are ambiguous without intonation information in a realistic human robot interaction scenario. Distortions present in the speech signal are room reverberations, background noise, and a music source at 60 degree. Despite the challenging task our system is able to classify the intonation pattern with good accuracy. With several experiments we evaluate the contribution of the different aspects of our system.



Download Bibtex file Download PDF

Search