Notice: Function _load_textdomain_just_in_time was called incorrectly. Translation loading for the acf domain was triggered too early. This is usually an indicator for some code in the plugin or theme running too early. Translations should be loaded at the init action or later. Please see Debugging in WordPress for more information. (This message was added in version 6.7.0.) in /var/www/html/roboticslab/wp-includes/functions.php on line 6121
Speaker identification using three signal voice domains during human-robot interaction - RoboticsLab

Speaker identification using three signal voice domains during human-robot interaction

Download file: sigproc-sp
Download: BibTeX | Plain Text

Description

This LBR describes a novel method for user recognition in HRI, based on analyzing the peculiarities of users voices, and specially focused at being used in a robotic system. The method is inspired by acoustic fingerprinting techniques, and is made of two phases: a)enrollment in the system: the fea- tures of the user’s voice are stored in files called voiceprints, b)searching phase: the features extracted in real time are compared with the voiceprints using a pattern matching method to obtain the most likely user (match).

The audio samples are described thanks to features in three different signal domains: time, frequency, and time-frequency. Using the combination of these three domains has enabled significant increases in the accuracy of user identification compared to existing techniques. Several tests using an in- dependent user voice database show that only half a second of user voice is enough to identify the speaker. The recogni- tion is text-independent: users do not need to say a specific sentence (key-pass) to get identified for the robot.