Inicio  /  Applied Sciences  /  Vol: 13 Par: 5 (2023)  /  Artículo
ARTÍCULO
TITULO

Voice Interaction Recognition Design in Real-Life Scenario Mobile Robot Applications

Shih-An Li    
Yu-Ying Liu    
Yun-Chien Chen    
Hsuan-Ming Feng    
Pi-Kang Shen and Yu-Che Wu    

Resumen

This paper designed a voice interactive robot system that can conveniently execute assigned service tasks in real-life scenarios. It is equipped without a microphone where users can control the robot with spoken commands; the voice commands are then recognized by a well-trained deep neural network model of automatic speech recognition (ASR), which enables the robot to execute and complete the command based on the navigation of a real-time simultaneous localization and mapping (SLAM) algorithm. The voice interaction recognition model is divided into two parts: (1) speaker separation and (2) ASR. The speaker separation is applied by a deep-learning system consisting of eight convolution layers, one LSTM layer, and two fully connected (FC) layers to separate the speaker?s voice. This model recognizes the speaker?s voice as a referrer that separates and holds the required voiceprint and removes noises from other people?s voiceprints. Its automatic speech recognition uses the novel sandwich-type conformer model with a stack of three layers, and combines convolution and self-attention to capture short-term and long-term interactions. Specifically, it contains a multi-head self-attention module to directly convert the voice data into text for command realization. The RGB-D vision-based camera uses a real-time appearance-based mapping algorithm to create the environment map and replace the localization with a visional odometer to allow the robot to navigate itself. Finally, the proposed ASR model was tested to check if the desired results will be obtained. Performance analysis was applied to determine the robot?s environment isolation and voice recognition abilities. The results showed that the practical robot system successfully completed the interactive service tasks in a real environment. This experiment demonstrates the outstanding performance with other ASR methods and voice control mobile robot systems. It also verified that the designed voice interaction recognition system enables the mobile robot to execute tasks in real-time, showing that it is a convenient way to complete the assigned service applications.

 Artículos similares

       
 
Jong-Gyu Shin, Ga-Young Choi, Han-Jeong Hwang and Sang-Ho Kim    
With the development of artificial intelligence technology, voice-based intelligent systems (VISs), such as AI speakers and virtual assistants, are intervening in human life. VISs are emerging in a new way, called human?AI interaction, which is different... ver más
Revista: Applied Sciences

 
Donghoon Oh, Jeong-Sik Park, Ji-Hwan Kim and Gil-Jin Jang    
Automatic speech recognition; chatbot; voice-assisted control; multimodal man-machine interaction systems.
Revista: Applied Sciences

 
I. I. Tolstikova,O. A. Ignatjeva,K. S. Kondratenko,A. V. Pletnev     Pág. 90 - 98
This paper explores the specifics of Generation Z perception of blended learning technology, which was actively introduced into the educational process during the pandemic caused by COVID-19. The motivation that prompted the team of authors to conduct su... ver más

 
Leonardo Bilic,Markus Ebner,Martin Ebner     Pág. pp. 226 - 232
An educational, interactive Amazon Alexa Skill called ?Österreich und Europa Spiel / Austria and Europe Game? was developed at Graz University of Technology for a German as well as English speaking audience. This Skills intent is to assist learning geogr... ver más

 
Sandi Ljubic,Damir Arbula     Pág. pp. 66 - 82
In the area of Human-Computer Interaction (HCI), the concept of Around Device Interaction (ADI) is gaining special attention nowadays. It provides the possibility of interacting with the device by making use of gestures made in the surrounding space, as ... ver más