Next Article in Journal
Neural Image Analysis for the Determination of Total and Volatile Solids in a Composted Sewage Sludge and Maize Straw Mixture
Next Article in Special Issue
Data-Driven Model for Identifying Factors Influencing Electric Vehicle Charging Demand: A Comparative Analysis of Early- and Maturity-Phases of Electric Vehicle Programs in Korea
Previous Article in Journal
Simplified Calculation of Shear Rotations for First-Order Shear Deformation Theory in Deep Bridge Beams
Previous Article in Special Issue
A Research on Fault Diagnosis of a USV Thruster Based on PCA and Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Voice Interaction Recognition Design in Real-Life Scenario Mobile Robot Applications

1
Department of Electrical and Computer Engineering, Tamkang University, Taipei City 10650, Taiwan
2
Department of Computer Science and Information Engineering, National Quemoy University, Kinmen County 89250, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(5), 3359; https://doi.org/10.3390/app13053359
Submission received: 4 February 2023 / Revised: 3 March 2023 / Accepted: 5 March 2023 / Published: 6 March 2023
(This article belongs to the Special Issue Selected Papers from IMETI 2021)

Abstract

:
This paper designed a voice interactive robot system that can conveniently execute assigned service tasks in real-life scenarios. It is equipped without a microphone where users can control the robot with spoken commands; the voice commands are then recognized by a well-trained deep neural network model of automatic speech recognition (ASR), which enables the robot to execute and complete the command based on the navigation of a real-time simultaneous localization and mapping (SLAM) algorithm. The voice interaction recognition model is divided into two parts: (1) speaker separation and (2) ASR. The speaker separation is applied by a deep-learning system consisting of eight convolution layers, one LSTM layer, and two fully connected (FC) layers to separate the speaker’s voice. This model recognizes the speaker’s voice as a referrer that separates and holds the required voiceprint and removes noises from other people’s voiceprints. Its automatic speech recognition uses the novel sandwich-type conformer model with a stack of three layers, and combines convolution and self-attention to capture short-term and long-term interactions. Specifically, it contains a multi-head self-attention module to directly convert the voice data into text for command realization. The RGB-D vision-based camera uses a real-time appearance-based mapping algorithm to create the environment map and replace the localization with a visional odometer to allow the robot to navigate itself. Finally, the proposed ASR model was tested to check if the desired results will be obtained. Performance analysis was applied to determine the robot’s environment isolation and voice recognition abilities. The results showed that the practical robot system successfully completed the interactive service tasks in a real environment. This experiment demonstrates the outstanding performance with other ASR methods and voice control mobile robot systems. It also verified that the designed voice interaction recognition system enables the mobile robot to execute tasks in real-time, showing that it is a convenient way to complete the assigned service applications.

1. Introduction

In recent years, artificial intelligence (AI) has led to breakthroughs in linear predictive coding (LPC) and dynamic time warp (DTW) in the field of speech.
The most important breakthrough in the history of speech technologies dealing with large amounts of data is the application of hidden Markov models (HMM). Today’s automatic speech recognition models not only have the framework of HMM, but other techniques related to artificial intelligence such as deep neural networks (DNN) or recurrent neural networks (RNN) are also heavily used in this field. The current popular technologies in speech-related fields are broadly classified as automatic speech recognition (ASR), natural language processing, text-to-speech (TTS), speaker verification, and speaker recognition. These can be combined to build more complex applications such as speech-to-speech translation. Voice technology is currently being used in many areas; it is utilized as voice assistants, in ticketing systems, chatting robots, real-time voice translation, etc.
The topic of human–computer interaction has always been of great interest to everyone in the applications of the mobile robot [1], and speech-related technologies are also heavily used in this topic. Human–computer interaction is the interaction and communication between humans and machines. The most intuitive and simplest communication of humans is through speech. In the past, humans have relied on interfaces to operate robots, and it takes time for the elderly or children to learn, and become familiar with how to operate them. If the robot can directly listen and follow the user’s voice command to perform service tasks, it will be a lot easier and more convenient for humans. For a more complete description in the advantages of human–robot interaction, please refer to [2].
This research withdrew from the traditional manual interface, but focused on directly controlling the robot through a voice command to perform specific tasks. A human–interactive robot is an inevitable trend. A service robot that has functions of automatic speech recognition (ASR), natural language understanding, image recognition, and interactive communication will approach future development.
The simplest case of the speaker separation model is to disentangle two mixed voice signals with the same dimension of input and output. In recent years, deep clustering [3], TasNet [4,5], and permutation invariant training (PIT) [6] models have become popular speaker separation models. Below is a brief discussion of these models.
Deep clustering uses the concept of k-means clustering, which first converts a sound signal into a spectrum and then, translates the spectrum into a vector. The vector is composed of the units calculated in the spectrum and the surrounding units, so the spectrum is transformed from a flat matrix at the beginning to a matrix-like a cube; thus, one dimension is added. Next, the transformed matrix is divided into two categories using k-means clustering, where the number of speakers is assumed to be two. The k-means clustering method is based on the principle of aggregating similar vectors and the result of clustering produces two sets of masks, which will be multiplied by the original input signal to produce two separate voice signals.
TasNet or the time-domain audio-separation network has two main frameworks: Bi-LSTM and CNN. The original one, which uses Bi-LSTM as the core has a long training time because of the LSTM; thus, CNN was later added to improve the original architecture. TasNet is divided into three main parts: encoder, separator, and decoder. Initially, the sound signal is generated as a matrix by the encoder, which is like a Fourier transform that converts the sound signal into a spectrum; the difference is that the three parts of TasNet are all trained. The encoder output matrix is generated by the separator mask, which is similar to the speaker separation classical framework. Finally, the mask is multiplied with the encoder output and input into the decoder to obtain the separated sound signal.
The idea of permutation invariant training is to first give the speaker separation model and calculate different losses through different output permutations and then, continue training the model after deducing a correct permutation from the losses. Therefore, a separation model is first initialized randomly to obtain a permutation; then, the model is updated and iterated until convergence. This method still needs to be executed with a known number of talkers, but it can already be used for end-to-end purposes.
Automatic speech recognition (ASR) is used to capture the acoustic signal of speech and determine the spoken words by a pattern-matching method. ASR is efficiently and accurately to convert voice signal content into its corresponding text. Its function is a little different from the above-mentioned speaker separation models since they deal with the speaker’s vocal signal, while the ASR analyzes the vocabulary content of the voice signal. The natural communication between machine and human can be enhanced by ASR.
The acoustic model (AM) is commonly thought of as the modeling of sound generation. In general, the AM is used as the basic unit for automatic speech recognition, converting speech signals into acoustic states. More precisely, it is the probability of converting a certain acoustic symbol. AM mostly uses the HMM [7], which has a good mathematical structure and is the theoretical basis for the formation of a large number of applications. The HMM is the internal state of this Markov model, while the external world only sees the output values at each moment, which are the acoustic symbols (acoustic features) for automatic speech recognition.
The pronunciation model contains all the sets of words that need to be processed and their pronunciations; thus, it is also known as the pronunciation dictionary. Normally, the Viterbi algorithm is an efficient approach to decode the acoustic feature probabilities output by the HMM and outputs them as appropriate word strings with a maximum probability.
The purpose of the language model is to combine the output of the pronunciation model into reasonable sentences according to the logic and probability of language usage. It converts a string of letters into a sentence that people can understand.
Traditional automatic speech recognition, as mentioned above, is trained separately using different training data, tuning parameters, optimization strategies, and loss of functions that are not directly transferable, which can be improved with machine learning.
In recent years, the development of machine learning has boomed significantly, which has led to its application in the field of speech. With research, the disadvantages of traditional automatic speech recognition model training have been improved and the end-to-end model and sequence-to-sequence model were found to be reasonable and effective solutions. For example, in 2012, Google developed the RNN-Architecture [8], which incorporates the memory feature of the recurrent neural network (RNN) into the ASR model. Not only did it eliminate the need to align audio and text, but it also achieved an end-to-end effect, jointly optimizing the parameters of the model and allowing the selection of the best phrase for the language model. In 2017, Google added the well-developed attention memory model to the ASR model and developed the LAS model, which can input a string of data and output the data together. Since its development, it has become popular in the field of speech [9]. The sequence-to-sequence deep-learning models are the leading architecture to generate voice recognition applications. Specifically, multiple layers of self-attention blocks are a powerful tool to encode acoustic features [10,11,12,13]. The encoded features are then decoded into word sequences using recurrent transducers [14] or attention-based models [8] for the ASR applications. This paper selected a novel sandwich-type conformer model to approach the generic function of the ASR system to control the mobile robot.
Simultaneous localization and mapping (SLAM) has been an important research topic in the field of robot navigation in the last decade. It offers map generation and self-localization abilities to greatly improve the working efficiency of robots in an unknown environment and to avoid collision problems.
Nowadays, many types of sensors are proposed to build maps. The most popular laser sensors, such as the two-dimension laser Hector SLAM [15], combine various algorithms to approach the target of SLAM. This paper used a camera-type sensor in the SLAM application called the vision-based simultaneous localization and mapping (vSLAM) [16,17]. The vSLAM has the advantage of a lower hardware cost and bountiful information. The monocular-type camera only catches a two-dimensional picture and cannot realize the distance of an object.
Due to the fixed axle spacing between two camera lenses, a binocular-type camera uses two camera lenses to capture two pictures and fuses them to calculate the depth of the object. However, it is difficult to perfectly position the camera lens and it consumes high computation resources.
The RGBD-type camera, which uses three color channels (RGB) to capture bountiful information, and the infrared structured light or time of fly (TOF) are used to directly obtain the three-dimensional information (including depth). The efficiency of the RGBD-type camera is very similar to the laser radar.
This paper applied the robot operating system (ROS) platform to develop the robot system. ROS features a distributed frame to generate multiple communications in a topic, service, etc. ROS is compatible with many programs and languages, such as C++, Python, and JavaScript, with a topic or service module to communicate and develop applications. Another advantage of ROS is that it offers a higher integration and systematically builds individual module examination and verification.
This paper designed an interactive human–robot system with a voice command to control a mobile robot and enable it to complete service tasks. The deep-learning neural network structure was employed for analysis and to allow the mobile robot equipped with an ASR to make an appropriate decision. The mobile robot monitors the current state of the task to approach the required action. A real-time appearance-based mapping (RTAB-MAP) algorithm [17] with a visual odometer was used to construct the three-dimensional maps and allow the robot to gradually move into the desired position in the service environment.
This paper is organized in five parts, where Section 2 explains the human–voice interface system to show the outstanding nature of the proposed methods, Section 3 presents environment map generation through VSLAM, Section 4 completes voice interactive robot control in a real-robot experiment, and Section 5 presents the conclusions.

2. Human–Voice Interface System

The audio signal is input through the microphone; the input audio, which can be from multiple people or other noise in the background, is then pre-processed at this stage (called noisy audio from hereon). Next, the speaker separation system is used to separate the audio part of a specific speaker from the noisy audio. The separated speaker-specific audio signals are transformed into speaker command strings through a speech recognition system to provide the robot with decision-making ability.

2.1. Speech Pre-Processing

Even if a microphone with a specified pattern is used, the sound signal may still be mixed with background sounds and noises; however, voice recognition only requires segments of human speech. If there is noise or background sound, the accuracy of the recognition will be greatly affected. To remove background sounds, noise, and silent signals for the subsequent ASR, this pre-processing step used voice activity detection (VAD)/speech activity detection (SAD). VAD is mainly used to detect the presence of speech sounds in the current incoming voice signal, and its classical design consists of three steps: first, the sound signal goes through a noise reduction process, such as spectral subtraction [18]; then, it undergoes feature extraction and classifier, which are the focus of VAD.
In this paper, we used the VAD algorithm in WebRTC, which is a multi-feature integrated evaluation that employs the Gaussian mixture model (GMM) [19]. The GMM solves the clustering problem of the samples by estimating the probability density distribution of the sample; thus, it is suitable for VAD:
f ( x ) = 1 σ 2 π e ( x μ ) 2 2 σ 2
f ( x k | Z , r k ) = 1 2 π e ( x k μ z ) 2 2 σ 2
Gaussian distribution is also known as a normal distribution. If the random variable Χ serves a Gaussian distribution with a location denominator μ and a scale denominator σ then its probability density function can be expressed as Equation (1). The Gaussian distribution in WebRTC is expressed as Equation (2), in which x k is the selected feature vector that is the energy feature of the six sub-bands just mentioned; and r k is the combination of the mean μ z and variance σ parameters, which determine the probability of the Gaussian distribution. Z = 1 calculates the probability of speech, Z = 0 calculates the probability of noise, and the final classifier uses the likelihood ratio test. If the likelihood ratio is greater than the threshold value, it is considered as speech, which is then output to the speaker separation system.

2.2. Speaker Separation System

This paper utilized the speaker separation system [20] to achieve speaker-specific voice separation. The speaker separation system architecture shown in Figure 1 is divided into two main parts: (1) a speaker encoder, which is responsible for extracting the voice pattern characteristics of a specific speaker set by the user; and (2) a voice filter model, which is responsible for retaining and outputting the voice of a specific speaker in a noisy or multi-person environment. The two are trained separately; the details are as follows:
(1)
Speaker encoder
The Speaker encoder extracts the embedding that represents the features of a particular speaker’s voice pattern, which is called the d-vector. The speaker encoder architecture is a three-layer long short-term memory (LSTM) with generalized end-to-End loss (GE2E) [21]. The input is the Mel-inverted spectrum (MFC) [22] with reference audio conversion, having one frame every 1.6 s with each frameshift at 50%, which is equivalent to 0.8 s. The output is a 256-dimensional d-vector. The speaker encoder architecture is shown in Figure 2.
(2)
Voice Filter model
This voice filter part is the main architecture of the system containing 8 layers of convolution layer, 1 layer of LSTM, and 2 layers of FC layer. The training and running process is shown in Figure 3. The input is a mix of audio and the d-vector. Before the mix enters the voice filter, it goes through short-time Fourier transform (SFTF), is converted into a spectrum map, and is input to the convolution layer. Meanwhile, the d-vector is used as an input for long-term memory (LSTM). There are two reasons for placing the d-vector in the long- and short-term memory (LSTM), but not in the convolutional layer as input. First, the d-vector already has good robustness to represent the vocal features of a particular speaker and does not need to be processed in the convolutional layer; and second, the convolutional layer assumes that time and frequency are aligned. Therefore, two completely different signals cannot be put together as inputs.
As shown in Table 1, except for the last FC layer (FC) where the activation function is sigmoid, all the other layers are commonly used ReLU. Additionally presented in Table 1 are the parameters of each layer; the loss function uses mean square error (MSE).
Table 1 shows the design of dilated convolutional layers to extract low-level acoustic features more efficiently. The time and frequency of the spectrum map in speech recognition were the same as the height and width of the image in image recognition.
(3)
Evaluation and Training in Voice Filter model
The Libri speech dataset [23] was used in our voice filter model and its training flow is shown in Figure 4. First, before the model started training, all the parameters and training data of the model were set up in advance; the training data include the training set and test set. Training of the model began after setting one of the parameter settings to “step”, which determines the time when the model weights are saved. After the training reached the number of times the step was set, the model weights were saved and immediately substituted into the model for evaluation of the goodness of fit.
In this paper, the signal-to-distortion Ratio (SDR) is used to evaluate the voice filter model. SDR is commonly used to evaluate signal separation systems [24] and is calculated as the sound intensity ratio in decibels (dB), which determines the error between the projection of the resulting audio and the energy of the clean audio. As shown in Figure 5, the more parallel the two sound vectors are (the more similar the sound is), the higher the signal distortion rate is; unlike the signal-to-noise ratio (SNR), which is not affected by the sound size.
As described in Figure 5 and Table 2, the output audio vector X *  is projected to the vertical direction of the target audio vector X ^ to obtain X T , which is parallel to the target audio vector X ^ . X T is determined using Equation (3). Finally, Equation (4) is calculated. If the output audio vector X * is parallel to the target audio vector X ^ , the larger the SDR is, which means that the sounds are more similar; if the output audio vector X * is perpendicular to the target audio vector X ^ , the smaller the SDR is, which means that the sounds are different.
X T = X * · X ^ | | X ^ | | 2 X ^
S D R = 10 log 10 | | X T | | 2 | | X E | | 2
In this paper, we used the SDR for model evaluation and employed the test set for SDR calculation. When the SDR after model evaluation did not continue to rise after several sets of steps, the model was stopped; conversely, when the SDR continued to rise, we retrained the model with that weight as the initial weight for model learning.

2.3. Automatic Speech Recognition (ASR)

In this paper, ASR used the conformer [11] model, which utilized the transformer that is specialized in capturing a large range of feature interaction information, and the CNN specialized in extracting local subtle features; both have contributed to the field of speech and other areas of machine learning in recent years. The conformer model combines the advantages of both for speech recognition.

2.3.1. Conformer Model Architecture

The conformer model is a modification of the encoder in the transformer framework, as it incorporates the conformer blocks of the CNN framework. The conformer architecture is shown in Figure 6. The left side is the classic transformer model architecture and the right side (red box) is the internal architecture after the modification. At the start, the spectrum of the audio signal output from the speaker separation system is inputted into the encoder; then, after spectrogram augmentation, it enters linearly in the convolution subsampling, dropout, and the main conformer blocks. Conformer blocks are executed N times according to the input size, just like the encoder in the transformer architecture.
The internal structure of the conformer blocks is shown in Figure 7. It consists of three modules: feed-forward module, multi-head self-attention module, and convolution module, each of which uses a residual connection. The structure of the conformer blocks is like a sandwich, where the feed-forward module is located between the multi-head self-attention module and the convolution module, connecting them with each other; however, each feed-forward module only contributes half of its weight. Finally, it enters the post-layer norm for data processing, that is, the post-norm residual unit is utilized. Figure 8 [25] presents the Macaron-net [26] structure, which is formed through Equations (5)–(9). Conformer blocks are represented by x_i for the ith input, y_i for the output, FFN denotes the feed-forward module, MHSA denotes the multi-head self-attention module, and the convolution module is symbolized by Convol.
x i ˜ = x i + 1 2 F F N ( x i )
x i = x i ˜ + M H S A ( x i ˜ )
x i = x i + C o n v ( x i )
x i ˜ = x i + 1 2 F F N ( x i )
y i = l a y e r n o r m ( x i ˜ )
Multi-head self-attention is the key structure in the transformer, and is a variation of self-attention. It is similar to an RNN, but it can input and output a whole set of data at the same time. Figure 9 shows the structure of the multi-head self-attention module, which uses the relative positional embedding and incorporates the pre-norm residual units with dropout (see Figure 10). Consequently, the convolution and feed-forward modules are also designed using the same structure. The transformer is unable to learn information on the sequence of the positions, so the developers of the transformer proposed to use position embedding for the multi-head self-attention to gain position information. Relative position coding provides better generalization over different input lengths and better robustness to variable speech lengths.

2.3.2. Convolution Module Design

The convolution module starts from the rating mechanism [27], which consists of a pointwise convolution layer and a gated linear unit (GLU), followed by a 1D depth-wise convolution layer. The batch-norm is arranged after the convolution layer to facilitate model training, and the pre-norm residual unit is used as in the multi-head self-attention module. The convolutional module architecture is shown in Figure 11.
As mentioned, the used depth-wise separable convolution [28] is a new computational structure for CNN developed by Google. It aims to reduce the computational effort of CNN by splitting the original convolutional computation into two parts: pointwise convolutions and depth-wise convolutions, each of which is performed without affecting the output structure to reduce the computation.

2.3.3. Feed-Forward Module

As proposed in reference [29], the feed-forward module of the transformer architecture appears after the multi-head self-attention layer. Although it is based on the feed-forward module, it has a nonlinear activation function called Swish activation [30] between the two linear transitions. A dropout is also added to normalize the network. The module follows the feed-forward module in the transformer architecture and uses the pre-norm residual unit as shown in Figure 12.
The conformer training data set for this paper is the entire training set of LibriSpeech, which is approximately 1000 h long. The audio is extracted one frame every 25 ms, with each frameshift at 10 ms. The 80-D audio spectrum is obtained by 80-channel filter banks and the speech signal is enhanced by using Spec Augment [31,32].
In this paper, the conformer used character error Rate (CER) and word error rate (WER) [33] for testing. We focused mainly on the WER, in which the lower the WER, the better the training of the model. The Equation for the calculation of the WER is shown in Equation (10):
W E R = ( S + D + I ) N × 100 % = S + D + I S + D + C × 100 %
where S denotes the number of substituted words, D is number of deleted words, I is the number of inserted words, N is the total words in original text, and C is the correct number of words.

2.3.4. Model Testing Result

In this section, the training results of the two models are presented: (1) the SDR with speaker separation model and (2) the WER with speech recognition mode.
(1)
Speaker Separation Model Testing
There are different training sets for evaluating the performance of the speaker separation model. Table 3 shows the results of the different training sets (i.e., train-clean-100, train-clean-360, train-clean-100, and 8 train-clean-360).
In the first experiment, the performance results of the train-clean-100 dataset with 251 speakers were not good with an SDR of only 3.2 dB. Further, the actual output audio was much different from the clean audio. Moreover, the output volume was about 80% less than the input volume and the specific speaker audio was not separated very successfully, and is similar to turning down the volume of the mixed audio. The training effect of the train-clean-360 with 921 speakers was slightly improved; the volume was still cut, but the magnitude was reduced and the separation effect was slightly present. As can be seen in Figure 13, the SDR was raised to 5.4 dB, and the volume of speaker A was greater than the volume of speaker B. Finally, the two training sets were summed up and the number of speakers reached about 1200. The training was conducted twice. The first time we set the batch size to 8, the SDR stopped at about 7.5 dB at the frequency step of 300 k; so, we lowered the batch size to 4, and the SDR continued to set at 300 k and approached a better SDR of 8.6 dB at 530 k. As shown in Figure 14, the separated clean audio was precisely matched to the actual output. The proposed effectiveness of the separated model was to reduce the influence of the outside environment. This experiment also illustrated that the more speakers are included in the training set, the better the performance of the separated trained model. This characteristic of the speaker separation model also supports the next ASR model to obtain the adaptive reality to fit the right voice command in a real environment.
(2)
Evaluations for Automatic Speech Recognition Model
The model discussed earlier was added to the ASR system. The training data were divided into several different sets to evaluate the goodness of WER. All the training cycles were set as epoch = 100 and batch size = 4. The audio was added to the noise to obtain the adaption of the speaker separation. The test set with a different amount of training dataset was mainly applied to evaluate the ASR model (see Table 4). When the total training cycle for train-100 audio data was finished, it obtained a not-so-good WER of 14%. In another experiment, the total training cycle was gradually added to the other training data set (i.e., train-360 and train-500) to approach the best WER, which was from 14% to 5.3%. The smallest unit in the English language is a single word, so most of the Speech Recognition model uses a WER index to detect the performance of the trained result. In this paper’s real speech recognition experiment, a total of 129 different English voice patterns were continuously collected from different individuals and the pre-trained speech recognition model built in this study was applied to finally approach the correct rate of 89.3% in the real environment.
Zhang et al. used the semi-supervised learning method with the large-scale YouTube-based data to pre-train the conformer model based on the original LibriSpeech database and the upstream/downstream to self-train the model size, so it will match the multiple public datasets. Experiential results reached the best WER of 7.7% as it approached multiple tasks by voice command [34]. Several more studies have registered and trained the deep neural network HMM to advance ASR in a real environment. A pretraining ASR system has been built with a novel learning stratagem and with a structure of the artificial neural network (ANN) model based on a large, but limited playback loudspeaker sub databases. It was aimed to outperform the famous and appropriated Google API, IBM API, and Bing API in terms of the WER [35] after completing an environment-based training cycle with the specific questionnaires. In another voice command with robot application study [36], researchers provided a cloud-based NLP platform and ROS-based mobile robot system for participants interacting with the robot. The lowest WER it obtained was 5.8% and the highest was 83.3% with the correct rate of voice entity detections in an office, without an ambient noise environment. The WER increased to 21.9% and the correct rate of entity detections was decreased to 61.7% with respect to the effect of noise. In the research paper [37], receiving and recognizing voice commands were presented using a LabVIEW machine learning model for a person to communicate with the robot. Its practical tests, voice command reached an average precision of 86% for “go forward and back”, and 78% correctness for “left and right” to control the robot’s actions. Its total average precision reached 82%, while people used voice to control the mobile robot forward and backward. A multi-stream HMM decision fusion [38] was also used to improve the voice recognition rate of the acoustic model to 88.5% in the close-talking type of ASR experiments. The other identification of voice correctness was approximately 72.4% when the throat microphone was used to test the same voice command.
Table 5 compares the evaluations of WER and speech recognition accuracy rates in many different methods. This data analysis illustrates that the proposed deep-learning type voice machine achieved the best results with the lowest WER of speech recognition testing and the highest accuracy in control command for the mobile robot. A generalized speech training database can result in more complete word comprehension, but requires more intensive training methods and more training time. Single-voice training is easier because it is based on a specific set of required commands. This research paper suggests that the performance of the general-purpose speech training database is the best result compared to other research papers.

3. Environment Map Generation through VSLAM

A real-time appearance-based mapping (RTAB-Map) algorithm based on the concept of SLAM to generate the map of the environment is proposed in this study. The RGB-D cameras concurrently obtain the image and distance information. The RGB-D maps do not consume much computational time to calculate the distance, and approach better accuracy of position than the binocular or monocular camera.
The proposed RTAB-Map SLAM algorithm is divided into three parts: (1) front terminal part, (2) rear terminal part, and (3) loop closure detection. The objective of visual odometry extracts the feature and information from the odometer’s captured image, and plots the local map. The rear part is used to optimize the original maps and robot posture estimation and other processes. Figure 15 shows the Flow Chart of the RTAB-based SLAM.
In the rear part, the speeded-up robust features (SURF) is used to extract and match the feature point purpose [33]. A Hessian matrix is applied to detect feature points; image integration is also proposed to speed up the calculation. The URF structure is a multiple-layer filter containing different filters in the same layer. Further, same-sized filters are contained in the different layers of URF. Therefore, the Vague coefficient of filter is gradually increased to reduce the captured process and improve the handing speed. SURF confirms the matching degree based on the Euclidean distance between two feature points, where the shorter distance presents the better matching degree.
To calculate the visual odometry, the reference frame is obtained from the previous time and the current frame is defined in the now time. The reference frame is selected as the benchmark of the coordinate system, where it is matched with the current frame. The random sample consensus-perspective-n-points (RANSAC-PnP) [39] is taken to calculate the coordinate position of the current RGB-D sensor. The previous cycle is then repeatedly performed while the whole movement track of the RGB-D sensor is achieved. This path is considered the track of the mobile robot.
The local map is selected as the matched feature point of the object of interest in the environment, and the real position and posture of the object are calculated through the matching procedure between the current frame and the feature point. The advantage of the local map method is that it can be modified based on the constructed information of each frame even when some of the resulting frames are wrong.
In the rear terminal part, the SLAM procedure meets the error noise problem, which is gradually accumulated when operating in the front part. The error is a big amount because it is accumulated over a long period. Therefore, it appears that there is a huge gap between the constructed and the real maps; thus, collecting all the possible map information is necessary to complete the whole optimization. The optimization is added to the load tags from the previous motion trajectory, which is obtained in the front part; the prediction of the motion state is then approached to the desired target. After that, the second rank in optimization is achieved to obtain the better pose graph, which is needed to optimize the global map after updating the motion state. Because the object is continuously detected by the RGB-D sensor, the same object is always constructed, which causes some errors in the map. This research extracted the primary feature point to match the object and then, the position value was attained through the motion relationship. Through this, we were able to efficiently construct the dynamic motion of the mobile robot while generating and keeping the trajectory of the primary feature point.
For the loop closure detection function, this study proposed the RTAB-map algorithm [15] to finish the real-time localization function of the robot. Figure 16 shows the flowchart of the RTAB-map algorithm.
In creating the robot’s localization point, the bag-of-words without the process of training the specific environment is used to obtain favorable localization data. The new localization weight is updated in the next stage. The current localization point makes a similarity comparison with the last localization point of the short-term memory (STM). The current localization is replaced by the last localization point of STM if the similarity is over the defined threshold.
The Bayesian filter is used to refresh the working memory (WM). It evaluates the probability of loop closure between the current localization point and the localization point of WM. The closure loop with the highest p is established, with the probability p less than the defined threshold. The connection between the new and old localization points is ensured, and the new weight is refreshed by adding the original weight to the old one.
The localization points with the lowest weight and the longest holding time are transferred into long-term memory (LTM) [40] with the graph processing time set at an interval time T. Thus, the inverse distance weighting method is used to reorganize the weight to greatly improve the efficiency of the closure loop detection. Figure 17 illustrates the simulation of the feature-matching result.

4. Voice Interactive Robot Control in Real-Robot Experiment

In the last real-world experiment, the speaker separation and ASR are combined to accomplish the voice interactive-based mobile robot applications. The speaker separation can percolate the voice into a specific signal and the ASR analyses the signal to transfer it as a control command. A detailed mobile robot with dual-arm architecture is built in our laboratory [41,42], which completes the object recognition and placement at the set position for different specific service tasks in a real environment. Detailed demonstration of this paper in a real environment is illustrated in the next paragraph.
Figure 18 shows the service tasks through a real robot that can correctly pick-up and deliver three types of objects based on the human voice control command. A mobile robot stands in front of the service desk, and is commanded by a human voice to pick up and deliver three types of objects: an alcohol bottle, a box, and a toilet paper. Figure 18a shows the mobile robot receives the control command “give me the alcohol”, and Figure 18b illustrates the mobile robot makes the next action decision by recognizing the control command, which is to pick up the alcohol bottle in Figure 18c. Figure 18d shows the robot delivers the bottle to the service request in Figure 18e. The sequential actions presented in Figure 18f–j exhibit that the service robot successfully picks up and delivers the box to the correct position after recognizing the voice command “give me the box”. Finally, Figure 18k–o show the mobile robot completing the service task as a response to the voice command “give me the toilet paper”. After giving the user toilet paper, Figure 18p shows that the mobile robot moves back to the original location and stands by. The demonstration for the service mobile robot systems in a real environment are fully illustrated in detail in the video presentation, indicating a successful human-interactive robot application in an external noise environment [43]. In addition to the service actions shown above, the dual-arm mobile robot can perform the other different behavioral actions, such as cup recognition and gripping, according to the specific actions that need to be controlled with human voice after completing the manipulator setting and training for these tasks.

5. Conclusions

This paper combined the speaker separation system and ASR to analyze the practical semantic word of human speech. The speaker separation system included a fast speaker encoder, which requires a lesser time to convert the speech spectrum and remove the ambient noise using a trained voice filter. Thus, a clear and correct control instruction can run through a noisy environment to automatically finish speech recognition. The conformer module used a block CNN-type encoder to form a transformer framework, which takes the advantage of the multi-head self-attention machine to note the affirmatory word, efficiently separate the concise word, and clearly acquire the correct command to handle the required actions for the mobile robot.
A real-time vision-based appearance-base mapping (RTAB-map) algorithm, which utilized the RGB-D device, was employed in this study to concurrently generate the environmental map and fit the matching boundary of the object in the working space. The RTAB-map with a visual odometer was employed to predict the motion of the mobile robot to appropriately guide it into the unloading zone and service area to complete the navigation process.
Finally, a real mobile robot with dual arm was used to test the study’s model. It was placed in front of a table and was asked to complete a series of human voice-activated commands. The results showed that the robot was able to correctly recognize the voice command and efficiently perform the service tasks.

Author Contributions

Conceptualization, S.-A.L. and H.-M.F.; methodology, S.-A.L., Y.-Y.L. and H.-M.F.; software, Y.-Y.L. and Y.-C.C.; validation, S.-A.L., Y.-Y.L. and Y.-C.C.; formal analysis, S.-A.L. and Y.-Y.L.; investigation, Y.-C.W. and P.-K.S.; resources, Y.-C.W. and P.-K.S.; data curation, S.-A.L. and H.-M.F.; writing—original draft preparation, Y.-Y.L. and Y.-C.C.; writing—review and editing, S.-A.L. and H.-M.F.; funding acquisition, S.-A.L. and H.-M.F. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was partly supported by the Ministry of Science and Technology of the Republic of China with contract number: MOST-110-2221-E-032-039 and MOST-110-2221-E-507-009.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lin, P.H.; Lin, C.Y.; Hung, C.T.; Chen, J.J.; Liang, J.M. The Autonomous Shopping-Guide Robot in Cashier-Less Convenience Stores. Proc. Eng. Technol. Innov. 2020, 14, 9–15. [Google Scholar] [CrossRef]
  2. Wuth, J.; Correa, P.; Núñez, T.; Saavedra, M.; Yoma, N.B. The Role of Speech Technology in User Perception and Context Acquisition in HRI. Int. J. Soc. Robot. 2021, 13, 949–968. [Google Scholar] [CrossRef]
  3. Hershey, J.R.; Chen, Z.; Roux, J.L.; Watanabe, S. Deep Clustering: Discriminative Embeddings for Segmentation and Separation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 31–35. [Google Scholar] [CrossRef] [Green Version]
  4. Luo, Y.; Mesgarani, N. TasNet: Time-Domain Audio Separation Network for Real-Time, Single-Channel Speech Separation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 696–700. [Google Scholar] [CrossRef] [Green Version]
  5. Luo, Y.; Mesgarani, N. Conv-TasNet: Surpassing Ideal Time–Frequency Magnitude Masking for Speech Separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 1256–1266. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Yu, D.; Kolbæk, M.; Tan, Z.H.; Jensen, J. Permutation Invariant Training of Deep Models for Speaker-Independent Multi-Talker Speech Separation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 241–245. [Google Scholar] [CrossRef] [Green Version]
  7. Rabiner, L.R. A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef] [Green Version]
  8. Graves, A. Sequence transduction with recurrent neural networks. arXiv 2012, arXiv:1211.3711. [Google Scholar] [CrossRef]
  9. Chan, W.; Jaitly, N.; Le, Q.; Vinyals, O. Listen, Attend and Spell: A Neural Network for Large Vocabulary Conversational Speech Recognition. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 4960–4964. [Google Scholar] [CrossRef]
  10. Narayanan, A.; Chiu, C.C.; O’Malley, T.; Wang, Q.; He, Y. Cross-Attention Conformer for Context Modeling in Speech Enhancement for ASR. In Proceedings of the 2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), Cartagena, Colombia, 13–17 December 2021; pp. 312–319. [Google Scholar] [CrossRef]
  11. Gulati, A.; Qin, J.; Chiu, C.C.; Parmar, N.; Zhang, Y.; Yu, J.; Han, W.; Wang, S.; Zhang, Z.; Wu, Y.; et al. Conformer: Convolution-Augmented Transformer for Speech Recognition. In Proceedings of the Interspeech, Shanghai, China, 25–29 October 2020; pp. 5036–5040. [Google Scholar] [CrossRef]
  12. Yeh, C.F.; Mahadeokar, J.; Kalgaonkar, K.; Wang, Y.; Le, D.; Jain, M.; Schubert, K.; Fuegen, C.; Seltzer, M.L. Transformer-Transducer: End-to-end speech Recognition with Self-attention. arXiv 2019, arXiv:1910.12977. [Google Scholar]
  13. Karita, S.; Chen, N.; Hayashi, T.; Hori, T.; Inaguma, H.; Jiang, Z.; Someki, M.; Soplin, N.E.Y.; Karita, S.; Chen, N.; et al. A Comparative Study on Transformer vs RNN in Speech Applications. In Proceedings of the 2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), Singapore, 14–18 December 2019; pp. 449–456. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, S.; Wu, Y.; Chen, Z.; Wu, J.; Li, J.; Yoshioka, T.; Wang, C.; Liu, S.; Zhou, M. Continuous Speech Separation with Conformer. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 5749–5753. [Google Scholar] [CrossRef]
  15. Kohlbrecher, S.; Stryk, O.V.; Meyer, J.; Klingauf, U. A Flexible and Scalable SLAM System with Full 3D Motion Estimation. In Proceedings of the IEEE International Symposium on Safety, Security, and Rescue Robotics, Kyoto, Japan, 1–5 November 2011; pp. 155–160. [Google Scholar] [CrossRef] [Green Version]
  16. Mur-Artal, R.; Tardós, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot 2017, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  17. Labbé, M.; Michaud, F. RTAB-Map as an Open-Source Lidar and Visual Simultaneous Localization and Mapping Library for Large-Scale and Long-Term Online Operation. J. Field Robot. 2019, 36, 416–444. [Google Scholar] [CrossRef]
  18. Boll, S.F. Suppression of Acoustic Noise in Speech Using Spectral Subtraction. IEEE/ACM Trans. Audio Speech Lang. Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  19. Reynolds, D.A. Gaussian Mixture Models. Encycl. Biom. 2009, 741, 659–663. [Google Scholar] [CrossRef]
  20. Wang, Q.; Muckenhirn, H.; Wilson, K.; Sridhar, P.; Wu, Z.; Hershey, J.R.; Saurous, R.A.; Weiss, R.J.; Jia, Y.; Moreno, I.L. VoiceFilter: Targeted Voice Separation by Speaker-Conditioned Spectrogram Masking. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019; pp. 2728–2732. [Google Scholar]
  21. Wan, L.; Wang, Q.; Papir, A.; Moreno, I.L. Generalized End-to-End Loss for Speaker Verification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 4879–4883. [Google Scholar] [CrossRef] [Green Version]
  22. Sahidullah, M.; Saha, G. Design, Analysis and Experimental Evaluation of Block Based Transformation in MFCC Computation for Speaker Recognition. Speech Commun. 2012, 54, 543–565. [Google Scholar] [CrossRef]
  23. Chung, J.S.; Nagrani, A.; Zisserman, A. VoxCeleb2: Deep Speaker Recognition. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018; pp. 1086–1090. [Google Scholar] [CrossRef] [Green Version]
  24. Vincent, E.; Gribonval, R.; Fevotte, C. Performance Measurement in Blind Audio Source Separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2006, 14, 1462–1469. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, Q.; Li, B.; Xiao, T.; Zhu, J.; Li, C.; Wong, D.F.; Chao, L.S. Learning Deep Transformer Models for Machine Translation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 1810–1822. [Google Scholar]
  26. Lu, Y.; Li, Z.; He, D.; Sun, Z.; Dong, B.; Qin, T.; Wang, L.; Liu, T.Y. Understanding and Improving Transformer from a Multi-Particle Dynamic System Point of View. arXiv 2019, arXiv:1906.02762. [Google Scholar]
  27. Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language Modeling with Gated Convolutional Networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; Volume 70, pp. 933–941. [Google Scholar]
  28. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar] [CrossRef] [Green Version]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4 December 2017; pp. 6000–6010. [Google Scholar]
  30. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  31. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition. In Proceedings of the Interspeech, Graz, Austria, 15–19 September 2019; pp. 2613–2617. [Google Scholar] [CrossRef]
  32. Park, D.S.; Zhang, Y.; Chiu, C.C.; Chen, Y.; Li, B.; Chan, W.; Le, Q.V.; Wu, Y. SpecAugment on Large Scale Datasets. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 6879–6883. [Google Scholar]
  33. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded Up Robust Feature. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 404–417. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Park, D.S.; Han, W.; Qin, J.; Gulati, A.; Shor, J.; Jansen, A.; Xu, Y.; Huang, Y.; Wang, S.; et al. BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition. IEEE J. Sel. Top. Signal Process. 2022, 16, 1519–1532. [Google Scholar] [CrossRef]
  35. Novoa, J.; Wuth, J.; Escudero, J.P.; Fredes, J.; Mahu, R.; Yoma, N.B. DNN-HMM based Automatic Speech Recognition for HRI Scenarios. In Proceedings of the 2018 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Chicago, IL, USA, 5–8 March 2018; pp. 150–159. [Google Scholar]
  36. Stuede, M.; Wilkening, J.; Tappe, S.; Ortmaier, T. Voice Recognition and Processing Interface for an Interactive Guide Robot in a University Scenario. In Proceedings of the 2019 19th International Conference on Control, Automation and Systems (ICCAS), Jeju, Republic of Korea, 15–18 October 2019; pp. 1238–1242. [Google Scholar] [CrossRef]
  37. Pleshkova, S.G.; Bekyarski, A.B.; Zahariev, Z.T. LabVIEW Model of Voice Commands for Mobile Robot Motion Control Using Internet of Thinks Module. In Proceedings of the 2019 X National Conference with International Participation (ELECTRONICA), Sofia, Bulgaria, 16–17 May 2019; pp. 1–4. [Google Scholar] [CrossRef]
  38. Heracleous, P.; Even, J.; Sugaya, F.; Hashimoto, M.; Yoneyama, A. Exploiting Alternative Acoustic Sensors for Improved Noise Robustness in Speech Communication. Pattern Recognit. Lett. 2018, 112, 191–197. [Google Scholar] [CrossRef]
  39. Lin, H.I.; Nanda, S. 6 DOF Pose Estimation for Efficient Robot Manipulation. In Proceedings of the IEEE Conference on Industrial Cyberphysical Systems (ICPS), Tampere, Finland, 10–12 June 2020; pp. 279–284. [Google Scholar] [CrossRef]
  40. Labbé, M.; Michaud, F. Long-Term Online Multi-Session Graph-Based SPLAM with Memory Management. Auton. Robot. 2018, 42, 1133–1150. [Google Scholar] [CrossRef]
  41. Wong, C.C.; Chien, S.Y.; Feng, H.M.; Aoyama, H. Motion Planning for Dual-Arm Robot Based on Soft Actor-Critic. IEEE Access 2021, 9, 26871–26885. [Google Scholar] [CrossRef]
  42. Li, S.A.; Chou, L.H.; Chang, T.H.; Wong, C.C.; Feng, H.M. Design and Implementation of an Autonomous Service Robot Based on Cyber Physical Modeling Systems. Proc. Inst. Mech. Eng. B J. Eng. Manuf. 2022, 1–15, Advance online publication. [Google Scholar] [CrossRef]
  43. Available online: https://youtu.be/6-bWdy5DG8A (accessed on 16 January 2022).
Figure 1. Speaker separation system architecture diagram.
Figure 1. Speaker separation system architecture diagram.
Applsci 13 03359 g001
Figure 2. Speaker-Encoder Architecture Diagram.
Figure 2. Speaker-Encoder Architecture Diagram.
Applsci 13 03359 g002
Figure 3. Voice filter architecture diagram.
Figure 3. Voice filter architecture diagram.
Applsci 13 03359 g003
Figure 4. Voice filter model train flow chart.
Figure 4. Voice filter model train flow chart.
Applsci 13 03359 g004
Figure 5. SDR vector diagram.
Figure 5. SDR vector diagram.
Applsci 13 03359 g005
Figure 6. Conformer model structure.
Figure 6. Conformer model structure.
Applsci 13 03359 g006
Figure 7. Conformer blocks.
Figure 7. Conformer blocks.
Applsci 13 03359 g007
Figure 8. Post-norm residual unit.
Figure 8. Post-norm residual unit.
Applsci 13 03359 g008
Figure 9. Multi-head self-attention module.
Figure 9. Multi-head self-attention module.
Applsci 13 03359 g009
Figure 10. Pre-norm residual unit.
Figure 10. Pre-norm residual unit.
Applsci 13 03359 g010
Figure 11. Convolution module.
Figure 11. Convolution module.
Applsci 13 03359 g011
Figure 12. Feed-forward module.
Figure 12. Feed-forward module.
Applsci 13 03359 g012
Figure 13. The related SDR value with the amount dataset of 921 people.
Figure 13. The related SDR value with the amount dataset of 921 people.
Applsci 13 03359 g013
Figure 14. The related SDR value with the amount dataset of 1200 people.
Figure 14. The related SDR value with the amount dataset of 1200 people.
Applsci 13 03359 g014
Figure 15. Flow chart of the RTAB-based SLAM.
Figure 15. Flow chart of the RTAB-based SLAM.
Applsci 13 03359 g015
Figure 16. Flow chart of the loop closure detection by RTAB-map.
Figure 16. Flow chart of the loop closure detection by RTAB-map.
Applsci 13 03359 g016
Figure 17. Simulation for feature-matching result.
Figure 17. Simulation for feature-matching result.
Applsci 13 03359 g017
Figure 18. Sequential actions of a real mobile robot for different service tasks.
Figure 18. Sequential actions of a real mobile robot for different service tasks.
Applsci 13 03359 g018aApplsci 13 03359 g018bApplsci 13 03359 g018c
Table 1. Parameters of voice filter network.
Table 1. Parameters of voice filter network.
LayerKernel SizeDilationFilters/Nodes
TimeFreqTimeFreq
Convolution Layer 1171164
Convolution Layer 2711164
Convolution Layer 3551164
Convolution Layer 4552164
Convolution Layer 5554164
Convolution Layer 6558164
Convolution Layer 75516164
Convolution Layer 811118
LSTM 400
FC 1 600
FC 2 600
Table 2. Meaning of SDR symbol.
Table 2. Meaning of SDR symbol.
SymbolMeaning
X * Vector of output audio
X ^ Vector of target audio
X T X *   project   X ^
X E X * X ^
Table 3. Evaluation of speaker separation model.
Table 3. Evaluation of speaker separation model.
Training SetSDRFrequency (Step)Batch Size
Train-1003.2 dB65 k8
Train-3605.4 dB65 k8
Train-100 and train-3607.5 dB30 k8
Train-100 and train-3608.6 dB530 k4
Table 4. Evaluations of automatic speech recognition model.
Table 4. Evaluations of automatic speech recognition model.
Training SetWER Test Result
Train-10014.1%
Train-36011.5%
Train-100 and train-36010.3%
Train-100 and train-360, and train-5005.3%
Table 5. Comparisons for WER and accuracy of speech recognition in different methods.
Table 5. Comparisons for WER and accuracy of speech recognition in different methods.
MethodsWER%Accuracy of Speech RecognitionFitting Type
Zhang et al. [34]7.7%XUniversal word type
Novoa et al. [35]11.62%XUniversal word type
Google API [35]15.79%XUniversal word type
IBM API [35]40.74%XUniversal word type
Stuede et al. [36]5.8%83.3%Single word type
Pleshkova et al. [37]X82%Single word type
Heracleous et al. [38]X88.5%Single word type
The proposed methods5.3%89.3%Universal word type
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.-A.; Liu, Y.-Y.; Chen, Y.-C.; Feng, H.-M.; Shen, P.-K.; Wu, Y.-C. Voice Interaction Recognition Design in Real-Life Scenario Mobile Robot Applications. Appl. Sci. 2023, 13, 3359. https://doi.org/10.3390/app13053359

AMA Style

Li S-A, Liu Y-Y, Chen Y-C, Feng H-M, Shen P-K, Wu Y-C. Voice Interaction Recognition Design in Real-Life Scenario Mobile Robot Applications. Applied Sciences. 2023; 13(5):3359. https://doi.org/10.3390/app13053359

Chicago/Turabian Style

Li, Shih-An, Yu-Ying Liu, Yun-Chien Chen, Hsuan-Ming Feng, Pi-Kang Shen, and Yu-Che Wu. 2023. "Voice Interaction Recognition Design in Real-Life Scenario Mobile Robot Applications" Applied Sciences 13, no. 5: 3359. https://doi.org/10.3390/app13053359

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop