Redirigiendo al acceso original de articulo en 20 segundos...
Inicio  /  Acoustics  /  Vol: 5 Par: 1 (2023)  /  Artículo
ARTÍCULO
TITULO

Visual Speech Recognition for Kannada Language Using VGG16 Convolutional Neural Network

Shashidhar Rudregowda    
Sudarshan Patil Kulkarni    
Gururaj H L    
Vinayakumar Ravi and Moez Krichen    

Resumen

Visual speech recognition (VSR) is a method of reading speech by noticing the lip actions of the narrators. Visual speech significantly depends on the visual features derived from the image sequences. Visual speech recognition is a stimulating process that poses various challenging tasks to human machine-based procedures. VSR methods clarify the tasks by using machine learning. Visual speech helps people who are hearing impaired, laryngeal patients, and are in a noisy environment. In this research, authors developed our dataset for the Kannada Language. The dataset contained five words, which are Avanu, Bagge, Bari, Guruthu, Helida, and these words are randomly chosen. The average duration of each video is 1 s to 1.2 s. The machine learning method is used for feature extraction and classification. Here, authors applied VGG16 Convolution Neural Network for our custom dataset, and relu activation function is used to get an accuracy of 91.90% and the recommended system confirms the effectiveness of the system. The proposed output is compared with HCNN, ResNet-LSTM, Bi-LSTM, and GLCM-ANN, and evidenced the effectiveness of the recommended system.

Palabras claves

 Artículos similares

       
 
Oliver Ohneiser, Jyothsna Adamala and Ioan-Teodor Salomea    
Assistant based speech recognition (ABSR) prototypes for air traffic controllers have demonstrated to reduce controller workload and aircraft flight times as a result. However, two aspects of ABSR could enhance benefits, i.e., (1) the predicted controlle... ver más
Revista: Aerospace

 
Jesús B. Alonso-Hernández, María Luisa Barragán-Pulido, José Manuel Gil-Bordón, Miguel Ángel Ferrer-Ballester and Carlos M. Travieso-González    
Currently, there are more and more frequent studies focused on the evaluation of Alzheimer?s disease (AD) from the automatic analysis of the speech of patients, in order to detect the presence of the disease in an individual or for the evolutionary contr... ver más
Revista: Applied Sciences

 
Pei Liu, Dezhong Peng and Ming Zhang    
In this work, we propose a novel priors-based attention neural network (PANN) for image captioning, which aims at incorporating two kinds of priors, i.e., the probabilities being mentioned for local region proposals (PBM priors) and part-of-speech clues ... ver más
Revista: Applied Sciences

 
Jing Qi, Xilun Ding, Weiwei Li, Zhonghua Han and Kun Xu    
Hand postures and speech are convenient means of communication for humans and can be used in human?robot interaction. Based on structural and functional characteristics of our integrated leg-arm hexapod robot, to perform reconnaissance and rescue tasks i... ver más
Revista: Applied Sciences

 
Sriraksha Nayak,Chandrakala C B     Pág. pp. 52 - 69
According to the World Health Organization estimation, globally the number of people with some visual impairment is estimated to be 285 million, of whom 39 million are blind.  The inability to use features such as sending and reading of email, sched... ver más