Dysphonia detection using a fully convolutional neural network adapted to dynamic speech lengths
Date
Authors
Type
Language
Reading access rights:
Rights Holder
Conference Date
Conference Place
Conference Title
ISBN, e-ISBN
Container Title
Version
Faculty
First Page
Subject (OSZKAR)
Voice pathology
Deep Learning
Convolutional Neural Network
Variable-length Speech
Gender
University
- Cite this item
- https://doi.org/10.3311/WINS2024-003
OOC works
Abstract
Various conditions affect human speech, leading individuals to produce speech that differs in terms of pitch, quality, and clarity. Numerous deep learning-based methods have been proposed to detect speech disorders. Deep learning (DL) detection methods based on speech require fixed-length dimensional input, which can sometimes be challenging to achieve, even for normal speakers, particularly when dealing with continuous speech rather than sustained vowels or one-word utterances. In this paper, we propose a fully convolutional approach for dysphonia detection. Our proposed method can accommodate any speech duration, unlike previous work that relies on fixed-length samples. The model incorporates exclusively convolutional layers without fully connected layers, enabling it to handle varying speech lengths. Our results demonstrate the superior performance of our proposed model in comparison to other DL approaches that used the same dataset for dysphonia detection. Specifically, our model showcased an accuracy of 91.69%. This represents a notable improvement of over 6% compared to the performance achieved by previous methodologies.