DSpace
 

iRepository at Perpustakaan UniMAP >
Theses & Dissertations >
School of Mechatronic Engineering (Theses) >

Please use this identifier to cite or link to this item: http://dspace.unimap.edu.my:80/dspace/handle/123456789/21614

Title: Implementation of feature extraction and classification for speech dysfluencies
Authors: Lim, Sin Chee
Keywords: Speech dysfluencies;Speech;Communication
Issue Date: 2011
Publisher: Universiti Malaysia Perlis (UniMAP)
???metadata.dc.publisher.department???: School of Mechatronic Engineering
Abstract: Speech is prone to disruption of involuntary dysfluent events especially repetitions and prolongations of sounds, syllables and words which lead to dysfluency in communication. Traditionally, speech language pathologists count and classify occurrence of dysfluencies in flow of speech manually. However, these types of assessment are subjective, inconsistent, time-consuming and prone to error. In the last three decades, many research works have been developed to automate the conventional assessments with various approaches such as speech signal analysis, personal variables, acoustic analysis of speech signal and artificial intelligence techniques. From the previous works, it can be concluded that feature extraction methods and classification techniques play important roles in this research field. Therefore, in this work, there are few feature extraction methods, namely, Short Time Fourier Transform (STFT), Mel-frequency Cepstral Coefficient (MFCC) and Linear Predictive Coding (LPC) based parameterization were proposed to extract the salient feature of the two types of dysfluencies. By applying the feature extraction methods on each signal, there are total of seven acoustical features extracted namely STFT, MFCC and five acoustical features from Linear Predictive Coding based parameterization, that is, Linear Predictive Coefficient (LPC), Linear Predictive Cepstral Coefficient (LPCC), Weighted Linear Predictive Cepstral Coefficient(WLPCC), First Order Temporal Derivatives (FOTD) and Second Order Temporal Derivatives (SOTD). Acoustical features are extracted from the signal are use as input parameters for classifiers. Both linear and nonlinear classifiers namely Linear Discriminant Analysis (LDA), k-Nearest Neighbor (kNN) and Least-Squares Support Vector Machines (LSSVM) with linear kernel (SLIN) and Radial Basis Function kernel (SRBF) were suggested to classify the two types of dysfluencies. In order to evaluate the effectiveness of the different feature extraction methods and classification techniques, a standard database named as University College London’s Archive of Stuttered Speech (UCLASS) is used. The reliability of the classification accuracy is achieved by adopting the two validation schemas, namely, conventional validation and ten-fold cross-validation. For further analysis, parameters selections of the respective classifiers and parameter variation namely order of Linear Predictive Coding based parameterization, parameter used to control the degree of preemphasis filtering, frame length and overlap percentages on the signal pre-processing techniques are investigated. Analysis results reported that the highest classification accuracy is achieved by STFT features and SLIN classifier. By observing the classification accuracy obtained from different acoustical features and classifiers, it can be concluded that it is necessary to evaluate correlation between acoustical features and different classifiers in order to achieve the best classification accuracy. As a conclusion, the proposed feature extraction methods and classifiers can be used in speech dysfluencies classification. Finally, a Graphical User Interface of this work is developed by using MATLAB® based on the results achieved in the experiments.
URI: http://hdl.handle.net/123456789/21614
Appears in Collections:School of Mechatronic Engineering (Theses)

Files in This Item:

File Description SizeFormat
Full text.pdfAccess is limited to UniMAP community6.34 MBAdobe PDFView/Open
p. 1-24.pdfThis item is protected by original copyright289.79 kBAdobe PDFView/Open
View Statistics

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

 

Valid XHTML 1.0! Perpustakaan Tuanku syed Faizuddin Putra, Kampus Pauh Putra, Universiti Malaysia Perlis, 02600, Arau Perlis
TEL: +604-9885420 | FAX: +604-9885405 | EMAIL: rujukan@unimap.edu.my Feedback