Paper
27 August 1993 Speaker recognition using neural network and adaptive wavelet transform
Mohammad Bodruzzaman, Xingkang Li, Kah Eng Kuah, Lamar Crowder, Mohan Malkani, Harold H. Szu, Brian A. Telfer
Author Affiliations +
Abstract
The same word uttered by different people has different waveforms. It has also been observed that the same word uttered by the same person has different waveform at different times. This difference can be characterized by some time domain dilation effects in the waveform. In our experiment a set of words was selected and each word was uttered eight times by five different speakers. The objective of this work is to extract a wavelet basis function for the speech data generated by each individual speaker. The wavelet filter coefficients are then used as a feature set and fed into a neural network-based speaker recognition system. This is an attempt to cascade a wavelet network (wavenet) and a neural network (neural-net) for feature extraction and classification respectively and applied for speaker recognition. The results show very high promise and good prospects to couple a wavelet network and neural networks.
© (1993) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Mohammad Bodruzzaman, Xingkang Li, Kah Eng Kuah, Lamar Crowder, Mohan Malkani, Harold H. Szu, and Brian A. Telfer "Speaker recognition using neural network and adaptive wavelet transform", Proc. SPIE 1961, Visual Information Processing II, (27 August 1993); https://doi.org/10.1117/12.150976
Lens.org Logo
CITATIONS
Cited by 8 scholarly publications.
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Wavelets

Neural networks

Speaker recognition

Wavelet transforms

Visual information processing

Acoustics

Binary data

Back to Top