KEYWORDS: Acoustics, Speaker recognition, Data hiding, Information fusion, Feature extraction, System identification, Data modeling, Biometrics, Expectation maximization algorithms, Performance modeling
Prior knowledge helps to make the speaker recognition system more reliable and robust. This paper presents a uniform framework of feature-level fusion to incorporate the prior knowledge for speaker recognition using gender information based on dynamic Bayesian network (DBN). DBNs are a new statistical approach, with the ability to handle hidden variables and missing data in a principled way with high extensibility. And thus, DBNs can describe the prior knowledge conveniently. Our contribution is to apply DBNs to construct a general feature-level fusion to combine the general acoustic feature like MFCC and prior information like gender into a single DBN for speaker identification. In our framework, gender information become additional observed data to influence both hidden variables and observed acoustic data. Experimental evaluation over a subnet of YOHO corpus show promising results.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.