Nonlinear spectro-temporal features based on a cochlear model for automatic speech recognition in a noisy situation

Cited 3 time in webofscience Cited 3 time in scopus
  • Hit : 435
  • Download : 0
DC FieldValueLanguage
dc.contributor.author최용선ko
dc.contributor.authorLee, Soo-Youngko
dc.date.accessioned2019-04-15T14:52:58Z-
dc.date.available2019-04-15T14:52:58Z-
dc.date.created2013-10-07-
dc.date.created2013-10-07-
dc.date.issued2013-09-
dc.identifier.citationNEURAL NETWORKS, v.45, pp.62 - 69-
dc.identifier.issn0893-6080-
dc.identifier.urihttp://hdl.handle.net/10203/254498-
dc.description.abstractA nonlinear speech feature extraction algorithm was developed by modeling human cochlear functions, and demonstrated as a noise-robust front-end for speech recognition systems. The algorithm was based on a model of the Organ of Corti in the human cochlea with such features as such as basilar membrane (BM), outer hair cells (OHCs), and inner hair cells (IHCs). Frequency-dependent nonlinear compression and amplification of OHCs were modeled by lateral inhibition to enhance spectral contrasts. In particular, the compression coefficients had frequency dependency based on the psychoacoustic evidence. Spectral subtraction and temporal adaptation were applied in the time-frame domain. With long-term and short-term adaptation characteristics, these factors remove stationary or slowly varying components and amplify the temporal changes such as onset or offset. The proposed features were evaluated with a noisy speech database and showed better performance than the baseline methods such as mel-frequency cepstral coefficients (MFCCs) and RASTA-PLP in unknown noisy conditions. (C) 2013 Elsevier Ltd. All rights reserved.-
dc.languageEnglish-
dc.publisherPERGAMON-ELSEVIER SCIENCE LTD-
dc.titleNonlinear spectro-temporal features based on a cochlear model for automatic speech recognition in a noisy situation-
dc.typeArticle-
dc.identifier.wosid000323589200006-
dc.identifier.scopusid2-s2.0-84880847951-
dc.type.rimsART-
dc.citation.volume45-
dc.citation.beginningpage62-
dc.citation.endingpage69-
dc.citation.publicationnameNEURAL NETWORKS-
dc.identifier.doi10.1016/j.neunet.2013.02.006-
dc.contributor.localauthorLee, Soo-Young-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorNonlinear auditory features-
dc.subject.keywordAuthorCochlear model-
dc.subject.keywordAuthorNonlinear amplification-
dc.subject.keywordAuthorNoise-robust speech recognition-
dc.subject.keywordAuthorAdaptive gain control-
dc.subject.keywordPlusMAGNITUDE-
dc.subject.keywordPlusMASKING-
dc.subject.keywordPlusSIGNALS-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 3 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0