Nothing Special   »   [go: up one dir, main page]

CN103000174A - Feature compensation method based on rapid noise estimation in speech recognition system - Google Patents

Feature compensation method based on rapid noise estimation in speech recognition system Download PDF

Info

Publication number
CN103000174A
CN103000174A CN2012104869360A CN201210486936A CN103000174A CN 103000174 A CN103000174 A CN 103000174A CN 2012104869360 A CN2012104869360 A CN 2012104869360A CN 201210486936 A CN201210486936 A CN 201210486936A CN 103000174 A CN103000174 A CN 103000174A
Authority
CN
China
Prior art keywords
noise
speech
estimation
gmm
gaussian
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012104869360A
Other languages
Chinese (zh)
Other versions
CN103000174B (en
Inventor
吕勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hohai University HHU
Original Assignee
Hohai University HHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hohai University HHU filed Critical Hohai University HHU
Priority to CN201210486936.0A priority Critical patent/CN103000174B/en
Publication of CN103000174A publication Critical patent/CN103000174A/en
Application granted granted Critical
Publication of CN103000174B publication Critical patent/CN103000174B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Circuit For Audible Band Transducer (AREA)

Abstract

The invention discloses a feature compensation method based on rapid noise estimation in a speech recognition system. The method is characterized in that noise parameter estimation in the feature compensation is separated from pure speech estimation, and noise estimation and pure speech estimation are achieved through different Gaussian mixture models (GMMs). A GMM containing less Gaussian units is used for extracting noise parameters from a noisy tested speech; another GMM containing more Gaussian units is used for being combined with an estimated single Gaussian noise model to obtain a noisy GMM matched with the current test environment; and finally the noisy GMM is used for calculating the posterior probability of the noisy tested speech and the pure speech feature vector is estimated from the noisy tested speech through the minimum mean square error method. According to the method, estimation accuracy of the pure speech can be guaranteed while the calculated amount is reduced.

Description

Feature compensation method based on rapid noise estimation in speech recognition system
Technical Field
The invention relates to a feature compensation method based on rapid noise estimation in a voice recognition system, in particular to a feature compensation method for rapidly estimating noise parameters by using a Gaussian mixture model with less Gaussian units and estimating pure voice feature vectors from noise-containing test voice by using a Gaussian mixture model with more Gaussian units, belonging to the technical field of voice recognition.
Background
At present, speech recognition systems have achieved good performance in laboratory ideal environments. However, in practical environments, background noise and channel distortion are often unavoidable, which may cause the extracted feature vectors in practical application environments to be severely mismatched with the pre-trained acoustic model, and the performance of the recognizer may be severely deteriorated or even may completely fail. Therefore, the research on the environment compensation technology of the voice recognition, the reduction of the influence of the environment mismatch on the voice recognition system and the improvement of the performance of the voice recognition system in the actual environment have very important significance.
In general, the environment compensation techniques can be divided into front-end feature compensation and back-end model compensation. The feature compensation compensates the speech features in the test environment to match the acoustic models in the training environment. And the model compensation adjusts the acoustic model in the training environment to be matched with the testing environment, and the testing voice is directly recognized. Compared with the back-end model compensation, the front-end characteristic compensation technology has the advantages of small calculated amount, flexible realization and independence with a back-end recognizer, so the application range of the front-end characteristic compensation technology is wider.
In practical applications, it is difficult to ensure that there are enough silence frames for each test speech to estimate the noise parameters. In order to track the change of the environment in time, noise parameters are often required to be extracted from noisy test speech. However, the environmental transformation relationship between the training environment and the test environment is non-linear, and the noise parameters have no closed form solution. Vector Taylor Series (VTS) is an effective noise robust technique, and can well approximate the nonlinear environment transformation relation caused by noise. However, noise parameter estimation based on VTS involves more matrix operations, and the amount of computation is proportional to the number of gaussian units of the speech model. Since in feature compensation the speech model used for noise estimation is also used for estimating the clean speech feature vectors. In order to fully describe the distribution of speech and to guarantee the accuracy of the clean speech estimation, the speech model used for feature compensation must contain enough gaussian units. Therefore, the feature compensation method based on VTS has a large calculation amount, and is difficult to implement in real time on an independent terminal such as an embedded system.
Disclosure of Invention
The purpose of the invention is as follows: aiming at the problems and the defects in the prior art, the invention provides a feature compensation method based on rapid noise estimation in a speech recognition system.
The technical scheme is as follows: a feature compensation method based on fast noise estimation in a speech recognition system is mainly characterized in that noise parameter estimation and pure speech estimation in feature compensation are separated, and the noise estimation and the pure speech estimation are realized by different Gaussian Mixture Models (GMMs). A Gaussian mixture model GMM with less Gaussian units is used for extracting noise parameters from the noisy test voice; the other Gaussian mixture model GMM with more Gaussian units is used for carrying out model combination with the estimated single Gaussian noise model to obtain a noise-containing GMM matched with the current test environment; and finally, calculating the posterior probability of the noise-containing test voice by using the noise-containing GMM, and estimating pure voice feature vectors from the noise-containing test voice by using a Minimum Mean Square Error (MMSE) method.
A feature compensation method based on rapid noise estimation in a voice recognition system specifically comprises a training stage and a testing stage;
the training phase comprises the following specific steps:
(1) extracting a pure voice feature vector from the pure training voice, and adopting Mel-Frequency Cepstral coeffients (MFCC) as a characteristic parameter of the voice;
(2) GMM training with MFCCs for all training voices generates two GMMs: the first GMM contains fewer gaussian units for noise estimation; the second GMM contains more Gaussian units and is used for model combination and pure voice estimation;
(3) performing acoustic Model training by using the training speech of each basic speech unit to generate a Hidden Markov Model (HMM) of each basic speech unit;
the specific steps of the test phase include:
(4) extracting a noisy speech MFCC from a noisy test speech;
(5) extracting noise parameters including a Gaussian mean vector and a covariance matrix of noise from the noisy speech MFCC by using a first GMM;
(6) performing parameter transformation on the mean value and the variance of the second GMM by using the estimated noise parameters, calculating the posterior probability of the noise-containing test voice, and estimating the MFCC of the pure voice by using an MMSE (minimum mean square error) method;
(7) and performing acoustic decoding on the MFCC of the pure voice by using the HMM of each voice unit to obtain a recognition result.
Has the advantages that: compared with the prior art, the characteristic compensation method based on the rapid noise estimation in the voice recognition system separates the noise parameter estimation and the pure voice estimation in the characteristic compensation, and the noise parameter estimation and the pure voice estimation are respectively realized by different voice models, so that the calculation amount can be reduced, and the precision of the pure voice estimation can be ensured.
Drawings
FIG. 1 is a feature compensation framework based on fast noise estimation according to an embodiment of the present invention;
FIG. 2 is a block diagram of a speech recognition system based on fast noise estimation according to an embodiment of the present invention.
Detailed Description
The present invention is further illustrated by the following examples, which are intended to be purely exemplary and are not intended to limit the scope of the invention, as various equivalent modifications of the invention will occur to those skilled in the art upon reading the present disclosure and fall within the scope of the appended claims.
As shown in fig. 1, a gaussian mixture model GMM1 with fewer gaussian units is used to extract noise parameters from noisy test speech; the other Gaussian mixture model GMM2 with more Gaussian units is used for carrying out model combination with the estimated single Gaussian noise model to obtain a noise-containing GMM matched with the current test environment; and finally, calculating the posterior probability of the noise-containing test voice by using the noise-containing GMM, and estimating a pure voice feature vector from the noise-containing test voice by using a minimum mean square error method.
As shown in fig. 2, the feature compensation method based on fast noise estimation mainly includes a training phase and a testing phase. The training stage mainly completes GMM training and HMM training; the testing stage mainly completes the noise parameter estimation and the pure voice estimation.
1. And GMM training:
the GMM is used to model the distribution of speech, and two GMMs are generated from all training speech: GMM1 and GMM 2. GMM1 contains fewer gaussian cells for noise estimation; the GMM2 contains more gaussian cells for pure speech estimation. The covariance matrices for GMM1 and GMM2 both take diagonal matrices.
2. Training by using an HMM:
the invention models each basic phonetic unit of speech recognition by using continuous density HMM, and generates the HMM of each basic phonetic unit by using the training speech of each basic phonetic unit. The number of HMMs depends on the number of speech units. The covariance matrix of all HMMs also takes the diagonal matrix.
3. Noise parameter estimation:
in the cepstral domain, the relationship between noisy speech feature vector y and clean speech feature vector x can be expressed as:
y=x+Clog(1+exp(C-1(n-x)))
(1)
where n represents the additive noise cepstrum feature vector, C and C-1Respectively, a Discrete Cosine Transform (DCT) matrix and an inverse matrix thereof. The mean value mu of x of formula (1)xAnd the initial mean value mu of nn0Nearby, unfolding with a first order VTS yields:
wherein,i denotes a unit matrix of the cell,
Figure GDA00002469320500042
and U is:
Figure GDA00002469320500043
U = Cdiag ( exp ( C - 1 ( μ n 0 - μ x ) ) 1 + exp ( C - 1 ( μ n 0 - μ x ) ) ) C - 1 - - - ( 4 )
in equation (4), diag () represents a diagonal matrix generated by using vector elements in parentheses as diagonal elements.
Taking the mean and variance on both sides of equation (2) can result in:
Figure GDA00002469320500045
Σy=(I-U)Σx(I-U)T+U∑n UT
(6)
wherein, muy、μxAnd munMean vectors, Σ, representing noisy speech y, clean speech x and additive noise n, respectivelyy、ΣxSum-sigmanRespectively representing their covariance matrices.
For the mth gaussian unit of GMM1, equations (5) and (6) are expressed as:
Figure GDA00002469320500051
σy,m=(Vm·Vmx,m+(Um·Umn
(8)
wherein, Vm=I-Um,σy,m、σx,mAnd σnRespectively represent ∑y,m、Σx,mSum-sigmanThe diagonal element vector of (2).
Substituting the formula (7) and the formula (8) into the auxiliary function respectively to obtain the noise parameter munAnd σnMaximum likelihood estimation of (2):
Figure GDA00002469320500052
σ n = [ Σ m = 1 M 1 Σ t = 1 T γ m ( t ) G m ( U m · U m ) ] - 1 [ Σ m = 1 M 1 Σ t = 1 T γ m ( t ) G m ( ( y t - μ y , m ) · ( y t - μ y , m ) - ( V m · V m ) σ x , m ) ] - - - ( 10 )
wherein M is1The number of Gaussian cells that are GMM 1; gamma raym(t)=P(kt=m|ytλ) represents the noisy speech feature vector y for the t-th frame given the a priori parameter λ of GMM1tPosterior probability of the mth gaussian unit belonging to GMM 1; gmGiven by:
G m = ( U m T · U m T ) diag [ ( ( V m · V m ) σ x , m + ( U m · U m ) σ n 0 ) - 2 ] - - - ( 11 )
in formula (11), σn0Is σnThe initial value of (c).
4. And (3) pure voice estimation:
estimating noiseAcoustic parameter munAnd σnThereafter, the mean and variance of GMM2 are first parametrically transformed using equations (7) and (8), at which time μy,mAnd σy,mMean and variance of noisy speech representing the mth gaussian unit of GMM 2; then using the noisy speech parameter muy,mAnd σy,mCalculating the posterior probability of the current test voice belonging to the Mth Gaussian unit of GMM2
Figure GDA00002469320500061
Finally, MMSE estimation of pure speech characteristic vector is obtained by the following formula
Figure GDA00002469320500062
x ^ t = E ( x t | y t ) ≈ y t - Σ m = 1 M 2 γ ‾ m ( t ) ( C log ( 1 + exp ( C - 1 ( μ n - μ x , m ) ) ) ) - - - ( 12 )
Wherein M is2Representing the number of gaussian cells of GMM 2.
First order dynamic coefficient of clean speech feature vector
Figure GDA00002469320500064
And second order dynamic coefficient
Figure GDA00002469320500065
Can be obtained by comparing the estimated static coefficient
Figure GDA00002469320500066
And carrying out time domain difference to obtain the target.

Claims (5)

1. A feature compensation method based on fast noise estimation in a speech recognition system is characterized by comprising the following steps:
(1) adopting a Mel frequency cepstrum coefficient as a characteristic parameter of a voice recognition system, wherein the characteristic compensation aims at extracting a pure voice MFCC from a noisy test voice;
(2) in the training phase, the distribution of speech is modeled using a gaussian mixture model, and two GMMs are generated using the entire training speech: a first GMM and a second GMM;
(3) modeling background noise by using a single Gaussian model, and extracting a mean vector and a covariance matrix of the single Gaussian noise model from noise-containing test voice in order to track the change of the environment in real time;
(4) extracting noise parameters including a Gaussian mean vector and a covariance matrix of noise from a noisy test speech MFCC by using a first GMM;
(5) performing parameter transformation on the mean value and the variance of the second GMM2 by using the estimated noise parameters, namely performing model combination on a single Gaussian noise model and the second GMM to obtain the mean value and the variance of the noisy speech of the second GMM;
(6) calculating the posterior probability of the noise-containing test voice by using the mean value and the variance of the noise-containing voice of the second GMM, and estimating the MFCC of the pure voice by using a minimum mean square error method;
(7) the first order dynamic coefficient and the second order dynamic coefficient of the clean voice feature vector are not directly estimated from the noise-containing test voice, but are obtained by carrying out time domain difference on the estimated static coefficient.
2. The method of fast noise estimation based feature compensation in a speech recognition system according to claim 1, wherein: the first gaussian mixture model for noise parameter estimation contains fewer gaussian units and thus is less computationally intensive and allows for fast estimation of the mean and variance of the noise from noisy test speech.
3. The method of fast noise estimation based feature compensation in a speech recognition system according to claim 1, wherein: the second Gaussian mixture model for pure speech estimation contains more Gaussian units, so that the distribution of speech can be fully described, and an accurate pure speech estimation value can be obtained.
4. The method of fast noise estimation based feature compensation in a speech recognition system according to claim 1, wherein: the covariance matrices of the gaussian mixture model for both the noise parameter estimation and the clean speech estimation take the diagonal matrix.
5. The method of fast noise estimation based feature compensation in a speech recognition system according to claim 1, wherein: the first GMM and the second GMM only model the static coefficient of the feature vector and do not consider the dynamic coefficient; the noise parameter estimation based on the first GMM and the clean speech estimation based on the second GMM also only calculate the static coefficients of the noise and the speech; the dynamic coefficients of the clean speech feature vector are obtained by time-domain differencing the estimated static coefficients.
CN201210486936.0A 2012-11-26 2012-11-26 Feature compensation method based on rapid noise estimation in speech recognition system Expired - Fee Related CN103000174B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210486936.0A CN103000174B (en) 2012-11-26 2012-11-26 Feature compensation method based on rapid noise estimation in speech recognition system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210486936.0A CN103000174B (en) 2012-11-26 2012-11-26 Feature compensation method based on rapid noise estimation in speech recognition system

Publications (2)

Publication Number Publication Date
CN103000174A true CN103000174A (en) 2013-03-27
CN103000174B CN103000174B (en) 2015-06-24

Family

ID=47928671

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210486936.0A Expired - Fee Related CN103000174B (en) 2012-11-26 2012-11-26 Feature compensation method based on rapid noise estimation in speech recognition system

Country Status (1)

Country Link
CN (1) CN103000174B (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103646648A (en) * 2013-11-19 2014-03-19 清华大学 Noise power estimation method
CN104464728A (en) * 2014-11-26 2015-03-25 河海大学 Speech enhancement method based on Gaussian mixture model (GMM) noise estimation
CN104485108A (en) * 2014-11-26 2015-04-01 河海大学 Noise and speaker combined compensation method based on multi-speaker model
CN104900232A (en) * 2015-04-20 2015-09-09 东南大学 Isolation word identification method based on double-layer GMM structure and VTS feature compensation
CN105355198A (en) * 2015-10-20 2016-02-24 河海大学 Multiple self-adaption based model compensation type speech recognition method
CN105355199A (en) * 2015-10-20 2016-02-24 河海大学 Model combination type speech recognition method based on GMM (Gaussian mixture model) noise estimation
CN105352541A (en) * 2015-10-14 2016-02-24 国网福建省电力有限公司 Transformer operation auxiliary monitoring system and method based on disaster prevention and reduction system of power grid
CN105448303A (en) * 2015-11-27 2016-03-30 百度在线网络技术(北京)有限公司 Voice signal processing method and apparatus
CN106331969A (en) * 2015-07-01 2017-01-11 奥迪康有限公司 Enhancement of noisy speech based on statistical speech and noise models
CN106356058A (en) * 2016-09-08 2017-01-25 河海大学 Robust speech recognition method based on multi-band characteristic compensation
CN106384588A (en) * 2016-09-08 2017-02-08 河海大学 Additive noise and short time reverberation combined compensation method based on vector Taylor series
CN106653048A (en) * 2016-12-28 2017-05-10 上海语知义信息技术有限公司 Method for separating sound of single channels on basis of human sound models
CN106663446A (en) * 2014-07-02 2017-05-10 微软技术许可有限责任公司 User environment aware acoustic noise reduction
CN107167669A (en) * 2017-04-28 2017-09-15 湘潭大学 A kind of electromagnetic radiation measuring modification method under the environment based on white Gaussian noise
CN107919115A (en) * 2017-11-13 2018-04-17 河海大学 A kind of feature compensation method based on nonlinear spectral conversion
CN107945795A (en) * 2017-11-13 2018-04-20 河海大学 A kind of accelerated model adaptive approach based on Gaussian classification
CN108053835A (en) * 2017-11-13 2018-05-18 河海大学 A kind of noise estimation method based on passage Taylor series
CN108877777A (en) * 2018-08-01 2018-11-23 云知声(上海)智能科技有限公司 A kind of audio recognition method and system
CN115662409A (en) * 2022-10-27 2023-01-31 亿铸科技(杭州)有限责任公司 Voice recognition method, device, equipment and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100262423A1 (en) * 2009-04-13 2010-10-14 Microsoft Corporation Feature compensation approach to robust speech recognition

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100262423A1 (en) * 2009-04-13 2010-10-14 Microsoft Corporation Feature compensation approach to robust speech recognition

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MORENO P J,RAJ B,STERN R M: "A vector Taylor series approach for environment-independent speech recognition", 《IEEE INT CONF ON ACOUSTICS,SPEECH,AND SIGNAL PROCESSING》 *
吕勇,吴镇扬: "基于矢量泰勒级数的模型自适应算法", 《电子与信息学报》 *
吕勇,吴镇扬: "基于矢量泰勒级数的鲁棒语音识别", 《天津大学学报》 *

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103646648A (en) * 2013-11-19 2014-03-19 清华大学 Noise power estimation method
CN103646648B (en) * 2013-11-19 2016-03-23 清华大学 A kind of noise power estimation method
CN106663446A (en) * 2014-07-02 2017-05-10 微软技术许可有限责任公司 User environment aware acoustic noise reduction
CN106663446B (en) * 2014-07-02 2021-03-12 微软技术许可有限责任公司 User environment aware acoustic noise reduction
CN104464728A (en) * 2014-11-26 2015-03-25 河海大学 Speech enhancement method based on Gaussian mixture model (GMM) noise estimation
CN104485108A (en) * 2014-11-26 2015-04-01 河海大学 Noise and speaker combined compensation method based on multi-speaker model
CN104900232A (en) * 2015-04-20 2015-09-09 东南大学 Isolation word identification method based on double-layer GMM structure and VTS feature compensation
CN106331969A (en) * 2015-07-01 2017-01-11 奥迪康有限公司 Enhancement of noisy speech based on statistical speech and noise models
CN105352541A (en) * 2015-10-14 2016-02-24 国网福建省电力有限公司 Transformer operation auxiliary monitoring system and method based on disaster prevention and reduction system of power grid
CN105352541B (en) * 2015-10-14 2017-12-08 国网福建省电力有限公司 A kind of transformer station high-voltage side bus auxiliary monitoring system and its monitoring method based on power network disaster prevention disaster reduction system
CN105355198B (en) * 2015-10-20 2019-03-12 河海大学 It is a kind of based on multiple adaptive model compensation audio recognition method
CN105355198A (en) * 2015-10-20 2016-02-24 河海大学 Multiple self-adaption based model compensation type speech recognition method
CN105355199A (en) * 2015-10-20 2016-02-24 河海大学 Model combination type speech recognition method based on GMM (Gaussian mixture model) noise estimation
CN105355199B (en) * 2015-10-20 2019-03-12 河海大学 A kind of model combination audio recognition method based on the estimation of GMM noise
CN105448303A (en) * 2015-11-27 2016-03-30 百度在线网络技术(北京)有限公司 Voice signal processing method and apparatus
CN106384588A (en) * 2016-09-08 2017-02-08 河海大学 Additive noise and short time reverberation combined compensation method based on vector Taylor series
CN106384588B (en) * 2016-09-08 2019-09-10 河海大学 The hybrid compensation method of additive noise and reverberation in short-term based on vector Taylor series
CN106356058B (en) * 2016-09-08 2019-08-20 河海大学 A kind of robust speech recognition methods based on multiband feature compensation
CN106356058A (en) * 2016-09-08 2017-01-25 河海大学 Robust speech recognition method based on multi-band characteristic compensation
CN106653048A (en) * 2016-12-28 2017-05-10 上海语知义信息技术有限公司 Method for separating sound of single channels on basis of human sound models
CN106653048B (en) * 2016-12-28 2019-10-15 云知声(上海)智能科技有限公司 Single channel sound separation method based on voice model
CN107167669B (en) * 2017-04-28 2019-08-06 湘潭大学 A kind of electromagnetic radiation measuring modification method under the environment based on white Gaussian noise
CN107167669A (en) * 2017-04-28 2017-09-15 湘潭大学 A kind of electromagnetic radiation measuring modification method under the environment based on white Gaussian noise
CN108053835A (en) * 2017-11-13 2018-05-18 河海大学 A kind of noise estimation method based on passage Taylor series
CN107945795A (en) * 2017-11-13 2018-04-20 河海大学 A kind of accelerated model adaptive approach based on Gaussian classification
CN107919115A (en) * 2017-11-13 2018-04-17 河海大学 A kind of feature compensation method based on nonlinear spectral conversion
CN107945795B (en) * 2017-11-13 2021-06-25 河海大学 Rapid model self-adaption method based on Gaussian classification
CN108053835B (en) * 2017-11-13 2021-07-09 河海大学 Noise estimation method based on channel Taylor series
CN108877777A (en) * 2018-08-01 2018-11-23 云知声(上海)智能科技有限公司 A kind of audio recognition method and system
CN115662409A (en) * 2022-10-27 2023-01-31 亿铸科技(杭州)有限责任公司 Voice recognition method, device, equipment and storage medium
CN115662409B (en) * 2022-10-27 2023-05-05 亿铸科技(杭州)有限责任公司 Voice recognition method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN103000174B (en) 2015-06-24

Similar Documents

Publication Publication Date Title
CN103000174A (en) Feature compensation method based on rapid noise estimation in speech recognition system
Lei et al. A noise robust i-vector extractor using vector taylor series for speaker recognition
Hasan et al. Duration mismatch compensation for i-vector based speaker recognition systems
Sehr et al. Reverberation model-based decoding in the logmelspec domain for robust distant-talking speech recognition
Prasad et al. Improved cepstral mean and variance normalization using Bayesian framework
CN102945670B (en) Multi-environment characteristic compensation method for voice recognition system
Frey et al. Algonquin-learning dynamic noise models from noisy speech for robust speech recognition
Gales Model-based approaches to handling uncertainty
Li et al. Improvements to VTS feature enhancement
CN104485108A (en) Noise and speaker combined compensation method based on multi-speaker model
Gong A method of joint compensation of additive and convolutive distortions for speaker-independent speech recognition
Sarkar et al. Stochastic feature compensation methods for speaker verification in noisy environments
Tran et al. Nonparametric uncertainty estimation and propagation for noise robust ASR
Dionelis et al. Modulation-domain speech enhancement using a Kalman filter with a Bayesian update of speech and noise in the log-spectral domain
Tran et al. Extension of uncertainty propagation to dynamic MFCCs for noise robust ASR
Faubel et al. On expectation maximization based channel and noise estimation beyond the vector Taylor series expansion
Xu et al. Vector taylor series based joint uncertainty decoding.
Chung Vector Taylor series based model adaptation using noisy speech trained hidden Markov models
Li et al. HMM adaptation using a phase-sensitive acoustic distortion model for environment-robust speech recognition
Kim et al. Application of VTS to environment compensation with noise statistics
Wu et al. An environment-compensated minimum classification error training approach based on stochastic vector mapping
Wang et al. Improving reverberant VTS for hands-free robust speech recognition
Dimitriadis et al. An alternative front-end for the AT&T WATSON LV-CSR system
Du et al. HMM-based pseudo-clean speech synthesis for splice algorithm
Sim et al. A trajectory-based parallel model combination with a unified static and dynamic parameter compensation for noisy speech recognition

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150624

Termination date: 20211126