The speech recognition system of the present invention includes: a sound source separating section which separates mixed speeches from multiple sound sources; a mask generating section which generates a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section; and a speech recognizing section which recognizes speeches separated by the sound source separating section using soft masks generated by the mask generating section.
|
1. A speech recognition system comprising:
multiple sound sources;
a sound source separating section which separates mixed speeches from the multiple sound sources; and
at least one processor configured to:
e####
generate a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section, and
recognize speeches separated by the sound source separating section using the soft masks,
wherein the reliability of separation R(f,t) is defined as
where Y is a sum of a speech Ŝm, a background noise bn and a leak.
5. A method for generating a soft mask for a speech recognition system, the method comprising:
separating, at a sound source separating section of the speech recognition system, mixed speeches from multiple sound sources;
generating, at a mask generating section of the speech recognition system, a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section;
recognizing, at a speech recognizing section of the speech recognition system, speeches separated by the sound source separating section using soft masks generated by the mask generating section, the soft mask being determined using a function of the reliability of separation, which has at least one parameter;
obtaining a histogram of the reliability of separation; and
determining a value of said at least one parameter from a form of the histogram of the reliability of separation,
wherein the reliability of separation R(f,t) is defined as
where Y is a sum of a speech Ŝm, a background noise bn and a leak.
4. A method for generating a soft mask for a speech recognition system, the method comprising:
separating, at a sound source separating section of the speech recognition system, mixed speeches from multiple sound sources;
generating, at a mask generating section of the speech recognition system, a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section;
recognizing, at a speech recognizing section of the speech recognition system, speeches separated by the sound source separating section using soft masks generated by the mask generating section, the soft mask being determined using a function of the reliability of separation, which has at least one parameter;
determining a search space of said at least one parameter;
obtaining a speech recognition rate of the speech recognition system while changing a value of the speech recognition system in the search space; and
setting the value which maximizes a speech recognition rate of the speech recognition system to said at least one parameter,
wherein the reliability of separation R(f,t) is defined as
where Y is a sum of a speech Ŝm, a background noise bn and a leak.
2. A speech recognition system according to
1/(1+exp(−a(R−b)) where R represents reliability of separation and a and b represent constants.
3. A speech recognition system according to
6. A method for generating a soft mask for a speech recognition system according to
μ1 and μ2 (μ1<μ2)
indicate mean values and
σ1 and σ2
indicate standard deviations and R indicates reliability of separation, the mean values and standard deviations
μ1, μ2, σ1 and σ2
are estimated by fitting the histogram of reliability of separation R with a first probability density function of normal distribution f1(R) which has
(μ1,σ1)
and a second probability density function of normal distribution f2(R) which has
(μ2,σ2)
and the soft mask is generated using f1(R), f2(R),
μ1 and μ2.
7. A method for generating a soft mask for a speech recognition system according to
S(R)=0 when R<μ1,
S(R)=f2(R)/f(R) when μ1≦R≦μ2
S(R)=1 when μ2<R.
8. A method for generating a soft mask for a speech recognition system according to
wherein SM(R) represents a soft missing feature mask (MFM).
9. A method for generating a soft mask for a speech recognition system according to
μ1<R<μ2 is set to b and a is determined such that
1/(1+exp(−a(R−b)) is fit to
f2(R)/f(R) and the value of the missing feature mask (MFM) S(R) is determined by
S(R)=1/(1+exp(−a(R−b)). 10. A speech recognition system according to
μ1 and μ2 (μ1<μ2)
indicate mean values,
σ1 and σ2
indicate standard deviations, and R indicates the reliability of separation, the mean values and standard deviations
μ1, μ2, σ1 and σ2 are estimated by fitting a histogram of the reliability of separation R with a first probability density function of normal distribution f1(R) which has
(μ1,σ1)
and a second probability density function of normal distribution f2(R) which has
(μ2,σ2)
and the soft mask is generated using f1(R), f2(R),
μ1 and μ2.
|
This application claims the benefit of U.S. Provisional Application No. 61/136,225 filed on Aug. 20, 2008, and claiming priority of Japanese patent application JP 2009-185164, filed on Aug. 7, 2009. The disclosure of the priority applications are hereby incorporated by reference herein in their entirety.
1. Field of the Invention
The present invention relates to a speech recognition system for simultaneous recognition of speeches from multiple sources and a method for generating a mask for the speech recognition system.
2. Description of the Related Art
The technology of simultaneous recognition of speeches from multiple sources is important for robots that work in the real environment. A speech recognition system for simultaneous recognition of speeches from multiple sources separates speeches according to sources and performs speech recognition using the acoustic feature of a separated speech. In speech recognition a mask prepared according to reliability of separation is used for each acoustic feature (for example, Reference 2). Conventionally hard masks which are binary, taking a value of 0 or 1 are used as such masks (for example, Reference 3). Although soft masks which take continuous values between 0 and 1 are known (for example, Reference 4), a soft mask used for a speech recognition system for simultaneous recognition of speeches from multiple sources has not been developed. The reason is that conventionally those skilled in the art thought that hard masks are more appropriate for a speech recognition system for simultaneous recognition of speeches from multiple sources than soft masks (for example, Reference 2). Thus, a speech recognition system provided with a soft mask appropriately designed for simultaneous recognition of speeches from multiple sources and having an increased speech recognition rate has not been developed.
Accordingly, there is a need for a speech recognition system provided with a soft mask appropriately designed for simultaneous recognition of speeches from multiple sources and having a higher speech recognition rate has not been developed.
A speech recognition system according to the invention includes a sound source separating section which separates mixed speeches from multiple sound sources; a mask generating section which generates a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section; and a speech recognizing section which recognizes speeches separated by the sound source separating section using soft masks generated by the mask generating section.
The speech recognition system according to the invention recognizes speeches using a soft mask which can take continuous values between 0 and 1 for each separated speech depending on reliability of separation to increase a speech recognition rate.
In a speech recognition system according to an embodiment of the invention, the soft masks are determined using a sigmoid function
1/(1+exp(−a(R−b))
where R represents reliability of separation and a and b represent constants.
In the speech recognition system according to the embodiment, the soft masks can be easily adjusted by changing constants a and b of the sigmoid function.
In a speech recognition system according to another embodiment of the invention, the soft masks are determined using a probability density function of a normal distribution, which has a variable R which represents reliability of separation.
In the speech recognition system according to the embodiment, the soft masks can be easily adjusted by changing a form of the probability density function of the normal distribution.
A method for generating a soft mask for a speech recognition system according to the invention, is used to generate a soft mask for the system including: a sound source separating section which separates mixed speeches from multiple sound sources; a mask generating section which generates a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section; and a speech recognizing section which recognizes speeches separated by the sound source separating section using soft masks generated by the mask generating section, the soft mask being determined using a function of reliability of separation, which has at least one parameter. The method includes the steps of: determining a search space of said at least one parameter; obtaining a speech recognition rate of the speech recognition system while changing a value of speech recognition system in the search space; and setting the value which maximizes a speech recognition rate of the speech recognition system to said at least one parameter.
In the method for generating a soft mask for a speech recognition system according to the invention, the soft mask is determined using a function of reliability of separation, which has at least one parameter. Accordingly, the at least one parameter can be determined such that the speech recognition rate is maximized by obtaining speech recognition rates for the soft mask with various values of the at least one parameter.
A method for generating a soft mask for a speech recognition system according to the invention, is used to generate a soft mask for the system including: a sound source separating section which separates mixed speeches from multiple sound sources; a mask generating section which generates a soft mask which can take continuous values between 0 and 1 for each separated speech according to reliability of separation in separating operation of the sound source separating section; and a speech recognizing section which recognizes speeches separated by the sound source separating section using soft masks generated by the mask generating section, the soft mask being determined using a function of reliability of separation, which has at least one parameter. The method includes the steps of: obtaining a histogram of reliability of separation; and determining a value of said at least one parameter from a form of the histogram of reliability of separation.
In the method for generating a soft mask for a speech recognition system according to the invention, the soft mask is determined using a function of reliability of separation, which has at least one parameter. Accordingly, the at least one parameter can be appropriately determined by obtaining a form of the histogram of reliability of separation.
In a method for generating a soft mask for a speech recognition system according to an embodiment of the invention, assuming that
In the method for generating a soft mask for a speech recognition system according to the embodiment, the soft mask can be easily generated by fitting the histogram of reliability of separation R with probability density functions of normal distributions.
In a method for generating a soft mask for a speech recognition system according to another embodiment of the invention, assuming that a value of the soft mask is S(R) and f(R)=f1(R)+f2(R),
In the method for generating a soft mask for a speech recognition system according to the embodiment, the soft mask can be easily determined by using the probability density functions of normal distributions, obtained from the histogram of reliability of separation R.
In a method for generating a soft mask for a speech recognition system according to another embodiment of the invention, assuming that a value of the soft mask is S(R),
In the method for generating a soft mask for a speech recognition system according to the embodiment, the soft mask can be easily determined by using the probability density functions of normal distributions, obtained from the histogram of reliability of separation R.
In a method for generating a soft mask for a speech recognition system according to another embodiment of the invention, a value of R at the intersection of f1(R) and f2(R) which satisfies
μ1<R<μ2
is set to b and a is determined such that
1/(1+exp(−a(R−b))
is fit to
f2(R)/f(R)
and the value of the MFM S(R) is determined by
S(R)=1/(1+exp(−a(R−b)).
In the method for generating a soft mask for a speech recognition system according to the embodiment, the soft mask can be easily determined by using the probability density functions of normal distributions, obtained from the histogram of reliability of separation R.
The speech recognition system 100 performs simultaneous recognition of speeches from multiple sound sources such as multiple speakers. The sound source separating section 101 receives mixed speeches from multiple sound sources, for example, through 8-channel microphone array. The sound source separating section 101 sends separated sounds to the speech recognizing section 105. Further, the sound source separating section 101 sends information which is used by the mask generating section 103 for generating masks as described below, to the mask generating section 103. The mask generating section 103 generates masks using the information form the sound source separating section 101 and sends the masks to the speech recognizing section 105. The speech recognizing section 105 obtains acoustic features of the separated sounds and performs speech recognition using the masks from the mask generating section 103. Functions of the speech recognizing section 105, the sound source separating section 101 and the mask generating section 103 will be described below.
Speech Recognizing Section
The speech recognizing section 105 outputs a sequence of phonemes from a sequence of acoustic feature sets of separated speech and the corresponding sequence of masks based on missing-feature theory. An acoustic feature set and a mask are calculated for each time frame. A sequence of acoustic feature sets means acoustic feature sets each of which is calculated for each time frame and which are arranged in time order while a sequence of masks means masks each of which is calculated for each time frame and which are arranged in time order. The speech recognizing section 105 is a hidden Markov model (HMM) based recognizer, which is commonly used in conventional Automatic Speech Recognition (ASR) systems. The difference between the ASR method of the speech recognizing section 105 according to the embodiment and conventional ASR method is described below.
In conventional ASR methods, estimation of a path with maximum likelihood is based on state transition and output probabilities in the HMM. This process of estimating output probability is modified in the speech recognizing section 105 according to the embodiment as follows. Let
M=[M(1), . . . M(F)]
be an Missing-Feature Mask (MFM) vector and M(f) represent the reliability of the f-th acoustic feature. F is the size of the MFM vector and a MFM vector for a time frame has F elements. The output probability
where P(|) is a probability operator, and L represents the number of distributions of mixture of normal distributions while l represents an index of the number.
x=[x(1), . . . , x(F)]
is an acoustic feature vector, F is the size of the acoustic feature vector. That is, an acoustic feature vector for a time frame has F elements.
For the speech recognizing section 105, Multiband Julius [References 5 and 6] is used, which is an extension of the Japanese real-time large vocabulary speech recognition engine Julius [Reference 7].
Sound Source Separating Section
The GSS approach of Reference 9 has been modified so as to provide faster adaptation using a stochastic gradient and shorter time frame estimations [Reference 11]. The initial separation using GSS is followed by the multi-channel post-filter based on a generalization of beamformer post-filtering [Reference 11] for multiple sources. This post-filter uses adaptive spectral estimation of background noise and interfering sources for enhancing the signal produced during the initial separation.
The essential feature of the sound source separating section 101 is that the noise estimate is decomposed into stationary and transient components, which are assumed to be due to the leakage between the output channels in the initial separation stage.
This GSS method operates in the frequency domain. Let
W(f,t)
The output of the GSS algorithm is then enhanced by a frequency-domain post-filter based on the optimal estimator originally proposed in Reference 12.
An input of the multi-channel post-filter is the output of GSS;
y(f,t)=(y1(f,t), . . . , yM(f,t)).
The stationary noise estimate, λmleak (f, t), is obtained using the minima controlled recursive average (MCRA) [Reference 10]. To estimate λmleak, it is assumed that the interference from other sources is reduced by factor η (typically −10 dB≦η≦−5 dB) by LSS. The leakage estimate is expressed as below.
Feature vector of 48 spectral-related features are used. The MFM is a vector corresponding to 24 static spectral features and 24 dynamic spectral features. Each element of a vector represents the reliability of each feature. In conventional MFM generation, a binary MFM (i.e., 1 for reliable and 0 for unreliable) was used. The mask generating section 103 generates a soft MFM whose element of vector ranges from 0.0 to 1.0. In this context, “generating a soft MFM” means determining a value of the soft MFM according to a formula defining the soft MFM.
The mask generating section 103 performs calculation of a MFM using input ym(f, t), output ŝm(f, t), and the estimated background noise, b(f, t), of the multi-channel post-filter. These parameters are calculated from the multi-channel input speech with object related transfer function (ORTF). The variables filtered by the Mel filter bank are Ym (f, t), Ŝm (f, t), and BN(f, t), respectively. The Mel filter bank is a group of filters arranged at regular intervals on the Mel frequency axis.
For each Mel-frequency band, the feature is considered reliable if the ratio of the output energy over the input energy is greater than a threshold, θhard. This assumes that the more noise present in a certain frequency band, the lower the post-filter gain will be for that band.
Let R(f, t) be the reliability of separation defined as
Y is a sum of speech Ŝm, background noise BN and leak. So, the reliability of separation becomes 1 when there exists no leak (when a speech is completely separated without blending of any other speeches) and approaches 0 as the leak becomes larger.
The hard MFM θhard (f, t) for the static spectral feature [x(1), . . . , x(24)] is defined as
where whard is weight factor (0.0≦whard≦1.0). The hard MFM HMd(f, t) for the dynamic spectral features
The unweighted hard mask (Qhard(f, t/θhard)) for the dynamic feature is 1 if only the hard masks for the static features within two contiguous frames are 1.
The soft MFM SMs (f, t) for the static spectral feature
where wsoft is weight factor (0.0≦wsoft≦1.0). Qsoft (•|k, θsoft) is a modified sigmoid function which has two tunable parameters. k and θsoft, correspond to the tilt and position of the sigmoid function. How to determine the parameters of the modified sigmoid function will be described later.
The dynamic spectral features are robust against leak noise and stationary background noise because the dynamic spectral feature defined as difference of contiguous static features can cancel leak noise and stationary background noise. The static spectral feature is less robust than dynamic spectral feature against such noises. Therefore, it is expected that recognition rate is improved when contribution of the dynamic spectral feature is higher than that of the static spectral feature. To increase the contribution of the dynamic spectral feature, it is effective to set a small value to w.
The soft MFM SMd(f, t) for the dynamic spectral feature is defined as below.
In the above, the soft MFM was created using the modified sigmoid function. In general a soft MFM can be created in various methods. Various methods to create a soft MFM will be described below.
In step S1010 of
In step S1020 of
In step S1030 of
Assuming that a value of the MFM is S(R) and f(R)=f1(R)+f2(R),
Assuming that a value of the MFM is S(R),
Third Method
A value of R at the intersection of f1(R) and f2(R) which satisfies
μ1<R<μ2
is set to b and a is determined such that
1/(1+exp(−a(R−b))
is fit to
f2(R)/f(R)
and the value of the MFM S(R) is determined by
S(R)=1/(1+exp(−a(R−b)).
Experiments
To evaluate the efficiency of the speech recognition system according to the embodiment, experiments on recognition of three simultaneous speech signals were performed. A humanoid robot (SIG2 robot) was used for the experiments with eight omnidirectional microphones symmetrically placed on the body. The transfer function of the robot's body affected the captured sound since the microphones were not in the air.
Three loudspeakers were used to generate three simultaneous speech signals and the simultaneous speech signals were recorded. The reverberation time was about 0.35 seconds.
To optimize the parameters, θhard, θsoft, k, and w, in Equations (9), (12), and (13), experiments were performed on recognition of three simultaneous speech signals.
In step S2010 of
In step S2020 of
TABLE 1
Parameters
Hard mask
Soft mask
Threshold θhard
0.0-0.4 (step 0.05)
—
Tilt k
—
−80-160 (step 20)
Center θsoft
—
−0.0 0.4 (step 0.05)
Weight w
0.0-1.0 (step 0.1)
0.0-1.0 (step 0.1)
In step S2030 of
In step S2040 of
The results show that the optimal threshold (the parameter maximizing the speech recognition rate) for the hard mask θhard was 0.1 and the optimal parameter set (the parameter set maximizing the speech recognition rate) for the soft mask was
{w,θsoft,k}={0.3,0.2,140}.
The soft mask performed better than the hard mask because the best recognition rates from the center speaker based on the hard and soft masks are 93% and 97%, respectively.
Multiband Julius was used as the ASR. In the experiments, a triphone acoustic model and a grammar-based language model were used to recognize isolated words. The triphone is an HMM which has 3 states and 4 mixtures in each state, and trained on 216 clean phonetically balanced words distributed by ATR. The size of the vocabulary was 200 words.
Thus, use of appropriately designed and adjusted soft masks has improved word recognition rate of the speech recognition system for simultaneous recognition of multiple sources.
In the embodiments described above, soft masks are determined using reliability of separation R. Instead of reliabilities of separation R, S/N ratios of input speeches obtained in the sound source separating section can be used for setting values of the soft masks.
Takahashi, Toru, Nakadai, Kazuhiro, Okuno, Hiroshi
Patent | Priority | Assignee | Title |
10013981, | Jun 06 2015 | Apple Inc | Multi-microphone speech recognition systems and related techniques |
10304462, | Jun 06 2015 | Apple Inc. | Multi-microphone speech recognition systems and related techniques |
10614812, | Jun 06 2015 | Apple Inc. | Multi-microphone speech recognition systems and related techniques |
11373672, | Jun 14 2016 | The Trustees of Columbia University in the City of New York | Systems and methods for speech separation and neural decoding of attentional selection in multi-speaker environments |
11456003, | Apr 12 2018 | Nippon Telegraph and Telephone Corporation | Estimation device, learning device, estimation method, learning method, and recording medium |
11961533, | Jun 14 2016 | The Trustees of Columbia University in the City of New York | Systems and methods for speech separation and neural decoding of attentional selection in multi-speaker environments |
11996091, | May 24 2018 | TENCENT TECHNOLOGY (SHENZHEN) COMPANY LIMITED | Mixed speech recognition method and apparatus, and computer-readable storage medium |
8538751, | Oct 15 2010 | HONDA MOTOR CO , LTD | Speech recognition system and speech recognizing method |
8849657, | Dec 14 2010 | Samsung Electronics Co., Ltd. | Apparatus and method for isolating multi-channel sound source |
9865265, | Jun 06 2015 | Apple Inc. | Multi-microphone speech recognition systems and related techniques |
9881631, | Oct 21 2014 | Mitsubishi Electric Research Laboratories, Inc. | Method for enhancing audio signal using phase information |
Patent | Priority | Assignee | Title |
6707910, | Sep 04 1997 | RPX Corporation | Detection of the speech activity of a source |
6967455, | Mar 09 2001 | Japan Science and Technology Agency | Robot audiovisual system |
7130705, | Jan 08 2001 | LinkedIn Corporation | System and method for microphone gain adjust based on speaker orientation |
20040252845, | |||
20060056647, | |||
20060136205, | |||
20060206315, | |||
20070025564, | |||
20070133811, | |||
20080071540, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 19 2009 | Honda Motor Co., Ltd. | (assignment on the face of the patent) | / | |||
Oct 20 2009 | NAKADAI, KAZUHIRO | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 023622 | /0512 | |
Oct 23 2009 | OKUNO, HIROSHI | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 023622 | /0512 | |
Oct 26 2009 | TAKAHASHI, TORU | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 023622 | /0512 |
Date | Maintenance Fee Events |
Jul 29 2013 | ASPN: Payor Number Assigned. |
Aug 25 2016 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Aug 21 2020 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Oct 21 2024 | REM: Maintenance Fee Reminder Mailed. |
Date | Maintenance Schedule |
Mar 05 2016 | 4 years fee payment window open |
Sep 05 2016 | 6 months grace period start (w surcharge) |
Mar 05 2017 | patent expiry (for year 4) |
Mar 05 2019 | 2 years to revive unintentionally abandoned end. (for year 4) |
Mar 05 2020 | 8 years fee payment window open |
Sep 05 2020 | 6 months grace period start (w surcharge) |
Mar 05 2021 | patent expiry (for year 8) |
Mar 05 2023 | 2 years to revive unintentionally abandoned end. (for year 8) |
Mar 05 2024 | 12 years fee payment window open |
Sep 05 2024 | 6 months grace period start (w surcharge) |
Mar 05 2025 | patent expiry (for year 12) |
Mar 05 2027 | 2 years to revive unintentionally abandoned end. (for year 12) |