An architecture and framework for speech/noise classification of an audio signal using multiple features with multiple input channels (e.g., microphones) are provided. The architecture may be implemented with noise suppression in a multi-channel environment where noise suppression is based on an estimation of the noise spectrum. The noise spectrum is estimated using a model that classifies each time/frame and frequency component of a signal as speech or noise by applying a speech/noise probability function. The speech/noise probability function estimates a speech/noise probability for each frequency and time bin. A speech/noise classification estimate is obtained by fusing (e.g., combining) data across different input channels using a layered network model. Individual feature data acquired at each channel and/or from a beam-formed signal is mapped to a speech probability, which is combined through layers of the model into a final speech/noise classification for use in noise estimation and filtering processes for noise suppression.
|
1. A method for noise estimation and filtering based on classifying an audio signal received at a noise suppression module via a plurality of input channels as speech or noise, the method comprising:
measuring signal classification features for a frame of the audio signal input from each of the plurality of input channels;
generating a feature-based speech probability for each of the measured signal classification features of each of the plurality of input channels;
generating a combined speech probability for the measured signal classification features over the plurality of input channels using a probabilistic layered network model, wherein an additive model is used for a top layer of the probabilistic layered network model;
classifying the audio signal as speech or noise based on the combined speech probability; and
updating an initial noise estimate for each of the plurality of input channels using the combined speech probability.
2. The method of
3. The method of
4. The method of
5. The method of
6. The method of
7. The method of
8. The method of
9. The method of
10. The method of
11. The method of
12. The method of
13. The method of
combining the frames of the audio signal input from the plurality of input channels;
measuring at least one signal classification feature of the combined frames of the audio signal;
calculating a feature-based speech probability for the combined frames using the measured at least one signal classification feature; and
combining the feature-based speech probability for the combined frames with the speech probabilities generated for each of the plurality of input channels.
14. The method of
15. The method of
16. The method of
17. The method of
18. The method of
19. The method of
20. The method of
21. The method of
22. The method of
23. The method of
24. The method of
25. The method of
26. The method of
27. The method of
28. The method of
|
This application is a Continuation of copending application Ser. No. 13/193,297 filed on Jul. 28, 2011. The entire contents of all of the above applications is hereby incorporated by reference.
The present disclosure generally relates to systems and methods for transmission of audio signals such as voice communications. More specifically, aspects of the present disclosure relate to estimating and filtering noise using speech probability modeling.
In audio communications (e.g., voice communications), excessive amounts of surrounding and/or background noise can interfere with intended exchanges of information and data between participants. Surrounding and/or background noise includes noise introduced from a number of sources, some of the more common of which include computers, fans, microphones, and office equipment. Accordingly, noise suppression techniques are sometimes implemented to reduce or remove such noise from audio signals during communication sessions.
When multiple input channels (e.g., microphones) are involved in audio communications, noise suppression processing becomes more complex. Conventional approaches to multi-channel noise suppression focus on a beam-forming component (e.g., a combined signal), which is a time-delayed sum of the two (or more) input channel/microphone signals. These conventional approaches use this combined input signal as the basis for noise estimation and speech enhancement processes that form part of the overall noise suppression. A problem with these conventional approaches is that the beam-forming may not be effective. For example, if a user moves around, or the room filter (and hence time-delays) are difficult to estimate, then relying on the beam-formed signal only is not effective in reducing noise.
This Summary introduces a selection of concepts in a simplified form in order to provide a basic understanding of some aspects of the present disclosure. This Summary is not an extensive overview of the disclosure, and is not intended to identify key or critical elements of the disclosure or to delineate the scope of the disclosure. This Summary merely presents some of the concepts of the disclosure as a prelude to the Detailed Description provided below.
One embodiment of the present disclosure relates to a method for noise estimation and filtering based on classifying an audio signal received at a noise suppression module via a plurality of input channels as speech or noise, the method comprising: measuring signal classification features for a frame of the audio signal input from each of the plurality of input channels; generating a feature-based speech probability for each of the measured signal classification features of each of the plurality of input channels; generating a combined speech probability for the measured signal classification features over the plurality of input channels; classifying the audio signal as speech or noise based on the combined speech probability; and updating an initial noise estimate for each of the plurality of input channels using the combined speech probability.
In another embodiment of the disclosure, the step of generating the combined speech probability in the method for noise estimation and filtering is performed using a probabilistic layered network model.
In another embodiment of the disclosure, the method for noise estimation and filtering further comprises determining a speech probability for an intermediate state of a layer of the probabilistic layered network model using data from a lower layer of the probabilistic layered network model.
In still another embodiment of the disclosure, the method for noise estimation and filtering further comprises applying an additive model or a multiplicative model to one of a set of state-conditioned transition probabilities to combine data from a lower layer of the probabilistic layered network model.
In another embodiment of the disclosure, the measured signal classification features from the plurality of input channels are input data to the probabilistic layered network model.
In another embodiment of the disclosure, the measured signal classification features from the plurality of input channels are input data to the probabilistic layered network model.
In yet another embodiment of the disclosure, the combined speech probability over the plurality of input channels is an output of the probabilistic layered network model.
In another embodiment of the disclosure, the probabilistic layered network model includes a set of intermediate states each denoting a class state of speech or noise for one or more layers of the probabilistic layered network model.
In another embodiment of the disclosure, the probabilistic layered network model further includes a set of state-conditioned transition probabilities.
In still a another embodiment of the disclosure, the feature-based speech probability for each of the measured signal classification features denotes a probability of a class state of speech or noise for a layer of the one or more layers of probabilistic layered network model
In another embodiment of the disclosure, the speech probability for the intermediate state of the layer of the probabilistic layered network model is determined using one or both of an additive model and a multiplicative model.
In another embodiment of the disclosure, the method for noise estimation and filtering further comprises generating, for each of the plurality of input channels, a speech probability for the input channel using the feature-based speech probabilities of the input channel.
In another embodiment of the disclosure, the speech probability for the input channel denotes a probability of a class state of speech or noise for a layer of the one or more layers of the probabilistic layered network model.
In yet another embodiment of the disclosure, the combined speech probability is generated as a weighted sum of the speech probabilities for the plurality of input channels.
In another embodiment of the disclosure, the weighted sum of the speech probabilities includes one or more weighting terms, the one or more weighting terms being based on one or more conditions.
In one embodiment of the disclosure the probabilistic layered network model is a Bayesian network model, while in another embodiment of the disclosure the probabilistic layered network model includes three layers.
In yet another embodiment of the disclosure, the step of classifying the audio signal as speech or noise based on the combined speech probability includes applying a threshold to the combined speech probability.
In another embodiment of the disclosure, the method for noise estimation and filtering further comprises determining an initial noise estimate for each of the plurality of input channels.
In still another embodiment of the disclosure, the method for noise estimation and filtering further comprises: combining the frames of the audio signal input from the plurality of input channels; measuring at least one signal classification feature of the combined frames of the audio signal; calculating a feature-based speech probability for the combined frames using the measured at least one signal classification feature; and combining the feature-based speech probability for the combined frames with the speech probabilities generated for each of the plurality of input channels.
In one embodiment of the disclosure the combined frames of the audio signal is a time-aligned superposition of the frames of the audio signal received at each of the plurality of input channels, while in another embodiment of the disclosure the combined frames of the audio signal is a signal generated using beam-forming on signals from the plurality of input channels.
In another embodiment of the disclosure, the combined frames of the audio signal is used as an additional input channel to the plurality of input channels.
In one or more other embodiments of the disclosure, the feature-based speech probability is a function of the measured signal classification feature, and the speech probability for each of the plurality of input channels is a function of the feature-based speech probabilities for the input channel.
In another embodiment of the disclosure, the speech probability for each of the plurality of input channels is obtained by combining the feature-based speech probabilities of the input channel using one or both of an additive model and a multiplicative model for a state-conditioned transition probability.
In still another embodiment of the disclosure, the feature-based speech probability is generated for each of the signal classification features by mapping each of the signal classification features to a probability value using a map function.
In other embodiments of the disclosure, the method for noise estimation and filtering described herein may optionally include one or more of the following additional features: the map function is a model with a set of width and threshold parameters; the feature-based speech probability is updated with a time-recursive average; the signal classification features include at least: average likelihood ratio over time, spectral flatness measure, and spectral template difference measure; at any layer and for any intermediate state, an additive model is used to generate a speech probability for the intermediate state, conditioned on the lower layer state; at any layer and for any intermediate state, a multiplicative model is used to generate a speech probability for the intermediate state, conditioned on the lower layer state; for a single input channel an additive model is used for a middle layer of the probabilistic layered network model to generate a speech probability for the single input channel; for a single input channel a multiplicative model is used for a middle layer of the probabilistic layered network model to generate a speech probability for the single input channel; a speech probability for an intermediate state at any intermediate layer of the probabilistic layered network model conditioned on a state on the previous layer is fixed off-line or determined adaptively on-line; for a set of two input channels an additive model is used for a top layer of the probabilistic layered network model to generate a speech probability for the two input channels; a beam-formed signal is another input to the probabilistic layered network model and an additive model is used for a top layer to generate a speech probability for the two input channels and the beam-formed signal; for each of the two input channels an additive model or a multiplicative model is used for a middle layer of the probabilistic layered network model to generate a speech probability for the intermediate layer; for the beam-formed signal, a speech probability conditioned on signal classification features of the beam-formed signal is obtained by mapping the signal classification features to a probability value using a map function and a time-recursive update; and/or, a time-recursive average is used to update the speech probability of the beam-formed signal.
Further scope of applicability of the present invention will become apparent from the Detailed Description given below. However, it should be understood that the Detailed Description and specific examples, while indicating preferred embodiments of the invention, are given by way of illustration only, since various changes and modifications within the spirit and scope of the invention will become apparent to those skilled in the art from this Detailed Description.
These and other objects, features and characteristics of the present disclosure will become more apparent to those skilled in the art from a study of the following Detailed Description in conjunction with the appended claims and drawings, all of which form a part of this specification. In the drawings:
The headings provided herein are for convenience only and do not necessarily affect the scope or meaning of the claimed invention.
In the drawings, the same reference numerals and any acronyms identify elements or acts with the same or similar structure or functionality for ease of understanding and convenience. The drawings will be described in detail in the course of the following Detailed Description.
Various examples of the invention will now be described. The following description provides specific details for a thorough understanding and enabling description of these examples. One skilled in the relevant art will understand, however, that the invention may be practiced without many of these details. Likewise, one skilled in the relevant art will also understand that the invention can include many other obvious features not described in detail herein. Additionally, some well-known structures or functions may not be shown or described in detail below, so as to avoid unnecessarily obscuring the relevant description.
Noise suppression aims to remove or reduce surrounding background noise to enhance the clarity of the intended audio thereby enhancing the comfort of the listener. In at least some embodiments of the present disclosure, noise suppression occurs in the frequency domain and includes both noise estimation and noise filtering processes. In scenarios involving high non-stationary noise levels, relying only on local speech-to-noise ratios (SNRs) to drive noise suppression often incorrectly biases a likelihood/probability determination of speech and noise presence. As will be described in greater detail herein, a process is provided for updating and adapting a speech/noise probability measure, for each input frame and frequency of an audio signal, that incorporates multiple speech/noise classification features (e.g., “signal classification features” or “noise-estimation features” as also referred to herein) from multiple input channels (e.g., microphones or similar audio capture devices) for an overall speech/noise classification determination. The architecture and framework for multi-channel speech/noise classification described herein provides for a more accurate and robust estimation of speech/noise presence in the frame. In the following description, “speech/noise classification features,” “signal classification features,” and “noise-estimation features” are interchangeable and refer to features of an audio signal that may be used (e.g., measured) to classify the signal, for each frame and frequency, into a state of either speech or noise.
Aspects and embodiments of the present disclosure relate to systems and methods for speech/noise classification using multiple features with multiple input channels (e.g., microphones). At least some embodiments described herein provide an architecture that may be implemented with methods and systems for noise suppression in a multi-channel environment where noise suppression is based on an estimation of the noise spectrum. In such noise suppression methods and systems, the noise spectrum may be estimated based on a model that classifies each time/frame and frequency component of a received input signal as speech or noise by using a speech/noise probability (e.g., likelihood) function. The speech/noise probability function estimates a speech/noise probability for each frequency and time bin of the received input signal, which is a measure of whether the received frame, at a given frequency, is likely speech (e.g., an individual speaking) or noise (e.g., office machine operating in the background). A good estimate of this speech/noise classification is important for robust estimation and update of background noise in noise suppression algorithms. The speech/noise classification can be estimated using various features of the received frame, such as spectral shape, average likelihood ratio (LR) factor, spectral template, peaks frequencies, local SNR, etc., all of which are good indicators as to whether a frequency/time bin is likely speech or noise.
For robust classification, multiple audio signal features should be incorporated into the speech/noise probability determination. When multiple input channels are involved, the difficulty lies in figuring out how to fuse (e.g., combine) the multiple features from the multiple channels. As described above, conventional approaches for multi-channel noise suppression focus on a beam-forming component (e.g., signal), which is a time-delayed sum of the two (or more) input channel signals. Noise estimation and speech enhancement process are then based on this combined/beam-formed input signal. A problem with these conventional approaches is that the beam-forming may not be effective. For example, if a user moves around, or the room filter (and hence time-delays) are difficult to estimate, then reliance on the beam-formed signal is not effective at reducing noise that may be present. Furthermore, conventional approaches to multi-channel noise suppression do not incorporate multiple audio signal features to estimate the speech/noise classification as is done in the numerous embodiments described herein.
In the methods and systems described herein, the beam-formed signal is used as only one input for the speech/noise classification determination. The direct input signals from the channels (e.g., the microphones) are also used. As will be further described below, the present disclosure provides a framework and architecture for combining information (e.g., feature measurements and speech/noise probability determinations) from all the channels involved, including the beam-formed signal.
In some embodiments, the noise suppression module 160 may be one component in a larger system for audio (e.g., voice) communications or audio processing. Although referred to herein as a “module,” noise suppression module 160 may also be referred to as a “noise suppressor” or, in the context of a larger system, a “noise suppression component.” The noise suppression module 160 may be an independent component in such a larger system or may be a subcomponent within an independent component (not shown) of the system. In the example embodiment illustrated in
Each of the capture devices 105A, 105B through 105N may be any of a variety of audio input devices, such as one or more microphones configured to capture sound and generate input signals. Render device 130 may be any of a variety of audio output devices, including a loudspeaker or group of loudspeakers configured to output sound of one or more channels. For example, capture devices 105A, 105B through 105N and render device 130 may be hardware devices internal to a computer system, or external peripheral devices connected to a computer system via wired and/or wireless connections. In some arrangements, capture devices 105A, 105B through 105N and render device 130 may be components of a single device, such as a speakerphone, telephone handset, etc. Additionally, capture devices 105A, 105B through 105N and/or render device 130 may include analog-to-digital and/or digital-to-analog transformation functionalities.
In at least the embodiment shown in
In some embodiments of the present disclosure, one or more other components, modules, units, etc., may be included as part of the noise suppression module 160, in addition to or instead of those illustrated in
The signal analysis unit 110 shown in
In various embodiments of the present disclosure, the methods, systems, and algorithms described herein for determining a speech/noise probability are implemented by the speech/noise classification unit 140. As shown in
Following the noise estimate update performed by the noise estimation update unit 135, an input frame is passed to the gain filter 145 for noise suppression. In one arrangement, the gain filter 145 may be a Wiener gain filter configured to reduce or remove the estimated amount of noise from the input frame. The gain filter may be applied on any one of the input (e.g., microphone) channels 105A, 105B, through 105N, on the beam-formed signal from beam-forming unit 120, or on any combination thereof.
The signal synthesis unit 155 may be configured to perform various post-noise suppression processes on the input frame following application of the gain filter 145. In at least one embodiment, upon receiving a noise-suppressed input frame from the gain filter 145, the signal synthesis unit 155 may use inverse DFT to convert the frame back to the time-domain, and then may perform energy scaling to help rebuild the frame in a manner that increases the power of speech present after suppression. For example, energy scaling may be performed on the basis that only input frames determined to be speech are amplified to a certain extent, while frames found to be noise are left alone. Because noise suppression may reduce the speech signal level, some amplification of speech segments via energy scaling by the signal synthesis unit 155 is beneficial. In one arrangement, the signal synthesis unit 155 is configured to perform scaling on a speech frame based on energy lost in the frame due to the noise estimation and filtering processes.
The example architecture shown in
The first (e.g., bottom) layer of the classification architecture, indicated as “Layer 1” in
The second (e.g., middle) layer of the classification architecture, indicated as “Layer 2” in
In the third (e.g., top) layer of the classification architecture, indicated as “Layer 3” in
According to embodiments described herein, the probability of a speech/noise state is obtained for each frequency k bin and time-frame t of an audio signal input from each of the channels 200A, 200B, through 200N. In one example arrangement, the received signal is processed in blocks (e.g., frames) of 10 milliseconds (ms), 20 ms, or the like. The discrete time index t may be used to index each of these blocks/frames. The audio signal in each of theses frames is then transformed into the frequency domain (e.g., using Discrete Fourier Transform (DFT) in the signal analysis unit 110 shown in
For purposes of notational simplicity, the following description of the layered network model shown in the example architecture of
A speech/noise probability function for a two-channel arrangement may be expressed as:
P(C|Y1(k,t),Y2(k,t),{Fi})=P(Y1(k,t),Y2(k,t)|C)P(C|{Fi})p({Fi})
where Yi(k,t) is the observed (noisy) frequency spectrum for the input channel (e.g., microphone) i, at time/frame index t, for frequency k, and C is the discrete classification state that denotes whether the time-frequency bin is speech (e.g., C=1) or noise (e.g., C=0). The quantities {Fi} are a set of features (e.g., “signal classification features,” which may include F1 through F6 shown in
The first term in the above expression, P(Y1(k,t), Y2(k,t)|C), can be determined based on, for example, a Gaussian assumption for the probability distribution of the observed spectrums {Yi(k,t)}, and an initial noise estimation. Other assumptions on the distribution of the spectrums {Yi(k,t)}, such as super-Gaussian, Laplacian, etc., may also be invoked. The initial noise estimation may be used to define one or more parameters of the probability distribution of the spectrums {Yi(k,t)}. An example method for computing the initial noise estimation is described in greater detail below. The second term in the expression, P(C|{Fi}), is the speech/noise probability, conditioned on the features derived from the channel inputs (e.g., the input signals from channels 200A and 200B shown in
In one or more embodiments described herein, an initial noise estimation may be derived based on a quantile noise estimation. In at least one example, the initial noise estimation may be computed by the noise estimation unit 115 shown in
Following the determination of speech/noise probability function P(C|Y1(k,t), Y2(k,t)), a noise estimation and update process is performed, as indicated by the noise estimation update unit 135 shown in
|N(k,t)|=γn|N(k,t−1)+(1−γn)A
A=P(C=1|Y1(k,t),Y2(k,t),{Fi})|N(k,t−1)|+P(C=0|Y1(k,t),Y2(k,t),{Fi})|Z(k,t)|
where |N(k,t)| is the estimate of the magnitude of the noise spectrum, for frame/time m and frequency bin k. The parameter γn controls the smoothing of the noise update, and the second term in the first expression above updates the noise with both the input spectrum and previous noise estimation, weighted according to the probability of speech/noise. The state C=1 denotes state of speech, and C=0 denotes state of noise. The quantity |Z(k,t)| is the magnitude of the input spectrum used for the noise update which, as described above for the gain filter, may be any one of the input (e.g., microphone) channel's magnitude spectrum (e.g., input channels 200A, 200B, through 200N shown in
The feature set {Fi} includes signal classification features for each channel input and, in at least some embodiments, an additional one or more signal classification features FBF derived from a combined/beam-formed signal 205 shown in
In other embodiments, numerous other features of the channel inputs may also be used in addition to or instead of these three example features. Furthermore, in various embodiments described herein the one or more features for the combined/beam-formed input, FBF, may include any of the same features as the channel inputs, or instead may include other feature quantities different from those of the channel inputs.
The P(C|{Fi}) term may be expressed as:
where the intermediate states {D1, D2, D3} denote the (internal) speech/noise state (e.g., D=1 for speech and D=0 for noise). The quantity P(Dj|{Fi}) is the probability of speech/noise given the set of features {Fi}. The quantity P(C|D1, D2, D3) is referred to as a state-conditioned transition probability in the following description below.
A model describing how the individual features from the channel inputs propagate to the Layer 3 (the top layer) speech/noise classifier may be expressed using another set of discrete states {Ei}, and corresponding state-conditioned transition probabilities (e.g., P(D1|E1, E2, E3) as follows:
The above expression corresponds to the three-layered network model shown in
The quantity P(C|D1, D2, D3), which is included in the above expression and illustrated in
In various embodiments of the present disclosure, the layered network model described herein may be implemented in one or more different user-scenarios or arrangements. For example, in a two-channel (e.g., two microphone) scenario, a first channel may be configured to sample (e.g., receive) noisy speech while a second channel is configured to sample only noise. In such an arrangement, P(C|D1, D2, D3) may only use information from the first channel input. In another example involving a two-channel scenario, both channels may be configured to sample speech and noise, in which case P(C|D1, D2, D3) may use information from both channel inputs, as well as information or data from a beam-formed input (e.g., beam-formed signal 205 shown in
Additionally, in the various scenarios and arrangements described above, a user may control how information or data from each channel is weighted when combined in the layered network model. For example, input from different channels (e.g., any of the channels 200A, 200B, up through 200N shown in
According to at least one embodiment, a structure for the fusion or combination term (e.g., the top layer of the network architecture, indicated as Layer 3 in
P(C|D1,D2,D3)=λ1δ(C−D1)+λ2δ(C−D2)+λ3δ(C−D3)
where δ(x) is defined as δ(x=0)=1, and otherwise δ(x)=0. As described above, λ1, λ2, and λ3 are weighting terms that may be controlled by a user, or based on a user's preferences or on the configuration/location of the input channels (e.g., microphones).
Single-Channel Scenario
In one example, the three signal classification features considered in the single-channel scenario of
In at least one embodiment described herein, the signal classification feature corresponding to the LR factor (e.g., F1) is the geometric average of a time-smoothened likelihood ratio (LR):
where N is the number of frequency bins used in the average, {tilde over (Δ)}(k,t) is the time-smoothened likelihood ratio, obtained as a recursive time-average from the LR factor, Δ(k,t),
log({tilde over (Δ)}(k,t))=γlrt log({tilde over (Δ)}(k,t−1))+(1−γlrt)log(Δ(k,t))
The LR factor is defined as the ratio of the probability of the input spectrum being in a state of speech over the probability of the input spectrum being in a state of noise, for a given frequency and time/frame index:
The two quantities in the second expression above denote the prior and post SNR, respectively, which may be defined as:
where |N(k,t)| is the estimated noise magnitude spectrum, |Y(k,t)| is the magnitude spectrum of the input (noisy) speech, and |X(k,t)| is the magnitude spectrum of the (unknown) clean speech. In one embodiment, the prior SNR may be estimated using a decision-directed update:
where H(k,t−1) is the gain filter (e.g., Wiener gain filter) for the previous processed frame, and |Y(k,t−1)| is the input magnitude spectrum of the noisy speech for the previous frame. In at least this example, the above expression may be taken as the decision-directed (DD) update of the prior SNR with a temporal smoothing parameter γdd.
In at least one embodiment, the spectral flatness feature is obtained as follows. For purposes of obtaining a spectral flatness measurement (F2), it is assumed that speech is likely to have more harmonic behavior than noise. Whereas the speech spectrum typically shows peaks at the fundamental frequency (pitch) and harmonics, the noise spectrum tends to be relatively flat in comparison. Accordingly, measures of local spectral flatness may collectively be used as a good indicator/classifier of speech and noise. In computing spectral flatness, N represents the number of frequency bins and B represents the number of bands. The index for a frequency bin is k and the index for a band is j. Each band will contain a number of bins. For example, the frequency spectrum of 128 bins can be divided into 4 bands (e.g., low band, low-middle band, high-middle band, and high band) each containing 32 bins. In another example, only one band containing all the frequencies is used. The spectral flatness may be computed as the ratio of the geometric mean to the arithmetic mean of the input magnitude spectrum:
where N represents the number of frequencies in the band. The computed quantity F2 will tend to be larger and constant for noise, and smaller and more variable for speech.
In at least one embodiment, the third signal classification feature (e.g., F3) may be determined as follows. In addition to the assumptions about noise described above for the spectral flatness measure (F2), another assumption that can be made about the noise spectrum is that it is more stationary than the speech spectrum. Therefore, it can be assumed that the overall shape of the noise spectrum will tend be the same during any given session. Proceeding under this assumption, a third signal classification feature, the spectral template difference measure (F3), can be said to be a measure of the deviation of the input spectrum from the shape of the noise spectrum.
In at least some embodiments, the spectral template difference measure (F3) may be determined by comparing the input spectrum with a template learned noise spectrum. For example, the template spectrum may be determined by updating the spectrum, which is initially set to zero, over segments that have strong likelihood of being noise or pause in speech. A result of the comparison is a conservative noise estimate, where the noise is only updated for segments where the speech probability is determined to be below a threshold. In other arrangements, the template spectrum may also be selected from a table of shapes corresponding to different noises. Given the input spectrum, Y(k,t), and the template spectrum, which may be denoted as α(k,t), the spectral template difference feature may be obtained by initially defining the spectral difference measure as:
where (ν,u) are shape parameters, such as linear shift and amplitude parameters, obtained by minimizing J. Parameters (ν,u) are obtained from a linear equation, and therefore are easily extracted for each frame. In some examples, the parameters account for any simple shift/scale changes of the input spectrum (e.g., if the volume increases). The feature is then the normalized measure,
where the normalization is the average input spectrum over all frequencies and over some time window of previous frames:
If the spectral template measure (F3) is small, then the input frame spectrum can be taken as being “close to” the template spectrum, and the frame is considered to be more likely noise. On the other hand, where the spectral template difference feature is large, the input frame spectrum is very different from the noise template spectrum, and the frame is considered to be speech. It is important to note that the spectral template difference measure (F3) is more general than the spectral flatness measure (F2). In the case of a template with a constant (e.g., near perfectly) flat spectrum, the spectral template difference feature reduces to a measure of the spectral flatness.
Referring again to the subset of the speech/noise classification model shown in
In any of the various embodiments described herein, one of two methods may be implemented for the flow of data and information for the middle and top layers, Layers 2 and 3, respectively, of the network architecture for the single-channel arrangement. These methods correspond to the following two models described below, where the first is an additive model and the second is a multiplicative model.
Method 1: Additive Middle Layer Model
For a single-channel arrangement, such as that illustrated in
P(C|D1)=δ(C−D1)
P(D1|E1,E2,E3)=τ1δ(D1−E1)+τ2δ(D1−E2)+τ3δ(D1−E3)
where {τi} are weight thresholds. The additive model refers to the structure used for the state-conditioned transition probability P(D1|E1, E2, E3) in the above equation.
The speech/noise probability conditioned on the features, P(C|{Fi}), then becomes the following, which is derived using the above two expressions:
P(C|F1,F2,F3)=τ1P(C|F1)+τ2P(C|F2)+τ3P(C|F3)
The individual terms P(C|Fi) in the above expression are computed and updated for each input (noisy) speech frame as
Pt(C|Fi)=γiPt−1(C|Fi)+(1−γi)M(zi,wi)
zi=Fi−Ti
where γi is the time-averaging factor defined for each feature, and parameters {wi} and {Ti} are thresholds that may be determined off-line or adaptively on-line. In at least one embodiment, the same time-averaging factor is used for all features, e.g., γi=γ.
Method 2: Multiplicative Middle Layer Model
In addition to the additive model for the middle layer described above, other embodiments involving a single-channel arrangement such as that illustrated in
P(C|D1)=δ(C−D1)
P(D1|E1,E2,E3)=P(D1|E1)P(D1|E2)P(D1|E3)
The multiplicative model refers to the structure used for the state-conditioned transition probability P(D1|E1, E2, E3) in the above equation.
The speech/noise probability conditioned on the features, P(C|{Fi}), then becomes the following, derived using the above two expressions:
The above expression is a product of three terms, each of which has two components: P(C|Ei) and P(Ei|Fi). For the P(Ei|Fi) components, the following model equations are used, which are the same as those described above for the additive model implementation:
Pt(Ei|Fi)=γiPt−1(Ei|Fi)+(1−γi)M(zi,wi)
zi=Fi−Ti
For the P(C|Ei) components, the following model equations are used:
P(C=0|Ei=0)=q
P(C=0|Ei=1)=1−q
P(C=1|Ei)=1−P(C=0|Ei)
The single parameter q may be used to characterize the quantity P(C|Ei), since the states {C, Ei} are binary (0 or 1). The parameter q as defined above determined the probability of the state C being in a noise state given that the state Ei in the previous layer is in a noise state. It may be determined off-line or may be determined adaptively on-line.
Multi-Channel Scenario
The following describes an implementation method for a multi-channel arrangement, such as that illustrated in
In at least one example involving a two-microphone channel scenario, three signal classification features may be considered for each of the two direct channel inputs (e.g., channels 200A and 200B shown in
The signal classification features F1, F2, F3, F4, F5, F6 may be measured for a frame of a (noisy) speech signal input from channels 1 and 2, along with signal classification feature FBF for the beam-formed signal, and may be used in Layer 1 to map the signal to a state of speech or noise for each input. In at least some embodiments described herein, the beam-formed signal (e.g., beam-formed signal 205 shown in
In the present example, the two-microphone channel implementation is based on three constraints, the first constraint being an additive weighted model for the top level (e.g., Layer 3) of the network architecture as follows:
P(C|D1,D2,D3)=λ1δ(C−D1)+λ2δ(C−D2)+λ3δ(C−D3)
where, as described above, δ(x) is defined as δ(x=0)=1, and otherwise δ(x)=0; and the weighting terms λ1, λ2, and λ3 (collectively {λi}) may be determined based on various user-scenarios and preferences. The second constraint is that each of the inputs from channels 1 and 2 use the same method/model as in the single-channel scenario described above. The third constraint is that the beam-formed signal uses a method/model derived from the time-recursive update according to the following equations presented in the single-channel scenario description and reproduced as follows:
Pt(C|Fi)=γiPt−1(C|Fi)+(1−γi)M(zi,wi)
zi=Fi−Ti
Given the first constraint/condition described above, the speech/noise probability is then derived from the sum of three terms, corresponding to each of the three inputs (e.g., the inputs from channel 1, channel 2, and the beam-formed signal). As such, the speech/noise probability for the two-microphone channel scenario may be expressed as follows:
P(C|F1,F2,F3,F4,F5,F6,FBF)=λ1P(C|F1,F2,F3)+λ2P(C|F4,F5,F6)+λ3P(C|FBF)
Using the second constraint/condition, where P(C|F1, F2, F3) and P(C|F4, F5, F6) are determined from either the additive middle layer model or the multiplicative middle layer model described above, depending on which method is used for the single-channel case, the speech/noise probability equations for the first two terms above are completely specified. The additive and multiplicative methods used for the second constraint/condition are reproduced (in that order) as follows:
P(C|F1,F2,F3)=τ1P(C|F1)+τ2P(C|F2)+τ3P(C|F3)
The same equations and set of parameters (adapted accordingly) would also be used for the P(C|F4, F5, F6) term (the second channel).
Finally, using the third constraint/condition for the two-microphone channel scenario, the third term P(C|FBF), based on the beam-formed input, is determined using the following:
Pt(C|FBF)=γBFPt−1(C|FBF)+(1−γBF)M(zBF,wBF)
zBF=FBF−TBF
where γBF is the time-averaging factor, wBF is a parameter for the sigmoid function, and TBF is a threshold. These parameter values are specific to the beam-forming input (e.g., there are generally different settings for the two direct input channels, which in some embodiments may be microphones or other audio capture devices).
The process begins at step 400 where signal classification features of an input frame are measured/extracted at each input channel (e.g., each of input channels 200A, 200B, through 200N shown in
In step 405, an initial noise estimate is computed for each of the input channels. As described above, in at least some embodiments an initial noise estimation may be derived (e.g., by the noise estimation unit 115 shown in
After the feature-based speech/noise probabilities are calculated in step 410, the process continues to step 415 where the feature-based speech/noise probabilities of each input channel are combined to generate a speech/noise probability (also sometimes referred to simply as “speech probability) for the channel. For example, referring again to the network architecture shown in
In step 420, an overall speech/noise probability for the input frame is calculated using the speech/noise probabilities of all the input channels (e.g., input channels 200A, 200B, through 200N, and also combined/beam-formed input 205 shown in
The overall speech/noise probability for the input frame calculated in step 420 is used in step 425 to classify the input frame as speech or noise. In at least some embodiments described herein, the speech/noise probability P(C|{Fi}) denotes the probabilistic classification state of the frame as either speech or noise, and depends on the best estimates combined from each of the input channels.
The final speech/noise probability function is therefore given as
P(C|Y1(k,t),Y2(k,t),{Fi})=P(Y1(k,t),Y2(k,t)|C)P(C|{Fi})p({Fi})
and is used in step 430 of the process to update the initial noise estimate, for each frame and frequency index of the received signal. In at least some embodiments of the disclosure, the noise estimate update is a soft-recursive update based on the following model, which is reproduced from above for convenience:
|N(k,t)|=γn|N(k,t−1)|+(1−γn)A
A=P(C=1|Y1(k,t),Y2(k,t),{Fi})|N(k,t−1)|+P(C=0|Y1(k,t),Y2(k,t),{Fi})|Z(k,t)|
Depending on the desired configuration, processor 510 can be of any type including but not limited to a microprocessor (μP), a microcontroller (μC), a digital signal processor (DSP), or any combination thereof. Processor 510 may include one or more levels of caching, such as a level one cache 511 and a level two cache 512, a processor core 513, and registers 514. The processor core 513 may include an arithmetic logic unit (ALU), a floating point unit (FPU), a digital signal processing core (DSP Core), or any combination thereof. A memory controller 515 can also be used with the processor 510, or in some embodiments the memory controller 515 can be an internal part of the processor 510.
Depending on the desired configuration, the system memory 520 can be of any type including but not limited to volatile memory (e.g., RAM), non-volatile memory (e.g., ROM, flash memory, etc.) or any combination thereof. System memory 520 typically includes an operating system 521, one or more applications 522, and program data 524. In at least some embodiments, application 522 includes a multipath processing algorithm 523 that is configured to pass a noisy input signal from multiple input channels (e.g., input channels 200A, 200B, through 200N shown in
Computing device 500 can have additional features and/or functionality, and additional interfaces to facilitate communications between the basic configuration 501 and any required devices and interfaces. For example, a bus/interface controller 540 can be used to facilitate communications between the basic configuration 501 and one or more data storage devices 550 via a storage interface bus 541. The data storage devices 550 can be removable storage devices 551, non-removable storage devices 552, or any combination thereof. Examples of removable storage and non-removable storage devices include magnetic disk devices such as flexible disk drives and hard-disk drives (HDD), optical disk drives such as compact disk (CD) drives or digital versatile disk (DVD) drives, solid state drives (SSD), tape drives and the like. Example computer storage media can include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, and/or other data.
System memory 520, removable storage 551 and non-removable storage 552 are all examples of computer storage media. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 500. Any such computer storage media can be part of computing device 500.
Computing device 500 can also include an interface bus 542 for facilitating communication from various interface devices (e.g., output interfaces, peripheral interfaces, communication interfaces, etc.) to the basic configuration 501 via the bus/interface controller 540. Example output devices 560 include a graphics processing unit 561 and an audio processing unit 562, either or both of which can be configured to communicate to various external devices such as a display or speakers via one or more A/V ports 563. Example peripheral interfaces 570 include a serial interface controller 571 or a parallel interface controller 572, which can be configured to communicate with external devices such as input devices (e.g., keyboard, mouse, pen, voice input device, touch input device, etc.) or other peripheral devices (e.g., printer, scanner, etc.) via one or more I/O ports 573. An example communication device 580 includes a network controller 581, which can be arranged to facilitate communications with one or more other computing devices 590 over a network communication (not shown) via one or more communication ports 582. The communication connection is one example of a communication media. Communication media may typically be embodied by computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. A “modulated data signal” can be a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media can include wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared (IR) and other wireless media. The term computer readable media as used herein can include both storage media and communication media.
Computing device 500 can be implemented as a portion of a small-form factor portable (or mobile) electronic device such as a cell phone, a personal data assistant (PDA), a personal media player device, a wireless web-watch device, a personal headset device, an application specific device, or a hybrid device that include any of the above functions. Computing device 500 can also be implemented as a personal computer including both laptop computer and non-laptop computer configurations.
There is little distinction left between hardware and software implementations of aspects of systems; the use of hardware or software is generally (but not always, in that in certain contexts the choice between hardware and software can become significant) a design choice representing cost versus efficiency tradeoffs. There are various vehicles by which processes and/or systems and/or other technologies described herein can be effected (e.g., hardware, software, and/or firmware), and the preferred vehicle will vary with the context in which the processes and/or systems and/or other technologies are deployed. For example, if an implementer determines that speed and accuracy are paramount, the implementer may opt for a mainly hardware and/or firmware vehicle; if flexibility is paramount, the implementer may opt for a mainly software implementation. In one or more other scenarios, the implementer may opt for some combination of hardware, software, and/or firmware.
The foregoing detailed description has set forth various embodiments of the devices and/or processes via the use of block diagrams, flowcharts, and/or examples. Insofar as such block diagrams, flowcharts, and/or examples contain one or more functions and/or operations, it will be understood by those within the art that each function and/or operation within such block diagrams, flowcharts, or examples can be implemented, individually and/or collectively, by a wide range of hardware, software, firmware, or virtually any combination thereof.
In one or more embodiments, several portions of the subject matter described herein may be implemented via Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs), digital signal processors (DSPs), or other integrated formats. However, those skilled in the art will recognize that some aspects of the embodiments described herein, in whole or in part, can be equivalently implemented in integrated circuits, as one or more computer programs running on one or more computers (e.g., as one or more programs running on one or more computer systems), as one or more programs running on one or more processors (e.g., as one or more programs running on one or more microprocessors), as firmware, or as virtually any combination thereof. Those skilled in the art will further recognize that designing the circuitry and/or writing the code for the software and/or firmware would be well within the skill of one of skilled in the art in light of the present disclosure.
Additionally, those skilled in the art will appreciate that the mechanisms of the subject matter described herein are capable of being distributed as a program product in a variety of forms, and that an illustrative embodiment of the subject matter described herein applies regardless of the particular type of signal-bearing medium used to actually carry out the distribution. Examples of a signal-bearing medium include, but are not limited to, the following: a recordable-type medium such as a floppy disk, a hard disk drive, a Compact Disc (CD), a Digital Video Disk (DVD), a digital tape, a computer memory, etc.; and a transmission-type medium such as a digital and/or an analog communication medium (e.g., a fiber optic cable, a waveguide, a wired communications link, a wireless communication link, etc.).
Those skilled in the art will also recognize that it is common within the art to describe devices and/or processes in the fashion set forth herein, and thereafter use engineering practices to integrate such described devices and/or processes into data processing systems. That is, at least a portion of the devices and/or processes described herein can be integrated into a data processing system via a reasonable amount of experimentation. Those having skill in the art will recognize that a typical data processing system generally includes one or more of a system unit housing, a video display device, a memory such as volatile and non-volatile memory, processors such as microprocessors and digital signal processors, computational entities such as operating systems, drivers, graphical user interfaces, and applications programs, one or more interaction devices, such as a touch pad or screen, and/or control systems including feedback loops and control motors (e.g., feedback for sensing position and/or velocity; control motors for moving and/or adjusting components and/or quantities). A typical data processing system may be implemented utilizing any suitable commercially available components, such as those typically found in data computing/communication and/or network computing/communication systems.
With respect to the use of substantially any plural and/or singular terms herein, those having skill in the art can translate from the plural to the singular and/or from the singular to the plural as is appropriate to the context and/or application. The various singular/plural permutations may be expressly set forth herein for sake of clarity.
While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
Patent | Priority | Assignee | Title |
10026415, | Mar 17 2014 | KONINKLIJKE PHILIPS N V | Noise suppression |
10032462, | Feb 26 2015 | Indian Institute of Technology Bombay | Method and system for suppressing noise in speech signals in hearing aids and speech communication devices |
10141003, | Jun 09 2014 | Dolby Laboratories Licensing Corporation | Noise level estimation |
10242677, | Aug 25 2015 | MALASPINA LABS BARBADOS , INC | Speaker dependent voiced sound pattern detection thresholds |
9449615, | Nov 07 2013 | Continental Automotive Systems, Inc | Externally estimated SNR based modifiers for internal MMSE calculators |
9520141, | Feb 28 2013 | GOOGLE LLC | Keyboard typing detection and suppression |
9659578, | Nov 27 2014 | Tata Consultancy Services Ltd. | Computer implemented system and method for identifying significant speech frames within speech signals |
9721580, | Mar 31 2014 | GOOGLE LLC | Situation dependent transient suppression |
Patent | Priority | Assignee | Title |
5185848, | Dec 14 1988 | Hitachi, Ltd. | Noise reduction system using neural network |
5251263, | May 22 1992 | Andrea Electronics Corporation | Adaptive noise cancellation and speech enhancement system and apparatus therefor |
5335312, | Sep 06 1991 | New Energy and Industrial Technology Development Organization | Noise suppressing apparatus and its adjusting apparatus |
5353376, | Mar 20 1992 | Texas Instruments Incorporated; TEXAS INSTRUMENTS INCORPORATED A CORP OF DELAWARE | System and method for improved speech acquisition for hands-free voice telecommunication in a noisy environment |
6363345, | Feb 18 1999 | Andrea Electronics Corporation | System, method and apparatus for cancelling noise |
6804651, | Mar 20 2001 | Swissqual AG | Method and device for determining a measure of quality of an audio signal |
6820053, | Oct 06 1999 | Analog Devices International Unlimited Company | Method and apparatus for suppressing audible noise in speech transmission |
6937980, | Oct 02 2001 | HIGHBRIDGE PRINCIPAL STRATEGIES, LLC, AS COLLATERAL AGENT | Speech recognition using microphone antenna array |
7031478, | May 26 2000 | KONINKLIJKE PHILIPS ELECTRONICS, N V | Method for noise suppression in an adaptive beamformer |
7565288, | Dec 22 2005 | Microsoft Technology Licensing, LLC | Spatial noise suppression for a microphone array |
7590530, | Sep 03 2005 | GN RESOUND A S | Method and apparatus for improved estimation of non-stationary noise for speech enhancement |
7620546, | Mar 23 2004 | BlackBerry Limited | Isolating speech signals utilizing neural networks |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Sep 26 2011 | Google Inc. | (assignment on the face of the patent) | / | |||
Sep 29 2017 | Google Inc | GOOGLE LLC | CHANGE OF NAME SEE DOCUMENT FOR DETAILS | 044695 | /0115 |
Date | Maintenance Fee Events |
Jan 25 2013 | ASPN: Payor Number Assigned. |
Feb 08 2016 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Mar 30 2020 | REM: Maintenance Fee Reminder Mailed. |
Sep 14 2020 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Aug 07 2015 | 4 years fee payment window open |
Feb 07 2016 | 6 months grace period start (w surcharge) |
Aug 07 2016 | patent expiry (for year 4) |
Aug 07 2018 | 2 years to revive unintentionally abandoned end. (for year 4) |
Aug 07 2019 | 8 years fee payment window open |
Feb 07 2020 | 6 months grace period start (w surcharge) |
Aug 07 2020 | patent expiry (for year 8) |
Aug 07 2022 | 2 years to revive unintentionally abandoned end. (for year 8) |
Aug 07 2023 | 12 years fee payment window open |
Feb 07 2024 | 6 months grace period start (w surcharge) |
Aug 07 2024 | patent expiry (for year 12) |
Aug 07 2026 | 2 years to revive unintentionally abandoned end. (for year 12) |