A speech enhancement method is disclosed. The method includes the steps of: receiving a plurality of frames of sound signals by a microphone array; calculating an inter-aural time difference for each frequency band of each frame of the sound signals corresponding to at least one two-microphone set of the microphone array; calculating a plurality of values of cumulative histograms according to the calculated inter-aural time differences, wherein each value of the cumulative histograms is associates with a sound signal intensity of a respective frame; determining a first inter-aural time difference threshold according to the calculated value of the cumulative histograms; and filtering the plurality of frames of sound signals according to the first inter-aural time difference threshold.
|
1. A speech enhancement method, comprising the following steps:
utilizing a two-microphone set of a microphone array to receive a plurality of frames of sound signals;
calculating an inter-aural time difference for each frequency band of each frame of the sound signals in accordance with the two-microphone set of the microphone array;
calculating a plurality of values of a cumulative histogram in accordance with the calculated inter-aural time differences, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent on the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences;
determining a first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of the variances;
and filtering a plurality of the frames of the sound signals in accordance with the first inter-aural time difference threshold.
23. A speech enhancement system, comprising:
a microphone module, having at least one two-microphone set of a microphone array;
an inter-aural time difference calculating module, calculating an inter-aural time difference for each frequency band of each frame of sound signals in accordance with the two-microphone set of the microphone array;
a cumulative histogram module, calculating a plurality of values of a cumulative histogram in accordance with an inter-aural time difference of each frame, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent on the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences of the frames in the cumulative histogram;
a first inter-aural time difference threshold calculating module, calculating the first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of the variances; and
a sound signal filtering module, filtering the sound signals in accordance with the first inter-aural time difference threshold.
8. A speech enhancement method, comprising the following steps:
utilizing a two-microphone set of a microphone array to receive a plurality of frames of sound signals;
calculating an inter-aural time difference for each frequency band of each frame of the sound signals in accordance with the two-microphone set of the microphone array;
calculating a plurality of values of a cumulative histogram and a histogram in accordance with the calculated inter-aural time differences, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent on the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences of the frames in the cumulative histogram;
determining a first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of the variances;
determining a second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold; and
filtering the frames of the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold;
wherein the second inter-aural time difference threshold is greater than the first inter-aural time difference threshold.
24. A speech enhancement system comprising:
a microphone module, having at least one two-microphone set of a microphone array;
an inter-aural time difference calculating module, calculating an inter-aural time difference for each frequency band of each frame of sound signals in accordance with the two-microphone set of the microphone array;
a cumulative histogram module, calculating a plurality of values of a cumulative histogram and a histogram in accordance with an inter-aural time difference for each frame, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent on the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences of the frames in the cumulative histogram;
a first inter-aural time difference threshold calculating module, determining the first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of the variances;
a second inter-aural time difference threshold calculating module, determining the second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold; and
a sound signal filtering module, filtering the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold.
25. A speech enhancement method, comprising the following steps:
utilizing a microphone array to receive a plurality of frames of sound signals, wherein the microphone array includes a plurality of microphones;
calculating an inter-aural time difference for each frequency band of each frame of the sound signals in accordance with at least one two-microphone set of the microphone array;
calculating a plurality of values of a cumulative histogram and a histogram in accordance with the calculated inter-aural time differences, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent in the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences of the frames in the cumulative histogram;
determining a first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of variances;
determining a second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold;
filtering the frames of the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold and obtaining at least one speech enhancement signal, wherein the second inter-aural time difference threshold is greater than the first inter-aural time difference threshold; and
weighting at least one of the speech enhancement signals to obtain a weighted speech enhancement signal.
26. A speech enhancement system, comprising:
a microphone module, having a plurality of microphones;
an inter-aural time difference calculating module, calculating an inter-aural time difference for each frequency band of each frame of sound signals in accordance with at least one two-microphone set of a plurality of microphones;
a cumulative histogram module, calculating a plurality of values of a cumulative histogram and a histogram in accordance with an inter-aural time difference for each frame, wherein each value of the cumulative histogram is associated with a sound signal intensity of a respective frame dependent on the inter-aural time difference of that frame, wherein variances in the cumulative histogram are calculated in accordance with different inter-aural time differences of the frames in the cumulative histogram;
a first inter-aural time difference threshold calculating module, determining the first inter-aural time difference threshold in accordance with the values of the cumulative histogram, wherein the first inter-aural time difference threshold is determined in accordance with a maximum of the variances;
a second inter-aural time difference threshold calculating module, determining the second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold;
a sound signal filtering module, filtering the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold to generate at least one speech enhancement signal; and
a weighting module, predetermining at least one weighting value and weighting at least one speech enhancement signal to obtain a weighted speech enhancement signal.
2. The speech enhancement method of
searching for a plurality of frequency bands whose inter-aural time differences are greater than the first inter-aural time difference threshold; and
removing the frequency bands from each frame of the sound signals.
3. The speech enhancement method of
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; and η is a minimum variable.
5. The speech enhancement method of
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; and β is a variable to control the filtering degree.
6. The speech enhancement method of
calculating a plurality of variances of each inter-aural time difference in accordance with the values of a cumulative histogram;
and determining the inter-aural time difference having a maximum variance to be the first inter-aural time difference threshold.
7. The speech enhancement method of
9. The speech enhancement method of
searching for a plurality of frequency bands whose inter-aural time differences are greater than the second inter-aural time difference threshold;
removing the frequency bands whose inter-aural time difference is greater than the second inter-aural time difference threshold;
searching for a plurality of frequency bands whose inter-aural time differences are between the second inter-aural time difference threshold and the first inter-aural time difference threshold; and
attenuating the frequency bands whose inter-aural time difference is between the second inter-aural time difference threshold and the first inter-aural time difference threshold.
10. The speech enhancement method of
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; α is a variable between 0 and 1 to control the filtering degree; and η is a minimum variable.
12. The speech enhancement method of
13. The speech enhancement method of
wherein SNR is the signal to noise ratio between the voice source and the noise source; and β is a variable to control the filtering degree.
15. The speech enhancement method of
calculating a signal to noise ratio of a voice source and a noise source in accordance with the values of the histogram; and
determining the second inter-aural time difference threshold in accordance with the signal to noise ratio of a voice source and a noise source, the inter-aural time difference of the noise source, and the first inter-aural time difference.
16. The speech enhancement method of
17. The speech enhancement method of
τ2=τ1+δ+R×SNR, wherein τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; R means that the inter-aural time difference of the noise source is reduced by subtracting the first inter-aural time difference threshold; SNR is the signal to noise ratio between the voice source and the noise source; and δ is a minimum angle variable.
19. The speech enhancement method of
wherein τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; R means that the inter-aural time difference of the noise source is reduced by subtracting the first inter-aural time difference threshold; SNR is the signal to noise ratio between the voice source and the noise source; β is a variable to control the filtering degree; and δ is a minimum angle variable.
21. The speech enhancement method of
calculating a plurality of variances of each inter-aural time difference in accordance with the values of a cumulative histogram; and
determining the inter-aural time difference having a maximum variance to be the first inter-aural time difference threshold.
22. The speech enhancement method of
|
The disclosure relates to a speech enhancement method and system thereof.
Speech enhancement technology can filter noise from received speech signals in order to enhance the speech signals. Speech enhancement technology can be applied to oral communication, voice user interface, voice input, and other applications. Currently, with rapid development of mobile devices, vehicle electronic devices, and robots, the requirements of oral communication, voice input, and human-machine voice user interface in the noisy environment are quickly increasing. Thus, the issues of how to filter noise, enhance speech signal, and increase the quality of oral communication and human-machine voice user interface has become more and more important.
Generally, the speech signals received from microphones include signals from voice sources and noise sources. Since noise sources decrease the quality of oral communication and human-machine voice user interface, it is essential to reduce noise in order to increase signal quality. Although traditional speech enhancement technology with a single microphone utilizes filters, adaptive filters, and statistical models to enhance signal quality, the efficiency of such technology is limited. In addition, although the speech enhancement system with multiple microphones has better efficiency than the speech enhancement system with a single microphone, the speech enhancement system with multiple microphones requires too much computation load to apply for mobile devices with limited computation capability.
The present disclosure provides a speech enhancement method that includes the steps of: utilizing a two-microphone set of a microphone array to receive a plurality of frames of sound signals; calculating an inter-aural time difference for each frequency band of each frame of the sound signals in accordance with the two-microphone set of the microphone array; calculating a plurality of values of a cumulative histogram in accordance with the calculated inter-aural time differences; determining a first inter-aural time difference threshold in accordance with the values of the cumulative histogram; and filtering a plurality of the frames of the sound signals in accordance with the first inter-aural time difference threshold.
The present disclosure provides a speech enhancement system comprising a microphone module, an inter-aural time difference calculating module, a cumulative histogram module, a first inter-aural time difference threshold calculating module, and a sound signal filtering module. The microphone module has at least one two-microphone set of a microphone array. The inter-aural time difference calculating module calculates an inter-aural time difference for each frequency band of each frame of sound signals in accordance with the two-microphone set of the microphone array. The cumulative histogram module calculates a plurality of values of a cumulative histogram in accordance with an inter-aural time difference for each frame. The first inter-aural time difference threshold calculating module calculates the first inter-aural time difference threshold in accordance with the values of the cumulative histogram. The sound signal filtering module filters the sound signals in accordance with the first inter-aural time difference threshold.
The present disclosure also provides a speech enhancement method comprising the following steps: utilizing a two-microphone set of a microphone array to receive a plurality of frames of sound signals; calculating an inter-aural time difference for each frequency band of each frame of the sound signals in accordance with the two-microphone set of the microphone array; calculating a plurality of values of a cumulative histogram and a histogram in accordance with the calculated inter-aural time differences; determining a first inter-aural time difference threshold in accordance with the values of the cumulative histogram; determining a second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold; and filtering the frames of the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold, wherein the second inter-aural time difference threshold is greater than the first inter-aural time difference threshold.
The present disclosure also provides a speech enhancement system comprising a microphone module, an inter-aural time difference calculating module, a cumulative histogram module, a first inter-aural time difference threshold calculating module, a second inter-aural time difference threshold calculating module, and an sound signal filtering module. The microphone module has at least one two-microphone set of a microphone array. The inter-aural time difference calculating module calculates an inter-aural time difference for each frequency band of each frame of sound signals in accordance with the two-microphone set of the microphone array. The cumulative histogram module calculates a plurality of values of a cumulative histogram in accordance with an inter-aural time difference for each frame. The first inter-aural time difference threshold calculating module calculates the first inter-aural time difference threshold in accordance with the values of the cumulative histogram. The second inter-aural time difference threshold calculating module calculates the second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold. The sound signal filtering module filters the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold.
The foregoing has outlined rather broadly the features and technical benefits of the disclosure in order that the detailed description of the invention that follows may be better understood. Additional features and benefits of the invention will be described hereinafter, and form the subject of the claims of the invention. It should be appreciated by those skilled in the art that the conception and specific embodiment disclosed may be readily utilized as a basis for modifying or designing other structures or processes for carrying out the same purposes of the disclosure. It should also be realized by those skilled in the art that such equivalent constructions do not depart from the spirit and scope of the invention as set forth in the appended claims.
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the disclosure and, together with the description, serve to explain the principles of the invention.
In the following description, numerous specific details are set forth. However, it should be understood that embodiments of the disclosure may be practiced without these specific details. In other instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description. References to “the embodiment,” “an embodiment,” “another embodiment,” “other embodiment,” etc. indicate that the embodiment(s) of the disclosure so described may include a particular feature, structure, or characteristic, but not every embodiment necessarily includes the particular feature, structure, or characteristic. Further, repeated use of the phrase “in the embodiment” does not necessarily refer to the same embodiment, although it may. Unless specifically stated otherwise, as apparent from the following discussions, it should be appreciated that, throughout the specification, discussions utilizing terms such as “searching,” “filtering,” “calculating,” “determining,” “implementing,” “removing,” “attenuating,” “generating,” or the like refer to the action and/or processes of a computer or computing system, or similar electronic computing device, state machine and the like that manipulate and/or transform data represented as physical, such as electronic, quantities, into other data similarly represented as physical quantities.
The present disclosure is directed to a speech enhancement method and a system thereof. In order to make the present disclosure completely comprehensible, detailed steps and structures are provided in the following description. Obviously, implementation of the present disclosure does not limit special details known by persons skilled in the art. In addition, known structures and steps are not described in details, so as not to limit the present disclosure unnecessarily. Preferred embodiments of the present disclosure will be described below in detail. However, in addition to the detailed description, the present disclosure may also be widely implemented in other embodiments. The scope of the present disclosure is not limited to the detailed description, and is defined by the claims.
In an embodiment of the present disclosure of a speech enhancement system shown in
Referring to
The speech enhancement system shown in
wherein ∠XR(k0,m0) and ∠XR(k0,m0) mean phase values of XR(k0;m0) and XL(k0;m0), respectively; 2πr is compensation item to control the phase of ∠XR(k0,m0) and ∠XR(k0,m0) to range between 0 and 2π; ωk
Step 203 calculates a plurality of values of a cumulative histogram in accordance with the calculated inter-aural time difference.
Step 204 determines a first inter-aural time difference threshold in accordance with the values of the cumulative histogram.
Step 205 filters a plurality of frames of the sound signal in accordance with the first inter-aural time difference threshold. The embodiment of the present disclosure searches for a plurality of frequency bands whose inter-aural time difference is greater than the first inter-aural time difference threshold and then removes the frequency bands from each frame of the sound signals.
In the embodiment of the present disclosure, Step 205 is implemented by the following formula:
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; and η is a minimum variable. In the embodiment of the present invention, η is 0.01. In the embodiment of the present invention, Step 205 can be implemented by the following formula:
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; and β is a variable to control the filtering degree. A greater value of β correlates to more sound signals being filtered.
As shown in the above-mentioned formulas, Step 205 will preserve the frequency bands whose inter-aural time difference are smaller than the first inter-aural time difference threshold, and Step 205 will filter the frequency bands whose inter-aural time difference is greater than the first inter-aural time difference threshold. In addition, the embodiment of the present disclosure utilizes the variance of the values of the cumulative histogram with different frames to determine the first inter-aural time difference threshold. The variance calculating step further includes a step of calculating an updated variance in a recurrence calculation based on the previous variance. Therefore, the speech enhancement method of the present disclosure can preserve previous frames of sound signals into hardware to reduce computation load. In other words, the present disclosure can preserve a previous variance and receive a new sound signal to update the first inter-aural time difference threshold.
The speech enhancement method shown in
Referring
Comparing the speech enhancement methods of
Step 605 determines a second inter-aural time difference threshold in accordance with the values of the histogram and the first inter-aural time difference threshold.
In the embodiment of the present disclosure, the second inter-aural time difference threshold is calculated by the following formula:
τ2=τ1+δ+R×SNR,
wherein τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; R means that the inter-aural time difference of the noise source 160 is reduced by subtracting the first inter-aural time difference threshold; SNR is the signal to noise ratio between the voice source 150 and the noise source 160; and δ is a minimum angle variable. In the embodiment of the present disclosure, δ is 0.1. Referring to
In another embodiment of the present disclosure, the second inter-aural time difference threshold is calculated by the following formula:
wherein τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; R means that the inter-aural time difference of the noise source 160 is reduced by subtracting the first inter-aural time difference threshold; SNR is the signal to noise ratio between the voice source 150 and the noise source 160; β is a variable to control the filtering degree; and δ is a minimum angle variable. In the embodiment of the present disclosure, δ is 0.1. If SNR of the voice source 150 and the noise source 160 is greater than 0.5, the minor region will be enlarged. In contrast, if SNR of the voice source 150 and the noise source 160 is less than 0.5, the minor region will be reduced.
Step 606 filters the frames of the sound signals in accordance with the first inter-aural time difference threshold and the second inter-aural time difference threshold. In the embodiment of present disclosure, the sound signals filtering step further includes the steps of: searching for a plurality of frequency bands whose inter-aural time differences are greater than the second inter-aural time difference threshold; removing the frequency bands whose inter-aural time difference is greater than the second inter-aural time difference threshold; searching for a plurality of frequency bands whose inter-aural time differences are between the second inter-aural time difference threshold and the first inter-aural time difference threshold; and attenuating the frequency bands whose inter-aural time difference is between the second inter-aural time difference threshold and the first inter-aural time difference threshold. In other words, after the frequency bands having inter-aural time differences greater than the second inter-aural time difference threshold are removed from the sound signals, the sound signals attenuating the frequency bands having inter-aural time differences between the second inter-aural time difference threshold and the first inter-aural time difference threshold are defined as speech enhancement signal. In the embodiment of the present disclosure, Step 606 (including the step of removing frequency bands and the step of attenuating frequency bands) is implemented by the following formula:
wherein γ(k0,m0) is a weighting value of frequency band k0 in the frame m0 of the sound signals; d(k0,m0) is an inter-aural time difference of frequency band k0 in the frame m0 of the sound signals; τ1 is the first inter-aural time difference threshold; τ2 is the second inter-aural time difference threshold; α is a variable between 0 and 1 to control the filtering degree; and η is a minimum variable. In the embodiment of the present disclosure, η is 0.01.
Based on the above-method steps, the present disclosure preserves the frequency bands of the main region, attenuates the frequency bands of the minor region, and removes the frequency bands of the filtering region to obtain the speech enhancement signal. In the embodiment of the present disclosure, α and the signal to noise ratio between the voice source and the noise source are in direct proportion. In addition, α is calculated by the following formula:
wherein SNR is the signal to noise ratio between the voice source 150 and the noise source 160 and can be determined by Smax/Nmax; and β is a variable to control the filtering degree. A greater value of β corresponds to a higher filtering degree.
Referring to the speech enhancement system 100 of
As shown in
wherein W1 and W2 are weighting factors of the enhanced speech signal 1 and the enhanced speech signal 2, respectively. As shown in
In summary, the speech enhancement method of the present disclosure utilizes the values of the cumulative histogram of the inter-aural time difference to determine a main region and a filtering region and filters the received sound signals in accordance with different filtering degrees. In addition, the speech enhancement method of the present disclosure can utilize a simple microphone array and a smaller computation load to obtain the speech enhancement signals.
The above-described embodiments of the present disclosure are intended to be illustrative only. Numerous alternative embodiments may be to devised by persons skilled in the art without departing from the scope of the following claims. Those skilled in the art may devise numerous alternative embodiments without departing from the scope of the following claims.
Patent | Priority | Assignee | Title |
9706299, | Mar 13 2014 | GM Global Technology Operations LLC | Processing of audio received at a plurality of microphones within a vehicle |
Patent | Priority | Assignee | Title |
6002776, | Sep 18 1995 | Interval Research Corporation | Directional acoustic signal processor and method therefor |
6266633, | Dec 22 1998 | Harris Corporation | Noise suppression and channel equalization preprocessor for speech and speaker recognizers: method and apparatus |
6937980, | Oct 02 2001 | HIGHBRIDGE PRINCIPAL STRATEGIES, LLC, AS COLLATERAL AGENT | Speech recognition using microphone antenna array |
7103541, | Jun 27 2002 | Microsoft Technology Licensing, LLC | Microphone array signal enhancement using mixture models |
7197146, | May 02 2002 | Microsoft Technology Licensing, LLC | Microphone array signal enhancement |
7426464, | Jul 15 2004 | BITWAVE PTE LTD. | Signal processing apparatus and method for reducing noise and interference in speech communication and speech recognition |
7443989, | Jan 17 2003 | Samsung Electronics Co., Ltd. | Adaptive beamforming method and apparatus using feedback structure |
7533015, | Mar 01 2004 | International Business Machines Corporation | Signal enhancement via noise reduction for speech recognition |
7619563, | Aug 26 2005 | Dolby Laboratories Licensing Corporation | Beam former using phase difference enhancement |
7783060, | May 10 2005 | The United States of America as represented by the Administrator of the National Aeronautics and Space Administration; United States of America as represented by the Administrator of the National Aeronautics and Space Administration | Deconvolution methods and systems for the mapping of acoustic sources from phased microphone arrays |
7881480, | Mar 17 2004 | Cerence Operating Company | System for detecting and reducing noise via a microphone array |
20050143989, | |||
20090264961, | |||
20090304203, | |||
20110182437, | |||
20120148069, | |||
CN101192411, | |||
CN101779476, | |||
CN101903948, | |||
CN102142259, | |||
CN1670823, | |||
CN1831554, | |||
CN1967658, | |||
TW200921645, | |||
TW200926150, | |||
TW201030733, | |||
WO2010091077, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Mar 22 2012 | LIAO, HSIEN CHENG | Industrial Technology Research Institute | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 027967 | /0085 | |
Mar 30 2012 | Industrial Technology Research Institute | (assignment on the face of the patent) | / | |||
Apr 26 2016 | Micron Technology, Inc | U S BANK NATIONAL ASSOCIATION, AS COLLATERAL AGENT | SECURITY INTEREST SEE DOCUMENT FOR DETAILS | 038669 | /0001 | |
Apr 26 2016 | Micron Technology, Inc | U S BANK NATIONAL ASSOCIATION, AS COLLATERAL AGENT | CORRECTIVE ASSIGNMENT TO CORRECT THE REPLACE ERRONEOUSLY FILED PATENT #7358718 WITH THE CORRECT PATENT #7358178 PREVIOUSLY RECORDED ON REEL 038669 FRAME 0001 ASSIGNOR S HEREBY CONFIRMS THE SECURITY INTEREST | 043079 | /0001 |
Date | Maintenance Fee Events |
Nov 05 2018 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Nov 07 2022 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
May 05 2018 | 4 years fee payment window open |
Nov 05 2018 | 6 months grace period start (w surcharge) |
May 05 2019 | patent expiry (for year 4) |
May 05 2021 | 2 years to revive unintentionally abandoned end. (for year 4) |
May 05 2022 | 8 years fee payment window open |
Nov 05 2022 | 6 months grace period start (w surcharge) |
May 05 2023 | patent expiry (for year 8) |
May 05 2025 | 2 years to revive unintentionally abandoned end. (for year 8) |
May 05 2026 | 12 years fee payment window open |
Nov 05 2026 | 6 months grace period start (w surcharge) |
May 05 2027 | patent expiry (for year 12) |
May 05 2029 | 2 years to revive unintentionally abandoned end. (for year 12) |