Methods and systems are provided for enhanced stereo audio recordings in electronic devices. stereophonic recording performance in an electronic device, using a first microphone and a second microphone in the electronic device, may be assessed; and processing of signals generated by the first microphone and the second microphone may be configured based on the assessed stereophonic recording performance. The configuring may comprises adaptively modifying the processing to enhance stereophonic recording performance, to match or approximate an ideal performance. The assessing of the stereophonic recording in the electronic device may be based on a type of each of the first microphone and the second microphone, and/or based on a spacing therebetween. The processing may be adaptively modified to simulate directional reception of signals by the first microphone and the second microphone when the microphones are omnidirectional.
|
13. A method for enhanced stereophonic audio, comprising:
in an electronic device comprising a first microphone and a second microphone:
assessing stereophonic recording performance in the electronic device using the first microphone and the second microphone; and
configuring processing of signals generated by the first microphone and the second microphone, based on the assessed stereophonic recording performance, wherein the configuring comprises adaptively modifying the processing to enhance stereophonic recording performance, to match or approximate a stereophonic recording performance obtained by at least one out of (a) increasing a spacing between the first and second microphones to a desired distance between the first and second microphones, the desired distance exceeds a current distance between the first and second microphones, and (b) increasing a directional reception characteristic of at least one of the first and second microphones.
1. A system for enhanced stereophonic audio, comprising:
an electronic device comprising one or more circuits and a first microphone and a second microphone, the one or more circuits are for:
assessing stereophonic recording performance in the electronic device using the first microphone and the second microphone; and
configuring processing of signals generated by the first microphone and the second microphone, based on the assessed stereophonic recording performance, wherein the configuring comprises adaptively modifying the processing to enhance stereophonic recording performance, to match a stereophonic recording performance obtained by-at least one out of (a) increasing a spacing between the first and second microphones to a desired distance between the first and second microphones, the desired distance exceeds a current distance between the first and second microphones, and (b) increasing a directional reception characteristic of at least one of the first and second microphones.
2. The system of
3. The system of
4. The system of
5. The system of
6. The system of
7. The system of
8. The system of
9. The system of
10. The system of
11. The system of
12. The system of
14. The method of
15. The method of
16. The method of
17. The method of
18. The method of
19. The method of
20. The method of
21. The method of
22. The method of
23. The method according to
24. The method according to
|
This patent application makes reference to, claims priority to and claims benefit from the U.S. Provisional Patent Application Ser. No. 61/723,797, filed on Nov. 8, 2012, and having the title “Enhanced Stereo Audio Recordings in Handheld Devices.” The above stated application is hereby incorporated herein by reference in its entirety.
Aspects of the present application relate to audio processing. More specifically, certain implementations of the present disclosure relate to enhanced stereophonic audio recordings in handheld devices.
Existing methods and systems for managing audio input/output components (e.g., speakers and microphones) in electronic devices may be inefficient and/or costly. Further limitations and disadvantages of conventional and traditional approaches will become apparent to one of skill in the art, through comparison of such approaches with some aspects of the present method and apparatus set forth in the remainder of this disclosure with reference to the drawings.
A system and/or method is provided for enhanced stereophonic audio recordings in handheld devices, substantially as shown in and/or described in connection with at least one of the figures, as set forth more completely in the claims.
These and other advantages, aspects and novel features of the present disclosure, as well as details of illustrated implementation(s) thereof, will be more fully understood from the following description and drawings.
Certain implementations may be found in method and system for enhanced stereophonic audio recordings in electronic devices, particularly in handheld devices. As utilized herein the terms “circuits” and “circuitry” refer to physical electronic components (i.e. hardware) and any software and/or firmware (“code”) which may configure the hardware, be executed by the hardware, and or otherwise be associated with the hardware. As used herein, for example, a particular processor and memory may comprise a first “circuit” when executing a first plurality of lines of code and may comprise a second “circuit” when executing a second plurality of lines of code. As utilized herein, “and/or” means any one or more of the items in the list joined by “and/or”. As an example, “x and/or y” means any element of the three-element set {(x), (y), (x, y)}. As another example, “x, y, and/or z” means any element of the seven-element set {(x), (y), (z), (x, y), (x, z), (y, z), (x, y, z)}. As utilized herein, the terms “block” and “module” refer to functions than can be performed by one or more circuits. As utilized herein, the term “example” means serving as a non-limiting example, instance, or illustration. As utilized herein, the terms “for example” and “e.g.,” introduce a list of one or more non-limiting examples, instances, or illustrations. As utilized herein, circuitry is “operable” to perform a function whenever the circuitry comprises the necessary hardware and code (if any is necessary) to perform the function, regardless of whether performance of the function is disabled, or not enabled, by some user-configurable setting.
The electronic device 100 may comprise suitable circuitry for performing or supporting various functions, operations, applications, and/or services. The functions, operations, applications, and/or services performed or supported by the electronic device 100 may be run or controlled based on user instructions and/or pre-configured instructions.
In some instances, the electronic device 100 may support communication of data, such as via wired and/or wireless connections, in accordance with one or more supported wireless and/or wired protocols or standards.
In some instances, the electronic device 100 may be a handheld device—i.e. intended to be held by a user during use of the device, allowing for use of the device on the move and/or at different locations. In this regard, the electronic device 100 may be designed and/or configured to allow for ease of movement, such as to allow it to be readily moved while being held by the user as the user moves, and the electronic device 100 may be configured to perform at least some of the operations, functions, applications and/or services supported by the device on the move. Examples of electronic devices that are handheld devices comprise communication mobile devices (e.g., cellular phones, smartphones, and/or tablets), computers (e.g., laptops), media devices (e.g., portable media players and cameras), and the like. The electronic device 100 may even be a wearable device—i.e., may be worn by the device's user rather than being held in the user's hands. Examples of wearable electronic devices may comprise digital watches and watch-like devices (e.g., iWatch). The disclosure, however, is not limited to any particular type of electronic device.
The electronic device 100 may support input and/or output of audio. The electronic device 100 may incorporate, for example, a plurality of speakers and microphones, for use in outputting and/or inputting (capturing) audio, along with suitable circuitry for driving, controlling and/or utilizing the speakers and microphones. As shown in
In some instances, it may be desirable to arrange microphones so that they are close to one another. For example, in mobile communication devices, the microphones that are intended for use in audio recording may also be used in supporting such functions as, for example, noise reduction. The use of advanced noise reduction techniques in mobile communication devices may incorporate, for example, use of two microphones that may be used in picking up ambient noise. In some instances, the performance of noise reduction would generally be best when the two microphones are placed close to each other (e.g., in the range of 1-2 cm), such as to ensure that correlation between the noise that is picked up in both microphones is significantly higher, and thus the performance of the noise reduction with the two microphones may be significantly better. Arrangements of microphones in such manner (e.g., by having the microphones placed close to one another), whether to enhance other functions like noise reduction or because of space limitation, may be particularly done in certain types of electronic devices—e.g., mobile communication devices and other handheld electronic devices. Examples of such devices are shown in, for example,
Such arrangements of microphones, however, may degrade performance of stereophonic recording—e.g., due to poor differentiation between the two microphones as a result of them being placed too close to one another for stereophonic recording purposes. Accordingly, in various implementations in accordance with the present disclosure, stereophonic recording may be enhanced in devices having microphones that are not optimally place—e.g., being too close to one another, such as in the range of 1-2 cm. The enhancing of stereophonic recording may be achieved by use of, for example, adaptive processing that may allow for simulating results that would normally be achieved by use of microphones in optimal arrangements—e.g., spaced apart and/or have directional reception characteristics. This is described in more detail in connection with the following figures.
Each of the smartphone 200 and the handheld camera 250 may incorporate multiple microphones (e.g., two) to support stereophonic audio recordings. For example, smartphone 200 comprises a pair of microphones 210 and 220 (arranged as right and left microphones, respectively), and handheld camera 250 comprises a pair of microphones 260 and 270 (arranged as right and left microphones, respectively). Nonetheless, while the two microphones in each of the smartphone 200 and the handheld camera 250 are shown as being on the same side, the disclosure is not so limited. Rather, it should be understood that in instances the two microphones may be located on different sides of the devices—e.g., be located such that one microphone (e.g., microphone 210) may be on the front side of the smartphone 200 while the other microphone (e.g., microphone 220) may be located on the back of the smartphone 200, but with the two microphones still being close to one another (e.g., both at the bottom portion of the phone). The microphones (microphones 210 and 220 in the smartphone 210 and microphones 260 and 270 in the handheld camera 250) may be used in generating audio recordings that are intended to capture environmental sounds that may come from various sources (e.g., at distances between zero to several meters). The recordings may be done in conjunction with other operations in the devices (e.g., during video capture).
In some instances, however, relatively small dimensions of certain handheld devices, as well as design considerations, may limit the physical spacing between the microphones, necessitating placement of the microphones close to one another. Because of limited physical space and/or a desire to optimize particular functions (e.g., noise reduction) in such handheld devices as smartphones and portable handheld cameras, for example, the spacing between the microphones in the smartphone 200 and the camera 250 (e.g., separation 230 between microphones 210 and 220 in the smartphone 210, and separation 280 between microphones 260 and 270 in the handheld camera 250) may be relatively small. For example, in both of the smartphone 200 and the camera 250, the microphones incorporated therein may be identical omnidirectional microphones that are located on the front plan of the device, at a small horizontal distance from each other. For example, microphones 210 and 220 of the smartphone 200 may be placed in the bottom of the front plane, aligned on an horizontal line with a separation distance (230) of 1 cm between them; while microphones 260 and 270 of the camera 250 may be located in a diagonal direction such that they may have horizontal separation distance (280) of 1 cm between them in both Portrait and Landscape shooting modes. The small spacing between two microphones in each of the smartphone 200 and the camera 250 (as well as their type—that is being ‘omnidirectional’ microphones) may cause poor differentiation between the two microphones.
Accordingly, in various implementations, devices supporting stereophonic recording but having microphone arrangements that may degrade stereophonic recording performance may incorporate adaptive architecture and/or functions for enhancing stereophonic recording. The stereophonic recording enhancement may be achieved by, for example, use of adaptively modified digital processing that may be applied to signals coming from close microphone pairs, to produce two new output signals with enhanced stereophonic effects. Thus, the use of the adaptive modified digital processing in this manner may allow use of two microphones that may be positioned too close to one another (e.g., about 1-2 cm) to produce audio with stereophonic effect that may be comparable to the stereophonic effect of a recording with two microphones that are positioned optimally far apart for stereophonic recording (e.g., 15 cm). In one example implementation, audio signals arriving from different directions and captured by the close microphone pairs may have appropriate intensity that depends on the direction of arrival on each one of the two output signals. Thus, the individual directions may be clearly recognized by human ears during playback. Due to the small distance between the microphones, the amplitudes of the two original input signals do not significantly differ from each other. Accordingly, a small phase difference of the input signals may be converted, with the application of adaptive processing, into a significant amplitude difference between the two output signals. An example architecture (and adaptive processing applicable thereby) is described in more detail with respect to
The electronic device 300 may be similar to the electronic device 100 of
The processer 310 may comprise suitable circuitry configurable to process data, control or manage operations (e.g., of the electronic device 300 or components thereof), perform tasks and/or functions (or control any such tasks/functions). The processor 310 may run and/or execute applications, programs and/or code, which may be stored in, for example, memory (not shown). Further, the processor 310 may control operations of electronic device 300 (or components or subsystems thereof) using one or more control signals. The processor 310 may comprise a general purpose processor, which may be configured to perform or support particular types of operations (e.g., audio related operations). The processor 310 may also comprise a special purpose processor. For example, the processor 310 may comprise a digital signal processor (DSP), a baseband processor, and/or an application processor (e.g., an ASIC).
The audio codec 320 may comprise suitable circuitry configurable to perform voice coding/decoding operations. For example, the audio codec 320 may comprise one or more analog-to-digital converters (ADCs), one or more digital-to-analog converters (DACs), and one or more multiplexers (mux), which may be used in directing signals handled in the audio codec 320 to appropriate input and output ports thereof.
In operations, the electronic device 300 may support inputting and/or outputting of audio signals. For example, the microphone 3301 and 3302 may capture audio, generating corresponding analog audio input signals (e.g., analog signals 342 and 344), which may be forwarded to the audio codec 320. The audio codec 320 may convert the analog audio input (e.g., via the ADCs) to a digital audio signals (e.g., signals 352 and 354), which may be transferred to the processor 310 (e.g., over I2S connections). In some instances, however, the analog-to-digital conversions (and thus the audio codec 320 if that was the only reason it was utilized) may be bypassed with the signals being fed directly from the microphone 3301 and 3302 to the processor 310—e.g., if the microphone 3301 and 3302 were digital microphones. The processor 310 may then apply digital processing to the digital audio signals.
In some instances, the processor 310 may be configured to support stereophonic recordings. Accordingly, in some instances the processor 310 may generate, based on processing on audio input signals generated by the microphones 3301 and 3302, left-side signal 362 and right-side signal 364 (i.e., signals intended for each of a listener's left and right ears, respectively, which when received by the ears allow for generating stereophonic effect in the brain). The stereophonic recording performed in the electronic device 300 may, however, be degraded due to microphone arrangements utilized thereon. For example, the microphone 3301 and 3302 may be implemented as omnidirectional microphones (i.e., configured for receiving ambient audio from wide range rather than over narrow beams), and/or may be placed too close to one another (e.g., only 1-2 cm apart)—e.g., due to lack of space in the electronic device 300 and/or to enable optimal noise reduction processing.
Accordingly, in various implementations, the electronic device 300 may be configured for supporting enhanced audio recordings. The enhanced stereophonic recording may be used to overcome shortcomings or deficiencies in stereophonic recording that may be caused by less-than-optimal placement of the microphones (e.g., microphones 3301 and 3302) or characteristics thereof. The enhanced stereophonic recording may be achieved by using, for example, adaptive enhancement functions that are performed (e.g., in the processor 310) during processing of input audio signals (i.e., signals captured by the microphones). Thus, the architecture of the electronic device 300 may be particularly modified to enable or support these functions, and/or to allow performing them when needed. An example of adaptive processing that may be implemented in the electronic device (e.g., via the processor 310) is described in more detail with respect to
Similar architecture and/or functions as described with respect to the electronic device 300 may be utilized in devices having microphone arrangements posing similar shortcomings with respect to stereophonic recording and such requiring enhanced stereophonic recording—e.g., handheld devices with closely placed (and typically omnidirectional) microphones, such as the smartphone 200 and the camera 250.
The omnidirectional microphones 410 and 420 may correspond to microphones in a handheld device (e.g., microphones 210 and 220 of the smartphone 200). Because the omnidirectional microphones 410 and 420 may be spaced too close for optimal stereophonic recording, the differentiation between signals received by these microphones from a single audio source (e.g., source 400) may not result in satisfactory stereophonic effect when subjected to normal processing. Accordingly, the signals may be processed using a processor (e.g., the processor 310) which may be configured to incorporate processing modified to provide enhanced stereophonic recording.
For example, as shown in
G=G0/R (1)
D=R/V (2)
Where ‘R’ corresponds to the actual distance from the source (i.e., R corresponds to each of R_right and R_left for each of the right and left microphones 410 and 420), and V is the applicable propagation speed of sound.
Accordingly, the audio channels corresponding to signals captured by each of the right and left microphones 410 and 420 may be represented as:
S_left(t)=G_left*S(t−D_left) (3)
S_right(t)=G_right*S(t−D_right) (4)
The processor (e.g., the processor 310) may then apply the enhanced stereophonic recording processing. The processor 310 may use the small phase difference between the microphones 410 and 420 to produce a noticeable gain difference between the two output signals, which may depend on the direction of arrival of the sound. Thus, the individual directions can be clearly recognized by the human ears during playback. Various enhancement processing schemes may be utilized. For example, in the example implementation shown in
S_left(t)=G0*M_left(t)−G1*M_right(t−d) (5)
S_right(t)=G0*M_right(t)−G1*M_left(t−d) (6)
Where M_left(t) and M_right(t) are the signals that are simultaneously captured by the two microphones; and constants G0, G1, and d may relate to a virtual audio source that comes from the right side (i.e., when θ=−90°).
For example, the delay d in this case depends only on the space h between the two microphones, and may be pre-calculated and used as a constant. The values G0 and G1 are also constants, and are pre-calculated assuming a certain ‘desired’ distance h′ that is much bigger than h (e.g., 100 cm). In an example use scenario, d may be determined as h/V (where V is the speed of sound). Thus for h=1 cm (and assuming V is 343.2 m/s), d would be ≈29 us. G0 may be set to 1, whereas G1 may be set to h′/(h+h′). Thus, with h of 1 cm and h′ set to 100 cm, G1 would be ≈0.99. The processing done in the manner described above may result in a directional effect in each channel (as shown in
The described process can be carried-out either in the time domain or in the spectral domain. In the time domain, the delay value d is implemented by applying an interpolation process on the sampled signal. This enables delays of sub-samples (e.g., in a 8000 sample/sec sampling rate, h=1 cm requires a delay of ˜0.25 sample). In the frequency domain, each bin of frequency ω within a time-frame is multiplied by Exp−(ω*T) to introduce a time-delay T.
One advantage of the described process is that the output stereophonic channel pair is almost of a common delay. Zero delay stereophonic pairs can be easily transferred into mono audio channels by just summing together the Left and Right channels. This is not possible in stereophonic channel pairs that introduce significant delays between the two channels (e.g. when the space between microphones is greater than 10 cm), where a simple summation usually results in a decay of certain frequencies in the audio signal. Another advantage of the described process is that multiple audio sources do not require separate processes. That is to say, a single process takes care of all simultaneous audio sources within the recorded scene. For example, with a common process an audio source from the left side will result in enhanced gain in the left channel (and low gain in the right channel), while a simultaneous second audio source from the right side will result in enhanced gain in the right channel.
In starting step 502, an electronic device (e.g., the electronic device 300) may be powered on and initialized. This may comprise powering on, activating and/or initializing various components of the electronic device, such that the electronic device may be ready to perform or execute functions or application supported thereby.
In step 504, the microphone arrangement in the electronic device may be assessed—e.g., particularly with respect to stereophonic recording. In this regard, certain microphone arrangements (e.g., two omnidirectional microphones that are spaced too close to one another) may degrade performance of stereophonic recordings. Therefore, assessing the microphone arrangement may comprise determining (or estimating) performance of stereophonic recording done using the microphones. The estimated performance may be estimated in terms of anticipated quality of stereophonic effects of audio content produced based on signals captured via the microphones.
The outcome of the assessment may be checked in step 506. In this regard, the checking may comprise comparing the assessed performance against one or more predefined thresholds, which may be related to (or calculated based on) quality of stereophonic effects in anticipated output audio. For example, quality of stereophonic effect may be expressed as a percentage (with 100% corresponding to ideal quality of stereophonic effect), with the thresholds being set as particular percentages (e.g., 50%, 75%, 90%, etc.). Thus, a minimal ‘acceptable’ quality may be set to, e.g., 90% to indicate that only recordings with stereophonic effect having quality of less than 90% would be considered degraded. In some implementations, however, the adaptive processing may be done at all time, being adjusted dynamically to always ensure achieving (or attempt to achieve) ideal performance. In instances where it may be determined that the microphone arrangement does not degrade stereophonic recording, the process may proceed to step 510. Alternatively, in instances where it may be determined that the microphone arrangement may degrade stereophonic recording, the process may proceed to step 508.
In step 508, signal processing may be adaptively configured (or modified), to enable enhancing stereophonic recording—e.g., to simulate performance corresponding to spaced microphones and/or directional reception. For example, the processing of input signals captured by the microphones may be adaptive modified similar to the processing described with respect to
In step 510, input signals captured (or generated) by the microphones may be processed. The resultant signals (corresponding to left and right channels) may provide desirable stereophonic effects, either based on the microphones suitable arrangement or as result of the adaptive processing performed when the microphone arrangement is less than optimal.
In some implementations, a method for enhancing stereophonic recording may be used in a system that may comprise an electronic device (e.g., electronic device 300), which may comprise one or more circuits (e.g., processor 310 and audio codec 320) and a first microphone and a second microphone (e.g., microphones 3301 and 3302). The method may comprise assessing stereophonic recording performance in the electronic device using the first microphone and the second microphone; and configuring processing of signals generated by the first microphone and the second microphone, based on the assessed stereophonic recording performance, wherein the configuring comprises adaptively modifying the processing to enhance stereophonic recording performance, to match or approximate an ideal performance. The method may further comprise generating, based on the processing of signals generated by the first microphone and the second microphone, a left channel signal and a right channel signal, for outputting to a listener's left and right ears, respectively. The method may comprise adaptively modifying the processing when the assessed stereophonic recording performance falls below a predetermined threshold. The method may comprise assessing the stereophonic recording in the electronic device based on a type of each of the first microphone and the second microphone, and/or based on a spacing between the first microphone and the second microphone. The electronic device may comprise a handheld device. The method may comprise adaptively modifying the processing based on a distance between the first microphone and the second microphone, a distance from a source of signals captured by the first microphone and the second microphone, an initial gain at a location of the source of signals, and/or audio propagation speed. The method may comprise generating, based on the adaptive modifying of the processing, noticeable gain difference between two output signals corresponding to signals captured by each of the first microphone and the second microphone. The method may comprise adaptively modifying the processing to simulate directional reception of signals by the first microphone and the second microphone when the microphones are omnidirectional. The simulating of directional reception may result in amplifying audio sources that are located in an appropriate channel side are amplified. The simulating of directional reception may result in fully decaying audio sources that are located in an opposite side of a channel.
In some implementations, stereophonic recording may be enhanced in a system that may comprise an electronic device (e.g., electronic device 300), which may comprise one or more circuits (e.g., processor 310 and audio codec 320) and a first microphone and a second microphone (e.g., microphones 3301 and 3302). The one or more circuits may be operable to assess stereophonic recording performance in the electronic device using the first microphone and the second microphone; and configure processing of signals generated by the first microphone and the second microphone, based on the assessed stereophonic recording performance, wherein the configuring comprises adaptively modifying the processing to enhance stereophonic recording performance, to match or approximate an ideal performance. The processing may comprise generating a left channel signal and a right channel signal, for outputting to a listener's left and right ears, respectively. The one or more circuits may be operable to adaptively modify the processing when the assessed stereophonic recording performance falls below a predetermined threshold. The one or more circuits may be operable to assess the stereophonic recording in the electronic device based on a type of each of the first microphone and the second microphone, and/or based on a spacing between the first microphone and the second microphone. The electronic device may comprise a handheld device (e.g., smartphone 200 or camera 250). The one or more circuits may be operable to adaptively modify the processing based on a distance between the first microphone and the second microphone, a distance from a source of signals captured by the first microphone and the second microphone, an initial gain at a location of the source of signals, and/or audio propagation speed. The one or more circuits may be operable to adaptively modify the processing to generate noticeable gain difference between two output signals corresponding to signals captured by each of the first microphone and the second microphone. The one or more circuits may be operable to adaptively modify the processing to simulate directional reception of signals by the first microphone and the second microphone when the microphones are omnidirectional. The simulating of directional reception may result in amplifying audio sources that are located in an appropriate channel side are amplified. The simulating of directional reception may result in fully decaying audio sources that are located in an opposite side of a channel.
Other implementations may provide a non-transitory computer readable medium and/or storage medium, and/or a non-transitory machine readable medium and/or storage medium, having stored thereon, a machine code and/or a computer program having at least one code section executable by a machine and/or a computer, thereby causing the machine and/or computer to perform the steps as described herein for enhanced stereophonic audio recordings in handheld devices.
Accordingly, the present method and/or system may be realized in hardware, software, or a combination of hardware and software. The present method and/or system may be realized in a centralized fashion in at least one computer system, or in a distributed fashion where different elements are spread across several interconnected computer systems. Any kind of computer system or other system adapted for carrying out the methods described herein is suited. A typical combination of hardware and software may be a general-purpose computer system with a computer program that, when being loaded and executed, controls the computer system such that it carries out the methods described herein. Another typical implementation may comprise an application specific integrated circuit or chip.
The present method and/or system may also be embedded in a computer program product, which comprises all the features enabling the implementation of the methods described herein, and which when loaded in a computer system is able to carry out these methods. Computer program in the present context means any expression, in any language, code or notation, of a set of instructions intended to cause a system having an information processing capability to perform a particular function either directly or after either or both of the following: a) conversion to another language, code or notation; b) reproduction in a different material form. Accordingly, some implementations may comprise a non-transitory machine-readable (e.g., computer readable) medium (e.g., FLASH drive, optical disk, magnetic storage disk, or the like) having stored thereon one or more lines of code executable by a machine, thereby causing the machine to perform processes as described herein.
While the present method and/or system has been described with reference to certain implementations, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the scope of the present method and/or system. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the present disclosure without departing from its scope. Therefore, it is intended that the present method and/or system not be limited to the particular implementations disclosed, but that the present method and/or system will include all implementations falling within the scope of the appended claims.
Heiman, Arie, Yehuday, Uri, Haiut, Moshe
Patent | Priority | Assignee | Title |
10873823, | May 09 2017 | SOCIONEXT INC. | Sound processing device and sound processing method |
9930467, | Oct 29 2015 | Xiaomi Inc. | Sound recording method and device |
Patent | Priority | Assignee | Title |
3710034, | |||
6675114, | Aug 15 2000 | YOSHIMASA ELECTRONIC INC | Method for evaluating sound and system for carrying out the same |
7020290, | Oct 07 1999 | Method and apparatus for picking up sound | |
7577262, | Nov 18 2002 | Panasonic Corporation | Microphone device and audio player |
7953233, | Mar 20 2007 | National Semiconductor Corporation | Synchronous detection and calibration system and method for differential acoustic sensors |
8649529, | Jun 20 2008 | Fraunhofer-Gesellschaft zur Foerderung der Angewandten Forschung E V | Apparatus, method and computer program for localizing a sound source |
20100100374, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Nov 07 2013 | DSP Group Ltd. | (assignment on the face of the patent) | / | |||
Nov 07 2013 | HEIMAN, ARIE | DSP Group | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 031919 | /0386 | |
Nov 07 2013 | HAIUT, MOSHE | DSP Group | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 031919 | /0386 | |
Nov 07 2013 | YEHUDAY, URI | DSP Group | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 031919 | /0386 | |
Feb 03 2016 | DSP Group | DSP Group Ltd | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 037858 | /0536 |
Date | Maintenance Fee Events |
Aug 14 2019 | M2551: Payment of Maintenance Fee, 4th Yr, Small Entity. |
Feb 01 2022 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Jul 21 2023 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
Feb 23 2019 | 4 years fee payment window open |
Aug 23 2019 | 6 months grace period start (w surcharge) |
Feb 23 2020 | patent expiry (for year 4) |
Feb 23 2022 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 23 2023 | 8 years fee payment window open |
Aug 23 2023 | 6 months grace period start (w surcharge) |
Feb 23 2024 | patent expiry (for year 8) |
Feb 23 2026 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 23 2027 | 12 years fee payment window open |
Aug 23 2027 | 6 months grace period start (w surcharge) |
Feb 23 2028 | patent expiry (for year 12) |
Feb 23 2030 | 2 years to revive unintentionally abandoned end. (for year 12) |