`
`\..
`
`PTO/SB/05 (08-08)
`Approved for use through 09/30/2010. 0MB 0651-0032
`U.S. Patent and Trademark Office. U.S. DEPARTMENT OF COMMERCE
`Under the Paoerwork Reduction Act of 1995 no oersons are re□ uired to resoond to a collection of information unless it disolavs a valid 0MB control number.
`
`UTILITY
`PATENT APPLICATION
`TRANSMITTAL
`
`Attorney Docket No.
`
`CreativeTech_01 NP _US
`
`First Inventor
`
`Title
`
`Manli Zhu
`
`Microphone Array System
`
`(Only for new nonprovisional applications under 37 CFR 1.53(b))
`
`Express Mail Label No.
`
`APPLICATION ELEMENTS
`See MPEP chapter 600 concerning utility patent application contents.
`
`ADDRESS TO:
`
`Commissioner for Patents
`P.O. Box 1450
`Alexandria VA 22313-1450
`
`3. 0
`
`1. □ Fee Transmittal Form (e.g., PTO/SB/17)
`Applicant claims small entity status.
`2. 0
`See 37 CFR 1.27.
`Specification
`[Total Pages
`43
`Both the claims and abstract must start on a new page
`(For information on the preferred arrangement, see MPEP 608.0t(a))
`Drawing(s) (35 U.S.C. 113)
`[Total Sheets
`33
`
`4. 0
`5. Oath or Declaration
`[Total Sheets
`4
`a. 0
`Newly executed (original or copy)
`b. D A copy from a prior application (37 CFR 1.63(d))
`_J!_or continuation/divisional with Box 18 completed)
`i. LJ DELETION OF INVENTOR(Sl
`Signed statement attached deleting inventor(s)
`name in the prior application, see 37 CFR
`1.63(d)(2) and 1.33(b).
`
`l
`
`I
`
`l
`
`ACCOMPANYING APPLICATION PARTS
`9. □ Assignment Papers (cover sheet & document(s))
`
`Name of Assignee
`
`10. □ 37 CFR 3.73(b) Statement
`(when there is an assignee)
`
`0Powerof
`Attorney
`
`11. □ English Translation Document (if applicable)
`12. 0 lnfoEj'tion Disclosure Statement (PTO/SB/08 or PTO-1449)
`Copies of citations attached
`
`6. □
`Application Data Sheet. See 37 CFR 1. 76
`7.0
`
`CD-ROM or CD-R in duplicate, large table or
`QQrr)puter Program (Appendix)
`LJ Landscape Table on CD
`
`8. Nucleotide and/or Amino Acid Sequence Submission
`(if applicable, items a. - c. are required)
`a. D Computer Readable Form (CRF)
`
`b.
`
`13. D Preliminary Amendment
`14. D Return Receipt Postcard (MPEP 503)
`(Should be specifically itemized)
`15. D Certified Copy of Priority Document(s)
`(if foreign priority is claimed)
`16. D Nonpublication Request under 35 U.S.C. 122(b)(2)(B)(i).
`
`Applicant must attach form PTO/SB/35 or equivalent.
`
`17. Dother:
`
`Specification Sequence Listing on:
`i. □
`CD-ROM or CD-R (2 copies); or
`ii. □
`Paper
`C. □ Statements verifying identity of above copies
`18. If a CONTINUING APPLICATION, check appropriate box, and supply the requisite information below and in the first sentence of the
`specification following the title, or in an Application Data Sheet under 37 CFR 1. 76:
`D Continuation
`D Divisional
`D Continuation-in-part (CIP)
`
`of prior application No.:
`
`Prior application information:
`
`Examiner ______________ _
`
`Art Unit: ______________ _
`
`19. CORRESPONDENCE ADDRESS
`
`I
`D The address associated with Customer Number: I
`~--------------~
`
`OR 0 Correspondence address below
`
`Name
`
`Address
`
`City
`
`Country
`
`Signature
`
`Name
`(Print/Tvoe l
`
`Ashok Tankha
`
`36 Greenleigh Drive
`
`Sewell
`
`USA
`
`/a tankha/
`
`Ashok Tankha
`
`I NJ
`I State
`I Telephone 1856-266-5145
`
`I Zip Code I 08080
`I Email
`I ash@ipprocurement.com
`
`I Date
`
`16 March 2011
`Registration No. I
`(Attornev/Aaentl 33802
`
`This collection of information is required by 37 CFR 1.53(b). The information is required to obtain or retain a benefit by the public which is to file (and by the
`USPTO to process) an application. Confidentiality is governed by 35 U.S.C. 122 and 37 CFR 1.11 and 1.14. This collection is estimated to take 12 minutes to
`complete, including gathering, preparing, and submitting the completed application form to the USPTO. Time will vary depending upon the individual case. Any
`comments on the amount of time you require to complete this form and/or suggestions for reducing this burden, should be sent to the Chief Information Officer,
`U.S. Patent and Trademark Office, U.S. Department of Commerce, P.O. Box 1450, Alexandria, VA 22313-1450. DO NOT SEND FEES OR COMPLETED
`FORMS TO THIS ADDRESS. SEND TO: Commissioner for Patents, P.O. Box 1450, Alexandria, VA 22313-1450.
`If you need assistance in completing the form, ca/11-800-PTO-9199 and select option 2.
`
`Page 1 of 200
`
`GOOGLE EXHIBIT 1004
`
`
`
`MICROPHONE ARRAY SYSTEM
`
`CROSS REFERENCE TO RELATED APPLICATIONS
`
`[0001] This application claims the benefit of provisional patent application number
`
`61/403,952 titled "Microphone array design and implementation for telecommunications
`
`and handheld devices", filed on September 24, 2010 in the United States Patent and
`
`Trademark Office.
`
`[0002] The specification of the above referenced patent application is incorporated
`
`herein by reference in its entirety.
`
`BACKGROUND
`
`[0003] Microphones constitute an important element in today's speech acquisition
`
`devices. Currently, most of the hands-free speech acquisition devices, for example,
`
`mobile devices, lapels, headsets, etc., convert sound into electrical signals by using a
`
`microphone embedded within the speech acquisition device. However, the paradigm of a
`
`single microphone often does not work effectively because the microphone picks up
`
`many ambient noise signals in addition to the desired sound, specifically when the
`
`distance between a user and the microphone is more than a few inches. Therefore, there is
`
`a need for a microphone system that operates under a variety of different ambient noise
`
`conditions and that places fewer constraints on the user with respect to the microphone,
`
`thereby eliminating the need to wear the microphone or be in close proximity to the
`
`microphone.
`
`[0004] To mitigate the drawbacks of the single microphone system, there is a need for a
`
`microphone array that achieves directional gain in a preferred spatial direction while
`
`suppressing ambient noise from other directions. Conventional microphone arrays
`
`include arrays that are typically developed for applications such as radar and sonar, but
`
`are generally not suitable for hands-free or handheld speech acquisition devices. The
`
`1
`
`Page 2 of 200
`
`
`
`main reason is that the desired sound signal has an extremely wide bandwidth relative to
`
`its center frequency, thereby rendering conventional narrowband techniques employed in
`
`the conventional microphone arrays unsuitable. In order to cater to such broadband
`
`speech applications, the array size needs to be vastly increased, making the conventional
`
`microphone arrays large and bulky, and precluding the conventional microphone arrays
`
`from having broader applications, for example, in mobile and handheld communication
`
`devices. There is a need for a microphone array system that provides an effective
`
`response over a wide spectrum of frequencies while being unobtrusive in terms of size.
`
`[0005] Hence, there is a long felt but unresolved need for a broadband microphone
`
`array and broadband beamforming system that enhances acoustics of a desired sound
`
`signal while suppressing ambient noise signals.
`
`SUMMARY OF THE INVENTION
`
`[0006] This summary is provided to introduce a selection of concepts in a simplified
`
`form that are further described in the detailed description of the invention. This summary
`
`is not intended to identify key or essential inventive concepts of the claimed subject
`
`matter, nor is it intended for determining the scope of the claimed subject matter.
`
`[0007] The method and system disclosed herein addresses the above stated need for
`
`enhancing acoustics of a target sound signal received from a target sound source, while
`
`suppressing ambient noise signals. As used herein, the term "target sound signal" refers
`
`to a sound signal from a desired or target sound source, for example, a person's speech
`
`that needs to be enhanced. A microphone array system comprising an array of sound
`
`sensors positioned in an arbitrary configuration, a sound source localization unit, an
`
`adaptive beamforming unit, and a noise reduction unit, is provided. The sound source
`
`localization unit, the adaptive beamforming unit, and the noise reduction unit are in
`
`operative communication with the array of sound sensors. The array of sound sensors is,
`
`for example, a linear array of sound sensors, a circular array of sound sensors, or an
`
`arbitrarily distributed coplanar array of sound sensors. The array of sound sensors herein
`
`2
`
`Page 3 of 200
`
`
`
`referred to as a "microphone array" receives sound signals from multiple disparate sound
`
`sources. The method disclosed herein can be applied on a microphone array with an
`
`arbitrary number of sound sensors having, for example, an arbitrary two dimensional (2D)
`
`configuration. The sound signals received by the sound sensors in the microphone array
`
`comprise the target sound signal from the target sound source among the disparate sound
`
`sources, and ambient noise signals.
`
`[0008] The sound source localization unit estimates a spatial location of the target
`
`sound signal from the received sound signals, for example, using a steered response
`
`power-phase transform. The adaptive beamforming unit performs adaptive beamforming
`
`for steering a directivity pattern of the microphone array in a direction of the spatial
`
`location of the target sound signal. The adaptive beamforming unit thereby enhances the
`
`target sound signal from the target sound source and partially suppresses the ambient
`
`noise signals. The noise reduction unit suppresses the ambient noise signals for further
`
`enhancing the target sound signal received from the target sound source.
`
`[0009]
`
`In an embodiment where the target sound source that emits the target sound
`
`signal is in a two dimensional plane, a delay between each of the sound sensors and an
`
`origin of the microphone array is determined as a function of distance between each of
`
`the sound sensors and the origin, a predefined angle between each of the sound sensors
`
`and a reference axis, and an azimuth angle between the reference axis and the target
`
`sound signal. In another embodiment where the target sound source that emits the target
`
`sound signal is in a three dimensional plane, the delay between each of the sound sensors
`
`and the origin of the microphone array is determined as a function of distance between
`
`each of the sound sensors and the origin, a predefined angle between each of the sound
`
`sensors and a first reference axis, an elevation angle between a second reference axis and
`
`the target sound signal, and an azimuth angle between the first reference axis and the
`
`target sound signal. This method of determining the delay enables beamforming for
`
`arbitrary numbers of sound sensors and multiple arbitrary microphone array
`
`configurations. The delay is determined, for example, in terms of number of samples.
`
`3
`
`Page 4 of 200
`
`
`
`Once the delay is determined, the microphone array can be aligned to enhance the target
`
`sound signal from a specific direction.
`
`[0010] The adaptive beamforming unit comprises a fixed beamformer, a blocking
`
`matrix, and an adaptive filter. The fixed beamformer steers the directivity pattern of the
`
`microphone array in the direction of the spatial location of the target sound signal from
`
`the target sound source for enhancing the target sound signal, when the target sound
`
`source is in motion. The blocking matrix feeds the ambient noise signals to the adaptive
`
`filter by blocking the target sound signal from the target sound source. The adaptive filter
`
`adaptively filters the ambient noise signals in response to detecting the presence or
`
`absence of the target sound signal in the sound signals received from the disparate sound
`
`sources. The fixed beamformer performs fixed beamforming, for example, by filtering
`
`and summing output sound signals from the sound sensors.
`
`[0011]
`
`In an embodiment, the adaptive filtering comprises sub-band adaptive filtering.
`
`The adaptive filter comprises an analysis filter bank, an adaptive filter matrix, and a
`
`synthesis filter bank. The analysis filter bank splits the enhanced target sound signal from
`
`the fixed beamformer and the ambient noise signals from the blocking matrix into
`
`multiple frequency sub-bands. The adaptive filter matrix adaptively filters the ambient
`
`noise signals in each of the frequency sub-bands in response to detecting the presence or
`
`absence of the target sound signal in the sound signals received from the disparate sound
`
`sources. The synthesis filter bank synthesizes a full-band sound signal using the
`
`frequency sub-bands of the enhanced target sound signal. In an embodiment, the adaptive
`
`beamforming unit further comprises an adaptation control unit for detecting the presence
`
`of the target sound signal and adjusting a step size for the adaptive filtering in response to
`
`detecting the presence or the absence of the target sound signal in the sound signals
`
`received from the disparate sound sources.
`
`[0012] The noise reduction unit suppresses the ambient noise signals for further
`
`enhancing the target sound signal from the target sound source. The noise reduction unit
`
`performs noise reduction, for example, by using a Wiener-filter based noise reduction
`
`4
`
`Page 5 of 200
`
`
`
`algorithm, a spectral subtraction noise reduction algorithm, an auditory transform based
`
`noise reduction algorithm, or a model based noise reduction algorithm. The noise
`
`reduction unit performs noise reduction in multiple frequency sub-bands employed for
`
`sub-band adaptive beamforming by the analysis filter bank of the adaptive beamforming
`
`unit.
`
`[0013] The microphone array system disclosed herein comprising the microphone array
`
`with an arbitrary number of sound sensors positioned in arbitrary configurations can be
`
`implemented in handheld devices, for example, the iPad® of Apple Inc., the iPhone® of
`
`Apple Inc., smart phones, tablet computers, laptop computers, etc. The microphone array
`
`system disclosed herein can further be implemented in conference phones, video
`
`conferencing applications, or any device or equipment that needs better speech inputs.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[0014] The foregoing summary, as well as the following detailed description of the
`
`invention, is better understood when read in conjunction with the appended drawings. For
`
`the purpose of illustrating the invention, exemplary constructions of the invention are
`
`shown in the drawings. However, the invention is not limited to the specific methods and
`
`instrumentalities disclosed herein.
`
`[0015] FIG. 1 illustrates a method for enhancing a target sound signal from multiple
`
`sound signals.
`
`[0016] FIG. 2 illustrates a system for enhancing a target sound signal from multiple
`
`sound signals.
`
`[0017] FIG. 3 exemplarily illustrates a microphone array configuration showing a
`
`microphone array having N sound sensors arbitrarily distributed on a circle.
`
`5
`
`Page 6 of 200
`
`
`
`[0018] FIG. 4 exemplarily illustrates a graphical representation of a filter-and-sum
`
`beamforming algorithm for determining output of the microphone array having N sound
`
`sensors.
`
`[0019] FIG. 5 exemplarily illustrates distances between an origin of the microphone
`
`array and sound sensor M 1 and sound sensor M3 in the circular microphone array
`
`configuration, when the target sound signal is at an angle 0 from the Y-axis.
`
`[0020] FIG. 6A exemplarily illustrates a table showing the distance between each sound
`
`sensor in a circular microphone array configuration from the origin of the microphone
`
`array, when the target sound source is in the same plane as that of the microphone array.
`
`[0021] FIG. 6B exemplarily illustrates a table showing the relationship of the position
`
`of each sound sensor in the circular microphone array configuration and its distance to
`
`the origin of the microphone array, when the target sound source is in the same plane as
`
`that of the microphone array.
`
`[0022] FIG. 7 A exemplarily illustrates a graphical representation of a microphone
`
`array, when the target sound source is in a three dimensional plane.
`
`[0023] FIG. 7B exemplarily illustrates a table showing delay between each sound
`
`sensor in a circular microphone array configuration and the origin of the microphone
`
`array, when the target sound source is in a three dimensional plane.
`
`[0024] FIG. 7C exemplarily illustrates a three dimensional working space of the
`microphone array, where the target sound signal is incident at an elevation angle 'P < Q.
`
`[0025] FIG. 8 exemplarily illustrates a method for estimating a spatial location of the
`
`target sound signal from the target sound source by a sound source localization unit using
`
`a steered response power-phase transform.
`
`6
`
`Page 7 of 200
`
`
`
`[0026] FIG. 9A exemplarily illustrates a graph showing the value of the steered
`
`response power-phase transform for every 10°.
`
`[0027] FIG. 9B exemplarily illustrates a graph representing the estimated target sound
`
`signal from the target sound source.
`
`[0028] FIG. 10 exemplarily illustrates a system for performing adaptive beamforming
`
`by an adaptive beamforming unit.
`
`[0029] FIG. 11 exemplarily illustrates a system for sub-band adaptive filtering.
`
`[0030] FIG. 12 exemplarily illustrates a graphical representation showing the
`
`performance of a perfect reconstruction filter bank.
`
`[0031] FIG. 13 exemplarily illustrates a block diagram of a noise reduction unit that
`
`performs noise reduction using a Wiener-filter based noise reduction algorithm.
`
`[0032] FIG. 14 exemplarily illustrates a hardware implementation of the microphone
`
`array system.
`
`[0033] FIGS. lSA-lSC exemplarily illustrate a conference phone comprising an eight(cid:173)
`
`sensor microphone array.
`
`[0034] FIG. 16A exemplarily illustrates a layout of an eight-sensor microphone array
`
`for a conference phone.
`
`[0035] FIG. 16B exemplarily illustrates a graphical representation of eight spatial
`
`regions to which the eight-sensor microphone array of FIG. 16A responds.
`
`7
`
`Page 8 of 200
`
`
`
`[0036] FIGS. 16C-16D exemplarily illustrate computer simulations showing the
`
`steering of the directivity patterns of the eight-sensor microphone array of FIG. 16A in
`
`the directions of 15° and 60° respectively, in the frequency range 300 Hz to 5 kHz.
`
`[0037] FIGS. 16E-16L exemplarily illustrate graphical representations showing the
`
`directivity patterns of the eight-sensor microphone array of FIG. 16A in each of the eight
`
`spatial regions, where each directivity pattern is an average response from 300Hz to
`
`5000Hz.
`
`[0038] FIG. 17 A exemplaril y illustrates a graphical representation of four spatial
`
`regions to which a four-sensor microphone array for a wireless handheld device responds.
`
`[0039] FIGS. 17B-171 exemplaril y illustrate computer simulations showing the
`
`directivity patterns of the four-sensor microphone array of FIG. 17 A with respect to
`
`azimuth and frequency.
`
`[0040] FIGS. lSA-18B exemplarily illustrate a microphone array configuration for a
`
`tablet computer.
`
`[0041] FIG. 18C exemplarily illustrates an acoustic beam formed using the microphone
`
`array configuration of FIGS. lSA-18B according to the method and system disclosed
`
`herein.
`
`[0042] FIGS. 18D-lSG exemplarily illustrate graphs showing processing results of the
`
`adaptive beamforming unit and the noise reduction unit for the microphone array
`
`configuration of FIG. 18B, in both a time domain and a spectral domain for the tablet
`
`computer.
`
`[0043] FIGS. 19A-19F exemplarily illustrate tables showing different microphone array
`
`configurations and the corresponding values of delay Tn for the sound sensors in each of
`
`the microphone array configurations.
`
`8
`
`Page 9 of 200
`
`
`
`DETAILED DESCRIPTION OF THE INVENTION
`
`[0044] FIG. 1 illustrates a method for enhancing a target sound signal from multiple
`
`sound signals. As used herein, the term "target sound signal" refers to a desired sound
`
`signal from a desired or target sound source, for example, a person's speech that needs to
`
`be enhanced. The method disclosed herein provides 101 a microphone array system
`
`comprising an array of sound sensors positioned in an arbitrary configuration, a sound
`
`source localization unit, an adaptive beamforming unit, and a noise reduction unit. The
`
`sound source localization unit, the adaptive beamforming unit, and the noise reduction
`
`unit are in operative communication with the array of sound sensors. The microphone
`
`array system disclosed herein employs the array of sound sensors positioned in an
`
`arbitrary configuration, the sound source localization unit, the adaptive beamforming
`
`unit, and the noise reduction unit for enhancing a target sound signal by acoustic beam
`
`forming in the direction of the target sound signal in the presence of ambient noise
`
`signals.
`
`[0045] The array of sound sensors herein referred to as a "microphone array" comprises
`
`multiple or an arbitrary number of sound sensors, for example, microphones, operating in
`
`tandem. The microphone array refers to an array of an arbitrary number of sound sensors
`
`positioned in an arbitrary configuration. The sound sensors are transducers that detect
`
`sound and convert the sound into electrical signals. The sound sensors are, for example,
`
`condenser microphones, piezoelectric microphones, etc.
`
`[0046] The sound sensors receive 102 sound signals from multiple disparate sound
`
`sources and directions. The target sound source that emits the target sound signal is one
`
`of the disparate sound sources. As used herein, the term "sound signals" refers to
`
`composite sound energy from multiple disparate sound sources in an environment of the
`
`microphone array. The sound signals comprise the target sound signal from the target
`
`sound source and the ambient noise signals. The sound sensors are positioned in an
`
`arbitrary planar configuration herein referred to as a "microphone array configuration",
`
`9
`
`Page 10 of 200
`
`
`
`for example, a linear configuration, a circular configuration, any arbitrarily distributed
`
`coplanar array configuration, etc. By employing beamforming according to the method
`
`disclosed herein, the microphone array provides a higher response to the target sound
`
`signal received from a particular direction than to the sound signals from other directions.
`
`A plot of the response of the microphone array versus frequency and direction of arrival
`
`of the sound signals is referred to as a directivity pattern of the microphone array.
`
`[0047] The sound source localization unit estimates 103 a spatial location of the target
`
`sound signal from the received sound signals. In an embodiment, the sound source
`
`localization unit estimates the spatial location of the target sound signal from the target
`
`sound source, for example, using a steered response power-phase transform as disclosed
`
`in the detailed description of FIG. 8.
`
`[0048] The adaptive beamforming unit performs adaptive beamforming 104 by steering
`
`the directivity pattern of the microphone array in a direction of the spatial location of the
`
`target sound signal, thereby enhancing the target sound signal, and partially suppressing
`
`the ambient noise signals. Beamforming refers to a signal processing technique used in
`
`the microphone array for directional signal reception, that is, spatial filtering. This spatial
`
`filtering is achieved by using adaptive or fixed methods. Spatial filtering refers to
`
`separating two signals with overlapping frequency content that originate from different
`
`spatial locations.
`
`[0049] The noise reduction unit performs noise reduction by further suppressing 105
`
`the ambient noise signals and thereby further enhancing the target sound signal. The
`
`noise reduction unit performs the noise reduction, for example, by using a Wiener-filter
`
`based noise reduction algorithm, a spectral subtraction noise reduction algorithm, an
`
`auditory transform based noise reduction algorithm, or a model based noise reduction
`
`algorithm.
`
`[0050] FIG. 2 illustrates a system 200 for enhancing a target sound signal from multiple
`
`sound signals. The system 200, herein referred to as a "microphone array system",
`
`10
`
`Page 11 of 200
`
`
`
`comprises the array 201 of sound sensors positioned in an arbitrary configuration, the
`
`sound source localization unit 202, the adaptive beamforming unit 203, and the noise
`
`reduction unit 207.
`
`[0051] The array 201 of sound sensors, herein referred to as the "microphone array" is
`
`in operative communication with the sound source localization unit 202, the adaptive
`
`beamforming unit 203, and the noise reduction unit 207. The microphone array 201 is,
`
`for example, a linear array of sound sensors, a circular array of sound sensors, or an
`
`arbitrarily distributed coplanar array of sound sensors. The microphone array 201
`
`achieves directional gain in any preferred spatial direction and frequency band while
`
`suppressing signals from other spatial directions and frequency bands. The sound sensors
`
`receive the sound signals comprising the target sound signal and ambient noise signals
`
`from multiple disparate sound sources, where one of the disparate sound sources is the
`
`target sound source that emits the target sound signal.
`
`[0052] The sound source localization unit 202 estimates the spatial location of the target
`
`sound signal from the received sound signals. In an embodiment, the sound source
`
`localization unit 202 uses, for example, a steered response power-phase transform, for
`
`estimating the spatial location of the target sound signal from the target sound source.
`
`[0053] The adaptive beamforming unit 203 steers the directivity pattern of the
`
`microphone array 201 in a direction of the spatial location of the target sound signal,
`
`thereby enhancing the target sound signal and partially suppressing the ambient noise
`
`signals. The adaptive beamforming unit 203 comprises a fixed beamformer 204, a
`
`blocking matrix 205, and an adaptive filter 206 as disclosed in the detailed description of
`
`FIG. 10. The fixed beamformer 204 performs fixed beamforming by filtering and
`
`summing output sound signals from each of the sound sensors in the microphone array
`
`201 as disclosed in the detailed description of FIG. 4. In an embodiment, the adaptive
`
`filter 206 is implemented as a set of sub-band adaptive filters. The adaptive filter 206
`
`comprises an analysis filter bank 206a, an adaptive filter matrix 206b, and a synthesis
`
`filter bank 206c as disclosed in the detailed description of FIG. 11.
`
`11
`
`Page 12 of 200
`
`
`
`[0054] The noise reduction unit 207 further suppresses the ambient noise signals for
`
`further enhancing the target sound signal. The noise reduction unit 207 is, for example, a
`
`Wiener-filter based noise reduction unit, a spectral subtraction noise reduction unit, an
`
`auditory transform based noise reduction unit, or a model based noise reduction unit.
`
`[0055] FIG. 3 exemplarily illustrates a microphone array configuration showing a
`
`microphone array 201 having N sound sensors 301 arbitrarily distributed on a circle 302
`
`with a diameter "d", where "N" refers to the number of sound sensors 301 in the
`
`microphone array 201. Consider an example where N = 4, that is, there are four sound
`
`sensors 301 M0, M 1, M2, and M3 in the microphone array 201. Each of the sound sensors
`301 is positioned at an acute angle "<Dn" from a Y-axis, where <Dn2: 0 and n=O, 1, 2, ... N-
`
`1. In an example, the sound sensor 301 Mo is positioned at an acute angle <1>0 from the Y(cid:173)
`axis; the sound sensor 301 M 1 is positioned at an acute angle <1> 1 from the Y-axis; the
`sound sensor 301 M2 is positioned at an acute angle <1>2 from the Y-axis; and the sound
`sensor 301 M3 is positioned at an acute angle <1>3 from the Y-axis. A filter-and-sum
`beamforming algorithm determines the output "y" of the microphone array 201 having N
`
`sound sensors 301 as disclosed in the detailed description of FIG. 4.
`
`[0056] FIG. 4 exemplarily illustrates a graphical representation of the filter-and-sum
`
`beamforming algorithm for determining the output of the microphone array 201 having N
`
`sound sensors 301. Consider an example where the target sound signal from the target
`
`sound source is at an angle 8 with a normalized frequency m. The microphone array
`
`configuration is arbitrary in a two dimensional plane, for example, a circular array
`
`configuration where the sound sensors 301 Mo, M1, M2, ... , MN, MN-l of the microphone
`array 201 are arbitrarily positioned on a circle 302. The sound signals received by each of
`
`the sound sensors 301 in the microphone array 201 are inputs to the microphone array
`
`201. The adaptive beamforming unit 203 employs the filter-and-sum beamforming
`
`algorithm that applies independent weights to each of the inputs to the microphone array
`
`201 such that directivity pattern of the microphone array 201 is steered to the spatial
`
`12
`
`Page 13 of 200
`
`
`
`location of the target sound signal as determined by the sound source localization unit
`
`202.
`
`[0057] The output "y" of the microphone array 201 having N sound sensors 301 is the
`
`filter-and-sum of the outputs of the N sound sensors 301. That is,
`
`'""N-l
`
`T
`
`y = L. n=o w n xn , where Xn 1s the output of the (n+ 1) sound sensor 301, and Wn
`th
`T
`
`,
`
`denotes a transpose of a length-L filter applied to the (n+ l)th sound sensor 301.
`
`[0058] The spatial directivity pattern H ( m, 8) for the target sound signal from angle 8
`
`with normalized frequency m is defined as:
`
`H(m,0)
`
`Y(m,0)
`X(m,0)
`
`I:-~Wn ( OJ)X n ( OJ, 0)
`
`X(m,0)
`
`(1)
`
`where X is the signal received at the origin of the circular microphone array 201 and Wis
`
`the frequency response of the real-valued finite impulse response (FIR) filter w. If the
`
`target sound source is far enough away from the microphone array 201, the difference
`between the signal received by the (n+ l)th sound sensor 301 "xn" and the origin of the
`microphone array 201 is a delay In; that is, X n ( OJ, r) = X ( OJ, 0)e - jOJTn •
`
`[0059] FIG. 5 exemplarily illustrates distances between an origin of the microphone
`
`array 201 and the sound sensor 301 M 1 and the sound sensor 301 M3 in the circular
`
`microphone array configuration, when the target sound signal is at an angle 0 from the Y(cid:173)
`
`axis. The microphone array system 200 disclosed herein can be used with an arbitrary
`
`directivity pattern for arbitrarily distributed sound sensors 301. For any specific
`
`microphone array configuration, the parameter that is defined to achieve beamformer
`
`coefficients is the value of delay Tn for each sound sensor 301. To define the value of Tn,
`
`an origin or a reference point of the microphone array 201 is defined; and then the
`
`distance dn between each sound sensor 301 and the origin is measured, and then the angle
`
`<Dn of each sound sensor 301 biased from a vertical axis is measured.
`
`13
`
`Page 14 of 200
`
`
`
`[0060] For example, the angle between the Y-axis and the line joining the origin and
`
`the sound sensor 301 M0 is <1>0, the angle between the Y-axis and the line joining the
`origin and the sound sensor 301 M 1 is <1> 1, the angle between the Y-axis and the line
`joining the origin and the sound sensor 301 M2 is <1>2, and the angle between the Y-axis
`and the line joining the origin and the sound sensor 301 M3 is <1>3_ The distance between
`the origin O and the sound sensor 301 M 1, and the origin O and the sound sensor 301 M3
`
`when the incoming target sound signal from the target sound source is at an angle 0 from
`
`the Y-axis is denoted as T 1 and T3, respectively.
`
`[0061] For purposes of illustration, the detailed description refers to a circular
`
`microphone array configuration; however, the scope of the microphone array system 200
`
`disclosed herein is not limited to the circular microphone array configuration but may be
`
`extended to include a linear array configuration, an arbitrarily distributed coplanar array
`
`configuration, or a microphone array configuration with any arbitrary geometry.
`
`[0062] FIG. 6A exemplarily illustrates a table showing the distance between each sound
`
`sensor 301 in a circular microphone array configuration from the origin of the
`
`microphone array 201, when the target sound source is in the same plane as that of the
`
`microphone array 201. The distance measured in meters and the corresponding delay (T)
`
`measured in number of samples is exemplarily illustrated in FIG. 6A. In an embodiment
`
`where the target sound source that emits the target sound signal is in a two dimensional
`
`plane, the delay ( T) between each