throbber
(12) United States Patent
`Burnett et al.
`
`(10) Patent No.:
`(45) Date of Patent:
`
`US 8.467,543 B2
`*Jun. 18, 2013
`
`USOO8467543B2
`
`(54)
`
`(75)
`
`(73)
`(*)
`
`(21)
`(22)
`(65)
`
`(60)
`
`(51)
`
`(52)
`
`(58)
`
`MCROPHONE AND VOICE ACTIVITY
`DETECTION (VAD) CONFIGURATIONS FOR
`USE WITH COMMUNICATION SYSTEMS
`
`Inventors: Gregory C. Burnett, Livermore, CA
`(US); Nicolas J. Petit, San Francisco,
`CA (US); Alexander M. Asseily, San
`Francisco, CA (US); Andrew E.
`Einaudi, San Francisco, CA (US)
`Assignee: AliphCom, San Francisco, CA (US)
`
`Notice:
`
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 977 days.
`This patent is Subject to a terminal dis
`claimer.
`
`Appl. No.: 10/400,282
`
`Filed:
`
`Mar. 27, 2003
`
`Prior Publication Data
`US 2003/0228O23 A1
`Dec. 11, 2003
`
`Related U.S. Application Data
`Provisional application No. 60/368,209, filed on Mar.
`27, 2002.
`
`Int. C.
`H04B I5/00
`U.S. C.
`USPC .......... 381/94.1: 381/92: 381/943: 381/94.7;
`704/226; 704/233
`
`(2006.01)
`
`Field of Classification Search
`USPC ................ 381/94.7, 71.6, 110, 71.7, 111, 11,
`381/91-92, 94.1-94.3: 704/226, 233
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`EP
`EP
`
`U.S. PATENT DOCUMENTS
`3,789,166 A
`1/1974 Sebesta
`4,006,318 A
`2f1977 Sebesta et al.
`(Continued)
`FOREIGN PATENT DOCUMENTS
`O 637 187 A
`2, 1995
`O 795 851 A2
`9, 1997
`(Continued)
`OTHER PUBLICATIONS
`Zhao Li et al: "Robust Speech Coding Using Microphone Arrays'.
`Signals Systems and Computers, 1997. Conf. record of 31st Asilomar
`Conf. Nov. 2-5, 1997, IEEE Comput. Soc. Nov. 2, 1997. USA.
`(Continued)
`Primary Examiner — Disler Paul
`(74) Attorney, Agent, or Firm — Kokka & Backus, PC
`(57)
`ABSTRACT
`Communication systems are described, including both por
`table handset and headset devices, which use a number of
`microphone configurations to receive acoustic signals of an
`environment. The microphone configurations include, for
`example, a two-microphone array including two unidirec
`tional microphones, and a two-microphone array including
`one unidirectional microphone and one omnidirectional
`microphone. The communication systems also include Voice
`Activity Detection (VAD) devices to provide information of
`human Voicing activity. Components of the communications
`systems receive the acoustic signals and Voice activity signals
`and, in response, automatically generate control signals from
`data of the Voice activity signals. Components of the commu
`nication systems use the control signals to automatically
`select a denoising method appropriate to data of frequency
`Subbands of the acoustic signals. The selected denoising
`method is applied to the acoustic signals to generate denoised
`acoustic signals when the acoustic signal includes speech and
`O1SC.
`
`26 Claims, 29 Drawing Sheets
`
`106 - ul-WAD
`
`Voicing Information
`
`s(n)
`
`H(z)
`
`H(z)
`
`n(n)
`
`s
`n(n) u-r Mic 1
`
`103
`\
`m(n)
`
`
`
`104
`
`m(n)
`s(n)
`Y. 1.
`Mic 2
`
`s
`(c)
`sign
`
`S
`
`t
`(i)
`Noise
`n(n)
`
`-100
`
`t
`Cleaned Speech
`
`
`
`105
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 1 of 44
`
`

`

`US 8,467,543 B2
`Page 2
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`. TO4,233
`
`U.S. PATENT DOCUMENTS
`5, 1986
`4,591,668
`Iwata
`2, 1990
`Gollmar et al.
`4,901,354
`3, 1992
`Baba
`5,097,515
`Ariyoshi
`5, 1993
`5,212,764
`* 10, 1994
`Oh et al.
`5,353,376
`3, 1995
`Linhard
`5,400.409
`Silverberg et al.
`4, 1995
`5,406,622
`Sims, Jr.
`5, 1995
`5,414,776
`* 12, 1995
`Yoshida et al.
`5,473,702
`5, 1996
`Scanlon et al.
`5,515,865
`Sugiyama
`5, 1996
`5,517,435
`T. 1996
`Robbe et al.
`5,539,859
`* 12, 1996
`Park et al. ..................... 704/227
`5,590,241
`4, 1997
`Matouk et al. ........... 379,392.01
`5,625,684
`ck
`5, 1997
`Kanamori et al.
`5,633,935
`Gupta et al.
`7, 1997
`5,649,055
`11, 1997
`Scanlon et al.
`5,684,460
`3, 1998
`Holzrichter et al.
`5,729,694
`5, 1998
`Hosoi et al.
`5,754,665
`11, 1998
`Warnaka et al.
`5,835,608
`12, 1998
`Scanlon
`5,853,005
`6, 1999
`Sasaki et al.
`5,917,921
`10, 1999
`5,966,090
`McEwan
`11, 1999
`5,986,600
`McEwan
`12, 1999
`Holzrichter
`6,006, 175
`Nagata
`12, 1999
`6,009,396
`5, 2000
`Martin et al.
`6,069,963
`6, 191,724
`2, 2001
`McEwan
`Ikeda
`6,266.422
`T/2001
`Handel et al.
`6,430,295
`8, 2002
`Housini ...................... 455,550.1
`9, 2004
`6,795,713
`Vaudrey et al.
`... 381 (94.7
`11/2005
`6,963,649
`12, 2005
`Turnbull et al.
`340/425.5
`6,980,092
`Yang et al. ...................... 381/92
`4, 2007
`7,206,418
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`.
`
`. 381 (94.7
`
`B1
`B1
`B1
`
`
`
`4/2002 Burnett et al.
`2002fOO39425 A1
`2003/0044025 A1* 3/2003 Ouyang et al. .................. 381/92
`2003/0130839 A1* 7/2003 Beaucoup et al. ............ TO4/226
`FOREIGN PATENT DOCUMENTS
`O984 660 A2
`3, 2000
`2000 3.12395
`11, 2000
`2001 1899.87
`T 2001
`WOO2 07151
`1, 2002
`
`EP
`JP
`JP
`WO
`
`OTHER PUBLICATIONS
`L.C. Ng et al.: “Denoising of Human Speech Using Combined
`Acoustic and EM Sensor Signal Processing”, 2000 IEEE Intl Confon
`Acoustics Speech and Signal Processing. Proceedings (Cat. No.
`00CH37100), Istanbul, Turkey, Jun. 5-9, 2000 XP002 186255, ISBN
`O-78O3-6293-4.
`S. Affes et al.: “A Signal Subspace Tracking Algorithm for Micro
`phone Array Processing of Speech”. IEEE Transactions on Speech
`and Audio Processing, N.Y. USA vol. 5, No. 5, Sep. 1, 1997.
`XPOOOTT4303. ISBN 1063-6676.
`Gregory C. Burnett: “The Physiological Basis of Glottal Electromag
`netic Micropower Sensors (GEMS) and Their Use in Defining an
`Excitation Function for the Human Vocal Tract’. Dissertation. Uni
`versity of California at Davis. Jan. 1999. USA.
`Todd J. Gable et al.: "Speaker Verification Using Combined Acoustic
`and EM Sensor Signal Processing”, ICASSP-2001, Salt Lake City,
`USA.
`A. Hussain: “Intelligibility Assessment of a Multi-Band Speech
`Enhancement Scheme'. Proceedings IEEE Intl. Conf. on Acoustics,
`Speech & Signal Processing (ICASSP-2000). Istanbul, Turkey. Jun.
`2000.
`
`* cited by examiner
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 2 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 1 of 29
`
`US 8.467,543 B2
`
`
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 3 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 2 of 29
`
`US 8.467,543 B2
`
`
`
`
`
`
`
`Microphone
`Configuration
`
`140
`
`150
`
`130
`
`120
`
`160
`
`170
`
`WAD
`Algorithm
`
`
`
`Pathfinder Noise
`Suppression
`
`
`
`Denoised
`Speech
`
`Communication
`Device
`
`FIG.1A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 4 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 3 of 29
`
`US 8,467,543 B2
`
`pouray)
`
`yosads
`
`sandepy
`
`TOY
`
`weer ee ie fee ee ee KE ee ee ne ee dee eeee
`
`
`
`
`
`J9]90087)ASTONaandepy
`
`ATewaIg
`
`JOUIOTOY
`
`dq]Old
`
`(LUVWord)
`
`(a)°u
`
`
`(u)0(«s))
`
`(2)H
`
`(u)!uYSION
`
`((*))
`
`HOS
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 5 of 44
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 5 of 44
`
`
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 4 of 29
`
`US 8.467,543 B2
`
`UNIDIRECTIONAL
`
`Omni-
`
`Super-
`
`Hyper-
`
`Bi
`
`
`
`
`
`
`
`is
`
`los I go"
`
`O
`
`126
`
`O
`
`110
`
`12 dB
`
`6 dB
`
`27%
`
`25%
`
`33%
`
`1.9
`
`2
`
`1.7
`
`Polar
`Response
`Pattern
`
`I
`
`Maxi
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`so
`
`lar
`
`O
`
`25 dB
`
`100%
`
`33%
`
`1
`
`1.7
`
`Angle of
`R 180
`(null angle)
`Rear
`Rejection
`(relative to front)
`Ambient Sound
`Sensitivity
`(relative to Omni)
`Distance
`Factor
`(relative to Omni)
`
`FIG.2
`(PRIOR ART)
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 6 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 5 Of 29
`
`US 8.467,543 B2
`
`300
`
`OMNI
`
`
`
`AWay
`from
`Speech
`
`Towards
`Speech
`
`FIG.3A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 7 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 6 of 29
`
`US 8.467,543 B2
`
`
`
`
`
`
`
`SpJBAOL
`
`?009d?
`
`{{{TOIH
`
`19SpuBH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 8 of 44
`
`

`

`U.S. Patent
`
`US 8.467,543 B2
`
`
`
`s
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 9 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 8 of 29
`
`US 8.467,543 B2
`
`400
`
`/
`
`N is
`
`- 1
`
`
`
`Towards
`Speech
`
`FIG.4A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 10 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 9 Of 29
`
`US 8.467,543 B2
`
`spJEA0J,
`
`?009dS
`
`SpI8AA0J,
`
`q000dS
`
`10spupH
`
`(INH
`INWO (€)
`
`
`
`
`
`ÁBAW
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 11 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 10 of 29
`
`US 8.467,543 B2
`
`
`
`
`
`spJBAOL
`
`?009dS
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 12 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 11 of 29
`
`US 8.467,543 B2
`
`
`
`500
`
`AWay
`from
`Speech
`
`Towards
`Speech
`
`FIG.5A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 13 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 12 of 29
`
`US 8.467,543 B2
`
`spIBMOL
`
`q009dS
`
`
`
`
`
`
`
`{{S'OIH
`
`19spueH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 14 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 13 of 29
`
`US 8.467,543 B2
`
`HCIIS
`
`INWO
`
`|----
`
`spJBAOL
`
`?009dS
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 15 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 14 of 29
`
`US 8.467,543 B2
`
`-
`
`
`
`N
`
`600
`
`/
`
`Towards
`Speech
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 16 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 15 Of 29
`
`US 8.467,543 B2
`
`(INH
`IN[] ©
`
`
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 17 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 16 of 29
`
`US 8.467,543 B2
`
`SpJBAOL |
`?000d?
`
`SpJEMOJ,
`
`?009dS
`
`09’OIH
`
`
`
`
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 18 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 17 Of 29
`
`US 8.467,543 B2
`
`
`
`700
`
`AWay
`from
`Speech
`
`Towards
`Speech
`
`FIG.7A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 19 of 44
`
`

`

`U.S. Patent
`
`US 8.467,543 B2
`
`
`
`
`
`{{L'OIH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 20 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 19 Of 29
`
`US 8.467,543 B2
`
`SpIBMOL
`
`?009d?
`
`OL'OIH
`
`
`
`s ".
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 21 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 20 Of 29
`
`US 8.467,543 B2
`
`
`
`800
`
`Towards
`Speech
`
`FIG.8A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 22 of 44
`
`

`

`U.S. Patent
`
`
`
`US 8.467,543 B2
`
`
`
`
`
`19SpuBH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 23 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 22 of 29
`
`US 8.467,543 B2
`
`SpI8AA0J, |
`?000dS
`
`O8’OIH
`
`
`
`
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 24 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 23 Of 29
`
`US 8.467,543 B2
`
`
`
`900
`
`(Orientation Irrelevant)
`
`FIG.9A
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 25 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 24 of 29
`
`US 8.467,543 B2
`
`
`
`
`
`
`
`19SpupH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 26 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 25 Of 29
`
`US 8.467,543 B2
`
`
`
`
`
`SpJBAOL
`
`q000dS
`
`SpJeAOL
`
`?000dS
`
`O6'OIH
`
`19Spe9H
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 27 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 26 of 29
`
`US 8.467,543 B2
`
`
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 28 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 27 Of 29
`
`US 8.467,543 B2
`
`
`
`O
`CN
`O
`y
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 29 of 44
`
`

`

`U.S. Patent
`
`US 8.467,543 B2
`
`
`
`VII’OIH
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 30 of 44
`
`

`

`U.S. Patent
`
`Jun. 18, 2013
`
`Sheet 29 Of 29
`
`US 8.467,543 B2
`
`
`
`O
`CN
`w
`va
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 31 of 44
`
`

`

`US 8,467,543 B2
`
`1.
`MCROPHONE AND VOICE ACTIVITY
`DETECTION (VAD) CONFIGURATIONS FOR
`USE WITH COMMUNICATION SYSTEMS
`
`RELATED APPLICATIONS
`
`This application claims priority from U.S. Patent Applica
`tion No. 60/368,209, entitled MICROPHONE AND VOICE
`ACTIVITY DETECTION (VAD) CONFIGURATIONS
`10
`FOR USE WITH PORTABLE COMMUNICATION SYS
`TEMS, filed Mar. 27, 2002.
`Further, this application relates to the following U.S. Patent
`Applications: Application Ser. No. 09/905,361, entitled
`METHOD AND APPARATUS FOR REMOVING NOISE
`FROM ELECTRONIC SIGNALS, filed Jul 12, 2001; appli
`cation Ser. No. 10/159,770, entitled DETECTING VOICED
`AND UNVOICED SPEECH USING BOTH ACOUSTIC
`AND NONACOUSTIC SENSORS, filed May 30, 2002;
`application Ser. No. 10/301.237, entitled METHOD AND
`APPARATUS FOR REMOVING NOISE FROM ELEC
`TRONICSIGNALS, filed Nov. 21, 2002; and application Ser.
`No. 10/383,162, entitled VOICE ACTIVITY DETECTION
`(VAD) DEVICES AND METHODS FOR USE WITH
`NOISE SUPPRESSION SYSTEMS, filed Mar. 5, 2003.
`
`15
`
`25
`
`TECHNICAL FIELD
`
`The disclosed embodiments relate to systems and methods
`for detecting and processing a desired acoustic signal in the
`presence of acoustic noise.
`
`30
`
`BACKGROUND
`
`Many noise Suppression algorithms and techniques have
`been developed over the years. Most of the noise suppression
`systems in use today for speech communication systems are
`based on a single-microphone spectral Subtraction technique
`first develop in the 1970s and described, for example, by S.
`F. Boll in “Suppression of Acoustic Noise in Speech using
`Spectral Subtraction.” IEEE Trans. on ASSP. pp. 113-120,
`1979. These techniques have been refined over the years, but
`the basic principles of operation have remained the same. See,
`for example, U.S. Pat. No. 5,687.243 of McLaughlin, et al.,
`and U.S. Pat. No. 4,811.404 of Vilmur, et al. Generally, these
`techniques make use of a single-microphone Voice Activity
`Detector (VAD) to determine the background noise charac
`teristics, where “voice' is generally understood to include
`human Voiced speech, unvoiced speech, or a combination of
`Voiced and unvoiced speech.
`The VAD has also been used in digital cellular systems. As
`an example of such a use, see U.S. Pat. No. 6,453,291 of
`Ashley, where a VAD configuration appropriate to the front
`end of a digital cellular system is described. Further, some
`Code Division Multiple Access (CDMA) systems utilize a
`VAD to minimize the effective radio spectrum used, thereby
`allowing for more system capacity. Also, Global System for
`Mobile Communication (GSM) systems can include a VAD
`to reduce co-channel interference and to reduce battery con
`sumption on the client or subscriber device.
`These typical single-microphone VAD systems are signifi
`cantly limited in capability as a result of the analysis of
`acoustic information received by the single microphone,
`wherein the analysis is performed using typical signal pro
`cessing techniques. In particular, limitations in performance
`of these single-microphone VAD systems are noted when
`processing signals having a low signal-to-noise ratio (SNR).
`and in settings where the background noise varies quickly.
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`2
`Thus, similar limitations are found in noise Suppression sys
`tems using these single-microphone VADs.
`Many limitations of these typical single-microphone VAD
`systems were overcome with the introduction of the Path
`finder noise Suppression system by Aliph of San Francisco,
`Calif. (http://www.aliph.com), described in detail in the
`Related Applications. The Pathfinder noise suppression sys
`tem differs from typical noise cancellation systems in several
`important ways. For example, it uses an accurate Voiced activ
`ity detection (VAD) signal along with two or more micro
`phones, where the microphones detecta mix of both noise and
`speech signals. While the Pathfinder noise suppression sys
`tem can be used with and integrated in a number of commu
`nication systems and signal processing systems, so can a
`variety of devices and/or methods be used to supply the VAD
`signal. Further, a number of microphone types and configu
`rations can be used to provide acoustic signal information to
`the Pathfinder system.
`
`BRIEF DESCRIPTION OF THE FIGURES
`
`FIG. 1 is a block diagram of a signal processing system
`including the Pathfinder noise removal or Suppression system
`and a VAD system, under an embodiment.
`FIG. 1A is a block diagram of a noise Suppression/com
`munication system including hardware for use in receiving
`and processing signals relating to VAD, and utilizing specific
`microphone configurations, under the embodiment of FIG.1.
`FIG.1B is a block diagram of a conventional adaptive noise
`cancellation system of the prior art.
`FIG. 2 is a table describing different types of microphones
`and the associated spatial responses in the prior art.
`FIG. 3A shows a microphone configuration using a unidi
`rectional speech microphone and an omnidirectional noise
`microphone, under an embodiment.
`FIG. 3B shows a microphone configuration in a handset
`using a unidirectional speech microphone and an omnidirec
`tional noise microphone, under the embodiment of FIG. 3A.
`FIG. 3C shows a microphone configuration in a headset
`using a unidirectional speech microphone and an omnidirec
`tional noise microphone, under the embodiment of FIG. 3A.
`FIG. 4A shows a microphone configuration using an omni
`directional speech microphone and a unidirectional noise
`microphone, under an embodiment.
`FIG. 4B shows a microphone configuration in a handset
`using an omnidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 4A.
`FIG. 4C shows a microphone configuration in a headset
`using an omnidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 4A.
`FIG. 5A shows a microphone configuration using an omni
`directional speech microphone and a unidirectional noise
`microphone, under an alternative embodiment.
`FIG. 5B shows a microphone configuration in a handset
`using an omnidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 5A.
`FIG. 5C shows a microphone configuration in a headset
`using an omnidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 5A.
`FIG. 6A shows a microphone configuration using a unidi
`rectional speech microphone and a unidirectional noise
`microphone, under an embodiment.
`FIG. 6B shows a microphone configuration in a handset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 6A.
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 32 of 44
`
`

`

`US 8,467,543 B2
`
`10
`
`15
`
`25
`
`30
`
`35
`
`40
`
`45
`
`3
`FIG. 6C shows a microphone configuration in a headset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 6A.
`FIG. 7A shows a microphone configuration using a unidi
`rectional speech microphone and a unidirectional noise
`microphone, under an alternative embodiment.
`FIG. 7B shows a microphone configuration in a handset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 7A.
`FIG. 7C shows a microphone configuration in a headset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 7A.
`FIG. 8A shows a microphone configuration using a unidi
`rectional speech microphone and a unidirectional noise
`microphone, under an embodiment.
`FIG. 8B shows a microphone configuration in a handset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 8A.
`FIG. 8C shows a microphone configuration in a headset
`using a unidirectional speech microphone and a unidirec
`tional noise microphone, under the embodiment of FIG. 8A.
`FIG.9A shows a microphone configuration using an omni
`directional speech microphone and an omnidirectional noise
`microphone, under an embodiment.
`FIG. 9B shows a microphone configuration in a handset
`using an omnidirectional speech microphone and an omnidi
`rectional noise microphone, under the embodiment of FIG.
`9A.
`FIG. 9C shows a microphone configuration in a headset
`using an omnidirectional speech microphone and an omnidi
`rectional noise microphone, under the embodiment of FIG.
`9A.
`FIG. 10A shows an area of sensitivity on the human head
`appropriate for receiving a GEMS sensor, under an embodi
`ment.
`FIG. 10B shows GEMS antenna placement on a generic
`handset or headset device, under an embodiment.
`FIG. 11A shows areas of sensitivity on the human head
`appropriate for placement of an accelerometer/SSM, under
`an embodiment.
`FIG. 11B shows accelerometer/SSM placement on a
`generic handset or headset device, under an embodiment.
`In the drawings, the same reference numbers identify iden
`tical or Substantially similar elements or acts. To easily iden
`tify the discussion of any particular element or act, the most
`significant digit or digits in a reference number refer to the
`Figure number in which that element is first introduced (e.g.,
`element 105 is first introduced and discussed with respect to
`FIG. 1).
`The headings provided herein are for convenience only and
`do not necessarily affect the scope or meaning of the claimed
`invention. The following description provides specific details
`for a thorough understanding of and enabling description for,
`embodiments of the invention. However, one skilled in the art
`will understand that the invention may be practiced without
`these details. In other instances, well-known structures and
`functions have not been shown or described in detail to avoid
`unnecessarily obscuring the description of the embodiments
`of the invention.
`
`50
`
`55
`
`60
`
`65
`
`DETAILED DESCRIPTION
`
`Numerous communication systems are described below,
`including both handset and headset devices, which use a
`variety of microphone configurations to receive acoustic sig
`nals of an environment. The microphone configurations
`include, for example, a two-microphone array including two
`
`4
`unidirectional microphones, and a two-microphone array
`including one unidirectional microphone and one omnidirec
`tional microphone, but are not so limited. The communication
`systems can also include Voice Activity Detection (VAD)
`devices to provide Voice activity signals that include infor
`mation of human Voicing activity. Components of the com
`munications systems receive the acoustic signals and Voice
`activity signals and, in response, automatically generate con
`trol signals from data of the Voice activity signals. Compo
`nents of the communication systems use the control signals to
`automatically select a denoising method appropriate to data
`of frequency Subbands of the acoustic signals. The selected
`denoising method is applied to the acoustic signals to gener
`ate denoised acoustic signals when the acoustic signals
`include speech and noise.
`Numerous microphone configurations are described below
`for use with the Pathfinder noise Suppression system. As such,
`each configuration is described in detail along with a method
`of use to reduce noise transmission in communication
`devices, in the context of the Pathfinder system. When the
`Pathfinder noise suppression system is referred to, it should
`be kept in mind that noise Suppression systems that estimate
`the noise waveform and Subtract it from a signal and that use
`or are capable of using the disclosed microphone configura
`tions and VAD information for reliable operation are included
`in that reference. Pathfinder is simply a convenient referenced
`implementation for a system that operates on signals com
`prising desired speech signals along with noise. Thus, the use
`of these physical microphone configurations includes but is
`not limited to applications such as communications, speech
`recognition, and Voice-feature control of applications and/or
`devices.
`The terms “speech” or “voice” as used herein generally
`refer to voiced, unvoiced, or mixed voiced and unvoiced
`human speech. Unvoiced speech or voiced speech is distin
`guished where necessary. However, the term "speech signal”
`or “speech', when used as a converse to noise, simply refers
`to any desired portion of a signal and does not necessarily
`have to be human speech. It could, as an example, be music or
`Some other type of desired acoustic information. As used in
`the Figures, “speech” is meant to mean any signal of interest,
`whether human speech, music, or anything other signal that it
`is desired to hear.
`In the same manner, “noise' refers to unwanted acoustic
`information that distorts a desired speech signal or makes it
`more difficult to comprehend. “Noise suppression' generally
`describes any method by which noise is reduced or eliminated
`in an electronic signal.
`Moreover, the term “VAD is generally defined as a vector
`or array signal, data, or information that in Some manner
`represents the occurrence of speech in the digital or analog
`domain. A common representation of VAD information is a
`one-bit digital signal sampled at the same rate as the corre
`sponding acoustic signals, with a Zero value representing that
`no speech has occurred during the corresponding time
`sample, and a unity value indicating that speech has occurred
`during the corresponding time sample. While the embodi
`ments described herein are generally described in the digital
`domain, the descriptions are also valid for the analog domain.
`The term “Pathfinder, unless otherwise specified, denotes
`any denoising system using two or more microphones, a VAD
`device and algorithm, and which estimates the noise in a
`signal and subtracts it from that signal. The Aliph Pathfinder
`system is simply a convenient reference for this type of
`denoising system, although it is more capable than the above
`definition. In some cases (such as the microphone arrays
`described in FIGS. 8 and 9), the “full capabilities” or “full
`
`Meta Platforms, Inc. - Exhibit 1001
`Page 33 of 44
`
`

`

`US 8,467,543 B2
`
`10
`
`15
`
`30
`
`35
`
`40
`
`5
`version of the Aliph Pathfinder system are used (as there is a
`significant amount of speech energy in the noise micro
`phone), and these cases will be enumerated in the text. “Full
`capabilities’ indicates the use of both H (Z) and H2(z) by the
`Pathfinder system in denoising the signal. Unless otherwise
`specified, it is assumed that only H(Z) is used to denoise the
`signal.
`The Pathfinder system is a digital signal processing—
`(DSP) based acoustic noise Suppression and echo-cancella
`tion system. The Pathfinder system, which can couple to the
`front-end of speech processing systems, uses VAD informa
`tion and received acoustic information to reduce or eliminate
`noise in desired acoustic signals by estimating the noise
`waveform and Subtracting it from a signal including both
`speech and noise. The Pathfinder system is described further
`below and in the Related Applications.
`FIG. 1 is a block diagram of a signal processing system 100
`including the Pathfinder noise removal or Suppression system
`105 and a VAD system 106, under an embodiment. The signal
`processing system 100 includes two microphones MIC 1103
`and MIC 2 104 that receive signals or information from at
`least one speech signal source 101 and at least one noise
`source 102. The path s(n) from the speech signal source 101
`to MIC 1 and the path n(n) from the noise source 102 to MIC
`25
`2 are considered to be unity. Further, H (Z) represents the path
`from the noise source 102 to MIC 1, and H2(z) represents the
`path from the speech signal source 101 to MIC 2.
`Components of the signal processing system 100, for
`example the noise removal system 105, couple to the micro
`phones MIC 1 and MIC 2 via wireless couplings, wired
`couplings, and/or a combination of wireless and wired cou
`plings. Likewise, the VAD system 106 couples to components
`of the signal processing system 100, like the noise removal
`system 105, via wireless couplings, wired couplings, and/or a
`combination of wireless and wired couplings. As an example,
`the VAD devices and microphones described below as com
`ponents of the VAD system 106 can comply with the Blue
`tooth wireless specification for wireless communication with
`other components of the signal processing system, but are not
`so limited.
`FIG. 1A is a block diagram of a noise Suppression/com
`munication system including hardware for use in receiving
`and processing signals relating to VAD, and utilizing specific
`microphone configurations, under an embodiment. Referring
`to FIG. 1A, each of the embodiments described below
`includes at least two microphones in a specific configuration
`110 and one voiced activity detection (VAD) system 130,
`which includes both a VAD device 140 and a VAD algorithm
`150, as described in the Related Applications. Note that in
`Some embodiments the microphone configuration 110 and
`the VAD device 140 incorporate the same physical hardware,
`but they are not so limited. Both the microphones 110 and the
`VAD 130 input information into the Pathfinder noise suppres
`sion system 120 which uses the received information to
`denoise the information in the microphones and output
`denoised speech 160 into a communications device 170.
`The communications device 170 includes both handset and
`headset communication devices, but is not so limited. Hand
`sets or handset communication devices include, but are not
`limited to, portable communication devices that include
`microphones, speakers, communications electronics and
`electronic transceivers, such as cellular telephones, portable
`or mobile telephones, satellite telephones, wireline tele
`phones, Internet telephones, wireless transceivers, wireless
`communication radios, personal digital assistants (PDAs),
`and personal computers (PCs).
`
`55
`
`45
`
`50
`
`60
`
`65
`
`6
`Headset or headset communication devices include, but are
`not limited to, self-contained devices including microphones
`and speakers generally attached to and/or worn on the body.
`Headsets often function with handsets via couplings with the
`handsets, where the couplings can be wired, wireless, or a
`combination of wired and wireless connections. However, the
`headsets can communicate independently with components
`of a communications network.
`The VAD device 140 includes, but is not limited to, accel
`erometers, skin surface microphones (SSMs), and electro
`magnetic devices, along with the associated Software or algo
`rithms. Further, the VAD device 140 includes acoustic
`microphones along with the associated software. The VAD
`devices and associated software are described in U.S. patent
`application Ser. No. 10/383,162, entitled VOICE ACTIVITY
`DETECTION (VAD) DEVICES AND METHODS FOR
`USE WITH NOISE SUPPRESSION SYSTEMS, filed Mar.
`5, 2003.
`The configurations described below of each handset/head
`set design include the location and orientation of the micro
`phones and the method used to obtain a reliable VAD signal.
`All other components (including the speaker and mounting
`hardware for headsets and the speaker, buttons, plugs, physi
`cal hardware, etc. for the handsets) are inconsequential for the
`operation of the Pathfinder noise Suppression algorithm and
`will not be discussed in great detail, with the exception of the
`mounting of unidirectional microphones in the handset or
`headset. The mounting is described to provide information
`for the proper ventilation of the directional microphones.
`Those familiar with the state of the art will not have difficulty
`mounting the unidirectional microphones correctly given the
`placement and orientation information in this application.
`Furthermore, the method of coupling (either physical or
`electromagnetic or otherwise) of the headsets described
`below is inconsequential. The headsets described work with
`any type of coupling, so they are not specified in this disclo
`sure. Finally, the microphone configuration 110 and the VAD
`130 are independent, so that any microphone configuration
`can work with any VAD device/method, unless it is desired to
`use the same microphones for both the VAD and the micro
`phone configuration. In this case the VAD can place certain
`requirements on the microphone configuration. These excep
`tions are noted in the text.
`Microphone Configurations
`The Pathfinder system, although using particular micro
`phone types (omnidirectional or unidirectional, including the
`amount of unidirectionality) and microphone orientations, is
`not sensitive to the typical distribution of responses of indi
`vidual microphones of a given type. Thus the microphones do
`not need to be matched in terms of frequency response nor do
`they need to be especially sensitive or expensive. In fact,
`configurations described herein have been constructed using
`inexpensive off-the-shelf microphones, which have proven to
`be very effective. As an aid to review, the Pathfinder setup is
`shown in FIG. 1 and is explained in detail below and in the
`Related Applications. The relative placement and orientation
`of the microphones in the Pathfinder system is described
`herein. Unlike classical adaptive noise cancellation (ANC),
`which specifies that there can be no speech signal in the

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket