throbber
UNITED STATES PATENT AND TRADEMARK OFFICE
`
`
`
`
`
`
`BEFORE THE PATENT TRIAL AND APPEAL BOARD
`
`
`
`
`
`
`
`
`GOOGLE LLC,
`
`Petitioner,
`
`v.
`
`VOCALIFE, LLC
`
`Patent Owner.
`
`
`
`
`
`
`
`
`Case No. IPR2022-00004
`U.S. Patent No. RE47,049
`
`
`
`
`
`
`
`
`Declaration of Shauna L. Wiest Regarding Briere
`
`
`
`
`
`
`
`
`Page 1 of 29
`
`GOOGLE EXHIBIT 1020
`
`

`

`
`I.
`
`Introduction
`
`Declaration of Shauna L. Wiest
`
`
`1.
`
`I have prepared this Declaration in connection with the Petition for
`
`Inter Partes Review of U.S. Patent No. RE47,049, which I understand will be
`
`filed concurrently with this Declaration.
`
`2.
`
`I am currently a contract research analyst with the Research &
`
`Information Services team at Finnegan, Henderson, Farabow, Garrett &
`
`Dunner, LLP, located at 901 New York Avenue, NW, Washington, DC 20001-
`
`4413.
`
`3.
`
`I am over eighteen years of age, and I am competent to make this
`
`Declaration. I make this Declaration based on my own personal knowledge,
`
`and my knowledge of library science practices.
`
`4.
`
`I earned a Master of Science in Library Science from the
`
`University of North Carolina at Chapel Hill in 1999, and a Bachelor of Arts in
`
`Political Science from the University of California at San Diego in 1985. I
`
`have worked as a librarian for over twenty years. I have been employed in the
`
`Research & Information Services Department at Finnegan Henderson since
`
`2021. Before that, from 2000-2015, I was employed as a Law Librarian at
`
`Stoel Rives LLP. And from 2015-2016, I was employed as a Competitive
`
`Intelligence Specialist for Nossaman LLP.
`
`
`
`2
`
`Page 2 of 29
`
`

`

`
`II.
`
`Standard Library Practice for Receiving, Cataloging, and Making
`Materials Publicly Available
`
`Declaration of Shauna L. Wiest
`
`
`5.
`
`I have knowledge of and experience with standard library practices
`
`regarding the receipt, cataloging, and making materials publicly available. I have
`
`knowledge of and experience with the Machine-Readable Cataloging (MARC)
`
`system, an industry-wide standard that libraries use to catalog materials. I also
`
`have knowledge of and experience with OCLC (previously called Online
`
`Computer Library Center, Inc.) and OCLC Control Numbers, which are unique
`
`accession numbers assigned by the OCLC system when a bibliographic record
`
`is added to WorldCat. I also have knowledge of and experience with WorldCat,
`
`which is an online database of records cooperatively built from the bibliographic
`
`and ownership information of OCLC contributing libraries. WorldCat is the
`
`foundation of many OCLC services that enable institutions to process, manage,
`
`and share information resources, including cataloging, resource sharing,
`
`reference, discovery, collection evaluation, and local holdings.
`
`6.
`
`By the mid-1970s, standard library practice involved cataloging
`
`items utilizing the MARC system. The MARC system was developed during the
`
`1960s to standardize bibliographic records so they could be read by computers and
`
`shared among libraries. By the mid-1970s, MARC and OCLC had become the
`
`international standards for bibliographic data and cataloguing. They are still used
`
`today.
`
`
`
`3
`
`Page 3 of 29
`
`

`

`
`
`7.
`
`Generally, after an item is cataloged, the public may access the item
`
`Declaration of Shauna L. Wiest
`
`
`by searching the catalog and either requesting or electronically accessing the item
`
`from the library. Standard library practice is to make the item available to the
`
`public within a few days or weeks of cataloging it.
`
`III. WorldCat, OCLC and Digital Libraries
`
`8. WorldCat is the world’s largest network of library content and
`
`services with 517,963,343 bibliographic records from over 72,000 libraries as of
`
`July 2021. (Source: https://www.oclc.org/en/worldcat/inside-worldcat.html). An
`
`Online Computer Library Center (“OCLC”) number is a unique control number
`
`given to all bibliographic records in the WorldCat catalog. OCLC control numbers
`
`are often used by the public to search for specific records within WorldCat.
`
`9.
`
`According to Digital Libraries by William Y. Arms, published by the
`
`MIT Press in January 2000 (and available here:
`
`https://www.cs.cornell.edu/wya/DigLib/text/Chapter1.html), an informal definition
`
`of a digital library is “a managed collection of information, with associated
`
`services, where the information is stored in digital formats and accessible over a
`
`network. A key part of this definition is that the information is managed.”
`
`Moreover, “Digital libraries contain diverse collections of information for use by
`
`many different users. Digital libraries range in size from tiny to huge. They can use
`
`any type of computing equipment and any suitable software. The unifying theme is
`
`that information is organized on computers and available over a network, with
`4
`
`
`Page 4 of 29
`
`

`

`
`procedures to select the material in the collections, to organize it, to make it
`
`Declaration of Shauna L. Wiest
`
`
`available to users, and to archive it.”
`
`IV. OCLC & Digital Library Records for Briere
`
`10. Appendix A to this declaration is a true and correct copy of
`
`Simon Briére, Dominic Létourneau, Maxime Fréchette, Jean-Marc Valin, and
`
`Francois Michaud, Embedded and Integrated Audition for a Mobile Robot
`
`(2006) (“Briere”), presented at AAAI 2006 Fall Symposium: Aurally
`
`Informed Performance: Integrating Machine Listening and Auditory
`
`Presentation in Robotic Systems (Symposium Track), held Friday through
`
`Sunday, October 13–15, 2006 at the Hyatt Regency Crystal City in
`
`Washington, DC. Technical Report FS-06-01. AAAI Press, Menlo Park,
`
`California, 2006. I understand that Briere has been submitted as Exhibit 1010
`
`in this proceeding.
`
`11. Appendix B to this declaration is a true and accurate copy of the
`
`OCLC record for the AAAI 2006 Fall Symposium papers containing Briere.
`
`(OCLC Control Number 166143839; Permalink:
`
`http://www.worldcat.org/oclc/166143839). Appendix B denotes the title of the
`
`OCLC record as “Aurally informed performance : integrating machine listening
`
`and auditory presentation in robotic systems : papers from the AAAI Fall
`
`Symposium.” The publisher and series information denotes that the referenced
`
`
`
`5
`
`Page 5 of 29
`
`

`

`
`papers are from the 2006 Symposium, and would include Briere. (OCLC Control
`
`Declaration of Shauna L. Wiest
`
`
`Number 166143839). (Appendix B at 1-2).
`
`12. Appendix C to this declaration is a true and accurate copy of the
`
`uniform resource locater (“URL”) and webpage for AAAI’s Digital Library of
`
`AAAI Fall Symposium Series Papers within the date range of 1992-2017 (URL:
`
`https://aaai.org/Library/Symposia/fallsymposia-library.php). The AAAI digital
`
`library provides the following directional advice for users seeking to locate and
`
`access individual fall symposia papers in the collection: “For individual papers or
`
`abstracts of fall symposium papers, you should consult the following contents.”
`
`Users can then either click on the year 2006 at the top of the page or scroll down
`
`and identify the symposium year for which they are seeking content. For year
`
`2006, users can click or scroll to “Aurally Informed Performance: Integrating
`
`Machine Listening and Auditory Presentation in Robotic Systems — Derek Brock,
`
`Ramani Duraiswami, and Alexander I. Rudnicky, Program Cochairs” and then
`
`click the hyperlink FS-06-01 to access all papers presented during the Fall 2006
`
`symposium track (Appendix C at 5).
`
`13. Appendix D to this declaration is a true and accurate copy of the
`
`URL and webpage of “Papers from the 2006 AAAI Fall Symposium” – “Aurally
`
`Informed Performance: Integrating Machine Listening And Auditory Presentation
`
`In Robotic Systems” (URL: https://aaai.org/Library/Symposia/Fall/fs06-01.php).
`
`On this digital library page, under “Contents,” at page 6, is a link to a true and
`6
`
`
`Page 6 of 29
`
`

`

`
`correct copy of the Briere paper, which provides a copyright date of 2006 by the
`
`Declaration of Shauna L. Wiest
`
`
`American Association for Artificial Intelligence.
`
`V. Conclusion
`
`14.
`
`In signing this Declaration, I understand it will be filed as
`
`evidence in a contested case before the Patent Trial and Appeal Board of the
`
`United States Patent and Trademark Office. I understand I may be subject to
`
`cross-examination in this case and that cross-examination will take place
`
`within the United States. If cross-examination is required of me, I will appear
`
`for cross-examination within the United States during the time allotted for
`
`cross-examination.
`
`15.
`
`I declare that all statements made herein of my knowledge are true,
`
`that all statements made on information and belief are believed to be true, and that
`
`these statements were made with the knowledge that willful false statements and
`
`the like so made are punishable by fine or imprisonment, or both, under Section
`
`1001 of Title 18 of the United States Code.
`
`
`
`Executed on October 7, 2021, in Washington, DC.
`
`
`
`
`
`
`
`Shauna L. Wiest
`
`7
`
`Page 7 of 29
`
`

`

`APPENDIX A
`APPENDIX A
`
`Page 8 of 29
`
`Page 8 of 29
`
`

`

`Embedded and Integrated Audition for a Mobile Robot
`
`S. Bri`ere, D. L´etourneau, M. Fr´echette, J.-M. Valin, F. Michaud
`Universit´e de Sherbrooke
`Sherbrooke (Qu´ebec) CANADA J1K 2R1
`{simon.briere,dominic.letourneau,maxime.frechette,jean-marc.valin,francois.michaud}@USherbrooke.ca
`
`Abstract
`
`”Aurally Informed Performance’ for mobile robots op-
`erating in natural environments brings difficult chal-
`lenges, such as:
`localizing sound sources all around
`the robot; tracking these sources as they or the robot
`move; separate the sources as a pre-processing step for
`recognition and processing, in real-time; dialogue man-
`agement and interaction in crowded conditions; evalu-
`ating performances of the different processing compo-
`nents in open conditions. In this paper, we present how
`we address these challenges by describing our eight mi-
`crophone system for sound source localization, track-
`ing and separation, our on-going work on its DSP im-
`plementation, and the use of the system on Spartacus,
`our mobile robot entry to AAAI Mobile Robot Compe-
`titions addressing human-robot interaction in open set-
`tings.
`
`Introduction
`To deal with the problems of auditory scene analysis on
`a mobile robot, we have developed a system that uses an
`array of eight microphones to localize, track and separate
`sound sources, in real-time and in noisy and reverberant en-
`vironments (Valin, Michaud, & Rouat 2006; Valin, Rouat,
`& Michaud 2004). The system was first validated on Spar-
`tacus, a robot participant to the AAAI 2005 Mobile Robot
`Challenge, making the robot attend the conference as a reg-
`ular attendee (Michaud et al. 2005). Spartacus uses three
`onboard computers to implement its software architecture,
`with one dedicated to the audition capabilities of the robot.
`Software integration is done using MARIE, our middleware
`framework for distributed component processing (Cote et al.
`2006). Using its audition skills, Spartacus is able to position
`its pan-tilt-zoom camera in the direction of sound sources, to
`navigate by following a sound source, to understand specific
`vocal requests (using NUANCE1) and to respond accord-
`ingly (using a preprogrammed dialogue manager and Festi-
`val for speech generation).
`Copyright c(cid:2) 2006, American Association for Artificial Intelli-
`gence (www.aaai.org). All rights reserved.
`1http://www.nuance.com
`
`Since then, we have pursued our work to extend the capa-
`bilities of the system by: 1) porting the audio extraction and
`sound source separation algorithms on a floating-point DSP
`(Digital Signal Processor), providing specialized processing
`resources to free the onboard computer for other decisional
`components, and also allowing the system to be used on dif-
`ferent robots and in various settings; and 2) improving the
`audition modalities on the robot for more natural interac-
`tion with people. In this paper, we present these two trusts
`and explain how Spartacus audition modalities (localization,
`tracking and separation of sound sources; dialogue man-
`agement and interaction) have improved for its participation
`to AAAI 2006 Mobile Robot Open Interaction competition
`(held in July 2006).
`
`DSP Implementation of our Sound Source
`Localization, Tracking and Separation
`(SSLTS) System
`It uses an
`Our SSLTS system is represented in Figure 1.
`array of eight microphones, spatially positioned on the front
`and the back of the robot’s torso. The system is made of
`three main modules: the localization module, the tracking
`module and the sound separation module.
`The localization module is used to localize multiples
`sources in the environment surrounding the robot. It uses
`a steered beamformer (Valin et al. 2004; Valin, Michaud, &
`Rouat 2006) approach (also known as SRP-PHAT) to find
`the position of the sources. By computing the weighted
`cross-correlation between each microphone pair in the sys-
`tem (28 pairs when using 8 microphones), the beamformer
`determines the direction of arrival of multiple sources using
`1024-sample frames at a 48 kHz sampling frequency.
`When the source positions are found, the tracking module
`is used to follow moving sources. By using particle filters
`(Arulampalam et al. 2002) for each source already being
`tracked, the system updates the state (position and velocity)
`of the detected sources as they move based on a probabilis-
`tic representation of the problem. This module also manages
`which sources are being tracked by determining if the local-
`ized source is a real source (not a false detection) and if the
`
`Page 9 of 29
`
`

`

`at frequencies from 32 kHz to 192 kHz. These ADC are
`used to capture the microphone outputs. The board also
`provides 64 Mbytes of external memory (SDRAM) and 8
`Mbytes of flash ROM memory. It also provides an USB2
`interface that can be used to control the system and to
`transfer the separated sound sources to another computer
`for post-processing (e.g., speech recognition).
`
`Figure 1: Block diagram of our SSLTS system.
`
`Figure 2: Lyrtech CPA-II, our development board.
`
`sources already tracked are still present.
`The information provided by the tracking module is used
`by the sound separation module to separate the detected
`sources. The separation is necessary for analyzing the au-
`dio content of a single source, for example when using a
`speech recognition system. In order to separate the audio
`sources, the separation module uses an modified version
`of the geometric sound source (GSS) separation algorithm.
`The modifications to the GSS were made necessary to en-
`sure that the algorithm could be used in real-time. A multi-
`source post-processing algorithm is used to remove interfer-
`ence and noise from the GSS-separated audio streams. The
`post-processing is optimized for speech recognition accu-
`racy (and not for human audibility quality).
`Evaluation and usability of this system will be limited
`by its portability. For instance, a small mobile robot to be
`used to study human-robot interaction with autistic children
`(Michaud, Duquette, & Nadeau 2003) needs to be small and
`cannot require computer-intensive processing. Currently,
`the system requires 25% of the processing power of a 1.6
`GHz Pentium M laptop for the localization and tracking
`modules (Valin, Michaud, & Rouat 2006) and 25% of the
`same processor for the separation module. So that is why
`we ported the implementation of these modules on a DSP,
`as explained in the following subsection.
`
`DSP Implementation
`a
`on
`done
`is
`Our
`SSLTS-DSP
`implementation
`TMS3206C713 Texas
`Instruments DSP. This proces-
`sor is a floating-point 225 MHz processor with 256 kbytes
`of internal RAM memory with L1 and L2 cache support.
`According to the specifications,
`the processor is rated
`at 225 MIPS and 1500 MFLOPS, and its architecture is
`optimized for audio processing. The DSP is built on a
`Lyrtech2) CPA-II board shown in Figure 2. This board
`has 24 bits analog-to-digital converters (ADC) able to run
`
`2http://www.lyrtech.com
`
`The transfer of the system on a dedicated processor is mo-
`tivated by some reasons. First of all, we want to make a sys-
`tem that is better suited to robotics applications in term of
`power consumption and size than a laptop. Having an exter-
`nal board removes particular software and OS dependency,
`thus allowing the system to be used in many different con-
`figurations and systems. Because the processing would be
`done by a different processor than the embedded robot one,
`we will have more CPU time available to other complex al-
`gorithm (e.g. image recognition). We also hope that such a
`system will cost less than a laptop.
`Work is currently in progress to port the actual system on
`the DSP board. The first phase of the project was to port the
`system from a Linux C++ based environment to a TI DSP
`C-based environment. Using recorded microphones outputs,
`we were able to obtain similar results between the DSP and
`the original system. We used eight simulated microphones
`inputs at a sampling frequency of 48 kHz with an overlap
`of 50% and 1024 sample frames. However, there was no
`memory management and the code was not optimized for
`DSP processing, which uses Very Large Instruction Words
`(VLIW) for parallel processing.
`To meet the original specifications of the system, the
`processing needs to be done under 21.33 ms for a frame size
`of 1024 samples frames with no overlap, and 10.66 ms for
`a frame size of 1024 samples with 50% overlap which is
`the case in the original system. We are trying to build the
`system using similar parameters as the original system: four
`simultaneous sources at 48 kHz and the same localization
`resolution. Currently, without optimization, the localization
`and tracking modules on the DSP require between 13 ms
`and 106 ms of processing cycle time, depending on the num-
`ber of sound sources being tracked. The separation module
`requires from 5 ms (for one sources) and 20 ms (for four
`sources) of the DSP processing cycle time. The worst case
`scenario is then 126 ms, which is about 12 times slower than
`the real-time requirement of 10.66 ms.
`We are currently working on different solutions to im-
`prove the system processing cycle time, and the optimiza-
`
`Page 10 of 29
`
`

`

`tion process is currently on-going. One is to make the nec-
`essary changes to the code in order to take advantage of the
`DSP parallel architecture. The DSP optimization tools are
`helping us pinpoint critical elements in the processing done
`by the DSP. Improving memory management will also bring
`improved results. The SSLTS has high memory require-
`ments because it uses eight microphones, with 1024 sample
`frames for each one, represented using floating point com-
`plex numbers. The DSP external memory is slow compared
`to its internal memory, and a more efficient use of the inter-
`nal memory may substantially improve the processing cycle
`time. However, it appears evident at this point that it will be
`impossible to build the original system on the DSP with the
`exact same parameters, because of memory limitations. The
`algorithm uses an array of 71736 floating-point elements in
`order to perform an acurate localization on a 2562 point grid
`around each microphone. Each elements in the array needs
`to be accessed one at the time in a loop, making the bene-
`fit of using the cache minimal because the values are read
`but not reused afterward. Because the array is about 280
`kilobytes in size and that we have 256 kilobytes of internal
`memory on the processor, we have to put the array in exter-
`nal memory, resulting in slower performance. To success-
`fully build the system on that particular processor, we prob-
`ably will have to reduce sampling rate, reduce the resolution
`of the localization system and reduce the maximum number
`of simultaneous sources in order to reach the targeted time
`requirements.
`
`Dialogue Management and Integration on a
`Mobile Robot
`Spartacus integrates planning and scheduling, sound source
`localization,
`tracking and separation, message reading,
`speech recognition and generation, and autonomous navi-
`gation capabilities onboard a custom-made interactive ro-
`bot.
`Integration of such a high number of capabilities
`revealed interesting new issues such as coordinating au-
`dio/visual/graphical capabilities, monitoring the impacts of
`the capabilities in usage by the robot, and inferring the ro-
`bot’s intentions and goals. Spartacus development is on-
`going and aims at adding new capabilities to the robot and
`improving our software and computational architectures, ad-
`dressing issues of human-robot interaction and the inte-
`grated challenges of designing an autonomous service ro-
`bot. Reported issues are: 1) the robust integration of differ-
`ent software packages and intelligent decision-making ca-
`pabilities; 2) natural interaction modalities in open settings;
`3) adaptation to environmental changes for localization; and
`4) monitoring/reporting decisions made by the robot (Gock-
`ley et al. 2004; Smart et al. 2003; Maxwell et al. 2004;
`Simmons et al. 2003).
`For our participation to the AAAI 2006 Mobile Robot
`Competition, Spartacus is designed to be a scientific robot
`reporter, in the sense of a human-robot interaction research
`assistant. The objective is to have Spartacus provide under-
`standable and configurable interaction, intention and infor-
`mation in unconstrained environmental conditions, report-
`ing the robot experiences for scientific data analysis.
`In-
`
`teractions occurring in open settings are rapid, diverse and
`context-related. Having an autonomous robot determining
`on its own when and what it has to do based on a variety of
`modalities (time constraints, events occurring in the world,
`requests from users, etc.) also makes it difficult to under-
`stand the robot’s behavior just by looking at it. Therefore,
`we concentrated our integration effort this year to design
`a robot that can interact and explain, through speech and
`graphical displays, its decisions and its experiences as they
`occur (for on-line and off-line diagnostics) in open settings.
`The robot is programmed to respond to requests from peo-
`ple, and with only one intrinsic goal of wanting to recharge
`when its energy is getting low (by either going to an outlet
`identified on the map or by searching for one, and then ask
`to be plugged in). Spartacus’ duty is to address these re-
`quests to the best of its capabilities and what is ‘robotically’
`possible. Such requests may be to deliver a written or a vo-
`cal message to a specific location or to a specific person, to
`meet at a specific time and place, to schmooze, etc. The ro-
`bot may receive multiple requests at different periods, and
`will have to manage on its own what, when and how it will
`satisfy them.
`Spartacus is shown in Figure 3. Spartacus is equipped
`with a SICK LMS200 laser range finder, a Sony SNC-
`RZ30N 25X PTZ color camera, the microphone array placed
`on the robot’s body, a touchscreen interface, an audio am-
`plifier and speakers, a business card dispenser and a LEDs
`electronic display. High-level processing is carried out us-
`ing an embedded Mini-ITX computer (Pentium M 1.7 GHz),
`and two laptop computers (Pentium M 1.6 GHz) installed on
`the platform. One is equipped with a RME Hammerfal DSP
`Multiface sound card using eight analog inputs to simulta-
`neously sample signals coming from the microphone array.
`It is also connected to the audio amplifier and speakers using
`the audio output port. The other laptop does video process-
`ing and is connected to the camera through a 100Mbps Eth-
`ernet link. Communication between the three on-board com-
`puters is accomplished with a 100Mbps Ethernet link. All
`computers are running Debian GNU Linux.
`Figure 4 illustrates Spartacus’ software architecture.
`integrates Player for sensor and actuator abstraction
`It
`layer (Vaughan, Gerkey, & Howard 2003), and CARMEN
`(Carnegie Mellon Robot Navigation Toolkit) for path plan-
`ning and localization (Montemerlo, Roy, & Thrun 2003).
`Also integrated but not shown on the figure is Stage/Gazebo
`for 2D and 3D simulators, and Pmap library3 for 2D map-
`ping, all designed at the University of Southern California.
`RobotFlow and FlowDesigner (FD) (Cote et al. 2004) are
`also used to implement the behavior-producing modules and
`SSLTS. For speech recognition and dialogue management,
`we interfaced this year the CSLU toolkit4. Software in-
`tegration of all these components are made possible using
`MARIE, a middleware framework oriented towards devel-
`oping and integrating new and existing software for robotic
`systems (Cote et al. 2006). NUANCE speech recognition
`software is also available through MARIE.
`
`3http://robotics.usc.edu/ ahoward/pmap/
`4http://cslu.cse.ogi.edu/toolkit/
`
`Page 11 of 29
`
`

`

`Figure 3: Spartacus (front view, back view), with the micro-
`phones located near the *.
`
`Figure 4: Spartacus software architecture.
`
`Figure 5: Track audio interface.
`
`Regarding the auditory and interaction capabilities of the
`robot, we came up with a set of modalities to improve
`our understanding of integrated modalities (vision, audition,
`graphical, navigation), as follows:
`• Visualization of the SSLTS results. In 2005, the SSLTS
`system was implemented on Spartacus and was directing
`the camera toward the loudest sound source around the
`robot. To fully represent what the robot is hearing, we
`needed a more appropriate interface. Figure 5 illustrates
`the interface developed. The upper part shows the angle
`of the perceived sound sources around the robot, in rela-
`tion to time. The interface shows in real-time the sound
`sources perceived, using dots of a distinct color. The
`sound sources are saved, and the user can select them and
`play back what the SSLTS generated. This interface re-
`veals to be very valuable for explaining what the robot can
`hear, to construct a database of typical audio streams in
`a particular setting and to analyze the performance of the
`entire audio block (allowing to diagnose potential difficul-
`ties that the speech recognition module has to face). More
`than 6600 audio streams were recorded over the twelve
`hours of operation of the robot at the AAAI 2006 con-
`ference, held at Seaport Hotel and World Trade Center in
`Boston. By constructing such database of audio streams,
`we will be able to evaluate in a more diverse set of condi-
`tions the performance of speech recognition algorithms.
`• Develop contextual interfaces and interaction capabilities
`according to the state and intentions of the robot.
`It is
`sometimes difficult to know exactly which task is priori-
`tized by the robot at a given time, making the users un-
`able to know what the robot is actually doing and eval-
`uate its performance. So we made the robot indicate its
`intention verbally and also graphically (in case, for what-
`ever reasons, the interlocutor is not able to understand the
`message). Figure 6 illustrates the case where Spartacus is
`requesting assistance to find a location in the convention
`center.
`The graphical interfaces of the robot were made so that
`users can indicate through push buttons (using the touch
`screen) what they want Spartacus to do. We also made it
`possible for the users to make these requests verbally. To
`facilitate the recognition process, a specific grammar and
`dialogue manager are loaded in CSLU for each graphical
`
`Page 12 of 29
`
`

`

`sue on Software Development and Integration in Robotics
`3(1):55–60.
`Gockley, R.; Simmons, R.; Wang, J.; Busquets, D.; DiS-
`alvo, C.; Caffrey, K.; Rosenthal, S.; Mink, J.; Thomas, S.;
`Adams, W.; Lauducci, T.; Bugajska, M.; Perzanowski, D.;
`and Schultz, A. 2004. Grace and george: Social robots
`at aaai. Technical Report WS-04-11, AAAI Mobile Robot
`Competition Workshop. pp. 15-20.
`Maxwell, B.; Smart, W.; Jacoff, A.; Casper, J.; Weiss, B.;
`Scholtz, J.; Yanco, H.; Micire, M.; Stroupe, A.; Stormont,
`D.; and Lauwers, T. 2004. 2003 aaai robot competition and
`exhibition. AI Magazine 25(2):68–80.
`Michaud, F.; Brosseau, Y.; Cote, C.; Letourneau, D.;
`Moisan, P.; Ponchon, A.; Raievsky, C.; Valin, J.-M.;
`Beaudry, E.; and Kabanza, F. 2005. Modularity and in-
`tegration in the design of a socially interactive robot.
`In
`Proceedings IEEE International Workshop on Robot and
`Human Interactive Communication, 172–177.
`Michaud, F.; Duquette, A.; and Nadeau, I. 2003. Character-
`istics of mobile robotic toys for children with pervasive de-
`velopmental disorders. In Proceedings IEEE International
`Conference on Systems, Man, and Cybernetics.
`Montemerlo, M.; Roy, N.; and Thrun, S. 2003. Perspec-
`tives on standardization in mobile robot programming: The
`carnegie mellon navigation (carmen) toolkit. In Proceed-
`ings IEEE/RSJ International Conference on Intelligent Ro-
`bots and Systems, 2436–2441.
`Simmons, R.; Goldberg, D.; Goode, A.; Montemerlo, M.;
`Roy, N.; Sellner, B.; Urmson, C.; Schultz, A.; Abramson,
`M.; Adams, W.; Atrash, A.; Bugajska, M.; Coblenz, M.;
`MacMahon, M.; Perzanowski, D.; Horswill, I.; Zubek, R.;
`Kortenkamp, D.; Wolfe, B.; Milam, T.; and Maxwell, B.
`2003. Grace : An autonomous robot for the aaai robot
`challenge. AI Magazine 24(2):51–72.
`Smart, W. D.; Dixon, M.; Melchior, N.; Tucek, J.; and
`Srinivas, A. 2003. Lewis the graduate student: An en-
`try in the aaai robot challenge. Technical report, AAAI
`Workshop on Mobile Robot Competition. p. 46-51.
`Valin, J.-M.; Michaud, F.; Hadjou, B.; and Rouat, J. 2004.
`Localization of simultaneous moving sound sources for
`mobile robot using a frequency-domaine steered beam-
`former approach. In Proceedings IEEE International Con-
`ference on Robotics and Automation, 1033–1038.
`Valin, J.-M.; Michaud, F.; and Rouat, J. 2006. Robust
`3d localization and tracking of sound sources using beam-
`forming and particle filtering. In Proceedings International
`Conference on Audio, Speech and Signal Processing, 221–
`224.
`Valin, J.-M.; Rouat, J.; and Michaud, F. 2004. Enhanced
`robot audition based on microphone array source separa-
`tion with post-filter. In Proceedings of the IEEE/RSJ Inter-
`national Conference on Intelligent Robots and Systems.
`Vaughan, R. T.; Gerkey, B. P.; and Howard, A. 2003. On
`device abstractions for portable, reusable robot code.
`In
`Proceedings IEEE/RSJ International Conference on Intel-
`ligent Robots and Systems, 2421–2427.
`
`Figure 6: Graphical display for an assistance request to find
`a location.
`
`mode. The audio and graphical interaction are therefore
`tighly integrated together.
`These modalities are functional and were demonstrated at
`AAAI 2006 Conference, showing our implementation of in-
`tegrated and dynamic representations of localization, vocal
`requests and menus. However, it is only a first step in com-
`ing up with the most efficient integration of these modal-
`ities. Our intents are to explore what can be learned us-
`ing this integration, and to use our setup to evaluate speech
`recognition (e.g., NUANCE, Sphinx) and dialogue manage-
`ment (e.g., CSLU, Collagen) packages, to study how they
`can be interfaced to the visual and mobile capabilities of
`the robot, and to assess how a stronger coupling between
`SSLTS and speech recognition algorithms can help improve
`speech recognition. We are also interested in adding new
`aural capabilities (e.g., speaker identification, sound recog-
`nition), and evaluate how the robot could be able to listen
`while speaking.
`
`Acknowledgments
`F. Michaud holds the Canada Research Chair (CRC) in Mo-
`bile Robotics and Autonomous Intelligent Systems. Support
`for this work is provided by the Natural Sciences and Engi-
`neering Research Council of Canada, the Canada Research
`Chair program and the Canadian Foundation for Innovation.
`
`References
`Arulampalam, M.; Maskell, S.; Gordon, N.; and Clapp, T.
`2002. A tutorial on particle filters for online nonlinear/non-
`gaussian bayesian tracking. IEEE Transactions on Signal
`Processing 50(2):174–188.
`Cote, C.; Letourneau, D.; Michaud, F.; Valin, J.-M.;
`Brosseau, Y.; Raievsky, C.; Lemay, M.; and Tran, V. 2004.
`Code reusability tools for programming mobile robots. In
`Proceedings of the IEEE/RSJ International Conference on
`Intelligent Robots and Systems, 1820–1825.
`Cote, C.; Brosseau, Y.; Letourneau, D.; Raievsky, C.; and
`Michaud, F. 2006. Using marie in software development
`and integration for autonomous mobile robotics. Interna-
`tional Journal of Advanced Robotic Systems, Special Is-
`
`Page 13 of 29
`
`

`

`jnoqe|yeas|asmoiq|awoyTHNLSo8Vasso1Hs
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`SL-||‘S4auUa}st]UPLUNYJeENpIAIpU]405Suol}e}UaSa1gYDVadg§II}OqGoYSAOAdUI|OFBuIyeyaatIedsiagAroypnyJoJdasu07ayyBuisy
`
`
`
`
`
`
`
`
`
`
`9007SaidWV'LO-90-S4HOdayjPIIULIALIVWWV"9002‘SL-EL4890930‘WSN‘Dd‘UOWuUIYseM‘UINIsoduAs|je4
`
`
`
`
`
`
`
`
`
`-PNEUIIIASIOSUBIY"UIIBADIB|A|-UBSl‘SVeUISs4SLUIXP‘Nesuino]JUILUOG‘49gUOLUIISBs+TFElii
`
`
`
`
`
`
`
`
`LZ-G|"SJOGOYSfIqo]UlUOIEIYISSE]}JUSLUUIOIAUII9SNODyJO)sIsAjeuyJUa}UOD
`
`
`
`G-|*S}JOqoY10jJayJOsuoIssaidxyjequan-uOKJayIOpueAJOJIpny
`a-.>~ae
`
`
`
`
`
`O

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket