throbber
TOMMANDATI U NI MIMI KUTOA MAONI
`
`US009884631B2
`
`( 12 ) United States Patent
`James et al .
`
`( 10 ) Patent No . :
`( 45 ) Date of Patent :
`
`US 9 , 884 , 631 B2
`Feb . 6 , 2018
`
`( 54 ) TRANSITIONING BETWEEN OPERATIONAL
`MODES OF AN AUTONOMOUS VEHICLE
`Applicants : Toyota Motor Engineering &
`Manufacturing North America , Inc . ,
`Erlanger , KY ( US ) ; Toyota Jidosha
`Kabushiki Kaisha , Toyota - shi ( JP )
`( 72 ) Inventors : Michael R . James , Northville , MI
`( US ) ; Katsuhiro Sakai , Ann Arbor , MI
`( US ) ; Toshiki Kindo , Yokohama ( JP ) ;
`Danil V . Prokhorov , Canton , MI ( US ) ;
`Masahiro Harada , Novi , MI ( US )
`( 73 ) Assignees : Toyota Motor Engineering &
`Manufacturing North America , Inc . ,
`Erlanger , KY ( US ) ; Toyota Jidosha
`Kabushiki Kaisha , Toyota - shi ,
`Aichi - ken ( JP )
`Subject to any disclaimer , the term of this
`patent is extended or adjusted under 35
`U . S . C . 154 ( b ) by 40 days .
`( 21 ) Appl . No . : 14 / 730 , 570
`( 22 ) Filed :
`Jun . 4 , 2015
`Prior Publication Data
`( 65 )
`US 2016 / 0355192 A1 Dec . 8 , 2016
`Int . Ci .
`( 2012 . 01 )
`B60W 50 / 08
`B60W 50 / 14
`( 2012 . 01 )
`( Continued )
`( 52 ) U . S . CI .
`8
`CPC . . . . . . . . . . . B60W 50 / 082 ( 2013 . 01 ) ; B60R 11 / 04
`( 2013 . 01 ) ; B60W 40 / 08 ( 2013 . 01 ) ; B60W
`50 / 14 ( 2013 . 01 ) ;
`( Continued )
`( 58 ) Field of Classification Search
`CPC . . . . . . . . . GO5D 1 / 00 ; G05D 1 / 02 ; GO5D 1 / 0088 ;
`GO5D 1 / 0289 ; GO5D 1 / 0259 ;
`( Continued )
`
`( * ) Notice :
`
`( 51 )
`
`( 71 )
`
`( 56 )
`
`B60Q 1 / 46
`340 / 471
`HO4N 7 / 183
`348 / E7 . 086
`
`References Cited
`U . S . PATENT DOCUMENTS
`4 , 346 , 365 A *
`8 / 1982 Ingram . . . . . . . . . . . . .
`5 , 877 , 897 A *
`3 / 1999 Schofield . . . . . . . . . .
`( Continued )
`OTHER PUBLICATIONS
`National Highway Traffic Safety Administration ; “ U . S . Department
`of Transportation Releases Policy on Automated Vehicle Develop
`ment ” ; May 30 , 2013 ; [ retrieved Jun . 3 , 2015 ) ; retrieved from the
`Internet :
`< http : / / www . nhtsa . gov / About + NHTSA / Press + Releases /
`U . S . + Department + of + Transportation + Releases + Policy + on +
`Automated + Vehicle + Development > ( 2 pages ) .
`National Highway Traffic Safety Administration ; “ National High
`way Traffic Safety Administration Preliminary Statement of Policy
`Concerning Automated Vehicles ” ; 2013 ; ( 14 pages ) .
`( Continued )
`Primary Examiner – Yonel Beaulieu
`Assistant Examiner — Martin Weeks
`( 74 ) Attorney , Agent , or Firm — Christopher G . Darrow ;
`Darrow Mustafa PC
`( 57 )
`ABSTRACT
`Arrangements relating to the transitioning of a vehicle
`between operational modes are described . The vehicle can
`transition between a first operational mode and a second
`operational mode . The second operational mode has a
`greater degree of manual involvement than the first opera
`tional mode . For instance , the first operational mode can be
`an unmonitored autonomous operational mode , and the
`second operational mode can be a monitored autonomous
`operational mode or a manual operational mode . It can be
`determined whether an operational mode transition event
`has occurred while the vehicle is operating in the first
`operational mode . In response to determining that an opera
`tional mode transition event has occurred , a time buffer for
`continuing in the first operational mode before switching to
`the second operational mode can be determined . A transition
`alert can be presented within the vehicle . The transition alert
`can represent the determined time buffer .
`23 Claims , 3 Drawing Sheets
`
`103
`
`101
`
`115
`
`Data Stone ( 3 )
`111
`Map Data
`Friving Scenes112
`Scene Markens
`113
`
`Sensor System
`Environnent
`Sensor ( s )
`
`125
`
`1100
`
`Output
`System
`
`Actuator ( s )
`Vehicle Systeme
`Propulsion System
`
`Braking System
`
`Stoering System
`
`Throttle System
`
`Transmission System
`
`Signaling System
`
`Navigation System
`
`145
`
`- 150
`155
`161
`
`165
`
`- 180
`
`Processor ( s )
`
`Autonomous
`Driving
`Module ( s )
`
`- - 130
`
`Imut
`System
`
`Transition Event
`Detection
`Madule
`
`137
`
`2012
`
`- 191
`
`Time Buller
`Deterrnination
`Module
`
`Display ( 3 )
`
`Speaker ( s )
`
`Haptic
`Actuators )
`
`102 _
`
`104
`
`IPR2025-00943
`Tesla EX1021 Page 1
`
`

`

`US 9 , 884 , 631 B2
`Page 2
`
`( 52 )
`
`( 56 )
`
`( 2006 . 01 )
`( 2012 . 01 )
`( 2012 . 01 )
`
`( 51 ) Int . CI .
`BOOR 11 / 04
`B60W 50 / 16
`B60W 40 / 08
`U . S . CI .
`CPC . . . . . B60W 50 / 16 ( 2013 . 01 ) ; B60W 2040 / 0818
`( 2013 . 01 ) ; B60W 2050 / 143 ( 2013 . 01 ) ; B60W
`2050 / 146 ( 2013 . 01 ) ; B60W 2420 / 42 ( 2013 . 01 ) ;
`B60W 2540 / 04 ( 2013 . 01 ) ; B60W 2550 / 402
`( 2013 . 01 )
`( 58 ) Field of Classification Search
`CPC . . . . GO5D 1 / 0246 ; GO5D 1 / 088 ; G05D 1 / 0061 ;
`B60W 50 / 08 ; B60W 50 / 14 ; B60W 50 / 16 ;
`B60W 40 / 08 ; B60W 2050 / 146 ; B6OR
`11 / 04
`See application file for complete search history .
`References Cited
`U . S . PATENT DOCUMENTS
`8 , 618 , 922 B2 12 / 2013 Debouk et al .
`2013 / 0131907 A1 5 / 2013 Green et al .
`2014 / 0244096 AL
`8 / 2014 An et al .
`2014 / 0303827 Al 10 / 2014 Dolgov et al .
`2016 / 0357186 A1 *
`12 / 2016 Dias . . . . . .
`
`. . . GO5D 1 / 0061
`
`OTHER PUBLICATIONS
`Kamata ; “ Background Map Format for Autonomous Driving " ; U . S .
`Appl . No . 14 / 574 , 151 , filed Dec . 17 , 2014 .
`* cited by examiner
`
`IPR2025-00943
`Tesla EX1021 Page 2
`
`

`

`U . S . Patent
`
`Feb . 6 , 2018
`
`Sheet 1 of 3
`
`US 9 , 884 , 631 B2
`
`FIG . 1
`FIG . 1
`
`103 _ ro
`
`103
`
`101
`
`- 115
`
`Data Store ( s )
`Map Data
`Driving Scenes
`Scene Markers
`
`Sensor System
`Environment
`Sensoris
`
`Processor ( s )
`
`Autonomous
`Driving
`Module ( s )
`
`- 130
`
`\ 125
`- 126
`
`131
`
`100
`
`106
`
`145
`
`T
`

`
`1052
`
`Input
`System
`
`Transition Event
`Detection
`Module
`
`Time Buffer
`Determination
`Module
`
`- 300
`
`Display ( s )
`
`Speaker ( s )
`
`Haptic
`Actuator ( s )
`
`Output
`System
`
`Actuator ( s ) 7140
`
`Vehicle Systems
`Propulsion System
`
`Braking System
`
`Steering System
`
`Throttle System
`
`Transmission System
`
`Signaling System
`
`Navigation System
`
`180
`
`102 _
`
`104
`
`IPR2025-00943
`Tesla EX1021 Page 3
`
`

`

`U . S . Patent
`
`Feb . 6 , 2018
`
`Sheet 2 of 3
`
`US 9 , 884 , 631 B2
`
`200
`
`Determining whether an operational mode transition
`event has occurred while a vehicle is operating
`in a first operational mode
`
`Responsive to determining that an operational mode
`transition event has occurred , determining a time
`buffer for continuing in the first operational mode
`before switching to a second operational mode
`
`210
`
`220
`
`Presenting a transition alert within the vehicle , the
`transition alert corresponding to the determined
`time buffer
`
`h
`
`230
`
`FIG . 2
`
`IPR2025-00943
`Tesla EX1021 Page 4
`
`

`

`U . S . Patent
`
`Feb . 6 , 2018
`
`Sheet 3 of 3
`
`US 9 , 884 , 631 B2
`
`310
`
`30
`
`- 300
`
`? ????
`
`FIG . 3
`
`????
`???? ( Time Left , ?
`00 : 10 ????
`
`315
`
`305
`
`- 300
`
`Minutes : Seconds
`
`FIG . 4
`
`IPR2025-00943
`Tesla EX1021 Page 5
`
`

`

`US 9 , 884 , 631 B2
`
`TRANSITIONING BETWEEN OPERATIONAL
`MODES OF AN AUTONOMOUS VEHICLE
`
`DETAILED DESCRIPTION
`
`FIG . 2 is an example of a method of transitioning a
`vehicle between a first operational mode and a second
`operational mode , wherein the second operational mode has
`a greater degree of manual involvement than the first opera
`5 tional mode .
`FIG . 3 is an example of a transition alert presented on a
`display , wherein the transition alert is a bar graph .
`FIG . 4 is another example of a transition alert presented
`on a display , wherein the transition alert is a countdown
`10 clock .
`
`FIELD
`The subject matter described herein relates in general to
`vehicles that have a plurality of operational modes including
`an autonomous operational mode and , more particularly , to
`the transitioning between different operational modes .
`BACKGROUND
`Some vehicles are configured to operate in a plurality of
`operational modes . An example of an operational mode is
`This detailed description relates to the transitioning of an
`one in which a computing system is used to navigate and / or
`maneuver the vehicle along a travel route with minimal or no 15 autonomous vehicle from
`a
`first operational mode to
`a
`second operational mode . This detailed description is more
`input from a human driver . Such vehicles are equipped with
`sensors that are configured to detect information about the
`particularly related to instances in which the second opera
`tional mode has a greater degree of manual involvement
`surrounding environment , including the presence of objects
`in the environment . The detected information can be sent to
`than the first operational mode . In response to determining
`the computing system . Other operational modes can include 20 that an operational mode transition event has occurred , a
`luding a manual mode in time buffer for continuing in the first operational mode
`different levels of human input , including a manual mode in
`which a human driver is responsible for navigating and / or
`before switching to the second operational mode can be
`determined . A transition alert corresponding to the deter
`maneuvering the vehicle through the surrounding environ
`ment . Vehicles with a plurality of operational modes are
`mined time buffer can be presented . The present detailed
`configured to allow switching between the various opera - 25 description relates to systems , methods and computer pro
`gram products that incorporate such features . In at least
`tional modes .
`some instances , such systems , methods and computer pro
`gram products can improve performance and / or safety of an
`SUMMARY
`autonomous vehicle .
`Detailed embodiments are disclosed herein ; however , it is
`In one respect , the present disclosure is directed to a 30
`to be understood that the disclosed embodiments are
`method of transitioning a vehicle between a first operational
`intended only as exemplary . Therefore , specific structural
`mode and a second operational mode . The second opera -
`and functional details disclosed herein are not to be inter
`tional mode has a greater degree of manual involvement
`preted as limiting , but merely as a basis for the claims and
`than the first operational mode . The method can include
`determining whether an operational mode transition event 35 as a representative basis for teaching one skilled in the art to
`has occurred while the vehicle is operating in the first
`variously employ the aspects herein in virtually any appro
`operational mode . The method can also include , responsive
`priately detailed structure . Further , the terms and phrases
`to determining that an operational mode transition event has
`used herein are not intended to be limiting but rather to
`occurred , determining a time buffer for continuing in the first
`provide an understandable description of possible imple
`operational mode before switching to the second operational 40 mentations . Various embodiments are shown in FIGS . 1 - 4 ,
`mode . The method can further include presenting a transi -
`but the embodiments are not limited to the illustrated
`tion alert within the vehicle . The transition alert can corre -
`structure or application .
`It will be appreciated that for simplicity and clarity of
`spond to the determined time buffer .
`In another respect , the present disclosure is directed to a
`illustration , where appropriate , reference numerals have
`system for transitioning a vehicle between a first operational 45 been repeated among the different figures to indicate corre
`mode and a second operational mode . The second opera -
`sponding or analogous elements . In addition , numerous
`tional mode has a greater degree of manual involvement
`specific details are set forth in order to provide a thorough
`than the first operational mode . The system can include a
`understanding of the embodiments described herein . How
`user interface located within the vehicle . The system can
`ever , it will be understood by those of ordinary skill in the
`also include a processor operatively connected to the sensor 50 art that the embodiments described herein can be practiced
`system . The processor can be programmed to initiate execut
`without these specific details .
`able operations . The executable operations can include
`Referring to FIG . 1 , an example a vehicle 100 is shown .
`determining whether an operational mode transition event
`As used herein , " vehicle ” means any form of motorized
`has occurred while the vehicle is operating in the first
`transport . In one or more implementations , the vehicle 100
`operational mode . The executable operations can also 55 can be an automobile . While arrangements will be described
`include , responsive to determining that an operational mode
`herein with respect to automobiles , it will be understood that
`transition event has occurred , determining a time buffer for
`embodiments are not limited to automobiles . In some imple
`continuing in the first operational mode before switching to
`mentations , the vehicle 100 may be a watercraft , an aircraft
`the second operational mode . The executable operations can
`or any other form of motorized transport .
`further include presenting a transition alert within the 60
`According to arrangements herein , the vehicle 100 can be
`vehicle , the transition alert corresponding to the determined
`an autonomous vehicle . As used herein , “ autonomous
`vehicle ” means a vehicle that configured to operate in an
`time buffer .
`autonomous mode . “ Autonomous mode ” means that one or
`BRIEF DESCRIPTION OF THE DRAWINGS
`more computing systems are used to navigate and / or maneu
`65 ver the vehicle along a travel route with minimal or no input
`FIG . 1 is an example of an autonomous vehicle configured
`from a human driver . In one or more arrangements , the
`to transition between different operational modes .
`vehicle 100 can be highly automated .
`
`IPR2025-00943
`Tesla EX1021 Page 6
`
`

`

`US 9 , 884 , 631 B2
`The vehicle 100 can have a plurality of operational
`vehicle 100 to a stop , keeping the vehicle 100 stopped , or
`modes . For instance , the vehicle 100 can have an unmoni -
`having the vehicle 100 take the next exit on a highway , just
`tored autonomous operational mode . “ Unmonitored autono -
`to name a few possibilities .
`mous operational mode ” means that one or more computing
`The vehicle 100 can be configured to be switched between
`systems are used to navigate and / or maneuver the vehicle 5 the various operational modes . Such switching can be imple
`mented in any suitable manner , now known or later devel
`along a travel route with no input or supervision required
`oped . The switching can be performed automatically , or it
`from a human driver . As an example , the unmonitored
`can be done responsive to receiving a manual input or
`autonomous operational mode can include Level 4 ( L4 ) , as
`request . It will be appreciated that the switching between the
`defined by the National Highway Traffic Safety Adminis
`its Preliminary Statement of Policy Concerning 2010 operational modes can raise concerns about reliable transfer
`tration in
`of operational authority .
`Automated Vehicles ( May 30 , 2013 ) ( “ NHTSA 2013
`In one or more arrangements , the switching can be done
`Policy ” ) , which is incorporated herein by reference . The
`from a first operational mode to a second operational mode .
`vehicle 100 can have a monitored autonomous operational
`In some instances , the second operational mode can have a
`mode . “ Monitored autonomous operational mode ” means
`means 15 greater degree of manual involvement than the first opera
`that one or more computing systems are used to navigate
`tional mode . A " greater degree of manual involvement ”
`and / or maneuver the vehicle with at least some human driver
`means that a human driver is required to or should increase
`his or her level of supervision and / or input with respect to
`supervision required . As an example , the monitored autono -
`mous operational mode can include Level 3 or L3 , as defined
`the control of at least the navigation and / or maneuvering of
`by the NHTSA 2013 Policy . In some instances , when the 20 the vehicle . One example of when the second operational
`vehicle 100 is in a monitored autonomous operational mode ,
`mode can have a greater degree of manual involvement than
`a signal ( e . g . , an audial signal , a visual signal , a haptic
`the first operational mode is when the first operational mode
`signal , etc . ) can be presented to a human driver to take an
`is an unmonitored autonomous operational mode and the
`action within a predetermined amount of time . If such action
`second operational mode is a monitored autonomous opera
`is not taken within the predetermined amount of time , one or 25 tional mode or a manual operational mode . Another example
`is when the first operational mode is a monitored operational
`more safety maneuvers can be implemented
`mode and the second operational mode is a semi - autono
`The vehicle 100 can have one or more semi - autonomous
`mous operational mode or a manual operational mode .
`operational modes . “ Semi - autonomous operational mode "
`In some instances , the second operational mode can have
`means that a portion of the navigation and / or maneuvering
`30 a lesser degree of manual involvement than the first opera
`of the vehicle along a travel route is performed by one or
`tional mode . A " lesser degree of manual involvement "
`more computing systems , and a portion of the navigation
`means that a human driver can decrease his or her level of
`and / or maneuvering of the vehicle along a travel route is
`supervision and / or input with respect to the control of at
`performed by a human driver . As an example , the semi
`least the navigation and / or maneuvering of the vehicle . One
`autonomous operational mode can include Levels 2 ( L2 ) 22 ) 35 example of when the second operational mode can have a
`and / or Level 1 ( L1 ) , as defined by the NHTSA 2013 Policy .
`lesser degree of manual involvement than the first opera
`One example of a semi - autonomous operational mode is
`tional mode is when the first operational mode is a manual
`when an adaptive cruise control system is activated . In such
`operational mode , and the second operational mode is
`a
`case , the speed of a vehicle can be automatically adjusted to
`semi - autonomous operational mode , a monitored autono
`maintain a safe distance from a vehicle ahead based on data 40 mous operational mode , or an unmonitored autonomous
`received from on - board sensors , but the vehicle is otherwise
`operational mode . Another example is when the first opera
`operated manually by a human driver . Upon receiving a
`tional mode is a monitored operational mode , and the second
`driver input to alter the speed of the vehicle ( e . g . by
`operational mode is an unmonitored operational mode .
`depressing the brake pedal to reduce the speed of the
`The vehicle 100 can have a forward end 101 and a
`vehicle ) , the adaptive cruise control system is deactivated 45 rearward end 102 . The vehicle 100 can have an associated
`longitudinal axis 103 , which can be the central axis of the
`and the speed of the vehicle is reduced .
`The vehicle 100 can have a manual operational mode .
`vehicle 100 . The vehicle 100 can have an associated longi
`“ Manual operational mode ” means that a substantial major -
`tudinal direction 104 . “ Longitudinal direction " means any
`ity or all of the navigation and / or maneuvering of the vehicle
`direction that is substantially parallel to and / or co - linear
`along a travel route is performed by a human driver with 50 with the longitudinal axis 103 . The vehicle 100 can have an
`minimal or no input from a computing system . As an
`associated lateral axis 105 , which can be substantially per
`example , the manual operational mode can include Level (
`pendicular to the longitudinal axis 103 . As used herein , the
`( LO ) , as defined by the NHTSA 2013 Policy .
`term “ substantially ” includes exactly the term
`it modifies
`The vehicle 100 can have a special operational mode .
`and slight variations therefrom . Thus , the term “ substantially
`“ Special operational mode ” means that , if a requested 55 perpendicular ” means exactly perpendicular and slight
`human driver action is not taken or confirmed within a
`variations therefrom . In this particular example , slight varia
`predetermined amount of time , the navigation and / or
`tions therefrom can include within normal manufacturing
`maneuvering of the vehicle can be controlled by one or more
`tolerances , within about 10 degrees or less , within about 5
`computing systems to implement one or more safety maneu -
`degrees or less , within about 4 degrees or less , within about
`vers . The safety maneuver can be a predetermined safety 60 3 degrees or less , within about 2 degrees or less , or within
`maneuver based on the current driving environment . For
`about 1 degree or less . The vehicle 100 can have an
`instance , if a human driver does not take control of the
`associated lateral direction 106 . “ Lateral direction ” means
`vehicle 100 within a predetermined amount of time , the
`any direction that is substantially parallel to and / or co - linear
`safety maneuver may include moving the vehicle 100 to the
`with the lateral axis 105 .
`side of the road , moving the vehicle 100 onto the shoulder 65
`The vehicle 100 can include various elements , some of
`of the road , reducing the speed of the vehicle 100 , turning
`which may be a part of an autonomous driving system . Some
`the vehicle 100 into the nearest parking lot , bringing the
`of the possible elements of the vehicle 100 are shown in FIG .
`
`IPR2025-00943
`Tesla EX1021 Page 7
`
`

`

`US 9 , 884 , 631 B2
`
`5
`
`driving scenes ” is defined as one or more driving scenes .
`1 and will now be described . It will be understood that it is
`“ Driving scenes ” means sensor system data of a location
`not necessary for the vehicle 100 to have all of the elements
`within a geographic area . As an example , the driving scenes
`shown in FIG . 1 or described herein . The vehicle 100 can
`can be images or videos . The driving scenes can include any
`have any combination of the various elements shown in FIG .
`suitable sensor system data of a road , an intersection ,
`1 . Further , the vehicle 100 can have additional elements to
`buildings , structures , traffic control devices , lane markers ,
`those shown in FIG . 1 . In some arrangements , vehicle 100
`landmarks , features . In some instances , the set of driving
`may not include one or more of the elements shown in FIG .
`scenes 112 can be located onboard the vehicle 100 . Alter
`1 . Further , while the various elements are shown as being
`natively , at least a portion of the set of driving scenes 112
`located within the vehicle 100 in FIG . 1 , it will be under
`stood that one or more of these elements can be located 10 can be located remote from the vehicle 100 .
`external to the vehicle 100 . Further , the elements shown may
`In one or more arrangements , the one or more data stores
`be physically separated by large distances .
`115 can include a set of scene markers 113 . The term “ set of
`The vehicle 100 can include one or more processors 110 .
`scene markers ” is defined as one or more scene markers . A
`“ Processor ” means any component or group of components
`" scene marker ” is an object or feature of interest located in
`that are configured to execute any of the processes described 15 and / or describing a driving scene . Examples of scene mark
`herein or any form of instructions to carry out such processes
`ers can include any suitable sensor system data of a road , an
`or cause such processes to be performed . The processor 110
`intersection , buildings , structures , traffic control devices ,
`may be implemented with one or more general - purpose
`lane markers , landmarks , road paint , signs , poles , curbs ,
`and / or one or more special - purpose processors . Examples of
`features . In some instances , the set of scene markers 113 can
`suitable processors include microprocessors , microcon - 20 be located onboard the vehicle 100 . Alternatively , at least a
`trollers , DSP processors , and other circuitry that can execute
`portion of the set of scene markers 113 can be located remote
`software . Further examples of suitable processors include ,
`from the vehicle 100 .
`but are not limited to , a central processing unit ( CPU ) , an
`In one or more arrangements , a priority can be established
`array processor , a vector processor , a digital signal processor
`with respect to the set of scene markers 113 . For instance ,
`( DSP ) , a field - programmable gate array ( FPGA ) , a program - 25 the priority can include sequentially ranking the set of scene
`mable logic array ( PLA ) , an application specific integrated
`markers 113 . Alternatively , the priority can include assign
`circuit ( ASIC ) , programmable logic circuitry , and a control -
`ing a priority level to the set of scene markers 113 . For
`ler . The processor 110 can include at least one hardware
`instance , the priority levels can include high , medium ,
`circuit ( e . g . , an integrated circuit ) configured to carry out
`and / or low . The ranking can be configured by a user or some
`instructions contained in program code . In arrangements in 30 other entity .
`which there is a plurality of processors 110 , such processors
`The vehicle 100 can include an autonomous driving
`can work independently from each other or one or more
`module 120 . The autonomous driving module 120 can be
`processors can work in combination with each other . In one
`implemented as computer readable program code that , when
`or more arrangements , the processor 110 can be a main
`executed by a processor , implement various processes
`processor of the vehicle 100 . For instance , the processor 110 35 described herein , including , for example , determining a
`can be an engine control unit ( ECU ) .
`travel route , implementing the determined travel route ,
`The vehicle 100 can include one or more data stores 115
`determining a modification to a current driving maneuver of
`for storing one or more types of data . The data store 115 can
`the vehicle 100 and / or causing , directly or indirectly , a
`include volatile and / or non - volatile memory . Examples of
`current driving maneuver of the vehicle 100 to be modified .
`suitable data stores 115 include RAM ( Random Access 40 The autonomous driving module 120 can be a component of
`Memory ) , flash memory , ROM ( Read Only Memory ) ,
`the processor 110 , or the autonomous driving module 120
`PROM ( Programmable Read - Only Memory ) , EPROM
`can be executed on and / or distributed among other process
`( Erasable Programmable Read - Only Memory ) , EEPROM
`ing systems to which the processor 110 is operatively
`( Electrically Erasable Programmable Read - Only Memory ) ,
`connected .
`registers , magnetic disks , optical disks , hard drives , or any 45
`The autonomous driving module 120 can include instruc
`other suitable storage medium , or any combination thereof .
`tions ( e . g . , program logic ) executable by the processor 110 .
`The data store 115 can be a component of the processor 110 ,
`Such instructions can include instructions to execute various
`or the data store 115 can be operatively connected to the
`vehicle functions and / or to transmit data to , receive data
`processor 110 for use thereby . The term “ operatively con -
`from , interact with , and / or control the vehicle 100 or one or
`nected , " as used throughout this description , can include 50 more systems thereof ( e . g . one or more of vehicle systems
`direct or indirect connections , including connections without
`145 ) . Alternatively or in addition , the data store 115 may
`direct physical contact .
`contain such instructions .
`In one or more arrangements , the one or more data stores
`As noted above , the vehicle 100 can include a sensor
`115 can include map data 111 . The map data 111 can include
`system 125 . The sensor system 125 can include one or more
`maps of one or more geographic areas or regions . The map 55 sensors . “ Sensor ” means any device , component and / or
`data 111 can include information or data on roads , traffic
`system that can detect , determine , assess , monitor , measure ,
`control devices , structures , features , landmarks in the one or
`quantify and / or sense something . The one or more sensors
`more geographic areas . The map data 111 can be in any
`can be configured to detect , determine , assess , monitor ,
`suitable form . In some instances , the map data 111 can
`measure , quantify and / or sense in real - time . As used herein ,
`include aerial views of an area . In some instances , the map 60 the term “ real - time ” means a level of processing respon
`data 111 can include ground views of an area , including 360
`siveness that a user or system senses as sufficiently imme
`degree ground views . The map data 111 can be highly
`diate for a particular process or determination to be made , or
`detailed . In some instances , the map data 111 can be located
`that enables the processor to keep up with some external
`onboard the vehicle 100 . Alternatively , at least a portion of
`process .
`the map data 111 can be located remote from the vehicle 100 . 65
`In arrangements in which the sensor system 125 includes
`In one or more arrangements , the one or more data stores
`a plurality of sensors , the sensors can work independently
`115 can include a set of driving scenes 112 . The term “ set of
`from each other . Alternatively , two or more of the sensors
`
`IPR2025-00943
`Tesla EX1021 Page 8
`
`

`

`US 9 , 884 , 631 B2
`
`one or more objects in the external environment of the
`can work in combination with each other . In such case , the
`vehicle 100 , the position of each detected object relative to
`two or more sensors can form a sensor network . The sensor
`the vehicle 100 , the distance between each detected object
`system 125 and / or the one or more sensors can be opera -
`and the vehicle 100 in one or more directions ( e . g . in the
`tively connected to the processor 110 , the data store 115 , the
`autonomous driving module 120 and / or other element of the 5 longitudinal direction 104 , the lateral direction 106 and / or
`vehicle 100 and / or the autonomous driving system .
`other direction ( s ) ) , the speed of each detected object and / or
`The sensor system 125 can include any suitable type of
`the movement of each detected object .
`sensor . For example , the sensor system 125 can include one
`In one or more arrangements , one or more of the envi
`or more sensors configured to detect , determine , assess ,
`ronment sensors 126 can use at least in part ultrasound . Such
`monitor , measure , quantify and / or sense information about 10 sensors can include an ultrasound source configured to emit
`the vehicle 100 . Alternatively or in addition , the sensor
`ultrasonic signals and a detector configured to detect reflec
`system 125 can include one or more sensors configured to
`tions of the ultrasonic signal . The one or more ultrasound
`detect , determine , assess , monitor , measure , quantify and / or
`based environment sensors 126 can be configured to detect ,
`sense information about the external environment in which
`determine , assess , monitor , measure , quantify and / or sense ,
`the vehicle 100 is located , including information about 15 directly or indirectly , the presence of one or more objects in
`objects in the external environment . Such objects may be
`the external environment of the vehicle 100 , the position of
`stationary object or moving objects . Alternatively or in
`each detected object relative to the vehicle 100 , the distance
`addition to one or more of the above examples , the sensor
`between each detected object and the vehicle 100 in one or
`system 125 can include one or more sensors configured to
`more directions ( e . g . in the longitudinal direction 104 , the
`detect , determine , assess , monitor , measure , quantify and / or 20 lateral direction 106 and / or other direction ( s ) ) , the speed of
`sense the location of the vehicle 100 and / or the location of
`each detected object and / or the movement of each detected
`objects in the environment relative to the vehicle 100 .
`object . Such detecting can be based on a characteristic ( e . g .
`Various examples of these and other types of sensors will be
`the intensity ) of a reflected ultrasonic signal .
`described herein . It will be understood that t

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket