throbber
(12) United States Patent
`Caveney
`
`(10) Patent No.:
`(45) Date of Patent:
`
`US 9.483,059 B2
`Nov. 1, 2016
`
`US009483059B2
`
`(54) METHOD TO GAIN DRIVER'S ATTENTION
`FOR AUTONOMOUS VEHICLE
`(71) Applicant: Toyota Motor Engineering &
`Manufacturing North America, Inc.,
`Erlanger, KY (US)
`Derek S. Caveney, Plymouth, MI (US)
`(72) Inventor:
`(73) Assignee: Toyota Motor Engineering &
`Manufacturing North America, Inc.,
`Erlanger, KY (US)
`Subject to any disclaimer, the term of this
`patent is extended or adjusted under 35
`U.S.C. 154(b) by 0 days.
`(21) Appl. No.: 14/554,291
`(22) Filed:
`Nov. 26, 2014
`
`(*) Notice:
`
`(65)
`
`Prior Publication Data
`US 2016/O1466 18 A1
`May 26, 2016
`
`(2006.01)
`(2006.01)
`
`(51) Int. Cl.
`GOIC 21/34
`G05D I/06
`(52) U.S. Cl.
`CPC ....................................... G05D I/06 (2013.01)
`(58) Field of Classification Search
`None
`See application file for complete search history.
`
`(56)
`
`References Cited
`
`U.S. PATENT DOCUMENTS
`
`5,813,993 A * 9/1998 Kaplan ................ A61B 5,0476
`600/26
`6,748,302 B2 * 6/2004 Kawazoe ............. GOSD 1.0246
`340/937
`
`3/2010 Beneker et al.
`7,679,496 B2
`8, 2012 Nieves
`8,248,270 B2
`8,554,408 B2 10/2013 Springer et al.
`8,698,639 B2 * 4/2014 Fung ...................... B6OK 28,06
`340,576
`8,909,428 B1* 12/2014 Lombrozo ........... B62D 15,025
`TO1/41
`2004/0164851 A1* 8, 2004 Crawshaw ............. B60Q 9/008
`340/.435
`
`
`
`116,...,
`
`116
`
`116
`
`2004/0262063 A1* 12/2004 Kaufmann ............. B62D 1,286
`180/169
`2005/O115753 A1* 6, 2005 Pemberton ............. G08G 1, 164
`18Of 167
`2008, 008074.0 A1* 4/2008 Kaufmann ............. G08G 1, 167
`382/104
`2008/0183342 A1* 7/2008 Kaufmann ........... B6OK 28,066
`TO1/1
`2008/0270.018 A1* 10, 2008 Citelli ................ B6OK 31,0008
`701,532
`4, 2009 Bolourch ............ B6OK 28,066
`340/.435
`2010/0172542 A1* 7, 2010 Stein .................. GO6K9/00798
`382,103
`
`2009/0091435 A1 *
`
`9/2010 Lee et al.
`2010/0228417 A1
`9/2012 Veen et al.
`2012fO226418 A1
`11/2012 Bobbitt, III
`2012,0283939 A1
`2013,0084847 A1* 4, 2013 TibbittS ................. HO4W 8.245
`455,419
`
`2013/0342365 A1 12/2013 Kiefer et al.
`2014/0309927 A1* 10/2014 Ricci........................ B60Q 1/00
`701 424
`
`* cited by examiner
`Primary Examiner — Truc M Do
`(74) Attorney, Agent, or Firm — Christopher G. Darrow;
`Darrow Mustafa PC
`
`ABSTRACT
`(57)
`A computer-implemented method for the automated driving
`of a vehicle. The method may include coordinating a
`planned vehicle path using a path planner application. The
`path planner application may receive information based on
`inputs to sensors disposed on the vehicle. The method may
`include sending a command to one or more vehicle systems
`to control the vehicle to follow the planned vehicle path.
`While the vehicle follows the planned vehicle path, the
`method may include receiving an indication that the path
`planner application is not meeting a threshold performance
`level. After receiving the indication that the path planner
`application is not meeting the threshold performance level.
`a command is sent to one or more vehicle systems to control
`the vehicle to follow a temporary and irregular full vehicle
`movement to alert a vehicle driver. The temporary and
`irregular full vehicle movement may be a full vehicle
`side-to-side wobbling movement.
`18 Claims, 4 Drawing Sheets
`
`Coordinate a planned vehicle path
`for use with an autonomous vehicle
`
`/n 410
`
`Direct the autonomous wehicle
`to follow the planned vehicle path
`
`/n 420
`
`Monitor and receive information from one or
`more sensors of the autonomous vehicle
`J.
`Periodically determine whether a Path Planner /n 440
`Application meets a threshold performance level
`
`/ra 430
`
`Receive an indication requesting manual
`control of the autonomous vehicle
`
`/n 450
`
`Initiate a temporary and irregular full vehicle /n 460
`movement to alert a driver of the wehicle
`
`IPR2025-00943
`Tesla EX1023 Page 1
`
`

`

`U.S. Patent
`
`Nov. 1, 2016
`
`Sheet 1 of 4
`
`US 9.483,059 B2
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`Exterial Storage
`
`
`
`Operating
`System
`
`114
`
`-10
`
`110
`
`Data Analyzer
`
`Sensors
`
`Path Piafer
`
`Vehicle
`Controle
`
`Applications
`
`Coalpuig Device
`
`Vehicle
`Systems
`
`FIG. 1
`
`IPR2025-00943
`Tesla EX1023 Page 2
`
`

`

`U.S. Patent
`
`Nov. 1, 2016
`
`Sheet 2 of 4
`
`US 9.483,059 B2
`
`rera in re r s
`
`:
`
`be retres in a s r. sent us r is is peg is an ess near sy
`
`
`
`2O2
`
`216
`
`FIG. 2B
`
`204
`
`206 208
`
`IPR2025-00943
`Tesla EX1023 Page 3
`
`

`

`U.S. Patent
`
`Nov. 1, 2016
`
`Sheet 3 of 4
`
`US 9.483,059 B2
`
`
`
`
`
`Z09
`
`IPR2025-00943
`Tesla EX1023 Page 4
`
`

`

`U.S. Patent
`
`Nov. 1, 2016
`
`Sheet 4 of 4
`
`US 9.483,059 B2
`
`
`
`Coordinate a planned vehicle path
`for use With an autonomous Vehicle
`
`Direct the autonomous vehicle
`to follow the planned vehicle path
`
`Monitor and receive information from One Or
`more SenSOrS Of the autonomous vehicle
`
`Periodically determine whether a Path Planner
`Application meets a threshold performance level
`
`Receive an indication requesting manual
`COntrol Of the autonomous Vehicle
`
`Initiate a temporary and irregular full vehicle
`movement to alert a driver Of the Vehicle
`
`410
`
`420
`
`430
`
`440
`
`450
`
`460
`
`FIG. 4
`
`IPR2025-00943
`Tesla EX1023 Page 5
`
`

`

`1.
`METHOD TO GAN DRIVERSATTENTION
`FOR AUTONOMOUS VEHICLE
`
`TECHNICAL FIELD
`
`The present disclosure generally relates to methods for
`autonomous driving and, more particularly, to methods for
`gaining the attention of a driver to indicate a need for manual
`control of a vehicle.
`
`BACKGROUND
`
`The background description provided herein is for the
`purpose of generally presenting the context of the disclo
`sure. Work of the presently named inventors, to the extent it
`may be described in this background section, as well as
`aspects of the description that may not otherwise qualify as
`prior art at the time of filing, are neither expressly nor
`impliedly admitted as prior art against the present technol
`Ogy.
`Partially-automated or monitored driving systems are
`designed to assist drivers in operating a vehicle safely and
`efficiently on the road. For example, they may use tech
`niques such as eye-tracking of the driver to send a warning
`when the driver becomes inattentive, lane tracking of the
`vehicle to send a warning to the driver when the vehicle is
`leaving its lane, and controlling vehicle Velocity based on
`distance to a vehicle ahead of the driver when adaptive
`cruise control is activated by the driver.
`Fully or highly automated driving systems are preferably
`designed to operate a vehicle on a road without driver
`interaction or other external control, for example, in self
`driving or autonomous vehicles. Advanced driver safety
`systems may monitor the situation of a vehicle, including its
`location, as well as a location of other vehicles in its vicinity.
`However, fully automated driving systems are not currently
`designed to notify the driver of upcoming vehicle operations
`in order to prepare the driver in terms of what to expect from
`the automated driving systems control of the vehicle
`While certain systems may provide haptic feedback in the
`form of a vibrating steering wheel, textual warnings within
`the interior of a vehicle, or audible warnings requesting
`attention from the driver, such warnings may, in fact, be
`insufficient to alert the driver. Still further, certain warnings
`may be overbearing or distracting to the point of causing
`alarm on the part of the driver.
`Accordingly, it would be desirable to provide improved
`and reliable warning means to gain the attention of a driver
`of an autonomous vehicle when the need arises for the
`manual control of the vehicle or for another action that needs
`input from the driver.
`
`SUMMARY
`
`10
`
`15
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`This section provides a general Summary of the disclo
`Sure, and is not a comprehensive disclosure of its full scope
`or all of its features.
`In various aspects, the present teachings provide a com
`puter-implemented method for the automated driving of a
`60
`vehicle. The method may include coordinating a planned
`vehicle path using a path planner application, and directing
`the vehicle to follow the planned vehicle path. In various
`aspects, the method includes receiving information from one
`or more sensors of the vehicle. Where it is determined that
`the information from the one or more sensors is not sufficient
`to meet a threshold performance level, the method includes
`
`65
`
`US 9,483,059 B2
`
`2
`initiating a temporary and irregular movement to alert a
`driver of the vehicle, such as a full vehicle side-to-side
`wobbling movement.
`In other aspects, the present teachings provide an auto
`mated driving system for a vehicle. The system may include
`one or more sensors disposed on a vehicle, and a computing
`device in communication with the one or more sensors. The
`computing device may include one or more processors for
`controlling the operations of the computing device, and
`memory for storing data and program instructions used by
`the one or more processors. The one or more processors may
`be configured to execute instructions stored in the memory.
`The system may determine, using a path planner application
`receiving information based on inputs to the one or more
`sensors, a planned vehicle path. The system may send a
`command to one or more vehicle systems to control the
`vehicle to follow the planned vehicle path. Periodically, the
`system may determine whether the path planner application
`is meeting a threshold performance level. If the path planner
`application is not meeting the threshold performance level.
`the system may send a command to the one or more vehicle
`systems to control the vehicle to initiate a temporary and
`irregular full vehicle movement to alert a driver of the
`vehicle.
`In still other aspects, the present teachings provide a
`computer-implemented method for the automated driving of
`a vehicle. The method may include coordinating a planned
`vehicle path using a path planner application. The path
`planner application may receive information based on inputs
`to one or more sensors disposed on the vehicle. The method
`may include sending a command to one or more vehicle
`systems to control the vehicle to follow the planned vehicle
`path. The method may also include monitoring the planned
`vehicle path, and periodically determining whether the path
`planner application is meeting a threshold performance
`level. If the path planner application is not meeting the
`threshold performance level, the method may include send
`ing a command to one or more vehicle systems to control the
`vehicle to initiate a temporary and irregular full vehicle
`movement to alert a driver of the vehicle.
`Further areas of applicability and various methods of
`enhancing mapping technology will become apparent from
`the description provided herein. The description and specific
`examples in this Summary are intended for purposes of
`illustration only and are not intended to limit the scope of the
`present disclosure.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`The present teachings will become more fully understood
`from the detailed description and the accompanying draw
`ings, wherein:
`FIG. 1 is a block diagram of an exemplary computing
`device that may be used in assisting an automated driving
`system;
`FIG. 2A is a schematic illustration of a vehicle including
`the computing device of FIG. 1;
`FIG. 2B is a schematic representation of the vehicle of
`FIG. 2A traveling in a forward direction along a roadway
`containing multiple lanes;
`FIG. 3 illustrates an exemplary portion of a navigation
`route being traversed by the vehicle of FIG. 2A and an
`example planned vehicle path along the portion of the
`navigation route; and
`FIG. 4 is a high-level flow chart illustrating a system and
`method for gaining the attention of a driver of an autono
`mous vehicle.
`
`IPR2025-00943
`Tesla EX1023 Page 6
`
`

`

`US 9,483,059 B2
`
`3
`It should be noted that the figures set forth herein are
`intended to exemplify the general characteristics of materi
`als, methods, and devices among those of the present
`technology, for the purpose of the description of certain
`aspects. These figures may not precisely reflect the charac
`teristics of any given aspect, and are not necessarily intended
`to define or limit specific embodiments within the scope of
`this technology. Further, certain aspects may incorporate
`features from a combination of figures.
`
`DETAILED DESCRIPTION
`
`10
`
`15
`
`4
`control the vehicle to follow the planned vehicle path. In
`certain aspects, if an indication is received that the path
`planner application is not meeting a threshold performance
`level, the automated driving system can be configured to
`provide a notification to a driver of the vehicle to take
`control of the vehicle if the path planner application is not
`meeting the threshold performance level. For example, the
`present technology provides the means for a driver to
`receive timely, non-intrusive instruction from the autono
`mous vehicle that manual control or input will shortly be
`required. As discussed in more detail below, the instruction
`may be provided in the form of a wobbling motion of the
`entire vehicle sufficient to alert or get the attention of a driver
`while minimizing unnecessary alarm on the part of the
`driver. It is envisioned that the present teachings provide
`safety features that improve the overall operation of the
`automated driving system as compared to prior art autono
`mous systems that may simply cancel operation of the
`automated driving system if the path planner application
`does not meet a threshold performance level.
`According to various aspects of the present technology,
`once a system or computing device of the autonomous
`vehicle has been informed or has determined that either the
`path planner is operating below a predetermined threshold
`level, or that information regarding the Surrounding envi
`ronment has been degraded such that autonomous operation
`may not be possible or recommended for much longer, the
`vehicle can be configured to initiate an irregular full vehicle
`moVement.
`In various aspects, the vehicle may be commanded to
`steer in a slight back and forth, or wobbling manner, for a
`very brief period of time. This may provide a slight jiggling
`feeling to the driver in order to indicate that the driver may
`need to take over manual control of the vehicle, or that
`certain vehicle systems may need additional information.
`The irregular full vehicle movement, or wobbling, would
`preferably occur while the vehicle generally remains on its
`planned path. In one example, as will be explained in more
`detail below, lane lines or lane markers of a roadway may be
`occluded or partially removed from the pavement such that
`the vehicle may not be able to accurately recognize lane
`information in order to fully or adequately support autono
`mous driving. In another example, a vehicle controller or
`computing device may determine that certain locating algo
`rithms are taking too long to process, and localization may
`not be occurring at a sufficient interval to accurately place
`the vehicle with respect to its surroundings. In yet another
`example, various systems may rely on the receipt of signals
`from satellites, and it may be determined that the reception
`is becoming sporadic, or the strength, quality, and/or reli
`ability of the signals may be decreasing. In all of the above
`non-limiting examples, the vehicle may still be able to
`provide Sufficient control means, but a degraded quality or
`clarity of the available information may soon require manual
`intervention to ensure future control. Thus in various
`aspects, the present teachings may also utilize the vehicle
`systems, controllers, or computing devices to determine a
`probability that a future threshold performance level can be
`met based on a current ability to analyze data or information
`from one or more vehicle sensors.
`FIG. 1 is a block diagram of an exemplary computing
`device 100 that may be used for implementing the auto
`mated driving system. The computing device 100 can be any
`type of vehicle-installed, handheld, desktop, or other form of
`single computing device, or can be composed of multiple
`computing devices. The processing unit in the computing
`device can be a conventional central processing unit (CPU)
`
`The following description is merely illustrative in nature
`and is in no way intended to limit the disclosure, its
`application, or uses. As used herein, the phrase at least one
`of A, B, and C should be construed to mean a logical (A or
`B or C), using a non-exclusive logical “or.” It should be
`understood that the various steps within a method may be
`executed in different order without altering the principles of
`the present disclosure. Disclosure of ranges includes disclo
`Sure of all ranges and Subdivided ranges within the entire
`range.
`The headings (such as “Background” and "Summary”)
`and Sub-headings used herein are intended only for general
`organization of topics within the present disclosure, and are
`not intended to limit the disclosure of the technology or any
`aspect thereof. The recitation of multiple embodiments
`having stated features is not intended to exclude other
`embodiments having additional features, or other embodi
`ments incorporating different combinations of the stated
`features.
`As used herein, the terms “comprise' and “include” and
`their variants are intended to be non-limiting, Such that
`recitation of items in Succession or a list is not to the
`exclusion of other like items that may also be useful in the
`devices and methods of this technology. Similarly, the terms
`“can and “may” and their variants are intended to be
`non-limiting, Such that recitation that an embodiment can or
`may comprise certain elements or features does not exclude
`other embodiments of the present technology that do not
`contain those elements or features.
`The broad teachings of the present disclosure can be
`implemented in a variety of forms. Therefore, while this
`disclosure includes particular examples, the true scope of the
`disclosure should not be so limited since other modifications
`will become apparent to the skilled practitioner upon a study
`of the specification and the following claims. Reference
`herein to one aspect, or various aspects means that a
`particular feature, structure, or characteristic described in
`connection with an embodiment is included in at least one
`embodiment or aspect. The appearances of the phrase "in
`one aspect' (or variations thereof) are not necessarily refer
`ring to the same aspect or embodiment.
`The present technology generally relates to an automated
`driving system of a vehicle that is able to gain the attention
`of a driver by initiating a temporary and irregular full vehicle
`movement. As used herein, it should be understood that the
`term vehicle should not be construed narrowly, and should
`include all types of vehicles, including a passenger car,
`truck, motorcycle, off-road vehicle, bus, boat, airplane,
`helicopter, lawn mower, recreational vehicle, amusement
`park vehicle, farm vehicle, construction vehicle, tram, golf
`cart, train, or trolley.
`Automated driving systems can be configured to deter
`mine or otherwise follow a planned vehicle path using a path
`planner application and send commands, for example,
`through a vehicle controller, to various vehicle systems to
`
`25
`
`30
`
`35
`
`40
`
`45
`
`50
`
`55
`
`60
`
`65
`
`IPR2025-00943
`Tesla EX1023 Page 7
`
`

`

`US 9,483,059 B2
`
`10
`
`15
`
`25
`
`30
`
`5
`102 or any other type of device, or multiple devices, capable
`of manipulating or processing information. A memory 104
`in the computing device can be a random access memory
`device (RAM) or any other suitable type of storage device.
`The memory 104 can include data 106 that is accessed by the
`CPU 102 using a bus 108.
`The memory 104 can also include an operating system
`110 and installed applications 112, with the installed appli
`cations 112 including programs that permit the CPU 102 to
`perform the automated driving methods described below.
`The computing device 100 can also include secondary,
`additional, or external storage 114, for example, a memory
`card, flash drive, or any other form of computer readable
`medium. The installed applications 112 can be stored in
`whole or in part in the external storage 114 and loaded into
`the memory 104 as needed for processing.
`The computing device 100 can also be in communication
`with one or more sensors 116. The sensors 116 may be
`disposed on a vehicle and can capture data and/or signals for
`processing by an inertial measurement unit (IMU), a lane
`keeping assist (LKA) system, a dead-reckoning system, an
`adaptive cruise control (ACC) system, a global navigation
`satellite system (GNSS) or global positioning system (GPS),
`a light detection and ranging (LIDAR) system, a radar
`system, a Sonar system, an image-based sensor system,
`simultaneous localization and mapping (SLAM), visual
`SLAM (VSLAM), or any other type of system capable of
`capturing information specific to the environment Surround
`ing a vehicle, including information specific to objects Such
`as other vehicles proximate to the navigation route of the
`vehicle, pedestrians, features of the route being traveled by
`the vehicle, landmarks, or other localized position data
`and/or signals and outputting corresponding data and/or
`signals to the CPU 102.
`35
`The sensors 116 can also capture data representative of
`changes in X, y, and Z-axis position, Velocity, acceleration,
`rotation angle, and rotational angular rate for the vehicle and
`similar data for objects or landmarks proximate to the
`navigation route of the vehicle. If the sensors 116 capture
`data for a dead-reckoning system, data relating to wheel
`revolution speeds, travel distance, steering angle, and steer
`ing angular rate of change can be captured. If the sensors 116
`capture signals for a GNSS or GPS, a receiver can calculate
`vehicle position and Velocity estimated in global coordi
`nates. A plurality of satellites can be used to estimate the
`vehicle's position and Velocity using three-dimensional tri
`angulation and time estimation.
`If the sensors 116 capture data for a LIDAR system,
`ranging data relating to intensity or reflectivity returns of the
`environment Surrounding the vehicle can be captured. In
`various examples, the sensors 116 can capture, at least: data
`for a dead-reckoning system, ACC system, or other system
`that estimates vehicle Velocity, acceleration, deceleration,
`position, and orientation; signals for a GNSS or other system
`that determines vehicle position and Velocity; and data for a
`LIDAR system, LKA System, or other system that measures
`vehicle distance from lane lines (e.g., route surface markings
`or route boundaries), obstacles, objects, or other environ
`mental features including traffic lights and road signs. The
`computing device 100 can also be in communication with
`one or more vehicle systems 118, such as vehicle braking
`systems, vehicle steering systems, vehicle propulsion sys
`tems, etc. The vehicle systems 118 can also be in commu
`nication with the sensors 116, the sensors 116 being con
`65
`figured to capture data indicative of performance of the
`vehicle systems 118.
`
`55
`
`6
`With respect to the example computing device 100 as
`described with reference to FIG. 1, the applications 112
`stored in the memory 104 may include at least a data
`analyzer 120, a path planner 122, and a vehicle controller
`124. In general, data captured by the sensors 116 can be used
`by one or more of these applications 112 to understand the
`environment Surrounding the vehicle, plan one or more
`potential vehicle paths for autonomous operation of the
`vehicle along a navigation route for the vehicle, improve
`positional accuracy of the vehicle, and send commands to
`the various vehicle systems 118 to change the current
`operating characteristics of the vehicle.
`FIG. 2A shows a schematic of a vehicle 200 including the
`computing device 100 described in FIG. 1. The computing
`device 100 can be located within the vehicle 200 as shown
`in FIG. 2A or can be located remotely from the vehicle 200
`in an alternate location (not shown). If the computing device
`100 is located remotely from the vehicle 200, the vehicle
`200 can include the capability of communicating with the
`computing device 100.
`The vehicle 200 can also include a plurality of sensors,
`such as the sensors 116 described in reference to FIG.1. One
`or more of the sensors 116 shown can be configured to
`capture changes in Velocity, acceleration, wheel revolution
`speed, and distance to objects within the Surrounding envi
`ronment for use by the computing device 100 to estimate
`position and orientation of the vehicle 200, steering angle for
`a dead-reckoning system, images for processing by an image
`sensor, vehicle position in global coordinates based on
`signals from a plurality of satellites, or any other data and/or
`signals that could be used to determine the current state of
`the vehicle or determine the position of the vehicle 200 in
`respect to its environment.
`For example, if the sensors 116 are configured to capture
`data for use by a LIDAR system, the sensors 116 can capture
`data related to laser returns from physical objects in the area
`Surrounding the vehicle 200 with ranging distances calcu
`lated by measuring the time it takes for a signal to return to
`the sensor 116. Laser returns can include the backscattered
`light reflected by objects hit by a source of light, e.g. laser
`light, being emitted by the sensors 116 or another source on
`or proximate to the vehicle 200. Once the light is reflected
`by an object, the sensors 116 can capture intensity values
`and reflectivity of each point on the object to be used for
`analyzing and classifying the object, for example, by the
`data analyzer 120, one of the applications 112 stored within
`or accessible to the computing device 100.
`The data analyzer 120 briefly described in FIG. 1 can
`analyze data and/or signals captured by the one or more
`sensors 116 by, for example, filtering noise, extracting
`features for clustering, and/or classifying and tracking
`objects. The data analyzer 120 can also process data from the
`one or more sensors 116 such that the data is configured for
`use by the other various applications 112 used to implement
`the automated driving system, Such as the path planner 122.
`In certain aspects, at least one of the sensors 116 can
`capture signals for use with a global navigation satellite
`system, a global positioning system, or a receiver for use
`with one of the systems. As such, the data analyzer 120 may
`be configured to analyze the quality of signals received from
`satellites for use with a GNSS or GPS. The vehicle may
`include one or more processors configured to determine that
`the path planner application is not capable of meeting a
`threshold performance level when the global navigation
`satellite system or a global positioning system fails to
`properly perform or meet predetermined minimum stan
`dards. The vehicle may also have a processor configured to
`
`40
`
`45
`
`50
`
`60
`
`IPR2025-00943
`Tesla EX1023 Page 8
`
`

`

`7
`determine a probability whether a future threshold perfor
`mance level can be met based on a current ability to analyze
`data from the one or more sensors receiving satellite signals.
`In various other aspects, the data analyzer 120 may be
`configured to analyze the quality and clarity of digital
`images that may be obtained from a camera sensor 115. For
`example, the data analyzer may be able to analyze and
`review data or information pertaining to the sharpness,
`contrast, patterns, spatial frequencies, quality, resolution,
`noise, and other features or metrics of objects, particularly
`proximate objects, landmarks, lane marking information,
`and lane markers or lane indicators, which are located within
`an image. When LIDAR, SLAM, or VSLAM systems are
`used, the data analyzer 120 may be configured to ascertain
`the accuracy of the systems to map-match the vehicle 200 to
`the environment. For example, where landmarks 322 (FIG.
`3) in the environment may be used to determine location. In
`various aspects, the vehicle 200 includes at least one pro
`cessor configured to determine that the path planner appli
`cation 122 is not capable of meeting a threshold perfor
`mance level when the LIDAR, SLAM, or VSLAM system
`fails to map-match the vehicle with landmarks according to
`predetermined minimum standards.
`Thus, the data analyzer may not only detect the presence
`or absence of indicators such as lane markings, but may be
`configured to analyze the quality of the indicators or mark
`ings. For example, the quality of data may be monitored
`and/or periodically analyzed in order to detect any occlu
`Sion, degradation, or loss of quality of the indicators that
`may signal a likelihood of a future failure mode or the
`probability that the quality, accuracy, and/or reliability of the
`data may get worse or ultimately not be able to support
`features or systems that may be required for autonomous
`driving.
`The path planner 122 can be configured to determine the
`navigation route for the vehicle 200 to follow based on the
`vehicle's 200 current location in respect to the surrounding
`environment as well as any points of destination chosen, for
`example, by the driver of the vehicle 200. The path planner
`122 can thus determine the navigation route for the vehicle
`200 based on data received from the data analyzer 120.
`The vehicle controller 124 can be configured to send
`commands to one or more vehicle systems 118 in order to
`maintain the navigation route indicated by the path planner
`122. In one example, the vehicle controller 124 can be a
`propulsion controller configured to send a command to the
`engine throttle to move the position of a throttle plate based
`on the position of an accelerator pedal or a brake pedal. In
`another example, the vehicle controller 124 can send com
`mands to a traction control system to implement steering
`50
`changes or a hybrid control system to redistribute a power
`ratio between electric and gas power sources. As another
`example, the vehicle controller 124 can be an electronic
`stability controller configured to send a command to activate
`one of the front or rear brakes if either more or less yaw
`motion (rotation around the vehicle's 200 vertical axis) is
`detected than optimum for the current angle of the steering
`wheel. In yet another aspect, the vehicle controller 124 can
`be a steering controller configured to initiate the temporary
`and irregular full vehicle movement, Such as side-to-side
`wobbling, to alert a driver of the vehicle.
`In FIG. 2B, the vehicle 200 is illustrated moving in a
`forward direction along an exemplary roadway 202 that
`contains multiple side-by-side lanes, with three lanes 204,
`206, and 208 being shown by example. It shown be under
`stood that the roadway 202 may contain as few as one lane
`and up four or more lanes.
`
`30
`
`35
`
`40
`
`45
`
`55
`
`60
`
`65
`
`US 9,483,059 B2
`
`10
`
`15
`
`25
`
`8
`In one aspect, a sensor may include a camera 115 having
`a field of view directed to the front, to the left side and to the
`right side of the vehicle as respectively shown by the
`directional arrows 210, 212 and 214. This arrangement
`allows the camera 115 to detect surface features of the
`roadway 202, such as lane markers. The roadway 202 can be
`a highway or freeway with typical lane markers, such as a
`solid continuous lane marker 216 at the left edge (in the
`direction of vehicle travel) of the left most lane 204, dashed
`lane markers 218 and 220 respectively defining the right
`edge of the left most lane 204 and the right edge of the
`middle lane 206. The right most lane 208 is delimited at a
`right edge by a solid continuous lane marker 222. The
`camera 115 can have the field of view shown in FIG. 2B
`where a camera 115 can obtain an image of the lane marker
`type to the immediate left side and to the immediate right
`side of the vehicle, such as lane markers 218 and 220 for the
`position of the vehicle 200 in FIG. 2B in the middle lane
`206. Alternately, when the camera 115 has a larger field of
`view, lane marker types at the far edges of the adjacent lanes,
`such as the lane markers 202 or 220 can also be obtained by
`the camera 115 or another suitable sensor.
`The camera 115 may be a black and white or color camera
`capable of sending images of the lane markers detected
`within the fiel

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket