throbber

`
`
`
`1111111111111111 IIIIII IIIII 111111111111111 lllll 111111111111111 1111111111 1111111111 11111111
`
`
`
`
`US 20160267335Al
`
`c19) United States
`c12) Patent Application Publication
`Hampiholi
`
`c10) Pub. No.: US 2016/0267335 Al
`Sep. 15, 2016
`(43) Pub. Date:
`
`(54) DRIVER DISTRACTION DETECTION
`SYSTEM
`
`(71) Applicant: Harman International Industries,
`Incorporated, Stamford, CT (US)
`
`(72)
`
`Inventor: Vallabha Vasant Hampiholi, Bangalore
`(IN)
`
`(21) Appl. No.: 14/657,070
`
`(22) Filed:
`
`Mar. 13, 2015
`
`Publication Classification
`
`(51)
`
`Int. Cl.
`G06K9/00
`H04N7/18
`B60R 161037
`G06T 7100
`G06T 7120
`G02B 27101
`G06F 3101
`(52) U.S. Cl.
`CPC ........ G06K 9/00845 (2013.01); G02B 2710172
`(2013.01); H04N 71181 (2013.01); H04N 71185
`
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`(2006.01)
`
`(2013.01); G06K9/00771 (2013.01); G06K
`9/00597 (2013.01); G06F 3/013 (2013.01);
`G06T 71004 (2013.01); G06T 7120 (2013.01);
`B60R 161037 (2013.01); G02B 2027/0138
`(2013.01); G02B 2027/014 (2013.01); G06T
`2207/30241 (2013.01); G06T 2207/30268
`(2013.01); G06T 2207/30252 (2013.01)
`
`ABSTRACT
`(57)
`Embodiments are described for determining and responding
`to driver distractions. An example in-vehicle computing sys(cid:173)
`tem of a vehicle includes an external device interface com(cid:173)
`municatively connecting the in-vehicle computing system to
`a mobile device, an inter-vehicle system communication
`module communicatively connecting the in-vehicle comput(cid:173)
`ing system to one or more vehicle systems of the vehicle, a
`processor, and a storage device storing instructions execut(cid:173)
`able by the processor to receive image data from the mobile
`device, and determine a driver state based on the received
`image data. The instructions are further executable to receive
`vehicle data from one or more of the vehicle systems, deter(cid:173)
`mine a vehicle state based on the vehicle data, determine a
`distraction severity level based on the driver state and the
`vehicle state, and control one or more devices of the vehicle to
`perform a selected action based on the distraction severity
`level.
`
`102
`
`150
`
`IPR2025-00943
`Tesla EX1007 Page 1
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 1 of 10
`
`US 2016/0267335 Al
`
`...---
`(9
`LL
`
`N
`0
`
`IPR2025-00943
`Tesla EX1007 Page 2
`
`

`

`....
`0 ....
`N
`('D .....
`rJJ =(cid:173)
`0 ....
`'? ....
`
`('D
`
`O'I
`
`N
`~Ul
`
`('D
`rJJ
`
`O'I --- 0
`0 ....
`N
`rJJ
`c
`
`Ul
`~
`~
`--....J
`O'I
`N
`
`> ....
`
`0
`
`0 =
`....
`~ .....
`
`(')
`
`....
`~ .....
`
`(')
`
`0 = ""O = O" -....
`('D = ..... t "e -....
`
`~ .....
`""O
`
`FIG. 2
`
`~
`
`262
`
`CONTROL
`STEERING
`
`'
`!
`
`CONTROLS 261
`
`VEHICLE
`
`I __________________
`l
`
`! TELECOMM. SYS. 236
`
`! CLIMATE CONTROL SYS. 234
`
`i
`i
`
`! AUDIO SYSTEM 232
`
`VEHICLE SYSTEM 231
`
`VEHICLE CONTROL SYSTEM 230
`
`EXTERNAL STORAGE DEVICES 254
`
`BLUETOOTH DEVICES 252
`
`APP. 248 ]°]
`
`.1.1
`
`l.t; EXTERNAL SERVICES 246
`
`I APP. 244 11
`
`l..~OBILE DEVICE 242
`
`EXTERNAL DEVICES 240
`
`'\.
`',r~~~
`.. --"',
`
`~
`
`__ .J--(A)
`
`----:'- ~:;:~
`-~
`/
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`r201
`I
`
`1
`
`TOUCH SCREEN
`
`218
`
`USER INTERFACE
`
`MODULE 222
`
`COMMUNICATION
`
`SYSTEM
`
`INTER-VEHICLE
`
`INTERFACE 212
`
`EXTERNAL DEVICE
`
`ANTENNA 2Q.6_
`
`INTERFACE PROCESSOR 22Q
`
`SUBSYSTEM 211.
`NAVIGATION
`
`204
`
`PROCESSING UNIT
`
`SPEECH
`
`APP(S)
`
`208
`
`STORAGE DEVICE
`
`MEMORY 216
`
`PROCESSOR 214
`
`! ........ CAMERA(S)······___i
`
`SYSTEM
`
`OPERATING
`
`SUBSYSTEM 210
`
`SENSOR
`
`MICROPHONE 202
`
`IN-VEHICLE COMPUTING SYSTEM 200
`
`,---
`
`IPR2025-00943
`Tesla EX1007 Page 3
`
`

`

`O'I --- 0
`0 ....
`N
`rJ'1
`c
`
`Ul
`(,H
`(,H
`-....J
`O'I
`N
`
`> ....
`
`....
`0 ....
`('D .....
`rJ'1 =(cid:173)
`0 ....
`'? ....
`
`0
`
`(,H
`
`('D
`
`O'I
`
`N
`~Ul
`
`('D
`rJ'1
`
`0 =
`....
`~ .....
`
`(')
`
`....
`~ .....
`
`(')
`
`0 = ""O = O" -....
`('D = ..... t "e -....
`
`~ .....
`""O
`
`I ~ r········1
`
`I
`
`,
`
`CAN INTERFACE
`
`AUDIO SUBSYSTEM
`
`318
`
`SUBSYSTEM 316
`
`DISPLAY
`
`DECOMPRESSOR 314
`
`VIDEO
`
`VISUAL ALERT I I AUDIO ALERT I I ii~~~~L
`
`.................. =--~--
`
`'
`
`~--~---··············-.
`
`i
`
`!
`
`! _____________
`
`!
`
`MOTION ANALYSIS
`
`IMAGE CORRECTION
`
`SEVERITY RANKING
`
`ENHANCEMENT
`
`VIDEO
`
`DISTRACTION SEVERITY ANALYSIS 308
`
`DISTRACTION ANALYSIS 306
`
`HEAD UNIT 304a
`
`,--------------------------------------
`
`,300a
`
`CONTROL
`ENGINE
`
`~
`
`~
`
`AUDIOIWARNING
`
`FIG. 3A
`
`SENSOR(S)
`
`OTHER
`
`FAC ING CAMERAS
`FRONT /OUTWARD(cid:173)
`
`WEARABLE DEVICE 302
`
`_,,f-:SI
`
`.... ,-- .. --
`
`~
`
`BLUETOOTH 312
`
`--
`
`,:,/-~:•'
`
`~
`
`WI-Fl 310
`
`-
`
`IPR2025-00943
`Tesla EX1007 Page 4
`
`

`

`....
`0 ....
`('D .....
`rJJ =(cid:173)
`0 ....
`'? ....
`
`0
`
`.i;...
`
`('D
`
`O'I
`
`N
`~Ul
`
`('D
`rJJ
`
`0 =
`....
`~ .....
`
`(')
`
`....
`~ .....
`
`(')
`
`0 = ""O = O" -....
`('D = ..... t "e -....
`
`~ .....
`""O
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-1-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`-
`
`HEAD UNIT INTERFACE 328 +--------~'
`
`'-----t
`
`-
`
`-
`
`-
`
`-
`
`DECOMPRESSOR 314
`
`...:..._
`! BLUETOOTH 312 !
`. ..... ,
`.
`
`-
`
`-
`
`-
`
`-
`
`VIDEO
`
`i.
`
`,300b
`
`L. .............................................
`
`:
`.i
`l l VEHICLE CONTROL l
`I
`
`: ...............................................
`.!
`i
`i AUDIO ALERT
`l VISUAL ALERT i
`:
`
`t ~
`
`1/ !
`!
`i
`! SEVERITY RANKING !
`:
`i
`
`DISTRACTION SEVERITY ANALYSIS 308
`
`CLOUD COMPUTING DEVICE -3.22_
`
`:
`l MOTION ANALYSIS
`·,
`! ENHANCEMENT
`!
`VIDEO
`~--,------_,
`DISTRACTION ANALYSIS .3Q.6.
`
`VIDEO SCENE
`
`ANALYSIS
`
`IMAGE CORRECTION l
`
`I
`I
`I.--__
`I
`I -----'
`I
`I ~--____,:
`I
`I
`I
`.----------------~
`,---------------------------------------
`
`O'I --- 0
`0 ....
`N
`rJJ
`c
`
`Ul
`(,H
`(,H
`-....J
`O'I
`N
`
`> ....
`
`CONTROL
`ENGINE
`
`VEH ICLEI CONTROL
`
`AUDIO WARNING
`
`~
`•
`
`~
`
`I,
`
`VISUAUWARNING
`
`FIG. 38
`
`FACING CAMERAs ... ..J SENSOR(S)
`FRONT/OUTWARD-l OTHER
`
`• ~ I +
`
`! DISPLAY SUBSYSTEM 316 ! l AUDIO SUBSYSTEM~ ! ! CAN INTERFACE~ i
`BLU;~~~~:ill:
`
`HEAD uN1T 304b
`
`_,..-:
`
`I v101::oco1v1PR1::ssORm ~ cLouD INTERFACE J2§ i
`
`~EARABLE DE~ICE ~ ~--·;:::~~:::
`f
`l,i::;
`• " " " "
`-J-
`,
`WI-Fl 310
`.,___---,
`
`I
`
`-
`
`1
`
`IPR2025-00943
`Tesla EX1007 Page 5
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 5 of 10
`
`US 2016/0267335 Al
`
`---~ 404
`
`PROCESS RECEIVED DATA
`
`r400
`
`START
`
`RECEIVE IMAGE DATA
`FROM MOBILE DEVICE 402
`,-------- --- --- ----,
`:
`IMAGE DATA FROM
`:
`: FRONT-FACING CAMERA :
`,----- ----- --- --- --,
`, _________________
`J
`:
`IMAGE DATA FROM
`:
`: DRIVER-FACING CAMERA :
`, _________________
`J
`
`,-------------------
`'
`I
`: SEND DATA TO CLOUD :
`'-------- ,---------]
`,--------- ----------
`: RECEIVE DATA FROM
`:
`CLOUD
`:
`:
`_________
`J
`----------
`
`DETERMINE IF DRIVER IS
`DISTRACTED (FIG. 5) 406
`
`CALCULATE SEVERITY
`RANK, R (FIG. 6) 410
`
`NO
`
`YES PRESENT VISUAL WARNING
`414
`>----1~
`
`YES PRESENT AUDIO WARNING
`418
`>----1~
`
`YES
`
`PERFORM ENGINE
`CONTROL422
`
`RETURN
`
`FIG. 4
`
`IPR2025-00943
`Tesla EX1007 Page 6
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 6 of 10
`
`US 2016/0267335 Al
`
`START
`
`r500
`
`SEND DRIVER-FACING
`CAMERA DATA 502
`r--------,------,
`: HEAD UNIT : : CLOUD :
`, ________
`JI
`-----1
`
`SEND DATA FROM
`OTHER SENSORS 504
`,--------,------,
`: HEAD UNIT : : CLOUD :
`-----1
`, ________
`JI
`
`SEND FRONT-FACING
`CAMERA DATA 506
`,--------,------,
`: HEAD UNIT:: CLOUD:
`, ________
`JI
`-----1
`
`I ..........
`
`••••••••••••••••••••••••••••••:
`
`DATA PROCESSING 508
`,------------------------------------
`!
`!
`PERFORM DAT A
`i
`i
`CORRECTION
`.
`.
`·-----------------------------------·
`j
`PERFORM DATA
`j
`
`
`.__ _____ L-----~~~~~-?.~-~-~-~-: _______ j ____ ___.
`! PERFORM SATURATION !
`_ __j
`! ___ LEVEL_CORRECTION
`,------------------------------------
`.
`.
`i
`:
`PERFORM DATA
`. ___________________________________
`i
`i
`SMOOTHING
`:
`
`DETERMINE DRIVER
`DISTRACTION 514
`
`ANALYZE DATA TO DETERMINE
`ANALYZE PROCESSED DAT A FOR
`DRIVER, VEHICLE, AND
`
`DISTRACTION INDICATORS 510
`ENVIRONMENTAL STATES 512
`i PERFORM COMPARISON WITH i
`-------------------------------------------·
`!
`i
`! PERFORM VIDEO SCENE !
`HISTORICAL DATA
`·----------------------------------------------·
`ANALYSIS
`:
`:
`·------------------------------------------·
`.
`.
`·------------------------------------------·
`! _ PERFORM _MOTION ANALYSIS _j
`.
`.
`.
`.
`·-------------------------------------------
`~~-:~~~-~~-~~~:-~ _________
`L_ _______
`j
`i DETERMINE OBJECT AND !
`
`! ____ VEHICLE_TRAJECTORIES ____ J
`.
`·------------------------------------------·
`! COMPARE OBJECT AND
`'
`j
`L_ __ ~-~~'.?.~-~-!-~-~~-?_:~-~!~~-----j
`! PERFORM COMPARISON WITH l
`!
`i
`HISTORICAL DATA
`..:
`:_ _________________________________________
`
`FIG. 5
`
`IPR2025-00943
`Tesla EX1007 Page 7
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 7 of 10
`
`US 2016/0267335 Al
`
`r600
`
`RETRIEVE ANALYZED DATA 602
`
`DETERMINE TYPES OF DRIVER
`DISTRACTION 604
`
`ANALYZE/PREDICT ENVIRONMENTAL
`DANGERS/RISKS 606
`
`DETERMINE NUMBER OF DRIVER
`DISTRACTIONS 608
`
`DETERMINE NUMBER OF
`ENVIRONMENTAL DANGERS/RISKS 610
`
`PRIORITIZE AND RANK SEVERITY
`BASED ON 604-610 612
`
`PERFORM ACTION BASED ON
`SEVERITY RANK (FIG.4) 614
`
`RETURN
`
`FIG. 6
`
`IPR2025-00943
`Tesla EX1007 Page 8
`
`

`

`O'I --- 0
`0 ....
`N
`rJJ
`c
`
`Ul
`~
`~
`-....J
`O'I
`N
`
`> ....
`
`....
`0 ....
`QO
`('D .....
`rJJ =(cid:173)
`0 ....
`'? ....
`
`O'I
`
`N
`~Ul
`
`('D
`rJJ
`
`0
`
`('D
`
`0 =
`....
`~ .....
`
`(')
`
`....
`~ .....
`
`(')
`
`0 = ""O = O" -....
`('D = ..... t "e -....
`
`~ .....
`""O
`
`FIG. 7
`
`8
`
`8
`
`5
`
`4
`
`1
`
`1
`
`turns within city limits, high speed
`Driver talking on the phone Not using hands free, upcoming
`
`city limits
`vehicle speed, upcoming turn within
`Vehicle not in reverese gear, high
`
`Driver not looking ahead
`
`Vehicle parked on side of freeeway
`
`Parking light off
`
`upcoming turns, medium speed
`
`Driver talking on the phone Using hands free, freeway, no
`
`Vehicle parked in garage
`
`Parking light off
`
`Vehicle stopped at traffic signal
`
`Driver not looking ahead
`
`Severity Rank ( R )
`
`Road/Vehicle Condition
`
`Distraction
`
`r700
`
`IPR2025-00943
`Tesla EX1007 Page 9
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 9 of 10
`
`US 2016/0267335 Al
`
`START
`
`,800
`
`RECEIVE IMAGES FROM
`CAMERAS/SENSORS 802
`
`NO
`
`MONITOR DRIVER DATA.
`VEHICLE DATA, CAN DATA
`806
`
`DETERMINE VEHICLE
`LOCATION AND SPEED 808
`
`NO
`
`NO
`
`NO
`
`INDICATE SEVERITY RANK
`R IN 1sr RANGE, PRESENT
`VISUAL WARNING fila
`
`INDICATE SEVERITY RANK
`IN 2ND RANGE, PRESENT
`AUDIO WARNING ll2Q
`
`INDICATE SEVERITY RANK
`IN 3RD RANGE, PERFORM
`ENGINE CONTROL822
`
`RETURN
`
`FIG. 8
`
`IPR2025-00943
`Tesla EX1007 Page 10
`
`

`

`Patent Application Publication
`
`Sep. 15, 2016 Sheet 10 of 10
`
`US 2016/0267335 Al
`
`START
`
`RECEIVE DATA FROM
`MOBILE DEVICE 902
`,------------------,
`: RECEIVE COMPRESSED :
`: DATA FROM HEAD UNIT :
`'-----------------~
`
`PROCESS RECEIVED DATA,
`DETERMINE VEHICLE DATA,
`OBJECT DATA, DRIVER
`DATA 904
`
`NO
`
`NO
`
`NO
`
`YES
`
`YES
`
`CALCULATE
`TRAJECTORIES OF
`VEHICLE, OBJECT 912
`
`R IN 2ND RANGE, SEND
`AUDIO ALERT FROM CLOUD
`916
`: SEND DATA TO HEAD :
`:
`: UNIT, SEND AUDIO
`'-----------------~
`ALERT
`I
`I
`
`NO
`
`R IN JRD RANGE, PERFORM
`ENGINE CONTROL FROM
`:--SENDC6~~2 {68
`HEAD--:
`: UNIT, PERFORM ENGINE :
`
`: ______ C_QN_T_R.9.!-______ :
`
`YES
`
`FIG. 9
`
`IPR2025-00943
`Tesla EX1007 Page 11
`
`

`

`US 2016/0267335 Al
`
`Sep. 15,2016
`
`1
`
`DRIVER DISTRACTION DETECTION
`SYSTEM
`
`FIELD
`
`[0001] The disclosure relates to assessing driver distraction
`based on the output of a wearable device and other sensors.
`
`BACKGROUND
`
`[0002] Distracted driving may include any activity that
`could divert a person's attention away from the primary task
`of driving. All distractions endanger driver, passenger, and
`bystander safety and could increase the chance of a motor
`vehicle crash. Some types of distraction include visual dis(cid:173)
`traction, where the driver takes his/her eyes off the road,
`manual distraction, where the driver takes his/her hands off
`the wheel, and cognitive distraction, where the driver takes
`his/her mind off of driving. The severity of the distraction
`could depend on both the level and duration of these distrac(cid:173)
`tions and may be compounded by external factors such as
`speed and location of vehicle and objects in the path of the
`vehicle, for example.
`
`SUMMARY
`
`[0003] Embodiments are described for determining and
`responding to driver distractions. An example in-vehicle
`computing system of a vehicle includes an external device
`interface communicatively connecting the in-vehicle com(cid:173)
`puting system to a mobile device, an inter-vehicle system
`communication module communicatively connecting the in(cid:173)
`vehicle computing system to one or more vehicle systems of
`the vehicle, a processor, and a storage device storing instruc(cid:173)
`tions executable by the processor to receive image data from
`the mobile device via the external device interface, the image
`data imaging a driver and a driver environment, and deter(cid:173)
`mine a driver state based on the received image data. The
`instructions are further executable to, responsive to determin(cid:173)
`ing that the driver state indicates that the driver is distracted,
`receive vehicle data from one or more of the vehicle systems
`via the inter-vehicle system communication module, deter(cid:173)
`mine a vehicle state based on the vehicle data, determine a
`distraction severity level based on the driver state and the
`vehicle state, and control one or more devices of the vehicle to
`perform a selected action based on the distraction severity
`level.
`[0004] An example method of determining driver distrac(cid:173)
`tion includes receiving driver data from a wearable device, the
`driver data including image data from a driver-facing camera,
`receiving object data from one or more imaging devices of at
`least one of the wearable device and the vehicle, the object
`data including image data of a vehicle environment, and
`receiving vehicle data from one or more vehicle systems, the
`vehicle data including an indication of an operating condition
`of the vehicle. The example method further includes deter(cid:173)
`mining whether a driver is distracted by correlating the driver
`data with the object data, responsive to determining that the
`driver is distracted, selecting an action based on correlating
`the driver data with the object data and the vehicle data and
`performing the selected action, and responsive to determining
`that the driver is not distracted, maintaining current operating
`parameters.
`[0005] An example distraction monitoring system includes
`a wearable device including a driver-facing camera, an out(cid:173)
`ward-facing camera, and one or more sensors, an in-vehicle
`
`computing system communicatively connected to the wear(cid:173)
`able device and one or more vehicle systems, the in-vehicle
`computing system comprising a first processor and a first
`storage device, and a cloud computing device remote from the
`in-vehicle computing system and communicatively con(cid:173)
`nected to the in-vehicle computing system via a network, the
`cloud computing device comprising a second processor and a
`second storage device. One or more of the first storage device
`and the second storage device storing first instructions
`executable by a respective one or more of the first processor
`and the second processor to receive image data from the
`driver-facing camera and sensor data from the one or more
`sensors of the wearable device indicating a driver state,
`receive image data from the outward-facing camera of the
`wearable device indicating object states of one or more
`objects, receive vehicle data from one or more vehicle sys(cid:173)
`tems to indicate vehicle state, and select an action to be
`performed based on the indicated driver state, object states,
`and vehicle state. The first storage device may store second
`instructions executable by the first processor to transmit a
`command to one or more of a display device of the in-vehicle
`computing system, an audio device of the vehicle, and an
`engine control unit of the vehicle to perform the selected
`action.
`
`BRIEF DESCRIPTION OF THE DRAWINGS
`
`[0006] The disclosure may be better understood from read(cid:173)
`ing the following description of non-limiting embodiments,
`with reference to the attached drawings, wherein below:
`[0007] FIG. 1 shows an example partial view of a vehicle
`cabin in accordance with one or more embodiments of the
`present disclosure;
`[0008] FIG. 2 shows an example in-vehicle computing sys(cid:173)
`tem in accordance with one or more embodiments of the
`present disclosure;
`[0009] FIG. 3A shows a block diagram of an example dis(cid:173)
`traction determination system in accordance with one or more
`embodiments of the present disclosure;
`[0010] FIG. 3B shows a block diagram of an example dis(cid:173)
`traction determination system in accordance with one or more
`embodiments of the present disclosure;
`[0011] FIG. 4 is a flow chart of an example method for
`generating warnings based on a calculated severity rank range
`from the perspective of a head unit in accordance with one or
`more embodiments of the present disclosure;
`[0012] FIG. 5 is a flow chart for an example method of
`determining a distraction level of a driver in accordance with
`one or more embodiments of the present disclosure;
`[0013] FIG. 6 is a flow chart for an example method of
`determining a distraction severity rank in accordance with
`one or more embodiments of the present disclosure;
`[0014] FIG. 7 shows an example table mapping severity
`rank R to driver states and vehicle conditions in accordance
`with one or more embodiments of the present disclosure;
`[0015] FIG. 8 is a flow chart for an example method of
`determining an action to be performed responsive to example
`driver states and vehicle conditions in accordance with one or
`more embodiments of the present disclosure; and
`[0016] FIG. 9 is a flow chart for an example method of
`determining an action to be performed responsive to example
`driver states, object states, and vehicle conditions in accor(cid:173)
`dance with one or more embodiments of the present disclo(cid:173)
`sure.
`
`IPR2025-00943
`Tesla EX1007 Page 12
`
`

`

`US 2016/0267335 Al
`
`Sep. 15,2016
`
`2
`
`DETAILED DESCRIPTION
`
`[0017] As described above, driver distraction may be dan(cid:173)
`gerous to occupants of a vehicle, as well as people in a vicinity
`of the vehicle. However, by only monitoring a driver state, or
`by monitoring only a small number of distraction-related
`behaviors of a driver, an inappropriate response to a distracted
`driver may be provided. For example, if a driver is nodding off
`to sleep, a visual warning may be insufficient to correct the
`distracted behavior. As another example, if a driver briefly
`looks away from the road, but there are no objects in the path
`of the vehicle, a loud, audible warning of driver distraction
`may be unnecessary and instead startle the driver, causing the
`driver to lose control of the vehicle. A distraction monitoring
`system that not only monitors the level of driver distraction,
`but also effectively responds to such distraction in a timely
`manner may address the issue of distracted driving and the
`major traffic safety issue that such driving poses.
`[0018] Accordingly, the disclosure provides a distraction
`monitoring system including one or more of an in-vehicle
`computing system and a cloud computing device that receives
`sensor data from a wearable device and/or other sensor
`devices to determine a driver state and a state of objects in a
`vehicle environment. The in-vehicle computing system and/
`or cloud computing device may also receive data from vehicle
`systems in order to determine a vehicle state. By correlating
`the driver, object(s), and vehicle state, the distraction moni(cid:173)
`toring system may first determine whether the driver is dis(cid:173)
`tracted, and then determine a severity of that distraction. The
`distraction monitoring system may perform a different action
`(e.g., provide a visual alert, provide an audible alert, and/or
`perform a vehicle control) responsive to different levels of
`distraction severity, as different types of distractions may
`benefit from different types of warnings/responses. In this
`way, a driver may be alerted to his/her distraction in an appro(cid:173)
`priate manner based on an intelligent combination of the
`different types of data.
`[0019] FIG. 1 shows an example partial view of one type of
`environment for a communication system: an interior of a
`cabin 100 of a vehicle 102, in which a driver and/or one or
`more passengers may be seated. Vehicle 102 of FIG. 1 may be
`a motor vehicle including drive wheels (not shown) and an
`internal combustion engine 104. Internal combustion engine
`104 may include one or more combustion chambers, which
`may receive intake air via an intake passage and exhaust
`combustion gases via an exhaust passage. Vehicle 102 may be
`a road automobile, among other types of vehicles. In some
`examples, vehicle 102 may include a hybrid propulsion sys(cid:173)
`tem including an energy conversion device operable to absorb
`energy from vehicle motion and/ or the engine and convert the
`absorbed energy to an energy form suitable for storage by an
`energy storage device. Vehicle 102 may include a fully elec(cid:173)
`tric vehicle, incorporating fuel cells, solar energy capturing
`elements, and/or other energy storage systems for powering
`the vehicle.
`[0020] As shown, an instrument panel 106 may include
`various displays and controls accessible to a driver (also
`referred to as the user) of vehicle 102. For example, instru(cid:173)
`ment panel 106 may include a touch screen 108 of an in(cid:173)
`vehicle computing system 109 ( e.g., an infotainment system),
`an audio system control panel, and an instrument cluster 110.
`While the example system shown in FIG. 1 includes audio
`system controls that may be performed via a user interface of
`in-vehicle computing system 109, such as touch screen 108
`without a separate audio system control panel, in other
`
`embodiments, the vehicle may include an audio system con(cid:173)
`trol panel, which may include controls for a conventional
`vehicle audio system such as a radio, compact disc player,
`MP3 player, etc. The audio system controls may include
`features for controlling one or more aspects of audio output
`via speakers 112 of a vehicle speaker system. For example,
`the in-vehicle computing system or the audio system controls
`may control a volume of audio output, a distribution of sound
`among the individual speakers of the vehicle speaker system,
`an equalization of audio signals, and/or any other aspect of
`the audio output. In further examples, in-vehicle computing
`system 109 may adjust a radio station selection, a playlist
`selection, a source of audio input ( e.g., from radio or CD or
`MP3), etc., based on user input received directly via touch
`screen 108, or based on data regarding the user (such as a
`physical state and/or environment of the user) received via
`external devices 150 and/or mobile device 128.
`[0021]
`In some embodiments, one or more hardware ele(cid:173)
`ments of in-vehicle computing system 109, such as touch
`screen 108, a display screen, various control dials, knobs and
`buttons, memory, processor( s ), and any interface elements
`(e.g., connectors or ports) may form an integrated head unit
`that is installed in instrument panel 106 of the vehicle. The
`head unit may be fixedly or removably attached in instrument
`panel 106. In additional or alternative embodiments, one or
`more hardware elements of the in-vehicle computing system
`may be modular and may be installed in multiple locations of
`the vehicle.
`[0022] The cabin 100 may include one or more sensors for
`monitoring the vehicle, the user, and/or the environment. For
`example, the cabin 100 may include one or more seat(cid:173)
`mounted pressure sensors configured to measure the pressure
`applied to the seat to determine the presence of a user, door
`sensors configured to monitor door activity, humidity sensors
`to measure the humidity content of the cabin, microphones to
`receive user input in the form of voice commands, to enable a
`user to conduct telephone calls, and/or to measure ambient
`noise in the cabin 100, etc. It is to be understood that the
`above-described sensors and/or one or more additional or
`alternative sensors may be positioned in any suitable location
`of the vehicle. For example, sensors may be positioned in an
`engine compartment, on an external surface of the vehicle,
`and/or in other suitable locations for providing information
`regarding the operation of the vehicle, ambient conditions of
`the vehicle, a user of the vehicle, etc. Information regarding
`ambient conditions of the vehicle, vehicle status, or vehicle
`driver may also be received from sensors external to/separate
`from the vehicle (that is, not part of the vehicle system), such
`as from sensors coupled to external devices 150 and/or
`mobile device 128.
`[0023] Cabin 100 may also include one or more user
`objects, such as mobile device 128, that are stored in the
`vehicle before, during, and/or after travelling. The mobile
`device may include a smart phone, a tablet, a laptop computer,
`a portable media player, and/or any suitable mobile comput(cid:173)
`ing device. The mobile device 128 may be connected to the
`in-vehicle computing system via communication link 130.
`The communication link 130 may be wired ( e.g., via Univer(cid:173)
`sal Serial Bus [USB], Mobile High-Definition Link [MHL],
`High-Definition Multimedia Interface [HDMI], Ethernet,
`etc.) or wireless (e.g., via BLUETOOTH, WI-FI, Near-Field
`Communication [NFC], cellular connectivity, etc.) and con(cid:173)
`figured to provide two-way communication between the
`mobile device and the in-vehicle computing system. For
`
`IPR2025-00943
`Tesla EX1007 Page 13
`
`

`

`US 2016/0267335 Al
`
`Sep. 15,2016
`
`3
`
`example, the communication link 130 may provide sensor
`and/or control signals from various vehicle systems (such as
`vehicle audio system, climate control system, etc.) and the
`touch screen 108 to the mobile device 128 and may provide
`control and/or display signals from the mobile device 128 to
`the in-vehicle systems and the touch screen 108. The com(cid:173)
`munication link 130 may also provide power to the mobile
`device 128 from an in-vehicle power source in order to charge
`an internal battery of the mobile device.
`[0024]
`In-vehicle computing system 109 may also be com(cid:173)
`municatively coupled to additional devices operated and/or
`accessed by the user but located external to vehicle 102, such
`as one or more external devices 150. In the depicted embodi(cid:173)
`ment, external devices 150 are located outside of vehicle 102
`though it will be appreciated that in alternate embodiments,
`external devices may be located inside cabin 100. The exter(cid:173)
`nal devices may include a server computing system, personal
`computing system, portable electronic device, electronic
`wrist band, electronic head band, portable music player, elec(cid:173)
`tronic activity tracking device, pedometer, smart-watch, GPS
`system, etc. External devices 150 may be connected to the
`in-vehicle computing system via communication link 136,
`which may be wired or wireless, as discussed with reference
`to communication link 130, and configured to provide two(cid:173)
`way communication between the external devices and the
`in-vehicle computing system. For example, external devices
`150 may include one or more sensors and communication
`link 136 may transmit sensor output from external devices
`150 to in-vehicle computing system 109 and touch screen
`108. External devices 150 may also store and/or receive infor(cid:173)
`mation regarding contextual data, user behavior/preferences,
`operating rules, etc. and may transmit such information from
`the external devices 150 to in-vehicle computing system 109
`and touch screen 108.
`[0025]
`In-vehicle computing system 109 may analyze the
`input received from external devices 150, mobile device 128,
`and/or other input sources and select settings for various
`in-vehicle systems (such as climate control system or audio
`system), provide output via touch screen 108 and/or speakers
`112, communicate with mobile device 128 and/or external
`devices 150, and/or perform other actions based on the
`assessment. In some embodiments, all or a portion of the
`assessment may be performed by the mobile device 128 and/
`or the external devices 150.
`[0026]
`In some embodiments, one or more of the external
`devices 150 may be communicatively coupled to in-vehicle
`computing system 109 indirectly, via mobile device 128 and/
`or another of the external devices 150. For example, commu(cid:173)
`nication link 136 may communicatively couple external
`devices 150 to mobile device 128 such that output from exter(cid:173)
`nal devices 150 is relayed to mobile device 128. Data received
`from external devices 150 may then be aggregated at mobile
`device 128 with data collected by mobile device 128, the
`aggregated data then transmitted to in-vehicle computing sys(cid:173)
`tem 109 and touch screen 108 via communication link 130.
`Similar data aggregation may occur at a server system and
`then transmitted to in-vehicle computing system 109 and
`touch screen 108 via communication link 136/130.
`[0027]
`In the example environment illustrated in FIG. 1, the
`in-vehicle computing system 109 may be connected to one or
`more vehicle systems, such as speakers 112, display 108,
`vehicle sensors, and/or other suitable vehicle systems via any
`suitable network. In some examples, the in-vehicle comput(cid:173)
`ing system 109 includes a talker device configured to transmit
`
`audio/video data to listener devices, such as speakers 112 and
`display 108 via a network. The network may be configured in
`accordance with Layer 2 of the Open Systems Interconnec(cid:173)
`tion (OSI) model, in which routing and forwarding decisions
`or determinations in the network may be performed on a
`media access control (MAC) addressing basis. An example
`Layer 2 network may be an Ethernet AudioNideo Bridging
`(AVB) network. For Layer 2 networks configured as AVB
`networks, the talkers and the listeners may be configured to
`communicate over the AVB network using various AVB stan(cid:173)
`dards and protocols, including the Institute of Electrical and
`Electronics Engineers (IEEE) 802.lAS-2011 (gPTP) for net(cid:173)
`work timing and synchronization, IEEE 802 .1Q-2011 clause
`34 for queuing and forwarding streaming data, IEEE 802 .1 Q-
`2011 clause 35 (Stream Reservation Protocol (SRP)) for
`reserving a network connection or path and/or resources such
`as bandwidth for communication over the network connec(cid:173)
`tion, and/or IEEE 1722-2011 related to a possible data
`streaming format. Other AVE-related standards and proto(cid:173)
`cols, and/or other versions of the AVB standards and proto(cid:173)
`cols, previously, currently, or later developed, may addition(cid:173)
`ally or alternatively be used.
`[0028]
`It is to be understood that FIG. 1 depicts one
`example environment, however the communication systems
`and methods described herein may be utilized in any suitable
`environment. Any suitable devices that transmit and/or
`receive information, sense data, and/or otherwise contribute
`to a driver distraction detection and/or alert system may be
`utilized as the systems and/or to perform the methods
`described herein.
`[0029] FIG. 2 shows a block diagram of an in-vehicle com(cid:173)
`puting system 200 configured and/or integrated inside vehicle
`201. In-vehicle computing system 200 may be an example of
`in-vehicle computing system 109 of FIG. 1 and/or may per(cid:173)
`form one or more of the methods described herein in some
`embodiments. In some examples, the in-vehicle computing
`system may be a vehicle infotainment system configured to
`provide information-based media content (audio and/or
`visual media content, including entertainment content, navi(cid:173)
`gational services, etc.) to a vehicle user to enhance the opera(cid:173)
`tor's in-vehicle experience. The vehicle infotainment system
`may include, or be coupled to, various vehicle systems, sub(cid:173)
`systems, hardware components, as well as software applica(cid:173)
`tions and systems that are integrated in, or integratable into,
`vehicle 201 in order to enhance an in-vehicle experience for a
`driver and/or a passenger.
`[0030]
`In-vehicle computing system 200 may include one
`or more processors including an operating sy

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket