`in View of
`
`U.S. Patent No. 9,392,212 to (“Ross”),
`Filed September 17, 2014 (claims benefit of provisional application 61/980,658 filed April 17, 2014).
`
`#
`1a
`
`U.S. Patent No. 10,406,432
`1. A computer program product
`embodied on a non-transitory
`computer-readable medium,
`comprising code executable by a
`virtual image display apparatus
`having at least a processor and a
`memory, the memory being
`configured to store an information
`providing condition of the virtual
`image display apparatus and being
`further configured to store to-be-
`provided information, to cause the
`virtual image display apparatus to
`carry out the following steps:
`
`Ross
`Ross expressly or inherently discloses these claim element(s).
`
`[A computer program product embodied on a non-transitory computer-readable medium,
`comprising code executable by a virtual image display apparatus having at least a processor and
`a memory,]
`
`“The processor may be configured to execute computer program components. The computer
`program components may be configured to enable an expert and/or user to interface with the system
`and/or provide other functionality attributed herein to the user interface, the sensor, the electronic
`storage, and/or the processor. The computer program components may include a direction
`component, a content component, a display component, an interaction component, and/or other
`components.” 3:57-65
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of A computer
`program product embodied on a non-transitory computer-readable medium (e.g., the computer
`program components), comprising code executable by a virtual image display apparatus (e.g., the
`display component) having at least a processor and a memory (e.g., the electronic storage and
`processor.
`
`[the memory being configured to store an information providing condition of the virtual image
`display apparatus]
`
`“Returning to FIG. 1, electronic storage 30 may comprise electronic storage media that electronically
`stores information. The electronic storage media of the electronic storage may include one or both of
`storage that is provided integrally (i.e., substantially non-removable) with the respective device and/or
`removable storage that is removably connectable to the respective device...Electronic storage may
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 1 of 24
`
`Supercell
`Exhibit 1013
`Page 1
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`1b
`
`detecting, with a sensor
`operationally linked to the virtual
`image display apparatus, a
`movement of a body part of a
`player, the body part comprising
`
`Ross
`store files, software algorithms, information determined by processor(s) 20, and/or other
`information that enables the respective devices to function as described herein.” 23:46-68
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of the memory
`(e.g., the electronic storage 30) being configured to store an information providing condition of the
`virtual image display apparatus (e.g., the information that enables the respective devices to function,
`such as any instructions related to an information providing condition).
`
`[and being further configured to store to-be-provided information, to cause the virtual image
`display apparatus to carry out the following steps:]
`
`“The display may be configured to present the virtual reality content to the user. The display may be
`controlled by the processor to present the virtual reality content to the user such that the presented
`virtual reality content corresponds to a view direction of the user. In some implementations, the display
`may be included in a virtual reality headset worn by the user.” 3:38-45
`
`“Display 16 may be controlled by processor 20 to present the virtual reality content to the user such
`that the presented virtual reality content corresponds to a view direction of the user…As described
`above, in some implementations, display 16 may be included in a virtual reality headset worn by the
`user.” 12:7-10
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of and being
`further configured to store to-be-provided information (e.g., the virtual reality content), to cause the
`virtual image display apparatus (e.g., the display 16) to carry out the following steps (e.g.,
`information stored in the electronic storage 30 executed by the processor 30 and enabling the
`display to function).
`
`
`Ross expressly or inherently discloses these claim element(s).
`
`
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 2 of 24
`
`Supercell
`Exhibit 1013
`Page 2
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`at least one of a head of the player
`and an eye of the player, and the
`sensor being at least one of the set
`of: a gyro sensor configured to
`measure movement of the head of
`the player, an acceleration sensor
`configured to measure movement
`of the head of the player, a
`geomagnetic sensor configured to
`measure movement of the head of
`the player and a line-of-sight
`sensor configured to measure
`movement of the eye of the
`player; and
`
`Ross
`[detecting, with a sensor operationally linked to the virtual image display apparatus, a movement
`of a body part of a player, the body part comprising at least one of a head of the player and an
`eye of the player,]
`
`“Returning to FIG. 1, sensors 18 may be configured to generate output signals conveying information
`related to a view direction of the user and/or other information” 11:4-6
`
`“In some implementations, system 10 may comprise one or more of a user interface 14 (which may
`include a display 16 and/or other components as described herein), a sensor 18, a processor 20,
`electronic storage 30, and/or other components.” 7:63-66
`
`“For example, the system may be configured to determine that a view direction of the user has moved
`across a field boundary between a first field of view (where the primary virtual reality content is
`displayed) and a second field of view (e.g., where the secondary virtual reality content is displayed).
`The view direction of the user may refer to a direction toward which the user's gaze is directed, an
`orientation of the user's gaze (e.g., the user may tilt his head and/or lean over), a position of the user
`within the virtual space, and/or other directional and/or positional information (e.g., a user may move
`his position in the virtual space across a boundary).” 2:41-52
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of detecting, with
`a sensor (e.g., sensor 18) operationally linked (e.g., via the system 10) to the virtual image display
`apparatus (e.g., display 16), a movement of a body part of a player (e.g., view direction of the user),
`the body part comprising at least one of a head of the player and an eye of the player (e.g., the user’s
`head tilt or user’s gaze corresponding to the view direction).
`
`
`[and the sensor being at least one of the set of: a gyro sensor configured to measure movement of
`the head of the player, an acceleration sensor configured to measure movement of the head of the
`player, a geomagnetic sensor configured to measure movement of the head of the player and a
`line-of-sight sensor configured to measure movement of the eye of the player;]
`
`“In some implementations, sensors 18 may include one or more of a GPS sensor, a gyroscope, an
`accelerometer, an altimeter, a compass, a camera based sensor, a magnetic sensor, an optical sensor, an
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 3 of 24
`
`Supercell
`Exhibit 1013
`Page 3
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`1c
`
`determining, based on the
`movement of the body part of the
`player, a position and direction of
`the body part of a player;
`
`Ross
`infrared sensor, a motion tracking sensor, an inertial sensor, a CCB sensor, an eye tracking sensor, a
`facial tracking sensor, a body tracking sensor, and/or other sensors.” 11:22-28
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of and the sensor
`being at least one of the set of: a gyro sensor configured to measure movement of the head of the
`player, an acceleration sensor configured to measure movement of the head of the player, a
`geomagnetic sensor configured to measure movement of the head of the player and a line-of-sight
`sensor configured to measure movement of the eye of the player (e.g., sensor 18 examples including
`a gyroscope, accelerometer, magnetic sensor, eye tracking sensor)
`
`Ross expressly or inherently discloses these claim element(s).
`
`“Returning to FIG. 1, sensors 18 may be configured to generate output signals conveying information
`related to a view direction of the user and/or other information. The view direction of the user may
`correspond to a physical direction toward which a gaze of the user is directed, an orientation of one or
`more parts of the user's body (e.g., the user's head may be tilted, the user may be leaning over), a
`position of a user within the virtual space, and/or other directional information.” 11:18-26
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of determining
`(e.g., using the sensors 18 to determine the view direction), based on the movement of the body part
`of the player (e.g., gaze, user body part orientation), a position (e.g., position of the user) and
`direction of the body part of a player (e.g., and other directional information);
`
`
`1d
`
`displaying, on a display
`operationally linked to the virtual
`image display apparatus, in
`accordance with the position and
`direction of the body part of the
`player, an image of a virtual
`space including a first area and a
`second area; and
`
`[displaying, on a display operationally linked to the virtual image display apparatus, in
`accordance with the position and direction of the body part of the player,]
`
`“Display 16 may be configured to present the virtual reality content to the user. Display 16 may be
`configured to present the virtual reality content to the user such that the presented virtual reality
`content corresponds to a view direction of the user. Display 16 may be controlled by processor 20 to
`present the virtual reality content to the user such that the presented virtual reality content corresponds
`to a view direction of the user.” 12:3-10
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 4 of 24
`
`Supercell
`Exhibit 1013
`Page 4
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`1e
`
`with the virtual image display
`apparatus, providing, when the
`information providing condition
`is satisfied, the to-be-provided
`information to the player by
`displaying the to-be-provided
`information in the second area;
`
`Ross
`Accordingly, in view of the above, the Ross reference anticipates the claim element of displaying, on
`a display operationally linked to the virtual image display apparatus (e.g., display 16), in accordance
`with the position and direction of the body part of the player (e.g., presented VR content corresponds
`to a view direction of the user).
`
`
`[an image of a virtual space including a first area and a second area;]
`
`“The present system may facilitate storytelling using virtual reality via boundaries that divide a virtual
`space into areas where primary content, secondary content, tertiary content, etc. may be presented
`to the user.” 5:52-55
`
`“Field boundary 404 may be visible, partially visible, and/or invisible to a user.” 19:43-44.
`
`As field boundaries may be invisible, the different areas are not visibly separate from each other.
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of an image of a
`virtual space (e.g., the virtual space) including a first area (e.g., the area including the primary
`content) and a second area (e.g., the area of the virtual space including the secondary content).
`
`
`Ross expressly or inherently discloses these claim element(s).
`
`Presented below are two separate interpretations of the claimed language using the Ross reference.
`Each interpretation by itself anticipates the claimed limitation. In the first comparison, the first area
`of the claims is corresponded to the secondary content of Ross, and the second area of the claims is
`corresponded to the primary content of Ross. In the second comparison, these are reversed, with the
`first area corresponded to the primary content in Ross, and the second area corresponded to the
`secondary content in Ross.
`
`Comparison 1) first area = secondary content in Ross, second area = primary content in Ross
`
`“The user may interact with the secondary, tertiary, etc., virtual reality content and then return his
`view back to the movie wherein the movie has been adjusted to reflect the user's interaction with the
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 5 of 24
`
`Supercell
`Exhibit 1013
`Page 5
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`Ross
`secondary, tertiary, etc., virtual reality content. In some implementations, while a user views primary
`virtual reality content, system 10 may be configured to display a “pop-up” cue indicating that
`additional virtual reality content (e.g., additional characters) is available in a different (e.g.,
`secondary, tertiary, etc.) content area.” 7:41-50 (the movie can include “2D and/or 3D displayed
`content, captured video, 2D and/or 3D generated content, digitally created characters, objects, and
`spaces, algorithmically created content” 7:14-17)
`
`
`“In this example, the primary virtual reality content may be a movie (2D and/or 3D displayed
`content, captured video, 2D and/or 3D generated content, digitally created characters, objects, and
`spaces, algorithmically created content), and/or any other primary virtual reality content displayed to
`the user.” 7:13-18
`
`“This enables data, cues, results, and/or instructions and any other communicable items, collectively
`referred to as "information," to be communicated between the user and system 10.” 11:31-35
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of with the
`virtual image display apparatus, providing (e.g., displaying the pop-up), when the information
`providing condition is satisfied (e.g., when the user returns the view from the secondary to the
`primary content), the to-be-provided information (e.g., the pop-up cue) to the player by displaying
`the to-be-provided information in the second area (e.g., the area of the primary content, when the
`user views the primary content).
`
`Comparison 2) first area = primary content in Ross, second area = secondary content in Ross
`
`“System 10 may be configured such that primary virtual reality content may be displayed to a user,
`and, responsive to the user turning his view and/or otherwise moving away from the primary
`virtual reality content, one or more sensory cues may be provided to the user that indicates to the
`user that his view is no longer directed toward the primary virtual reality content…System 10
`may be configured to determine that a view direction of the user has moved across a field boundary
`between a first field of view where the primary virtual reality content is displayed and a second
`field of view where, for example, secondary virtual reality content may be displayed.” 6:46-60.
`
`“Other sensory cues may include other visual cues (e.g., a flashing visual indicator that is part of user
`interface 14), auditory cues (e.g., sounds produced by user interface 14 responsive to the user's view
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 6 of 24
`
`Supercell
`Exhibit 1013
`Page 6
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`Ross
`direction moving across a field boundary such as audio from the primary virtual reality content slowing
`to a stop, speaker vibration cues)...In some implementations, display component 26 may be configured
`to cause display 16 and/or other components of system 10 to vary an intensity of an individual cue
`and/or provide more than one cue at a time.” 16:39-61
`
`“This enables data, cues, results, and/or instructions and any other communicable items, collectively
`referred to as "information," to be communicated between the user and system 10.” 11:32-35
`
`“The secondary (tertiary, etc.) virtual reality content may be adjusted based on a user's progression
`through the primary virtual reality content, interaction with the primary virtual reality content, and/or
`other information.” 7:6-10
`
`“Content component 24 may be configured such that primary virtual reality content is presented to the
`user responsive to the user's view direction being within a first field of view (e.g., a forward looking
`field of view), and secondary virtual reality content may be presented to the user when the user's view
`direction is within a second field of view (e.g., a rearward looking field of view), and so on for tertiary,
`etc., virtual reality content. In some implementations, content component 24 may be configured to
`change the virtual reality content displayed to the user responsive to the user's view direction
`moving across a field boundary (e.g., between two different fields of view), based on an amount of
`view direction rotation, and/or based on other information” 15:29-41
`
`“Responsive to the user turning his view from the primary virtual reality content (e.g., left or right, up
`or down, and/or other directions) and/or otherwise moving away from the primary virtual reality
`content in the virtual space, a sensory cue comprising a visually perceptible darkening of the primary
`virtual reality content (e.g., a movie), a brightening of related (e.g., movie characters, objects seen in
`the movie, a setting of the movie, an extension of the setting of the movie, etc.) secondary, tertiary,
`etc.” 7: 21-33
`
`Therefore, as the user’s view moving from the area of the primary content to the secondary content, a
`visual cue is provided which Ross refers to as “information.” In comparison 1) above, Ross also
`refers to the pop-up with the additional information notice as a “cue” as well. Ross further motes that
`the content displayed to the user in the secondary area may change based on a variety of factors, and
`also notes that the content displayed to the user based on the user’s view direction moving. Thus, it is
`inferred that the visual cue or other information may be presented in the secondary area. An example
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 7 of 24
`
`Supercell
`Exhibit 1013
`Page 7
`
`
`
`
`
`#
`
`1f
`
`U.S. Patent No. 10,406,432
`
`Ross
`provided in Ross is a cue that involves at least the brightening of the secondary content area, and thus,
`it can be inferred that other visual cues/information are presented at the secondary area as well.
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of with the
`virtual image display apparatus, providing (e.g., providing the visual cue), when the information
`providing condition is satisfied (e.g., when the user’s view moves from the primary to the secondary
`area), the to-be-provided information (e.g., the visual cue) to the player by displaying the to-be-
`provided information in the second area (e.g., the secondary content area).
`
`wherein the information
`providing condition is a condition
`of a gaze position moving to the
`second area from the first area,
`the gaze position being specified
`by at least one of the body part of
`the player being in a specified
`position or the direction of the
`body part of the player being at
`least a specified direction.
`
`Ross expressly or inherently discloses these claim element(s).
`
`[wherein the information providing condition is a condition of a gaze position moving to the
`second area from the first area]
`
`See above comparisons showing the user’s view moving from the primary to the secondary or vice
`versa in two separate comparisons.
`
`[the gaze position being specified by at least one of the body part of the player being in a
`specified position or the direction of the body part of the player being at least a specified
`direction.]
`
`“Returning to FIG. 1, sensors 18 may be configured to generate output signals conveying information
`related to a view direction of the user and/or other information. The view direction of the user may
`correspond to a physical direction toward which a gaze of the user is directed, an orientation of one or
`more parts of the user's body (e.g., the user's head may be tilted, the user may be leaning over), a
`position of a user within the virtual space, and/or other directional information.” 11:6-12
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of the gaze
`position (e.g., view direction) being specified by at least one of the body part of the player (e.g.,
`user’s body, head, gaze) being in a specified position (e.g., position of user) or the direction of the
`body part of the player (e.g., directional information) being at least a specified direction (e.g.,
`moving to the primary or secondary area).
`
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 8 of 24
`
`Supercell
`Exhibit 1013
`Page 8
`
`
`
`
`
`#
`2a
`
`U.S. Patent No. 10,406,432
`2. The computer program product
`of claim 1, wherein
`the virtual space includes a target
`object selectable by a gaze of the
`player,
`wherein an area including the
`target object is recorded in the
`memory,
`
`
`Ross
`Ross expressly or inherently discloses these claim element(s).
`
`“Primary virtual reality content may include movies, performances, virtual events, video games,
`interactive games, simulated physical world settings, characters, objects…Secondary, tertiary, etc.,
`virtual reality content may include individual characters, objects that are part of a specific setting,
`objects that are usable within the virtual space such as tools, food, a virtual currency, virtual
`clothing (e.g., shirt, hat, pants, etc.), a vehicle, a pet, movies, performances, virtual events, characters,
`objects, settings…” 14:60-15:28
`
`“Interaction may include participating in an instance of the virtual space (e.g., a scene in a movie) by
`controlling one or more available user controlled entities in the virtual space. Such control may be
`exercised through control inputs and/or commands input by the user through computing device 12, for
`example. Interaction may include repeatedly looking in a specific direction within a field of view
`(e.g., at a specific character). Interaction may include repeatedly looking at and/or otherwise
`directing attention toward a specific character, object, and/or location in the virtual space; making
`facial expressions that express an attitude toward a specific character, object, and/or location in the
`virtual space; and/or other interaction (e.g., determined based on output signals from eye tracking,
`facial tracking, brain wave, body tracking sensors, and/or other sensors included in sensors 18).”
`17:49-63
`
`“The virtual reality content (e.g., the primary, secondary, tertiary, etc., virtual reality content,
`information that describes the virtual space itself) for display to the user may be obtained from
`electronic storage, via external resources not included in system 10 (e.g., a server), and/or from other
`sources.” 24:59-64
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of the virtual
`space includes a target object (e.g., objects, characters) selectable by a gaze of the player (e.g.,
`interaction by repeatedly looking at an object), wherein an area including the target object (e.g., a
`primary VR content area, secondary VR content area) is recorded in the memory (e.g., electronic
`storage).
`
`
`2b
`
`wherein the information
`providing condition includes a
`
`Ross expressly or inherently discloses these claim element(s).
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 9 of 24
`
`Supercell
`Exhibit 1013
`Page 9
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`condition wherein a gaze position
`of the player, the gaze position
`being identified from the position
`and the direction of the body part
`of the player, becomes directed
`outside the area, and
`further comprising providing,
`with the virtual image display
`apparatus, the to-be-provided
`information to the player when
`the gaze position of the player
`becomes directed outside the
`area.
`
`Ross
`See comparison above for 1e, where the primary/secondary content areas in Ross are compared
`against the first and second areas in the ‘432 claims. Here, the area with the target object can be
`matched with either the primary or secondary VR content area of Ross, and the region outside this
`area can be the secondary or primary VR content areas in Ross, respectively. When the user’s view
`direction in Ross moves from one area to another, as shown in the citations in 1e (not reproduced here
`for sake of clarity), information, such as a visual cue or a pop up, is shown to the user in the other
`area.
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of wherein the
`information providing condition includes a condition wherein a gaze position of the player (e.g., the
`user’s view direction), the gaze position being identified from the position (e.g., position) and the
`direction (e.g., directional information) of the body part of the player (e.g., the user’s head, body
`part), becomes directed outside the area (e.g., outside the primary or secondary VR content areas),
`and
`further comprising providing, with the virtual image display apparatus, the to-be-provided
`information (e.g., the visual cue or popup cue) to the player when the gaze position of the player
`becomes directed outside the area (e.g., moving from the primary/secondary to secondary/primary
`areas).
`
`
`3a
`
`3b
`
`3. The computer program product
`of claim 1, wherein
`the virtual space includes a target
`object selectable by a gaze of the
`player,
`
`
`wherein the information
`providing condition includes a
`condition wherein a gaze position
`of the player, the gaze position
`being identified from the position
`
`Ross expressly or inherently discloses these claim element(s).
`
`See 2a, which includes the same limitation.
`
`Accordingly, in view of the above, the Ross reference anticipates the claim element of the virtual
`space includes a target object (e.g., objects, characters) selectable by a gaze of the player (e.g.,
`interaction by repeatedly looking at an object).
`
`Two different combination of references are presented below.
`
`1) The limitations are obvious in view of Ross.
`
`“By way of a non-limiting example, the secondary, tertiary, etc., virtual reality content may be
`determined by system 10 based on the user's progression through a movie (e.g., as characters are
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 10 of 24
`
`Supercell
`Exhibit 1013
`Page 10
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`and direction of the body part of
`the player, moves outside the
`target object before the player
`completes selecting the target
`object, and
`further comprising providing,
`with the virtual image display
`apparatus, the to-be-provided
`information to the player when
`the gaze position of the player
`moves outside the target object
`before the player completes
`selecting the target object.
`
`Ross
`introduced in the movie they may be added to the secondary, tertiary, etc., virtual reality content),
`based on a user's interaction with the movie (e.g., a user may repeatedly look over to a particular
`character in the movie which the user is then able to interact with as secondary, tertiary, etc.,
`virtual reality content), and/or based on other information. The user may interact with the
`secondary, tertiary, etc., virtual reality content (e.g., socialize with a movie character) and then
`return his view back to the movie wherein the movie has been adjusted by interaction component
`28 to reflect the user's interaction with the secondary, tertiary, etc., virtual reality content movie
`character.” 18:58-19:5
`
`“The user may interact with the secondary, tertiary, etc., virtual reality content and then return his
`view back to the movie wherein the movie has been adjusted to reflect the user's interaction with the
`secondary, tertiary, etc., virtual reality content. In some implementations, while a user views primary
`virtual reality content, system 10 may be configured to display a “pop-up” cue indicating that
`additional virtual reality content (e.g., additional characters) is available in a different (e.g.,
`secondary, tertiary, etc.) content area.” 7:41-50
`
`Ross discloses that a user may interact with objects, e.g., a movie character, in a secondary content
`area, after which the primary area will be modified after the user’s view returns to the primary area.
`Ross also discloses that the primary VR content area may be configured to display a pop-up cue
`indicating additional content such as characters are available in a different content area. This clearly
`implies that if a user does not complete selection by repeatedly looking at an object/character, when
`the user subsequently moves outside the area of the object/character (i.e., the secondary area), the
`system provides a pop-up cue to the user which has information informing the user that more content is
`available.
`
`While it may be debatable as to whether Ross is actually checking whether the user has not completed
`selection of the object/character by repeatedly looking at it enough times, and only providing the pop-
`up in response to such a condition, such a modification of Ross, if necessary, would be obvious to try
`as it applies a known technique of warning a user that something the user selected is not completed
`(see, e.g., US Pat. Pub. 2004/0242322 at [0085] providing an error when a user has not completed a
`game session after initiating the session) to the established process in Ross of notifying the user that
`additional content, e.g., content that the user did not finish selecting, is available.
`
`
`Invalidity of U.S. Patent No. 10,406,432 in View of Ross
`
`Page 11 of 24
`
`Supercell
`Exhibit 1013
`Page 11
`
`
`
`
`
`#
`
`U.S. Patent No. 10,406,432
`
`Ross
`Accordingly, in view of the above, the Ross reference teaches the claim element of wherein the
`information providing condition includes a condition wherein a gaze position of the player (e.g., the
`view direction of the user), the gaze position being identified from the position and direction of the
`body part of the player (e.g., the user’s body, head, position, directional information, see above),
`moves outside the target object (e.g., returns to the primary area) before the player completes
`selecting the target object (e.g., by repeatedly looking at the object/character), and
`further comprising providing (e.g., displaying), with the virtual image display apparatus, the to-be-
`provided information (e.g., the pop up cue) to the player when the gaze position of the player moves
`outside the target object (e.g., to the primary area) before the player completes selecting the target
`object (e.g., as the user has not repeatedly looked at the object to select it).
`
`2) The limitations are obvious in view of Ross and secondary reference Lankford Nov. 2000 “Effective
`eye-gaze input into Windows.”
`
`“The Eye-gaze Response Interface Computer Aid (ERICA) is a computer that tracks eye movement.
`The system noninvasively tracks where a user is looking by monitoring the user's eye movements
`through a camera mounted underneath the computer monitor. ERICA was originally developed to
`assist the disabled by providing them with a means to communicate. The device has helped numerous
`disabled individuals since its inception in 1983. ERICA recently moved to the Windows platform. The
`mouse cursor now points to where the user is looking on the screen. To fully take advantage of the
`capabilities Windows has to offer, a method to click and to type with the eye needed to be incorporated
`into the graphical user interface (GUI).” Section 1.
`
`“The system uses dwell time to provide the user with access to various mouse clicking actions. When
`the user fixates (focuses at a point on the screen and keeps the mouse cursor stationary) for a
`predetermined amount of time on the computer display, a red rectangle appears, centered on the
`point of fixation. This rectangle begins collapsing in on itself. The rectangle serves as a visual cue to
`the user that if they keep fixating at that point, then they will perform a mouse control action at th