throbber
www.uspto.gov
`
`UNITED STATES DEPARTMENT OF COMMERCE
`United States Patent and Trademark Office
`Address: COMMISSIONER FOR PATENTS
`P.O. Box 1450
`Alexandria, Virginia 22313-1450
`
`17/922,919
`
`11/02/2022
`
`Junji OTSUKA
`
`1946-1828
`
`4998
`
`Paratus Law Group, PLLC
`1765 Greensboro Station Place
`Suite 320
`Tysons Corner, VA 22102
`
`MCCULLEY, RYAN D
`
`2611
`
`MAIL DATE
`
`11/08/2024
`
`PAPER NUMBER
`
`DELIVERY MODE
`
`PAPER
`
`Please find below and/or attached an Office communication concerning this application or proceeding.
`
`The time period for reply, if any, is set in the attached communication.
`
`PTOL-90A (Rev. 04/07)
`
`

`

`Application No.
`Applicant(s)
`171922919
`OTSUKA etal.
`
`Office Action Summary Art Unit|AIA (FITF)StatusExaminer
`Ryan McCulley
`2611
`Yes
`
`
`
`-- The MAILING DATEof this communication appears on the cover sheet with the correspondence address --
`Period for Reply
`
`A SHORTENED STATUTORYPERIOD FOR REPLYIS SET TO EXPIRE 3 MONTHS FROM THE MAILING
`DATE OF THIS COMMUNICATION.
`Extensionsof time may be available underthe provisions of 37 CFR 1.136(a). In no event, however, may a reply betimely filed after SIX (6) MONTHSfrom the mailing
`date of this communication.
`If NO period for reply is specified above, the maximum statutory period will apply and will expire SIX (6) MONTHSfrom the mailing date of this communication.
`-
`- Failure to reply within the set or extended period for reply will, by statute, cause the application to become ABANDONED (35 U.S.C. § 133).
`Any reply received by the Office later than three months after the mailing date of this communication, evenif timely filed, may reduce any earned patent term
`adjustment. See 37 CFR 1.704(b).
`
`Status
`
`
`
`1)™) Responsive to communication(s)filed on 09 September 2024.
`C) A declaration(s)/affidavit(s) under 37 CFR 1.130(b) was/werefiled on
`2a)[¥) This action is FINAL.
`2b) (J This action is non-final.
`3) An election was madeby the applicant in responseto a restriction requirement set forth during the interview
`on
`; the restriction requirement and election have been incorporated into this action.
`4)(2) Since this application is in condition for allowance except for formal matters, prosecution as to the merits is
`closed in accordance with the practice under Exparte Quayle, 1935 C.D. 11, 453 O.G. 213.
`
`Disposition of Claims*
`1-20 is/are pending in the application.
`)
`Claim(s)
`5a) Of the above claim(s) _ is/are withdrawn from consideration.
`C} Claim(s)__ is/are allowed.
`Claim(s) 1-20 is/are rejected.
`(] Claim(s)__ is/are objectedto.
`C] Claim(s
`are subjectto restriction and/or election requirement
`)
`* If any claims have been determined allowable, you maybeeligible to benefit from the Patent Prosecution Highway program at a
`participating intellectual property office for the corresponding application. For more information, please see
`http://www.uspto.gov/patents/init_events/pph/index.jsp or send an inquiry to PPHfeedback@uspto.gov.
`
`) ) ) )
`
`Application Papers
`10) The specification is objected to by the Examiner.
`11)0) The drawing(s) filedon__ is/are: a)(J accepted or b)( objected to by the Examiner.
`Applicant may not request that any objection to the drawing(s) be held in abeyance. See 37 CFR 1.85(a).
`Replacement drawing sheet(s) including the correction is required if the drawing(s) is objected to. See 37 CFR 1.121(d).
`
`Priority under 35 U.S.C. § 119
`12)7) Acknowledgment is made of a claim for foreign priority under 35 U.S.C. § 119(a)-(d)or (f).
`Certified copies:
`c)Z None ofthe:
`b)() Some**
`a)C All
`1.1.) Certified copies of the priority documents have been received.
`2.2) Certified copies of the priority documents have been received in Application No.
`3.1.) Copies of the certified copies of the priority documents have been receivedin this National Stage
`application from the International Bureau (PCT Rule 17.2(a)).
`*“ See the attached detailed Office action for a list of the certified copies not received.
`
`Attachment(s)
`
`1) [[] Notice of References Cited (PTO-892)
`
`2) (J Information Disclosure Statement(s) (PTO/SB/08a and/or PTO/SB/08b)
`Paper No(s)/Mail Date
`U.S. Patent and Trademark Office
`
`3)
`
`4)
`
`(LJ Interview Summary (PTO-413)
`Paper No(s)/Mail Date
`(Qj Other:
`
`PTOL-326 (Rev. 11-13)
`
`Office Action Summary
`
`Part of Paper No./Mail Date 20241107
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 2
`
`DETAILED ACTION
`
`Notice of Pre-AlA or AIA Status
`
`The present application, filed on or after March 16, 2013,
`
`is being examined
`
`underthefirst inventor to file provisions of the AIA.
`
`Response to Amendment
`
`This Office Action is in response to Applicant's amendment/response filed on 09
`
`September 2024, which has been entered and made of record.
`
`Abstract
`
`The objection to the abstract is withdrawn in view of the replacement abstract.
`
`Claim Interpretation
`
`The claims are no longer interpreted under 35 U.S.C. 112(f) in view of the claim
`
`amendments.
`
`Claim Rejections - 35 USC § 112
`
`The previous rejections under 35 U.S.C. 112(b) are withdrawn in view of the
`
`claim amendments.
`
`Claim Rejections - 35 USC § 101
`
`The previous rejection of claim 20 under 35 U.S.C. 101 is withdrawn in view of
`
`the claim amendments.
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 3
`
`Response to Arguments
`
`Applicant's arguments filed 09 September 2024 have beenfully considered but
`
`they are not persuasive.
`
`Applicant argues “there is nothing in Kiemele that would fairly teach or suggest
`
`the elements presently recited” and “Kiemele clearly does not teach or suggest” the
`
`claimed features (Remarks, pg. 15). However, Applicant does not provide any
`
`reasoning as to how thecitations from Kiemele (provided in the previous Office Action
`
`and updated here) fail to teach the claim limitations. Kiemele recites “a visible-light
`
`camera configured to collect a visible-light
`
`image of a physical space” (para. 53).
`
`Kiemele further describes using a touch input, hand gestures, and even physical
`
`movementof the input device to create and update a virtual object in both displays (see
`
`Fig. 4). All of these types of input represent 3D information of areal space around a
`
`user. Therefore, Kiemele teaches acquiring three-dimensional
`
`information of a real
`
`space around a user, and dynamically changing a parameterof a virtual object based
`
`on the 3D information.
`
`Any remaining arguments are considered moot based on the foregoing.
`
`Claim Rejections - 35 USC § 102
`
`The following is a quotation of the appropriate paragraphs of 35 U.S.C. 102 that
`
`form the basis for the rejections under this section madein this Office action:
`
`A person shall be entitled to a patent unless —
`
`(a)(1) the claimed invention was patented, described in a printed publication, orin public use,
`on sale, or otherwise available to the public beforethe effectivefiling date of the claimed
`invention.
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 4
`
`Claims 1-4 and 18-20 are rejected under 35 U.S.C. 102(a)(1) as being
`
`anticipated by Kiemele et al. (US 2019/0065026; hereinafter “Kiemele’).
`
`Regarding claim 1, Kiemele discloses An information processing apparatus
`
`comprising: circuitry (“one or more processors,” para. 64) configured to acquire
`
`three-dimensional information of a real space around a user (“User input received
`
`by the touch sensor mayinclude touch, hover, and/or hand gesture input,” para. 15;
`
`“employ a visible light and/or depth camera to locate and segment image data
`
`corresponding to input device 110,” para. 17; “movementof touch-sensitive input device
`
`may effect the rendering of visual content,” para. 30), and dynamically change at
`
`least one parameter related to display of a virtual object in an image (“a user can
`
`draw 2D or 3D shapes and imagesvia the application of touch input,” para. 27; “edit a
`
`representation of visual content,” para. 33), the at least one parameter controlling
`
`display of the virtual object on each display device of a plurality of display
`
`devices according to an image expression method of expressing the image, the
`
`method being assigned for display of the image by each display device of the
`
`plurality of display devices that display the image including the virtual object (see
`
`Figs. 4-5), wherein the at least one parameter is changed according to the
`
`acquired three-dimensional information of the real space around the user (Fig. 4
`
`illustrates a user input, which reflects a real space 3D position of a user’s finger/nand or
`
`a real space 3D position of the handheld device, resulting in changesto the virtual
`
`object 404 in both the handheld device and the HMD viewpoint).
`
`Regarding claim 2, Kiemele discloses wherein the plurality of display devices
`
`includea first display device that is controlled to display scenery of a real space
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 5
`
`in which the virtual object is virtually arranged, the scenery being viewed from a
`
`first viewpoint defined as a viewpoint of the user in the real space (“an augmented-
`
`reality experience ... to give the wearer a view of the surrounding real-world physical
`
`space ... the user will see augmented-reality objects that are not actually present in the
`
`physical space,” para. 45; ), and a second display device that is controlled to
`
`display an image of the virtual object (“a representation of visual content rendered on
`
`touch-sensitive input device,” para. 20).
`
`Regarding claim 3, Kiemele discloses wherein the circuitry dynamically
`
`changes the parameterfor controlling the first display device according to the
`
`three-dimensional information of the real space around the user acquired froma
`
`real space information acquisition device (“representation 204 [of Fig. 2] is rendered
`
`in a 3D location proximate to input device 110 where the representation appears to float
`
`over the input device,” para. 20; “varying visual content rendered in response to
`
`movementof input device,” para. 30; “any suitable input, such as a hand gesturein
`
`which the user swipes from a location proximate to the input deviceinto the virtual
`
`reality experience,” para. 32; “The one or more outward facing image sensors 812A,
`
`812B may be configured to measure physical environmentattributes of a physical
`
`space,” para. 53).
`
`Regarding claim 4, Kiemele discloses wherein the real space information
`
`acquisition device includes at least one imaging device configured to image the
`
`real space around the user or a distance measuring device configured to acquire
`
`depth information of the real space around the user (“HMD device 104 [of Fig. 1]
`
`includes a sensor system 116 for tracking touch-sensitive input device 110 ... sensor
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 6
`
`system 116 may employ a visible light and/or depth camera to locate and segment
`
`image data corresponding to input device 110,” para. 17).
`
`Regarding claim 18, Kiemele discloses wherein thefirst display device
`
`superimposes and displays an imageof the virtual object on an imageof thereal
`
`space, projects and displays the image of the virtual object in the real space, or
`
`projects and displays the image of the virtual object on a retina of the user (“he
`
`virtual-reality experience includes an augmented-reality experience in which the near-
`
`eye display is wholly or partially transparent from the perspective of the wearer, to give
`
`the wearer a view of the surrounding real-world physical space,” para. 45).
`
`Regarding claims 19 and 20, they are rejected using the same citations and
`
`rationales described in the rejection of claim 1.
`
`Claim Rejections - 35 USC § 103
`
`The following is a quotation of 35 U.S.C. 103 which forms the basis forall
`
`obviousness rejections set forth in this Office action:
`
`Apatent fora claimed invention may notbe obtained, notwithstanding thatthe claimed
`invention is not identically disclosed as set forth in section 102, if the differences between the
`claimed invention and the prior artare suchthat the claimed invention as a whole would have
`been obvious beforethe effective filing date of the claimed inventionto a person having
`ordinary skill in the art to which the claimed invention pertains. Patentability shall notbe
`negated by the manner in whichthe invention was made.
`
`Claim 5 is rejected under 35 U.S.C. 103 as being unpatentable over Kiemele in
`
`view of Panseet al. (US 2021/0398316; hereinafter “Panse’”).
`
`Regarding claim 5, Kiemele does not disclose wherein whena region in which
`
`a shielding object located betweenthevirtual object and the user is present in the
`
`real space or a region in which the three-dimensional information cannot be
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 7
`
`acquired is detected on the basis of the three-dimensional information, the
`
`circuitry is further configured to set the region as an occlusion region, and
`
`change a display position or a display form of the virtual object or a movement
`
`amountof the virtual object in moving image display on thefirst display device
`
`so as to reduce a region in whichthevirtual object and the occlusion region are
`
`superimposed.
`
`In the same art of augmented reality, Panse teaches wherein whena region in
`
`which a shielding object located between the virtual object and the useris
`
`presentin the real space ... the circuitry is further configured to set the region as
`
`an occlusion region, and changea display position or a display form of the virtual
`
`object ... on the first display device so as to reduce a region in whichthevirtual
`
`object and the occlusion region are superimposed (“as shown in FIG. 5E, a portion
`
`or an entirety of virtual object 54 may be positioned behind physical object 20 whereby
`
`physical object 20 blocks a visualization of such portion of virtual object 54 or an entirety
`
`of virtual object 54 ... virtual object 54 may alternatively be moved within augmented
`
`reality display,” para. 92).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Panse to
`
`Kiemele. The motivation would have been “fora functional visualization by a user of the
`
`AR device’ (Panse, para. 135).
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 8
`
`Claim 6 is rejected under 35 U.S.C. 103 as being unpatentable over the
`
`combination of Kiemele and Panse, and further in view of Dedonato et al. (US
`
`2021/0150818; hereinafter “Dedonato’).
`
`Regarding claim 6, the combination of Kiemele and Panse doesnotdisclose
`
`wherein the circuitry controls the first display device so as to display another
`
`virtual object in an indefinite region where the three-dimensional information
`
`cannot be acquired.
`
`In the same art of augmented reality, Dedonato teaches wherein thecircuitry
`
`controls the first display device so as to display anothervirtual object in an
`
`indefinite region where the three-dimensional information cannot be acquired
`
`(“update the virtual content to include a graphic associated with the unmapped location;
`
`direct the user to observe the graphic,” para. 7).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Dedonato to the
`
`combination of Kiemele and Panse. The motivation would have been “forfacilitating
`
`virtual or augmented reality interaction” (Dedonato, para. 4).
`
`Claims 7 and 8 are rejected under 35 U.S.C. 103 as being unpatentable over
`
`Kiemele in view of Kimura et al. (US 2021/0142552; hereinafter “Kimura’).
`
`Regarding claim 7, Kiemele does not disclose acquire position information
`
`including distance information and positional relationship information between
`
`the virtual object and the user in the real space, and wherein the circuitry
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 9
`
`dynamically changes the parameterfor controlling thefirst display device
`
`according to the position information.
`
`In the same art of augmented reality, Kimura teaches acquire position
`
`information including distance information and positional relationship
`
`information between the virtual object and the user in the real space, and wherein
`
`the circuitry dynamically changes the parameterfor controlling the first display
`
`device according to the position information (“calculates the display distance from
`
`the user to the display position of the virtual object ... when the display format
`
`determination unit has determined that the calculated display distance is equal to or
`
`larger than a... sets the display resolution of the virtual object,” paras. 114-116).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Kimura to
`
`Kiemele. The motivation would have been to improve a user’s visual understanding of
`
`the virtual object.
`
`Regarding claim 8, the combination of Kiemele and Kimura renders obvious
`
`wherein the circuitry performs control such that a display area of the virtual
`
`object to be displayed on the first display device increases as a distance between
`
`the virtual object and the user increases (“a second distance shorter than thefirst
`
`distance ... when the display distance is equal to thefirst distance, the control unit
`
`increases display area of the virtual object as comparedto a casein which the display
`
`distance is equal to the second distance,” para. 229; see claim 7 for motivation to
`
`combine).
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 10
`
`Claim 9 is rejected under 35 U.S.C. 103 as being unpatentable over the
`
`combination of Kiemele and Kimura, and further in view of Shinohara (US
`
`2011/0261076).
`
`Regarding claim 9, the combination of Kiemele and Kimura does disclose
`
`wherein the circuitry performs control such that a display change amountin
`
`moving image display of the virtual object to be displayed on thefirst display
`
`device increases as a distance betweenthevirtual object and the userincreases.
`
`In the same art of displaying virtual objects, Shinohara teaches wherein the
`
`circuitry performs control such that a display change amount in moving image
`
`display of the virtual object to be displayed on thefirst display device increases
`
`as a distance betweenthevirtual object and the user increases (“the viewpoint
`
`movement speed determination means determines the movement speed so that the
`
`movement speed increases in value as the virtual object-to-viewpoint distance
`
`increases in value. With this configuration, the view field flows at a lower speed as the
`
`distance between the virtual object and the viewpoint becomes smaller,” para. 24).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Shinohara to the
`
`combination of Kiemele and Kimura. The motivation would have beento display
`
`movement“more precisely” (Shinohara, para. 24).
`
`Claim 10 is rejected under 35 U.S.C. 103 as being unpatentable over the
`
`combination of Kiemele and Kimura, and further in view of Mathey-Owensetal. (US
`
`2018/0005438; hereinafter “Mathey-Owens’).
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 11
`
`Regarding claim 10, the combination of Kiemele and Kimura doesnot disclose
`
`wherein the circuitry performs control to further smooth a trajectory in moving
`
`image display of the virtual object to be displayed on thefirst display device as a
`
`distance betweenthe virtual object and the user increases.
`
`In the same art of displaying virtual objects, Mathey-Owens teaches wherein the
`
`control section performs control to further smootha trajectory in moving image
`
`display of the virtual object to be displayed on thefirst display device as a
`
`distance betweenthe virtual object and the user increases (“the smoothing
`
`algorithm is configured to ‘smooth’ the motion and/or rotation of the virtual element,”
`
`para. 27; “perform less smoothing when the user focuses on the large virtual balloon
`
`that is closer .... perform more smoothing when the user focuses on the small virtual
`
`balloon that is further away,” para. 28).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Mathey-Owensto
`
`the combination of Kiemele and Kimura. The motivation would have been to “improve
`
`the user experience” (Mathey-Owens,para. 21).
`
`Claim 11 is rejected under 35 U.S.C. 103 as being unpatentable over the
`
`combination of Kiemele and Kimura, and further in view of Chang et al. (US
`
`2016/0163110; hereinafter “Chang”).
`
`Regarding claim 11, the combination of Kiemele and Kimura doesnot disclose
`
`the circuitry dynamically changes a display change amountof the virtual object
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 12
`
`to be displayed on thefirst display device, the display change amount being
`
`changed byan input operation of the user, according to the position information.
`
`In the same art of displaying virtual object, Chang teaches thecircuitry
`
`dynamically changes a display change amountofthevirtual object to be
`
`displayed on thefirst display device, the display change amount being changed
`
`by an input operation of the user, according to the position information (“the
`
`adjustment speedin the size of the virtual object is determined according to the
`
`movement speed of the user ... and the distance between the user andthevirtual
`
`object,” para. 38).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Chang to the
`
`combination of Kiemele and Kimura. The motivation would have beento provide
`
`additional control to a user.
`
`Claim 12 is rejected under 35 U.S.C. 103 as being unpatentable over Kiemele in
`
`view of Lindeman (US 2016/0196694).
`
`Regarding claim 12, Kiemele does not disclose the circuitry controls the
`
`second display device so as to display an imageof the virtual object visually
`
`recognized from a second viewpoint different from the first viewpoint in the real
`
`space.
`
`In the same art of augmented reality, Lindeman teaches thecircuitry controls
`
`the second display device so as to display an image of the virtual object visually
`
`recognized from a second viewpoint different from the first viewpoint in the real
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 13
`
`space(“the virtual world on the tablet can be generated from a different point of view
`
`than the virtual world of the HMD,” para. 57).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Lindeman to
`
`Kiemele. The motivation would have been to allow a user more control of the
`
`visualization.
`
`Claim 13 is rejected under 35 U.S.C. 103 as being unpatentable over the
`
`combination of Kiemele and Lindeman, and further in view of Shpigelman (US
`
`2015/0302651).
`
`Regarding claim 13, the combination of Kiemele and Lindeman does not disclose
`
`wherein the second viewpointis virtually arranged on the virtual object.
`
`In the same art of augmented reality, Shpigelman teaches wherein the second
`
`viewpointis virtually arranged on thevirtual object (“A camera position may be
`
`associated with the displaying of the first or second of the two or more scenes, and the
`
`position maybe a point of view of a character within the respective first or second of the
`
`two or more scenes. The method mayfurther include selecting a chapter by default or
`
`receiving a selection from a user of a character whose point of view is to be adopted,
`
`and the displaying may include placing the virtual camera at the location of the selected
`
`character. The method may further include displaying an avatar associated with a user
`
`in the displayed first or second of the two or more scenes,” para.8).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Shpigelman to
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 14
`
`the combination of Kiemele and Lindeman. The motivation would have been “the user
`
`may advantageously employ the VR/AR functionality to become immersed”
`
`(Shpigelman, para. 33).
`
`Claims 14-17 are rejected under 35 U.S.C. 103 as being unpatentable over
`
`Kiemele in view of Zhu et al. (“BISHARE: Exploring Bidirectional Interactions Between
`
`Smartphones and Head-Mounted Augmented Reality”; hereinafter “Zhu’).
`
`Regarding claim 14, Kiemele does not disclose wherein the circuitry changes
`
`a display change amountofthe virtual object to be displayed on eachofthefirst
`
`and second display devices in moving image display according to the method of
`
`expressing the image assigned to each of the first and second display devices for
`
`displaying the image.
`
`In the same art of multi-device augmented reality, Zhu teaches wherein the
`
`circuitry changes a display change amountof thevirtual object to be displayed
`
`on each of the first and second display devices in moving image display
`
`according to the method of expressing the image assigned to eachofthefirst
`
`and second display devices for displaying the image (“to support content transfer in
`
`a phone-centric context (D4P), content on the phone can be dragged using a touch
`
`operation, and once reaching the edge, the movement can continue as a freehand
`
`pinch gesture [38] (D5P). In a HMD-centric setting, a user could combine spatial
`
`movementsof the phoneitself, with 2D touchscreen gestures on the phone, to perform
`
`manipulations of spatial content,” pg. 5, sec. D5).
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 15
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Zhu to Kiemele.
`
`The motivation would have been “The combination of mobile + AR as an entire system
`
`is very good, intuitive” (Zhu, pg. 9, col. 1).
`
`Regarding claim 15, Kiemele does not disclose acquire a selection result
`
`indicating whether the user has selected oneof thefirst display device and the
`
`second display device as an input device, and wherein the circuitry dynamically
`
`changes a display change amountof the virtual object changed by an input
`
`operation of the user according to the selection result.
`
`In the same art of multi-device augmented reality, Zhu teaches acquire a
`
`selection result indicating whether the user has selected oneof thefirst display
`
`device and the second display device as an input device, and wherein the
`
`circuitry dynamically changes a display change amountofthe virtual object
`
`changed byan input operation of the user according to the selection result (“In a
`
`spatial-centric task, a user may wish to use the smartphone as a 6DOF device to
`
`perform ray casting for object selection of 3D components [41] (D1H). Using the
`
`touchscreen, recasting could be enhanced with gestures, for example, to cycle between
`
`overlapping objects, or define a subsequent manipulation,” pg. 5, col. 1; “Alternatively,
`
`objects in the spatial environment can be accessed using freehand gestures. An air tap
`
`and dragging gesture can be used to select and move an object in 3D space,” pg. 6, col.
`
`2).
`
`Before the effectivefiling date of the claimed invention,
`
`it would have been
`
`obvious to one having ordinary skill
`
`in the art to apply the teachings of Zhu to Kiemele.
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 16
`
`The motivation would have been “The combination of mobile + AR as an entire system
`
`is very good, intuitive” (Zhu, pg. 9, col. 1).
`
`Regarding claim 16, the combination of Kiemele and Zhu renders obvious
`
`wherein the circuitry acquires the selection result on a basis of a detection result
`
`of a line-of-sight of the user from a line-of-sight detection device (“Another
`
`commonform of input in AR is Gaze [35], which is commonly combined with freehand
`
`gestures [35, 44, 54], such as the air tap technique used for selection,” Zhu, pg. 2, col.
`
`2; see claim 15 for motivation to combine).
`
`Regarding claim 17, the combination of Kiemele and Zhu renders obvious
`
`wherein the circuitry acquires the selection result on a basis of a detection result
`
`of a gesture of the user from a gesture detection device (“The use of free-hand
`
`gestures,
`
`local touch events, and controller-based input events can all be combined to
`
`create new forms of cross-platform gestures that incorporate both input platforms,” Zhu,
`
`pg. 5, sec. D5; see claim 15 for motivation to combine).
`
`Conclusion
`
`Applicant's amendment necessitated any new ground(s) of rejection presented in
`
`this Office action. Accordingly, THIS ACTION IS MADE FINAL. See MPEP § 706.07(a).
`
`Applicant is reminded of the extension of time policy as set forth in 37 CFR 1.136(a).
`
`A shortened statutory period for reply to this final action is set to expire THREE
`
`
`
`MONTHS from the mailing date of this action. In the eventafirst reply is filed within
`
`TWO MONTHS of the mailing date of this final action and the advisory action is not
`
`mailed until after the end of the THREE-MONTHshortened statutory period, then the
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 17
`
`shortened statutory period will expire on the date the advisory action is mailed, and any
`
`extension fee pursuant to 37 CFR 1.136(a) will be calculated from the mailing date of
`
`the advisory action. In no event, however, will the statutory period for reply expire later
`
`than SIX MONTHS from the date of this final action.
`
`Anyinquiry concerning this communication or earlier communications from the
`
`examiner should be directed to Ryan McCulley whose telephone numberis (571)270-
`
`3754. The examiner can normally be reached Monday through Friday, 8:00am -
`
`4:30pm.
`
`Examinerinterviews are available via telephone,
`
`in-person, and video
`
`conferencing using a USPTO supplied web-based collaboration tool. To schedule an
`
`interview, applicant is encouraged to use the USPTO AutomatedInterview Request
`
`(AIR) at http:/Awww.uspto.gov/interviewpractice.
`
`If attempts to reach the examiner by telephone are unsuccessful, the examiner's
`
`supervisor, Kee Tung can be reached on (571) 272-7794. The fax phone numberfor the
`
`organization where this application or proceeding is assigned is 571-273-8300.
`
`Information regarding the status of published or unpublished applications may be
`
`obtained from Patent Center. Unpublished application information in Patent Centeris
`
`available to registered users. To file and manage patent submissions in Patent Center,
`
`visit: httos://patentcenter.uspto.gov. Visit https:/Avww.uspto.gov/patents/apply/patent-
`
`center for more information about Patent Center and
`
`https://www.uspto.gov/patents/docx for information aboutfiling in DOCX format. For
`
`additional questions, contact the Electronic Business Center (EBC) at 866-217-9197
`
`

`

`Application/Control Number: 17/922,919
`Art Unit: 2611
`
`Page 18
`
`(toll-free).
`
`If you would like assistance from a USPTO Customer Service
`
`Representative, call 800-786-9199 (IN USA OR CANADA)or 571-272-1000.
`
`/RYAN MCCULLEY/
`Primary Examiner, Art Unit 2611
`
`

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket