Prosecution Insights
Last updated: April 19, 2026
Application No. 17/653,034

MOVING A DIRECTION OF GAZE OF AN AVATAR

Non-Final OA §103§112
Filed
Mar 01, 2022
Examiner
PATEL, SHIVANG I
Art Unit
2615
Tech Center
2600 — Communications
Assignee
Cavendish Capital LLC
OA Round
7 (Non-Final)
74%
Grant Probability
Favorable
7-8
OA Rounds
2y 4m
To Grant
93%
With Interview

Examiner Intelligence

Grants 74% — above average
74%
Career Allow Rate
309 granted / 415 resolved
+12.5% vs TC avg
Strong +18% interview lift
Without
With
+18.5%
Interview Lift
resolved cases with interview
Typical timeline
2y 4m
Avg Prosecution
22 currently pending
Career history
437
Total Applications
across all art units

Statute-Specific Performance

§101
10.3%
-29.7% vs TC avg
§103
57.8%
+17.8% vs TC avg
§102
16.7%
-23.3% vs TC avg
§112
13.5%
-26.5% vs TC avg
Black line = Tech Center average estimate • Based on career data from 415 resolved cases

Office Action

§103 §112
DETAILED ACTION Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Continued Examination Under 37 CFR 1.114 A request for continued examination under 37 CFR 1.114, including the fee set forth in 37 CFR 1.17(e), was filed in this application after final rejection. Since this application is eligible for continued examination under 37 CFR 1.114, and the fee set forth in 37 CFR 1.17(e) has been timely paid, the finality of the previous Office action has been withdrawn pursuant to 37 CFR 1.114. Applicant's submission filed on 1/9/2026 has been entered. Response to Arguments Applicant’s arguments, see page 7-9, filed 1/9/2026 with respect to 35 USC §103 rejection of claims 1-2,4-12,14-20 have been fully considered but were not persuasive. Applicant has amended claims and argues previously cited references do not disclose amended claim limitations. Applicant argues Du in general discloses changing the gaze direction of a person from image data and animate an avatar to change a gaze direction, including rotating a mesh representation for the avatar. Applicant points to Grant to discloses improving eye movements in the virtual world including a description of managing eye movements of an avatar to redirect viewing angle and a system for redirecting a head position if the redirection of the eyes surpasses a threshold. Applicant argues amended claim language avatar gaze movement is based on a change in direction of the gaze of the avatar from the first virtual item to the second virtual item and avatar gaze movement is then separated into a set of components, including 1) an angular eye movement, 2) an angular head movement, and 3) an angular torso movement. Applicant argues at best Du discloses rotating a mesh of a head to change direction of a gaze and Grant describes changing the eye position of the avatar and if necessary changing the head position. Applicant argues Du nor Grant disclose component movement for the avatar gaze movement that includes an angular torso movement. Applicant further argues Grant at best describes a posture selection, which may involve directional movement of a torso, but is silent regarding that torso movement being a movement component based on a avatar gaze movement. Applicant argues Du and Grant are silent to determining the avatar gaze movement, and then separating the gaze movement into the individual component movements. In response, examiner first notes claimed separating the avatar gaze movement into a set of component movements is not described in instant application specification, close found paragraph is [0324] to disclose machine learning training process and lateral movements of humans are separated into four components. However paragraph [034] does not disclose separating avatar gaze movement as argued. Based on paragraph [0324] and what is known to one of ordinary skill in the art, examiner points to paragraph [0073-0074] of Grant to disclose avatar model includes posture model for posture. Grant’s posture includes body configuration\change including position of various body parts including torso as disclosed in paragraph [0074]. Applicant arguments that Du nor Grant does not disclose amended claim language of separating the gaze movement into the individual component movements is not persuasive as Grant’s posture discloses moveable body parts of avatar including torso. Instant application specification does not provide another definition of separating the gaze movement into the individual component movements and one of ordinary skill in the art would understand in the context of Grant’s movable body parts of avatar to include eye, head and torso movement. Claim Rejections - 35 USC § 112 The following is a quotation of the first paragraph of 35 U.S.C. 112(a): (a) IN GENERAL.—The specification shall contain a written description of the invention, and of the manner and process of making and using it, in such full, clear, concise, and exact terms as to enable any person skilled in the art to which it pertains, or with which it is most nearly connected, to make and use the same, and shall set forth the best mode contemplated by the inventor or joint inventor of carrying out the invention. The following is a quotation of the first paragraph of pre-AIA 35 U.S.C. 112: The specification shall contain a written description of the invention, and of the manner and process of making and using it, in such full, clear, concise, and exact terms as to enable any person skilled in the art to which it pertains, or with which it is most nearly connected, to make and use the same, and shall set forth the best mode contemplated by the inventor of carrying out his invention. Claim 1,11 rejected under 35 U.S.C. 112(a) or 35 U.S.C. 112 (pre-AIA ), first paragraph, as failing to comply with the written description requirement. The claim(s) contains subject matter which was not described in the specification in such a way as to reasonably convey to one skilled in the relevant art that the inventor or a joint inventor, or for applications subject to pre-AIA 35 U.S.C. 112, the inventor(s), at the time the application was filed, had possession of the claimed invention. Applicant has amended claims and recite “separating the avatar gaze movement into a set of component movements” however specifications do not disclose processing for separating movement into set of component movements. Closet found paragraph is [0325] which discloses “a machine learning process can be trained to learn how lateral movements of humans are separated into these four components”. Claim Rejections - 35 USC § 103 In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis (i.e., changing from AIA to pre-AIA ) for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status. The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. Claim(s) 1-2, 4-6,8-12,14-16,18-20 is/are rejected under 35 U.S.C. 103 as being unpatentable over Du et al (US 20230051409 A1) and Grant (US 20180268589 A1). Regarding claim 1, Du discloses a method for changing a direction of gaze of an avatar within a virtual environment of a video conference ([0033] virtual conferencing that includes rendering a conference display with animated avatars), the method comprises: sensing that a certain participant of a video conference changes his direction of gaze from a first virtual item to a second virtual item ([0057] gaze direction (gaze point) information can be determined for a user associated with the avatar to be animated); wherein the first virtual item and the second virtual item appear in a version of a virtual video conference environment that is displayed on a display of the certain participant ([0063] when a user (Alice) is looking at another person, e.g., another participant's avatar, in the VC display 330, as can be determined from provided eye-tracking gaze point information, Alice's avatar can then be animated to look at that person in the VC display 330. For example, as shown in FIG. 3, gaze information provided for Alice can indicate she is looking, within a threshold distance in the VC display 330, at Dave. That gaze information can then be used (e.g., transmitted to other VC participants) and an avatar for Alice can be animated to look at Dave in the VC display); wherein the sensing comprises image processing of consecutive images acquired by a camara that captures the participant ([0075] a sensing system (such as including sensors such as those described herein), a control system, at least one processor, and/or an outward facing image sensor, or camera, the smartglasses may include a gaze tracking device including, for example, one or more sensors, to detect and track eye gaze direction and movement. e.g., which information can be provided to virtual conference participants ), wherein different participants of the video conference receive different versions of the virtual video conference environment ([0065] Each VC device can send current gaze point location, as determined by the eye tracking algorithm, and/or an audio signal including speech of an associated VC participant to the VC server for broadcast to other VC participant(s), The VC devices can then use the provided information (e.g., images, photo, depth map, 3D mesh, gaze point and/or audio signal) to render gaze-aware 3D photorealistic avatars): determining an avatar gaze movement corresponding to the change in direction of the gaze by the certain participant based on a change in the direction of the gaze of the avatar from the first virtual item to second virtual item ([0005] he indication of the current gaze direction can include changes in the current gaze direction, [0057] captured gaze information and/or speech information can, also at block 230, be sent (continuously sent to reflect changes in gaze and audio stream) to other VC participants.) wherein, once applied, the set of component movements cause the avatar of the certain participant to shift his gaze from the first virtual item to the second virtual item ([0064] animation of gaze redirection for the VC participant Alice); wherein a combination of the set of component movements equals the virtual movement ([0094] Positioning of the computing device/virtual object by the user when incorporated into the AR space can allow the user to position the computing device so as to view the virtual object in certain manners in the AR space), Grant discloses separating the avatar gaze movement into a set of component movements comprising ([0073] avatar model (135) includes a posture model (241) that is configured to predict a posture (e.g., body configuration and/or its changes) of an avatar) an angular eye movement ([0048] the eye movements change the rendering of the eyes of the avatar (141) relative to the face of the avatar, the eye movements also change the viewing angle and/or focus of the avatar) an angular head movement ([0034] the adjustment can be made via an animation of the head movements of the avatar relative to the body of the avatar to simulate the turning of the head of the avatar). and an angular torso movement ([0074] The posture may include the relative positions of various movable body parts of the avatar (e.g., limbs, torso) with respect to each other and/or their relative movements with respect to each other.), Du and Grant are combinable because they are from the same field of invention. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify animated avatars of Du to include determining separating the avatar gaze movement into a set of component movements comprising an angular eye movement , an angular head movement and an angular torso movement as described by Grant. The motivation for doing so would have been to present eye movement and eye contact in virtual reality using a computer-based eye movement model (Grant, [0025]). Therefore, it would have been obvious to combine Du and Grant to obtain the invention as specified in claim 1. Regarding claim 2, Du discloses further comprising generating, by an alpha channel machine learning process, alpha channel information related to the certain participant, and updating the avatar using the alpha channel information ([0067] the selected images are referenced as Img2 and Img3 444, which can be blended using alpha blending with a blending ratio, [0032] machine learning techniques can be used to generate a set of synthesized images, e.g., based on a photograph of a person.) Regarding claim 4, Du discloses wherein the set of component movements are determined by a machine learning process and wherein the machine learning process is trained by a training process that comprises receiving videos of the certain participant while performing actual movement ([0061] The synthesized images 326 and 328 can be generated using respective ML models (neural networks) that are trained using an image animation model) and wherein the actual movements comprises at least one out of (a) an actual movement of eyes of the certain participant ([0061] an ML model can be trained, using a first order motion model, to generate synthesized mouth shape images and gaze direction image), and (b) one or more additional actual movements that differ from the gaze movement of eyes of the certain participant. Regarding claim 5, Du discloses wherein the set of component movements are determined by a machine learning process and wherein the machine learning process is trained by a training process that comprises receiving videos of multiple persons while performing actual movement ([0061] The synthesized images 326 and 328 can be generated using respective ML models (neural networks) that are trained using an image animation model) and wherein the actual movement comprise (a) an actual movement of eyes of the multiple persons, and (b) one or more additional gaze movement that differ from the actual movement of eyes of the multiple persons ([0040] an ML model for generating gaze direction images can be trained using images of people gazing in different directions, while another ML model for generating different mouth shapes can be trained using videos of people speaking, where synthesized mouth shapes can be based on pitch and amplitude of associated speech). Regarding claim 6, Du discloses wherein there is a difference between at least one actual movement of the certain participant that led to the change to the direction of gaze and between the set of component movements ([0069] Rotation of a 3D mesh of an associated animated avatar can also follow the current gaze point by rotating the 3D mesh.) Regarding claim 8, Du discloses wherein the first virtual item is a first participant of the video conference ([0058] Alice during a virtual conference) Regarding claim 9, Du discloses wherein the second virtual item is a second participant of the video conference ([0063] , gaze information provided for Alice can indicate she is looking, within a threshold distance in the VC display 330, at Dave). Regarding claim 10, Du discloses wherein the first virtual item differs from any participant of the video conference ([0058] Alice during a virtual conference with Charlie, Bob and Dave) Regarding claim 11, Du discloses a non-transitory computer readable medium for changing a direction of gaze of an avatar within a virtual environment of a video conference ([0078] the storage device 706 may be or contain a computer-readable medium, such as a floppy disk device, a hard disk device, an optical disk device, or a tape device, a flash memory or other similar solid state memory device, or an array of devices), the non-transitory computer readable medium stores instructions for: sensing that a certain participant of a video conference changes his direction of gaze from a first virtual item to a second virtual item ([0057] gaze direction (gaze point) information can be determined for a user associated with the avatar to be animated) wherein the first virtual item and the second virtual item appear in a version of a virtual video conference environment that is displayed on a display of the certain participant ([0063] when a user (Alice) is looking at another person, e.g., another participant's avatar, in the VC display 330, as can be determined from provided eye-tracking gaze point information, Alice's avatar can then be animated to look at that person in the VC display 330. For example, as shown in FIG. 3, gaze information provided for Alice can indicate she is looking, within a threshold distance in the VC display 330, at Dave. That gaze information can then be used (e.g., transmitted to other VC participants) and an avatar for Alice can be animated to look at Dave in the VC display); the sensing comprises image processing of consecutive images acquired by a camara that captures the participant ([0075] a sensing system (such as including sensors such as those described herein), a control system, at least one processor, and/or an outward facing image sensor, or camera, the smartglasses may include a gaze tracking device including, for example, one or more sensors, to detect and track eye gaze direction and movement. e.g., which information can be provided to virtual conference participants ), different participants of the video conference receive different versions of the virtual video conference environment ([0065] Each VC device can send current gaze point location, as determined by the eye tracking algorithm, and/or an audio signal including speech of an associated VC participant to the VC server for broadcast to other VC participant(s), The VC devices can then use the provided information (e.g., images, photo, depth map, 3D mesh, gaze point and/or audio signal) to render gaze-aware 3D photorealistic avatars) determining an avatar gaze movement corresponding to the change in direction of the gaze by the certain participant based on a change in the direction of the gaze of the avatar from the first virtual item to second virtual item ([0005] he indication of the current gaze direction can include changes in the current gaze direction, [0057] captured gaze information and/or speech information can, also at block 230, be sent (continuously sent to reflect changes in gaze and audio stream) to other VC participants.) wherein, once applied, the set of component movements cause the avatar of the certain participant to shift his gaze from the first virtual item to the second virtual item ([0064] animation of gaze redirection for the VC participant Alice); wherein a combination of the set of component movements equals the virtual movement ([0094] Positioning of the computing device/virtual object by the user when incorporated into the AR space can allow the user to position the computing device so as to view the virtual object in certain manners in the AR space), Grant discloses separating the avatar gaze movement into a set of component movements comprising ([0073] avatar model (135) includes a posture model (241) that is configured to predict a posture (e.g., body configuration and/or its changes) of an avatar) an angular eye movement ([0048] the eye movements change the rendering of the eyes of the avatar (141) relative to the face of the avatar, the eye movements also change the viewing angle and/or focus of the avatar) an angular head movement ([0034] the adjustment can be made via an animation of the head movements of the avatar relative to the body of the avatar to simulate the turning of the head of the avatar). and an angular torso movement ([0074] The posture may include the relative positions of various movable body parts of the avatar (e.g., limbs, torso) with respect to each other and/or their relative movements with respect to each other.), Du and Grant are combinable because they are from the same field of invention. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify animated avatars of Du to include determining separating the avatar gaze movement into a set of component movements comprising an angular eye movement , an angular head movement and an angular torso movement as described by Grant. The motivation for doing so would have been to present eye movement and eye contact in virtual reality using a computer-based eye movement model (Grant, [0025]). Therefore, it would have been obvious to combine Du and Grant to obtain the invention as specified in claim 11. Regarding claim 12, Du discloses stores instructions for generating, by an alpha channel machine learning process, alpha channel information related to the certain participant, and updating the avatar using the alpha channel information ([0067] the selected images are referenced as Img2 and Img3 444, which can be blended using alpha blending with a blending ratio, [0032] machine learning techniques can be used to generate a set of synthesized images, e.g., based on a photograph of a person.) Regarding claim 14, Du discloses wherein the set of component movements are determined by a machine learning process and wherein the machine learning process is trained by a training process that comprises receiving videos of the certain participant while performing actual movement ([0061] The synthesized images 326 and 328 can be generated using respective ML models (neural networks) that are trained using an image animation model) and wherein the actual movement comprise at least one out of (a) an actual movement of eyes of the certain participant ([0061] an ML model can be trained, using a first order motion model, to generate synthesized mouth shape images and gaze direction image), or (b) one or more additional actual movements that differ from the gaze movement of the certain participant. Regarding claim 15, Du discloses wherein the set of component movements are determined by a machine learning process and wherein the machine learning process is trained by a training process that comprises receiving videos of multiple persons while performing at least one actual movement ([0061] The synthesized images 326 and 328 can be generated using respective ML models (neural networks) that are trained using an image animation model) and wherein the at least one actual movement comprise (a) an actual movement of eyes of the multiple persons, and (b) one or more additional gaze movement that differ from the actual movement of eyes of the multiple persons ([0040] an ML model for generating gaze direction images can be trained using images of people gazing in different directions, while another ML model for generating different mouth shapes can be trained using videos of people speaking, where synthesized mouth shapes can be based on pitch and amplitude of associated speech). Regarding claim 16, Du discloses wherein there is a difference between at least one actual movement of the certain participant that led to the change to the direction of gaze and between the set of component movements ([0069] Rotation of a 3D mesh of an associated animated avatar can also follow the current gaze point by rotating the 3D mesh.) Regarding claim 17, Du discloses wherein the determining of the virtual movement of the avatar is made regardless of the at least one actual movement of the certain participant ([0075] a microphone of the smartglasses 696 can be used to capture speech of a virtual conference participant, and an audio signal corresponding with that speech can be provided to other participants for avatar animation using the approaches described herein). Regarding claim 18, Du discloses wherein the first virtual item is a first participant of the video conference ([0058] Alice during a virtual conference) Regarding claim 19, Du discloses wherein the second virtual item is a second participant of the video conference ([0063] , gaze information provided for Alice can indicate she is looking, within a threshold distance in the VC display 330, at Dave). Regarding claim 20, Du discloses wherein the first virtual item differs from any participant of the video conference ([0058] Alice during a virtual conference with Charlie, Bob and Dave) Conclusion Any inquiry concerning this communication or earlier communications from the examiner should be directed to SHIVANG I PATEL whose telephone number is (571)272-8964. The examiner can normally be reached on M-F 9-5am. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Alicia Harrington can be reached on (571) 272-2330. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of an application may be obtained from the Patent Application Information Retrieval (PAIR) system. Status information for published applications may be obtained from either Private PAIR or Public PAIR. Status information for unpublished applications is available through Private PAIR only. For more information about the PAIR system, see https://ppair-my.uspto.gov/pair/PrivatePair. Should you have questions on access to the Private PAIR system, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative or access to the automated information system, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. /SHIVANG I PATEL/Primary Examiner, Art Unit 2615
Read full office action

Prosecution Timeline

Mar 01, 2022
Application Filed
Mar 24, 2023
Non-Final Rejection — §103, §112
Jul 31, 2023
Response Filed
Aug 21, 2023
Final Rejection — §103, §112
Dec 26, 2023
Request for Continued Examination
Jan 04, 2024
Response after Non-Final Action
Feb 09, 2024
Non-Final Rejection — §103, §112
Aug 14, 2024
Response Filed
Sep 27, 2024
Final Rejection — §103, §112
Mar 28, 2025
Request for Continued Examination
Apr 02, 2025
Response after Non-Final Action
Apr 03, 2025
Non-Final Rejection — §103, §112
Aug 27, 2025
Interview Requested
Sep 03, 2025
Examiner Interview Summary
Sep 03, 2025
Applicant Interview (Telephonic)
Sep 09, 2025
Response Filed
Oct 08, 2025
Final Rejection — §103, §112
Jan 09, 2026
Request for Continued Examination
Jan 23, 2026
Response after Non-Final Action
Mar 04, 2026
Non-Final Rejection — §103, §112 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12602847
SYSTEMS AND METHODS FOR LAYERED IMAGE GENERATION
2y 5m to grant Granted Apr 14, 2026
Patent 12599838
APPARATUS AND METHODS FOR RECORDING AND REPORTING ABUSIVE ONLINE INTERACTIONS
2y 5m to grant Granted Apr 14, 2026
Patent 12592004
IMAGE PROCESSING DEVICE AND IMAGE PROCESSING METHOD
2y 5m to grant Granted Mar 31, 2026
Patent 12591947
DISTORTION-BASED IMAGE RENDERING
2y 5m to grant Granted Mar 31, 2026
Patent 12584296
Work Machine Display Control System, Work Machine Display System, Work Machine, Work Machine Display Control Method, And Work Machine Display Control Program
2y 5m to grant Granted Mar 24, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

7-8
Expected OA Rounds
74%
Grant Probability
93%
With Interview (+18.5%)
2y 4m
Median Time to Grant
High
PTA Risk
Based on 415 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month