Prosecution Insights
Last updated: April 19, 2026
Application No. 17/159,977

IMAGE SYNTHESIS USING ONE OR MORE NEURAL NETWORKS

Non-Final OA §103
Filed
Jan 27, 2021
Examiner
MEROUAN, ABDERRAHIM
Art Unit
2683
Tech Center
2600 — Communications
Assignee
Nvidia Corporation
OA Round
7 (Non-Final)
73%
Grant Probability
Favorable
7-8
OA Rounds
3y 2m
To Grant
90%
With Interview

Examiner Intelligence

Grants 73% — above average
73%
Career Allow Rate
481 granted / 659 resolved
+11.0% vs TC avg
Strong +17% interview lift
Without
With
+17.4%
Interview Lift
resolved cases with interview
Typical timeline
3y 2m
Avg Prosecution
21 currently pending
Career history
680
Total Applications
across all art units

Statute-Specific Performance

§101
8.3%
-31.7% vs TC avg
§103
55.4%
+15.4% vs TC avg
§102
20.9%
-19.1% vs TC avg
§112
8.7%
-31.3% vs TC avg
Black line = Tech Center average estimate • Based on career data from 659 resolved cases

Office Action

§103
DETAILED ACTION Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Continued Examination Under 37 CFR 1.114 1. A request for continued examination under 37 CFR 1.114, including the fee set forth in 37 CFR 1.17(e), was filed in this application after final rejection. Since this application is eligible for continued examination under 37 CFR 1.114, and the fee set forth in 37 CFR 1.17(e) has been timely paid, the finality of the previous Office action has been withdrawn pursuant to 37 CFR 1.114. Applicant's submission filed on 10/14/2025 has been entered. Claim Rejections - 35 USC § 103 2. In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status. 3. The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102 of this title, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. 4. Claims 1, 2, 7, 8, 11, 13, 14, 17, 19, 20, 23, 25, 26 and 29 are rejected under 35 U.S.C. 103 as being unpatentable over Li et al., US 2019/0244329 A1, and further in view of Park et al. US 2022/0148241 A1. 5. As per claim 1, Li discloses: One or more processors, comprising: circuitry to use one or more neural networks to generate one or more images depicting an object by modifying one or more features of the object based, at least in part, on: (Li, [0027]:” The photo style transfer neural network model 110 receives a photorealistic content image I.sub.C and a photorealistic style image I.sub.S and generates a stylized photorealistic image Y that includes the content of the content image modified according to the style image.”, and [0029]:”FIG. 1B illustrates a style image, content image, and stylized photorealistic image, in accordance with an embodiment. The photorealistic style image I.sub.S is input to the encoder 165 and the photorealistic content image I.sub.C is input to the second encoder 165. The photorealistic style image and the photorealistic content image I.sub.C are processed by the photo style transfer neural network model 110 to produce the stylized photorealistic image Y. The cloud pattern in the photorealistic content image is retained in the stylized photorealistic image while a blue color of the sky and the green color of the landscape in the photorealistic style image appear in the stylized photorealistic image—the color of the sky and the landscape areas is changed compared with the photorealistic content image. The shape of the road in the stylized photorealistic image is consistent with the road in the photorealistic content image, while the color is changed to be similar to the road in the photorealistic style image. In addition to transferring color, the photo style transfer neural network model 110 may also be configured to synthesize patterns contained in the photorealistic content image, such as a cloud, snow, rain, and the like, in the stylized photorealistic image to be consistent with the content of the photorealistic content image.”) 6. Li doesn’t expressly disclose: one or more appearance features, extracted by a first encoder of the one or more neural networks, of one or more first objects from one or more first images and one or more structural features, extracted by a second encoder of the one or more neural networks, of one or more second objects from one or more second images; and a decoding of the one or more appearance features of the first object and the one or more structural features of the second object, that have been encoded into one or more latent spaces of the one or more neural networks. 7. Park discloses: one or more appearance features, extracted by a first encoder of the one or more neural networks, of one or more first objects from one or more first images (Park, [0063], “In particular, the deep image manipulation system 102 utilizes the encoder neural network 306 to extract a structure code 308 and a texture code 310 from the first digital image 302. Indeed, the deep image manipulation system 102 applies the encoder neural network 306 to the first digital image 302 to extract structural features for the structure code 308 and textural features for the texture code 310.”, and [0074], “In some embodiments, the encoder neural network 306 (E) includes or represents two different encoders: a structural encoder neural network E.sub.s and a textural encoder neural network E.sub.t that extract structure codes and texture codes, respectively.”,and [0016], “In other embodiments, the encoder manager 1002 extracts a structure code from a first digital image and extracts a texture code from a second digital image.”) and one or more structural features, extracted by a second encoder of the one or more neural networks, of one or more second objects from one or more second images; (Park ,[0064],” In a similar fashion, the deep image manipulation system 102 utilizes the encoder neural network 306 to extract the structure code 312 and the texture code 314 from the second digital image 304. More specifically, the deep image manipulation system 102 extracts structural features from the second digital image 304 for the structure code 312. In addition, the deep image manipulation system 102 extract textural features from the second digital image 304 for the texture code 314.”, and [0074], “In some embodiments, the encoder neural network 306 (E) includes or represents two different encoders: a structural encoder neural network E.sub.s and a textural encoder neural network E.sub.t that extract structure codes and texture codes, respectively.” and [0016], “In other embodiments, the encoder manager 1002 extracts a structure code from a first digital image and extracts a texture code from a second digital image.”) and a decoding of the one or more appearance features of the first object and the one or more structural features of the second object, that have been encoded into one or more latent spaces of the one or more neural networks. (Park, [0070], “where x.sup.1 represents a latent code representation of the first digital image 402, x.sup.2 represents a latent code representation of the second digital image 404, z.sub.s.sup.1 represents the structure code 406 from the first digital image 402, z.sub.t.sup.2 represents the texture code 412 from the second digital image 404, z.sub.l.sup.1 represents the scene layout of x.sup.1, and the other terms are defined above.”, and [0074], “where x represents the first digital image 402, H represents the height of the image, W represents the width of the image, and 3 is the number of channels in an RGB image (i.e. red, green, and blue). For example, the encoder neural network 306 maps the first digital image 402 to a latent space Z, and the generator neural network 318 generates the reconstructed digital image 418 from the encoding in the latent space Z.”) 8. Park is analogous art with respect to Li because they are from the same field of endeavor, namely image processing. At the time the application was filed, it would have been obvious to a person of ordinary skill in the art to include the process:” one or more appearance features, extracted by a first encoder of the one or more neural networks, of one or more first objects from one or more first images and one or more structural features, extracted by a second encoder of the one or more neural networks, of one or more second objects from one or more second images; and a decoding of the one or more appearance features of the first object and the one or more structural features of the second object, that have been encoded into one or more latent spaces of the one or more neural networks.”, as taught by Park into the teaching of Li. The suggestion for doing so would provide a flexibility when applied to editing real images. Therefore, it would have been obvious to combine Park with Li. 9. As per claim 2, Li in view of Park discloses: The one or more processors of claim 1, wherein individual appearance features, of 2the one or more appearance features, are associated with respective semantic regions of the one or first mages. (Li, [0029]:”FIG. 1B illustrates a style image, content image, and stylized photorealistic image, in accordance with an embodiment. The photorealistic style image I.sub.S is input to the encoder 165 and the photorealistic content image I.sub.C is input to the second encoder 165. The photorealistic style image and the photorealistic content image I.sub.C are processed by the photo style transfer neural network model 110 to produce the stylized photorealistic image Y. The cloud pattern in the photorealistic content image is retained in the stylized photorealistic image while a blue color of the sky and the green color of the landscape in the photorealistic style image appear in the stylized photorealistic image—the color of the sky and the landscape areas is changed compared with the photorealistic content image. The shape of the road in the stylized photorealistic image is consistent with the road in the photorealistic content image, while the color is changed to be similar to the road in the photorealistic style image. In addition to transferring color, the photo style transfer neural network model 110 may also be configured to synthesize patterns contained in the photorealistic content image, such as a cloud, snow, rain, and the like, in the stylized photorealistic image to be consistent with the content of the photorealistic content image.”, and [0058]) 10. As per claim 5, Li in view of Park discloses: The processor of claim 1, The one or more processors of claim 1, wherein the one or more neural networks include one or more first second convolutional neural networks (CNNs) to extract the one or more structural features from the one or more second objects, using a second encoder of the one or more second CNNs, and one or more second first convolutional neural networks (CNNs) to extract the one or more appearance features from the one or more first objects, using a first encoder of the one or more first CNNs. (Li, [0029]. “FIG. 1B illustrates a style image, content image, and stylized photorealistic image, in accordance with an embodiment. The photorealistic style image I.sub.S is input to the encoder 165 and the photorealistic content image I.sub.C is input to the second encoder 165. The photorealistic style image and the photorealistic content image I.sub.C are processed by the photo style transfer neural network model 110 to produce the stylized photorealistic image Y.”, [0030], and [0035]) 11. Claims 7, 13, 19, and 25 which are similar in scope to claim 1, thus rejected under the same rationale. 12. Claims 8, 14, 20, and 26, which are similar in scope to claim 2, thus rejected under the same rationale. 13. Claims 11, 17, 23, and 29 which are similar in scope to claim 5, thus rejected under the same rationale. 14. Claims 3, 9, 15, 21, and 27 are rejected under 35 U.S.C. 103 as being unpatentable over Li et al., US 2019/0244329 A1, in view of Park et al. US 2022/0148241 A, and further in view of Sekharan et al. US 2021/0004370 15. As per claim 3, Li in view of Park discloses: The one or more processors of claim 1, (See rejection of claim 1 above.) 16. Li doesn’ t expressly discloses: the one or more structural features of the one or more second objects 2and the one or more appearance features of the one or more first objects, are transformed into one or more transformed 3feature vectors using a slot attention transformer. 17. Sekharan discloses: the one or more structural features of the one or more second objects 2and the one or more appearance features are transformed into one or more transformed 3feature vectors using a slot attention transformer. (Sekharan, [0050], “The jointly trained machine learning model may be trained to map a named entity in the input sequence x =(x.sub.1, . . . , x.sub.T) to a corresponding slot in a template query denoted as y=(y.sub.1.sup.2, . . . , y.sub.T.sup.s). Accordingly, for each hidden state h.sub.i, the jointly trained machine learning model may compute a slot context vector c.sub.i.sup.s as a sum of the hidden states h.sub.1, . . . , h.sub.T weighted by the learned slot attention weights a.sub.i,j.sup.s in accordance with Equation (1) below.”) 18. Sekharan is analogous art with respect to Li in view of Park because they are from the same field of endeavor, namely image processing. At the time the application was filed, it would have been obvious to a person of ordinary skill in the art to include the process of that the one or more structural features of the one or more second objects 2and the one or more appearance features are transformed into one or more transformed 3feature vectors using a slot attention transformer, as taught by Sekharan into the teaching of Li in view of Park. The suggestion for doing so would maximize the value of the conditional probability based on adjusting the slot attention weights. Therefore, it would have been obvious to combine Sekharan with Li in view of Park. 19. Claims 9, 15, 21, and 27, which are similar in scope to claim 2, thus rejected under the same rationale. 20. Claims 4, 6, 10, 12, 16, 18, 22, 24, 28 and 30 are rejected under 35 U.S.C. 103 as being unpatentable over Li et al., US 2019/0244329 A1, in view of Park et al. US 2022/0148241 A1, and in view of Sekharan US 2021/0004370 et al., and further in view of Barzelay et al. US 2021/0048931 et al. 21. As per claim 4, Li in view of Park, and in view of Sekharan discloses: Theo or more processors of claim 3, (See rejection of claim 3 above.) 22. Li in view of Park, and in view of Sekharan doesn’ t expressly discloses: the one or more neural networks 2include a generative adversarial network (GAN) for receiving the one or more transformed 3feature vectors and generate one or more representations. 23. Barzelay discloses: the one or more neural networks 2include a generative adversarial network (GAN) for receiving the one or more transformed 3feature vectors and generate one or more representations. (Barzelay, [0035], “In accordance with at least one embodiment a generative adversarial network (GAN) may be used to reconstruct images given a first image. As will be apparent to one of skill in the art, a generative adversarial network has an aspect that encodes objects (e.g., images) as feature vectors and an aspect that decodes objects from feature vectors (e.g., generates images from feature vector.”) 24. Barzelay is analogous art with respect to Li in view of Park, and in view of Sekharan because they are from the same field of endeavor, namely image processing. At the time the application was filed, it would have been obvious to a person of ordinary skill in the art to include the process of that the one or more neural networks 2include a generative adversarial network (GAN) for receiving the one or more transformed 3feature vectors and generate one or more representations, as taught by Barzelay into the teaching of Li in view of Park, and in view of Sekharan. The suggestion for doing so would provide a more accurate interpretation of an image of an object. Therefore, it would have been obvious to combine Barzelay with Li in view of Park, and in view of Park, and in view of Sekharan. 25120. As per claim 6, Li in view of Park, and in view of Sekharan, and in view of Barzelay discloses: The one or more processors of claim 1, wherein the structural features are sampled from 2one or more feature distributions for one or more object types of the one or more second images. (Barzelay, [0030],” In embodiments, the neural network may classify a set of images as according to their shape, color, and texture, as a vector of continuous valued numbers (feature vectors). However, the shape, color, and texture of the image are not explicitly represented by the vector of continuous valued numbers. Instead, the feature vectors are used to identify similar images based on their respective vectors or values having a small Euclidean distance from the source image.”). The proposed combination as well as the motivation for combining the references presented in the rejection of the claim 4 apply to this claim and are incorporated herein by reference. 26. Claims 10, 16, 22, and 28 which are similar in scope to claim 4, thus rejected under the same rationale. 27. Claims 12, 18, 24, and 30, which are similar in scope to claim 6, thus rejected under the same rationale. Response to Arguments 28. Applicant’s arguments with respect to claims 1-30 filed 10/14/2025 have been considered but are moot because Applicant submitted new amended claims. Accordingly, new grounds of rejection are set forth above. The new grounds of rejection conclusion have been necessitated by Applicant's amendments to the claims. Conclusion 29. Any inquiry concerning this communication or earlier communications from the examiner should be directed to ABDERRAHIM MEROUAN whose telephone number is (571)270-5254. The examiner can normally be reached 9 AM -- 5 PM. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Kent Chang can be reached on 571-272-7667. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. /ABDERRAHIM MEROUAN/Primary Examiner, Art Unit 2614
Read full office action

Prosecution Timeline

Jan 27, 2021
Application Filed
Sep 23, 2021
Non-Final Rejection — §103
Dec 14, 2021
Response Filed
Mar 25, 2022
Final Rejection — §103
Sep 27, 2022
Notice of Allowance
Apr 06, 2023
Response after Non-Final Action
Apr 24, 2023
Applicant Interview (Telephonic)
Apr 27, 2023
Request for Continued Examination
May 04, 2023
Response after Non-Final Action
May 06, 2023
Examiner Interview Summary
Jun 02, 2023
Non-Final Rejection — §103
Sep 14, 2023
Applicant Interview (Telephonic)
Sep 25, 2023
Examiner Interview Summary
Oct 10, 2023
Response Filed
Jan 11, 2024
Final Rejection — §103
Mar 04, 2024
Applicant Interview (Telephonic)
Mar 09, 2024
Examiner Interview Summary
Jun 18, 2024
Request for Continued Examination
Jun 20, 2024
Response after Non-Final Action
Oct 04, 2024
Non-Final Rejection — §103
Apr 09, 2025
Response Filed
Jul 11, 2025
Final Rejection — §103
Jul 21, 2025
Interview Requested
Aug 12, 2025
Applicant Interview (Telephonic)
Aug 14, 2025
Examiner Interview Summary
Oct 14, 2025
Request for Continued Examination
Oct 16, 2025
Response after Non-Final Action
Oct 18, 2025
Non-Final Rejection — §103
Jan 26, 2026
Examiner Interview Summary
Jan 26, 2026
Applicant Interview (Telephonic)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12586300
DISTRIBUTED LIGHT TRANSPORT SIMULATION WITH EFFICIENT RAY FORWARDING
2y 5m to grant Granted Mar 24, 2026
Patent 12586037
SYSTEM AND METHOD FOR AUTONOMOUSLY GENERATING DIGITAL PROCEDURES AND PROCEDURE GUIDANCE WITHIN A FACILITY
2y 5m to grant Granted Mar 24, 2026
Patent 12579769
MAPPING & SEGMENTATION OF ORAL MAXILLOFACIAL COMPLEX & EXTRACTING CLINICAL DATA
2y 5m to grant Granted Mar 17, 2026
Patent 12573125
MESH TEXTURE METHOD FOR GENERATING TRUE ORTHOPHOTO USING GEOMETRIC INFORMATION AND THE SYSTEM THEREOF
2y 5m to grant Granted Mar 10, 2026
Patent 12567398
Systems and Methods for Performing Video Communication Using Text-Based Compression
2y 5m to grant Granted Mar 03, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

7-8
Expected OA Rounds
73%
Grant Probability
90%
With Interview (+17.4%)
3y 2m
Median Time to Grant
High
PTA Risk
Based on 659 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month