Prosecution Insights
Last updated: April 19, 2026
Application No. 18/569,528

ENHANCED ARCHITECTURE FOR DEEP LEARNING-BASED VIDEO PROCESSING

Non-Final OA §102§103
Filed
Dec 12, 2023
Examiner
SUN, JIANGENG
Art Unit
2671
Tech Center
2600 — Communications
Assignee
Intel Corporation
OA Round
1 (Non-Final)
82%
Grant Probability
Favorable
1-2
OA Rounds
2y 9m
To Grant
96%
With Interview

Examiner Intelligence

Grants 82% — above average
82%
Career Allow Rate
330 granted / 403 resolved
+19.9% vs TC avg
Moderate +14% lift
Without
With
+14.0%
Interview Lift
resolved cases with interview
Typical timeline
2y 9m
Avg Prosecution
22 currently pending
Career history
425
Total Applications
across all art units

Statute-Specific Performance

§101
6.4%
-33.6% vs TC avg
§103
45.3%
+5.3% vs TC avg
§102
25.7%
-14.3% vs TC avg
§112
20.4%
-19.6% vs TC avg
Black line = Tech Center average estimate • Based on career data from 403 resolved cases

Office Action

§102 §103
DETAILED ACTION Claim Rejections - 35 USC § 102 The following is a quotation of the appropriate paragraphs of 35 U.S.C. 102 that form the basis for the rejections under this section made in this Office action: A person shall be entitled to a patent unless – (a)(1) the claimed invention was patented, described in a printed publication, or in public use, on sale, or otherwise available to the public before the effective filing date of the claimed invention. (a)(2) the claimed invention was described in a patent issued under section 151, or in an application for patent published or deemed published under section 122(b), in which the patent or application, as the case may be, names another inventor and was effectively filed before the effective filing date of the claimed invention. Claim(s) 1, 2, 6, 11, 12, 16, 20 is/are rejected under 35 U.S.C. 102(a)(1) as being anticipated by MA( CN 111901592). Regarding claim 1, MA teaches a system ( [0031], computer) for deep learning-based video processing (DLVP), the system comprising: a first neural network associated with generating kernel weights ( [0056], S102 uses reconstructed pixels as labels to synchronously update the neural network weights of the predictive coding module ); and a second neural network associated with filtering image pixels ([0054], S101, with the help of a neural network, generates predicted pixels based on the original pixels or further enhances the predicted pixels after obtaining them) for the DLVP, the second neural network using a second hardware device, wherein the second neural network is configured to receive a first image and the kernel weights, and generate filtered image data based on the first image and the kernel weights(([0054], S101, with the help of a neural network, generates predicted pixels based on the original pixels) , and wherein the first neural network is configured to receive a second image and generate the kernel weights based on the second image, the first image preceding the second image in a series of images([0056], S102 uses reconstructed pixels as labels to synchronously update the neural network weights of the predictive coding module ). Regarding claim 2, MA teaches the system of claim 1, wherein the first neural network comprises a plurality of encoders, a plurality of decoders, and a weight predictor associated with generating the kernel weights based on image data decoded by the plurality of decoders( [0056], S102 … synchronously update the neural network weights of the predictive coding module at the encoder-decoder end). Regarding claim 6, MA teaches the system of any of claim 2, wherein the second neural network comprises a first plurality of filtering layers and a second plurality of filtering layers ( [0025], updating the weights of all layers and updating only the weights of a few layers). Claims 11, 12, 16 recite the methods in claims 1, 2, 6 , thus are also rejected. Claim 20 recites the device for claim 1. Since MA also teaches a device (( [0031], computer). Claim Rejections - 35 USC § 103 The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. Claim(s) 3-5, 7-10, 13-15, 17-19 is/are rejected under 35 U.S.C. 103 as being unpatentable over MA. Regarding claim 3, MA teaches the system of claim 2. MA does not expressly teach wherein the plurality of encoders comprises a convolution layer, a parametric rectified linear unit (PReLU) layer, and a pooling layer. However, official notice is taken that it is routine and conventional to implement neural network encoders comprising a convolution layer, a parametric rectified linear unit (PReLU) layer, and a pooling layer. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to implement the encoders in MA following industry standard practice, with motivation for fast implementation and cost savings. Regarding claim 4, MA teaches the system of claim 2. MA does not expressly teach wherein the plurality of decoders comprises a upsampling layer, a convolution layer, and a PReLU layer. However, official notice is taken that it is routine and conventional to implement neural network decoders comprising a upsampling layer, a convolution layer, and a PReLU layer. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to implement the encoders in MA following industry standard practice, with motivation for fast implementation and cost savings. Regarding claim 5, MA teaches the system of claim 2. MA does not expressly teach wherein the weight predictor comprises a 3x3 convolution layer associated with generating the kernel weights. However, official notice is taken that it is routine and conventional to implement neural network layers with 3x3 patches. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to implement the encoders in MA following industry standard practice, with motivation for fast implementation and cost savings. Regarding claim 7, MA teaches the system of claim 6, wherein the first plurality of filtering layers comprises a convolution layer ([0024], updating the weights of convolutional layers, pooling layers, activation layers, etc.). MA does not teach an average pooling layer. However, official notice is taken that it is routine and conventional to implement a pooling layer with an average pooling layer. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to implement the encoders in MA following industry standard practice, with motivation for fast implementation and cost savings. Regarding claim 8, MA teaches the system of claim 7, wherein the convolution layer receives the kernel weights ( [0025], updating the weights of all layers and updating only the weights of a few layers) . Regarding claim 9, MA teaches the system of claim 6, wherein the second plurality of filtering layers comprises a convolution layer ([0024], updating the weights of convolutional layers, pooling layers, activation layers, etc.). MA does not teach an upsampling layer. However, official notice is taken that it is routine and conventional to implement a convolutional layer with an upsampling area. It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to implement the encoders in MA following industry standard practice, with motivation for higher resolution images, fast implementation and cost savings. Regarding claim 10, MA teaches the system of claim 9, wherein the convolution layer receives the kernel weights([0024], updating the weights of convolutional layers, pooling layers, activation layers, etc.). Claims 13-15, 17-19 recite the methods in claims 3-5, 7-10, thus are also rejected. Conclusion Any inquiry concerning this communication or earlier communications from the examiner should be directed to JIANGENG SUN whose telephone number is (571)272-3712. The examiner can normally be reached 8am to 5pm, EST, M-F. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Randolph Vincent can be reached at 571 272 8243. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. JIANGENG SUN Examiner Art Unit 2671 /Jiangeng Sun/Examiner, Art Unit 2671
Read full office action

Prosecution Timeline

Dec 12, 2023
Application Filed
Jan 21, 2026
Non-Final Rejection — §102, §103
Apr 15, 2026
Examiner Interview Summary
Apr 15, 2026
Applicant Interview (Telephonic)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12591973
Histological Image Analysis
2y 5m to grant Granted Mar 31, 2026
Patent 12561872
METHOD OF TRAINING IMAGE DECOMPOSITION MODEL, METHOD OF DECOMPOSING IMAGE, ELECTRONIC DEVICE, AND STORAGE MEDIUM
2y 5m to grant Granted Feb 24, 2026
Patent 12561757
IMAGE SUPER-RESOLUTION NEURAL NETWORKS
2y 5m to grant Granted Feb 24, 2026
Patent 12548122
METHOD FOR FILTERING PERIODIC NOISE AND FILTER USING THE METHOD
2y 5m to grant Granted Feb 10, 2026
Patent 12524859
System And Method For The Visualization And Characterization Of Objects In Images
2y 5m to grant Granted Jan 13, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
82%
Grant Probability
96%
With Interview (+14.0%)
2y 9m
Median Time to Grant
Low
PTA Risk
Based on 403 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month