Prosecution Insights
Last updated: April 19, 2026
Application No. 17/800,553

AUTOMATED IDENTIFICATION OF TRAINING DATA CANDIDATES FOR PERCEPTION SYSTEMS

Non-Final OA §102§103§112
Filed
Aug 18, 2022
Examiner
GONZALES, VINCENT
Art Unit
2124
Tech Center
2100 — Computer Architecture & Software
Assignee
Edge Case Research Inc.
OA Round
1 (Non-Final)
78%
Grant Probability
Favorable
1-2
OA Rounds
3y 6m
To Grant
89%
With Interview

Examiner Intelligence

Grants 78% — above average
78%
Career Allow Rate
410 granted / 522 resolved
+23.5% vs TC avg
Moderate +10% lift
Without
With
+10.5%
Interview Lift
resolved cases with interview
Typical timeline
3y 6m
Avg Prosecution
26 currently pending
Career history
548
Total Applications
across all art units

Statute-Specific Performance

§101
21.2%
-18.8% vs TC avg
§103
39.9%
-0.1% vs TC avg
§102
13.2%
-26.8% vs TC avg
§112
14.6%
-25.4% vs TC avg
Black line = Tech Center average estimate • Based on career data from 522 resolved cases

Office Action

§102 §103 §112
DETAILED ACTION This action is written in response to the election dated 1/6/26. The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Claims 11-15 were not elected, and are withdrawn from consideration. Subject Matter Eligibility In determining whether the claims are subject matter eligible, the examiner has considered and applied the 2019 USPTO Patent Eligibility Guidelines, as well as guidance in the MPEP chapter 2106. The examiner finds that the independent claims are directed to the practical application of improving the labeling of a machine-learning perception system, ie object recognition. Claim Rejections - 35 USC § 112(b) - Indefiniteness The following is a quotation of the second paragraph of 35 U.S.C. 112: (b) CONCLUSION.—The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the inventor or a joint inventor regards as the invention. Claim 7 is rejected under 35 U.S.C. 112(b), as being indefinite for failing to particularly point out and distinctly claim the subject matter which applicant regards as the invention. The claim recites in part: “filtering, by a computing device, the results of the processing step to concentrate on specific perception weaknesses of engineering interest.” The phrase “of engineering interest” is a subjective term, and the Applicant provides no standard for measuring the scope of the claim. Therefore, the claim is indefinite. See MPEP 2173.05(b)(IV) (“Subjective terms”). Claim Rejections - 35 USC § 102 The following is a quotation of the appropriate paragraphs of 35 U.S.C. 102 that form the basis for the rejections under this section made in this Office action: A person shall be entitled to a patent unless – (a)(1) the claimed invention was patented, described in a printed publication, or in public use, on sale or otherwise available to the public before the effective filing date of the claimed invention. Claims 1-3, 5-10 and 16-17 are rejected under 35 U.S.C. 102(a)(1) as being anticipated by Mishra (US 2017/0103267 A1). Regarding claims 1, 16 and 17, Mishra discloses a computer-implemented method (and a related computer system and non-transitory computer-readable medium) for improving the labeling performance of a machine-learned (ML) perception system on an input dataset, the method comprising processing, by a computing device, the input dataset through the perception system; [0031] “It has been recognized that video imaging systems and platforms which analyze image and video content for detection and feature extraction can accumulate significant amounts of data suitable for training and learning analytics that can be leveraged to improve over time the classifiers used to perform the detection and feature extraction by employing a larger search space and generating additional and more complex classifiers through distributed processing.” identifying, by a computing device, one or more perception weaknesses; [0032] “The results of the feature analysis stage 22 are validated in a validation stage 24, which can be performed manually by analysts responsible for accepting or rejecting the results of the feature analyses (e.g., to identify misclassified objects). “ isolating, by a computing device, at least one scene from the input dataset that includes the one or more perception weaknesses; [0034] “A human user can validate the output and provide negative feedback to the algorithm when the algorithm performs poorly. This feedback is stored in a dataset and the classifier is retrained using the learning platform 12. In supervised learning, the goal is typically to label scenic elements and perform object detection.” relabeling the one or more perception weaknesses in the at least one scene to obtain at least one relabeled scene; and [0034] “A human user can validate the output and provide negative feedback to the algorithm when the algorithm performs poorly.” [0074] “ One method is to use an independent observer to classify objects independently from the trained classifier, or to use two trained classifiers with independent implementation and algorithms. The observer can generate validation data which can be compared against the output from the trained classifier.” retraining, by a computing device, the perception system with the at least one relabeled scene; [0034] “A human user can validate the output and provide negative feedback to the algorithm when the algorithm performs poorly. This feedback is stored in a dataset and the classifier is retrained using the learning platform 12. In supervised learning, the goal is typically to label scenic elements and perform object detection.” (Emphasis added.) whereby the labeling performance of the ML perception system is improved. [The Examiner notes that this limitation is an intended result.] [0032] “FIG. 1(a) illustrates a video analysis environment 10 which includes a learning platform 12 to generate and/or improve a set of classifiers 14 used in analyzing video content, e.g., for detection and extraction of features within image frames of a video.“ Regarding independent claims 16 and 17, Mishra also discloses the generic computer components recited therein. (See eg [0006] ‘processors’ and [0069] ‘memory’.) Regarding claim 2, Mishra discloses the further limitation wherein the step of identifying one or more perception weaknesses is performed by a defect detection engine. [0083] “video analysis module 342”, further discussed at [0085]-[0086]. Regarding claim 3, Mishra discloses the further limitation wherein the one or more perception weaknesses comprise labeled objects in bounding boxes. See fig. 9 (reproduced below) and 10. PNG media_image1.png 346 984 media_image1.png Greyscale [0067] “FIG. 9 illustrates how a classifier determines if an object of interest is located within a context. For this case, measurements are extracted from an image frame extracted from a traffic intersection video. The sliding window (box in leftmost image) is centered at each pixel within the frame and the pixel intensities are extracted (top 2nd column).” Regarding claim 5, Mishra discloses the further limitation wherein the step of relabeling the at least one scene comprises sending the at least one scene to be labeled by a human. [0032] “The results of the feature analysis stage 22 are validated in a validation stage 24, which can be performed manually by analysts responsible for accepting or rejecting the results of the feature analyses (e.g., to identify misclassified objects).” Regarding claim 6, Mishra discloses the further limitation wherein the input dataset is comprised of one or more of baseline data, augmented data, ground truth data, and simulated data. [The Examiner notes that this is a Markush group.] ‘baseline data’ :: “training data and labelled ground truth in a supervised learning mode”. Regarding claim 7, Mishra discloses the further limitation comprising filtering, by a computing device, the results of the processing step to concentrate on specific perception weaknesses of engineering interest. The Examiner interprets “engineering interest” according to its broadest reasonable interpretation as encompassing every perception weakness. See also §112(b) rejection supra. Regarding claim 8, Mishra discloses the further limitation comprising repeating the method and comparing the perception weaknesses with the perception weaknesses identified in the processing step that occurred prior to the relabeling to determine an amount of improvement in the performance of the perception system. [0069] “From the distributed processing, a learning and training analysis stage 112 is performed using the results of each partially computed algorithm in aggregation to iteratively improve the classifier. For example, the aggregate results could be used to determine the best feature to distinguish two categories, then that feature could be suppressed to find the next best feature, and the process can continue iteratively until acceptable at least one termination criterion is achieved.” Regarding claim 9, Mishra discloses the further limitation comprising repeating the steps of processing, identifying, isolating, relabeling and retraining until retraining the ML perception system with the newly labeled scenes does not meaningfully improve the labeling performance. [0069] “From the distributed processing, a learning and training analysis stage 112 is performed using the results of each partially computed algorithm in aggregation to iteratively improve the classifier. For example, the aggregate results could be used to determine the best feature to distinguish two categories, then that feature could be suppressed to find the next best feature, and the process can continue iteratively until acceptable at least one termination criterion is achieved.” (Emphasis added.) Regarding claim 10, Mishra discloses the further limitation comprising reengineering the ML perception system when the labeling performance does not meaningfully improve with the newly labeled scenes. [0069] “From the distributed processing, a learning and training analysis stage 112 is performed using the results of each partially computed algorithm in aggregation to iteratively improve the classifier. For example, the aggregate results could be used to determine the best feature to distinguish two categories, then that feature could be suppressed to find the next best feature, and the process can continue iteratively until acceptable at least one termination criterion is achieved.” Claim Rejections - 35 USC § 103 The following is a quotation of 35 U.S.C. 103(a) which forms the basis for all obviousness rejections set forth in this Office action: (a) A patent may not be obtained though the invention is not identically disclosed or described as set forth in section 102 of this title, if the differences between the subject matter sought to be patented and the prior art are such that the subject matter as a whole would have been obvious at the time the invention was made to a person having ordinary skill in the art to which said subject matter pertains. Patentability shall not be negatived by the manner in which the invention was made. The following are the references relied upon in the rejections below: Mishra (US 2017/0103267 A1) Farabet (US 2019/0303759 A1) Claim 4 is rejected under 35 U.S.C. 103 as being unpatentable over Mishra and Farabet. Regarding claim 4, Farabet discloses the further limitation which Mishra does not disclose wherein the step of processing the input dataset further comprises creating, by a computing device, an augmented dataset, processing, by a computing device, the augmented dataset through the perception system, and comparing, by a computing device, results of the input data set with results of the augmented dataset. The Examiner interprets “augmented dataset” in view of the applicant’s specification at [0006] as encompassing a dataset that is augmented with additive noise. This technique is disclosed in Farabet [0099] “GNNS sensors (e.g., GPS sensors) may be simulated within the simulation space to generate real-world coordinates. In order to this, noise functions may be used to approximate inaccuracy.” [0108] “In some examples, the KPI evaluation component 802 may also provide for more complex comparisons. For example, the KPI evaluation component 802 may be as complex as running analytics on the two differing CAN streams to find deviations.” [0175] “The processor(s) 1110 may further include a safety cluster engine that includes a dedicated processor subsystem to handle safety management for automotive applications. The safety cluster engine may include two or more processor cores, a tightly coupled RAM, support peripherals (e.g., timers, an interrupt controller, etc.), and/or routing logic. In a safety mode, the two or more cores may operate in a lockstep mode and function as a single core with comparison logic to detect any differences between their operations.” At the time of filing, it would have been obvious to a person of ordinary skill to implement the data augmentation technique disclosed by Farabet with the Mishra system because this can provide for models which generalize better, thus improving accuracy on new (ie previously unseen) data. Both disclosures pertain to machine vision systems. Additional Relevant Prior Art The following references were identified by the Examiner as being relevant to the disclosed invention, but are not relied upon in any particular prior art rejection: Wang discloses a computer vision system comprising techniques for image area classification. (US 2020/0242424 A1) Conclusion Any inquiry concerning this communication or earlier communications from the examiner should be directed to Vincent Gonzales whose telephone number is (571) 270-3837. The examiner can normally be reached on Monday-Friday 7 a.m. to 4 p.m. MT. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Miranda Huang, can be reached at (571) 270-7092. Information regarding the status of an application may be obtained from the USPTO Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. /Vincent Gonzales/Primary Examiner, Art Unit 2124
Read full office action

Prosecution Timeline

Aug 18, 2022
Application Filed
Feb 26, 2026
Non-Final Rejection — §102, §103, §112 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12585920
PREDICTING OPTIMAL PARAMETERS FOR PHYSICAL DESIGN SYNTHESIS
2y 5m to grant Granted Mar 24, 2026
Patent 12580040
DIFFUSION MODEL FOR GENERATIVE PROTEIN DESIGN
2y 5m to grant Granted Mar 17, 2026
Patent 12566984
METHODS AND SYSTEMS FOR EXPLAINING ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING
2y 5m to grant Granted Mar 03, 2026
Patent 12561402
IDENTIFICATION OF A SECTION OF BODILY TISSUE FOR PATHOLOGY TESTS
2y 5m to grant Granted Feb 24, 2026
Patent 12547647
Unsupervised Machine Learning System to Automate Functions On a Graph Structure
2y 5m to grant Granted Feb 10, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
78%
Grant Probability
89%
With Interview (+10.5%)
3y 6m
Median Time to Grant
Low
PTA Risk
Based on 522 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month