Prosecution Insights
Last updated: April 19, 2026
Application No. 17/155,458

Training and Deploying Model Frontiers to Automatically Adjust to Business Realities

Non-Final OA §103§112
Filed
Jan 22, 2021
Examiner
MARU, MATIYAS T
Art Unit
2148
Tech Center
2100 — Computer Architecture & Software
Assignee
Aible Inc.
OA Round
3 (Non-Final)
58%
Grant Probability
Moderate
3-4
OA Rounds
4y 6m
To Grant
70%
With Interview

Examiner Intelligence

Grants 58% of resolved cases
58%
Career Allow Rate
23 granted / 40 resolved
+2.5% vs TC avg
Moderate +12% lift
Without
With
+12.5%
Interview Lift
resolved cases with interview
Typical timeline
4y 6m
Avg Prosecution
39 currently pending
Career history
79
Total Applications
across all art units

Statute-Specific Performance

§101
35.9%
-4.1% vs TC avg
§103
50.9%
+10.9% vs TC avg
§102
1.9%
-38.1% vs TC avg
§112
11.3%
-28.7% vs TC avg
Black line = Tech Center average estimate • Based on career data from 40 resolved cases

Office Action

§103 §112
Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Continued Examination Under 37 CFR 1.114 A request for continued examination under 37 CFR 1.114, including the fee set forth in 37 CFR 1.17(e), was filed in this application after final rejection. Since this application is eligible for continued examination under 37 CFR 1.114, and the fee set forth in 37 CFR 1.17(e) has been timely paid, the finality of the previous Office action has been withdrawn pursuant to 37 CFR 1.114. Applicant's submission filed on 09/12/2025 has been entered. Response to argument Applicant's arguments filed 09/12/2025 ("Arguments/Remarks") have been fully considered but they are not persuasive. Argument – 1: page: 10, regarding 35 U.S.C 112 rejection, applicant contends: “Claims 7, 9, 18 and 20 are rejected under 35 U.S.C. 112(b) as allegedly being indefinite for failing to particularly point out and distinctly claim the subject matter which the inventors or applicant regards as the invention. The Applicant respectfully disagrees. Nevertheless, in an effort to move prosecution forwards Claims 7, 9, 18 and 20 are amended to address the Examiner’s comments. In particular, the Examiner alleges that “global feasible performance region,” “balance,” and “first feasible performance region” are unsupported by the Applicant’s specification. Applicant submits that “global feasible performance region,” “balance,” and “first feasible performance region” are fully supported at least at Paras. [0003], [0007]-[0008], [003 1]-[0033], [0042]-[0045] of the Applicant’s specification.” Regarding argument – 1, the Examiner notes that, claim 7 and its analogues claim 18 recite determining a “global feasible performance region.” The claim further states that the global feasible performance region is associated with “balanced values of the resource level.” However, the term “balanced value” has not been previously introduced or defined in the claims, nor is there a clear and definite description in the specification of how such “balanced values” are determined. While the specification may reference allocation or efficiency concepts, it does not provide sufficient guidance to inform one of ordinary skill in the art with reasonable certainty as to the meaning of “balanced values” in the context of the claims. Moreover, “the global feasible performance region” could represent set of all achievable performance outcomes, one of ordinary skill in the art would not be able to ascertain the intended scope. Further clarification is needed. Similarly, claim 9 and analogous claim 20 recite a “first feasible performance region.” these are terms that one of ordinary skill in the art would not be able to ascertain the intended scope because the terms are not terms of art and a criterion for determining them as claimed in unclear from the claim language. Further clarification is needed. Argument – 2, Page: 12 applicant contends: “Qin fails to cure the deficiencies of Guodong and Zhang because Qin also does not teach or suggest “a first set of resourcing levels corresponding to a first ratio or value per action or cost per action associated with the first set of models.” At best Qin describes an advertising setup where, for example, a company (e.g., a shoe manufacturer) that buys advertisements on a website (e.g., search engine and advertisement company) pays the website when a potential customer performs an action such as completing a sale (e.g., buys the shoes). This is referred to as a “cost per action” because in this setup the company pays the advertiser (i.e., incurs a cost) when the customer takes an action. As such, Qin’s cost per action is entirely different from the claimed “the first model including a first plurality of submodels trained at differing resource levels corresponding to a first ratio or value per action or cost per action associated with the first plurality of submodels, wherein the first set of resourcing levels comprises an allocation of one or more resources to the first sub-organization.” Accordingly, no combination of Qin, Guodong, and Qin teaches or suggests all of the elements of amended Claim 1 and so Claim 1 is fully patentable over the cited prior art. Claims 12 and 21 are similarly amended and also allowable over the cited prior art.” Regarding argument – 2, Qin ¶[0004], describes a cost per action model in which advertiser resources are expended only when a measurable action occurs. In particular, Qin explains that advertises submit ads with associated actions, ad auctions are conducted and advertisers are charged a fee for each reported action attributed to a winning ad. Thus expressly discloses a system that relates resources consumed (advertiser spend) to actions obtained (user conversions). Applicant’s specification ¶[0047] describes the use of “cost per action” and “value per action” in the context of allocating resource across different function, such as pay-per-click and direct marketing. ¶[0135] … A linear regression model is then used to estimate the action rate for the cost-per-action ads data (i.e.: using a model to estimate the action rate for cost per action ads. (i.e.: cost per action associated with the model)). In both Qin and Applicant’s disclosure, the fundamental principle is the same: determining the efficiency of resource use by relating the cost expended to the number or value of actions achieved, and then using this relationship to guide allocation decisions. Claim Rejections - 35 USC § 112: New Matter The following is a quotation of the first paragraph of 35 U.S.C. 112(a): (a) IN GENERAL.—The specification shall contain a written description of the invention, and of the manner and process of making and using it, in such full, clear, concise, and exact terms as to enable any person skilled in the art to which it pertains, or with which it is most nearly connected, to make and use the same, and shall set forth the best mode contemplated by the inventor or joint inventor of carrying out the invention. The following is a quotation of the first paragraph of pre-AIA 35 U.S.C. 112: The specification shall contain a written description of the invention, and of the manner and process of making and using it, in such full, clear, concise, and exact terms as to enable any person skilled in the art to which it pertains, or with which it is most nearly connected, to make and use the same, and shall set forth the best mode contemplated by the inventor of carrying out his invention. Claim(s) 7 and 18, are rejected under 35 U.S.C. 112(a) or 35 U.S.C. 112 (pre-AlA), first paragraph, as failing to comply with the written description requirement. The claim(s) contains subject matter which was not described in the specification in such a way as to reasonably convey to one skilled in the relevant art that the inventor or a joint inventor, or for applications subject to pre-AlA 35 U.S.C. 112, the inventor(s), at the time the application was filed, had possession of the claimed invention. Claim(s) 7 and 18 recite “… wherein the first and second set of resourcing levels are compatible;” that is considered new matter because the original disclosure does not appear to support for … wherein the first and second set of resourcing levels are compatible, the specification does not describe the first and second resourcing levels are compatible. Although the claims are interpreted in light of the specification, limitations from the specification are not read into the claims. See In re Van Geuns, 988 F.2d 1181, 26 USPQ2d 1057 (Fed. Cir. 1993). Claim Rejections - 35 USC § 112 The following is a quotation of 35 U.S.C. 112(b): (b) CONCLUSION.—The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the inventor or a joint inventor regards as the invention. The following is a quotation of 35 U.S.C. 112 (pre-AIA ), second paragraph: The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the applicant regards as his invention. Claim(s) 7, 9, 18 and 20 are rejected under 35 U.S.C. 112(b) or 35 U.S.C. 112 (pre-AIA ), second paragraph, as being indefinite for failing to particularly point out and distinctly claim the subject matter which the inventor or a joint inventor, or for pre-AIA the applicant regards as the invention. Regarding claims 7 and 18 recite limitations “a global feasible performance region”, “balanced values” and “wherein the first and second set of resourcing levels are compatible”. There is insufficient antecedent basis for the limitations in the claims, which arises from the ambiguity of reference. “a global feasible performance region” is indefinite because the specification does not explain what constitutes the boundaries of this “region” or how its determined. “balanced values” is indefinite because the term has not been previously introduced or defined in the claims, nor is there a clear and definite description in the specification of how such “balanced values” are determined. “wherein the first and second set of resourcing levels are compatible” the specification does not define “compatible” or provide any guidance on how to determine compatibility. These are terms that one of ordinary skill in the art would not be able to ascertain the intended scope because they are not term of art, lack clear definition in the specification and provide no objective boundaries for one ordinary skill in the art to ascertain their scope. Claim 7 and 18 also recite limitations “wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization”. There is insufficient antecedent basis for the limitation in the claim, which arises from the ambiguity of reference. Specifically, the term "the global performance region " lack clarity because it is not clearly defined or linked to a prior element in the claim. For a more precise understanding, the claim should explicitly define or identify terms before refereeing it. For examination purposes, the examiner will interpret “the global performance region” as the global feasible performance region. Regarding claim(s) 9 and 20 recite limitation “a first feasible performance region”. There is insufficient antecedent basis for the limitation in the claims, which arises from the ambiguity of reference. The term that one of ordinary skill in the art would not be able to ascertain the intended scope because the term is not terms of art and a criterion for determining it as claimed in unclear from the claim language. For examination purposes, the examiner interprets a performance region of a model with the scope of the claim limitation. Claim Rejections - 35 USC § 103 The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. Claim(s) 1 – 3, 5, 12 – 14 and 16 are rejected under 35 U.S.C. 103 as being unpatentable over Guodong, et al. "Federated learning for open banking", (hereafter Guodong) in view of Zhang et al., "Personalized federated learning with first order model optimization", (hereafter Zhang), Qin et al., Pub. No.: US20130246167A1, (hereafter Qin) and Achin et al., Pub. No.: US9489630B2, (hereafter Achin). Regarding claim 1, Guodong teaches: A method comprising: receiving data characterizing a first output of one or more of a first set of models associated with a first sub-organization of an organization, (Guodong, page: 6, “4 Statistical heterogeneity in federated learning One challenge of federated learning is statistical heterogeneity in which users have different data distribution. Statistical heterogeneity is an inherent characteristic of a user’s behavior. It is also identified as a non-IID problem. Conventional machine learning is built upon the IID assumption of a uniform dataset. The stochastic gradient descent (SGD) optimization used in vanilla federated learning is not specifically designed and optimized for tackling non-IID data. As described in Fig. 1, each participant’s data are generated from different distributions [receiving data characterizing a first output of one or more of a first set of models associated with a first sub-organization of an organization]. Each local model should then be initialized by the global model that represents a particular distribution.”) the one or more of the first set of models trained on a first dataset; training one or more of a second set of models associated with a second organization based on a second dataset, (Guodong, page: 9, “4.2 Personalized modelling: When a service provider wants to provide a service that is the best for each individual customer [a second organization], the model trained in the central server needs to be personalized or customized. The simplest solution is to treat the global model as a pre-trained model [the one or more of the first set of models trained on a first dataset] (i.e.: the global model is trained with global dataset before distributed to local models), and then use local data [based on a second dataset] to fine-tune the global model, which will derive a personalized model [training one or more of a second set of models associated with a second organization]. However, in most cases, each participant just has a limited number of instances, and the fine-tuning operation will cause over-fitting or increase the generalization error. Another solution is to treat each customer as a target task and the pre-trained global model as a source task, then to apply transfer learning [28] or domain adaptation [3] methods to fine-tune each personalized model. These methods will further leverage the global information to improve the fine-tuning process for each participant. [25] discusses two approaches, namely Data Interpolation and Model Interpolation, to learn a personalized federated learning model by weighting two components between local and global servers in terms of data distributions or models respectively. Personalization layers In general, a model can be decomposed into two parts: a representation learning part and a decisive part. For example, CNN is composed of convolution layers for representation extraction and fully-connected layers for classification decision. In a federated learning setting, heterogeneity could impact one of the two parts. [2] proposes to share representation layers across participants, and then keep decision layers as a personalized part. [24] thinks representation layers should be the personalized part, and then the decision layers could be shared across participants.”) global constraints, and the first output; (Guodong, page: 6, “3 Problem formulation: The learning process of federated learning is decomposed into two parts that occur in different places: server (coordinator) and nodes (participants). These two parts are linked to each other via a specifically designed mechanism. In particular, the participant i can train a local model hi using its own dataset Di = {(xi· , yi·)}. The model hi is initialized by a globally shared model parameter W [global constraints, and the first output] (i.e.: initial parameters set by a global model and shared to participants models) which is then fine-tuned to a new local model with parameters Wi using the data from node i. It is proposed that the coordinator in federated learning can learn a global model controlled by W that could be shared with all participants on distributed nodes. Through a few rounds of communication, the global model has been gradually improved to better suit all participants, and the final global model is an optimal solution that could directly be deployed on each participant for further use.”) retraining the first set of models or a subset thereof. (Guodong, page: 12, “In an open banking data marketplace, the use of participants’ data may be charged by times. For example, in a federated learning process, if the coordinator asks the participant to train the local model three times, they should be charged by three times as well. Moreover, the participant’s data is a dynamically changing profile including its banking-related activities. To capture the dynamic changes of the customers, the federated learning may take an incremental or lifelong learning strategy to regularly use a participant’s data to refine the global model [retraining the first set of models or a subset thereof] (i.e.: the global model is updated and retrained by data updates from participants model training output). This will bring a new challenge for continuously sharing data in a long-term model training framework.”) Guodong do not teach: assessing, based on a second output of the one or more of the second set of models, performance of the one or more of second set of models; a first set of resourcing levels corresponding to a first ratio or value per action or cost per action associated with the first set of models wherein the first set of resourcing levels comprises an allocation of one or more resources to the first sub-organization; Zhang teaches: assessing, based on a second output of the one or more of the second set of models, performance of the one or more of second set of models; and (Zhang, page: 2, “Key here is that after each federating round, we maintain the client-uploaded parameters individually, allowing clients in the next round to download these copies independently of each other. Each federated update is then a two-step process: given a local objective, clients (1) evaluate [performance of the one or more of second set of models;] how well their received models perform on their target task [assessing, based on a second output of the one or more of the second set of models] and (2) use these respective performances to weight each model’s parameters in a personalized update. We show that this intuitive process can be thought of as a particularly coarse version of popular iterative optimization algorithms such as SGD, where instead of directly accessing other clients’ data points and iteratively training our model with the granularity of gradient decent, we limit ourselves to working with their uploaded models. We hence propose an efficient method to calculate these optimal combinations for each client, calling it FedFomo, as (1) each client’s federated update is calculated with a simple first-order model optimization approximating a personalized gradient step, and (2) it draws inspiration from the “fear of missing out”, every client no longer necessarily factoring in contributions from all active clients during each federation round. In other words, curiosity can kill the cat. Each model’s personalized performance can be saved however by restricting unhelpful models from each federated update.”) Zhang and Guodong are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Zhang with teachings of Guodong to enhance customizability and adaptability of a machine learning model. Moreover, it facilitates transfer learning outside of local data distributions, ensuring improved model performance while maintaining privacy (Zhang, Abstract). Guodong in view of Zhang do not teach: a first set of resourcing levels corresponding to a first ratio or value per action or cost per action associated with the first set of models; wherein the first set of resourcing levels comprises an allocation of one or more resources to the first sub-organization; Qin teaches: a first set of resourcing levels corresponding to a first ratio or value per action or cost per action associated with the first set of models; (Qin, “[0135] The process described with reference to FIG. 10 relies on the search engine provider's ability to estimate the actual action rate associated with a particular ad. A variety of techniques may be employed to estimate the action rate. In an example implementation, various features associated with the cost-per-action ads are extracted [a first set of resourcing levels corresponding to a first ratio or value per action or cost per action] (i.e.: various features related to (cost-per-action) CPA ads are extracted (e.g., historical engagement metrics, ad attributes, contextual factors)). A linear regression model is then used to estimate the action rate for the cost-per-action ads data [associated with the first set of models]. The linear regression model may first be trained using a set of training for which the action rates are known. For instance, in one example implementation, given a training set of n ads for which action rates are known, for ad i, yi represents the action rate and xi represents the ad's feature vector. Using the training data, a weight vector β* is learned, which can be used to predict yi using xi according to:”) Qin, Guodong and Zhang are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Qin with teachings of Guodong and Zhang to enable a cost-per-action (CPA) model, ensuring advertisers pay based on actual user engagement. Contextual ad auctions optimize placements, while periodic reports verify actions, enhancing transparency and ROI, (Qin, Abstract). Guodong in view of Zhang and Qin does not teach: wherein the first set of resourcing levels comprises an allocation of one or more resources to the first sub-organization; Achin teaches: wherein the first set of resourcing levels comprises an allocation of one or more resources to the first sub-organization; ((Achin, col. 20 line[45 – 55]), “The allocated processing resources may include temporal resources (e.g., execution cycles of one or more processing nodes, execution time on one or more processing nodes, etc.), physical resources (e.g., a number of processing nodes, an amount of machine-readable storage (e.g., memory and/or secondary storage), etc.), and/or other allocable processing resources. In some embodiments, the allocated processing resources may be processing resources of a distributed computing system and/or a cloud-based computing system. In some embodiments, costs may be incurred when processing resources are allocated [wherein the first set of resourcing levels comprises an allocation] and/or used (e.g., fees may be collected by an operator of a data center in exchange for using the data center's resources) [of one or more resources to the first sub-organization].”) Achin, Guodong, Zhang and Qin are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Achin with teachings of Guodong, Zhang and Qin to allocate resources more efficiently and improve accuracy in forecasting advertiser actions and costs. (Achin, Abstract). Regarding claim 2, Guodong in view of Zhang, Qin and Achin teach the method of claim 1. Guodong further teaches: further comprising providing information associated with the assessment and/or the second output to the first set of models. PNG media_image1.png 284 574 media_image1.png Greyscale [AltContent: textbox ([further comprising providing information associated with the assessment])][AltContent: textbox ([the second output to the first set of models])][AltContent: rect] (Guodong, Fig. 1) Claim 13 recites analogous limitations as claim 2, so is rejected under similar rationale. Regarding claim 3, Guodong in view of Zhang, Qin and Achin teach the method of claim 1. Guodong further teaches: wherein the received data characterizing the first output of the one or more of the first set of models includes the global constraints and/or a first set of resourcing levels. (Guodong, page: 6, “3 Problem formulation: The learning process of federated learning is decomposed into two parts that occur in different places: server (coordinator) and nodes (participants). These two parts are linked to each other via a specifically designed mechanism. In particular, the participant i can train a local model hi using its own dataset Di = {(xi· , yi·)}. The model hi is initialized [wherein the received data characterizing the first output of the one or more of the first set of models includes] (i.e.: first set of models (global models) initialize participants models (second sets of models)) by a globally shared model parameter W [includes the global constraints] (i.e.: initial parameters set by a global model and shared to participants models) which is then fine-tuned to a new local model with parameters Wi using the data from node i. It is proposed that the coordinator in federated learning can learn a global model controlled by W that could be shared with all participants on distributed nodes. Through a few rounds of communication, the global model has been gradually improved to better suit all participants, and the final global model is an optimal solution that could directly be deployed on each participant for further use.”) Claim 14 recites analogous limitations as claim 3, so is rejected under similar rationale. Regarding claim 5, Guodong in view of Zhang, Qin and Achin teach the method of claim 3. Guodong further teaches: further comprising: training one or more of the first set of models, wherein the training is based on one or more of the global constraint, the second output of the one or more of the second set of models, (Guodong, page: 6, “3 Problem formulation: The learning process of federated learning is decomposed into two parts that occur in different places: server (coordinator) and nodes (participants). These two parts are linked to each other via a specifically designed mechanism. In particular, the participant i can train a local model hi using its own dataset Di = {(xi· , yi·)}. The model hi is initialized by a globally shared model parameter W which is then fine-tuned to a new local model with parameters Wi using the data from node i. It is proposed that the coordinator in federated learning can learn a global model [training one or more of the first set of models] controlled by W [wherein the training is based on one or more of the global constraint] that could be shared with all participants on distributed nodes. Through a few rounds of communication, the global model has been gradually improved [the second output of the one or more of the second set of models] (i.e.: output form each participants send back to the global model for update) to better suit all participants, and the final global model is an optimal solution that could directly be deployed on each participant for further use.”) the first set of resource levels, and training data associated with the first set of models. (Guodong, Fig. 1) Claim 16 recites analogous limitations as claim 5, so is rejected under similar rationale. Regarding claim 12, Guodong teaches: A system comprising: at least one data processor; and memory storing computer executable instructions which, when executed by the at least one data processor causes the at least one data processor to perform operations comprising: (Guodong, page: 4, “2.3 Federated learning for open banking Federated learning is a decentralized machine learning framework that can train a model [at least one data processor; and memory storing computer executable instructions] (i.e.: a federated learning framework involves components such as processors and storage units to facilitate decentralized training and data management) without direct access to users’ private data. The model coordinator and user/participant exchange model parameters that can avoid sending user data. However, the exchanging of model parameters, or gradients in machine learning terminology, may cause data leakage [12]. Therefore, differential privacy [11] technology is essential for federated learning to protect privacy from gradient based cyber-attack [1].”) The rest of the limitations are analogous to claim 1, so is rejected under similar rationale. Claim(s) 4 and 15 are rejected under 35 U.S.C. 103 as being unpatentable over Guodong in view of Zhang, Qin, Achin and in further view of Sui et al., "Feded: Federated learning via ensemble distillation for medical relation extraction", (hereafter Sui). Regarding claim 4, Guodong in view of Zhang, Qin and Achin teach the method of claim 3. Guodong in view of Zhang, Qin and Achin do not teach: wherein a second set of resourcing levels are determined based on the first set of resourcing levels. Sui teaches: wherein a second set of resourcing levels are determined based on the first set of resourcing levels. (Sui, page: 2, “In this paper, we introduce a privacy-preserving medical relation extraction model, named FedED. To prevent private information leakage, we leverage federated learning without sharing raw privacy sensitive medical texts. To overcome the communication bottleneck in federated relation extraction, we focus on reducing the size of transmitted messages at each communication round. To this end, we formulate the central aggregation process in federated learning as learning a compact central model (student) from the ensemble (Dietterich, 2000; Breiman, 2001) of multiple local models (teacher). From this perspective, only the predicted labels on a small dataset need to be uploaded to the central server [wherein a second set of resourcing levels] (i.e.: determination of the second set of resource levels, which involves uploading a small dataset to the central server is influenced by the characteristics of the first set of resource levels, which entail learning from a "teacher" model), because learning from a “teacher” model only requires the behavior of the “teacher” rather than the entire “teacher” network [are determined based on the first set of resourcing levels] (i.e.: the decision to upload only the predicted labels on a small dataset is based on the understanding that learning from a "teacher" model necessitates only the behavior of the "teacher" rather than the entire "teacher" network.) (Hinton et al., 2015). Besides, the ensemble model (teacher) is powerful, which defines the upper extreme of aggregating when limited to a single communication in federated learning (Yurochkin et al., 2019). To transfer the knowledge in the ensemble model to the central model, we leverage a strategy based on knowledge distillation (Hinton et al., 2015), which trains the central model by forcing it to have a similar prediction with the ensemble model.”) Sui, Guodong, Zhang, Qin and Achin are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Sui with teachings of Guodong, Zhang, Qin and Achin to mitigate privacy risks in relation to data extraction systems. By training local models with private data stored at individual platforms and aggregating model parameters centrally, the system ensures data privacy without uploading or storing sensitive information on a central server (Sui, Page: 2, “Previous relation extraction methods require centralizing the underlying training data from different medical platforms, such as hospitals and healthcare centers, on one server for training, while holding the centralized privacy-sensitive data puts patients’ privacy at risk. This is one of the reasons that hinder the use of relation extraction in clinical practice. As a possible solution, federated learning (McMahan et al., 2016) is proposed to make full use of privacy-sensitive data. Training local models with private data at local platforms and aggregating local models in the central server compose the federated learning process. In the framework of federated learning, no single piece of private data is uploaded to or stored on the central server, and only local models’ parameters are sent to the server for updating the central model.”) Claim 15 recites analogous limitations as claim 4, so is rejected under similar rationale. Claim(s) 6 and 17 are rejected under 35 U.S.C. 103 as being unpatentable over Guodong in view of Zhang, Qin, Achin and in further view of Rajkumar et al., Pub. No.: US20200311616A1, (hereafter Rajkumar). Regarding claim 6, Guodong in view of Zhang, Qin and Achin teach the method of claim 5. Guodong in view of Zhang, Qin and Achin do not teach: receiving a user input from a user associated with the second set of models, the input indicative of user constraints on the first output of the one or more of the first set of models; and training, the one or more of the first set of models, based on the user input. Rajkumar teaches: receiving a user input from a user associated with the second set of models, the input indicative of user constraints on the first output of the one or more of the first set of models; and training, the one or more of the first set of models, based on the user input. (Rajkumar, “[0003] In some implementations, robots can perform processes to learn new information and new abilities from situations they encounter. Techniques disclosed herein can enable a robot to incorporate new information so that the robot learns almost instantaneously from user feedback [receiving a user input from a user associated with the second set of models] (i.e.: when a user provides a robot with feedback for a previously unknown object, the robot may store the information and apply it to current and future tasks) or other sources of information. For example, a robot can have a local cache where certain types of learned information is stored. When the robot acquires new information, such as the classification for a previously unknown object, the robot can store a representation of the new information in the cache to make the information immediately available for the robot to use. This technique can allow near-instantaneous learning by a robot since the computational demands for incorporating new information, e.g., generating the representation and saving it in a cache accessed by robot systems, are extremely low. For example, the representation may have already been computed as part of the continuous onboard inference process of the robot. As a result, when a user provides a robot with a classification [the input indicative of user constraints on the first output of the one or more of the first set of models] (i.e.: the user's classification serves as a constraint or specification on how the robot (which can be considered a model) should interpret or output its understanding of the previously unknown object. The classification provided by the user dictates how the robot should adjust its behavior or output when it encounters that object in the future) for a previously unknown object, the robot may store the information and apply it to current and future tasks with minimal delay [training, the one or more of the first set of models, based on the user input] (i.e.: the robot incorporates the new information (user input) almost instantaneously and uses it for current and future tasks by training the models based on the user input, as the robot updates its knowledge base and improves its functionality based on the new information provided by the user).”) Rajkumar, Guodong, Zhang, Qin and Achin are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Rajkumar with teachings of Guodong, Zhang, Qin and Achin to add continuous and incremental learning to the machine learning model. By periodically updating the machine learning model and frequently providing new learned information, the system enables individual units to fine-tune their behavior regularly (Rajkumar, “[0006] In some implementations, the server can transmit the updated machine learning model to the fleet at a weekly or a fortnightly basis. The server can also provide update learned information more frequently, as embeddings or in other representations, each time a new set of robot learning is received by the server. Robots in the fleet can store this learned information in a cache and use it alongside the most recent version of the machine learning model. The receipt of new information for the cache can fine-tune the behavior of an overall perception system for each robot in the fleet on a daily basis or each time a new set of robot learning is provided to the robot. Thus, robots can use the updated machine learning model to incorporate the learning from the fleet, and can clear their local caches since the cached information has been incorporated into the model.”) Claim 17 recites analogous limitations as claim 6, so is rejected under similar rationale. Claim(s) 7 and 18 are rejected under 35 U.S.C. 103 as being unpatentable over Guodong in view of Zhang, Qin, Achin and in further view of Zhu et al., "Comparison of individual, ensemble and integrated ensemble machine learning methods to predict China’s SME credit risk in supply chain finance.", (hereafter Zhu) Featonby et al., Pub. No.: US20200310876A1, (hereafter Featonby) and Kaushik et al., Pub. No.: US10809936B1, (hereafter Kaushik). Regarding claim 7, Guodong in view of Zhang, Qin and Achin teach the method of claim 5. Guodong in view of Zhang, Qin and Achin do not teach: assessing a combined performance of the first set of models and the second set of models; determining, using the combined performance, a global feasible performance region, wherein the global feasible performance region is associated with balanced values of the first and a second set of resourcing levels; and displaying the global feasible performance region. wherein the first and second set of resourcing levels are compatible; wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization. Zhu teaches: assessing a combined performance of the first set of models and the second set of models; determining, using the combined performance, a global feasible performance region, wherein the global feasible performance region is associated with balanced values of the first and a second set of resourcing levels; and (Zhu, page: 6, “3.3 Experimental procedure: The experiments are performed on a PC with a 3.19 GHz Intel Core i3 CPU and 1.92 GB RAM, using Windows XP operating system. Data mining toolkit Waikato Environment for Knowledge Analysis (WEKA) version 3.6.12 is sued for experiment. WEKA is a popular and free available suite of machine learning and data mining software which is written with Java and developed at the University of Waikato, New Zealand. We firstly compare average accuracies of IEML methods (i.e., multi-boosting and RS–boosting) with other three EML methods (i.e., bagging, boosting and RS) and an IML method (i.e., DT), for predicting SMEs’ credit risk in SCF [assessing a combined performance of the first set of models and the second set of models] (i.e.: combine these performance metrics into a single composite score for each model. For instance, a weighted average or another aggregation method could be used to synthesize these metrics into one). Secondly, we compare type I and II errors of IEML methods with that of other three EML methods and an IML method. Finally, we compare area under ROC curves of IEML methods (i.e., multi-boosting and RS–boosting) with that of their base EML methods (i.e., bagging, boosting and RS) and IML (i.e., DT) method, respectively [determining, using the combined performance, a global feasible performance region, wherein the global feasible performance region is associated with balanced values of the first and a second set of resourcing levels] (i.e.: define a performance region where all models have acceptable combined scores. This region can be represented in a multidimensional space where each axis corresponds to a different performance metric or resource constraint). For implementation of bagging, boosting, RS and DT, we choose WEKA bagging module, i.e., WEKA ADBoostM1 module, WEKA random subspace module and WEKA J48, respectively. For implementation of RS–boosting, we use WEKA Package, i.e., WEKA.JAR and implement in Eclipse according to Wang and Ma [8]. For implementation of multi-boosting, we use WEKA MultiBoostAB. Meanwhile, we employ DT as base classifier of multi-boosting and RS–boosting according to Maclin and Opitze [20], Fu et al. [21] and Wang and Ma [8].”) displaying the global feasible performance region. (Zhu, page: 6, “4 Empirical results: In this section, we firstly evaluate two IEML methods on 377 data sets using both NN (i.e., Multilayer Perceptron) and DT (i.e., C4.5) as the base classifier for choosing the appropriate one. Then, we show [displaying the global feasible] that IEML methods compete quite outstanding against EML methods and IML method by analyzing prediction evaluation criteria [performance region] (i.e., average accuracy, type I error, II error, ‘precision’ rate, ‘recall’ rate, ‘F-Measure’ rate and ROC curve), which are excellent methods for predicting SMEs credit risk in SCF. Meanwhile, we compare the prediction evaluation criteria of multi-boosting with RS–boosting in order to find the better IEML method in predicting SMEs’ credit risk.”) Zhu, Guodong, Zhang, Qin and Achin are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Zhu with teachings of Guodong, Zhang, Qin and Achin to include individual, ensemble, and integrated ensemble techniques, to accurately predict risks and enhance reliability and decision-making, supporting smoother and more secure operations (Zhu, Abstract). Guodong, Zhang, Qin, Achin and Zhu: wherein the first and second set of resourcing levels are compatible; wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization. Featonby: wherein the first and second set of resourcing levels are compatible; (Featonby, “[0206] The optimization component 126 may then determine that the computational biases of the corresponding VM instances 114(1) and 114(2), and/or the workload categories 220(1) and 220(2), are computationally complimentary such that it is advantageous to have a same computing device 112 host the two VM instances 114(1) and 114(2). For example, and as illustrated, the compute-dimension utilizations 2004 for each of the resource-utilization models 224(1) and 224(2) may be compatible [wherein the first and second set of resourcing levels are compatible] such that the dimensions of compute utilized by workloads in the two workload categories 220(1) and 220(2) combine well to maximize the use of the resources provided by the computing device 112. As shown, five dimensions of compute 2006, 2008, 2010, 2012, 2014 may complement each other such that one dimension of compute 2006(1) for the resource-utilization model 224(1) may be relatively high, but the same dimension of compute 2006(2) for the resource-utilization model 224(2) may be relatively low.”) Featonby, Guodong, Zhang, Qin, Achin and Zhu are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Featonby with teachings of Guodong, Zhang, Qin, Achin and Zhu to maintain efficient performance while minimizing resource and operation cost (Featonby, Abstract). Guodong, Zhang, Qin, Achin, Zhu and Featonby do not teach: wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization Kaushik teaches: wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization (Kaushik, () “Storage systems may include storage processors, storage devices, front-end and back-end adapters, communication busses, etc. The storage processors receive and process 10 requests from hosts (e.g., client devices 104) that write data to and read data from the storage devices of the storage systems. Storage systems may be monitored during operation to display various performance characteristics. The output may be in the form of graphs plotting changes in one or more performance characteristics over time [wherein the global performance region comprises a plot of a range of performance indices as a function of constraints for the organization], or other visualizations of the monitored performance characteristics. The performance characteristics may include, by way of example, IOP S, latency over time, read percentage, etc. While a human operator can view the displayed output to try to detect trends corresponding to performance-impacting events, it is difficult to train a human operator to spot, diagnose and correct such issues. Further, some performance-impacting events may be difficult to discern via simple viewing of the displayed output.”) Kaushik, Guodong, Zhang, Qin, Achin, Zhu and Featonby are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Kaushik with teachings of Guodong, Zhang, Qin, Achin, Zhu and Featonby to monitor workload and to classify performance impacting event to help maintain efficiency, (Kaushik, Abstract). Claim 18 recites analogous limitations as claim 7, so is rejected under similar rationale. Claim(s) 8, 11 and 19 are rejected under 35 U.S.C. 103 as being unpatentable over Guodong in view of Zhang, Qin, Achin and in further view of Rajkumar and Sui. Regarding claim 8, Guodong in view of Zhang, Qin and Achin teach the method of claim 1. Guodong in view of Zhang, Qin and Achin do not teach: further comprising: determining a first set of resourcing levels; Sui teaches: further comprising: determining a first set of resourcing levels; (Sui, page: 2, “In this paper, we introduce a privacy-preserving medical relation extraction model, named FedED. To prevent private information leakage, we leverage federated learning without sharing raw privacy sensitive medical texts. To overcome the communication bottleneck in federated relation extraction, we focus on reducing the size of transmitted messages at each communication round. To this end, we formulate the central aggregation process in federated learning as learning a compact central model (student) from the ensemble (Dietterich, 2000; Breiman, 2001) of multiple local models (teacher). From this perspective, only the predicted labels on a small dataset need to be uploaded to the central server [wherein a second set of resourcing levels] (i.e.: determination of the second set of resource levels, which involves uploading a small dataset to the central server is influenced by the characteristics of the first set of resource levels, which entail learning from a "teacher" model), because learning from a “teacher” model only requires the behavior of the “teacher” rather than the entire “teacher” network [determining a first set of resourcing levels] (Hinton et al., 2015). Besides, the ensemble model (teacher) is powerful, which defines the upper extreme of aggregating when limited to a single communication in federated learning (Yurochkin et al., 2019). To transfer the knowledge in the ensemble model to the central model, we leverage a strategy based on knowledge distillation (Hinton et al., 2015), which trains the central model by forcing it to have a similar prediction with the ensemble model.”) Sui, Guodong, Zhang, Qin and Achin are related to the same field of endeavor (i.e.: training machine learning models). It would have been obvious to a person having ordinary skill in the art before the effective filing date of the invention to combine the teaching of Sui with teachings of Guodong, Zhang, Qin and Achin to mitigate privacy risks in relation to data extraction systems. By training local models with private data stored at individual platforms and aggregating model parameters centrally, the system ensures data privacy without uploading or storing sensitive information on a central server (Sui, Page: 2, “Previous relation extraction methods require centralizing the underlying training data from different medical platforms, such as hospitals and healthcare centers, on one server for training, while holding the centralized privacy-sensitive data puts patients’ privacy at risk. This is one of the reasons that hinder the use of relation extraction in clinical practice. As a possible solution, federated learning (McMahan et al., 2016) is proposed to make full use of privacy-sensitive data. Training local models with privat
Read full office action

Prosecution Timeline

Jan 22, 2021
Application Filed
Jun 20, 2024
Non-Final Rejection — §103, §112
Nov 27, 2024
Applicant Interview (Telephonic)
Nov 27, 2024
Examiner Interview Summary
Dec 23, 2024
Response Filed
Mar 12, 2025
Final Rejection — §103, §112
Aug 20, 2025
Applicant Interview (Telephonic)
Aug 20, 2025
Examiner Interview Summary
Sep 12, 2025
Request for Continued Examination
Sep 23, 2025
Response after Non-Final Action
Oct 14, 2025
Non-Final Rejection — §103, §112
Apr 03, 2026
Examiner Interview Summary
Apr 03, 2026
Applicant Interview (Telephonic)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12586114
GENERATING DIGITAL RECOMMENDATIONS UTILIZING COLLABORATIVE FILTERING, REINFORCEMENT LEARNING, AND INCLUSIVE SETS OF NEGATIVE FEEDBACK
2y 5m to grant Granted Mar 24, 2026
Patent 12572796
METHODS AND SYSTEMS FOR GENERATING RECOMMENDATIONS FOR COUNTERFACTUAL EXPLANATIONS OF COMPUTER ALERTS THAT ARE AUTOMATICALLY DETECTED BY A MACHINE LEARNING ALGORITHM
2y 5m to grant Granted Mar 10, 2026
Patent 12567004
METHOD OF MACHINE LEARNING TRAINING FOR DATA AUGMENTATION
2y 5m to grant Granted Mar 03, 2026
Patent 12561588
Methods and Systems for Generating Example-Based Explanations of Link Prediction Models in Knowledge Graphs
2y 5m to grant Granted Feb 24, 2026
Patent 12561584
TEACHING DATA PREPARATION DEVICE, TEACHING DATA PREPARATION METHOD, AND PROGRAM
2y 5m to grant Granted Feb 24, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

3-4
Expected OA Rounds
58%
Grant Probability
70%
With Interview (+12.5%)
4y 6m
Median Time to Grant
High
PTA Risk
Based on 40 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month