Prosecution Insights
Last updated: April 19, 2026
Application No. 17/729,585

METHOD FOR ENHANCED SAMPLING FROM A PROBABILITY DISTRIBUTION

Non-Final OA §101§102§103§112
Filed
Apr 26, 2022
Examiner
SMITH, KEVIN LEE
Art Unit
2122
Tech Center
2100 — Computer Architecture & Software
Assignee
Multiverse Computing S L
OA Round
1 (Non-Final)
37%
Grant Probability
At Risk
1-2
OA Rounds
4y 8m
To Grant
55%
With Interview

Examiner Intelligence

Grants only 37% of cases
37%
Career Allow Rate
49 granted / 134 resolved
-18.4% vs TC avg
Strong +18% interview lift
Without
With
+18.0%
Interview Lift
resolved cases with interview
Typical timeline
4y 8m
Avg Prosecution
45 currently pending
Career history
179
Total Applications
across all art units

Statute-Specific Performance

§101
30.7%
-9.3% vs TC avg
§103
36.4%
-3.6% vs TC avg
§102
10.1%
-29.9% vs TC avg
§112
17.3%
-22.7% vs TC avg
Black line = Tech Center average estimate • Based on career data from 134 resolved cases

Office Action

§101 §102 §103 §112
DETAILED ACTION 1. The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . 2. This communication is in response to the Applicant’s submission filed 26 April 2022, where: Claims 1-20 are pending. Claims 1-20 are rejected. Claim Rejections – 35 U.S.C. § 112 3. The following is a quotation of 35 U.S.C. § 112(b): (b) CONCLUSION.—The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the inventor or a joint inventor regards as the invention. 4. Claims 5-7, 9, and 16-18 are rejected under 35 U.S.C. § 112(b) as being indefinite for failing to particularly point out and distinctly claim the subject matter which the inventor or a joint inventor regards as the invention. Claim 5, line 2, and claim 16, lines 2-3, each recite “comprises minimizing the following negative log-likelihood function.” There is insufficient antecedent basis for this limitation in the respective claim. Claim 9, line 2, recites “any one of the first, second, third and fourth moments.” There is insufficient antecedent basis for this limitation in the claim. Claims 6 and 7 depend directly or indirectly from claim 5. Claims 17 and 18 depend directly or indirectly from claim 16. Claims 6, 7, 17, and 18 are rejected as depending from a rejected claim; further, the claims fail to cure the deficiencies of claims 5 and 17, respectively. Claim Rejections - 35 U.S.C. § 101 5. 35 U.S.C. § 101 reads as follows: Whoever invents or discovers any new and useful process, machine, manufacture, or composition of matter, or any new and useful improvement thereof, may obtain a patent therefor, subject to the conditions and requirements of this title. 6. Claims 1-20 are rejected under 35 U.S.C. § 101 because the claimed invention is directed to an abstract idea without significantly more. Claim 1 recites a device or system, which is a product, and thus one of the statutory categories of patentable subject matter. (35 U.S.C. § 101). However, under Step 2A Prong One, the claim recites the limitations of “[(b)] providing a tensor codifying the probability distribution such that each configuration of the plurality of discrete random variables has its respective probability codified therein,” “[(c)] encoding the tensor into a tensor network in the form of a matrix product state,” and “[(d)] computing at least one moment of the probability distribution by processing the tensor network for sampling of the probability distribution.” The limitations of “[(b)] . . . a tensor codifying,” “[(c)] encoding,” and “[(d)] computing at least one moment,” contain activities that can practically be performed in the human mind, including, for example, observations, evaluations, judgments, and opinions, and accordingly, are a mental process, (MPEP § 2106.04(a)(2) sub III), which is one of the groups of abstract ideas. (MPEP § 2106.04(a)(2)). Also, because “tensor networks” are mathematical frameworks that represents high-dimensional, correlated data as a network of smaller, interconnected “tensors.” “Tensors” are mathematical objects generalizing scalars, vectors, and matrices into a multilinear object, and accordingly, are mathematical concepts, (MPEP § 2106.04(a)(2) sub I), and is one of the groups of abstract ideas. The claim recites more details or specifics to the abstract idea of “[(b)] providing a tensor codifying,” where “[(b.1)] where all probabilities are greater than or equal to zero and a sum of all probabilities is equal to one,” and accordingly, is merely more specific to the abstract idea. Therefore, claim 1 is directed to an abstract idea. Under Step 2A Prong Two, the claim as a whole is not integrated into a practical application, because the additional elements recited in the claim beyond the identified abstract idea include “at least one processor,” “at least one memory comprising computer program code for one or more programs,” which are generic computer components using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not serve to integrate the abstract idea into a practical application. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is an pre-processing insignificant extra-solution activity of mere data gathering, (MPEP § 2106.05(g)), that does not serve to integrate the abstract idea into a practical application. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not serve to integrate the abstract idea into a practical application. Therefore, claim 1 is directed to the abstract idea. Finally, under Step 2B, the additional elements, taken alone or in combination, do not represent significantly more than the abstract idea itself. The additional elements include “at least one processor,” “at least one memory comprising computer program code for one or more programs,” which are generic computer components using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not amount to significantly more than the abstract idea. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is a well-understood, routine, and conventional activity of retrieving information in memory, (MPEP § 2106.05(d) sub II.iv), which does not amount to significantly more than the abstract idea. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not amount to significantly more than the abstract idea. Therefore, claim 1 is subject-matter ineligible. Claim 12 recites a computer-implemented method, which is a process, and thus one of the statutory categories of patentable subject matter. (35 U.S.C. § 101). However, under Step 2A Prong One, the claim recites the limitations of “[(b)] providing a tensor codifying the probability distribution such that each configuration of the plurality of discrete random variables has its respective probability codified therein,” “[(c)] encoding the tensor into a tensor network in the form of a matrix product state,” and “[(d)] computing at least one moment of the probability distribution by processing the tensor network for sampling of the probability distribution.” The limitations of “[(b)] . . . a tensor codifying,” “[(c)] encoding,” and “[(d)] computing at least one moment,” contain activities that can practically be performed in the human mind, including, for example, observations, evaluations, judgments, and opinions, and accordingly, are a mental process, (MPEP § 2106.04(a)(2) sub III), which is one of the groups of abstract ideas. (MPEP § 2106.04(a)(2)). Also, because “tensor networks” are mathematical frameworks that represents high-dimensional, correlated data as a network of smaller, interconnected “tensors.” “Tensors” are mathematical objects generalizing scalars, vectors, and matrices into a multilinear object, and accordingly, are mathematical concepts, (MPEP § 2106.04(a)(2) sub I), and is one of the groups of abstract ideas. The claim recites more details or specifics to the abstract idea of “[(b)] providing a tensor codifying,” where “[(b.1)] where all probabilities are greater than or equal to zero and a sum of all probabilities is equal to one,” and accordingly, is merely more specific to the abstract idea. Therefore, claim 1 is directed to an abstract idea. Under Step 2A Prong Two, the claim as a whole is not integrated into a practical application, because the additional elements recited in the claim beyond the identified abstract idea include a “computer-implemented method,” which is a generic computer component using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not serve to integrate the abstract idea into a practical application. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is an pre-processing insignificant extra-solution activity of mere data gathering, (MPEP § 2106.05(g)), that does not serve to integrate the abstract idea into a practical application. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables about a target,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not serve to integrate the abstract idea into a practical application. Therefore, claim 12 is directed to the abstract idea. Finally, under Step 2B, the additional elements, taken alone or in combination, do not represent significantly more than the abstract idea itself. The additional elements include “computer-implemented method,” which are generic computer components using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not amount to significantly more than the abstract idea. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is a well-understood, routine, and conventional activity of retrieving information in memory, (MPEP § 2106.05(d) sub II.iv), which does not amount to significantly more than the abstract idea. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not amount to significantly more than the abstract idea. Therefore, claim 12 is subject-matter ineligible. Claim 20 recites a non-transitory computer-readable medium, which is a product, and thus one of the statutory categories of patentable subject matter. (35 U.S.C. § 101). However, under Step 2A Prong One, the claim recites the limitations of “[(b)] providing a tensor codifying the probability distribution such that each configuration of the plurality of discrete random variables has its respective probability codified therein,” “[(c)] encoding the tensor into a tensor network in the form of a matrix product state,” and “[(d)] computing at least one moment of the probability distribution by processing the tensor network for sampling of the probability distribution.” The limitations of “[(b)] . . . a tensor codifying,” “[(c)] encoding,” and “[(d)] computing at least one moment,” contain activities that can practically be performed in the human mind, including, for example, observations, evaluations, judgments, and opinions, and accordingly, are a mental process, (MPEP § 2106.04(a)(2) sub III), which is one of the groups of abstract ideas. (MPEP § 2106.04(a)(2)). Also, because “tensor networks” are mathematical frameworks that represents high-dimensional, correlated data as a network of smaller, interconnected “tensors.” “Tensors” are mathematical objects generalizing scalars, vectors, and matrices into a multilinear object, and accordingly, are mathematical concepts, (MPEP § 2106.04(a)(2) sub I), and is one of the groups of abstract ideas. The claim recites more details or specifics to the abstract idea of “[(b)] providing a tensor codifying,” where “[(b.1)] where all probabilities are greater than or equal to zero and a sum of all probabilities is equal to one,” and accordingly, is merely more specific to the abstract idea. Therefore, claim 20 is directed to an abstract idea. Under Step 2A Prong Two, the claim as a whole is not integrated into a practical application, because the additional elements recited in the claim beyond the identified abstract idea include a “non-transitory computer-readable medium encoded with instructions that, when executed by at least one processor or hardware, perform or make a device,” which is a generic computer component using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not serve to integrate the abstract idea into a practical application. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is an pre-processing insignificant extra-solution activity of mere data gathering, (MPEP § 2106.05(g)), that does not serve to integrate the abstract idea into a practical application. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not serve to integrate the abstract idea into a practical application. Therefore, claim 20 is directed to the abstract idea. Finally, under Step 2B, the additional elements, taken alone or in combination, do not represent significantly more than the abstract idea itself. The additional elements include “non-transitory computer-readable medium encoded with instructions that, when executed by at least one processor or hardware, perform or make a device,” which are generic computer components using instructions to cause a system to implement the abstract idea, (MPEP § 2106.05(f)), that does not amount to significantly more than the abstract idea. The claim also recites the limitation of “[(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target,” which is a well-understood, routine, and conventional activity of retrieving information in memory, (MPEP § 2106.05(d) sub II.iv), which does not amount to significantly more than the abstract idea. The claim also recites more details or specifics of the additional element of “[(a)] receiving,” where “[(a.1)] the probability distribution relating to a plurality of discrete random variables,” and accordingly, is merely more specific to the additional element. Further, that the “probability distribution . . . is about a target” is generally linking the use of the judicial exception to a particular technological environment or field of use (MPEP § 2106.05(h)), that does not amount to significantly more than the abstract idea. Therefore, claim 12 is subject-matter ineligible. Claim 2 depends from claim 1. Claim 13 depends from claim 12. The claims further recite “[(e)] at least carry out the following: providing a predetermined command at least based on the computed at least one moment.” The activity of “[(e)] providing a predetermined command” is a post-processing insignificant extra-solution of transmitting data, (MPEP § 2106.05(f)), that does not serve to integrate the abstract idea into a practical application. Also, the activity of “[(e)] providing a predetermined command” is a well-understood, routine, and conventional activity of transmitting a command over a network, (MPEP § 2106.05(d) sub II.i), that does not amount to significantly more than the abstract idea. The abstract idea of these claims are not integrated into a practical application, (see MPEP § 2106.04(d)), nor do they amount to significantly more than the abstract idea, (MPEP § 2106.05 sub I; see also MPEP § 2106.05(a) – (h)), because the claims recite no more than the abstract idea. Therefore, claims 2 and 13 are subject-matter ineligible. Claim 3 depends directly or indirectly from claim 1. Claim 14 depends directly or indirectly from claim 12. The claims recite more details or specifics to the additional element of ““[(e)] providing a predetermined command” where “[(e.1)] wherein the predetermined command comprises one or both of: [(e.1.1)] providing a notification indicative of the computed at least one moment to an electronic device; and [(e.1.2)] providing a command to a controlling device or system associated with the target or to the target itself when the target is either a machine or a system, the predetermined command being for changing a behavior of the target,” and accordingly, are merely more specific to the additional element. The abstract idea of these claims are not integrated into a practical application, (see MPEP § 2106.04(d)), nor do they amount to significantly more than the abstract idea, (MPEP § 2106.05 sub I; see also MPEP § 2106.05(a) – (h)), because the claims recite no more than the abstract idea. Therefore, claims 3 and 14 are subject-matter ineligible. Claim 4, 5, and 8 depend directly or indirectly from claim 1. Claims 15, 16, and 19 depend directly or indirectly from claim 12. The claims recite more details or specifics to the abstract idea of “[(c)] encoding the tensor,” (claims 4 and 15: “[(c)] wherein encoding the tensor into the tensor network comprises [(c.3)] factorizing the tensor into the tensors of the tensor network by processing the tensor so that the following equation is solved: PNG media_image1.png 35 92 media_image1.png Greyscale where P is the resulting normalized factorization into the tensors of the tensor network, T is the encoded tensor, and ZT is a predetermined normalization factor ZT=ΣX1, . . . , xN TX 1 , . . . , xN, with X1, . . . , XN being respective N configurations of the plurality of discrete random variables of the probability distribution, TX 1 , . . . , xN being the tensor for the respective configuration,” and N being the number of discrete random variables in the plurality of discrete random variables”) and (claims 5 and 16: “[(c)] wherein encoding the tensor into the tensor network further comprises [(c.4)] minimizing the following negative log-likelihood function for each sample xi of a discrete multivariate distribution: PNG media_image2.png 50 174 media_image2.png Greyscale where each sample xi has values for each of the discrete random variables, and TXi is the tensor for the sample xi”), and (claims 8 and 19: “[(c)] wherein encoding the tensor into the tensor network further comprises [(c.4)] compressing a probability mass function into a tensor that is not negative, and minimizing the following Kullback-Leibler divergence equation: PNG media_image2.png 50 174 media_image2.png Greyscale where PX 1 , . . . , xN is a probability mass function corresponding to the probability distribution”), and accordingly, are merely more specific to the abstract idea. Therefore, claims 4, 5, 8, 15, 16, and 19 are subject-matter ineligible. Claim 6 depends directly or indirectly from claim 1. Claim 17 depends directly or indirectly from claim 12. The claims recite more details or specifics to the abstract idea of “[(c)] encoding the tensor,” “[(c.4.1)] wherein the minimization of the negative log-likelihood function for each sample xi is calculated with local gradient-descent in which the gradient of the function is computed for all tensors of the tensor network,” and accordingly, are merely more specific to the abstract idea. Therefore, claims 6 and 17 are subject-matter ineligible. Claim 7 depends directly or indirectly from claim 1. Claim 18 depends directly or indirectly from claim 12. The claims recite more details or specifics to the abstract idea of “[(c)] encoding the tensor,” “wherein values of the tensors of the tensor network are modified iteratively to approximate the probability distribution therein,” and accordingly, are merely more specific to the abstract idea. Therefore, claims 6 and 17 are subject-matter ineligible. Claims 9 and 10 depend directly or indirectly from claim 1. The claims recite more details or specifics to the abstract idea of “[(d)] computing at least one moment,” (claim 9: “[(d)] wherein computing the at least one moment comprises [(d.1)] computing any one of the first, second, third and fourth moments of the probability distribution by processing the tensor network; claim 10: “[(d)] wherein computing the at least one moment comprises [(d.1)] computing a contraction of the tensor network”), and accordingly, are merely more specific to the abstract idea. Therefore, claims 9 and 10 are subject-matter ineligible. Claim 11 depends from claim 1. The claim recites more details or specifics to the additional element of “[(a)] receiving data,” “[(a.2)] wherein the target comprises: an electrical grid, an electricity network, a portfolio of financial derivatives, a stock market, a set of patients of a hospital unit, or a system comprising one of: one or more devices, one or more machines, or a combination thereof,” and accordingly, is merely more specific to the additional element. The abstract idea of these claims are not integrated into a practical application, (see MPEP § 2106.04(d)), nor do they amount to significantly more than the abstract idea, (MPEP § 2106.05 sub I; see also MPEP § 2106.05(a) – (h)), because the claims recite no more than the abstract idea. Therefore, claim 11 is subject-matter ineligible. Claim Rejections – 35 U.S.C. § 102 7. The following is a quotation of the appropriate paragraphs of 35 U.S.C. § 102 that form the basis for the rejections under this section made in this Office action: A person shall be entitled to a patent unless – (a)(1) the claimed invention was patented, described in a printed publication, or in public use, on sale, or otherwise available to the public before the effective filing date of the claimed invention. (a)(2) the claimed invention was described in a patent issued under section 151, or in an application for patent published or deemed published under section 122(b), in which the patent or application, as the case may be, names another inventor and was effectively filed before the effective filing date of the claimed invention. 8. Claims 1-3, 9, 10, 12-14, and 20 are rejected under 35 U.S.C. § 102(a)(1) as being anticipated by Gillman et al., "A Tensor Network Approach to Finite Markov Decision Processes," arXiv (2020) [hereinafter Gillman]. Regarding claims 1, 12, and 20, Gillman teaches [a] device or system (Gillman, left column of p. 2, “2.1. Finite Markov Decision Processes,” first paragraph, teaches “the state of the system is a terminal state [(that is, a device or system)]”) of claim 1, [a] computer-implemented method (Gillman, right column of p. 1, “1. Introduction,” second paragraph, teaches “[g]iven the nature of the problems tackled by TNs in physics, it is natural to consider them for machine learning problems [(that is, “machine learning problems” are inherently a computer-implemented method)]”) of claim 12, and [a] non-transitory computer-readable medium encoded with instructions that, when executed by at least one processor or hardware (Gillman, right column of p. 1, “1. Introduction,” second paragraph, teaches “[g]iven the nature of the problems tackled by TNs in physics, it is natural to consider them for machine learning problems [(that is, “machine learning problems” inherently includes a non-transitory computer-readable medium encoded with instructions, when executed by at least one processor or hardware )]”) of claim 20, comprising: at least one processor; and at least one memory comprising computer program code for one or more programs; the at least one processor, the at least one memory, and the computer program code (Gillman, right column of p. 1, “1. Introduction,” second paragraph, teaches “[g]iven the nature of the problems tackled by TNs in physics, it is natural to consider them for machine learning problems [(that is, “machine learning problems” inherently includes the at least one processor, the at least one memory, and the computer program code)]”) being configured to cause the device or system (Gillman, left column of p. 2, “2.1. Finite Markov Decision Processes,” first paragraph, teaches “the state of the system is a terminal state [(that is, a device or system)]”) to at least carry out the following: [(a)] receiving data including a probability distribution of a dataset or a multivariate probability distribution about a target (Gillman, Abstract, teaches “[a]s an application [of a general tensor network formulation of finite, episodic and discrete Markov decision processes (MDPs),] we consider the issue - formulated as an RL problem - of finding a stochastic evolution that satisfies specific dynamical conditions [(that is, a probability distribution of a dataset . . . about a target)], using the simple example of random walk excursions as an illustration”; Gillman, left column of p. 3, “2.2 Tensor Networks for Hidden Markov Models,” first paragraph, teaches “[a]s suggested by the chosen notation, Eq. (11) [(that is, since the matrices are rank-2 tensors and the vector a rank-1 tensor, this can be considered a TN consisting of T + 1 tensors, where the contraction pattern is given by the usual matrix products. Performing such a contraction produces a new vector, | p T , with components, PNG media_image3.png 41 364 media_image3.png Greyscale where | p 0 = ∑ s c s | s for some coefficients cs and the vectors | s associated to each state s form a basis of a vector space VS),] is exactly the [tensor network representation (TNR)] for a probability distribution over states produced by a Markovian dynamics [(that is, receiving data including a probability distribution of a dataset)]. In that case, the components of Mt are equal to the probabilities of state transitions, PNG media_image4.png 38 360 media_image4.png Greyscale while the components of | p 0 give the initial probabilities distribution over states [(that is, receiving data including a probability distribution of a dataset)]”; Gillman, right column of p. 1, “1. Introduction,” second paragraph, teaches “using the formalism of Markov decision processes (MDPs), consisting of repeated updates according to an agent’s decision making policy and the dynamics of an environment it inhabits [(that is, “dynamics of an environment” is receiving data including a probability distribution of a dataset . . . about a target)]”; [Examiner notes that the plain meaning of term “target” pertains to an environment in which the system is deployed, and accordingly, the broadest reasonable interpretation of the term “target” covers the teachings of Gillman, which is not inconsistent with the Applicant’s disclosure. (MPEP § 2111.01).]), [(a.1)] the probability distribution relating to a plurality of discrete random variables (Gillman, left column of p. 2, “2.1 Finite Markov Decisions Processes,” first paragraph, teaches “[i]n a discrete-time, episodic, MDP, individual trajectories take the form, PNG media_image5.png 53 435 media_image5.png Greyscale where St; At and Rt are random variables for the state, action and reward, taking values st, at, and rt respectively [(that is, the probability distribution relating to a plurality of discrete random variables)]”); [(b)] providing a tensor codifying the probability distribution such that each configuration of the plurality of discrete random variables has its respective probability codified therein (Gillman, right column of p. 2, “2.2 Tensor Networks for Hidden Markov Models,” second paragraph, teaches in a diagrammatic notation, “rank-K tensors are represented as shapes with K legs, and contractions are indicated by joining the appropriate legs together. In this notation, Eq. (11) for T = 3 [(that is, each “termination time T” is each configuration of the plurality of discrete random variables)] reads, PNG media_image6.png 189 519 media_image6.png Greyscale As suggested by the chosen notation, Eq. (11) is exactly the TNR for a probability distribution over states produced by a Markovian dynamics. In that case, the components of Mt are equal to the probabilities of state transitions, PNG media_image7.png 44 377 media_image7.png Greyscale while the components of |   p 0 give the initial probability distribution over states”), [(b.1)] where all probabilities are greater than or equal to zero and a sum of all probabilities is equal to one (Gillman, left column of p. 3, “2.2 Tensor Networks for Hidden Markov Models,” first paragraph, teaches “Any vector | p t ∈ V S , that is a convex combination of the particular basis | s , can be interpreted as a probability distribution over states via their components s p t = p t s . This implies their normalization as - s p t = 1 , where - s | is the flat-vector for the basis, - s | = ∑ s s | ”) [(that is, “normalization” is where all probabilities are greater than or equal to zero)]”; [Examiner notes the plain meaning of the term “greater than or equal to zero and a sum of all probabilities is equal to one” is that of a normalized data where data values are transformed to a specific range, such as 0 to 1, based on the minimum and maximum values in the dataset, and accordingly, the broadest reasonable interpretation of “greater than or equal to zero and a sum of all probabilities is equal to one” covers the teachings of Gillman, which is not inconsistent with the Applicant’s disclosure. (MPEP § 2111.01)]; [(c)] encoding the tensor into a tensor network in the form of a matrix product state (Gillman, left column of p. 3, “2.2 Tensor Networks for Hidden Markov Models,” second paragraph, teaches “[a] more complicated TNR relevant for dynamics is offered by the matrix product state (MPS) representation - also known as the tensor train decomposition (Oseledets, 2011) – of HMMs”), [(c.1)] where an external index of each tensor of the tensor network represents one discrete random variable of the plurality discrete random variables, and [(c.2)] an internal index or internal indices of each tensor of the tensor network represents correlation between the tensor and the corresponding adjacent tensor of the tensor network (Specification at p. 2, lines 20-23, teaches “[a]s known in the art, the tensors of an MPS have an external index, and one or two internal indices, depending on whether the tensor is at an end of the MPS or not. The external index, also referred to as physical dimension, of each tensor is representative of a respective discrete random variable, hence the MPS has as many tensors as discrete random variables are in the probability distribution. Further, the internal index or indices, also referred to as virtual dimension or dimensions, are representative of the correlation between adjacent tensors; [Examiner notes that, in view of Applicant’s disclosure, the term “matrix product state” inherently includes “an external index” and “an internal index or internal indices”]); and [(d)] computing at least one moment of the probability distribution (Gillman, right column of p. 3, “2.3 Tensor Networks for Time Integrated Observables,” first paragraph, teaches that “[w]hen considering dynamics described by an [Hidden Markov Model (HMM)], one is often interested in time integrated observables. Such objects can be represented easily in terms of [tensor networks (TNs)], which in turn allows for the [tensor network representation (TNR)] of averages or higher-order moments [(that is, computing at least one moment of the probability distribution)]”). [Examiner notes that the Specification at p. 2, lines 28-29, teaches “[b]y operating the tensor network as known in the art, different data can be sampled from the probability distribution since it is encoded in the tensor network itself [(that is, processing the tensor network for sampling of the probability distribution)]”). Regarding claims 2 and 13, Gillman teaches all of the limitations of claims 1 and 12, respectively, as described above in detail. Gillman teaches - wherein the at least one processor, the at least one memory, and the computer program code are configured to further cause the device or system to [(e)] at least carry out the following: providing a predetermined command at least based on the computed at least one moment (Gillman, left column of p. 7, “5.1 Conditioned Dynamics and FMDPs,” first & second paragraphs, teaches that “[a]n elementary example of rare events are \stochastic excursions’ (Majumdar & Orland, 2015), where a simple random walker is conditioned to stay above a certain line and at a given time must return to this line. . . . For an episode with fixed termination time, T, the positions of the random walker are encoded in S={-T, . . . , -1, 0, 1, +T} such that |S| = 2T+1. The action space is Α = {0; 1}, where a = 0, 1 [(that is, “action a = 0, 1” is at least carry out the following: providing a predetermined command at least based on the computer at least one moment)] correspond to a down/up move of the walker, respectively”). Regarding claims 3 and 14, Gillman teaches all of the limitations of claims 2 and 13, respectively, as described above in detail. Gillman teaches - [(e.1)] wherein the predetermined command comprises one or both of: [(e.1.1)] providing a notification indicative of the computed at least one moment to an electronic device; and [(e.1.2)] providing a command to a controlling device or system associated with the target or to the target itself when the target is either a machine or a system, the predetermined command being for changing a behavior of the target (Gillman, left column of p. 7, “5.1 Conditioned Dynamics and FMDPs,” first & second paragraphs, teaches that “[a]n elementary example of rare events are \stochastic excursions’ (Majumdar & Orland, 2015), where a simple random walker is conditioned to stay above a certain line and at a given time must return to this line. . . . For an episode with fixed termination time, T, the positions of the random walker are encoded in S={-T, . . . , -1, 0, 1, +T} such that |S| = 2T+1. The action space is Α = {0; 1}, where a = 0, 1 [(that is, “action a = 0, 1” is providing a command to a controlling device or system associated with the target itself)] correspond to a down/up move of the walker, respectively [(that is, “the down/up move of the walker” is the predetermined command being for changing a behavior of the target)]”). Regarding claim 9, Gillman teaches all of the limitations of claim 1, as described above in detail. Gillman teaches - [(d)] wherein computing the at least one moment comprises [(d.1)] computing any one of the first, second, third and fourth moments of the probability distribution by processing the tensor network (Gillman, right column of p. 3, “2.3 Tensor Networks for Time Integrated Observables,” first paragraph, teaches “When considering dynamics described by an HMM, one is often interested in time integrated observables. Such objects can be represented easily in terms of TNs, which in turn allows for the TNR of averages or higher-order moments”). Regarding claim 10, Gillman teaches all of the limitations of claim 1, as described above in detail. Gillman teaches - [(d)] wherein computing the at least one moment comprises [(d.1)] computing a contraction of the tensor network (Gillman, right column of p. 2, “2.2. Tensor Networks for Hidden Markov Models,” first paragraph, teaches “A TN is a collection of tensors contracted together in a given pattern, typically specified by a graph. . . . Since the matrices are rank-2 tensors and the vector a rank-1 tensor, this can be considered a [tensor network (TN)] consisting of T + 1 tensors, where the contraction pattern is given by the usual matrix products. Performing such a contraction produces a new vector, | P T [(that is, computing a contraction of the tensor network )]”). Claim Rejections – 35 U.S.C. § 103 9. The following is a quotation of 35 U.S.C. § 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. 10. The factual inquiries for establishing a background for determining obviousness under 35 U.S.C. § 103 are summarized as follows: 1. Determining the scope and contents of the prior art. 2. Ascertaining the differences between the prior art and the claims at issue. 3. Resolving the level of ordinary skill in the pertinent art. 4. Considering objective evidence present in the application indicating obviousness or nonobviousness. 11. This application currently names joint inventors. In considering patentability of the claims the Examiner presumes that the subject matter of the various claims was commonly owned as of the effective filing date of the claimed invention(s) absent any evidence to the contrary. Applicant is advised of the obligation under 37 CFR 1.56 to point out the inventor and effective filing dates of each claim that was not commonly owned as of the effective filing date of the later invention in order for the Examiner to consider the applicability of 35 U.S.C. § 102(b)(2)(C) for any potential 35 U.S.C. § 102(a)(2) prior art against the later invention. 12. Claims 4-8 and 15-19 are rejected under 35 U.S.C. § 103 as being unpatentable over Gillman et al., "A Tensor Network Approach to Finite Markov Decision Processes," arXiv (2020) [hereinafter Gillman] in view of Glasser et al., “Expressive power of tensor-network factorizations for probabilistic modeling,” arXiv (2019) [hereinafter Glasser]. Regarding claims 4 and 15, Gillman teaches all of the limitations of claims 1 and 12, respectively, as described above in detail. Though Gillman teaches a tensor network representation for dynamics of a matrix product state (MPS) representation (also known as the tensor train decomposition), Gillman, however, does not explicitly teach - [(c)] wherein encoding the tensor into the tensor network comprises [(c.3)] factorizing the tensor into the tensors of the tensor network by processing the tensor so that the following equation is solved: PNG media_image1.png 35 92 media_image1.png Greyscale where P is the resulting normalized factorization into the tensors of the tensor network, T is the encoded tensor, and ZT is a predetermined normalization factor ZT=ΣX1, . . . , xN TX 1 , . . . , xN, with X1, . . . , XN being respective N configurations of the plurality of discrete random variables of the probability distribution, TX 1 , . . . , xN being the tensor for the respective configuration, and N being the number of discrete random variables in the plurality of discrete random variables But Glasser teaches - [(c)] wherein encoding the tensor into the tensor network comprises [(c.3)] factorizing the tensor into the tensors of the tensor network by processing the tensor so that the following equation is solved: PNG media_image1.png 35 92 media_image1.png Greyscale where P is the resulting normalized factorization into the tensors of the tensor network, T is the encoded tensor, and ZT is a predetermined normalization factor ZT=ΣX1, . . . , xN TX 1 , . . . , xN, with X1, . . . , XN being respective N configurations of the plurality of discrete random variables of the probability distribution, TX 1
Read full office action

Prosecution Timeline

Apr 26, 2022
Application Filed
Nov 25, 2025
Non-Final Rejection — §101, §102, §103 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12591815
METHOD AND SYSTEM FOR UPDATING MACHINE LEARNING BASED CLASSIFIERS FOR RECONFIGURABLE SENSORS
2y 5m to grant Granted Mar 31, 2026
Patent 12585917
REINFORCEMENT LEARNING USING ADVANTAGE ESTIMATES
2y 5m to grant Granted Mar 24, 2026
Patent 12547759
PRIVACY PRESERVING MACHINE LEARNING MODEL TRAINING
2y 5m to grant Granted Feb 10, 2026
Patent 12530613
SYSTEMS AND METHODS FOR PERFORMING QUANTUM EVOLUTION IN QUANTUM COMPUTATION
2y 5m to grant Granted Jan 20, 2026
Patent 12518214
DISTRIBUTED MACHINE LEARNING SYSTEMS INCLUDING GENERATION OF SYNTHETIC DATA
2y 5m to grant Granted Jan 06, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
37%
Grant Probability
55%
With Interview (+18.0%)
4y 8m
Median Time to Grant
Low
PTA Risk
Based on 134 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month