DETAILED ACTION
Notice of Pre-AIA or AIA Status
The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA .
Information Disclosure Statement
The information disclosure statement (IDS), submitted on 12/15/2023, is being considered by the examiner.
Claim Rejections - 35 USC § 101
35 U.S.C. 101 reads as follows:
Whoever invents or discovers any new and useful process, machine, manufacture, or composition of matter, or any new and useful improvement thereof, may obtain a patent therefor, subject to the conditions and requirements of this title.
35 U.S.C. 101 requires that a claimed invention must fall within one of the four eligible categories of invention (i.e. process, machine, manufacture, or composition of matter) and must not be directed to subject matter encompassing a judicially recognized exception as interpreted by the courts. Three categories of subject matter are found to be judicially recognized exceptions to 35 U.S.C. § 101 (i.e. patent ineligible) (1) laws of nature, (2) physical phenomena, and (3) abstract ideas. To be patent-eligible, a claim directed to a judicial exception must as whole be directed to significantly more than the exception itself. Hence, the claim must describe a process or product that applies the exception in a meaningful way, such that it is more than a drafting effort designed to monopolize the exception.
Claim 15 is rejected under 35 U.S.C. 101 as not falling within one of the four statutory categories of invention because the broadest reasonable interpretation of the instant claim in view of the specification encompasses transitory signals. But transitory signals are not within one of the four statutory categories (i.e. non-statutory subject matter). However, claims directed toward a non-transitory computer readable medium may qualify as a manufacture and make the claim patent-eligible subject matter. Therefore, amending the claims to recite a "non-transitory computer-readable medium" would resolve this issue.
Claim Rejections - 35 USC § 102
The following is a quotation of the appropriate paragraphs of 35 U.S.C. 102 that form the basis for the rejections under this section made in this Office action:
A person shall be entitled to a patent unless –
(a)(1) the claimed invention was patented, described in a printed publication, or in public use, on sale or otherwise available to the public before the effective filing date of the claimed invention.
Claim 15 is rejected under 35 U.S.C. 102(a)(1) as being anticipated by Rosewarne (AU Patent Application Publication 2021202140 A1), (“Rosewarne”).
Regarding claim 15, it is rejected under 35 U.S.C. 102(a)(1) with the following reasons.
Claim 15 recites “a bit stream generated by a method of encoding feature information of an image, the encoding method comprising ….”, which is a product by process claim limitation where the product is the bit stream and the process is the method steps to generate the bitstream. MPEP §2113 recites “Product-by-Process claims are not limited to the manipulations of the recited steps, only the structure implied by the steps”. Thus, the scope of the claim is the storage medium storing the bitstream (with the structure implied by the method steps). The structure includes the modes, coding block, flags and values manipulated by the steps.
“To be given patentable weight, the printed matter and associated product must be in a functional relationship. A functional relationship can be found where the printed matter performs some function with respect to the product to which it is associated”. MPEP §2111.05(I)(A). When a claimed computer-readable medium merely serves as a support for information or data, no functional relationship exists. MPEP §2111.05(III). The storage medium storing the claimed bitstream in claim 15 merely services as a support for the storage of the bitstream and provides no functional relationship between the stored bitstream and storage medium. Therefor the structure bitstream, which scope is implied by the method steps, is non-functional descriptive material and given no patentable weight. MPEP §2111.05(III). Thus, the claim scope is just a storage medium storing data, and it is anticipated by Rosewarne’s recitation of a storage medium storing a bitstream in paragraphs [00054, 00057, 00059].
Claim Rejections - 35 USC § 103
The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action:
A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102 of this title, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made.
In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status.
The factual inquiries set forth in Graham v. John Deere Co., 383 U.S. 1, 148 USPQ 459 (1966), that are applied for establishing a background for determining obviousness under pre-AIA 35 U.S.C. 103(a) are summarized as follows:
1. Determining the scope and contents of the prior art.
2. Ascertaining the differences between the prior art and the claims at issue.
3. Resolving the level of ordinary skill in the pertinent art.
4. Considering objective evidence present in the application indicating obviousness or nonobviousness.
This application currently names joint inventors. In considering patentability of the claims under pre-AIA 35 U.S.C. 103(a), the examiner presumes that the subject matter of the various claims was commonly owned at the time any inventions covered therein were made absent any evidence to the contrary. Applicant is advised of the obligation under 37 CFR 1.56 to point out the inventor and invention dates of each claim that was not commonly owned at the time a later invention was made in order for the examiner to consider the applicability of pre-AIA 35 U.S.C. 103(c) and potential pre-AIA 35 U.S.C. 102(e), (f) or (g) prior art under pre-AIA 35 U.S.C. 103(a).
Claims 1-8 and 12-15 are rejected under 35 U.S.C. 103 as being unpatentable over Rosewarne (AU Patent Application Publication 2021202140 A1), (“Rosewarne”), in view of Yang et al. (US Patent 11,501,447 B2), (“Yang”).
Regarding claim 1, Rosewarne meets the claim limitations as follow.
A method of decoding feature information (A method of decoding feature maps from encoded data) [Rosewarne: 00016, Abstract] of an image (an image or video frame) [Rosewarne: para. 0006] performed by a decoding apparatus (an apparatus
for decoding feature maps) [Rosewarne: para. 00017], the method comprising:obtaining at least one feature map (obtained from the decoded metadata 155, such that each feature map belongs to one group and one or more groups are indicated in the feature map groups 820) [Rosewarne: para. 000137] for a first image (obtained from a previously decoded frame) [Rosewarne: para. 000114]; determining at least one feature transform matrix for the feature map (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122]: and inversely transforming (the residual coefficients 636 are passed through an inverse secondary transform module 644, operating in accordance with the secondary transform index 688 to produce intermediate inverse transform coefficients) [Rosewarne: para. 000121] a plurality of features included in the feature map ((The intermediate inverse transform coefficients 642 are inverse quantised by a dequantiser module 640 according to the quantisation parameter 692 to produce inverse transform coefficients) [Rosewarne: para. 000121]; (inverse quantisation for samples decoded from the encoded data to the feature maps, wherein, for at least a part of the samples decoded from the encoded data, one parameter is shared by a group of samples corresponding to a plurality of feature maps) [Rosewarne: para. 00018] – Rosewarne teaches that feature map data are processed by the inverse quantize and the inverse transformation modules) based on the determined feature transform matrix (The inverse primary transform module 648 applies DCT-2 transforms horizontally and vertically, constrained by the maximum available transform size as described with reference to the forward primary transform module 626. The types of inverse transform performed by the inverse secondary transform module 644 correspond with the types of forward transform performed by the forward secondary transform module 630. The types of inverse transform performed by the inverse primary transform module 648 correspond with the types of primary transform performed by the primary transform module 626) [Rosewarne: para. 000121], wherein the at least one feature transform matrix (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122] comprises a global feature transform matrix commonly applied to two or more features ((The 'channels' dimension indicates the number of concurrent 'feature maps' for a given tensor and the height and width dimensions indicate the size of the feature maps at the particular stage of the CNN. Channel count varies through a CNN according to the network architecture. Feature map size also varies, depending on subsampling occurring in specific network layers.) [Rosewarne: para. 0005, 00056]; (In an arrangement of the CNN backbone 310, the dimensionality of the tensors and thus the size of the resulting feature maps is selected to be aligned to the block size of the VVC standard. With generally rectangular video and a default CTU size of 128x128, the feature map widths and heights may be powers of two, e.g. the sizes for the three layers may be 128x64, 64x32, and 32x 16. Feature map sizes being powers of two results in greater alignment of the packed features with available block sizes in the VVC standard resulting from quadtree, binary, or ternary splits and reduced potential for coding artefacts in one feature map being caused by the contents of an adjacent feature map) [Rosewarne: para. 000197] – Rosewarne discloses that dimensions of the tensor with associated with the transform matrix are selected to be aligned with the block sizes of the VVC standard and resulting in the greater alignment of the packed features. It can be considered as a global matrix of the claim limitation), and wherein the global feature transform matrix is generated in advance based on a predetermined feature data set (The group determiner module 510 assigns the feature maps (channels) of the input tensors 115 into feature map groups 512, based either on a predetermined criteria or on some measure of the data present in the tensors 115) [Rosewarne: para. 00087] obtained from a second image.
Rosewarne does not explicitly disclose the following claim limitations (Emphasis added).
the at least one feature transform matrix is generated in advance based on a predetermined feature data set obtained from a second image.
However, in the same field of endeavor Yang further discloses the claim limitations and the deficient claim limitations as follows:
wherein the at least one feature transform matrix comprises a global feature transform matrix commonly applied to two or more features (applying the transform matrix to the object features resulting in transformed object features, and obtaining a predicted object mask associated with the second image frame by decoding the transformed object features.) [Yang: col. 2, line 7-11], and
wherein the global feature transform matrix is generated in advance based on a predetermined feature data set obtained from a second image (computing a transform matrix based on the image features of the first image frame and the image features of the second image frame) [Yang: col. 2, line 5-7].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claims 2 and 13, Rosewarne meets the claim limitations as set forth in claims 1 and 12. Rosewarne further meets the claim limitations as follow.
wherein the number of global feature transform matrix is differently determined ((The backbone layers of the CNN may produce multiple tensors as output, for example, corresponding to different spatial scales of an input image represented by the video frame data 113. A 'feature pyramid network' (FPN) architecture may result in three tensors, corresponding to three layers, output from the backbone 114, with varying spatial resolution and channel count. The feature map quantiser and packer 116 receives tensors 115, which are output from the CNN backbone 114) [Rosewarne: para. 00087] – Note: Multiple tensors with varying spatial resolution indicate that the number of global feature transform matrix is differently determined) according to a type of a target task performed based on the feature map ((Fig. 1 is a schematic block diagram showing a distributed machine task system) [Rosewarne: 00024]; (The 'task performance' of a CNN is measured by comparing the result of the CNN in performing a task using specific input with a provided ground truth (i.e., 'training data'), generally prepared by humans and intended to indicate a 'correct' result) [Rosewarne: 0007] – Note: Paragraph [0290] of the application spec describes different tasks such as a machine task versus a hybrid of a machine task versus and a human task. As a result, Rosewarne discloses the claim limitation).
Regarding claims 3 and 14, Rosewarne meets the claim limitations as set forth in claims 1 and 12. Rosewarne further meets the claim limitations as follow.
wherein the global feature transform matrix is generated by applying a predetermined dimensionality reduction algorithm to the feature data set (Where a convolution stage has a 'stride' greater than one, an output tensor from the convolution has a lower spatial resolution than a corresponding input tensor. Operations such as 'max pooling' also reduce spatial size of the output tensor compared to the input tensor.) [Rosewarne: para. 0004].
Regarding claim 4, Rosewarne meets the claim limitations as set forth in claim 1. Rosewarne further meets the claim limitations as follow.
wherein the feature data set comprises at least one feature selected from among a plurality of features obtained from the second image (feature maps are extracted from each frame according to a packing format to produce unpacked feature maps 812. The unpacked feature maps 812 include sample values as present in the decoded frame 14 7. Packing formats are described further with reference to Figs. 11-13. The sets of feature maps in the unpacked feature maps 812 are assigned to groups according to feature map groups 820, obtained from the decoded metadata 155, such that each feature map belongs to one group and one or more groups are indicated in the feature map groups 820) [Rosewarne: para. 000137; Fig. 8].
Rosewarne does not explicitly disclose the following claim limitations (Emphasis added).
a plurality of features obtained from a second image.
However, in the same field of endeavor Yang further discloses the claim limitations and the deficient claim limitations as follows:
a plurality of features obtained from the second image (obtaining a predicted object mask associated with the second image frame by decoding the transformed object features) [Yang: col. 2, line 7-11].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claim 5, Rosewarne meets the claim limitations as set forth in claim 4. Rosewarne further meets the claim limitations as follow.
wherein the selected feature has a modified data structure within the feature data set (Sample-wise interleaving results in the higher structural detail of the four feature maps being shared by the same coding tree structure, with detail between the four feature maps varying from sample to sample. Accordingly, a common coding
tree structure and shared residual (except for local differences necessary to code the adjacent samples of different feature maps) is achieved, resulting in an increase in compression efficiency. Once all groups of size four have been packed into the monochrome frame 1202 for a given layer, the remaining feature maps, such as feature map 1214, are packed adjacently, based on grouping but not in an interleaved manner. The remaining feature maps may be assigned to groups of any size as their group composition does not affect the packing process, aside from the order of packing. For the next layer, groups of four, such as group 1220 are packed in a sample-wise interleaved manner followed by feature maps belonging to groups of other sizes, such as feature map 1224. For the final layer, groups of four, such as group 1230 are packed in a sample-wise interleaved manner followed by feature maps belonging to groups of other sizes, such as feature map 1234) [Rosewarne: para. 000149; Fig. 12] – Please see details of the modified data structure in Fig. 12).
In addition, in the same field of endeavor Yang further discloses the claim limitations and the deficient claim limitations as follows:
a modified data structure (Communication media may be embodied by computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. The term "modulated data signal" may describe a signal that has one or more characteristics set or changed in such a manner as to encode information in the signal) [Yang: col. 14, line 60-67].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claim 6, Rosewarne meets the claim limitations as set forth in claim 1. Rosewarne further meets the claim limitations as follow.
wherein the feature data set comprises only a plurality of regions of interest (ROls) obtained from the second image (The RPN head module 1020 performs a convolution on the input tensors, producing an intermediate tensor which is fed into two subsequent sibling layers, one for classifications and one for bounding box, or 'region of interest' (ROI), regression, as classification and bounding boxes 1022. The classification and bounding boxes 1022 are passed to an NMS module 1024 which prunes out redundant bounding boxes by removing overlapping boxes with a lower score to produce pruned bounding boxes 1026. The bounding boxes 1026 are passed to a region of interest (ROI) pooler 1028. The ROI pooler 1028 produces fixed-size feature maps from various input size maps using max pooling operations, where a subsampling takes the maximum value in each group of input values to produce one output value in the output tensor. Input to the ROI pooler 1028 are the P2-P5 feature maps 1010, 1012, 1014, and 1016, and region of interest proposals 1026. Each proposal (ROI) from 1026 is associated with a portion of the feature maps (1010-1016) to produce a fixed-size map.) [Rosewarne: para. 000143-000144; Fig. 10]).
Rosewarne does not explicitly disclose the following claim limitations (Emphasis added).
obtained from a second image.
However, in the same field of endeavor Yang further discloses the deficient claim limitations as follows:
obtained from the second image (obtaining a predicted object mask associated with the second image frame by decoding the transformed object features) [Yang: col. 2, line 7-11].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claim 7, Rosewarne meets the claim limitations as set forth in claim 1. Rosewarne further meets the claim limitations as follow.
wherein the feature data set is individually generated for ROI features and non-ROI features obtained from the second image (The RPN head module 1020 performs a convolution on the input tensors, producing an intermediate tensor which is fed into two subsequent sibling layers, one for classifications and one for bounding box, or 'region of interest' (ROI)) [Rosewarne: para. 000143; Fig. 10]– Note: Rosewarne discloses that his invention generates both ROI (i.e. one for bounding box, or 'region of interest' (ROI)) and non-ROI (i.e. one for classifications)).
Rosewarne does not explicitly disclose the following claim limitations (Emphasis added).
obtained from a second image.
However, in the same field of endeavor Yang further discloses the deficient claim limitations as follows:
obtained from the second image (obtaining a predicted object mask associated with the second image frame by decoding the transformed object features) [Yang: col. 2, line 7-11].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claim 8, Rosewarne meets the claim limitations as set forth in claim 1. Rosewarne further meets the claim limitations as follow.
wherein the feature transform matrix is determined based on matrix index information obtained from a bitstream (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122].
Regarding claim 11, Rosewarne meets the claim limitations as set forth in claim 1. Rosewarne further meets the claim limitations as follow.
wherein the feature transform matrix is determined based on matrix index information obtained from a bitstream (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122].
Regarding claim 12, Rosewarne meets the claim limitations as follow.
A method of encoding feature information (A method of decoding feature maps from encoded data) [Rosewarne: 00016, Abstract] of an image (an image or video frame) [Rosewarne: para. 0006] performed by a encoding apparatus (an apparatus
for decoding feature maps) [Rosewarne: para. 00020], the method comprising:obtaining at least one feature map (obtained from the decoded metadata 155, such that each feature map belongs to one group and one or more groups are indicated in the feature map groups 820) [Rosewarne: para. 000137] for a first image (obtained from a previously decoded frame) [Rosewarne: para. 000114]; determining at least one feature transform matrix for the feature map (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122]: and transforming (The transform is applied) [Rosewarne: para. 00013] a plurality of features included in the feature map ((samples to be encoded into the encoded data to encode the feature maps, wherein, for at least a part of the samples to be encoded into the encoded data, one parameter is shared by a group of samples corresponding to a plurality of feature maps) [Rosewarne: para. 00019]) based on the determined feature transform matrix (The inverse primary transform module 648 applies DCT-2 transforms horizontally and vertically, constrained by the maximum available transform size as described with reference to the forward primary transform module 626. The types of inverse transform performed by the inverse secondary transform module 644 correspond with the types of forward transform performed by the forward secondary transform module 630. The types of inverse transform performed by the inverse primary transform module 648 correspond with the types of primary transform performed by the primary transform module 626) [Rosewarne: para. 000121], wherein the at least one feature transform matrix (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122] comprises a global feature transform matrix commonly applied to two or more features ((The 'channels' dimension indicates the number of concurrent 'feature maps' for a given tensor and the height and width dimensions indicate the size of the feature maps at the particular stage of the CNN. Channel count varies through a CNN according to the network architecture. Feature map size also varies, depending on subsampling occurring in specific network layers.) [Rosewarne: para. 0005, 00056]; (In an arrangement of the CNN backbone 310, the dimensionality of the tensors and thus the size of the resulting feature maps is selected to be aligned to the block size of the VVC standard. With generally rectangular video and a default CTU size of 128x128, the feature map widths and heights may be powers of two, e.g. the sizes for the three layers may be 128x64, 64x32, and 32x 16. Feature map sizes being powers of two results in greater alignment of the packed features with available block sizes in the VVC standard resulting from quadtree, binary, or ternary splits and reduced potential for coding artefacts in one feature map being caused by the contents of an adjacent feature map) [Rosewarne: para. 000197] – Rosewarne discloses that dimensions of the tensor with associated with the transform matrix are selected to be aligned with the block sizes of the VVC standard and resulting in the greater alignment of the packed features. It can be considered as a global matrix of the claim limitation), and wherein the global feature transform matrix is generated in advance based on a predetermined feature data set (The group determiner module 510 assigns the feature maps (channels) of the input tensors 115 into feature map groups 512, based either on a predetermined criteria or on some measure of the data present in the tensors 115) [Rosewarne: para. 00087] obtained from a second image.
Rosewarne does not explicitly disclose the following claim limitations (Emphasis added).
the at least one feature transform matrix is generated in advance based on a predetermined feature data set obtained from a second image.
However, in the same field of endeavor Yang further discloses the claim limitations and the deficient claim limitations as follows:
wherein the at least one feature transform matrix comprises a global feature transform matrix commonly applied to two or more features (applying the transform matrix to the object features resulting in transformed object features, and obtaining a predicted object mask associated with the second image frame by decoding the transformed object features.) [Yang: col. 2, line 7-11], and
wherein the global feature transform matrix is generated in advance based on a predetermined feature data set obtained from a second image (computing a transform matrix based on the image features of the first image frame and the image features of the second image frame) [Yang: col. 2, line 5-7].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne with Yang to program the system to implement of Yang’s method.
Therefore, the combination of Rosewarne with Yang will enable the system to improve feature coding efficiency [Yang: col. 6, line 55-67].
Regarding claim 15, this claim lists essentially the same elements as claim 12, but is drawn to a product-by-process claim rather than a method. Therefore, the rationales for the rejection of claim 12 apply to the instant claim. Moreover, a storage medium storing data is taught by a storage medium storing a bitstream, disclosed in paragraphs [00054, 00057, 00059] of Rosewarne.
Claim 9 is rejected under 35 U.S.C. 103 as being unpatentable over Rosewarne (AU Patent Application Publication 2021202140 A1), (“Rosewarne”), in view of Yang et al. (US Patent 11,501,447 B2), (“Yang”), in view of Afrouzi et al. (US Patent 11,153,503 B1), (“Afrouzi”).
Regarding claim 9, Rosewarne and Yang meet the claim limitations as set forth in claim 1. Rosewarne and Yang further meet the claim limitations as follow.
wherein features (A method of decoding feature maps from encoded data) [Rosewarne: 00016, Abstract] obtained from the second image (obtaining a predicted object mask associated with the second image frame by decoding the transformed object features) [Yang: col. 2, line 7-11] are clustered and included in the feature data set (feature maps are extracted from each frame according to a packing format to produce unpacked feature maps 812. The unpacked feature maps 812 include sample values as present in the decoded frame 14 7. Packing formats are described further with reference to Figs. 11-13. The sets of feature maps in the unpacked feature maps 812 are assigned to groups according to feature map groups 820, obtained from the decoded metadata 155, such that each feature map belongs to one group and one or more groups are indicated in the feature map groups 820) [Rosewarne: para. 000137; Figs. 8, 11-13] – Rosewarne discloses that the features are packed in the packing format).
Rosewarne and Yang do not explicitly disclose the following claim limitations (Emphasis added).
wherein features are clustered.
However, in the same field of endeavor Afrouzi further the deficient claim limitations as follows:
wherein features are clustered (In some embodiments, the processor may employ unsupervised learning or clustering to organize unlabeled data into groups based on their similarities. Clustering may involve assigning data points to clusters wherein data points in the same cluster are as similar as possible. In some
embodiments, clusters may be identified using similarity measures, such as distance.) [Afrouzi: col. 47, line 54-60].
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne and Yang with Afrouzi to program the system to implement of Afrouzi’s method.
Therefore, the combination of Rosewarne and Yang with Afrouzi will enable the system to improve accuracy of the map [Afrouzi: col. 33, line 4-7].
Claim 10 is rejected under 35 U.S.C. 103 as being unpatentable over Rosewarne (AU Patent Application Publication 2021202140 A1), (“Rosewarne”), in view of Yang et al. (US Patent 11,501,447 B2), (“Yang”), in view of Rosewarne et al. (US Patent 12,309,389 B2), (“Rosewarne_389”).
Regarding claim 10, Rosewarne and Yang meet the claim limitations as set forth in claim 1. Rosewarne and Yang further meet the claim limitations as follow.
wherein the feature transform matrix is determined ((applying the transform matrix to the object features resulting in transformed object features, and obtaining a predicted object mask associated with the second image frame by decoding the transformed object features.) [Yang: col. 2, line 7-11]; (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122]) based on a most probable matrix (MPM) list for a current feature, andwherein the MPM list comprises a feature transform matrix (selected matrix signalled in the bitstream 120 using an index to identify which matrix of the set of matrices is to be used) [Rosewarne: para. 000122] for a feature reconstructed before the current feature as a MPM candidate (The decoded syntax elements are used to reconstruct parameters within the video decoder 144. Parameters include residual coefficients (represented by an arrow 724), a quantisation parameter 774, a secondary transform index 770, and mode selection information such as an intra prediction mode (represented by an arrow 758). The mode selection information also includes information) [Rosewarne: 000129]; (The reconstructed transform coefficients 7 40 are passed to an inverse primary transform module 744. The module 744 transforms the coefficients 740 from the frequency domain back to the spatial domain. The inverse primary transform module 744 applies inverse DCT-2 transforms horizontally and vertically, constrained by the maximum available transform size as described with reference to the forward primary transform module 626. The result of operation of the module 744 is a block ofresidual samples, represented by an arrow 748. The block of residual samples 748 is equal in size to the corresponding CB. The residual samples 748 are supplied to a summation module 750. At the summation module 750 the residual samples 748 are added to a decoded PB (represented as 752) to produce a block of reconstructed samples) [Rosewarne: para. 000131-000132; Fig. 7]).
Rosewarne and Yang do not explicitly disclose the following claim limitations (Emphasis added).
based on a most probable matrix (MPM) list.
However, in the same field of endeavor Afrouzi further the deficient claim limitations as follows:
a most probable matrix (MPM) list ((Blocks of size 4x4 have 35 possible matrix intra prediction modes, otherwise blocks not exceeding 8x8 in size (i.e. 4x8, 8x4, and 8x8) have 19 possible matrix intra prediction modes. For other block sizes there are 11 possible matrix intra prediction modes. Typically, a list of 'most probable matrix modes' (MPMs) is generated in the video encoder 114.) [Rosewarne_389: col. 15, line 61-67]; (For 4x4 blocks, the 35 possible MW modes can be encoded using a 5- or 6-bit code, with 5 bits used for MW modes 0-28 and 6 bits used for MW modes 29-34. For 4x8, 8x4, and 8x8 blocks, the 19 possible MIP modes may be encoded using a 4- or 5-bit code, with 4 bits used for MW modes 0-12 and 5 bits used for MIP modes 13-18. For other sized blocks, the 11 possible MW modes may be encoded using a 3- or 4-bit code, with 3 bits used for MW modes 0-4 and 4 bits used for MIP modes 5-10) [Rosewarne_389: col. 16, line 43-51]).
It would have been obvious to one with an ordinary skill in the art before the effective filing date of the claimed invention to modify the teachings of Rosewarne and Yang with Rosewarne_389 to program the system to implement of Rosewarne_389’s method.
Therefore, the combination of Rosewarne and Yang with Rosewarne_389 will enable the system to reduce the worst case memory bandwidth of the coefficient memory, reducing complexity and hardware cost without resulting in a proportionate degradation in coding efficiency [Rosewarne_389: col. 19, line 20-23].
Reference Notice
Additional prior arts, included in the Notice of Reference Cited, made of record and not relied upon is considered pertinent to applicant's disclosure.
Schmidt; Arthur W. (US Patent US-3720912-A) teaches most probable matrix.
Rosewarne, Christopher James (AU 2021202142 A1) discloses an apparatus that generates encoded data including encoded data of a feature map based on a neural network.
Contact Information
Any inquiry concerning this communication or earlier communications from the examiner should be directed to Philip Dang whose telephone number is (408) 918-7529. The examiner can normally be reached on Monday-Thursday between 8:30 am - 5:00 pm (PST).
Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice.
If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Sath Perungavoor can be reached on 571-272-7455. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300.
Information regarding the status of an application may be obtained from the Patent Application Information Retrieval (PAIR) system. Status information for published applications may be obtained from either Private PAIR or Public PAIR. Status information for unpublished applications is available through Private PAIR only. For more information about the PAIR system, see http://pair-direct.uspto.gov. Should you have questions on access to the Private PAIR system, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative or access to the automated information system, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000./Philip P. Dang/Primary Examiner, Art Unit 2488