Prosecution Insights
Last updated: April 19, 2026
Application No. 18/678,716

Apparatus, Methods and Computer Programs for Encoding Spatial Metadata

Non-Final OA §102§DP
Filed
May 30, 2024
Examiner
KRZYSTAN, ALEXANDER J
Art Unit
2694
Tech Center
2600 — Communications
Assignee
Nokia Technologies Oy
OA Round
1 (Non-Final)
81%
Grant Probability
Favorable
1-2
OA Rounds
3y 1m
To Grant
88%
With Interview

Examiner Intelligence

Grants 81% — above average
81%
Career Allow Rate
913 granted / 1121 resolved
+19.4% vs TC avg
Moderate +7% lift
Without
With
+6.9%
Interview Lift
resolved cases with interview
Typical timeline
3y 1m
Avg Prosecution
38 currently pending
Career history
1159
Total Applications
across all art units

Statute-Specific Performance

§101
2.7%
-37.3% vs TC avg
§103
37.1%
-2.9% vs TC avg
§102
24.3%
-15.7% vs TC avg
§112
21.0%
-19.0% vs TC avg
Black line = Tech Center average estimate • Based on career data from 1121 resolved cases

Office Action

§102 §DP
Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Double Patenting The nonstatutory double patenting rejection is based on a judicially created doctrine grounded in public policy (a policy reflected in the statute) so as to prevent the unjustified or improper timewise extension of the “right to exclude” granted by a patent and to prevent possible harassment by multiple assignees. A nonstatutory double patenting rejection is appropriate where the conflicting claims are not identical, but at least one examined application claim is not patentably distinct from the reference claim(s) because the examined application claim is either anticipated by, or would have been obvious over, the reference claim(s). See, e.g., In re Berg, 140 F.3d 1428, 46 USPQ2d 1226 (Fed. Cir. 1998); In re Goodman, 11 F.3d 1046, 29 USPQ2d 2010 (Fed. Cir. 1993); In re Longi, 759 F.2d 887, 225 USPQ 645 (Fed. Cir. 1985); In re Van Ornum, 686 F.2d 937, 214 USPQ 761 (CCPA 1982); In re Vogel, 422 F.2d 438, 164 USPQ 619 (CCPA 1970); In re Thorington, 418 F.2d 528, 163 USPQ 644 (CCPA 1969). A timely filed terminal disclaimer in compliance with 37 CFR 1.321(c) or 1.321(d) may be used to overcome an actual or provisional rejection based on nonstatutory double patenting provided the reference application or patent either is shown to be commonly owned with the examined application, or claims an invention made as a result of activities undertaken within the scope of a joint research agreement. See MPEP § 717.02 for applications subject to examination under the first inventor to file provisions of the AIA as explained in MPEP § 2159. See MPEP § 2146 et seq. for applications not subject to examination under the first inventor to file provisions of the AIA . A terminal disclaimer must be signed in compliance with 37 CFR 1.321(b). The filing of a terminal disclaimer by itself is not a complete reply to a nonstatutory double patenting (NSDP) rejection. A complete reply requires that the terminal disclaimer be accompanied by a reply requesting reconsideration of the prior Office action. Even where the NSDP rejection is provisional the reply must be complete. See MPEP § 804, subsection I.B.1. For a reply to a non-final Office action, see 37 CFR 1.111(a). For a reply to final Office action, see 37 CFR 1.113(c). A request for reconsideration while not provided for in 37 CFR 1.113(c) may be filed after final for consideration. See MPEP §§ 706.07(e) and 714.13. The USPTO Internet website contains terminal disclaimer forms which may be used. Please visit www.uspto.gov/patent/patents-forms. The actual filing date of the application in which the form is filed determines what form (e.g., PTO/SB/25, PTO/SB/26, PTO/AIA /25, or PTO/AIA /26) should be used. A web-based eTerminal Disclaimer may be filled out completely online using web-screens. An eTerminal Disclaimer that meets all requirements is auto-processed and approved immediately upon submission. For more information about eTerminal Disclaimers, refer to www.uspto.gov/patents/apply/applying-online/eterminal-disclaimer. Claims 1-20 are rejected on the ground of nonstatutory double patenting as being unpatentable over claim 11 of U.S. Patent No. 12027174. Although the claims at issue are not identical, they are not patentably distinct from each other because the patent and application claim 1 each claim: An apparatus comprising: at least one processor; and at least one memory storing instructions that, when executed with the at least one processor, cause the apparatus at least to: obtain spatial audio content (patent claim 1,11, the decoding device receiving from the encoding device of claim 1); decode encoded spatial metadata associated with the spatial audio content based (the audio decoder of claim 11), at least partially, on a configuration parameter (patent, from the encoder of claim 1) indicative of a codec configuration used to encode, at least in part, spatial metadata (via the codec of claim 1); determine at least one prototype audio signal based, at least partially, on the spatial audio content and a configuration of at least one output type (any of the subsequent signaling recovered once the decoder of claim 11 receives the signaling from the encoder claim 1); determine one or more spatial audio signals based, at least partially, on the at least one prototype audio signal and the decoded spatial metadata (the output audio signals from the decoder of claim 1 that is based on the encoder of claim 1 of the patent); and provide the one or more spatial audio signals (the function of the decoder of claim 11) . Claim Rejections - 35 USC § 102 The following is a quotation of the appropriate paragraphs of 35 U.S.C. 102 that form the basis for the rejections under this section made in this Office action: A person shall be entitled to a patent unless – (a)(2) the claimed invention was described in a patent issued under section 151, or in an application for patent published or deemed published under section 122(b), in which the patent or application, as the case may be, names another inventor and was effectively filed before the effective filing date of the claimed invention. Claim(s) 1-20 is/are rejected under 35 U.S.C. 102(a)(2) as being anticipated by Setiawan et al (US 11632549 B2). As per claim 1, Setiawan discloses an apparatus comprising: at least one processor; and at least one memory storing instructions that, when executed with the at least one processor (processor, software and memory are required to implement the following functions), cause the apparatus at least to: obtain spatial audio content (the encoder per para 26 encodes multichannel audio into a bitstream which is received by the decoder of fig. 9, quantized bits); decode encoded spatial metadata associated with the spatial audio content based, at least partially, on a configuration parameter indicative of a codec configuration used to encode, at least in part, spatial metadata (para 70: obtain mode information F/A (fixed/adaptive) defining the adaptive mode or a fixed mode. ); determine at least one prototype audio signal based, at least partially, on the spatial audio content and a configuration of at least one output type (any of the signaling associated with any of the subsequent processing as shown in fig. 9, 902-908; determine one or more spatial audio signals based, at least partially, on the at least one prototype audio signal and the decoded spatial metadata (part of block 909); and provide the one or more spatial audio signals (stage 910). As per claim 2, The apparatus of claim 1, wherein obtaining the spatial audio content comprises the instructions, when executed with the at least one processor, cause the apparatus to: obtain at least one encoded transport audio signal (the bitstream in 901); decode the at least one obtained encoded transport audio signal for rendering (902); and separate the at least one decoded transport audio signal and the decoded spatial metadata(steps 903-907). As per claim 3, the apparatus of claim 1, wherein determining the at least one prototype audio signal comprises the instructions, when executed with the at least one processor, cause the apparatus to: obtain at least one transport audio signal (902); and determine the at least one prototype audio signal based (per the claim 1 rejection), at least partially, on the at least one transport audio signal. As per claim 4, the apparatus of claim 3, wherein the instructions, when executed with the at least one processor, cause the apparatus to: determine a type of the at least one transport audio signal (based on the mode of the claim 1 rejection), wherein the one or more spatial audio signals are determined based, at least partially, on the determined type of the at least one transport audio signal (the mode determines the subsequent processing of fig. 9). As per claim 5, the apparatus of claim 3, wherein the at least one prototype audio signal comprises at least one of: the at least one transport audio signal (the processing between 901 and 902), or a processed version of the at least one transport audio signal. As per claim 6, the apparatus of claim 1, wherein the codec configuration used to encode, at least in part, the spatial metadata is based, at least partially, on a source format of the spatial audio content (per the mode of the claim 1 rejection). As per claim 7, the apparatus of claim 1, wherein a format of the spatial audio content is at least partially different from a format the at least one output type supports (in the situations where the bitstream comprises content and metadata for the first mode, and then later comprises content and metadata for the second mode). As per claim 8, the apparatus of claim 1, wherein a number of channels the spatial audio content comprises is at least partially different from a number of channels the at least one output type supports (in the situations where the bitstream comprises content and metadata for the first mode, and then later comprises content and metadata for the second mode and or different number of channels as part of the bitstream). As per claim 9, the apparatus of claim 1, wherein determining the one or more spatial audio signals comprises the instructions, when executed with the at least one processor, cause the apparatus to: determine a direct stream (output of 902) based, at least partially, on a direct prototype audio signal (signaling of 902) of the at least one prototype audio signal (input to 902) and at least a first part of the decoded spatial metadata (the mode per the claim 1 rejection); determine a diffuse stream (the bitstream can support known audio formats including 5.1 per para 2, where 5.1 comprises rear/diffuse channels) based, at least partially, on a diffuse prototype audio signal of the at least one prototype audio signal and at least a second part of the decoded spatial metadata (the mode as implemented by one of stage 903-907; and combine the direct stream (the center channel of 5.1) and the diffuse stream to determine the one or more spatial audio signals (to create the 910 signal). As per claim 10, the apparatus of claim 1, wherein the at least one prototype audio signal comprises at least one of: at least one diffuse prototype audio signal (per claim 9 rejection), or at least one direct prototype audio signal. As per claim 11, the apparatus of claim 1, wherein the decoded spatial metadata comprises at least one of: at least one direction parameter (the designation of a particular channel in the context of a multichannel signal is a parameter that indicates direction, for example the 5.1 standard cited above has a direction inherent to each defined channel), at least one direction of arrival of audio, at least one distance to an audio source, at least one coherence parameter, at least one energy ratio, at least one direct-to-total energy ratio, or at least one diffuse-to-total energy ratio. As per claim 12, the claim 1 rejection discloses a method comprising: obtaining spatial audio content; decoding encoded spatial metadata associated with the spatial audio content based, at least partially, on a configuration parameter indicative of a codec configuration used to encode, at least in part, spatial metadata; determining at least one prototype audio signal based, at least partially, on the spatial audio content and a configuration of at least one output type; determining one or more spatial audio signals based, at least partially, on the at least one prototype audio signal and the decoded spatial metadata; and providing the one or more spatial audio signals (per claim 1 rejection). As per claim 13, The method of claim 12, wherein the obtaining of the spatial audio content comprises: obtaining at least one encoded transport audio signal; decoding the at least one obtained encoded transport audio signal for rendering; and separating the at least one decoded transport audio signal and the decoded spatial metadata (per claim 2 rejection). As per claim 14, the method of claim 12, wherein the determining of the at least one prototype audio signal comprises: obtaining at least one transport audio signal; and determining the at least one prototype audio signal based, at least partially, on the at least one transport audio signal (per claim 3 rejection). As per claim 15, the method of claim 14, further comprising: determining a type of the at least one transport audio signal, wherein the one or more spatial audio signals are determined based, at least partially, on the determined type of the at least one transport audio signal (per claim 4 rejection). As per claim 16, the method of claim 14, wherein the at least one prototype audio signal comprises at least one of: the at least one transport audio signal, or a processed version of the at least one transport audio signal (per claim 5 rejection). As per claim 17, the method of claim 12, wherein the codec configuration used to encode, at least in part, the spatial metadata is based, at least partially, on a source format of the spatial audio content (per claim 6 rejection). As per claim 18, the method of claim 12, wherein the determining of the one or more spatial audio signals comprises: determining a direct stream based, at least partially, on a direct prototype audio signal of the at least one prototype audio signal and at least a first part of the decoded spatial metadata; determining a diffuse stream based, at least partially, on a diffuse prototype audio signal of the at least one prototype audio signal and at least a second part of the decoded spatial metadata; and combining the direct stream and the diffuse stream to determine the one or more spatial audio signals (per claim 9 rejection). As per claim 19, the method of claim 12, wherein the decoded spatial metadata comprises at least one of: at least one direction parameter, at lest one direction of arrival of audio, at least one distance to an audio source, at least one coherence parameter, at least one energy ratio, at least one direct-to-total energy ratio, or at least one diffuse-to-total energy ratio (per claim 11 rejection). As per claim 20, a non-transitory computer-readable medium comprising program instructions stored thereon for performing at least the following: causing obtaining of spatial audio content; decoding encoded spatial metadata associated with the spatial audio content based, at least partially, on a configuration parameter indicative of a codec configuration used to encode, at least in part, spatial metadata; determining at least one prototype audio signal based, at least partially, on the spatial audio content and a configuration of at least one output type; determining one or more spatial audio signals based, at least partially, on the at least one prototype audio signal and the decoded spatial metadata; and causing providing of the one or more spatial audio signals (required by the processor of the system of the claim 12 and 1 rejections). Any inquiry concerning this communication or earlier communications from the examiner should be directed to ALEXANDER KRZYSTAN whose telephone number is 571-272-7498, and whose email address is alexander.krzystan@uspto.gov The examiner can usually be reached on m-f 7:30-4:00 est. If attempts to reach the examiner by telephone or email are unsuccessful, the examiner’s supervisor, Fan Tsang can be reached on (571) 272-7547. The fax phone numbers for the organization where this application or proceeding is assigned are 571-273-8300 for regular communications and 571-273-8300 for After Final communications. /ALEXANDER KRZYSTAN/Primary Examiner, Art Unit 2653 Examiner Alexander Krzystan March 9, 2026
Read full office action

Prosecution Timeline

May 30, 2024
Application Filed
Mar 02, 2026
Non-Final Rejection — §102, §DP (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12598440
RENDERING OF OCCLUDED AUDIO ELEMENTS
2y 5m to grant Granted Apr 07, 2026
Patent 12593170
SWITCHING METHOD FOR AUDIO OUTPUT CHANNEL, AND DISPLAY DEVICE
2y 5m to grant Granted Mar 31, 2026
Patent 12573410
DECODER, ENCODER, AND METHOD FOR INFORMED LOUDNESS ESTIMATION IN OBJECT-BASED AUDIO CODING SYSTEMS
2y 5m to grant Granted Mar 10, 2026
Patent 12574675
Acoustic Device and Method
2y 5m to grant Granted Mar 10, 2026
Patent 12541554
TRANSCRIPT AGGREGATON FOR NON-LINEAR EDITORS
2y 5m to grant Granted Feb 03, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
81%
Grant Probability
88%
With Interview (+6.9%)
3y 1m
Median Time to Grant
Low
PTA Risk
Based on 1121 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month