Prosecution Insights
Last updated: April 19, 2026
Application No. 19/074,087

AUDIO PROCESSING METHOD AND ELECTRONIC DEVICE

Non-Final OA §103
Filed
Mar 07, 2025
Examiner
SHIBRU, HELEN
Art Unit
2484
Tech Center
2400 — Computer Networks
Assignee
Hisense Visual Technology Co. Ltd.
OA Round
1 (Non-Final)
59%
Grant Probability
Moderate
1-2
OA Rounds
3y 8m
To Grant
62%
With Interview

Examiner Intelligence

Grants 59% of resolved cases
59%
Career Allow Rate
443 granted / 756 resolved
+0.6% vs TC avg
Minimal +4% lift
Without
With
+3.7%
Interview Lift
resolved cases with interview
Typical timeline
3y 8m
Avg Prosecution
36 currently pending
Career history
792
Total Applications
across all art units

Statute-Specific Performance

§101
8.7%
-31.3% vs TC avg
§103
42.6%
+2.6% vs TC avg
§102
31.3%
-8.7% vs TC avg
§112
10.2%
-29.8% vs TC avg
Black line = Tech Center average estimate • Based on career data from 756 resolved cases

Office Action

§103
DETAILED ACTION Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Claim Rejections - 35 USC § 103 In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis (i.e., changing from AIA to pre-AIA ) for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status. The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. Claims 1-3, 9, 11-13 and 19 are rejected under 35 U.S.C. 103 as being unpatentable over Oughriss et al. (US PG PUB 2014/0118616 hereinafter referred as Oughriss) in view of Kim et al. (US Pat. No. 10, 972, 853). Regarding claim 1, Oughriss discloses an audio processing method, comprising: based on that a to-be-played multimedia file comprises multiple channels of audio tracks, parsing the to-be-played multimedia file to obtain the multiple channels of audio tracks of the to-be-played multimedia file (see figure 3 step 310 receiving a program streams with video packets and audio packets, the audio packets comprising first and second language; see figure 2 and paragraph 0022 program stream decoder separates the video streams and the audio streams; the streams identified by audio PIDS are processed by audio processor; first and second audio output identified by audio PID; see also paragraph 0015); determining at least two channels of target audio tracks from the multiple channels of audio tracks, and establishing a corresponding relationship between each channel of target audio track and each playback earphone (see figure 2, units 225 and 230, see paragraph 0022 the streams identified by audio PIDS are processed by audio processor; first and second audio output identified by audio PID; first audio sent to speaker and second audio sent to audio wireless transmitter; see figure 3, step 330 audio packets use by first user and second user; see paragraph 0023); decoding each channel of target audio track based on a preset decoder corresponding to each channel of target audio track (see figure 3, step 330 outputting the audio packets; see figure 2 audio output 245 and audio transmitter 250; audio sent to speaker and wireless transmitter; see paragraph 0023 streaming the second audio for the second language via any wireless technology); playing each channel of target audio track through the playback earphone corresponding to each channel of target audio track based on the corresponding relationship between each channel of target audio track and each playback earphone (see figure 2, AUDIO PID-SPN playing via wireless transmitter such as WIFI, Bluetooth and headphone, for instance; AUDIO-PID-ENG playing via speaker to be heard by group 280; see paragraphs 0022-0023). Claim 1 differs from Oughriss in that the claim further requires obtaining pulse modulation code corresponding to each channel and playing each channel based on pulse modulation code corresponding to each channel of target audio track. In the same field of endeavor Kim discloses obtaining pulse modulation code corresponding to each channel and playing each channel based on pulse modulation code corresponding to each channel of target audio track (see col. 2 lines 59-65 audio encoders receive input including in object-based audio which involves pulse-code-modulation data for single audio objects with associated metadata containing their location coordinates; see col. 7 lines 14-26 decoding device configured to decode object based audio; the decoding device including format generation unit and rendering unit; the rendering unit may be implemented in a separate device, such as loudspeaker, headphone unit, audio base or satellite device; see col. 7 lines 58-65 format generation unit configured to generate a sound field in a specified format based on multi-channel audio data; see col. 8 lines 4-21 generate audio signals for playback at a plurality of loudspeakers; the audio signals may include channels C1 through CL; see figure 5). Therefore, in light of the teaching of Kim, it is notoriously well known before the effective filing date of the claimed invention to obtain a pulse modulation code when decoding encoded signals and playing audio according to that in order to reduce signal to noise ratio, to deliver high-resolution sound, and to obtain and output one or more speaker feeds. Regarding claim 2, Kim discloses wherein before decoding each channel of target audio track based on the preset decoder corresponding to each channel of target audio track, the method further comprises: obtaining parameter information corresponding to each channel of target audio track; establishing the preset decoder corresponding to each channel of target audio track based on the parameter information corresponding to each channel of target audio track (see figures 2-3; col. 5 line 53-col. 6 line 5; col. 6 lines 28-35 and col. 7 lines 37-65). The motivation to combine the references is discussed in claim 1 above. Regarding claim 3, Kim discloses the preset decoder comprises a hard decoder and a soft decoder; the parameter information comprises: an audio sampling rate, a quantity of soundtracks and a bit rate; the establishing the preset decoder corresponding to each channel of target audio track based on the parameter information corresponding to each channel of target audio track, comprises: performing a product operation on the audio sampling rate, the quantity of soundtracks, and the bit rate of each channel of target audio track to obtain a product operation result of each channel of target audio track; establishing the hard decoder for a target audio track corresponding to a maximum product operation result, and establishing the soft decoder for other target audio tracks (see figure 5; col. 1 lines 48-64; col. 4 lines 14-35; col. 6 line 36-col. 7 line 13; col. 7 lines 37-57; col. 11 lines 25-43 and col. 11 line 50-col. 12 line 40). The motivation to combine the references is discussed in claim 1 above. Regarding claim 9, Oughriss discloses based on that a switching instruction input from a user is received, switching the target audio track currently being played (see paragraphs 0023 and 0025). Regarding claim 11, the limitation of claim 11 can be found in claim 1 above. Therefore claim 11 is analyzed and rejected for the same reasons as discussed in claim 1 above. See also Kim’s claim 30. Claims 12-13 and 19 are rejected for the same reasons as discussed in claims 2-3 and 9 respectively above. Claims 4-8, 10, 14-18 and 20 are rejected under 35 U.S.C. 103 as being unpatentable over Oughriss et al. (US PG PUB 2014/0118616 hereinafter referred as Oughriss) in view of Kim et al. (US Pat. No. 10, 972, 853) and further in view of Erickson (US Pat. No. 10, 129, 587). Regarding claim 4, although the applied prior art of Oughriss discloses based on that each channel of target audio track is played through the playback earphone corresponding to each channel of target audio track, synchronously playing videos contained in the to-be-played multimedia file based on a target synchronization clock (see paragraphs 0020-0022); the combination of Oughriss and Kim fails to specifically discloses playing the captions included in the multimedia file. In the same field of endeavor Erickson discloses playing the captions included in the multimedia file (see col. 6 lines 9-20; the second media streams 327 may correspond to alternative subtitle streams synchronized to the first media stream 324, where the first media stream 324 may correspond to audio and/or video; see also col. 2 lines 52-67 and figures 1-3). Therefore, in light of the teaching in Erickson, it would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the above combination by adding closed captioning data in the multimedia stream and output the media file simultaneously in order to facilitate fast switching of synchronized media using time-stamp management, to capture user data and to perform time-stamp management function. Regarding claim 5, Erickson the synchronously playing the videos and captions contained in the to-be-played multimedia file based on the target synchronization clock comprises: decoding elementary streams corresponding to the videos and captions contained in the to-be-played multimedia file to obtain initial data respectively corresponding to the videos and captions; synchronously playing the videos and captions contained in the to-be-played multimedia file based on the target synchronization clock and the initial data respectively corresponding to the videos and the captions (see figures 1-3; col. 2 lines 52-67; col. 4 lines 13-23; col. 5 lines 7-33 and col. 6 lines 9-20). The motivation to combine the references is discussed in claim 4 above. Regarding claim 6, Erickson discloses before parsing elementary streams corresponding to the videos and captions contained in the to-be-played multimedia file to obtain initial data respectively corresponding to the videos and captions, the method further comprises: determining an audio clock corresponding to each channel of target audio track; determining the target synchronization clock from a plurality of audio clocks, wherein the target synchronization clock is used to synchronously play the videos, captions and at least two channels of target audio tracks contained in the to-be-played multimedia file (see col. 5 lines 7-33; col. 6 lines 9-20; col. 6 line 51-col. 7 line 7; and col. 7 line 50-col. 8 line 6). The motivation to combine the references is discussed in claim 4 above. Regarding claim 7, Oughriss discloses the determining the target synchronization clock from the plurality of audio clocks, comprises: determining a first target audio track corresponding to a maximum product operation result based on a product operation result of each channel of target audio track; determining an audio clock of the first target audio track as the target synchronization clock (see paragraphs 0016-0017 and 0024). Regarding claim 8, Kim discloses the said determining the target synchronization clock from the plurality of audio clocks, comprises: based on that a corresponding preset decoder is established for each channel of target audio track, determining a second target audio track corresponding to a preset decoder that is a last audio track to finish the establishment; determining an audio clock of the second target audio track as the target synchronization clock (see col. 11 lines 11-24; col. 12 line 40-col. 13 line 5). See also Oughriss’s paragraph 0022 and Erickson’s col. 3 lines 49-67. The motivation to combine the references is discussed in claims 1 and 4 above. Regarding claim 10, Erickson discloses based on that each channel of target audio track is played through the playback earphone corresponding to each channel of target audio track, playing the captions contained in the to-be-played multimedia file, comprises: receiving a selection operation from a user; in response to the selection operation, determining one or more channels of to-be-output captions from the captions contained in the to-be-played multimedia file; obtaining caption data of each channel of to-be-output caption; obtaining a global clock of a playback pipeline of the to-be-played multimedia file and a synchronous rendering logic corresponding to each channel of to-be-output caption; based on the global clock and the synchronous rendering logic corresponding to each channel of to-be-output caption, synchronously rendering the caption data of each channel of to-be-output caption, so as to synchronously display each channel of to-be-output caption while playing the to-be-played multimedia file (see figure 4; col. 2 lines 52-67; col. 4 lines 44-53; col. 5 lines 33-48; col. 6 line 9-col. 7 line 29 and also col. 7 line 20-col. 8 line 6). The motivation to combine the references is discussed in claim 4 above. Claims 14-18 and 20 are rejected for the same reasons as discussed in claims 4-8 and 10 respectively. Conclusion Any inquiry concerning this communication or earlier communications from the examiner should be directed to HELEN SHIBRU whose telephone number is (571)272-7329. The examiner can normally be reached M-TR 8:00AM-5:00PM. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, THAI TRAN can be reached at 571 272 7382. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. /HELEN SHIBRU/Primary Examiner, Art Unit 2484 February 5, 2026
Read full office action

Prosecution Timeline

Mar 07, 2025
Application Filed
Feb 06, 2026
Non-Final Rejection — §103 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12603970
METHOD, APPARATUS, ELECTRONIC DEVICE, AND COMPUTER-READABLE MEDIUM FOR DISPLAYING LYRIC EFFECTS
2y 5m to grant Granted Apr 14, 2026
Patent 12603112
VIDEO GENERATION METHOD, APPARATUS, DEVICE, AND STORAGE MEDIUM
2y 5m to grant Granted Apr 14, 2026
Patent 12586375
Content Validation Using Scene Modification
2y 5m to grant Granted Mar 24, 2026
Patent 12563277
USER-CUSTOMIZED ALREADY VIEWED VIDEO SUMMARY
2y 5m to grant Granted Feb 24, 2026
Patent 12562194
METHOD, APPARATUS, DEVICE AND MEDIUM FOR GENERATING A VIDEO
2y 5m to grant Granted Feb 24, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
59%
Grant Probability
62%
With Interview (+3.7%)
3y 8m
Median Time to Grant
Low
PTA Risk
Based on 756 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month