Prosecution Insights
Last updated: April 19, 2026
Application No. 17/837,684

VOICE COMMUNICATION BETWEEN A SPEAKER AND A RECIPIENT OVER A COMMUNICATION NETWORK

Final Rejection §103
Filed
Jun 10, 2022
Examiner
ARMSTRONG, ANGELA A
Art Unit
2659
Tech Center
2600 — Communications
Assignee
Elektrobit Automotive GmbH
OA Round
4 (Final)
75%
Grant Probability
Favorable
5-6
OA Rounds
3y 11m
To Grant
84%
With Interview

Examiner Intelligence

Grants 75% — above average
75%
Career Allow Rate
478 granted / 641 resolved
+12.6% vs TC avg
Moderate +10% lift
Without
With
+9.5%
Interview Lift
resolved cases with interview
Typical timeline
3y 11m
Avg Prosecution
25 currently pending
Career history
666
Total Applications
across all art units

Statute-Specific Performance

§101
21.9%
-18.1% vs TC avg
§103
43.7%
+3.7% vs TC avg
§102
14.8%
-25.2% vs TC avg
§112
7.7%
-32.3% vs TC avg
Black line = Tech Center average estimate • Based on career data from 641 resolved cases

Office Action

§103
DETAILED ACTION Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . This Office Action is in response to the amendment filed November 21, 2025. Claims 1, 3, 8, 11, 13, 18, 21, 23, and 28 have been amended. Claims 2, 12, and 22 have been cancelled. Claims 1, 3-5, 8-11, 13-15, 18-21, 23-25, and 28-33 remain pending. Claim Rejections - 35 USC § 103 The text of those sections of Title 35, U.S. Code not included in this action can be found in a prior Office action. Claims 1, 3-5, 8-11, 13-15, 18-21, 23-25, and 28-33 are rejected under 35 U.S.C. 103 as being unpatentable over Cutler et al (US Patent Application No. 2018/0218727), hereinafter Cutler, in view of Thenthiruperai et al (US Patent No. 7,069,014), hereinafter Thenthiruperai and further in view of Chae et al (US Patent Application Publication No. 2020/0058290), hereinafter, Chae. Cutler discloses a system for artificially generated speech for a communication session. Regarding claims 1 and 11, Cutler discloses [Figure 2; para 0041; 0043] a method and a non-transitory computer-readable medium having stored thereon computer-executable instructions, which, when executed by at least one processor, cause the at least one processor to provide voice communication between a speaker and a recipient over a communication network by performing operations comprising: receiving an input speech utterance from the speaker [speech captured by microphone at transmission side -- para 0024; 0037]; converting the input speech utterance to text [ASR used to convert speech to text – para 0024; 0038-0040]; transmitting at least the text over the communication network [transmission of text with audio – para 0024; 0038-0040; 0044]; converting the transmitted text into an output speech utterance that simulates a voice of the speaker [conversion of TTS based on user-specific model specific to the speech captured at transmission side -- para 0024; 0042; 0045-0048]; and providing the output speech utterance to the recipient [conversion of TTS based on user-specific model specific to the speech captured at transmission side -- para 0024; 0042; 0045-0048] and wherein the transmitted text is converted into the output speech utterance by one or more trained artificial intelligence models [Cutler’s text-to-speech converter 218; user specific model specific to the user from whom speech is captured –para 0024; 0004-0006; 0009-0010; 0035; dynamically updating/training receiver side and transmission side models - 0010; 0038; 0056; 0058-0059; Cutler’s voice models used for multiparty scenarios – para 0063]. Cutler fails to teach evaluating a bandwidth of a connection to the communication network at the side of the speaker and further processing data if the bandwidth is sufficient for transmitting voice. In a similar field of endeavor, Thenthiruperai teaches bandwidth-determined selection for wireless devices, providing for determining the currently available bandwidth with a bandwidth manager for transmissions over a wireless network; provides for rendering only appropriate data or content in a specific format that is appropriate for the current bandwidth; and allows for changing the interaction medium (audio+video; audio; or audio+text) [col. 4, lines 6-16; col. 9, lines 14-39; col. 10,line 33 to col. 11, line 33; col. 12, line 65 to col. 13, line 13; col. 13, lines15-26; col. 14, line 61 to col. 15, line 10]. One having ordinary skill in the art at the time of the invention would have recognized the advantages of implementing the bandwidth evaluation techniques and appropriate content transmission depending on the current bandwidth suggested by Thenthiruperai, in the system of Cutler/Chae, so as to ensure the recipients receive all the necessary transmitted data with minimal loss, errors or disruption, thereby improving system performance and the users’ experience. Cutler fails to teach a first trained artificial intelligence model transforms the transmitted text into an intermediate speech utterance and a second trained artificial intelligence model transforms the intermediate speech utterance into the output speech utterance. In a similar field of endeavor, Chae teaches artificial intelligence apparatus for correcting synthesized speech by receiving a first text, generating a first synthesized speech signal, acquiring corrected speech features using speech correction models and generating a second synthesized speech using the corrected speech features [Fig. 9; para 0286-0295]. Chae teaches the system provides for improvement of the quality of synthesized speech [para 0012]. One having ordinary skill in the art at the time of the invention would have recognized the advantages of implementing correction models speech synthesis, suggested by Chae, in the system of Cutler, for the purpose of improving the quality of the generated synthetic speech, as suggested by Chae. Regarding claim 3, the combination of Cutler, Chae and Thenthiruperai teaches the input speech utterance is transmitted as voice and as text if the connection has sufficient bandwidth [Cutler at para 0049; Thenthiruperai’s rendering only appropriate data or content in a specific format that is appropriate for the current bandwidth and changing the interaction medium (audio+video; audio; or audio+text) col. 4, lines 6-16; col. 9, lines 14-39; col. 10,line 33 to col. 11, line 33; col. 12, line 65 to col. 13, line 13; col. 13, lines15-26; col. 14, line 61 to col. 15, line 10]. Regarding claim 4, the combination of Cutler, Chae and Thenthiruperai teaches the transmitted text is converted into the output speech utterance by a text-to-speech algorithm [Cutler’s text-to-speech converter 218; Chae at Fig 9]. Regarding claim 5, the combination of Cutler, Chae and Thenthiruperai teaches the text-to-speech algorithm uses a phoneme library suitable for simulating different speakers [Cutler’s voice models used for multiparty scenarios – para 0063]. Regarding claims 8-10, the combination of Cutler, Chae and Thenthiruperai teaches a bank of trained models [Cutler’s text-to-speech converter 218; user specific model specific to the user from whom speech is captured –para 0024; 0004-0006; 0009-0010; 0035; dynamically updating/training receiver side and transmission side models -- 0010; 0038; 0056; 0058-0059; Cutler’s voice models used for multiparty scenarios – para 0063; Chae’s speech correction models ]. Regarding claim 31, the combination of Cutler, Chae and Thenthiruperai teaches a third artificial intelligence model synthesizes an emotion of the speaker based on tags that are received along with the transmitted text, wherein the tags specify a characteristic of the input speech utterance, including at least one characteristic of a group of characteristics, the group of characteristics including: an intonation, a speed of speech, one or more detected emotions, and a respective duration of each individual word of the input speech utterance [Chae para 0007; 0218; 0240; 0251; 0277-0278; 0301-0302; 0309-0310 – where providing a third model to provide Chae’s emotion synthesis is an obvious requiring only routine skill in the art]. Regarding claim 21, Cutler discloses [Figure 2; para 0041; 0043] a non-transitory computer-readable medium having stored thereon computer-executable instructions, which, when executed by at least one processor, cause the at least one processor to provide voice communication between a speaker and a recipient over a communication network by performing operations comprising: receiving an input speech utterance from the speaker [speech captured by microphone at transmission side -- para 0024; 0037]; converting the input speech utterance to text [ASR used to convert speech to text – para 0024; 0038-0040]; transmitting at least the text over the communication network [transmission of text with audio – para 0024; 0038-0040; 0044]; converting the transmitted text into an output speech utterance that simulates a voice of the speaker [conversion of TTS based on user-specific model specific to the speech captured at transmission side -- para 0024; 0042; 0045-0048]; and providing the output speech utterance to the recipient [conversion of TTS based on user-specific model specific to the speech captured at transmission side -- para 0024; 0042; 0045-0048] and wherein the transmitted text is converted into the output speech utterance by one or more trained artificial intelligence models [Cutler’s text-to-speech converter 218; user specific model specific to the user from whom speech is captured –para 0024; 0004-0006; 0009-0010; 0035; dynamically updating/training receiver side and transmission side models - 0010; 0038; 0056; 0058-0059; Cutler’s voice models used for multiparty scenarios – para 0063]. Cutler fails to teach evaluating a bandwidth of a connection to the communication network at the side of the speaker and further processing data if the bandwidth is sufficient for transmitting voice. In a similar field of endeavor, Thenthiruperai teaches bandwidth-determined selection for wireless devices, providing for determining the currently available bandwidth with a bandwidth manager for transmissions over a wireless network; provides for rendering only appropriate data or content in a specific format that is appropriate for the current bandwidth; and allows for changing the interaction medium (audio+video; audio; or audio+text) [col. 4, lines 6-16; col. 9, lines 14-39; col. 10,line 33 to col. 11, line 33; col. 12, line 65 to col. 13, line 13; col. 13, lines15-26; col. 14, line 61 to col. 15, line 10]. One having ordinary skill in the art at the time of the invention would have recognized the advantages of implementing the bandwidth evaluation techniques and appropriate content transmission depending on the current bandwidth suggested by Thenthiruperai, in the system of Cutler/Chae, so as to ensure the recipients receive all the necessary transmitted data with minimal loss, errors or disruption, thereby improving system performance and the users’ experience. Cutler fails to teach the system is implemented in a vehicle. In a similar field of endeavor, Chae teaches artificial intelligence apparatus for correcting synthesized speech by receiving a first text, generating a first synthesized speech signal, acquiring corrected speech features using speech correction models and generating a second synthesized speech using the corrected speech features [Fig. 9; para 0286-0295] and suggests the system can be implemented in a vehicle [para 0056]. Chae teaches the system provides for improvement of the quality of synthesized speech [para 0012]. One having ordinary skill in the art at the time of the invention would have recognized the advantages of implementing the system of Cutler in a vehicle, so as to provide enhanced communication sessions for drivers while in a vehicle, with minimal distractions or disruptions for the vehicle driver, thereby improving safety during communications while driving. Claims 13-15, 18-20, 23-25, 28-30 and 32-33 are rejected under similar rationale as claims 3-5, 8-10 and 31. Response to Arguments Applicant's arguments filed November 21, 2025 have been fully considered but they are not persuasive. Applicant argues the references cited by the examiner, even if properly combinable, do not disclose, teach, or suggest the features for "evaluating a bandwidth of a connection to the communication network at the side of the speaker" and "converting the input speech utterance to text if the bandwidth is sufficient for transporting voice signals.” The Examiner respectfully disagrees. As indicated in the rejection above, Thenthiruperai teaches bandwidth-determined selection for wireless devices, providing for determining the currently available bandwidth with a bandwidth manager for transmissions over a wireless network; provides for rendering only appropriate data or content in a specific format that is appropriate for the current bandwidth; and allows for changing the interaction medium (audio+video; audio; or audio+text) [col. 4, lines 6-16; col. 9, lines 14-39; col. 10,line 33 to col. 11, line 33; col. 12, line 65 to col. 13, line 13; col. 13, lines15-26; col. 14, line 61 to col. 15, line 10]. Given the teachings of Cutler and Thenthiruperai, one having ordinary skill in the art at the time of the invention would have recognized the advantages of implementing the bandwidth evaluation techniques and appropriate content transmission depending on the current bandwidth suggested by Thenthiruperai, in the system of Cutler/Chae, so as to ensure the recipients receive all the necessary transmitted data with minimal loss, errors or disruption, thereby improving system performance and the users’ experience. The teachings of Cutler in combination with the teachings of Thenthiruperai provide adequate support for the limitation as claimed. Conclusion Applicant's amendment necessitated the new ground(s) of rejection presented in this Office action. Accordingly, THIS ACTION IS MADE FINAL. See MPEP § 706.07(a). Applicant is reminded of the extension of time policy as set forth in 37 CFR 1.136(a). A shortened statutory period for reply to this final action is set to expire THREE MONTHS from the mailing date of this action. In the event a first reply is filed within TWO MONTHS of the mailing date of this final action and the advisory action is not mailed until after the end of the THREE-MONTH shortened statutory period, then the shortened statutory period will expire on the date the advisory action is mailed, and any nonprovisional extension fee (37 CFR 1.17(a)) pursuant to 37 CFR 1.136(a) will be calculated from the mailing date of the advisory action. In no event, however, will the statutory period for reply expire later than SIX MONTHS from the mailing date of this final action. Any inquiry concerning this communication or earlier communications from the examiner should be directed to ANGELA A ARMSTRONG whose telephone number is (571)272-7598. The examiner can normally be reached M,T,TH,F 11:30-8:00. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Pierre Desir can be reached at 571-272-7799. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. ANGELA A. ARMSTRONG Primary Examiner Art Unit 2659 /ANGELA A ARMSTRONG/Primary Examiner, Art Unit 2659
Read full office action

Prosecution Timeline

Jun 10, 2022
Application Filed
Aug 24, 2024
Non-Final Rejection — §103
Nov 22, 2024
Response Filed
Jan 25, 2025
Final Rejection — §103
Apr 30, 2025
Request for Continued Examination
May 06, 2025
Response after Non-Final Action
May 17, 2025
Non-Final Rejection — §103
Nov 21, 2025
Response Filed
Mar 06, 2026
Final Rejection — §103 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12602547
DOMAIN ADAPTING GRAPH NETWORKS FOR VISUALLY RICH DOCUMENTS
2y 5m to grant Granted Apr 14, 2026
Patent 12596879
METHOD AND SYSTEM FOR IDENTIFYING CITATIONS WITHIN REGULATORY CONTENT
2y 5m to grant Granted Apr 07, 2026
Patent 12585892
AUTO-TRANSLATION OF CUSTOMIZED ASSISTANT
2y 5m to grant Granted Mar 24, 2026
Patent 12555491
Inclusive Intelligence for Digital Workplace
2y 5m to grant Granted Feb 17, 2026
Patent 12547843
SYSTEMS AND METHODS FOR GENERALIZED ENTITY MATCHING
2y 5m to grant Granted Feb 10, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

5-6
Expected OA Rounds
75%
Grant Probability
84%
With Interview (+9.5%)
3y 11m
Median Time to Grant
High
PTA Risk
Based on 641 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month