Prosecution Insights
Last updated: April 19, 2026
Application No. 17/806,900

METHOD AND APPARATUS FOR DETECTING ATTACK IN CAN BUS

Final Rejection §103
Filed
Jun 14, 2022
Examiner
BEAN, GRIFFIN TANNER
Art Unit
2121
Tech Center
2100 — Computer Architecture & Software
Assignee
Autocrypt Co., Ltd.
OA Round
2 (Final)
21%
Grant Probability
At Risk
3-4
OA Rounds
4y 4m
To Grant
50%
With Interview

Examiner Intelligence

Grants only 21% of cases
21%
Career Allow Rate
4 granted / 19 resolved
-33.9% vs TC avg
Strong +28% interview lift
Without
With
+28.4%
Interview Lift
resolved cases with interview
Typical timeline
4y 4m
Avg Prosecution
45 currently pending
Career history
64
Total Applications
across all art units

Statute-Specific Performance

§101
37.7%
-2.3% vs TC avg
§103
40.4%
+0.4% vs TC avg
§102
11.2%
-28.8% vs TC avg
§112
9.7%
-30.3% vs TC avg
Black line = Tech Center average estimate • Based on career data from 19 resolved cases

Office Action

§103
DETAILED ACTION This Action is responsive to Claims filed 10/03/2025. Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . Status of the Claims Claims 1-8 and 15 have been cancelled. Claims 9-14 have been amended. Claims 9-14 and 16 are pending. Response to Arguments Applicant’s arguments, see Pages 5-10, filed 10/03/2025, with respect to Claims 1-16 have been fully considered and are persuasive. The 35 U.S.C. 101 Rejection of Claims 1-16 has been withdrawn. Applicant’s arguments, see Pages 10-13, regarding the 35 U.S.C. 103 Rejection(s) of Claims 1-16 have been considered but are moot because the new ground of rejection does not rely on any reference applied in the prior rejection of record for any teaching or matter specifically challenged in the argument. On Page 13, the Applicant alleges distinctions between the cited reference Song and the instant Application based on what the loss function of Song trains. It is unclear to the Examiner the distinction the Applicant attempts to draw between training the LSTM generator model, which is a part of Song’s intrusion/anomaly detection system or model, versus training the intrusion-detection model. The Examiner notes Song was not relied on to teach a bi-directional GPT2 network, but that a combination of at least Song, Chen, and Debjit, used as an intrusion detection system, would reasonably be trained similarly. See the updated 35 U.S.C. 103 Rejection below. Claim Rejections - 35 USC § 103 The text of those sections of Title 35, U.S. Code not included in this action can be found in a prior Office action. The factual inquiries for establishing a background for determining obviousness under 35 U.S.C. 103 are summarized as follows: 1. Determining the scope and contents of the prior art. 2. Ascertaining the differences between the prior art and the claims at issue. 3. Resolving the level of ordinary skill in the pertinent art. 4. Considering objective evidence present in the application indicating obviousness or nonobviousness. This application currently names joint inventors. In considering patentability of the claims the examiner presumes that the subject matter of the various claims was commonly owned as of the effective filing date of the claimed invention(s) absent any evidence to the contrary. Applicant is advised of the obligation under 37 CFR 1.56 to point out the inventor and effective filing dates of each claim that was not commonly owned as of the effective filing date of the later invention in order for the examiner to consider the applicability of 35 U.S.C. 102(b)(2)(C) for any potential 35 U.S.C. 102(a)(2) prior art against the later invention. Claim(s) 9-14 and 16 is/are rejected under 35 U.S.C. 103 as being unpatentable over Song et al. (Self-Supervised Anomaly Detection for In-Vehicle Network Using Noised Pseudo Normal Data, 2021), hereinafter Song; Chen et al. (DATA CURATION AND QUALITY ASSURANCE FOR MACHINE LEARNING-BASED CYBER INTRUSION DETECTION, 2021), hereinafter Chen; and Debjit et al. (COINS: Dynamically Generating COntextualized Inference Rules for Narrative Story Completion, 2021), hereinafter Debjit. In regards to claim 9: The present invention claims: “An intrusion detection apparatus for a controller area network (CAN), the intrusion detection apparatus comprising: a processor; a memory storing one or more instructions executable by the processor, and a transceiver connected to a bus of the CAN to perform communications, wherein when executed by the processor, the one or more instructions cause the intrusion detection apparatus to:” Song teaches “Thus, the objectives of this paper are to provide an anomaly detection method in such a limited data environment [CAN bus and CAN traffic] where attack data is deficient while normal data is relatively easy to obtain.” (Page 1098, right column). “collect CAN identifiers (IDs) from the CAN in an order of occurrence;” Song teaches “In this study, we use only CAN ID sequence for modeling sequential patterns of CAN traffic except payload of messages.” (Page 1103, left column). “generate a sequence of a predetermined number L of CAN IDs from the collected CAN IDs, L being an integer greater than or equal to 1;” Song teaches “The input to the generator model is a CAN ID or a sequence of CAN IDs.” (Page 1103, right column). “input the sequence into a… network and calculating a value of a loss function corresponding to the sequence…” Song teaches “The input to the generator model is a CAN ID or a sequence of CAN IDs. The LSTM-based model is trained to guess which CAN ID is the most probable as the following CAN ID at each time step based on a given CAN ID or a sequence of CAN IDs.” (Page 1103, right column) and “The categorical cross-entropy loss function is used because the last dense layer outputs a probability over the CAN IDs. The categorical cross-entropy can be implemented by adding Softmax activation before calculating the cross-entropy.” (Page 1104, left column). “wherein when a number of allowed CAN IDs is K, the allowed CAN IDs in the sequence are sorted in an order of magnitude and converted into values 0 to K - 1,” Song teaches “For the generator model, each CAN ID represented as a hexadecimal string is mapped to an integer representation as an indexes from 0 to the number of CAN IDs in CAN traffic.” (Page 1103, left column). “wherein the intrusion detection apparatus is included in a head unit or a gateway of a vehicle, and is implemented as a diagnostic device existing inside or outside the vehicle,” Song teaches “This paper proposes a novel method of training anomaly detection model based on self-supervised learning for in-vehicle network security.” (Introduction, right column). While Song teaches an LSTM as the generator model, Song fails to explicitly teach: “bi-directional generative pretrained transformer 2 (GPT2)…” and “CAN IDs that are not allowed in the sequence are converted to K, the sequence is input to the bidirectional GPT2 network, and K is an integer equal to or greater than 1.” However, Chen, in a similar field of endeavor of anomaly detection, teaches “Since the goal of this case study is to develop a host-based intrusion detection system (HIDS), we conduct machine learning experiments on the UNM, MIT, and ADFA-LD datasets. A detailed introduction of these datasets can be found in our GitHub repository. Regardless of the slight difference in ADFA and UNM data formats, we use similar data cleaning and augmentation techniques to create a dataset for each class. Processing data for pre-trained language models vectors, such as BERT [62] and GPT-2 [63], is similar to normal machine learning algorithms.” (GPT2, Page 9, Section 4.1.1) and “row with normal sequence is labeled 0, whereas the one with intrusion sequence is labeled 1. We use normal data and intrusion data from each dataset to create a sample pool.” (Page 9, Section 4.1.1, given the Applicant’s Specification recites the “converted to K” step without specifics, the Examiner interprets this as a highly general means of delineating an abnormal sequence, which Chen reads on). Chen highlights the need to improve the data quality when building a machine learning based intrusion detection system (Abstract). It would have been obvious to one of ordinary skill in the art at the time of the Applicant’s filing to incorporate the improved data-work teachings of Chen in an intrusion detection system like Song’s in order to improve the overall model quality. The combination of Song and Chen fails to explicitly teach “the GPT2 network including a forward GPT module, a backward GPT module, and a fully-connected layer,” However, Debjit teaches “We adopt a pretrained GPT-2 (base)…transformer model with multiple Transformer blocks of multi-head self-attention and fully connected layers.” (Page 4, Inference Step) and “While we designed our sentence generation model in such a way that it can utilize inference rules from both forward and backward directions for each sentence…” (Page 5, left column, inference rules in both directions would require the network to handle both directions (modules)). Debjit highlights the strengths of GPT2 networks on sequential inference tasks (Introduction). It would have been obvious to one of ordinary skill in the art at the time of the Applicant’s filing to utilize these known GPT2 structures or functionalities if they were to use GPT2 in a combination of Song and Chen to realize the sequential inferencing power of GPT2. “wherein the loss function is defined as PNG media_image1.png 69 504 media_image1.png Greyscale , and wherein the forward GPT module, the backward GPT module, and the fully-connected layer are trained to minimize the value of the loss function, xn) is an l-th variable of an n-th normal CAN ID sequence used for training, and yn) is an actual generated CAN ID for x(') which is a ground truth value.” Per the Applicant’s Specification (Equation 5), the output of this loss is based on a probability matrix, which Song reads on in “The categorical cross-entropy loss function is used because the last dense layer outputs a probability over the CAN IDs. The categorical cross-entropy can be implemented by adding Softmax activation before calculating the cross-entropy. The softmax activation normalizes a C-dimensional vector s to a C-dimensional vector σ(s) in the range (0, 1) of which the sum is 1, which is calculated as:” (Page 1104, left column). In regards to claim 10: The present invention claims: “wherein the bi-directional GPT2 network further includes a concatenation unit and a softmax function layer.” Similar to the structure recited above, Debjit teaches both concatenation and the use of a softmax layer in the Inference Step Section on Pages 4 and 5. In regards to claim 11: The present invention claims: “wherein to calculate the value of the loss function, the one or more instructions further cause the intrusion detection apparatus to input the sequence to the forward GPT module in an original order, and to input the sequence to the backward GPT module in a reverse order.” Debjit teaches “While we designed our sentence generation model in such a way that it can utilize inference rules from both forward and backward directions for each sentence…” (Page 5, left column). In regards to claim 12: The present invention claims: “wherein to calculate the value of the loss function, the one or more instructions further cause the intrusion detection apparatus to input to the forward GPT module embedding vectors corresponding to CAN IDs from 0-th CAN ID to (L-2)-th CAN ID belonging to the sequence, and the forward GPT module outputs E-dimensional vectors having a same dimensionality as the embedding vectors, which correspond to 1st to (L-1)-th CAN IDs.” Song teaches “The embedding layer is the input layer that maps a given index, an integer representation of a CAN ID, to a vector of fixed-length, which is set to 256 in this study. The converted CANID vector is fed to the next LSTM layer with 256 units. The LSTM layer extracts the context of a given sequence. Finally, the dense layer then outputs logits predicting the log-likelihood of the next CAN ID. Therefore, the output size of the dense layer is the same as the number of CAN IDs.” (Pages 1103-1104). In regards to claim 13: The present invention claims: “wherein to calculate the value of the loss function, the one or more instructions further cause the intrusion detection apparatus to input to the forward GPT module embedding vectors corresponding to CAN IDs from (L-1)-th CAN ID to 1st CAN ID belonging to the sequence, and the backward GPT module outputs E-dimensional vectors having a same dimensionality as the embedding vectors, which correspond to (L-2)-th to 0-th CAN IDs.” Song teaches “The embedding layer is the input layer that maps a given index, an integer representation of a CAN ID, to a vector of fixed-length, which is set to 256 in this study. The converted CANID vector is fed to the next LSTM layer with 256 units. The LSTM layer extracts the context of a given sequence. Finally, the dense layer then outputs logits predicting the log-likelihood of the next CAN ID. Therefore, the output size of the dense layer is the same as the number of CAN IDs.” (Pages 1103-1104) and Debjit teaches “While we designed our sentence generation model in such a way that it can utilize inference rules from both forward and backward directions for each sentence…” (Page 5, left column). A combination of Song, Chen, and Debjit would reasonably read on vector embeddings and outputs in the forwards and backwards directions. In regards to claim 14: The present invention claims: “wherein to calculate the value of the loss function, the one or more instructions further cause the intrusion detection apparatus to concatenate the output of the forward GPT module and the output of the backward GPT module to generate a 2E x L matrix, the 2E x L matrix is converted to a (K + 1) x L matrix by the fully-connected layer, and the (K + 1) x L matrix is transformed into a probability matrix by a softmax layer.” Debjit teaches “We concatenate the generated inference rules (Ii = Ei ⊕ Ci)7 and store the last hidden representation in MemIR 2 IRN_L_H, where N is the number of sentences, L the maximum inference sequence length and H the hidden state dimensions.” (Page 5, right column) and “The softmax layer defines the model to output the most probable target sequence: the most likely inference rules (Ei and Ci) for each relation type (cf. Algorithm Line 4-5).” (Page 5, left column). Song also teaches “The categorical cross-entropy loss function is used because the last dense layer outputs a probability over the CAN IDs. The categorical cross-entropy can be implemented by adding Softmax activation before calculating the cross-entropy.” (Page 1104, left column). In regards to claim 16: The present invention claims: “compare the value of the loss function to a threshold; and when the value of the loss function is equal to or greater than the threshold, determining a period corresponding to the sequence as a period in which an intrusion exists, wherein the loss function is defined as PNG media_image2.png 65 260 media_image2.png Greyscale xf") is an l-th variable of an m-th CAN ID sequence corresponding to a detection target sequence, and y,"n is an actual generated CAN ID for xi", which is a ground truth value.” Per the Applicant’s Specification (Equation 7), the output of this loss can be interpreted as an anomaly score, which Song reasonably reads on in “The categorical cross-entropy loss function is used because the last dense layer outputs a probability over the CAN IDs.” (Page 1104, left column) and Chen reasonably reads on in “We use small GPT-2 to train on processed data from both classes of each dataset using 16 epochs and 16 batch size. Unlike BERT, GPT-2 does not require an additional output layer; instead, it outputs the sequence’s likelihood in each class.” (Page 11, GPT-2, mapping the output of a probability to a general anomaly score). Conclusion Applicant's amendment necessitated the new ground(s) of rejection presented in this Office action. Accordingly, THIS ACTION IS MADE FINAL. See MPEP § 706.07(a). Applicant is reminded of the extension of time policy as set forth in 37 CFR 1.136(a). A shortened statutory period for reply to this final action is set to expire THREE MONTHS from the mailing date of this action. In the event a first reply is filed within TWO MONTHS of the mailing date of this final action and the advisory action is not mailed until after the end of the THREE-MONTH shortened statutory period, then the shortened statutory period will expire on the date the advisory action is mailed, and any nonprovisional extension fee (37 CFR 1.17(a)) pursuant to 37 CFR 1.136(a) will be calculated from the mailing date of the advisory action. In no event, however, will the statutory period for reply expire later than SIX MONTHS from the mailing date of this final action. Any inquiry concerning this communication or earlier communications from the examiner should be directed to GRIFFIN T BEAN whose telephone number is (703)756-1473. The examiner can normally be reached M - F 7:30 - 4:30. Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Li Zhen can be reached at (571) 272-3768. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. /GRIFFIN TANNER BEAN/Examiner, Art Unit 2121 /Li B. Zhen/Supervisory Patent Examiner, Art Unit 2121
Read full office action

Prosecution Timeline

Jun 14, 2022
Application Filed
Jul 01, 2025
Non-Final Rejection — §103
Oct 03, 2025
Response Filed
Jan 14, 2026
Final Rejection — §103 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12424302
ACCELERATED MOLECULAR DYNAMICS SIMULATION METHOD ON A QUANTUM-CLASSICAL HYBRID COMPUTING SYSTEM
2y 5m to grant Granted Sep 23, 2025
Patent 12314861
SYSTEMS AND METHODS FOR SEMI-SUPERVISED LEARNING WITH CONTRASTIVE GRAPH REGULARIZATION
2y 5m to grant Granted May 27, 2025
Patent 12261947
LEARNING SYSTEM, LEARNING METHOD, AND COMPUTER PROGRAM PRODUCT
2y 5m to grant Granted Mar 25, 2025
Study what changed to get past this examiner. Based on 3 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

3-4
Expected OA Rounds
21%
Grant Probability
50%
With Interview (+28.4%)
4y 4m
Median Time to Grant
Moderate
PTA Risk
Based on 19 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month