DETAILED ACTION
This Final Office Action is in response to the arguments and amendments filed December 03, 2025.
Claims 1-20 are Originals.
Notice of Pre-AIA or AIA Status
The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA .
Claim Rejections - 35 USC § 103
In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis (i.e., changing from AIA to pre-AIA ) for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status.
The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action:
A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made.
The factual inquiries for establishing a background for determining obviousness under 35 U.S.C. 103 are summarized as follows:
1. Determining the scope and contents of the prior art.
2. Ascertaining the differences between the prior art and the claims at issue.
3. Resolving the level of ordinary skill in the pertinent art.
4. Considering objective evidence present in the application indicating obviousness or nonobviousness.
Claim(s) 1-3, 7-11, 13, 15-19 are rejected under 35 U.S.C. 103 as being unpatentable over Jose et al [US2022,017,2147A1] hereafter Jose, in view of Yong-bo et al, [CN115481283A] hereafter Yong-bo; in view of Dangi et al [US2021,015,8302A1] hereafter Dangi.
As per claim 1, 9 and 17 (Similar scope and language)
Jose teaches;
A method for evaluating candidates through Artificial Intelligence (AI) models, the method comprising: receiving, by a computing device, input data comprising video data and audio data corresponding to an interview of a candidate, wherein the video data comprises a plurality of frames;
{[0027] In an embodiment, the computing system 112 is configured to receive the one or more interviews captured by the one or more image capturing devices 108 and the one or more microphones 110. The computing system 112 extracts audio and video data from the received one or more interviews between the interviewer and the candidate. Further, the computing system 112 also identifies one or more key segments from a plurality of segments.}
Jose discloses generating parameters;
and generating, by the computing device, a score corresponding to each of the set of predefined parameters of the candidate using the first self-learning AI model and the second self-learning AI model based on the comparison.
{[0038] The score card generation module 218 is configured to generate a score card associated with the interviewer including one or more interviewer profile parameters based on the determined one or more attributes and predefined criteria by using the interview optimization-based AI model.
{[0045] At step 310, a score card associated with the interviewer including one or more interviewer profile parameters is generated based on the determined one or more attributes and predefined criteria by using the interview optimization-based AI mode.
Jose does not explicitly disclose the self-learning AI model, comparing an information to reach a threshold, however; Yong-bo discloses;
extracting in near real-time, by the computing device, a set of video features from each of the plurality of frames of the video data using a first self-learning AI model, and a set of audio features from the audio data using a second self-learning AI model, wherein the set of video features and the set of audio features correspond to a set of predefined parameters;
{[Page 4]; Further, the self-supervised learning of the multimodal feature vector includes: whether the video frame sequence obtained by extracting frames from the video and the audio information extracted from the video come from the same video Binary classification supervised learning, and simultaneously perform binary classification supervised learning on whether the video frame sequence obtained by extracting frames from the video and the audio information extracted from the video are aligned.
[page 6] The fusion process includes: first, feature analysis is performed according to the video feature vector and audio feature vector extracted from the original video, the video refers to a smooth video stream, because the video stream contains rich visual, auditory and alphabetic information, so These video features and audio features can be color, texture, shape, tone, and text, etc., and then use multi-modal analysis methods, that is, use two or more modal information for processing at the same time.
[Page 4] An extraction module is used to extract video features and audio features;
The first self-supervised module is used for self-supervised learning of the extracted video features and audio features;
The second self-supervised module is used to perform self-supervised learning on the multimodal feature vector.}
Yong-bo discloses;
comparing, by the computing device, the set of video features and the set of audio features with self-adjusting threshold values corresponding to the set of predefined parameters;
{[ Page 3] Further, the video data preprocessing of the video includes: performing frame extraction on the video to obtain a video frame sequence, and setting a video frame sequence length threshold; if the sequence length is greater than the length threshold, perform equal interval Extracting the number of frames corresponding to the length threshold; if the sequence length is less than the length threshold, interpolating to the number of frames corresponding to the length threshold.
[Page 6] Carry out self-supervised learning on the extracted audio features. In the embodiment of the present invention, the audio unsupervised learning uses the method in wav2vec2.0 to block the audio signal, and train a comparison task to distinguish the real quantized hidden variable representation from other negative examples Come out to optimize the parameters of the audio feature extraction model.}
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose with the inclusion of analyzing data in real-time as taught by Yong-bo. Jose consisted references teaching receiving interviews, extracting audio and videos in real time using Artificial intelligence, determining attributes associated to the interviews and generating score cards associated with the interview parameter. Yong-bo also included analyzing in real time and comparing the information to reach a threshold, extracting in real time using a first and second self-supervised module.
The combination of Jose and Yong-bo does not disclose the screening method; however, Dangi discloses the following;
a score corresponding to each of the set of predefined parameters of the candidate using the first self-learning AI model and the second self-learning AI model based on the comparison.
{[0063] As set forth above, AI/ML engine 27 may assist with the screening and selection of candidate profiles. The AI/ML engine 27 may include an algorithm that parses resumes to extract information relevant to the position requisition. Alternatively, the AI/ML engine may access and assess parsed resume data received through an API. The AI/ML engine 27 may also access database 10 to access data that may include similar position requisitions and data associated with successful candidates that filled such similar position, such as educational background, years of experience, industry classifications and other data. The algorithm may be seeded with such data and continually learn which candidate profiles tend to be most successful. The AI/ML engine 27, after selecting such candidate profiles, may then receive additional inputs such as those described above with reference to FIG. 1e , to form a candidate skills model 50 for each candidate profile. The AI/ML engine 27 may then use predictive analytics models, such as forecast models, to predict which candidates will likely be most successful and to rank those candidate profiles based on a numerical score value. Other predictive analytics models may also be used, including, for example, a classification model in which the AI/ML engine 27 separates the data into categories that are applicable to previously successful candidates and then compares the data associated with the current selection of candidate profiles to the data classifications to determine whether a particular candidate is likely to be successful. Clustering models and others may also be used by the AI/ML engine 27.
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose and Yong-bo with the inclusion of a screening method as taught by Dangi.
The combination of Jose and Yong-bo, references receiving interviews, extracting audio and videos in real time using Artificial intelligence, determining attributes associated to the interviews and generating score cards associated with the interview parameter and comparing the information to reach a threshold. Dangi added to the references by screening the candidates.
As per claim 2, 10 and 18 (Similar scope and language)
Jose discloses;
The method of claim 1, further comprising generating a report for the candidate, wherein the report comprises the set of predefined parameters and the score corresponding to each of the set of predefined parameters.
{[0006] Furthermore, the plurality of modules include a score card generation module configured to generate a score card associated with the interviewer including one or more interviewer profile parameters based on the determined one or more attributes and predefined criteria by using the interview optimization-based AI model.
[0007] Also, the method includes generating a score card associated with the interviewer including one or more interviewer profile parameters based on the determined one or more attributes and predefined criteria by using the interview optimization-based AI model.
[0027] The computing system 112 generates a score card associated with the interviewer including one or more interviewer profile parameters based on the determined one or more attributes and predefined criteria by using the interview optimization-based AI model.}
As per claim 3, 11 and 19 (Similar scope and language)
Jose discloses;
The method of claim 1, wherein the set of predefined parameters comprises soft skill attributes, communication skill attributes, body language attributes, and knowledge attributes.
{[0034] The one or more key segments are sections of the plurality of segments in which relevant topics are discussed, such as qualification, experience, soft skills of the candidate and the like.
[0035] Body language and communication effectiveness is analyzed.}
As per claim 7 and 15 (Similar scope and language)
Jose discloses;
The method of claim 1, further comprising: extracting the video data from the input data; and extracting the audio data from the input data.
{[0027] The computing system 112 extracts audio and video data from the received one or more interviews between the interviewer and the candidate. Further, the computing system 112 also identifies one or more key segments from a plurality of segments. The plurality of segments are identified from the extracted audio data corresponding to the interviewer and the candidate. The computing system 112 determines one or more sentiment parameters for the interviewer and the candidate by analyzing the extracted video data. The one or more sentiment parameters include emotion, attitude, thought of the interviewer and the candidate and the like. Furthermore, the computing system 112 determines one or more attributes associated with the one or more interviews based on the extracted audio data, the extracted video data, the one or more key segments, the one or more sentiment parameters, job description, resume of the candidate or any combination thereof by using an interview optimization based Artificial Intelligence (AI) model.
Claim(s) 5, 8, 13, 16 are rejected under 35 U.S.C. 103 as being unpatentable over Jose et al, in view of Yong-bo et al, in view of Dangi et al; in further view of Olshansky et al [US2022,009,2548A1], hereafter Olshansky.
As per claim 5 and 13 (Similar scope and language)
Jose does not disclose receiving of data in real-time; however, Olshansky discloses the following;
The method of claim 1, further comprising receiving in real-time, the input data from a camera.
{[0015] In an embodiment, a method of connecting an employer with a candidate, is provided. The method can include receiving, at a system server, criteria data from the employer regarding a job opening, wherein the criteria data from the employer includes minimum attributes and real-time connection attributes. The method can include receiving, at the system server, background data from the candidate. The method can include recording audio data and video data of the candidate in a video interview of the candidate in a booth with a first camera, a second camera, and a microphone. The method can include recording behavioral data of the candidate with at least one depth sensor disposed in the booth. The method can include analyzing prior to an end of the video interview, at the system server, the audio data of the candidate with speech-to-text analysis to identify textual interview data, wherein candidate data includes the textual interview data and the background data. The method can include analyzing prior to the end of the video interview, at the system server, the behavioral data of the candidate to identify behavioral interview data, wherein the candidate data further includes the behavioral data.
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose, Yong-bo, and Dangi with the inclusion of receiving data in real-time as taught by Olshansky. Olshansky also included using cameras and microphone to receive data in real time, to enable analyzing the candidate behavioral data.
As per claim 8 and 16 (Similar scope and language)
The combination of Jose, Yong-bo, Dangi does not disclose the storing of video and audio data separately; however, Olshansky discloses the following;
The method of claim 1, further comprising: storing the video data in a video repository; and storing the audio data in an audio repository.
{[0085] To save storage space, audio and video compression formats can be utilized when storing data 862. These can include, but are not limited to, H.264, AVC, MPEG-4 Video, MP3, AAC, ALAC, and Windows Media Audio. Note that many of the video formats encode both visual and audio data. To the extent the microphones 220 are integrated into the cameras, the received audio and video data from a single integrated device can be stored as a single file. However, in some embodiments, audio data is stored separately the video data.
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose, Dangi and Mowbray with the inclusion of storing the video and audio data separately as taught by Olshansky. The combination of Jose and Dangi, references receiving interviews, extracting audio and videos in real time using Artificial intelligence, determining attributes associated to the interviews and generating score cards associated with the interview parameter, comparing the information to reach a threshold and screening the candidates. Olshansky also included a storage medium, to enable storing the audio and video formats in separate repository.
Claim(s) 4, 12, 20 are rejected under 35 U.S.C. 103 as being unpatentable over Jose et al, in view of Yong-bo et al, in view of Dangi et al; in further view of Mowbray et al [US2024,021,1888], hereafter Mowbray.
As per claim 4, 12 and 20 (Similar scope and language)
The combination of Jose, Yong-bo and Dangi does not disclose the Natural Language Processing (NLP) model; however, Mowbray discloses the following;
The method of claim 1, wherein the second self-learning AI model is a Natural Language Processing (NLP) model.
{[0024] In embodiments, the employment matching system 102 can provide one or more job matching and/or searching algorithms to identify potential employment opportunities for candidates, e.g., the user 104. The employment matching system 102 can utilize one or more weighted scoring algorithms for determining matches for potential employment opportunities, as described below. The employment matching system 102 can utilize one or more trained machine learning algorithms. The one or more trained machine learning algorithms can utilize natural language processing (NLP) techniques to match jobs to candidates.
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose, Yong-bo and Dangi with the inclusion of a Natural Language Processing (NLP) model as taught by Mowbray. The combination of Jose, Yong-bo and Dangi, references receiving interviews, extracting audio and videos in real time using Artificial intelligence, determining attributes associated to the interviews and generating score cards associated with the interview parameter, comparing the information to reach a threshold and screening the candidates. Mowbray included an employment matching system using a Natural Language Processing (NLP) model.
Claim(s) 6, 14 are rejected under 35 U.S.C. 103 as being unpatentable over Jose et al, in view of Yong-bo et al, in view of Dangi et al; in view of Olshansky et al, in further view of Mowbray et al [US2024,021,1888], hereafter Mowbray.
As per claim 6 and 14 (Similar scope and language)
The combination of Jose, Yong-bo, Olshansky and Dangi does not disclose the training of the AI system; however, Mowbray discloses the following;
The method of claim 5, further comprising training the first self-learning AI model and the second self-learning AI model using the input data received in real-time.
{[0024] In embodiments, the employment matching system 102 can provide one or more job matching and/or searching algorithms to identify potential employment opportunities for candidates, e.g., the user 104. The employment matching system 102 can utilize one or more weighted scoring algorithms for determining matches for potential employment opportunities, as described below. The employment matching system 102 can utilize one or more trained machine learning algorithms. The one or more trained machine learning algorithms can utilize natural language processing (NLP) techniques to match jobs to candidates. For example, various data points (e.g. job title and job description) embeddings are generated and then a cosine similarity is used to score the similarity between user and job profiles. Jobs with the highest scores will be recommended to the candidates. The one or more machine learning algorithms can utilize self-learning processes to tune the one or more machine learning algorithms to specific users, specific employers, specific job fields, specific demographics, and the like.
Motivation: It would have been obvious to one of ordinary skill in the art before the effective filing date of the claimed invention to modify the system and method of evaluating candidates as disclosed by Jose Olshansky and Dangi with the inclusion of training the AI system as taught by Mowbray. The combination of Jose, Olshansky and Dangi, references receiving interviews, extracting audio and videos in real time using Artificial intelligence, determining attributes associated to the interviews and generating score cards associated with the interview parameter, comparing the information to reach a threshold and screening the candidates. Mowbray included a trained machine learning algorithms using a Natural Language Processing (NLP) model to achieve matching the employment system.
Response to Arguments
Applicant’s arguments, see response, filed 3 December 2025, with respect to the rejection(s) of claim(s) 1-3, 5, 7-11, 13, 15-19 under 103 have been fully considered and are persuasive. Therefore, the rejection has been withdrawn. However, upon further consideration, a new ground(s) of rejection is made in view of newly found prior art reference(s), Yong-bo.
Conclusion
A shortened statutory period for reply to this final action is set to expire THREE MONTHS from the mailing date of this action. In the event a first reply is filed within TWO MONTHS of the mailing date of this final action and the advisory action is not mailed until after the end of the THREE-MONTH shortened statutory period, then the shortened statutory period will expire on the date the advisory action is mailed, and any nonprovisional extension fee (37 CFR 1.17(a)) pursuant to 37 CFR 1.136(a) will be calculated from the mailing date of the advisory action. In no event, however, will the statutory period for reply expire later than SIX MONTHS from the mailing date of this final action.
Any inquiry concerning this communication or earlier communications from the examiner
should be directed to VICTOR CHIGOZIRIM ESONU whose telephone number is (571)272 -
4883. The examiner can normally be reached Monday - Friday 9:00 am - 5pm.
Examiner interviews are available via telephone, in-person, and video conferencing using
a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is
encouraged to use the USPTO Automated Interview Request (AIR) at
http://www.uspto.gov/interviewpractice.
If attempts to reach the examiner by telephone are unsuccessful, the examiner’s
supervisor, Sarah Monfeldt can be reached on (571) 270-1833. The fax phone number for the
organization where this application or proceeding is assigned is 571-273-8300.
Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, vis it: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000.
/VICTOR CHIGOZIRIM ESONU/
Examiner, Art Unit 3629