Prosecution Insights
Last updated: April 19, 2026
Application No. 18/882,295

INTERACTIVE VIRTUAL OBJECT PLACEMENT WITH CONSISTENT PHYSICAL REALISM

Non-Final OA §103
Filed
Sep 11, 2024
Examiner
WU, MING HAN
Art Unit
2618
Tech Center
2600 — Communications
Assignee
Rembrand Inc.
OA Round
1 (Non-Final)
76%
Grant Probability
Favorable
1-2
OA Rounds
2y 8m
To Grant
99%
With Interview

Examiner Intelligence

Grants 76% — above average
76%
Career Allow Rate
282 granted / 370 resolved
+14.2% vs TC avg
Strong +23% interview lift
Without
With
+23.3%
Interview Lift
resolved cases with interview
Typical timeline
2y 8m
Avg Prosecution
35 currently pending
Career history
405
Total Applications
across all art units

Statute-Specific Performance

§101
7.8%
-32.2% vs TC avg
§103
68.3%
+28.3% vs TC avg
§102
2.1%
-37.9% vs TC avg
§112
12.6%
-27.4% vs TC avg
Black line = Tech Center average estimate • Based on career data from 370 resolved cases

Office Action

§103
DETAILED ACTION Notice of Pre-AIA or AIA Status The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA . In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status. Claim Rejections - 35 USC § 103 The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action: A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102 of this title, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made. The factual inquiries set forth in Graham v. John Deere Co., 383 U.S. 1, 148 USPQ 459 (1966), that are applied for establishing a background for determining obviousness under 35 U.S.C. 103 are summarized as follows: 1. Determining the scope and contents of the prior art. 2. Ascertaining the differences between the prior art and the claims at issue. 3. Resolving the level of ordinary skill in the pertinent art. 4. Considering objective evidence present in the application indicating obviousness or nonobviousness. Claims 1, 2, 3, 4, 7, 9, 10 – 14, 17, 19 and 20 are rejected under 35 U.S.C. 103 as being unpatentable over Ahmed et al. (Publication: US 2022/0180898 A1) in view of Tomaru (Publication: US 2021/0256760 A1). Regarding claim 1, see rejection on claim 19. Regarding claim 2, see rejection on claim 20. Regarding claim 3, see rejection on claim 13. Regarding claim 4, see rejection on claim 14. Regarding claim 7, see rejection on claim 17. Regarding claim 9, Ahmed in view of Tomaru disclose all the limitation of claim 1. Ahmed discloses further comprising generating a relative pixel depth map associated with a frame included in the input video sequence ([0029] , [0065] - Replacement media generation constraints 410 may include rules that can be used to match secondary content items with candidate insertion points of media content items. More particularly, the rules may include or be used to generate a “mapping” between compatible scene types (e.g., kitchen counter in comedy, billboard outside in drama) and attributes of compatible secondary content (e.g., product type, product category, brand). Constraints 410 can specify secondary content items (or associated attributes) that are permitted (or, conversely, not permitted) to be “placed” in primary content. For example, the primary content can include a specific media item, a specific clip/scene, or a specific candidate placement. As another example, the primary content may be a video title, clip, or candidate placement having specific attributes. Each secondary content item represents a three-dimensional object, 3D information indicating a depth of the first object, having a flat surface that is suitable for placement on a kitchen counter. In this example, secondary content items 128 include a beer bottle, a soda can, and an Amazon Echo. Replacement clips 130 are generated using original clip 122 and secondary content items 128. Each replacement clip 130 is the same length and corresponds to the same segment of time. Thus Mapping involves 3D information indicating a depth of the secondary content item in the clip, video sequence.). Regarding claim 10, Ahmed in view of Tomaru disclose all the limitation of claim 1. Ahmed discloses further comprising generating an ordered list of combinations of virtual objects and planar surfaces based at least on the suitability metric ([0055] - Attributes of a candidate placement can further include a duration of time that the candidate placement is visible on screen. a candidate placement can include attributes of object(s) on which secondary content can be “placed” (e.g., categorization, relative size of the object relative to other content, pixel coordinates, etc.) and/or attribute(s) surface(s) on which the secondary content can be “placed” (e.g., whether the surface is vertical or horizontal, whether the surface is flat or curved, relative size of the surface, etc.). [0016] - corresponding to surfaces of object(s) represented within the clip on which a virtual product (e.g. Amazon logo), object, or branding can be placed .). Regarding claim 11, see rejection on claim 19. Regarding claim 12, see rejection on claim 20. Regarding claim 13, Ahmed in view of Tomaru disclose all the limitation of claim 12. Ahmed discloses wherein the suitability metric is further based on a size or a duration associated with the virtual object ( [0055] - Attributes of a candidate placement can further include a duration of time that the candidate placement is visible on screen. a candidate placement can include attributes of object(s) on which secondary content can be “placed” (e.g., categorization, relative size of the object relative to other content, pixel coordinates, etc.) and/or attribute(s) surface(s) on which the secondary content can be “placed” (e.g., whether the surface is vertical or horizontal, whether the surface is flat or curved, relative size of the surface, etc.). [0016] - corresponding to surfaces of object(s) represented within the clip on which a virtual product (e.g. Amazon logo), object, or branding can be placed .). Regarding claim 14, Ahmed in view of Tomaru disclose all the limitation of claim 11. Ahmed discloses wherein the one or more machine learning models include [[a rendering generator or a diffusion generator] -] - ([0086] - Machine learning algorithms used to ascertain media attributes may be automatically updated as new data is collected. This data can include additional attributes that are ascertained, for example, based on analysis of frames of video titles as they are added to video title catalog 702. In addition, the data can include human input that validates (or invalidates) the accuracy of attributes determined via the machine learning algorithms, as well as human input that modifies previously identified attributes of video titles or adds additional attributes to those previously identified.). Tomaru discloses include a rendering generator or a diffusion generator ([0123] - The generation of the image of the virtual object in the three-dimensional drawer 45 “rendering generator”can be implemented by performing rendering by using three-dimensional graphics.) Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed in view Tomaru with include a rendering generator or a diffusion generator as taught by Tomaru. The motivation for doing so the user can institutively perceive the position of an object. Regarding claim 17, Ahmed in view of Tomaru disclose all the limitation of claim 11. Tomaru discloses to perform the step of generating a polygon defining a boundary of the planar surface ([0119] - As illustrated in FIGS. 8 and 9A, a pair of straight lines Lθa and Lθb forming the angles θa and θb with the Zv-axis in the Xv-Zv plane in the user viewpoint coordinate system are drawn, and boundary surfaces Sθa and Sθb are formed by trajectories obtained by moving the straight lines Lθa and Lθb in the Yv-axis direction. Although the straight lines Lθa and Lθb and boundary surfaces Sθa and Sθb extend infinitely, FIG. 9A illustrates only part of the straight lines Lθa and Lθb and boundary surfaces Sθa and Sθb. The boundary surfaces Sθa and Sθb can be said to be surfaces that define the area of the reference horizontal view angle.). Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed with select a virtual object included in an object library as taught by Tomaru. The motivation for doing so the user can institutively perceive the position of an object. Regarding claim 19, Ahmed discloses a system comprising ([0038] , [0040] - FIG. 3, a computing environment in which a video content service 302 provides streaming content (e.g., video and/or audio) via network 304 to a variety of client devices (306-1 through 306-5).): one or more memories storing instructions ([0038] - FIG. 3, memory stores instructions. ); and one or more processors for executing the instructions to ([0038] - FIG. 3, the processor to execute instructions.) : identify a planar surface depicted in an input video sequence ([0020], [0073] - surface of box is identified as a candidate placement for secondary content in the clip, frames of the movie “sequence”. [0073] - the clip, frames of the movie, is retrieved from primary content information, “input video sequence, frames” .); generate, for a combination of the planar surface and the virtual object, a suitability metric associated with the combination ( [0065] - rules that can be used to match secondary content items “a suitability metric” with candidate insertion points of media content items. More particularly, the rules may include or be used to generate a “mapping” between compatible scene types (e.g., kitchen counter in comedy, billboard outside in drama) and attributes of compatible secondary content (e.g., product type, product category, brand), “generate a suitability metric”. [0026] - As shown in FIG. 1A, replacement clip is generated using original clip and a secondary content overlay “combination” including an Amazon logo based on the rules used to match. In this example, vertical surface 116 of Amazon box 118 represented in replacement clip contains an Amazon logo, which replaces the text previously shown on vertical surface of box in original clip. Since the actor is moving box in the scene, pixel coordinates of candidate placement may vary among frames in replacement clip, “for a combination of the planar surface and the virtual object; a suitability metric associated with the combination”. [0016] - corresponding to surfaces of object(s) represented within the clip on which a virtual product (e.g. Amazon logo), object, or branding can be placed, “virtual object” PNG media_image1.png 338 612 media_image1.png Greyscale ), wherein the suitability metric is based at least on a semantic compatibility between the virtual object and the planar surface ( [0051], [0065] - rules that can be used to match secondary content items with candidate insertion points of media content items. More particularly, the rules may include or be used to generate a “mapping” between compatible scene types (e.g., kitchen counter in comedy, billboard outside in drama) and attributes of compatible secondary content (e.g., product type, product category, brand), “suitability metric”. For a given title, clips corresponding to candidate insertion points may be scored according to a set of scoring algorithms, where each candidate insertion point clip includes a representation of an virtual object or associated object surface on which secondary content can be virtually placed. For example, the clip can correspond to a scene in a movie. Candidate insertion points at which segments of a video content item can be replaced with a corresponding replacement clip may be selected based, at least in part, on the scores, “suitability metric is based at least on a semantic compatibility between the virtual object and the planar surface”. [0016] - candidate replacement, e.g. Amazon logo, corresponding to surfaces of object(s) represented within the clip on which a virtual product (e.g. Amazon logo), object, or branding can be placed, “virtual object”); and generate, via one or more machine learning models, a modified video sequence based on the suitability metric ( [0026] - As shown in FIG. 1A, replacement clip “video” is generated using original clip and a secondary content overlay including an Amazon logo, “generate”. Machine learning techniques can be used to identify candidate insertion points , as well as attributes of an identified candidate insertion point. A candidate insertion point may be defined by a clip (e.g., scene) within the media content item, as well as candidate placement(s) “modified” corresponding to surfaces of physical object(s) represented within the clip on which a virtual product, (e.g. Amazon logo) or branding can be placed. Attributes of a candidate insertion point can include a start time of the pertinent clip, an end time and/or duration of the pertinent clip, and attributes of pertinent physical object(s) represented within the clip, “generate, via one or more machine learning models, a modified video sequence”. [0065] - include rules that can be used to match secondary content items with candidate insertion points of media content items. More particularly, the rules may include or be used to generate a “mapping” between compatible scene types (e.g., kitchen counter in comedy, billboard outside in drama) and attributes of compatible secondary content (e.g., product type, product category, brand), (e.g. Amazon logo), “generate a modified video sequence based on the suitability metric”. [0016] - The information that is collected and stored for a media content item may be analyzed to identify “candidate insertion points” at which secondary content may be “inserted” to achieve virtual product placements “modified” or branding. A candidate insertion point is a potential two-dimensional (2D) placement for a virtual product or branding. ), wherein the modified video sequence depicts the virtual object placed on the planar surface ([0016] - A candidate insertion point may be defined by a clip “video” (e.g., scene) within the media content item, as well as candidate placement(s) “modified” corresponding to surfaces of physical object(s) represented within the clip on which a virtual product or branding can be placed, “modified video sequence depicts the virtual object placed on the planar surface”. [0026] - As shown in FIG. 1A, replacement clip “video” is generated using original clip and a secondary content overlay including an Amazon logo. [0016] - candidate replacement “modified”, e.g. Amazon logo, corresponding to surfaces of object(s) represented within the clip on which a virtual product (e.g. Amazon logo), object, or branding can be placed, “virtual object” Attributes of a candidate insertion point can include a start time of the pertinent clip, an end time and/or duration of the pertinent clip, and attributes of pertinent physical object(s) represented within the clip. Attributes of an object represented within a clip can include a classification of the object (e.g., a desk, a bus, an empty wall, a billboard). Attributes of a candidate insertion point can also include attributes of the pertinent clip (e.g., scene) and/or media content item such as a setting or genre. PNG media_image2.png 314 612 media_image2.png Greyscale ). Ahmed does not however Tomaru discloses select a virtual object included in an object library ([0094] - a virtual object selector to select one of the plurality of virtual objects “library” stored in the information storage.). Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed with select a virtual object included in an object library as taught by Tomaru. The motivation for doing so the user can institutively perceive the object. Regarding claim 20, Ahmed in view of Tomaru disclose all the limitation of claim 19. Ahmed discloses wherein the suitability metric is further based on a semantic compatibility between the virtual object and a scene depicted in the input video sequence ([0055] - Attributes of a candidate placement can further include a duration of time that the candidate placement is visible on screen. a candidate placement can include attributes of object(s) on which secondary content can be “placed” (e.g., categorization, relative size of the object relative to other content, pixel coordinates, etc.) and/or attribute(s) surface(s) on which the secondary content can be “placed” (e.g., whether the surface is vertical or horizontal, whether the surface is flat or curved, relative size of the surface, etc.). [0016] - corresponding to surfaces of object(s) represented within the clip, “input video sequence” , on which a virtual product (e.g. Amazon logo), object, or branding can be placed, “virtual object” .). Claims 8 and 18 are rejected under 35 U.S.C. 103 as being unpatentable over Ahmed et al. (Publication: US 2022/0180898 A1) in view of Tomaru (Publication: US 2021/0256760 A1) and Border et al. (Publication: US 2012/0019928 A1). . Regarding claim 8, see rejection on claim 18. Regarding claim 18, Ahmed in view of Tomaru disclose all the limitation of claim 11. Ahmed in view of Tomaru do not however Border discloses to perform the step of generating, for each of one or more pixels included in the planar surface, a normal vector describing an orientation of the pixel ([0026] , [0029] , [0038] - generate the norma map and the normal map 244 indicates surface normals in the rendered image 240. For example, the normal map 244 could include a respective vector for each pixel of the character indicating a surface normal direction.). Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed in view of Tomaru with to perform the step of generating, for each of one or more pixels included in the planar surface, a normal vector describing an orientation of the pixel as taught by Border. The motivation for doing so to effectively improve rendering. Claims 5, 6, 15 and 16 are rejected under 35 U.S.C. 103 as being unpatentable over Ahmed et al. (Publication: US 2022/0180898 A1) in view of Tomaru (Publication: US 2021/0256760 A1) and Gusmao et al. (NPL: DeepPlacer: A custom integrated OpAmp placement tool using deep models, Applied Soft Computing, Volume 115, January 2022, 108188) Regarding claim 5, see rejection on claim 15. Regarding claim 6, see rejection on claim 16. Regarding claim 15, Ahmed in view of Tomaru disclose all the limitation of claim 11. Ahmed discloses perform the step of iteratively modifying one or more input parameters associated with the one or more machine learning models based on a function ([0016] , [0086] - Machine learning algorithms used to ascertain media attributes may be automatically updated as new data is collected, “modifying one or more input parameter”. Computer vision and machine learning techniques can be used to identify candidate insertion points, as well as attributes of an identified candidate insertion point. A candidate insertion point may be defined by a clip (e.g., scene) within the media content item, as well as candidate placement(s) corresponding to surfaces of physical object(s) represented within the clip on which a virtual product or branding can be placed. Attributes of a candidate insertion point can include a start time of the pertinent clip, an end time and/or duration of the pertinent clip, and attributes of pertinent physical object(s) represented within the clip. Attributes of an object represented within a clip can include a classification of the object (e.g., a desk, a bus, an empty wall, a billboard). Attributes of a candidate insertion point can also include attributes of the pertinent clip (e.g., scene) and/or media content item such as a setting or genre. [0089] - the candidate placement identification process may be applied, repeatedly, “iteratively”. ). Ahmed in view of Tomaru do not however Border discloses Perform based on a generation loss function (Page 1 – innovative formulation of a generation loss function). Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed in view of Tomaru with Perform based on a generation loss function as taught by Border. The motivation for doing so to effectively improve the design by automation. Regarding claim 16, Ahmed in view of Tomaru disclose all the limitation of claim 11. Ahmed discloses to perform the step of iteratively modifying one or more placement parameters based on a placement function ([0106] , Fig. 9 - , replacement clips may be selected for more than one candidate insertion point with an emphasis on repeated exposure to the same product or brand. If there are further candidate insertion points for the video (918), the replacement clip selection process may repeat “iteratively” for subsequent candidate insertion points (910). Step 918 in Fig. 9 is “placement parameters based on a placement function”). Ahmed in view of Tomaru do not however Border discloses Perform based on a generation loss function (Page 1 – innovative formulation of a generation loss function). Before the effective filing date of the claimed invention, it would have been obvious to one of ordinary skill in the art to modify Ahmed in view of Tomaru with Perform based on a generation loss function as taught by Border. The motivation for doing so to effectively improve the design by automation. Conclusion The prior art made of record and not relied upon is considered pertinent to applicant's disclosure. Any inquiry concerning this communication or earlier communications from the examiner should be directed to DEVONA E FAULK whose telephone number is (571)272-7515. The examiner can normally be reached on M-Th on campus: 9:30am - 6:00pm EST . Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice. If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, Devona Faulk can be reached on 571-272-7515. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300. Information regarding the status of an application may be obtained from the Patent Application Information Retrieval (PAIR) system. Status information for published applications may be obtained from either Private PAIR or Public PAIR. Status information for unpublished applications is available through Private PAIR only. For more information about the PAIR system, see http://pair-direct.uspto.gov. Should you have questions on access to the Private PAIR system, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative or access to the automated information system, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000. /MING WU/ Primary Examiner, Art Unit 2618
Read full office action

Prosecution Timeline

Sep 11, 2024
Application Filed
Oct 24, 2025
Response after Non-Final Action
Feb 24, 2026
Non-Final Rejection — §103 (current)

Precedent Cases

Applications granted by this same examiner with similar technology

Patent 12597109
SYSTEMS AND METHODS FOR GENERATING THREE-DIMENSIONAL MODELS USING CAPTURED VIDEO
2y 5m to grant Granted Apr 07, 2026
Patent 12579702
METHOD AND SYSTEM FOR ADAPTING A DIFFUSION MODEL
2y 5m to grant Granted Mar 17, 2026
Patent 12579623
IMAGE PROCESSING METHOD AND APPARATUS, ELECTRONIC DEVICE, AND READABLE STORAGE MEDIUM
2y 5m to grant Granted Mar 17, 2026
Patent 12567185
Method and system of creating and displaying a visually distinct rendering of an ultrasound image
2y 5m to grant Granted Mar 03, 2026
Patent 12548202
TEXTURE COORDINATE COMPRESSION USING CHART PARTITION
2y 5m to grant Granted Feb 10, 2026
Study what changed to get past this examiner. Based on 5 most recent grants.

AI Strategy Recommendation

Get an AI-powered prosecution strategy using examiner precedents, rejection analysis, and claim mapping.
Powered by AI — typically takes 5-10 seconds

Prosecution Projections

1-2
Expected OA Rounds
76%
Grant Probability
99%
With Interview (+23.3%)
2y 8m
Median Time to Grant
Low
PTA Risk
Based on 370 resolved cases by this examiner. Grant probability derived from career allow rate.

Sign in with your work email

Enter your email to receive a magic link. No password needed.

Personal email addresses (Gmail, Yahoo, etc.) are not accepted.

Free tier: 3 strategy analyses per month