DETAILED ACTION
Notice of Pre-AIA or AIA Status
The present application, filed on or after March 16, 2013, is being examined under the first inventor to file provisions of the AIA .
Claim Rejections - 35 USC § 112
The following is a quotation of 35 U.S.C. 112(b):
(b) CONCLUSION.—The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the inventor or a joint inventor regards as the invention.
The following is a quotation of 35 U.S.C. 112 (pre-AIA ), second paragraph:
The specification shall conclude with one or more claims particularly pointing out and distinctly claiming the subject matter which the applicant regards as his invention.
Claims 2 – 10, 12, 13, and 18 are rejected under 35 U.S.C. 112(b) or 35 U.S.C. 112 (pre-AIA ), second paragraph, as being indefinite for failing to particularly point out and distinctly claim the subject matter which the inventor or a joint inventor (or for applications subject to pre-AIA 35 U.S.C. 112, the applicant), regards as the invention.
Claims 2 and 12 recite the limitation “a reference frame… the reference frame comprising at least one of the first still image or the second still image” in lines 6 - 7. Claims 1 and 11, the claims upon which claims 2 and 12 respectively depend, however, recite a limitation “a reference frame…the reference frame comprising the first still image” in lines 6 - 7. It is therefore unclear whether the reference frame recitation of claims 2 and 12 is the same reference frame of claims 1 and 11 or a different reference frame. For the purposes of examination, examiner interprets “a reference frame…the reference frame comprising the first still image” of claims 1 and 11 to be “a first reference frame, the first reference frame comprising the first still image”, and “a reference frame… the reference frame comprising at least one of the first still image or the second still image” of claims 2 and 12 to be “a second reference frame, the second reference frame comprising at least one of the first still image or the second still image”.
With respect to claim 5, the claim recites the limitation “the reference frame.” Claim 1, a claim upon which claim 5 properly depends however, recites the limitation “a reference frame…the reference frame comprising the first still image”, and claim 2, a claim upon which claim 5 properly depends, recites the limitation “a reference frame… the reference frame comprising at least one of the first still image or the second still image.” It is therefore unclear whether the reference claim of claim 1 or the reference frame of claim 2 is being referred to. For the purposes of examination, examiner interprets “the reference frame” to mean “the second reference frame” [see discussion above].
With respect to claim 9, the claim recites the limitation “the reference frame” in line 2. Claim 1, a claim upon which claim 9 properly depends however, recites the limitation “a reference frame…the reference frame comprising the first still image”, and claim 2, a claim upon which claim 9 properly depends, recites the limitation “a reference frame… the reference frame comprising at least one of the first still image or the second still image.” It is therefore unclear whether the reference claim of claim 1 or the reference frame of claim 2 is being referred to. For the purposes of examination, examiner interprets “the reference frame” to mean “the first reference frame and the second reference frame” [see discussion above].
Claims 4, 6 – 8 and 10 properly depend on claim 2, and are therefore likewise rejected.
Claims 13 and 18 properly depend on claim 12, and are therefore likewise rejected.
Claim Rejections - 35 USC § 103
In the event the determination of the status of the application as subject to AIA 35 U.S.C. 102 and 103 (or as subject to pre-AIA 35 U.S.C. 102 and 103) is incorrect, any correction of the statutory basis (i.e., changing from AIA to pre-AIA ) for the rejection will not be considered a new ground of rejection if the prior art relied upon, and the rationale supporting the rejection, would be the same under either status.
The following is a quotation of 35 U.S.C. 103 which forms the basis for all obviousness rejections set forth in this Office action:
A patent for a claimed invention may not be obtained, notwithstanding that the claimed invention is not identically disclosed as set forth in section 102, if the differences between the claimed invention and the prior art are such that the claimed invention as a whole would have been obvious before the effective filing date of the claimed invention to a person having ordinary skill in the art to which the claimed invention pertains. Patentability shall not be negated by the manner in which the invention was made.
The factual inquiries for establishing a background for determining obviousness under 35 U.S.C. 103 are summarized as follows:
1. Determining the scope and contents of the prior art.
2. Ascertaining the differences between the prior art and the claims at issue.
3. Resolving the level of ordinary skill in the pertinent art.
4. Considering objective evidence present in the application indicating obviousness or nonobviousness.
Claim(s) 1 – 13, 18, and 51 - 56 is/are rejected under 35 U.S.C. 103 as being unpatentable over Chen et al. (US 2013/0038686) (hereinafter Chen) in view of Furlan (US 2010/0295966) (hereinafter Furlan).
Regarding claims 1 and 11, Chen teaches a method and a system comprising control circuitry configured to perform the method, the metho comprising:
receiving, by control circuitry, image data comprising a first video having a capture duration (e.g. Fig. 1 and par. 42: depicting and describing that the system receives image data, the image data including a video);
image encoding, by control circuitry, the first still image for storage (Chen, e.g. Figs. 2 and 4, and pars. 42, 72 – 85 and 105: depicting and describing that a first frame of the video is encoded, wherein a frame of video is the equivalent of a still image); and
video encoding, by control circuitry, the first video via inter-prediction for storage, said inter-prediction using a reference frame as a surrogate intra-coded (I) frame, the reference frame comprising the first still image (e.g. Figs. 2 and 4, and pars. 72 – 85 and 105: depicting and describing that the first frame of the video [I-frame] is used as a reference frame for inter-prediction coding of other frames of the video, wherein the frame of video is the equivalent of the still image).
Chen does not explicitly teach:
wherein the image data further comprising a first still image captured during the capture duration.
Furlan, however, teaches a method and system:
wherein the image data further comprising a first still image captured during the capture duration (e.g. Figs. 4, and pars. 38 – 42 and 62: depicting and describing that high-resolution still images are captured at the same time as lower resolution video images are captured, the still images are encoded as I-frames while the video data are P-frames and B-frames).
It therefore would have been obvious to one of ordinary skill in the art to modify the teachings of Chen by adding the teachings of Furlan in order for the image data to further comprise a first still image captured during the capture duration. One of ordinary skill in the art would have been motivated to make such a modification because the modification allows for both video images and high-resolution still images to be available for playback (Furlan, e.g. par. 34: describing a desire to provide both lower resolution video images and high-resolution still images for display automatically).
Turning to claims 2 and 12, Chen and Furlan teach all of the limitations of claims 1 and 11, respectively, as discussed above. Chen further teaches:
wherein the image data further comprises a second video captured simultaneous with the first video, and a second still image captured simultaneous with the first still image (e.g. Fig. 4 and pars. 103 – 104: depicting and describing that the system receives image data containing two views such that a stereoscopic pair is formed, wherein image data containing two views is the equivalent of a second video and a second still image [wherein a frame of the second video is the equivalent of a still image, see discussion above], wherein it is known to those of ordinary skill in the art that in order to form a stereo pair, slightly displaced images must necessarily be captured at the same time); and
wherein: the image encoding further comprises, image encoding the second still image; and the video encoding further comprises, video encoding the second video via inter-prediction for storage, said inter-prediction using a reference frame as a surrogate I-frame, the reference frame comprising at least one of: the first still image; or the second still image (e.g. Figs. 4 and 5A, and pars. 105 – 106 and 117 – 119: depicting and describing that the system encodes the second view video using a frame from the first view video to predict the second view video, the frame of the first view video being an I-frame, wherein the frame of the first view video is the equivalent of the first still image [see discussion above]).
Regarding claims 3 and 13, Chen and Furlan teach all of the limitations of claims 1 and 2, and claims 11 and 12, respectively, as discussed above. Chen further teaches:
wherein image encoding the second still image comprises inter-view prediction using the first still image as a reference picture (e.g. Fig. 5A and pars. 117 – 119: depicting and describing that encoding a frame of the second view comprises inter-view prediction using a frame of the first view as a reference picture, wherein the frame of the first view is the equivalent of the first still image and the frame of the second view is the equivalent of the second still image).
Turning to claim 4, Chen and Furlan teach all of the limitations of claims 1 and 2, as discussed above. Chen further teaches:
wherein the first still image is captured using one or more first optical parameters and the second still image is captured using one or more second optical parameters different to the first optical parameters (e.g. pars. 103 – 104: describing that the at least two videos are videos of two different fields of view, wherein a first video captured with a first field of view and a second video captured with a second field of view is the equivalent of the first image being captured using one or more first optical parameters and the second image being captured using one or more second optical parameters different from the first optical parameters).
Regarding claim 5, Chen and Furlan teach all of the limitations of claims 1 and 2, as discussed above. Chen further teaches:
wherein the video encoding further comprises: generating the reference frame, said generating comprising:
decoding at least one of: the encoded first still image; or the encoded second still image (e.g. Fig. 5A, element 106, and pars. 117 – 119: depicting and describing that the system decodes an encoded picture of the first view, wherein a picture of the first view is the equivalent of the first still image); and
adjusting at least one of: the decoded first still image; or the decoded second still image, said adjusting using one or more selected from: spatial alignment; cropping; scaling; resampling (e.g. Fig. 5A, element 108, and pars. 117 – 119: depicting and describing that the system upsamples the decoded picture of the first view, wherein upsampling the picture is the equivalent of resampling, and wherein the picture of the first view is the equivalent of the first still image).
Turning to claim 6, Chen and Furlan teach all of the limitations of claims 1, 2, and 5, as discussed above. Chen further teaches:
wherein said adjusting is based on a video frame being encoded from at least one of: the first video; or the second video, said adjusting comprising: identifying a matched feature between: at least one of: the first still image; or the second still image; and the video frame being encoded; and adjusting at least one of: the decoded first still image; or the decoded second still image, such that the generated reference frame comprises the matched feature (e.g. Fig. 5A and pars. 117 – 122: depicting and describing that the system adjusts the first view picture based on the resolution of the picture to be encoded, the picture of the first view either upsampled or downsampled to match the resolution of the picture to be encoded, wherein upsampling or downsampling an image is the equivalent of resampling, wherein the picture to be encoded is the equivalent of either the first video or the second video, wherein the first view picture is the equivalent of the first still image, wherein determining the resolution of the picture to be encoded is the equivalent of identifying the matched feature, and wherein resampling the picture of the first view based on whether the picture of the second view is at full resolution or half resolution is the equivalent of adjusting the decoded first still image such that the generated reference frame comprises the matched feature).
Regarding claim 7, Chen and Furlan teach all of the limitations of claims 1, 2, and 5, as discussed above. Chen further teaches:
wherein the first video and the first still image share a common first perspective, and wherein the second video and the second still image share a common second perspective; and wherein generating the reference frame further comprises forming a stereoscopic still image from the adjusted first still image and second still image (e.g. Fig. 4 and pars. 101 – 104: depicting and describing a first video comprising a plurality of frames, the first video having a first view [S0] and a second video comprising a plurality of frames, the second video having a second view [S1], frames of the first video and frames of the second video forming stereo view pairs, wherein a frame of the first video is the equivalent of the first still image and a frame of the second video is the equivalent of the second still image [see discussion above]).
Turning to claims 8 and 18, Chen and Furlan teach all of the limitations of claims 1 and 2, and claims 11 and 12, respectively, as discussed above. Chen further teaches:
wherein the video encoding further comprises: generating video frames to be encoded, said generating comprising: adjusting frames of at least one of: the first video; or the second video, said adjusting using one or more selected from: spatial alignment; cropping; scaling; resampling; frame rate adjustment; aspect ratio adjustment; letter-boxing; pillar-boxing (e.g. Fig. 5A and pars. 117 – 122: depicting and describing that frames of the first view video and frames of the second view video are downsampled, wherein downsampling is the equivalent of resampling).
Chen does not explicitly teach:
excluding, for said video encoding, a video frame of the first video captured at a same time instance as the first still image and a video frame of the second video captured at a same time instance as the second still image.
Furlan, however, teaches a method and system:
excluding, for said video encoding, a video frame of the first video captured at a same time instance as the first still image and a video frame of the second video captured at a same time instance as the second still image (e.g. par. 41: describing that a captured still image takes the place of a frame of a video when captured at the same time as the video [captured still image replaces frames 1 and 21 of the video, the system then encoding only frames 2 – 20 of the captured video], wherein the still image taking the place of a frame of video captured at the same time is the equivalent of excluding from encoding a video frame of the first video captured at the a same time instance as the first still image and a video frame of the second video captured at a same time instance of the second still image).
It therefore would have been obvious to one of ordinary skill in the art to modify the teachings of Chen by adding the teachings of Furlan in order to exclude, for said video encoding, a video frame of the first video captured at a same time instance as the first still image and a video frame of the second video captured at a same time instance as the second still image. One of ordinary skill in the art would have been motivated to make such a modification because the modification allows for both video images and high-resolution still images to be available for playback (Furlan, e.g. par. 34: describing a desire to provide both lower resolution video images and high-resolution still images for display automatically).
Regarding claim 9, Chen and Furlan teach all of the limitations of claims 1, 2 and 8, as discussed above. Chen further teaches:
wherein the video encoding further comprises: resampling the reference frame, such that the resampled reference frame comprises a resolution matching a resolution of the video frames to be video encoded; video encoding the video frames to be video encoded using the resampled reference frame (e.g. Fig. 5A, and pars. 117 – 122: depicting and describing that the system either downsamples or upsamples a picture of the first view video to match the resolution of the other view video to be encoded, the system encoding pictures of the other view using the resampled picture of the first view [downsampled to match the resolution of the second view video and upsampled to match the resolution of the third view video], wherein the picture of the first view video is the equivalent of the reference frame).
Turning to claim 10, Chen and Furlan teach all of the limitations of claims 1, 2, and 8, as discussed above. Chen further teaches:
wherein the video encoding further comprises: encoding the video frames to be encoded via inter-prediction using the reference frame in a reverse display order from a time instance of the reference frame; and encoding the video frames to be encoded via inter-prediction using the reference frame in a forward display order from a time instance of the reference frame (e.g. Fig. 4 and par. 105: depicting and describing that the system encodes frames using inter-prediction with the reference frame in a reverse display order from a time instance of the reference frame [I-frame at T8 used as a reference frame for inter-prediction of frames T4, T6, and T7, frames at T4, T6, and T7 occurring earlier in time than I-frame at T8], and the system encodes frames using inter-prediction using the reference frame in a forward display order [I-frame at T0 used as reference frame for frames at T1, T2, and T4, frames at T1, T2, and T4 occurring at time instances after the I-frame at T0]).
Regarding claims 51 and 54, Chen and Furlan teach all of the limitations of claims 1 and 11, respectively, as discussed above. Chen further teaches:
wherein the video encoding generates an encoded video wherein all frames of the encoded video are predicted frames (e.g. Figs. 2 and 4, and pars. 71 – 80: depicting and describing that the system generates encoded video using prediction).
Turning to claims 52 and 55, Chen and Furlan teach all of the limitations of claims 1 and 11, respectively, as discussed above. Chen further teaches:
storing the encoded video, wherein all frames of the encoded video are predicted frames (e.g. Fig. 1, par. 44: depicting and describing that the system stores the encoded video, the encoded video being generated using prediction [see, e.g. Fig. 2 and pars. 71 – 80: depicting and describing that the system encodes video data using prediction]).
Regarding claims 53 and 56, Chen and Furlan teach all of the limitations of claims 1 and 11, respectively, as discussed above. Chen does not explicitly teach:
removing from the first video, prior to the video encoding, a video frame that corresponds to time instance when the first still image was captured; wherein the video encoding comprises generating an encoded video based on the first video that has been modified to remove the video frame that corresponds to the time instance when the first still image was captured; and storing the encoded video.
Furlan, however, teaches a method and system:
removing from the first video, prior to the video encoding, a video frame that corresponds to time instance when the first still image was captured; wherein the video encoding comprises generating an encoded video based on the first video that has been modified to remove the video frame that corresponds to the time instance when the first still image was captured; and storing the encoded video (e.g. par. 41: describing that a captured still image takes the place of a frame of a video when captured at the same time as the video [captured still image replaces frames 1 and 21 of the video, the system then encoding only frames 2 – 20 of the captured video], wherein the still image taking the place of a frame of video captured at the same time is the equivalent of removing from the first video prior to encoding, a video frame that corresponds to the time instances when the first still image was captured).
It therefore would have been obvious to one of ordinary skill in the art to modify the teachings of Chen by adding the teachings of Furlan in order to remove from the first video, prior to the video encoding, a video frame that corresponds to time instance when the first still image was captured, wherein the video encoding comprises generating an encoded video based on the first video that has been modified to remove the video frame that corresponds to the time instance when the first still image was captured; and storing the encoded video. One of ordinary skill in the art would have been motivated to make such a modification because the modification allows for both video images and high-resolution still images to be available for playback (Furlan, e.g. par. 34: describing a desire to provide both lower resolution video images and high-resolution still images for display automatically).
Conclusion
Any inquiry concerning this communication or earlier communications from the examiner should be directed to SHANIKA M BRUMFIELD whose telephone number is (571)270-3700. The examiner can normally be reached M-F 8:30 - 5 PM AWS.
Examiner interviews are available via telephone, in-person, and video conferencing using a USPTO supplied web-based collaboration tool. To schedule an interview, applicant is encouraged to use the USPTO Automated Interview Request (AIR) at http://www.uspto.gov/interviewpractice.
If attempts to reach the examiner by telephone are unsuccessful, the examiner’s supervisor, David Czekaj can be reached at 571-272-7327. The fax phone number for the organization where this application or proceeding is assigned is 571-273-8300.
Information regarding the status of published or unpublished applications may be obtained from Patent Center. Unpublished application information in Patent Center is available to registered users. To file and manage patent submissions in Patent Center, visit: https://patentcenter.uspto.gov. Visit https://www.uspto.gov/patents/apply/patent-center for more information about Patent Center and https://www.uspto.gov/patents/docx for information about filing in DOCX format. For additional questions, contact the Electronic Business Center (EBC) at 866-217-9197 (toll-free). If you would like assistance from a USPTO Customer Service Representative, call 800-786-9199 (IN USA OR CANADA) or 571-272-1000.
SHANIKA M. BRUMFIELD
Examiner
Art Unit 2487
/SHANIKA M BRUMFIELD/Examiner, Art Unit 2487
/Dave Czekaj/Supervisory Patent Examiner, Art Unit 2487