Patent application number | Description | Published |
20120062756 | Method and System for Processing Multiview Videos for View Synthesis Using Skip and Direct Modes - Multiview videos are acquired by overlapping cameras. Side information is used to synthesize multiview videos. A reference picture list is maintained for current frames of the multiview videos, the reference picture indexes temporal reference pictures and spatial reference pictures of the acquired multiview videos and the synthesized reference pictures of the synthesized multiview video. Each current frame of the multiview videos is predicted according to reference pictures indexed by the associated reference picture list with a skip mode and a direct mode, whereby the side information is inferred from the synthesized reference picture. Alternatively, the depth images corresponding to the multiview videos of the input data, and this data are encoded as part of the bitstream depending on a SKIP type. | 03-15-2012 |
20120206442 | Method for Generating Virtual Images of Scenes Using Trellis Structures - An image for a virtual view of a scene is generated based on a set of texture images and a corresponding set of depth images acquired of the scene. A set of candidate depth values associated with each pixel of a selected image is determined. For each candidate depth value, a cost that estimates a synthesis quality of the virtual image is determined. The candidate depth value with a least cost is selected to produce an optimal depth value for the pixel. Then, the virtual image is synthesized based on the optimal depth value of each pixel and the texture images. | 08-16-2012 |
20120206451 | Method for Enhancing Depth Images of Scenes Using Trellis Structures - An image for a virtual view of a scene is generated based on a set of texture images and a corresponding set of depth images acquired of the scene. A set of candidate depths associated with each pixel of a selected image is determined. For each candidate depth, a cost that estimates a synthesis quality of the virtual image is determined. The candidate depth with a least cost is selected to produce an optimal depth for the pixel. Then, the virtual image is synthesized based on the optimal depth of each pixel and the texture images. The method also applies first and second depth enhancement before, and during view synthesis to correct errors or suppress noise due to the estimation or acquisition of the dense depth images and sparse depth features. | 08-16-2012 |
20120269458 | Method for Generating High Resolution Depth Images from Low Resolution Depth Images Using Edge Layers - A method interpolates and filters a depth image with reduced resolution to recover a high resolution depth image using edge information, wherein each depth image includes an array of pixels at locations and wherein each pixel has a depth. The reduced depth image is first up-sampled, interpolating the missing positions by repeating the nearest-neighboring depth value. Next, a moving window is applied to the pixels in the up-sampled depth image. The window covers a set of pixels centred at each pixel. The pixels covered by the window are selected according to their relative offset to the depth edge, and only pixels that are within the same side of the depth edge of the centre pixel are used for the filtering procedure. | 10-25-2012 |
20120314027 | Method and System for Processing Multiview Videos for View Synthesis Using Motion Vector Predictor List - Multiview videos are acquired by overlapping cameras. Side information is used to synthesize multiview videos. A reference picture list is maintained for current frames of the multiview videos, the reference picture indexes temporal reference pictures and spatial reference pictures of the acquired multiview videos and the synthesized reference pictures of the synthesized multiview video. Each current frame of the multiview videos is predicted according to reference pictures indexed by the associated reference picture list with a skip mode and a direct mode, whereby the side information is inferred from the synthesized reference picture. In addition, the skip and merge modes for single view video coding are modified to support multiview video coding by generating a motion vector prediction list by also considering neighboring blocks that are associated with synthesized reference pictures. | 12-13-2012 |
20130162773 | COMPRESSION METHODS AND APPARATUS FOR OCCLUSION DATA - Methods and apparatuses for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data , the original format selected from one of a sparse occlusion data format and a filled occlusion data format; decoding the received occlusion data to produce decoded occlusion data ; and when the indicator indicates the original format as a filled occlusion data format, converting the decoded occlusion data from a sparse occlusion data format to the filled occlusion data format , the converting further including; replacing non-occlusion area data, which is represented with a defined characteristic, by respective collocated samples from 2D data in the video data frame associated with the occlusion data ; outputting the decoded occlusion data and, when present, converted decoded occlusion data . | 06-27-2013 |
20130162774 | COMPRESSION METHODS AND APPARATUS FOR OCCLUSION DATA - Methods and apparatuses for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data, the original format selected from a one of a sparse occlusion data format and a filled occlusion data format; arranging 2D data, which is associated with said occlusion data, at location after temporal and inter-view pictures in a reference picture list; identifying at least one of an occlusion area macroblock and a non-occlusion area macroblock for the occlusion data; decoding said occlusion data to produce decoded occlusion data, wherein said decoding includes: for each non-occlusion macroblock, when said indicator indicates the filled occlusion data format, replacing the occlusion data in said non-occlusion macroblock with a corresponding macroblock of associated 2D data to produce a decoded occlusion data; and when said indicator indicates the sparse occlusion data format, filling said non-occlusion macroblock with data indicative of a defined characteristic to produce decoded occlusion data; and otherwise for each occlusion macroblock, decoding said occlusion macroblock to produce decoded occlusion data; and outputting the decoded occlusion data. | 06-27-2013 |
20130176394 | COMPRESSION METHODS AND APPARATUS FOR OCCLUSION DATA - Methods and apparatus for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data, the original format selected from one of a sparse occlusion data format and a filled occlusion data format; arranging 2D data, which is associated with the occlusion data, at location | 07-11-2013 |
20130194511 | REMOTE CONTROL DEVICE FOR 3D VIDEO SYSTEM - A remote control device is operative to enable and facilitate user control of video systems that are operative to provide one or more three-dimensional (3D) viewing effects. According to an exemplary embodiment, the remote control device includes a user input terminal having an input element operative to receive user inputs to adjust at least one of a volume setting and a channel setting of a video system, and further operative to receive user inputs to adjust a three-dimensional (3D) viewing effect of the video system. A transmitter is operative to transmit control signals to the video system in response to the user inputs. | 08-01-2013 |
20130201177 | Method for Modeling and Estimating Rendering Errors in Virtual Images - A quality of a virtual image for a synthetic viewpoint in a 3D scene is determined. The 3D scene is acquired by texture images, and each texture image is associated with a depth image acquired by a camera arranged at a real viewpoint. A texture noise power is based on the acquired texture images and reconstructed texture images corresponding to a virtual texture image. A depth noise power is based on the depth images and reconstructed depth images corresponding to a virtual depth image. The quality of the virtual image is based on a combination of the texture noise power and the depth noise power, and the virtual image is rendered from the reconstructed texture images and the reconstructed depth images. | 08-08-2013 |
20130202194 | Method for generating high resolution depth images from low resolution depth images using edge information - A method interpolates and filters a depth image with reduced resolution to recover a high resolution depth image using edge information, wherein each depth image includes an array of pixels at locations and wherein each pixel has a depth. The reduced depth image is first up-sampled, interpolating the missing positions by repeating the nearest-neighboring depth value. Next, a moving window is applied to the pixels in the up-sampled depth image. The window covers a set of pixels centred at each pixel. The pixels covered by the window are selected according to their relative position to the edge, and only pixels that are within the same side of the edge of the centre pixel are used for the filtering procedure. A single representative depth from the set of selected pixel in the window is assigned to the pixel to produce a processed depth image. | 08-08-2013 |
20130287289 | Synthetic Reference Picture Generation - A synthetic image block in a synthetic picture is generated for a viewpoint based on a texture image and a depth image. A subset of samples from the texture image are warped to the synthetic image block. Disoccluded samples are marked, and the disoccluded samples in the synthetic image block are filled based on samples in a constrained area. The method and system enables both picture level and block level processing for synthetic reference picture generation. The method can be used for power limited devices, and can also refine the synthetic reference picture quality at a block level to achieve coding gains. | 10-31-2013 |
20140092208 | Method and System for Backward 3D-View Synthesis Prediction using Neighboring Blocks - Videos of a scene are processed for view synthesis. The videos are acquired by corresponding cameras arranged so that a view of each camera overlaps with the view of at least one other camera. For each current block, motion or disparity vector is obtained from neighboring blocks. A depth block is based on a corresponding reference depth image and the motion or disparity vector. A prediction block is generated based on the depth block using backward warping. Then, predictive coding for the current block using the prediction block. | 04-03-2014 |
20140092210 | Method and System for Motion Field Backward Warping Using Neighboring Blocks in Videos - Videos of a scene are processed for view synthesis. The videos are acquired by corresponding cameras arranged so that a view of each camera overlaps with the view of at least one other camera. For each current block, motion or disparity vector is obtained from neighboring blocks. A depth block is based on a corresponding reference depth image and the motion or disparity vector. A prediction block is generated based on the depth block using backward warping of a motion field. Then, predictive coding for the current block using the prediction block. Backward mapping can also be performed in the spatial domain. | 04-03-2014 |
20140147031 | Disparity Estimation for Misaligned Stereo Image Pairs - A disparity vector for a pixel in a right image corresponding to a pixel in a left image in a pair of stereo images is determined. The disparity vector is based on a horizontal disparity and a vertical disparity and the pair of stereo images is unrectified. First, a set of candidate horizontal disparities is determined. For each candidate horizontal disparity, a cost associated with a particular horizontal disparity and corresponding vertical disparities is determined. The vertical disparity associated with a first optimal cost is assigned to each candidate horizontal disparity, so that the candidate horizontal disparity and the vertical disparity yield a candidate disparity vector. Lastly, the candidate disparity vector with a second optimal cost is selected as the disparity vector of the pixel in the right image. | 05-29-2014 |
20140219330 | Method and System for Encoding Collections of Images and Videos - An input segment of an input video is encoded by first extracting and storing, for each segment of previously encoded videos, a set of reference features. The set of input features are matched with each set of the reference features to produce a set of scores. The reference segments having largest scores are selected to produce a first reduced set of reference segments. A rate-distortion cost for each reference segment in the first reduced set of reference segments is estimated. The reference segments in the first reduced set of reference segments is selected to produce a second reduced set of reference segments. Then, the input segment are encoded based on second reduced set of reference segments. | 08-07-2014 |
20140301479 | TILING IN VIDEO ENCODING AND DECODING - Implementations are provided that relate, for example, to view tiling in video encoding and decoding. A particular method includes accessing a video picture that includes multiple pictures combined into a single picture ( | 10-09-2014 |