Patent application number | Description | Published |
20100020871 | Method and Device for Video Coding and Decoding - Embodiments of the present invention relate to video coding for multi-view video content. It provides a coding system enabling scalability for the multi-view video content. In one embodiment, a method is provided for encoding at least two views representative of a video scene, each of the at least two views being encoded in at least two scalable layers, wherein one of the at least two scalable layers representative of one view of the at least two views is encoded with respect to a scalable layer representative of the other view of the at least two views. | 01-28-2010 |
20100189173 | METHOD AND APPARATUS FOR VIDEO CODING AND DECODING - A method comprises encoding a first view component of a first view of a multiview bitstream; and encoding a second view component of a second view; wherein the encoding of the second view component enables generating of a reference picture list for the second view component to include at least one of the following: (a) a first field view component based on the first view component or (b) a first complementary field view component pair including the first view component. | 07-29-2010 |
20100309286 | ENCODING OF THREE-DIMENSIONAL CONVERSION INFORMATION WITH TWO-DIMENSIONAL VIDEO SEQUENCE - This disclosure describes techniques for encoding a two-dimensional (2D) video sequence of video frames along with three-dimensional (3D) conversion information comprising a set of parameters that can be applied to each of the video frames of the 2D sequence to generate 3D video data. The set of parameters may comprise a relatively small amount of data that can be applied to each of the original video frames of the 2D sequence to generate secondary views of each of the original video frames. The original video frames and the secondary views may collectively define a stereoscopic 3D video sequence. The 2D sequence and the set of parameters may comprise significantly less data than would otherwise be needed to communicate a 3D sequence. This disclosure also describes some exemplary syntax that may be used to encode the set of parameters in an effective and efficient manner. | 12-09-2010 |
20100316122 | MULTIVIEW VIDEO CODING OVER MPEG-2 SYSTEMS - A multiplexer may produce an MPEG-2 (Motion Picture Experts Group) System standard bitstream comprising views with nonconsecutive view order indexes. In one example, an apparatus comprises a video encoder that encodes a plurality of views of a scene, a multiplexer that constructs a data structure for signaling that a corresponding MPEG-2 (Motion Picture Experts Group) System standard bitstream comprises a first view of the plurality of views of the scene associated with a first view order index and a second view of the plurality of views of the scene associated with a second view order index, wherein the first view order index and the second view order index are non-consecutive, and an output interface that outputs the data structure. | 12-16-2010 |
20100316134 | ASSEMBLING MULTIVIEW VIDEO CODING SUB-BISTREAMS IN MPEG-2 SYSTEMS - A demultiplexer may assemble view components of sub-bitstreams. In one example, an apparatus comprises a demultiplexer that produces a multiview video coding (MVC) standard compliant bitstream from a received bitstream comprising a primary sub-bitstream and an embedded sub-bitstream. To produce the MVC standard compliant bitstream, the demultiplexer determines whether a view component of the primary sub-bitstream has a view order index that is greater than a view order index of a view component of the embedded sub-bitstream, and to add the view component from the sub-bitstream for which the view order index is lower to the produced bitstream. The received bitstream may comprise delimiter network abstraction layer (NAL) units between each view component to differentiate the view components. The apparatus may further comprise a video decoder to decode the bitstream produced by the demultiplexer. | 12-16-2010 |
20110032999 | SIGNALING CHARACTERISTICS OF AN MVC OPERATION POINT - Source and destination video devices may use data structures that signal details of an operation point for an MPEG-2 (Motion Picture Experts Group) System bitstream. In one example, an apparatus includes a multiplexer that constructs a data structure corresponding to a multiview video coding (MVC) operation point of an MPEG-2 (Motion Picture Experts Group) System standard bitstream, wherein the data structure signals a rendering capability value that describes a rendering capability to be satisfied by a receiving device to use the MVC operation point, a decoding capability value that describes a decoding capability to be satisfied by the receiving device to use the MVC operation point, and a bitrate value that describes a bitrate of the MVC operation point, and that includes the data structure as part of the bitstream, and an output interface that outputs the bitstream comprising the data structure. | 02-10-2011 |
20110047374 | METHOD AND APPARATUS FOR A CONFIGURABLE ONLINE PUBLIC KEY INFRASTRUCTURE (PKI) MANAGEMENT SYSTEM - A method and apparatus are provided for generating identity data to be provisioned in product devices that are a part of a project. The method includes establishing a template associated with each CA in a hierarchical chain of CAs having a root CA at a highest level in the chain and a signing CA at a lowest level in the chain. The template associated with the signing CA inherits mandatory attribute fields specified in the root CA and any intermediate CA in the hierarchical chain. The mandatory attribute fields are user-specifiable fields to be populated with PKI data. A configuration file is generated upon receipt of an order for digital certificates using PKI data provided by a user to populate the mandatory attribute fields of the template associated with the signing CA. The digital certificates requested in the order are generated using the PKI data in the configuration file. | 02-24-2011 |
20110058021 | RENDERING MULTIVIEW CONTENT IN A 3D VIDEO SYSTEM - There is disclosed methods and apparatuses for multi-view video encoding, decoding and display. A depth map is provided for each of the available views. The depth maps of the available views are used to synthesize a target view for rendering an image from the perspective of the target view based on images of the available views. | 03-10-2011 |
20110064146 | MEDIA EXTRACTOR TRACKS FOR FILE FORMAT TRACK SELECTION - A video coding apparatus may be configured to utilize media extractors in a media extractor track that reference two or more non-consecutive network access layer (NAL) units of a separate track. An example apparatus includes a multiplexer to construct a first track including a video sample comprising NAL units, based on encoded video data, wherein the video sample is included in an access unit, construct a second track including an extractor that identifies at least first one of the NAL units in the video sample of the first track, and wherein the extractor identifies a second NAL unit of the access unit, wherein the first identified NAL unit and the second identified NAL unit are non-consecutive, and include the first track and the second track in a video file conforming at least in part to ISO base media file format. The identified NAL units may be in separate tracks. | 03-17-2011 |
20110096828 | ENHANCED BLOCK-REQUEST STREAMING USING SCALABLE ENCODING - A block-request streaming system provides for improvements in the user experience and bandwidth efficiency of such systems, typically using an ingestion system that generates data in a form to be served by a conventional file server (HTTP, FTP, or the like), wherein the ingestion system intakes content and prepares it as files or data elements to be served by the file server. A client device can be adapted to take advantage of the ingestion process as well as including improvements that make for a better presentation independent of the ingestion process. The files or data elements are organized as blocks that are transmitted and decoded as a unit, and the system is configured to provide and consume scalable blocks such that the quality of the presentation increases as more of the block is downloaded. Encoding and decoding blocks with multiple independent scalability layers can be done as well. | 04-28-2011 |
20110096832 | DEPTH MAP GENERATION TECHNIQUES FOR CONVERSION OF 2D VIDEO DATA TO 3D VIDEO DATA - This disclosure describes techniques for generating depth maps for video units, such as video frames or slices video frames. The techniques may be performed by a video encoder in order to convert two-dimensional (2D) video to three-dimensional (3D) video. The techniques may alternatively be performed by a video decoder in order to convert received 2D video to 3D video. The techniques may use a combination of motion and color considerations in the depth map generation process. | 04-28-2011 |
20110099594 | STREAMING ENCODED VIDEO DATA - A source device may signal characteristics of a media presentation description (MPD) file such that a destination device may select one of a number of presentations corresponding to the MPD file and retrieve one or more video files of the selected presentation. In one example, an apparatus for transporting encoded video data includes a management unit configured to receive encoded video data comprising a number of video segments and forms a presentation comprising a number of video files, each of the video files corresponding to a respective one of the video segments, and a network interface configured to, in response to a request specifying a temporal section of the video data, output at least one of the video files corresponding to the number of video segments of the requested temporal section. A client may request temporally sequential fragments from different ones of the presentations. | 04-28-2011 |
20110154426 | METHOD AND SYSTEM FOR CONTENT DELIVERY - A method and system of content delivery provide availability of at least two versions of content by delivering data for a first version of content, a difference data representing at least one difference between the first version and a second version of content, and metadata derived from two transformation functions that relate the first version and the second version of content respectively to a master version. | 06-23-2011 |
20110170001 | Method of Predicting Motion and Texture Data - The invention relates to a method for generating for at least one block of pixels of a picture of a sequence of progressive pictures at least one motion predictor and at least one texture predictor from motion data, respectively texture data, associated with the pictures of a sequence of low resolution interlaced pictures. | 07-14-2011 |
20110194613 | VIDEO CODING WITH LARGE MACROBLOCKS - A video coder may utilize large macroblocks having more than 16×16 pixels. Syntax for the large macroblocks may define whether a bitstream includes large macroblocks, such as superblocks having 64×64 pixels or bigblocks having 32×32 pixels. The syntax may be included in a slice header or a sequence parameter set. The large macroblocks may also be encoded according to a large macroblock syntax. The bitstream may further include syntax data that indicates a level value based on whether the bitstream includes any of the large macroblocks, for example, as a smallest-sized luminance prediction block. A decoder may use the level value to determine whether the decoder is capable of decoding the bitstream. | 08-11-2011 |
20110196982 | STREAMING ENCODED VIDEO DATA - A source device may signal characteristics of a media presentation description (MPD) file such that a destination device may select one of a number of presentations corresponding to the MPD file and retrieve one or more video files of the selected presentation. In one example, an apparatus for transporting encoded video data includes a management unit configured to receive encoded video data comprising a number of video segments and forms a presentation comprising a number of video files, each of the video files corresponding to a respective one of the video segments, and a network interface configured to, in response to a request specifying a temporal section of the video data, output at least one of the video files corresponding to the number of video segments of the requested temporal section. A client may request temporally sequential fragments from different ones of the presentations. | 08-11-2011 |
20110197061 | CONFIGURABLE ONLINE PUBLIC KEY INFRASTRUCTURE (PKI) MANAGEMENT FRAMEWORK - A method and apparatus is provided for establishing a process for provisioning a digital certificate service delivered by a PKI system. The method includes receiving a request for a digital certificate service and receiving data specifying a project that includes at least one product to be provisioned with a digital certificate. Data specifying an identification of an owner organization of the project and at least one participant organization participating in the project is also received. Attributes with which PKI data to be included in the digital certificates is to comply is received from the owner organization. Based on the received data and attributes, an account is established for each of the organizations associated with the project through which users associated with each of the organizations can respectively request digital certificates for the at least one product in accordance with the attributes received from the owner organization. | 08-11-2011 |
20110216833 | SHARING OF MOTION VECTOR IN 3D VIDEO CODING - Joint coding of depth map video and texture video is provided, where a motion vector for a texture video is predicted from a respective motion vector of a depth map video or vice versa. For scalable video coding, depth map video is coded as a base layer and texture video is coded as an enhancement layer(s). Inter-layer motion prediction predicts motion in texture video from motion in depth map video. With more than one view in a bit stream (for multi view coding), depth map videos are considered monochromatic camera views and are predicted from each other. If joint multi-view video model coding tools are allowed, inter-view motion skip issued to predict motion vectors of texture images from depth map images. Furthermore, scalable multi-view coding is utilized, where interview prediction is applied between views in the same dependency layer, and inter-layer (motion) prediction is applied between layers in the same view. | 09-08-2011 |
20110238789 | ENHANCED BLOCK-REQUEST STREAMING SYSTEM USING SIGNALING OR BLOCK CREATION - A block-request streaming system provides for improvements in the user experience and bandwidth efficiency of such systems, typically using an ingestion system that generates data in a form to be served by a conventional file server (HTTP, FTP, or the like), wherein the ingestion system intakes content and prepares it as files or data elements to be served by the file server. The system might include controlling the sequence, timing and construction of block requests, time based indexing, variable block sizing, optimal block partitioning, control of random access point placement, including across multiple presentation versions, dynamically updating presentation data, and/or efficiently presenting live content and time shifting. | 09-29-2011 |
20110280311 | ONE-STREAM CODING FOR ASYMMETRIC STEREO VIDEO - An asymmetric frame of a coded video bitstream may include a first resolution picture of a left view and a reduced resolution picture of a right view, where the left and right views form a stereo view pair for three-dimensional video playback. In addition, the reduced resolution frame may be predicted relative to a picture of the left view. In one example, an apparatus includes a video encoder configured to encode a first picture of a first view of a scene to produce an encoded picture with a first resolution, encode at least a portion of a second picture of a second view of the scene relative to a reference picture of the first view to produce an encoded picture with a reduced resolution relative to the first resolution, and output the encoded first resolution picture and the encoded reduced resolution picture in a common bitstream. | 11-17-2011 |
20110280316 | FRAME PACKING FOR ASYMMETRIC STEREO VIDEO - An asymmetric frame of a coded video bitstream may include a full resolution picture of a left view and a reduced resolution picture of a right view, where the left and right views form a stereo view pair for three-dimensional video playback. In one example, an apparatus includes a video encoder configured to receive a first picture of a first view of a scene having a first resolution, receive a second picture of a second view of the scene having a reduced resolution relative to the first resolution, form an asymmetric frame comprising the first picture and the second picture, and encode the asymmetric frame. In this manner, decoders of varying capabilities may receive the same bitstream, and the bitstream may consume less bandwidth than one or more bitstreams having full resolution pictures of a stereo view pair. The bitstream may have better quality than a bitstream having subsampled pictures. | 11-17-2011 |
20110304618 | CALCULATING DISPARITY FOR THREE-DIMENSIONAL IMAGES - An apparatus may calculate disparity values for pixels of a two-dimensional image based on depth information for the pixels and generate a second image using the disparity values. The calculation of the disparity value for a pixel may correspond to a linear relationship between the depth of the pixel and a corresponding disparity range. In one example, an apparatus for rendering three-dimensional image data includes a view synthesizing unit configured to calculate disparity values for a plurality of pixels of a first image based on depth information associated with the plurality of pixels and disparity ranges to which the depth information is mapped, wherein the disparity values describe horizontal offsets for corresponding ones of a plurality of pixels for a second image. The apparatus may receive the first image and depth information from a source device. The apparatus may produce the second image using the first image and disparity values. | 12-15-2011 |
20110317760 | SIGNALING VIDEO SAMPLES FOR TRICK MODE VIDEO REPRESENTATIONS - This disclosure describes techniques relevant to HTTP streaming of media data. According to these techniques, a server device may signal a byte range for at least one intra-decodable (I-frame) of a video fragment. According to the techniques of this disclosure, a client device may communicate a request to a server device to retrieve the at least one I-frame based on the signaled byte range, and use the retrieved I-frame to provide a high speed version of a video presentation that includes the at least one I-frame. A high speed version of a video presentation may be a trick mode of the video presentation, such as a fast forward or fast rewind version of the video presentation. | 12-29-2011 |
20110317771 | SIGNALING RANDOM ACCESS POINTS FOR STREAMING VIDEO DATA - This disclosure describes techniques relevant to HTTP streaming of media data. According to these techniques, a server device may signal an open decoding refresh (ODR) random access point (RAP) for a movie segmentation of a movie representation. At least one frame of the media segmentation following the ODR RAP frame in decoding order may not be correctly decoded, and wherein each frame of the media segmentation following the ODR RAP frame in display order can be correctly decoded without relying on content of frames prior to the ODR RAP in display order. According to the techniques of this disclosure, a client device may communicate a request to a server device for the streaming of media data based on signaling of the ODR RAP. Also according to the techniques of this disclosure, a client device may commence decoding and/or playback of the movie representation based on signaling of the ODR RAP. | 12-29-2011 |
20120013746 | SIGNALING DATA FOR MULTIPLEXING VIDEO COMPONENTS - A server may provide information describing characteristics of audio and video components to a client, separately from encoded samples of the audio and video components themselves. The client may use the information to select components, and then to request the selected components, e.g., in accordance with a streaming network protocol. In one example, an apparatus for sending encapsulated video data includes a processor configured to determine characteristics for components of a plurality of representations of video content, wherein the characteristics comprise at least one of a frame rate, a profile indicator, a level indicator, and dependencies between the components, and one or more interfaces configured to send the characteristics to a client device, receive a request from the client device for at least one of the components after sending the characteristics, and send the requested components to the client device in response to the request. | 01-19-2012 |
20120016965 | VIDEO SWITCHING FOR STREAMING VIDEO DATA - In one example, an apparatus includes a processor configured to provide information to a client device for accessing data for a first representation of a video sequence and a second representation of the video sequence, and to provide information to the client device indicating that the client device can, during transmission of the data for the first representation, switch to the second representation without experiencing a presentation gap when displaying the video sequence and without simultaneously executing more than one decoder to decode the data for the first representation and the second representation during the switch, in response to a first request from the client device to retrieve data from the first representation. In this manner, the client device may use the information to perform seamless switching between the representations. | 01-19-2012 |
20120020413 | PROVIDING FRAME PACKING TYPE INFORMATION FOR VIDEO CODING - This disclosure includes techniques for signaling characteristics of a representation of multimedia content at a representation layer, such as frame packing arrangement information for the representation. In one example, an apparatus for receiving video data includes a processing unit configured to receive information indicative of whether a bitstream includes a packed frame of video data, wherein the packed frame comprises two frames corresponding to different views of a scene for three-dimensional playback, and wherein the information is present in a representation layer external to a codec layer of the bitstream, automatically determine whether the apparatus is capable of decoding and rendering the bitstream based on an analysis of the received information and decoding and rendering capabilities of the device, and retrieve the bitstream when the processing unit determines that the device is capable of decoding and rendering the bitstream. | 01-26-2012 |
20120023249 | PROVIDING SEQUENCE DATA SETS FOR STREAMING VIDEO DATA - A device may encapsulate video data such that Supplemental Enhancement Information (SEI) messages are stored separately from a sequence of coded video pictures described by the SEI messages. An example device includes a control unit configured to generate one or more SEI messages separate from the coded video pictures, wherein the SEI messages describe respective ones of the sequence of coded video pictures and include elements common to more than one of the coded video pictures, and an output interface configured to output the SEI messages separately from the sequence of coded video pictures. An example destination device may receive the SEI messages separately from the coded video pictures and render the coded video pictures using the SEI messages. | 01-26-2012 |
20120023250 | ARRANGING SUB-TRACK FRAGMENTS FOR STREAMING VIDEO DATA - A video file may include movie fragments divided into sub-track fragments that store all pictures of common hierarchical levels for the respective hierarchical levels. In one example, an apparatus includes an interface configured to output data according to a streaming protocol, and a control unit configured to assemble encoded video data into a plurality of sub-track fragments, each of the sub-track fragments comprising a plurality of hierarchically related video pictures of the encoded video data, wherein the plurality of hierarchically related video pictures each correspond to a common hierarchical layer receiving a request in accordance with the streaming protocol, wherein the request specifies at least one of the plurality of sub-track fragments, and, in response to the request, cause the interface to output the plurality of hierarchically related video pictures of the at least one of the plurality of sub-track fragments. | 01-26-2012 |
20120027088 | CODING MOTION PREDICTION DIRECTION IN VIDEO CODING - This disclosure relates to techniques for reducing a cost of coding prediction information in video coding. Video blocks in a generalized P/B (GPB) frame are encoded using up to two motion vectors calculated from reference pictures in two separate reference picture lists that are identical. When one of the reference picture lists is preferred over the other reference picture list, the preferred reference picture list may be used for unidirectional prediction, by default. When a GPB frame is enabled such that the first and second reference picture lists are identical, either of the first and second reference picture lists may be used for unidirectional prediction. The techniques include coding one or more syntax elements indicating that a video block is coded using one of the unidirectional prediction mode with respect to a reference picture in a reference picture list and the bidirectional prediction mode using less than two bits. | 02-02-2012 |
20120033037 | Signaling Attributes for Network-Streamed Video Data - In one example, an apparatus includes a processor configured to receive video data for two or more views of a scene, determine horizontal locations of camera perspectives for each of the two or more views, assign view identifiers to the two or more views such that the view identifiers correspond to the relative horizontal locations of the camera perspectives, form a representation comprising a subset of the two or more views, and, in response to a request from a client device, send information indicative of a maximum view identifier and a minimum view identifier for the representation to the client device. | 02-09-2012 |
20120036544 | Signaling Attributes for Network-Streamed Video Data - In one example, an apparatus for signaling information for video data includes a processor configured to receive video data for two or more views of a scene, form a representation comprising a subset of the two or more views, and send, to a client device, as a part of a manifest of the representation, information indicative of a maximum number of views in the representation that can be targeted for output. An apparatus for receiving information for video data may receive the manifest including the information indicating the maximum number of views and request at least a portion of the video data of the representation based at least in part on a maximum number of views that can be output by the apparatus and the information indicative of the maximum number of views in the representation that can be targeted for output. | 02-09-2012 |
20120042050 | Representation groups for network streaming of coded multimedia data - In one example, a device for receiving information for multimedia data, the device comprising one or more processors configured to analyze at least a portion of a manifest file for multimedia content, wherein the portion of the manifest file includes information indicative of sets of representations of the multimedia content and information indicative of common characteristics for each of the sets of representations, select one of the sets of representations based on the common characteristics for the one of the sets of representations, select one of the representations of the selected one of the sets of representations based on one or more coding characteristics of the one of the representations of the one of the sets, and generate a request for data of the one of the representations based on the selection. | 02-16-2012 |
20120042089 | TRICK MODES FOR NETWORK STREAMING OF CODED MULTIMEDIA DATA - In one example, a device for retrieving multimedia data, the device comprising one or more processors configured to analyze information of a manifest file for multimedia content, wherein the information of the manifest file indicates that at least one representation of the multimedia content includes a temporal sub-sequence, determine one or more locations of data for the temporal sub-sequence, and submit one or more requests for the data for the temporal sub-sequence. | 02-16-2012 |
20120042090 | MANIFEST FILE UPDATES FOR NETWORK STREAMING OF CODED MULTIMEDIA DATA - In one example, a device for retrieving multimedia data, the device comprising one or more processors configured to retrieve data of a first segment of a representation of multimedia content in accordance with data of a copy of a manifest file stored by the device, retrieve a portion of a second segment of the representation in accordance with the manifest file, wherein the second segment occurs after the first segment in the representation, and wherein the portion of the second segment indicates that the manifest file is to be updated, update the copy of the manifest file stored by the device based on the indication that the manifest file is to be updated, and retrieve media data of the second segment in accordance with the updated manifest file. | 02-16-2012 |
20120075436 | CODING STEREO VIDEO DATA - In one example, a method of decoding video data comprising base layer data having a first resolution and enhancement layer data having the first resolution includes decoding the base layer data, wherein the base layer data comprises a reduced resolution version of a left view relative to the first resolution and a reduced resolution version of a right view relative to the first resolution. The method also includes decoding enhancement layer data comprising enhancement data for exactly one of the left view and the right view, wherein the enhancement data has the first resolution, and wherein decoding the enhancement layer data comprises decoding the enhancement layer data relative to at least a portion of the base layer data. | 03-29-2012 |
20120121017 | REFERENCE PICTURE LIST CONSTRUCTION FOR GENERALIZED P/B FRAMES IN VIDEO CODING - This disclosure relates to techniques for indicating that a video frame is coded as a generalized P/B (GPB) frame in order to reduce a cost of coding and constructing a second reference picture list in video coding. For a GPB frame, which has identical reference picture lists, signaling and constructing both the first and second reference picture lists may be redundant. The techniques of this disclosure may include coding syntax elements indicating that the video frame is coded as the GPB frame, and coding syntax elements indicating a number of reference pictures and reference picture list construction commands for only a first reference picture list. The techniques also include constructing the first reference picture list based on the syntax elements, and then creating the second reference picture list as a duplicate of the first reference picture list. | 05-17-2012 |
20120127267 | DEPTH ESTIMATION BASED ON GLOBAL MOTION - This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, an initial indication of depth (e.g., an optical flow) may be determined for a 2D view. The initial indication of depth may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. The initial indication of depth may be modified based on the estimation of global motion to create a global motion-adjusted indication of depth. The global motion-adjusted depth indication may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation. | 05-24-2012 |
20120127270 | DEPTH ESTIMATION BASED ON GLOBAL MOTION - This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, a plurality of feature points may be determined for a 2D view. The plurality of feature points may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. For example, the plurality of feature points may be used to generate a global motion frame difference. The global motion frame difference may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation. | 05-24-2012 |
20120170648 | FRAME SPLITTING IN VIDEO CODING - In one example, this disclosure describes a method of decoding a frame of video data comprising a plurality of block-sized coding units including one or more largest coding units (LCUs) that include a hierarchically arranged plurality of relatively smaller coding units. In this example, the method includes determining a granularity at which the hierarchically arranged plurality of smaller coding units has been split when forming independently decodable portions of the frame. The method also includes identifying an LCU that has been split into a first section and a second section using the determined granularity. The method also includes decoding an independently decodable portion of the frame that includes the first section of the LCU without the second section of the LCU. | 07-05-2012 |
20120189053 | COMBINED REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - This disclosure relates to techniques for constructing a combined reference picture list, List C, based on List | 07-26-2012 |
20120189058 | SINGLE REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - The example techniques described in this disclosure provide for an efficient manner to encode or decode a video block of a picture using a single reference picture list. The single reference picture list may include identifiers for reference picture or pictures used to encode or decode the video block. In some examples, a video encoder or decoder may encode or decode a video block that is predicted from two reference pictures using the single reference picture list, and encode or decode a video block that is predicted from one reference picture using the same, single reference picture list. | 07-26-2012 |
20120229602 | CODING MULTIVIEW VIDEO PLUS DEPTH CONTENT - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to receive one or more texture components from at least a portion of an image representing a view of three dimensional video data, receive a depth map component for at least the portion of the image, code a block unit indicative of pixels of the one or more texture components for a portion of the image and the depth map component. The coding comprises receiving texture data for a temporal instance of a view of video data, receiving depth data corresponding to the texture data for the temporal instance of the view of video data, and encapsulating the texture data and the depth data in a view component for the temporal instance of the view, such that the texture data and the depth data are encapsulated within a common bitstream. | 09-13-2012 |
20120230401 | BUFFER MANAGEMENT IN VIDEO CODECS - In one example, a video decoder is configured to determine, from data associated with an open decoding refresh (ODR) picture of video data, an identifier for a previous picture of the video data, determine whether the previous picture is currently stored in a reference picture memory, and decode only data for pictures of the video data having display order values that are greater than a display order value of the ODR picture, when the previous picture is not stored in the reference picture memory. In another example, a video encoder is configured to encode an open decoding refresh (ODR) picture, determine a previously coded picture having a display order value less than a display order value of the ODR picture and having a temporal identifier value equal to zero, and signal syntax data for the ODR picture representative of an identifier of the determined previously coded picture. | 09-13-2012 |
20120230409 | DECODED PICTURE BUFFER MANAGEMENT - The example techniques described in this disclosure are generally related to decoded picture buffer management. One or more pictures stored in the decoded picture buffer may be usable for prediction, and others may not. Pictures that are usable for prediction may be referred to as reference pictures. The example techniques described herein may determine whether a reference picture, that is currently indicated to be usable for inter-prediction, should be indicated to be unusable for inter-prediction. | 09-13-2012 |
20120230421 | TRANSFORMS IN VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes determining a first residual quadtree (RQT) depth at which to apply a first transform to luma information associated with a block of video data, wherein the RQT represents a manner in which transforms are applied to luma information and chroma information. The method also includes determining a second RQT depth at which to apply a second transform to the chroma information associated with the block of video data, wherein the second RQT depth is different than the first RQT depth. The method also includes coding the luma information at the first RQT depth and the chroma information at the second RQT depth. | 09-13-2012 |
20120230433 | VIDEO CODING TECHNIQUES FOR CODING DEPENDENT PICTURES AFTER RANDOM ACCESS - In general, this disclosure describes techniques for coding video data for random access. In particular, this disclosure proposes to code a syntax element that indicates if a dependent picture may be successfully decoded in the event of a random access request to a clean decoding refresh (CDR) picture and may be required for decoding the pictures following the clean decoding refresh (CDR) picture in display order. | 09-13-2012 |
20120236115 | POST-FILTERING IN FULL RESOLUTION FRAME-COMPATIBLE STEREOSCOPIC VIDEO CODING - Stereoscopic video data encoded according to a full resolution frame-compatible stereoscopic vide coding process. Such stereoscopic video data consists of a right view and a left that are encoded in half resolution versions in an interleaved base layer and an interleaved enhancement layer. When decoded, the right view and left view are filtered according to two sets of filter coefficients, one set for the left view and one set for the right view. The sets of filter coefficients are generated by an encoder by comparing the original left and right views to decoded versions of the left and right views. | 09-20-2012 |
20120236934 | SIGNALING OF MULTIVIEW VIDEO PLUS DEPTH CONTENT WITH A BLOCK-LEVEL 4-COMPONENT STRUCTURE - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to receive one or more texture components from at least a portion of an image representing a view of three dimensional video data, receive a depth map component for at least the portion of the image, code a block unit indicative of pixels of the one or more texture components for a portion of the image and the depth map component. The coding comprises coding the depth map component relative to at least one of the texture components, and signalling an attribute of the depth map component relative to the one or more texture components. | 09-20-2012 |
20120250773 | COMBINED REFERENCE PICTURE LIST CONSTRUCTION AND MAPPING - The example techniques of this disclosure are directed to default construction techniques for the construction of a combined reference picture list, and default mapping techniques for the combined reference picture list. In some examples, a video coder may construct first and second reference picture lists from frame number values, and construct the combined reference picture list from the frame number values of the first and second reference picture lists. In some examples, a video coder may construct first and second reference picture lists from picture order count (POC) values, and construct the combined reference picture list from the POC values of the first and second reference picture lists. In some examples, a video coder may construct a combined reference picture list from received information for the construction, and map the pictures of the combined reference picture list to one of a first or second reference picture list. | 10-04-2012 |
20120262542 | DEVICES AND METHODS FOR WARPING AND HOLE FILLING DURING VIEW SYNTHESIS - Implementations include methods and systems for a converting reference images or video to 3D images or video. A two-step conversion is described which accomplishes warping and hole filling on a pixel-by-pixel basis. In one implementation, of a plurality of pixel values of a reference image at a plurality of first collinear pixels locations are successively mapped to a respective plurality of second pixel locations of a destination image. Between two of the mappings, a location of a hole between two of the second pixel locations may be identified and filled. | 10-18-2012 |
20120269270 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to, in an example, a method that includes identifying a first block of video data in a first temporal location from a first view, wherein the first block is associated with a first disparity motion vector. The method also includes determining a motion vector predictor for a second motion vector associated with a second block of video data, wherein the motion vector predictor is based on the first disparity motion vector. When the second motion vector comprises a disparity motion vector, the method includes determining the motion vector predictor comprises scaling the first disparity motion vector to generate a scaled motion vector predictor, wherein scaling the first disparity motion vector comprises applying a scaling factor comprising a view distance of the second disparity motion vector divided by a view distance of the first motion vector to the first disparity motion vector. | 10-25-2012 |
20120269271 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes identifying a first block of video data in a first temporal location from a first view, wherein the first block of video data is associated with a first temporal motion vector. The method also includes determining, when a second motion vector associated with a second block of video data comprises a temporal motion vector and the second block is from a second view, a motion vector predictor for the second motion vector based on the first temporal motion vector. The method also includes coding prediction data for the second block using the motion vector predictor. | 10-25-2012 |
20120314026 | INTERNAL BIT DEPTH INCREASE IN VIDEO CODING - In an example aspects of this disclosure generally relate to a method of coding video data that includes determining a first bit depth for outputting video data and a second bit depth for coding the video data, wherein the first bit depth is less than the second bit depth. The method also includes determining whether the video data will be used as reference data when coding other video data. The method also includes storing, based on the determination, the video data at the first bit depth when the video data is not used as reference data, and the video data at the second bit depth when the video data is used as reference data. | 12-13-2012 |
20120314767 | BORDER PIXEL PADDING FOR INTRA PREDICTION IN VIDEO CODING - A video coder performs a padding operation that processes a set of border pixels according to an order. The order starts at a bottom-left border pixel and proceeds through the border pixels sequentially to a top-right border pixel. When the padding operation processes an unavailable border pixel, the padding operation predicts a value of the unavailable border pixel based on a value of a border pixel previously processed by the padding operation. The video coder may generate an intra-predicted video block based on the border pixels. | 12-13-2012 |
20130007223 | ENHANCED BLOCK-REQUEST STREAMING SYSTEM FOR HANDLING LOW-LATENCY STREAMING - A block-request streaming system provides for low-latency streaming of a media presentation. A plurality of media segments are generated according to an encoding protocol. Each media segment includes a random access point. A plurality of media fragments are encoded according to the same protocol. The media segments are aggregated from a plurality of media fragments. | 01-03-2013 |
20130016769 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 01-17-2013 |
20130022104 | ADAPTATION PARAMETER SETS FOR VIDEO CODING - In one example, a video decoder is configured to store an adaptation parameter set (APS) data structure in a buffer of a coding device, wherein the APS data structure includes signaling data applicable to one or more slices of video data, decode at least one of the slices based on the signaling data of the APS data structure, after decoding the slice, determine that the APS data structure can be removed from the buffer, remove the APS data structure from the buffer based on the determination, and decode one or more additional slices of the video data after removing the APS data structure from the buffer. | 01-24-2013 |
20130022111 | CODING MOTION DEPTH MAPS WITH DEPTH RANGE VARIATION - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to determine a first real-world depth range for a first depth view component comprising a reference view component, determine a second real-world depth range for a second depth view component comprising a current view component, wherein the current view component is predicted relative to the reference view component, determine a predictive block for a portion of the current view component from the reference view component, adjust values of the predictive block based on a difference between the first real-world depth range and the second real-world depth range, and predict the portion of the current view based on the adjusted values of the predictive block. | 01-24-2013 |
20130022113 | SLICE HEADER PREDICTION FOR DEPTH MAPS IN THREE-DIMENSIONAL VIDEO CODECS - In one example, a video coder is configured to code a first slice, wherein the first slice comprises one of a texture slice and a corresponding depth slice, wherein the first slice has a slice header comprising complete syntax elements representative of characteristics of the first slice. The video coder is further configured to determine common syntax elements for a second slice from the slice header of the first slice. The video coder is also configured to code the second slice after coding the first slice at least partially based on the determined common syntax elements, wherein the second slice comprises one of the texture slice and the depth slice that is not the first slice, wherein the second slice has a slice header comprising syntax elements representative of characteristics of the second slice, excluding values for syntax elements that are common to the first slice. | 01-24-2013 |
20130034170 | CODING PARAMETER SETS FOR VARIOUS DIMENSIONS IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code, for a bitstream, information representative of which of a plurality of video coding dimensions are enabled for the bitstream, and code values for each of the enabled video coding dimensions, without coding values for the video coding dimensions that are not enabled, in a network abstraction layer (NAL) unit header of a NAL unit comprising video data coded according to the values for each of the enabled video coding dimensions. In this manner, NAL unit headers may have variable lengths, while still providing information for scalable dimensions to which the NAL units correspond. | 02-07-2013 |
20130038686 | THREE-DIMENSIONAL VIDEO WITH ASYMMETRIC SPATIAL RESOLUTION - A video coding device may be configured to code a bitstream including multiple views plus depth information. Two of the views may have reduced resolutions, while a third view may have a full resolution. The third view may be predicted relative to upsampled versions of the two reduced-resolution views. Each view may include texture data and depth data, such that a view component may include a texture component and a depth component. Moreover, the texture and depth components may be arranged within an access unit according to a particular order, which may simplify component extraction from the access unit. | 02-14-2013 |
20130057646 | SLICE HEADER THREE-DIMENSIONAL VIDEO EXTENSION FOR SLICE HEADER PREDICTION - In one example, a video coder is configured to code one or more blocks of video data representative of texture information of at least a portion of a frame of video data, process a texture slice for a texture view component of a current view associated, the texture slice comprising the coded one or more blocks and a texture slice header comprising a set of syntax elements representative of characteristics of the texture slice, code depth information representative of depth values for at least the portion of the frame, and process a depth slice for a depth view component corresponding to the texture view component of the view, the depth slice comprising the coded depth information and a depth slice header comprising a set of syntax elements representative of characteristics of the depth slice, wherein process the texture slice or the depth slice comprises predict at least one syntax element. | 03-07-2013 |
20130077677 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to constructing reference picture lists. The reference picture lists may be constructed from reference picture subsets of a reference picture set. In some examples, the reference picture subsets may be ordered in a particular manner to form the reference picture lists. | 03-28-2013 |
20130077678 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to constructing reference picture lists. The reference picture lists may be constructed from reference picture subsets of a reference picture set. In some examples, the techniques may repeatedly list reference pictures identified in the reference picture subsets until the number of entries in the reference picture list is equal to the maximum number of allowable entries in the reference picture list. | 03-28-2013 |
20130077679 | VIDEO CODING WITH SUBSETS OF A REFERENCE PICTURE SET - Techniques are described related to deriving a reference picture set. A reference picture set may identify reference pictures that can potentially be used to inter-predict a current picture and picture following the current picture in decoding order. In some examples, deriving the reference picture set may include constructing a plurality of reference picture subsets that together form the reference picture set. | 03-28-2013 |
20130077680 | DECODED PICTURE BUFFER MANAGEMENT - Techniques are described related to output and removal of decoded pictures from a decoded picture buffer (DPB). The example techniques may remove a decoded picture from the DPB prior to coding a current picture. For instance, the example techniques may remove the decoded picture if that decoded picture is not identified in the reference picture set of the current picture. | 03-28-2013 |
20130077681 | REFERENCE PICTURE SIGNALING AND DECODED PICTURE BUFFER MANAGEMENT - Techniques are described related to performing random access starting from a random access point picture that is not an instantaneous decoder refresh picture. Some techniques are also related to reducing the amount of information that is signaled for long-term reference pictures of a reference picture set. Additional techniques are also related to decoded picture buffer management, such as removing decoded pictures based on a temporal identification value. | 03-28-2013 |
20130077685 | REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - Techniques are described related to modifying an initial reference picture list. The example techniques may identify a reference picture in at least one of the reference picture subsets used to construct the initial reference picture. The example techniques may list the identified reference picture in a current entry of the initial reference picture list to construct a modified reference picture list. | 03-28-2013 |
20130077687 | CODING REFERENCE PICTURES FOR A REFERENCE PICTURE SET - Techniques are described related to coding of long-term reference pictures for a reference picture set. In some examples, a video coder may code candidate long-term reference pictures in a parameter set. The video coder also code syntax elements that indicate which long-term reference pictures from the candidate long-term reference pictures belong in the reference picture set. | 03-28-2013 |
20130089134 | EFFICIENT SIGNALING OF REFERENCE PICTURE SETS - A video coder can select which reference pictures should be signaled in a parameter set such as a picture parameter set (PPS) and which reference pictures should be signaled in a slice header such that when a video decoder constructs a reference picture set, the video decoder does not need to reorder the reference picture set to construct an initial reference picture list for a slice of video data. | 04-11-2013 |
20130089135 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving a first decoded frame of video data, wherein the first decoded frame is associated with a first resolution, determining whether a decoded picture buffer is available to store the first decoded frame based on the first resolution, and in the event the decoded picture buffer is available to store the first decoded frame, storing the first decoded frame in the decoded picture buffer, and determining whether the decoded picture buffer is available to store a second decoded frame of video data, wherein the second decoded frame is associated with a second resolution, based on the first resolution and the second resolution, wherein the first decoded frame is different than the second decoded frame. | 04-11-2013 |
20130089152 | SIGNALING PICTURE IDENTIFICATION FOR VIDEO CODING - In one example, a video coder, such as a video encoder or video decoder, is configured to determine a number of least significant bits of picture identifying information for a picture of video data, determine a value of the picture identifying information for the picture, and code information indicative of the determined number of least significant bits of the value of the picture identifying information for the picture. | 04-11-2013 |
20130089154 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving first and second sub-sequences of video, wherein the first sub-sequence includes one or more frames each having a first resolution, and the second sub-sequence includes one or more frames each having a second resolution, receiving a first sequence parameter set and a second sequence parameter set for the coded video sequence, wherein the first sequence parameter set indicates the first resolution of the one or more frames of the first sub-sequence, and the second sequence parameter set indicates the second resolution of the one or more frames of the second sub-sequence, and wherein the first sequence parameter set is different than the second sequence parameter set, and using the first sequence parameter set and the second sequence parameter set to decode the coded video sequence. | 04-11-2013 |
20130101035 | GROUPING OF TILES FOR VIDEO CODING - Techniques described herein for coding video data include techniques for coding pictures partitioned into tiles, in which each of the plurality of tiles in a picture is assigned to one of a plurality of tile groups. One example method for coding video data comprising a picture that is partitioned into a plurality tiles comprises coding video data in a bitstream, and coding, in the bitstream, information that indicates one of a plurality of tile groups to which each of the plurality of tiles is assigned. The techniques for grouping tiles described herein may facilitate improved parallel processing for both encoding and decoding of video bitstreams, improved error resilience, and more flexible region of interest (ROI) coding. | 04-25-2013 |
20130107942 | FRAGMENTED PARAMETER SET FOR VIDEO CODING | 05-02-2013 |
20130107953 | RANDOM ACCESS WITH ADVANCED DECODED PICTURE BUFFER (DPB) MANAGEMENT IN VIDEO CODING | 05-02-2013 |
20130114670 | MULTIVIEW VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes obtaining, from an encoded bitstream, one or more network abstraction layer (NAL) units for each view component of a plurality of view components of encoded video data, where each view component of the plurality of view components corresponds to a common temporal location, and where the one or more NAL units encapsulate at least a portion of the encoded video data for the respective view components and include information indicative of a decoding order of the respective view components. The method also includes obtaining information, separate from the NAL units, indicating relationships between view identifiers for the views and the decoding order of the view components. The method also includes decoding the encoded video data of the plurality of view components in the decoding order based on the received information. | 05-09-2013 |
20130114694 | PARAMETER SET GROUPS FOR CODED VIDEO DATA - A video coding device, such as a video encoder or a video decoder, may be configured to code a parameter set group representing a first parameter set of a first type and a second parameter set of a second, different type, and code a slice of video data using information of the parameter set group, information of the first parameter set, and information of the second parameter set, wherein the slice includes information referring to the parameter set group. The video coding device may further code the first and second parameter sets. | 05-09-2013 |
20130114705 | MULTIVIEW VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes obtaining, from an encoded bitstream and for any view component of a first view, reference view information indicating one or more reference views for predicting view components of the first view. The method also includes including, for decoding a first view component in an access unit and in the first view, one or more reference candidates in a reference picture list, where the one or more reference candidates comprise view components in the access unit and in the reference views indicated by the reference view information, where the number of reference candidates is equal to the number of reference views. The method also includes decoding the first view component based on the one or more reference candidates in the reference picture list. | 05-09-2013 |
20130127987 | SIGNALING DEPTH RANGES FOR THREE-DIMENSIONAL VIDEO CODING - In one example, a video coder, such as a video encoder or a video decoder, is configured to code a first set of one or more depth range values for a first set of video data, wherein the first set of one or more depth range values have respective first precisions, code a second set of one or more depth range values for a second set of video data, wherein the second set of one or more depth range values have respective second precisions different than the respective first precisions, and code at least a portion of the second set of video data using the second set of one or more depth range values. In this manner, the video coder may update precisions (e.g., numbers of bits) used to represent depth range values for coding multiview plus depth video data. | 05-23-2013 |
20130128965 | INSIDE VIEW MOTION PREDICTION AMONG TEXTURE AND DEPTH VIEW COMPONENTS - The techniques of this disclosure may be generally related to using motion information for a corresponding block from a texture view component that corresponds with a block in a depth view component in coding the block in the depth view component. In some examples, for coding purposes, the techniques may use motion information when the spatial resolution of the texture view component is different than the spatial resolution of the depth view component. | 05-23-2013 |
20130135431 | SEQUENCE LEVEL INFORMATION FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device for processing the video data comprising one or more processors may perform the techniques. The one or more processors may be configured to determine first sequence level information describing characteristics of the depth components, and determine second sequence level information describing characteristics of an operation point of the video data. | 05-30-2013 |
20130135433 | DEPTH COMPONENT REMOVAL FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device configured to code video data may perform the techniques. The video coding device may comprise a decoded picture buffer and a processor configured to store a depth component in the decoded picture buffer, analyze a view dependency to determine whether the depth component is used for inter-view prediction and remove the depth component from the decoded picture buffer in response to determining that the depth component is not used for inter-view prediction. for processing video data including a view component comprised of a depth component and a texture component | 05-30-2013 |
20130135434 | NESTED SEI MESSAGES FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately processing depth and texture components of video data. A device configured to process video data including a view component comprised of a depth component and a texture component may perform various aspects of the techniques. The device may comprise a processor configured to determine a supplemental enhancement information message that applies when processing the view component of the video data, and determine a nested supplemental enhancement information message that applies in addition to the supplemental enhancement information message when processing the depth component of the view component. | 05-30-2013 |
20130136176 | ACTIVATION OF PARAMETER SETS FOR MULTIVIEW VIDEO CODING (MVC) COMPATIBLE THREE-DIMENSIONAL VIDEO CODING (3DVC) - In general, techniques are described for separately coding depth and texture components of video data. A video coding device for coding video data that includes a view component comprised of a depth component and a texture component may perform the techniques. The video coding device may comprise, as one example, a processor configured to activate a parameter set as a texture parameter set for the texture component of the view component, and code the texture component of the view component based on the activated texture parameter set. | 05-30-2013 |
20130142256 | CODING LEAST SIGNFICANT BITS OF PICTURE ORDER COUNT VALUES IDENTIFYING LONG-TERM REFERENCE PICTURES - In general, techniques are described for coding picture order count values identifying long-term reference pictures. A video decoding device comprising a processor may perform the techniques. The processor may be configured to determine a number of bits used to represent least significant bits of the picture order count value that identifies a long-term reference picture to be used when decoding at least a portion of a current picture and parse the determined number of bits from a bitstream representative of the encoded video data. The parsed bits represent the least significant bits of the picture order count value. The processor retrieves the long-term reference picture from a decoded picture buffer based on the least significant bits, and decodes at least the portion of the current picture using the retrieved long-term reference picture. | 06-06-2013 |
20130142257 | CODING PICTURE ORDER COUNT VALUES IDENTIFYING LONG-TERM REFERENCE FRAMES - In general, techniques are described for coding picture order count values identifying long-term reference pictures. A video decoding device comprising a processor may perform the techniques. The processor may determine least significant bits (LSBs) of a picture order count (POC) value that identifies a long-term reference picture (LTRP). The LSBs do not uniquely identify the POC value with respect to the LSBs of any other POC value identifying any other picture in a decoded picture buffer (DPB). The processor may determine most significant bits (MSBs) of the POC value. The MSBs combined with the LSBs is sufficient to distinguish the POC value from any other POC value that identifies any other picture in the DPB. The processor may retrieve the LTRP from the decoded picture buffer based on the LSBs and MSBs of the POC value, and decode a current picture of the video data using the retrieved LTRP. | 06-06-2013 |
20130148722 | REFERENCE PICTURE LIST MODIFICATION FOR VIEW SYNTHESIS REFERENCE PICTURES - A video encoder generates a bitstream that includes a reference picture list modification (RPLM) command. The RPLM command belongs to a type of RPLM commands for inserting short-term reference pictures into reference picture lists. The RPLM command instructs a video decoder to insert a synthetic reference picture into the reference picture list. The video decoder decodes, based at least in part on syntax elements parsed from the bitstream, one or more view components and generates, based at least in part on the one or more view components, the synthetic reference picture. The video decoder modifies, in response to the RPLM commands, a reference picture list to include the synthetic reference picture. The video decoder may use one or more pictures in the reference picture list as reference pictures to perform inter prediction on one or more video blocks of a picture. | 06-13-2013 |
20130155184 | REFERENCE PICTURE LIST CONSTRUCTION FOR MULTI-VIEW AND THREE-DIMENSIONAL VIDEO CODING - A video encoder generates, based on a reference picture set of a current view component, a reference picture list for the current view component. The reference picture set includes an inter-view reference picture set. The video encoder encodes the current view component based at least in part on one or more reference pictures in the reference picture list. In addition, the video encoder generates a bitstream that includes syntax elements indicating the reference picture set of the current view component. A video decoder parses, from the bitstream, syntax elements indicating the reference picture set of the current view component. The video decoder generates, based on the reference picture set, the reference picture list for the current view component. In addition, the video decoder decodes at least a portion of the current view component based on one or more reference pictures in the reference picture list. | 06-20-2013 |
20130176389 | SIGNALING VIEW SYNTHESIS PREDICTION SUPPORT IN 3D VIDEO CODING - In one example, a video coder is configured to code information indicative of whether view synthesis prediction is enabled for video data. When the information indicates that view synthesis prediction is enabled for the video data, the video coder may generate a view synthesis picture using the video data and code at least a portion of a current picture relative to the view synthesis picture. The at least portion of the current picture may comprise, for example, a block (e.g., a PU, a CU, a macroblock, or a partition of a macroblock), a slice, a tile, a wavefront, or the entirety of the current picture. On the other hand, when the information indicates that view synthesis prediction is not enabled for the video data, the video coder may code the current picture using at least one of intra-prediction, temporal inter-prediction, and inter-view prediction without reference to any view synthesis pictures. | 07-11-2013 |
20130176390 | MULTI-HYPOTHESIS DISPARITY VECTOR CONSTRUCTION IN 3D VIDEO CODING WITH DEPTH - A method and apparatus for decoding and encoding multiview video data is described. An example method may include coding a block of video data using a motion vector prediction process, determining a motion vector candidate list, determining a disparity vector candidate list for the motion prediction process, wherein the disparity vector candidate list includes at least two types of disparity vectors from a plurality of disparity vector types, the plurality including a spatial disparity vector (SDV), a smooth temporal-view (STV) disparity vector, a view disparity vector (VDV), and a temporal disparity vector (TDV), and performing the motion vector prediction process using one of the disparity vector candidate list and the motion vector candidate list. | 07-11-2013 |
20130182755 | CODING PARAMETER SETS AND NAL UNIT HEADERS FOR VIDEO CODING - In one example, a video coder, such as a video encoder or video decoder, is configured to code a video parameter set (VPS) for one or more layers of video data, wherein each of the one or more layers of video data refer to the VPS, and code the one or more layers of video data based at least in part on the VPS. The video coder may code the VPS for video data conforming to High-Efficiency Video Coding, Multiview Video Coding, Scalable Video Coding, or other video coding standards or extensions of video coding standards. The VPS may include data specifying parameters for corresponding sequences of video data within various different layers (e.g., views, quality layers, or the like). The parameters of the VPS may provide indications of how the corresponding video data is coded. | 07-18-2013 |
20130188013 | MVC BASED 3DVC CODEC SUPPORTING INSIDE VIEW MOTION PREDICTION (IVMP) MODE - This disclosure describes features and techniques applicable to three-dimensional (3D) video coding. In one example, a technique may include coding a texture view video block, and coding a depth view video block, wherein the depth view video block is associated with the texture view video block. Coding the depth view video block may include coding a syntax element to indicate whether or not motion information associated with the texture view video block is adopted as motion information associated with the depth view video block. | 07-25-2013 |
20130188717 | MOTION PREDICTION IN SVC USING PARTITION MODE WITHOUT SPLIT FLAG - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data. The video data includes a base layer and an enhancement layer. The base layer includes a coding unit tree co-located with an enhancement layer coding unit in the enhancement layer. The coding unit tree includes a plurality of nodes arranged in a tree structure and motion vectors. The enhancement layer coding unit is inter-mode coded. A processor is configured to split the enhancement layer coding unit into a plurality of nodes arranged in a tree structure that is the same as the tree structure of the coding unit tree. The processor is also configured to perform motion prediction for the enhancement layer coding unit based on the motion vectors of the coding unit tree. | 07-25-2013 |
20130188718 | MOTION PREDICTION IN SVC WITHOUT INCLUDING A TEMPORALLY NEIGHBORING BLOCK MOTION VECTOR IN A CANDIDATE LIST - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data and a candidate list. The video data includes a base layer and an enhancement layer. The base layer includes a base layer prediction unit co-located with an enhancement layer prediction unit in the enhancement layer. The candidate list includes a list of motion vectors for use by the enhancement layer prediction unit. A processor is configured to store motion vectors originating from spatial neighbors of the enhancement layer prediction unit, and not motion vectors originating from temporal neighbors of the enhancement layer prediction unit, in the candidate list. The processor is also configured to store motion vectors originating from the base layer prediction unit in the candidate list. | 07-25-2013 |
20130188719 | MOTION PREDICTION IN SVC USING MOTION VECTOR FOR INTRA-CODED BLOCK - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data. The video data may include a base layer and an enhancement layer. The base layer may include a base layer coding unit co-located with a first enhancement layer coding unit in the enhancement layer. A processor may be configured to construct one or more motion vectors based at least in part on one or more base layer motion vectors available at the co-located base layer coding unit. The one or more motion vectors may be associated with the first enhancement layer coding unit. The processor may also be configured to determine pixel values of a neighbor enhancement layer coding unit based at least in part on the one or more motion vectors. | 07-25-2013 |
20130195171 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include a dedicated array for parameter sets. | 08-01-2013 |
20130195172 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include an indicator identifying a number of temporal layers of the video stream. | 08-01-2013 |
20130195173 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include an indicator identifying a number of parameter sets stored within one or more access units of the video stream. | 08-01-2013 |
20130195205 | METHOD OF CODING VIDEO AND STORING VIDEO CONTENT - A device comprising a video file creation module is configured to obtain a plurality of slices of coded video content. Parameter sets are associated with the coded video content. The video creation module encapsulates the plurality of slices of coded video content within one or more access units of a video stream. A first type of parameter set may be encapsulated within one or more access units of the video stream. A second type of parameter set may be encapsulated within a sample description. The sample description may include stream properties associated with the video stream. | 08-01-2013 |
20130202035 | REFERENCE PICTURE LIST MODIFICATION FOR VIDEO CODING - A video coder may, in some cases, signal whether one or more initial reference picture lists are to be modified. When an initial list is to be modified, the video coder can signal information indicating a starting position in the initial reference picture list. When the starting position signaled by the video coder is less than a number of pictures included in the initial reference picture list, then the video coder signals the number of pictures to be inserted into the initial reference picture list, and a reference picture source from which a picture can be retrieved to insert into the initial reference picture list to construct a modified reference picture list. | 08-08-2013 |
20130222537 | BITSTREAM EXTRACTION IN THREE-DIMENSIONAL VIDEO - To extract a sub-bitstream from a 3-dimensional video (3DV) bitstream, a device determines a texture target view list that indicates views in the 3DV bitstream that have texture view components that are required for decoding pictures in a plurality of target views. The target views are a subset of the views in the bitstream that are to be decodable from the sub-bitstream. In addition, the device determines a depth target view list that indicates views in the 3DV bitstream that have depth view components that are required for decoding pictures in the plurality of target views. The device determines the sub-bitstream based at least in part on the texture target view list and the depth target view list. | 08-29-2013 |
20130222538 | NETWORK ABSTRACTION LAYER (NAL) UNIT HEADER DESIGN FOR THREE-DIMENSIONAL VIDEO CODING - A video encoder generates a network abstraction layer (NAL) unit that includes at least a first syntax element and a second syntax element. The first syntax element indicates that the NAL unit belongs to a particular NAL unit type. Coded slices of texture view components and depth view components are encapsulated within NAL units that belong to the particular NAL unit type. The second syntax element indicates whether a NAL unit header of the NAL unit includes an Advanced Video Coding (AVC)-compatible 3-dimensional video (3DV) header extension or includes a Multiview Video Coding (MVC)-compatible 3DV header extension. The video encoder outputs a bitstream that includes the NAL unit. A video decoder receives the NAL unit and determines whether the second syntax element indicates that the NAL unit header of the NAL unit includes the AVC-compatible 3DV header extension or the MVC-compatible 3DV header extension. | 08-29-2013 |
20130242046 | DISPARITY VECTOR PREDICTION IN VIDEO CODING - Techniques are described for determining a disparity vector for a current block based on disparity motion vectors of one or more spatially and temporally neighboring regions to a current block to be predicted. The spatially and temporally neighboring regions include one or a plurality of blocks, and the disparity motion vector represents a single vector in one reference picture list for the plurality of blocks within the spatially or temporally neighboring region. The determined disparity vector could be used to coding tools which utilize the information between different views such as merge mode, advanced motion vector prediction (AMVP) mode, inter-view motion prediction, and inter-view residual prediction. | 09-19-2013 |
20130243081 | HIGH-LEVEL SYNTAX EXTENSIONS FOR HIGH EFFICIENCY VIDEO CODING - In one example, a device includes a video coder configured to code a picture order count (POC) value for a first picture of video data, code a second-dimension picture identifier for the first picture, and code, in accordance with a base video coding specification or an extension to the base video coding specification, a second picture based at least in part on the POC value and the second-dimension picture identifier of the first picture. The video coder may comprise a video encoder or a video decoder. The second-dimension picture identifier may comprise, for example, a view identifier, a view order index, a layer identifier, or other such identifier. The video coder may code the POC value and the second-dimension picture identifier during coding of a motion vector for a block of the second picture, e.g., during advanced motion vector prediction or merge mode coding. | 09-19-2013 |
20130243093 | MOTION VECTOR CODING AND BI-PREDICTION IN HEVC AND ITS EXTENSIONS - In one example, a device includes a video coder (e.g., a video encoder or a video decoder) configured to determine that a block of video data is to be coded in accordance with a three-dimensional extension of High Efficiency Video Coding (HEVC), and, based the determination that the block is to be coded in accordance with the three-dimensional extension of HEVC, disable temporal motion vector prediction for coding the block. The video coder may be further configured to, when the block comprises a bi-predicted block (B-block), determine that the B-block refers to a predetermined pair of pictures in a first reference picture list and a second reference picture list, and, based on the determination that the B-block refers to the predetermined pair, equally weight contributions from the pair of pictures when calculating a predictive block for the block. | 09-19-2013 |
20130258052 | INTER-VIEW RESIDUAL PREDICTION IN 3D VIDEO CODING - In general, this disclosure describes techniques for improved inter-view residual prediction (IVRP) in three-dimensional video coding. These techniques include determining IVRP availability based on coded block flags and coding modes of residual reference blocks, disallowing IVRP coding when a block is inter-view predicted, using picture order count (POC) values to determine whether IVRP is permitted, applying IVRP to prediction units (PUs) rather than coding units (CUs), inferring values of IVRP flags when a block is skip or merge mode coded, using an IVRP flag of a neighboring block to determine context for coding an IVRP flag of a current block, and avoiding resetting of samples of a residual reference block to zeros during generation. | 10-03-2013 |
20130265388 | DISPARITY VECTOR CONSTRUCTION METHOD FOR 3D-HEVC - When coding multiview video data, a video coder can code one or more pictures in one or more reference views, including a first reference view and determine a disparity vector for a current block based on motion information of one or more neighboring blocks of the current block, wherein the current block is in a second view, wherein the disparity vector points from the current block to a corresponding block in a picture of the same time instance in one of the one or more reference views. | 10-10-2013 |
20130266075 | LOW-DELAY VIDEO BUFFERING IN VIDEO CODING - As one example, a method of coding video data includes storing one or more decoding units of video data in a picture buffer. The method further includes obtaining a respective buffer removal time for the one or more decoding units, wherein obtaining the respective buffer removal time comprises receiving a respective signaled value indicative of the respective buffer removal time for at least one of the decoding units. The method further includes removing the decoding units from the picture buffer in accordance with the obtained buffer removal time for each of the decoding units. The method further includes coding video data corresponding to the removed decoding units, wherein coding the video data comprises decoding the at least one of the decoding units. | 10-10-2013 |
20130266076 | LOW-DELAY VIDEO BUFFERING IN VIDEO CODING - As one example, a method of coding video data includes storing one or more decoding units of video data in a coded picture buffer (CPB). The method further includes obtaining a respective buffer removal time for the one or more decoding units. The method further includes removing the decoding units from the CPB in accordance with the obtained buffer removal time for each of the decoding units. The method further includes determining whether the CPB operates at access unit level or sub-picture level. The method further includes coding video data corresponding to the removed decoding units. If the CPB operates at access unit level, coding the video data comprises coding access units comprised in the decoding units. If the CPB operates at sub-picture level, coding the video data comprises coding subsets of access units comprised in the decoding units. | 10-10-2013 |
20130271565 | VIEW SYNTHESIS BASED ON ASYMMETRIC TEXTURE AND DEPTH RESOLUTIONS - An apparatus for processing video data includes a processor configured to associate, in a minimum processing unit (MPU), one pixel of a depth image of a reference picture with one or more pixels of a first chroma component of a texture image of the reference picture, associate, in the MPU, the one pixel of the depth image with one or more pixels of a second chroma component of the texture image, and associate, in the MPU, the one pixel of the depth image with a plurality of pixels of a luma component of the texture image. The number of the pixels of the luma component is different than the number of the one or more pixels of the first chroma component and the number of the one or more pixels of the second chroma component. | 10-17-2013 |
20130271566 | VIEW SYNTHESIS MODE FOR THREE-DIMENSIONAL VIDEO CODING - A video encoder signals, in a bitstream, a syntax element that indicates whether a current video unit is predicted from a VSP picture. The current video unit is a macroblock or a macroblock partition. The video encoder determines, based at least in part on whether the current video unit is predicted from the VSP picture, whether to signal, in the bitstream, motion information for the current video unit. A video decoder decodes the syntax element from the bitstream and determines, based at least in part on the syntax element, whether the bitstream includes the motion information. | 10-17-2013 |
20130272403 | REFERENCE PICTURE SET PREDICTION FOR VIDEO CODING - In one example, a device for decoding video data includes a video decoder configured to decode one or more syntax elements of a current reference picture set (RPS) prediction data structure, wherein at least one of the syntax elements represents a picture order count (POC) difference between a POC value associated with the current RPS and a POC value associated with a previously decoded RPS, form a current RPS based at least in part on the RPS prediction data structure and the previously decoded RPS, and decode one or more pictures using the current RPS. A video encoder may be configured to perform a substantially similar process during video encoding. | 10-17-2013 |
20130272408 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to, in an example, a method that includes identifying a first block of video data in a first temporal location from a first view, wherein the first block is associated with a first disparity motion vector. The method also includes determining a motion vector predictor for a second motion vector associated with a second block of video data, wherein the motion vector predictor is based on the first disparity motion vector. When the second motion vector comprises a disparity motion vector, the method includes determining the motion vector predictor comprises scaling the first disparity motion vector to generate a scaled motion vector predictor, wherein scaling the first disparity motion vector comprises applying a scaling factor comprising a view distance of the second disparity motion vector divided by a view distance of the first motion vector to the first disparity motion vector. | 10-17-2013 |
20130279576 | VIEW DEPENDENCY IN MULTI-VIEW CODING AND 3D CODING - This disclosure described techniques for coding layer dependencies for a block of video data. According to these techniques, a video encoder generates layer dependencies associated with a given layer. The video encoder also generates a type of prediction associated with one or more of the layer dependencies. In some examples, the video encoder generates a first syntax element to signal layer dependencies and a second syntax element to signal a type of prediction associated with one or more of the layer dependencies. A video decoder may obtain the layer dependencies associated with a given layer and the type of prediction associated with one or more of the layer dependencies. | 10-24-2013 |
20130287108 | DISPARITY VECTOR GENERATION FOR INTER-VIEW PREDICTION FOR VIDEO CODING - In one example, a video coder (e.g., a video encoder or a video decoder) is configured to determine that a current block of video data is coded using a disparity motion vector, wherein the current block is within a containing block, based on a determination that a neighboring block to the current block is also within the containing block, substitute a block outside the containing block and that neighbors the containing block for the neighboring block in a candidate list, select a disparity motion vector predictor from one of a plurality of blocks in the candidate list, and code the disparity motion vector based on the disparity motion vector predictor. In this manner, the techniques of this disclosure may allow blocks within the containing block to be coded in parallel. | 10-31-2013 |
20130304935 | Providing Sequence Data Sets for Streaming Video Data - A device may encapsulate video data such that Supplemental Enhancement Information (SEI) messages are stored separately from a sequence of coded video pictures described by the SEI messages. An example device includes a control unit configured to generate one or more SEI messages separate from the coded video pictures, wherein the SEI messages describe respective ones of the sequence of coded video pictures and include elements common to more than one of the coded video pictures, and an output interface configured to output the SEI messages separately from the sequence of coded video pictures. An example destination device may receive the SEI messages separately from the coded video pictures and render the coded video pictures using the SEI messages. | 11-14-2013 |
20130322531 | EXTERNAL PICTURES IN VIDEO CODING - A video encoder generates a syntax element that indicates whether a video unit of a current picture is predicted from an external picture. The external picture is in a different layer than the current picture. Furthermore, the video encoder outputs a video data bitstream that includes a representation of the syntax element. The video data bitstream may or may not include a coded representation of the external picture. A video decoder obtains the syntax element from the video data bitstream. The video decoder uses the syntax element in a process to reconstruct video data of a portion of the video unit. | 12-05-2013 |
20130329007 | REDUNDANCY REMOVAL FOR ADVANCED MOTION VECTOR PREDICTION (AMVP) IN THREE-DIMENSIONAL (3D) VIDEO CODING - In general, techniques are described for performing motion vector prediction in 3D video coding and, more particularly for managing a candidate list of motion vector predictors (MVPs) for a block of video data. In some examples, a video coder, such as video encoder or video decoder, includes at least three motion vector predictors (MVPs) in a candidate list of MVPs for a current block in a first view of a current access unit of the video data, wherein the at least three MVPs comprise an inter-view motion vector predictor (IVMP), which is a temporal motion vector derived from a block in a second view of the current access unit or a disparity motion vector derived from a disparity vector. | 12-12-2013 |
20130329787 | SIGNALING DATA FOR LONG TERM REFERENCE PICTURES FOR VIDEO CODING - A video coder codes a slice header for a slice of video data. The slice header includes a syntax element comprising identifying information for a long term reference picture, wherein the identifying information is explicitly signaled in the slice header or derived from a sequence parameter set corresponding to the slice. When the syntax element indicates that the identifying information for the long term reference picture is explicitly signaled, to code the slice header, the video coder is further configured to code a value for the identifying information for the long term reference picture in the slice header. | 12-12-2013 |
20130335522 | DERIVATION OF DEPTH MAP ESTIMATE - In some example techniques for generating and updating depth map estimates used for inter-view motion prediction and/or inter-view residual prediction for coding multiview video data, each of a plurality of dependent views is associated with a dependent depth view estimate, which may be generated or updated based on coding of the texture data of the dependent view relative to a base view. In such examples, each of the dependent depth map estimates may be warped to the base view to produce a respective one of a plurality of base depth map estimates. Each dependent depth map estimate and the respective base depth map estimate for a depth map estimate pair associated with the respective one of the plurality of dependent views. | 12-19-2013 |
20130336405 | DISPARITY VECTOR SELECTION IN VIDEO CODING - A video coder determines a first disparity vector using a first disparity vector derivation process. In addition, the video coder determines a second disparity vector using a second disparity vector derivation process. The first disparity vector derivation process is different than the second disparity vector derivation process. The video coder uses the first disparity vector to determine a motion vector prediction (MVP) candidate in a set of MVP candidates for a current prediction unit (PU). The video coder uses the second disparity vector to determine residual data. | 12-19-2013 |
20130336406 | REDUNDANCY REMOVAL FOR MERGE/SKIP MODE MOTION INFORMATION CANDIDATE LIST CONSTRUCTION - In general, techniques are described for constructing a merging candidate list for coding video data according to a merge mode and/or a skip mode. In some examples, the techniques include identifying one or more spatial merging candidates (SMCs) and an inter-view merging candidate (IVMC) for inclusion in a merging candidate list, and comparing the motion information of at least one of the SMCs to the motion information of the IVMC. In such examples, if the SMC has the same motion information as the IVMC, the techniques may further include pruning the merging candidate list to exclude the one of the merging candidates from the merging candidate list. | 12-19-2013 |
20130336407 | TEMPORAL MOTION VECTOR PREDICTION IN HEVC AND ITS EXTENSIONS - In one example, a device includes a video coder configured to determine, for each reference picture in one or more reference picture lists for a current picture, whether the reference picture is to be included in a plurality of reference pictures based on types for the reference pictures in the reference picture lists, compare picture order count (POC) values of each of the plurality of reference pictures to a POC value of the current picture to determine a motion vector predictor for a current block based on motion vectors of a co-located block of video data in a reference picture of the plurality of reference pictures, determine whether a forward motion vector or a backward motion vector of the co-located block is to be initially used to derive the motion vector predictor, and code a motion vector for the current block of video data relative to the motion vector predictor. | 12-19-2013 |
20130343465 | HEADER PARAMETER SETS FOR VIDEO CODING - An example method of decoding video data includes determining a header parameter set that includes one or more syntax elements specified individually by each of one or more slice headers, the header parameter set being associated with a header parameter set identifier (HPS ID), and determining one or more slice headers that reference the header parameter set to inherit at least one of the syntax elements included in the header parameter set, where the slice headers are each associated with a slice of the encoded video data, and where the slice headers each reference the header parameter set using the HPS ID. | 12-26-2013 |
20140003491 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003492 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003493 | VIDEO PARAMETER SET FOR HEVC AND EXTENSIONS | 01-02-2014 |
20140003536 | STREAMING ADAPTION BASED ON CLEAN RANDOM ACCESS (CRA) PICTURES | 01-02-2014 |
20140003537 | RANDOM ACCESS AND SIGNALING OF LONG-TERM REFERENCE PICTURES IN VIDEO CODING | 01-02-2014 |
20140003538 | SIGNALING LONG-TERM REFERENCE PICTURES FOR VIDEO CODING | 01-02-2014 |
20140010277 | SUPPLEMENTAL ENHANCEMENT INFORMATION (SEI) MESSAGES HAVING A FIXED-LENGTH CODED VIDEO PARAMETER SET (VPS) ID - Systems, methods, and devices are disclosed that code a supplemental enhancement information (SEI) message. In some examples, the SEI message may contain an identifier of an active video parameter set (VPS). In some examples, the identifier may be fixed-length coded. | 01-09-2014 |
20140016699 | REFERENCE PICTURE LIST MODIFICATION FOR VIDEO CODING - A video decoder generates an initial reference picture list (RPL). Furthermore, the video decoder determines that an ordered set of reference picture list modification (RPLM) syntax elements does not include any additional syntax elements when a syntax element in the ordered set of RPLM syntax elements has a particular value. Furthermore, the video decoder generates a final RPL. For each respective RPLM syntax element in the ordered set of syntax elements, when the respective RPLM syntax element does not have the particular value, the final RPL includes, at an insertion position for the respective RPLM syntax element, a particular reference picture. The respective syntax element indicates a position in the initial RPL of the particular RPLM reference picture. The insertion position for the respective RPLM syntax element corresponds to a position in the ordered set of RPLM syntax elements of the respective RPLM syntax element. | 01-16-2014 |
20140016701 | TEMPORAL MOTION VECTOR PREDICTION IN VIDEO CODING EXTENSIONS - A first reference index value indicates a position, within a reference picture list associated with a current prediction unit (PU) of a current picture, of a first reference picture. A reference index of a co-located PU of a co-located picture indicates a position, within a reference picture list associated with the co-located PU of the co-located picture, of a second reference picture. When the first reference picture and the second reference picture belong to different reference picture types, a video coder sets a reference index of a temporal merging candidate to a second reference index value. The second reference index value is different than the first reference index value. | 01-16-2014 |
20140022343 | PARAMETER SETS IN VIDEO CODING - A video parameter set (VPS) is associated with one or more coded video sequences (CVSs). The VPS includes a VPS extension for a video coding extension. The VPS extension includes a syntax element that indicates whether a video coding tool associated with the video coding extension is enabled for a set of applicable layers of a bitstream. When the syntax element indicates that the coding tool is enabled for the applicable layers, at least a portion of the video data that is associated with the CVSs and that is associated with the applicable layers is coded using the coding tool. When the syntax element indicates that the coding tool is not enabled for the applicable layers, the video data that is associated with the CVSs and that is associated with the applicable layers is not coded using the coding tool. | 01-23-2014 |
20140023138 | REUSING PARAMETER SETS FOR VIDEO CODING - In one example, a device includes a video coder (e.g., a video encoder or a video decoder) configured to code parameter set information for a video bitstream, code video data of a base layer of the video bitstream using the parameter set information, and code video data of an enhancement layer of the video bitstream using at least a portion of the parameter set information. The parameter set information may include, for example, profile and level information and/or hypothetical reference decoder (HRD) parameters. For example, the video coder may code a sequence parameter set (SPS) for a video bitstream, code video data of a base layer of the video bitstream using the SPS, and code video data of an enhancement layer of the video bitstream using at least a portion of the SPS, without using any other SPS for the enhancement layer. | 01-23-2014 |
20140049604 | CONSTRUCTING REFERENCE PICTURE LISTS FOR MULTI-VIEW OR 3DV VIDEO CODING - In one example, a video coder, such as a video encoder or a video decoder, is configured to code a value for a layer identifier in a slice header for a current slice in a current layer of multi-layer video data, and, when the value for the layer identifier is not equal to zero, code a first set of syntax elements in accordance with a base video coding standard, and code a second set of one or more syntax elements in accordance with an extension to the base video coding standard. The second set of syntax elements may include a syntax element representative of a position for an identifier of an inter-layer reference picture of a reference layer in a reference picture list, and the video coder may construct the reference picture list such that the identifier of the inter-layer reference picture is located in the determined position. | 02-20-2014 |
20140049605 | INTER-VIEW PREDICTED MOTION VECTOR FOR 3D VIDEO - A video coder determines a first picture order count (POC) value of a first reference picture associated with a first motion vector of a corresponding block that points in a first direction and determines whether a first reference picture list for the current block includes a reference picture having the first POC value; in response to the reference picture list not including the reference picture having the first POC value, determines a second POC value of a second reference picture associated with a second motion vector of the corresponding block that points in a second direction, determines whether the first reference picture list for the current block includes a reference picture having the second POC value and in response to the first reference picture list including the reference picture having the second POC value, decodes the current motion vector using the second motion vector of the corresponding block. | 02-20-2014 |
20140071235 | INTER-VIEW MOTION PREDICTION FOR 3D VIDEO - This disclosure describes techniques for improving coding efficiency of motion prediction in multiview and 3D video coding. In one example, a method of decoding video data comprises deriving one or more disparity vectors for a current block, the disparity vectors being derived from neighboring blocks relative to the current block, converting a disparity vector to one or more of inter-view predicted motion vector candidates and inter-view disparity motion vector candidates, adding the one or more inter-view predicted motion vector candidates and the one or more inter-view disparity motion vector candidates to a candidate list for a motion vector prediction mode, and decoding the current block using the candidate list. | 03-13-2014 |
20140078250 | ADVANCED INTER-VIEW RESIDUAL PREDICTION IN MULTIVIEW OR 3-DIMENSIONAL VIDEO CODING - Systems and methods for coding video information for a current view based on a residual prediction from video information for a reference view are described. In one innovative aspect, an apparatus for coding digital video is provided. The apparatus includes a memory configured to store current view video information and reference view video information. The apparatus also includes a processor configured to determine a value of a current video unit of the current view based at least on a motion compensated block of the reference view. The motion compensated block may be determined based at least in part on motion information and a disparity vector associated with the current video unit. Decoding devices and methods as well as corresponding encoding devices and methods are described. | 03-20-2014 |
20140078251 | SELECTION OF PICTURES FOR DISPARITY VECTOR DERIVATION - When coding multiview video data, a video encoder and video decoder may select a candidate picture from one of one or more random access point view component (RAPVC) pictures and one or more pictures having a lowest temporal identification value. The video encoder and video decoder may determine whether a block in the selected candidate picture is inter-predicted with a disparity motion vector and determine a disparity vector for a current block of a current picture based on the disparity motion vector. The video encoder and video decoder may inter-prediction encode or decode, respectively, the current block based on the determined disparity vector. | 03-20-2014 |
20140086325 | SCALABLE EXTENSIONS TO HEVC AND TEMPORAL MOTION VECTOR PREDICTION - In one example, a device includes a video coder configured to determine a first co-located reference picture for generating a first temporal motion vector predictor candidate for predicting a motion vector of a current block, determine a second co-located reference picture for generating a second temporal motion vector predictor candidate for predicting the motion vector of the current block, determine a motion vector predictor candidate list that includes at least one of the first temporal motion vector predictor candidate and the second temporal motion vector predictor candidate, select a motion vector predictor from the motion vector predictor candidate list, and code the motion vector of the current block relative to the selected motion vector predictor. | 03-27-2014 |
20140086328 | SCALABLE VIDEO CODING IN HEVC - In one example, a device includes a video coder configured to determine a temporal motion vector predictor for a motion vector associated with a block of video data of a current picture of a first, non-base layer of a plurality of layers of video data using a temporal motion vector prediction process, wherein the temporal motion vector prediction process includes identifying a co-located picture from which to derive the temporal motion vector predictor, and restrict the temporal motion vector prediction process such that the co-located picture used to derive the temporal motion vector predictor is not located in a layer other than the first layer of the plurality of layers of video data. | 03-27-2014 |
20140092213 | SUB-BITSTREAM EXTRACTION FOR MULTIVIEW, THREE-DIMENSIONAL (3D) AND SCALABLE MEDIA BITSTREAMS - Techniques are described for modal sub-bitstream extraction. For example, a network entity may select a sub-bitstream extraction mode from a plurality of sub-bitstream extraction modes. Each sub-bitstream extraction mode may define a particular manner in which to extract coded pictures from views or layers to allow a video decoder to decode target output views or layers for display. In this manner, the network entity may adaptively select the appropriate sub-bitstream extraction technique, rather than a rigid, fixed sub-bitstream extraction technique. | 04-03-2014 |
20140092967 | USING BASE LAYER MOTION INFORMATION - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory is configured to store the video data associated with a base layer and an enhancement layer. The base layer may comprise a reference block and base layer motion information associated with the reference block. The enhancement layer may comprise a current block. A processor operationally coupled to the memory is configured to determine a position of the base layer motion information in a candidate list based on a prediction mode in a plurality of prediction modes used at the enhancement layer. The processor is further configured to perform a prediction of the current block based at least in part on the candidate list. | 04-03-2014 |
20140098851 | INDICATION OF VIDEO PROPERTIES - In one example, a method of decoding video data includes receiving, by a video decoder, a coded video sequence and decoding one or more bits of a reserved bits syntax element for the coded video sequence as one or more coding tool enable bits, wherein the reserved bit syntax element is part of a syntax structure that includes profile and level information, and wherein the one or more coding tool enable bits indicate whether one or more coding tools are enabled for use by the video decoder in decoding the video sequence. In some examples, the syntax structure is a profile_tier_level syntax structure. In additional examples, one or more coding tool enable bits are not included elsewhere in a sequence parameter set (SPS) syntax information. | 04-10-2014 |
20140098880 | PREDICTION MODE INFORMATION UPSAMPLING FOR SCALABLE VIDEO CODING - In one embodiment, an apparatus configured to code video data includes a processor and a memory unit. The memory unit stores video data associated with a first layer having a first spatial resolution and a second layer having a second spatial resolution. The video data associated with the first layer includes at least a first layer block and first layer prediction mode information associated with the first layer block, and the first layer block includes a plurality of sub-blocks where each sub-block is associated with respective prediction mode data of the first layer prediction mode information. The processor derives the predication mode data associated with one of the plurality of sub-blocks based at least on a selection rule, upsamples the derived prediction mode data and the first layer block, and associates the upsampled prediction mode data with each upsampled sub-block of the upsampled first layer block. | 04-10-2014 |
20140098881 | MOTION FIELD UPSAMPLING FOR SCALABLE CODING BASED ON HIGH EFFICIENCY VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a first layer having a first spatial resolution and a corresponding second layer having a second spatial resolution, wherein the first spatial resolution is less than the second spatial resolution. The video information includes at least motion field information associated with the first layer. The processor upsamples the motion field information associated with the first layer. The processor further adds an inter-layer reference picture including the upsampled motion field information in association with an upsampled texture picture of the first layer to a reference picture list to be used for inter prediction. The processor may encode or decode the video information. | 04-10-2014 |
20140098882 | INTER-VIEW PREDICTED MOTION VECTOR FOR 3D VIDEO - For a depth block in a depth view component, a video coder derives a motion information candidate that comprises motion information of a corresponding texture block in a decoded texture view component, adds the motion information candidate to a candidate list for use in a motion vector prediction operation, and codes the current block based on a candidate in the candidate list. | 04-10-2014 |
20140105299 | PERFORMING RESIDUAL PREDICTION IN VIDEO CODING - In general, techniques are described for performing residual prediction in video coding. As one example, a device configured to code scalable or multi-view video data may comprise one or more processors configured to perform the techniques. The processors may determine a difference picture, for a current picture, based on a first reference picture in a same layer or view as the current picture and a decoded picture in a different layer or view as the current picture. The decoded picture may be in a same access unit as the first reference picture. The processors may perform bi-prediction based on the difference picture to code at least a portion of the current picture. | 04-17-2014 |
20140119437 | TARGET OUTPUT LAYERS IN VIDEO CODING - In one example, a device includes a video coder configured to code a multilayer bitstream comprising a plurality of layers of video data, where the plurality of layers of video data are associated with a plurality of layer sets, and where each layer set contains one or more layers of video data of the plurality of layers, and to code on one or more syntax elements of the bitstream indicating one or more output operation points, where each output operation point is associated with a layer set of the plurality of layer sets and one or more target output layers of the plurality of layers. | 05-01-2014 |
20140119447 | REFERENCE PICTURE STATUS FOR VIDEO CODING - The techniques of this disclosure may be generally related to reference statues of pictures. The techniques may store the reference status information of reference pictures of a picture, at an instance when the picture is being coded. The techniques may then utilize the reference status information of the reference pictures of the picture, at the instance when the picture was coded, to inter-predict video blocks of a subsequent picture. | 05-01-2014 |
20140133556 | MPEG FRAME COMPATIBLE VIDEO CODING - In an example, a video coder may determine a first layer component of a first layer of video data, wherein the first layer of video data is associated with a layer identifier. The video coder may generate at least one filtered layer component by filtering the first layer component, and assign the layer identifier of the first layer and a filtered layer component index to the at least one filtered layer component, where the filtered layer component index is different than a layer component index of the first layer component. The video coder may also add the at least one filtered layer component to a reference picture set for performing inter-layer prediction of a layer other than the first layer of video data. | 05-15-2014 |
20140139627 | ADAPTIVE LUMINANCE COMPENSATION IN THREE DIMENSIONAL VIDEO CODING - A method of coding video data includes deriving prediction weights for illumination compensation of luma samples of a video block partition once for the video block partition such that the video block partition has a common set of prediction weights for performing illumination compensation of the luma samples regardless of a transform size for the video block partition, calculating a predicted block for the video block partition using the prediction weights using illumination compensation, and coding the video block partition using the predicted block. | 05-22-2014 |
20140140399 | LOW-COMPLEXITY SUPPORT OF MULTIPLE LAYERS FOR HEVC EXTENSIONS IN VIDEO CODING - A method of coding video data includes receiving video information associated with a first layer and a second layer and determining whether at least one of the first layer and the second layer is a default layer. The method can include at least partially restricting inter-layer prediction when neither the first layer nor the second layer is the default layer. A default layer can be a base layer or an enhancement layer. A flag can be received that indicates that inter-layer prediction is to be restricted. In addition, the method can include determining whether inter-layer prediction is allowed for the video information associated with the first layer, and determining whether inter-layer prediction is partially allowed for the video information associated with the second layer such that motion compensation is not used with the second layer video information. | 05-22-2014 |
20140141762 | GENERIC FEATURE-LICENSING FRAMEWORK - Disclosed is an arrangement to enable customers to provision devices with feature licenses that enable specified features in the devices. The arrangement includes a feature-licensing system for performing feature-licensing processes to provision the devices with feature licenses and a feature-licensing process configuration system. | 05-22-2014 |
20140161175 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining a partition mode for coding a block of video data, where the partition mode indicates a division of the block of video data for predictive coding. The method also includes determining whether to code a weighting factor for an inter-view residual prediction process based on the partition mode, where, when the weighting factor is not coded, the inter-view residual prediction process is not applied to predict a residual for the block. The method also includes coding the block of video data with the determined partition mode. | 06-12-2014 |
20140161179 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION BASED ON HIGH EFFICIENCY VIDEO CODING - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with an enhancement layer having a first block and a base layer having a second block, the second block in the base layer corresponding to the first block in the enhancement layer. The processor is configured to predict, by inter layer prediction, the first block in the enhancement layer based on information derived from the second block in the base layer. At least a portion of the second block is located outside of a reference region of the base layer, the reference region being available for use for the inter layer prediction of the first block. The processor may encode or decode the video information. | 06-12-2014 |
20140161186 | ADVANCED MERGE/SKIP MODE AND ADVANCED MOTION VECTOR PREDICTION (AMVP) MODE FOR 3D VIDEO - Techniques are described where if an inter-view predicted motion vector candidate (IPMVC) and an inter-view disparity motion vector candidate (IDMVC) are derived based on a shifted disparity vector, where the amount by which the disparity vector is shifted for the IPMVC and IDMVC is different. The techniques also prioritize the inclusion of the IPMVC over the IDMVC in a candidate list, and prune the IPMVC and the IDMVC if there is a duplicated IPMVC or IDMVC in the candidate list. | 06-12-2014 |
20140161187 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding multi-layer video data includes determining, for a first block of video data at a first temporal location, whether one or more reference picture lists for coding the first block contain at least one reference picture at a second, different temporal location. The method also includes coding the first block of video data relative to at least one reference block of video data of a reference picture in the one or more reference picture lists, where coding includes disabling an inter-view residual prediction process when the one or more reference picture lists do not include at least one reference picture at the second temporal location. | 06-12-2014 |
20140161188 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining a location of a temporal reference block indicated by a temporal motion vector to a current block of video data, where the current block and the temporal reference block are located in a first layer of video data. The method also includes interpolating, with a first type of interpolation, a location of a disparity reference block indicated by a disparity vector of the current block, where the disparity reference block is located in a second, different layer, and where the first type of interpolation comprises a bi-linear filter. The method also includes determining a temporal-disparity reference block of the disparity reference block indicated by a combination of the temporal motion vector and the disparity vector, and coding the current block based on the temporal reference block, the disparity reference block, and the temporal-disparity reference block. | 06-12-2014 |
20140161189 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining, for a first block of video data in a first layer of video data, a temporal motion vector and associated temporal reference picture for predicting the first block, where the temporal reference picture has a picture order count value. The method also includes determining a disparity reference block in a disparity reference picture indicated by a disparity vector associated with the first block, and determining whether a decoded picture buffer contains a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture. When the decoded picture buffer does not contain a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture, the method includes modifying an inter-view residual prediction process for predicting residual data of the first block. | 06-12-2014 |
20140168363 | DISPARITY VECTOR DERIVATION - A parent block is partitioned into the plurality of blocks and a disparity vector derivation process is performed to derive a disparity vector for a representative block in the plurality of blocks. A video encoder generates a bitstream that includes a coded representation of the video data in part by performing, based on the derived disparity vector and without separately deriving disparity vectors for any block in the plurality of blocks other than the representative block, inter-view prediction for two or more blocks in the plurality of blocks. A video decoder reconstructs sample blocks for two or more blocks in the plurality of blocks in part by performing, based on the derived disparity vector and without separately deriving disparity vectors for any block in the plurality of blocks other than the representative block, inter-view prediction for the two or more blocks in the plurality of blocks. | 06-19-2014 |
20140169474 | DISPARITY VECTOR DERIVATION - A device performs a disparity vector derivation process to determine a disparity vector for a current block. As part of performing the disparity vector derivation process, when either a first or a second spatial neighboring block has a disparity motion vector or an implicit disparity vector, the device converts the disparity motion vector or the implicit disparity vector to the disparity vector for the current block. The number of neighboring blocks that is checked in the disparity vector derivation process is reduced, potentially resulting in decreased complexity and memory bandwidth requirements. | 06-19-2014 |
20140169475 | MOTION VECTOR PREDICTION IN VIDEO CODING - During a process to derive an inter-view predicted motion vector candidate (IPMVC) for an Advanced Motion Vector Prediction (AMVP) candidate list, a video coder determines, based on a disparity vector of a current prediction unit (PU), a reference PU for the current PU. Furthermore, when a first reference picture of the reference PU has the same picture order count (POC) value as a target reference picture of the current PU, the video coder determines an IPMVC based on a first motion vector of the reference PU. Otherwise, when a second reference picture of the reference PU has the same POC value as the target reference picture of the current PU, the video coder determines the IPMVC based on a second motion vector of the reference PU. | 06-19-2014 |
20140176674 | DISPARITY VECTOR DERIVATION IN THREE-DIMENSIONAL VIDEO CODING - A video coder searches a set of neighbor blocks to generate a plurality of disparity vector candidates. Each of the neighbor blocks is a spatial or temporal neighbor of a current block. The video coder determines, based at least in part on the plurality of disparity vector candidates, a final disparity vector for the current block. | 06-26-2014 |
20140177720 | CONSTRAINTS ON NEIGHBORING BLOCK BASED DISPARITY VECTOR (NBDV) TECHNIQUES FOR 3D VIDEO - Techniques are described for determining whether a block in a candidate reference picture is available. A video coder may determine a location of a co-located largest coding unit (CLCU) in the candidate reference picture, where the CLCU is co-located with a LCU in a current picture, and the LCU includes a current block that is to be inter-predicted. The video coder may determine whether a block in the candidate reference picture is available based on a location of the block in the candidate reference picture relative to the location of the CLCU. If the block in the candidate reference picture is unavailable, the video coder may derive a disparity vector for the current block from a block other than the block determined to be unavailable. | 06-26-2014 |
20140184740 | PARSING SYNTAX ELEMENTS IN THREE-DIMENSIONAL VIDEO CODING - When a current view is a dependent texture view, a current coding unit (CU) is not intra coded, and a partitioning mode of the current CU is equal to PART_2N×2N, a video coder obtains, from a bitstream that comprises an encoded representation of the video data, a weighting factor index for the current CU, wherein the current CU is in a picture belonging to a current view. When the current view is not a dependent texture view, or the current CU is intra coded, or the partitioning mode of the current CU is not equal to PART_2N×2N, the video decoder assumes that the weighting factor index is equal to a particular value that indicates that residual prediction is not applied with regard to the current CU. | 07-03-2014 |
20140185682 | MOTION VECTOR PREDICTION FOR VIDEO CODING - In one example, a device for coding video data includes a video coder configured to determine a first target reference index representing a reference picture to which a candidate motion vector predictor for a current motion vector refers, determine, for the candidate motion vector predictor and during temporal motion vector prediction (TMVP) of the current motion vector, a value for an additional target reference index that is equal to a predetermined value, and code the current motion vector based at least in part on the value of the additional reference index. | 07-03-2014 |
20140192149 | NON-NESTED SEI MESSAGES IN VIDEO CODING - A device obtains, from a bitstream that includes an encoded representation of the video data, a non-nested Supplemental Enhancement Information (SEI) message that is not nested within another SEI message in the bitstream. Furthermore, the device determines a layer of the bitstream to which the non-nested SEI message is applicable. The non-nested SEI message is applicable to layers for which video coding layer (VCL) network abstraction layer (NAL) units of the bitstream have layer identifiers equal to a layer identifier of a SEI NAL unit that encapsulates the non-nested SEI message. A temporal identifier of the SEI NAL unit is equal to a temporal identifier of an access unit containing the SEI NAL unit. Furthermore, the device processes, based in part on one or more syntax elements in the non-nested SEI message, video data of the layer of the bitstream to which the non-nested SEI message is applicable. | 07-10-2014 |
20140192151 | INDICATION OF PRESENCE OF TEXTURE AND DEPTH VIEWS IN TRACKS FOR MULTIVIEW CODING PLUS DEPTH - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of video data, wherein the track includes one or more views. The method further includes parsing information to determine whether the track includes only texture views, only depth views, or both texture and depth views. Another example method includes composing a track of video data, wherein the track includes one or more views and composing information that indicates whether the track includes only texture views, only depth views, or both texture and depth views. | 07-10-2014 |
20140192152 | INDICATION OF CURRENT VIEW DEPENDENCY ON REFERENCE VIEW IN MULTIVIEW CODING FILE FORMAT - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of video data, wherein the track includes one or more views. The method further includes parsing information to determine whether a texture view or a depth view of a reference view is required for decoding at least one of the one or more views in the track. Another example method includes composing a track of video data, wherein the track includes one or more views and composing information that indicates whether a texture view or a depth view of a reference view is required for decoding at least one of the one or more views in the track. | 07-10-2014 |
20140192153 | SIGNALING OF SPATIAL RESOLUTION OF DEPTH VIEWS IN MULTIVIEW CODING FILE FORMAT - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of multiview video data, wherein the track includes at least one depth view. The method further includes parsing information to determine a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view. Another example method includes composing a track of multiview video data, wherein the track includes the one or more views. The example method further includes composing information to indicate a spatial resolution associated with the depth view, wherein decoding the spatial resolution does not require parsing of a sequence parameter set of the depth view. | 07-10-2014 |
20140192157 | VIEW SYNTHESIS IN 3D VIDEO - In an example, a method of decoding video data includes determining whether a reference index for a current block corresponds to an inter-view reference picture, and when the reference index for the current block corresponds to the inter-view reference picture, obtaining, from an encoded bitstream, data indicating a view synthesis prediction (VSP) mode of the current block, where the VSP mode for the reference index indicates whether the current block is predicted with view synthesis prediction from the inter-view reference picture. | 07-10-2014 |
20140192882 | VIDEO BUFFERING OPERATIONS FOR RANDOM ACCESS IN VIDEO CODING - In an example, the present disclosure provides for receiving in a video bitstream an access unit having a first random access point (RAP) picture and receiving in the video bitstream, after the access unit in the bitstream, a subsequent access unit having a second RAP picture. In a case that one or more random access skipped leading (RASL) pictures for the subsequent access unit are not present in the received video bitstream, shifting a picture buffer removal time earlier based on a picture buffer removal delay offset. Another example provides for receiving an access unit after an earlier initialization of the hypothetical reference decoder (HRD), the access unit having a RAP picture, wherein associated access units containing RASL pictures are not received and initializing a picture buffer removal time and a picture buffer removal delay offset in response to receiving the access unit and not receiving the associated access units containing RASL pictures. | 07-10-2014 |
20140192885 | BITSTREAM CONSTRAINTS AND MOTION VECTOR RESTRICTION FOR INTER-VIEW OR INTER-LAYER REFERENCE PICTURES - Techniques are described for motion vector restriction where information in a bitstream ensures that a derived motion vector from a motion vector predictor is compliant with a motion vector restriction. Techniques are also described for indicating the motion vector restriction for parallel decoding. | 07-10-2014 |
20140192895 | MULTI-RESOLUTION DECODED PICTURE BUFFER MANAGEMENT FOR MULTI-LAYER VIDEO CODING - This disclosure describes various methods and techniques for decoded picture buffer (DPB) management when multiple decoded layer components with different resolutions need to be stored. In one example, a method of coding video data comprises decoding video data to produce a plurality of decoded layer components, storing the decoded layer components in one or more sub-units of a DPB, and performing a DPB management process on the one or more sub-units, wherein the DPB management process is managed separately for each of the one or more sub-units. | 07-10-2014 |
20140193139 | SEPARATE TRACK STORAGE OF TEXTURE AND DEPTH VIEWS FOR MULTIVIEW CODING PLUS DEPTH - Techniques for encapsulating video streams containing multiple coded views in a media file are described herein. In one example, a method includes parsing a track of multiview video data, wherein the track includes one or more views, including only one of a texture view of a particular view and a depth view of the particular view. The method further includes parsing a track reference to determine a dependency of the track to a referenced track indicated in the track reference. Track reference types include ‘deps’ that indicates that the track includes the depth view of the particular view and the reference track includes the texture, ‘tref’ that indicates that the track depends on the texture view which is stored in the referenced track, and ‘dref’ that indicates that the track depends on the depth view which is stored in the referenced track. | 07-10-2014 |
20140198181 | DISABLING INTER-VIEW PREDICTION FOR REFERENCE PICTURE LIST IN VIDEO CODING - A video coder signals, in a bitstream, a syntax element that indicates whether inter-view/layer reference pictures are ever included in a reference picture list for a current view component/layer representation. A video decoder obtains, from the bitstream, the syntax element that indicates whether inter-view/layer reference pictures are ever included in a reference picture list for a current view component/layer representation. The video decoder decodes the current view component/layer representation. | 07-17-2014 |
20140219363 | INTER-LAYER SYNTAX PREDICTION CONTROL - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer, an enhancement layer, or both. The processor is operationally coupled to the memory unit and is configured to provide a first indicator configured to enable or disable inter-layer syntax prediction and to provide a second indicator configured to enable or disable inter-layer texture prediction. The first and second indicators can be provided separately from each other. In one embodiment, the processor is further configured to disable inter-layer syntax prediction based at least in part upon information indicative of a codec associated with the base layer. | 08-07-2014 |
20140240456 | NEIGHBOR BLOCK-BASED DISPARITY VECTOR DERIVATION IN 3D-AVC - Techniques are described for deriving a disparity vector for a current block based on a disparity motion vector of a neighboring block in a 3D-AVC video coding process. The disparity vector derivation allows for texture-first coding where a depth view component of a dependent view is coded subsequent to the coding of the corresponding texture component of the dependent view. | 08-28-2014 |
20140241430 | NEIGHBORING BLOCK DISPARITY VECTOR DERIVATION IN 3D VIDEO CODING - In one example of the disclosure, a method of coding video data comprises coding video data using texture-first coding, and performing an NBDV derivation process for a block of the video data using a plurality of neighboring blocks. The NBDV derivation process comprises designating a motion vector associated with a neighboring block of the plurality of neighboring blocks coded with a block-based view synthesis prediction (BVSP) mode as an available disparity motion. | 08-28-2014 |
20140241431 | NEIGHBORING BLOCK DISPARITY VECTOR DERIVATION IN 3D VIDEO CODING - In one example of the disclosure, a method of coding video data comprises coding video data using texture-first coding, and performing an NBDV derivation process for a block of the video data using a plurality of neighboring blocks. The NBDV derivation process comprises designating a motion vector associated with a neighboring block of the plurality of neighboring blocks coded with a block-based view synthesis prediction (BVSP) mode as an available disparity motion. | 08-28-2014 |
20140253681 | INTER-VIEW RESIDUAL PREDICTION IN MULTI-VIEW OR 3-DIMENSIONAL VIDEO CODING - A video coder scales a motion vector of a current prediction unit (PU) of a current picture in order to compensate for a difference in temporal distance. In addition, the video coder determines a predictive block for the current PU, determines a disparity reference block based on samples of the disparity reference picture at a location indicated by a disparity vector of the current PU, and determines, based on samples of the fixed reference picture at a location indicated by the scaled motion vector, a temporal-disparity reference block for the current PU. The video coder then determines a residual predictor for the current PU. Each sample of the residual predictor for the current PU indicates a difference between a sample of the temporal-disparity reference block for the current PU and a corresponding sample of the disparity reference block. | 09-11-2014 |
20140253682 | SIMPLIFIED DEPTH CODING - In an example, a method of coding video data includes determining a first depth value of a depth look up table (DLT), where the first depth value is associated with a first pixel of the video data. The method also includes determining a second depth value of the DLT, where the second depth value is associated with a second pixel of the video data, The method also includes coding the DLT including coding the second depth value relative to the first depth value. | 09-11-2014 |
20140254679 | INTER-LAYER REFERENCE PICTURE CONSTRUCTION FOR SPATIAL SCALABILITY WITH DIFFERENT ASPECT RATIOS - A method of coding video data includes upsampling at least a portion of a reference layer picture to an upsampled picture having an upsampled picture size. The upsampled picture size has a horizontal upsampled picture size and a vertical upsampled picture size. At least one of the horizontal or vertical upsampled picture sizes may be different than a horizontal picture size or vertical picture size, respectively, of an enhancement layer picture. In addition, position information associated with the upsampled picture may be signaled. An inter-layer reference picture may be generated based on the upsampled picture and the position information. | 09-11-2014 |
20140254682 | DERIVED DISPARITY VECTOR IN 3D VIDEO CODING - A video coder stores only one derived disparity vector (DDV) for a slice of a current picture of the video data. The video coder uses the DDV for the slice in a Neighboring Block Based Disparity Vector (NBDV) derivation process to determine a disparity vector for a particular block. Furthermore, the video coder stores, as the DDV for the slice, the disparity vector for the particular block. | 09-11-2014 |
20140267605 | SIMPLIFICATIONS ON DISPARITY VECTOR DERIVATION AND MOTION VECTOR PREDICTION IN 3D VIDEO CODING - A device for coding three-dimensional video data includes a video coder configured to determine a first block of a first texture view is to be coded using a block-based view synthesis mode; locate, in a depth view, a first depth block that corresponds to the first block of the first texture view; determine depth values of two or more corner positions of the first depth block; based on the depth values, derive a disparity vector for the first block; using the disparity vector, locate a first block of a second texture view; and, inter-predict the first block of the first texture view using the first block of the second texture view. | 09-18-2014 |
20140269898 | SIMPLIFICATIONS ON DISPARITY VECTOR DERIVATION AND MOTION VECTOR PREDICTION IN 3D VIDEO CODING - A video coder can be configured to perform texture first coding for a first texture view, a first depth view, a second texture view, and a second depth view; for a macroblock of the second texture view, locate a depth block of the first depth view that corresponds to the macroblock; based on at least one depth value of the depth block, derive a disparity vector for the macroblock; code a first sub-block of the macroblock based on the derived disparity vector; and, code a second sub-block of the macroblock based on the derived disparity vector. | 09-18-2014 |
20140286421 | DISPARITY VECTOR REFINEMENT IN VIDEO CODING - A video coding device performs a disparity vector derivation process for a current block of multi-view video data. The current block is in a current view. An availability value indicates that a disparity vector for the current block is unavailable when the disparity vector derivation process is unable to derive the disparity vector for the current block. When the availability value indicates that the disparity vector derivation process has not derived the disparity vector for the current block, the video coding device generates a disparity vector for the current block in another manner. | 09-25-2014 |
20140286423 | DISPARITY VECTOR DERIVATION IN 3D VIDEO CODING FOR SKIP AND DIRECT MODES - A video decoder performs a neighboring-block based disparity vector (NBDV) derivation process to determine a disparity vector or performs a NBDV refinement (NBDV-R) process to determine the disparity vector. The video decoder uses the disparity vector as a disparity vector for a current block without using a median filtering process on multiple disparity motion vectors, wherein the current block is coded in either a skip mode or a direct mode. Furthermore, the video coder determines pixel values for the current block. | 09-25-2014 |
20140294061 | DEPTH CODING MODES SIGNALING OF DEPTH DATA FOR 3D-HEVC - Techniques are described for encoding and decoding depth data for three-dimensional (3D) video data represented in a multiview plus depth format using depth coding modes that are different than high-efficiency video coding (HEVC) coding modes. Examples of additional depth intra coding modes available in a 3D-HEVC process include at least two of a Depth Modeling Mode (DMM), a Simplified Depth Coding (SDC) mode, and a Chain Coding Mode (CCM). In addition, an example of an additional depth inter coding mode includes an Inter SDC mode. In one example, the techniques include signaling depth intra coding modes used to code depth data for 3D video data in a depth modeling table that is separate from the HEVC syntax. In another example, the techniques of this disclosure include unifying signaling of residual information of depth data for 3D video data across two or more of the depth coding modes. | 10-02-2014 |
20140294097 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code (e.g., encode or decode) video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a reference layer and an enhancement layer, the enhancement layer comprising an enhancement layer (EL) picture and the reference layer comprising a reference layer (RL) picture. The processor is configured to generate an inter-layer reference picture (ILRP) by resampling the RL picture; and determine whether, at a predetermined time, a reference picture of the ILRP was a short-term or long-term reference picture with respect to the ILRP. The processor may encode or decode the video information. | 10-02-2014 |
20140301439 | CROSS-LAYER POC ALIGNMENT FOR MULTI-LAYER BITSTREAMS THAT MAY INCLUDE NON-ALIGNED IRAP PICTURES - In one example, a video coder is configured to code a value for a syntax element indicating whether at least a portion of a picture order count (POC) value of a picture is to be reset to a value of zero, when the value for the syntax element indicates that the portion of the POC value is to be reset to the value of zero, reset at least the portion of the POC value such that the portion of the POC value is equal to zero, and code video data using the reset POC value. Coding video data using the reset POC value may include inter-predicting a block of a subsequent picture relative to the picture, where the block may include a motion parameter that identifies the picture using the reset POC value. The block may be coded using temporal inter-prediction or inter-layer prediction. | 10-09-2014 |
20140301441 | NON-ENTROPY ENCODED SET OF PROFILE, TIER, AND LEVEL SYNTAX STRUCTURES - Systems, methods, and devices for coding multilayer video data are disclosed that may include, encoding, decoding, transmitting, or receiving a non-entropy encoded set of profile, tier, and level syntax structures, potentially at a position within a video parameter set (VPS) extension. The systems, methods, and devices may refer to one of the profile, tier, and level syntax structures for each of a plurality of output layer sets. The systems, methods, and devices may encode or decode video data of one of the output layer sets based on information from the profile, tier, and level syntax structure referred to for the output layer set. | 10-09-2014 |
20140301454 | DEPTH CODING MODES SIGNALING OF DEPTH DATA FOR 3D-HEVC - Techniques are described for encoding and decoding depth data for three-dimensional (3D) video data represented in a multiview plus depth format using depth coding modes that are different than high-efficiency video coding (HEVC) coding modes. Examples of additional depth intra coding modes available in a 3D-HEVC process include at least two of a Depth Modeling Mode (DMM), a Simplified Depth Coding (SDC) mode, and a Chain Coding Mode (CCM). In addition, an example of an additional depth inter coding mode includes an Inter SDC mode. In one example, the techniques include signaling depth intra coding modes used to code depth data for 3D video data in a depth modeling table that is separate from the HEVC syntax. In another example, the techniques of this disclosure include unifying signaling of residual information of depth data for 3D video data across two or more of the depth coding modes. | 10-09-2014 |
20140301456 | INTER-LAYER PICTURE SIGNALING AND RELATED PROCESSES - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory configured to store inter-layer reference pictures associated with a current picture that is being coded. The apparatus further comprises a processor operationally coupled to the memory. In one embodiment, the processor is configured to indicate a number of inter-layer reference pictures to use to predict the current picture using inter-layer prediction. The processor is also configured to indicate which of the inter-layer reference pictures to use to predict the current picture using inter-layer prediction. The processor is also configured to determine an inter-layer reference picture set associated with the current picture using the indication of the number of inter-layer reference pictures and the indication of which of the inter-layer reference pictures to use to predict the current picture using inter-layer prediction. | 10-09-2014 |
20140301466 | GENERALIZED RESIDUAL PREDICTION IN HIGH-LEVEL SYNTAX ONLY SHVC AND SIGNALING AND MANAGEMENT THEREOF - An apparatus for decoding video information according to certain aspects includes a memory unit and a processor operationally coupled to the memory unit. The memory unit is configured to store at least one reference picture list of an enhancement layer, the at least one reference picture list comprising residual prediction reference picture information. The processor is configured to: decode signaled information about residual prediction reference picture generation; generate a residual prediction reference picture based on an enhancement layer reference picture and the decoded signaled information such that the generated residual prediction reference picture has the same motion field and the same picture order count (POC) as the enhancement layer reference picture from which it is generated; and store the generated residual prediction reference picture in the at least one reference picture list of the enhancement layer. | 10-09-2014 |
20140301467 | ADVANCED MERGE MODE FOR THREE-DIMENSIONAL (3D) VIDEO CODING - As part of a video encoding process or a video decoding process, a video coder may determine a first available disparity motion vector among spatial neighboring blocks of a current block of the video data. Furthermore, the video coder may shift a horizontal component of the first available disparity motion vector to derive a shifted disparity motion vector candidate (DSMV). The video coder may add the DSMV into a merge candidate list. | 10-09-2014 |
20140301469 | CODING VIDEO DATA FOR AN OUTPUT LAYER SET - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may receive or transmit a first output layer set for a layer set and receive or transmit a second output layer set for the layer set. The systems, methods, and devices may code (encode or decode) video data for at least one of the first output layer set and the second output layer set. | 10-09-2014 |
20140301483 | NON-ENTROPY ENCODED REPRESENTATION FORMAT - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may receive or transmit a non-entropy coded representation format within a video parameter set (VPS). The systems, methods, and devices may code (encode or decode) video data based on the non-entropy coded representation format within the VPS, wherein the representation format includes one or more of chroma format, whether different colour planes are separately coded, picture width, picture height, luma bit depth, and chroma bit depth. | 10-09-2014 |
20140301484 | PARAMETER SET DESIGNS FOR VIDEO CODING EXTENSIONS - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may transmit or receive a video parameter set (VPS) including information for a series of layers, each layer including visual signal information. The systems, methods, and devices may code (encode or decode) video data based on the visual signal information signaled per layer in the VPS. | 10-09-2014 |
20140301485 | IRAP ACCESS UNITS AND BITSTREAM SWITCHING AND SPLICING - In one example, a device for coding video data includes a video coder configured to code an intra random access point (IRAP) picture of a partially aligned IRAP access unit of video data, and code data that indicates, when performing random access from the partially aligned IRAP access unit, at least one picture of a video coding layer that is not correctly decodable. When the video coder comprises a video decoder, the video decoder may skip decoding of the pictures that are not correctly decodable, assuming random access has been performed starting from the partially aligned IRAP access unit. | 10-09-2014 |
20140307795 | BACKWARD VIEW SYNTHESIS PREDICTION - In one example, a device for coding video data includes a video coder configured to code motion information for a block of multiview video data, wherein the motion information includes a reference index that identifies a reference picture comprising a source for backward-warping view synthesis prediction (BVSP), perform BVSP on a portion of the reference picture to produce a BVSP reference block, and predict the block using the BVSP reference block. | 10-16-2014 |
20140307803 | NON-ENTROPY ENCODED LAYER DEPENDENCY INFORMATION - Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving a non-entropy encoded layer dependency information at a position within a video parameter set (VPS) extension prior to syntax elements of the VPS extension that are entropy encoded. The systems, methods, and devices may encode or decode the non-entropy encoded layer dependency information before an entropy encoded syntax element. The systems, methods, and devices may encode or decode video data of one or more of the layers of video data based on the non-entropy encoded layer dependency information. The layer dependency information indicates whether one of the layers is a direct reference layer for another of the layers. | 10-16-2014 |
20140341275 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 11-20-2014 |
20140355666 | ADVANCED DEPTH INTER CODING BASED ON DISPARITY OF DEPTH BLOCKS - In one example, the disclosure is directed to techniques that include, for each prediction unit (PU) of a respective coding unit (CU) of a slice of a picture of the video data, determining at least one disparity value based at least in part on at least one depth value of at least one reconstructed depth sample of at least one neighboring sample. The techniques further include determining at least one disparity vector based at least in part on the at least one disparity value, wherein the at least one disparity vector is for the respective CU for each PU. The techniques further include reconstructing, based at least in part on the at least one disparity vector, a coding block for the respective CU for each PU. | 12-04-2014 |
20140355685 | PARALLEL DERIVED DISPARITY VECTOR FOR 3D VIDEO CODING WITH NEIGHBOR-BASED DISPARITY VECTOR DERIVATION - For each respective coding unit (CU) of a slice of a picture of the video data, a video coder may set, in response to determining that the respective CU is the first CU of a coding tree block (CTB) row of the picture or the respective CU is the first CU of the slice, a derived disparity vector (DDV) to an initial value. Furthermore, the video coder may perform a neighbor-based disparity vector derivation (NBDV) process that attempts to determine a disparity vector for the respective CU. When performing the NBDV process does not identify an available disparity vector for the respective CU, the video coder may determine that the disparity vector for the respective CU is equal to the DDV. | 12-04-2014 |
20140355692 | HYPOTHETICAL REFERENCE DECODER MODEL AND CONFORMANCE FOR CROSS-LAYER RANDOM ACCESS SKIPPED PICTURES - A device may determine, based on a value, whether all cross-layer random access skipped (CL-RAS) pictures of an intra random access point (IRAP) access unit are present in a video data bitstream. In addition, the device may reconstruct pictures of the video data based at least in part on syntax elements decoded from the video data bitstream. | 12-04-2014 |
20140376633 | MORE ACCURATE ADVANCED RESIDUAL PREDICTION (ARP) FOR TEXTURE CODING - Techniques for advanced residual prediction (ARP) for coding video data may include inter-view ARP. Inter-view ARP may include identifying a disparity motion vector (DMV) for a current video block. The DMV is used for inter-view prediction of the current video block based on an inter-view reference video block. The techniques for inter-view ARP may also include identifying temporal reference video blocks in the current and reference views based on a temporal motion vector (TMV) of the inter-view reference video block, and determining a residual predictor block based on a difference between the temporal reference video blocks. | 12-25-2014 |
20150016500 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a first layer and a second layer. The processor is configured to decode first layer pictures of the first layer, store the decoded first layer pictures in a decoded picture buffer, determine whether second layer pictures having no corresponding first layer pictures are to be coded, and in response to determining that second layer pictures having no corresponding first layer pictures are to be coded, process an indication that one or more decoded first layer pictures stored in the decoded picture buffer are to be removed. The processor may encode or decode the video information. | 01-15-2015 |
20150016532 | SELECTION OF TARGET OUTPUT LAYERS IN HIGH EFFICIENCY VIDEO CODING EXTENSIONS - A method, apparatus, and manufacture for processing video data. A list of output layer sets in a video bitstream is received, and an index to at least one target output layer set in the list of output layer sets is received. Next, target output layers in the at least one target output layer set is determined based on the index. At least the target output layers from the video bitstream are decoded. Then, the decoded target output layers are output without outputting layers that are not targeted for output. | 01-15-2015 |
20150016545 | DECODED PICTURE BUFFER OPERATIONS FOR VIDEO CODING - A method of decoding video data comprising partitioning a decoded picture buffer (DPB) into a plurality of sub-DPBs, receiving at least one indication of a sub-DPB size for the plurality of sub-DPBs for one or more operation points of a multi-layer video coding process, and allocating memory space for the plurality of sub-DPBs based on the at least one indication. | 01-15-2015 |
20150023422 | PROCESSING ILLUMINATION COMPENSATION FOR VIDEO CODING - In one example, a device for coding (e.g., encoding or decoding) video data includes a memory configured to store video data and a video coder configured to determine a value for an advanced residual prediction (ARP) weighting factor of a current block of the video data and to skip coding of an illumination compensation syntax element for the current block and to code the current block when the value of the ARP weighting factor is not equal to zero. The video coder may further be configured to code the illumination compensation syntax element for the current block and code the current block based at least in part on the value of the illumination compensation syntax element when the value of the ARP weighting factor is equal to zero. | 01-22-2015 |
20150023423 | BLOCK IDENTIFICATION USING DISPARITY VECTOR IN VIDEO CODING - Techniques are described for determining a block in a reference picture in a reference view based on a disparity vector for a current block. The techniques start the disparity vector from a bottom-right pixel in a center 2×2 sub-block within the current block, and determine a location within the reference picture to which the disparity vector refers. The determined block covers the location referred to by the disparity vector based on the disparity vector starting from the bottom-right pixel in the center 2×2 sub-block within the current block. | 01-22-2015 |
20150030073 | SUB-PU MOTION PREDICTION FOR TEXTURE AND DEPTH CODING - In accordance with one or more techniques of this disclosure, a video coder may divide a current prediction unit (PU) into a plurality of sub-PUs. Each of the sub-PUs may have a size smaller than a size of the PU. Furthermore, the current PU may be in a depth view of the multi-view video data. For each respective sub-PU from the plurality of sub-PUs, the video coder may identify a reference block for the respective sub-PU. The reference block may be co-located with the respective sub-PU in a texture view corresponding to the depth view. The video coder may use motion parameters of the identified reference block for the respective sub-PU to determine motion parameters for the respective sub-PU. | 01-29-2015 |
20150030087 | USE OF A DEPTH CONDITION IN 3DV CODEC - In general, the disclosure relates to encoding and decoding a block of video data associated with three-dimensional (3D) video. A video coding device determines whether a depth condition associated with the block of video data should be set to valid within a coded bitstream. When the depth condition should be set to valid, the video coding device sets the depth condition to valid and encodes the block of video data using at least one camera parameter. The video coding device then determines whether the depth condition is valid. When the depth condition is valid, the video coding device decodes the block of video data using at least one camera parameter. | 01-29-2015 |
20150049821 | IN-LOOP DEPTH MAP FILTERING FOR 3D VIDEO CODING - This disclosure describes techniques for in-loop depth map filtering for 3D video coding processes. In one example, a method of decoding video data comprises decoding a depth block corresponding to a texture block, receiving a respective indication of one or more offset values for the decoded depth block, and performing a filtering process on edge pixels of the depth block using at least one of the one or more offset values to create a filtered depth block. | 02-19-2015 |
20150055704 | SUB-PU-LEVEL ADVANCED RESIDUAL PREDICTION - A prediction unit (PU) of a coding unit (CU) is split into two or more sub-PUs including a first sub-PU and a second sub-PU. A first motion vector of a first type is obtained for the first sub-PU and a second motion vector of the first type is obtained for the second sub-PU. A third motion vector of a second type is obtained for the first sub-PU and a fourth motion vector of the second type is obtained for the second sub-PU, such that the second type is different than the first type. A first portion of the CU corresponding to the first sub-PU is coded according to advanced residual prediction (ARP) using the first and third motion vectors. A second portion of the CU corresponding to the second sub-PU is coded according to ARP using the second and fourth motion vectors. | 02-26-2015 |
20150063464 | LOOKUP TABLE CODING - In general, techniques are described for lookup table coding. A device comprising one or more processors and a memory may be configured to perform the techniques. The processors are configured to receive at least one difference table including a set of values, each value of the set being included or not included in the reference lookup table and generate a current lookup table based on the reference lookup table and the difference table. The current lookup table may include at least one of a value from the difference table that is not included in the reference table or a value from the reference table that is not included in the difference table. The one or more processors may then decode the video data based on a set of values of the current lookup table. The memory may be configured to store the current lookup table. | 03-05-2015 |
20150078450 | VIDEO CODING TECHNIQUES USING ASYMMETRIC MOTION PARTITIONING - Techniques for decoding video data include receiving residual data corresponding to a block of video data, wherein the block of video data is encoded using asymmetric motion partitioning, is uni-directionally predicted using backward view synthesis prediction (BVSP), and has a size of 16×12, 12×16, 16×4 or 4×16, partitioning the block of video data into sub-blocks, each sub-block having a size of 8×4 or 4×8, deriving a disparity motion vector for each of the sub-blocks from a corresponding depth block in a depth picture corresponding to a reference picture, synthesizing a respective reference block for each of the sub-blocks using the respective derived disparity motion vector, and decoding the block of video data by performing motion compensation on each of the sub-blocks using the residual data and the synthesized respective reference blocks. | 03-19-2015 |
20150085929 | SUB-PREDICTION UNIT (PU) BASED TEMPORAL MOTION VECTOR PREDICTION IN HEVC AND SUB-PU DESIGN IN 3D-HEVC - Techniques are described for sub-prediction unit (PU) based motion prediction for video coding in HEVC and 3D-HEVC. In one example, the techniques include an advanced temporal motion vector prediction (TMVP) mode to predict sub-PUs of a PU in single layer coding for which motion vector refinement may be allowed. The advanced TMVP mode includes determining motion vectors for the PU in at least two stages to derive motion information for the PU that includes different motion vectors and reference indices for each of the sub-PUs of the PU. In another example, the techniques include storing separate motion information derived for each sub-PU of a current PU predicted using a sub-PU backward view synthesis prediction (BVSP) mode even after motion compensation is performed. The additional motion information stored for the current PU may be used to predict subsequent PUs for which the current PU is a neighboring block. | 03-26-2015 |
20150085930 | COMBINED BI-PREDICTIVE MERGING CANDIDATES FOR 3D VIDEO CODING - A video coder generates a list of merging candidates for coding a video block of the 3D video. A maximum number of merging candidates in the list of merging candidates may be equal to 6. As part of generating the list of merging candidates, the video coder determines whether a number of merging candidates in the list of merging candidates is less than 5. If so, the video coder derives one or more combined bi-predictive merging candidates. The video coder includes the one or more combined bi-predictive merging candidates in the list of merging candidates. | 03-26-2015 |
20150085935 | SUB-PREDICTION UNIT (PU) BASED TEMPORAL MOTION VECTOR PREDICTION IN HEVC AND SUB-PU DESIGN IN 3D-HEVC - Techniques are described for sub-prediction unit (PU) based motion prediction for video coding in HEVC and 3D-HEVC. In one example, the techniques include an advanced temporal motion vector prediction (TMVP) mode to predict sub-PUs of a PU in single layer coding for which motion vector refinement may be allowed. The advanced TMVP mode includes determining motion vectors for the PU in at least two stages to derive motion information for the PU that includes different motion vectors and reference indices for each of the sub-PUs of the PU. In another example, the techniques include storing separate motion information derived for each sub-PU of a current PU predicted using a sub-PU backward view synthesis prediction (BVSP) mode even after motion compensation is performed. The additional motion information stored for the current PU may be used to predict subsequent PUs for which the current PU is a neighboring block. | 03-26-2015 |