Patent application number | Description | Published |
20080260027 | MODE UNIFORMITY SIGNALING FOR INTRA-CODING - Techniques for efficiently signaling one or more prediction modes for blocks in a macroblock. A mode uniformity indicator is provided for each macroblock to indicate whether all blocks in the macroblock are to be predicted using the same prediction mode. In an embodiment, an encoder signals a mode uniformity indicator and a prediction mode. In another embodiment, a decoder receives a mode uniformity indicator and a prediction mode, and predicts at least two blocks of the macroblock using the prediction mode if the mode uniformity indicator is true. | 10-23-2008 |
20080260030 | DIRECTIONAL TRANSFORMS FOR INTRA-CODING - Techniques for transforming the prediction error of intra-coded blocks using mode-dependent transform functions. In an embodiment, an encoder selects a set of transform functions to represent prediction error based on the spatial mode used for prediction. In an alternative embodiment, a decoder reconstructs an image block by using the signaled spatial mode to derive the corresponding set of transform functions. No additional signaling between encoder and decoder is required as compared to prior art implementations. | 10-23-2008 |
20080260031 | PIXEL-BY-PIXEL WEIGHTING FOR INTRA-FRAME CODING - Techniques for improving the accuracy of prediction in intra-frame coding. A prediction mode can specify a pixel along a direction independently of other pixels along the same direction. In an embodiment, an encoder selects a prediction mode to best represent the image block. In an alternative embodiment, a decoder reconstructs each pixel in the image block by weighting neighboring pixels according to a weight matrix specified by the prediction mode. | 10-23-2008 |
20080310504 | ADAPTIVE COEFFICIENT SCANNING FOR VIDEO CODING - This disclosure describes techniques for scanning coefficients of video blocks. In particular, the techniques of this disclosure adapt a scan order used to scan a two-dimensional block of coefficients into a one-dimensional coefficient vector based on statistics associated with one or more previously coded blocks. For example, statistics that indicate the likelihood that a given coefficient value in each position of a two-dimensional block is zero or non-zero may be collected for one or more previously coded blocks. At some point, an adjustment to the scan order can be made in order to better ensure that non-zero coefficients are grouped together near the front of the one-dimensional coefficient vector, which can improve the effectiveness of entropy coding. The collection of statistics and adjustment of scan order may be made separately for each possible prediction mode. | 12-18-2008 |
20080310507 | ADAPTIVE CODING OF VIDEO BLOCK PREDICTION MODE - This disclosure describes techniques for coding of header information of video blocks. In particular, the techniques of this disclosure select one of a plurality of prediction modes for use in generating a prediction block of a video block of a coding unit, the plurality of prediction modes including unidirectional prediction modes and multi-directional prediction modes that combine at least two unidirectional prediction modes. An encoding device encodes the prediction mode of the current video block based on prediction modes of one or more previously encoded video blocks of the coding unit. Likewise, a decoding unit receives encoded video data of a video block of a coding unit and decodes the encoded video data to identify one of a plurality of prediction modes for use in generating a prediction block of the video block based on prediction modes of one or more previously decoded video blocks of the coding unit. | 12-18-2008 |
20080310512 | SEPARABLE DIRECTIONAL TRANSFORMS - This disclosure describes techniques for transforming residual blocks of video data. In particular, a plurality of different transforms selectively applied to the residual blocks based on the prediction mode of the video blocks. At least a portion of the plurality of transforms are separable directional transform specifically trained for a corresponding prediction mode to provide better energy compaction for the residual blocks of the given prediction mode. Using separable directional transforms offers the benefits of lower computation complexity and storage requirement than use of non-separable directional transforms. Additionally, a scan order used to scan the coefficients of the residual block may be adjusted when applying separable directional transforms. In particular, the scan order may be adjusted based on statistics associated with one or more previously coded blocks to better ensure that non-zero coefficients are grouped near the front of the one-dimensional coefficient vector to improve the effectiveness of entropy coding. | 12-18-2008 |
20080310745 | ADAPTIVE COEFFICIENT SCANNING IN VIDEO CODING - This disclosure describes techniques for scanning coefficients of video blocks, e.g., quantized and transformed coefficients. Rather than use conventional zig-zag scanning, the techniques of this disclosure adapt the scanning order based on statistics associated with previously coded blocks that were coded in the same prediction mode. For each prediction mode, statistics of the coefficients are stored, e.g., indicating probabilities that given coefficients are zero or non-zero. Periodically, adjustments to the scanning order can be made in order to better ensure that non-zero coefficients are grouped together and zero value coefficients are grouped together, which can improve the effectiveness of entropy coding. The techniques of this disclosure provide thresholds and threshold adjustments that can reduce the frequency that the scanning order adjustments occur, yet still achieve desired improvements in compression due to such scanning order adjustments. | 12-18-2008 |
20090003440 | SUB-BAND SCANNING TECHNIQUES FOR ENTROPY CODING OF SUB-BANDS - This disclosure describes techniques useful in the encoding and/or decoding of video data of a video sequence. In general, this disclosure sets forth scanning techniques useful in the context of sub-band coding, which may improve the level of compression that can be achieved by entropy coding following sub-band coding. In one example, a method of encoding video data of a video sequence comprises sub-band encoding the video data to generate a plurality of sub-bands, scanning each of the sub-bands from two-dimensional blocks into one-dimensional vectors based on scan orders defined for each of the sub-bands, and entropy encoding each of the scanned sub-bands. | 01-01-2009 |
20090097548 | ENHANCEMENT LAYER CODING FOR SCALABLE VIDEO CODING - This disclosure describes scalable video coding techniques. In particular, the techniques may be used to encode refinements of a video block for enhancement layer bit streams in a single coding pass, thereby reducing coding complexity, coding delay and memory requirements. In some instances, the techniques encode each nonzero coefficient of a coefficient vector of the enhancement layer without knowledge of any subsequent coefficients. Coding the enhancement layer in a single pass may eliminate the need to perform a first pass to analyze the coefficient vector and a second pass for coding the coefficient vector based on the analysis. | 04-16-2009 |
20090097558 | SCALABLE VIDEO CODING TECHNIQUES FOR SCALABLE BITDEPTHS - This disclosure describes video coding techniques that are executed in a bitdepth-based scalable video coding (SVC) scheme. Base layer video data and enhancement layer video data may be defined at different bitdepths. The techniques of this disclosure concern the coding of enhancement layer video blocks using a unique combination of inter-layer predictive coding modes after determining that inter-layer prediction should be used. In particular, two or more different inter-layer predictive coding modes may be utilized block corresponding to inter base layer blocks, and one of these inter-layer predictive coding modes may also be utilized for blocks corresponding to intra base layer blocks. Syntax information can be generated at the encoder and communicated with the bitstream, so that a decoder can use the proper inter-layer predictive coding mode in the decoding. | 04-16-2009 |
20090097568 | ENTROPY CODING OF INTERLEAVED SUB-BLOCKS OF A VIDEO BLOCK - This disclosure describes techniques for entropy coding of video blocks, and proposes a syntax element that may promote coding efficiency. The syntax element may identify a number of non-zero value sub-blocks within a video block, wherein the non-zero value sub-blocks comprise sub-blocks within the video block that include at least one non-zero coefficient. A method of coding a video block may comprise coding the syntax element, generating the non-zero value sub-blocks of the video block, and entropy coding the non-zero value sub-blocks. | 04-16-2009 |
20090175331 | TWO PASS QUANTIZATION FOR CABAC CODERS - This disclosure describes techniques for quantizing coefficients of a video block for a video coding process that supports context-based adaptive binary arithmetic coding (CABAC). A method may comprise estimating a last non-zero coefficient of the video block, and generating a set of quantized coefficients for the video block based on an assumption that the estimated last non-zero coefficient is actually the last non-zero coefficient of the video block. | 07-09-2009 |
20090175332 | QUANTIZATION BASED ON RATE-DISTORTION MODELING FOR CABAC CODERS - This disclosure describes techniques for quantizing coefficients of a video block for a video coding process that supports context-based adaptive binary arithmetic coding (CABAC). A method may comprise generating a plurality of sets of quantized coefficients for the video block, wherein each of the sets of quantized coefficients is defined based on an assumption that a particular one of the coefficients of the video block corresponds to a last non-zero coefficient of the video block, estimating a plurality of costs associated with coding the video block via the CABAC, wherein the estimated costs are based respectively on different ones of the sets of quantized coefficients, and selecting a given one of the sets of quantized coefficients associated with a lowest cost. | 07-09-2009 |
20090175334 | ADAPTIVE CODING OF VIDEO BLOCK HEADER INFORMATION - This disclosure describes techniques for adaptively coding video block header information based on previously encoded video blocks. A video encoder may adaptively select a coding table for use in encoding a syntax element of a current video block based on corresponding syntax elements of one or more previously encoded blocks. In one aspect, the video encoder may adaptively select the coding table for use in encoding a block type of the current block based on block types of one or more video blocks adjacent to the current video block, i.e., neighboring video blocks. The video encoder may also predict one or more other header syntax elements of the current block based on at least one of the previously encoded video blocks. If prediction is successful, the video encoder may encode a flag to indicate the success of prediction. | 07-09-2009 |
20090175336 | VIDEO CODING OF FILTER COEFFICIENTS BASED ON HORIZONTAL AND VERTICAL SYMMETRY - This disclosure recognizes and exploits the fact that some of the filter coefficients defined at the encoder may possess symmetry relative to other filter coefficients. Accordingly, this disclosure describes techniques in which a first set of the filter coefficients are used to predictively encode a second set of the filter coefficients, thereby exploiting any symmetry between filter coefficients. Rather than communicate all of the filter coefficients to the decoding device, the encoding device may communicate the first set of filter coefficients and difference values associated with the second set of filter coefficients. Using this information, the decoder may be able to reconstruct all of the filter coefficients. In some cases, if exact symmetry is imposed, the need to send the difference values may be eliminated and the decoder may be able to derive the second set of filter coefficients from the first set of filter coefficients. | 07-09-2009 |
20090175349 | LAYERED ENCODED BITSTREAM STRUCTURE - This disclosure describes techniques for generating a layered encoded bitstream structure that exploits correlation in header information among video blocks of a coded unit of video data. A video encoder configured to operate in accordance with the techniques of this disclosure separates header information of the video blocks of a slice (or other coded unit) from the residual information of the video blocks of the slice. The video encoder run-length encodes the header information of the video blocks to better exploit the correlation of the header information among the blocks of the slice. After encoding the header information of the blocks of the slice, the video encoder encodes the residual information for each of the blocks of the slice. The video encoder transmits the encoded header information as a first sequence of bits and transmits the encoded residual information as second sequence of bits. | 07-09-2009 |
20090213930 | FAST MACROBLOCK DELTA QP DECISION - A system and method for encoding multimedia video is described. As video is encoded a quantization parameter is selected for each macroblock. As described herein, the quantization parameter for each macroblock may be selected by limiting the universe of all possible quantization parameters to a particular range of possible quantization parameter values. This increases the speed of video encoding by reducing the number of quantization parameters that are tested for each video macroblock. | 08-27-2009 |
20090257489 | RATE-DISTORTION QUANTIZATION FOR CONTEXT-ADAPTIVE VARIABLE LENGTH CODING (CAVLC) - In general, this disclosure provides techniques for quantization of the coefficients of video blocks in a manner that can achieve a desirable balance of rate and distortion. The described techniques may analyze a plurality of quantization levels associated with each individual coefficient to select the quantization level for the individual coefficients that results in a lowest coding cost. Since CAVLC does not encode each coefficient independently, the techniques may compute the coding costs for each of the candidate quantization levels associated with the individual coefficients based on quantization levels selected for previously quantized coefficients and estimated (or predicted) quantization levels for subsequent coefficients of a coefficient vector. The quantization levels for each of the coefficients are selected based on computed coding costs to obtain a set of quantized coefficients that minimize a rate-distortion model. | 10-15-2009 |
20090257494 | SYMMETRY FOR INTEROPLATION FILTERING OF SUB-PIXEL POSITIONS IN VIDEO CODING - This disclosure describes filtering techniques applied by an encoder and a decoder during the prediction stage of a video encoding and/or decoding process. The filtering techniques may enhance the accuracy of predictive data used during fractional interpolation, and may improve predictive data of integer blocks of pixels. There are several aspects to this disclosure, including a useful twelve-pixel filter support that may be used for interpolation, techniques that use coefficient symmetry and pixel symmetry to reduce the amount of data needed to be sent between an encoder and a decoder to configure the filter support for interpolation, and techniques for filtering data at integer pixel locations in a manner that is similar to sub-pixel interpolation. Other aspects of this disclosure concern techniques for encoding information in the bitstream to convey the type of filter used, and possibly the filter coefficients used. Predictive coding of filter coefficients is also described. | 10-15-2009 |
20090257499 | ADVANCED INTERPOLATION TECHNIQUES FOR MOTION COMPENSATION IN VIDEO CODING - This disclosure describes various interpolation techniques performed by an encoder and a decoder during the motion compensation process of video coding. In one example, an encoder interpolates pixel values of reference video data based on a plurality of different pre-defined interpolation filters. In this example, the decoder receives a syntax element that identifies an interpolation filter, and interpolates pixel values of reference video data based on the interpolation filter identified by the syntax element. In another example, a method of interpolating predictive video data includes generating half-pixel values based on integer pixel values, rounding the half-pixel values to generate half-pixel interpolated values, storing the half-pixel values as non-rounded versions of the half-pixel values, and generating quarter-pixel values based on the non-rounded versions of the half-pixel values and the integer pixel values. | 10-15-2009 |
20090257500 | OFFSETS AT SUB-PIXEL RESOLUTION - This disclosure describes techniques applied during video encoding and decoding processes. In one example, a method of encoding video data comprises calculating a plurality of offset values for a coded unit of the video data, wherein the offset values are associated with a plurality of different integer and sub-integer pixel locations, applying the offset values to predictive video blocks to generate offset predictive video blocks, and encoding video blocks of the coded unit based on the offset predictive video blocks. In another example, a method of decoding video data comprises receiving a plurality of offset values for a coded unit of the video data, wherein the offset values are associated with a plurality of different integer and sub-integer pixel locations, applying the offset values to predictive video blocks to generate offset predictive video blocks, and decoding video blocks of the coded unit based on the offset predictive video blocks. | 10-15-2009 |
20090257501 | INTERPOLATION-LIKE FILTERING OF INTEGER-PIXEL POSITIONS IN VIDEO CODING - This disclosure describes filtering techniques applied by an encoder and a decoder during the prediction stage of a video encoding and/or decoding process. The filtering techniques may enhance the accuracy of predictive data used during fractional interpolation, and may improve predictive data of integer blocks of pixels. There are several aspects to this disclosure, including a useful twelve-pixel filter support that may be used for interpolation, techniques that use coefficient symmetry and pixel symmetry to reduce the amount of data needed to be sent between an encoder and a decoder to configure the filter support for interpolation, and techniques for filtering data at integer pixel locations in a manner that is similar to sub-pixel interpolation. Other aspects of this disclosure concern techniques for encoding information in the bitstream to convey the type of filter used, and possibly the filter coefficients used. Predictive coding of filter coefficients is also described. | 10-15-2009 |
20090257502 | RATE-DISTORTION DEFINED INTERPOLATION FOR VIDEO CODING BASED ON FIXED FILTER OR ADAPTIVE FILTER - This disclosure describes filtering techniques applied by an encoder and a decoder during the prediction stage of a video encoding and/or decoding process. The filtering techniques may enhance the accuracy of predictive data used during fractional interpolation, and may improve predictive data of integer blocks of pixels. There are several aspects to this disclosure, including a useful twelve-pixel filter support that may be used for interpolation, techniques that use coefficient symmetry and pixel symmetry to reduce the amount of data needed to be sent between an encoder and a decoder to configure the filter support for interpolation, and techniques for filtering data at integer pixel locations in a manner that is similar to sub-pixel interpolation. Other aspects of this disclosure concern techniques for encoding information in the bitstream to convey the type of filter used, and possibly the filter coefficients used. Predictive coding of filter coefficients is also described. | 10-15-2009 |
20090257503 | ADVANCED INTERPOLATION TECHNIQUES FOR MOTION COMPENSATION IN VIDEO CODING - This disclosure describes various interpolation techniques performed by an encoder and a decoder during the motion compensation process of video coding. In one example, an encoder interpolates pixel values of reference video data based on a plurality of different pre-defined interpolation filters. In this example, the decoder receives a syntax element that identifies an interpolation filter, and interpolates pixel values of reference video data based on the interpolation filter identified by the syntax element. In another example, a method of interpolating predictive video data includes generating half-pixel values based on integer pixel values, rounding the half-pixel values to generate half-pixel interpolated values, storing the half-pixel values as non-rounded versions of the half-pixel values, and generating quarter-pixel values based on the non-rounded versions of the half-pixel values and the integer pixel values. | 10-15-2009 |
20090257668 | PREDICTION TECHNIQUES FOR INTERPOLATION IN VIDEO CODING - This disclosure describes filtering techniques applied by an encoder and a decoder during the prediction stage of a video encoding and/or decoding process. The filtering techniques may enhance the accuracy of predictive data used during fractional interpolation, and may improve predictive data of integer blocks of pixels. There are several aspects to this disclosure, including a useful twelve-pixel filter support that may be used for interpolation, techniques that use coefficient symmetry and pixel symmetry to reduce the amount of data needed to be sent between an encoder and a decoder to configure the filter support for interpolation, and techniques for filtering data at integer pixel locations in a manner that is similar to sub-pixel interpolation. Other aspects of this disclosure concern techniques for encoding information in the bitstream to convey the type of filter used, and possibly the filter coefficients used. Predictive coding of filter coefficients is also described. | 10-15-2009 |
20100002770 | VIDEO ENCODING BY FILTER SELECTION - A method and a device are described for selecting between multiple available filters in an encoder to provide a frame having a low error and distortion rate. For each full and sub pixel position, determining whether to use an alternative filter over the default filter during interpolation by estimating the rate distortion gain of using each filter and signaling to the decoder the optimal filter(s) applied to each full and sub-pixel position. In one embodiment, identifying a reference frame and a current frame, interpolating the reference frame using a default filter to create a default interpolated frame, interpolating the reference frame using an alternative filter to create an alternative interpolated frame, determining for each sub-pixel position whether to use the default filter or the alternative filter based on a minimal cost to generate a final reference frame. | 01-07-2010 |
20100008430 | FILTERING VIDEO DATA USING A PLURALITY OF FILTERS - Systems and methods of filtering video data using a plurality of filters are disclosed. In an embodiment, a method includes receiving and decoding a plurality of filters embedded in a video data bitstream at a video decoder. The method includes selecting, based on information included in the video data bitstream, a particular filter of the plurality of filters. The method further includes applying the particular filter to at least a portion of decoded video data of the video data bitstream to produce filtered decoded video data. | 01-14-2010 |
20100074332 | OFFSET CALCULATION IN SWITCHED INTERPOLATION FILTERS - This disclosure describes techniques for adding offset to predictive video blocks during video coding. In one example, a method of encoding a video block includes interpolating a first block of predictive values based on a first reference video unit within a first list of reference data, and a second block of predictive values based on a second reference video unit within a second list of reference data, calculating, for sub-integer pixel positions, a first offset value based on the first block and the current video block, and a second offset value based on the first offset value and the second block, determining a final block of offset values based on the first block of predictive values, the second block of predictive values, the first offset values, and the second offset values, and encoding the current video block based on the final block of offset values. | 03-25-2010 |
20100086025 | QUANTIZATION PARAMETER SELECTIONS FOR ENCODING OF CHROMA AND LUMA VIDEO BLOCKS - This disclosure describes rules that may be applied during block-based video coding to ensure that quantization parameter selections for luma blocks will not adversely affect the quality of chroma blocks. In accordance with this disclosure, rate-controlled video encoding occurs in which quantization parameter changes in luma blocks are pre-evaluated to determine whether such quantization parameter changes in luma blocks will also cause quantization changes for chroma blocks. If quantization parameter changes in the luma blocks will also cause quantization changes for chroma blocks, then that quantization parameter change for luma blocks may be skipped and not evaluated. In this way, secondary effects of quantization parameter changes in the luma blocks (with respect to the chroma blocks) can be avoided. | 04-08-2010 |
20100086027 | EFFICIENT PREDICTION MODE SELECTION - In generally, techniques are described for efficiently selecting a prediction mode by which to predict predictive video data from reference video data. In particular, an apparatus may include a memory that stores at least a first and second reference coded unit that each includes a first and second reference video data unit, respectively. The apparatus may further comprise a motion compensation unit that performs default weighted prediction to predict a first version of a predictive video data unit from the first and second reference video data units and calculates an offset value for the first version of the predictive video data unit. The motion compensation unit may then perform, based on the calculated offset value, either implicit weighted prediction or explicit weighted prediction to predict a second version of the predictive video data unit and encode the predictive video data unit as either the first or second version. | 04-08-2010 |
20100086029 | VIDEO CODING WITH LARGE MACROBLOCKS - Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels, for example, 64×64 pixels. In one example, an apparatus includes a video encoder configured to encode a video block having a size of more than 16×16 pixels, generate block-type syntax information that indicates the size of the block, and generate a coded block pattern value for the encoded block, wherein the coded block pattern value indicates whether the encoded block includes at least one non-zero coefficient. The encoder may set the coded block pattern value to zero when the encoded block does not include at least one non-zero coefficient or set the coded block pattern value to one when the encoded block includes a non-zero coefficient. | 04-08-2010 |
20100086030 | VIDEO CODING WITH LARGE MACROBLOCKS - Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels, for example, 64×64 pixels. Each macroblock may be partitioned into two or more partitions, and two or more of the partitions may be encoded using different modes. In one example, an apparatus includes a video encoder configured to receive a video block having a size of more than 16×16 pixels, partition the block into partitions, encode one of the partitions using a first encoding mode, encode another of the partitions using a second encoding mode different from the first encoding mode, and generate block-type syntax information that indicates the size of the block and identifies the partitions and the encoding modes used to encode the partitions. | 04-08-2010 |
20100086031 | VIDEO CODING WITH LARGE MACROBLOCKS - Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels, for example, 64×64 pixels. In one example, an apparatus includes a video encoder configured to receive a video coding unit, determine a first rate-distortion metric for encoding the video coding unit using first video blocks with sizes of 16×16 pixels, determine a second rate-distortion metric for encoding the video coding unit using second video blocks with sizes of more than 16×16 pixels, encode the video coding unit using the first video blocks when the first rate-distortion metric is less than second rate-distortion metric, and encode the video coding unit using the second video blocks when the second rate-distortion metric is less than the first rate-distortion metric. | 04-08-2010 |
20100086032 | VIDEO CODING WITH LARGE MACROBLOCKS - Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels. In one example, an apparatus includes a video encoder configured to encode a coded unit comprising a plurality of video blocks, wherein at least one of the plurality of video blocks comprises a size of more than 16×16 pixels and to generate syntax information for the coded unit that includes a maximum size value, wherein the maximum size value indicates a size of a largest one of the plurality of video blocks in the coded unit. The syntax information may also include a minimum size value. In this manner, the encoder may indicate to a decoder the proper syntax decoder to apply to the coded unit. | 04-08-2010 |
20100086049 | VIDEO CODING USING TRANSFORMS BIGGER THAN 4X4 AND 8X8 - In a video processing system, a method and system for applying transforms larger than 8×8 and non-rectangular transforms, and generating transform size syntax elements indicative of the transforms for video decoding are provided. The transform size syntax element may be generated by an encoder based on a prediction block size of a video block and the contents of the video block. Further, the transform size syntax element may be generated according to a set of rules to select from 4×4, 8×8, and larger transform sizes during an encoding process. A decoder may perform an inverse transform based on the transform size syntax element and the rules used by the encoder. The transform size syntax element may be transmitted to the decoder as part of the encoded video bitstream. | 04-08-2010 |
20100098156 | WEIGHTED PREDICTION BASED ON VECTORIZED ENTROPY CODING - This disclosure describes methods that control the selection of predictive coding techniques for enhancement layer video blocks based on characteristics of vectorized entropy coding for such enhancement layer video blocks. In accordance with this disclosure, the predictive techniques used for predictive-based video coding of enhancement layer video blocks are dependent upon the vectorized entropy coding used for such enhancement layer the video blocks. For each coded unit, predictive coding techniques (e.g. weighted or non-weighted prediction) may be selected depending upon whether the vectorized entropy coding defines a single vector for the video blocks of that coded unit or multiple vectors for the video blocks of that coded unit. | 04-22-2010 |
20100111182 | DIGITAL VIDEO CODING WITH INTERPOLATION FILTERS AND OFFSETS - This disclosure describes techniques for encoding digital video data using interpolation filters and offsets. An encoder may be configured to select interpolation filters for sub-pixel precision motion estimation based on historical interpolation results obtained for previously encoded video units, such as frames or slices. The encoder also may be configured to compute and assign offsets to the sub-pixel positions after interpolation based on differences between a reference unit and the unit to be coded. The computation and assignment of offsets may be performed before motion estimation. Motion estimation may be refined so that the motion search considers sub-pixel positions to which offsets have been previously added and evaluates sub-pixel positions that have a non-zero offset. In some cases, interpolation filter selection, offset computation, and/or refined motion estimation for a given unit may be performed in a single encoding pass. | 05-06-2010 |
20100158103 | COMBINED SCHEME FOR INTERPOLATION FILTERING, IN-LOOP FILTERING AND POST-LOOP FILTERING IN VIDEO CODING - In one example, this disclosure describes filtering techniques for filtering of video blocks of a video unit. The filtering techniques may select one or more different types of filtering for each video block of the video unit based on various factors such as whether the video block is inter-coded or intra-coded, and whether adaptive interpolations were preformed during a motion compensation process during the encoding of the video block. When adaptive interpolations were performed, the adaptive interpolations may provide a level of filtering that renders additional filtering unnecessary or undesirable in some cases. | 06-24-2010 |
20100260262 | EFFICIENT TRANSCODING OF B-FRAMES TO P-FRAMES - This disclosure describes techniques for efficient transcoding from a first format that supports I-units, P-units and B-units to a second format that supports I-units and P-units but does not support the B-units. In particular, this disclosure provides techniques for converting B-frames or B-slices of the first format into P-frames or P-slices of the second format. The techniques avoid the need to decode and re-encode that frames or slices. Instead, residuals associated with the B-video blocks in the first format are augmented and made to be dependent upon only one of the two lists associated with the B-video blocks so that such B-video blocks in the first format can be redefined as P-video blocks in the second format. | 10-14-2010 |
20100296579 | ADAPTIVE PICTURE TYPE DECISION FOR VIDEO CODING - A video encoding apparatus determines whether to encode a key frame of a group of pictures using a bi-directional prediction mode. In one example, a video encoding apparatus includes a mode select unit configured to generate a virtual key frame for a current group of pictures based on a previous key frame of a previous group of pictures and a next key frame of a next group of pictures, calculate an error value representing error between a current key frame of the current group of pictures and the virtual key frame, and determine whether the error value exceeds a threshold value, and a video encoder configured to encode the current key frame using a bi-directional prediction encoding mode when the error value does not exceed the threshold value. The video encoder may comprise the mode select unit, or a preprocessing unit of the apparatus may comprise the mode select unit. | 11-25-2010 |
20100309286 | ENCODING OF THREE-DIMENSIONAL CONVERSION INFORMATION WITH TWO-DIMENSIONAL VIDEO SEQUENCE - This disclosure describes techniques for encoding a two-dimensional (2D) video sequence of video frames along with three-dimensional (3D) conversion information comprising a set of parameters that can be applied to each of the video frames of the 2D sequence to generate 3D video data. The set of parameters may comprise a relatively small amount of data that can be applied to each of the original video frames of the 2D sequence to generate secondary views of each of the original video frames. The original video frames and the secondary views may collectively define a stereoscopic 3D video sequence. The 2D sequence and the set of parameters may comprise significantly less data than would otherwise be needed to communicate a 3D sequence. This disclosure also describes some exemplary syntax that may be used to encode the set of parameters in an effective and efficient manner. | 12-09-2010 |
20100316134 | ASSEMBLING MULTIVIEW VIDEO CODING SUB-BISTREAMS IN MPEG-2 SYSTEMS - A demultiplexer may assemble view components of sub-bitstreams. In one example, an apparatus comprises a demultiplexer that produces a multiview video coding (MVC) standard compliant bitstream from a received bitstream comprising a primary sub-bitstream and an embedded sub-bitstream. To produce the MVC standard compliant bitstream, the demultiplexer determines whether a view component of the primary sub-bitstream has a view order index that is greater than a view order index of a view component of the embedded sub-bitstream, and to add the view component from the sub-bitstream for which the view order index is lower to the produced bitstream. The received bitstream may comprise delimiter network abstraction layer (NAL) units between each view component to differentiate the view components. The apparatus may further comprise a video decoder to decode the bitstream produced by the demultiplexer. | 12-16-2010 |
20100329329 | 8-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing an 8-point inverse discrete cosine transform (IDCT). An apparatus comprising an 8-point inverse discrete cosine transform (IDCT) hardware unit may implement these techniques to transform media data from a frequency domain to a spatial domain. The 8-point IDCT hardware unit includes an even portion comprising factors A, B that are related to a first scaled factor (μ) in accordance with a first relationship. The 8-point IDCT hardware unit also includes an odd portion comprising third, fourth, fifth and sixth internal factors (G, D, E, Z) that are related to a second scaled factor (η) in accordance with a second relationship. The first relationship relates the first scaled factor to the first and second internal factors. The second relationship relates the second scaled factor to the third, fourth, fifth and sixth internal factors. | 12-30-2010 |
20100329338 | LOW COMPLEXITY B TO P-SLICE TRANSCODER - A system and method for transcoding compressed multimedia video is described. Particularly, a system and method for converting Bi-Predictive frame to transcoded Predictive frames, is disclosed. Present embodiments accomplish this conversion with minimal additional error, thereby providing an efficient means for maintaining video quality even after transcoding. | 12-30-2010 |
20100329342 | VIDEO CODING BASED ON FIRST ORDER PREDICTION AND PRE-DEFINED SECOND ORDER PREDICTION MODE - This disclosure describes video encoding and decoding techniques in which a first order prediction process and a second order prediction process are used in combination to generate predictive video blocks for video coding. First order prediction may be similar to conventional motion estimation and motion compensation that generates residual video blocks. The second order prediction may involve a process similar to conventional intra-prediction, but is performed on the residual video blocks. The techniques of this disclosure may pre-define the second order prediction to a specific mode, such as a mode similar to the intra-DC mode used in intra coding. In addition, the techniques of this disclosure may combine aspects of the first order and second order prediction into a single process so that the effects of second order prediction on the residuals are taken into account during the first order prediction process, which may improve compression. | 12-30-2010 |
20110002388 | TEMPLATE MATCHING FOR VIDEO CODING - In one aspect of this disclosure, template matching motion prediction is applied to B-frames. In another aspect of this disclosure, template matching motion prediction as applied to video block coding may include generating a template offset, generating a weighted sum of absolute differences, selecting a number of hypotheses used to encode video blocks based on the cost associated with the number of hypotheses and signaling, with a new syntax, to a decoder, the number of hypotheses used in encoding, rejecting hypotheses if the difference in value between a hypothesis and a reference hypothesis is greater than a threshold value, and/or generating the content of a sub-block that does not have reconstructed data available by combining motion-compensated prediction and luma residuals. | 01-06-2011 |
20110007799 | NON-ZERO ROUNDING AND PREDICTION MODE SELECTION TECHNIQUES IN VIDEO ENCODING - In one aspect of this disclosure, techniques are described for selecting among default weighted prediction, implicit weighted prediction, and explicit weighted prediction. In this context, techniques are also described for adding offset to prediction data, e.g., using the format of explicit weighted prediction to allow for offsets to predictive data that is otherwise determined by implicit or default weighted prediction. | 01-13-2011 |
20110007802 | NON-ZERO ROUNDING AND PREDICTION MODE SELECTION TECHNIQUES IN VIDEO ENCODING - In one aspect of this disclosure, rounding adjustments to bi-directional predictive data may be purposely eliminated to provide predictive data that lacks any rounding bias. In this case, rounded and unrounded predictive data may both be considered in a rate-distortion analysis to identify the best data for prediction of a given video block. In another aspect of this disclosure, techniques are described for selecting among default weighted prediction, implicit weighted prediction, and explicit weighted prediction. In this context, techniques are also described for adding offset to prediction data, e.g., using the format of explicit weighted prediction to allow for offsets to predictive data that is otherwise determined by implicit or default weighted prediction. | 01-13-2011 |
20110007803 | DIFFERENT WEIGHTS FOR UNI-DIRECTIONAL PREDICTION AND BI-DIRECTIONAL PREDICTION IN VIDEO CODING - In one aspect of this disclosure, techniques are described for the decoupling of uni-directional and bi-directional prediction weights, particularly for explicit weighted predictions of video blocks within a B-unit. According to this disclosure, explicit weights communicated in the bitstream may be applied by a decoder for explicit bi-directional prediction, but different weights (which may be default weights or separately defined explicit unidirectional weights) may be used for explicit uni-directional prediction. The described techniques may improve video quality relative to techniques that use the same explicit weights for explicit bi-directional prediction and explicit uni-directional prediction within a B-unit. | 01-13-2011 |
20110032999 | SIGNALING CHARACTERISTICS OF AN MVC OPERATION POINT - Source and destination video devices may use data structures that signal details of an operation point for an MPEG-2 (Motion Picture Experts Group) System bitstream. In one example, an apparatus includes a multiplexer that constructs a data structure corresponding to a multiview video coding (MVC) operation point of an MPEG-2 (Motion Picture Experts Group) System standard bitstream, wherein the data structure signals a rendering capability value that describes a rendering capability to be satisfied by a receiving device to use the MVC operation point, a decoding capability value that describes a decoding capability to be satisfied by the receiving device to use the MVC operation point, and a bitrate value that describes a bitrate of the MVC operation point, and that includes the data structure as part of the bitstream, and an output interface that outputs the bitstream comprising the data structure. | 02-10-2011 |
20110064146 | MEDIA EXTRACTOR TRACKS FOR FILE FORMAT TRACK SELECTION - A video coding apparatus may be configured to utilize media extractors in a media extractor track that reference two or more non-consecutive network access layer (NAL) units of a separate track. An example apparatus includes a multiplexer to construct a first track including a video sample comprising NAL units, based on encoded video data, wherein the video sample is included in an access unit, construct a second track including an extractor that identifies at least first one of the NAL units in the video sample of the first track, and wherein the extractor identifies a second NAL unit of the access unit, wherein the first identified NAL unit and the second identified NAL unit are non-consecutive, and include the first track and the second track in a video file conforming at least in part to ISO base media file format. The identified NAL units may be in separate tracks. | 03-17-2011 |
20110096832 | DEPTH MAP GENERATION TECHNIQUES FOR CONVERSION OF 2D VIDEO DATA TO 3D VIDEO DATA - This disclosure describes techniques for generating depth maps for video units, such as video frames or slices video frames. The techniques may be performed by a video encoder in order to convert two-dimensional (2D) video to three-dimensional (3D) video. The techniques may alternatively be performed by a video decoder in order to convert received 2D video to 3D video. The techniques may use a combination of motion and color considerations in the depth map generation process. | 04-28-2011 |
20110099594 | STREAMING ENCODED VIDEO DATA - A source device may signal characteristics of a media presentation description (MPD) file such that a destination device may select one of a number of presentations corresponding to the MPD file and retrieve one or more video files of the selected presentation. In one example, an apparatus for transporting encoded video data includes a management unit configured to receive encoded video data comprising a number of video segments and forms a presentation comprising a number of video files, each of the video files corresponding to a respective one of the video segments, and a network interface configured to, in response to a request specifying a temporal section of the video data, output at least one of the video files corresponding to the number of video segments of the requested temporal section. A client may request temporally sequential fragments from different ones of the presentations. | 04-28-2011 |
20110150078 | 8-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing an 8-point discrete cosine transform (DCT). An apparatus comprising an 8-point discrete cosine transform (DCT) hardware unit may implement these techniques to transform media data from a spatial domain to a frequency domain. The 8-point DCT hardware unit includes an even portion comprising factors A, B that are related to a first scaled factor (μ) in accordance with a first relationship. The 8-point DCT hardware unit also includes an odd portion comprising third, fourth, fifth and sixth internal factors (G, D, E, Z) that are related to a second scaled factor (η) in accordance with a second relationship. The first relationship relates the first scaled factor to the first and second internal factors. The second relationship relates the second scaled factor to the third internal factor and a fourth internal factor, as well as, the fifth internal factor and a sixth internal factor. | 06-23-2011 |
20110150079 | 16-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing a 16-point discrete cosine transform (DCT) that is capable of applying multiple IDCT of different sizes. For example, an apparatus comprising a 16-point discrete cosine transform of type II (DCT-II) unit may implement the techniques of this disclosure. The 16-point DCT-II unit performs these DCTs-II of different sizes to transform data from a spatial to a frequency domain. The 16-point DCT-II unit includes an 8-point DCT-II unit that performs one of the DCTs-II of size 8 and a first 4-point DCT-II unit that performs one of the DCTs-II of size 4. The 8-point DCT-II unit includes the first 4-point DCT-II unit. The 16-point DCT-II unit also comprises an 8-point DCT-IV unit that includes a second 4-point DCT-II unit and a third 4-point DCT-II unit. Each of the second and third 4-point DCT-II units performs one of the DCTs-II of size 4. | 06-23-2011 |
20110153699 | 16-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing a 16-point inverse discrete cosine transform (IDCT) that is capable of applying multiple IDCTs of different sizes. For example, an apparatus comprising a 16-point inverse discrete cosine transform of type II (IDCT-II) unit may implement the techniques of this disclosure. The 16-point IDCT-II unit performs these IDCTs-II of different sizes to transform data from a spatial to a frequency domain. The 16-point IDCT-II unit includes an 8-point IDCT-II unit that performs one of the IDCTs-II of size 8 and a first 4-point IDCT-II unit that performs one of the IDCTs-II of size 4. The 8-point IDCT-II unit includes the first 4-point DCT-II unit. The 16-point IDCT-II unit also comprises an inverse 8-point DCT-IV unit that includes a second 4-point IDCT-II unit and a third 4-point IDCT-II unit. Each of the second and third 4-point IDCT-II units performs one of the IDCTs-II of size 4. | 06-23-2011 |
20110194613 | VIDEO CODING WITH LARGE MACROBLOCKS - A video coder may utilize large macroblocks having more than 16×16 pixels. Syntax for the large macroblocks may define whether a bitstream includes large macroblocks, such as superblocks having 64×64 pixels or bigblocks having 32×32 pixels. The syntax may be included in a slice header or a sequence parameter set. The large macroblocks may also be encoded according to a large macroblock syntax. The bitstream may further include syntax data that indicates a level value based on whether the bitstream includes any of the large macroblocks, for example, as a smallest-sized luminance prediction block. A decoder may use the level value to determine whether the decoder is capable of decoding the bitstream. | 08-11-2011 |
20110196982 | STREAMING ENCODED VIDEO DATA - A source device may signal characteristics of a media presentation description (MPD) file such that a destination device may select one of a number of presentations corresponding to the MPD file and retrieve one or more video files of the selected presentation. In one example, an apparatus for transporting encoded video data includes a management unit configured to receive encoded video data comprising a number of video segments and forms a presentation comprising a number of video files, each of the video files corresponding to a respective one of the video segments, and a network interface configured to, in response to a request specifying a temporal section of the video data, output at least one of the video files corresponding to the number of video segments of the requested temporal section. A client may request temporally sequential fragments from different ones of the presentations. | 08-11-2011 |
20110200097 | ADAPTIVE TRANSFORM SIZE SELECTION FOR GEOMETRIC MOTION PARTITIONING - In one example, an apparatus includes a video encoder configured to partition a block of video data into a first geometric partition and a second geometric partition using a geometric motion partition line, wherein the block comprises N×N pixels, divide the block of video data into four equally-sized, non-overlapping (N/2)×(N/2) sub-blocks, and encode at least one of the sub-blocks through which the geometric motion partition line passes using a transform size smaller than (N/2)×(N/2). The video encoder may determine transform sizes for the sub-blocks based on whether the geometric motion partition line passes through the sub-blocks. In one example, a video decoder may inverse transform the sub-blocks, and may determine transform sizes for the sub-blocks based on whether the geometric motion partition line passes through the sub-blocks. | 08-18-2011 |
20110200108 | CHROMINANCE HIGH PRECISION MOTION FILTERING FOR MOTION INTERPOLATION - A video coding unit may be configured to encode or decode chrominance blocks of video data by reusing motion vectors for corresponding luminance blocks. A motion vector may have greater precision for chrominance blocks than luminance blocks, due to downsampling of chrominance blocks relative to corresponding luminance blocks. The video coding unit may interpolate values for a reference chrominance block by selecting interpolation filters based on the position of the pixel position pointed to by the motion vector. For example, a luminance motion vector may have one-quarter-pixel precision and a chrominance motion vector may have one-eighth-pixel precision. There may be interpolation filters associated with the quarter-pixel precisions. The video coding unit may use interpolation filters either corresponding to the pixel position or neighboring pixel positions to interpolate a value for the pixel position pointed to by the motion vector. | 08-18-2011 |
20110200109 | FIXED POINT IMPLEMENTATION FOR GEOMETRIC MOTION PARTITIONING - In one example, an apparatus includes a video encoder configured to partition a block of video data into a first partition and a second partition using a geometric motion partition line, calculate a slope value and a y-intercept value of the geometric motion partition line, wherein the slope value and the y-intercept value comprise integer values, calculate a mask indicative of pixels of the block in the first partition and pixels of the block in the second partition, encode the first partition and the second partition based on the mask, and output the encoded first partition, the encoded second partition, the slope value, and the y-intercept value. This may allow for a fixed point implementation. A video decoder may receive the slope and y-intercept values to calculate the mask and decode the block based on the mask. | 08-18-2011 |
20110200110 | SMOOTHING OVERLAPPED REGIONS RESULTING FROM GEOMETRIC MOTION PARTITIONING - In one example, an apparatus includes a video encoder configured to partition a block of video data into a first partition and a second partition using a geometric motion partition line, calculate a prediction value of a pixel in a transition region of the block using a filter that applies a value for at least one neighboring pixel from the first partition and a value for at least one neighboring pixel from the second partition, calculate a residual value of the pixel in the transition region of the block based on the prediction value of the pixel in the transition region, and output the residual value of the pixel. In one example, a video decoder may use a similar filter to decode an the encoded block after receiving the residual value for the encoded block, and using a definition of the geometric motion partition line. | 08-18-2011 |
20110200111 | ENCODING MOTION VECTORS FOR GEOMETRIC MOTION PARTITIONING - In one example, an apparatus includes a video encoder configured to partition a block of video data into a first partition and a second partition using a geometric motion partition line, determine a first motion vector for the first partition and a second motion vector for the second partition, encode the first motion vector based on a first motion predictor selected from motion vectors for blocks neighboring the first partition, encode the second motion vector based on a second motion predictor selected from motion vectors for blocks neighboring the second partition, wherein the blocks neighboring the second partition are determined independently of the blocks neighboring the first partition, and output the encoded first and second motion vectors. A video decoder may similarly decode the motion vectors based on determining the first and second motion predictors for the first and second partitions. | 08-18-2011 |
20110206123 | BLOCK TYPE SIGNALLING IN VIDEO CODING - In one example, this disclosure describes video encoding and decoding techniques applicable to bi-directional prediction. For signaling from an encoder to a decoder a block type for a video block, an encoder can transmit to the decoder two or more separate syntax elements, where one of the two or more separate syntax elements identifies a partition size for the video block and another of the two or more separate syntax elements identifies a prediction direction for a partition of the video block. The separate syntax elements can be individually encoded and transmitted from the encoder to the decoder. | 08-25-2011 |
20110206125 | ADAPTIVE MOTION RESOLUTION FOR VIDEO CODING - A video encoder may encode video data by adaptively selecting between one-eighth-pixel and one-quarter-pixel precision motion vectors, and signal the selected precision. In one example, an apparatus includes a video encoder to encode a block of video data using a one-eighth-pixel precision motion vector when use of the one-eighth-pixel precision motion vector is determined to be preferable for the block over a one-quarter-pixel precision motion vector, and to generate a signal value indicative of the use of the one-eighth-pixel precision motion vector for the block, and an output interface to output the encoded block and the signal value. A video decoder may be configured to receive the signal value and the encoded block, analyze the signal value to determine whether the block was encoded using one-eighth-pixel precision or one-quarter-pixel precision, and decode the block based on the determination. | 08-25-2011 |
20110248873 | VARIABLE LENGTH CODES FOR CODING OF VIDEO DATA - A method and system for entropy coding can comprise, in response to detecting a first symbol combination comprising first run information indicating a first number of contiguous zero coefficients is greater than a cut-off-run value, assigning a first codeword to a first symbol combination, wherein the first codeword comprises an escape code from a first-level VLC table; and in response to a second symbol combination comprising second run information indicating a second number of contiguous zero coefficients is less than or equal to the cut-off-run value, assigning a second codeword to the second symbol combination, wherein the second codeword is from the first-level VLC table. The system and method can further comprise collecting coding statistics for a set of candidate symbol combinations and adjusting a mapping between codewords of the first-level VLC table and a subset of the set of candidate symbol combinations based on the coding statistics. | 10-13-2011 |
20110249721 | VARIABLE LENGTH CODING OF CODED BLOCK PATTERN (CBP) IN VIDEO COMPRESSION - This disclosure describes techniques for coding video data. As one example, this disclosure describes a coded block pattern (CBP) for a coding unit (CU) of video data that indicates whether or not each of a luminance component (Y), a first chrominance component (U), and a second chrominance component (V) include at least one non-zero coefficient. According to another example, this disclosure describes a CBP that indicates whether respective blocks of a CU include at least on non-zero coefficient. The CBP described herein may be mapped to a single variable length code (VLC) code word. The VLC code word may be used by a coder to code the CU of video data. | 10-13-2011 |
20110249737 | MIXED TAP FILTERS - During the prediction stage of a video encoding and/or decoding process, a video coder can use relatively longer filters for certain motion vectors pointing to certain sub-pixel positions and relatively shorter filters for motion vectors pointing to other sub-pixel positions, where a longer filter generally refers to an interpolation filter with a greater number of filter coefficients, also called taps, while a shorter filter generally refers to an interpolation filter with fewer taps. | 10-13-2011 |
20110249745 | BLOCK AND PARTITION SIGNALING TECHNIQUES FOR VIDEO CODING - A video block syntax element indicates whether all of the partitions of a video block are predicted based on a same reference list and no greater than quarter-pixel accuracy is used. If the video block syntax element is set, partition-level signaling of the reference lists is avoided. If the video block syntax element is not set, partition-level signaling of the reference lists occurs. If the video block syntax element is set, partition-level syntax elements may be used for each of the partitions of the video block, wherein the partition-level syntax elements each identify one of the reference lists and motion vector accuracy for a given one of the partitions. | 10-13-2011 |
20110249754 | VARIABLE LENGTH CODING OF CODED BLOCK PATTERN (CBP) IN VIDEO COMPRESSION - In one example, this disclosure describes method of coding video data. The method comprises coding a block of video data as one or more luminance blocks of transform coefficients and one or more chrominance blocks of transform coefficients, and coding a coded block pattern (CBP) for the block of video data. The CBP comprises syntax information that identifies whether non-zero data is included in each of the luminance blocks and each of the chrominance blocks. Coding the CBP includes selecting one or more variable length coding (VLC) tables based on a transform size used in performing one or more transforms on the one or more luminance blocks. | 10-13-2011 |
20110280311 | ONE-STREAM CODING FOR ASYMMETRIC STEREO VIDEO - An asymmetric frame of a coded video bitstream may include a first resolution picture of a left view and a reduced resolution picture of a right view, where the left and right views form a stereo view pair for three-dimensional video playback. In addition, the reduced resolution frame may be predicted relative to a picture of the left view. In one example, an apparatus includes a video encoder configured to encode a first picture of a first view of a scene to produce an encoded picture with a first resolution, encode at least a portion of a second picture of a second view of the scene relative to a reference picture of the first view to produce an encoded picture with a reduced resolution relative to the first resolution, and output the encoded first resolution picture and the encoded reduced resolution picture in a common bitstream. | 11-17-2011 |
20110280316 | FRAME PACKING FOR ASYMMETRIC STEREO VIDEO - An asymmetric frame of a coded video bitstream may include a full resolution picture of a left view and a reduced resolution picture of a right view, where the left and right views form a stereo view pair for three-dimensional video playback. In one example, an apparatus includes a video encoder configured to receive a first picture of a first view of a scene having a first resolution, receive a second picture of a second view of the scene having a reduced resolution relative to the first resolution, form an asymmetric frame comprising the first picture and the second picture, and encode the asymmetric frame. In this manner, decoders of varying capabilities may receive the same bitstream, and the bitstream may consume less bandwidth than one or more bitstreams having full resolution pictures of a stereo view pair. The bitstream may have better quality than a bitstream having subsampled pictures. | 11-17-2011 |
20110304618 | CALCULATING DISPARITY FOR THREE-DIMENSIONAL IMAGES - An apparatus may calculate disparity values for pixels of a two-dimensional image based on depth information for the pixels and generate a second image using the disparity values. The calculation of the disparity value for a pixel may correspond to a linear relationship between the depth of the pixel and a corresponding disparity range. In one example, an apparatus for rendering three-dimensional image data includes a view synthesizing unit configured to calculate disparity values for a plurality of pixels of a first image based on depth information associated with the plurality of pixels and disparity ranges to which the depth information is mapped, wherein the disparity values describe horizontal offsets for corresponding ones of a plurality of pixels for a second image. The apparatus may receive the first image and depth information from a source device. The apparatus may produce the second image using the first image and disparity values. | 12-15-2011 |
20110310976 | Joint Coding of Partition Information in Video Coding - In one example, a video decoder is configured to receive a value for a coding unit of video data, wherein the coding unit is partitioned into a plurality of sub-coding units, determine whether the sub-coding units are partitioned into further sub-coding units based on the value, and decode the sub-coding units and the further sub-coding units. In another example, a video encoder is configured to partition a coding unit of video data into a plurality of sub-coding units, determine whether to partition the sub-coding units into further sub-coding units, and encode the coding unit to include a value that indicates whether the sub-coding units are partitioned into the further sub-coding units. | 12-22-2011 |
20110317757 | INTRA PREDICTION MODE SIGNALING FOR FINER SPATIAL PREDICTION DIRECTIONS - A video encoder selects a prediction mode for a current video block from a plurality of prediction modes that includes both main modes and finer directional intra spatial prediction modes, also referred to as non-main modes. The video encoder may be configured to encode the selection of the prediction mode of the current video block based on prediction modes of one or more previously encoded video blocks of the series of video blocks. The selection of a non-main mode can be coded as a combination of a main mode and a refinement to that main mode. A video decoder may also be configured to perform the reciprocal decoding function of the encoding performed by the video encoder. Thus, the video decoder uses similar techniques to decode the prediction mode for use in generating a prediction block for the video block. | 12-29-2011 |
20110317760 | SIGNALING VIDEO SAMPLES FOR TRICK MODE VIDEO REPRESENTATIONS - This disclosure describes techniques relevant to HTTP streaming of media data. According to these techniques, a server device may signal a byte range for at least one intra-decodable (I-frame) of a video fragment. According to the techniques of this disclosure, a client device may communicate a request to a server device to retrieve the at least one I-frame based on the signaled byte range, and use the retrieved I-frame to provide a high speed version of a video presentation that includes the at least one I-frame. A high speed version of a video presentation may be a trick mode of the video presentation, such as a fast forward or fast rewind version of the video presentation. | 12-29-2011 |
20110317764 | INCLUSION OF SWITCHED INTERPOLATION FILTER COEFFICIENTS IN A COMPRESSED BIT-STREAM - In one example, an encoder may apply a plurality of pre-defined interpolation filters to units of video data, such as frames of reference video, in order to generate a plurality of different interpolated prediction data. The encoder may also at times determine that a new interpolation filter or set of interpolation filters might improve coding quality by either improving video compression or improving reconstructed image quality. The encoder may also signal to a video decoder whether one of the pre-defined interpolation filters was used or a new set of interpolation filters was used. The encoder may also signal to a video decoder whether to continue using the new set of interpolation filters, or whether to revert back to using the pre-defined set of interpolation filters. A video decoder can decode video data based on data received from the video encoder. | 12-29-2011 |
20110317771 | SIGNALING RANDOM ACCESS POINTS FOR STREAMING VIDEO DATA - This disclosure describes techniques relevant to HTTP streaming of media data. According to these techniques, a server device may signal an open decoding refresh (ODR) random access point (RAP) for a movie segmentation of a movie representation. At least one frame of the media segmentation following the ODR RAP frame in decoding order may not be correctly decoded, and wherein each frame of the media segmentation following the ODR RAP frame in display order can be correctly decoded without relying on content of frames prior to the ODR RAP in display order. According to the techniques of this disclosure, a client device may communicate a request to a server device for the streaming of media data based on signaling of the ODR RAP. Also according to the techniques of this disclosure, a client device may commence decoding and/or playback of the movie representation based on signaling of the ODR RAP. | 12-29-2011 |
20120008675 | CODING SYNTAX ELEMENTS FOR ADAPTIVE SCANS OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - In one example, an apparatus for encoding video data includes a video encoder configured to scan a two-dimensional block of transform coefficients to produce a one-dimensional vector of the transform coefficients, determine values indicative of whether the transform coefficients in the one-dimensional vector are significant; and entropy encode at least one of the values using a context model selected based on at least a percentage of significant coefficients in a predetermined number of the values encoded before the at least one of the values. | 01-12-2012 |
20120008682 | VIDEO CODING USING DIRECTIONAL TRANSFORMS - In one example, an apparatus for encoding video data includes a video encoder configured to calculate a residual block for a block of video data based on a predicted block formed using an intra-prediction mode, and transform the residual block using a transform mapped from the intra-prediction mode. In another example, an apparatus includes video encoder configured to receive an indication of a first intra-prediction mode in a first set of intra-prediction modes for a block of video data, determine a second intra-prediction mode from a second set of intra-prediction modes, smaller than the first set of intra-prediction modes, to which the first intra-prediction mode is mapped, determine a directional transform to which the second intra-prediction mode is mapped, and apply the directional transform to residual data of the block. | 01-12-2012 |
20120008683 | SIGNALING SELECTED DIRECTIONAL TRANSFORM FOR VIDEO CODING - In one example, an apparatus for encoding video data includes a video encoder configured to select an intra-prediction mode to use to encode a block of video data, determine whether the block includes a sub-block of a size for which multiple transforms are possible based on the size of the sub-block and the selected intra-prediction mode, when the block includes the sub-block of the size for which multiple transforms are possible based on the size of the sub-block and the selected intra-prediction mode, select one of the multiple possible transforms, transform the sub-block using the selected one of the multiple possible transforms, and provide an indication of the selected one of the multiple possible transforms for the size of the block. | 01-12-2012 |
20120013746 | SIGNALING DATA FOR MULTIPLEXING VIDEO COMPONENTS - A server may provide information describing characteristics of audio and video components to a client, separately from encoded samples of the audio and video components themselves. The client may use the information to select components, and then to request the selected components, e.g., in accordance with a streaming network protocol. In one example, an apparatus for sending encapsulated video data includes a processor configured to determine characteristics for components of a plurality of representations of video content, wherein the characteristics comprise at least one of a frame rate, a profile indicator, a level indicator, and dependencies between the components, and one or more interfaces configured to send the characteristics to a client device, receive a request from the client device for at least one of the components after sending the characteristics, and send the requested components to the client device in response to the request. | 01-19-2012 |
20120014433 | ENTROPY CODING OF BINS ACROSS BIN GROUPS USING VARIABLE LENGTH CODEWORDS - This disclosure describes techniques for entropy coding bins representing video data symbols with reduced bottlenecks in the entropy coding process. The techniques of this disclosure enable an entropy coding device to perform entropy coding of bins grouped into bin subsets from across different bin groups, e.g., context groups or probability groups, using variable length codewords. In one example, the bins may be assigned to context groups with no context dependencies between the context groups. In another example, the bins may be assigned to probability groups associated with different intervals of probability states. The bins may be grouped into the bin subsets according to determined formations of the bin subsets. In this way, the entropy coding device may reduce an amount of bin and codeword buffering by efficiently forming the bin subsets and designing variable length codewords for each of the bin subsets. | 01-19-2012 |
20120014455 | Variable Localized Bit-Depth Increase for Fixed-Point Transforms in Video Coding - This disclosure describes techniques for mitigating rounding errors in a fixed-point transform associated with video coding by applying a variable localized bit-depth increase at the transform. More specifically, the techniques include selecting a constant value based on a size of a fixed-point transform in a video coding device and applying a variable localized bit-depth increase at the transform with a value equal to the constant value. Applying the variable localized bit-depth increase includes left-shifting a transform input signal by a number of bits equal to the constant value before the fixed-point transform, and right-shifting a transform output signal by a number of bits equal to the constant value after the fixed-point transform. The constant value is selected from a plurality of constant values stored on the video coding device. Each of the constant values is pre-calculated for one of a plurality of different transform sizes supported by the video coding. | 01-19-2012 |
20120016965 | VIDEO SWITCHING FOR STREAMING VIDEO DATA - In one example, an apparatus includes a processor configured to provide information to a client device for accessing data for a first representation of a video sequence and a second representation of the video sequence, and to provide information to the client device indicating that the client device can, during transmission of the data for the first representation, switch to the second representation without experiencing a presentation gap when displaying the video sequence and without simultaneously executing more than one decoder to decode the data for the first representation and the second representation during the switch, in response to a first request from the client device to retrieve data from the first representation. In this manner, the client device may use the information to perform seamless switching between the representations. | 01-19-2012 |
20120020413 | PROVIDING FRAME PACKING TYPE INFORMATION FOR VIDEO CODING - This disclosure includes techniques for signaling characteristics of a representation of multimedia content at a representation layer, such as frame packing arrangement information for the representation. In one example, an apparatus for receiving video data includes a processing unit configured to receive information indicative of whether a bitstream includes a packed frame of video data, wherein the packed frame comprises two frames corresponding to different views of a scene for three-dimensional playback, and wherein the information is present in a representation layer external to a codec layer of the bitstream, automatically determine whether the apparatus is capable of decoding and rendering the bitstream based on an analysis of the received information and decoding and rendering capabilities of the device, and retrieve the bitstream when the processing unit determines that the device is capable of decoding and rendering the bitstream. | 01-26-2012 |
20120023249 | PROVIDING SEQUENCE DATA SETS FOR STREAMING VIDEO DATA - A device may encapsulate video data such that Supplemental Enhancement Information (SEI) messages are stored separately from a sequence of coded video pictures described by the SEI messages. An example device includes a control unit configured to generate one or more SEI messages separate from the coded video pictures, wherein the SEI messages describe respective ones of the sequence of coded video pictures and include elements common to more than one of the coded video pictures, and an output interface configured to output the SEI messages separately from the sequence of coded video pictures. An example destination device may receive the SEI messages separately from the coded video pictures and render the coded video pictures using the SEI messages. | 01-26-2012 |
20120023250 | ARRANGING SUB-TRACK FRAGMENTS FOR STREAMING VIDEO DATA - A video file may include movie fragments divided into sub-track fragments that store all pictures of common hierarchical levels for the respective hierarchical levels. In one example, an apparatus includes an interface configured to output data according to a streaming protocol, and a control unit configured to assemble encoded video data into a plurality of sub-track fragments, each of the sub-track fragments comprising a plurality of hierarchically related video pictures of the encoded video data, wherein the plurality of hierarchically related video pictures each correspond to a common hierarchical layer receiving a request in accordance with the streaming protocol, wherein the request specifies at least one of the plurality of sub-track fragments, and, in response to the request, cause the interface to output the plurality of hierarchically related video pictures of the at least one of the plurality of sub-track fragments. | 01-26-2012 |
20120027088 | CODING MOTION PREDICTION DIRECTION IN VIDEO CODING - This disclosure relates to techniques for reducing a cost of coding prediction information in video coding. Video blocks in a generalized P/B (GPB) frame are encoded using up to two motion vectors calculated from reference pictures in two separate reference picture lists that are identical. When one of the reference picture lists is preferred over the other reference picture list, the preferred reference picture list may be used for unidirectional prediction, by default. When a GPB frame is enabled such that the first and second reference picture lists are identical, either of the first and second reference picture lists may be used for unidirectional prediction. The techniques include coding one or more syntax elements indicating that a video block is coded using one of the unidirectional prediction mode with respect to a reference picture in a reference picture list and the bidirectional prediction mode using less than two bits. | 02-02-2012 |
20120027089 | CODING MOTION VECTORS IN VIDEO CODING - This disclosure relates to techniques for reducing a cost of coding prediction information in video coding. Video blocks in a generalized P/B (GPB) frame are encoded using up to two motion vectors calculated from reference pictures in two separate reference picture lists that are identical. Video blocks of a GPB frame may, therefore, be encoded using a bidirectional prediction mode with a first motion vector from a reference picture in a first reference picture list and a second motion vector from the same or substantially similar reference picture in a second reference picture list. The techniques include jointly coding the first and second motion vectors for a video block of a GPB frame. The techniques include coding the first motion vector relative to a first motion predictor generated from a motion vector of a neighboring block, and coding the second motion vector relative to the first motion vector. | 02-02-2012 |
20120033037 | Signaling Attributes for Network-Streamed Video Data - In one example, an apparatus includes a processor configured to receive video data for two or more views of a scene, determine horizontal locations of camera perspectives for each of the two or more views, assign view identifiers to the two or more views such that the view identifiers correspond to the relative horizontal locations of the camera perspectives, form a representation comprising a subset of the two or more views, and, in response to a request from a client device, send information indicative of a maximum view identifier and a minimum view identifier for the representation to the client device. | 02-09-2012 |
20120036544 | Signaling Attributes for Network-Streamed Video Data - In one example, an apparatus for signaling information for video data includes a processor configured to receive video data for two or more views of a scene, form a representation comprising a subset of the two or more views, and send, to a client device, as a part of a manifest of the representation, information indicative of a maximum number of views in the representation that can be targeted for output. An apparatus for receiving information for video data may receive the manifest including the information indicating the maximum number of views and request at least a portion of the video data of the representation based at least in part on a maximum number of views that can be output by the apparatus and the information indicative of the maximum number of views in the representation that can be targeted for output. | 02-09-2012 |
20120044986 | LOW COMPLEXITY ADAPTIVE FILTER - For a first series of video blocks, an encoder determines two filters, a first decoding filter that is to be transmitted to a decoder and a first interim filter that is not to be transmitted to the decoder. The first interim filter is used to determine which coded units of a second series of video blocks are to be filtered. After a decision is made as to which coded units of the second series of video blocks are to be filtered, the encoder determines a second decoding filter for the second series of video blocks and transmits the second decoding filter to the decoder. In addition to determining the second decoding filter, the encoder also determines a second interim filter, which the encoder uses to determine which coded units of a third series of video blocks are to be filtered. This process may repeat for many series of video blocks. | 02-23-2012 |
20120044992 | LOW COMPLEXITY ADAPTIVE FILTER - An interim filter determined for a previous series of video blocks can be applied to a current series of video blocks to determine an interim filter map for the current series of video blocks. Based on the interim filter map, a decoding filter can be determined. By applying the decoding filter to the current series of video blocks, a decoding filtering map can be determined. Based on CUs identified as having filtering off by the decoding filtering map, an interim filter for the current series of video blocks can be determined. The decoding filter and decoding filtering map can be transmitted to a decoder, while the interim filter and interim filter map may not be transmitted to a decoder. The interim filter for the current series of video blocks can be used to generate an interim filter map for a next series of video blocks. | 02-23-2012 |
20120063515 | Efficient Coding of Video Parameters for Weighted Motion Compensated Prediction in Video Coding - This disclosure relates to techniques for efficient coding of video parameters for weighted motion compensated prediction in video encoding and decoding. A video coding device may code a video block using weighted motion compensated prediction with respect to prediction data generated based on at least one motion vector and video parameter values. The video parameter values may include scale and/or offset parameter values. The techniques reduce signaling overhead by only signaling video parameter values when the motion vector points to a predefined sub-pixel position of a reference block. The techniques include storing a list of predefined sub-pixels associated with the video parameters. When the motion vector points to a sub-pixel position included in the list of predefined sub-pixels, the video coding device may code the video parameter values. The list of predefined sub-pixels may be signaled to a video decoder at a video coding unit or higher level. | 03-15-2012 |
20120075436 | CODING STEREO VIDEO DATA - In one example, a method of decoding video data comprising base layer data having a first resolution and enhancement layer data having the first resolution includes decoding the base layer data, wherein the base layer data comprises a reduced resolution version of a left view relative to the first resolution and a reduced resolution version of a right view relative to the first resolution. The method also includes decoding enhancement layer data comprising enhancement data for exactly one of the left view and the right view, wherein the enhancement data has the first resolution, and wherein decoding the enhancement layer data comprises decoding the enhancement layer data relative to at least a portion of the base layer data. | 03-29-2012 |
20120082210 | CODING PREDICTION MODES IN VIDEO CODING - A video encoder can maintain, by generating, storing, adjusting, altering, and/or updating, one or more variable length coding (VLC) tables that represent a mapping of prediction modes to codewords. One or more codewords representing a selected prediction mode can be communicated to the decoder for a CU of a frame. The decoder maintains one or more VLC tables that match the VLC tables maintained by the video encoder. Thus, based on the one or more codewords received from the video encoder, the video decoder can determine the prediction mode used to encode a CU. | 04-05-2012 |
20120082222 | VIDEO CODING USING INTRA-PREDICTION - In general, techniques of this disclosure are related to determining a prediction characteristic associated with a coding unit of video data, wherein determining the prediction characteristic includes determining a prediction type that defines a number of prediction units associated with the coding unit. Techniques of this disclosure may also be related to generating a set of available intra-prediction modes for the coding unit based on the prediction characteristic, selecting an intra-prediction mode from the available intra-prediction modes, and applying one of the available intra-prediction modes to code the coding unit. | 04-05-2012 |
20120082223 | INDICATING INTRA-PREDICTION MODE SELECTION FOR VIDEO CODING - For a block of video data, a video encoder can signal to a video decoder a selected intra-prediction mode using a codeword that is mapped to a modified intra-prediction mode index. The video decoder can receive the codeword, determine the modified intra-prediction mode index corresponding to the codeword, determine most probable modes based on a context, map the modified intra-prediction mode index to an intra-prediction mode index by comparing the modified intra-prediction mode index to the mode indexes of the most probable modes, and determine the selected intra-prediction mode used to encode the block of video data based on the intra-prediction mode index. | 04-05-2012 |
20120082224 | INTRA SMOOTHING FILTER FOR VIDEO CODING - This disclosure relates to techniques for reducing the amount of additional data encoded with a block encoded using intra-predictive coding. Particularly, the techniques provide apparatus and methods of applying a smoothing filter to prediction samples used in intra-predictive coding. For example, in fixed mode-dependent intra-predictive coding, a video encoder may determine the type of smoothing filter applied to prediction samples based on block size and intra-prediction mode combination associated with the current block, where the combination is used to look up a filter in a first filter table. In adaptive mode-dependent intra-predictive coding, the encoder uses two filters, one from the first filter table and another from a second filter table, applies both filters, and determines which yields better results. When the second filter table filter yields better results, the encoder encodes a filtering indication. When a filter from the first filter table is used, no filtering indication is encoded. | 04-05-2012 |
20120082225 | SELECTIVE INDICATION OF TRANSFORM SIZES - During a video encoding process, a video encoder may transform the residual data of a coding unit (CU) using a single transform if the CU only has a single prediction unit (PU). If the CU has multiple PUs, the video encoder may transform the residual data using multiple transforms. The video encoder outputs an indication of the size of a transform used to transform residual data of the CU only when the CU has more than one PU. If a video decoder receives such an indication, the video decoder may reconstruct residual data of the CU using a transform of the indicated size. Otherwise, the video decoder may reconstruct the residual data of the CU using a transform having same size as the CU. | 04-05-2012 |
20120082230 | VARIABLE LENGTH CODING OF VIDEO BLOCK COEFFICIENTS - This disclosure describes techniques for coding transform coefficients for a block of video data. According to one aspect of this disclosure, a coder (e.g., an encoder or decoder) may map between a code number cn and a level_ID value and an run value based on a structured mapping. According to other aspects of this disclosure, the coder may map between a code number cn and a level_ID value and an run value for the current transform coefficient using a first technique or a second technique based on a coded block type of a block of video data being coded. For example, if the coded block type is a first coded block type, the coder may use a structured mapping. However, if the coded block type is a second coded block type different than the first coded block type, the coder may access one or more mapping tables stored in memory to perform the mapping. | 04-05-2012 |
20120082231 | ZERO-OUT OF HIGH FREQUENCY COEFFICIENTS AND ENTROPY CODING RETAINED COEFFICIENTS USING A JOINT CONTEXT MODEL - This disclosure describes techniques for performing entropy encoding and decoding of video coefficients using a joint context model shared between transform units having different sizes. For example, the joint context model may be shared between transform units having a first size of 32×32 and transform units having a second size of 16×16. Performing entropy coding using a joint context model shared between transform units having different sizes may reduce an amount of memory necessary to store contexts and probabilities, and reduce computational costs of maintaining context models. In one example, the joint context model may be shared between transform units having the first size with coefficients zeroed out to generate a retained coefficient block having the second size and transform units having the second size. In another example, the joint context model may be shared between transform units having the first size and transform units having the second size. | 04-05-2012 |
20120082232 | ENTROPY CODING COEFFICIENTS USING A JOINT CONTEXT MODEL - This disclosure describes techniques for performing entropy encoding and decoding of video coefficients using a joint context model shared between transform units having different sizes. For example, the joint context model may be shared between transform units having a first size of 32×32 and transform units having a second size of 16×16. Performing entropy coding using a joint context model shared between transform units having different sizes may reduce an amount of memory necessary to store contexts and probabilities, and reduce computational costs of maintaining context models. In one example, the joint context model may be shared between transform units having the first size with coefficients zeroed out to generate a retained coefficient block having the second size and transform units having the second size. In another example, the joint context model may be shared between transform units having the first size and transform units having the second size. | 04-05-2012 |
20120093226 | ADAPTIVE MOTION VECTOR RESOLUTION SIGNALING FOR VIDEO CODING - A video encoder may be configured to adaptively select a sub-pixel precision for motion vectors used to encode video data. The video encoder may further entropy encode an indication of the sub-pixel precision using context adaptive binary arithmetic coding, where the context may correspond to the size of a block of video data for the motion vector. For example, the size may correspond to the depth of a coding unit, the size of a prediction unit of the coding unit, and/or a type for the prediction unit. The video encoder may also interpolate values for one-sixteenth pixel positions of chrominance data using bilinear interpolation. The video encoder may further encode a motion vector difference value for the motion vector using an encoding scheme corresponding to the sub-pixel precision of the motion vector. A video decoder may use similar, reciprocal techniques for decoding the video data. | 04-19-2012 |
20120099646 | ADAPTIVE SCANNING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - In an example, an apparatus for decoding video data includes a video decoder that is configured to determine a number of transform coefficients associated with a block of video data, and to determine whether the number of transform coefficients exceeds a predetermined threshold. The video decoder is also configured to identify a scan order when the number of transform coefficients exceeds the predetermined threshold, wherein the scan order indicates an order in which the transform coefficients have been serialized from a two-dimensional array to a one-dimensional array, and to apply the scan order to inverse scan the number of transform coefficients associated with a block of video data. | 04-26-2012 |
20120106649 | JOINT CODING OF SYNTAX ELEMENTS FOR VIDEO CODING - In one example, a video decoder is configured to determine whether a component of a transform unit of a coding unit of video data includes at least one non-zero coefficient based on a codeword for the transform unit, determine whether the transform unit is split into sub-transform units based on the codeword, and decode the transform unit based on the determinations. In another example, a video encoder is configured to determine whether a component of a transform unit of a coding unit of video data includes at least one non-zero coefficient, determine whether the transform unit is split into sub-transform units, select a codeword from a variable length code table, wherein the variable length code table provides an indication that the codeword corresponds to the determinations, and provide the codeword for the transform unit. | 05-03-2012 |
20120121011 | PARALLEL CONTEXT CALCULATION IN VIDEO CODING - In one example, a method for coding video data includes identifying a scan path for scanning significance information associated with a quantized transform coefficient. The method also includes determining a context support neighborhood for entropy coding the significance information associated with the quantized transform coefficient, wherein the context support neighborhood excludes one or more context support elements that are located in the scan path. The method also includes coding the significance information using the modified context support neighborhood. | 05-17-2012 |
20120121017 | REFERENCE PICTURE LIST CONSTRUCTION FOR GENERALIZED P/B FRAMES IN VIDEO CODING - This disclosure relates to techniques for indicating that a video frame is coded as a generalized P/B (GPB) frame in order to reduce a cost of coding and constructing a second reference picture list in video coding. For a GPB frame, which has identical reference picture lists, signaling and constructing both the first and second reference picture lists may be redundant. The techniques of this disclosure may include coding syntax elements indicating that the video frame is coded as the GPB frame, and coding syntax elements indicating a number of reference pictures and reference picture list construction commands for only a first reference picture list. The techniques also include constructing the first reference picture list based on the syntax elements, and then creating the second reference picture list as a duplicate of the first reference picture list. | 05-17-2012 |
20120127267 | DEPTH ESTIMATION BASED ON GLOBAL MOTION - This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, an initial indication of depth (e.g., an optical flow) may be determined for a 2D view. The initial indication of depth may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. The initial indication of depth may be modified based on the estimation of global motion to create a global motion-adjusted indication of depth. The global motion-adjusted depth indication may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation. | 05-24-2012 |
20120127270 | DEPTH ESTIMATION BASED ON GLOBAL MOTION - This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, a plurality of feature points may be determined for a 2D view. The plurality of feature points may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. For example, the plurality of feature points may be used to generate a global motion frame difference. The global motion frame difference may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation. | 05-24-2012 |
20120140813 | SEPARATELY CODING THE POSITION OF A LAST SIGNIFICANT COEFFICIENT OF A VIDEO BLOCK IN VIDEO CODING - In one example, an apparatus is disclosed for coding coefficients associated with a block of video data during a video coding process, wherein the apparatus includes a video coder configured to code information that identifies a position of a last non-zero coefficient within the block according to a scanning order associated with the block prior to coding information that identifies positions of other non-zero coefficients within the block, including at least one of the following: coding a one-dimensional position within the block that identifies the position of the last non-zero coefficient; coding a two-dimensional position within the block that identifies the position of the last non-zero coefficient, and coding a flag that indicates whether the last non-zero coefficient is located within a range of positions within the block, and coding the one-dimensional position when the last non-zero coefficient is located within the range, and otherwise coding the two-dimensional position. | 06-07-2012 |
20120140814 | CODING THE POSITION OF A LAST SIGNIFICANT COEFFICIENT WITHIN A VIDEO BLOCK BASED ON A SCANNING ORDER FOR THE BLOCK IN VIDEO CODING - In one example, an apparatus is disclosed for coding coefficients associated with a block of video data during a video coding process, wherein the apparatus includes a video coder configured to code x- and y-coordinates that indicate a position of a last non-zero coefficient within the block according to a scanning order associated with the block when the scanning order comprises a first scanning order, and code interchanged x- and y-coordinates that indicate the position of the last non-zero coefficient within the block according to the scanning order when the scanning order comprises a second scanning order, wherein the second scanning order is different than the first scanning order. | 06-07-2012 |
20120140822 | VIDEO CODING USING FUNCTION-BASED SCAN ORDER FOR TRANSFORM COEFFICIENTS - Video coding devices and methods use a function-based definition of scan order to scan transform coefficients associated with a block of residual video data. A video coder may define a scan order for coefficients based on a predefined function and one or more parameter values. A video encoder may use a function-based scan order to scan a two-dimensional array of coefficients to produce a one-dimensional array of coefficients for use in producing encoded video data. The video encoder may signal the parameters to a video decoder, or the video decoder may infer one or more of the parameters. The video decoder may use the function-based scan order to scan a one-dimensional array of coefficients to reproduce the two-dimensional array of coefficients for use in producing decoded video data. In each case, the scan order may vary according to the parameter values, which may include block size, orientation, and/or orientation strength. | 06-07-2012 |
20120147947 | CODEWORD ADAPTATION FOR VARIABLE LENGTH CODING - In one example, this disclosure describes a method of codeword adaptation for variable length coding. The method includes applying a first codeword adaptation scheme to groups of codewords in a variable length coding (VLC) table to change mappings of codewords within the groups to events in the VLC table; and applying a second codeword adaptation scheme to individual codewords within the groups of codewords in the VLC table to change mappings of the codewords to the events within the groups in the VLC table. | 06-14-2012 |
20120147961 | USE OF MOTION VECTORS IN EVALUATING GEOMETRIC PARTITIONING MODES - During a video encoding process, rectangular prediction units (PUs) for a coding unit (CU) are generated. Furthermore, a geometric partitioning mode is used to generate a first and a second geometric PU for the CU. The first and second geometric PUs are associated with different geometric partitions of a sample block of the CU. One of the rectangular PUs is identified as overlapping the first geometric PU. The motion vector of the identified rectangular PU is used to identify a given area of a reference frame. The given area of a reference frame is then used as a starting point of a search to identify a reference sample for the first geometric PU. A motion vector for the first geometric PU indicates a position of the reference sample relative to a position of the first geometric PU. A prediction block is generated using the motion vector for first geometric PU. | 06-14-2012 |
20120147967 | ADAPTIVE SUPPORT FOR INTERPOLATING VALUES OF SUB-PIXELS FOR VIDEO CODING - This disclosure describes techniques for calculating values of sub-integer pixels applied by an encoder and a decoder to encode blocks of video data. In one example, a video encoder is configured to receive values for a full integer pixel positions of a reference sample, apply an interpolation filter to a first set of the values for the full integer pixel positions to calculate a value for a first sub-integer pixel of one of the full integer pixel positions, apply the interpolation filter to a second, different set of the values for the full integer pixel positions to calculate a value for a second, different sub-integer pixel of the one of the full integer pixel positions, encode a current block of pixels using a motion vector that points to one of the first sub-integer pixel and the second sub-integer pixel. | 06-14-2012 |
20120147970 | CODEWORD ADAPTATION FOR VARIABLE LENGTH CODING - In one example, this disclosure describes a method of codeword adaptation for variable length coding. The method comprises determining if a number codewords stored in a variable length coding (VLC) table satisfies a threshold; selecting a codeword adaptation scheme from a group of two or more codeword adaptation schemes based on whether the number of codewords satisfies the threshold; and applying the selected adaptation scheme to the codewords stored in the VLC table. | 06-14-2012 |
20120147971 | CODEWORD ADAPTATION FOR VARIABLE LENGTH CODING - In one example, this disclosure describes a method of codeword adaptation for variable length coding. The method comprises applying a first codeword adaptation scheme to a first group of codewords of a variable length coding (VLC) table to change a mapping of codewords to events in the VLC table; and applying a second codeword adaptation scheme to a second group of codewords of the VLC table to change the mapping of the codewords to the events in the VLC table. | 06-14-2012 |
20120163448 | CODING THE POSITION OF A LAST SIGNIFICANT COEFFICIENT OF A VIDEO BLOCK IN VIDEO CODING - In one example, an apparatus is disclosed for coding coefficients associated with a block of video data during a video coding process, wherein the apparatus includes a video coder configured to code information that identifies a position of a last non-zero coefficient within the block according to a scanning order associated with the block, wherein to code the information, the video coder is configured to perform a context adaptive entropy coding process that includes the video coder applying a context model based on at least three contexts, wherein the at least three contexts include a size associated with the block, a position of a given one of the coefficients within the block according to the scanning order, and the scanning order. | 06-28-2012 |
20120163455 | MODE DEPENDENT SCANNING OF COEFFICIENTS OF A BLOCK OF VIDEO DATA - This disclosure describes apparatuses and methods of encoding coefficients associated with a block of video data. In one example, a method may comprise selecting a scan order for the coefficients based on an intra coding mode used to predict the block of video data and a transform block size used in transforming the block of video data, and generating a syntax element to communicate the selected scan order for the block of video data. | 06-28-2012 |
20120163456 | USING A MOST PROBABLE SCANNING ORDER TO EFFICIENTLY CODE SCANNING ORDER INFORMATION FOR A VIDEO BLOCK IN VIDEO CODING - In one example, an apparatus is disclosed for coding coefficients associated with a block of video data during a video coding process, the apparatus comprising a video coder configured to code information that identifies a scanning order associated with the block, wherein to code the information that identifies the scanning order associated with the block, the video coder is configured to determine a most probable scanning order for the block, and code an indication of whether the scanning order associated with the block is the most probable scanning order. In another example, to code the information that identifies the scanning order associated with the block, the video coder is further configured to, in the event the scanning order associated with the block is not the most probable scanning order, code an indication of the scanning order associated with the block. | 06-28-2012 |
20120163460 | SUB-PIXEL INTERPOLATION FOR VIDEO CODING - In one example, an apparatus includes a video coder configured to determine a first set of support pixels used to interpolate a value for a first sub-integer pixel position of a pixel of a reference block of video data; determine a second, different set of support pixels used to interpolate a value for a second sub-integer pixel position of the pixel; determine a third, different set of support pixels used to interpolate a value for a third sub-integer pixel position of the pixel; combine corresponding values from the first, second, and third sets of support pixels; apply an interpolation filter to the combined values to calculate a value for a fourth sub-integer-pixel comprising a one-eighth-integer position of the pixel and code a portion of a current block of the video data relative to the fourth one-eighth-integer pixel position of the reference block. | 06-28-2012 |
20120163471 | VARIABLE LENGTH CODING OF VIDEO BLOCK COEFFICIENTS - This disclosure describes techniques for coding transform coefficients for a block of video data. According to some aspects of this disclosure, an encoder or decoder may map between a code number cn and last_pos and level_ID syntax elements associated with a block of video data based on a scaling factor S. The scaling factor S may be based on a size of the block of video data being coded. | 06-28-2012 |
20120163472 | EFFICIENTLY CODING SCANNING ORDER INFORMATION FOR A VIDEO BLOCK IN VIDEO CODING - An apparatus is disclosed for coding coefficients associated with a block of video data, including a video coder configured to code information that identifies a first scanning order associated with the block if a position of any of one or more of the coefficients within the block, starting with a first coefficient within the block and ending with a last non-zero coefficient within the block according to the first scanning order, and proceeding according to the first scanning order, according to the first scanning order is different than a position of the respective coefficient within the block according to a second scanning order, and avoid coding the information if the position of each of the one or more of the coefficients within the block according to the first scanning order is the same as the position of the respective coefficient within the block according to the second scanning order. | 06-28-2012 |
20120170645 | VIDEO FILTERING USING A COMBINATION OF ONE-DIMENSIONAL SWITCHED FILTER AND ONE-DIMENSIONAL ADAPTIVE FILTER - In one example, this disclosure describes a method of loop filtering of reconstructed video data during a video coding process. The method may comprising applying a one-dimensional (1D) switched filter to the video data in a first dimension that is either horizontal or vertical, and applying a 1D adaptive filter to the video data in a second dimension that is perpendicular to the first dimension. | 07-05-2012 |
20120170648 | FRAME SPLITTING IN VIDEO CODING - In one example, this disclosure describes a method of decoding a frame of video data comprising a plurality of block-sized coding units including one or more largest coding units (LCUs) that include a hierarchically arranged plurality of relatively smaller coding units. In this example, the method includes determining a granularity at which the hierarchically arranged plurality of smaller coding units has been split when forming independently decodable portions of the frame. The method also includes identifying an LCU that has been split into a first section and a second section using the determined granularity. The method also includes decoding an independently decodable portion of the frame that includes the first section of the LCU without the second section of the LCU. | 07-05-2012 |
20120170649 | VIDEO CODING USING MAPPED TRANSFORMS AND SCANNING MODES - A video encoder may transform residual data by using a transform selected from a group of transforms. The transform is applied to the residual data to create a two-dimensional array of transform coefficients. A scanning mode is selected to scan the transform coefficients in the two-dimensional array into a one-dimensional array of transform coefficients. The combination of transform and scanning mode may be selected from a subset of combinations that is based on an intra-prediction mode. The scanning mode may also be selected based on the transform used to create the two-dimensional array. The transforms and/or scanning modes used may be signaled to a video decoder. | 07-05-2012 |
20120170650 | LOW COMPLEXITY INTERPOLATION FILTERING WITH ADAPTIVE TAP SIZE - During a prediction stage of video coding, a video coder may use relatively longer interpolation filters to generate predictive sub-pixel values using values of reference integer pixels of a reference block of video data positioned in parallel relative to a scanning order associated with the block and may use relatively shorter interpolation filters to generate predictive sub-pixel values using values of reference integer pixels of the block positioned perpendicular relative to the scanning order, wherein a longer interpolation filter generally refers to a filter with relatively more filter coefficients, or “taps,” and a shorter filter generally refers to a filter with relatively fewer taps. | 07-05-2012 |
20120170662 | VARIABLE LENGTH CODING OF VIDEO BLOCK COEFFICIENTS - This disclosure describes techniques for coding transform coefficients for a block of video data. According to some aspects of this disclosure, a coder (e.g., an encoder or decoder) may map between a code number cn and level_ID and run values associated with a first transform coefficient of the block of video data according to a first technique (e.g., a structured mapping), and map between a code number cn and level_ID and run values associated with a second coefficient of the block using a second technique. According to other aspects of this disclosure, the coder may map between a code number cn and level_ID and run syntax elements using different mathematical relationships, depending on a determined value of the code number cn or the level_ID syntax element. For example, the coder may access a mapping table of a plurality of mapping tables differently, dependent on the determined value. | 07-05-2012 |
20120177108 | 32-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing a 32-point discrete cosine transform (DCT) that is capable of applying multiple DCTs of different sizes. For example, an apparatus comprising a 32-point discrete cosine transform of type II (DCT-II) unit may implement the techniques of this disclosure. The 32-point DCT-II unit performs these DCTs-II of different sizes to transform data from a spatial to a frequency domain. The 32-point DCT-II unit includes an 16-point DCT-II unit that performs one of the DCTs-II of size 16 and at least one 8-point DCT-II unit that performs one of the DCTs-II of size 8. The 16-point DCT-II unit includes another 8-point DCT-II unit. The 16-point DCT-II unit also comprises at least one 4-point DCTs-II unit. Two or more of these DCTs-II units may concurrently perform DCTs-II of different sizes to various portions of the content data. | 07-12-2012 |
20120177114 | ADAPTIVELY PERFORMING SMOOTHING OPERATIONS - During a video coding process, a prediction block is generated for a CU. The CU has two or more prediction units. A computing device determines, based on sizes of one or more of the prediction units, whether to perform a smoothing operation on samples in a transition zone of the prediction block. The transition zone is located at a boundary between samples of the prediction block associated with different prediction units. If the computing device makes the determination to perform the smoothing operation, the smoothing operation is performed to smooth samples of the prediction block in the transition zone. | 07-12-2012 |
20120177118 | INDICATING INTRA-PREDICTION MODE SELECTION FOR VIDEO CODING USING CABAC - For a block of video data, a video encoder can signal to a video decoder, using a context-based adaptive binary arithmetic coding (CABAC) process, a selected intra-prediction mode using a codeword that is mapped to a modified intra-prediction mode index. The video decoder can perform a context-based adaptive binary arithmetic coding (CABAC) process to determine the codeword signaled by the video encoder, determine the modified intra-prediction mode index corresponding to the codeword, determine most probable modes based on a context, map the modified intra-prediction mode index to an intra-prediction mode index by comparing the modified intra-prediction mode index to the mode indexes of the most probable modes, and determine the selected intra-prediction mode used to encode the block of video data based on the intra-prediction mode index. | 07-12-2012 |
20120177120 | IDENTIFICATION OF SAMPLES IN A TRANSITION ZONE - During a video encoding or decoding process, a predicted prediction block is generated for a CU. The CU may have two or more prediction units (PUs). A computing device selects a neighbor region size. After the computing device selects the neighbor region size, samples in a transition zone of the prediction block are identified. Samples associated with a first PU are in the transition zone if neighbor regions that contain the samples also contain samples associated with a second PU. Samples associated with the second PU may be in the transition zone if neighbor regions that contain the samples also contain samples associated with the first PU. The neighbor regions have the selected neighbor region size. A smoothing operation is then performed on the samples in the transition zone. | 07-12-2012 |
20120189049 | SUB-SLICES IN VIDEO CODING - A video encoder partitions a slice of a picture into a plurality of sub-slices. When the video encoder encodes the slice, the video encoder generates a coded slice that includes coded sub-slices that correspond to the sub-slices. Each of the sub-slices can be parsed independently of each other one of the sub-slices and can be reconstructed independently of each other one of the sub-slices. Accordingly, a video decoder may parse two or more of the sub-slices in parallel and decode two or more of the sub-slices in parallel. | 07-26-2012 |
20120189052 | SIGNALING QUANTIZATION PARAMETER CHANGES FOR CODED UNITS IN HIGH EFFICIENCY VIDEO CODING (HEVC) - In one example, this disclosure describes a method of decoding video data. The method comprises receiving a coding unit (CU) of encoded video data. The CU is partitioned into a set of block-sized coded units (CUs) according to a quadtree partitioning scheme, and decoding one or more syntax elements for the CU to indicate a change in a quantization parameter for the CU relative to a predicted quantization parameter for the CU only if the CU includes any non-zero transform coefficients. The one or more syntax elements are decoded from a position within the encoded video data after an indication that the CU will include at least some non-zero transform coefficients, and before the transform coefficients for the CU. | 07-26-2012 |
20120189053 | COMBINED REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - This disclosure relates to techniques for constructing a combined reference picture list, List C, based on List | 07-26-2012 |
20120189055 | MOTION VECTOR PREDICTION - A method of coding video data includes determining a candidate motion vector for each of one or more candidate portions of a video frame and determining a current motion vector for a current portion of a current frame. The current motion vector identifies a portion of a reference frame that at least partially matches the current portion of the current frame. The method also includes calculating a motion vector difference between the current motion vector and each of the candidate motion vectors, selecting one of the candidate motion vectors based on the calculated motion vector differences, signaling an index identifying the candidate portion having the selected one of the candidate motion vectors, and signaling the corresponding motion vector difference calculated with respect to the selected one of the candidate motion vectors. | 07-26-2012 |
20120189058 | SINGLE REFERENCE PICTURE LIST CONSTRUCTION FOR VIDEO CODING - The example techniques described in this disclosure provide for an efficient manner to encode or decode a video block of a picture using a single reference picture list. The single reference picture list may include identifiers for reference picture or pictures used to encode or decode the video block. In some examples, a video encoder or decoder may encode or decode a video block that is predicted from two reference pictures using the single reference picture list, and encode or decode a video block that is predicted from one reference picture using the same, single reference picture list. | 07-26-2012 |
20120195368 | PERFORMING MOTION VECTOR PREDICTION FOR VIDEO CODING - In general, techniques are described for performing motion vector prediction for video coding. An apparatus comprising a motion compensation unit may implement the techniques. The motion compensation unit determines spatial candidate motion vectors (MVPs) associated with a current portion of a video frame and prunes the spatial candidate motion vectors to remove duplicates without removing a temporal candidate motion vector. The motion compensation unit selects one of the temporal candidate motion vector or one of the spatial candidate motion vectors remaining after pruning as a selected candidate motion vector based on a motion vector predictor (MVP) index signaled in a bitstream and performs motion compensation based on the selected candidate motion vector. | 08-02-2012 |
20120195378 | PIXEL LEVEL ADAPTIVE INTRA-SMOOTHING - This disclosure describes intra-smoothing techniques for intra-coding of video data. In one example, a video encoder is configured to determine a plurality of prediction samples for intra-predictive coding of a block to be coded in a video frame, calculate a local statistic for a first prediction sample of the prediction samples, select a filter from a plurality of filters based on the local statistic, and apply the selected filter to the first prediction sample. The video encoder may also be configured to calculate at least one local statistic for each prediction sample of the prediction samples, select a filter from a plurality of different filters based on the at least one local statistic for each prediction sample, and apply each selected filter to the corresponding prediction sample to generate the filtered version of the prediction sample. Embodiments further include video decoders and method of decoding encoded video signals. | 08-02-2012 |
20120213271 | MULTI-METRIC FILTERING - An encoder can generate a series of codewords to signal to a decoder a mapping of range combinations for two or more metrics. The encoder can generate a first codeword to indicate a current combination being decoded maps to the same filter as the most recently decoded combination that shares the same range for the first metric. The encoder can generate a second codeword to indicate that the current combination maps to the same filter as the most recently decoded combination that shares the same range for the second metric. If the current combination does not map to the same filter as either of these most recently decoded combinations, then the encoder can generate a third codeword that indicates the filter that maps to the current combination. | 08-23-2012 |
20120213291 | MULTI-METRIC FILTERING - A filter unit of a video encoder or video decoder can determine a first metric for a group of pixels within a block of pixels, determine a second metric for the group of pixels, determine a filter based on the first metric and the second metric, and generate a filtered image by applying the filter to the group of pixels. The first metric and second metric can be an activity metric and a direction metric, respectively, or can be other metrics such as an edge metric, horizontal activity metric, vertical activity metric, or diagonal activity metric. | 08-23-2012 |
20120213292 | MULTI-METRIC FILTERING - A filter unit determines a mapping of range combinations to filters. Each range combination includes a range for a first metric and a range for a second metric. The filter unit determines a unique range combination identification (ID) for each range combination, with each unique range combination ID corresponding to a sequential value. The filter unit assigns a unique group ID to each group of range combinations. A group of range combinations includes range combinations mapped to the same filter, and the unique group IDs correspond to a set of sequential values. The filter unit signals or reconstructs coefficients for the filters in an order based on the sequential values of the group IDs. | 08-23-2012 |
20120213293 | MULTI-METRIC FILTERING - A filter unit of a video encoder or video decoder can determine a first metric for a group of pixels within a block of pixels based on a comparison of a subset of the pixels in the block to other pixels in the block; determine a filter based on the first metric; and generate a filtered image by applying the filter to the group of pixels. The subset of pixels can be selected to not include pixels on the boundary of the block of pixels. | 08-23-2012 |
20120219064 | HIERARCHY OF MOTION PREDICTION VIDEO BLOCKS - A video decoder is configured to obtain an index value for a current video block. The video decoder obtains a partition type for a current video block. The video decoder selects one of a plurality of defined sets of ordered candidate predictive video blocks based on the partition type of the current video block. A video decoder selects a predictive video block from a selected one of the plurality of defined sets of ordered candidate predictive video blocks based on an index value. A video decoder generates a motion vector for a current video block based on motion information of the predictive video block. | 08-30-2012 |
20120224640 | QUANTIZED PULSE CODE MODULATION IN VIDEO CODING - A quantized PCM mode includes the step of quantizing samples so that distortion is added to coded video. In this way, video blocks coded with quantized PCM appear more uniformly with adjacent video blocks that have been coded with other lossy compression techniques, such as inter-prediction coding or intra-prediction coding. A video encoder may first quantize input video samples with a quantization step prior to PCM coding the quantized samples. This quantization step may be signaled to a decoder in the encoded video bitstream. A video decoder may receive the encoded video bitstream that includes quantized PCM coded video samples. These samples are first decoded using pulse code demodulation and then are inverse quantized with the same quantization step used to encode the video. The video decoder may extract this quantization step from the output bit-depth in the encoded video bitstream. | 09-06-2012 |
20120229602 | CODING MULTIVIEW VIDEO PLUS DEPTH CONTENT - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to receive one or more texture components from at least a portion of an image representing a view of three dimensional video data, receive a depth map component for at least the portion of the image, code a block unit indicative of pixels of the one or more texture components for a portion of the image and the depth map component. The coding comprises receiving texture data for a temporal instance of a view of video data, receiving depth data corresponding to the texture data for the temporal instance of the view of video data, and encapsulating the texture data and the depth data in a view component for the temporal instance of the view, such that the texture data and the depth data are encapsulated within a common bitstream. | 09-13-2012 |
20120230392 | MOTION VECTOR PREDICTORS (MVPs) FOR BI-PREDICTIVE INTER MODE IN VIDEO CODING - This disclosure describes video coding techniques applicable to a bi-predictive inter mode that uses adaptive motion vector prediction (AMVP). In particular, this disclosure describes techniques for encoding or decoding video data in which AMVP is used to code a first motion vector and a second motion vector associated with a current video block that is coded in the bi-predictive inter mode. More specifically, this disclosure describes techniques in which one motion vector of a bi-predictive video block can be used to define a candidate motion vector predictor (MVP) for predicting another motion vector of the bi-predictive video block. In many examples, a scaled version of the first motion vector of the bi-predictive video block is used as the MVP for the second motion vector of the bi-predictive video block. By defining an additional MVP candidate for the second motion vector of a bi-predictive video block, improved compression may be achieved. | 09-13-2012 |
20120230401 | BUFFER MANAGEMENT IN VIDEO CODECS - In one example, a video decoder is configured to determine, from data associated with an open decoding refresh (ODR) picture of video data, an identifier for a previous picture of the video data, determine whether the previous picture is currently stored in a reference picture memory, and decode only data for pictures of the video data having display order values that are greater than a display order value of the ODR picture, when the previous picture is not stored in the reference picture memory. In another example, a video encoder is configured to encode an open decoding refresh (ODR) picture, determine a previously coded picture having a display order value less than a display order value of the ODR picture and having a temporal identifier value equal to zero, and signal syntax data for the ODR picture representative of an identifier of the determined previously coded picture. | 09-13-2012 |
20120230409 | DECODED PICTURE BUFFER MANAGEMENT - The example techniques described in this disclosure are generally related to decoded picture buffer management. One or more pictures stored in the decoded picture buffer may be usable for prediction, and others may not. Pictures that are usable for prediction may be referred to as reference pictures. The example techniques described herein may determine whether a reference picture, that is currently indicated to be usable for inter-prediction, should be indicated to be unusable for inter-prediction. | 09-13-2012 |
20120230417 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes techniques for coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 09-13-2012 |
20120230418 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes techniques for coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 09-13-2012 |
20120230419 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes techniques for coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 09-13-2012 |
20120230420 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes techniques for coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 09-13-2012 |
20120230421 | TRANSFORMS IN VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes determining a first residual quadtree (RQT) depth at which to apply a first transform to luma information associated with a block of video data, wherein the RQT represents a manner in which transforms are applied to luma information and chroma information. The method also includes determining a second RQT depth at which to apply a second transform to the chroma information associated with the block of video data, wherein the second RQT depth is different than the first RQT depth. The method also includes coding the luma information at the first RQT depth and the chroma information at the second RQT depth. | 09-13-2012 |
20120230433 | VIDEO CODING TECHNIQUES FOR CODING DEPENDENT PICTURES AFTER RANDOM ACCESS - In general, this disclosure describes techniques for coding video data for random access. In particular, this disclosure proposes to code a syntax element that indicates if a dependent picture may be successfully decoded in the event of a random access request to a clean decoding refresh (CDR) picture and may be required for decoding the pictures following the clean decoding refresh (CDR) picture in display order. | 09-13-2012 |
20120236115 | POST-FILTERING IN FULL RESOLUTION FRAME-COMPATIBLE STEREOSCOPIC VIDEO CODING - Stereoscopic video data encoded according to a full resolution frame-compatible stereoscopic vide coding process. Such stereoscopic video data consists of a right view and a left that are encoded in half resolution versions in an interleaved base layer and an interleaved enhancement layer. When decoded, the right view and left view are filtered according to two sets of filter coefficients, one set for the left view and one set for the right view. The sets of filter coefficients are generated by an encoder by comparing the original left and right views to decoded versions of the left and right views. | 09-20-2012 |
20120236931 | TRANSFORM COEFFICIENT SCAN - This disclosure describes techniques for coding transform coefficients for a block of video data. According to these techniques, a video encoder may adaptively scan a first plurality of coefficients of a two-dimensional matrix of coefficients, and use a fixed scan technique for a second plurality of coefficients of the two-dimensional matrix, to generate a one-dimensional vector of transform coefficients. Also according to these techniques, a video decoder may adaptively scan a first plurality of coefficients of a one-dimensional vector of coefficients, and use a fixed scan technique for a second plurality of coefficients of the one-dimensional vector, to generate a two-dimensional matrix of transform coefficients. | 09-20-2012 |
20120236934 | SIGNALING OF MULTIVIEW VIDEO PLUS DEPTH CONTENT WITH A BLOCK-LEVEL 4-COMPONENT STRUCTURE - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to receive one or more texture components from at least a portion of an image representing a view of three dimensional video data, receive a depth map component for at least the portion of the image, code a block unit indicative of pixels of the one or more texture components for a portion of the image and the depth map component. The coding comprises coding the depth map component relative to at least one of the texture components, and signalling an attribute of the depth map component relative to the one or more texture components. | 09-20-2012 |
20120243609 | BI-PREDICTIVE MERGE MODE BASED ON UNI-PREDICTIVE NEIGHBORS IN VIDEO CODING - This disclosure describes a bi-predictive merge mode in which a bi-predictive video block inherits motion information from two different neighboring blocks, wherein the two different neighboring blocks were each encoded in a uni-predictive mode. Bi-predictive coding may improve the ability to achieve compression in video coding. The described bi-predictive merge mode may increase the number of bi-predictive candidates that can be used in the context of merge mode coding by allowing two separate uni-predicted neighbors to be used to define bi-predictive motion information for a video block. | 09-27-2012 |
20120250773 | COMBINED REFERENCE PICTURE LIST CONSTRUCTION AND MAPPING - The example techniques of this disclosure are directed to default construction techniques for the construction of a combined reference picture list, and default mapping techniques for the combined reference picture list. In some examples, a video coder may construct first and second reference picture lists from frame number values, and construct the combined reference picture list from the frame number values of the first and second reference picture lists. In some examples, a video coder may construct first and second reference picture lists from picture order count (POC) values, and construct the combined reference picture list from the POC values of the first and second reference picture lists. In some examples, a video coder may construct a combined reference picture list from received information for the construction, and map the pictures of the combined reference picture list to one of a first or second reference picture list. | 10-04-2012 |
20120262542 | DEVICES AND METHODS FOR WARPING AND HOLE FILLING DURING VIEW SYNTHESIS - Implementations include methods and systems for a converting reference images or video to 3D images or video. A two-step conversion is described which accomplishes warping and hole filling on a pixel-by-pixel basis. In one implementation, of a plurality of pixel values of a reference image at a plurality of first collinear pixels locations are successively mapped to a respective plurality of second pixel locations of a destination image. Between two of the mappings, a location of a hole between two of the second pixel locations may be identified and filled. | 10-18-2012 |
20120269270 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to, in an example, a method that includes identifying a first block of video data in a first temporal location from a first view, wherein the first block is associated with a first disparity motion vector. The method also includes determining a motion vector predictor for a second motion vector associated with a second block of video data, wherein the motion vector predictor is based on the first disparity motion vector. When the second motion vector comprises a disparity motion vector, the method includes determining the motion vector predictor comprises scaling the first disparity motion vector to generate a scaled motion vector predictor, wherein scaling the first disparity motion vector comprises applying a scaling factor comprising a view distance of the second disparity motion vector divided by a view distance of the first motion vector to the first disparity motion vector. | 10-25-2012 |
20120269271 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes identifying a first block of video data in a first temporal location from a first view, wherein the first block of video data is associated with a first temporal motion vector. The method also includes determining, when a second motion vector associated with a second block of video data comprises a temporal motion vector and the second block is from a second view, a motion vector predictor for the second motion vector based on the first temporal motion vector. The method also includes coding prediction data for the second block using the motion vector predictor. | 10-25-2012 |
20120287988 | OFFSET TYPE AND COEFFICIENTS SIGNALING METHOD FOR SAMPLE ADAPTIVE OFFSET - This disclosure describes techniques for performing sample adaptive offset signaling and coding in a video coding process. Techniques of the disclosure include both a merge-based and prediction-based signaling process for sample adaptive offset information (i.e., offset values and offset type). The techniques includes determining offset information for a current partition, comparing the offset information of the current partition with offset information of one or more neighbor partitions, coding a merge instruction in the case that the offset information of one of the one or more neighbor partitions is the same as the offset information of the current partition, and coding one of a plurality of prediction instructions in the case that the offset information of the one or more neighbor partitions is not the same as the offset information of the current partition. | 11-15-2012 |
20120287994 | FILTERING BLOCKINESS ARTIFACTS FOR VIDEO CODING - In one example, an apparatus for coding video data includes a video coder configured to determine a first prediction mode for a first block of video data and a second prediction mode for a second block of video data, wherein the first block and the second block share a common edge, decode the first block using the first prediction mode and the second block using the second prediction mode, and determine whether to deblock the common edge between the first block and the second block based at least in part on whether at least one of the first prediction mode and the second prediction mode comprises short distance intra-prediction (SDIP). | 11-15-2012 |
20120300835 | PIXEL-BY-PIXEL WEIGHTING FOR INTRA-FRAME CODING - Techniques for improving the accuracy of prediction in intra-frame coding. A prediction mode can specify a pixel along a direction independently of other pixels along the same direction. In an embodiment, an encoder selects a prediction mode to best represent the image block. In an alternative embodiment, a decoder reconstructs each pixel in the image block by weighting neighboring pixels according to a weight matrix specified by the prediction mode. | 11-29-2012 |
20120307888 | RUN-MODE BASED COEFFICIENT CODING FOR VIDEO CODING - A video coding device is configured to code coefficients of residual blocks of video data. When a coefficient of a transform unit of video data has a scan order value that is less than a threshold and when the coefficient is the last significant coefficient in a scan order in the transform unit, the video coding device may execute a function to determine a mapping between data for the coefficient and a codeword index value, and code the data for the coefficient using a codeword associated with the codeword index value. The video coding device may comprise a video encoder or a video decoder, in some examples. | 12-06-2012 |
20120307894 | INTRA PREDICTION MODE CODING WITH DIRECTIONAL PARTITIONS - A video coder can determine a most probable mode for a block of video data using non-square partitions based on the direction of the non-square partitions. When the direction of the non-square partitions is vertical, an intra prediction mode of a left-neighboring block can be selected as the most probable intra prediction mode, and when the direction of the non-square partitions is horizontal, an intra prediction mode of an above-neighboring block can be selected as the most probable intra prediction mode. | 12-06-2012 |
20120314026 | INTERNAL BIT DEPTH INCREASE IN VIDEO CODING - In an example aspects of this disclosure generally relate to a method of coding video data that includes determining a first bit depth for outputting video data and a second bit depth for coding the video data, wherein the first bit depth is less than the second bit depth. The method also includes determining whether the video data will be used as reference data when coding other video data. The method also includes storing, based on the determination, the video data at the first bit depth when the video data is not used as reference data, and the video data at the second bit depth when the video data is used as reference data. | 12-13-2012 |
20120314766 | ENHANCED INTRA-PREDICTION MODE SIGNALING FOR VIDEO CODING USING NEIGHBORING MODE - This disclosure describes techniques for intra-prediction mode signaling for video coding. In one example, a video coder is configured to determine, for a block of video data, a set of most probable intra-prediction modes such that the set of most probable intra-prediction modes has a size that is equal to a predetermined number that is greater than or equal to two. The video coder is also configured to code a value representative of an actual intra-prediction mode for the block based at least in part on the set of most probable intra-prediction modes and code the block using the actual intra-prediction mode. The video coder may further be configured to code the block using the actual intra-prediction mode, e.g., to encode or decode the block. Video encoders and video decoders may implement these techniques. | 12-13-2012 |
20120314767 | BORDER PIXEL PADDING FOR INTRA PREDICTION IN VIDEO CODING - A video coder performs a padding operation that processes a set of border pixels according to an order. The order starts at a bottom-left border pixel and proceeds through the border pixels sequentially to a top-right border pixel. When the padding operation processes an unavailable border pixel, the padding operation predicts a value of the unavailable border pixel based on a value of a border pixel previously processed by the padding operation. The video coder may generate an intra-predicted video block based on the border pixels. | 12-13-2012 |
20120320968 | UNIFIED MERGE MODE AND ADAPTIVE MOTION VECTOR PREDICTION MODE CANDIDATES SELECTION - A unified candidate block set for both adaptive motion vector prediction (AMVP) mode and merge mode for use in inter-prediction is proposed. In general, the same candidate block set is used regardless of which motion vector prediction mode (e.g., merge mode or AMVP mode) is used. In other examples of this disclosure, one candidate block in a set of candidate blocks is designated as an additional candidate block. The additional candidate block is used if one of the other candidate blocks is unavailable. Also, the disclosure proposes a checking pattern where the left candidate block is checked before the below left candidate block. Also, the above candidate block is checked before the right above candidate block. | 12-20-2012 |
20120320969 | UNIFIED MERGE MODE AND ADAPTIVE MOTION VECTOR PREDICTION MODE CANDIDATES SELECTION - A unified candidate block set for both adaptive motion vector prediction (AMVP) mode and merge mode for use in inter-prediction is proposed. In general, the same candidate block set is used regardless of which motion vector prediction mode (e.g., merge mode or AMVP mode) is used. In other examples of this disclosure, one candidate block in a set of candidate blocks is designated as an additional candidate block. The additional candidate block is used if one of the other candidate blocks is unavailable. Also, the disclosure proposes a checking pattern where the left candidate block is checked before the below left candidate block. Also, the above candidate block is checked before the right above candidate block. | 12-20-2012 |
20120328003 | MEMORY EFFICIENT CONTEXT MODELING - In an example, aspects of this disclosure relate to a method of coding video data that includes determining context information for a block of video data, where the block is included within a coded unit of video data, where the block is below a top row of blocks in the coded unit, and where the context information does not include information from an above-neighboring block in the coded unit. That method also includes entropy coding data of the block using the determined context information. | 12-27-2012 |
20120328004 | QUANTIZATION IN VIDEO CODING - In an example, aspects of this disclosure relate to a method of coding video data that includes identifying a plurality of quantization parameter (QP) values associated with a plurality of reference blocks of video data. The method also includes generating a reference QP for the plurality of reference blocks based on the plurality of QPs. The method also includes storing the reference QP, and coding a block of video data based on the stored reference QP. | 12-27-2012 |
20120328026 | CONTEXT-ADAPTIVE CODING VIDEO DATA - In an example, aspects of this disclosure relate to a method of coding data that includes coding a sequence of bins according to a context adaptive entropy coding process. A current coding cycle used to code at least one bin of the sequence of bins includes determining a context for the bin; selecting a probability model based on the context, wherein the probability model is updated based on a value of a previous bin coded with the context and coded at least two coding cycles prior to the current coding cycle; applying the probability model to code the bin; and updating the probability model based on a value of the bin | 12-27-2012 |
20130003821 | SIGNALING SYNTAX ELEMENTS FOR TRANSFORM COEFFICIENTS FOR SUB-SETS OF A LEAF-LEVEL CODING UNIT - This disclosure describes techniques for coding transform coefficients for a block of video data. According to these techniques, a video encoder divides a leaf-level unit of video data into a plurality of transform coefficient sub-sets. The video encoder generates, for a sub-set of the plurality of transform coefficient sub-sets, a syntax element that indicates whether or not the sub-set includes any non-zero coefficients. In some examples, the video encoder may selectively determine whether to generate the syntax element for each sub-set. A decoder may read an entropy encoded bit stream that includes the syntax element, and determine whether to decode the sub-set based on the syntax element. | 01-03-2013 |
20130003824 | APPLYING NON-SQUARE TRANSFORMS TO VIDEO DATA - In one example, a device for coding video data includes a video coder, such as a video encoder or a video decoder, that is configured to code information indicative of whether a transform unit of the video data is square or non-square, and code data of the transform unit based at least in part on whether the transform unit is square or non-square. In this manner, the video coder may utilize non-square transform units. The video coder may be configured to use non-square transform units for certain situations, such as only for chrominance or luminance components or only when a corresponding prediction unit is non-square. The video coder may further be configured to perform an entropy coding process that selects context for coding data of the transform unit based on whether the transform unit is square or non-square. | 01-03-2013 |
20130003834 | DERIVATION OF THE POSITION IN SCAN ORDER OF THE LAST SIGNIFICANT TRANSFORM COEFFICIENT IN VIDEO CODING - A video decoder converts a block-based LSC indicator into a scan-based LSC indicator. The block-based LSC indicator indicates coordinates of a last significant coefficient of a block of transform coefficients according to a scanning order. The scan-based LSC indicator indicates an ordinal position of the LSC according to the scanning order. The video decoder may decode, based on the scan-based LSC indicator, a significance map for the block. | 01-03-2013 |
20130003835 | CODING OF LAST SIGNIFICANT TRANSFORM COEFFICIENT - A video encoder determines that the last significant coefficient (LSC) of a transform coefficient block occurs at a given ordinal position according to a coding scanning order. The video encoder generates a coordinate indicator that specifies the coordinates of a given transform coefficient in the transform coefficient block. The given transform coefficient occurs at the same ordinal position according to an assumed scanning order. A video decoder receives the coordinate indicator and converts the coordinate indicator into a scan-based LSC indicator. The scan-based LSC indicator indicates the ordinal position of the LSC. | 01-03-2013 |
20130003849 | VIDEO CODING USING ADAPTIVE MOTION VECTOR RESOLUTION - In one example, a device for coding video data includes a video coder configured to configured to code information representative of whether an absolute value of an x-component of a motion vector difference value for a current block of video data is greater than zero, code information representative of whether an absolute value of a y-component of the motion vector difference value is greater than zero, when the absolute value of the x-component is greater than zero, code information representative of the absolute value of the x-component, when the absolute value of the y-component is greater than zero, code information representative of the absolute value of the y-component, when the absolute value of the x-component is greater than zero, code a sign of the x-component, and when the absolute value of the y-component is greater than zero, code a sign of the y-component. | 01-03-2013 |
20130003859 | TRANSITION BETWEEN RUN AND LEVEL CODING MODES - This disclosure describes techniques for coding transform coefficients for a block of video data. According to some aspects of this disclosure, a video coder (e.g., encoder, decoder) may code a first coefficient of a leaf-level unit of video data using a run encoding mode. The coder may code a second coefficient of the leaf-level unit of video data using a level encoding mode. After coding at least one coefficient using the level coding mode, the coder may use the run coding mode to code a third other coefficient of the leaf-level unit of video data. According to other aspects, an encoder may signal, to a decoder, at least one indication of a transition between level and run coding modes. According to still other aspects, a coder may automatically determine when to transition between the level and run coding modes. | 01-03-2013 |
20130010865 | REDUCED RESOLUTION PIXEL INTERPOLATION - An offset can be applied to intermediate values obtained while performing an interpolation filtering operation such that applying the offset reduces the bitdepth of the intermediate value. The intermediate value can be stored with the reduced bitdepth, and when retrieved, the offset can be added back such that future calculation can be performed using the intermediate value with the original bitdepth. | 01-10-2013 |
20130016769 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 01-17-2013 |
20130022104 | ADAPTATION PARAMETER SETS FOR VIDEO CODING - In one example, a video decoder is configured to store an adaptation parameter set (APS) data structure in a buffer of a coding device, wherein the APS data structure includes signaling data applicable to one or more slices of video data, decode at least one of the slices based on the signaling data of the APS data structure, after decoding the slice, determine that the APS data structure can be removed from the buffer, remove the APS data structure from the buffer based on the determination, and decode one or more additional slices of the video data after removing the APS data structure from the buffer. | 01-24-2013 |
20130022107 | DEBLOCKING OF NON-SQUARE BLOCKS FOR VIDEO CODING - In one example, a video coding device is configured to decode a first block of video data and a second block of video data, wherein the first block and the second block share a common edge, and determine whether to deblock the common edge between the first block and the second block based at least in part on whether at least one of the first block and the second block is non-square. Non-square blocks of video data may comprise, for example, partitions (e.g., prediction units or transform units) resulting from asymmetric motion partitioning, non-square transforms for symmetric motion partitioning, or other quadtree transform splitting procedures. The video coding device may apply adapted deblocking decision functions and/or adapted deblocking filters to the common edge when either or both of the blocks is non-square. | 01-24-2013 |
20130022111 | CODING MOTION DEPTH MAPS WITH DEPTH RANGE VARIATION - This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to determine a first real-world depth range for a first depth view component comprising a reference view component, determine a second real-world depth range for a second depth view component comprising a current view component, wherein the current view component is predicted relative to the reference view component, determine a predictive block for a portion of the current view component from the reference view component, adjust values of the predictive block based on a difference between the first real-world depth range and the second real-world depth range, and predict the portion of the current view based on the adjusted values of the predictive block. | 01-24-2013 |
20130022113 | SLICE HEADER PREDICTION FOR DEPTH MAPS IN THREE-DIMENSIONAL VIDEO CODECS - In one example, a video coder is configured to code a first slice, wherein the first slice comprises one of a texture slice and a corresponding depth slice, wherein the first slice has a slice header comprising complete syntax elements representative of characteristics of the first slice. The video coder is further configured to determine common syntax elements for a second slice from the slice header of the first slice. The video coder is also configured to code the second slice after coding the first slice at least partially based on the determined common syntax elements, wherein the second slice comprises one of the texture slice and the depth slice that is not the first slice, wherein the second slice has a slice header comprising syntax elements representative of characteristics of the second slice, excluding values for syntax elements that are common to the first slice. | 01-24-2013 |
20130022119 | BUFFERING PREDICTION DATA IN VIDEO CODING - In an example, aspects of this disclosure relate to a method of coding video data that generally includes determining prediction information for a block of video data, where the block is included in a coded unit of video data and positioned below a top row of above-neighboring blocks in the coded unit, and where the prediction information for the block is based on prediction information from one or more other blocks in the coded unit but not based on prediction information from any of the top row of blocks in the coded unit. The method also generally includes coding the block based on the determined prediction information. | 01-24-2013 |
20130034170 | CODING PARAMETER SETS FOR VARIOUS DIMENSIONS IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to code, for a bitstream, information representative of which of a plurality of video coding dimensions are enabled for the bitstream, and code values for each of the enabled video coding dimensions, without coding values for the video coding dimensions that are not enabled, in a network abstraction layer (NAL) unit header of a NAL unit comprising video data coded according to the values for each of the enabled video coding dimensions. In this manner, NAL unit headers may have variable lengths, while still providing information for scalable dimensions to which the NAL units correspond. | 02-07-2013 |
20130038686 | THREE-DIMENSIONAL VIDEO WITH ASYMMETRIC SPATIAL RESOLUTION - A video coding device may be configured to code a bitstream including multiple views plus depth information. Two of the views may have reduced resolutions, while a third view may have a full resolution. The third view may be predicted relative to upsampled versions of the two reduced-resolution views. Each view may include texture data and depth data, such that a view component may include a texture component and a depth component. Moreover, the texture and depth components may be arranged within an access unit according to a particular order, which may simplify component extraction from the access unit. | 02-14-2013 |
20130044809 | APPLYING PARTITION-BASED FILTERS - In general, techniques are described for applying partition-based filters when coding video data. A device comprising at least one processor may be configured to implement the techniques. The processor selects a filter to apply near a boundary of a first portion of the video data and determines at least one of the plurality of filter coefficients of the selected filter for which the video data will not be available to be filtered. Based on the determination, the processor determines a partial filter that does not include the at least one of the plurality of filter coefficients for which the video data will not be available to be filtered. The processor renormalizes the plurality of filter coefficients included within the partial filter and applies the renormalized partial filter near the boundary of the first portion of the video data to generate a filtered first portion of the video data. | 02-21-2013 |
20130044812 | ADAPTIVE CODING OF VIDEO BLOCK PREDICTION MODE - This disclosure describes techniques for coding of header information of video blocks. In particular, the techniques of this disclosure select one of a plurality of prediction modes for use in generating a prediction block of a video block of a coding unit, the plurality of prediction modes including unidirectional prediction modes and multi-directional prediction modes that combine at least two unidirectional prediction modes. An encoding device encodes the prediction mode of the current video block based on prediction modes of one or more previously encoded video blocks of the coding unit. Likewise, a decoding unit receives encoded video data of a video block of a coding unit and decodes the encoded video data to identify one of a plurality of prediction modes for use in generating a prediction block of the video block based on prediction modes of one or more previously decoded video blocks of the coding unit. | 02-21-2013 |
20130057646 | SLICE HEADER THREE-DIMENSIONAL VIDEO EXTENSION FOR SLICE HEADER PREDICTION - In one example, a video coder is configured to code one or more blocks of video data representative of texture information of at least a portion of a frame of video data, process a texture slice for a texture view component of a current view associated, the texture slice comprising the coded one or more blocks and a texture slice header comprising a set of syntax elements representative of characteristics of the texture slice, code depth information representative of depth values for at least the portion of the frame, and process a depth slice for a depth view component corresponding to the texture view component of the view, the depth slice comprising the coded depth information and a depth slice header comprising a set of syntax elements representative of characteristics of the depth slice, wherein process the texture slice or the depth slice comprises predict at least one syntax element. | 03-07-2013 |
20130058407 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 03-07-2013 |
20130064294 | CONTEXT ADAPTIVE ENTROPY CODING FOR NON-SQUARE BLOCKS IN VIDEO CODING - Disclosed are techniques for coding coefficients of a video block having a non-square shape defined by a width and a height, comprising coding one or more of x- and y-coordinates that indicate a position of a last non-zero coefficient within the block according to an associated scanning order, including coding each coordinate by determining one or more contexts used to code the coordinate based on one of the width and the height that corresponds to the coordinate, and coding the coordinate by performing a context adaptive entropy coding process based on the contexts. Also disclosed are techniques for coding information that identifies positions of non-zero coefficients within the block, including determining one or more contexts used to code the information based on one or more of the width and the height, and coding the information by performing a context adaptive entropy coding process based on the contexts. | 03-14-2013 |
20130070848 | LINE BUFFER REDUCTION FOR SHORT DISTANCE INTRA-PREDICTION - A video coder, such as a video encoder or a video decoder, identifies an entropy coding context in a set of one or more entropy coding contexts. The video coder identifies the entropy coding context without reference to a neighboring coding unit that is above a current coding unit in a current picture. The video coder then entropy codes a short distance intra-prediction (SDIP) syntax element of a coding unit (CU) using the identified entropy coding context. The SDIP syntax element at least partially defines a mode by which the CU is partitioned into a set of one or more transform units. | 03-21-2013 |
20130070854 | MOTION VECTOR DETERMINATION FOR VIDEO CODING - For each prediction unit (PU) belonging to a coding unit (CU), a video coder generates a candidate list. The video coder generates the candidate list such that each candidate in the candidate list that is generated based on motion information of at least one other PU is generated without using motion information of any of the PUs belonging to the CU. After generating the candidate list for a PU, the video coder generates a predictive video block for the PU based on one or more reference blocks indicated by motion information of the PU. The motion information of the PU is determinable based on motion information indicated by a selected candidate in the candidate list for the PU. | 03-21-2013 |
20130070855 | HYBRID MOTION VECTOR CODING MODES FOR VIDEO CODING - In one example, a device for coding video data includes a video coder (such as a video decoder or a video encoder) configured to code motion information for a current block of video data using a hybrid motion information coding mode, wherein to code the motion information, the video coder is configured to code a merge index syntax element of the motion information in a manner substantially conforming to a merge mode, and code at least one additional syntax element of the motion information in a manner substantially conforming to an advanced motion vector prediction (AMVP) mode, and wherein the video coder is configured to code the current block using the motion information. The hybrid mode may comprise a partial merge mode or a partial AMVP mode. | 03-21-2013 |
20130077691 | PARALLELIZATION FRIENDLY MERGE CANDIDATES FOR VIDEO CODING - This disclosure presents methods and systems for coding video in merge mode of a motion vector prediction process. A method of coding video data may determining a merge candidate set for a current prediction unit of a current coding unit, wherein the merge candidate set is determined without comparing motion information of a merge candidate in the merge candidate set to motion information of any other prediction units, and performing a merge motion vector prediction process for the current prediction unit using the merge candidate set. The method may further comprise excluding merge candidates from the merge candidate set that are within another prediction unit of the current coding unit. | 03-28-2013 |
20130083844 | COEFFICIENT CODING FOR SAMPLE ADAPTIVE OFFSET AND ADAPTIVE LOOP FILTER - Techniques for coding both edge and band offset values are described. Offset values may be predicted such that one offset value in a group of offset values is predicted from another offset value in the group. In addition, offset values of a partition may be predicted from offset values of a neighboring partition. Offset values may also be right shifted to be at a lower precision before signaling in the encoded video bitstream. A video decoding device may apply the techniques to filter a current partition based on offset values associated with a neighboring partition. | 04-04-2013 |
20130083853 | MOTION VECTOR PREDICTOR CANDIDATE CLIPPING REMOVAL FOR VIDEO CODING - This disclosure describes techniques for coding a video block based on an unclipped version of a motion vector predictor candidate. The techniques include determining a motion vector predictor candidate list including motion vector predictor candidates from neighboring video blocks without clipping the motion vector predictor candidates. More specifically, if one of the motion vector predictor candidates points to a prediction block located outside of a reference picture boundary relative to the current video block, the techniques allow an unclipped version of the motion vector predictor candidate to be included in the candidate list. The current video block is then coded based on a determined unclipped motion vector predictor candidate of the candidate list. Elimination of the motion vector predictor candidate clipping process reduces complexity at both the video encoder and the video decoder. | 04-04-2013 |
20130083856 | CONTEXTS FOR COEFFICIENT LEVEL CODING IN VIDEO COMPRESSION - This disclosure describes techniques for coding video data. In particular, this disclosure describes techniques for entropy coding of residual transform coefficients generated by a video coding process. In one example, a method selects a bin 2 context for coding a bin 2 level of one or more transform coefficients in the vector according to the entropy coding process. The method further codes the bin 2 level of one or more transform coefficients in the vector according to the selected bin 2 context. Selecting the bin 2 context comprises selecting the bin 2 context for a current transform coefficient in the vector based on the bin 2 level of one or more previously coded transform coefficients in the vector. | 04-04-2013 |
20130083857 | MULTIPLE ZONE SCANNING ORDER FOR VIDEO CODING - A method for encoding transform coefficients in a video encoding process includes dividing a block of transform coefficients into a plurality of zones, determining a scan order for each of the plurality of zones, and performing a scan on each of the transform coefficients in each of the plurality of zones according to their respective determined scan order. In another example, a method for decoding transform coefficients in a video encoding process includes receiving a one-dimensional array of transform coefficients, determining a scan order for each of a plurality of sections of the one-dimensional array, wherein each section of the one-dimensional array corresponds to one of a plurality of zones defining a block of transform coefficients, and performing a scan on each of the transform coefficients in each of the section of the one dimensional array of zones according to their respective determined scan order. | 04-04-2013 |
20130089135 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving a first decoded frame of video data, wherein the first decoded frame is associated with a first resolution, determining whether a decoded picture buffer is available to store the first decoded frame based on the first resolution, and in the event the decoded picture buffer is available to store the first decoded frame, storing the first decoded frame in the decoded picture buffer, and determining whether the decoded picture buffer is available to store a second decoded frame of video data, wherein the second decoded frame is associated with a second resolution, based on the first resolution and the second resolution, wherein the first decoded frame is different than the second decoded frame. | 04-11-2013 |
20130089138 | CODING SYNTAX ELEMENTS USING VLC CODEWORDS - This disclosure describes techniques for coding transform coefficients for a block of video data. For example, according to one embodiment, a video encoder determines an lrg1Pos value associated with the transform coefficient based on the noTr1 value and a position k of the transform in the scan order of the block of video data based on using at least one table that defines an lrg1Pos value for more than one potential noTr1 value for the scan order of the block of video data. In one embodiment, the video decoder uses the determined lrg1Pos value associated with the transform coefficient to perform a structured mapping to determine a code number cn based on a determined value for the level_ID syntax element and a determined value for the run syntax element. | 04-11-2013 |
20130089145 | MOST PROBABLE TRANSFORM FOR INTRA PREDICTION CODING - A video coder can be configured to determine an intra-prediction mode for a block of video data, identify a most probable transform based on the intra-prediction mode determined for the block of video data, and code an indication of whether the most probable transform is a transform used to encode the block of video data. The most probable transform can be a non-square transform. | 04-11-2013 |
20130089154 | ADAPTIVE FRAME SIZE SUPPORT IN ADVANCED VIDEO CODECS - Techniques are described related to receiving first and second sub-sequences of video, wherein the first sub-sequence includes one or more frames each having a first resolution, and the second sub-sequence includes one or more frames each having a second resolution, receiving a first sequence parameter set and a second sequence parameter set for the coded video sequence, wherein the first sequence parameter set indicates the first resolution of the one or more frames of the first sub-sequence, and the second sequence parameter set indicates the second resolution of the one or more frames of the second sub-sequence, and wherein the first sequence parameter set is different than the second sequence parameter set, and using the first sequence parameter set and the second sequence parameter set to decode the coded video sequence. | 04-11-2013 |
20130094569 | SAMPLE ADAPTIVE OFFSET MERGED WITH ADAPTIVE LOOP FILTER IN VIDEO CODING - Techniques for performing sample adaptive offset (SAO) and adaptive loop filter (ALF) processes in a video coding process are described. The SAO and ALF processes may be combined. In one example, the determination of offset values for a SAO filter process may be based on classifications used in an ALF process. In one example, an ALF classification that indicates a particular directional characteristic of a video block may be used to determine how and whether an SAO filter process is applied to each sample within the video block. | 04-18-2013 |
20130094572 | PERFORMING TRANSFORM DEPENDENT DE-BLOCKING FILTERING - In general, techniques are described for performing transform dependent de-blocking filtering, which may be implemented by a video encoding device. The video encoding device may apply a transform to a video data block to generate a block of transform coefficients, apply a quantization parameter to quantize the transform coefficients and reconstruct the block of video data from the quantized transform coefficients. The video encoding device may further determine at least one offset used in controlling de-blocking filtering based on the size of the applied transform, and perform de-blocking filtering on the reconstructed block of video data based on the determined offset. Additionally, the video encoder may specify a flag in a picture parameter set (PPS) that indicates whether the offset is specified in one or both of the PPS and a header of an independently decodable unit. | 04-18-2013 |
20130101016 | LOOP FILTERING AROUND SLICE BOUNDARIES OR TILE BOUNDARIES IN VIDEO CODING - The techniques of this disclosure apply to loop filtering across slice or tile boundaries in a video coding process. In one example, a method for performing loop filtering in a video coding process includes determining that pixels corresponding to filter coefficients of a filter mask for a loop filter are across a slice or tile boundary, removing filter coefficients corresponding to the pixels across the slice or tile boundary from the filter mask, renormalizing the filter mask without the removed filter coefficients, performing loop filtering using the renormalized filter mask. | 04-25-2013 |
20130101018 | ADAPTIVE LOOP FILTERING FOR CHROMA COMPONENTS - This disclosure proposes techniques to allow more flexibility in filtering chroma components in the adaptive loop filter. In one example, a method for adaptive loop filtering includes performing luma adaptive loop filtering based for luma components of a block of pixels, and performing chroma adaptive loop filtering for chroma components of the block of pixels, wherein filter coefficients for both the luma adaptive loop filtering and chroma adaptive loop filtering are derived from a block-based mode or a region-based mode. The method may further include determining to perform luma adaptive loop filtering on the block of pixels, and determining to perform chroma adaptive loop filtering on the block of pixels, wherein the determining to perform chroma adaptive loop filtering is performed independently of determining to perform luma adaptive loop filtering. | 04-25-2013 |
20130101024 | DETERMINING BOUNDARY STRENGTH VALUES FOR DEBLOCKING FILTERING FOR VIDEO CODING - A video coder associates a first boundary strength value with an edge in response to determining that a first video block or a second video block is associated with an intra-predicted coding unit (CU), where the edge occurs at a boundary between the first video block and the second video block. The video coder may associate a second or a third boundary strength value with the edge when neither the first video block nor the second video block is associated with an intra-predicted CU. The video coder may apply one or more deblocking filters to samples associated with the edge when the edge is associated with the first boundary strength value or the second boundary strength value. The third boundary strength value indicates that the deblocking filters are turned off for the samples associated with the edge. | 04-25-2013 |
20130101025 | INTRA PULSE CODE MODULATION (IPCM) AND LOSSLESS CODING MODE DEBLOCKING FOR VIDEO CODING - Techniques for coding video data include coding a plurality of blocks of video data, wherein at least one block of the plurality of blocks of video data is coded using a coding mode that is one of an intra pulse code modulation (IPCM) coding mode and a lossless coding mode. In some examples, the lossless coding mode may use prediction. The techniques further include assigning a non-zero quantization parameter (QP) value for the at least one block coded using the coding mode. The techniques also include performing deblocking filtering on one or more of the plurality of blocks of video data based on the coding mode used to code the at least one block and the assigned non-zero QP value for the at least one block. | 04-25-2013 |
20130101031 | DETERMINING QUANTIZATION PARAMETERS FOR DEBLOCKING FILTERING FOR VIDEO CODING - A video coder determines a deblocking quantization parameter (QP) value based on at least one of a first QP value and a second QP value. Subsequently, the video coder applies a deblocking filter that is based on the deblocking filter to an edge associated with a first video block. The edge occurs at a boundary between the first video block and a second video block. The first video block is associated with a current coding unit (CU) and the second video block is associated with a neighboring CU. The current CU is included in a first quantization group and the neighboring CU is included in a second quantization group. The first QP value is defined for the first quantization group. The second QP value is defined for the second quantization group. | 04-25-2013 |
20130101033 | CODING NON-SYMMETRIC DISTRIBUTIONS OF DATA - This disclosure describes techniques for coding non-symmetric distributions of data and techniques for quantization matrix compression. The techniques for coding non-symmetric distributions of data may use a mapping that is configured to bias either positive data values or negative data values of a signed integer source towards shorter codewords of a variable length code that codes non-negative integers. This may allow signed integer data sources that have non-symmetric distributions of data to be coded in a more efficient manner. The quantization matrix compression techniques of this disclosure may use a predictor that is configured to generate prediction residuals for a quantization matrix that are skewed in favor of positive values. This may allow entropy coding techniques that favor data distributions which are skewed toward positive data values (e.g., the techniques for coding non-symmetric distributions described above) to be used to increase the coding efficiency of the quantization matrix. | 04-25-2013 |
20130101035 | GROUPING OF TILES FOR VIDEO CODING - Techniques described herein for coding video data include techniques for coding pictures partitioned into tiles, in which each of the plurality of tiles in a picture is assigned to one of a plurality of tile groups. One example method for coding video data comprising a picture that is partitioned into a plurality tiles comprises coding video data in a bitstream, and coding, in the bitstream, information that indicates one of a plurality of tile groups to which each of the plurality of tiles is assigned. The techniques for grouping tiles described herein may facilitate improved parallel processing for both encoding and decoding of video bitstreams, improved error resilience, and more flexible region of interest (ROI) coding. | 04-25-2013 |
20130107942 | FRAGMENTED PARAMETER SET FOR VIDEO CODING | 05-02-2013 |
20130107950 | NON-SQUARE TRANSFORMS IN INTRA-PREDICTION VIDEO CODING | 05-02-2013 |
20130107951 | MAPPING STATES IN BINARY ARITHMETIC CODER FOR VIDEO CODING | 05-02-2013 |
20130107952 | UNIFIED DESIGN FOR PICTURE PARTITIONING SCHEMES | 05-02-2013 |
20130107970 | TRANSFORM UNIT PARTITIONING FOR CHROMA COMPONENTS IN VIDEO CODING | 05-02-2013 |
20130107973 | LOOP FILTERING CONTROL OVER TILE BOUNDARIES | 05-02-2013 |
20130114669 | VLC COEFFICIENT CODING FOR LARGE CHROMA BLOCK - This disclosure describes techniques for coding transform coefficients for a block of video data. According to these techniques, a video coder (a video encoder or video decoder) determines whether a block of video data is a luma block or a chroma block. If the block of video data is a luma block, the video coder adaptively updates a VLC table index value based on a code number cn and value of a scaling factor. However, if the block of video data is a chroma block, the video coder adaptively updates the VLC table index value based on the code number cn and without using the scaling factor. The video coder uses the updated VLC table index value to select a VLC table of a plurality of VLC tables that are used to encode or decode the block of video data. | 05-09-2013 |
20130114670 | MULTIVIEW VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes obtaining, from an encoded bitstream, one or more network abstraction layer (NAL) units for each view component of a plurality of view components of encoded video data, where each view component of the plurality of view components corresponds to a common temporal location, and where the one or more NAL units encapsulate at least a portion of the encoded video data for the respective view components and include information indicative of a decoding order of the respective view components. The method also includes obtaining information, separate from the NAL units, indicating relationships between view identifiers for the views and the decoding order of the view components. The method also includes decoding the encoded video data of the plurality of view components in the decoding order based on the received information. | 05-09-2013 |
20130114671 | CONTEXT REDUCTION FOR CONTEXT ADAPTIVE BINARY ARITHMETIC CODING - A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56. | 05-09-2013 |
20130114672 | CONTEXT REDUCTION FOR CONTEXT ADAPTIVE BINARY ARITHMETIC CODING - A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56. | 05-09-2013 |
20130114673 | CONTEXT REDUCTION FOR CONTEXT ADAPTIVE BINARY ARITHMETIC CODING - A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56. | 05-09-2013 |
20130114674 | ADAPTIVE CENTER BAND OFFSET FILTER FOR VIDEO CODING - A video coder configured to perform sample adaptive offset filtering can determine a center value for a set of pixels based on values of pixels in the set, divide bands of pixels values into groups based on the center value, and determine offset values for the bands based on the groups. | 05-09-2013 |
20130114675 | CONTEXT STATE AND PROBABILITY INITIALIZATION FOR CONTEXT ADAPTIVE ENTROPY CODING - In one example, an apparatus for context adaptive entropy coding may include a coder configured to determine one or more initialization parameters for a context adaptive entropy coding process based on one or more initialization parameter index values. The coder may be further configured to determine one or more initial context states for initializing one or more contexts of the context adaptive entropy coding process based on the initialization parameters. The coder may be still further configured to initialize the contexts based on the initial context states. In some examples, the initialization parameters may be included in one or more tables, wherein, to determine the initialization parameters, the coder may be configured to map the initialization parameter index values to the initialization parameters in the tables. Alternatively, the coder may be configured to calculate the initialization parameters using the initialization parameter index values and one or more formulas. | 05-09-2013 |
20130114676 | CONTEXT OPTIMIZATION FOR LAST SIGNIFICANT COEFFICIENT POSITION CODING - A video encoder is configured to encode a binary sting indicating a position of a last significant coefficient within a video block. A video decoder is configured to decode the encoded binary string. The string may be coded using context adaptive binary arithmetic coding (CABAC). Binary indices of the binary string may be assigned a context. The context may be determined according to a mapping function. A context may be a assigned to one or more binary indices where each index is associated with a different block size. The last binary index of a 16×16 video block may share a context with the last binary index of a 32×32 video block. | 05-09-2013 |
20130114691 | ADAPTIVE INITIALIZATION FOR CONTEXT ADAPTIVE ENTROPY CODING - In one example, an apparatus for context adaptive entropy coding a video unit comprises a coder configured to code a syntax element, wherein a first value of the syntax element indicates that one or more of a plurality of context states are initialized using an adaptive initialization mode for the video unit, and a second value of the syntax element indicates that each of the plurality of context states is initialized using a default initialization mode for the video unit. In some examples, when the syntax element has the first value, the coder is further configured to code a map that indicates which of the context states are initialized using the adaptive initialization mode, and to further code either an initial state value for those contexts, or information from which the initial state values of those adaptively initialized context may be derived. | 05-09-2013 |
20130114694 | PARAMETER SET GROUPS FOR CODED VIDEO DATA - A video coding device, such as a video encoder or a video decoder, may be configured to code a parameter set group representing a first parameter set of a first type and a second parameter set of a second, different type, and code a slice of video data using information of the parameter set group, information of the first parameter set, and information of the second parameter set, wherein the slice includes information referring to the parameter set group. The video coding device may further code the first and second parameter sets. | 05-09-2013 |
20130114695 | SIGNALING QUANTIZATION MATRICES FOR VIDEO CODING - The techniques of this disclosure may be generally related to signaling values of a quantization matrix. In some examples, coefficient values in the quantization matrix may be downsampled with different factors based on where the coefficient values are located in the quantization matrix. | 05-09-2013 |
20130114705 | MULTIVIEW VIDEO CODING - Aspects of this disclosure relate to a method of coding video data. In an example, the method includes obtaining, from an encoded bitstream and for any view component of a first view, reference view information indicating one or more reference views for predicting view components of the first view. The method also includes including, for decoding a first view component in an access unit and in the first view, one or more reference candidates in a reference picture list, where the one or more reference candidates comprise view components in the access unit and in the reference views indicated by the reference view information, where the number of reference candidates is equal to the number of reference views. The method also includes decoding the first view component based on the one or more reference candidates in the reference picture list. | 05-09-2013 |
20130114707 | INTRA-MODE VIDEO CODING - This disclosure relates to methods and apparatus for encoding and decoding video data. In an example an intra-mode for predicting a current block of video data and most probable intra-modes (MPMs) for predicting the current block of video data may be determined. MPMs for predicting the current block of video data may be determined. An index for each of the MPMs based on an order in which the intra-mode for predicting the current block is compared to the MPMs may be determined. When one of the MPMs for predicting the current block matches the intra-mode for predicting the current block, the index of the matching MPM. | 05-09-2013 |
20130114708 | SECONDARY BOUNDARY FILTERING FOR VIDEO CODING - In one example, a video coding device is configured to intra-predict a block of video data, using values of pixels along a primary boundary of the block, to form a predicted block, determine whether to filter the predicted block using data of a secondary boundary of the block, and filter the predicted block using data of the secondary boundary in response to determining to filter the predicted block. The video coding device may determine whether to filter the predicted block based on a comparison of a Laplacian value or a gradient difference value to a threshold. The determination of whether to filter the predicted block may be based at least in part on a boundary relationship, e.g., the relationship of one boundary to another, or of a boundary to pixel values of the predicted block. | 05-09-2013 |
20130114717 | GENERATING ADDITIONAL MERGE CANDIDATES - In generating a candidate list for inter prediction video coding, a video coder can perform pruning operations when adding spatial candidates and temporal candidates to a candidate list while not performing pruning operations when adding an artificially generated candidate to the candidate list. The artificially generated candidate can have motion information that is the same as motion information of a spatial candidate or temporal candidate already in the candidate list. | 05-09-2013 |
20130114730 | CODING SIGNIFICANT COEFFICIENT INFORMATION IN TRANSFORM SKIP MODE - This disclosure describes techniques for coding significant coefficient information for a video block in a transform skip mode. The transform skip mode may provide a choice of a two-dimensional transform mode, a horizontal one-dimensional transform mode, a vertical one-dimensional transform mode, or a no transform mode. In other cases, the transform skip mode may provide a choice between a two-dimensional transform mode and a no transform mode. The techniques include selecting a transform skip mode for a video block, and coding significant coefficient information for the video block using a coding procedure defined based at least in part on the selected transform skip mode. Specifically, the techniques include using different coding procedures to code one or more of a position of a last non-zero coefficient and a significance map for the video block in the transform skip mode. | 05-09-2013 |
20130114734 | CODING SYNTAX ELEMENTS USING VLC CODEWORDS - This disclosure describes techniques for coding transform coefficients for a block of video data. According to these techniques, a video coder (a video encoder or video decoder) stores a first VLC table array selection table in memory, and an indication of at least one difference between the first VLC table array selection table and a second VLC table array selection table. The video coder reconstructs at least one entry of the second VLC table array selection table based on the first VLC table array selection table using the stored indication of the difference between the first VLC table array selection table and a second VLC table array selection table. The video coder uses the reconstructed at least one entry of the second VLC table array selection table to code at least one block of video data. | 05-09-2013 |
20130114736 | PADDING OF SEGMENTS IN CODED SLICE NAL UNITS - A video encoder divides a picture into a plurality of picture partitions, such as tiles or wavefront parallel processing (WPP) waves. The picture partitions are associated with non-overlapping subsets of the treeblocks of the picture. The video encoder generates a coded slice network abstraction layer (NAL) unit that includes encoded representations of the treeblocks associated with a slice of the picture. The coded treeblocks are grouped within the coded slice NAL unit into segments associated with different ones of the picture partitions. The video encoder pads one or more of the segments such that each of the segments begins on a byte boundary. | 05-09-2013 |
20130114738 | PROGRESSIVE CODING OF POSITION OF LAST SIGNIFICANT COEFFICIENT - A video encoder is configured to determine a first and second binary string for a value indicating the position of the last significant coefficient, within a video block of size T. A video decoder is configured to determine a value indicating the position of a last significant coefficient within a video block of size T based on a first and second binary string. In one example, the first binary string is based on a truncated unary coding scheme defined by a maximum bit length defined by 2 log | 05-09-2013 |
20130121406 | 8-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing an 8-point discrete cosine transform (DCT). An apparatus comprising an 8-point discrete cosine transform (DCT) hardware unit may implement these techniques to transform media data from a spatial domain to a frequency domain. The 8-point DCT hardware unit includes an even portion comprising factors A, B that are related to a first scaled factor (μ) in accordance with a first relationship. The 8-point DCT hardware unit also includes an odd portion comprising third, fourth, fifth and sixth internal factors (G, D, E, Z) that are related to a second scaled factor (η) in accordance with a second relationship. The first relationship relates the first scaled factor to the first and second internal factors. The second relationship relates the second scaled factor to the third internal factor and a fourth internal factor, as well as, the fifth internal factor and a sixth internal factor. | 05-16-2013 |
20130121417 | CONSTRAINED REFERENCE PICTURE SETS IN WAVE FRONT PARALLEL PROCESSING OF VIDEO DATA - A video encoder determines reference blocks for each inter-predicted prediction unit (PU) of a tree block group such that each of the reference blocks is in a reference picture that is in a reference picture subset for the tree block group. The reference picture subset for the tree block group includes less than all reference pictures in a reference picture set of the current picture. The tree block group comprises a plurality of concurrently-coded tree blocks in the current picture. For each inter-predicted PU of the tree block group, the video encoder indicates, in a bitstream that includes a coded representation of video data, a reference picture that includes the reference block for the inter-predicted PU. A video decoder receives the bitstream, determines the reference pictures of the inter-predicted PUs of the tree block group, and generates decoded video blocks using the reference blocks of the inter-predicted PUs. | 05-16-2013 |
20130127987 | SIGNALING DEPTH RANGES FOR THREE-DIMENSIONAL VIDEO CODING - In one example, a video coder, such as a video encoder or a video decoder, is configured to code a first set of one or more depth range values for a first set of video data, wherein the first set of one or more depth range values have respective first precisions, code a second set of one or more depth range values for a second set of video data, wherein the second set of one or more depth range values have respective second precisions different than the respective first precisions, and code at least a portion of the second set of video data using the second set of one or more depth range values. In this manner, the video coder may update precisions (e.g., numbers of bits) used to represent depth range values for coding multiview plus depth video data. | 05-23-2013 |
20130128964 | REFERENCE MODE SELECTION IN INTRA MODE CODING - A video coder can determine a first most probable intra prediction mode and a second most probable intra prediction mode for a current video block, and based on the first and second most probable intra prediction modes, determine an estimate of an actual intra prediction mode used to code the current video block. The estimate of the actual intra prediction mode can be determined by selecting one of the first most probable mode and the second most probable mode as the estimate of the actual intra prediction in response to both the first most probable mode and the second most probable mode being different angular prediction modes. The estimate of the actual intra prediction mode plus difference information can be used to identify the actual intra prediction mode used to code the current video block. | 05-23-2013 |
20130128965 | INSIDE VIEW MOTION PREDICTION AMONG TEXTURE AND DEPTH VIEW COMPONENTS - The techniques of this disclosure may be generally related to using motion information for a corresponding block from a texture view component that corresponds with a block in a depth view component in coding the block in the depth view component. In some examples, for coding purposes, the techniques may use motion information when the spatial resolution of the texture view component is different than the spatial resolution of the depth view component. | 05-23-2013 |
20130128971 | TRANSFORMS IN VIDEO CODING - Aspects of this disclosure relate to coding video data. In an example, a method of coding video data includes determining a first residual quadtree (RQT) depth at which to apply one or more first transforms to residual video data based on at least one characteristic of the residual of video data. The method also includes determining a second RQT depth at which to apply one or more second transforms to the residual video data based on the at least one characteristic. The method also includes coding the residual video data using the one or more first transforms and the one or more second transforms. | 05-23-2013 |
20130128974 | ADAPTIVE OVERLAPPED BLOCK MOTION COMPENSATION - In general, techniques are described for performing adaptive overlapped block motion compensation when coding video data. A video coding device configured to code video data may implement the techniques. The video coding device may comprise one or more processors configured to determine an adaptation metric from a region of support for one of a first partition and a second partition of one or more blocks of the video data separated by a partition boundary. The region of support may not be the entire one or more blocks of video data. The one or more processors may further be configured to adapt, based on the adaptation metric, application of an overlapped block motion compensation process to determine predicted pixel values near the partition boundary for a first predicted block representative of the first partition of the video block. | 05-23-2013 |
20130136167 | LARGEST CODING UNIT (LCU) OR PARTITION-BASED SYNTAX FOR ADAPTIVE LOOP FILTER AND SAMPLE ADAPTIVE OFFSET IN VIDEO CODING - This disclosure relates to techniques for performing sample adaptive offset (SAO) processes in a video coding process. A video coder may store sets of SAO information. The SAO information may include data indicative of offset values. The video coder may also store mapping information that maps at least some of the sets of SAO information for one or more sequence partitions of a frame of video data. Additionally, the video coder may perform the SAO processes for one of the partitions of the frame based on the stored SAO information and the stored mapping information. | 05-30-2013 |
20130136175 | NON-SQUARE TRANSFORM UNITS AND PREDICTION UNITS IN VIDEO CODING - This disclosure proposes techniques for transform partitioning in an intra-prediction video coding process. In one example, for a given intra-predicted block, a reduced number of transform unit partition options is allowed, based on certain conditions. In another example, transform units are decoupled from prediction units for intra-predicted block. For a given prediction unit, transforms of different sizes and shapes from the prediction unit may be applied. In another example, a reduced number of intra-prediction modes are allowed for a prediction unit having a non-square shape. | 05-30-2013 |
20130148718 | 8-POINT TRANSFORM FOR MEDIA DATA CODING - In general, techniques are described for implementing an 8-point inverse discrete cosine transform (IDCT). An apparatus comprising an 8-point inverse discrete cosine transform (IDCT) hardware unit may implement these techniques to transform media data from a frequency domain to a spatial domain. The 8-point IDCT hardware unit includes an even portion comprising factors A, B that are related to a first scaled factor (μ) in accordance with a first relationship. The 8-point IDCT hardware unit also includes an odd portion comprising third, fourth, fifth and sixth internal factors (G, D, E, Z) that are related to a second scaled factor (η) in accordance with a second relationship. The first relationship relates the first scaled factor to the first and second internal factors. The second relationship relates the second scaled factor to the third, fourth, fifth and sixth internal factors. | 06-13-2013 |
20130148722 | REFERENCE PICTURE LIST MODIFICATION FOR VIEW SYNTHESIS REFERENCE PICTURES - A video encoder generates a bitstream that includes a reference picture list modification (RPLM) command. The RPLM command belongs to a type of RPLM commands for inserting short-term reference pictures into reference picture lists. The RPLM command instructs a video decoder to insert a synthetic reference picture into the reference picture list. The video decoder decodes, based at least in part on syntax elements parsed from the bitstream, one or more view components and generates, based at least in part on the one or more view components, the synthetic reference picture. The video decoder modifies, in response to the RPLM commands, a reference picture list to include the synthetic reference picture. The video decoder may use one or more pictures in the reference picture list as reference pictures to perform inter prediction on one or more video blocks of a picture. | 06-13-2013 |
20130163664 | UNIFIED PARTITION MODE TABLE FOR INTRA-MODE CODING - In an example, aspects of this disclosure relate to a method for coding video data that includes predicting a first non-square partition of a current block of video data using a first intra-prediction mode, where the first non-square partition has a first size. The method also includes predicting a second non-square partition of the current block of video data using a second intra-prediction mode, where the second non-square partition has a second size different than the first size. The method also includes coding the current block based on the predicted first and second non-square partitions. | 06-27-2013 |
20130163668 | PERFORMING MOTION VECTOR PREDICTION FOR VIDEO CODING - In general, techniques are described for performing motion vector prediction for video coding. A video coding device comprising a processor may perform the techniques. The processor may be configured to determine a plurality of candidate motion vectors for a current block of the video data so as to perform the motion vector prediction process and scale one or more of the plurality of candidate motion vectors determined for the current block of the video data to generate one or more scaled candidate motion vectors. The processor may then be configured to modify the scaled candidate motion vectors to be within a specified range. | 06-27-2013 |
20130170553 | CODING MOTION VECTOR DIFFERENCE - The techniques described in this disclosure may be generally related to identifying when motion vector difference (MVD) is skipped for one or both reference picture lists. The techniques may further relate to contexts for signaling MVD values. The techniques may also be related to syntax that indicates when at least one of the MVD values is zero. | 07-04-2013 |
20130170562 | DEBLOCKING DECISION FUNCTIONS FOR VIDEO CODING - In one example, a video coding device is configured to decode four blocks of video data, wherein the four blocks are non-overlapping and share one common point such that four edge segments are formed by the four blocks, for each of the four edge segments, determine whether to deblock the respective edge segment based on a first analysis of at least one line of pixels that is perpendicular to the respective edge segment and that intersects the respective edge segment, for each of the four edge segments that was determined to be deblocked, determine whether to apply a strong filter or a weak filter to the respective edge segment based on a second analysis of the at least one line of pixels for the respective edge, and deblock the four edge segments based on the determinations. | 07-04-2013 |
20130176389 | SIGNALING VIEW SYNTHESIS PREDICTION SUPPORT IN 3D VIDEO CODING - In one example, a video coder is configured to code information indicative of whether view synthesis prediction is enabled for video data. When the information indicates that view synthesis prediction is enabled for the video data, the video coder may generate a view synthesis picture using the video data and code at least a portion of a current picture relative to the view synthesis picture. The at least portion of the current picture may comprise, for example, a block (e.g., a PU, a CU, a macroblock, or a partition of a macroblock), a slice, a tile, a wavefront, or the entirety of the current picture. On the other hand, when the information indicates that view synthesis prediction is not enabled for the video data, the video coder may code the current picture using at least one of intra-prediction, temporal inter-prediction, and inter-view prediction without reference to any view synthesis pictures. | 07-11-2013 |
20130176390 | MULTI-HYPOTHESIS DISPARITY VECTOR CONSTRUCTION IN 3D VIDEO CODING WITH DEPTH - A method and apparatus for decoding and encoding multiview video data is described. An example method may include coding a block of video data using a motion vector prediction process, determining a motion vector candidate list, determining a disparity vector candidate list for the motion prediction process, wherein the disparity vector candidate list includes at least two types of disparity vectors from a plurality of disparity vector types, the plurality including a spatial disparity vector (SDV), a smooth temporal-view (STV) disparity vector, a view disparity vector (VDV), and a temporal disparity vector (TDV), and performing the motion vector prediction process using one of the disparity vector candidate list and the motion vector candidate list. | 07-11-2013 |
20130177070 | SIGNIFICANCE MAP SUPPORT FOR PARALLEL TRANSFORM COEFFICIENT PROCESSING IN VIDEO CODING - In an example, aspects of this disclosure relate to a process for video coding that includes determining that a set of support for selecting a context model to code a current significant coefficient flag of a transform coefficient of a block of video data includes at least one significant coefficient flag that is not available. The process also includes, based on the determination, modifying the set of support, and calculating a context for the current significant coefficient flag using the modified set of support. The process also includes applying context-adaptive binary arithmetic coding (CABAC) to code the current significant coefficient flag based on the calculated context. | 07-11-2013 |
20130177083 | MOTION VECTOR CANDIDATE INDEX SIGNALING IN VIDEO CODING - A video encoder generates a first and a second candidate list. The first candidate list includes a plurality of motion vector (MV) candidates. The video encoder selects, from the first candidate list, a MV candidate for a first prediction unit (PU) of a coding unit (CU). The second MV candidate list includes each of the MV candidates of the first MV candidate list except the MV candidate selected for the first PU. The video encoder selects, from the second MV candidate list, a MV candidate for a second PU of the CU. A video decoder generates the first and second MV candidate lists in a similar way and generates predictive sample blocks for the first and second PUs based on motion information of the selected MV candidates. | 07-11-2013 |
20130182757 | THROUGHPUT IMPROVEMENT FOR CABAC COEFFICIENT LEVEL CODING - This disclosure proposes various techniques for limiting the number of bins that are coded using an adaptive context model with context adaptive binary arithmetic coding (CABAC). In particular, this disclosure proposes to limit the number of bins that use CABAC for coding level information of transform coefficients in a video coding process. | 07-18-2013 |
20130182758 | DETERMINING CONTEXTS FOR CODING TRANSFORM COEFFICIENT DATA IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to determine values for coded sub-block flags of one or more neighboring sub-blocks to a current sub-block, determine a context for coding a transform coefficient of the current sub-block based on the values for the coded sub-block flags, and entropy code the transform coefficient using the determined context. | 07-18-2013 |
20130182772 | DETERMINING CONTEXTS FOR CODING TRANSFORM COEFFICIENT DATA IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to determine a context for coding a transform coefficient of a video block based on a region of the video block in which the transform coefficient occurs, and entropy code the transform coefficient using the determined context. The region may comprise one of a first region comprising one or more upper-left 4×4 sub-blocks of transform coefficients of the video block and a second region comprising transform coefficients of the video block outside the first region. | 07-18-2013 |
20130182773 | DETERMINING CONTEXTS FOR CODING TRANSFORM COEFFICIENT DATA IN VIDEO CODING - In one example, a device for coding video data includes a video coder configured to determine whether a transform coefficient of a video block is a DC transform coefficient, when the transform coefficient is determined to be the DC transform coefficient of the video block, determine a context for coding the transform coefficient based on the transform coefficient being the DC transform coefficient without regard for a size of the video block, and entropy code the transform coefficient using the determined context. | 07-18-2013 |
20130188013 | MVC BASED 3DVC CODEC SUPPORTING INSIDE VIEW MOTION PREDICTION (IVMP) MODE - This disclosure describes features and techniques applicable to three-dimensional (3D) video coding. In one example, a technique may include coding a texture view video block, and coding a depth view video block, wherein the depth view video block is associated with the texture view video block. Coding the depth view video block may include coding a syntax element to indicate whether or not motion information associated with the texture view video block is adopted as motion information associated with the depth view video block. | 07-25-2013 |
20130188698 | COEFFICIENT LEVEL CODING - In one example, a device includes a video coder configured to code a first set of syntax elements for the coefficients of a residual block of video data, and code, using at least a portion of the first set of syntax elements as context data, a second set of syntax elements for the coefficients, wherein the first set of syntax elements each correspond to a first type of syntax element for the coefficients, and wherein the second set of syntax elements each correspond to a second, different type of syntax element for the coefficients. For example, the first set of syntax elements may comprise values indicating whether the coefficients are significant (that is, have non-zero level values), and the second set of syntax elements may comprise values indicating whether level values for the coefficients have absolute values greater than one. | 07-25-2013 |
20130188699 | CODING OF COEFFICIENTS IN VIDEO CODING - A video encoder performs multiple coding passes on coefficients in a coefficient block. During each coding pass, the video encoder encodes a different set of syntax elements for coefficients in the coefficient block. A video decoder uses the syntax elements for a coefficient to determine the value of the coefficient. When the video encoder performs a coding pass, the video encoder selects a coding context for a syntax element based at least in part on a syntax element generated in an earlier coding pass for a non-causal coefficient and based at least in part on a value of a syntax element generated during the current coding pass for a causal coefficient. The video encoder entropy encodes the syntax element based on the selected coding context. A video decoder performs a similar series of coding passes in which the video decoder selects coding contexts and entropy decodes the syntax elements. | 07-25-2013 |
20130188700 | CONTEXT ADAPTIVE ENTROPY CODING WITH A REDUCED INITIALIZATION VALUE SET - Techniques for coding data, such as, e.g., video data, include coding a first syntax element, conforming to a particular type of syntax element, of a first slice of video data, conforming to a first slice type, using an initialization value set. The techniques further include coding a second syntax element, conforming to the same type of syntax element, of a second slice of video data, conforming to a second slice type, using the same initialization value set. In this example, the first slice type may be different from the second slice type. Also in this example, at least one of the first slice type and the second slice type may be a temporally predicted slice type. For example, the at least one of the first and second slice types may be a unidirectional inter-prediction (P) slice type, or a bi-directional inter-prediction (B) slice type. | 07-25-2013 |
20130188701 | SUB-BLOCK LEVEL PARALLEL VIDEO CODING - The techniques of this disclosure are generally related to parallel coding of video units that reside along rows or columns of blocks in largest coding units. For example, the techniques include removing intra-prediction dependencies between two video units in different rows or columns to allow for parallel coding of rows or columns of the video units. | 07-25-2013 |
20130188715 | DEVICE AND METHODS FOR MERGE LIST REORDERING IN VIDEO CODING - A video coding device configured according to some aspects of this disclosure includes a memory configured to store an initial list of motion vector candidates and a temporal motion vector predictor (TMVP). The video coding device also includes a processor in communication with the memory. The processor is configured to obtain a merge candidate list size value (N) and identify motion vector candidates to include in a merge candidate list having a list size equal to the merge candidate list size value. The merge candidate list may be a merge motion vector (MV) candidate list or a motion vector predictor (MVP) candidate list (also known as an AMVP candidate list). The processor generates the merge candidate list such that the merge candidate list includes the TMVP, regardless of the list size. | 07-25-2013 |
20130188716 | TEMPORAL MOTION VECTOR PREDICTOR CANDIDATE - The techniques of this disclosure may be generally related to temporal motion vector prediction candidate. A video coder may determine a temporal motion vector prediction candidate for a plurality of blocks only once. Each of the plurality of blocks may include different spatial motion vector prediction candidates, but the temporal motion vector prediction candidate for the plurality of blocks may be the same. | 07-25-2013 |
20130188717 | MOTION PREDICTION IN SVC USING PARTITION MODE WITHOUT SPLIT FLAG - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data. The video data includes a base layer and an enhancement layer. The base layer includes a coding unit tree co-located with an enhancement layer coding unit in the enhancement layer. The coding unit tree includes a plurality of nodes arranged in a tree structure and motion vectors. The enhancement layer coding unit is inter-mode coded. A processor is configured to split the enhancement layer coding unit into a plurality of nodes arranged in a tree structure that is the same as the tree structure of the coding unit tree. The processor is also configured to perform motion prediction for the enhancement layer coding unit based on the motion vectors of the coding unit tree. | 07-25-2013 |
20130188718 | MOTION PREDICTION IN SVC WITHOUT INCLUDING A TEMPORALLY NEIGHBORING BLOCK MOTION VECTOR IN A CANDIDATE LIST - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data and a candidate list. The video data includes a base layer and an enhancement layer. The base layer includes a base layer prediction unit co-located with an enhancement layer prediction unit in the enhancement layer. The candidate list includes a list of motion vectors for use by the enhancement layer prediction unit. A processor is configured to store motion vectors originating from spatial neighbors of the enhancement layer prediction unit, and not motion vectors originating from temporal neighbors of the enhancement layer prediction unit, in the candidate list. The processor is also configured to store motion vectors originating from the base layer prediction unit in the candidate list. | 07-25-2013 |
20130188719 | MOTION PREDICTION IN SVC USING MOTION VECTOR FOR INTRA-CODED BLOCK - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory unit is configured to store the video data. The video data may include a base layer and an enhancement layer. The base layer may include a base layer coding unit co-located with a first enhancement layer coding unit in the enhancement layer. A processor may be configured to construct one or more motion vectors based at least in part on one or more base layer motion vectors available at the co-located base layer coding unit. The one or more motion vectors may be associated with the first enhancement layer coding unit. The processor may also be configured to determine pixel values of a neighbor enhancement layer coding unit based at least in part on the one or more motion vectors. | 07-25-2013 |
20130188720 | VIDEO CODING USING PARALLEL MOTION ESTIMATION - An example video encoder is configured to receive an indication of merge mode coding of a block within a parallel motion estimation region (PMER), generate a merge mode candidate list comprising one or more spatial neighbor motion vector (MV) candidates and one or more temporal motion vector prediction (TMVP) candidates, wherein motion information of at least one of the spatial neighbor MV candidates is known to be unavailable during coding of the block at an encoder, determine an index value identifying, within the merge mode candidate list, one of the TMVP candidates or the spatial neighbor MV candidates for which motion information is available during coding of the particular block, and merge mode code the block using the identified MV candidate. | 07-25-2013 |
20130188733 | SIGNALING OF DEBLOCKING FILTER PARAMETERS IN VIDEO CODING - This disclosure describes techniques for signaling deblocking filter parameters for a current slice of video data with reduced bitstream overhead. Deblocking filter parameters may be coded in one or more of a picture layer parameter set and a slice header. The techniques reduce a number of bits used to signal the deblocking filter parameters by coding a first syntax element that indicates whether deblocking filter parameters are present in both the picture layer parameter set and the slice header, and only coding a second syntax element in the slice header when both sets of deblocking filter parameters are present. Coding the second syntax element is eliminated when deblocking filter parameters are present in only one of the picture layer parameter set or the slice header. The second syntax element indicates which set of deblocking filter parameters to use to define a deblocking filter applied to a current slice. | 07-25-2013 |
20130188744 | DEBLOCKING CHROMA DATA FOR VIDEO CODING - A video coding device is configured to obtain an array of sample values. The sample values may be formatted according to a 4:2:0, 4:2:2, or 4:4:4 chroma format. The video coding device determines whether to apply a first filter to rows of chroma sample values associated with defined horizontal edges within the array. The video coding device determines whether to apply a second filter to columns of chroma sample values associated with defined vertical edges. The horizontal and vertical edges may be separated by a number of chroma samples according to a deblocking grid. | 07-25-2013 |
20130195189 | IMPLICIT DERIVATION OF PARALLEL MOTION ESTIMATION RANGE SIZE - A method for decoding video data is described. The method may comprise receiving an indication of a size of a parallel motion estimation (PME) area, performing a motion vector prediction process on coding units having a size smaller than or equal to the PME area using a PME style candidate list construction process and the PME area, deriving an implicit PME area for coding units having a size larger than the PME area, and performing the motion vector prediction process on coding units having the size larger than the PME area using the PME style candidate list construction process and the implicit PME area. | 08-01-2013 |
20130195199 | RESIDUAL QUAD TREE (RQT) CODING FOR VIDEO CODING - A video decoding device receives an array of transform coefficients for a chroma component of video data. The video decoding device receives entropy encoded data representing the value of a split flag associated with the chroma component. The value of the split flag indicates whether the array of transform coefficients is divided into smaller transform blocks. The video decoding device determines a context for the entropy encoded data representing the split flag. The context is based on the value of a split flag associated with another component of video data. The video decoding device entropy decodes the data representing the value of the split flag based on the determined context using context adaptive binary arithmetic coding (CABAC). The luma and chroma components have independent residual quadtree (RQT) structures. | 08-01-2013 |
20130202037 | RESTRICTION OF PREDICTION UNITS IN B SLICES TO UNI-DIRECTIONAL INTER PREDICTION - A computing device determines whether a prediction unit (PU) in a B slice is restricted to uni-directional inter prediction. In addition, the computing device generates a merge candidate list for the PU and determines a selected merge candidate in the merge candidate list. If the PU is restricted to uni-directional inter prediction, the computing device generates a predictive video block for the PU based on no more than one reference block associated with motion information specified by the selected merge candidate. If the PU is not restricted to uni-directional inter prediction, the computing device generates the predictive video block for the PU based on one or more reference blocks associated with the motion information specified by the selected merge candidate. | 08-08-2013 |
20130202038 | RESTRICTION OF PREDICTION UNITS IN B SLICES TO UNI-DIRECTIONAL INTER PREDICTION - A video coding device generates a motion vector (MV) candidate list for a prediction unit (PU) of a coding unit (CU) that is partitioned into four equally-sized PUs. The video coding device converts a bi-directional MV candidate in the MV candidate list into a uni-directional MV candidate. In addition, the video coding device determines a selected MV candidate in the merge candidate list and generates a predictive video block for the PU based at least in part on one or more reference blocks indicated by motion information specified by the selected MV candidate. | 08-08-2013 |
20130215974 | CODING OF LOOP FILTER PARAMETERS USING A CODEBOOK IN VIDEO CODING - Techniques for coding video data include coding sample adaptive offset (SAO) offset values as part of performing a video coding process. In particular, the techniques include determining the SAO offset values according to a SAO process. The techniques further include storing a codebook defining a plurality of codes for coding different variations of SAO offset values. The techniques also include coding the SAO offset values in accordance with the codebook so as to specify the SAO offset values as one of the plurality of codes defined by the codebook. | 08-22-2013 |
20130222537 | BITSTREAM EXTRACTION IN THREE-DIMENSIONAL VIDEO - To extract a sub-bitstream from a 3-dimensional video (3DV) bitstream, a device determines a texture target view list that indicates views in the 3DV bitstream that have texture view components that are required for decoding pictures in a plurality of target views. The target views are a subset of the views in the bitstream that are to be decodable from the sub-bitstream. In addition, the device determines a depth target view list that indicates views in the 3DV bitstream that have depth view components that are required for decoding pictures in the plurality of target views. The device determines the sub-bitstream based at least in part on the texture target view list and the depth target view list. | 08-29-2013 |
20130222538 | NETWORK ABSTRACTION LAYER (NAL) UNIT HEADER DESIGN FOR THREE-DIMENSIONAL VIDEO CODING - A video encoder generates a network abstraction layer (NAL) unit that includes at least a first syntax element and a second syntax element. The first syntax element indicates that the NAL unit belongs to a particular NAL unit type. Coded slices of texture view components and depth view components are encapsulated within NAL units that belong to the particular NAL unit type. The second syntax element indicates whether a NAL unit header of the NAL unit includes an Advanced Video Coding (AVC)-compatible 3-dimensional video (3DV) header extension or includes a Multiview Video Coding (MVC)-compatible 3DV header extension. The video encoder outputs a bitstream that includes the NAL unit. A video decoder receives the NAL unit and determines whether the second syntax element indicates that the NAL unit header of the NAL unit includes the AVC-compatible 3DV header extension or the MVC-compatible 3DV header extension. | 08-29-2013 |
20130230097 | SCAN-BASED SLIDING WINDOW IN CONTEXT DERIVATION FOR TRANSFORM COEFFICIENT CODING - A video coding process that includes defining a context derivation neighborhood for one of a plurality of transform coefficients based on a transform coefficient scan order. The process also includes determining a context for the one of the plurality of transform coefficients based on the context derivation neighborhood. The process also includes coding the one of the plurality of transform coefficients based on the determined context. | 09-05-2013 |
20130242046 | DISPARITY VECTOR PREDICTION IN VIDEO CODING - Techniques are described for determining a disparity vector for a current block based on disparity motion vectors of one or more spatially and temporally neighboring regions to a current block to be predicted. The spatially and temporally neighboring regions include one or a plurality of blocks, and the disparity motion vector represents a single vector in one reference picture list for the plurality of blocks within the spatially or temporally neighboring region. The determined disparity vector could be used to coding tools which utilize the information between different views such as merge mode, advanced motion vector prediction (AMVP) mode, inter-view motion prediction, and inter-view residual prediction. | 09-19-2013 |
20130251026 | CODED BLOCK FLAG INFERENCE IN VIDEO CODING - A video decoder determines whether a first condition is satisfied. The first condition is satisfied when any sibling residual quad tree (RQT) node of a current RQT node is associated with a significant luma coefficient block. In addition, the video decoder determines whether a second condition is satisfied. The second condition is satisfied when a chroma transform block of a parent RQT node of the current RQT node is significant and is split along with a luma transform block of the parent RQT node. In response to determining that neither the first nor the second condition is satisfied, the video decoder determines that a luma coded block flag (CBF) for the current RQT node is omitted from the bitstream. | 09-26-2013 |
20130251030 | INTER LAYER TEXTURE PREDICTION FOR VIDEO CODING - An apparatus for coding video information according to certain aspects is disclosed. Multi-layer video steams including a base layer and an enhancement layer can be coded. Predictors generated for the base layer and the enhancement layer can be combined to form a final predictor of the enhancement layer. Each predictor can be weighted such that those predictors which are more likely to produce high quality results can be factored more heavily in the final predictor. The conditions upon which the respective weights for enhancement layer predictors and base layer predictors are determined may be implicitly derived from the predictors or characteristics thereof. Alternatively, data may be generated explicating indicating the weights or providing information from which the weights can be determined. | 09-26-2013 |
20130251041 | DERIVING CONTEXT FOR LAST POSITION CODING FOR VIDEO CODING - In one example, a device includes a video coder configured to determine a context for entropy coding a bin of a value indicative of a last significant coefficient of a block of video data using a function of an index of the bin, and code the bin using the determined context. The video coder may encode or decode the bin using context-adaptive binary arithmetic coding (CABAC). The function may also depend on a size of the block. In this manner, a table indicating context indexes for the contexts need not be stored by the device. | 09-26-2013 |
20130258049 | MERGE SIGNALING AND LOOP FILTER ON/OFF SIGNALING - Systems, methods, and devices are disclosed that encode video, decode video, or both. These systems, methods, and devices generate and/or receive an enable syntax element in an encoded bitstream, wherein the enable syntax element indicates whether a loop filter is turned on or turned off for a group of video blocks. They also generate or receive one or more additional syntax elements identifying parameters for the loop filter for the group of video blocks in response to the enable syntax element indicating the loop filter is turned on for the group of video blocks. These systems, methods, and devices also perform the loop filter for the group of video blocks based on the received enable syntax element. | 10-03-2013 |
20130258052 | INTER-VIEW RESIDUAL PREDICTION IN 3D VIDEO CODING - In general, this disclosure describes techniques for improved inter-view residual prediction (IVRP) in three-dimensional video coding. These techniques include determining IVRP availability based on coded block flags and coding modes of residual reference blocks, disallowing IVRP coding when a block is inter-view predicted, using picture order count (POC) values to determine whether IVRP is permitted, applying IVRP to prediction units (PUs) rather than coding units (CUs), inferring values of IVRP flags when a block is skip or merge mode coded, using an IVRP flag of a neighboring block to determine context for coding an IVRP flag of a current block, and avoiding resetting of samples of a residual reference block to zeros during generation. | 10-03-2013 |
20130259120 | QUANTIZATION MATRIX AND DEBLOCKING FILTER ADJUSTMENTS FOR VIDEO CODING - A device may include a video coder configured to determine an equivalent quantization parameter (QP) for a decoded block of video data using a quantization matrix for the decoded block of video data, determine deblocking parameters based on the determined equivalent QP, and deblock an edge of the decoded block based on the determined deblocking parameters. In particular, the video coder may determine equivalent QPs for two neighboring blocks defining a common edge, and deblock the common edge based on the equivalent QPs. The video coder may determine deblocking parameters, such as β and t | 10-03-2013 |
20130259130 | WEIGHTED PREDICTION PARAMETER CODING - The techniques are generally related to the coding of weighted prediction parameters. A video coder may determine the weighted prediction parameters for a reference picture list based on coded weighted prediction parameters for another reference picture list. Examples of the reference picture list include reference picture lists constructed for coding purposes, including a combined reference picture list. | 10-03-2013 |
20130259141 | CHROMA SLICE-LEVEL QP OFFSET AND DEBLOCKING - In one example, an apparatus for processing video data comprises a video coder configured to, for each of the one or more chrominance components, calculate a chrominance quantization parameter for a common edge between two blocks of video data based on a first luminance quantization parameter for the first block of video data, a second luminance quantization parameter for the second block of video data, and a chrominance quantization parameter offset value for the chrominance component. The video coder is further configured to determine a strength for a deblocking filter for the common edge based on the chrominance quantization parameter for the chrominance component, and apply the deblocking filter according to the determined strength to deblock the common edge. | 10-03-2013 |
20130265388 | DISPARITY VECTOR CONSTRUCTION METHOD FOR 3D-HEVC - When coding multiview video data, a video coder can code one or more pictures in one or more reference views, including a first reference view and determine a disparity vector for a current block based on motion information of one or more neighboring blocks of the current block, wherein the current block is in a second view, wherein the disparity vector points from the current block to a corresponding block in a picture of the same time instance in one of the one or more reference views. | 10-10-2013 |
20130266059 | LCU-BASED ADAPTIVE LOOP FILTERING FOR VIDEO CODING - This disclosure describes techniques associated with filtering reconstructed video data in a video encoding and/or video decoding processes and, more particularly, this disclosure describes techniques related to adaptive loop filtering. According to one technique of this disclosure, instead of implementing region adaptive (RA) classification using a fixed set of regions, a variable number of regions in a picture can be used. According to another technique of this disclosure, a global set of filters can be signaled in a parameter set from an encoder to a decoder, but instead of selecting filters from the global set based on an RA classification or a block adaptive (BA) classification, a filter from the global set can be explicitly signaled, for a given block, using a filter ID associated with the filter. | 10-10-2013 |
20130266074 | CODED BLOCK FLAG CODING - A video encoder generates a bitstream that includes a residual quad tree (RQT) for a coding unit (CU). The CU is larger than a maximum-allowable transform unit (TU) size and the RQT includes a hierarchy of nodes. A root node of the RQT corresponds to the CU as a whole and leaf nodes of the RQT correspond to TUs of the CU. The root node is associated with a coded block flag (CBF) for a chroma component. The CBF for the chroma component indicates whether any of the TUs of the CU are associated with a significant coefficient block that is based on samples of the particular chroma component. A video decoder receives the bitstream and determines, based on the CBF, whether coefficient blocks associated with TUs that correspond to the leaf nodes include non-zero coefficients. | 10-10-2013 |
20130271566 | VIEW SYNTHESIS MODE FOR THREE-DIMENSIONAL VIDEO CODING - A video encoder signals, in a bitstream, a syntax element that indicates whether a current video unit is predicted from a VSP picture. The current video unit is a macroblock or a macroblock partition. The video encoder determines, based at least in part on whether the current video unit is predicted from the VSP picture, whether to signal, in the bitstream, motion information for the current video unit. A video decoder decodes the syntax element from the bitstream and determines, based at least in part on the syntax element, whether the bitstream includes the motion information. | 10-17-2013 |
20130272370 | WAVEFRONT PARALLEL PROCESSING FOR VIDEO CODING - In one example, a video coder may be configured to determine that a slice of a picture of video data begins in a row of coding tree units (CTUs) in the picture at a position other than a beginning of the row. Based on the determination, the video coder may be further configured to determine that the slice ends within the row of CTUs. The video coder may be further configured to code the slice based on the determination that the slice ends within the row of CTUs. | 10-17-2013 |
20130272377 | BYPASS BINS FOR REFERENCE INDEX CODING IN VIDEO CODING - In an example, aspects of this disclosure relate to a method for decoding a reference index syntax element in a video decoding process that includes decoding at least one bin of a reference index value with a context coding mode of a context-adaptive binary arithmetic coding (CABAC) process. The method also includes decoding, when the reference index value comprises more bins than the at least one bin coded with the context coded mode, at least another bin of the reference index value with a bypass coding mode of the CABAC process, and binarizing the reference index value. | 10-17-2013 |
20130272378 | COEFFICIENT GROUPS AND COEFFICIENT CODING FOR COEFFICIENT SCANS - Techniques are described for a video coder (e.g., video encoder or video decoder) that is configured to select a context pattern from a plurality of context patterns that are the same for a plurality of scan types. Techniques are also described for a video coder that is configured to select a context pattern that is stored as a one-dimensional context pattern and identifies contexts for two or more scan types. | 10-17-2013 |
20130272379 | COEFFICIENT GROUPS AND COEFFICIENT CODING FOR COEFFICIENT SCANS - Techniques are described for a video coder (e.g., video encoder or video decoder) that is configured to select a context pattern from a plurality of context patterns that are the same for a plurality of scan types. Techniques are also described for a video coder that is configured to select a context pattern that is stored as a one-dimensional context pattern and identifies contexts for two or more scan types. | 10-17-2013 |
20130272380 | GROUPING BYPASS CODED SYNTAX ELEMENTS IN VIDEO CODING - A video encoding device is configured to generate a first group of syntax elements. Each syntax element in the first group indicates whether a prediction mode of a respective prediction unit (PU) is based on an index into a list of most probable modes. A second group of syntax elements is generated that correspond to respective syntax elements in the first group. The syntax elements in the second group identify either an index into the list of most probable modes or an intra-prediction mode. The first group of syntax elements are context adaptive binary arithmetic coding (CABAC) encoded, and the second group of syntax elements are bypass encoded. A video decoder is configured to receive the entropy encoded first and second groups of syntax elements. The video decoder CABAC decodes the first group of flags and bypass decodes the second group of flags. | 10-17-2013 |
20130272381 | SIMPLIFIED NON-SQUARE QUADTREE TRANSFORMS FOR VIDEO CODING - In an example, a method of decoding video data includes determining a prediction partitioning structure for predicting pixel values associated with a block of video data. The method also includes determining a transform partitioning structure for applying one or more transforms to the predicted pixel values. Determining the transform split structure includes splitting a parent transform unit, upon determining the transform partitioning structure comprises splitting the parent transform unit into one or more square transforms, determining one or more square transforms such that each of the one or more square transforms correspond to exactly one prediction partition, and upon determining the transform partitioning structure comprises splitting the parent transform unit into one or more non-square transforms, determining whether to split the one or more non-square transforms based at least in part on the one or more non-square transforms being non-square. | 10-17-2013 |
20130272385 | LEVEL DECISION IN RATE DISTORTION OPTIMIZED QUANTIZATION - A computing device, such as a video encoder, determines an initial quantized level for a coefficient of a coefficient block and determines whether the coefficient is less than the product of the initial quantized level and a quantization step size value. In response to determining that the coefficient is less than the product of the initial quantized level and the quantization step size value, the computing device determines rate-distortion costs of quantizing the coefficient to be the initial quantized level for the coefficient, the initial quantized level minus one, and in some circumstances, 0. The computing device determines an actual quantized level for the coefficient based at least in part on the calculated rate-distortion costs and includes the actual quantized level in a quantized version of the coefficient block. | 10-17-2013 |
20130272386 | LOOKUP TABLE FOR RATE DISTORTION OPTIMIZED QUANTIZATION - A computing device, such as a video encoder, uses respective positions of respective coefficients in a coefficient block to look up, in a lookup table, respective quantization offsets for the respective coefficients. Furthermore, the computing device determines, based at least in part on the quantization offsets for the one or more coefficients, respective quantized levels for the respective coefficients. | 10-17-2013 |
20130272390 | UNIFORM GRANULARITY FOR QUANTIZATION MATRIX IN VIDEO CODING - The techniques of this disclosure are directed toward the use of modified quantization parameter (QP) values to calculate quantized and dequantized transform coefficients of a video block with uniform QP granularity. Conventionally, when a quantization matrix is used during quantization and dequantization of transform coefficients, the quantization matrix entries act as scale factors of a quantizer step-step corresponding to a base QP value, which results in non-uniform QP granularity. To provide uniform QP granularity across all quantization matrix entries, the techniques include calculating modified QP values for transform coefficients based on associated quantization matrix entries used as offsets to a base QP value. At a video decoder, the techniques include calculating dequantized transform coefficients from quantized transform coefficients based on the modified QP values. At a video encoder, the techniques include calculating quantized transform coefficients from transform coefficients based on the modified QP values. | 10-17-2013 |
20130272402 | INTER-LAYER MODE DERIVATION FOR PREDICTION IN SCALABLE VIDEO CODING - In some embodiments of a video coder, if some prediction information is not available for a first block in a current layer, the video coder uses corresponding information (e.g., intra prediction direction and motion information), if available, from the first block's co-located second block in the base layer as if it were the prediction information for the first block. The corresponding information can then be used in the current layer to determine the prediction information of succeeding blocks in the current layer. | 10-17-2013 |
20130272408 | MOTION VECTOR PREDICTION IN VIDEO CODING - Aspects of this disclosure relate to, in an example, a method that includes identifying a first block of video data in a first temporal location from a first view, wherein the first block is associated with a first disparity motion vector. The method also includes determining a motion vector predictor for a second motion vector associated with a second block of video data, wherein the motion vector predictor is based on the first disparity motion vector. When the second motion vector comprises a disparity motion vector, the method includes determining the motion vector predictor comprises scaling the first disparity motion vector to generate a scaled motion vector predictor, wherein scaling the first disparity motion vector comprises applying a scaling factor comprising a view distance of the second disparity motion vector divided by a view distance of the first motion vector to the first disparity motion vector. | 10-17-2013 |
20130272409 | BANDWIDTH REDUCTION IN VIDEO CODING THROUGH APPLYING THE SAME REFERENCE INDEX - Techniques for encoding and decoding video data are described. A method of coding video may include determining a plurality of motion vector candidates for a block of video data for use in a motion vector prediction process, wherein each of the motion vector candidates points to a respective reference frame index, performing the motion vector prediction process using the motion vector candidates to determine a motion vector for the block of video data, and performing motion compensation for the block of video data using the motion vector and a common reference frame index, wherein the common reference frame index is used regardless of the respective reference frame index associated with the determined motion vector. | 10-17-2013 |
20130272410 | MOTION VECTOR ROUNDING - A video decoder determines, based at least in part on a size of a prediction unit (PU), whether to round either or both a horizontal or a vertical component of a motion vector of the PU from sub-pixel accuracy to integer-pixel accuracy. The video decoder generates, based at least in part on the motion vector, a predictive sample block for the PU and generates, based in part on the predictive sample block for the PU, a reconstructed sample block. | 10-17-2013 |
20130272411 | SCALABLE VIDEO CODING PREDICTION WITH NON-CAUSAL INFORMATION - This disclosure pertains to video coding. Prediction information for a current block in an enhancement layer may be determined based at least in part on base layer information obtained by coding a base block in a base layer beneath the enhancement layer. This base block may occur in a position in the base layer such that it is co-located with a non-causal block in the enhancement layer (e.g., a block that occurs after the current block in the coding order of the enhancement layer). The prediction information determined for the current block may be used to code the current block (e.g., encoding or decoding the current block). | 10-17-2013 |
20130272412 | COMMON MOTION INFORMATION CANDIDATE LIST CONSTRUCTION PROCESS - In one example, an apparatus for coding video data comprises a video coder configured to generate first and second lists of motion information candidates, respectively, for first and second video block using a common list construction process, wherein the common list construction process is common to at least a first motion information prediction mode and a second motion information prediction mode. The video coder is further configured to code the first video block using the first motion information prediction mode based on a first motion information candidate selected from the first list, and code the second video block using the second motion information prediction mode based on a second motion information candidate selected from the second list. | 10-17-2013 |
20130272413 | COMMON SPATIAL CANDIDATE BLOCKS FOR PARALLEL MOTION ESTIMATION - In one example, an apparatus for coding video data comprises a video coder configured to, for a parallel motion estimation (PME) region comprising a plurality of blocks of video data within the PME region, identify a common set of spatial candidate blocks outside of and adjacent to the PME region, each of the common set of spatial candidate blocks at a respective, predefined location relative to the PME region and, for each of the blocks within the PME region for which motion information prediction is performed, generate a respective motion information candidate list, wherein, for at least some of the blocks within the PME region for which motion information prediction is performed, generating the motion information candidate list comprises evaluating motion information of at least one of the common set of spatial candidate blocks for inclusion in the motion information candidate list for the block. | 10-17-2013 |
20130272414 | GROUP FLAG IN TRANSFORM COEFFICIENT CODING FOR VIDEO CODING - A method of decoding data indicative of a subset of transform coefficients is described. The coefficients are indicative of a block of video data. The method may include determining that no transform coefficient in the subset of transform coefficients has an absolute value greater than one, and, based on the determining, skipping one or more decoding passes on the subset of transform coefficients, the decoding passes relating to decoding level information associated with the subset of transform coefficients. | 10-17-2013 |
20130272423 | TRANSFORM COEFFICIENT CODING - Techniques are described for determining a scan order for transform coefficients of a block. The techniques may determine context for encoding or decoding significance syntax elements for the transform coefficients based on the determined scan order. A video encoder may encode the significance syntax elements and a video decoder may decode the significance syntax elements based on the determined contexts. | 10-17-2013 |
20130272424 | SIGN HIDING TECHNIQUES FOR QUANTIZED TRANSFORM COEFFICIENTS IN VIDEO CODING - This disclosure proposes techniques for encoding and decoding transform coefficients in a video coding process. In particular, this disclosure proposes techniques determining whether or not to apply a sign data hiding process for a group of transform coefficients, and techniques for applying the sign data hiding process. In one example, this disclosure describes a method for decoding video data comprising determining a block of transform coefficients, determining whether to perform a sign data hiding process for at least one transform coefficient in the block of transform coefficients based on a single variable compared to a threshold, and decoding sign information for the block based on the determination of whether to perform the sign data hiding process. | 10-17-2013 |
20130272425 | BETA OFFSET CONTROL FOR DEBLOCKING FILTERS IN VIDEO CODING - Techniques are described for providing continuous control of a deblocking filter for a video block using a beta offset parameter. Deblocking filters are defined based on one or more deblocking decisions. Conventionally, a quantization parameter and a beta offset parameter are used to identify a beta parameter (“β”) value that determines threshold values of the deblocking decisions. The value of the beta offset parameter results in a change or increment of the β value. For small increments of the β value, rounding of the threshold values may result in no change and discontinuous control of the deblocking decisions. The techniques include calculating at least one deblocking decision for the deblocking filter according to a threshold value that has been modified based on a multiplier value of the beta offset parameter. The multiplier value applied to the beta offset parameter causes an integer change in the modified threshold value. | 10-17-2013 |
20130287108 | DISPARITY VECTOR GENERATION FOR INTER-VIEW PREDICTION FOR VIDEO CODING - In one example, a video coder (e.g., a video encoder or a video decoder) is configured to determine that a current block of video data is coded using a disparity motion vector, wherein the current block is within a containing block, based on a determination that a neighboring block to the current block is also within the containing block, substitute a block outside the containing block and that neighbors the containing block for the neighboring block in a candidate list, select a disparity motion vector predictor from one of a plurality of blocks in the candidate list, and code the disparity motion vector based on the disparity motion vector predictor. In this manner, the techniques of this disclosure may allow blocks within the containing block to be coded in parallel. | 10-31-2013 |
20130287109 | INTER-LAYER PREDICTION THROUGH TEXTURE SEGMENTATION FOR VIDEO CODING - An apparatus for coding video data according to certain aspects includes a memory and a processor in communication with the memory. The memory stores the video data. The video data may include a base layer and an enhancement layer, the base layer including a base layer block and the enhancement layer including an enhancement layer block. The base layer block may be located at a position in the base layer corresponding to a position of the enhancement layer block in the enhancement layer. The processor determines, based on information associated with the base layer block, a partitioning mode of the enhancement layer block. The partitioning mode may indicate that the enhancement layer block is to be partitioned into a first partition and a second partition. The processor further performs motion compensation for the first partition and the second partition of the enhancement layer block. | 10-31-2013 |
20130294524 | TRANSFORM SKIPPING AND LOSSLESS CODING UNIFICATION - An example method includes determining whether an encoded block of residual video data was encoded losslessly in accordance with a lossless coding mode, based on whether transform operations were skipped during encoding of the block of residual video data, and if the block of residual video data was encoded losslessly, then decoding the encoded block of residual video data according to the lossless coding mode to form a reconstructed block of residual video data, where decoding the encoded block of residual data comprises bypassing quantization and sign hiding while decoding the encoded block of residual video data, and bypassing all loop filters with respect to the reconstructed block of residual video data. | 11-07-2013 |
20130308701 | VIDEO CODING WITH LARGE MACROBLOCKS - Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels. In one example, an apparatus includes a video encoder configured to encode a coded unit comprising a plurality of video blocks, wherein at least one of the plurality of video blocks comprises a size of more than 16×16 pixels and to generate syntax information for the coded unit that includes a maximum size value, wherein the maximum size value indicates a size of a largest one of the plurality of video blocks in the coded unit. The syntax information may also include a minimum size value. In this manner, the encoder may indicate to a decoder the proper syntax decoder to apply to the coded unit. | 11-21-2013 |
20130329007 | REDUNDANCY REMOVAL FOR ADVANCED MOTION VECTOR PREDICTION (AMVP) IN THREE-DIMENSIONAL (3D) VIDEO CODING - In general, techniques are described for performing motion vector prediction in 3D video coding and, more particularly for managing a candidate list of motion vector predictors (MVPs) for a block of video data. In some examples, a video coder, such as video encoder or video decoder, includes at least three motion vector predictors (MVPs) in a candidate list of MVPs for a current block in a first view of a current access unit of the video data, wherein the at least three MVPs comprise an inter-view motion vector predictor (IVMP), which is a temporal motion vector derived from a block in a second view of the current access unit or a disparity motion vector derived from a disparity vector. | 12-12-2013 |
20130329782 | ADAPTIVE UPSAMPLING FILTERS - In one embodiment, a video coder for coding video data includes a processor and a memory. The processor selects a filter set from a multiple filter sets for upsampling reference layer video data based at least on a prediction operation mode for enhanced layer video data and upsamples the reference layer video data using the selected filter set. Some of the multiple filter sets have some different filter characteristics from one another, and the upsampled reference layer video data has the same spatial resolution as the enhanced layer video data. The processor further codes the enhanced layer video data based at least on the upsampled reference layer video data and the prediction operation mode. The memory stores the upsampled reference layer video data. | 12-12-2013 |
20130329806 | BI-LAYER TEXTURE PREDICTION FOR VIDEO CODING - In one example, an apparatus is configured to code video data. The apparatus comprises a processor configured to determine a base layer reference block for a current block. The base layer reference block may be located in the base layer. The processor is further configured to determine an enhancement layer reference block for the current block. The enhancement layer reference block may comprise a weighted sum of a first reference block located in the enhancement layer and a second reference block located in the enhancement layer. The processor is further configured to determine a reference block from the base layer reference block and the enhancement layer reference block. | 12-12-2013 |
20130335522 | DERIVATION OF DEPTH MAP ESTIMATE - In some example techniques for generating and updating depth map estimates used for inter-view motion prediction and/or inter-view residual prediction for coding multiview video data, each of a plurality of dependent views is associated with a dependent depth view estimate, which may be generated or updated based on coding of the texture data of the dependent view relative to a base view. In such examples, each of the dependent depth map estimates may be warped to the base view to produce a respective one of a plurality of base depth map estimates. Each dependent depth map estimate and the respective base depth map estimate for a depth map estimate pair associated with the respective one of the plurality of dependent views. | 12-19-2013 |
20130336382 | GROUPING OF BYPASS-CODED BINS FOR SAO SYNTAX ELEMENTS - A video encoder generates a sequence of sample adaptive offset (SAO) syntax elements for a coding tree block. The SAO syntax elements include regular context-adaptive binary arithmetic coding (CABAC) coded bins for a color component and bypass-coded bins for the color component. None of the bypass-coded bins is between two of the regular CABAC-coded bins in the sequence. The video encoder uses regular CABAC to encode the regular CABAC-coded bins and uses bypass coding to encode the bypass-coded bins. The video encoder outputs the SAO syntax elements in a bitstream. A video decoder receives the bitstream, uses regular CABAC to decode the regular CABAC-coded bins, uses bypass coding to decode the bypass-coded bins, and modifies a reconstructed picture based on the SAO syntax elements. | 12-19-2013 |
20130336386 | SAMPLE ADAPTIVE OFFSET (SAO) CODING - A video coder according to the techniques of the present disclosure may code a prefix value and code a suffix value, such that the combination of the suffix value and the prefix value identify an offset value determined for a sample adaptive offset filtering (SAO) operation. | 12-19-2013 |
20130336394 | INFERRED BASE LAYER BLOCK FOR TEXTURE_BL MODE IN HEVC BASED SINGLE LOOP SCALABLE VIDEO CODING - An apparatus for coding video data using a single-loop decoding approach may include a memory unit and a processor in communication with the memory unit. In an embodiment, the memory unit stores the video data, the video data including a base layer and an enhancement layer. The base layer includes a base layer block, a non-constrained INTRA mode block, and an INTER mode block. The base layer block includes a sub-block located at least partially within one of the non-constrained INTRA mode block or the INTER mode block. The enhancement layer includes an enhancement layer block located at a position in the enhancement layer corresponding to a position of the base layer block in the base layer. The processor approximates pixel values of the sub-block and determines, based at least in part on the approximated pixel values, pixel values of the enhancement layer block. | 12-19-2013 |
20130336395 | UNIFICATION OF SIGNALING LOSSLESS CODING MODE AND PULSE CODE MODULATION (PCM) MODE IN VIDEO CODING - Techniques described herein are related to harmonizing the signaling of coding modes and filtering in video coding. In one example, a method of decoding video data is provided that includes decoding a first syntax element to determine whether PCM coding mode is used for one or more video blocks, wherein the PCM coding mode refers to a mode that codes pixel values as PCM samples. The method further includes decoding a second syntax element to determine whether in-loop filtering is applied to the one or more video blocks. Responsive to the first syntax element indicating that the PCM coding mode is used, the method further includes applying in-loop filtering to the one or more video blocks based at least in part on the second syntax element and decoding the one or more video blocks based at least in part on the first and second syntax elements. | 12-19-2013 |
20130336406 | REDUNDANCY REMOVAL FOR MERGE/SKIP MODE MOTION INFORMATION CANDIDATE LIST CONSTRUCTION - In general, techniques are described for constructing a merging candidate list for coding video data according to a merge mode and/or a skip mode. In some examples, the techniques include identifying one or more spatial merging candidates (SMCs) and an inter-view merging candidate (IVMC) for inclusion in a merging candidate list, and comparing the motion information of at least one of the SMCs to the motion information of the IVMC. In such examples, if the SMC has the same motion information as the IVMC, the techniques may further include pruning the merging candidate list to exclude the one of the merging candidates from the merging candidate list. | 12-19-2013 |
20130343464 | TRANSFORM SKIP MODE - A device for coding video data includes a video coder configured to: determine for a chroma transform block (TB) a sub-sampling format for the chroma TB; based on the sub-sampling format for the chroma TB, identify one or more corresponding luma TBs; determine, for each of the one or more corresponding luma TBs, if the corresponding luma TB is coded using a transform skip mode; and, based on a number of the one or more corresponding luma TBs coded using the transform skip mode being greater than or equal to a threshold value, determine that the chroma TB is coded in the transform skip mode. | 12-26-2013 |
20140003529 | CODING OF SIGNIFICANCE FLAGS | 01-02-2014 |
20140003530 | SIGN HIDING TECHNIQUES FOR QUANTIZED TRANSFORM COEFFICIENTS IN VIDEO CODING | 01-02-2014 |
20140003532 | WAVEFRONT PARALLEL PROCESSING FOR VIDEO CODING | 01-02-2014 |
20140010280 | INTRA MODE EXTENSIONS FOR DIFFERENCE DOMAIN INTRA PREDICTION - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores difference video information associated with a difference video layer of pixel information derived from a difference between an enhancement layer and a corresponding base layer of the video information. The processor determines a DC prediction value for a video unit associated with the difference video layer while refraining from using pixel information from a neighboring area of the video unit, wherein the DC prediction value is equal to zero or is offset by an offset value. The DC prediction value is a prediction value used in intra prediction based at least on an average of neighboring video units of the video unit. The processor further determines a value of the video unit based at least in part on the DC prediction value. | 01-09-2014 |
20140010292 | SKIP TRANSFORM AND RESIDUAL CODING MODE EXTENSION FOR DIFFERENCE DOMAIN INTRA PREDICTION - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores difference video information associated with a difference video layer of pixel information derived from a difference between an enhancement layer and a corresponding base layer of the video information. The processor determines a value of a video unit based on the difference video layer. The processor is further configured to refrain from performing a transform and residual coding of the determined value of the video unit. | 01-09-2014 |
20140010300 | SMOOTHING OF DIFFERENCE REFERENCE PICTURE - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores difference video information associated with a difference video layer of pixel information derived from a difference between an enhancement layer and a corresponding base layer of the video information. The processor determines a value of a video unit based on a reference video unit or spatial neighboring video unit within the difference video layer and applies a smoothing filter to the reference video unit or spatial neighboring video unit. | 01-09-2014 |
20140010310 | MOTION VECTOR DIFFERENCE CODING EXTENSION FOR ENHANCEMENT LAYER - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores difference video information associated with a difference video layer of pixel information derived from a difference between an enhancement layer and a corresponding base layer of the video information. The processor determines pixel accuracy of motion predictor information, determines a motion vector based on the pixel accuracy of the motion predictor information, and determines a value of a video unit based at least in part on the motion vector. | 01-09-2014 |
20140015925 | GENERALIZED RESIDUAL PREDICTION FOR SCALABLE VIDEO CODING AND 3D VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information of a reference layer. The processor determines a value of a video unit based at least in part on a prediction value and an adjusted residual prediction value associated with the reference layer. The adjusted residual prediction value is equal to a residual prediction from the reference layer multiplied by a weighting factor that is different from 1. | 01-16-2014 |
20140016698 | ROTATION OF PREDICTION RESIDUAL BLOCKS IN VIDEO CODING WITH TRANSFORM SKIPPING - Techniques are described for coding residual data of a prediction residual block with transform skipping. A transform may be skipped for a residual block when the residual block is coded using either a lossless coding mode or a lossy coding mode in a transform skip mode. According to the techniques, based on a transform being skipped for a residual block, a rotation unit included in a video encoder or a video decoder determines whether to rotate the residual block prior to coding residual data of the residual block. In some examples, a rotation value may be explicitly signaled between the video encoder and the video decoder. In other examples, the video encoder and the video decoder may each independently determine whether to rotate the residual block based on a type of boundary at two or more edges of the residual block. | 01-16-2014 |
20140016701 | TEMPORAL MOTION VECTOR PREDICTION IN VIDEO CODING EXTENSIONS - A first reference index value indicates a position, within a reference picture list associated with a current prediction unit (PU) of a current picture, of a first reference picture. A reference index of a co-located PU of a co-located picture indicates a position, within a reference picture list associated with the co-located PU of the co-located picture, of a second reference picture. When the first reference picture and the second reference picture belong to different reference picture types, a video coder sets a reference index of a temporal merging candidate to a second reference index value. The second reference index value is different than the first reference index value. | 01-16-2014 |
20140044161 | ADATIVE UP-SAMPLING FILTER FOR SCALABLE VIDEO CODING - Methods and systems for video image coding are provided. Sets of filters may be selected and applied to video information at least partially based on the type of inter layer prediction implemented in coding the video information. Different filters, or filter sets, may be used for inter layer intra prediction, difference domain intra prediction, and/or difference domain inter prediction. Filter selection information may be embedded in the video bit stream. | 02-13-2014 |
20140044162 | ADAPTIVE INFERENCE MODE INFORMATION DERIVATION IN SCALABLE VIDEO CODING - Systems and methods for determining information about an enhancement layer of digital video based on information included in a base layer of digital video are described. In one innovative aspect, an apparatus for coding digital video is provided. The apparatus includes a memory for storing a base layer of digital video information and an enhancement layer of digital video information. The apparatus determines a syntax element value for a portion of the enhancement layer based on a syntax element value for a corresponding portion of the base layer. Decoding devices and methods as well as corresponding encoding devices and methods are described. | 02-13-2014 |
20140044178 | WEIGHTED DIFFERENCE PREDICTION UNDER THE FRAMEWORK OF GENERALIZED RESIDUAL PREDICTION - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information associated with a reference layer. The processor determines a value of a current video unit based on, at least in part, a reconstruction value associated with the reference layer and an adjusted difference prediction value. The adjusted difference prediction value is equal to a difference between a prediction of a current layer and a prediction of the reference layer multiplied by a weighting factor that is different from 1. | 02-13-2014 |
20140044179 | MULTI-HYPOTHESIS MOTION COMPENSATION FOR SCALABLE VIDEO CODING AND 3D VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information of a reference layer and an enhancement layer. The processor determines a value of a current video unit of the enhancement layer based on, at least in part, explicit hypotheses and implicit hypotheses calculated from movement information from the reference layer. | 02-13-2014 |
20140049605 | INTER-VIEW PREDICTED MOTION VECTOR FOR 3D VIDEO - A video coder determines a first picture order count (POC) value of a first reference picture associated with a first motion vector of a corresponding block that points in a first direction and determines whether a first reference picture list for the current block includes a reference picture having the first POC value; in response to the reference picture list not including the reference picture having the first POC value, determines a second POC value of a second reference picture associated with a second motion vector of the corresponding block that points in a second direction, determines whether the first reference picture list for the current block includes a reference picture having the second POC value and in response to the first reference picture list including the reference picture having the second POC value, decodes the current motion vector using the second motion vector of the corresponding block. | 02-20-2014 |
20140050261 | ENHANCEMENT LAYER SCAN ORDER DERIVATION FOR SCALABLE VIDEO CODING - An apparatus for coding video information includes a memory unit configured to store video information associated with a reference block; and a processor in communication with the memory unit, wherein the processor is configured to determine a value of a current video unit associated with the reference block based on, at least in part, a classification of the reference block and a scan order selected by the processor based upon the classification. The scan order indicates an order in which values within the reference block are processed to at least partially determine the value of the current video unit. | 02-20-2014 |
20140050265 | VIDEO CODING BASED ON FIRST ORDER PREDICTION AND PRE-DEFINED SECOND ORDER PREDICTION MODE - This disclosure describes video encoding and decoding techniques in which a first order prediction process and a second order prediction process are used in combination to generate predictive video blocks for video coding. First order prediction may be similar to conventional motion estimation and motion compensation that generates residual video blocks. The second order prediction may involve a process similar to conventional intra-prediction, but is performed on the residual video blocks. The techniques of this disclosure may pre-define the second order prediction to a specific mode, such as a mode similar to the intra-DC mode used in intra coding. In addition, the techniques of this disclosure may combine aspects of the first order and second order prediction into a single process so that the effects of second order prediction on the residuals are taken into account during the first order prediction process, which may improve compression. | 02-20-2014 |
20140056361 | ALTERNATIVE TRANSFORM IN SCALABLE VIDEO CODING - An apparatus for coding video data according to certain aspects includes a memory and a processor in communication with the memory. The memory stores video block information. The video block information includes reference layer block information. The processor determines, based on a parameter of the video block information, a transform function that may be used to code the video block information. The processor may encode or decode the video block information. The transform function may be an alternative transform when the parameter is a predetermined value and a primary transform when the parameter is not the predetermined value. The alternative transform includes one of: a discrete-sine-transform (DST), a Type-I DST, a Type-III DST, a Type-IV DST, a Type-VII DST, a discrete-cosine-transform (DCT), a DCT of different types, and a Karhunen-Loeve transform (KLT). | 02-27-2014 |
20140064359 | INTRA PREDICTION MOST PROBABLE MODE ORDER IMPROVEMENT FOR SCALABLE VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information associated with a reference layer and a corresponding enhancement layer. The processor receives a mode list associated with the enhancement layer, the mode list comprising three entities, each entity identifying a different mode for determining a value of a video unit located at a position within the enhancement layer. The processor changes the mode list when a mode associated with a co-located video unit in the reference layer is not stored as the first entity in the mode list. The co-located video unit is located at a position in the reference layer corresponding to the position of the video unit in the enhancement layer. | 03-06-2014 |
20140064360 | INTRA PREDICTION IMPROVEMENTS FOR SCALABLE VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information associated with a reference layer and a corresponding enhancement layer. The processor determines a value of a video unit positioned at a position within the enhancement layer based at least in part on an intra prediction value weighted by a first weighting factor, wherein the intra prediction value is determined based on at least one additional video unit in the enhancement layer, and a value of a co-located video unit in the reference layer weighted by a second weighting factor, wherein the co-located video unit is located at a position in the reference layer corresponding to the position of the video unit in the enhancement layer. In some embodiments, the at least one of the first and second weighting factors is between 0 and 1. | 03-06-2014 |
20140064361 | TRANSFORM BASIS ADJUSTMENT IN SCALABLE VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information associated with a reference layer and a corresponding enhancement layer. The processor obtains residue block information based at least in part on video information associated with the reference layer and the enhancement layer. The processor determines an adjustment transform function based on a transform function associated with the video information. The processor determines a transform block based on the adjusted transform function and the residue block information. | 03-06-2014 |
20140064386 | SIGNALING OF DOWN-SAMPLING LOCATION INFORMATION IN SCALABLE VIDEO CODING - Methods and systems for video image coding are provided. Sets of filters may be selected and applied to video information at least partially based on phase displacement information between a first and second layer of video information. For example, the phase displacement information may correspond to a difference between a position of a pixel in the first layer and a corresponding position of the pixel in the second layer. The selected filter set can be an up-sampling filter or a down-sampling filter. The phase displacement information may be encoded as a syntax element embedded in the video bit stream. | 03-06-2014 |
20140071235 | INTER-VIEW MOTION PREDICTION FOR 3D VIDEO - This disclosure describes techniques for improving coding efficiency of motion prediction in multiview and 3D video coding. In one example, a method of decoding video data comprises deriving one or more disparity vectors for a current block, the disparity vectors being derived from neighboring blocks relative to the current block, converting a disparity vector to one or more of inter-view predicted motion vector candidates and inter-view disparity motion vector candidates, adding the one or more inter-view predicted motion vector candidates and the one or more inter-view disparity motion vector candidates to a candidate list for a motion vector prediction mode, and decoding the current block using the candidate list. | 03-13-2014 |
20140078250 | ADVANCED INTER-VIEW RESIDUAL PREDICTION IN MULTIVIEW OR 3-DIMENSIONAL VIDEO CODING - Systems and methods for coding video information for a current view based on a residual prediction from video information for a reference view are described. In one innovative aspect, an apparatus for coding digital video is provided. The apparatus includes a memory configured to store current view video information and reference view video information. The apparatus also includes a processor configured to determine a value of a current video unit of the current view based at least on a motion compensated block of the reference view. The motion compensated block may be determined based at least in part on motion information and a disparity vector associated with the current video unit. Decoding devices and methods as well as corresponding encoding devices and methods are described. | 03-20-2014 |
20140078251 | SELECTION OF PICTURES FOR DISPARITY VECTOR DERIVATION - When coding multiview video data, a video encoder and video decoder may select a candidate picture from one of one or more random access point view component (RAPVC) pictures and one or more pictures having a lowest temporal identification value. The video encoder and video decoder may determine whether a block in the selected candidate picture is inter-predicted with a disparity motion vector and determine a disparity vector for a current block of a current picture based on the disparity motion vector. The video encoder and video decoder may inter-prediction encode or decode, respectively, the current block based on the determined disparity vector. | 03-20-2014 |
20140079135 | PERFORMING QUANTIZATION TO FACILITATE DEBLOCKING FILTERING - A method of encoding video data includes encoding a quantization parameter delta value in a coding unit (CU) of the video data before coding a version of a block of the CU in a bitstream so as to facilitate deblocking filtering. Coding the quantization parameter delta value may comprise coding the quantization parameter delta value based on the value of a no_residual_syntax flag that indicates whether no blocks of the CU have residual transform coefficients. | 03-20-2014 |
20140086307 | CONTEXT DERIVATION FOR CONTEXT-ADAPTIVE, MULTI-LEVEL SIGNIFICANCE CODING - A device for coding video data includes a video coder configured to code first significance information for transform coefficients associated with residual data, wherein the first significance information indicates if a first sub-block comprises at least one non-zero coefficient, wherein the first sub-block is a sub-block of an entire transform block; and, code second significance information, wherein the second significance information indicates if a second sub-block comprises at least one non-zero coefficient, wherein the second sub-block is a sub-block of the first sub-block, wherein coding the second significance information comprises performing an arithmetic coding operation on the second significance information, wherein a context for the arithmetic coding operation is determined based on one or more neighboring sub-blocks of a same size as the first sub-block. | 03-27-2014 |
20140086325 | SCALABLE EXTENSIONS TO HEVC AND TEMPORAL MOTION VECTOR PREDICTION - In one example, a device includes a video coder configured to determine a first co-located reference picture for generating a first temporal motion vector predictor candidate for predicting a motion vector of a current block, determine a second co-located reference picture for generating a second temporal motion vector predictor candidate for predicting the motion vector of the current block, determine a motion vector predictor candidate list that includes at least one of the first temporal motion vector predictor candidate and the second temporal motion vector predictor candidate, select a motion vector predictor from the motion vector predictor candidate list, and code the motion vector of the current block relative to the selected motion vector predictor. | 03-27-2014 |
20140086328 | SCALABLE VIDEO CODING IN HEVC - In one example, a device includes a video coder configured to determine a temporal motion vector predictor for a motion vector associated with a block of video data of a current picture of a first, non-base layer of a plurality of layers of video data using a temporal motion vector prediction process, wherein the temporal motion vector prediction process includes identifying a co-located picture from which to derive the temporal motion vector predictor, and restrict the temporal motion vector prediction process such that the co-located picture used to derive the temporal motion vector predictor is not located in a layer other than the first layer of the plurality of layers of video data. | 03-27-2014 |
20140092965 | INTRA-CODING FOR 4:2:2 SAMPLE FORMAT IN VIDEO CODING - Systems, methods, and devices for video coding that may receive a rectangular chroma block are disclosed. The rectangular chroma block may be divided into a first square sub-block and a second square sub-block. The systems, methods, and devices may reconstruct the first square sub-block and intra-predict the second square sub-block using reference samples from the reconstructed first square sub-block. | 04-03-2014 |
20140092967 | USING BASE LAYER MOTION INFORMATION - Systems, methods, and devices for coding video data are described herein. In some aspects, a memory is configured to store the video data associated with a base layer and an enhancement layer. The base layer may comprise a reference block and base layer motion information associated with the reference block. The enhancement layer may comprise a current block. A processor operationally coupled to the memory is configured to determine a position of the base layer motion information in a candidate list based on a prediction mode in a plurality of prediction modes used at the enhancement layer. The processor is further configured to perform a prediction of the current block based at least in part on the candidate list. | 04-03-2014 |
20140092983 | CODED BLOCK FLAG CODING FOR 4:2:2 SAMPLE FORMAT IN VIDEO CODING - Systems, methods, and devices for video coding that may obtain a rectangular chroma block having first and second square sub-blocks are disclosed. These systems, methods, and devices may also decode a first coded block flag (CBF) for the first square sub-block to indicate whether the first square sub-block includes at least one nonzero transform coefficient. These systems, methods, and devices may also decode a second CBF for the second square sub-block to indicate whether the second square sub-block includes at least one nonzero transform coefficient and not decoding a CBF for the rectangular chroma block. | 04-03-2014 |
20140098880 | PREDICTION MODE INFORMATION UPSAMPLING FOR SCALABLE VIDEO CODING - In one embodiment, an apparatus configured to code video data includes a processor and a memory unit. The memory unit stores video data associated with a first layer having a first spatial resolution and a second layer having a second spatial resolution. The video data associated with the first layer includes at least a first layer block and first layer prediction mode information associated with the first layer block, and the first layer block includes a plurality of sub-blocks where each sub-block is associated with respective prediction mode data of the first layer prediction mode information. The processor derives the predication mode data associated with one of the plurality of sub-blocks based at least on a selection rule, upsamples the derived prediction mode data and the first layer block, and associates the upsampled prediction mode data with each upsampled sub-block of the upsampled first layer block. | 04-10-2014 |
20140105299 | PERFORMING RESIDUAL PREDICTION IN VIDEO CODING - In general, techniques are described for performing residual prediction in video coding. As one example, a device configured to code scalable or multi-view video data may comprise one or more processors configured to perform the techniques. The processors may determine a difference picture, for a current picture, based on a first reference picture in a same layer or view as the current picture and a decoded picture in a different layer or view as the current picture. The decoded picture may be in a same access unit as the first reference picture. The processors may perform bi-prediction based on the difference picture to code at least a portion of the current picture. | 04-17-2014 |
20140112387 | SEPARABLE DIRECTIONAL TRANSFORMS - This disclosure describes techniques for transforming residual blocks of video data. In particular, a plurality of different transforms selectively applied to the residual blocks based on the prediction mode of the video blocks. At least a portion of the plurality of transforms are separable directional transform specifically trained for a corresponding prediction mode to provide better energy compaction for the residual blocks of the given prediction mode. Using separable directional transforms offers the benefits of lower computation complexity and storage requirement than use of non-separable directional transforms. Additionally, a scan order used to scan the coefficients of the residual block may be adjusted when applying separable directional transforms. In particular, the scan order may be adjusted based on statistics associated with one or more previously coded blocks to better ensure that non-zero coefficients are grouped near the front of the one-dimensional coefficient vector to improve the effectiveness of entropy coding. | 04-24-2014 |
20140140393 | METHOD AND SYSTEM FOR INTRA BASE LAYER (BL) TRANSFORM IN VIDEO CODING - An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer and an enhancement layer. The processor is configured to, in response to determining that the video information associated with the enhancement layer is to be determined based upon the video information associated with the base layer, select between a first transform and a second transform based at least in part on at least one of a transform unit (TU) size and a color component type of the enhancement layer video information. | 05-22-2014 |
20140140399 | LOW-COMPLEXITY SUPPORT OF MULTIPLE LAYERS FOR HEVC EXTENSIONS IN VIDEO CODING - A method of coding video data includes receiving video information associated with a first layer and a second layer and determining whether at least one of the first layer and the second layer is a default layer. The method can include at least partially restricting inter-layer prediction when neither the first layer nor the second layer is the default layer. A default layer can be a base layer or an enhancement layer. A flag can be received that indicates that inter-layer prediction is to be restricted. In addition, the method can include determining whether inter-layer prediction is allowed for the video information associated with the first layer, and determining whether inter-layer prediction is partially allowed for the video information associated with the second layer such that motion compensation is not used with the second layer video information. | 05-22-2014 |
20140140405 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer and an enhancement layer. The video information comprises at least one enhancement layer (EL) block and at least one co-located base layer (BL) block. The co-located BL block has motion information associated therewith. The processor is configured to, in response to determining that the size of the EL block is smaller than a threshold size, either (1) use less than all of the motion information associated with the co-located BL block to code the EL block, or (2) refrain from using any motion information associated with the co-located BL block to code the EL block. The processor may encode or decode the video information. | 05-22-2014 |
20140146875 | LOOP FILTERING ACROSS CONSTRAINED INTRA BLOCK BOUNDARIES IN VIDEO CODING - This disclosure describes techniques for coding video data. In particular, this disclosure describes techniques for loop filtering for video coding. The techniques of this disclosure may apply to loop filtering and/or partial loop filtering across block boundaries in scalable video coding processes. Loop filtering may include, for example, one or more of adaptive loop filtering (ALF), sample adaptive offset (SAO) filtering, and deblocking filtering. | 05-29-2014 |
20140161175 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining a partition mode for coding a block of video data, where the partition mode indicates a division of the block of video data for predictive coding. The method also includes determining whether to code a weighting factor for an inter-view residual prediction process based on the partition mode, where, when the weighting factor is not coded, the inter-view residual prediction process is not applied to predict a residual for the block. The method also includes coding the block of video data with the determined partition mode. | 06-12-2014 |
20140161186 | ADVANCED MERGE/SKIP MODE AND ADVANCED MOTION VECTOR PREDICTION (AMVP) MODE FOR 3D VIDEO - Techniques are described where if an inter-view predicted motion vector candidate (IPMVC) and an inter-view disparity motion vector candidate (IDMVC) are derived based on a shifted disparity vector, where the amount by which the disparity vector is shifted for the IPMVC and IDMVC is different. The techniques also prioritize the inclusion of the IPMVC over the IDMVC in a candidate list, and prune the IPMVC and the IDMVC if there is a duplicated IPMVC or IDMVC in the candidate list. | 06-12-2014 |
20140161187 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding multi-layer video data includes determining, for a first block of video data at a first temporal location, whether one or more reference picture lists for coding the first block contain at least one reference picture at a second, different temporal location. The method also includes coding the first block of video data relative to at least one reference block of video data of a reference picture in the one or more reference picture lists, where coding includes disabling an inter-view residual prediction process when the one or more reference picture lists do not include at least one reference picture at the second temporal location. | 06-12-2014 |
20140161188 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining a location of a temporal reference block indicated by a temporal motion vector to a current block of video data, where the current block and the temporal reference block are located in a first layer of video data. The method also includes interpolating, with a first type of interpolation, a location of a disparity reference block indicated by a disparity vector of the current block, where the disparity reference block is located in a second, different layer, and where the first type of interpolation comprises a bi-linear filter. The method also includes determining a temporal-disparity reference block of the disparity reference block indicated by a combination of the temporal motion vector and the disparity vector, and coding the current block based on the temporal reference block, the disparity reference block, and the temporal-disparity reference block. | 06-12-2014 |
20140161189 | ADVANCED RESIDUAL PREDICTION IN SCALABLE AND MULTI-VIEW VIDEO CODING - In an example, a method of coding video data includes determining, for a first block of video data in a first layer of video data, a temporal motion vector and associated temporal reference picture for predicting the first block, where the temporal reference picture has a picture order count value. The method also includes determining a disparity reference block in a disparity reference picture indicated by a disparity vector associated with the first block, and determining whether a decoded picture buffer contains a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture. When the decoded picture buffer does not contain a temporal-disparity reference picture in the second view and having the picture order count value of the temporal reference picture, the method includes modifying an inter-view residual prediction process for predicting residual data of the first block. | 06-12-2014 |
20140168363 | DISPARITY VECTOR DERIVATION - A parent block is partitioned into the plurality of blocks and a disparity vector derivation process is performed to derive a disparity vector for a representative block in the plurality of blocks. A video encoder generates a bitstream that includes a coded representation of the video data in part by performing, based on the derived disparity vector and without separately deriving disparity vectors for any block in the plurality of blocks other than the representative block, inter-view prediction for two or more blocks in the plurality of blocks. A video decoder reconstructs sample blocks for two or more blocks in the plurality of blocks in part by performing, based on the derived disparity vector and without separately deriving disparity vectors for any block in the plurality of blocks other than the representative block, inter-view prediction for the two or more blocks in the plurality of blocks. | 06-19-2014 |
20140169474 | DISPARITY VECTOR DERIVATION - A device performs a disparity vector derivation process to determine a disparity vector for a current block. As part of performing the disparity vector derivation process, when either a first or a second spatial neighboring block has a disparity motion vector or an implicit disparity vector, the device converts the disparity motion vector or the implicit disparity vector to the disparity vector for the current block. The number of neighboring blocks that is checked in the disparity vector derivation process is reduced, potentially resulting in decreased complexity and memory bandwidth requirements. | 06-19-2014 |
20140169475 | MOTION VECTOR PREDICTION IN VIDEO CODING - During a process to derive an inter-view predicted motion vector candidate (IPMVC) for an Advanced Motion Vector Prediction (AMVP) candidate list, a video coder determines, based on a disparity vector of a current prediction unit (PU), a reference PU for the current PU. Furthermore, when a first reference picture of the reference PU has the same picture order count (POC) value as a target reference picture of the current PU, the video coder determines an IPMVC based on a first motion vector of the reference PU. Otherwise, when a second reference picture of the reference PU has the same POC value as the target reference picture of the current PU, the video coder determines the IPMVC based on a second motion vector of the reference PU. | 06-19-2014 |
20140169483 | DEBLOCKING FILTER WITH REDUCED LINE BUFFER - An apparatus configured to filter video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video information comprising at least two adjacent video blocks, each video block comprising a plurality of video samples, and each video sample having a bit depth. The processor determines a filtered video sample based at least in part on a video sample and an adjustment value. The processor determines the adjustment value at least in part from an input with a limited bit depth. The input is determined from a set of one or more video samples, and its bit depth is limited such that it is less than the bit depth of the one or more video samples. | 06-19-2014 |
20140176674 | DISPARITY VECTOR DERIVATION IN THREE-DIMENSIONAL VIDEO CODING - A video coder searches a set of neighbor blocks to generate a plurality of disparity vector candidates. Each of the neighbor blocks is a spatial or temporal neighbor of a current block. The video coder determines, based at least in part on the plurality of disparity vector candidates, a final disparity vector for the current block. | 06-26-2014 |
20140177704 | MULTI-TYPE PARALLELIZED SAMPLE ADAPTIVE OFFSET IN VIDEO CODING - In general, techniques are described for performing multiple passes of sample adaptive offset (SAO) filtering when coding video data. A video decoding device comprising one or more processors may perform the techniques. The processors may determine a first SAO pixel classification for a block of video data and determine a first offset value based on the first SAO pixel classification and one or more pixel values of the block. The one or more processors may also determine a second SAO pixel classification for the block and determine a second offset value based on the second SAO pixel classification and the one or more pixel values of block of video data. The processors may then apply the first offset value and the second offset value to the block of video data to generate a block of SAO filtered video data. | 06-26-2014 |
20140185664 | INTER-LAYER PREDICTION USING SAMPLE-ADAPTIVE ADJUSTMENTS FOR BIT DEPTH SCALABLE VIDEO CODING - An apparatus for coding video data according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores video data. The video data may include a base layer comprising samples with a lower bit depth and an enhancement layer comprising samples with a higher bit depth. The processor predicts the values of samples in the enhancement layer based on the values of samples in the base layer. The prediction performed by the processor includes applying a preliminary mapping to the base layer samples to obtain preliminary predictions, and then applying adaptive adjustments to the preliminary predictions to obtain refined predictions. Parameters used for the adaptive adjustments may depend on the values and distribution of base layer samples. The processor may encode or decode the video data. | 07-03-2014 |
20140185665 | HIGH-FREQUENCY-PASS SAMPLE ADAPTIVE OFFSET IN VIDEO CODING - This disclosure presents techniques for sample adaptive offset (SAO) filtering that focus on filtering the changed region of a picture or layer (e.g., the high frequency component of video blocks of a picture or layer). In general, the techniques of this disclosure may be referred to as High-Frequency-Pass SAO (HSAO) filtering. In one example, a method for coding video data comprises obtaining a decoded block of video data, separating the decoded block of video data into a low-frequency component and a high-frequency component, applying an SAO filter to the high-frequency component to generate a SAO filtered high-frequency component, and combining the SAO filtered high-frequency component and the low-frequency component to generate a filtered decoded block of video data. | 07-03-2014 |
20140185680 | DEVICE AND METHOD FOR SCALABLE AND MULTIVIEW/3D CODING OF VIDEO INFORMATION - An apparatus configured to code (e.g., encode or decode) video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer and an enhancement layer. The processor is configured to up-sample a base layer reference block by using an up-sampling filter when the base and enhancement layers have different resolutions; perform motion compensation interpolation by filtering the up-sampled base layer reference block; determine base layer residual information based on the filtered up-sampled base layer reference block; determine weighted base layer residual information by applying a weighting factor to the base layer residual information; and determine an enhancement layer block based on the weighted base layer residual information. The processor may encode or decode the video information. | 07-03-2014 |
20140192868 | INTER-LAYER REFERENCE PICTURE GENERATION FOR HLS-ONLY SCALABLE VIDEO CODING - An apparatus for coding video information may include a memory unit configured to store video information associated with a picture and a processor in communication with the memory unit configured to resample video information of a reference picture to obtain a resampled picture having a plurality of slices and a different picture size than a picture to be encoded. Further, the processor may determine slice definitions for slices in the resampled picture. The slices of the resampled picture may correspond to slices of the reference picture. The processor may determine, based on the slice definitions, whether a slice of the resampled picture satisfies one or more slice definition rules. In response to determining that the slice of the resampled picture does not satisfy at least one slice definition rule, the processor can modify the slice definition for the slice so as to satisfy the slice definition rule. | 07-10-2014 |
20140192892 | BLOCKINESS METRIC FOR LARGE HEVC BLOCK ARTIFACTS - In one example, a method for encoding video data includes determining a first value for a line of a first plurality of lines, wherein the first plurality of lines are perpendicular to a first edge of a block of the video data; in response to determining that the value satisfies at least one threshold: determining a second value for the line of the first plurality of lines; and determining a second value for a line of a second plurality of lines, wherein the second plurality of lines are perpendicular to a second edge of the block of video data; determining, based on the second value for the line of the first plurality of lines and the second value for the line of the second plurality of lines, an average value; and determining, based on the average value, whether or not to encode one or more deblocking parameter offsets. | 07-10-2014 |
20140198846 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code (e.g., encode or decode) video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a base layer and an enhancement layer, the enhancement layer comprising an enhancement layer (EL) block and the base layer comprising a base layer (BL) block that is co-located with the enhancement layer block. The processor is configured to determine predicted pixel information of the EL block by applying a prediction function to pixel information of the BL block, and to determine the EL block using the predicted pixel information. The processor may encode or decode the video information. | 07-17-2014 |
20140198855 | SQUARE BLOCK PREDICTION - Systems, devices, and methods for coding video data may limit an intra-prediction angle to predict a chroma component from a reference array. The limited intra-prediction angle used varies between a value that is less than or equal to a maximum intra-prediction angle of a luma component. The systems, devices, and methods for coding video data may code a chroma intra-coded current block based on the limited intra-prediction angle. In another example, systems devices, and methods for coding video data may extend the reference array based on reference values that are outside the reference array in a video coding scheme including a number of intra-prediction angles, store prediction values in the extended reference array, and intra-coding a current block based on at least the prediction values in the extended reference array. | 07-17-2014 |
20140219342 | MODE DECISION SIMPLIFICATION FOR INTRA PREDICTION - In general, techniques are described for reducing the complexity of mode selection when selecting from multiple, different prediction modes. A video coding device comprising a processor may perform the techniques. The processor may compute approximate costs for a pre-defined set of intra-prediction modes identified in a current set. The current set of intra-prediction modes may include fewer modes than a total number of intra-prediction modes. The processor may compare approximate costs computed for one or more most probable intra-prediction modes to a threshold and replace one or more of the intra-prediction modes of the current set with one or more most probable intra-prediction modes. The processor may perform rate distortion analysis with respect to each intra-prediction mode identified in the current set and perform intra-prediction coding with respect to the current block using a mode of the current set. | 08-07-2014 |
20140219349 | INTRA PREDICTION MODE DECISION WITH REDUCED STORAGE - In general, techniques are described for reducing the space required to store rate distortion values when selecting from multiple, different prediction modes. A video coding device comprising a processor may perform the techniques. The processor may determine first and second sets of intra-prediction modes for a current block of video data. The first and second sets of intra-prediction modes may include less intra-prediction modes, collectively, than a total number of intra-prediction modes. The processor may compute an approximate cost for each intra-prediction mode included in the first and second sets of intra-prediction modes. The processor may store the approximate cost for each intra-prediction mode identified in the first and second sets of intra-prediction modes to a memory. The processor may perform intra-prediction to encode the current block using a mode identified in at least one of the first or second set. | 08-07-2014 |
20140241417 | OFFSET TYPE AND COEFFICIENTS SIGNALING METHOD FOR SAMPLE ADAPTIVE OFFSET - This disclosure describes techniques for performing sample adaptive offset signaling and coding in a video coding process. Techniques of the disclosure include both a merge-based and prediction-based signaling process for sample adaptive offset information (i.e., offset values and offset type). The techniques includes determining offset information for a current partition, comparing the offset information of the current partition with offset information of one or more neighbor partitions, coding a merge instruction in the case that the offset information of one of the one or more neighbor partitions is the same as the offset information of the current partition, and coding one of a plurality of prediction instructions in the case that the offset information of the one or more neighbor partitions is not the same as the offset information of the current partition. | 08-28-2014 |
20140253681 | INTER-VIEW RESIDUAL PREDICTION IN MULTI-VIEW OR 3-DIMENSIONAL VIDEO CODING - A video coder scales a motion vector of a current prediction unit (PU) of a current picture in order to compensate for a difference in temporal distance. In addition, the video coder determines a predictive block for the current PU, determines a disparity reference block based on samples of the disparity reference picture at a location indicated by a disparity vector of the current PU, and determines, based on samples of the fixed reference picture at a location indicated by the scaled motion vector, a temporal-disparity reference block for the current PU. The video coder then determines a residual predictor for the current PU. Each sample of the residual predictor for the current PU indicates a difference between a sample of the temporal-disparity reference block for the current PU and a corresponding sample of the disparity reference block. | 09-11-2014 |
20140253682 | SIMPLIFIED DEPTH CODING - In an example, a method of coding video data includes determining a first depth value of a depth look up table (DLT), where the first depth value is associated with a first pixel of the video data. The method also includes determining a second depth value of the DLT, where the second depth value is associated with a second pixel of the video data, The method also includes coding the DLT including coding the second depth value relative to the first depth value. | 09-11-2014 |
20140267605 | SIMPLIFICATIONS ON DISPARITY VECTOR DERIVATION AND MOTION VECTOR PREDICTION IN 3D VIDEO CODING - A device for coding three-dimensional video data includes a video coder configured to determine a first block of a first texture view is to be coded using a block-based view synthesis mode; locate, in a depth view, a first depth block that corresponds to the first block of the first texture view; determine depth values of two or more corner positions of the first depth block; based on the depth values, derive a disparity vector for the first block; using the disparity vector, locate a first block of a second texture view; and, inter-predict the first block of the first texture view using the first block of the second texture view. | 09-18-2014 |
20140269898 | SIMPLIFICATIONS ON DISPARITY VECTOR DERIVATION AND MOTION VECTOR PREDICTION IN 3D VIDEO CODING - A video coder can be configured to perform texture first coding for a first texture view, a first depth view, a second texture view, and a second depth view; for a macroblock of the second texture view, locate a depth block of the first depth view that corresponds to the macroblock; based on at least one depth value of the depth block, derive a disparity vector for the macroblock; code a first sub-block of the macroblock based on the derived disparity vector; and, code a second sub-block of the macroblock based on the derived disparity vector. | 09-18-2014 |
20140269939 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information comprises a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a reference layer (RL) and an enhancement layer (EL). The EL comprises an EL video unit and the RL comprises an RL video unit corresponding to the EL video unit. The processor is configured to perform upsampling and bit-depth conversion on pixel information of the RL video unit in a single combined process to determine predicted pixel information of the EL video unit, and determine the EL video unit using the predicted pixel information. | 09-18-2014 |
20140286400 | INTRA PREDICTION MODES FOR LOSSY CODING WHEN TRANSFORM IS SKIPPED - A video coder applies a residual differential pulse code modulation technique to a residual data of a block coded using lossy coding. The block may be coded without application of a transform to the residual data. | 09-25-2014 |
20140286412 | INTRA DC PREDICTION FOR LOSSLESS CODING IN VIDEO CODING - A video coder may generate a predictive block. As part of generating the predictive block, the video coder may use at least one of a losslessly reconstructed sample to left of a current sample in a current row of a predictive block and a losslessly reconstructed sample for a row of the predictive block above the current row for DC prediction of the current sample. | 09-25-2014 |
20140286413 | DISABLING SIGN DATA HIDING IN VIDEO CODING - A video coder may determine that sign data hiding is disabled for a current block if the current block is generated using lossy coding without application of a transform to residual data and the current block is intra predicted using an intra prediction mode in which a residual differential pulse code modulation (DPCM) technique is used. | 09-25-2014 |
20140294062 | INTER-LAYER REFERENCE PICTURE RESTRICTION FOR HIGH LEVEL SYNTAX-ONLY SCALABLE VIDEO CODING - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory unit configured to store reference layer pictures associated with a reference layer, an enhancement layer, or both. The apparatus further comprises a processor operationally coupled to the memory unit. In one embodiment, the processor is configured to restrict usage of at most one reference layer pictures that has been resampled as an inter-layer reference picture, and predict a current picture using inter-layer prediction and the inter-layer reference picture. | 10-02-2014 |
20140294063 | INTER-LAYER REFERENCE PICTURE RESTRICTION FOR HIGH LEVEL SYNTAX-ONLY SCALABLE VIDEO CODING - In one implementation, an apparatus is provided for encoding or decoding video information. The apparatus comprises a memory unit configured to store reference layer pictures associated with a reference layer, an enhancement layer, or both. The apparatus further comprises a processor operationally coupled to the memory unit. In one embodiment, the processor is configured to restrict usage of at most one reference layer pictures that has a different spatial resolution than a current picture as an inter-layer reference picture, and predict the current picture using inter-layer prediction and the inter-layer reference picture. | 10-02-2014 |
20140301438 | SAMPLE ADAPTIVE OFFSET SCALING BASED ON BIT-DEPTH - This disclosure provides systems, methods and apparatus for sample adaptive offset (SAO) scaling. For example, the apparatus may include a processor configured to determine an offset value for an SAO filter applied to video data to improve reconstruction of signal amplitudes in the video data. The processor may be further configured to determine a first value indicative of a bit depth and a second value indicative of a scale factor for the video data, to provide a scaled offset value based on applying the scale factor to the offset value, and to scale at least one color component of the video data according to the scaled offset value. The processor may also be configured to identify an edge offset category for a scaled group of neighboring pixel values, and to adjust the SAO filter based on the identified edge offset category. | 10-09-2014 |
20140301457 | MULTIPLE BASE LAYER REFERENCE PICTURES FOR SHVC - A method of coding video data can include receiving video information associated with a reference layer, an enhancement layer, or both, and generating a plurality of inter-layer reference pictures using a plurality of inter-layer filters and one or more reference layer pictures. The generated plurality of inter-layer reference pictures may be inserted into a reference picture list. A current picture in the enhancement layer may be coded using the reference picture list. The inter-layer filters may comprise default inter-layer filters or alternative inter-layer filters signaled in a sequence parameter set, video parameter set, or slice header. | 10-09-2014 |
20140301458 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code (e.g., encode or decode) video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a video layer comprising one or more temporal sub-layers. The processor is configured to determine presence information for a coded video sequence in a bitstream, the presence information indicating whether said one or more temporal sub-layers of the video layer are present in the bitstream. The processor may encode or decode the video information. | 10-09-2014 |
20140301460 | INTRA RATE CONTROL FOR VIDEO ENCODING BASED ON SUM OF ABSOLUTE TRANSFORMED DIFFERENCE - This disclosure describes techniques for rate control for intra coded frames. In one example of the disclosure, a rate control parameter may be calculated using a target bit rate and a complexity measure. In one example, the complexity measure is calculated with a sum of absolute transformed differences (SATD) calculation of an intra-coded frame. | 10-09-2014 |
20140301466 | GENERALIZED RESIDUAL PREDICTION IN HIGH-LEVEL SYNTAX ONLY SHVC AND SIGNALING AND MANAGEMENT THEREOF - An apparatus for decoding video information according to certain aspects includes a memory unit and a processor operationally coupled to the memory unit. The memory unit is configured to store at least one reference picture list of an enhancement layer, the at least one reference picture list comprising residual prediction reference picture information. The processor is configured to: decode signaled information about residual prediction reference picture generation; generate a residual prediction reference picture based on an enhancement layer reference picture and the decoded signaled information such that the generated residual prediction reference picture has the same motion field and the same picture order count (POC) as the enhancement layer reference picture from which it is generated; and store the generated residual prediction reference picture in the at least one reference picture list of the enhancement layer. | 10-09-2014 |
20140301474 | DETERMINING PALETTES IN PALETTE-BASED VIDEO CODING - In an example, a method of coding video data includes determining a first palette having first entries indicating first pixel values, determining, based on the first entries of the first palette, one or more second entries indicating second pixel values of a second palette, and coding pixels of a block of video data using the second palette. | 10-09-2014 |
20140301475 | DETERMINING PALETTE INDICES IN PALETTE-BASED VIDEO CODING - In an example, a method of coding video data includes determining a first index value associated with a first pixel in a block of video data, wherein the first index value relates a position of the first pixel to an entry of a palette of pixel values, determining, based on the first index value, one or more second index values associated with one or more second pixels in the block of video data, wherein the second index values relate the positions of the one or more second pixels to one or more entries of the palette of pixel values, and coding the first pixel and the one or more second pixels of the block of video data. | 10-09-2014 |
20140307777 | CODING OF TRANSFORM COEFFICIENTS FOR VIDEO CODING - This disclosure describes devices and methods for coding transform coefficients associated with a block of residual video data in a video coding process. Aspects of this disclosure include the selection of a scan order for both significance map coding and level coding, as well as the selection of contexts for entropy coding consistent with the selected scan order. This disclosure proposes a harmonization of the scan order to code both the significance map of the transform coefficients as well as to code the levels of the transform coefficient. It is proposed that the scan order for the significance map should be in the inverse direction (i.e., from the higher frequencies to the lower frequencies). This disclosure also proposes that transform coefficients be scanned in sub-sets as opposed to fixed sub-blocks. In particular, transform coefficients are scanned in a sub-set consisting of a number of consecutive coefficients according to the scan order. | 10-16-2014 |
20140307800 | RICE PARAMETER UPDATE FOR COEFFICIENT LEVEL CODING IN VIDEO CODING PROCESS - Techniques are described for updating a value of a Rice parameter used to define Golomb codes for coefficient level coding. The Golomb codes defined by the Rice parameter may be used to code a remaining absolute value of a coefficient level for at least one coefficient in a coefficient group (CG). According to the techniques, the value of the Rice parameter is updated based on a selection of a minimum of either a maximum value of the Rice parameter or a variable increment of the value of the Rice parameter. The variable increment is determined based on the value of the Rice parameter and an absolute value of the coefficient level for the current coefficient being coded. Techniques are also described for adjusting the value of the Rice parameter used to define Golomb codes for coefficient level coding based on whether a transform is applied to a transform block. | 10-16-2014 |
20140314140 | INDICATION OF CROSS-LAYER PICTURE TYPE ALIGNMENT IN MULTI-LAYER VIDEO CODING - In one example, the disclosure is directed to techniques that include receiving a bitstream comprising at least a syntax element, a first network abstraction layer unit type, and a coded access unit comprising a plurality of pictures. The techniques further include determining a value of the syntax element which indicates whether the access unit was coded using cross-layer alignment. The techniques further include determining the first network abstraction layer unit type for a picture in the access unit and determining whether the first network abstraction layer unit type equals a value in a range of type values. The techniques further include setting a network abstraction layer unit type for all other pictures in the coded access unit to equal the value of the first network abstraction layer unit type if the first network abstraction layer unit type is equal to a value in the range of type values. | 10-23-2014 |
20140328398 | CONDITIONALLY INVOKING A RESAMPLING PROCESS IN SHVC - An apparatus for coding video information according to certain aspects includes a memory and a processor. The memory unit is configured to store video information associated with an interlayer reference picture for a current picture to be coded. The processor is configured to: receive information relating to a plurality of interlayer reference offsets that are configured to define a region of a resampled version of the interlayer reference picture, wherein the region is used to generate a prediction of the current picture, and wherein the plurality of interlayer reference offsets include a left offset, a top offset, a right offset, and a bottom offset that are each specified relative to the current picture; determine based at least in part on the plurality of interlayer reference offsets whether to resample the interlayer reference picture; and in response to determining to resample the interlayer reference picture, resample the interlayer reference picture. | 11-06-2014 |
20140341274 | CODING THE POSITION OF A LAST SIGNIFICANT COEFFICIENT WITHIN A VIDEO BLOCK BASED ON A SCANNING ORDER FOR THE BLOCK IN VIDEO CODING - In one example, an apparatus is disclosed for coding coefficients associated with a block of video data during a video coding process, wherein the apparatus includes a video coder configured to code x- and y-coordinates that indicate a position of a last non-zero coefficient within the block according to a scanning order associated with the block when the scanning order comprises a first scanning order, and code interchanged x- and y-coordinates that indicate the position of the last non-zero coefficient within the block according to the scanning order when the scanning order comprises a second scanning order, wherein the second scanning order is different than the first scanning order. | 11-20-2014 |
20140341275 | SIGNALING PICTURE SIZE IN VIDEO CODING - A video encoder is configured to determine a picture size for one or more pictures included in a video sequence. The picture size associated with the video sequence may be a multiple of an aligned coding unit size for the video sequence. In one example, the aligned coding unit size for the video sequence may comprise a minimum coding unit size where the minimum coding unit size is selected from a plurality of smallest coding unit sizes corresponding to different pictures in the video sequence. A video decoder is configured to obtain syntax elements to determine the picture size and the aligned coding unit size for the video sequence. The video decoder decodes the pictures included in the video sequence with the picture size, and stores the decoded pictures in a decoded picture buffer. | 11-20-2014 |
20140341297 | VIDEO CODING USING ADAPTIVE MOTION VECTOR RESOLUTION - In one example, a device for coding video data includes a video coder configured to configured to code information representative of whether an absolute value of an x-component of a motion vector difference value for a current block of video data is greater than zero, code information representative of whether an absolute value of a y-component of the motion vector difference value is greater than zero, when the absolute value of the x-component is greater than zero, code information representative of the absolute value of the x-component, when the absolute value of the y-component is greater than zero, code information representative of the absolute value of the y-component, when the absolute value of the x-component is greater than zero, code a sign of the x-component, and when the absolute value of the y-component is greater than zero, code a sign of the y-component. | 11-20-2014 |
20140348240 | VIDEO CODING USING SAMPLE PREDICTION AMONG COLOR COMPONENTS - A video coder may reconstruct a residual signal of a predictor color component generated using motion prediction. The reconstructed residual signal of the predictor color component may include reconstructed residual sample values of the predictor color component. Additionally, the video coder may use the reconstructed residual sample values of the predictor color component to predict residual sample values of a different, predicted color component. | 11-27-2014 |
20140355666 | ADVANCED DEPTH INTER CODING BASED ON DISPARITY OF DEPTH BLOCKS - In one example, the disclosure is directed to techniques that include, for each prediction unit (PU) of a respective coding unit (CU) of a slice of a picture of the video data, determining at least one disparity value based at least in part on at least one depth value of at least one reconstructed depth sample of at least one neighboring sample. The techniques further include determining at least one disparity vector based at least in part on the at least one disparity value, wherein the at least one disparity vector is for the respective CU for each PU. The techniques further include reconstructing, based at least in part on the at least one disparity vector, a coding block for the respective CU for each PU. | 12-04-2014 |
20140355669 | CONTEXT REDUCTION FOR CONTEXT ADAPTIVE BINARY ARITHMETIC CODING - A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56. | 12-04-2014 |
20140355681 | CONTEXT REDUCTION FOR CONTEXT ADAPTIVE BINARY ARITHMETIC CODING - A reduction in the number of binarizations and/or contexts used in context adaptive binary arithmetic coding (CABAC) for video coding is proposed. In particular, this disclosure proposes techniques that may lower the number contexts used in CABAC by up to 56. | 12-04-2014 |
20140362909 | DYNAMIC RANGE CONTROL OF INTERMEDIATE DATA IN RESAMPLING PROCESS - An apparatus for coding video information according to certain aspects includes a memory and a processor. The memory unit is configured to store video information. The processor is configured to: obtain reference layer video information; upsample the reference layer video information in a first dimension to generate an intermediate output; constrain the intermediate output to a predetermined bit depth; and upsample the constrained intermediate output in a second dimension, wherein the second dimension is orthogonal to the first dimension. | 12-11-2014 |
20140362917 | RESIDUAL DIFFERENTIAL PULSE CODE MODULATION (DPCM) EXTENSIONS AND HARMONIZATION WITH TRANSFORM SKIP, ROTATION, AND SCANS - Techniques are described for harmonizing coding techniques when residual differential pulse code modulation (RDPCM) is applied to a residual block. In some examples, a scan order used for such a residual block may be required to be the same as when the residual block is generated from intra-predicting the current block and when the residual block is generated from inter-predicting or intra block copy predicting the current block. | 12-11-2014 |
20140369404 | SIGNALING OF DEBLOCKING FILTER PARAMETERS IN VIDEO CODING - This disclosure describes techniques for signaling deblocking filter parameters for a current slice of video data with reduced bitstream overhead. Deblocking filter parameters may be coded in one or more of a picture layer parameter set and a slice header. The techniques reduce a number of bits used to signal the deblocking filter parameters by coding a first syntax element that indicates whether deblocking filter parameters are present in both the picture layer parameter set and the slice header, and only coding a second syntax element in the slice header when both sets of deblocking filter parameters are present. Coding the second syntax element is eliminated when deblocking filter parameters are present in only one of the picture layer parameter set or the slice header. The second syntax element indicates which set of deblocking filter parameters to use to define a deblocking filter applied to a current slice. | 12-18-2014 |
20140369426 | INTER-COMPONENT FILTERING - In one embodiment, an apparatus configured to encode video information includes a memory unit and a processor. The memory unit is configured to store an inter-layer reference picture comprising video information. The processor is operationally coupled to the memory unit. In addition, the processor is configured to partition the inter-layer reference picture into a quadtree structure having a plurality of leafs; determine inter-component filter parameters specific to each individual leaf based upon the video information in each individual leaf and signal the inter-component filter parameters for each of the leafs. | 12-18-2014 |
20140376611 | ADAPTIVE COLOR TRANSFORMS FOR VIDEO CODING - A device for coding video data includes a memory and at least one processor configured to determine a cost associated with a plurality of color transforms associated with a coding unit, determine a cost associated with a plurality of color transforms associated with a coding unit, select a color transform of the plurality of color transforms having a lowest associated cost, transform a first block of video data having a first, Red, Green, Blue (RGB) color space to produce a second block of video data having a second color space using the selected color transform of the plurality of color transforms, and encode the second video block having the second color space. | 12-25-2014 |
20140376633 | MORE ACCURATE ADVANCED RESIDUAL PREDICTION (ARP) FOR TEXTURE CODING - Techniques for advanced residual prediction (ARP) for coding video data may include inter-view ARP. Inter-view ARP may include identifying a disparity motion vector (DMV) for a current video block. The DMV is used for inter-view prediction of the current video block based on an inter-view reference video block. The techniques for inter-view ARP may also include identifying temporal reference video blocks in the current and reference views based on a temporal motion vector (TMV) of the inter-view reference video block, and determining a residual predictor block based on a difference between the temporal reference video blocks. | 12-25-2014 |
20140376634 | INTRA PREDICTION FROM A PREDICTIVE BLOCK - Techniques coding video data, including a mode for intra prediction of blocks of video data from predictive blocks of video data within the same picture, may include determining a predictive block of video data for the current block of video data, wherein the predictive block of video data is a reconstructed block of video data within the same picture as the current block of video data. A two-dimensional vector, which may be used by a video coder to identify the predictive block of video data, includes a horizontal displacement component and a vertical displacement component relative to the current block of video data. The mode for intra prediction of blocks of video data from predictive blocks of video data within the same picture may be referred to as Intra Block Copy or Intra Motion Compensation. | 12-25-2014 |
20150016501 | PALETTE PREDICTION IN PALETTE-BASED VIDEO CODING - In palette-based coding, a video coder may form a so-called “palette” as a table of colors representing the video data of a given block. The video coder may code index values for one or more pixels values of a current block of video data, where the index values indicate entries in the palette that represent the pixel values of the current block. According to the techniques, a video coder determines one or more palette entries in a predictive palette that are copied to the current palette, and a number of new palette entries not in the predictive palette that are included in the current palette. The video coder calculates a size of the current palette equal to the sum of the number of the copied palette entries and the number of the new palette entries, and generates the current palette including the copied palette entries and the new palette entries. | 01-15-2015 |
20150016502 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory unit and a processor in communication with the memory unit. The memory unit is configured to store video information associated with a current layer and an enhancement layer, the current layer having a current picture. The processor is configured to determine whether the current layer may be coded using information from the enhancement layer, determine whether the enhancement layer has an enhancement layer picture corresponding to the current picture, and in response to determining that the current layer may be coded using information from the enhancement layer and that the enhancement layer has an enhancement layer picture corresponding to the current picture, code the current picture based on the enhancement layer picture. The processor may encode or decode the video information. | 01-15-2015 |
20150016512 | INTER-COLOR COMPONENT RESIDUAL PREDICTION - A device for decoding video data includes a memory configured to store video data, and at least one processor. The at least one processor may be configured to: determine luma residual samples for a block of video data, determine predictive chroma residual samples for the block of video data, scale the luma residual samples with a scale factor to produce scaled luma residual samples, and determine updated chroma residual samples based on the predictive chroma residual samples and the scaled luma residual samples. | 01-15-2015 |
20150016533 | INTRA MOTION COMPENSATION EXTENSIONS - A video coder comprising one or more processors determines that a current block of the video data is encoded using an intra motion compensation (IMC) mode, wherein the current block is in a frame of video; determines an offset vector for a first color component of the current block of the video data; locates, in the frame of video, a reference block of the first color component using the offset vector; modifies the offset vector to generate a modified offset vector in response to the offset vector pointing to a sub-pixel position for a second color component of the current block of video data; locates, in the frame of video, a reference block for the second color component using the modified offset vector; and codes the current block based on the reference block for the first color component and the reference block for the second color component. | 01-15-2015 |
20150016537 | RICE PARAMETER INITIALIZATION FOR COEFFICIENT LEVEL CODING IN VIDEO CODING PROCESS - Techniques are described for initializing a Rice parameter used to define codes for coefficient level coding. According to the techniques, the initial value of the Rice parameter is determined for a coefficient group (CG) in a transform block of video data based on statistics of coefficient levels that are gathered for previously coded coefficients of the video data. The statistics may be statistics of absolute values of coefficient levels or remaining absolute values of coefficient levels of previously coded coefficients. A value of the statistics may be initialized to zero at a beginning of a video slice and updated based on coefficient levels coded in each CG of the slice. The statistics may be updated once per CG. In some cases, statistics may be gathered separately for each of a plurality of different categories of CGs that are defined based on characteristics of transform blocks that include the CGs. | 01-15-2015 |
20150016550 | ADAPTIVE FILTERING IN VIDEO CODING - Performing deblock filtering on video data may include determining, for a first non-luma color component of the video data, whether to perform deblock filtering based on a first deblock filtering process or a second deblock filtering process. Next, deblock filtering may be performed on the first non-luma color component in accordance with the determined deblock filtering process. | 01-15-2015 |
20150023405 | DISABLING INTRA PREDICTION FILTERING - A video decoder may adaptively disable, based on a syntax element, one or more filters used for intra-prediction. In addition, the video decoder may perform intra-prediction to generate prediction data for a current block of a current video slice. Furthermore, a video encoder may adaptively disable one or more filters used for intra-prediction. Furthermore, the video encoder may signal a syntax element that controls the one or more filters. In addition, the video encoder may perform intra prediction to generate prediction data for a current video block the video data. | 01-22-2015 |
20150023419 | DEVICE AND METHOD FOR SCALABLE CODING OF VIDEO INFORMATION - An apparatus configured to code video information includes a memory and a processor in communication with the memory. The memory is configured to store video information associated with a reference layer and an enhancement layer, the reference layer comprising a reference layer (RL) picture having a first slice and a second slice, and the enhancement layer comprising an enhancement layer (EL) picture corresponding to the RL picture. The processor is configured to generate an inter-layer reference picture (ILRP) by upsampling the RL picture, the ILRP having a single slice associated therewith, set slice information of the single slice of the ILRP equal to slice information of the first slice, and use the ILRP to code at least a portion of the EL picture. The processor may encode or decode the video information. | 01-22-2015 |
20150023422 | PROCESSING ILLUMINATION COMPENSATION FOR VIDEO CODING - In one example, a device for coding (e.g., encoding or decoding) video data includes a memory configured to store video data and a video coder configured to determine a value for an advanced residual prediction (ARP) weighting factor of a current block of the video data and to skip coding of an illumination compensation syntax element for the current block and to code the current block when the value of the ARP weighting factor is not equal to zero. The video coder may further be configured to code the illumination compensation syntax element for the current block and code the current block based at least in part on the value of the illumination compensation syntax element when the value of the ARP weighting factor is equal to zero. | 01-22-2015 |
20150023423 | BLOCK IDENTIFICATION USING DISPARITY VECTOR IN VIDEO CODING - Techniques are described for determining a block in a reference picture in a reference view based on a disparity vector for a current block. The techniques start the disparity vector from a bottom-right pixel in a center 2×2 sub-block within the current block, and determine a location within the reference picture to which the disparity vector refers. The determined block covers the location referred to by the disparity vector based on the disparity vector starting from the bottom-right pixel in the center 2×2 sub-block within the current block. | 01-22-2015 |
20150055703 | DETERMINING REGIONS WHEN PERFORMING INTRA BLOCK COPYING - In general, techniques are described for performing an intra block copying process to code video data. A video decoding device that includes a memory and one or more processors may perform the techniques. The memory may be configured to store a current block of a picture. The processors may be configured to perform an intra block copying process to decode the current block using a prediction block that is from a same slice or a same tile as that in which the coded current block resides, the prediction block restricted to be within a search region that only includes the same slice or the same tile as that in which the coded current block resides. | 02-26-2015 |
20150055704 | SUB-PU-LEVEL ADVANCED RESIDUAL PREDICTION - A prediction unit (PU) of a coding unit (CU) is split into two or more sub-PUs including a first sub-PU and a second sub-PU. A first motion vector of a first type is obtained for the first sub-PU and a second motion vector of the first type is obtained for the second sub-PU. A third motion vector of a second type is obtained for the first sub-PU and a fourth motion vector of the second type is obtained for the second sub-PU, such that the second type is different than the first type. A first portion of the CU corresponding to the first sub-PU is coded according to advanced residual prediction (ARP) using the first and third motion vectors. A second portion of the CU corresponding to the second sub-PU is coded according to ARP using the second and fourth motion vectors. | 02-26-2015 |
20150063440 | CONSTRAINED INTRA PREDICTION IN VIDEO CODING - A method of encoding video data includes determining that a current block of video data is to be encoded using an intra block copy (BC) mode and constrained intra prediction, determining one or more reference blocks that may be used for encoding the current block using the intra BC mode, wherein each of the one or more reference blocks is encoded with an intra prediction mode, and encoding the current block of video data using the intra BC mode and at least one of the determined one or more reference blocks. | 03-05-2015 |
20150063454 | RESIDUAL PREDICTION FOR INTRA BLOCK COPYING - In an example, a method of decoding video data includes generating a residual block of a picture based on a predicted residual block including reconstructing one or more residual values of the residual block based on one or more predicted residual values of the residual block. The method also includes generating a current block of the picture based on a combination of the residual block and a prediction block of the picture. | 03-05-2015 |
20150063464 | LOOKUP TABLE CODING - In general, techniques are described for lookup table coding. A device comprising one or more processors and a memory may be configured to perform the techniques. The processors are configured to receive at least one difference table including a set of values, each value of the set being included or not included in the reference lookup table and generate a current lookup table based on the reference lookup table and the difference table. The current lookup table may include at least one of a value from the difference table that is not included in the reference table or a value from the reference table that is not included in the difference table. The one or more processors may then decode the video data based on a set of values of the current lookup table. The memory may be configured to store the current lookup table. | 03-05-2015 |
20150071357 | PARTIAL INTRA BLOCK COPYING FOR VIDEO CODING - In general, techniques are described for coding a current video block within a current picture based on a predictor block within the current picture, the predictor block identified by a block vector. The techniques include identifying an unavailable pixel of the predictor block, obtaining a value for the unavailable pixel based on at least one neighboring reconstructed pixel of the unavailable pixel, and coding the current video block based on a version of the predictor block that includes the obtained value for the unavailable pixel. The unavailable pixel may be located outside of a reconstructed region of the current picture. | 03-12-2015 |
20150071359 | TWO LEVEL LAST SIGNIFICANT COEFFICIENT (LSC) POSITION CODING - In some examples, a video coder employs a two-level technique to code information that identifies a position within the block of transform coefficients of one of the coefficients that is a last significant coefficient (LSC) for the block according to a scanning order associated with the block of transform coefficients. For example, a video coder may code a sub-block position that identifies a position of one of the sub-blocks that includes the LSC within the block, and code a coefficient position that identifies a position of the LSC within the sub-block that includes the LSC. | 03-12-2015 |