Abstract:
In general, this disclosure describes techniques for coding video blocks using a color-space conversion process. A video coder, such as a video encoder or a video decoder, may determine whether to use color-space conversion for a coding unit and set a value of a syntax element of the coding unit to indicate the use of color-space conversion. The video coder may apply a color-space transform process in encoding the coding unit. The video coder may decode the syntax element of the coding unit. The video coder may determine whether a value of the syntax element indicates that the coding unit was encoded using color-space conversion. The video coder may apply a color-space inverse transform process in decoding the coding unit in response to determining that the syntax element indicates that the coding unit was coded using color-space conversion.
Abstract:
An apparatus for coding video data according to certain aspects includes a memory and a processor in communication with the memory. The memory is configured to store video information, such as base layer video information and enhancement layer video information. The processor is configured to determine a value of a current video unit of enhancement layer video information based at least on a weighted inter-layer predictor and a weighted intra-layer predictor of at least one color component of the current video unit.
Abstract:
In one embodiment, a video coder for coding video data includes a processor and a memory. The processor selects a filter set from a multiple filter sets for upsampling reference layer video data based at least on a prediction operation mode for enhanced layer video data and upsamples the reference layer video data using the selected filter set. Some of the multiple filter sets have some different filter characteristics from one another, and the upsampled reference layer video data has the same spatial resolution as the enhanced layer video data. The processor further codes the enhanced layer video data based at least on the upsampled reference layer video data and the prediction operation mode. The memory stores the upsampled reference layer video data.
Abstract:
An example video encoder is configured to receive an indication of merge mode coding of a block within a parallel motion estimation region (PMER), generate a merge mode candidate list comprising one or more spatial neighbor motion vector (MV) candidates and one or more temporal motion vector prediction (TMVP) candidates, wherein motion information of at least one of the spatial neighbor MV candidates is known to be unavailable during coding of the block at an encoder, determine an index value identifying, within the merge mode candidate list, one of the TMVP candidates or the spatial neighbor MV candidates for which motion information is available during coding of the particular block, and merge mode code the block using the identified MV candidate.
Abstract:
A video coding device configured according to some aspects of this disclosure includes a memory configured to store an initial list of motion vector candidates and a temporal motion vector predictor (TMVP). The video coding device also includes a processor in communication with the memory. The processor is configured to obtain a merge candidate list size value (N) and identify motion vector candidates to include in a merge candidate list having a list size equal to the merge candidate list size value. The merge candidate list may be a merge motion vector (MV) candidate list or a motion vector predictor (MVP) candidate list (also known as an AM VP candidate list). The processor generates the merge candidate list such that the merge candidate list includes the TMVP, regardless of the list size.
Abstract:
A video decoder determines, based at least in part on a size of a prediction unit (PU), whether to round either or both a horizontal or a vertical component of a motion vector of the PU from sub-pixel accuracy to integer-pixel accuracy. The video decoder generates, based at least in part on the motion vector, a predictive sample block for the PU and generates, based in part on the predictive sample block for the PU, a reconstructed sample block.
Abstract:
Un método para decodificar datos de vídeo que incluye recibir un flujo de bits que incluye una secuencia de bits que forma una representación de una imagen codificada de los datos de vídeo, dividir la imagen codificada de los datos de vídeo en una pluralidad de bloques utilizando tres o más estructuras de partición diferentes y reconstruir la pluralidad de bloques de la imagen codificada de los datos de vídeo. La división de la imagen codificada de los datos de vídeo puede incluir la división de la imagen codificada de los datos de vídeo en la pluralidad de bloques utilizando las tres o más estructuras de partición diferentes, en donde al menos tres de las tres o más estructuras de partición diferentes pueden utilizarse en cada profundidad de una estructura de árbol que representa cómo se divide un bloque particular de la imagen codificada de los datos de vídeo. (Traducción automática con Google Translate, sin valor legal)
Abstract:
A video coder may determine a motion vector of a non-adjacent block of a current picture of the video data. The non-adjacent block is non-adjacent to a current block of the current picture. Furthermore, the video coder determines, based on the motion vector of the non-adjacent block, a motion vector predictor (MVP) for the current block. The video coder may determine a motion vector of the current block. The video coder may also determine a predictive block based on the motion vector of the current block.
Abstract:
A video decoder can be configured to determine that a block of video data is encoded using a bi-directional inter prediction mode; determine that the block of video data is encoded using a bi-directional optical flow (BIO) process; inter predict the block of video data according to the bi-directional inter prediction mode; perform the BIO process for the block, wherein performing the BIO process for the block comprises determining a single motion vector refinement for a group of pixels in the block, wherein the group of pixels comprises at least two pixels; refine the group of pixels based on the single motion vector refinement; and output a BIO refined predictive block of video data comprising the refined group of pixels.