Abstract:
Systems, devices and methods are described including performing scalable video coding using inter-layer pixel sample prediction. Inter-layer pixel sample prediction in an enhancement layer coding unit, prediction unit, or transform unit may use reconstructed pixel samples obtained from a base layer or from a lower enhancement layer. The pixel samples may be subjected to upsample filtering and/or refinement filtering. The upsample or refinement filter coefficients may be predetermined or may be adaptively determined.
Abstract:
Systems, apparatus, articles, and methods are described including operations for 3D video coding including depth based disparity vector calibration.
Abstract:
A three-dimensional (3D) video codec encodes multiple views of a 3D video, each including texture and depth components. The encoders of the codec encode video blocks of their respective views based on a set of prediction parameters, such as quad-tree split flags, prediction modes, partition sizes, motion fields, inter directions, reference indices, luma intra modes, and chroma intra modes. The prediction parameters may be inherited across different views and different ones of the texture and depth components.
Abstract:
Systems, mediums, and methods for simplified depth coding for 3D video coding comprises performing modified intra DC coding or modified intra planar coding of at least one coding unit associated with a plurality of pixels and of at least one depth map, and comprising forming a prediction value of the at least one coding unit.
Abstract:
Systems, devices and methods are described including performing scalable video coding using inter-layer pixel sample prediction. Inter-layer pixel sample prediction in an enhancement layer coding unit, prediction unit, or transform unit may use reconstructed pixel samples obtained from a base layer or from a lower enhancement layer. The pixel samples may be subjected to upsample filtering and/or refinement filtering. The upsample or refinement filter coefficients may be predetermined or may be adaptively determined.
Abstract:
Systems, apparatus, articles, and methods are described including operations for size based transform unit context derivation. In an example encoder, first circuitry is to encode video input data into a bitstream according to a bitstream syntax, wherein the video input data includes one or more pictures, the one or more pictures are partitioned into one or more coding tree blocks, the one or more coding tree blocks are partitioned into slices including one or more coding tree blocks, the one or more coding tree blocks include one or more transform blocks according to a transform tree including a split_transform_flag indicative of the split of a given coding block into corresponding one or more transform blocks, the split_transform_flag is coded using CABAC, and a context index associated with the CABAC coding of the split_transform_flag is based on a value. Second circuitry of the encoder is to output the bitstream.
Abstract:
Systems, apparatus, articles, and methods are described including operations for size based transform unit context derivation. In an example encoder, first circuitry is to encode video input data into a bitstream according to a bitstream syntax, wherein the video input data includes one or more pictures, the one or more pictures are partitioned into one or more coding tree blocks, the one or more coding tree blocks are partitioned into slices including one or more coding tree blocks, the one or more coding tree blocks include one or more transform blocks according to a transform tree including a split_transform_flag indicative of the split of a given coding block into corresponding one or more transform blocks, the split_transform_flag is coded using CABAC, and a context index associated with the CABAC coding of the split_transform_flag is based on a value. Second circuitry of the encoder is to output the bitstream.
Abstract:
Methods, systems, and computer program products for the generation of multiple layers of scaled encoded video data compatible with the HEVC standard. Residue from prediction processing may be transformed into coefficients in the frequency domain. The coefficients may then be sampled to create a layer of encoded data. The coefficients may be sampled in different ways to create multiple respective layers. The layers may then be multiplexed and sent to a decoder. There, one or more of the layers may be chosen. The choice of certain layer(s) may be dependent on the desired attributes of the resulting video. A certain level of video quality, frame rate, resolution, and/or bit depth may be desired, for example. The coefficients in the chosen layers may then be assembled to create a version of the residue to be used in video decoding.