Abstract:
A three-dimensional (3D) video codec encodes multiple views of a 3D video, each including texture and depth components. The encoders of the codec encode video blocks of their respective views based on a set of prediction parameters, such as quad-tree split flags, prediction modes, partition sizes, motion fields, inter directions, reference indices, luma intra modes, and chroma intra modes. The prediction parameters may be inherited across different views and different ones of the texture and depth components.
Abstract:
Systems, apparatus and methods are described related to real-time automatic conversion of 2-dimensional images or video to 3-dimensional stereo images or video.
Abstract:
Systems, devices and methods are described including performing scalable video coding using inter-layer pixel sample prediction. Inter-layer pixel sample prediction in an enhancement layer coding unit, prediction unit, or transform unit may use reconstructed pixel samples obtained from a base layer or from a lower enhancement layer. The pixel samples may be subjected to upsample filtering and/or refinement filtering. The upsample or refinement filter coefficients may be predetermined or may be adaptively determined.
Abstract:
Systems, apparatus, articles, and methods are described including operations for size based transform unit context derivation. In an example encoder, first circuitry is to encode video input data into a bitstream according to a bitstream syntax, wherein the video input data includes one or more pictures, the one or more pictures are partitioned into one or more coding tree blocks, the one or more coding tree blocks are partitioned into slices including one or more coding tree blocks, the one or more coding tree blocks include one or more transform blocks according to a transform tree including a split_transform_flag indicative of the split of a given coding block into corresponding one or more transform blocks, the split_transform_flag is coded using CABAC, and a context index associated with the CABAC coding of the split_transform_flag is based on a value. Second circuitry of the encoder is to output the bitstream.
Abstract:
Techniques related to accelerated video enhancement using deep learning selectively applied based on video codec information are discussed. Such techniques include applying a deep learning video enhancement network selectively to decoded non-skip blocks that are in low quantization parameter frames, bypassing the deep learning network for decoded skip blocks in low quantization parameter frames, and applying non-deep learning video enhancement to high quantization parameter frames.
Abstract:
An apparatus for edge aware upscaling is described herein. The apparatus comprises a potential edge detector, a thin-edge detector, a one-directional edge detector, a correlation detector, and a corrector. The potential edge detector identifies potential edge pixels in an input image, and the thin-edge detector detects thin edges in the potential edge pixels of the input image. The one-directional edge detector detects one-directional edges in the potential edge pixels of the input image, and the correlation detector detects strongly correlated edges in the potential edge pixels of the input image. The corrector derives a target output value based on an edge type and classification of a corresponding input pixel as identified by a source map point.
Abstract:
An apparatus for super resolution imaging includes a convolutional neural network (104) to receive a low resolution frame (102) and generate a high resolution illuminance component frame. The apparatus also includes a hardware scaler (106) to receive the low resolution frame (102) and generate a second high resolution chrominance component frame. The apparatus further includes a combiner (108) to combine the high resolution illuminance component frame and the high resolution chrominance component frame to generate a high resolution frame (110).
Abstract:
Systems, apparatus, articles, and methods are described including operations for size based transform unit context derivation. In an example encoder, first circuitry is to encode video input data into a bitstream according to a bitstream syntax, wherein the video input data includes one or more pictures, the one or more pictures are partitioned into one or more coding tree blocks, the one or more coding tree blocks are partitioned into slices including one or more coding tree blocks, the one or more coding tree blocks include one or more transform blocks according to a transform tree including a split_transform_flag indicative of the split of a given coding block into corresponding one or more transform blocks, the split_transform_flag is coded using CABAC, and a context index associated with the CABAC coding of the split_transform_flag is based on a value. Second circuitry of the encoder is to output the bitstream.
Abstract:
Systems, apparatus, articles, and methods are described below including operations for scalable real-time face beautification of video images.