Abstract:
A system and method for enabling parallel decoder implementation for different views, even when there are existing dependencies between views. In various embodiments of the present invention, information is signaled to a decoder that slices are coded using certain constraints, so that parallel decoding of slices is possible. This signaling can be performed at a sequence parameter set level, or it could be performed at picture parameter set level, slice header level or macroblock header level. Additionally, the delay between various views is also signaled to the decoder, at the same location as the constraint information in various embodiments. Various algorithms can also be used to improve the coding efficiency of the system.
Abstract:
Joint coding of depth map video and texture video is provided, where a motion vector for a texture video is predicted from a respective motion vector of a depth map video or vice versa. For scalable video coding, depth map video is coded as a base layer and texture video is coded as an enhancement layer(s). Inter-layer motion prediction predicts motion in texture video from motion in depth map video. With more than one view in a bit stream (for multi view coding), depth map videos are considered monochromatic camera views and are predicted from each other. If joint multi-view video model coding tools are allowed, inter-view motion skip issued to predict motion vectors of texture images from depth map images. Furthermore, scalable multi-view coding is utilized, where interview prediction is applied between views in the same dependency layer, and inter-layer (motion) prediction is applied between layers in the same view.
Abstract:
The time until a media decoder is refreshed to produce correct output samples is minimized if an IDR picture is the first picture in a decoding order in each MPE-FEC frame. A system enabling minimization of tune-in delay is described. Aspects of the system include a content encoder for receiving a source signal including first and second bit streams and for encoding the first and second bit streams, the encoded first bit stream including a decoder refresh bit stream and the encoded second bit stream including a media bit stream. A server receives the encoded first and second bit streams and encapsulates the encoded first and second bit streams into packets. An encapsulator then receives the packets and composes a time-sliced multi-protocol encapsultation (MPE) stream by replacing a spliceable picture from the encoded second bit stream.
Abstract:
In a start code based bistream structure for use in video coding, a picture header is used simultaneously with the parameter set in the start code. The picture header is present before the slice header in the same picture. As such, it is possible to allow the presence of at least one sequence parameter set or picture parameter set before the slice header. In a NAL unit parameter set bitstream structure, the picture head NAL unit is present before the slice NAL unit. In either structure, the picture header information must remain unchanged during the decoding of the entire picture, but should change more frequently between pictures as compared to the information in the picture parameter set.
Abstract:
A method of encoding a plurality of views of a scene comprises constructing an initial reference picture list based at least in part on intra-view reference pictures and inter-view reference pictures; and providing a signalling element for reordering of inter- view reference pictures relative to the initial reference picture list, the signalling element being derived based at least in part on a view identifier value.
Abstract:
"método de codificação/decodificação de um sinal de vídeo representando uma seqüência de imagens, e, codificador/decodificador de vídeo". método para codificar um sinal de vídeo representando uma sequência de imagens, o método compreendendo receber a imagem atual para codificar, formando uma predição temporal da imagem atual de uma imagem de referência default para a imagem atual, comparando a imagem de referência default com ao menos uma outra imagem de referência, calculando a medida de similaridade entre a imagem de referência default e cada outra imagem de referência e, se a medida de similaridade encontra um critério pré-determinado, produzir um indicador identificando a outra imagem de referência.