Abstract:
Described embodiments include methods and techniques for creating media container files. An exemplary method comprises steps wherein, instead of estimating the size of a media segment header based on the segment duration, identified media data boxes are used. In the method, an identified media data element is written in a container file. Included into the identified data element, an identifier is used as a reference for the identified media data of the segment by other elements. The identifier can be a first movie fragment sequence number, or a first track fragment decode time.
Abstract:
A method, apparatus and computer program product are provided to include information within a container that also includes a video bitstream as to whether individual image items corresponding to the video frames of the video bitstream bitstreams are self-decodable or, alternatively, are dependent upon one or more other image items. In an instance in which a respective image item is dependent upon one or more other image items, the method, apparatus and computer program product also include dependence information within the container identifying the other image item(s)upon which the decodability of the respective image item is dependent. As such, the method, apparatus and computer program product permit decoding relationships to be defined in the container between a predictively coded frame and other image item(s)upon which the predictively coded frame is dependent, thereby facilitating the decoding of the frame and, in turn, the video bitstream.
Abstract:
There are disclosed various methods, apparatuses and computer program products for video encoding and decoding. In some embodiments a bitstream comprising a coded first-view picture and a coded second-view picture is encoded or encapsulated. The coded second-view picture represents a smaller field of view than the coded first-view picture, wherein decoding of the coded first-view picture results in a decoded first-view picture, and decoding of the coded second-view picture results in a decoded second-view picture. An indication is inserted in or along the bitstream that a reconstructed second-view picture comprises the decoded second-view picture and at least one region of the decoded first-view picture, wherein the reconstructed second-view picture represents the same field of view as the decoded first-view picture.
Abstract:
Joint coding of depth map video and texture video is provided, where a motion vector for a texture video is predicted from a respective motion vector of a depth map video or vice versa. For scalable video coding, depth map video is coded as a base layer and texture video is coded as an enhancement layer(s). Inter-layer motion prediction predicts motion in texture video from motion in depth map video. With more than one view in a bitstream (for multiview coding), depth map videos are considered monochromatic camera views and are predicted from each other. If joint multiview video model coding tools are allowed, inter-view motion skip is used to predict motion vectors of texture images from depth map images. Furthermore, scalable multiview coding is utilized, where inter-view prediction is applied between views in the same dependency layer, and inter-layer (motion) prediction is applied between layers in the same view.
Abstract:
Joint coding of depth map video and texture video is provided, where a motion vector for a texture video is predicted from a respective motion vector of a depth map video or vice versa. For scalable video coding, depth map video is coded as a base layer and texture video is coded as an enhancement layer(s). Inter-layer motion prediction predicts motion in texture video from motion in depth map video. With more than one view in a bitstream (for multiview coding), depth map videos are considered monochromatic camera views and are predicted from each other. If joint multiview video model coding tools are allowed, inter-view motion skip is used to predict motion vectors of texture images from depth map images. Furthermore, scalable multiview coding is utilized, where inter-view prediction is applied between views in the same dependency layer, and inter-layer (motion) prediction is applied between layers in the same view.
Abstract:
A method, apparatus and computer readable medium are provided for facilitating encoding of media content. In the context of a method, information is included in a manifest of available media content. The method also indicates in the manifest one of the following: a viewport or a spatial region. The method further indicates in the manifest at least one value indicative of a respective quality of the viewport or the spatial region.
Abstract:
It is inter alia disclosed a method comprising: projecting a distance value and position associated with each of a plurality of pixels of an array of pixel sensors in an image sensor of a time of flight camera system onto a three dimensional world coordinate space as a plurality of depth sample points; and merging pixels from a two dimensional camera image with the plurality of depth sample points of the three dimensional world coordinate space to produce a fused two dimensional depth mapped image.