Abstract:
A method for edge correction of images of a three-dimensional video content, the video content including at least one original view image and at least one depth or disparity map, the method including the following steps: detecting edges in at least one original view image for obtaining original edges; warping the original edges according to the depth or disparity map; detecting a set of warped edges altered by the warping process; and correcting the altered edges for obtaining corrected edges.
Abstract:
A method for edge correction of images of a three-dimensional video content, the video content including at least one original view image and at least one depth or disparity map, the method including the following steps: detecting edges in at least one original view image for obtaining original edges; warping the original edges according to the depth or disparity map; detecting a set of warped edges altered by the warping process; and correcting the altered edges for obtaining corrected edges.
Abstract:
A method is described for generating a color image composed of a plurality of components (Y, U, V) by starting from at least one depth or disparity map (DM1, DM2), wherein a first set of pixels of said at least one depth or disparity map (DM1, DM2) is entered into the luminance component (Y) of said color image, and wherein a second and a third sets of pixels of said at least one depth or disparity map (DM1, DM2) are entered into the two chrominance components (U, V) of said color image.
Abstract:
A method is described for generating a colour image composed of a plurality of components (Y, U, V) by starting from at least one depth or disparity map (DM1, DM2), wherein a first set of pixels of said at least one depth or disparity map (DM1, DM2) is entered into the luminance component (Y) of said colour image, and wherein a second and a third sets of pixels of said at least one depth or disparity map (DM1, DM2) are entered into the two chrominance components (U, V) of said colour image.
Abstract:
A method for learned image compression implemented in an autoencoder including a learnable encoder and a decoder, the method including: a) extracting from an image a latent space by the learnable encoder; b) quantizing the latent space by a quantizer to obtain a quantized latent space; c) entropy coding the quantized latent space by an entropy encoder to obtain a bitstream, wherein an entropy model used to encode the latent space is represented by a probability distribution; d) entropy decoding the bitstream by an entropy decoder to obtain an entropy decoded bitstream; e) feeding the entropy decoded bitstream to the decoder; f) recover a reconstructed image by the decoder; g) training the autoencoder via standard gradient descent of the backpropagated error gradient by finding learnable parameters of the learnable encoder and of the decoder that minimize a rate distortion cost function, wherein the entropy encoder is based on a differentiable formulation of a soft frequency counter.
Abstract:
A method for encoding digital images or video streams, includes a receiving phase, wherein a portion of an image is received; a graph weights prediction phase, wherein the elements of a weights matrix associated to the graph related to the blocks of the image (predicted blocks) are predicted on the basis of reconstructed, de-quantized and inverse-transformed pixel values of at least one previously coded block (predictor block) of the image, the weights matrix being a matrix comprising elements denoting the level of similarity between a pair of pixels composing said image, a graph transform computation phase, wherein the graph Fourier transform of the blocks of the image is performed, obtaining for the blocks a set of coefficients determined on the basis of the predicted weights; a coefficients quantization phase, wherein the coefficients are quantized an output phase wherein a bitstream comprising the transformed and quantized coefficients is transmitted and/or stored.
Abstract:
The present invention relates to a method for encoding the borders of pixel regions of an image, wherein the borders contain a sequence of vertices subdividing the image into regions of pixels (superpixels), by generating a sequence of symbols from an alphabet including the step of: defining for each superpixel a first vertex for coding the borders of the superpixel according to a criterion common to all superpixels; defining for each superpixel the same coding order of the border vertices, either clockwise or counter-clockwise; defining the order for coding the superpixels on the base of a common rule depending on the relative positions of the first vertices; defining a set of vertices as a known border, wherein the following steps are performed for selecting a symbol of the alphabet, for encoding the borders of the superpixels: a) determining the first vertex of the next superpixel border individuated by the common criterion; b) determining the next vertex to be encoded on the basis of the coding direction; c) selecting a first symbol (“0”) for encoding the next vertex if the next vertex of a border pertains to the known border, d) selecting a symbol (“1”; “2”) different from the first symbol (“0”) if the next vertex is not in the known border; e) repeating steps b), c), d) and e) until all vertices of the superpixel border have been encoded; f) adding each vertex of the superpixel border that was not in the known border to the set; g) determining the next superpixel whose border is to be encoded according to the common rule, if any; i) repeating steps a)-g) until the borders of all the superpixels of the image have being added to the known border.
Abstract:
A method and an apparatus for encoding and/or decoding digital images or video streams are provided, wherein the encoding apparatus includes a processor configured for reading at least a portion of the image, segmenting the portion of the image in order to obtain groups of pixels identified by borders information and containing at least two pixels having one or more homogeneous characteristics, computing, for each group of pixels, a weight map on the basis of the borders information associated to the group of pixels, a graph transform matrix on the basis of the weight map, and transform coefficients on the basis of the graph transform matrix (U) and of the pixels contained in the group of pixels.