Generative image filling using a reference image

    公开(公告)号:GB2631163A

    公开(公告)日:2024-12-25

    申请号:GB202407721

    申请日:2024-05-31

    Applicant: ADOBE INC

    Abstract: This application discloses a method comprising: obtaining an input image, a reference image, and a text prompt; encoding, using an image encoder, the reference image to obtain an image embedding; encoding, using a text encoder, the text prompt to obtain a text embedding; and generating, using an image generation model, a composite image based on the input image, the image embedding, and the text embedding. The composite image may depict the input image with a region filled with the content or style from the reference image. The image and text embeddings may be used as guidance for the image generation model. A mask may be received from a user. The image may be generated based on the mask. The text prompt may describe an object in the reference image. The reference image may comprise a portion or a style from the input image. The image generation model may be a diffusion model comprising a self-attention layer. The image and text encoders may be multimodal encoders. Also disclosed is an apparatus to implement the method.

    Generating modified digital images using deep visual guided patch match models for image inpainting

    公开(公告)号:GB2606836A

    公开(公告)日:2022-11-23

    申请号:GB202203239

    申请日:2022-03-09

    Applicant: ADOBE INC

    Abstract: The present disclosure relates to systems, methods, and computer readable media for accurately, efficiently, and flexibly generating modified digital images utilizing a guided inpainting approach that implements a patch match model informed by a deep visual guide. In particular, the disclosed systems can utilize a visual guide algorithm to automatically generate guidance maps to help identify replacement pixels for inpainting regions of digital images utilizing a patch match model. For example, the disclosed systems can generate guidance maps in the form of structure maps, depth maps, or segmentation maps that respectively indicate the structure, depth, or segmentation of different portions of digital images. Additionally, the disclosed systems can implement a patch match model to identify replacement pixels for filling regions of digital images according to the structure, depth, and/or segmentation of the digital images.

    Generative image congealing
    3.
    发明专利

    公开(公告)号:GB2601220A

    公开(公告)日:2022-05-25

    申请号:GB202113083

    申请日:2021-09-14

    Applicant: ADOBE INC

    Abstract: A method of training a generator network 1012 to produce synthetic images comprises receiving an unaligned set of real image data at a spatial transformer 1010 and aligning the real image data set and providing at least one image of the set to a discriminator network 1014. Synthetic images are produced by the generator network and provided to the discriminator network. A training manager 1004 is then used to adversarially train the generator network. The training of the generator is performed using the output of the discriminator network, which determines whether the synthetic image(s) resemble the real image(s). The method may further comprise adversarial training of the spatial transformer using the training manager, the generator, and the discriminator, where the generator produces synthetic aligned images for comparison with the real aligned images produced by the spatial transformer. An image congealing system which performs the disclosed method is also disclosed.

    Video inpainting via confidence-weighted motion estimation

    公开(公告)号:GB2578354B

    公开(公告)日:2021-12-29

    申请号:GB201911506

    申请日:2019-08-12

    Applicant: ADOBE INC

    Abstract: A method of accessing a video having a target region 306, updating the video content in the target region based on confidence-weighted motion estimation for the target region, and presenting the updated video content on a display device. In particular, the video has a scene comprising a first 112a and a second 112b frame, the scene having an annotation identifying a target region to be modified in one or more video frames. A boundary motion for a boundary of the target region in the scene is computed, the boundary including boundary pixels 404a, 405a neighbouring the target region. Confidence values are assigned to the boundary pixels, wherein a confidence value is based on: a difference between a forward and reverse motion with respect to a particular boundary pixel, and/or a texture in a region that includes the particular boundary pixel. A target motion 412 of a target pixel is interpolated from the boundary motion, wherein the confidence value of the boundary pixel controls a contribution of a motion of the boundary pixel to the target motion. The colour of the target pixel is updated to correspond to the interpolated target motion.

    Automatic synthesis of a content-aware sampling region for a content-aware fill

    公开(公告)号:GB2589389A

    公开(公告)日:2021-06-02

    申请号:GB202003400

    申请日:2020-03-09

    Applicant: ADOBE INC

    Abstract: Embodiments of the present disclosure provide systems, methods, and computer storage media for automatically synthesizing a content-aware sampling region for a hole-filling algorithm such as content-aware fill. Given a source image and a hole (or other target region to fill), a sampling region can be synthesized by identifying a band of pixels surrounding the hole, clustering these pixels based on one or more characteristics (e.g., colour, x/y coordinates, depth, focus, etc.), passing each of the resulting clusters as foreground pixels to a segmentation algorithm, and unioning the resulting pixels to form the sampling region. The sampling region can be stored in a constraint mask and passed to a hole-filling algorithm such as content-aware fill to synthesize a fill for the hole (or other target region) from patches sampled from the synthesized sampling region.

    Interactive system for automatically synthesizing a content-aware fill

    公开(公告)号:GB2580596A

    公开(公告)日:2020-07-29

    申请号:GB201900485

    申请日:2019-01-14

    Applicant: ADOBE INC

    Abstract: Embodiments of the present disclosure provide systems, methods, and computer storage media for automatically synthesizing a content-aware fill using similarity transformed patches. A user interface receives a user-specified hole and a user-specified sampling region, both of which may be stored in a constraint mask. A brush tool can be used to interactively brush the sampling region and modify the constraint mask. The mask is passed to a patch-based synthesizer configured to synthesize the fill using similarity transformed patches sampled from the sampling region. Fill properties such as similarity transform parameters can be set to control the manner in which the fill is synthesized. A live preview can be provided with gradual updates of the synthesized fill prior to completion. Once a fill has been synthesized, the user interface presents the original image, replacing the hole with the synthesized fill.

    Learning parameters for neural networks using a semantic discriminator and an object-level discriminator

    公开(公告)号:GB2623162B

    公开(公告)日:2025-01-01

    申请号:GB202311936

    申请日:2023-08-03

    Applicant: ADOBE INC

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for panoptically guiding digital image inpainting utilizing a panoptic inpainting neural network. In some embodiments, the disclosed systems utilize a panoptic inpainting neural network to generate an inpainted digital image according to panoptic segmentation map that defines pixel regions corresponding to different panoptic labels. In some cases, the disclosed systems train a neural network utilizing a semantic discriminator that facilitates generation of digital images that are realistic while also conforming to a semantic segmentation. The disclosed systems generate and provide a panoptic inpainting interface to facilitate user interaction for inpainting digital images. In certain embodiments, the disclosed systems iteratively update an inpainted digital image based on changes to a panoptic segmentation map.

    Utilizing a critical edge detection neural network and a geometric model to determine camera parameters from a single digital image

    公开(公告)号:GB2585396A

    公开(公告)日:2021-01-13

    申请号:GB201916984

    申请日:2019-11-21

    Applicant: ADOBE INC

    Abstract: Th present disclosure relates to a system, computer-readable media and method for utilizing a critical edge detection neural network and a geometric model to determine camera parameter from a single digital image 802. In particular, in one or more embodiments, the disclosed systems can train and utilize a critical edge detection neural network to generate a vanishing edge map indicating vanishing lines from the digital image 804. The system can then utilize the vanishing edge map to determine camera parameters by applying a geometric model to the vanishing edge map 806. Further, the system can generate ground truth vanishing line data from a set of training digital images for training the critical edge detection neural network (Fig.9, 904). A critical edge detection neural network is trained by determining the distances between a vanishing point and a training line (Fig 9. 906). Predicted vanishing lines are generated to modify and train the neural network. The neural network may be a convolutional neural network (CNN) (Fig.9, 912). The camera parameters may be pitch, yaw, roll or focal length and the edge map may contain weights or confidence values.

Patent Agency Ranking