-
公开(公告)号:GB2573849B
公开(公告)日:2021-10-13
申请号:GB201901759
申请日:2019-02-08
Applicant: ADOBE INC
Inventor: ZHE LIN , XIAOHUI SHEN , MINGYANG LING , JIANMING ZHANG , JASON KUEN , BRETT BUTTERFIELD
IPC: G06F16/532 , G06F16/583 , G06K9/46 , G06K9/62
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for utilizing a deep neural network-based model to identify similar digital images for query digital images. For example, the disclosed systems utilize a deep neural network-based model to analyze query digital images to generate deep neural network-based representations of the query digital images. In addition, the disclosed systems can generate results of visually-similar digital images for the query digital images based on comparing the deep neural network-based representations with representations of candidate digital images. Furthermore, the disclosed systems can identify visually similar digital images based on user-defined attributes and image masks to emphasize specific attributes or portions of query digital images.
-
公开(公告)号:GB2571825A
公开(公告)日:2019-09-11
申请号:GB201900596
申请日:2019-01-16
Applicant: ADOBE INC
Inventor: ZHE LIN , YUFEI WANG , XIAOHUI SHEN , SCOTT D COHEN , JIANMING ZHANG
Abstract: Computer-implemented segmentation techniques are disclosed for identifying features in an image based on text tags. The method comprises: converting a tag 118 into a vector representation 202, the tag defining a semantic class 120 to be located in a digital image 106; generating an attention map 210 by an embedding neural network 208 based on the digital image and the vector representation, the attention map defining a location in the digital image that corresponds to the semantic class, where the embedding neural network has been trained using image-level tags of respective semantic classes; refining 212 the location of the semantic class in the attention map by a refinement neural network 214 , the refinement neural network trained using localized tags of respective semantic classes; and indicating the refined location of the semantic class in the digital image using the refined attention map 216. The method thus makes use of tags for similar items in a machine learning environment in order to provide more examples for training.
-
3.
公开(公告)号:GB2623402B
公开(公告)日:2025-04-23
申请号:GB202311871
申请日:2023-08-02
Applicant: ADOBE INC
Inventor: ZHE LIN , HAITIAN ZHENG , ELYA SHECHTMAN , JIANMING ZHANG , JINGWAN LU , NING XU , QING LIU , SCOTT COHEN , SOHRAB AMIRGHODSI
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for panoptically guiding digital image inpainting utilizing a panoptic inpainting neural network. In some embodiments, the disclosed systems utilize a panoptic inpainting neural network to generate an inpainted digital image according to panoptic segmentation map that defines pixel regions corresponding to different panoptic labels. In some cases, the disclosed systems train a neural network utilizing a semantic discriminator that facilitates generation of digital images that are realistic while also conforming to a semantic segmentation. The disclosed systems generate and provide a panoptic inpainting interface to facilitate user interaction for inpainting digital images. In certain embodiments, the disclosed systems iteratively update an inpainted digital image based on changes to a panoptic segmentation map.
-
公开(公告)号:GB2626398B
公开(公告)日:2025-02-19
申请号:GB202317072
申请日:2023-11-07
Applicant: ADOBE INC
Inventor: JONATHAN BRANDT , SCOTT COHEN , ZHE LIN , ZHIHONG DING , DARSHAN G PRASAD , MATTHEW JOSS , CELSO GOMES , JIANMING ZHANG , OLENA SOROKA , KLAAS STOECKMANN , MICHAEL ZIMMERMANN , THOMAS MUEHRKE
Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that modify digital images via scene-based editing using image understanding facilitated by artificial intelligence. For instance, in one or more embodiments, the disclosed systems generate utilizing a segmentation neural network, an object mask for each object of a plurality of objects of a digital image. The disclosed systems detect a first user interaction with an object in the digital image displayed via a graphical user interface. The disclosed systems surface, via the graphical user interface, the object mask for the object in response to the first user interaction. The disclosed systems perform an object-aware modification of the digital image in response to a second user interaction with the object mask for the object.
-
公开(公告)号:GB2623162B
公开(公告)日:2025-01-01
申请号:GB202311936
申请日:2023-08-03
Applicant: ADOBE INC
Inventor: ZHE LIN , HAITIAN ZHENG , ELYA SHECHTMAN , JIANMING ZHANG , JINGWAN LU , NING XU , QING LIU , SCOTT COHEN , SOHRAB AMIRGHODSI
Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for panoptically guiding digital image inpainting utilizing a panoptic inpainting neural network. In some embodiments, the disclosed systems utilize a panoptic inpainting neural network to generate an inpainted digital image according to panoptic segmentation map that defines pixel regions corresponding to different panoptic labels. In some cases, the disclosed systems train a neural network utilizing a semantic discriminator that facilitates generation of digital images that are realistic while also conforming to a semantic segmentation. The disclosed systems generate and provide a panoptic inpainting interface to facilitate user interaction for inpainting digital images. In certain embodiments, the disclosed systems iteratively update an inpainted digital image based on changes to a panoptic segmentation map.
-
公开(公告)号:GB2578950B
公开(公告)日:2021-03-10
申请号:GB201912054
申请日:2019-08-22
Applicant: ADOBE INC
Inventor: ZHE LIN , XIAOHUI SHEN , MINGYANG LING , JIANMING ZHANG , JASON KUEN
Abstract: In implementations of object detection in images, object detectors are trained using heterogeneous training datasets. A first training dataset is used to train an image tagging network to determine an attention map of an input image for a target concept. A second training dataset is used to train a conditional detection network that accepts as conditional inputs the attention map and a word embedding of the target concept. Despite the conditional detection network being trained with a training dataset having a small number of seen classes (e.g., classes in a training dataset), it generalizes to novel, unseen classes by concept conditioning, since the target concept propagates through the conditional detection network via the conditional inputs, thus influencing classification and region proposal. Hence, classes of objects that can be detected are expanded, without the need to scale training databases to include additional classes.
-
公开(公告)号:GB2571825B
公开(公告)日:2021-01-13
申请号:GB201900596
申请日:2019-01-16
Applicant: ADOBE INC
Inventor: JIANMING ZHANG , ZHE LIN , YUFEI WANG , XIAOHUI SHEN , SCOTT D COHEN
Abstract: Semantic segmentation techniques and systems are described that overcome the challenges of limited availability of training data to describe the potentially millions of tags that may be used to describe semantic classes in digital images. In one example, the techniques are configured to train neural networks to leverage different types of training datasets using sequential neural networks and use of vector representations to represent the different semantic classes.
-
公开(公告)号:GB2587841B
公开(公告)日:2022-09-21
申请号:GB202007986
申请日:2020-05-28
Applicant: ADOBE INC
Inventor: ZHE LIN , JIANMING ZHANG , HE ZHANG , FEDERICO PERAZZI
Abstract: The present disclosure relates to utilizing a neural network having a two-stream encoder architecture to accurately generate composite digital images that realistically portray a foreground object from one digital image against a scene from another digital image. For example, the disclosed systems can utilize a foreground encoder of the neural network to identify features from a foreground image and further utilize a background encoder to identify features from a background image. The disclosed systems can then utilize a decoder to fuse the features together and generate a composite digital image. The disclosed systems can train the neural network utilizing an easy-to-hard data augmentation scheme implemented via self-teaching. The disclosed systems can further incorporate the neural network within an end-to-end framework for automation of the image composition process.
-
公开(公告)号:GB2578950A
公开(公告)日:2020-06-03
申请号:GB201912054
申请日:2019-08-22
Applicant: ADOBE INC
Inventor: ZHE LIN , XIAOHUI SHEN , MINGYANG LING , JIANMING ZHANG , JASON KUEN
Abstract: In implementations of object detection in images, object detectors are trained using heterogeneous training datasets. A first training dataset is used to train an image tagging network 202 to determine an attention map 208 of an input image for a target concept. A second training dataset is used to train a conditional detection network that accepts as conditional inputs the attention map and a word-embedding of the target concept. Despite the conditional detection network being trained with a training dataset having a small number of seen classes (e.g., classes in a training dataset), it generalizes to novel, unseen classes by concept conditioning, since the target concept propagates through the conditional detection network via cite conditional inputs, thus influencing classification and region proposal. Hence, classes of objects that can be detected are expanded without the need to scale training databases to include additional classes.
-
公开(公告)号:GB2624754B
公开(公告)日:2025-02-05
申请号:GB202314582
申请日:2023-09-22
Applicant: ADOBE INC
Inventor: HAITIAN ZHENG , ZHE LIN , CONNELLY STUART BARNES , ELYA SHECHTMAN , JINGWAN LU , QING LIU , SOHRAB AMIRGHODSI , YUQIAN ZHOU , SCOTT COHEN , JIANMING ZHANG
IPC: G06T5/77
Abstract: Systems and methods for image processing are described. Embodiments of the present disclosure receive an image comprising a first region that includes content and a second region to be inpainted. Noise is then added to the image to obtain a noisy image, and a plurality of intermediate output images are generated based on the noisy image using a diffusion model trained using a perceptual loss. The intermediate output images predict a final output image based on a corresponding intermediate noise level of the diffusion model. The diffusion model then generates the final output image based on the intermediate output image. The final output image includes inpainted content in the second region that is consistent with the content in the first region.
-
-
-
-
-
-
-
-
-