DEPTH COMPLETION USING ATTENTION-BASED REFINEMENT OF FEATURES

    公开(公告)号:US20250148628A1

    公开(公告)日:2025-05-08

    申请号:US18633302

    申请日:2024-04-11

    Abstract: Systems and techniques are provided for generating depth information from one or more images. For example, a process can include obtaining a first depth map corresponding to an input comprising an image of the one or more images and a sparse depth measurement. A three-dimensional (3D) point cloud can be generated based on the first depth map and multi-scale visual features of the input, wherein the 3D point cloud includes a plurality of 3D point features uplifted from the multi-scale visual features. At least a portion of the plurality of 3D point features can be processed using one or more self-attention layers to generate refined 3D point features. A two-dimensional (2D) projection of the refined 3D point features can be generated and a second depth map can be generated based on the 2D projection of the refined 3D point features.

    ATTENTION-BASED REFINEMENT FOR DEPTH COMPLETION

    公开(公告)号:US20250054168A1

    公开(公告)日:2025-02-13

    申请号:US18448845

    申请日:2023-08-11

    Abstract: A processor-implemented method for attention-based depth completion includes receiving, by an artificial neural network (ANN), an input. The input includes an image and a sparse depth measurement. The ANN extracts multi-scale visual features of the input. The ANN applies a self-attention mechanism to the multi-scale visual features to generate a set of attended multi-scale visual features. The ANN generates a dense depth map based on the set of attended multi-scale visual features.

    DISPARITY-BASED DEPTH REFINEMENT USING CONFIDENCE INFORMATION AND STEREOSCOPIC DEPTH INFORMATION

    公开(公告)号:US20240404093A1

    公开(公告)日:2024-12-05

    申请号:US18327380

    申请日:2023-06-01

    Abstract: Systems and techniques are provided for generating disparity information from two or more images. For example, a process can include obtaining first disparity information corresponding to a pair of images, the pair of images including a first image of a scene and a second image of the scene. The process can include obtaining confidence information associated with the first disparity information. The process can include processing, using a machine learning network, the first disparity information and the confidence information to generate second disparity information corresponding to the pair of images. The process can include combining, based on the confidence information, the first disparity information with the second disparity information to generate a refined disparity map corresponding to the pair of images.

    PHYSICALLY-BASED EMITTER ESTIMATION FOR INDOOR SCENES

    公开(公告)号:US20240303913A1

    公开(公告)日:2024-09-12

    申请号:US18180797

    申请日:2023-03-08

    CPC classification number: G06T15/506 G06T7/593

    Abstract: Systems and techniques are provided for physical-based light estimation for inverse rendering of indoor scenes. For example, a computing device can obtain an estimated scene geometry based on a multi-view observation of a scene. The computing device can further obtain a light emission mask based on the multi-view observation of the scene. The computing device can also obtain an emitted radiance field based on the multi-view observation of the scene. The computing device can then determine, based on the light emission mask and the emitted radiance field, a geometry of at least one light source of the estimated scene geometry.

Patent Agency Ranking