SHARED VISION SYSTEM BACKBONE
    21.
    发明申请

    公开(公告)号:US20250037478A1

    公开(公告)日:2025-01-30

    申请号:US18917905

    申请日:2024-10-16

    Abstract: A method for generating a dense light detection and ranging (LiDAR) representation by a vision system of a vehicle includes generating, at a depth estimation network, a depth estimate of an environment depicted in an image captured by an image capturing sensor integrated with the vehicle. The method also includes generating, via a sparse depth network, one or more sparse depth estimates of the environment, each sparse depth estimate associated with a respective sparse representation of one or more sparse representations. The method further includes generating the dense LiDAR representation based on a dense depth estimate that is generated based on the depth estimate and the one or more sparse depth estimates. The method still further includes controlling an action of the vehicle based on the dense LiDAR representation.

    SYSTEM AND METHOD FOR ESTIMATING DEPTH UNCERTAINTY FOR SELF-SUPERVISED 3D RECONSTRUCTION

    公开(公告)号:US20220262068A1

    公开(公告)日:2022-08-18

    申请号:US17734899

    申请日:2022-05-02

    Abstract: A method for three-dimensional (3D) scene reconstruction by an agent includes estimating an ego-motion of the agent based on a current image from a sequence of images and a previous image from the sequence of images. Each image in the sequence of images may be a two-dimensional (2D) image. The method also includes estimating a depth of the current image via a depth estimation model comprising a group of encoder layers and a group of decoder layers. The method further includes generating a 3D reconstruction of the current image based on the estimated ego-motion and the estimated depth. The method still further includes controlling an action of the agent based on the 3D reconstruction.

    CAMERA AGNOSTIC DEPTH NETWORK
    25.
    发明申请

    公开(公告)号:US20220148206A1

    公开(公告)日:2022-05-12

    申请号:US17581743

    申请日:2022-01-21

    Abstract: A method for monocular depth/pose estimation in a camera agnostic network is described. The method includes projecting lifted 3D points onto an image plane according to a predicted ray vector based on a monocular depth model, a monocular pose model, and a camera center of a camera agnostic network. The method also includes predicting a warped target image from a predicted depth map of the monocular depth model, a ray surface of the predicted ray vector, and a projection of the lifted 3D points according to the camera agnostic network.

    PIXEL-WISE RESIDUAL POSE ESTIMATION FOR MONOCULAR DEPTH ESTIMATION

    公开(公告)号:US20210398302A1

    公开(公告)日:2021-12-23

    申请号:US16908442

    申请日:2020-06-22

    Abstract: A method for scene reconstruction includes generating a depth estimate and a first pose estimate from a current image. The method also includes generating a second pose estimate based on the current image and one or more previous images in a sequence of images. The method further includes generating a warped image by warping each pixel in the current image based on the depth estimate, the first pose estimate, and the second pose estimate. The method still further includes controlling an action of an agent based on the second warped image.

    KEYPOINT MATCHING USING GRAPH CONVOLUTIONS

    公开(公告)号:US20210326601A1

    公开(公告)日:2021-10-21

    申请号:US17231905

    申请日:2021-04-15

    Abstract: A method for keypoint matching includes determining a first set of keypoints corresponding to a current environment of the agent. The method further includes determining a second set of keypoints from a pre-built map of the current environment. The method still further includes identifying matching pairs of keypoints from the first set of keypoints and the second set of keypoints based on geometrical similarities between respective keypoints of the first set of keypoints and the second set of keypoints. The method also includes determining a current location of the agent based on the identified matching pairs of keypoints. The method further includes controlling an action of the agent based on the current location.

    ADVERSARIAL OBJECT-AWARE NEURAL SCENE RENDERING FOR 3D OBJECT DETECTION

    公开(公告)号:US20240135721A1

    公开(公告)日:2024-04-25

    申请号:US17964827

    申请日:2022-10-12

    CPC classification number: G06V20/58 G06T7/70 G06V10/82

    Abstract: A method for improving 3D object detection via object-level augmentations is described. The method includes recognizing, using an image recognition model of a differentiable data generation pipeline, an object in an image of a scene. The method also includes generating, using a 3D reconstruction model, a 3D reconstruction of the scene from the image including the recognized object. The method further includes manipulating, using an object level augmentation model, a random property of the object by a random magnitude at an object level to determine a set of properties and a set of magnitudes of an object manipulation that maximizes a loss function of the image recognition model. The method also includes training a downstream task network based on a set of training data generated based on the set of properties and the set of magnitudes of the object manipulation, such that the loss function is minimized.

Patent Agency Ranking