-
公开(公告)号:US20210398301A1
公开(公告)日:2021-12-23
申请号:US16904444
申请日:2020-06-17
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Vitor GUIZILINI , Sudeep PILLAI , Adrien David GAIDON , Rares A. AMBRUS , Igor VASILJEVIC
Abstract: A method for monocular depth/pose estimation in a camera agnostic network is described. The method includes training a monocular depth model and a monocular pose model to learn monocular depth estimation and monocular pose estimation based on a target image and context images from monocular video captured by the camera agnostic network. The method also includes lifting 3D points from image pixels of the target image according to the context images. The method further includes projecting the lifted 3D points onto an image plane according to a predicted ray vector based on the monocular depth model, the monocular pose model, and a camera center of the camera agnostic network. The method also includes predicting a warped target image from a predicted depth map of the monocular depth model, a ray surface of the predicted ray vector, and a projection of the lifted 3D points according to the camera agnostic network.
-
公开(公告)号:US20240320844A1
公开(公告)日:2024-09-26
申请号:US18734906
申请日:2024-06-05
Inventor: Vitor GUIZILINI , Rares Andrei AMBRUS , Igor VASILJEVIC , Gregory SHAKHNAROVICH
IPC: G06T7/55 , B60R1/27 , B60W60/00 , G05D1/248 , G05D1/646 , G06F18/214 , G06N3/08 , G06T3/04 , G06T3/18 , G06T3/40 , G06T3/4046 , G06T7/11 , G06T7/292 , G06T7/579 , G06V20/56 , H04N23/90
CPC classification number: G06T7/55 , B60R1/27 , B60W60/001 , G05D1/248 , G05D1/646 , G06F18/214 , G06F18/2148 , G06N3/08 , G06T3/04 , G06T3/18 , G06T3/40 , G06T3/4046 , G06T7/11 , G06T7/292 , G06T7/579 , G06V20/56 , H04N23/90 , B60R2300/102 , B60W2420/403 , G06T2207/10028 , G06T2207/20081 , G06T2207/20084 , G06T2207/30244 , G06T2207/30252
Abstract: A method for scale-aware depth estimation using multi-camera projection loss is described. The method includes determining a multi-camera photometric loss associated with a multi-camera rig of an ego vehicle. The method also includes training a scale-aware depth estimation model and an ego-motion estimation model according to the multi-camera photometric loss. The method further includes predicting a 360° point cloud of a scene surrounding the ego vehicle according to the scale-aware depth estimation model and the ego-motion estimation model. The method also includes planning a vehicle control action of the ego vehicle according to the 360° point cloud of the scene surrounding the ego vehicle.
-
公开(公告)号:US20240271959A1
公开(公告)日:2024-08-15
申请号:US18645240
申请日:2024-04-24
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Jiexiong TANG , Rares Andrei AMBRUS , Hanme KIM , Adrien David GAIDON , Vitor GUIZILINI , Xipeng WANG , Jeffrey WALLS , Sudeep PILLAI
IPC: G01C21/00
CPC classification number: G01C21/3837 , G01C21/3826 , G01C21/3896
Abstract: A method for labeling keypoints includes labeling, via a keypoint model, a first set of keypoints in a first image associated with a three-dimensional (3D) map of an environment, the first image having been captured during a first time period. The method also includes labeling, via the keypoint model, a second set of keypoints in a second image associated with the 3D map, the second image having been captured during a second time period. The method further includes updating, via the keypoint model, one of more of the first set of keypoints based on labeling the second set of keypoints.
-
公开(公告)号:US20240217538A1
公开(公告)日:2024-07-04
申请号:US18091872
申请日:2022-12-30
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Vitor GUIZILINI
CPC classification number: B60W60/001 , B60W40/02 , G06T7/521 , B60W2420/42 , G06T2207/10016 , G06T2207/10028 , G06T2207/20081 , G06T2207/30252
Abstract: A method estimating a depth of an environment includes generating, via a cross-attention model, a cross-attention cost volume based on a current image of the environment and a previous image of the environment in a sequence of images. The method also includes generating, via the cross-attention model, a depth estimate of the current image based on the cross-attention cost volume, the cross-attention model having been trained using a photometric loss associated with a single-frame depth estimation model. The method further includes controlling an action of the vehicle based on the depth estimate.
-
15.
公开(公告)号:US20210407115A1
公开(公告)日:2021-12-30
申请号:US16913214
申请日:2020-06-26
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Vitor GUIZILINI , Adrien David GAIDON , Rares A. AMBRUS
IPC: G06T7/50
Abstract: Systems and methods for generating depth models and depth maps from images obtained from an imaging system are presented. A self-supervised neural network may be capable of regularizing depth information from surface normals. Rather than rely on separate depth and surface normal networks, surface normal information is extracted from the depth information and a smoothness function is applied to the surface normals instead of a depth gradient. Smoothing the surface normal may provide improved representation of environmental structures by both smoothing texture-less areas while preserving sharp boundaries between structures.
-
公开(公告)号:US20210390718A1
公开(公告)日:2021-12-16
申请号:US16899425
申请日:2020-06-11
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Vitor GUIZILINI , Adrien David GAIDON
Abstract: A method for estimating depth is presented. The method includes generating, at each decoding layer of a neural network, decoded features of an input image. The method also includes upsampling, at each decoding layer, the decoded features to a resolution of a final output of the neural network. The method still further includes concatenating, at each decoding layer, the upsampled decoded features with features generated at a convolution layer of the neural network. The method additionally includes sequentially receiving the concatenated upsampled decoded features at a long-short term memory (LSTM) module of the neural network from each decoding layer. The method still further includes generating, at the LSTM module, a depth estimate of the input image after receiving the concatenated upsampled inverse depth estimate from a final layer of a decoder of the neural network. The method also includes controlling an action of an agent based on the depth estimate.
-
公开(公告)号:US20210319577A1
公开(公告)日:2021-10-14
申请号:US17230941
申请日:2021-04-14
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Jiexiong TANG , Rares Andrei AMBRUS , Vitor GUIZILINI , Adrien David GAIDON
Abstract: A method for depth estimation performed by a depth estimation system of an autonomous agent includes determining a first pose of a sensor based on a first image captured by the sensor and a second image captured by the sensor. The method also includes determining a first depth of the first image and a second depth of the second image. The method further includes generating a warped depth image based on at least the first depth and the first pose. The method still further includes determining a second pose based on the warped depth image and the second depth image. The method also includes updating the first pose based on the second pose and updating a first warped image based on the updated first pose.
-
公开(公告)号:US20210319236A1
公开(公告)日:2021-10-14
申请号:US17230947
申请日:2021-04-14
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Jiexiong TANG , Rares Andrei AMBRUS , Vitor GUIZILINI , Adrien David GAIDON
Abstract: A method for keypoint matching includes receiving an input image obtained by a sensor of an agent. The method also includes identifying a set of keypoints of the received image. The method further includes augmenting the descriptor of each of the keypoints with semantic information of the input image. The method also includes identifying a target image based on one or more semantically augmented descriptors of the target image matching one or more semantically augmented descriptors of the input image. The method further includes controlling an action of the agent in response to identifying the target.
-
公开(公告)号:US20210318140A1
公开(公告)日:2021-10-14
申请号:US17230942
申请日:2021-04-14
Applicant: TOYOTA RESEARCH INSTITUTE, INC.
Inventor: Jiexiong TANG , Rares Andrei AMBRUS , Hanme KIM , Vitor GUIZILINI , Adrien David GAIDON , Xipeng WANG , Jeff WALLS, SR. , Sudeep PILLAI
IPC: G01C21/00
Abstract: A method for localization performed by an agent includes receiving a query image of a current environment of the agent captured by a sensor integrated with the agent. The method also includes receiving a target image comprising a first set of keypoints matching a second set of keypoints of the query image. The first set of keypoints may be generated based on a task specified for the agent. The method still further includes determining a current location based on the target image.
-
公开(公告)号:US20250118094A1
公开(公告)日:2025-04-10
申请号:US18982985
申请日:2024-12-16
Applicant: TOYOTA RESEARCH INSTITUTE, INC. , THE BOARD OF TRUSTEES OF THE LELAND STANFORD JUNIOR UNIVERSITY
Inventor: Rares Andrei AMBRUS , Or LITANY , Vitor GUIZILINI , Leonidas GUIBAS , Adrien David GAIDON , Jie LI
Abstract: A method for 3D object detection is described. The method includes predicting, using a trained monocular depth network, an estimated monocular input depth map of a monocular image of a video stream and an estimated depth uncertainty map associated with the estimated monocular input depth map. The method also includes feeding back a depth uncertainty regression loss associated with the estimated monocular input depth map during training of the trained monocular depth network to update the estimated monocular input depth map. The method further includes detecting 3D objects from a 3D point cloud computed from the estimated monocular input depth map based on seed positions selected from the 3D point cloud and the estimated depth uncertainty map. The method also includes selecting 3D bounding boxes of the 3D objects detected from the 3D point cloud based on the seed positions and an aggregated depth uncertainty.
-
-
-
-
-
-
-
-
-