-
公开(公告)号:US20210325882A1
公开(公告)日:2021-10-21
申请号:US17363986
申请日:2021-06-30
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Mengye Ren , Andrei Pokrovsky , Bin Yang
IPC: G05D1/00 , G01S17/89 , G01S17/86 , G01S17/931
Abstract: The present disclosure provides systems and methods that apply neural networks such as, for example, convolutional neural networks, to sparse imagery in an improved manner. For example, the systems and methods of the present disclosure can be included in or otherwise leveraged by an autonomous vehicle. In one example, a computing system can extract one or more relevant portions from imagery, where the relevant portions are less than an entirety of the imagery. The computing system can provide the relevant portions of the imagery to a machine-learned convolutional neural network and receive at least one prediction from the machine-learned convolutional neural network based at least in part on the one or more relevant portions of the imagery. Thus, the computing system can skip performing convolutions over regions of the imagery where the imagery is sparse and/or regions of the imagery that are not relevant to the prediction being sought.
-
公开(公告)号:US10768628B2
公开(公告)日:2020-09-08
申请号:US15992498
申请日:2018-05-30
Applicant: UATC, LLC
Inventor: R. Lance Martin , Zac Vawter , Andrei Pokrovsky
Abstract: Systems and methods are directed to object detection at various ranges for autonomous vehicles. In one example, a system includes a camera providing a first field of view; a machine-learned model that has been trained to generate object detection range estimates based at least in part on labeled training data representing image data having a second field of view different from the first field of view; and a computing system including one or more processors; and memory including instructions that, when executed by the one or more processors, cause the one or more processors to perform operations. The operations include obtaining image data from the camera. inputting the image data from the camera to the machine-learned model; obtaining a first range estimate as an output of the machine-learned model, wherein the first range estimate represents estimates for the second field of view; generating transformed range estimate by applying a range estimate transform to the first range estimate output from the machine-learned model; and providing the transformed range estimate for use in controlling operation of an autonomous vehicle.
-
公开(公告)号:US11726208B2
公开(公告)日:2023-08-15
申请号:US16437827
申请日:2019-06-11
Applicant: UATC, LLC
Inventor: Shenlong Wang , Andrei Pokrovsky , Raquel Urtasun Sotil , Ioan Andrei Bârsan
IPC: G01S17/42 , G01S7/4861 , G05D1/00 , G01S19/51 , G05D1/02 , G01S17/931
CPC classification number: G01S17/42 , G01S7/4861 , G01S17/931 , G01S19/51 , G05D1/0088 , G05D1/024
Abstract: Aspects of the present disclosure involve systems, methods, and devices for autonomous vehicle localization using a Lidar intensity map. A system is configured to generate a map embedding using a first neural network and to generate an online Lidar intensity embedding using a second neural network. The map embedding is based on input map data comprising a Lidar intensity map, and the Lidar sweep embedding is based on online Lidar sweep data. The system is further configured to generate multiple pose candidates based on the online Lidar intensity embedding and compute a three-dimensional (3D) score map comprising a match score for each pose candidate that indicates a similarity between the pose candidate and the map embedding. The system is further configured to determine a pose of a vehicle based on the 3D score map and to control one or more operations of the vehicle based on the determined pose.
-
公开(公告)号:US11061402B2
公开(公告)日:2021-07-13
申请号:US15890886
申请日:2018-02-07
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Mengye Ren , Andrei Pokrovsky , Bin Yang
IPC: G05D1/00 , G01S17/89 , G01S17/86 , G01S17/931 , G05D1/02
Abstract: The present disclosure provides systems and methods that apply neural networks such as, for example, convolutional neural networks, to sparse imagery in an improved manner. For example, the systems and methods of the present disclosure can be included in or otherwise leveraged by an autonomous vehicle. In one example, a computing system can extract one or more relevant portions from imagery, where the relevant portions are less than an entirety of the imagery. The computing system can provide the relevant portions of the imagery to a machine-learned convolutional neural network and receive at least one prediction from the machine-learned convolutional neural network based at least in part on the one or more relevant portions of the imagery. Thus, the computing system can skip performing convolutions over regions of the imagery where the imagery is sparse and/or regions of the imagery that are not relevant to the prediction being sought.
-
公开(公告)号:US20210200212A1
公开(公告)日:2021-07-01
申请号:US16825049
申请日:2020-03-20
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Abbas Sadat , Mengye Ren , Andrei Pokrovsky , Yen-Chen Lin , Ersin Yumer
Abstract: Systems and methods for generating motion plans for autonomous vehicles are provided. An autonomous vehicle can include a machine-learned motion planning system including one or more machine-learned models configured to generate target trajectories for the autonomous vehicle. The model(s) include a behavioral planning stage configured to receive situational data based at least in part on the one or more outputs of the set of sensors and to generate behavioral planning data based at least in part on the situational data and a unified cost function. The model(s) includes a trajectory planning stage configured to receive the behavioral planning data from the behavioral planning stage and to generate target trajectory data for the autonomous vehicle based at least in part on the behavioral planning data and the unified cost function.
-
公开(公告)号:US11755014B2
公开(公告)日:2023-09-12
申请号:US16825049
申请日:2020-03-20
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Abbas Sadat , Mengye Ren , Andrei Pokrovsky , Yen-Chen Lin , Ersin Yumer
CPC classification number: G05D1/0088 , G05D1/0214 , G05D1/0221 , G05D2201/0213
Abstract: Systems and methods for generating motion plans for autonomous vehicles are provided. An autonomous vehicle can include a machine-learned motion planning system including one or more machine-learned models configured to generate target trajectories for the autonomous vehicle. The model(s) include a behavioral planning stage configured to receive situational data based at least in part on the one or more outputs of the set of sensors and to generate behavioral planning data based at least in part on the situational data and a unified cost function. The model(s) includes a trajectory planning stage configured to receive the behavioral planning data from the behavioral planning stage and to generate target trajectory data for the autonomous vehicle based at least in part on the behavioral planning data and the unified cost function.
-
公开(公告)号:US20240085908A1
公开(公告)日:2024-03-14
申请号:US18513119
申请日:2023-11-17
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Mengye Ren , Andrei Pokrovsky , Bin Yang
IPC: G05D1/00 , G01S17/86 , G01S17/89 , G01S17/931
CPC classification number: G05D1/0088 , G01S17/86 , G01S17/89 , G01S17/931 , G05D1/0246
Abstract: The present disclosure provides systems and methods that apply neural networks such as, for example, convolutional neural networks, to sparse imagery in an improved manner. For example, the systems and methods of the present disclosure can be included in or otherwise leveraged by an autonomous vehicle. In one example, a computing system can extract one or more relevant portions from imagery, where the relevant portions are less than an entirety of the imagery. The computing system can provide the relevant portions of the imagery to a machine-learned convolutional neural network and receive at least one prediction from the machine-learned convolutional neural network based at least in part on the one or more relevant portions of the imagery. Thus, the computing system can skip performing convolutions over regions of the imagery where the imagery is sparse and/or regions of the imagery that are not relevant to the prediction being sought.
-
公开(公告)号:US11860629B2
公开(公告)日:2024-01-02
申请号:US17363986
申请日:2021-06-30
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Mengye Ren , Andrei Pokrovsky , Bin Yang
IPC: G05D1/00 , G01S17/89 , G01S17/86 , G01S17/931 , G05D1/02
CPC classification number: G05D1/0088 , G01S17/86 , G01S17/89 , G01S17/931 , G05D1/0246
Abstract: The present disclosure provides systems and methods that apply neural networks such as, for example, convolutional neural networks, to sparse imagery in an improved manner. For example, the systems and methods of the present disclosure can be included in or otherwise leveraged by an autonomous vehicle. In one example, a computing system can extract one or more relevant portions from imagery, where the relevant portions are less than an entirety of the imagery. The computing system can provide the relevant portions of the imagery to a machine-learned convolutional neural network and receive at least one prediction from the machine-learned convolutional neural network based at least in part on the one or more relevant portions of the imagery. Thus, the computing system can skip performing convolutions over regions of the imagery where the imagery is sparse and/or regions of the imagery that are not relevant to the prediction being sought.
-
公开(公告)号:US20230359202A1
公开(公告)日:2023-11-09
申请号:US18355188
申请日:2023-07-19
Applicant: UATC, LLC
Inventor: Raquel Urtasun , Yen-Chen Lin , Andrei Pokrovsky , Mengye Ren , Abbas Sadat , Ersin Yumer
CPC classification number: G05D1/0088 , G05D1/0214 , G05D1/0221 , G05D2201/0213
Abstract: Systems and methods for generating motion plans for autonomous vehicles are provided. An autonomous vehicle can include a machine-learned motion planning system including one or more machine-learned models configured to generate target trajectories for the autonomous vehicle. The model(s) include a behavioral planning stage configured to receive situational data based at least in part on the one or more outputs of the set of sensors and to generate behavioral planning data based at least in part on the situational data and a unified cost function. The model(s) includes a trajectory planning stage configured to receive the behavioral planning data from the behavioral planning stage and to generate target trajectory data for the autonomous vehicle based at least in part on the behavioral planning data and the unified cost function.
-
-
-
-
-
-
-
-