-
公开(公告)号:US11893750B2
公开(公告)日:2024-02-06
申请号:US16732274
申请日:2019-12-31
Applicant: Zoox, Inc.
Inventor: Kratarth Goel , James William Vaisey Philbin , Praveen Srinivasan , Sarah Tariq
IPC: G06T7/11 , G06T7/50 , G06T7/579 , G06N20/00 , G05D1/00 , G05D1/02 , G06F18/21 , G06V10/25 , G06V10/764 , G06V20/56 , G06V20/64 , G06T7/207
CPC classification number: G06T7/207 , G05D1/0038 , G05D1/0238 , G05D1/0253 , G06F18/217 , G06N20/00 , G06T7/11 , G06T7/50 , G06T7/579 , G06V10/25 , G06V10/764 , G06V20/56 , G06V20/64 , G06T2207/20081 , G06T2207/20104 , G06T2207/30252
Abstract: A machine-learning (ML) architecture for determining three or more outputs, such as a two and/or three-dimensional region of interest, semantic segmentation, direction logits, depth data, and/or instance segmentation associated with an object in an image. The ML architecture may output these outputs at a rate of 30 or more frames per second on consumer grade hardware.
-
公开(公告)号:US20230266771A1
公开(公告)日:2023-08-24
申请号:US18114531
申请日:2023-02-27
Applicant: Zoox, Inc.
Inventor: Sarah Tariq , Kratarth Goel , James William Vaisey Philbin
IPC: G05D1/02 , G05D1/00 , G06T3/40 , G06N3/08 , G06V10/25 , G06F18/214 , G06N7/01 , G06V10/774 , G06V10/82 , G06V20/58
CPC classification number: G05D1/0246 , G05D1/0088 , G05D1/0231 , G06T3/40 , G06N3/08 , G06V10/25 , G06F18/214 , G06N7/01 , G06V10/774 , G06V10/82 , G06V20/58
Abstract: Techniques for utilizing multiple scales of images as input to machine learning (ML) models are discussed herein. Operations can include providing an image associated with a first scale to a first ML model. An output of the first ML model can include a first bounding box indicative of a first region of the image representing a first object, with the first bounding box falling within a first range of sizes. Next, a scaled image can be generated by scaling the image. The scaled image can be provided to a second ML model, which can output a second bounding box indicative of a second region of the image representing a second object, the second bounding falling within a second range of sizes. Thus, inputting a scaled image to a same ML model (or to different ML models) can result in different detected features in the images.
-
公开(公告)号:US20230091924A1
公开(公告)日:2023-03-23
申请号:US17478602
申请日:2021-09-17
Applicant: Zoox, Inc.
Inventor: Jonathan Tyler Dowdall , Kratarth Goel , Adam Edward Pollack , Scott M. Purdy , Bharadwaj Raghavan
Abstract: Techniques for utilizing a depth completion algorithm to determine dense depth data are discussed are discussed herein. Two-dimensional image data representing an environment can be captured or otherwise received. Depth data representing the environment can be captured or otherwise received. The depth data can be projected into the image data and processed using the depth completion algorithm. The depth completion algorithm can be utilized to determine the dense depth values based on intensity values of pixels, and other depth values. A vehicle can be controlled based on the determined depth values.
-
公开(公告)号:US20230033177A1
公开(公告)日:2023-02-02
申请号:US17390174
申请日:2021-07-30
Applicant: Zoox, Inc.
Inventor: Kratarth Goel
Abstract: Techniques are discussed herein for generating three-dimensional (3D) representations of an environment based on two-dimensional (2D) image data, and using the 3D representations to perform 3D object detection and other 3D analyses of the environment. 2D image data may be received, along with depth estimation data associated with the 2D image data. Using the 2D image data and associated depth data, an image-based object detector may generate 3D representations, including point clouds and/or 3D pixel grids, for the 2D image or particular regions of interest. In some examples, a 3D point cloud may be generated by projecting pixels from the 2D image into 3D space followed by a trained 3D convolutional neural network (CNN) performing object detection. Additionally or alternatively, a top-down view of a 3D pixel grid representation may be used to perform object detection using 2D convolutions.
-
15.
公开(公告)号:US11163990B2
公开(公告)日:2021-11-02
申请号:US16457524
申请日:2019-06-28
Applicant: Zoox, Inc.
Inventor: Kratarth Goel
Abstract: Techniques described herein relate to using head detection to improve pedestrian detection. In an example, a head can be detected in sensor data received from a sensor associated with a vehicle using a machine learned model. Based at least partly on detecting the head in the sensor data, a pedestrian can be determined to be present in an environment within which the vehicle is positioned. In an example, an indication of the pedestrian can be provided to at least one system of the vehicle, for instance, for use by the at least one system to make a determination associated with controlling the vehicle.
-
公开(公告)号:US20210150279A1
公开(公告)日:2021-05-20
申请号:US16684568
申请日:2019-11-14
Applicant: Zoox, Inc.
Inventor: Thomas Oscar Dudzik , Kratarth Goel , Praveen Srinivasan , Sarah Tariq
Abstract: Techniques for training a machine learned (ML) model to determine depth data based on image data are discussed herein. Training can use stereo image data and depth data (e.g., lidar data). A first (e.g., left) image can be input to a ML model, which can output predicted disparity and/or depth data. The predicted disparity data can be used with second image data (e.g., a right image) to reconstruct the first image. Differences between the first and reconstructed images can be used to determine a loss. Losses may include pixel, smoothing, structural similarity, and/or consistency losses. Further, differences between the depth data and the predicted depth data and/or differences between the predicted disparity data and the predicted depth data can be determined, and the ML model can be trained based on the various losses. Thus, the techniques can use self-supervised training and supervised training to train a ML model.
-
公开(公告)号:US12056934B2
公开(公告)日:2024-08-06
申请号:US17390234
申请日:2021-07-30
Applicant: Zoox, Inc.
Inventor: Kratarth Goel
Abstract: Techniques are discussed herein for generating three-dimensional (3D) representations of an environment based on two-dimensional (2D) image data, and using the 3D representations to perform 3D object detection and other 3D analyses of the environment. 2D image data may be received, along with depth estimation data associated with the 2D image data. Using the 2D image data and associated depth data, an image-based object detector may generate 3D representations, including point clouds and/or 3D pixel grids, for the 2D image or particular regions of interest. In some examples, a 3D point cloud may be generated by projecting pixels from the 2D image into 3D space followed by a trained 3D convolutional neural network (CNN) performing object detection. Additionally or alternatively, a top-down view of a 3D pixel grid representation may be used to perform object detection using 2D convolutions.
-
公开(公告)号:US12051276B1
公开(公告)日:2024-07-30
申请号:US18205651
申请日:2023-06-05
Applicant: Zoox, Inc.
Inventor: Oytun Ulutan , Xin Wang , Kratarth Goel , Vasiliy Karasev , Sarah Tariq , Yi Xu
IPC: G06V40/20 , B60W60/00 , G05D1/00 , G06F18/21 , G06F18/214 , G06F18/24 , G06N20/00 , G06T7/70 , G06V20/58 , G06V40/10
CPC classification number: G06V40/28 , G06F18/2148 , G06F18/217 , G06F18/24 , G06T7/70 , G06V20/582 , G06V40/103 , G06V40/23 , B60W60/001 , B60W2420/403 , B60W2540/041 , G05D1/0088 , G05D1/0231 , G06N20/00 , G06T2207/20081 , G06T2207/30196 , G06T2207/30252
Abstract: Techniques for detecting attributes and/or gestures associated with pedestrians in an environment are described herein. The techniques may include receiving sensor data associated with a pedestrian in an environment of a vehicle and inputting the sensor data into a machine-learned model that is configured to determine a gesture and/or an attribute of the pedestrian. Based on the input data, an output may be received from the machine-learned model that indicates the gesture and/or the attribute of the pedestrian and the vehicle may be controlled based at least in part on the gesture and/or the attribute of the pedestrian. The techniques may also include training the machine-learned model to detect the attribute and/or the gesture of the pedestrian.
-
公开(公告)号:US11710352B1
公开(公告)日:2023-07-25
申请号:US17320690
申请日:2021-05-14
Applicant: Zoox, Inc.
Inventor: Oytun Ulutan , Xin Wang , Kratarth Goel , Vasiliy Karasev , Sarah Tariq , Yi Xu
IPC: G06V40/20 , G06T7/70 , G06V20/58 , G06V40/10 , G06F18/24 , G06F18/21 , G06F18/214 , G06N20/00 , G05D1/02 , G05D1/00 , B60W60/00
CPC classification number: G06V40/28 , G06F18/217 , G06F18/2148 , G06F18/24 , G06T7/70 , G06V20/582 , G06V40/103 , G06V40/23 , B60W60/001 , B60W2420/42 , B60W2540/041 , G05D1/0088 , G05D1/0231 , G06N20/00 , G06T2207/20081 , G06T2207/30196 , G06T2207/30252
Abstract: Techniques for detecting attributes and/or gestures associated with pedestrians in an environment are described herein. The techniques may include receiving sensor data associated with a pedestrian in an environment of a vehicle and inputting the sensor data into a machine-learned model that is configured to determine a gesture and/or an attribute of the pedestrian. Based on the input data, an output may be received from the machine-learned model that indicates the gesture and/or the attribute of the pedestrian and the vehicle may be controlled based at least in part on the gesture and/or the attribute of the pedestrian. The techniques may also include training the machine-learned model to detect the attribute and/or the gesture of the pedestrian.
-
公开(公告)号:US11681046B2
公开(公告)日:2023-06-20
申请号:US17508482
申请日:2021-10-22
Applicant: Zoox, Inc.
Inventor: Thomas Oscar Dudzik , Kratarth Goel , Praveen Srinivasan , Sarah Tariq
IPC: G01S17/86 , G06T3/00 , G06T7/593 , G01S7/48 , G01S17/931 , G06V20/58 , G06F18/214 , G06F18/25 , G06V10/774 , G06V10/776
CPC classification number: G01S17/86 , G01S7/4808 , G01S17/931 , G06F18/2155 , G06F18/254 , G06T3/0093 , G06T7/593 , G06V10/776 , G06V10/7753 , G06V20/58 , G06T2207/10012 , G06T2207/20081 , G06V2201/07
Abstract: Techniques for training a machine learned (ML) model to determine depth data based on image data are discussed herein. Training can use stereo image data and depth data (e.g., lidar data). A first (e.g., left) image can be input to a ML model, which can output predicted disparity and/or depth data. The predicted disparity data can be used with second image data (e.g., a right image) to reconstruct the first image. Differences between the first and reconstructed images can be used to determine a loss. Losses may include pixel, smoothing, structural similarity, and/or consistency losses. Further, differences between the depth data and the predicted depth data and/or differences between the predicted disparity data and the predicted depth data can be determined, and the ML model can be trained based on the various losses. Thus, the techniques can use self-supervised training and supervised training to train a ML model.
-
-
-
-
-
-
-
-
-