-
公开(公告)号:US20210042929A1
公开(公告)日:2021-02-11
申请号:US16968573
申请日:2019-04-19
Inventor: Xin ZHAO , Kaiqi HUANG , Zhe LIU
Abstract: A three-dimensional object detection method includes: extracting a target in a two-dimensional image by a pre-trained deep convolutional neural network to obtain a plurality of target objects; determining a point cloud frustum in a corresponding three-dimensional point cloud space based on each target object; segmenting the point cloud in the frustum based on a point cloud segmentation network to obtain a point cloud of interest; and estimating parameters of a 3D box in the point cloud of interest based on a network with the weighted channel features to obtain the parameters of the 3D box for three-dimensional object detection. According to the present invention, the features of the image can be learned more accurately by the deep convolutional neural network and the parameters of the 3D box in the point cloud of interest are estimated based on the network with the weighted channel features.
-
公开(公告)号:US20210065371A1
公开(公告)日:2021-03-04
申请号:US16968575
申请日:2019-04-19
Inventor: Xin ZHAO , Kaiqi HUANG , Yupei WANG
Abstract: A refined segmentation system, method and device of an image shadow area are provided. The system of the present invention includes: a feature extraction network, a reverse fusion network, and a weighted fusion network. The feature extraction network includes a plurality of sampling layers which are arranged sequentially, a plurality of segmentation features of the shadow areas in the input images are obtained through the sampling layers sequentially. The reverse fusion network includes a plurality of layered reverse fusion branches, each of which includes a plurality of feature fusion layers arranged in sequence, and two input features are fused in sequence through each feature fusion layer. The weighted fusion network is configured to perform weighted fusion on outputs of the plurality of reverse fusion branches to obtain a final segmentation result of the shadow area of the input image.
-