-
公开(公告)号:US20210360252A1
公开(公告)日:2021-11-18
申请号:US17125370
申请日:2020-12-17
Inventor: Chao LI , Yukang DING , Dongliang HE , Fu LI , Hao SUN , Shilei WEN , Hongwu ZHANG , Errui DING
IPC: H04N19/132 , H04N19/172 , G06N3/04 , G06K9/62
Abstract: A method for video frame interpolation, a related electronic device and a storage medium is disclosed. A video is obtained. An (i−1)th frame and an ith frame of the video are obtained. Visual semantic feature maps and depth maps of the (i−1)th frame and the ith frame are obtained. Frame interpolation information is obtained based on the visual semantic feature maps and the depth maps. An interpolated frame between the (i−1)th frame and the ith frame is generated based on the frame interpolation information and the (i−1)th frame and is inserted between the (i−1)th frame and the ith frame.
-
公开(公告)号:US20210209731A1
公开(公告)日:2021-07-08
申请号:US17024253
申请日:2020-09-17
Inventor: Chao LI , Shilei WEN , Errui DING
Abstract: Embodiments of the present disclosure provide a video processing method, a video processing device and a related non-transitory computer readable storage medium. The method includes the following. Frame sequence data of a low-resolution video to be converted is obtained. Pixel tensors of each frame in the frame sequence data are inputted into a pre-trained neural network model to obtain high-resolution video frame sequence data corresponding to the video to be converted output by the neural network model. The neural network model obtains the high-resolution video frame sequence data based on high-order pixel information of each frame in the frame sequence data.
-
公开(公告)号:US20200342271A1
公开(公告)日:2020-10-29
申请号:US16817419
申请日:2020-03-12
Inventor: Zhigang WANG , Jian WANG , Shilei WEN , Errui DING , Hao SUN
Abstract: The present disclosure provides a pedestrian re-identification method and apparatus, computer device and readable medium. The method comprises: collecting a target image and a to-be-identified image including a pedestrian image; obtaining a feature expression of the target image and a feature expression of the to-be-identified image respectively, based on a pre-trained feature extraction model; wherein the feature extraction model is obtained by training based on a self-attention feature of a base image as well as a co-attention feature of the base image relative to a reference image; identifying whether a pedestrian in the to-be-identified image is the same pedestrian as that in the target image according to the feature expression of the target image and the feature expression of the to-be-identified image. According to the pedestrian re-identification method of the present disclosure, the accuracy of the pedestrian re-identification can be effectively improved when the feature extraction model is used to perform the pedestrian re-identification.
-
公开(公告)号:US20210390731A1
公开(公告)日:2021-12-16
申请号:US17201665
申请日:2021-03-15
Inventor: Jian WANG , Zipeng LU , Hao SUN , Hongwu ZHANG , Shilei WEN , Errui DING
Abstract: A method and apparatus for positioning a key point, a device, and a storage medium are provided. The method may include: extracting a first feature map and a second feature map of a to-be-positioned image, the first feature map and the second feature map being different feature maps; determining, based on the first feature map, an initial position of a key point in the to-be-positioned image; determining, based on the second feature map, an offset of the key point; and adding the initial position of the key point with the offset of the key point to obtain a final position of the key point.
-
5.
公开(公告)号:US20210216782A1
公开(公告)日:2021-07-15
申请号:US17144205
申请日:2021-01-08
Inventor: Tianwei LIN , Xin LI , Dongliang HE , Fu LI , Hao SUN , Shilei WEN , Errui DING
Abstract: A method and apparatus for detecting a temporal action of a video, an electronic device and a storage medium are disclosed, which relates to the field of video processing technologies. An implementation includes: acquiring an initial temporal feature sequence of a video to be detected; acquiring, by a pre-trained video-temporal-action detecting module, implicit features and explicit features of a plurality of configured temporal anchor boxes based on the initial temporal feature sequence; and acquiring, by the video-temporal-action detecting module, the starting position and the ending position of a video clip containing a specified action, the category of the specified action and the probability that the specified action belongs to the category from the plural temporal anchor boxes according to the explicit features and the implicit features of the plural temporal anchor boxes.
-
6.
公开(公告)号:US20210390682A1
公开(公告)日:2021-12-16
申请号:US17116597
申请日:2020-12-09
Inventor: Shufei LIN , Jianfeng ZHU , Pengcheng YUAN , Bin ZHANG , Shumin HAN , Yingbo XU , Yuan FENG , Ying XIN , Xiaodi WANG , Jingwei LIU , Shilei WEN , Hongwu ZHANG , Errui DING
Abstract: A method for detecting a surface defect, a method for training model, an apparatus, a device, and a medium, are provided. The method includes: inputting a surface image of the article for detection into a defect detection model to perform a defect detection, and acquiring a defect detection result output by the defect detection model; inputting a surface image of a defective article determined to be defective into an image discrimination model based on the defect detection result to determine whether the surface image of the defective article is defective, wherein the image discrimination model is a trained generative adversarial networks model, and the generative adversarial networks model is obtained by training using a surface image of a defect-free good article; and adjusting the defect detection result of the surface image of the defective article according to a determination result of the image discrimination model.
-
公开(公告)号:US20210374415A1
公开(公告)日:2021-12-02
申请号:US17194131
申请日:2021-03-05
Inventor: Dongliang HE , Xiao TAN , Shilei WEN , Hao SUN
Abstract: Embodiments of the present disclosure disclose a method and apparatus for identifying a video. A specific embodiment of the method includes: acquiring a predetermined number of video frames from a video to be identified to obtain a video frame sequence; performing the following processing step: importing the video frame sequence into a pre-trained video identification model to obtain a classification tag probability corresponding to the video frame sequence, wherein the classification tag probability is used to characterize a probability of identifying a corresponding tag category of the video to be identified; and setting, in response to the classification tag probability being greater than or equal to a preset identification accuracy threshold, a video tag for the video to be identified according to the classification tag probability, or else increasing the number of video frames in the video frame sequence and continuing to perform the above processing step.
-
公开(公告)号:US20210227152A1
公开(公告)日:2021-07-22
申请号:US17025255
申请日:2020-09-18
Inventor: Henan ZHANG , Xin LI , Fu LI , Tianwei LIN , Hao SUN , Shilei WEN , Hongwu ZHANG , Errui DING
Abstract: Embodiments of the present disclosure provide a method and apparatus for generating an image. The method may include: receiving a first image including a face input by a user in an interactive scene; presenting the first image to the user; inputting the first image into a pre-trained generative adversarial network in a backend to obtain a second image output by the generative adversarial network; where the generative adversarial network uses face attribute information generated based on the input image as a constraint; and presenting the second image to the user in response to obtaining the second image output by the generative adversarial network in the backend.
-
9.
公开(公告)号:US20210216783A1
公开(公告)日:2021-07-15
申请号:US17144523
申请日:2021-01-08
Inventor: Xiang LONG , Dongliang HE , Fu LI , Xiang ZHAO , Tianwei LIN , Hao SUN , Shilei WEN , Errui DING
Abstract: A method includes screening, by a video-clip screening module in a video description model, a plurality of video proposal clips acquired from a video to be analyzed, to acquire a plurality of video clips suitable for description. The plural video proposal clips acquired from the video to be analyzed may be screened by the video-clip screening module to acquire the plural video clips suitable for description; and then, each video clip is described by a video-clip describing module, thus avoiding description of all the video proposal clips, only describing the screened video clips which have strong correlation with the video and are suitable for description, removing the interference of the description of the video clips which are not suitable for description in the description of the video, guaranteeing the accuracy of the final descriptions of the video clips, and improving the quality of the descriptions of the video clips.
-
10.
公开(公告)号:US20200372609A1
公开(公告)日:2020-11-26
申请号:US16810986
申请日:2020-03-06
Inventor: Chao LI , Dongliang HE , Xiao LIU , Yukang DING , Shilei WEN , Errui DING , Henan ZHANG , Hao SUN
IPC: G06T3/40
Abstract: A super-resolution video reconstruction method, device, apparatus and a computer-readable storage medium are provided. The method includes: extracting a hypergraph from consecutive frames of an original video; inputting a hypergraph vector of the hypergraph into a residual convolutional neural network to obtain an output result of the residual convolutional neural network; and inputting the output result of the residual convolutional neural network into a spatial upsampling network to obtain a super-resolution frame, wherein a super-resolution video of the original video is formed by multiple super-resolution frames.
-
-
-
-
-
-
-
-
-