IMAGE CONVERSION DEVICE, IMAGE CONVERSION MODEL LEARNING DEVICE, METHOD, AND PROGRAM

    公开(公告)号:US20220188975A1

    公开(公告)日:2022-06-16

    申请号:US17604307

    申请日:2020-04-20

    Abstract: A low-resolution image can be converted into a high-resolution image in consideration of differential values of the images.
    A learning conversion unit 22 inputs a first image for learning to a conversion processing model for converting the first image into a second image having a higher resolution than the first image to acquire the second image for learning corresponding to the first image for learning. Then, a differential value calculation unit 24 calculates a differential value from the acquired second image for learning, and calculates a differential value from a correct second image corresponding to the first image for learning. Then, the learning unit 26 causes the conversion processing model to learn by associating the calculated differential value of the second image for learning with the differential value of the correct second image.

    ACTION RECOGNITION LEARNING DEVICE, ACTION RECOGNITION LEARNING METHOD, ACTION RECOGNITION LEARNING DEVICE, AND PROGRAM

    公开(公告)号:US20220398868A1

    公开(公告)日:2022-12-15

    申请号:US17774113

    申请日:2020-10-30

    Abstract: The present invention makes it possible to cause an action recognizer capable of recognizing actions with high accuracy and with a small quantity of learning data to learn. An input unit 101 receives input of a learning video and an action label indicating an action of an object, a detection unit 102 detects a plurality of objects included in each frame image included in the learning video, a direction calculation unit 103 calculates a direction of a reference object, which is an object to be used as a reference among the plurality of detected objects, a normalization unit 104 normalizes the learning video so that a positional relationship between the reference object and another object becomes a predetermined relationship, and an optimization unit 106 optimizes parameters of an action recognizer to estimate the action of the object in the inputted video based on the action estimated by inputting the normalized learning video to the action recognizer and the action indicated by the action label.

    CANDIDATE REGION ESTIMATION DEVICE, CANDIDATE REGION ESTIMATION METHOD, AND PROGRAM

    公开(公告)号:US20210304415A1

    公开(公告)日:2021-09-30

    申请号:US17265166

    申请日:2019-08-01

    Abstract: The present invention makes it possible to estimate, with high precision, a candidate region indicating each of multiple target objects included in an image. A parameter determination unit 11 determines parameters to be used when detecting a boundary line of an image 101 based on a ratio between a density of boundary lines included in an image 101 and a density of boundary lines in a region indicated by region information 102 indicating the region including at least one of the multiple target objects included in the image 101. A boundary line detection unit 12 detects the boundary line in the image 101 using the parameter. For each of the multiple target objects included in the image 101, the region estimation unit 13 estimates the candidate region of the target object based on the detected boundary line.

    OBJECT DETECTION DEVICE, METHOD, AND PROGRAM

    公开(公告)号:US20210209403A1

    公开(公告)日:2021-07-08

    申请号:US17251172

    申请日:2019-05-07

    Abstract: Even if an object to be detected is not remarkable in images, and the input includes images including regions that are not the object to be detected and have a common appearance on the images, a region indicating the object to be detected is accurately detected. A local feature extraction unit 20 extracts a local feature of a feature point from each image included in an input image set. An image-pair common pattern extraction unit 30 extracts, from each image pair selected from images included in the image set, a common pattern constituted by a set of feature point pairs that have similar local features extracted by the local feature extraction unit 20 in images constituting the image pair, the set of feature point pairs being geometrically similar to each other. A region detection unit 50 detects, as a region indicating an object to be detected in each image included in the image set, a region that is based on a common pattern that is omnipresent in the image set, of common patterns extracted by the image-pair common pattern extraction unit 30.

    GEOMETRIC TRANSFORMATION MATRIX ESTIMATING DEVICE, GEOMETRIC TRANSFORMATION MATRIX ESTIMATING METHOD, AND PROGRAM

    公开(公告)号:US20210201440A1

    公开(公告)日:2021-07-01

    申请号:US17058089

    申请日:2019-05-28

    Abstract: An object is to make it possible to precisely infer a geometric transformation matrix for transformation between an image and a reference image representing a plane region even if correspondence to the reference image cannot be obtained. A first line segment group extraction unit 120 extracts, out of a line segment group in an image, line segments that correspond to a direction that is parallel or perpendicular to a side of a rectangle included in the image, from the inside of the rectangle, takes the extracted line segments to be a first line segment group, and extracts a plurality of line segments different from the first line segment group out of the line segment group. An endpoint detection unit 150 detects four intersection points between ends of the image and two line segments that are selected from line segments that correspond to a direction that is parallel or perpendicular to a side of the rectangle and are extracted from a plurality of line segments obtained by transforming the different line segments using an affine transformation matrix in which an angle of the first line segment group relative to a reference direction of the image is used as a rotation angle. A homography matrix inferring unit 160 computes a geometric transformation matrix based on the affine transformation matrix and a homography matrix computed based on correspondence between the four intersection points and the four vertexes of the rectangle in a reference image.

    SEGMENT RECOGNITION METHOD, SEGMENT RECOGNITION DEVICE AND PROGRAM

    公开(公告)号:US20230186478A1

    公开(公告)日:2023-06-15

    申请号:US17928851

    申请日:2020-06-05

    CPC classification number: G06T7/11 G06T7/194 G06T2210/12

    Abstract: A segmentation recognition method includes: an object detection step of detecting an object image in a target image by inputting bounding box information including a coordinate and category information of each bounding box defined in the target image to an object detection model that uses a machine learning approach; a filtering step of selecting effective training mask information from training mask information associated with foregrounds in the target image based on the bounding box information; a bounding box branch step of recognizing the object image using weight information of the object detection model as an initial value of weight information of an object recognition model that recognizes an object of the object image; and a mask branch step of generating mask information having a shape of the object image using the selected effective training mask information as training data and using weight information of the object recognition model as an initial value of weight information of a segmentation shape model that segments the target image according to a shape of the object image.

    OBJECT LIKELIHOOD ESTIMATION DEVICE, METHOD, AND PROGRAM

    公开(公告)号:US20210216829A1

    公开(公告)日:2021-07-15

    申请号:US15733883

    申请日:2019-05-31

    Abstract: Objectness indicating a degree of accuracy of a single object is accurately estimated. An edge detection unit 30 detects an edge for a depth image, an edge density/uniformity calculation unit 40 calculates an edge density on the periphery of a candidate region, an edge density inside the candidate region, and edge uniformity on the periphery of the candidate region. An objectness calculation unit 42 calculates the objectness of the candidate region based on the edge density on the periphery of the candidate region, the edge density inside the candidate region, and the edge uniformity on the periphery of the candidate region.

Patent Agency Ranking