Abstract:
A learning device makes an object detection device learn how to detect an object from an input image. A feature extraction unit performs feature extraction from input images including real images and pseudo images to generate feature maps, and the object detection unit detects objects included in the input images based on the feature maps. The domain identification unit identifies the domains forming the input images and generates domain identifiability information. Then, the feature extraction unit and the object detection unit learn common features that do not depend on the difference in domains, based on the domain identifiability information.
Abstract:
An object tracking apparatus, method and computer-readable medium for detecting an object from output information of sensors, tracking the object on a basis of a plurality of detection results, generating tracking information of the object represented in a common coordinate system, outputting the tracking information, and detecting the object on a basis of the tracking information.
Abstract:
Provided is a technique for enhancing operability of a mobile apparatus. An information processing apparatus (2000) includes a first processing unit (2020), a second processing unit (2040), and a control unit (2060). The first processing unit (2020) generates information indicating an event detection position in accordance with a position on a surveillance image set in a first operation. The first operation is an operation with respect to the surveillance image displayed on a display screen. The second processing unit (2040) performs a display change process with respect to the surveillance image or a window including the surveillance image. The control unit (2060) causes any one of the first processing unit (2020) and the second processing unit (2040) to process the first operation on the basis of a second operation.
Abstract:
In a foreground extraction apparatus, an extraction result generation unit performs a foreground extraction using a plurality of foreground extraction models for an input image, and generates foreground extraction results. A selection unit selects one or more foreground extraction models among the plurality of foreground extraction models using respective foreground results acquired by the plurality of foreground extraction models. A foreground region generation unit extracts each foreground region based on the input image using the selected one or more foreground extraction models.
Abstract:
Provided is a technique for enhancing operability of a mobile apparatus. An information processing apparatus (2000) includes a first processing unit (2020), a second processing unit (2040), and a control unit (2060). The first processing unit (2020) generates information indicating an event detection position in accordance with a position on a surveillance image set in a first operation. The first operation is an operation with respect to the surveillance image displayed on a display screen. The second processing unit (2040) performs a display change process with respect to the surveillance image or a window including the surveillance image. The control unit (2060) causes any one of the first processing unit (2020) and the second processing unit (2040) to process the first operation on the basis of a second operation.
Abstract:
Provided is an image processing apparatus (2000) including an index value calculation unit (2020) and a presentation unit (2040). The index value calculation unit (2020) acquires a plurality of images captured by a camera (3000) (captured images), and calculates an index value indicating the degree of change in the state of a monitoring target in the captured image, using the acquired captured image. The presentation unit (2040) presents an indication based on the index value calculated by the index value calculation unit (2020) on the captured image captured by the camera (3000).
Abstract:
Provided are an information processing system, an information processing method and a program capable of suitably monitoring a moving body related to a plurality of imaging apparatuses. The information processing system of the present invention includes: an input unit 110 for receiving inputs of pictures imaged by a plurality of video cameras 200; a human figure detection unit 120 for detecting a moving body appearing in a picture input from the input unit 110 and imaged by a video camera 200A among the plurality of video cameras 200; a time prediction unit 140 for predicting a time period in which the moving body appears in a picture imaged by a video camera 200N, based on a time transition of a probability of the moving body, which is detected in the picture of the video camera 200A, appearing in the picture of the video camera 200N and a time passage after the moving body detected in the picture of the video camera 200A leaves an imaging range; and a UI generation unit 153 for providing a notice regarding a time period in which the moving body is predicted to appear in the picture of the video camera 200N.
Abstract:
An information processing apparatus (2000) generates likelihood data for each of a plurality of partial regions (12) in image data (10). The likelihood data are data being associated with a position and a size on the image data (10) and indicating a likelihood that a target object exists in an image region at the position with the size. The information processing apparatus (2000) computes a distribution (probability hypothesis density: PHD) of an existence likelihood of a target object with respect to a position and a size by computing the total sum of likelihood data each piece of which is generated for each partial region (12). The information processing apparatus (2000) extracts, from the PHD, partial distributions each of which relates to one target object. For each extracted partial distribution, the information processing apparatus (2000) outputs a position and a size of a target object represented by the partial distribution, based on a statistic of the partial distribution.
Abstract:
An information processing system is provided. The information processing system comprises at least one processor configured to receive inputs of videos captured by a plurality of imaging apparatuses, detect a first person image appearing in a first video captured by a first imaging apparatus and a second person image appearing in a second video captured by a second imaging apparatus, associate a first person with a second person based on a first feature value of the first person and a second feature value of the second person, and provide an elapsed time from a first time point when the first person disappears from the first video to a second time point when the second person appears in the second video. The first feature value and the second feature value may include information based on histogram.
Abstract:
A guidance processing apparatus (100) includes an information acquisition unit (101) that acquires a plurality of different pieces of guidance information on the basis of states of a plurality of people within one or more images, and a control unit (102) that performs control of a plurality of target devices present in different spaces or time division control of a target device so as to set a plurality of different states corresponding to the plurality of pieces of guidance information.