-
公开(公告)号:US20240163558A1
公开(公告)日:2024-05-16
申请号:US18495849
申请日:2023-10-27
Applicant: CANON KABUSHIKI KAISHA
Inventor: TAKUYA TOYODA
CPC classification number: H04N23/69 , G06T7/20 , G06V10/25 , H04N23/61 , G06V2201/07
Abstract: An information processing apparatus includes a detection unit configured to detect a predetermined target object from an image that has been captured by an image capturing unit, a setting unit configured to set a predetermined region with respect to the image, a determination unit configured to perform a determination as to whether or not the target object has exited from the inside of the predetermined region, a control unit configured to control an image capturing range of the image capturing unit according to a result that has been determined by the determination unit, and a zoom magnification setting unit configured to set the zoom magnification of the image capturing unit to the same zoom magnification as that before the target object enters the predetermined region, in a case in which the determination unit determines that the target object has exited from the inside of the predetermined region.
-
公开(公告)号:US20240161442A1
公开(公告)日:2024-05-16
申请号:US18451287
申请日:2023-08-17
Inventor: Sujin JANG , Sangpil KIM , Jinkyu KIM , Wonseok ROH , Gyusam CHANG , Dongwook LEE , Dae Hyun JI
CPC classification number: G06V10/25 , G01S17/86 , G01S17/89 , G06V10/44 , G06V10/761 , G06V10/764 , G06V10/82 , G06V2201/07
Abstract: A method and apparatus with object detector training is provided. The method includes obtaining first input data and second input data from a target object; obtaining second additional input data by performing data augmentation on the second input data; extracting a first feature to a shared embedding space by inputting the first input data to a first encoder; extracting a second feature to the shared embedding space by inputting the second input data to a second encoder; extracting a second additional feature to the shared embedding space by inputting thesecond additional input data to the second encoder; identifying a first loss function based on the first feature, the second feature, and the second additional feature; identifying a second loss function based on the second feature and the second additional feature; and updating a weight of the second encoder based on the first loss function and the second loss function.
-
公开(公告)号:US20240161324A1
公开(公告)日:2024-05-16
申请号:US18500293
申请日:2023-11-02
Applicant: CANON KABUSHIKI KAISHA
Inventor: KYOKO MIYAMAE
CPC classification number: G06T7/62 , G06T3/40 , G06T7/20 , G06V10/761 , G06V2201/07
Abstract: An information processing device which detects a predetermined subject from an image captured, determines a speed at which to change an imaging range in accordance with a subject size of the predetermined subject detected, controls the change in the imaging range at the determined speed, determines the speed such that the imaging range is not changed while the subject size detected takes a value within a first range based on a target size, and when a first difference between a detection limit subject size at which the predetermined subject can be detected and a subject size associated with an edge of the first range is lower than a first predetermined value, and changes the subject size associated with the edge of the first range such that the first difference becomes greater than or equal to the first predetermined value.
-
公开(公告)号:US11982540B2
公开(公告)日:2024-05-14
申请号:US17398475
申请日:2021-08-10
Applicant: Mobileye Vision Technologies Ltd.
Inventor: Jonathan Abramson
IPC: G01C21/00 , B60W10/18 , B60W30/18 , G01C21/36 , G06T7/00 , G06T7/246 , G06T7/32 , G06T7/70 , G06T7/73 , G06V20/56 , G06V20/58 , G06V20/64 , G06V40/10 , G06V10/46
CPC classification number: G01C21/3807 , B60W10/18 , B60W30/181 , B60W30/18154 , G01C21/3602 , G01C21/3815 , G01C21/3841 , G06T7/246 , G06T7/32 , G06T7/70 , G06T7/73 , G06T7/97 , G06V20/582 , G06V20/584 , G06V20/588 , G06V20/64 , G06V40/103 , B60W2554/802 , B60W2556/40 , B60W2556/50 , G06T2207/30252 , G06T2207/30256 , G06T2207/30261 , G06V10/462 , G06V20/58 , G06V2201/07 , G06V2201/08
Abstract: A system may include a processor configured to receive a first image captured during a drive of a first vehicle along a road segment and receive a second image captured during a drive of the second vehicle along the road segment; analyze the first and second images to identify representations of objects; analyze the first and second images to determine position indicators for each of the objects relative to the road segment; correlate the position indicators for each of the objects, wherein the correlating includes determining refined positions of each object based on the determined position indicators; and generate, based on the refined positions of objects belonging to a particular predetermined category of objects, a map including representations of the refined positions of one or more of the objects that belong to the particular predetermined category of objects.
-
公开(公告)号:US20240153228A1
公开(公告)日:2024-05-09
申请号:US17980322
申请日:2022-11-03
Applicant: Black Sesame Technologies Inc.
Inventor: HungTing Liu , Bo Li , Shuen Lyu
IPC: G06V10/25 , G06T7/13 , G06V10/44 , G06V10/74 , G06V10/762 , G06V10/771
CPC classification number: G06V10/25 , G06T7/13 , G06V10/44 , G06V10/761 , G06V10/762 , G06V10/771 , G06T2207/20132 , G06T2207/30201 , G06V2201/07
Abstract: Disclosed is a system for automatic cropping of an image of interest from a video sample using smart systems. The image of interest is an image representative of the video sample, which includes desirable characteristics as required by the user, such as a person or object of focus, a specific aspect-ratio, preferred landmarks, information/time-stamps etc. The system for automatic cropping analyzes the video sample and its content to detect at least one image feature. The image feature is then classified based on importance and a potential test cropping area is determined based on the cumulative importance of features detected within each frame. The smart cropping systems and methods disclosed ensure that the most relevant aspects of a video sample are included within the image of interest.
-
公开(公告)号:US20240152159A1
公开(公告)日:2024-05-09
申请号:US18406729
申请日:2024-01-08
Applicant: ELBIT SYSTEMS C4I AND CYBER LTD.
Inventor: Yair KAHN , Nir BURSHTEIN , Ohad VOLVOVITCH , Alon FARAJ , Amit NATIV
IPC: G05D1/644 , G05D1/246 , G05D101/15 , G05D109/20 , G05D111/10 , G06T7/70 , G06V20/10 , G06V20/17
CPC classification number: G05D1/644 , G05D1/246 , G06T7/70 , G06V20/17 , G06V20/176 , G05D2101/15 , G05D2109/20 , G05D2111/10 , G06T2207/10032 , G06T2207/30184 , G06V2201/07
Abstract: Aspects of embodiments to systems and methods for navigating a mobile platform using an imaging device on the platform, from a point of origin towards a target located in a scene, and without requiring a Global Navigation Satellite system (GNSS), by employing the following steps: acquiring, by the imaging device, an image of the scene comprising the target; determining, based on analysis of the image, a direction vector pointing from the mobile platform to the target; advancing the mobile platform in accordance with the direction vector to a new position; and generating, by a distance sensing device, an output as a result of attempting to determine, with the distance sensing device, a distance between the mobile platform and the target. The mobile platform advanced towards the target until the output produced by the distance sensing device is descriptive of a distance which meets a low-distance criterion.
-
公开(公告)号:US20240144689A1
公开(公告)日:2024-05-02
申请号:US18280136
申请日:2022-03-03
Applicant: EVERSEEN LIMITED
Inventor: Dan Alin CRISFALUSI , Alan O'HERLIHY , Cristina TODORAN , Vasile GUI , Dan PESCARU , Ciprian Petru DAVID , Cosmin CERNAZANU , Arion ALEXANDRU
CPC classification number: G06V20/52 , G06T7/20 , G06T7/73 , G06V10/25 , G06V10/764 , G06V40/20 , G06T2207/30241 , G06V2201/07
Abstract: A method and apparatus for the identification of suspect behaviour in a retail environment, the method comprising: detecting a person in a frame of said stream of video data; extracting a set of activities of the identified person from the stream of video data; assigning a numeric value to each extracted activity in the set of extracted activities, said numeric value representative of a threat level of the activity; accumulating said numeric values to provide a behaviour score; and identifying a behaviour as being suspect when the behaviour score reaches a target threshold value associated with the behaviour.
-
78.
公开(公告)号:US20240144613A1
公开(公告)日:2024-05-02
申请号:US18406105
申请日:2024-01-06
Applicant: IMMERSIV
Inventor: STÉPHANE GUERIN , EMMANUELLE ROGER
CPC classification number: G06T19/006 , G06V10/761 , G06V20/52 , G06V2201/07
Abstract: An augmented reality method for monitoring an event in space includes acquisition of a plurality of images of space having at least two landmarks by a camera of a portable device. Space being associated with a three-dimensional reference frame and the portable device being associated with a two-dimensional reference frame. A three-dimensional position and orientation of the space in relation to the camera is determined. The instantaneous position, within the reference frame of the space, of a mobile element moving in the space is received. The position of the mobile element in the two-dimensional reference frame is calculated from transformation parameters calculated from the three-dimensional position and orientation of the space in relation to the camera. An overlay at a predetermined distance in relation to the position of the mobile element in the two-dimensional reference frame is displayed on the screen. Also, a portable electronic device implements the method.
-
公开(公告)号:US20240144509A1
公开(公告)日:2024-05-02
申请号:US18067657
申请日:2022-12-16
Applicant: Maxar International Sweden AB
Inventor: Ola NYGREN , Folke ISAKSSON
CPC classification number: G06T7/60 , G06T7/80 , G06V10/761 , G06V20/13 , G06V2201/07
Abstract: The disclosure relates to a computer implemented method for determining 3D coordinates describing a scene, obtaining first, second and third images covering the scene captured from different viewpoints, said images being associated to a timing of capture, wherein the scene is formed in an overlapping geographical area of the first, second and third images. The method further comprises obtaining intrinsic and/or extrinsic parameter values for at least one imaging device used for capture of the images, and determining the 3D coordinates describing the scene, said determination comprising performing bundle-adjustments based on pair-wise measurements on the images to minimize a re-projection error of reconstructed 3D points in the overlapping geographical area, wherein time is used as a parameter in the determination of the 3D coordinates describing the scene, and wherein the re-projection error is minimized allowing at least a height to be locally time dependent in reconstruction of 3D points at least for a part of the overlapping geographical area.
-
80.
公开(公告)号:US20240135722A1
公开(公告)日:2024-04-25
申请号:US18165857
申请日:2023-02-07
Applicant: NavInfo Europe B.V.
Inventor: Deepan Chakravarthi Padmanabhan , Shruthi Gowda , Elahe Arani , Bahram Zonooz
CPC classification number: G06V20/58 , G06V10/82 , G06V20/41 , G06V2201/07
Abstract: A computer-implemented method that provides a novel shape aware FSL framework, referred to as LSFSL. In addition to the inductive biases associated with deep learning models, the method of the current invention introduces meaningful shape bias. The method of the current invention comprises the step of capturing the human behavior of recognizing objects by utilizing shape information. The shape information is distilled to address the texture bias of CNN-based models. During training, the model has two branches: RIN-branch, network with colored images as input, preferably RGB images, and SIN-branch, network with shape semantic-based input. Each branch incorporates a CNN backbone followed by a fully connected layer performing classification. RIN-branch and SIN-branch receive the RGB input image and shape information enhanced RGB input image, respectively. The training objective is to improve the classification performance of the RIN-branch and SIN-branch as well as to distill shape semantics from SIN-branch to RIN-branch. The features of the RIN-branch and SIN-branch are aligned to distill shape representation into RIN-branch. This feature alignment implicitly achieves a bias-alignment between the RIN and SIN. The learned representations are generic and remain invariant to common attributes.
-
-
-
-
-
-
-
-
-