Abstract:
A method of generating an output video stream is provided, including receiving a first input video stream from at least one first camera capturing a scene, receiving a second input video stream from a second camera only partially capturing the scene, and generating an output video stream from the first input video stream, including, in response to determining that a particular area of the output video stream depicts a part of the scene depicted in the second input video stream, refraining from, or at least reducing a level of, a masking of the particular area of the output video stream. A corresponding device, camera monitoring system, computer program and computer program product are also provided.
Abstract:
A method of processing a video stream comprises receiving a first/second set of input image frames comprising image data originating from a first/second source of input image frames; encoding the sets of input image frames into first/second sets of encoded image frames, forming a combined video stream, wherein an encoded image frame of the first set of encoded image frames exclusively references one or more other encoded image frames of the first set of encoded image frames, and an encoded image frame of the second set of encoded image frames references both an encoded image frame of the first set of encoded image frames and another encoded image frame of the second set of encoded image frames; and signing, using a first/second key, a first/second signing group comprising one or more encoded image frames of the first/second set of encoded image frames, wherein each key pertains to the respective source.
Abstract:
A method for processing an image in a video device, comprises reading an image and combining the image with metadata related to the image by embedding the metadata in or with the image. The method further includes combining transforming the image and extracting the metadata from the image, before encoding the image in an encoder and utilizing the metadata as input in further processing.
Abstract:
This disclosure relates to methods systems and non-transitory computer-readable storage mediums for distributing load in a multi-chip image processing unit for processing image data into processed image data. An example method comprises receiving first image data, analysing the first image data using a first algorithm, the first algorithm performing a set number of operations for a given size of image data input to the first algorithm, and outputs at least one characteristic of the first image data; using the at least one characteristic to estimate use of memory bandwidth in the first and second chip when processing the first image data into processed image data; and distributing processing of the first image data between the first and the second chip such that the estimated use of memory bandwidth is distributed evenly.
Abstract:
A device and a method mask an object in a video stream. The camera is arranged in a system including the camera and another device. A location and field of view is known for the device and the camera. Furthermore, the field of view of the device and the camera are non-overlapping. Information indicating that an object is approaching the field of view of the camera is obtained. The obtained information is determined from the device indicating a location and a direction of movement of the object and the known locations and fields of view of the camera and the device. In response to the information, a threshold for detecting objects to be masked in the video stream captured by the camera is reduced. An object to be masked in the video stream is detected using the reduced threshold, and masking of the object in the video stream is inserted.
Abstract:
A method of distortion correction in an image captured by a non-rectilinear camera includes obtaining multiple images of a scene captured by the camera over time, determining where bottom portions of objects having moved over a horizontal surface in the scene are located in the images, determining a boundary of the horizontal surface in the scene based on the determined locations of the bottom portions, generating a three-dimensional model of the scene by defining one or more vertical surfaces around the determined boundary of the horizontal surface of the scene, and correcting a distortion of at least one of the images by projecting the image onto the three-dimensional model of the scene. A corresponding device, computer program and computer program product are also provided.
Abstract:
An image processing device, a camera system, a non-transitory computer-readable storage medium, and methods for encoding two video image frames captured by one of two image sensors, wherein each of the video image frames depicts a respective portion of a scene. A respective overlapping area is identified in each of the video image frames, which overlapping areas both depict a same sub-portion of the scene, and a video image frame of the video image frames is selected. Compression levels are then set for the image frames, wherein respective compression levels are set for pixel blocks in the selected video image frame based on a given principle. Respective compression levels for pixel blocks in the overlapping area in the selected video image frame are selectively set higher or lower than respective compression levels that would have been set based on the given principle. The video image frames are then encoded.
Abstract:
An image processing device, a non-transitory computer-readable storage medium and a method for dividing an array of pixels into slices and tiles, and a system for encoding the slices and tiles into a video stream by means of a first encoder arranged in a first image processing circuitry and a second encoder arranged in a second image processing circuitry are disclosed. The array of pixels consists of a first subset of pixels having pixel data stored in a first buffer arranged in the first image processing circuitry and a second subset of pixels having pixel data stored in a second buffer arranged in the second image processing circuitry, wherein a first border between the first subset of pixels and the second subset of pixels is tilted at a tilt angle in relation to the array of pixels. The array of pixels is divided into two or more slices, and each slice is divided into two or more tiles. The two or more tiles of each slice consists of a first subset of tiles and a second subset of tiles, wherein the first subset of tiles and the second subset of tiles are disjunct and the first subset of tiles is to be encoded by the first encoder and the second subset of tiles is to be encoded by the second encoder. Dividing each slice into two or more tiles comprises setting a second border between the first subset of tiles and the second subset of tiles at an offset between two adjacent slices of the two or more slices, wherein the offset is based on the tilt angle.
Abstract:
A method for identifying stationary regions in frames of a video sequence comprises receiving an encoded version of the video sequence, wherein the encoded version of the video sequence includes an intra-coded frame followed by a plurality of inter-coded frames; reading coding-mode information in the inter-coded frames of the encoded version of the video sequence, wherein the coding-mode information is indicative of blocks of pixels in the inter-coded frames being skip-coded; finding, using the read coding-mode information, one or more blocks of pixels that each was skip-coded in a respective plurality of consecutive frames in the encoded version of the video sequence; and designating each found block of pixels as a stationary region in the respective plurality of consecutive frames.
Abstract:
A method for determining authenticity of a video in a surveillance system, whereby a sequence of image frames of a scene is captured, and an object is tracked. A current image quality measure in an image area corresponding to the tracked object is determined in at least a first and second image frame. chosen such that the object has moved at least a predetermined distance between the first and second image frames. A current image quality measure variation for the object is determined, the image quality measure variation describing the image quality measure as a function of position of the object in the image frames. The current image quality measure variation is compared to a known image quality measure variation. In response to the current image quality measure variation deviating from the known pixel density variation by less than a predetermined amount, it is determined that the video is authentic.