Abstract:
A method for processing an image including: identifying a first group of keypoints in the image; for each keypoint of the first group, identifying at least one corresponding keypoint local feature related to the each keypoint; for the at least one keypoint local feature, calculating a corresponding local feature relevance probability; calculating a keypoint relevance probability based on the local feature relevance probabilities of the at least one local feature; selecting keypoints, among the keypoints of the first group, having the highest keypoint relevance probabilities to form a second group of keypoints, and exploiting the keypoints of the second group for analyzing the image. The local feature relevance probability calculated for a local feature of a keypoint is obtained by comparing the value assumed by the local feature with a corresponding reference statistical distribution of values of the local feature.
Abstract:
A method includes providing a neural network having a set of weights. The neural network receives an input data structure for generating a corresponding output array according to values of the set of weights. The neural network is trained to obtain a trained neural network. The training includes setting values of the set of weights with a gradient descent algorithm which exploits a cost function including a loss term and a regularization term. The trained neural network is deployed on a device through a communication network, and used by the device. The regularization term is based on a rate of change of elements of the output array caused by variations of the set of weights values.
Abstract:
A method for identifying keypoints in a digital image including a set of pixels. Each pixel has associated thereto a respective value of an image representative parameter. The method includes approximating a filtered image. The filtered image depends on a filtering parameter and includes for each pixel of the image a filtering function that depends on the filtering parameter to calculate a filtered value of the value of the representative parameter of the pixel. The approximating includes: a) generating a set of base filtered images; each base filtered image is the image filtered with a respective value of the filtering parameter; b) for each pixel of at least a subset of the set of pixels, approximating the filtering function by a respective approximation function based on the base filtered images; the approximation function is a function of the filtering parameter within a predefined range of the filtering parameter.
Abstract:
A method for processing an image, including: identifying a group of keypoints in the image; for each keypoint, calculating a corresponding descriptor array including plural array elements, each array element storing values taken by a corresponding color gradient histogram of a respective sub-region of the image in the neighborhood of the keypoint; for each keypoint, subdividing the descriptor array in at least two sub-arrays each including a respective number of elements of the descriptor array, and generating a compressed descriptor array including a corresponding compressed sub-array for each of the at least two sub-arrays, each compressed sub-array obtained by compressing the corresponding sub-array by vector quantization using a respective codebook; exploiting the compressed descriptor arrays of the keypoints for image analysis. For each keypoint of the group, the subdividing is based on correlation relationships among color gradient histograms with values stored in the elements of the descriptor array of each keypoint.
Abstract:
A method for processing an image includes: identifying a group of keypoints in the image; for each keypoint of the group; a) calculating a corresponding descriptor array including a plurality of array elements, each array element storing values taken by a corresponding color gradient histogram of a respective sub-region of the image in the neighborhood of the keypoint; b) generating at least one compressed descriptor array by compressing at least one portion of the descriptor array by vector quantization using a codebook including a plurality of codewords.
Abstract:
A method for processing an image including: identifying a first group of keypoints in the image; for each keypoint of the first group, identifying at least one corresponding keypoint local feature related to the each keypoint; for the at least one keypoint local feature, calculating a corresponding local feature relevance probability; calculating a keypoint relevance probability based on the local feature relevance probabilities of the at least one local feature; selecting keypoints, among the keypoints of the first group, having the highest keypoint relevance probabilities to form a second group of keypoints, and exploiting the keypoints of the second group for analyzing the image. The local feature relevance probability calculated for a local feature of a keypoint is obtained by comparing the value assumed by the local feature with a corresponding reference statistical distribution of values of the local feature.
Abstract:
A method for processing an image, including: identifying a group of keypoints in the image; for each keypoint, calculating a corresponding descriptor array including plural array elements, each array element storing values taken by a corresponding color gradient histogram of a respective sub-region of the image in the neighborhood of the keypoint; for each keypoint, subdividing the descriptor array in at least two sub-arrays each including a respective number of elements of the descriptor array, and generating a compressed descriptor array including a corresponding compressed sub-array for each of the at least two sub-arrays, each compressed sub-array obtained by compressing the corresponding sub-array by vector quantization using a respective codebook; exploiting the compressed descriptor arrays of the keypoints for image analysis. For each keypoint of the group, the subdividing is based on correlation relationships among color gradient histograms with values stored in the elements of the descriptor array of each keypoint.
Abstract:
A method in which a convolutional neural network is configured to receive an input data structure including a group of values corresponding to signal samples and to generate a corresponding classification output indicative of a selected one among plural predefined classes. The convolutional neural network includes an ordered sequence of layers, each configured to receive a corresponding layer input data structure including a group of input values, and generate a corresponding layer output data structure including a group of output values by convolving the layer input data structure with at least one corresponding filter including a corresponding group of weights. The layer input data structure of the first layer of the sequence corresponds to the input data structure. The layer input data structure of a generic layer of the sequence different from the first layer corresponds to the layer output data structure generated by a previous layer in the sequence.
Abstract:
A method for processing an image is proposed. The method comprises identifying a first group of keypoints in the image. For each keypoint of the first group, the method provides for identifying at least one corresponding keypoint local feature related to said each keypoint; for said at least one keypoint local feature, calculating a corresponding local feature relevance probability; calculating a keypoint relevance probability based on the local feature relevance probabilities of said at least one local feature. The method further comprises selecting keypoints, among the keypoints of the first group, having the highest keypoint relevance probabilities to form a second group of keypoints, and exploiting the keypoints of the second group for analyzing the image. The local feature relevance probability calculated for a local feature of a keypoint is obtained by comparing the value assumed by said local feature with a corresponding reference statistical distribution of values of said local feature.
Abstract:
A method (100) for comparing a first video shot (Vs1) comprising a first set of first images (I1(s)) with a second video shot (Vs2) comprising a second set of second images (I2(t)), at least one between the first and the second set comprising at least two images. The method comprises pairing (110) each first image of the first set with each second image of the second set to form a plurality of images pairs (IP(m)), and, for each image pair, carrying out the operations a)-g): a) identifying (120) first interest points in the first image and second interest points in the second image; b) associating (120) first interest points with corresponding second interest points in order to form corresponding interest point matches; c) for each pair of first interest points, calculating (130) the distance therebetween for obtaining a corresponding first length; d) for each pair of second interest points, calculating (130) the distance therebetween for obtaining a corresponding second length; e) calculating a plurality of distance ratios (130), each distance ratio corresponding to a selected pair of interest point matches and being based on a ratio of a first term and a second term or on a ratio of the second term and the first term, said first term corresponding to the distance between the first interest points of said pair of interest point matches and said second term corresponding to the distance between the second interest points of said pair of interest point matches; f) computing (140) a first representation of the statistical distribution of the plurality of calculated distance ratios; g) computing (150) a second representation of the statistical distribution of distance ratios obtained under the hypothesis that all the interest point matches in the image pair are outliers. The method further comprises generating (160) a first global representation of the statistical distribution of the plurality of calculated distance ratios computed for all the image pairs based on the first representations of all the image pairs; generating (170) a second global representation of the statistical distribution of distance ratios obtained under the hypothesis that all the interest point matches in all the image pairs are outliers based on the second representations of all the image pairs; comparing (180) said first global representation with said second global representation, and assessing (190) whether the first video shot contains a view of an object depicted in the second video shot based on said comparison.