Abstract:
Generating notifications comprising text and image data for client devices with limited display screens is disclosed. An image to be included in the notification is resized and reshaped using image processing techniques. The resized image is further analyzed to identify optimal portions for placing the text data. The text data can also be analyzed and shortened for including at the identified portion of resized image to generate a notification. The resulting notification displays the text and image data optimally within the limited screen space of the client device so that a user observing the notification can obtain the information at a glance.
Abstract:
In an embodiment, a method comprises obtaining a frequency domain representation associated with an image; obtaining one or more frequency domain representations of one or more object detection filters; generating a composite frequency domain representation based on the frequency domain representation associated with the image and the one or more frequency domain representations of the one or more object detection filters; and detecting one or more objects in the image based on the composite frequency domain representation. The frequency domain representation associated with the image may be obtained based on a forward transform performed on an image feature description. The image feature description may be obtained based on a feature extraction performed on the image. The one or more frequency domain representations of the one or more object detection filters may be obtained based on one or more Fourier transforms performed on the one or more object detection filters.
Abstract:
Briefly, embodiments of methods and/or systems of generating preference indices for contiguous portions of digital images are disclosed. For one embodiment, as an example, parameters of a neural network may be developed to generate object labels for digital images. The developed parameters may be transferred to a neural network utilized to generate signal sample value levels corresponding to preference indices for contiguous portions of digital images.
Abstract:
Briefly, embodiments of methods and/or systems of training multiclass convolutional neural networks (CNNs) are disclosed. For one embodiment, as an example, an auxiliary CNN may be utilized to form an ensemble with the collection as a linear combination. The linear combination may be based, at least in part, on boost prediction error encountered during the training process.
Abstract:
At least one label prediction model is trained, or learned, using training data that may comprise training instances that may be missing one or more labels. The at least one label prediction model may be used in identifying a content item's ground-truth label set comprising an indicator for each label in the label set indicating whether or not the label is applicable to the content item.
Abstract:
System, method and architecture for providing improved visual recognition by modeling visual content, semantic content and an implicit social network representing individuals depicted in a collection of content, such as visual images, photographs, etc. which network may be determined based on co-occurrences of individuals represented by the content, and/or other data linking the individuals. In accordance with one or more embodiments, using images as an example, a relationship structure may comprise an implicit structure, or network, determined from co-occurrences of individuals in the images. A kernel jointly modeling content, semantic and social network information may be built and used in automatic image annotation and/or determination of relationships between individuals, for example.
Abstract:
Techniques are provided for efficiently identifying relevant product images based on product items detected in a query image. In general, a query image may represent a digital image in any format that depicts a human body and one or more product items. For example, a query image may be an image for display on a webpage, an image captured by a user using a camera device, or an image that is part of a media content item, such as a frame from a video. Product items may be detected in a query image by segmenting the query image into a plurality of image segments and clustering one or more of the plurality image segments into one or more image segment clusters. The resulting image segments and image segment clusters may be used to search for visually similar product images.
Abstract:
In an embodiment, a method comprises obtaining a frequency domain representation associated with an image; obtaining one or more frequency domain representations of one or more object detection filters; generating a composite frequency domain representation based on the frequency domain representation associated with the image and the one or more frequency domain representations of the one or more object detection filters; and detecting one or more objects in the image based on the composite frequency domain representation. The frequency domain representation associated with the image may be obtained based on a forward transform performed on an image feature description. The image feature description may be obtained based on a feature extraction performed on the image. The one or more frequency domain representations of the one or more object detection filters may be obtained based on one or more Fourier transforms performed on the one or more object detection filters.
Abstract:
Disclosed herein is a system and method that facilitate searching and/or browsing of images by &lustering, or grouping, the images into a set of image clusters using facets, such as without limitation visual properties or visual characteristics, of the images, and representing each image cluster by a representative image selected for the image cluster. A map-reduce based probabilistic topic model may be used to identify one or more images belonging to each image cluster and update model parameters.
Abstract:
An approach is provided for acquiring images with camera-enabled mobile devices using objects of interest recognition. A mobile device is configured to acquire an image represented by image data and process the image data to identify a plurality of candidate objects of interest in the image. The plurality of candidate objects of interest may be identified based upon a plurality of low level features or “cues” in the image data. Example cues include, without limitation, color contrast, edge density and superpixel straddling. A particular candidate object of interest is selected from the plurality of candidate objects of interest and a graphical symbol is displayed on a screen of the mobile device to identify the particular candidate object of interest. The particular candidate object of interest may be located anywhere on the image. Passive auto focusing is performed at the location of the particular candidate object of interest.