Abstract:
A method performed by an electronic device is described. The method includes obtaining a combined image. The combined image includes a combination of images captured from one or more image sensors. The method also includes obtaining depth information. The depth information is based on a distance measurement between a depth sensor and at least one object in the combined image. The method further includes adjusting a combined image visualization based on the depth information.
Abstract:
Techniques and systems are provided for performing predictive random access using a background picture. For example, a method of decoding video data includes obtaining an encoded video bitstream comprising a plurality of pictures. The plurality of pictures include a plurality of predictive random access pictures. A predictive random access picture is at least partially encoded using inter-prediction based on at least one background picture. The method further includes determining, for a time instance of the video bitstream, a predictive random access picture of the plurality of predictive random access pictures with a time stamp closest in time to the time instance. The method further includes determining a background picture associated with the predictive random access picture, and decoding at least a portion of the predictive random access picture using inter-prediction based on the background picture.
Abstract:
In a particular embodiment, a method includes evaluating, at a mobile device, a first area of pixels to generate a first result. The method further includes evaluating, at the mobile device, a second area of pixels to generate a second result. Based on comparing a threshold with a difference between the first result and the second result, a determination is made that the second area of pixels corresponds to a background portion of a scene or a foreground portion of the scene.
Abstract:
An apparatus includes an object detector configured to receive image data of a scene viewed from the apparatus and including an object. The image data is associated with multiple scale space representations of the scene. The object detector is configured to detect the object responsive to location data and a first scale space representation of the multiple scale space representations.
Abstract:
A method performed by an electronic device is described. The method includes determining overlapping areas from neighboring images. The method also includes determining a difference measure between the overlapping areas. The method further includes determining a constraint measure corresponding to at least one of the overlapping areas. The method additionally includes determining a seam based on a combination of the difference measure and the constraint measure.
Abstract:
A method for interactive image caricaturing by an electronic device is described. The method includes detecting at least one feature location of an image. The method further includes generating, based on the at least one feature location, an image mesh that comprises a grid of at least one horizontal line and at least one vertical line. The method additionally includes obtaining a gesture input. The method also includes determining at least one caricature action based on the at least one gesture input. The method further includes generating a caricature image based on the image mesh, the at least one caricature action and the image.
Abstract:
Apparatus and methods for facial detection are disclosed. A plurality of images of an observed face is received for identification. Based at least on two or more selected images of the plurality of images, a template of the observed face is generated. In some embodiments, the template is a subspace generated based on feature vectors of the plurality of received images. A database of identities and corresponding facial data of known persons is searched based at least on the template of the observed face and the facial data of the known persons. One or more identities of the known persons are selected based at least on the search.
Abstract:
Embodiments include methods and systems for context-adaptive pixel processing based, in part, on a respective weighting-value for each pixel or a group of pixels. The weighting-values provide an indication as to which pixels are more pertinent to pixel processing computations. Computational resources and effort can be focused on pixels with higher weights, which are generally more pertinent for certain pixel processing determinations.
Abstract:
Systems and techniques are provided for performing video-based activity recognition. For example, a process can include extracting, using a first machine learning model, first one or more features from a first frame and second one or more features from a second frame. The first one or more features and the second one or more features are associated with a person driving a vehicle. The process can include processing, using a second machine learning model, the first one or more features and the second one or more features. The process can include determining, based on processing of the first one or more features and the second one or more features using the second machine learning model, at least one activity associated with the person driving the vehicle.
Abstract:
A method for interactive image caricaturing by an electronic device is described. The method includes detecting at least one feature location of an image. The method further includes generating, based on the at least one feature location, an image mesh that comprises a grid of at least one horizontal line and at least one vertical line. The method additionally includes obtaining a gesture input. The method also includes determining at least one caricature action based on the at least one gesture input. The method further includes generating a caricature image based on the image mesh, the at least one caricature action and the image.