Abstract:
An electronic device, method, and computer readable medium for 3D association of detected objects are provided. The electronic device includes at least one image sensor, an inertial measurement sensor, a memory, and at least one processor coupled to the at least one image sensor, the inertial measurement sensor, and the memory. The at least one processor is configured to capture an image of an environment using the at least one image sensor, detect an object in the captured image, define a bounded area in the image around the detected object, receive head pose data from the inertial measurement sensor, and determine a location of the detected object in a 3D space using the head pose data and the bounded area in the captured image.
Abstract:
An image and vision processing architecture included a plurality of image processing hardware accelerators each configured to perform a different one of a plurality of image processing operations on image data. A multi-port memory shared by the hardware accelerators stores the image data and is configurably coupled by a sparse crossbar interconnect to one or more of the hardware accelerators depending on a use case employed. The interconnect processes accesses of the image data by the hardware accelerators. Two or more of the hardware accelerators are chained to operate in sequence in a first order for a first use case, and at least one of the hardware accelerators is set to operate for a second use case. Portions of the memory are allocated to the hardware accelerators based on the use case employed, with an allocated portion of the memory configured as a circular buffer.
Abstract:
An electronic device, method, and computer readable medium for foveated storage and processing are provided. The electronic device includes a memory, and a processor coupled to the memory. The processor performs head tracking and eye tracking; generates a foveated image from an original image based on the head tracking and the eye tracking; and stores the foveated image using one of: a tile-based method or a frame-based method.
Abstract:
A wearable device is provided for authentication that includes a memory element and processing circuitry coupled to the memory element. The memory element configured to store a plurality of user profiles. The processing circuitry is configured to identify a pairing between the wearable device and a device. The processing circuitry is configured to identify a user of the wearable device. The processing circuitry also is configured to determine if the identified user matches a profile of the plurality of user profiles. The processing circuitry is also configured to responsive to the identified user matching the profile, determine if the profile provides authorization to access the device. The processing circuitry is also configured to responsive to the profile providing authorization to the device, send a message to the device authorizing access to the device.
Abstract:
A method to map a plurality of feature maps of a neural network onto a memory hierarchy includes mapping a first feature map of the plurality of feature maps to a memory in a memory hierarchy having available memory space and providing quickest access to the first feature map. The method also includes, when the first feature map expires, removing the first feature map from the memory used to store the first feature map.
Abstract:
A method to map a plurality of feature maps of a neural network onto a memory hierarchy includes mapping a first feature map of the plurality of feature maps to a memory in a memory hierarchy having available memory space and providing quickest access to the first feature map. The method also includes, when the first feature map expires, removing the first feature map from the memory used to store the first feature map.
Abstract:
A wearable device is provided for authentication that includes a memory element and processing circuitry coupled to the memory element. The memory element configured to store a plurality of user profiles. The processing circuitry is configured to identify a pairing between the wearable device and a device. The processing circuitry is configured to identify a user of the wearable device. The processing circuitry also is configured to determine if the identified user matches a profile of the plurality of user profiles. The processing circuitry is also configured to responsive to the identified user matching the profile, determine if the profile provides authorization to access the device. The processing circuitry is also configured to responsive to the profile providing authorization to the device, send a message to the device authorizing access to the device.
Abstract:
A method includes, in a first mode, positioning first and second tiltable image sensor modules of an image sensor array of an electronic device so that a first optical axis of the first tiltable image sensor module and a second optical axis of the second tiltable image sensor module are substantially perpendicular to a surface of the electronic device, and the first and second tiltable image sensor modules are within a thickness profile of the electronic device. The method also includes, in a second mode, tilting the first and second tiltable image sensor modules so that the first optical axis of the first tiltable image sensor module and the second optical axis of the second tiltable image sensor module are not perpendicular to the surface of the electronic device, and at least part of the first and second tiltable image sensor modules are no longer within the thickness profile of the electronic device.
Abstract:
An electronic device, method, and computer readable medium for 3D association of detected objects are provided. The electronic device includes at least one image sensor, an inertial measurement sensor, a memory, and at least one processor coupled to the at least one image sensor, the inertial measurement sensor, and the memory. The at least one processor is configured to capture an image of an environment using the at least one image sensor, detect an object in the captured image, define a bounded area in the image around the detected object, receive head pose data from the inertial measurement sensor, and determine a location of the detected object in a 3D space using the head pose data and the bounded area in the captured image.
Abstract:
An image and vision processing architecture included a plurality of image processing hardware accelerators each configured to perform a different one of a plurality of image processing operations on image data. A multi-port memory shared by the hardware accelerators stores the image data and is configurably coupled by a sparse crossbar interconnect to one or more of the hardware accelerators depending on a use case employed. The interconnect processes accesses of the image data by the hardware accelerators. Two or more of the hardware accelerators are chained to operate in sequence in a first order for a first use case, and at least one of the hardware accelerators is set to operate for a second use case. Portions of the memory are allocated to the hardware accelerators based on the use case employed, with an allocated portion of the memory configured as a circular buffer.