Abstract:
An example apparatus for detecting objects in video frames includes a receiver to receive a plurality of video frames from a video camera. The apparatus also includes a first still image object detector to receive a first frame of the plurality of video frames and calculate localization information and confidence information for each potential object patch in the first frame. The apparatus further includes a second still image object detector to receive an adjacent frame of the plurality of video frames adjacent to the first frame and calculate localization information and confidence information for each potential object patch in the adjacent frame. The apparatus includes a similarity detector trained to detect paired patches between the first frame and the adjacent frame based on a comparison of the detected potential object patches. The apparatus further includes an enhancer to modify a prediction result for a paired patch in the adjacent frame to a prediction result of a corresponding paired patch in the first frame including a higher confidence score than the prediction result of the paired patch in the adjacent frame.
Abstract:
Methods and apparatus relating to an adaptive partition mechanism with arbitrary tile shape for tile based rendering GPU (Graphics Processing Unit) architecture are described. In an embodiment, the primitive intersection cost value for each atomic tile of an image are determined at least partially based on a vertex element size, a vertex shader length, and a number of the vertices of a primitive of the image. Other embodiments are also disclosed and claimed.
Abstract:
An energy aware framework for computation and communication devices (CCDs) is disclosed. CCDs may support applications, which may participate in energy aware optimization. Such applications may be designed to support execution modes, which may be associated with different computation and communication demands or requirements. An optimization block may collect computation requirement values (CRVM), communication demand values (CDVM), and such other values of each execution mode to perform a specific task(s). The optimization block may collect computation energy cost information (CECIM) and multi-radio communication energy cost information (MCECIM) for each execution mode. Also, the optimization block may collect the workload values of a cloud-side processing device. The optimization block may determine power estimation values (PEV), based on the energy cost values (CECIM), (MCECIM), CRVM, and CDVM. The optimization block may then determine the execution mode or the apparatus best suited to perform the tasks.
Abstract:
An example apparatus for detecting objects in video frames includes a receiver to receive a plurality of video frames from a video camera. The apparatus also includes a first still image object detector to receive a first frame of the plurality of video frames and calculate localization information and confidence information for each potential object patch in the first frame. The apparatus further includes a second still image object detector to receive an adjacent frame of the plurality of video frames adjacent to the first frame and calculate localization information and confidence information for each potential object patch in the adjacent frame. The apparatus includes a similarity detector trained to detect paired patches between the first frame and the adjacent frame based on a comparison of the detected potential object patches. The apparatus further includes an enhancer to modify a prediction result for a paired patch in the adjacent frame to a prediction result of a corresponding paired patch in the first frame including a higher confidence score than the prediction result of the paired patch in the adjacent frame.
Abstract:
Techniques related to training and implementing a bidirectional pairing architecture for object detection are discussed. Such techniques include generating a first enhanced feature map for each frame of a video sequence by processing the frames in a first direction, generating a second enhanced feature map for frame by processing the frames in a second direction opposite the first, and determining object detection information for each frame using the first and second enhanced feature map for the frame.
Abstract:
System and techniques are provided for three-dimension (3D) semantic segmentation. A device for 3D semantic segmentation includes: an interface, to obtain a point cloud data set for a time-ordered sequence of 3D frames, the 3D frames including a current 3D frame and one or more historical 3D frames previous to the current 3D frame; and processing circuitry, to: invoke a first artificial neural network (ANN) to estimate a 3D scene flow field for each of the one or more historical 3D frames by taking the current 3D frame as a reference frame; and invoke a second ANN to: produce an aggregated feature map, based on the reference frame and the estimated 3D scene flow field for each of the one or more historical 3D frames; and perform the 3D semantic segmentation based on the aggregated feature map.
Abstract:
A disclosed example includes selecting, by a mobile computing device, a model description for a predictive analytics model in response to a user-level application request including input data from an application of the mobile computing device, the model description created with a predictive analytics model description language, the model description received from a predictive analytics provider; comparing, by the mobile computing device, first data associated with the user-level application request with second data indicative of digital rights permissions associated with the model description; and executing, by the mobile computing device, an executable associated with the model description without providing the processor circuitry access to the executable and without providing the input data to the predictive analytics provider.
Abstract:
Techniques related to training and implementing a bidirectional pairing architecture for object detection are discussed. Such techniques include generating a first enhanced feature map for each frame of a video sequence by processing the frames in a first direction, generating a second enhanced feature map for frame by processing the frames in a second direction opposite the first, and determining object detection information for each frame using the first and second enhanced feature map for the frame.