Abstract:
Apparatus and methods for a Motion Picture Experts Group (MPEG) media transport (MMT) device for sending or receiving media content is provided. The MMT device comprises a memory unit and a controller. The controller is configured to measure an available bandwidth, receive fragments of a first encoded stream of the media content and fragments of a second encoded stream of the same media content based on the available bandwidth and a current transmission bandwidth. A mixed encoded stream comprises the fragments received from the first encoded stream and the fragments received from the second encoded stream.
Abstract:
Apparatus and methods for a Motion Picture Experts Group (MPEG) media transport (MMT) device for sending or receiving media content is provided. The MMT device comprises a memory unit and a controller. The controller is configured to measure an available bandwidth, receive fragments of a first encoded stream of the media content and fragments of a second encoded stream of the same media content based on the available bandwidth and a current transmission bandwidth. A mixed encoded stream comprises the fragments received from the first encoded stream and the fragments received from the second encoded stream.
Abstract:
A wireless communication device includes a processor configured to execute an image query. The image query utilizes cluster selection criteria for a cluster-aggregation based vectorization of a set of local features based on a quantity of top local features having the highest posteriori probability values. The cluster selection criterion is measured as the summation of the posteriori probability values of the top local features. The quantity of top local features is determined by a predetermined integer value greater than one.
Abstract:
A video decoder performs a method of dynamic voltage/frequency scaling (DVFS) for video processing. The method includes parsing a bitstream associated with a video to determine embedded information associated with a complexity of a first portion of the video, estimating the complexity of the first portion of the video using the embedded information, determining a voltage and frequency to be used for decoding the first portion of the video based on the estimated complexity, and decoding the first portion of the video at the determined voltage and frequency.
Abstract:
A method includes calculating a complexity value for each segment or version of multimedia content. The multimedia content has multiple versions associated with common content and different encoding parameters, and each version is divided into segments. The method also includes adaptively streaming the multimedia content to a client device based on at least some of the calculated complexity values. The complexity value for each segment or version is calculated by summing multiple operation complexity values defining complexities of operations for decoding the segment or version. The operation complexity values could include values for a bit parsing operation, a buffer initialization operation, an intra-prediction operation, an inverse quantization and transform operation, a reference block fetch operation, a fractional interpolation operation, and in-loop deblocking operation, a sample adaptive offset correction operation, an adaptive loop filtering operation, and a de-ringing filtering operation.
Abstract:
A video decoder performs a method of dynamic voltage/frequency scaling (DVFS) for video processing. The method includes parsing a bitstream associated with a video to determine embedded information associated with a complexity of a first portion of the video, estimating the complexity of the first portion of the video using the embedded information, determining a voltage and frequency to be used for decoding the first portion of the video based on the estimated complexity, and decoding the first portion of the video at the determined voltage and frequency.
Abstract:
An image-encoding system that is configured to generate an output stream based on an input image is provided that includes an encoder and a metadata extractor. The encoder is configured to encode a low-resolution image to generate a quantized, low-resolution image. The low-resolution image is generated based on the input image. The metadata extractor is configured to extract super-resolution (SR) metadata from the input image. The output stream comprises the quantized, low-resolution image and the SR metadata. An image-decoding system is configured to receive the output stream. The image-decoding system includes a decoder and an SR processor. The decoder is configured to decode the quantized, low-resolution image to generate a decoded image. The super-resolution processor is configured to perform super-resolution on the decoded image based on the SR metadata to generate a super-resolved image.
Abstract:
A method includes calculating a complexity value for each segment or version of multimedia content. The multimedia content has multiple versions associated with common content and different encoding parameters, and each version is divided into segments. The method also includes adaptively streaming the multimedia content to a client device based on at least some of the calculated complexity values. The complexity value for each segment or version is calculated by summing multiple operation complexity values defining complexities of operations for decoding the segment or version. The operation complexity values could include values for a bit parsing operation, a buffer initialization operation, an intra-prediction operation, an inverse quantization and transform operation, a reference block fetch operation, a fractional interpolation operation, and in-loop deblocking operation, a sample adaptive offset correction operation, an adaptive loop filtering operation, and a de-ringing filtering operation.
Abstract:
Apparatus and methods are provided to process a downsampled image. The downsampled image is encoded. The downsampled image is upsampled. The downsampled image is filtered in combination with the upsampling to form predictor image. Weights of a spatial weight matrix are based on a spatial scaling ratio.
Abstract:
Segments for a video are transmitted in payload units with an extended network abstraction layer unit (NALU) header within which is embedded display adaptation information that may be employed to control display brightness and thereby reduce power consumption during display of the respective segment. The display adaptation information includes at least a maximum pixel brightness that may be used to scale pixel brightness to maximum and correspondingly reduce backlighting for liquid crystal displays, or to adjust the supply voltage for OLED displays. The display adaptation information may optionally include a minimum pixel brightness, a pixel histogram step size, and an indicator of scaling method.