Abstract:
A multi-party control unit (MCU) generates, based on audio data streams that represent sounds associated terminal devices, a mixed audio data stream. In addition, the MCU modifies the mixed mono audio data to steganographically embed sub-streams that include representations of the mono audio data streams. A terminal device receives the modified mixed audio data stream. When the terminal device is configured for stereo playback, the terminal device performs an inverse steganographic process to extract, from the mixed audio data stream, the sub-streams. The terminal device generates and outputs multi-channel audio data based on the extracted sub-streams and the mixed audio data stream. When the terminal device is not configured for stereo playback, the terminal device outputs sound based on the mixed audio data stream without extracting the embedded sub-streams.
Abstract:
A particular method includes determining, at a device, a voicing classification of an input signal. The input signal corresponds to an audio signal. The method also includes controlling an amount of an envelope of a representation of the input signal based on the voicing classification. The method further includes modulating a white noise signal based on the controlled amount of the envelope. The method also includes generating a high band excitation signal based on the modulated white noise signal.
Abstract:
Compressibility-based reallocation of initial bit allocations for frames of an audio signal is described. Applications to redundancy-based retransmission of critical frames (e.g., for fixed-bit-rate modes of speech codec operation) are also described.
Abstract:
A method includes extracting a voicing classification parameter of an audio signal and determining a filter coefficient of a low pass filter based on the voicing classification parameter. The method also includes filtering a low-band portion of the audio signal to generate a low-band audio signal and controlling an amplitude of a temporal envelope of the low-band audio signal based on the filter coefficient. The method also includes modulating a white noise signal based on the amplitude of the temporal envelope to generate a modulated white noise signal and scaling the modulated white noise signal based on a noise gain to generate a scaled modulated white noise signal. The method also includes mixing a scaled version of the low-band audio signal with the scaled modulated white noise signal to generate a high-band excitation signal that is used to generate a decoded version of the audio signal.
Abstract:
A method for applying error resiliency based on speed is disclosed. A speed of the wireless communication device is detected. Source-based error resiliency is dynamically applied to a signal based on the speed. The signal is sent on an uplink. In addition to the speed, the source-based error resiliency may be dynamically applied to the signal based on a characteristic of the signal.
Abstract:
Systems, methods, and apparatus as disclosed herein may be implemented to adjust criticality thresholds for speech frames, based on channel conditions. Such a threshold may be used to control retransmission frequency in response to changes in channel state.
Abstract:
A method includes extracting a voicing classification parameter of an audio signal and determining a filter coefficient of a low pass filter based on the voicing classification parameter. The method also includes filtering a low-band portion of the audio signal to generate a low-band audio signal and controlling an amplitude of a temporal envelope of the low-band audio signal based on the filter coefficient. The method also includes modulating a white noise signal based on the amplitude of the temporal envelope to generate a modulated white noise signal and scaling the modulated white noise signal based on a noise gain to generate a scaled modulated white noise signal. The method also includes mixing a scaled version of the low-band audio signal with the scaled modulated white noise signal to generate a high-band excitation signal that is used to generate a decoded version of the audio signal.
Abstract:
A particular method includes determining, at a device, a voicing classification of an input signal. The input signal corresponds to an audio signal. The method also includes controlling an amount of an envelope of a representation of the input signal based on the voicing classification. The method further includes modulating a white noise signal based on the controlled amount of the envelope. The method also includes generating a high band excitation signal based on the modulated white noise signal.
Abstract:
A method for applying error resiliency based on speed is disclosed. A speed of the wireless communication device is detected. Source-based error resiliency is dynamically applied to a signal based on the speed. The signal is sent on an uplink. In addition to the speed, the source-based error resiliency may be dynamically applied to the signal based on a characteristic of the signal.
Abstract:
Systems and methods of performing blind bandwidth extension are disclosed. In an embodiment, a method includes determining, based on a set of low-band parameters of an audio signal, a first set of high-band parameters and a second set of high-band parameters. The method further includes generating a predicted set of high-band parameters based on a weighted combination of the first set of high-band parameters and the second set of high-band parameters.