Abstract:
A method of detecting a user's voice activity in a mobile device is described herein. The method starts with a voice activity detector (VAD) generating a VAD output based on (i) acoustic signals received from microphones included in the mobile device and (ii) data output by an inertial sensor that is included in an earphone portion of the mobile device. The inertial sensor may detect vibration of the user's vocal chords modulated by the user's vocal tract based on vibrations in bones and tissue of the user's head. A noise suppressor may then receive the acoustic signals from the microphones and the VAD output and suppress the noise included in the acoustic signals received from the microphones based on the VAD output. The method may also include steering one or more beamformers based on the VAD output. Other embodiments are also described.
Abstract:
Electronic system for audio noise processing and noise reduction comprises: first and second noise estimators, selector and attenuator. First noise estimator processes first audio signal from voice beamformer (VB) and generate first noise estimate. VB generates first audio signal by beamforming audio signals from first and second audio pick-up channels. Second noise estimator processes first and second audio signal from noise beamformer (NB), in parallel with first noise estimator and generates second noise estimate. NB generates second audio signal by beamforming audio signals from first and second audio pick-up channels. First and second audio signals include frequencies in first and second frequency regions. Selector's output noise estimate may be a) second noise estimate in the first frequency region, and b) first noise estimate in the second frequency region. Attenuator attenuates first audio signal in accordance with output noise estimate. Other embodiments are also described.
Abstract:
A method of improving voice quality in a mobile device starts by receiving acoustic signals from microphones included in earbuds and the microphone array included on a headset wire. The headset may include the pair of earbuds and the headset wire. An output from an accelerometer that is included in the pair of earbuds is then received. The accelerometer may detect vibration of the user's vocal chords filtered by the vocal tract based on vibrations in bones and tissue of the user's head. A spectral mixer included in the mobile device may then perform spectral mixing of the scaled output from the accelerometer with the acoustic signals from the microphone array to generate a mixed signal. Performing spectral mixing includes scaling the output from the inertial sensor by a scaling factor based on a power ratio between the acoustic signals from the microphone array and the output from the inertial sensor. Other embodiments are also described.
Abstract:
A conversation detector processes microphone signals and other sensor signals of a headphone to declare a conversation and configures a filter block to activate a transparency audio signal. It then declares an end to the conversation based on processing one or more of the microphone signals and the other sensor signals, and in response deactivates the transparency audio signal. The conversation detector monitors an idle duration in which an OVAD and a TVAD are both or simultaneously indicating no activity and declares the end to the conversation in response to the idle duration being longer than an idle threshold. Other aspects are also described and claimed.
Abstract:
System for automatic right-left ear detection for headphone comprising: first earcup and second earcup that are identical. Each of first and second earcups includes: first microphone located on perimeter of each earcup, when first earcup is worn on user's right ear first microphone of first earcup is at location farthest from user's mouth when headphone is worn in normal wear position; second microphone located on perimeter of each earcup, when first earcup is worn on user's right ear, second microphone of first earcup is at location closer than first microphone of first earcup to user's mouth; third microphone located inside each earcup facing user's ear cavity, fourth microphone located at perimeter and bottom center portion of each earcup and facing exterior of each earcup, and fifth microphone located on perimeter of each earcup above and to left of second microphone when looking at outside housing of each earcup.
Abstract:
Systems, methods, devices and non-transitory, computer-readable storage mediums are disclosed for location-tracking wireless devices. In an embodiment, a method performed by an electronic device comprises: playing, or initiating the playing of, a sound through a loudspeaker of an accessory device via a communication link. The sound is played at a specified frequency that utilizes a frequency response of the loudspeaker (or loudspeaker plus speaker enclosure). The sound is received through two or more microphones of the electronic device and filtered by one or more filters. The one or more filters are configured to pass the sound at or around the specified frequency and to reduce masking of the sound by ambient noise. The filtered sound is associated with direction data generated from sensor data provided by one or more inertial sensors of the electronic device. In another embodiment, the specified frequency is higher than the maximum human hearing range.
Abstract:
A speaker recognition algorithm is trained (one or more of its models are tuned) with samples of a microphone signal produced by an inside microphone of a headphone, while the headphone is worn by a speaker. The trained speaker recognition algorithm then tests other samples of the inside microphone signal and produces multiple speaker identification scores for its given models, or a single speaker verification likelihood score for a single given model. Other embodiments are also described and claimed.
Abstract:
An electronic device that can be worn by a user can include a processing unit and one or more sensors operatively connected to the processing unit. The processing unit can be adapted to determine an installation position of the electronic device based on one or more signals received from at least one sensor.
Abstract:
Systems, methods, devices and non-transitory, computer-readable storage mediums are disclosed for location-tracking wireless devices. In an embodiment, a method performed by an electronic device comprises: playing, or initiating the playing of, a sound through a loudspeaker of an accessory device via a communication link. The sound is played at a specified frequency that utilizes a frequency response of the loudspeaker (or loudspeaker plus speaker enclosure). The sound is received through two or more microphones of the electronic device and filtered by one or more filters. The one or more filters are configured to pass the sound at or around the specified frequency and to reduce masking of the sound by ambient noise. The filtered sound is associated with direction data generated from sensor data provided by one or more inertial sensors of the electronic device. In another embodiment, the specified frequency is higher than the maximum human hearing range.
Abstract:
An audio system has a housing in which are integrated a number of microphones. A programmed processor accesses the microphone signals and produces a number of acoustic pick up beams based groups of microphones, an estimation of voice activity and an estimation of noise characteristics on each beam. Two or more beams including a voice beam that is used to pick up a desired voice and a noise beam that is used to provide information to estimate ambient noise are adaptively selected from among the plurality of beams, based on thresholds for voice separation and thresholds for noise-matching. Other embodiments are also described and claimed.