Abstract:
A drone system and method. Audio signals are received via one or more microphones positioned relative to a location on a drone and one or more of the audio signals are identified as of interest. Flight characteristics of the drone are then controlled based on the audio signals that are of interest.
Abstract:
A multi-channel sound (MCS) system features intelligent calibration (e.g., of acoustic echo cancellation (AEC)) for use in dynamic acoustic environments. A sensor subsystem is utilized to detect and identify changes in the acoustic environment and determine a “scene” corresponding to the resulting acoustic characteristics for that environment. This detected scene is compared to predetermined scenes corresponding to the acoustic environment. Each predetermined scene has a corresponding pre-tuned filter configuration for optimal AEC performance. Based on the results of the comparison, the pre-tuned filter configuration corresponding to the predetermined scene that most closely matches the detected scene is utilized by the AEC subsystem of the multi-channel sound system.
Abstract:
A multichannel acoustic system (MAS) comprises an arrangement of microphones, loudspeakers, and filters along with a multichannel acoustic processor (MAP) and other components to together provide and enhance the auditory experience of persons in a shared acoustic space such as, for example, the driver and other passengers in an automobile. Driver-specific features such as navigation and auditory feedback cues are described, as individual auditory customizations and collective communications both within the shared acoustic space as well as with other individuals not located in the space via enhanced conference call facilities.
Abstract:
A drone system and method. Audio signals are received via one or more microphones positioned relative to a location on a drone and one or more of the audio signals are identified as of interest. Flight characteristics of the drone are then controlled based on the audio signals that are of interest.
Abstract:
A method of generating audio output includes displaying a graphical user interface (GUI) at a user device. The GUI represents an area having multiple regions and multiple audio capture devices are located in the area. The method also includes receiving audio data from the multiple audio capture devices. The method further includes receiving an input indicating a selected region of the multiple regions. The method also includes generating, at the user device, audio output based on audio data from a subset of the multiple audio capture devices. Each audio capture device in the subset is located in the selected region.
Abstract:
Disclosed is a feature extraction and classification methodology wherein audio data is gathered in a target environment under varying conditions. From this collected data, corresponding features are extracted, labeled with appropriate filters (e.g., audio event descriptions), and used for training deep neural networks (DNNs) to extract underlying target audio events from unlabeled training data. Once trained, these DNNs are used to predict underlying events in noisy audio to extract therefrom features that enable the separation of the underlying audio events from the noisy components thereof.
Abstract:
Techniques for processing directionally-encoded audio to account for spatial characteristics of a listener playback environment are disclosed. The directionally-encoded audio data includes spatial information indicative of one or more directions of sound sources in an audio scene. The audio data is modified based on input data identifying the spatial characteristics of the playback environment. The spatial characteristics may correspond to actual loudspeaker locations in the playback environment. The directionally-encoded audio may also be processed to permit focusing/defocusing on sound sources or particular directions in an audio scene. The disclosed techniques may allow a recorded audio scene to be more accurately reproduced at playback time, regardless of the output loudspeaker setup. Another advantage is that a user may dynamically configure audio data so that it better conforms to the user's particular loudspeaker layouts and/or the user's desired focus on particular subjects or areas in an audio scene.
Abstract:
A method of selectively authorizing access includes obtaining, at an authentication device, first information corresponding to first synthetic biometric data. The method also includes obtaining, at the authentication device, first common synthetic data and second biometric data. The method further includes generating, at the authentication device, second common synthetic data based on the first information and the second biometric data. The method also includes selectively authorizing, by the authentication device, access based on a comparison of the first common synthetic data and the second common synthetic data.
Abstract:
Systems, devices, and methods are described for recognizing and focusing on at least one source of an audio communication as part of a communication including a video image and an audio communication derived from two or more microphones when a relative position between the microphones is known. In certain embodiments, linked audio and video focus areas providing location information for one or more sound sources may each be associated with different user inputs, and an input to adjust a focus in either the audio or video domain may automatically adjust the focus in the another domain.
Abstract:
Apparatus and methods for audio noise attenuation are disclosed. An audio signal analyzer can determine whether an input audio signal received from a microphone device includes a noise signal having identifiable content. If there is a noise signal having identifiable content, a content source is accessed to obtain a copy of the noise signal. An audio canceller can generate a processed audio signal, having an attenuated noise signal, based on comparing the copy of the noise signal to the input audio signal. Additionally or alternatively, data may be communicated on a communication channel to a separate media device to receive at least a portion of the copy of the noise signal from the separate media device, or to receive content-identification data corresponding to the content source.