File format for spatial audio
    42.
    发明授权

    公开(公告)号:US11272308B2

    公开(公告)日:2022-03-08

    申请号:US16645422

    申请日:2018-09-25

    Applicant: Apple Inc.

    Abstract: An audio asset library containing audio assets formatted in accordance with a file format for spatial audio includes asset metadata that enables simulated reality (SR) application developers to compose sounds for use in SR applications. The audio assets are formatted to include audio data encoding a sound capable of being composed into a SR application along with asset metadata describing not only how the sound was encoded, but also how a listener in SR environment experiences the sound. A SR developer platform is configured so that developers can compose sound for SR objects using audio assets stored in the audio library, including editing the asset metadata to include transformation parameters that support dynamic transformation of the asset metadata in the SR environment to alter how the SR listener experiences the composed sound. Other embodiments are also described and claimed.

    Feedback coordination for a virtual interaction

    公开(公告)号:US10782779B1

    公开(公告)日:2020-09-22

    申请号:US16580372

    申请日:2019-09-24

    Applicant: Apple Inc.

    Abstract: In some implementations, a method includes: obtaining user movement information, wherein the user movement information characterizes real-world body pose and trajectory information of the user; generating, from real-world user movement information and a predetermined placement of the virtual instrument in the computer generated reality (CGR) environment, a predicted virtual instrument interaction time for a virtual instrument interaction prior to the virtual instrument interaction occurring; determining whether or not the predicted virtual instrument interaction time falls within an acceptable temporal range around one of a plurality of temporal sound markers; and in response to determining that the predicted virtual instrument interaction time falls within the acceptable temporal range around a particular temporal sound marker of the plurality of temporal sound markers, quantizing the virtual instrument interaction by presenting play of the virtual instrument to match the particular temporal sound marker of the plurality of temporal sound markers.

    Processing Sound in an Enhanced Reality Environment

    公开(公告)号:US20240236610A1

    公开(公告)日:2024-07-11

    申请号:US18611407

    申请日:2024-03-20

    Applicant: Apple Inc.

    CPC classification number: H04S7/305 H04S2420/01

    Abstract: Processing sound in an enhanced reality environment can include generating, based on an image of a physical environment, an acoustic model of the physical environment. Audio signals captured by a microphone array, can capture a sound in the physical environment. Based on these audio signals, one or more measured acoustic parameters of the physical environment can be generated. A target audio signal can be processed using the model of the physical environment and the measured acoustic parameters, resulting in a plurality of output audio channels having a virtual sound source with a virtual location. The output audio channels can be used to drive a plurality of speakers. Other aspects are also described and claimed.

    Processing sound in an enhanced reality environment

    公开(公告)号:US11956623B2

    公开(公告)日:2024-04-09

    申请号:US17360823

    申请日:2021-06-28

    Applicant: Apple Inc.

    CPC classification number: H04S7/305 H04S2420/01

    Abstract: Processing sound in an enhanced reality environment can include generating, based on an image of a physical environment, an acoustic model of the physical environment. Audio signals captured by a microphone array, can capture a sound in the physical environment. Based on these audio signals, one or more measured acoustic parameters of the physical environment can be generated. A target audio signal can be processed using the model of the physical environment and the measured acoustic parameters, resulting in a plurality of output audio channels having a virtual sound source with a virtual location. The output audio channels can be used to drive a plurality of speakers. Other aspects are also described and claimed.

    Correcting depth estimations derived from image data using acoustic information

    公开(公告)号:US11947005B2

    公开(公告)日:2024-04-02

    申请号:US17978859

    申请日:2022-11-01

    Applicant: Apple Inc.

    CPC classification number: G01S15/89 G06T7/50 G06T2207/10028

    Abstract: In one implementation, a method includes: obtaining a first depth estimation characterizing a distance between the device and a surface in a real-world environment, wherein the first depth estimation is derived from image data including a representation of the surface; receiving, using the audio transceiver, an acoustic reflection of an acoustic wave, wherein the acoustic wave is transmitted in a known direction relative to the device; and determining a second depth estimation based on the acoustic reflection, wherein the second depth estimation characterizes the distance between the device and the surface in the real-world environment; and determining a confirmed depth estimation characterizing the distance between the device and the surface based on resolving any mismatch between the first depth estimation and the second depth estimation.

Patent Agency Ranking