Rendering scene-aware audio using neural network-based acoustic analysis

    公开(公告)号:US11190898B2

    公开(公告)日:2021-11-30

    申请号:US16674924

    申请日:2019-11-05

    Applicant: Adobe Inc.

    Abstract: Methods, systems, and non-transitory computer readable storage media are disclosed for rendering scene-aware audio based on acoustic properties of a user environment. For example, the disclosed system can use neural networks to analyze an audio recording to predict environment equalizations and reverberation decay times of the user environment without using a captured impulse response of the user environment. Additionally, the disclosed system can use the predicted reverberation decay times with an audio simulation of the user environment to optimize material parameters for the user environment. The disclosed system can then generate an audio sample that includes scene-aware acoustic properties based on the predicted environment equalizations, material parameters, and an environment geometry of the user environment. Furthermore, the disclosed system can augment training data for training the neural networks using frequency-dependent equalization information associated with measured and synthetic impulse responses.

    Generating synthetic acoustic impulse responses from an acoustic impulse response

    公开(公告)号:US11074925B2

    公开(公告)日:2021-07-27

    申请号:US16682961

    申请日:2019-11-13

    Applicant: Adobe Inc.

    Inventor: Nicholas Bryan

    Abstract: The disclosure describes one or more embodiments of an impulse response system that generates accurate and realistic synthetic impulse responses. For example, given an acoustic impulse response, the impulse response system can generate one or more synthetic impulse responses that modify the direct-to-reverberant ratio (DRR) of the acoustic impulse response. As another example, the impulse response system can generate one or more synthetic impulse responses that modify the reverberation time (e.g., T60) of the acoustic impulse response. Further, utilizing the synthetic impulse responses, the impulse response system can perform a variety of functions to improve a digital audio recording or acoustic measurement or prediction model.

    RENDERING SCENE-AWARE AUDIO USING NEURAL NETWORK-BASED ACOUSTIC ANALYSIS

    公开(公告)号:US20210136510A1

    公开(公告)日:2021-05-06

    申请号:US16674924

    申请日:2019-11-05

    Applicant: Adobe Inc.

    Abstract: Methods, systems, and non-transitory computer readable storage media are disclosed for rendering scene-aware audio based on acoustic properties of a user environment. For example, the disclosed system can use neural networks to analyze an audio recording to predict environment equalizations and reverberation decay times of the user environment without using a captured impulse response of the user environment. Additionally, the disclosed system can use the predicted reverberation decay times with an audio simulation of the user environment to optimize material parameters for the user environment. The disclosed system can then generate an audio sample that includes scene-aware acoustic properties based on the predicted environment equalizations, material parameters, and an environment geometry of the user environment. Furthermore, the disclosed system can augment training data for training the neural networks using frequency-dependent equalization information associated with measured and synthetic impulse responses.

    Generating scene-aware audio using a neural network-based acoustic analysis

    公开(公告)号:US11812254B2

    公开(公告)日:2023-11-07

    申请号:US17515918

    申请日:2021-11-01

    Applicant: Adobe Inc.

    CPC classification number: H04S7/305 G06N3/04 G06N3/08 H04S7/307 H04S2400/11

    Abstract: Methods, systems, and non-transitory computer readable storage media are disclosed for rendering scene-aware audio based on acoustic properties of a user environment. For example, the disclosed system can use neural networks to analyze an audio recording to predict environment equalizations and reverberation decay times of the user environment without using a captured impulse response of the user environment. Additionally, the disclosed system can use the predicted reverberation decay times with an audio simulation of the user environment to optimize material parameters for the user environment. The disclosed system can then generate an audio sample that includes scene-aware acoustic properties based on the predicted environment equalizations, material parameters, and an environment geometry of the user environment. Furthermore, the disclosed system can augment training data for training the neural networks using frequency-dependent equalization information associated with measured and synthetic impulse responses.

    GENERATING SCENE-AWARE AUDIO USING A NEURAL NETWORK-BASED ACOUSTIC ANALYSIS

    公开(公告)号:US20220060842A1

    公开(公告)日:2022-02-24

    申请号:US17515918

    申请日:2021-11-01

    Applicant: Adobe Inc.

    Abstract: Methods, systems, and non-transitory computer readable storage media are disclosed for rendering scene-aware audio based on acoustic properties of a user environment. For example, the disclosed system can use neural networks to analyze an audio recording to predict environment equalizations and reverberation decay times of the user environment without using a captured impulse response of the user environment. Additionally, the disclosed system can use the predicted reverberation decay times with an audio simulation of the user environment to optimize material parameters for the user environment. The disclosed system can then generate an audio sample that includes scene-aware acoustic properties based on the predicted environment equalizations, material parameters, and an environment geometry of the user environment. Furthermore, the disclosed system can augment training data for training the neural networks using frequency-dependent equalization information associated with measured and synthetic impulse responses.

    Audio production assistant for style transfers of audio recordings using one-shot parametric predictions

    公开(公告)号:US11082789B1

    公开(公告)日:2021-08-03

    申请号:US15931505

    申请日:2020-05-13

    Applicant: Adobe Inc.

    Abstract: One example method involves operations for receiving input to transform audio to a target style. Operations further include providing the audio to a predictive model trained to transform the audio into produced audio. Training the predictive model includes accessing representations of audios and unpaired audios. Further, training includes generating feature embeddings by extracting features from representations of an audio and an unpaired audio. The unpaired audio includes a reference production style, and the feature embeddings correspond to their representations. Training further includes generating a feature vector by comparing the feature embeddings using a comparison model. Further, training includes computing prediction parameters using a learned function. The prediction parameters can transform the feature vector into the reference style. Training further includes updating the predictive model with the prediction parameters. In addition, operations include generating the produced audio by modifying audio effects of the audio into the target style.

Patent Agency Ranking