Abstract:
Computer-based systems, devices, and methods for assigning mood labels to musical compositions are described. A mood classifier is trained based on mood-labeled musically-coherent segments of musical compositions and subsequently applied to automatically assign mood labels to musically-coherent segments of musical compositions. In both cases, the musically-coherent segments are generated using automated segmentation algorithms.
Abstract:
Methods, systems, and apparatus, including medium-encoded computer program products, for receiving outputs from a plurality of models that are each informed by real-time data provided by one or more sensors that are present in an aquaculture environment. An input is generated for an algorithmic music composer for algorithmically composing music that reflects multiple current conditions within the aquaculture environment, based at least on the received outputs from the plurality of models. The input is provided to the algorithmic music composer to algorithmically compose the music that reflects the multiple current conditions within the aquaculture environment.
Abstract:
Conventionally, significant time and effort are required to construct audio tracks. Disclosed embodiments enable automation of audio tracks using templates that associate sound generator(s) with template section(s). Each template enables a model to automatically generate unique audio tracks in which the sections and/or sounds are probabilistically determined. Certain embodiments introduce additional variability into the automated generation of audio tracks. In addition, the model may generate the audio tracks, note by note, to ensure that no copyrights are infringed.
Abstract:
A sound signal processing apparatus which is capable of correctly detecting expression modes and expression transitions of a song or performance from an input sound signal. A sound signal produced by performance or singing of musical tones is input and divided into frames of predetermined time periods. Characteristic parameters of the input sound signal are detected on a frame-by-frame basis. An expression determining process is carried out in which a plurality of expression modes of a performance or song are modeled as respective states, the probability that a section including a frame or a plurality of continuous frames lies in a specific state is calculated with respect to a predetermined observed section based on the characteristic parameters, and the optimum route of state transition in the predetermined observed section is determined based on the calculated probabilities so as to determine expression modes of the sound signal and lengths thereof.
Abstract:
A metaverse application performs an audio rollback of a local game state by receiving user input from a user during gameplay of a virtual experience. The metaverse application renders a first game state of gameplay of the virtual experience on the user device based on the user input. The metaverse application receives information about a second game state of gameplay of the virtual experience from a server. The metaverse application determines that there is a discrepancy between the first game state and the second game state. The metaverse application determines an audio gap in the first game state where a modification to game audio is to be inserted. The metaverse application generates replacement audio, wherein a duration of the replacement audio matches a duration of the audio gap. The metaverse application renders a corrected game state on the user device that includes the replacement audio.
Abstract:
Musical catalog amplification services that leverage or deploy a computer-based musical composition system are described. The computer-based musical composition system employs algorithms and, optionally, artificial intelligence to generate new music based on analyses of existing music. The new music may be wholly distinctive from, or may include musical variations of, the existing music. Rights in the new music generated by the computer-based musical composition system are granted to the rights holder(s) of the existing music. In this way, the musical catalog(s) of the rights holder(s) is/are amplified to include additional music assets. The computer-based musical composition system may be tuned so that the new music sounds more like, or less like, the existing music of the rights holder(s). Revenues generated from the new music are shared between the musical catalog amplification service provider and the rights holder(s).
Abstract:
A “Music Mapper” automatically constructs a set coordinate vectors for use in inferring similarity between various pieces of music. In particular, given a music similarity graph expressed as links between various artists, albums, songs, etc., the Music Mapper applies a recursive embedding process to embed each of the graphs music entries into a multi-dimensional space. This recursive embedding process also embeds new music items added to the music similarity graph without reembedding existing entries so long a convergent embedding solution is achieved. Given this embedding, coordinate vectors are then computed for each of the embedded musical items. The similarity between any two musical items is then determined as either a function of the distance between the two corresponding vectors. In various embodiments, this similarity is then used in constructing music playlists given one or more random or user selected seed songs or in a statistical music clustering process.
Abstract:
A “Music Mapper” automatically constructs a set coordinate vectors for use in inferring similarity between various pieces of music. In particular, given a music similarity graph expressed as links between various artists, albums, songs, etc., the Music Mapper applies a recursive embedding process to embed each of the graphs music entries into a multi-dimensional space. This recursive embedding process also embeds new music items added to the music similarity graph without reembedding existing entries so long a convergent embedding solution is achieved. Given this embedding, coordinate vectors are then computed for each of the embedded musical items. The similarity between any two musical items is then determined as either a function of the distance between the two corresponding vectors. In various embodiments, this similarity is then used in constructing music playlists given one or more random or user selected seed songs or in a statistical music clustering process.
Abstract:
A sound generation method that is realized by a computer includes receiving a representative value of a musical feature amount for each of a plurality of sections of a musical note, and using a trained model to process a first feature amount sequence in accordance with the representative value for each section, thereby generating a sound data sequence corresponding to a second feature amount sequence in which the musical feature amount changes continuously.
Abstract:
The present disclosure describes techniques for controllable music generation. The techniques comprise extracting latent vectors from unlabelled data, the unlabelled data comprising a plurality of music note sequences, the plurality of music note sequences indicating a plurality of pieces of music; clustering the latent vectors into a plurality of classes corresponding to a plurality of music styles; generating a plurality of labelled latent vectors corresponding to the plurality of music styles, each of the plurality labelled latent vectors comprising information indicating features of a corresponding music style; and generating a first music note sequence indicating a first piece of music in a particular music style among the plurality of music styles based at least in part on a particular labelled latent vector among the plurality of labelled latent vectors, the particular labelled latent vector corresponding to the particular music style.