Abstract:
The disclosed generative composition system produces a composition to a briefing that describes a musical journey in emotional descriptions. The composition is assembled from concatenated interchangeable Form Atoms “FAs” selectable by tags aligning emotional descriptions with respective compositional heuristics. Each FA has self-contained constructional properties representative of an historical musical corpus. These heuristics support generation of chords, in chord schemes of musical tonics, achieving an equivalent form function. Each FA also includes chord spacer heuristics that temporally space generated chords across a defined musical window, and a chord list in a local tonic defining branching structures giving options for generating different chords. A progression descriptor, in combination with a form function, expresses musically a question, an answer or a statement, with each FA creating a meta-map of a chord scheme for a musical section. Musical transitions between FA reflect groupings in which FA have similar tags but different constructional properties.
Abstract:
A generative composition system reduces existing musical artefacts to constituent elements termed “Form Atoms”. These Form Atoms may each be of varying length and have musical properties and associations that link together through Markov chains. To provide myriad new composition, a set of heuristics ensures that musical textures between concatenated musical sections follow a supplied and defined briefing narrative for the new composition whilst contiguous concatenated Form Atoms are also automatically selected to see that similarities in respective and identified attributes of musical textures for those musical sections are maintained to support maintenance of musical form. Independent aspects of the disclosure further ensure that, within the composition work, such as a media product or a real-time audio stream, chord spacing determination and control are practiced to maintain musical sense in the new composition. Further, a structuring of primitive heuristics operates to maintain pitch and permit key transformation. The system and its functionality provides signal analysis and music generation through allowing emotional connotations to be specified and reproduced from cross-referenced Form-Atoms.
Abstract:
Embodiments of the present invention relate to automatically identifying structures of a music stream. A segment structure may be generated that visually indicates repeating segments of a music stream. To generate a segment structure, a feature that corresponds to a music attribute from a waveform corresponding to the music stream is extracted from a waveform, such as an input signal. Utilizing a signal segmentation algorithm, such as a Variable Markov Oracle (VMO) algorithm, a symbolized signal, such as a VMO structure, is generated. From the symbolized signal, a matrix is generated. The matrix may be, for instance, a VMO-SSM. A segment structure is then generated from the matrix. The segment structure illustrates a segmentation of the music stream and the segments that are repetitive.
Abstract:
Digital signal processing and machine learning techniques can be employed in a vocal capture and performance social network to computationally generate vocal pitch tracks from a collection of vocal performances captured against a common temporal baseline such as a backing track or an original performance by a popularizing artist. In this way, crowd-sourced pitch tracks may be generated and distributed for use in subsequent karaoke-style vocal audio captures or other applications. Large numbers of performances of a song can be used to generate a pitch track. Computationally determined pitch trackings from individual audio signal encodings of the crowd-sourced vocal performance set are aggregated and processed as an observation sequence of a trained Hidden Markov Model (HMM) or other statistical model to produce an output pitch track.
Abstract:
Music recognition is carried out by accepting a musical score of musical elements in a digital format, transforming the digital format into a composite musical data object that models the musical score, defining the key signatures in the composite musical data object probabilistically, computing start times to play musical elements in respective measures of the composite musical data object without regard to rhythmic values of other musical elements in the respective measures, and generating an output including the defined key signatures and computed start times.
Abstract:
A sound signal analysis apparatus 10 includes sound signal input portion for inputting a sound signal indicative of a musical piece, tempo detection portion for detecting a tempo of each of sections of the musical piece by use of the input sound signal, judgment portion for judging stability of the tempo, and control portion for controlling a certain target in accordance with a result judged by the judgment portion.
Abstract:
Methods and systems for non-negative hidden Markov modeling of signals are described. For example, techniques disclosed herein may be applied to signals emitted by one or more sources. In some embodiments, methods and systems may enable the separation of a signal's various components. As such, the systems and methods disclosed herein may find a wide variety of applications. In audio-related fields, for example, these techniques may be useful in music recording and processing, source extraction, noise reduction, teaching, automatic transcription, electronic games, audio search and retrieval, and many other applications.
Abstract:
A music analysis apparatus has a feature extractor and an analysis processor. The feature extractor generates a time series of feature values from a sequence of notes which is designated as an object of analysis. The analysis processor computes an evaluation index value which indicates a probability that the designated sequence of notes is present in each of a plurality of reference music pieces by applying a probabilistic model to the time series of the feature values generated from the designated sequence of notes. The probabilistic model is generated by machine learning of the plurality of reference music pieces using time series of feature values obtained from the reference music pieces.
Abstract:
The method according to the invention includes the creation of a reference multimedia sequence structure, the breaking down of this structure into basic components (tracks P1, P2, Pn) each containing a series of basic subcomponents (bricks B11-Bn4), the association to each one of these basic subcomponents of a plurality of homologous subcomponents (homologous bricks B11 Hi, B21 Hj, B″1 Hk) to each of which are assigned attributes and an automatic composition phase of a new multimedia sequence containing the maintaining of the subcomponents or their replacing with homologous subcomponents chosen algorithmically according to an algorithm determining the probability of the subcomponents of being chosen, considering its attributes, then by performing a random choice in respect of these probabilities.
Abstract:
Variation over time in fundamental frequency in singing voices is separated into a melody-dependent component and a phoneme-dependent component, modeled for each of the components and stored into a singing synthesizing database. In execution of singing synthesis, a pitch curve indicative of variation over time in fundamental frequency of the melody is synthesized in accordance with an arrangement of notes represented by a singing synthesizing score and the melody-dependent component, and the pitch curve is corrected, for each of pitch curve sections corresponding to phonemes constituting lyrics, using a phoneme-dependent component model corresponding to the phoneme. Such arrangements can accurately model a singing expression, unique to a singing person and appearing in a melody singing style of the person, while taking into account phoneme-dependent pitch variation, and thereby permits synthesis of singing voices that sound more natural.