We investigate a data-driven approach to the analysis and transcription of polyphonic music, using a probabilistic model which is able to find sparse linear decompositions of a sequence of short-term Fourier spectra. The resulting system represents each input spectrum as a weighted sum of a small number of "atomic" spectra chosen from a larger dictionary; this dictionary is, in turn, learned from the data in such a way as to represent the given training set in an (information theoretically) efficient way. When exposed to examples of polyphonic music, most of the dictionary elements take on the spectral characteristics of individual notes in the music, so that the sparse decomposition can be used to identify the notes in a polyphonic mixture. Our approach differs from other methods of polyphonic analysis based on spectral decomposition by combining all of the following: (a) a formulation in terms of an explicitly given probabilistic model, in which the process estimating which notes are present corresponds naturally with the inference of latent variables in the model; (b) a particularly simple generative model, motivated by very general considerations about efficient coding, that makes very few assumptions about the musical origins of the signals being processed; and (c) the ability to learn a dictionary of atomic spectra (most of which converge to harmonic spectral profiles associated with specific notes) from polyphonic examples alone-no separate training on monophonic examples is required.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TNN.2005.861031 | DOI Listing |
IEEE Trans Cybern
November 2024
Predominant instrument recognition plays a vital role in music information retrieval. This task involves identifying and categorizing the dominant instruments present in a piece of music based on their distinctive time-frequency characteristics and harmonic distribution. Existing predominant instrument recognition approaches mainly focus on learning implicit mappings (such as deep neural networks) from time-domain or frequency-domain representations of music audio to instrument labels.
View Article and Find Full Text PDFPLoS One
April 2024
MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Penrith, Australia.
Music ensemble performance provides an ecologically valid context for investigating leadership dynamics in small group interactions. Musical texture, specifically the relative salience of simultaneously sounding ensemble parts, is a feature that can potentially alter leadership dynamics by introducing hierarchical relationships between individual parts. The present study extended previous work on quantifying interpersonal coupling in musical ensembles by examining the relationship between musical texture and leader-follower relations, operationalised as directionality of influence between co-performers' body motion in concert video recordings.
View Article and Find Full Text PDFFront Psychol
February 2024
Rhythmic Music Department, Faculty of Arts and Creative Technologies, Vilnius University of Applied Sciences, Vilnius, Lithuania.
Lithuanian traditional polyphonic songs, known as , are characterized by a distinctive musical language and have almost no analogues in world music. The aim of this article is to explore the peculiarities of their musical language and the socio-cultural context of their performance tradition in order to reveal their archaic origins. The archaic nature of songs is shown not by individual features of their musical language, but by the totality of these features, the peculiarities of their poetics, and performance traditions.
View Article and Find Full Text PDFJ Acoust Soc Am
October 2023
School of Physics, Engineering and Technology (Retired), University of York, York, YO10 5DD, United Kingdom.
Multiple fundamental frequency estimation has been extensively used in applications such as melody extraction, music transcription, instrument identification, and source separation. This paper presents an approach based on the iterative detection and extraction of note events, which are considered to be harmonic sounds characterised by a continuous pitch trajectory. Note events are assumed to be associated with musical notes being played by a single instrument, and their pitch trajectories are iteratively estimated.
View Article and Find Full Text PDFNeurosci Lett
January 2024
School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China.
Music-oriented auditory attention detection (AAD) aims at determining which instrument in polyphonic music a listener is paying attention to by analyzing the listener's electroencephalogram (EEG). However, the existing linear models cannot effectively mimic the nonlinearity of the human brain, resulting in limited performance. Thus, a nonlinear music-oriented AAD model is proposed in this paper.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!