Hi, I'm Matthias. Welcome to my website and blog!

I'm a lecturer and researcher in the field of music informatics. I currently work as a Royal Academy of Engineering Research Fellow with the Centre for Digital Music at Queen Mary, University of London (see my Queen Mary web page).

Past work places include the Internet music platform Last.fm, where I worked as Research Fellow, the Japanese research centre AIST in Tsukuba, and, as a research student, the Centre for Digital Music. Find more info on my biography page.

My main research interest (and the subject of my PhD thesis) has been the automatic transcription of chords from audio, but I've also done work on segmentation, harpsichord tuning estimation and, recently, lyrics-to-audio alignment. Please do have a look at my publications website to learn more about my work, ask Google Scholar directly, or visit my Software site if you're more interested in just using it.

The Pop Files »

[23 Mrz 2015 | Comments Off | 160 views]
Why not listen to Dancing Queen again? A masterpiece of pop, by one of the most prolific and successful songwriting partnerships the world has ever seen: Benny Andersson and Bjoern Ulvaeus from ABBA. Get your headphones out. So what is Dancing Queen about? It’s about getting nerdy boys onto the dance floor. How does it do that?
Firstly, very clever, suggestive lyrics, secondly an onslaught of never-ending, interlocking melody madness. Well, and yes, there’s some other bits going on as well. But let’s listen together first and then I’ll tell you more.

Listen-through

0:00 Straight in! Instantly recognisable piano [zoom], then into the intro with that instantly recognisable synth tune. Not any old borin…

Conference Paper, Publication »

[16 Feb 2015 | Comments Off | 153 views]
http://schall-und-mauch.de/artificialmusicality/wp-content/uploads/2015/02/notemodelling.png     Publication authored by Tian Cheng and Simon Dixon and Matthias Mauch. We investigate piano acoustics and compare the theoretical temporal decay of individual partials to recordings of real-world piano notes from the RWC Music Database. We first describe the theory behind double decay and beats, known phenomena caused by the interaction between strings and soundboard. Then we fit the decay of the first 30 partials to a standard linear model and two physically-motivated non-linear models that take into account the coupling of strings and soundboard. We show that the use of non-linear models provides a better fit to the data. We use these estimated decay rates to parameterise the characteristic decay response (decay rates along frequencies) of the piano under investigation. The results also show…

Conference Paper, Publication »

[16 Jan 2015 | Comments Off | 189 views]
http://cdn1.editmysite.com/uploads/3/2/1/8/32182799/background-images/1927180276.png     Publication authored by Rachel Bittner and Justin Salamon and Mike Tierney and Matthias Mauch and Chris Cannam and Juan Bello. We introduce MedleyDB: a dataset of annotated, royalty-free multitrack recordings. The dataset was primarily developed to support research on melody extraction, addressing important shortcomings of existing collections. For each song we provide melody f0 annotations as well as instrument activations for evaluating automatic instrument recognition. The dataset is also useful for research on tasks that require access to the individual tracks of a song such as source separation and automatic mixing. In this paper we provide a detailed description of MedleyDB, including curation, annotation, and musical content. To gain insight into the new challenges presented by the dataset, we run a set of experiments using a state-of…

Done and Liked, Featured, Other, Publication »

[11 Jan 2015 | Comments Off | 144 views]
http://schall-und-mauch.de/anatomy-of-the-charts/aotc1_files/137.png   Publication authored by Matthias Mauch. I’ve finally found a new home for the series of 5 blog posts called the “Anatomy of the Charts”, which I wrote in 2011 during my stay as a Research Fellow at Last.fm. In 2011 I’d started thinking about the evolution of music, and Last.fm provided the chance to analyse some really exciting data: more than 15,000 recordings from 50 years of UK charts. There was little scope for properly scientific work then, but Last.fm gave me the freedom to write about this on their blog, and my team lead at the time, Mark Levy, helped as well. This resulted in the series of 5 blog posts investigating different musical dimensions in a data-driven, but entertaining way—or so I hope. Sadly, as time went on, some image and text on La…

Conference Paper, Publication »

[9 Jan 2015 | Comments Off | 135 views]
    Publication authored by Tian Cheng and Simon Dixon and Matthias Mauch.
Recently, we have witnessed an increasing use of the source-filter model in music analysis, which is achieved by integrating the source filter model into a non-negative matrix factorisation (NMF) framework or statistical models. The combination of the source-filter model and NMF framework reduces the number of free parameters needed and makes the model more flexible to extend. This paper compares four extended source-filter models: the source-filter-decay (SFD) model, the NMF with time- frequency activations (NMF-ARMA) model, the multi-excitation (ME) model and the source-filter model based on β-divergence (SFbeta model). The first two models represent the time-varying spectra by adding a loss filter and a tim

Conference Paper, Publication »

[8 Jan 2015 | Comments Off | 141 views]
    Publication authored by Christof Weiß and Matthias Mauch and Simon Dixon.
We propose a novel set of chroma-based audio features inspired by pitch class set theory and show their utility for style analysis of classical music by using them to classify recordings into historical periods. Musicologists have long studied how composers’ styles develop and influence each other, but usually based on manual analyses of the score or, more recently, automatic analyses on symbolic data, both largely independent from timbre. Here, we investigate whether such musical style analyses can be realised using audio features. Based on chroma, our features describe the use of intervals and triads on multiple time scales. To test the efficacy of this approach we use a 1600 track balanced corpus that covers the Baroque, Cl

Journal Paper, Publication »

[1 Jan 2015 | Comments Off | 132 views]
    Publication authored by Peter Foster and Matthias Mauch and Simon Dixon. We propose string compressibility as a descriptor of temporal structure in audio, for the purpose of determining musical similarity. Our descriptors are based on computing track-wise compression rates of quantised audio features, using multiple temporal resolutions and quantisation granularities. To verify that our descriptors capture musically relevant information, we incorporate our descriptors into similarity rating prediction and song year prediction tasks. We base our evaluation on a dataset of 15,500 track excerpts of Western popular music, for which we obtain 7,800 web-sourced pairwise similarity ratings. To assess the agreement among similarity ratings, we perform an evaluation under controlled condit…

from me to you »

[10 Dez 2014 | Comments Off | 150 views]
I’m delighted to be able to give a seminar talk as part of the Statistics Seminar Series in the School of Mathematical Sciences at Queen Mary. It’s happening at 1600 for a start at 1630 on Thursday 11th December, 2014 (location: Mathematics Seminar Room (203) on Level 2 of the Mathematics Building). You can find the abstract here. In essence, it’s an overview of much of my research since 2007, with some emphasis on the recent evolutionary stuff. I think anyone can come along, so if you want to get up to speed with what I do, then do come along.

Seen and Liked »

[3 Nov 2014 | Comments Off | 239 views]
http://www.bbc.co.uk/rd/images/dynamic/W1siZmYiLCJwdWJsaWMvcmQvc2l0ZXMvNTAzMzVmZjM3MGI1YzI2MmFmMDAwMDA0L2NvbnRlbnRfZW50cnk1MDRlMTgxNTcwYjVjMjBhMGMwMDEzNDQvNTQ1NzY3NGVhY2ZiYWIwNDRkMTNkZDZlL2ZpbGVzL2NvbWJpbmVkLmpwZyJdLFsicCIsInRodW1iIiwiNTkyeCJdXQ/combined.jpg I’m happy that Segmentino, our Vamp implementation of my music segmentation algorithm is being used at the Beeb. BBC R&D did some research to find efficient ways of editing down a piece of music to 30 seconds, as Chris Baume describes in this blog post. Their tests suggest that the more sophisticated automatic methods (including the one using Segmentino) perform similarly to manual edits. The research was carried out—under Chris’s supervision—by Adib Mehrabi. Incidentally, Segmentino has recently been used in other research, too, in Wang’s cool paper about automatic segmentation of full-length concert videos, as seen at ISMIR last week.

Seen and Liked »

[25 Okt 2014 | Comments Off | 382 views]
Looking at the ISMIR 2014 programme I discovered that Emilio Molina and colleagues have tested our pYIN pitch tracker against a range of other methods (SWIPE, YIN, MELODIA, Boersma/Lei Wang, …). Here’s a link to their paper entitled “The importance of F0 tracking in Query-by-Singing-Humming“. Their results suggest that pYIN is among the best trackers on clean data, and the most robust against noise and distortion. Nice.