News and Events

PhD positions at the MTG

The MTG is opening 5 funded PhD positions to work within some of its research areas, with a starting date of September 2013. The candidates have to apply to the PhD program of the Department of Information and Communication Technologies. They have to demonstrate an academic and research background in the area they are applying for and have to submit a research proposal on a specific topic. The areas and topics for which we offer the funded positions are:

Sound and Music Communities (responsible faculty: Xavier Serra): Within the CompMusic project we have two open positions to work on computational approaches for the understanding of the relationship between lyrics and music in one or several of the following music traditions: Hindustani (North India), Carnatic (South India), Turkish-makam (Turkey), Andalusian (Maghreb) and Beijing Opera (China). In the context of Freesound we also have one open position to work on issues related to community profiling, linked data, automatic data structuring and sound ontologies.

Sound and Music Description (responsible faculty: Emilia Gomez): We have one open position to work on one of these two topics: (1) Computer-assisted transcription, similarity and classification of flamenco singing, using signal processing, machine learning and user modeling methodologies. Spanish funded project, SIGMUS. (2) Music & Autobiographical Memory, dealing with audio feature extraction, music recommendation and user preferences' modeling.

Musical and Advanced Interaction (responsible faculty: Sergi Jordà): We have one open position to work on tangible and tabletop interaction.

Before making the application the candidate needs the support of the faculty member responsible for the research area chosen. Interested people should first send a CV and a motivation letter to the faculty member identified.

3 Jan 2013 - 13:05 | view
Application open for the Master in Sound and Music Computing

The application for the Master in Sound and Music Computing, program 2012-2013, is open on-line. There are 4 application periods (deadlines: January 16th, March 14th, May 16th, June 28th). For more information on the UPF master programs and specifically on the SMC Master check here.

2 Jan 2013 - 18:29 | view
MTG-QBH: new dataset of sung melodies

As a little gift for the holiday season (be it Christmas, Hannukah, Tenno no Tanjobi or any other festivity you celebrate!), we're glad to announce the release of a new dataset: MTG-QBH.

The dataset includes 118 recordings of a cappella sung melody excerpts. The recordings were made as part of the experiments on Query-by-Humming (QBH) reported in:

J. Salamon, J. Serrà and E. Gómez, "Tonal Representations for Music Retrieval: From Version Identification to Query-by-Humming", International Journal of Multimedia Information Retrieval, special issue on Hybrid Music Information Retrieval, In Press (Nov. 2012).

In addition to the query recordings, three meta-data files are included, one describing the queries and two describing the music collections against which the queries were tested in the experiments described in the aforementioned article.

Whilst the query recordings are included in this dataset, audio files for the music collections listed in the meta-data files are not included in this datas

et, as they are protected by copyright law (sorry!). Nonetheless, all tracks are commercially available and we hope that those interested in using this dataset for QBH should be able to acquire them easily.

Further information about the queries, how they were recorded and by who is available on the dataset website, where you can of course download the audio and metadata files.

We hope that you find this dataset useful, whether for QBH or any other research topic (e.g. monophonic transcription), and would be very interested to receive your feedback.

13 Dec 2012 - 15:36 | view
New VST plug-in by Yamaha based on a previous joint research with the MTG

Yamaha Corporation has released through Steinberg Media Technologies a new VST plug-in known as 'sonote beat re:edit' based on the achievements of a research project in collaboration with the Music Technology Group.

On the top of the concept and know-how gained during the previous joint research with the MTG, Yamaha has fully conceptualised the product idea behind this novel application which is powered by Yamaha's proprietary technologies.

13 Dec 2012 - 14:03 | view
MELODIA downloaded over 250 times and HPCP reaches 100!

MELODIA, our melody extraction vamp plug-in by Justin Salamon reached its 250th download yesterday! Also, our recently released HPCP vamp plug-in by Emilia Gómez and Jordi Bonada has just reached 100 downloads!

Apart from obviously being excited about the interest in both plug-ins, we were also really surprised by the wide range of uses people have found for them. For MELODIA, in addition to the perhaps more expected research purposes (transcription, query-by-humming, computational musicology and ethnomusicology, music similarity, structure analysis, etc.), people have downloaded it for educational use in schools and universities, for music composition (for example for synthesizing natural sounding vibrato by using the pitch curve generated by a real singer, or for vocaloid compositions), for checking out the current state-of-the-art (including some commercial companies), and even just to "view music in a different way" and "for fun".

HPCP has also been downloaded for a variety of purposes including composition, analysis, education, alignment of different audio recordings, comparison of chroma-related features for retrieval in musical heritage collections, to analyze recordings of electronic music, to study song structure and even just to "Have fun with HPCP".

So... what next?

If you haven't already, you can try out MELODIA and HPCP for yourself:
http://mtg.upf.edu/technologies/melodia
http://mtg.upf.edu/technologies/hpcp

5 Dec 2012 - 18:20 | view
Article in the Information Processing & Management journal

A group of researchers from the Sound and Music Description research area @MTG (Dmitry Bogdanov, Martín Haro, Ferdinand Fuhrmann, Emilia Gómez and Perfecto Herrera), in collaboration with Open University (Anna Xambó) are publishing a paper on music recommendation and music preference visualization at the Information Processing & Management journal edited by Elsevier.

This work is part of their “The Musical Avatar“ project, a system that provides an iconic representation of one's musical preferences. The idea behind is to use computational tools to automatically describe your music (in audio format) in terms of melody, instrumentation, rhythm, etc and use this information to build an iconic representation of one’s musical preferences and to recommend you new music. All the system is only based on content description, i.e. on the signal itself and not on information about the music (context) as found on web sites, etc.

This is the reference:

The paper is also available at the MTG web page

29 Nov 2012 - 11:45 | view
Seminar by Dan Stowell on tracking sound sources in noise
22 Nov 2012

Dan Stowell, from Queen Mary, University of London, will give a seminar on "Tracking multiple intermittent sources in noise: inferring a mixture of Markov renewal processes" on Thursday November 22nd at 3:30pm in room 52.321.

Abstract: Consider the sound of birdsong, or footsteps. They are intermittent sounds, having as much structure in the gaps between events as in the events themselves. And often there's more than one bird, or more than one person - so the sound is a mixture of intermittent sources. Standard tracking techniques (e.g. Markov models, autoregressive models) are a poor fit to such situations. We describe a simple signal model (the Markov renewal process (MRP)) for these intermittent data, and introduce a novel inference technique that can infer the presence of multiple MRPs even in heavy noise. We illustrate the technique via a simulation of auditory streaming phenomena, and an experiment to track a mixture of singing birds.

19 Nov 2012 - 11:55 | view
The MTG takes part in "Programa Professors i Ciència" (Fundació Catalunya Caixa)

The MTG collaborates in the "Programa Professors i Ciència" (Teachers & Science program), funded by Fundació Catalunya Caixa.

The program offers high-school teachers the opportunity of taking part in scientific specialization courses at research centers in Catalonia. In this way, the program aims to bring research closer to educational institutions at the secondary level. The MTG organizes a course on sound & nature that takes place  at Poblenou Campus on November 9th and 16th, 2012.

The MTG organizes a workshop on "Sounds of Nature: The Nature of Sound" which is devoted to study natural sounds, their acoustic behavior and how to describe and generate them by means of a computer. The course provides a specific set of educational resources based on free tools and sounds so that the workshop content can be directly used in educational contexts. MTG researchers involved in this initiative are Agustín Martorell, Sonia Espí, Jaume Ferrete and Emilia Gómez.

13 Nov 2012 - 15:13 | view
MusicBrainz Summit at the UPF

The MTG-UPF will be hosting the 12th MusicBrainz Summit on November 9-11, 2012. MusicBrainz is an open music encyclopedia that collects music metadata and that the CompMusic project uses to collect all the metadata of the music collections that are being studied.

The MusicBrainz Summit is a meeting of the editors and developers of MusicBrainz to discuss its future and to do some group hacking. To learn more about this MusicBrainz Summit visit the official website.

7 Nov 2012 - 10:21 | view
HPCP vamp plug-in available for download!

Following the great success of our MELODIA - Melody Extraction vamp plugin, we are very pleased to announce the launch of the HPCP - Harmonic Pitch Class Profile vamp plug-in.

The plug-in provides a simple implementation of our chroma feature extraction algorithm which has been used in different applications, e.g. chord detection, key estimation, cover version identification and music classification. Full details of the algorithm can be found in the following papers:

NOTE: The main difference between this implementation and the original algorithm is that this implementation does not perform automatic tuning frequency estimation. The reference tuning frequency is defined as an input parameter.

The plug-in is available online for free download (non-commercial purposes). We hope it will serve the research community for evaluating different approaches for chroma feature extraction and for its further exploitation in higher-level music information retrieval tasks.

We are very interested in receiving feedback from the community, please let us know what you think!

19 Oct 2012 - 13:07 | view
intranet