News and Events

Seminar by Andre Holzapfel on beat tracking

Andre Holzapfel, from INESC Porto, gives a talk entitled "Selective sampling for beat tracking evaluation" on Friday March 2nd at 4pm in room 55.410.

Abstract: An approach is presented that identifies music samples which are difficult for current state-of-the-art beat trackers. In order to estimate this difficulty even for examples without ground truth, a method motivated by selective sampling is applied. This method assigns a degree of difficulty to a sample based on the mutual dis-agreement between the output of various beat tracking systems. On a large beat annotated dataset we show that this mutual agreement is correlated with the mean performance of the beat trackers evaluated against the ground truth, and hence can be used to identify difficult examples by predicting poor beat tracking performance. Towards the aim of advancing future beat tracking systems, we form a new dataset containing a high proportion of challenging music examples based on our method. We analyze the relations between perceptual difficulty and difficulty for automatic beat tracking using these data, and propose which signal properties are characterized by the highest potential of improvement of automatic beat tracking.

24 Feb 2012 - 09:58 | view
Phonos: Concert Michal Rataj & Ladislav Železný
On Wednesday February 29th 2012 at 19:30h in the Espai Polivalent of the Communication Campus of the UPF, Phonos is organizing a concert of Michal Rataj & Ladislav Železný.
23 Feb 2012 - 15:40 | view
Mohamed Sordo defends his PhD thesis on February 27th
Mohamed Sordo defends his PhD thesis entitled "Semantic Annotation of Music Collections: A Computational Approach" on Monday 27th of February 2012 at 11:00h in room 55.309.

The members of the jury's defense are:  Rafael Ramirez (UPF), Josep Lluis Arcos (IIIA-CSIC), Markus Schedl (Johannes Kepler University Linz).

Thesis abstract: Music consumption has changed drastically in the last few years. With the arrival of digital music, the cost of production has substantially dropped. The expansion of the World Wide Web has helped to promote the exploration of many more music content. Online stores, such as iTunes or Amazon, own music collections in the order of millions of songs. Accessing these large collections in an effective manner is still a big challenge.

In this dissertation we focus on the problem of annotating music collections with semantic words, also called tags. The foundations of all the methods used in this dissertation are based on techniques from the fields of information retrieval, machine learning, and signal processing. We propose an automatic music annotation algorithm that uses content-based audio similarity to propagate tags among songs. The algorithm is evaluated extensively using multiple music collections of varying size and quality of the data, including a large music collection of more than a half million songs, annotated with social tags derived from a music community. We assess the quality of our proposed algorithm by comparing it with several state of the art approaches. We also discuss the importance of using evaluation measures that cover different dimensions; per–song and per–tag evaluation. Our proposal achieves state of the art results, and has ranked high in the MIREX 2011 evaluation campaign. The obtained results also show some limitations of automatic tagging, related to data inconsistencies, correlation of concepts and the difficulty to capture some personal tags with content information. This is more evident in music communites, where users can annotate songs with any free text word. In order to tackle these issues, we present an in-depth study of the nature of music folksonomies. We concretely study whether tag annotations made by a large community (i.e. a folksonomy) correspond with a more controlled, structured vocabulary by experts in the music and the psychology fields. Results reveal that some tags are clearly defined and understood both by the experts and the wisdom of crowds, while it is difficult to achieve a common consensus on the meaning of other tags. Finally, we extend our previous work to a wide range of semantic concepts. We present a novel way to uncover facets implicit in social tagging, and classify the tags with respect to these semantic facets. The latter findings can help to understand the nature of social tags, and thus be beneficial for further improvement of semantic tagging of music.

Our findings have significant implications for music information retrieval systems that assist users to explore large music collections, digging for content they might like.

20 Feb 2012 - 10:08 | view
Seminar by Volker Hohmann on hearing device technology

Volker Hohmann, from the University of Oldenburg in Germany, gives a talk on "Trends in Hearing Device Technology" on Tuesday February 21st 2012 at 12:00pm in room 55.410.

Abstract: Currently about 18% of the European population suffers from hearing loss and this percentage is steadily increasing due to the demographic aging trend. Hearing loss is associated with minor to severe limitations in acoustic communication and therefore limits participation in social life. Most complaints allude to difficulties in understanding speech in noisy and reverberant environments, limited communication over the phone, and almost all hearing impaired report problems with perceiving and enjoying music. Due to the technological progress of the last decades, hearing devices have improved significantly, and the ever-growing computational capabilities of digital hearing devices offers a promising future. Still, the current technology does not allow a complete hearing loss rehabilitation: In mild hearing losses, an individual audio amplification would be helpful to alleviate most communication problems, but current consumer technology such as TVs, mobile phones, MP3 players and public announcement systems does not support even basic individualized audio amplification. In cases of mild to moderate hearing losses, which are commonly treated with acoustic hearing aids, the challenge is to provide a sufficient improvement in the signal-to-noise ratio to restore speech communication in acoustically challenging situations and to individually modify music with the aim of making music consumption enjoyable again. In the less frequent cases of severe to profound hearing losses or deafness, which are usually treated with cochlear implants (a biomedical device that activates the acoustic nerve electrically by implanted electrodes), all previously mentioned communication problems are more pronounced. This calls for further research into device technology and improved signal processing. In my talk, I will give an overview of the current technology of hearing devices and discuss the major research challenges and promising approaches for a further improvement of the rehabilitation of hearing loss.

Biography: Volker Hohmann received the doctorate degree in Physics (Dr. rer. nat.) from the University of Göttingen, Germany in 1993. Since then, he has been a faculty member of the Physics Institute at the University of Oldenburg, Germany, and member of the Medical Physics Group (head: B. Kollmeier). He is active in undergraduate and graduate teaching in Physics, and his research expertise is in Acoustics and Digital Signal Processing with applications to signal processing in speech processing devices, e. g., Hearing Aids. In the private sector, he is head of R&D at HoerTech gGmbH, a leading non-profit research institute in the field of audiological and acoustical developments related to hearing systems. He was a guest researcher at Boston University, Boston, USA in 2000 and at the Technical University of Catalonia, Barcelona, Spain in 2008/2009. Dr. Hohmann obtained the Habilitation (maximum academic degree in Germany) in Applied Physics in 2007 and received the Lothar-Cremer price of the German acoustical society (DEGA) in 2008 for outstanding contributions to audiological acoustics and signal processing.

17 Feb 2012 - 18:44 | view
Phonos: Barcelona Laptop Orchestra Concert

On Thursday February 23rd 2012 at 19:30h in the Espai Polivalent of the Communication Campus of the UPF, Phonos is organizing a concert of the Barcelona Laptop Orchestra.

16 Feb 2012 - 17:19 | view
2nd CompMusic Workshop: call for participation

Dates: July 12th-13th, 2012
Venue: Bahçeşehir Üniversitesi, Istanbul
Scientific committee: Xavier Serra, Preeti Rao, Hema Murthy, Bariş Bozkurt
More info: http://compmusic.upf.edu

CompMusic is a research project funded by the European Research Council and coordinated by Xavier Serra from the Music Technology Group of the Universitat Pompeu Fabra in Barcelona (Spain). It aims to advance in the automatic description of music, reducing the gap between audio signal features and semantically meaningful music concepts by taking a culture-specific perspective. It focuses on the study of five art-music traditions around the world: Hindustani and Carnatic (India), Turkish-makam, Andalusian (Arab countries) and Han (China).

This second workshop will cover all the topics of relevance to CompMusic, bringing together the researchers already working in the project plus researchers willing to contribute to the specific problems identified. We are specially interested in musicological contributions from any of the identified cultures. A selection of the articles presented in the workshop will be published as a special issue in the Journal of New Music Research.

All proposed papers should relate to one of the five art-music traditions in aspects like: general musicological studies; culture specific semantic analysis; melodic and rhythmic analysis; culture specific music similarity; or community profiling.

Interested participants should send an abstract of the proposed papers to xavier [dot] serra [at] upf [dot] edu (Xavier Serra) before March 30th 2012.

13 Feb 2012 - 17:31 | view
Seminar by Cédric Mesnage on Social Shuffle

Cédric Mesnage, from the University of Lugano, will give a research seminar on "Social Shuffle" on Friday February 10th at 13:00h in room 55.410.

Abstract: In this talk I present my understanding on Music Discovery. First the work carried out during my PhD studies in web engineering and social media, particularly web based experiments on the concepts of tag navigation and social diffusion for Music Discovery using Last.fm data, Facebook and Youtube. Second I list problems I see with the discovery of world music from an inter-cultural perspective. Third I give potential directions and outcomes for future work.

6 Feb 2012 - 11:14 | view
Open positions in CompMusic

The Music Technology Group of the Universitat Pompeu Fabra, Barcelona, has open research positions at the PhD and post-doc levels to work on the CompMusic project. Interested applicants should send a CV and a letter of motivation, expressing their research interests in relation to the CompMusic project, to Xavier Serra.

We are specially interested in candidates with an engineering background in areas such as semantic analysis, machine learning or signal processing and knowledge in one of the art-music traditions that are the focus of the project, more specially in the Arab (Andalusi) or the Chinese (Han) ones.

25 Jan 2012 - 14:17 | view
Lectures by Markus Schedl on context-based MiR

Markus Schedl, from the Department of Computational Perception of Johannes Kepler University in Linz (Austria) will give a series of lectures on context-based Music Information Retrieval.

* 23.02: 15:30-16:30 (room 52.321) DTIC Research seminar on "Geo-Aware Music Information Extraction from Social Media"

Abstract: The abundance of data present in Social Media opens an unprecedented source for information about every topic of our daily lives. Since music plays a vital role in many persons' lives, information about music items is found in large amounts in data sources such as social networks or microblogs. In this talk, I will report on latest findings in Social Media Mining to extract meaningful musical information from microblogs. Specifically, I will address the topics of similarity measurement, popularity estimation, and cultural-aware music taste and trend detection. In addition to elaborating on the methodological background, I will present some application scenarios and demonstrator systems that strive to illustrate some application domains of this interesting research field.

* 28.02-01.03 (room 52.S27) Lectures on: "Context-based Music Information Retrieval"

These lectures give an introduction to Music Information Retrieval (MIR), with a focus on context-based methods. MIR is concerned with the extraction, processing, and use of various music-related information from various musical data sources (scores, digital audio, live concerts, collaborative tags, video clips, album covers, etc.). I will focus on feature extraction (context- and Web-based), similarity measurement, and applications of MIR. I will also strive to include my latest research on Social Media Mining for MIR.

  • 28.02.: 12:00-14:00 Introduction to MIR

What is MIR? - definitions, important key aspects, subfields and typical tasks, basic scheme of an MIR system, basics in different retrieval approaches, feature extraction (audio and contextual), and similarity measurement

  • 29.02.: 12:00-14:00 Context-based Feature Extraction

motivation, data sources for contextual features, specific biases and problems of contextual features, term vector-based (web terms, tags, lyrics) and co-occurrence-based (play lists, page counts, P2P networks) approaches

  • 01.03.: 10:00-12:00 Similarity Measurement and Applications

similarity measurement on different kinds of music-related data (from scalar to multi-instance, multi-dimensional data), selected applications developed by the Department of Computational Perception / Johannes Kepler University, Linz, Austria (for instance, user interfaces to music)

19 Jan 2012 - 17:39 | view
CompMusic Workshop at KIIT-Gurgaon

The CompMusic project is organizing this workshop as a satellite event of the International Symposium on Frontiers of Research on Speech, Music and Allied Signal Processing (FRSM 2012) with the aim to give an overview of the Music Information Research of relevance to Hindustani and Carnatic music.

Date: January 20th, 2012, from 9:30am to 5:30pm
Venue: College of Engineering, Kamrah International Institute of Technology (KIIT), KIIT Campus, Sohna Road, Near Bhondsi, Gurgaon, Haryana, India

  • 09:30am: CompMusic: Current research and initial results (Xavier Serra)
  • 10:00am: Hindustani Music: A case for computational modeling (Preeti Rao)
  • 10:30am: Carnatic Music: A signal processing perspective (Hema Murthy)
  • 11:00am: tea
  • 11:30am: Carnatic Music: A musicological perspective (T. M. Krishna)
  • 12:30pm: Hindustani Music: A musicians perspective (Pt. Buddhadev Dasgupta)
  • 01:30pm: lunch
  • 02:30pm: Distribution based computational analysis of Makam Music (Bariş Bozkurt)
  • 03:30pm: Machine learning for music discovery (Joan Serrà)
  • 04:00pm: tea
  • 04:30pm: Panel discussion (moderator: Xavier Serra; panelists: Preeti Rao, Hema Murthy, Bariş Bozkurt, T. M. Krishna, Pt. Buddhadev Dasgupta, Mallika Banerjee)
6 Jan 2012 - 05:58 | view
intranet