Subscribe: Comments on MIR Research: Myths about tags
Added By: Feedage Forager Feedage Grade C rated
Language: English
audio  audition system  based  model  music  paul talk  query semantic  query  retrieve  search  semantic  system  talk  watch video  watch  words 
Rate this Feed
Rate this feedRate this feedRate this feedRate this feedRate this feed
Rate this feed 1 starRate this feed 2 starRate this feed 3 starRate this feed 4 starRate this feed 5 star

Comments (0)

Feed Details and Statistics Feed Statistics
Preview: Comments on MIR Research: Myths about tags

Comments on MIR Research: Myths about tags

Updated: 2016-10-23T12:45:43.719+01:00


Paul,The talk is by Doug Turnbull. He also held a ...



The talk is by Doug Turnbull. He also held a similar talk at QMUL's C4DM which records all talks. The direct link is here. However, it's hard to read the slides and the audio quality is not as good as the Microsoft version (and it doesn't seem to include the reference to tags).

Title is "Desining a content-based music search engine". The abstract for his talk is:

If you go to or the Apple Itunes store, your ability to search for new music will largely be limited by the query-by-metadata' paradigm: search by song, artist or album name. However, when we talk or write about music, we use a rich vocabulary of semantic concepts to convey our listening experience. If we can model a relationship
between these concepts and the audio content, then we can produce a more flexible music search engine based on a 'query-by-semantic-
description' paradigm.

In this talk, I will present a computer audition system that can both annotate novel audio tracks with semantically meaningful words and retrieve relevant tracks from a database of unlabeled audio content
given a text-base query. I consider the related tasks of content-
based audio annotation and retrieval as one supervised multi-class, multi-label problem in which we model the joint probability of acoustic features and words. For each word in a vocabulary, we use an annotated corpus of songs to train a Gaussian mixture model (GMM)
over an audio feature space. We estimate the parameters of the model using the weighted mixture hierarchies Expectation Maximization algorithm. This algorithm is more scalable to large data sets and produces better density estimates than standard parameter estimation techniques. The quality of the music annotations produced by our
system is comparable with the performance of humans on the same task. Our `query-by-semantic-description' system can retrieve appropriate
songs for a large number of musically relevant words. I also show that our audition system is general by learning a model that can annotate and retrieve sound effects.

Lastly, I will discuss three techniques for collecting the semantic annotations of music that are needed to train such a computer
audition system. They include text-mining web documents, conducting surveys, and deploying human computation games.

I wish I could watch the video, but unfortunately ...


I wish I could watch the video, but unfortunately it requires
Microsoft Internet Explorer with windows media player, so I can't watch it. Who was the talk by? What was the abstract?