User Tools

Site Tools


glad
A PCRE internal error occured. This might be caused by a faulty plugin

====== Differences ====== This shows you the differences between two versions of the page.

Link to this comparison view

glad [2013/06/30 22:15]
dawenl [2 Preliminary results]
glad [2013/07/01 18:31] (current)
dawenl [3 Future work]
Line 5: Line 5:
 | Code | [[https://​github.com/​dawenl/​glad_cal500|Github Link]] | | Code | [[https://​github.com/​dawenl/​glad_cal500|Github Link]] |
  
-[[http://​cosmal.ucsd.edu/​cal/​projects/​AnnRet/​|Cal500]] is a widely used dataset for music tagging. The tags it contains include instrumentation("​Electric Guitar"​),​ genre("​Jazz"​),​ emotion("​Happy"​),​ usage("​For a Party"​),​ etc. They were collected from human annotators and integrated by "​majority voting"​. However, considering the expertise from different annotators and the difficulty of different pieces, we can come up with a better statistical model for optimal label integration,​ which would ideally infer the label, as well as the expertise of the annotators and the difficulty of the songs. This work is primarily based on [[http://​mplab.ucsd.edu/​~jake/​OptimalLabeling.pdf|this paper]] in NIPS 2009.+[[http://​cosmal.ucsd.edu/​cal/​projects/​AnnRet/​|Cal500]] is a widely used dataset for music tagging. The tags it contains include instrumentation ("​Electric Guitar"​),​ genre ("​Jazz"​),​ emotion ("​Happy"​),​ usage ("For a Party"​),​ etc. They were collected from human annotators and integrated by "​majority voting" ​(The tags that most people annotated are kept). However, considering the expertise from different annotators and the difficulty of different pieces, we can come up with a better statistical model for optimal label integration,​ which would ideally infer the label, as well as the expertise of the annotators and the difficulty of the songs. This work is primarily based on [[http://​mplab.ucsd.edu/​~jake/​OptimalLabeling.pdf|this paper]] in NIPS 2009.
 ===== - Model ===== ===== - Model =====
  
Line 40: Line 40:
 I fit the model to instrument-based labels and genre-based labels as they are simple and easy to understand (plus for now the model I implemented only support binary labels). ​ I fit the model to instrument-based labels and genre-based labels as they are simple and easy to understand (plus for now the model I implemented only support binary labels). ​
  
-==== - Solo v.s. Instrument ​====+==== - Instruments as solo v.s. background ​====
  
-One thing interesting is how the annotators are good at labeling "​Solo",​ as opposed to just labeling ​"​Instrument"​ (as background).+One thing which is interesting ​to see is how the annotators are good at labeling ​instruments as "​Solo" ​(e.g. "Piano Solo", "​Electric Guitar Solo"), as opposed to just labeling ​instrument ​as background.
  
 {{:​comp.png?​200|}} {{:​comp.png?​200|}}
  
-The histogram above shows both the distribution of average expertise of labeling instrument as background and as solo. We can that there is a huge gap, indicating the annotators are way better at annotating solo.  ​+The histogram above shows both the distribution of average expertise ​$\hat{\alpha}$ ​of labeling instrument as background and as solo. We can that there is no overlapping, indicating the annotators are significantly ​better at annotating ​instruments as solo than as background.  ​
  
-==== - Difficulty of different instruments ====+==== - Difficulty of labeling ​different instruments ====
  
-We can average ​the expertise based on instrument ​to see the difficulty for labeling, from the annotator'​s point of view. Below is the top 5 simplest v.s. the top 5 hardest:+We can interpret ​the average ​expertise ​$\hat{\alpha}$ to label instrument-based tags as a reflection ​on how difficult it is to label the corresponding instruments correctly. Below is the top 5 simplest ​instruments ​v.s. the top 5 hardest ​instruments in terms of $\hat{\alpha}$:
  
 ^ Top 5 simplest ^ Top 5 hardest ^ ^ Top 5 simplest ^ Top 5 hardest ^
Line 59: Line 59:
 | Violin | Sequencer | | Violin | Sequencer |
  
 +The top 5 simplest instruments make a lot of sense as those are usually standing out clearly in music. On the other hand, the top 5 hardest is arguable, but still those are definitely not easy to label in general. ​
  
 +==== - Genre ====
  
 +We can take the similar approach on the genre-based tags:
  
-==== - Genre ====+^ Genre (from the simplest to the hardest) ^ 
 +| Rock | 
 +| World | 
 +| Folk | 
 +| Electronica | 
 +| R&B | 
 +| Pop | 
 +| Bluegrass | 
 +| Blues | 
 +| Hip-hop/Rap | 
 +| Country | 
 +| Jazz | 
 + 
 +Not surprising, Jazz is hard.  
 + 
 +===== - Future work ====
 +- At the moment, only binary labels are supported. But in fact, the model is easily extended to handle multinomial labels. 
 + 
 +- Now each individual label is treated completely independent. However, in the real world, it's easy to consider the correlation between different tags (e.g. "​Rock"​ is definitely more positively-correlated to "​Electric Guitar (Distortion)"​ than "​Sampler"​). This can be done by the similar idea from Correlated Topic Model ([[http://​machinelearning.wustl.edu/​mlpapers/​paper_files/​NIPS2005_774.pdf|CTM]]).  
 + 
 +- An interesting yet challenging problem would be to integrate the noisy beat annotations to create better ground truth data for beat tracking tasks. The main difference is that in beat annotation, the labels are no longer discretized categories, instead they are temporally-dependent series, which makes the problem much more difficult. ​
glad.1372644938.txt.gz · Last modified: 2013/06/30 22:15 by dawenl