Skip to content

List of automatic music tagging research articles that are evaluated against MagnaTagATune Dataset

Notifications You must be signed in to change notification settings

keunwoochoi/magnatagatune-list

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 

Repository files navigation

magnatagatune-list

List of automatic music tagging (using audio) research articles that are evaluated against MagnaTagATune Dataset

Download

MagnaTagATune is kindly hosted by MIRG of City University London. Visit here to download mp3s and others.

Folders and files

After download three files, $ cat mp3.zip.* > mp3_all.zip to merge them, then unzip it by unzip mp3_all.zip. You then got 16 folders, '0' to '9' and then 'a' to 'f'.

Many works based on splitting the folders as 13:1:3 for training/validatin/testing. However it might be the best choice since each folders do NOT have same tag distributions. Minimum line is to shuffle training set, and would like to recommend to shuffle all of them and then split. Still you might need the same splitting to comparison, as in my case :-(

Tags (sorted by popularity)

Top 50

guitar, classical, slow, techno, strings, drums, electronic, rock, fast, piano, ambient, beat, violin, vocal, synth, female, indian, opera, male, singing, vocals, no vocals, harpsichord, loud, quiet, flute, woman, male vocal, no vocal, pop, soft, sitar, solo, man, classic, choir, voice, new age, dance, male voice, female vocal, beats, harp, cello, no voice, weird, country, metal, female voice, choral

Top 50 by categories

  • genre: classical, techno, electronic, rock, indian, opera, pop, classic, new age, dance, country, metal
  • instrument: guitar, strings, drums, piano, violin, vocal, synth, female, male, singing, vocals, no vocals, harpsichord, flute, no vocal, sitar, man, choir, voice, male voice, female vocal, harp, cello, femal voice, choral
  • mood: slow, fast, ambient, loud, quiet, soft, weird
  • etc: beat, solo, beats

Top 51-188

electro, drum, male vocals, jazz, violins, eastern, female vocals, instrumental, bass, modern, no piano, harpsicord, jazzy, string, baroque, foreign, orchestra, hard rock, electric, trance, folk, chorus, chant, voices, classical guitar, spanish, heavy, upbeat, no guitar, acoustic, male singer, electric guitar, electronica, oriental, funky, tribal, banjo, dark, medieval, man singing, organ, blues, irish, no singing, bells, percussion, no drums, woman singing, noise, spacey, singer, female singer, middle eastern, chanting, no flute, low, strange, calm, wind, lute, heavy metal, different, punk, oboe, celtic, sax, flutes, talking, women, arabic, hard, mellow, funk, fast beat, house, rap, not english, no violin, fiddle, female opera, water, india, guitars, no beat, chimes, drone, male opera, trumpet, duet, birds, industrial, sad, plucking, girl, silence, men, operatic, horns, repetitive, airy, world, eerie, deep, hip hop, space, light, keyboard, english, not opera, not classical, not rock, clapping, horn, acoustic guitar, disco, orchestral, no strings, old, echo, lol, soft rock, no singer, jungle, bongos, reggae, monks, clarinet, scary, synthesizer, female singing, piano solo, no voices, woodwind, happy, viola, soprano, quick, clasical

histogram of tags

So the dataset is unbalanced.

histogram

Proposed tag preprocessing

I wrote code to merge these synonyms.

synonyms = [['beat', 'beats'],
			['chant', 'chanting'],
			['choir', 'choral'],
			['classical', 'clasical', 'classic'],
			['drum', 'drums'],
			['electro', 'electronic', 'electronica', 'electric'],
			['fast', 'fast beat', 'quick'],
			['female', 'female singer', 'female singing', 'female vocals', 'female voice', 'woman', 'woman singing', 'women'],
			['flute', 'flutes'],
			['guitar', 'guitars'],
			['hard', 'hard rock'],
			['harpsichord', 'harpsicord'],
			['heavy', 'heavy metal', 'metal'],
			['horn', 'horns'],
			['india', 'indian'],
			['jazz', 'jazzy'],
			['male', 'male singer', 'male vocal', 'male vocals', 'male voice', 'man', 'man singing', 'men'],
			['no beat', 'no drums'],
			['no singer', 'no singing', 'no vocal','no vocals', 'no voice', 'no voices', 'instrumental'],
			['opera', 'operatic'],
			['orchestra', 'orchestral'],
			['quiet', 'silence'],
			['singer', 'singing'],
			['space', 'spacey'],
			['string', 'strings'],
			['synth', 'synthesizer'],
			['violin', 'violins'],
			['vocal', 'vocals', 'voice', 'voices'],
			['strange', 'weird']]`

I'm not 100% sure if these should be merged.

			['opera', 'operatic'],
			['hard', 'hard rock'],

Papers

This list is based on google scholar, list of papers that cited the dataset, google scholar search result

2016

not yet

2015

According to this,

  • U Sandouk et al., arxiv: text-processing.
  • U Sandouk et al., arxiv: text-processing.
  • E Colautti et al. : it's not relevant.

(and,)

  • J Nam et al., arxiv: pdf
    • title: A deep bag-of-features model for music auto-tagging
    • algorithm: bag-of-features (mel-specgram, onsets) --> PCA --> RBM for feature extraction --> DNN + stackedRBM
    • set: 14600/1629/6499 for training/validation/test
    • tags: used 160 tags
    • (selected) result
algorithm Deep-BoF (proposed) P 2011 P Hamel 2011 P Hamel 2012
AUC tag 0.888 0.845 0.861 0.870
AUC clip 0.956 0.938 0.943 0.949
Precision@3 0.511 0.449 0.467 0.481
Precision@6 0.358 0.430 0.327 0.339
Precision@9 0.275 0.249 0.255 0.263
Precision@12 0.225 0.205 0.211 0.216
Precision@15 0.190 0.175 0.181 0.184

2014

According to this,

  • D Lim at al., JMLR: about metrics
  • S Duan et al.: survey of tagging techniques
  • G Sageder et al.: not used. onlysubset of MagnaTagATune is used to verity the proposed feature selection that is based on ismir2014 DB.

(and,)

  • S Dieleman et al.: pdf
    • title: End-to-end learning for music audio
    • algorithm: end-to-end setting from audio (and spectrogram for comparison), 1d conv - MP - 1d conv - MP - fc layers
    • set: 16-folds, 1-12/13/14-6 for training/validation/test
    • (selected) result
filter length stride AUC (spectrograms) AUC (raw audio)
1024 1024 0.8690 0.8366
256 256 0.8815 0.8487
  • A Oord et al.: pdf
    • title: Transfer learning by supervised pre-training for audio-based music classification
    • algorithm: transfering learned MLP into problems based on other dataset.
    • network: k-means feature learning --> MLP
    • result

| model | nmse | mean average precision (mAP) | | ------------- |:-------------:|:-------------:|:-------------:| | linear regression | 0.965 | 0.823 | 0.0099 | | MLP (1 hidden layer) | 0.939 | 0.841 | 0.0179 | | MLP (2 hidden layers) | 0.924 | 0.837 | 0.0179 |

task AUC
tag (top 50 tags) <0.88
tag (all 188 tags) <0.86
  • SN Tran et al.: pdf: about similarity
    • title: feature preprocessing with RBMs for music similarity learning
  • F Gouton et al.: pdf
    • title: on evaluation validity in music autotagging
    • used MagTag5k, which is subset of MagnaTagATune. Now the link is broken.

2013

According to this,

  • D Wolff et al. : similarity
  • J Watson et al., arxiv: proposing an algorithm for ranking prediction.
    • set: 16k for training, 160 tags (probably top-160 popular tags)
    • feature: 104-dim MFCC, 100-dim embedding dimension
    • result
algorithm precision@1 precision@3
k-nn 39.4% 28.6%
k-nn (Wsabie space) 45.2% 31.9%
Wsabie 48.7% 37.5%
Affinity Weighted Embedding 52.7% 39.2%
  • JA Burgoyne et al., ismir: just mentioning.

(and,)

  • S Dieleman et al., ismir: pdf
    • title: Multiscale Approaches To Music Audio Feature Learning
    • result
algorithm average AUC
Laplacian 1 frame 0.898
Multiresolution spectrograms 0.888
  • J Stastny et al.: pdf
    • title: audio data classification by means of new algorithms
    • Not on tag prediction

2012

According to this,

  • D wolff et al.: just mentioning.

  • Y Li et al.: not about the DB

  • M Levy: can't get the pdf of it, seems like internal document in the school.

  • L Barrington, PhD thesis: probably not focusing on automatic tagging

  • P Hamel, PhD thesis: written in french. according to its list of contents it would be same results as in the author's ismir paper. (and,)

  • J NAM et al.,: mirex 2012 submission

    • title: mirex 2012 submission: audio classification using sparse feature learning
    • algorithm: RBM + SVM
    • result
    • AROC (ranking): 0.7244,
    • F-measure (annotation): 0.1123

2011

According to this,

  • AJ Quinn et al., SIGCHI: a general survey on human computation
  • YH Yang et al. IEEE ASLP: just mentining. experiment was performed on smaller private dataset
  • P Hamel et al.:
    • title: temporal pooling and multiscale learning for automatic annotation and ranking of music audio
    • features: MFCC
    • set: 14600/1629/6499 for training/validation/test
    • tags: not specified, probably using all 160 tags.
    • comparison: against four algirithms [19][17][18][9] (see the references in the paper), all of which use MFCC or 'cepstral transform that is closely related to MFCCs' + spectral features (spectrao centroid, rolloff, flux). They also used GMM of tags, VQ, PAMIR, or individual SVM.
    • features: 128 MFCC --> 120 principle components, which is called PMSC
      • then mean,var,max,min, 3rd and 4th centered moments -pooling
  • model 1: MLP with L2 regularisation and cross-entropy for loss function.
  • model 2: multi-time-scale learning
  • evaluation metric: AUC
  • result:

table 1.

algorithm AUC average training time
MFCC(20) 0.77+-0.04 5.9gh
MEL-spectrum(128) 0.853+-0.008 5.2h
PMSC(120) 0.876+-0.004 1.5h

table 2.

measure Manzagol Zhi Mandel Marsyas mel-spec+PFC PMSC+PFC PSMC+MTSL
average AUC-tag 0.750 0.673 0.821 0.831 0.820 0.845 0.861
average AUC-clip 0.810 0.748 0.886 0.933 0.930 0.938 0.943
precision@3 0.255 0.224 0.323 0.440 0.430 0.449 0.467
precision@6 0.194 0.192 0.245 0.314 0.305 0.320 0.327
precision@9 0.159 0.168 0.197 0.244 0.240 0.249 0.255
precision@12 0.136 0.146 0.167 0.201 0.198 0.205 0.211
precision@15 0.119 0.127 0.145 0.172 0.170 0.175 0.181

left four: rsults from MIREX 2009

(and,)

  • S Dieleman et al., ismir 2011:
    • title: audio-based music classification with a pretrained convolutional network

2010

According to this,

  • E Law et al., MLKD: learning to tag from open voca labels
  • M Mandel et al.: learning tags that vary withnin a song
  • A Quinn et al.: not related.
  • E Law et al., ECML: learning to tag using noisy labels
  • Y Panagakis et al.: sparse multi-label linear embedding NNTF for automatic music tagging
  • F Maillet: about music recommendation

(and,)

  • K Seyerlehner et al.: pdf
  • title: Automatic music tag classification based on block-level
  • metric: f-score and G-mean (geometric mean of true negative rate and recall)
  • result:
feature set f-score g-mean f-score (S2) g-mean(S20
SAF 0.3775 0.6101 0.3962 0.6252
BLF-PCA 0.4163 0.6410 0.4201 0.6439
feature set avg. f-score avg. g-mean avg. f-score (S2) avg. g-mean(S20
SAF 0.1777 0.3573 0.1932 0.3784
BLF-PCA 0.2136 0.4019 0.2185 0.4081
  • J Bergstra et al., ismir 2010: pdf

    • used few selected tags only
  • SR Ness et al.: [pdf](Improving automatic music tag annotation using stacked generalization of probabilistic svm outputs)

    • title: Improving automatic music tag annotation using stacked generalization of probabilistic svm outputs
    • audio features (+Affinity) + SVM
    • result

global average (table 2)

algorithm precision recall accuracy F-score
Audio SVM 0.307 0.315 0.969 0.311
Affinity SVM 0.351 0.354 0.971 0.353

affinity svm - per-tag evaluation (table 4) (table 3 is ommited as table 4 outperforms in overall)

number of tags precision recall accuracy F-score
20 0.418 0.691 0.856 0.518
30 0.346 0.671 0.862 0.453
40 0.394 0.397 0.914 0.395
50 0.369 0.372 0.923 0.371
100 0.259 0.262 0.951 0.260
all (188) 0.184 0.186 0.971 0.185

2009

N/A

About

List of automatic music tagging research articles that are evaluated against MagnaTagATune Dataset

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published