ace: a framework for optimizing music classification cory mckay rebecca fiebrink daniel mcennis...

25
ACE: A Framework for optimizing music classification Cory McKay Rebecca Fiebrink Daniel McEnnis Beinan Li Ichiro Fujinaga Music Technology Area Faculty of Music McGill University

Upload: myra-patrick

Post on 31-Dec-2015

217 views

Category:

Documents


0 download

TRANSCRIPT

ACE: A Framework for optimizing music classification

Cory McKayRebecca FiebrinkDaniel McEnnis

Beinan LiIchiro Fujinaga

Music Technology AreaFaculty of MusicMcGill University

22/25/25

Goals

Highlight limitations of existing pattern recognition software when applied to MIR Present solutions to these limitations

Stress importance of standardized classification and feature extraction software Ease of use, portability and extensibility

Present the ACE software framework Uses meta-learning Uses classification ensembles

33/25/25

Existing music classification systems

Systems often implemented with specific tasks in mind Not extensible to general tasks Often difficult to use for those not involved in project

Need standardized systems for a variety of MIR problems No need to reimplement existing algorithms More reliable code More usable software Facilitates comparison of methodologies

Important foundations Marsyas (Tzanetakis & Cook 1999) M2K (Downie 2004)

44/25/25

Existing general classification systems

Available general-purpose systems: PRTools (van der Heijden et al. 2004 ) Weka (Witten & Frank 2005)

Other meta-learning systems: AST (Lindner and Studer 1999) Metal (www.metal-kdd.org)

55/25/25

Problems with existing systems

Distribution problems Proprietary software Not open source Limited licence

Music-specific systems are often limited None use meta-learning Classifier ensembles rarely used Interfaces not oriented towards end users

General-purpose systems not designed to meet the particular needs of music

66/25/25

Special needs of music classification (1)

Assign multiple classes to individual recordings A recording may belong to multiple genres, for example

Allow classification of sub-sections and of overall recordings Audio features often windowed Useful for segmentation problems

Maintain logical grouping of multi-dimensional features Musical features often consist of vectors (e.g. MFCC’s) This relatedness can provide classification opportunities

77/25/25

Special needs of music classification (2)

Maintain identifying meta-data about instances Title, performer, composer, date, etc.

Take advantage of hierarchically structured taxonomies Humans often organize music hierarchically Can provide classification opportunities

Interface for any user

88/25/25

Standardized file formats

Existing formats such as Weka’s ARFF format cannot represent needed information

Important to enable classification systems to communicate with arbitrary feature extractors

Four XML file formats that meet the above needs are described in proceedings

99/25/25

The ACE framework

ACE (Autonomous Classification Engine) is a classification framework that can be applied to arbitrary types of music classification

Meets all requirements presented above

Java implementation makes ACE portable and easy to install

1010/25/25

ACE and meta-learning

Many classification methodologies available Each have different strengths and weaknesses

Uses meta-learning to experiment with a variety of approaches Finds approaches well suited to each problem Makes powerful pattern recognition tools available to non-

experts Useful for benchmarking new classifiers and features

1111/25/25

ACE

Feature Extraction

System

Classification Methodology n

Dimensionality Reduction

Classification Methodology 1

Dimensionality Reduction

Model Classifications

MusicRecordings

Taxonomy Feature Settings

Extracted Features

Experiment Coordinator

Classifier Evaluator

Trained ClassifiersStatistical Comparison of Classification Methodologies

1212/25/25

Algorithms used by ACE

Uses Weka class libraries Makes it easy to add or develop new algorithms

Candidate classifiers Induction trees, naive Bayes, k-nearest neighbour, neural

networks, support vector machines Classifier parameters are also varied automatically

Dimensionality reduction Feature selection using genetic algorithms, principal

component analysis, exhaustive searches Classifier ensembles

Bagging, boosting

1313/25/25

Classifier ensembles

Multiple classifiers operating together to arrive at final classifications e.g. AdaBoost (Freund and Shapire 1996)

Success rates in many MIR areas are behaving asymptotically (Aucouturier and Pachet 2004) Classifier ensembles could provide some

improvement

1414/25/25

Musical evaluation experiments

Achieved a 95.6% success with a five-class beatbox recognition experiment (Sinyor et al. 2005)

Repeated Tindale’s percussion recognition experiment (2004) ACE achieved 96.3% success, as compared to Tindale’s

best rate of 94.9% A reduction in error rate of 27.5%

1515/25/25

General evaluation experiments

Applied ACE to six commonly used UCI datasets

Compared results to recently published algorithm (Kotsiantis and Pintelas 2004)

1616/25/25

Results of UCI experiments (1)UCI Experiments

50

55

60

65

70

75

80

85

90

95

100

autos diabetes ionosphere iris labor vote

Dataset

Succ

ess

Kotsiantis ACE

DataSet

ACE's Selected Classifier

Kotsiantis' Success

Rate

ACE's Success Rate

autos AdaBoost 81.70% 86.30%

diabetes Naïve Bayes 76.60% 78.00%

ionosphere AdaBoost 90.70% 94.30%

iris FF Neural Net 95.60% 97.30%

labor k-NN 93.40% 93.00%

vote Decision Tree 96.20% 96.30%

1717/25/25

Results of UCI experiments (2)

ACE performed very well Statistical uncertainty makes it difficult to say

that ACE’s results are inherently superior ACE can perform at least as well as a state of

the art algorithm with no tweaking ACE achieved these results using only one

minute per learning scheme for training and testing

1818/25/25

Results of UCI experiments (3)

Different classifiers performed better on different datasets Supports ACE’s experimental meta-learning

approach

Effectiveness of AdaBoost (chosen 2 times out of 6) demonstrates strength of classifier ensembles

1919/25/25

Feature extraction

ACE not tied to any particular feature extraction system Reads Weka ARFF as well as ACE XML files

Does include two powerful and extensible feature extractors are bundled with ACE Write Weka ARFF as well as ACE XML

2020/25/25

jAudio

Reads: .mp3 .wav .aiff .au .snd

2121/25/25

jSymbolic

Reads MIDI Uses 111

Bodhidharma features

2222/25/25

ACE’s interface

Graphical interface Includes an on-line manual

Command-line interface Batch processing External calls

Java API Open source Well documented Easy to extend

2323/25/25

Current status of ACE

In alpha release Full release scheduled for January 2006

Finalization of GUI User constraints on training, classification and meta-

learning times Feature weighting Expansion of candidate algorithms

Long-term Distributed processing, unsupervised learning, blackboard

systems, automatic cross-project optimization

2424/25/25

Conclusions

Need standardized classification software able to deal with the special needs of music

Techniques such as meta-learning and classifier ensembles can lead to improved performance

ACE designed to address these issues

Web site:Web site: coltrane.music.mcgill.ca/ACEcoltrane.music.mcgill.ca/ACE

E-mail:E-mail: [email protected]@mail.mcgill.ca