deep learning penmiir - sebastian stober · 1 chim chim cheree 3/4 210 13.3 13.5 2 take me out to...

1
Research Focus Cognitive Sciences Learning Discriminative Features from EEG Recordings by Encoding Similarity Constraints Sebastian Stober <[email protected]> MLC g Research Focus Cognitive Sciences Machine Learning in Cognitive Science Lab MLC g songs with / without lyrics: meter tempo length (s) 1 Chim Chim Cheree 3/4 210 13.3 13.5 2 Take me out to the Ballgame 3/4 186 7.7 7.7 3 Jingle Bells 4/4 200 9.7 9.0 4 Mary Had a Little Lamb 4/4 160 11.6 12.2 instrumental pieces: 1 Emperor Waltz 3/4 178 8.3 2 Harry Potter Theme 3/4 166 16.0 3 Imperial March (Star Wars Theme) 4/4 104 9.2 4 Eine Kleine Nachtmusik 4/4 140 6.9 penMIIR public domain dataset of EEG recordings for music imagery information retrieval This research was supported by the donation of a Geforce GTX Titan X graphics card. Classifier Features Accuracy McNemar’s test (mid-p) vs. * SVC raw EEG 18.52% 0.0002 SVC raw EEG channel mean 12.41% < 0.0001 End-to-end NN raw EEG 18.15% 0.0001 Neural Net (NN) SCE output 27.22% 0.8200 *SVC SCE output 27.59% Stimulus Recognition (12-class classification) • minimalistic encoder: 1 spatial convolution filter • nested cross-validation scheme: Deep Learning for EEG Analysis? • high-dimensional data • limited amount of trials • bad signal-to-noise ratio https://github.com/sstober/openmiir • passive listening condition • 64 EEG channels @ 512 Hz • 9 subjects x 12 stimuli x 5 trials • all trials cropped at 6.9 s • 12 music stimuli from 8 music pieces: 540 trials (~ 62 min total) with > 225k dimensions 9-fold subject cross-validation train on data from 8 subjects (8x5x12=480 trials) test on remaining subject (1x5x12=60 trials) Pre-Training 5-fold trial block cross-validation 8x4x12=384 training trials from 4 trial blocks 8x1x12=96 validation trials from remaining trial block train on 50688 triplets from 384 training trials select model (early stopping) based on 21120 validation triplets (a,b,c) with a from 96 validation trials and b,c from 480 training and validation trials encoder layer (L1) = average over folds Supervised Training 5-fold trial block cross-validation same training/validation splits as in pre-training phase train with C on 480 training trials classifier layer (L2) = mean over folds Linear Support Vector Classifier (SVC) grid search for C with highest mean validation accuracy Simple Neural Network (NN) select fold model (early stopping) based on highest validation accuracy Chim Chim Cheree (lyrics) Take Me Out to the Ballgame (lyrics) Jingle Bells (lyrics) Mary Had a Little Lamb (lyrics) Chim Chim Cheree Take Me Out to the Ballgame Jingle Bells Mary Had a Little Lamb Emperor Waltz Hedwig’s Theme (Harry Potter) Imperial March (Star Wars Theme) Eine Kleine Nachtmusik minimize constraint violations Feature Extraction (signal filter) Input Triplet (shared parameters) Pairwise Similarity (dot product) Prediction (binary) Encoder Pipeline Encoder Pipeline Encoder Pipeline Similarity Similarity Output Reference Paired Trial Other Trial network structure A B C • systematic errors: confusing lyrics/non-lyrics pairs • little difference between classifiers Average Weights of the Neural Network Classifier (including pre-trained encoder filter) Network Activation Analysis http://www.uni-potsdam.de/mlcog • temporal patterns in weights and activations are picking up musical events (down beats) forward model: (regression) • dimensionality reduction by factor 64 • significant improvement of signal-to-noise ratio • more complex encoders and classifiers possible • pre-training is key to reduce over-fitting goo.gl/4xzma3 Comparison against Baselines Confusion Analysis • common encoder pipeline extracts features that are representative and allow to distinguish classes sim(A,B) > sim(A,C) Challenge Dataset Method Experiment Results for all A,B from same class and C from other class Similarity-Constraint Encoding (SCE) Pre-Training • motivated by metric learning • find features that satisfy similarity constraints

Upload: others

Post on 27-Jan-2021

2 views

Category:

Documents


0 download

TRANSCRIPT

  • Research FocusCognitive Sciences

    Learning Discriminative Features from EEGRecordings by Encoding Similarity ConstraintsSebastian Stober

    MLC g Research Focus Cognitive SciencesMachine Learning in Cognitive Science Lab

    MLC g

    songs with / without lyrics: meter tempo length (s)1 Chim Chim Cheree 3/4 210 13.3 13.52 Take me out to the Ballgame 3/4 186 7.7 7.73 Jingle Bells 4/4 200 9.7 9.04 Mary Had a Little Lamb 4/4 160 11.6 12.2

    instrumental pieces:1 Emperor Waltz 3/4 178 8.32 Harry Potter Theme 3/4 166 16.03 Imperial March (Star Wars Theme) 4/4 104 9.24 Eine Kleine Nachtmusik 4/4 140 6.9

    penMIIRpublic domain dataset of EEG recordingsfor music imagery information retrieval

    This research was supported by the donation of a Geforce GTX Titan X graphics card.

    Classifier Features Accuracy McNemar’s test (mid-p) vs. *

    SVC raw EEG 18.52% 0.0002SVC raw EEG channel mean 12.41% < 0.0001End-to-end NN raw EEG 18.15% 0.0001Neural Net (NN) SCE output 27.22% 0.8200*SVC SCE output 27.59%

    Stimulus Recognition (12-class classification)• minimalistic encoder: 1 spatial convolution filter• nested cross-validation scheme:

    Deep Learning for EEG Analysis?

    • high-dimensional data

    • limited amount of trials

    • bad signal-to-noise ratio

    https://github.com/sstober/openmiir

    • passive listening condition• 64 EEG channels @ 512 Hz• 9 subjects x 12 stimuli x 5 trials• all trials cropped at 6.9 s

    • 12 music stimuli from 8 music pieces:

    540 trials (~ 62 min total) with > 225k dimensions

    9-foldsubjectcross-validationtrainondatafrom8subjects(8x5x12=480trials)

    testonremainingsubject(1x5x12=60trials)

    Pre-Training

    5-foldtrialblockcross-validation8x4x12=384trainingtrialsfrom4trialblocks

    8x1x12=96validationtrialsfromremainingtrialblock

    trainon50688tripletsfrom384trainingtrials

    selectmodel(earlystopping)basedon21120validationtriplets(a,b,c)witha from96validationtrialsandb,c from480trainingandvalidationtrials

    encoderlayer(L1)=averageoverfolds

    Supe

    rvise

    dTraining 5-foldtrialblockcross-validation

    sametraining/validationsplitsasinpre-trainingphase

    trainwithCon480trainingtrials classifierlayer(L2)=meanoverfolds

    LinearSupportVectorClassifier(SVC)gridsearch forCwithhighest

    meanvalidationaccuracy

    SimpleNeuralNetwork(NN)selectfoldmodel(earlystopping)

    basedonhighestvalidationaccuracy

    Chim Chim Cheree (lyrics)Take Me Out to the Ballgame (lyrics)

    Jingle Bells (lyrics)Mary Had a Little Lamb (lyrics)

    Chim Chim ChereeTake Me Out to the Ballgame

    Jingle BellsMary Had a Little Lamb

    Emperor WaltzHedwig’s Theme (Harry Potter)

    Imperial March (Star Wars Theme)Eine Kleine Nachtmusik

    minimizeconstraintviolations

    FeatureExtraction(signalfilter)

    InputTriplet

    (sharedparameters)

    PairwiseSimilarity(dotproduct)

    Prediction(binary)

    EncoderPipeline

    EncoderPipeline

    EncoderPipeline

    Similarity

    Similarity

    Output

    Reference

    PairedTrial

    OtherTrial

    netw

    ork

    stru

    ctur

    e

    A

    B

    C

    • systematic errors: confusing lyrics/non-lyrics pairs• little difference between classifiers

    Average Weights of the Neural Network Classifier(including pre-trained encoder filter)

    Network Activation Analysis

    http://www.uni-potsdam.de/mlcog

    • temporal patterns in weights and activations are picking up musical events (down beats)

    forward model:(regression)

    • dimensionality reduction by factor 64• significant improvement of signal-to-noise ratio• more complex encoders and classifiers possible• pre-training is key to reduce over-fitting

    goo.gl/4xzma3

    Comparison against Baselines

    Confusion Analysis

    • common encoder pipeline extracts features that are representative and allow to distinguish classes

    sim(A,B) > sim(A,C)

    Cha

    lleng

    e

    Dat

    aset

    Met

    hod

    Expe

    rimen

    t

    Res

    ults

    for all A,B from same class and C from other class

    Similarity-Constraint Encoding (SCE) Pre-Training• motivated by metric learning• find features that satisfy similarity constraints