confidence preserving machine for ... - dualplus.github.io

1
Problem Confidence Preserving Machine (CPM) Facial Action Unit (AU) detection Diagram of CPM Iterative CPM (iCPM) C onfidence P reserving M achine for Facial Action Unit Detection Jiabei Zeng 1 , Wen-Sheng Chu 2 , Fernando De la Torre 2 , Jeffrey F. Cohn 2,3 , and Zhang Xiong 1 SANTIAGO, CHILE DECEMBER 11-18, 2015 AU12 presence AU12 absence ? Experiments Settings SIFT descriptors on predetermined facial landmarks 10-fold with disjoint training and test sets Evaluate CPM componants Comparisons Standard supervised methods training subjects test subject (a) (b) easy negative samples hard samples easy positive samples test samples confident classifiers QSS classifier Main idea Training subjects Test subject classifiers confident (iterative CPM) easy test samples Prediction Train a QSS classifier Train confident classifiers 0 1 2 3 0 1 2 1 2 3 4 5 0 1 2 0 1 2 3 4 5 -1 0 1 2 3 0 1 2 3 4 5 0 1 2 0 1 2 3 4 5 0 1 2 0 1 2 3 4 5 -1 0 1 2 3 0 1 2 3 4 5 -1 0 1 2 3 0 1 2 3 4 5 0 1 2 3 0 1 2 3 4 -1 0 1 2 3 acc=0.97 acc=0.99 acc=1.00 (w - , w + ) on the top of each other w + w - w + w - w t w t training data 1 training data 2 test data update easy (hard) samples identify easy test samples identify easy test samples add easy test samples update easy (hard) samples retrain (w - , w + ) retrain (w - , w + ) train (w - , w + ) train w t train w t iter#1 iter#2 iter#0 Boosting × × × × × × TW-SVM [6] × × × × × Self-paced learning [7] × × × × × × RO-SVM [8] × × × × × × Co-training [9] × × × × Lap-SVM [2] × × × × × DAM [3] × × CPMs Multiple classifiers Indentify easy/hard Unlabeled data Different distribution Smoothness assumption Non-parallel hyperplane Progressive labeling Confident classifiers (w+, w - ), or w y to confidently predict on positive and neg- ative samples, respectively × Previous work Alternating algorithm for training confident classifiers Single hyperplane Perform well on easy samples, e.g., high intensity AUs, frontal head pose or on particular subjects Perform badly on hard samples, e.g. , subtle AUs Easy-to-hard strategy Address specific and known sources - E.g. subtle AUs, head pose, individual differences [1] Semi-supervised learning methods (SSL) [2] - Smoothness/Cluster/Manifold assumptions on rela- tionships between input and label space - Not suitable for AU Transfer learning methods [3,4,5] - Only individual differences - Absence of spatial-temporal smoothness - Less efficient than CPM Quasi-semi-supervised (QSS) learning Generalization on unseen test subjects −1 1 −1 1 −1 1 100 200 300 400 frame 100 200 300 400 frame 100 200 300 400 frame truth w/o S w/ S A person-specific classifier w t for a test subject Quasi-semi-supervised (QSS) classifier Objective (4) Effectiveness of spatial-temporal smoothness Identify easy and hard samples (1) easy: (w+, w - ) have same predictions hard: (w+, w - ) have different predictions Objective (2) Relabeling strategies Holistic relabeling: Relabel all the hard samples + - hard data negative positive noise instance Localized relabeling: Relabel part of the hard samples Comparison with alternative methods (w+, w - ) might not generalize well due to mismatch between training and test. iCPM jointly learns (w+, w - ) and QSS. iCPM on a synthetic example Datasets GFT [10]: 50 2-min spontaneous videos from 50 participants BP4D [11]: 328 spontaneous videos from 41 participants DISFA [12]: 27 spontaneous videos from 27 participants Metric F1-frame 45 47 49 51 53 55 GFT F1-event 36 39 42 45 48 45 46 47 48 49 SVM Conf CPM iCPM DISFA 26 30 34 38 42 BP4D 42 44 46 48 50 57 56 55 Frame-based F1 Event-based F1 baseline methods (SVM, Adaboost), SSL (Laplacian SVM [2]), transfer learning methods (DAM [3], MDA [4], GFK [5]) [1] W.-S. Chu, et al. Selective transfer machine for personalized facial action unit detection. CVPR 2013. [2] S. Melacci, et al. Laplacian support vector machines trained in the primal, JMLR, 12:1149–1184, 2011. [3] Duan L, et al. Domain adaptation from multiple sources: A domain-dependent regularization approach, TNNLS, 23(3): 504-518, 2012. [4] B. Gong, et al. Geodesic flow kernel for unsupervised domain adaptation, CVPR 2012. [5] Q. Sun, et al. A two-stage weighting framework for multi-source domain adaptation, NIPS 2011. [6] R. Khemchandani, et al. Twin support vector machines for pattern classification, TPAMI, 29:905-910, 2007. [7] M. P. Kumar, et al. Self-paced learning for latent variable models, NIPS 2010. [8] Y. Grandvalet, et al. Support vector machines with a reject option, NIPS 2009. [9] A. Blum, et al.Combining labeled and unlabeled data with co-training, CoLT 1998. [10] J. F. Cohn, et al. Spontaneous facial expression in a small group can be automatically measur- ed: An initial demonstration, Behavior Research Methods, 42(4):1079–1086, 2010. [11] X. Zhang, et al. A high-resolution spontaneous 3d dynamic facial expression database, FG 2013. [12] S. Mavadati, et al. Disfa:A spontaneous facial action intensity database,TAC,4:151–160, 2013. 1 2 3 Easy samples Hard samples Data A Data B Holistic on A Holistic on B Localized on B To get the person-specific classifier w t Virtual labels for easy sampels from (w+, w - ) Propagate virtual labels from easy to hard test samples under semi-supervided manner. loss spatial-temporal smoothness value of index of j i i+1 i-1 i-2 i-3 i+2 i+3 i+4 i+5 i-4 i-5 spatial disimilarity exclude preserve temporal similarity truth labels relabel MH096951, Supported in part by 61272350, 2013AA01A603

Upload: others

Post on 16-Jul-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Confidence Preserving Machine for ... - dualplus.github.io

Problem Confidence Preserving Machine (CPM)Facial Action Unit (AU) detection Diagram of CPM

Iterative CPM (iCPM)

Confidence Preserving Machine for Facial Action Unit DetectionJiabei Zeng1, Wen-Sheng Chu2, Fernando De la Torre2, Jeffrey F. Cohn2,3, and Zhang Xiong1

SANTIAGO, CHILE DECEMBER 11-18, 2015

AU12presence

AU12absence

?

ExperimentsSettingsSIFT descriptors on predetermined facial landmarks10-fold with disjoint training and test sets

Evaluate CPM componants

Comparisons

Standard supervised methods

trai

ning

sub

ject

ste

st s

ubje

ct

(a)

(b)

easy negative samples hard samples easy positive samples

test samplescon�dent classi�ers

QSS classi�er

Main idea

Training subjects

Test subject

classifiersconfident

(iterative CPM)easy test samples

PredictionTrain a QSS

classifier

Train confident classifiers

0 1 2 30

1

2

1 2 3 4 5

0

1

2

0 1 2 3 4 5-1

0

1

2

3

0 1 2 3 4 5

0

1

2

0 1 2 3 4 5

0

1

2

0 1 2 3 4 5-1

0

1

2

3

0 1 2 3 4 5-1

0

1

2

3

0 1 2 3 4 5

0

1

2

3

0 1 2 3 4-1

0

1

2

3

acc=0.97 acc=0.99 acc=1.00

(w− , w+) on the top of each other

w+w−w+

w−

wt wt

training data 1

training data 2

test data

upda

te e

asy

(har

d)

sam

ples

identify easy test samples identify easy test samples

add

easy

test

sam

ples

upda

te e

asy

(har

d)

sam

ples

retrain (w− , w+) retrain (w− , w+) train (w− , w+)

train wt train wt

iter#1 iter#2iter#0

Boosting × × × × × ×TW-SVM [6] × × × × ×Self-paced learning [7] × × × × × ×RO-SVM [8] × × × × × ×Co-training [9] × × × ×Lap-SVM [2] × × × × ×DAM [3] × ×CPMs

Multiple classifiers

Indentify easy/hard

Unlabeleddata

Different distribution

Smoothnessassumption

Non-parallel hyperplane

Progressivelabeling

Confident classifiers(w+, w-), or wy to confidently predict on positive and neg-ative samples, respectively

×

Previous work

Alternating algorithm for training confident classifiers

Single hyperplanePerform well on easy samples, e.g., high intensity AUs, frontal head pose or on particular subjectsPerform badly on hard samples, e.g. , subtle AUs

Easy-to-hard strategy

Address specific and known sources- E.g. subtle AUs, head pose, individual differences [1]Semi-supervised learning methods (SSL) [2]- Smoothness/Cluster/Manifold assumptions on rela-tionships between input and label space- Not suitable for AUTransfer learning methods [3,4,5]- Only individual differences- Absence of spatial-temporal smoothness- Less efficient than CPM

Quasi-semi-supervised (QSS) learningGeneralization on unseen test subjects

−1

1

−1

1

−1

1

100 200 300 400 frame 100 200 300 400 frame 100 200 300 400 frame

truthw/o Sw/ S

A person-specific classifier wt for a test subjectQuasi-semi-supervised (QSS) classifier

Objective (4)

Effectiveness of spatial-temporal smoothnessIdentify easy and hard samples (1)

easy: (w+, w-) have same predictionshard: (w+, w-) have different predictions

Objective (2)

Relabeling strategiesHolistic relabeling: Relabel all the hard samples

+-

hard datanegativepositive

noise instance

Localized relabeling: Relabel part of the hard samples

Comparison with alternative methods

(w+, w-) might not generalize well due to mismatch between training and test.iCPM jointly learns (w+, w-) and QSS.

iCPM on a synthetic example

DatasetsGFT [10]: 50 2-min spontaneous videos from 50 participants BP4D [11]: 328 spontaneous videos from 41 participants DISFA [12]: 27 spontaneous videos from 27 participants Metric

F1-fr

ame

454749515355

GFT

F1-e

vent

36

39

424548

4546

47

4849

SVM Conf CPM iCPM

DISFA2630

34

3842

BP4D4244

464850

57

56

55

Frame-based F1

Event-based F1

baseline methods (SVM, Adaboost), SSL (Laplacian SVM [2]), transfer learning methods (DAM [3], MDA [4], GFK [5])

[1] W.-S. Chu, et al. Selective transfer machine for personalized facial action unit detection. CVPR 2013.[2] S. Melacci, et al. Laplacian support vector machines trained in the primal, JMLR, 12:1149–1184, 2011.[3] Duan L, et al. Domain adaptation from multiple sources: A domain-dependent regularization approach, TNNLS, 23(3): 504-518, 2012.[4] B. Gong, et al. Geodesic flow kernel for unsupervised domain adaptation, CVPR 2012. [5] Q. Sun, et al. A two-stage weighting framework for multi-source domain adaptation, NIPS 2011.

[6] R. Khemchandani, et al. Twin support vector machines for pattern classification, TPAMI, 29:905-910, 2007.[7] M. P. Kumar, et al. Self-paced learning for latent variable models, NIPS 2010.[8] Y. Grandvalet, et al. Support vector machines with a reject option, NIPS 2009.[9] A. Blum, et al.Combining labeled and unlabeled data with co-training, CoLT 1998.

[10] J. F. Cohn, et al. Spontaneous facial expression in a small group can be automatically measur- ed: An initial demonstration, Behavior Research Methods, 42(4):1079–1086, 2010.[11] X. Zhang, et al. A high-resolution spontaneous 3d dynamic facial expression database, FG 2013.[12] S. Mavadati, et al. Disfa:A spontaneous facial action intensity database,TAC,4:151–160, 2013.

1 2 33

Easy samples

Hard samples

Data A Data B Holistic on A Holistic on B Localized on B

To get the person-specific classifier wt

Virtual labels for easy sampels from (w+, w-) Propagate virtual labels from easy to hard test samples under semi-supervided manner.

loss spatial-temporalsmoothness

valu

e of

index of ji i+1i-1i-2i-3 i+2 i+3i+4 i+5i-4i-5

spatialdisimilarity

excludepreservetemporalsimilarity

truth labels

relabel

MH096951,Supported in part by 61272350, 2013AA01A603