kullback-leibler boosting

21
Kullback- Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem

Upload: avian

Post on 07-Feb-2016

56 views

Category:

Documents


0 download

DESCRIPTION

Kullback-Leibler Boosting. Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem. RealBoost Review. Start with some candidate feature set Initialize training sample weights Loop: Add feature to minimize error bound - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Kullback-Leibler Boosting

Kullback-Leibler BoostingCe Liu, Hueng-Yeung ShumMicrosoft Research AsiaCVPR 2003

Presented by Derek Hoiem

Page 2: Kullback-Leibler Boosting

RealBoost Review Start with some candidate feature set Initialize training sample weights Loop:

Add feature to minimize error bound Reweight training examples, giving more weight to

misclassified examples Assign weight to weak classifier according to

weighted error of training samples Exit loop after N features have been added

Page 3: Kullback-Leibler Boosting

The Basic Idea of KLBoosting

Similar to RealBoost except:Features are general linear projectionsGenerates optimal features Uses KL divergence to select featuresFiner tuning on coefficients

Page 4: Kullback-Leibler Boosting

Linear Features

KLBoosting:

VJ Adaboost:

Page 5: Kullback-Leibler Boosting

What makes a feature good?

KLBoosting:

RealBoost:Minimize upper bound on classification error

Page 6: Kullback-Leibler Boosting

Creating the feature set

Sequential 1-D OptimizationBegin with large initial set of features (linear

projections)Choose top L features according to KL-Div Initial feature = weighted sum of L featuresSearch for optimal feature in directions of L

features

Page 7: Kullback-Leibler Boosting

Example

Initial feature set:

xx

xx

x

x

x

x

Page 8: Kullback-Leibler Boosting

Example

Top two features (by KL-Div):

xx

xx

x

x

x

x

w1 w2

Page 9: Kullback-Leibler Boosting

Example

Initial feature (weighted combo by KL):

xx

xx

x

x

x

x

w1 w2f0

Page 10: Kullback-Leibler Boosting

Example

Optimize over w1

xx

xx

x

x

x

x

w1 w2f1

f1= f0 + B* w1

B = -a1..a1

Page 11: Kullback-Leibler Boosting

Example

Optimize over w2

xx

xx

x

x

x

x

w1 w2f2

f2= f1 + B* w2

B = -a2..a2

(and repeat…)

Page 12: Kullback-Leibler Boosting

Creating the feature set

First three features

Selecting the first feature

Page 13: Kullback-Leibler Boosting

Creating feature set

Page 14: Kullback-Leibler Boosting

Classification

= ½ in RealBoost

Page 15: Kullback-Leibler Boosting

Parameter Learning

With each added feature k:Set first a1..ak-1 to current optimal value Set ak to 0Minimize recognition error on training:

Solve using greedy algorithm

Page 16: Kullback-Leibler Boosting

KLBoost vs AdaBoost

1024 candidate features for AdaBoost

Page 17: Kullback-Leibler Boosting

Face detection: candidate features

52,400 2,800450

Page 18: Kullback-Leibler Boosting

Face detection: training samples

8760 faces + mirror images 2484 non-face images 1.34B patches Cascaded classifier allows bootstrapping

Page 19: Kullback-Leibler Boosting

Face detection: final features

top ten

global semantic

global not semantic

local

Page 20: Kullback-Leibler Boosting

Results

xx

xx

8 85 853

Schneiderman (2003)

Test time: .4 sec per 320x240 image

Page 21: Kullback-Leibler Boosting

Comments

Training time?

Which improves performance:Generating optimal features?KL feature selection?Optimizing alpha coefficients?