limitations of cotemporary classification algorithms major limitations of classification algorithms...

1
Limitations of Cotemporary Classification Algorithms Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include, Requirement of a large amount of labeled training data. Fixed parameters after the end of training phase, i.e., these classifiers can not attune themselves to particular detection scenarios after deployment. Synopsis A boosted classification framework, in which, Separate views (features) of the data used for online collection of training examples through co-training. Combined view (all features) used to make classification decisions. Background modeling used to prune away stationary regions to speed up the classification process. Global feature representations used for robustness. Online Detection and Classification of Moving Objects Using Progressively Improving Detectors Omar Javed, Saad Ali, Mubarak Shah Computer Vision Lab University of Central Florida Initial Training The Bayes Classifier is used as the base (weak) classifier for boosting. Each feature vector component v q ,where q ranges from 1,.., m1+m2 (for two object classes + background class), is used to learn the pdf for each class. The classification decision by the qth base classifier h q is taken as c i , Adaboost.M1 (Freund and Schapire, 96) is used to learn the strong classifier, from the initial training data and the base classifiers. Results Initial Training: 50 examples of each class All examples scaled to 30x30 vector Validation Set :20 images per class Testing on three sequences (a) Change in performance with increase in time for sequence 1,2 and 3. The performance was measure over two minute intervals. Over 150 to 200 detection decisions were usually made in this time period. Properties of the Proposed Algorithm Requirement of minimal training data Automated online selection of training examples after deployment, for continuous improvement. •. Near real time performance (4-5 frames/sec) The Online Co-training Framework During the test phase, select example for training, if more than 10% of the classifiers confidently predict the label of an example. Example’s margin is less than the computed thresholds. Once an example has been labeled by the co- training mechanism, an online boosting algorithm by [Oza and Russel,02] is used to update the base classifiers and the boosting coefficients. Feature Extraction Features for classification are derived from Principal Component Analysis of the appearance templates of the training examples. For each object class c i (excluding background) an appearance subspace, represented by d x m i projection matrix S i , is constructed. m chosen such that eigenvectors represent 99% of respective subspace. Appearance features for base learners are obtained by projecting a training example ‘r’ into appearance subspace of each object class. for two object classes the feature vector v of an example will be, 1 1 1 [ ,..., ] T m v v rS 11 1 2 2 [ ,..., ] T m m m v v rS Probabili ty Vehicle Person Clutter Histograms of a feature coefficient from the appearance subspace Row 1: Top 3 eigenvectors for person appearance subspace. Row 2: Vehicle appearance subspace (|) ( |) i q j q Pc v Pc v i j If :() 1 () argm ax log i cC ih x c i Hx Where β {i=1…N} are the boosting parameters, C is the set of classes The Online Co-training Framework : Key Observations Boosting mechanism selects the least correlated base classifiers. Ideal for co-training! Examples confidently labeled by one classifier are used to train the other. Only the observations lying close to the decision boundary of the boosted classifier are useful for improving classification performance. Use examples with small margins for online training. The Online Co-training Framework: Implementation Steps Determine confidence thresholds for each base classifier, using a validation data set. For class c i and jth base classifier h j set the confidence threshold, T j,ci base ( highest probability achieved by a negative example). Compute Margin thresholds T ci base are from the validation data set. Foreground Models Feature Extractio n ROIs Background Background Models Updated parameter s Color Classifier Edge Classifier Base Classifiers Boosted Classifier Classification Output Updated Boosted Parameters Co-Training Decision (if confident prediction by one set) Updated weak learners Information flow for the real-time object classification system Performance vs. the number of co-trained examples for the three sequences. Relatively few examples are required for improving the detection rates since these examples are from the same scene in which the classifier is being evaluated.

Upload: diana-charles

Post on 05-Jan-2016

238 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Limitations of Cotemporary Classification Algorithms Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include, Requirement

Limitations of Cotemporary Classification Algorithms

• Major limitations of classification algorithms like Adaboost, SVMs, or Naïve Bayes include,

• Requirement of a large amount of labeled training data.

• Fixed parameters after the end of training phase, i.e., these classifiers can not attune themselves to particular detection scenarios after deployment.

Synopsis

• A boosted classification framework, in which,

• Separate views (features) of the data used for online collection of training examples through co-training.

• Combined view (all features) used to make classification decisions.

• Background modeling used to prune away stationary regions to speed up the classification process.

• Global feature representations used for robustness.

Online Detection and Classification of Moving Objects Using Progressively Improving Detectors

Omar Javed, Saad Ali, Mubarak Shah Computer Vision Lab

University of Central FloridaInitial Training

• The Bayes Classifier is used as the base (weak) classifier for boosting.

• Each feature vector component vq ,where q ranges from 1,.., m1+m2 (for two object classes + background class), is used to learn the pdf for

each class.

• The classification decision by the qth base classifier hq is taken as ci,

• Adaboost.M1 (Freund and Schapire, 96) is used to learn the strong classifier, from the initial training data and the base classifiers.

Results

• Initial Training: 50 examples of each class

• All examples scaled to 30x30 vector

• Validation Set :20 images per class

• Testing on three sequences

(a) Change in performance with increase in time for sequence 1,2 and 3. The performance was measure over two minute intervals. Over 150 to 200 detection decisions were usually made in this time period.

Properties of the Proposed Algorithm

• Requirement of minimal training data

• Automated online selection of training examples after deployment, for continuous improvement.

•. Near real time performance (4-5 frames/sec)

The Online Co-training Framework

• During the test phase, select example for training, if

• more than 10% of the classifiers confidently predict the label of an example.

• Example’s margin is less than the computed thresholds.

• Once an example has been labeled by the co-training mechanism, an online boosting algorithm by [Oza and Russel,02] is used to update the base classifiers and the boosting coefficients.

Feature Extraction

• Features for classification are derived from Principal Component Analysis of the appearance templates of the training examples.

• For each object class ci (excluding background) an appearance subspace, represented by d x mi projection matrix Si, is constructed.

• m chosen such that eigenvectors represent 99% of respective subspace.

• Appearance features for base learners are obtained by projecting a training example ‘r’ into appearance subspace of

each object class.

• for two object classes the feature vector v of an example will be,

1 1 1[ ,..., ] Tmv v r S

1 1 1 2 2[ ,..., ] Tm m mv v r S

Pro

babi

lity

Vehicle

Person

Clutter

Histograms of a feature coefficient from the appearance subspace

Row 1: Top 3 eigenvectors for person appearance subspace. Row 2: Vehicle appearance subspace

( | ) ( | )i q j qP c v P c v i j If

: ( )

1( ) arg max log

ic C i h x c i

H x

Where β {i=1…N} are the boosting parameters, C is the set

of classesThe Online Co-training Framework : Key Observations

• Boosting mechanism selects the least correlated base classifiers.

• Ideal for co-training!

• Examples confidently labeled by one classifier are used to train the other.

• Only the observations lying close to the decision boundary of the boosted classifier are useful for improving classification performance.

• Use examples with small margins for online training.

The Online Co-training Framework: Implementation Steps

• Determine confidence thresholds for each base classifier, using a validation data set.

• For class ci and jth base classifier hj set the confidence threshold, Tj,ci

base ( highest probability achieved by a negative example).

• Compute Margin thresholds Tcibase are from the validation data

set.

Foreground Models

Featu

reExtra

ctio

n

ROIs

Background

Background Models

Updated parameters

Color Classifier

Edge Classifier

Base Classifiers

Boosted Classifier

Classification Output

Updated Boosted Parameters

Co-Training Decision (if confident prediction by one set)

Updated weak learners

Information flow for the real-time object classification system

Performance vs. the number of co-trained examples

for the three sequences. Relatively few examples

are required for improving the detection rates since

these examples are from the same scene in which the

classifier is being evaluated.