cse 473 ensemble learning. © cse ai faculty 2 ensemble learning sometimes each learning technique...

28
CSE 473 Ensemble Learning

Upload: allen-floyd

Post on 21-Jan-2016

215 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

CSE 473

Ensemble Learning

CSE 473

Ensemble Learning

Page 2: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 2

Ensemble Learning

• Sometimes each learning technique yields a different hypothesis (or function)

• But no perfect hypothesis…

• Could we combine several imperfect hypotheses to get a better hypothesis?

Page 3: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 3

Example

this line is one simple classifier saying that everything to the left is + and everything to the right is -

Combining 3 linear classifiers

More complex classifier

Page 4: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 4

• Analogies:  Elections combine voters’ choices to pick a good

candidate (hopefully)  Committees combine experts’ opinions to make

better decisions  Students working together on Othello project

• Intuitions:Individuals make mistakes but the “majority” may be less likely to

  Individuals often have partial knowledge; a committee can pool expertise to make better decisions

Ensemble Learning: Motivation

Page 5: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 5

Technique 1: Bagging

• Combine hypotheses via majority voting

Page 6: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 6

Bagging: Analysis

Error probability went down from 0.1 to 0.01!

Page 7: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 7

Weighted Majority Voting

• In practice, hypotheses rarely independent

• Some hypotheses have less errors than others all votes are not equal!

• Idea: Let’s take a weighted majority

Page 8: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 8

Technique 2: Boosting

• Most popular ensemble learning technique  Computes a weighted majority of hypotheses  Can “boost” performance of a “weak learner”

• Operates on a weighted training set  Each training example (instance) has a “weight”  Learning algorithm takes weight of input into account

• Idea: when an input is misclassified by a hypothesis, increase its weight so that the next hypothesis is more likely to classify it correctly

Page 9: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 9

Boosting Example with Decision Trees (DTs)

training casecorrectlyclassified

training casehas large weightin this round

this DT has a strong vote.

Output of hfinal is weighted majority of outputs of h1,…,h4

Page 10: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 10

AdaBoost Algorithm

Page 11: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 11

AdaBoost Example

Original training set D1 : Equal weights to all training inputsGoal: In round t, learn classifier ht that minimizes error with respect to weighted training setht maps input to True (+1) or False (-1)

Taken from “A Tutorial on Boosting” by Yoav Freund and Rob Schapire

Page 12: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 12

AdaBoost ExampleROUND 1

Misclassified Increase weights

z1 = 0.42

Page 13: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 13

AdaBoost Example

ROUND 2

z2 = 0.65

Page 14: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 14

AdaBoost Example

ROUND 3

z3 = 0.92

Page 15: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

© CSE AI Faculty 15

AdaBoost Example

hfinal

sign(x) = +1 if x > 0 and -1 otherwise

Page 16: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Example 1: Semantic Mapping

Page 17: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Motivation

• Human-Robot interaction:• User: “Go to the corridor”

RoomRoom

CorridorCorridor

DoorwayDoorway

Page 18: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Shape

RoomRoom

Page 19: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Observations

RoomRoom DoorwayDoorway

Page 20: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Observations

RoomRoomCorridorCorridor DoorwayDoorway

Page 21: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Simple Features

• Gap = d > θ• f = # Gaps

•Minimum

• f =Area • f =Perimeter • f = d

•d•di

N

1f

•d

• f = d•

•d

•Σ di

Page 22: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Experiments

Training (top) # examples: 16045

•Test (bottom)•# examples: •18726•classification:

•93.94%

•Building 079 Uni. Freiburg

•Room•Room•Corridor•Corridor •Doorway•Doorway

Page 23: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Application to a New Environment

Training map

Intel Research Lab in Seattle

Page 24: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Application to a New Environment

Training map

Intel Research Lab in Seattle

•Room•Room•Corridor•Corridor •Doorway•Doorway

Page 25: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Example 2: Wearable Multi-Sensor Unit

© CSE AI Faculty

•Battery

•Camera (on ribbon cable)

•GPS receiver

•iMote2 +two sensor boards

•Microphone •Camera

•Light •sensors

•2 GB •SD card

•Indicator LEDs

• Records 4 hours of audio, images (1/sec), GPS, and sensor data (accelerometer, barometric pressure, light intensity, gyroscope, magnetometer)

Page 26: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Data Stream

© CSE AI Faculty

Courtesy of G. Borriello

Page 27: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Activity Recognition Model

© CSE AI Faculty

• Boosted cassifiers [Lester-Choudhury-etAl: IJCAI-05]• Virtual evidence boosting [Liao-Choudhury-Fox-Kautz:

IJCAI-07]

•Accuracy: 88% activities, 93% environment

Page 28: CSE 473 Ensemble Learning. © CSE AI Faculty 2 Ensemble Learning Sometimes each learning technique yields a different hypothesis (or function) But no perfect

Boosting• Extremely flexible framework• Handles high-dimensional continuous data• Easy to implement

• Limitation:  Only models local classification problems

28© CSE AI Faculty