Top Banner
Kullback- Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem
21

Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Dec 24, 2015

Download

Documents

Cody Sparks
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Kullback-Leibler Boosting

Ce Liu, Hueng-Yeung ShumMicrosoft Research AsiaCVPR 2003

Presented by Derek Hoiem

Page 2: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

RealBoost Review

Start with some candidate feature set Initialize training sample weights Loop:

Add feature to minimize error bound Reweight training examples, giving more weight to

misclassified examples Assign weight to weak classifier according to

weighted error of training samples Exit loop after N features have been added

Page 3: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

The Basic Idea of KLBoosting

Similar to RealBoost except:Features are general linear projectionsGenerates optimal features Uses KL divergence to select featuresFiner tuning on coefficients

Page 4: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Linear Features

KLBoosting:

VJ Adaboost:

Page 5: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

What makes a feature good?

KLBoosting:

RealBoost:Minimize upper bound on classification error

Page 6: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Creating the feature set

Sequential 1-D OptimizationBegin with large initial set of features (linear

projections)Choose top L features according to KL-Div Initial feature = weighted sum of L featuresSearch for optimal feature in directions of L

features

Page 7: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Example

Initial feature set:

xx

xx

x

x

x

x

Page 8: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Example

Top two features (by KL-Div):

xx

xx

x

x

x

x

w1 w2

Page 9: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Example

Initial feature (weighted combo by KL):

xx

xx

x

x

x

x

w1 w2f0

Page 10: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Example

Optimize over w1

xx

xx

x

x

x

x

w1 w2f1

f1= f0 + B* w1

B = -a1..a1

Page 11: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Example

Optimize over w2

xx

xx

x

x

x

x

w1 w2f2

f2= f1 + B* w2

B = -a2..a2

(and repeat…)

Page 12: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Creating the feature set

First three features

Selecting the first feature

Page 13: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Creating feature set

Page 14: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Classification

= ½ in RealBoost

Page 15: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Parameter Learning

With each added feature k:Set first a1..ak-1 to current optimal value

Set ak to 0

Minimize recognition error on training:

Solve using greedy algorithm

Page 16: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

KLBoost vs AdaBoost

1024 candidate features for AdaBoost

Page 17: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Face detection: candidate features

52,400 2,800450

Page 18: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Face detection: training samples

8760 faces + mirror images 2484 non-face images 1.34B patches Cascaded classifier allows bootstrapping

Page 19: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Face detection: final features

top ten

global semantic

global not semantic

local

Page 20: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Results

xx

xx

8 85 853

Schneiderman (2003)

Test time: .4 sec per 320x240 image

Page 21: Kullback-Leibler Boosting Ce Liu, Hueng-Yeung Shum Microsoft Research Asia CVPR 2003 Presented by Derek Hoiem.

Comments

Training time?

Which improves performance:Generating optimal features?KL feature selection?Optimizing alpha coefficients?