Top Banner
Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation Victoria, Austalia
22

Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Dec 27, 2015

Download

Documents

Eleanor Bond
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Extreme Re-balancing for SVMs and other classifiers

Presenter: Cui, Shuoyang

2005/03/02

Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation Victoria, Austalia

Page 2: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Imbalance makes the minority-classes samples farther from the true boundary than the majority-class samples.

Majority-class samples dominate the penalty introduced by soft margin.

ideal

Majority

Minority

Page 3: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Data Balancing• up/down samplings• No convincing evidence for how the balanced data

sampled

Imbalance-free algorithm design• Objective function should not be accuracy any longer

Reference:Machine Learning from Imbalanced Data Sets 101http://pages.stern.nyu.edu/~fprovost/Papers/skew.PDF

Page 4: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

In this paper

Exploring the characters of two class learning and analyses situations with supervised learning.

In the experiments offered later, comparing one-class learning with two class learning and list different forms of imbalance compensation.

Page 5: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Two class discrimination

to take examples from these two classes generate a model for discriminating them

for many machine learning algorithms, the training data should include the example form two classes.

Page 6: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

When the data has heavily unbalanced representatives of these two class.

• design re-balancing

• ignore the large pool of negative examples

• learn from positive examples only

Page 7: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Why extreme re-balancing

• Extreme imbalance in very high dimensional input spaces

• Minority class consisting of 1-3% of the total data

• Learning sample size is much below the dimensionality of the input space

• Data site has more than 10,000 features

Page 8: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

The kernel machine

The kernel machine is solved iteratively using the conjugate gradient method.

Designing a kernel machine is to take a standardalgorithm and massage it so that all references to the original data vectors x appear only in dot products ( xi; xj).

Given a training sequence(xi,yi) of binary n—vectors and bipolar labels

Page 9: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Two different cases of kernel machines used here

Page 10: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Two forms of imbalance compensation

• Sample balancing

• Weight balancing

Page 11: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Sample balancing

1:0------the case of 1-class learner using all of the negative examples1:1------ the case of 2-class learner using all training examples 0:1------ the case of 1-class learner using all of the positive examples

Page 12: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Weight balancing

Using different values of the regularisation of the regulation constants for both the minority and majority class data

B is a parameter called a balance factor

Page 13: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Experiments

Real world data collections

• AHR-data

• Reuters data

Page 14: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

• Combined training and test data set

• Each training instance labeled with

“control”, “change” or “nc”

• Convert all of the info from different files to a sparse matrix containing 18330 features

AHR-data

Page 15: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Reuters data

• A collection of 12902 documents

• Each document has been converted to a vector of 20197 dimensional word-presence feature space

Page 16: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

AROC is used as performance measure

AROC is the Area under the ROC

Receiver operating characteristic (ROC) curves are used to describe and compare the performance of diagnostic technology and diagnostic algorithms.

Page 17: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Experiments with Real World Data

• Impact of regularisation constant

• Experiment with sample balancing

• Experiments with weight balancing

Page 18: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Impact of regularisation constant

Page 19: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Experiment with sample balancing

The AROC with 2-class learners is close to 1 for all categories indicating that this categorization problem is easy to learn

Page 20: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Experiments with weight balancing

1.Test on AHRdata

Page 21: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

Experiments with weight balancing

2.Test on ReutersTo observe the performance ouf 1-class and 2-class SVMs when the most features are moved

Page 22: Extreme Re-balancing for SVMs and other classifiers Presenter: Cui, Shuoyang 2005/03/02 Authors: Bhavani Raskutti & Adam Kowalczyk Telstra Croporation.

The characters of test on Reuters

• The accuracy of all classifiers is very high

• SVM models start degenerating, the drop in performance for 2-class SVM is larger.

• 1-class SVM models start outperforming 2-class models

• Similar trends

• AROC is always bigger than 0.5