PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

29
PRESENTED BY: SAUPTIK DHAR PRACTICAL CONDITIONS FOR EFFECTIVENESS OF THE UNIVERSUM LEARNING 1

Transcript of PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

Page 1: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

PRESENTED BY: SAUPTIK DHAR

PRACTICAL CONDITIONS FOR EFFECTIVENESS OF THE UNIVERSUM

LEARNING

1

Page 2: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

AGENDA

HISTOGRAM OF PROJECTION

UNIVERSUM LEARNING

RESULTS

CONCLUSION

FUTURE IDEAS/REFERENCE

2

Page 3: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

HISTOGRAM OF PROJECTION3

MOTIVATION

BASICS FOR HISTOGRAM OF PROJECTION

UTILITY

Page 4: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

MOTIVATION FOR UNIVARIATE HISTOGRAM OF PROJECTION

4

Many applications in Machine Learning involve sparse high-dimensional

data low sample size (HDLSS) where ,n << d where, n=No. of Samples and d= No. of Dimensions

• Medical imaging (i.e., sMRI, fMRI).• Object and face recognition• Text categorization and retrieval• Web search.

Need a way to visualize the high dimensional data.

Page 5: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

UNIVARIATE HISTOGRAM OF PROJECTIONS

5

Project training data onto normal vector w of the trained SVM

The projection is , so we can also have projections for nonlinear SVM.

( )f x

( ( )) ( )y sign f sign b x w x

W0-1

+1

0-1 +1

( )f b x w x

Page 6: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

(SYNTHETIC) HYPERBOLA DATA6

Coordinate x1 = ((t-0.4)*3)2+0.225 Coordinate x2 = 1-((t-0.6)*3)2-0.225.

for class 1. (Uniformly distributed) for class 2. (Uniformly distributed)

Gaussian noise is added to both x1 and x2 co-ordinates, with standard deviation(σ) = 0.025

[0.2,0.6]t [0.4,0.8]t

• No. of Training samples = 500. (250 per class).• No. of Validation samples = 500.(This independent validation set is used for

Model selection).• Dimension of each sample = 2.

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.90.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Page 7: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

MODEL SELECTION7

MODEL SELECTION

[STEP 1] Build the SVM model for each (C, γ) values using the training data samples.

[STEP 2] Select the SVM model parameter (C*, γ*) that provides the smallest classification error on the validation data samples.

Page 8: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

TYPICAL HISTOGRAM OF PROJECTION

8

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

-4 -3 -2 -1 0 1 2 3 40

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.90.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

-2.5 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.50

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

( ( )) ( )y sign f sign b x w x ( ) ( )k kf b x x wHistogram for

( ( )) ( ( , ) )i i iy sign f sign y K b x x x ( ) ( ( , ))k i i i kf y K b x x xHistogram for

Page 9: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

MNIST Data (Handwritten 0-9 digit data set)

9

TASK :- Binary classification of digit “5” vs. digit “8”

• No. of Training samples = 1000. (500 per class).• No. of Validation samples = 1000.(This independent validation set is used for

Model selection).• No. of Test samples = 1866.• Dimension of each sample = 784(28 x 28).

28 pixel

28 pixel

28 pixel

28 pixel

Digit “5” Digit “8”

Page 10: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

TYPICAL HISTOGRAM OF PROJECTION

10

-2.5 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.50

50

100

150

200

250

-2.5 -2 -1.5 -1 -0.5 0 0.5 1 1.5 20

50

100

150

-3 -2 -1 0 1 2 30

50

100

150

200

250

(a)Histogram of projections of MNIST training data onto normal direction of RBF SVM decision boundary. Training set size ~ 1,000 samples. Training error(%)=0 (0/1000)

(c)Histogram of projections of MNIST Test data onto normal direction of RBF SVM decision boundary. Test set size ~ 1866 samples. Test error (%)=1.2326(23/1866)

(b)Histogram of projections of MNIST validation data onto normal direction of RBF SVM decision boundary. Validation set size ~ 1,000 samples. Validation error (%)=1.7 (17/1000)

Page 11: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

TYPICAL HISTOGRAM FOR HDLSS DATA11

-1.5 -1 -0.5 0 0.5 1 1.50

2

4

6

8

10

12

14

16

-3 -2 -1 0 1 2 30

50

100

150

200

250

-1.5 -1 -0.5 0 0.5 1 1.510

0

101

102

103

CASE 1 CASE 2

CASE 3

Page 12: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

UNIVERSUM LEARNING

MOTIVATION OF UNIVERSUM LEARNING

BASICS FOR UNIVERSUM LEARNING

OPTIMIZATION FORMULATION

EFFECTIVENESS FOR UNIVERSUM

12

Page 13: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

MOTIVATION OF UNIVERSUM LEARNING

MOTIVATION Inductive learning usually fails with high-dimensional, low

sample size (HDLSS) data: n << d .

POSSIBLE MODIFICATIONS Predict only for given test points transduction A priori knowledge in the form of additional ‘typical’

samples learning through contradiction Additional (group) info about training data Learning with

structured data Additional (group) info about training + test data Multi-

task learning

13

Page 14: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

Universum Learning (Vapnik, 1998)

14

Motivation: include a priori knowledge about the data

Example: Handwritten digit recognition 5 vs. 8 we may Incorporate priori

knowledge about the data space by using:- Data samples: digits other than 5 or 8

Data samples: randomly mixing pixels from images 5 or 8

Data samples: average of randomly selected examples of 5 and 8

Page 15: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

UNIVERSUM LEARNING FOR DUMMIES

Which boundary is better?

15

CLASS 1CLASS 2

UNIVERSUM

Page 16: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

OPTIMIZATION FORMULATIONGIVEN (Labeled samples + unlabeled Universum samples)

Primal Problem

minimize where

subject to

slack variable for Labeled samples

slack variable for Universum samples

NOTE

Universum samples use -insensitive loss

control the trade-off between min error and max

number of contradictions

When standard soft margin SVM

m

jj

n

ii CCbR

1

**

1

)(2

1),( www 0, * CC

iii by 1])[( xw 0, 1,...,i i n

*( )j jb w x mjj ,...,1,0*

i*j

( ) 0yf x 11

i*j

0, * CC

* 0C

Page 17: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

EFFECTIVENESS OF UNIVERSUM LEARNING

17

• Random Averaging (RA) Universum:– RA Universum does not depend on application domain– RA samples expected to fall inside the margin borders

• Properties of RA Universum depend on characteristics of labeled training data.

• Use the new form of model representation: univariate histograms

Average

Class 1

Class -1 Hyper-plane

Page 18: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

CONDITION FOR EFFECTIVENESS OF RA U-SVM

18

RA U-SVM is effective only for this Type 2 of histogram

-3 -2 -1 0 1 2 30

50

100

150

200

250

Page 19: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

EXPERIMENTAL SETUP19

DATASETS USED Synthetic 1000-dimensional hypercube data set. X~ U[0,1] dimension 1000 of which

200 are significant i.e y=sign(x1+x2+…+x200 – 100).(We use only Linear SVM)

No. of Training samples= 1000No. of Validation samples = 1000No. of Test samples= 5000

Real-life MNIST handwritten digit data set, where data samples represent handwritten digits 5 and 8. Each sample is represented as a real-valued vector of size 28*28=784. No. of Training samples= 1000

No. of Validation samples = 1000No. of Test samples= 1866

Real-life ABCDETC data set, where data samples represent handwritten lower case letters ‘a’ and ‘b’. Each sample is represented as a real-valued vector of size 100*100=10000.

No. of Training samples = 150 (75 per class).No. of Validation samples = 150 (75 per class).No. of Test samples = 209 (105 class ‘a’ , 104 class ‘b’)

Page 20: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

MODEL SELECTION20

[1]Perform model selection for standard SVM classifier, i.e. choose parameter and kernel parameter. Most practical applications use RBF kernel of the form where possible values of parameter C=[0.01, 0.1, 1, 10, 100, 1000] and γ = [2-8, 2-6, …, 22, 24] during model selection.

[2]Using fixed values of and , as selected above, tune additional parameters specific to U-SVM, as follows:For the ratio C*/C , try all values in the range ~ [0.01, 0.03, 0.1, 0.3, 1, 3, 10] parameter , try all values in the range ε ~ [0,0.02,0.05,0.1,0.2] for the number of Universum, it is suggested to use the number in the range of .If the dimensionality of the data is large, smaller number of samples will be used due to the computational consideration. where, n= No. of samples in Class 1. m= No. of samples in Class 2.

Note: steps 1 and 2 above is done by using an independent validation data set.

n m

Page 21: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

HISTOGRAM OF PROJECTIONS21

-2.5 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.50

50

100

150

200

250

-6 -4 -2 0 2 4 60

20

40

60

80

100

120

-3 -2 -1 0 1 2 30

50

100

150

200

250

300

Histogram of projections of MNIST training data onto normal direction of RBF SVM decision boundary. Training set size ~ 1,000 samples.

Histogram of projections of ABCDETC training data onto normal direction of Polynomial SVM decision boundary with d=3. Training set size ~ 150 samples.

(a) MNIST data set (b) synthetic data set

Histogram of projections onto normal direction of linear SVM hyperplane.

Page 22: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

RESULTS22

SVM U-SVM(RA) Synthetic data (Linear Kernel) 26.63% (1.54%) 26.89% (1.55%) MNIST(Linear Kernel) 4.58%(0.34%) 4.62%(0.37%) MNIST (RBF Kernel) 1.37% (0.22%) 1.20% (0.19%) ABCDETC (Poly Kernel d=3) 20.48%(2.60%) 18.85 %(2.81%)

TABLE : Average percent of Test error over 10 partitioning of dataset.(with the standard deviation in parenthesis).

Page 23: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

INSIGHTS23

FOR EFFECTIVE PERFORMANCE OF RANDOM AVERAGING

Training data is well-separable (in some optimally chosen kernel space).

The fraction of training data samples that project inside the margin borders is small.

QUESTIONS What are good universum samples?

Can we identify good universum samples using the univariate histogram of projection?

Page 24: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

Conditions for Effectiveness of the Universum

24

The histogram projection of the Universum samples is symmetric relative to (standard) SVM decision boundary.

The histogram projection of the Universum samples has wide distribution between margin borders denoted as points -1/+1 in the projection space.

Page 25: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

RESULTS25

MNIST DATA binary classification ‘5’ vs. ‘8’. UNIVERSUM :- Digit ‘1’, ‘3’ and ‘6’

TABLE : Average percent of Test error over 10 partitioning of dataset.(with the standard deviation in parenthesis). Training /Validation set size is 1000 samples.

SVM U-SVM (digit 1) U-SVM(digit 3) U-SVM(digit 6) Test error 1.47% (0.32%) 1.31% (0.31%) 1.01% (0.28%) 1.12% (0.27%)

-3 -2 -1 0 1 2 30

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

-3 -2 -1 0 1 2 30

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

-3 -2 -1 0 1 2 30

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

0.45

0.5

Digit ‘1’ Digit ‘3’ Digit ‘6’

Page 26: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

RESULTS26

ABCDETC DATA binary classification ‘a’ vs. ‘b’ .UNIVERSUM:- ‘A-Z’ , ‘0-9’, RA Universum samples.

TABLE : Average percent of Test error over 10 partitioning of dataset.(with the standard deviation in parenthesis). Training /Validation set size is 150 samples.

SVM U-SVM(upper case)

U-SVM(all digits) U-SVM(RA)

Test error 20.47%( 2.60%) 18.42 %( 2.97%) 18.37 %( 3.47%) 18.85 %( 2.81%)

-2 -1.5 -1 -0.5 0 0.5 1 1.50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

-2 -1.5 -1 -0.5 0 0.5 1 1.50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

-2 -1.5 -1 -0.5 0 0.5 1 1.50

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

A-Z(uppercase)

0-9(Digits) Random Averaging

Page 27: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

CONCLUSIONS27

PRACTICAL CONDITIONS Training data is well-separable (in some optimally chosen kernel space). The histogram projection of the Universum samples is symmetric relative to

(standard) SVM decision boundary. The histogram projection of the Universum samples has wide distribution

between margin borders denoted as points -1/+1 in the projection space.

ESSENSE(SIMPLE RULE)

Estimate standard SVM classifier for a given (labeled) training data set

Generate low-dimensional representation of training data by projecting it onto the normal direction vector of the SVM hyper plane estimated in (a);

Project the Universum data onto the normal direction vector of SVM hyper plane, and analyze projected Universum data in relation to projected training data. Specifically, the Universum is expected to yield improved prediction accuracy (over standard SVM) only if the conditions stated above are satisfied.

Page 28: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

REFERENCE[1] Vapnik, V.N., Statistical Learning Theory, Wiley, NY 1998.

[2] Cherkassky, V., and Mulier, F. (2007), Learning from Data Concepts: Theory and Methods, Second Edition, NY: Wiley.

[3] Weston, J., Collobert, R., Sinz, F., Bottou, L. and Vapnik, V., Inference with Universum, Proc. ICML 2006

[4] Vladimir Cherkassky and Wuyang Dai,'Empirical Study of the Universum SVM Learning for High-Dimensional Data',ICANN 2009.

[5] Sinz, F. H., O. Chapelle, A. Agarwal and B. Schölkopf, ‘An Analysis of Inference with the Universum.’ Advances in Neural Information Processing Systems 20: Proceedings of the 2007 Conference, 1369-1376. (Eds.) Platt, J. C., D. Koller, Y. Singer, S. Roweis, Curran, Red Hook, NY, USA (09 2008)

[6] Vladimir Cherkassky , Sauptik Dhar and Wuyang Dai,"Practical Conditions for Effectiveness of the Universum Learning“,IEEE Trans. on Neural Networks,May 2010.(submitted).

[7] Vladimir Cherkassky , Sauptik Dhar,"Simple Method for Interpretation of High-Dimensional Nonlinear SVM Classification Models",The 6th International Conference on Data Mining 2010.(submitted).

FUTURE IDEAS Devise a scheme to generate the Universum samples that are uniformly spread out within the soft-margin.{-1,+1}

Clever Feature selection using the Universum samples.

Extend Universum for Non Standard Setting.

Extend Universum for Multi-Category case.

Page 29: PRESENTED BY: SAUPTIK DHAR P RACTICAL C ONDITIONS FOR E FFECTIVENESS OF THE U NIVERSUM L EARNING 1.

THEORETICAL INSIGHTS29

PROBLEM 1

PROBLEM 2