Improving Support Vector Machine through Parameter Optimized

19
Improving Support Vector Machine through Parameter Optimized Rujiang Bai, Junhua Liao Shandong University of Technology Library Zibo 255049, China { brj, ljhbrj}@sdut.edu.cn

description

Rujiang Bai, Junhua Liao Shandong University of Technology Library Zibo 255049, China { brj, ljhbrj}@sdut.edu.cn. Improving Support Vector Machine through Parameter Optimized. Outline. Motivation and Background RGSC(Rough sets and Genetic algorithms for SVM classifier) Experiment - PowerPoint PPT Presentation

Transcript of Improving Support Vector Machine through Parameter Optimized

Page 1: Improving Support Vector Machine through Parameter Optimized

Improving Support Vector Machine through Parameter Optimized

Rujiang Bai, Junhua LiaoShandong University of Technology Library

Zibo 255049, China{ brj, ljhbrj}@sdut.edu.cn

Page 2: Improving Support Vector Machine through Parameter Optimized

OutlineOutline• Motivation and BackgroundMotivation and Background• RGSC(Rough sets and Genetic RGSC(Rough sets and Genetic

algorithms for SVM classifier)algorithms for SVM classifier)• ExperimentExperiment• ConclusionConclusion

Page 3: Improving Support Vector Machine through Parameter Optimized

Motivation and BackgroundMotivation and Background

Three problemsThree problems

1 、 the high dimensionality of input feature vectors impacts on the classification speed.

2、 The kernel parameters setting for SVM in a training process impacts on the classification accuracy.

3、 Feature selection also impacts classification accuracy.

Page 4: Improving Support Vector Machine through Parameter Optimized

Motivation and BackgroundMotivation and Background

GoalGoal

The objective of this work is to reduce the dimension of feature vectors, optimizing the parameters to improve the SVM classification accuracy and speed.•MethodMethod

In order to improve classification speed we spent rough sets theory to reduce the feature vector space. We present a genetic algorithm approach for feature selection and parameters optimization to improve classification accuracy.

Page 5: Improving Support Vector Machine through Parameter Optimized

Dataset

Preprocessi ng

Features reduce usi ngRough Set Theory

Trai ni ng set Testi ng set

Sel ected feature subset

Trai ni ng setwi th sel ectedfeature subset

Testi ng set wi thsel ected feature

subset

Trai ni ng SVMcl assi fi er

Trai ned SVMcl assi fi er

Fi tness eval uati on

Termi nati on aresati si f ed?

Opti mi zed(C, )andfeature subset

Opti mi zed SVMcl assi fi er

Cl assi fi cati onResul ts

Web pages

Preprocessi ng

Features reduce

.

.

Popul ati on

. . .

. . .

Converti nggenotype tophenotype

Parametergenes

Featuregenes

Phenotype offeature genes

Phenotype ofparameter genes

Cl assi fi cati on accuracyfor testi ng set

Geneti coperatati on

No

Yes

Experi ment

Preprocessi ng andFeatures reduce

Feature sel ecti on andparameters opti mi zati on

Page 6: Improving Support Vector Machine through Parameter Optimized

(1) Preprocessing: preprocessing includes remove HTML tags, segment word and construct Vector Space Model.

(2) Feature reduction by rough sets. Our objective is to find a reduction with minimal number of attributes. Described in Alg. 1.

(3) Converting genotype to phenotype. This step will convert each parameter and feature chromosome from its genotype into a phenotype.

(4) Feature subset. After the genetic operation and converting each feature subset chromosome from the genotype into the phenotype, a feature subset can be determined.

RGSCRGSC

Page 7: Improving Support Vector Machine through Parameter Optimized

(5) Fitness evaluation. For each chromosome representing C, γ and selected features, training dataset is used to train the SVM classifier, while the testing dataset is used to calculate classification accuracy. When the classification accuracy is obtained, each chromosome is evaluated by fitness function— formula (8).

RGSCRGSC

Page 8: Improving Support Vector Machine through Parameter Optimized

(6) Termination criteria. When the termination criteria are satisfied, the process ends; otherwise, we proceed with the next generation.

(7) Genetic operation. In this step, the system searches for better solutions by genetic operations, including selection, crossover, mutation, and replacement.

(8) Input the preprocessed data sets into the obtained optimized SVM classifier.

RGSCRGSC

Page 9: Improving Support Vector Machine through Parameter Optimized

Algorithm of RST-based feature Algorithm of RST-based feature reducereduce

RGSCRGSC

Page 10: Improving Support Vector Machine through Parameter Optimized

fVDCUT ,,,

mxxxU ,, 21

nCCCC ,, 21

)( M

jc

ic

ic

ic

krrr ,, 21

kr

1r

fVdduU ,},{Re,

Algorithm : Rough Sets Attribute Reduction algorithm

Input: a decision table

,

,

Output: a reduction of T , denoted as Redu.1. construct the binary discernibility matrix M of

T ;2. delete the rows in the M which are all 0’s,

Redu=

/* delete pairs of inconsistent objects*/

3. while

4. {(1) select an attribute ci in the M with the highest discernibility degree (if there are several

(j=1,2,…,m) with the same highest discernibility degree, choose randomly an attribute from them);

5. (2) Redu

Redu

;6. (3) remove the rows which have ‘‘1’’ in

the

column from M;7. (4) remove

the

column from M; }endwhile/* the following steps remove redundant attributes from

Redu */8. suppose that Redu =

contains k attributes which are sorted by the order of entering Redu,

is the first attributes chosen into Redu,

is the last one chosen into Redu.9. get the binary discernibility matrix MR of decision table

TR=

;10. delete the rows in the MR which are all

0’s;

Page 11: Improving Support Vector Machine through Parameter Optimized

Chromosome designChromosome designTo implement our proposed approach, this

research used the RBF kernel function for the SVM classifier .

The chromosome comprises three parts, C, γ, and the features mask.

RGSCRGSC

1Cg i

Cg CnCg

ng fn

fg1g jg k

fg1fg

Page 12: Improving Support Vector Machine through Parameter Optimized

Fitness functionFitness function

WA SVM classification accuracy weightSVM_accuracy SVM classification accuracyWF weight for the number of featuresCi cost of feature iFi ‘1’ represents that feature i is selected; ‘0’ represents that feature i is not

selected

RGSCRGSC

fn

iiiFA FCWaccuracySVMWfitness

1

1)(_

Page 13: Improving Support Vector Machine through Parameter Optimized

ExperimentsExperiments

• Experiment environmentExperiment environment

Our implementation was carried out on the YALE(Yet Another Learning Environment) 3.3 development environment(Available at:http://rapid-i.com/).

Feature reduction by Rough Sets Theory carried out on ROSETTA(you can download it from http://rosetta.sourceforge.net/) .

The empirical evaluation was performed on Intel Pentium IV CPU running at 3.0 GHz and 1GB RAM.

Page 14: Improving Support Vector Machine through Parameter Optimized
Page 15: Improving Support Vector Machine through Parameter Optimized
Page 16: Improving Support Vector Machine through Parameter Optimized
Page 17: Improving Support Vector Machine through Parameter Optimized
Page 18: Improving Support Vector Machine through Parameter Optimized

ConclusionConclusion

In this paper, we have proposed a document classification method using an SVM based on Rough Sets Theory and Genetic Algorithms. The feature vectors are reduced by Rough Set Theory. The feature vectors are selected and parameters optimization by Genetic Algorithms. The experimental results show that the RGSC we proposed yields the best result of these three methods.

Page 19: Improving Support Vector Machine through Parameter Optimized

Thank you!