The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University...

43
The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University

Transcript of The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University...

Page 1: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

The Knowledge Discovery Process;Data Preparation & Preprocessing

Bamshad MobasherDePaul University

Bamshad MobasherDePaul University

Page 2: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

2

The Knowledge Discovery Process

- The KDD Process

Page 3: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Types of Data Sets

i Record4 Relational records

4 Data matrix, e.g., numerical matrix, crosstabs

4 Document data: text documents: term-frequency vector

4 Transaction data

i Graph and network4 World Wide Web

4 Social or information networks

4 Molecular Structures

i Ordered

4 Video data: sequence of images

4 Temporal data: time-series

4 Sequential Data: transaction sequences

4 Genetic sequence data

i Spatial and multimedia:4 Spatial data: maps

4 Image data

4 Video data

3

Page 4: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

4

Data Objects

i Data sets are made up of data objects.

i A data object represents an entity.

i Examples:

4 sales database: object customers, store items, sales

4 medical database: object patients, treatments

4 university database: object students, professors, courses

i Also called samples , examples, instances, data points,

objects, tuples, vectors.

i Data objects are described by attributes.

i Database rows data objects; columns attributes.

Page 5: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

5

Attributes

i Attribute (or dimensions, features, variables): a data field representing a characteristic or property of a data object

4 E.g., customer _ID, name, address, income, GPA, ….i Types:

4 Nominal (Categorical)4 Ordinal4 Numeric: quantitative

h Interval-scaledh Ratio-scaled

Page 6: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Attribute Types i Nominal (Categorical): categories, states, or “names of things”

4 Hair_color = {auburn, black, blond, brown, grey, red, white}4 marital status, occupation, ID numbers, zip codes4 Often attributes with “yes” and “no” as values4 Binary

h Nominal attribute with only 2 states (0 and 1)

i Ordinal4 Values have a meaningful order (ranking) but magnitude between

successive values is not known.4 Size = {small, medium, large}, grades, army rankings4 Month = {jan, feb, mar, … }

i Numeric4 Quantity (integer or real-valued)4 Could also be intervals or ratios

6

Page 7: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

7

Discrete vs. Continuous Attributes

i Discrete Attribute4 Has only a finite or countably infinite set of values

h E.g., zip codes, profession, or the set of words in a collection of documents

4 Sometimes, represented as integer variables4 Note: Binary attributes are a special case of discrete attributes

i Continuous Attribute4 Has real numbers as attribute values

h E.g., temperature, height, or weight4 Practically, real values can only be measured and represented using

a finite number of digits4 Continuous attributes are typically represented as floating-point

variables

Page 8: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

8

The Knowledge Discovery Process

- The KDD Process

Page 9: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

9

Data Preprocessingi Why do we need to prepare the data?

4 In real world applications data can be inconsistent, incomplete and/or noisyh Data entry, data transmission, or data collection problemsh Discrepancy in naming conventionsh Duplicated recordsh Incomplete or missing datah Contradictions in data

i What happens when the data can not be trusted?4 Can the decision be trusted? Decision making is jeopardized

4 Better chance to discover useful knowledge when data is clean

Page 10: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

10

Data Preprocessing

Page 11: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

11

Data Preprocessing

-2,32,100,59,48 -0.02,0.32,1.00,0.59,0.48

Data CleaningData Cleaning

Data IntegrationData Integration

Data TransformationData Transformation

Data ReductionData Reduction

Page 12: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

12

Data Cleaningi Real-world application data can be incomplete, noisy, and

inconsistent4 No recorded values for some attributes4 Not considered at time of entry4 Random errors4 Irrelevant records or fields

i Data cleaning attempts to:4 Fill in missing values4 Smooth out noisy data4 Correct inconsistencies4 Remove irrelevant data

Page 13: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

13

Dealing with Missing Valuesi Data is not always available (missing attribute values in records)

4 equipment malfunction4 deleted due to inconsistency or misunderstanding4 not considered important at time of data gathering

i Solving Missing Data4 Ignore the record with missing values;4 Fill in the missing values manually;4 Use a global constant to fill in missing values (NULL, unknown, etc.);4 Use the attribute value mean to filling missing values of that attribute;4 Use the attribute mean for all samples belonging to the same class to fill in the

missing values;4 Infer the most probable value to fill in the missing value

h may need to use methods such as Bayesian classification or decision trees to automatically infer missing attribute values

Page 14: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

14

Smoothing Noisy Datai The purpose of data smoothing is to eliminate noise and

“smooth out” the data fluctuations.

Partition into equidepth binsBin1: 4, 8, 15Bin2: 21, 21, 24Bin3: 25, 28, 34

meansBin1: 9, 9, 9Bin2: 22, 22, 22Bin3: 29, 29, 29

boundariesBin1: 4, 4, 15Bin2: 21, 21, 24Bin3: 25, 25, 34

Binning

Ex: Original Data for “price” (after sorting): 4, 8, 15, 21, 21, 24, 25, 28, 34Ex: Original Data for “price” (after sorting): 4, 8, 15, 21, 21, 24, 25, 28, 34

Each value in a bin is replaced by the mean value of the bin.

Min and Max values in each bin are identified (boundaries). Each value in a bin is replaced with the closest boundary value.

Page 15: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

15

Smoothing Noisy Datai Other Methods

Clustering

Regression

Similar values are organized into groups (clusters). Values falling outside of clusters may be considered “outliers” and may be candidates for elimination.

Fit data to a function. Linear regression finds the best line to fit two variables. Multiple regression can handle multiple variables. The values given by the function are used instead of the original values.

Page 16: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

16

Smoothing Noisy Data - Example

ID Outlook Temperature Humidity Windy1 sunny 85 85 FALSE2 sunny 80 90 TRUE3 overcast 83 78 FALSE4 rain 70 96 FALSE5 rain 68 80 FALSE6 rain 65 70 TRUE7 overcast 58 65 TRUE8 sunny 72 95 FALSE9 sunny 69 70 FALSE

10 rain 71 80 FALSE11 sunny 75 70 TRUE12 overcast 73 90 TRUE13 overcast 81 75 FALSE14 rain 75 80 TRUE

ID Temperature7 586 655 689 694 70

10 718 72

12 7311 7514 752 80

13 813 831 85

Bin5

Bin1

Bin2

Bin3

Bin4

Want to smooth “Temperature” by bin means with bins of size 3:

1. First sort the values of the attribute (keep track of the ID or key so that the transformed values can be replaced in the original table.

2. Divide the data into bins of size 3 (or less in case of last bin).3. Convert the values in each bin to the mean value for that bin4. Put the resulting values into the original table

Page 17: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

17

Smoothing Noisy Data - Example

ID Temperature7 586 655 689 694 70

10 718 72

12 7311 7514 752 80

13 813 831 85

Bin5

Bin1

Bin2

Bin3

Bin4

ID Temperature7 646 645 649 704 70

10 708 73

12 7311 7314 792 79

13 793 841 84

Bin5

Bin1

Bin2

Bin3

Bin4

Value of every record in each bin is changed to the mean value for that bin. If it is necessary to keep the value as an integer, then the mean values are rounded to the nearest integer.

Page 18: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

18

Smoothing Noisy Data - Example

The final table with the new values for the Temperature attribute.

ID Outlook Temperature Humidity Windy1 sunny 84 85 FALSE2 sunny 79 90 TRUE3 overcast 84 78 FALSE4 rain 70 96 FALSE5 rain 64 80 FALSE6 rain 64 70 TRUE7 overcast 64 65 TRUE8 sunny 73 95 FALSE9 sunny 70 70 FALSE

10 rain 70 80 FALSE11 sunny 73 70 TRUE12 overcast 73 90 TRUE13 overcast 79 75 FALSE14 rain 79 80 TRUE

Page 19: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

19

Data Integrationi Data analysis may require a combination of data from multiple

sources into a coherent data store

i Challenges in Data Integration:4 Schema integration: CID = C_number = Cust-id = cust#4 Semantic heterogeneity4 Data value conflicts (different representations or scales, etc.)4 Synchronization (especially important in Web usage mining)4 Redundant attributes (redundant if it can be derived from other attributes) --

may be able to identify redundancies via correlation analysis:

i Meta-data is often necessary for successful data integration

Pr(A,B) / (Pr(A).Pr(B))= 1: independent,> 1: positive correlation,< 1: negative correlation.

Page 20: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

20

Data Transformation: Normalizationi Min-max normalization: linear transformation from v to v’

4 v’ = [(v - min)/(max - min)] x (newmax - newmin) + newmin4 Note that if the new range is [0..1], then this simplifies to

v’ = [(v - min)/(max - min)]4 Ex: transform $30000 between [10000..45000] into [0..1] ==>

[(30000 – 10000) / 35000] = 0.514

i z-score normalization: normalization of v into v’ based on attribute value mean and standard deviation4 v’ = (v - Mean) / StandardDeviation

i Normalization by decimal scaling4 moves the decimal point of v by j positions such that j is the minimum number

of positions moved so that absolute maximum value falls in [0..1].4 v’ = v / 10j

4 Ex: if v in [-56 .. 9976] and j=4 ==> v’ in [-0.0056 .. 0.9976]

Page 21: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

21

Normalization: Examplei z-score normalization: v’ = (v - Mean) / Stdevi Example: normalizing the “Humidity” attribute:

Humidity8590789680706595708070907580

Mean = 80.3Stdev = 9.84

Humidity0.480.99-0.231.60-0.03-1.05-1.551.49-1.05-0.03-1.050.99-0.54-0.03

Page 22: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

22

Normalization: Example II

i Min-Max normalization on an employee database4 max distance for salary: 100000-19000 = 810004 max distance for age: 52-27 = 254 New min for age and salary = 0; new max for age and salary = 1

ID Gender Age Salary1 F 27 19,0002 M 51 64,0003 M 52 100,0004 F 33 55,0005 M 45 45,000

ID Gender Age Salary1 1 0.00 0.002 0 0.96 0.563 0 1.00 1.004 1 0.24 0.445 0 0.72 0.32

Page 23: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

23

Data Transformation: Discretizationi 3 Types of attributes

4 nominal - values from an unordered set (also “categorical” attributes)4 ordinal - values from an ordered set4 numeric/continuous - real numbers (but sometimes also integer values)

i Discretization is used to reduce the number of values for a given continuous attribute4 usually done by dividing the range of the attribute into intervals4 interval labels are then used to replace actual data values

i Some data mining algorithms only accept categorical attributes and cannot handle a range of continuous attribute value

i Discretization can also be used to generate concept hierarchies4 reduce the data by collecting and replacing low level concepts (e.g., numeric

values for “age”) by higher level concepts (e.g., “young”, “middle aged”, “old”)

Page 24: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

24

Discretization - Example

Humidity8590789680706595708070907580

Low = 60-69Normal = 70-79High = 80+

HumidityHighHigh

NormalHighHigh

NormalLowHigh

NormalHigh

NormalHigh

NormalHigh

i Example: discretizing the “Humidity” attribute using 3 bins.

Page 25: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

25

Data Discretization Methods

i Binning 4 Top-down split, unsupervised

i Histogram analysis4 Top-down split, unsupervised

i Clustering analysis4 Unsupervised, top-down split or bottom-up merge

i Decision-tree analysis4 Supervised, top-down split

i Correlation (e.g., 2) analysis 4 Unsupervised, bottom-up merge

Page 26: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Simple Discretization: Binningi Equal-width (distance) partitioning

4 Divides the range into N intervals of equal size: uniform grid4 if A and B are the lowest and highest values of the attribute, the

width of intervals will be: W = (B –A)/N.4 The most straightforward, but outliers may dominate

presentation4 Skewed data is not handled well

i Equal-depth (frequency) partitioning4 Divides the range into N intervals, each containing

approximately same number of samples4 Good data scaling4 Managing categorical attributes can be tricky

26

Page 27: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

27

Discretization Without Using Class Labels(Binning vs. Clustering)

Data

Equal interval width (binning)

Equal frequency (binning) K-means clustering leads to better results

Page 28: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

28

Discretization by Classification & Correlation Analysis

i Classification (e.g., decision tree analysis)4 Supervised: Given class labels, e.g., cancerous vs. benign4 Using entropy to determine split point (discretization point)4 Top-down, recursive split

i Correlation analysis (e.g., Chi-merge: χ2-based discretization)4 Supervised: use class information4 Bottom-up merge: merge the best neighboring intervals (those

with similar distributions of classes, i.e., low χ2 values)4 Merge performed recursively, until a predefined stopping

condition

Page 29: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

29

Converting Categorical Attributes to Numerical Attributes

Attributes:Outlook (overcast, rain, sunny)Temperature realHumidity realWindy (true, false)

Attributes:Outlook (overcast, rain, sunny)Temperature realHumidity realWindy (true, false)

OutLook OutLook OutLook Temp Humidity Windy Windyovercast rain sunny TRUE FALSE

0 0 1 85 85 0 10 0 1 80 90 1 01 0 0 83 78 0 10 1 0 70 96 0 10 1 0 68 80 0 10 1 0 65 70 1 01 0 0 64 65 1 0. . . . . . .. . . . . . .

Standard Spreadsheet Format

ID Outlook Temperature Humidity Windy1 sunny 85 85 FALSE2 sunny 80 90 TRUE3 overcast 83 78 FALSE4 rain 70 96 FALSE5 rain 68 80 FALSE6 rain 65 70 TRUE7 overcast 58 65 TRUE8 sunny 72 95 FALSE9 sunny 69 70 FALSE10 rain 71 80 FALSE11 sunny 75 70 TRUE12 overcast 73 90 TRUE13 overcast 81 75 FALSE14 rain 75 80 TRUE

Create separate columns for each value of a categorical attribute (e.g., 3 values for the Outlook attribute and two values of the Windy attribute). There is no change to the numerical attributes.

Page 30: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

30

Data Reduction

i Data is often too large; reducing data can improve performance

i Data reduction consists of reducing the representation of the data set while producing the same (or almost the same) results

i Data reduction includes:4 Data cube aggregation4 Dimensionality reduction4 Discretization4 Numerosity reduction

h Regressionh Histogramsh Clusteringh Sampling

Page 31: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

31

Data Cube Aggregationi Reduce the data to the concept level needed in the analysis

4 Use the smallest (most detailed) level necessary to solve the problem

i Queries regarding aggregated information should be answered using data cube when possible

Page 32: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Dimensionality Reductioni Curse of dimensionality

4 When dimensionality increases, data becomes increasingly sparse4 Density and distance between points, which is critical to clustering, outlier

analysis, becomes less meaningful4 The possible combinations of subspaces will grow exponentially

i Dimensionality reduction4 Avoid the curse of dimensionality4 Help eliminate irrelevant features and reduce noise4 Reduce time and space required in data mining4 Allow easier visualization

i Dimensionality reduction techniques4 Principal Component Analysis4 Attribute subset selection4 Attribute or feature generation

32

Page 33: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Principal Component Analysis (PCA)

i Find a projection that captures the largest amount of variation in data

i The original data are projected onto a much smaller space, resulting in dimensionality reduction4 Done by finding the eigenvectors of the covariance matrix, and these

eigenvectors define the new space

33

x2

x1

e

Page 34: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Principal Component Analysis (Steps)

i Given N data vectors (rows in a table) from n dimensions (attributes), find k ≤ n orthogonal vectors (principal components) that can be best used to represent data 4 Normalize input data: Each attribute falls within the same range4 Compute k orthonormal (unit) vectors, i.e., principal components4 Each input data (vector) is a linear combination of the k principal

component vectors4 The principal components are sorted in order of decreasing

“significance” or strength4 The size of the data can be reduced by eliminating the weak

components, i.e., those with low variance h Using the strongest principal components, it is possible to

reconstruct a good approximation of the original data

i Works for numeric data only

34

Page 35: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

35

Attribute Subset Selection

i Another way to reduce dimensionality of data

i Redundant attributes 4 Duplicate much or all of the information contained in one or

more other attributes

4 E.g., purchase price of a product and the amount of sales tax paid

i Irrelevant attributes4 Contain no information that is useful for the data mining task at

hand

4 E.g., students' ID is often irrelevant to the task of predicting students' GPA

Page 36: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

36

Heuristic Search in Attribute Selection

i There are 2d possible attribute combinations of d attributesi Typical heuristic attribute selection methods:

4 Best single attribute under the attribute independence assumption: choose by significance tests

4 Best step-wise feature selection:h The best single-attribute is picked first. Then next best attribute condition to

the first, ...h {}{A1}{A1, A3}{A1, A3, A5}

4 Step-wise attribute elimination:h Repeatedly eliminate the worst attribute: {A1, A2, A3, A4, A5}{A1, A3, A4,

A5} {A1, A3, A5}, ….

4 Combined attribute selection and elimination4 Decision Tree Induction

Page 37: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

37

Decision Tree InductionUse information theoretic techniques to select the most “informative” attributes

Page 38: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

38

Attribute Creation (Feature Generation)

i Create new attributes (features) that can capture the important information in a data set more effectively than the original ones

i Three general methodologies4 Attribute extraction

h Domain-specific

4 Mapping data to new space (see: data reduction)h E.g., Fourier transformation, wavelet transformation, etc.

4 Attribute construction h Combining featuresh Data discretization

Page 39: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Data Reduction: Numerosity Reduction

i Reduce data volume by choosing alternative, smaller forms of data representation

i Parametric methods (e.g., regression)4 Assume the data fits some model, estimate model

parameters, store only the parameters, and discard the data (except possible outliers)

4 Ex.: Log-linear models—obtain value at a point in m-D space as the product on appropriate marginal subspaces

i Non-parametric methods 4 Do not assume models4 Major families: histograms, clustering, sampling, …

39

Page 40: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

Regression Analysis

i Collection of techniques for the modeling and analysis of numerical data consisting of values of a dependent variable (also response variable or measurement) and of one or more independent variables (aka. explanatory variables or predictors)

i The parameters are estimated to obtains a "best fit" of the data

i Typically the best fit is evaluated by using the least squares method, but other criteria have also been used

i Used for prediction (including forecasting of time-series data), inference, hypothesis testing, and modeling of causal relationships

40

y

x

y = x + 1

X1

Y1

Y1’

Page 41: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

41

i Linear regression: Y = w X + b4 Two regression coefficients, w and b, specify the line and are to be

estimated by using the data at hand

4 Using the least squares criterion on known values of Y1, Y2, …, X1, X2, ….

i Multiple regression: Y = b0 + b1 X1 + b2 X2

4 Many nonlinear functions can be transformed into the above

i Log-linear models4 Approximate discrete multidimensional probability distributions

4 Estimate the probability of each point in a multi-dimensional space for a

set of discretized attributes, based on a smaller subset of dimensions

4 Useful for dimensionality reduction and data smoothing

Regression Analysis

Page 42: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

42

Numerocity Reductioni Reduction via histograms:

4 Divide data into buckets and store representation of buckets (sum, count, etc.)

i Reduction via clustering4 Partition data into clusters based on

“closeness” in space4 Retain representatives of clusters (centroids)

and outliers

i Reduction via sampling4 Will the patterns in the sample represent the

patterns in the data?4 Random sampling can produce poor results4 Stratified sample (stratum = group based on

attribute value)

Page 43: The Knowledge Discovery Process; Data Preparation & Preprocessing Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.

43

Sampling Techniques

SRSWOR

(simple random

sample without

replacement)

SRSWRRaw Data

Raw Data

Cluster/Stratified Sample