Handling Outliers and Missing Data in Statistical Data Models

48
Handling Outliers and Missing Data in Statistical Data Models Kaushik Mitra Date: 17/1/2011 ECSU Seminar, ISI

description

Handling Outliers and Missing Data in Statistical Data Models. Kaushik Mitra Date: 17/1/2011 ECSU Seminar, ISI. Statistical Data Models. Goal: Find structure in data Applications Finance Engineering Sciences Biological Wherever we deal with data Some examples Regression - PowerPoint PPT Presentation

Transcript of Handling Outliers and Missing Data in Statistical Data Models

Page 1: Handling Outliers and Missing Data in Statistical Data Models

Handling Outliers and Missing Data in Statistical Data Models

Kaushik MitraDate: 17/1/2011ECSU Seminar, ISI

Page 2: Handling Outliers and Missing Data in Statistical Data Models

Statistical Data Models• Goal: Find structure in data• Applications– Finance– Engineering– Sciences

• Biological– Wherever we deal with data

• Some examples– Regression– Matrix factorization

• Challenges: Outliers and Missing data

Page 3: Handling Outliers and Missing Data in Statistical Data Models

Outliers Are Quite CommonGoogle search results for `male faces’

Page 4: Handling Outliers and Missing Data in Statistical Data Models

Need to Handle Outliers Properly

Noisy image Gaussian filtered image Desired result

Removing salt-and-pepper (outlier) noise

Page 5: Handling Outliers and Missing Data in Statistical Data Models

Missing Data Problem

Completing missing tracks

Incomplete tracksCompleted tracks by a sub-optimal method

Desired result

Missing tracks in structure from motion

Page 6: Handling Outliers and Missing Data in Statistical Data Models

Our Focus

• Outliers in regression– Linear regression– Kernel regression

• Matrix factorization in presence of missing data

Page 7: Handling Outliers and Missing Data in Statistical Data Models

Robust Linear Regression for High Dimension Problems

Page 8: Handling Outliers and Missing Data in Statistical Data Models

What is Regression?

• Regression– Find functional relation between y and x• x: independent variable• y: dependent variable

– Given • data: (yi,xi) pairs

• Model y = f(x, w)+n

– Estimate w– Predict y for a new x

Page 9: Handling Outliers and Missing Data in Statistical Data Models

Robust Regression

• Real world data corrupted with outliers• Outliers make estimates unreliable

• Robust regression– Unknown

• Parameter, w• Outliers

– Combinatorial problem• N data and k outliers

• C(N,k) ways

Page 10: Handling Outliers and Missing Data in Statistical Data Models

Prior Work

• Combinatorial algorithms– Random sample consensus (RANSAC)– Least Median Squares (LMedS)• Exponential in dimension

• M-estimators– Robust cost functions– local minima

Page 11: Handling Outliers and Missing Data in Statistical Data Models

Robust Linear Regression model

• Linear regression model : yi=xiTw+ei

– ei, Gaussian noise

• Proposed robust model: ei=ni+si

– ni, inlier noise (Gaussian)

– si, outlier noise (sparse)

• Matrix-vector form– y=Xw+n+s

• Estimate w, s

y1

y2

.

.yN

x1T

x2T

.

.xN

T

n1

n2

. .nN

s1

s2

. .sN

= + +

w1

w2

.wD

Page 12: Handling Outliers and Missing Data in Statistical Data Models

Simplification• Objective (RANSAC): Find w that minimizes the

number of outliers

• Eliminate w • Model: y=Xw+n+s • Premultiple by C : CX=0, N ≥ D– Cy=CXw+Cs+Cn– z=Cs+g

– g Gaussian• Problem becomes: • Solve for s -> identify outliers -> LS -> w

20 ||||||||min Cszss

tosubject

20 ||||||||min sXwysws,

tosubject

Page 13: Handling Outliers and Missing Data in Statistical Data Models

Relation to Sparse Learning

• Solve: – Combinatorial problem

• Sparse Basis Selection/ Sparse Learning• Two approaches :– Basis Pursuit (Chen, Donoho, Saunder 1995)– Bayesian Sparse Learning (Tipping 2001)

20 ||||||||min Cszss

tosubject

Page 14: Handling Outliers and Missing Data in Statistical Data Models

Basis Pursuit Robust regression (BPRR)

• Solve – Basis Pursuit Denoising (Chen et. al. 1995)– Convex problem– Cubic complexity : O(N3)

• From Compressive Sensing theory (Candes 2005)– Equivalent to original problem if

• s is sparse• C satisfy Restricted Isometry Property (RIP)

• Isometry: ||s1 - s2|| = ||C(s1 – s2)||• Restricted: to the class of sparse vectors

• In general, no guarantees for our problem

Cszss

thatsuch1

min

Page 15: Handling Outliers and Missing Data in Statistical Data Models

Bayesian Sparse Robust Regression (BSRR)

• Sparse Bayesian learning technique (Tipping 2001)– Puts a sparsity promoting prior on s : – Likelihood : p(z/s)=Ν(Cs,εI)– Solves the MAP problem p(s/z) – Cubic Complexity : O(N3)

N

i issp

1

1)(

Page 16: Handling Outliers and Missing Data in Statistical Data Models

Setup for Empirical Studies• Synthetically generated data

• Performance criteria– Angle between ground truth and estimated hyper-planes

Page 17: Handling Outliers and Missing Data in Statistical Data Models

Vary Outlier Fraction

BSRR performs well in all dimensions Combinatorial algorithms like RANSAC, MSAC, LMedS not

practical in high dimensions

Dimension = 2 Dimension = 8 Dimension = 32

Page 18: Handling Outliers and Missing Data in Statistical Data Models

Facial Age Estimation • Fgnet dataset : 1002 images of 82 subjects

• Regression– y : Age

– x: Geometric feature vector

Page 19: Handling Outliers and Missing Data in Statistical Data Models

Outlier Removal by BSRR

• Label data as inliers and outliers• Detected 177 outliers in 1002 images

BSRR

Inlier MAE 3.73

Outlier MAE 19.14

Overall MAE 6.45

•Leave-one-out testing

Page 20: Handling Outliers and Missing Data in Statistical Data Models

Summary for Robust Linear Regression

• Modeled outliers as sparse variable • Formulated robust regression as Sparse

Learning problem– BPRR and BSRR

• BSRR gives the best performance• Limitation: linear regression model– Kernel model

Page 21: Handling Outliers and Missing Data in Statistical Data Models

Robust RVM Using Sparse Outlier Model

Page 22: Handling Outliers and Missing Data in Statistical Data Models

Relevance Vector Machine (RVM)

• RVM model:– : kernel function

• Examples of kernels– k(xi, xj) = (xi

Txj)2 : polynomial kernel

– k(xi, xj) = exp( -||xi - xj||2/2σ2) : Gaussian kernel

• Kernel trick: k(xi,xj) = ψ(xi)Tψ(xj)– Map xi to feature space ψ(xi)

N

ii ewkwy

10),()( ixxx

),( ixxk

Page 23: Handling Outliers and Missing Data in Statistical Data Models

RVM: A Bayesian Approach• Bayesian approach– Prior distribution : p(w)– Likelihood :

• Prior specification– p(w) : sparsity promoting prior p(wi) = 1/|wi|– Why sparse?

• Use a smaller subset of training data for prediction• Support vector machine

• Likelihood – Gaussian noise

• Non-robust : susceptible to outliers

),|( wxyp

Page 24: Handling Outliers and Missing Data in Statistical Data Models

Robust RVM model

• Original RVM model– e, Gaussian noise

• Explicitly model outliers, ei= ni + si

– ni, inlier noise (Gaussian)

– si, outlier noise (sparse and heavy-tailed)

• Matrix vector form– y = Kw + n + s

• Parameters to be estimated: w and s

N

ijj ewkwy

10)( xx,

Page 25: Handling Outliers and Missing Data in Statistical Data Models

Robust RVM Algorithms

• y = [K|I]ws + n– ws = [wT sT]T : sparse vector

• Two approaches– Bayesian– Optimization

Page 26: Handling Outliers and Missing Data in Statistical Data Models

Robust Bayesian RVM (RB-RVM)

• Prior specification– w and s independent : p(w, s) = p(w)p(s)– Sparsity promoting prior for s: p(si)= 1/|si|

• Solve for posterior p(w, s|y) • Prediction: use w inferred above • Computation: a bigger RVM– ws instead of w

– [K|I] instead of K

Page 27: Handling Outliers and Missing Data in Statistical Data Models

Basis Pursuit RVM (BP-RVM)

• Optimization approach

– Combinatorial

• Closest convex approximation

• From compressive sensing theory– Same solution if [K|I] satisfies RIP• In general, can not guarantee

20 ||]|[||||||min ssw

wIKyws

tosubject

21 ||]|[||||||min ssw

wIKyws

tosubject

Page 28: Handling Outliers and Missing Data in Statistical Data Models

Experimental Setup

Page 29: Handling Outliers and Missing Data in Statistical Data Models

Prediction : Asymmetric Outliers Case

Page 30: Handling Outliers and Missing Data in Statistical Data Models

Image Denoising

• Salt and pepper noise– Outliers

• Regression formulation– Image as a surface over 2D grid• y: Intensity • x: 2D grid • Denoised image obtained by prediction

Page 31: Handling Outliers and Missing Data in Statistical Data Models

Salt and Pepper Noise

Page 32: Handling Outliers and Missing Data in Statistical Data Models

Some More ResultsRVM RB-RVM Median Filter

Page 33: Handling Outliers and Missing Data in Statistical Data Models

Age Estimation from Facial Images

• RB-RVM detected 90 outliers

• Leave-one-person-out testing

Page 34: Handling Outliers and Missing Data in Statistical Data Models

Summary for Robust RVM

• Modeled outliers as sparse variables• Jointly estimated parameter and outliers• Bayesian approach gives very good result

Page 35: Handling Outliers and Missing Data in Statistical Data Models

Limitations of Regression

• Regression: y = f(x,w)+n– Noise in only “y”– Not always reasonable

• All variables have noise– M = [x1 x2 … xN]– Principal component analysis (PCA)

• [x1 x2 … xN] = ABT – A: principal components– B: coefficients

– M = ABT: matrix factorization (our next topic)

Page 36: Handling Outliers and Missing Data in Statistical Data Models

Matrix Factorization in the presence of Missing Data

Page 37: Handling Outliers and Missing Data in Statistical Data Models

Applications in Computer Vision

• Matrix factorization: M=ABT

• Applications: build 3-D models from images– Geometric approach (Multiple views)

– Photometric approach (Multiple Lightings)

37

Structure from Motion (SfM)

Photometric stereo

Page 38: Handling Outliers and Missing Data in Statistical Data Models

Matrix Factorization

• Applications in Vision– Affine Structure from Motion (SfM)– Photometric stereo

• Solution: SVD– M=USVT

– Truncate S to rank r• A=US0.5, B=VS0.5

38

M =xij

yij= CST

Rank 4 matrix

M = NST, rank = 3

Page 39: Handling Outliers and Missing Data in Statistical Data Models

Missing Data Scenario

• Missed feature tracks in SfM

• Specularities and shadow in photometric stereo

39

Incomplete feature tracks

Page 40: Handling Outliers and Missing Data in Statistical Data Models

Challenges in Missing Data Scenario

• Can’t use SVD• Solve:

• W: binary weight matrix, λ: regularization parameter• Challenges

– Non-convex problem– Newton’s method based algorithm (Buchanan et. al. 2005)

• Very slow• Design algorithm

– Fast (handle large scale data) – Flexible enough to handle additional constraints

• Ortho-normality constraints in ortho-graphic SfM

)||||||(||||)(||min 222FFF

T BAABMWBA,

Page 41: Handling Outliers and Missing Data in Statistical Data Models

Proposed Solution

• Formulate matrix factorization as a low-rank semidefinite program (LRSDP)– LRSDP: fast implementation of SDP (Burer, 2001) • Quasi-Newton algorithm

• Advantages of the proposed formulation:– Solve large-scale matrix factorization problem– Handle additional constraints

41

Page 42: Handling Outliers and Missing Data in Statistical Data Models

Low-rank Semidefinite Programming (LRSDP)

• Stated as:• Variable: R• Constants• C: cost • Al, bl: constants

• Challenge• Formulating matrix factorization as LRSDP• Designing C, Al, bl

klbtosubject lT

lT ,...,2,1,min RRARRC

R

Page 43: Handling Outliers and Missing Data in Statistical Data Models

Matrix factorization as LRSDP: Noiseless Case

• We want to formulate:

• As:• LRSDP formulation:

),()(||||||||min ,,22

,jifortosubject jiji

TFF

BAMABBA

)(||||||||

)(||||),(||||22

22

TFF

TF

TF

trace

tracetrace

RRBA

BBBAAA

jimjiT

jijiT

,,,, )()( MRRMAB

||,...,2,1, lbtosubject lTT RRARRC l

C identity matrix,Al indicator matrix

Page 44: Handling Outliers and Missing Data in Statistical Data Models

Affine SfM

• Dinosaur sequence

• MF-LRSDP gives the best reconstruction

72% missing data

Page 45: Handling Outliers and Missing Data in Statistical Data Models

Photometric Stereo

• Face sequence

• MF-LRSDP and damped Newton gives the best result

42% missing data

Page 46: Handling Outliers and Missing Data in Statistical Data Models

Additional Constraints:Orthographic Factorization

• Dinosaur sequence

Page 47: Handling Outliers and Missing Data in Statistical Data Models

Summary

• Formulated missing data matrix factorization as LRSDP– Large scale problems– Handle additional constraints

• Overall summary– Two statistical data models• Regression in presence of outliers

– Role of sparsity• Matrix factorization in presence of missing data

– Low rank semidefinite program

Page 48: Handling Outliers and Missing Data in Statistical Data Models

Thank you! Questions?

48