Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori...

30
Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International Conference on Pattern Recognition November 14, 2012

Transcript of Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori...

Page 1: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

Multiple Instance Real Boosting with Aggregation Functions

Hossein Hajimirsadeghi and Greg MoriSchool of Computing Science

Simon Fraser University

International Conference on Pattern RecognitionNovember 14, 2012

Page 2: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

2

Multiple Instance Learning

• Traditional supervised learning gets Instance/label pairs

• A kind of weak learning to handle ambiguity in training data

• Standard Definitions:– Positive Bag: At least one of the instances is positive– Negative Bag: All the instances are negative

2x1x

5x4x

3x7x

9x8x

6x

1x10x

2x

7x

9x8x

6x

4x3x

11x12x

• Multiple Instance Learning (MIL) gets bag of instances/label pairs

Page 3: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

3

Applications of MIL

• Image Categorization– [e.g., chen et al., IEEE-TPAMI 2006]

• Content-Based Image Retrieval– [e.g., Li et al., ICCV11]

Page 4: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

4

Applications of MIL

• Text Categorization– [e.g., Andrews et al., NIPS02]

• Object Tracking– [e.g., Babenko et al., IEEE-TPAMI 2011]

Page 5: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

5

Problem & Objective

• The information “At least one of the instances is positive” is very weak and ambiguous.– There are examples of MIL datasets where most

instances in the positive bags are positive. • We aim to mine through different levels of

ambiguity in the data:– For example: a few instances are positive, some

instances are positive, many instances are positive, most instances are positive, …

Page 6: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

6

Approach

• Using the ideas in Boosting:– Finding a bag-level classifier by maximizing the

expected log-likelihood of the training bags– Finding an instance-level strong classifier as a

combination of weak classifiers like RealBoost Algorithm (Friedman et al. 2000), modified by the information from the bag-level classifier

• Using aggregation functions with different degrees of or-ness:– Aggregate the probability of instances to define

probability of a bag be positive

Page 7: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

7

Ordered Weighted Averaging (OWA)

• OWA is an aggregation function:]1,0[]1,0[:owa n

n

i

iin bwaaa1

21 ),...,,owa(

1],1,0[1

n

iii ww

ji aib oflargest th theis

Yager et al. IEEE-TSMC, 1988

Page 8: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

8

OWA: Example

?)6.0,1.0,9.0,5.0owa(

1) Sort the values:0.9, 0.6, 0.5, 0.1

1.05.06.09.0 4321 wwww

Ex: uniform aggregation (mean):

1.04

15.0

4

16.0

4

19.0

4

1

2) Compute the weighted sum:

Page 9: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

9

OWA: Linguistic Quantifiers

• Regular Increasing Monotonic (RIM) Quantifiers– All, Many, Half, Some, At Least One, …

)1

()(n

iQ

n

iQwi

ppQ )(

Page 10: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

10

OWA: RIM Quantifiers

• RIM Quantifier : All ppQ )(

1

1Q

ni

niwi 1

0

)6.0,1.0,9.0,5.0owa(

1.01.0*15.0*06.0*09.0*0

)1

()(n

iQ

n

iQwi

Ex:

Page 11: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

11

OWA: RIM Quantifiers

• RIM Quantifier : At Least One ppQ )(0

10

11

i

iwi

)6.0,1.0,9.0,5.0owa(

9.01.0*05.0*06.0*09.0*1

)1

()(n

iQ

n

iQwi

Ex:

1

1Q

Page 12: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

12

OWA: RIM Quantifiers

• RIM Quantifier : At Least Some ppQ )(5.0

Gives higher weights to the largest arguments

So, some high values are enough to make the result high

1

1

Q

n

1

n

2

n

3

1w

2w3w

Page 13: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

13

OWA: RIM Quantifiers

• RIM Quantifier : Many ppQ )(2

Gives lower weights to the largest arguments

So, many arguments should have high values to make the result high

1

1 Q

n

1

n

2

n

31w

3w2w

Page 14: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

14

OWA: Linguistic Quantifiers Linguistic Quantifier Degree of orness

At least one of them (Max function)

0.999

Few of them 0.1 0.909

Some of them 0.5 0.667

Half of them 1 0.5

Many of them 2 0.333

Most of them 10 0.091

All of them (Min Function)

0.001

0

Page 15: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

15

MIRealBoost

4x2x

3x1x

5x

6x

)( 1xp

)( 3xp

)( 2xp

)( 4xp

)( 5xp

)( 6xp

)( 2Xpk

)( 3Xpk

)( 1Xpk

)( 2XF b

)( 3XF b

)( 1XF b

)(xF

)( 1xF

)( 3xF

)( 2xF

)( 4xF

)( 5xF

)( 6xF OWA

InstanceProbabilities

BagProbabilities

TrainingBags

InstanceClassifier

)(XF b BagClassifier

Page 16: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

16

MIRealBoost

• MIL training input:

• Objective to find the bag classifier:

]1,1[:

)(sign)(

b

bb

H

XFXH

bags possible all ofset

NN YXYXYX ,,...,,,, 212111

imiii xxxX ,...,, 21

]1,1[iY

Page 17: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

17

MIRealBoost: Learning Bag Classifier

• Objective: Maximize the Expected Binomial Log-Likelihood:

)()(

)(

)(XFXF

XF

bb

b

ee

eXp

)(1log)|1()(log)|1(max)(

XpXYPXpXYPXF b

)|1(1

)|1(log

2

1)(

XYP

XYPXF b

• Proved: ?

Page 18: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

18

MIRealBoost: Estimate Bag Prob.

?)|1( XYP

4x1x

3x2x )|( 2xyP

)|( 4xyP

)|( 1xyP

)|( 3xyP)|( XYP

Estimate probability of each instance

Aggregate

• Aggregation functions:• Noisy-OR

• OWA

X

j

jXx xyPxyP1

)|1(11|1(NOR

?

Page 19: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

19

MIRealBoost: Estimate Instance Prob.

• Estimate Instance Probabilities by training the standard RealBoost classifier:

)(sign)( xFxH

M

m

m XfxF1

)()(

)()(

)(

)|1(xFxF

xF

ee

exyP

?)|1( xyP

• Then:

Page 20: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

20

MIRealBoost: Learning Instance Classifier• RealBoost classifier :

)()(1E min xfxFy

f

mm

m

e

)(1),(by weighted

,given of PDF: xyF

w

meyxw

yxP

)1|(

)1|(log

2

1)(

yxP

yxPxf

w

wm

• Proved:

?

Page 21: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

MIRealBoost: Estimate Weak Classifiers

?)|( yxPw

yxPyxPijijij wyxw |)|( ,,

)1|( yxPw)1|( yxPw

Page 22: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

22

MIRealBoost: Estimate Weak Classifiers

ipij Yy

)( ib

i XFYpij ew

)1|(

)1|(log

2

1)(

,,

,,

yxP

yxPxf

pijw

pijyijx

pij

pijij wyx

m

• We do not know true instance labels.

• Estimate the instance label by the bag label, weighted by the bag confidence

yxPijijij wyx |,,

?ijy

?

Page 23: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

23

MIRealBoost Algorithm

4x2x

3x1x

5x6x

)( 1xf km

)( 3xf km

)( 2xf km

)( 4xf km

)( 6xf km )( 5xf km

)( 1xpk

)( 3xpk

)( 2xpk

)( 4xpk)( 5xpk)( 6xpk

)( 2Xpk )( 3Xpk)( 1Xpk

kkk argmax*

)( 2XF b )( 3XF b)( 1XF b

For each feature k=1:K, compute the weak classifier

Compute the instance probabilities

Aggregate the instance probabilities to find bag probabilities

Compute the experimental log likelihood

)(1

)(log

2

1)( *

*

Xp

XpXF

k

kb

Page 24: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

24

Experiments

• Popular MIL datasets:– Image categorization: Elephant, Fox, and Tiger– Drug activity prediction: Musk1 and Musk2

Page 25: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

25

Results

• MIRealBoost classification accuracy with Different Aggregation functions

agg Elephant Fox Tiger Musk1 Musk2

NOR 83 63 72 85 74

Max 77 58 68 85 74

Few 75 58 70 83 72

Some 75 57 73 85 75

Half 72 54 70 90 77

Many 67 52 67 91 75

Most 54 50 51 83 69

All 50 50 50 84 69

Page 26: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

26

Results

• Comparison with MILBoost Algorithm

Method Elephant Fox Tiger Musk1 Musk2

MIRealBoost 83 63 73 91 77

MILBoost 73 58 56 71 61

MILBoost results are reported from Leistner et al. ECCV10

Page 27: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

27

Results

• Comparison between state-of-the-art MIL methodsMethod Elephant Fox Tiger Musk1 Musk2MIRealBoost 83 63 73 91 77MIForest 84 64 82 85 82MI-Kernel 84 60 84 88 89MI-SVM 81 59 84 78 84mi-SVM 82 58 79 87 84MILES 81 62 80 88 83AW-SVM 82 64 83 86 84AL-SVM 79 63 78 86 83EM-DD 78 56 72 85 85MIGraph 85 61 82 90 90miGraph 87 62 86 90 90

Page 28: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

28

Conclusion

• Proposed MIRealBoost algorithm

• Modeling different levels of ambiguity in data– Using OWA aggregation functions which can realize a

wide range of orness in aggregation

• Experimental results showed:– encoding degree of ambiguity can improve the accuracy– MIRealBoost outperforms MILBoost and comparable

with state-of-the art methds

Page 29: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

29

Thanks!

• supported by grants from the Natural Sciences and Engineering Research Council of Canada (NSERC).

Page 30: Multiple Instance Real Boosting with Aggregation Functions Hossein Hajimirsadeghi and Greg Mori School of Computing Science Simon Fraser University International.

30

MIRealBoost: Learning Instance Classifier

• Implementation details:– Each weak classifier is a stump (i.e., built from only

one feature).

– At each step, the best feature is selected as the feature which leads to the bag probabilities, which maximize the empirical log-likelihood of the bags.

)1|][(

)1|][(log

2

1)(

,],[

,],[

ykxP

ykxPxf

pij

pijij

pij

pijij

wykx

wykxkm