Learning to Detect Important People in Unlabelled Images for … · 2020. 8. 9. · ness weighting...

1
Methodology Experimental Results Datasets * EMS Dataset: 10, 687 images from multiple scenes. * ENCAA Dataset: 28, 798 frames of basketball game video. Comparisons with related methods on both datasets Evaluations of different components and techniques used to estimate importance score in proposed method Visualize results Introduction Learning to Detect Important People in Unlabelled Images for Semi-supervised Important People Detection FaTing Hong ! , WeiHong Li # and WeiShi Zheng ! ! Sun Yat-sen University # University of Edinburgh Introduction & Motivation * Existing methods of important people detection require massive quantities of labelled data and detecting important people in unlabeled images has not yet been developed. * The imbalance between the number of important people and non-important people in the picture will cause pseudo-labelling imbalance problem. * Not all unlabelled images contain important people; images without such people represent noisy unlabelled samples during learning. Contributions * The proposed approach is the first to study on learning important people detection from partially labelled data. * we contribute two large datasets called Extended-MS (EMS) and Extended- NCAA (ENCAA) for evaluation of semi-supervised important people detection by augmenting existing datasets with a large number of unlabelled images collected from the internet * Extensive experiments verify the efficacy of our proposed method on important people detection of semi-supervised phase. Detecting Noisy Unlabelled Images * Image-specific effectiveness weight : =1− * * Effectiveness weight reflect the confidence that an unlabelled image features important people = / 0 01! / 3 ( 7 3 , ̃ 3 ) < = @ Overview of Proposed Method * To alleviate the pseudo-labelling imbalance problem, we introduce a ranking strategy for pseudo-label estimation, and also introduce two weighting strategies applied to unlabelled data loss. * The final objective function can be expressed as: = + = 1 / / ℓ( 7 3 , 3 ) < = @ 01! + / 0 01! / 3 ( 7 3 , ̃ 3 ) < = @ . . 3 ∈= * , 0 =1− I J I K Pseudo-labelling by Ranking-based Sampling: * Ranking-based sampling procedure: 0 , ̃ = ( R , 3 < = ∈T @ , , ) * Replacing the unlabelled people and corresponding pseudo-labels with those sampled by RankS: = / / ( 7 3 , ̃ 3 ) < = @ 01! Balancing Loss via Importance Score Weighting: * Person-specific importance score weight: V = / / 3 ( 7 3 , ̃ 3 ) < = @ 01! , s. t. / 3 Y 31! = 1, 3 > 0. labelled images unlabelled images ! " stop using noisy image input labelled image input unlabelled image pseudo-labelling via Ranking-based Sampling noisy image labelled important person non-important person from unlabelled image classification space adapting hyperplane using unlabelled data " ( , ) " ( , ) " ( , ) ' ( , ) ' ( , ) feature encoder feature encoder Importance Score Ranking & Labelling Sampling Pseudo-labelling via Ranking-based Sampling prediction prediction Sampling individuals according to the Ranking- based Sampling Procedure ( ) ' ( , ) ‘important' 'non-important + " , - /(( ) ) 1− 3(( 4 ) 3(5) , 6 , 7 8 Unlabelled Image Labelled Image 9=9 " + <9 ' importance feature importance feature relation graph relation graph + = " + = ' > " > ' unlabelled individuals unlabelled individuals (a) Score of “important” class in an unlabelled image 0.5 0 1 Score of “important” class (b) Pseudo-labelling in Current Semi- supervised Learning approach 0.5 0 1 Score of “important” class (c) Our Ranking-based Sampling 0 1 Ranking Score (d) Labelled data in classification space (e) Adding unlabelled data via current semi-supervised methods ( f ) Adding unlabelled data via our Ranking-based Sampling softmax classifier threshold Adapting hyperplane Dataset EMS ENCAA #labelled images 33 % 66 % 100 % 33 % 66 % 100 % POINT (fully supervised) 83.36 85.97 88.48 84.60 88.21 89.75 Pseudo Label (PL) 83.37 85.35 88.57 85.70 88.43 90.56 Label Propagation (LP) 82.34 86.33 86.66 85.36 88.61 90.18 Mean Teacher (MT) 84.50 86.29 87.55 83.33 84.66 87.55 Ours 87.81 88.44 89.79 88.75 90.86 92.03 Table 1. Comparison with related methods on both datasets. Dataset EMS ENCAA #labelled images 33 % 66 % 100 % 33 % 66 % 100 % Ours w/o Ranks, ISW and EW 83.70 86.81 87.67 84.35 87.66 89.93 Ours w/o ISW and EW 85.55 87.25 88.53 87.13 90.53 91.49 Ours w/o EW 86.34 87.45 89.67 87.68 90.60 92.00 Ours 87.81 88.44 89.79 88.75 90.86 92.03 Table 2. Ablation study on both datasets. RankS repre- sents ranking-based sampling while ISW and EW indicate impor- tance score weighting and effectiveness weighting, respectively. Ours w/o ISW and EW means our model without using ISW and EW. Dataset EMS ENCAA Dataset EMS ENCAA #labelled images 33 % 66 % 100 % 33 % 66 % 100 % Ours LP 87.51 88.10 89.65 88.95 91.06 91.98 Ours MT 87.23 88.56 90.72 88.97 90.93 91.62 Ours 87.81 88.44 89.79 88.75 90.86 92.03 Table 3. Evaluation of different techniques (i.e., LP and MT) when used for instantiating pseudo-label estimation function (i.e., g (·)) instead of using Softmax function. (") ($) Fgiure (a) is the distribution of top 8 importance score in testing set in EMS datasets and Figure (b) is the statistics of unlabelled data’s pseudo-labels on EMS dataset. Better view in color. Matching Rate (%) Rank Matching Rate (%) Rank (a) CMC curve on EMS dataset (b) CMC curve on ENCAA dataset 0.890 0.339 0.044 0.001 (a) Importance scores estimated by our method. 0.082 0.034 0.021 0.005 (c) Pseudo-labels predicted by Mean Teacher. 1 0 0 0 (b) Pseudo-labels generated by our method. 0.278 0.250 0.035 0.098 (d) Pseudo-labels estimated by Label Propagation. ! = 0.98 ! = 0.96 ! = 0.13 ! = 0.34 Examples of unlabelled images and their effectiveness weights estimated automatically by our method. Examples of our method’s results and pseudo-labels estimated by different methods during training

Transcript of Learning to Detect Important People in Unlabelled Images for … · 2020. 8. 9. · ness weighting...

Page 1: Learning to Detect Important People in Unlabelled Images for … · 2020. 8. 9. · ness weighting strategy is able to detect and neglect noisy unlabelled images with no important

Methodology Experimental Results★ Datasets

* EMS Dataset: 10, 687 images from multiple scenes. * ENCAA Dataset: 28, 798 frames of basketball game video.

★ Comparisons with related methods on both datasets

★ Evaluations of different components and techniques used to estimate importancescore in proposed method

★ Visualize results

Introduction

Learning to Detect Important People in Unlabelled Imagesfor Semi-supervised Important People Detection

Fa−Ting Hong!, Wei−Hong Li# and Wei−Shi Zheng!!Sun Yat-sen University #University of Edinburgh

★ Introduction & Motivation* Existing methods of important people detection require massive quantities of labelled data and detecting important people in unlabeled images has not yet been developed.* The imbalance between the number of important people and non-important people in the picture will cause pseudo-labelling imbalance problem.* Not all unlabelled images contain important people; images without such people represent noisy unlabelled samples during learning.

★ Contributions* The proposed approach is the first to study on learning important people detection from partially labelled data.* we contribute two large datasets called Extended-MS (EMS) and Extended- NCAA (ENCAA) for evaluation of semi-supervised important people detection by augmenting existing datasets with a large number of unlabelled images collected from the internet * Extensive experiments verify the efficacy of our proposed method on important people detection of semi-supervised phase.

★ Detecting Noisy Unlabelled Images * Image-specific effectiveness weight : 𝜀

𝜀 = 1 −𝐻 𝓏*

𝐻 𝑀* Effectiveness weight reflect the confidence that an unlabelled image features important people:

𝐿𝒰 =𝜆𝒰 /𝜀0

𝒰

01!

/ 𝓌3ℓ𝒰(𝓎73𝒰, �̃�3)�

<=𝒰∈𝒮@

𝒰

★Overview of Proposed Method* To alleviate the pseudo-labelling imbalance problem, we introduce a ranking strategy for pseudo-label estimation, and also introduce two weighting strategies applied to unlabelleddata loss.* The final objective function can be expressed as:

𝐿 = 𝐿𝒯 + 𝜆𝐿𝒰

=1𝒯 𝐾/ / ℓ(𝓎73𝒯, 𝓎3𝒯)

<=𝒯∈𝒮@

𝒯

𝒯

01!

+𝜆𝒰 /𝜀0

𝒰

01!

/ 𝓌3ℓ𝒰(𝓎73𝒰, �̃�3)�

<=𝒰∈𝒮@

𝒰

𝑠. 𝑡. 𝓌3 ∈ 𝔀 = 𝜎 𝓏* , 𝜀0 = 1 − I 𝓏J

I K★ Pseudo-labelling by Ranking-based Sampling:

* Ranking-based sampling procedure:𝒮0𝒰, �̃� = 𝑅𝑎𝑛𝑘𝑆(𝑓R, 𝑥3𝒰 <=

𝒰∈T@𝒰, 𝛼, 𝐾)

* Replacing the unlabelled people and correspondingpseudo-labels with those sampled by RankS:

𝐿𝒰 =𝜆𝒰 / / ℓ𝒰(𝓎73𝒰, �̃�3)

<=𝒰∈𝒮@

𝒰

𝒰

01!

★ Balancing Loss via Importance Score Weighting:* Person-specific importance score weight: 𝓌V

𝐿𝒰 =𝜆𝒰 / / 𝓌3ℓ𝒰(𝓎73𝒰, �̃�3)

<=𝒰∈𝒮@

𝒰

𝒰

01!

, s. t./𝓌3

Y

31!

= 1,𝓌3 > 0.

labelled images

unlabelled images

!"

stop using noisy image

input labelled image

input unlabelled image

pseudo-labelling via Ranking-based Sampling

noisy image

labelled important person

non-important person from unlabelled image

classification space

adapting hyperplane using unlabelled data

ℓ"(, )ℓ"(, )

ℓ"(, )

ℓ'(, )ℓ'(, )

feature encoder

featureencoder

Importance Score

Ranking & Labelling

Sampling

Pseudo-labelling via Ranking-based Sampling

prediction

prediction

Sampling individuals according to the Ranking-based Sampling Procedure

()

ℓ'(, )

(̃‘important'

‘'non-important

+"

,-⨂

/(())

1 − 3((4)3(5)

,6

,7⨂

8

Unlabelled Image

Labelled Image

9 = 9" + <9'

importancefeature

importancefeature

relation graph

relation graph

+="

+='

>"

>'

unlabelledindividuals

unlabelledindividuals

(a) Score of “important” class in an unlabelled image

0.5

0

1

Scor

e of

impo

rtan

t” c

lass

(b) Pseudo-labelling in Current Semi-supervised Learning approach

0.5

0

1

Scor

e of

impo

rtan

t” c

lass

(c) Our Ranking-based Sampling0

1

Ran

king

Sco

re

(d) Labelled data in classification space

(e) Adding unlabelled data via current semi-supervised methods

( f ) Adding unlabelled data via our Ranking-based Sampling

softmaxclassifier threshold

Adapting hyperplane

POINT. We adopt the POINT [17] method, a state-of-the-art method of important people detection, as the baseline,which we train only on labelled data using a fully super-vised learning approach.Pseudo Label is a simple yet efficient semi-supervisedlearning approach for ordinary classification tasks, whichchooses the class with the maximum predicted probabilityas the true label for each unlabelled sample.Mean Teacher maintains two models: student and teacher.Given unlabelled samples, the outputs of the teacher modelare used as pseudo-labels. The consistency loss is deter-mined over the predictions of unlabelled images predictedby student model and the pseudo-labels generated by theteacher model such that the learned model can be invariantto stochastic noise between student and teacher models.Label Propagation infers the pseudo-labels of unlabelledsamples from the nearest neighbour graph, which is con-structed based on the embeddings of both labelled and un-labelled samples.

4.3. Implementation Details

We implement all methods in PyTorch. For a fair com-parison, we adopt POINT (we have detailed it in Sup-plementary Material) as the essential network with SGDused as the optimizer in our method as well as other semi-supervised baselines (i.e., PL, MT and LP). We run allmethods for 200 epochs and use the same hyper-parametersfor all methods. The hyper-parameter ↵ is learned on thevalidation data and is set to 0.99 for all the experiments.The weight decay is 0.0005 and the momentum is 0.9 inall experiments. The learning rate is initialized to 0.001,and we follow the learning rate update strategy of [17],i.e., the learning rate is scaled by a factor of 0.5 every 20epochs. We adopt the commonly used linear schedule toupdate weight �, i.e., we increase � linearly from 0 to itsmaximum (i.e., 1) over 35 epochs. We follow the standardevaluation metric in [17], i.e., the mean average precision isreported to measure the performance of all methods.

4.4. Comparisons with Related Methods

We first compare our method with current semi-supervised learning methods adapted for important peopledetection and the fully supervised learning baseline. FromTable 1, it is worth noting that the recent semi-supervisedlearning approaches attain comparable results (e.g., the re-sults of LP vs. those of POINT are 88.61 % vs. 88.21 %on the ENCAA dataset if 66 % of labelled images are used)but sometimes underperform the fully supervised baseline(e.g., the results of LP vs. those of POINT are 86.66 % vs.88.48 % on the EMS dataset if all labelled images are used).In contrast, our method achieves a significant and consis-tent improvement over the baseline; e.g., After adding unla-balled images, our method outperforms the fully supervisedbaseline by 4.45 % and 4.15 % on the EMS and ENCAA

Dataset EMS ENCAA#labelled images 33 % 66 % 100 % 33 % 66 % 100 %

POINT (fully supervised) 83.36 85.97 88.48 84.60 88.21 89.75

Pseudo Label (PL) 83.37 85.35 88.57 85.70 88.43 90.56Label Propagation (LP) 82.34 86.33 86.66 85.36 88.61 90.18Mean Teacher (MT) 84.50 86.29 87.55 83.33 84.66 87.55

Ours 87.81 88.44 89.79 88.75 90.86 92.03

Table 1. Comparison with related methods on both datasets.

datasets, respectively, in the regime with fewer labels (33%). These results of PL, LP and MT clearly demonstratethat treating each person independently are unable to lever-age valuable information from unlabelled images to helptraining. On the contrary, the results of our method indi-cate that three proposed strategies enable our method to ef-fectively leverage the information of unlabelled images toassist in training on a limited quantity of labelled data andsignificantly boost performance.

4.5. Effect of the Proportion of Labelled Images

To further understand the factors that affect the perfor-mance of semi-supervised important people detection, weevaluate our method using different portions of labelled im-ages. We randomly select 33 %, 66 % and 100 % of la-belled images, and the remaining labelled images togetherwith unlabelled images are used WITHOUT labels. Wereport the results in Table 1, Table 2 and Table 3. It isclearly observed that using more labelled data can boost theoverall performance of important people detection, whichalso enables the semi-supervised model to estimate moreaccurate pseudo-labels for unlabelled images and furtherboost performance. It also indicates that developing a semi-supervised model that can correctly predict pseudo-labelsand combine them with the labelled training set is neces-sary. From another point of view, the results shown in Table2 imply that our method can consistently outperform thefully supervised approach as well as related baselines andclearly demonstrate the consistent efficacy of the three pro-posed strategies.

4.6. Ablation Study

We conduct ablation study to investigate the effectof three proposed strategies (i.e., ranking-based sampling(RankS), importance score weighting (ISW) and effective-ness weighting (EW)) on important people detection andshown the results in Table 2, where “Oursw/o ISW and EW” in-dicates our method using RankS only.

In Table 2, it is evident that all strategies can improveperformance in most label regimes, and the ranking-basedsampling strategy attains the greatest improvement; for in-stance, on the ENCAA dataset, if 33 % of labelled imagesare used, the method “Oursw/o ISW and EW” outperforms the“Oursw/o RankS, ISW and EW” by 2.78 %. This result clearlyshows that the ranking-based sampling enables that the rel-atively high score should be labelled as “important” and

Dataset EMS ENCAA#labelled images 33 % 66 % 100 % 33 % 66 % 100 %

Oursw/o Ranks, ISW and EW 83.70 86.81 87.67 84.35 87.66 89.93Oursw/o ISW and EW 85.55 87.25 88.53 87.13 90.53 91.49Oursw/o EW 86.34 87.45 89.67 87.68 90.60 92.00Ours 87.81 88.44 89.79 88.75 90.86 92.03

Table 2. Ablation study on both datasets. RankS repre-sents ranking-based sampling while ISW and EW indicate impor-tance score weighting and effectiveness weighting, respectively.Oursw/o ISW and EW means our model without using ISW and EW.

Dataset EMS ENCAA#labelled images 33 % 66 % 100 % 33 % 66 % 100 %

OursLP 87.51 88.10 89.65 88.95 91.06 91.98OursMT 87.23 88.56 90.72 88.97 90.93 91.62Ours 87.81 88.44 89.79 88.75 90.86 92.03

Table 3. Evaluation of different techniques (i.e., LP and MT)when used for instantiating pseudo-label estimation function (i.e.,g(·)) instead of using Softmax function.

the rest remain “non-important” when predicting pseudo-labels within each unlabelled image, preventing assigningall “non-important” or all “important” pseudo-labels duringlabel guessing in an image. This is also verified by Figure2, where our method correctly predicts pseudo-labels for allindividuals (Figure 2(b)) during training and estimate accu-rate importance scores at the end (e.g., Figure 2(a)) whilecurrent semi-supervised learning approaches (i.e., LP andMT) assign all individuals as “non-important” samples

From Table 2, we also observe that adding importancescore weighting (ISW) can consistently albeit slightly boostthe performance (e.g., the results of “Oursw/o EW” vs. thoseof “Oursw/o ISW and EW” are 89.67 % vs. 88.53 % on the EMSif all labelled images are used). This indicates that ISWis able to alleviate the problem of data imbalance and ulti-mately benefits the training of important people detection.

In addition, comparing the full model and our model us-ing both RankS and ISW, we clearly observe that the esti-mated effectiveness weight (EW, defined in Eq. 6) improvesthe performance (e.g., “Ours” improves the performance of“Oursw/o EW” from 86.34 % to 87.81 % on EMS if 33 % oflabelled images are used). This implies that our effective-ness weighting strategy is able to detect and neglect noisyunlabelled images with no important people, and this ben-efits important people detection. To further better under-stand how the effectiveness weight works, we visualize EWof several unlabelled images and present them in Figure 6.We clearly observe that if there are no important people inthe unlabelled image, EW is small (as shown in the secondrow in Figure 6), while if important people are present, EWis nearly 1 (as shown in the first row in Figure 6). This re-sult again clearly demonstrates the efficacy of our proposedEW on detecting noisy images and neglecting noisy sam-ples during training.

Additionally, we also evaluate the effect of differenttechniques (i.e., LP and MT) used to estimate importance

! = 0.98 ! = 0.96

! = 0.13 ! = 0.34Figure 6. Examples of unlabelled images and their effectivenessweights estimated automatically by our method.

score in our method during pseudo-labelling in Table 3,where “OursLP” implys our method using Label Propa-gation technique for importance score estimation duringpseudo-labelling. It is clearly shown in Table 3 that the vari-ants of method using different techniques for importancescore estimation yield similar results, which demonstratesthe stableness of our methods.

5. Conclusion

In this work, we study semi-supervised learning in thecontext of important people detection and propose a semi-supervised learning method for this task. Compared withrecent semi-supervised learning approaches, our method isshown to be able to effectively leverage the informationof unlabelled images to assist in model training. We alsoconduct extensive experiments on important people detec-tion by a semi-supervised learning method, and the re-sults confirm that 1) the pseudo-labels of individuals in agiven unlabelled image should have the special pattern inimportant people detection (i.e., the relatively high scoreshould be labelled as “important” and the rest remains “non-important”), and our proposed ranking-based sampling isable to achieve this; 2) our importance score weighting canalleviate the imbalance problem and boost performance;and 3) enabling the model to neglect the noisy unlabelledimages with no important people is important during semi-supervised learning. By our learning, we are able to avoidcostly labelling on important people detection and achievesatisfactory performance.

6. Acknowledgement

This work was supported partially by the Na-tional Key Research and Development Program ofChina (2018YFB1004903), NSFC(U1911401,U1811461),Guangdong Province Science and Technology Innova-tion Leading Talents (2016TX03X157), Guangdong NSFProject (No. 2018B030312002), Guangzhou ResearchProject (201902010037), and Research Projects of ZhejiangLab (No. 2019KD0AB03).

Dataset EMS ENCAA#labelled images 33 % 66 % 100 % 33 % 66 % 100 %

Oursw/o Ranks, ISW and EW 83.70 86.81 87.67 84.35 87.66 89.93Oursw/o ISW and EW 85.55 87.25 88.53 87.13 90.53 91.49Oursw/o EW 86.34 87.45 89.67 87.68 90.60 92.00Ours 87.81 88.44 89.79 88.75 90.86 92.03

Table 2. Ablation study on both datasets. RankS repre-sents ranking-based sampling while ISW and EW indicate impor-tance score weighting and effectiveness weighting, respectively.Oursw/o ISW and EW means our model without using ISW and EW.

Dataset EMS ENCAA#labelled images 33 % 66 % 100 % 33 % 66 % 100 %

OursLP 87.51 88.10 89.65 88.95 91.06 91.98OursMT 87.23 88.56 90.72 88.97 90.93 91.62Ours 87.81 88.44 89.79 88.75 90.86 92.03

Table 3. Evaluation of different techniques (i.e., LP and MT)when used for instantiating pseudo-label estimation function (i.e.,g(·)) instead of using Softmax function.

the rest remain “non-important” when predicting pseudo-labels within each unlabelled image, preventing assigningall “non-important” or all “important” pseudo-labels duringlabel guessing in an image. This is also verified by Figure2, where our method correctly predicts pseudo-labels for allindividuals (Figure 2(b)) during training and estimate accu-rate importance scores at the end (e.g., Figure 2(a)) whilecurrent semi-supervised learning approaches (i.e., LP andMT) assign all individuals as “non-important” samples

From Table 2, we also observe that adding importancescore weighting (ISW) can consistently albeit slightly boostthe performance (e.g., the results of “Oursw/o EW” vs. thoseof “Oursw/o ISW and EW” are 89.67 % vs. 88.53 % on the EMSif all labelled images are used). This indicates that ISWis able to alleviate the problem of data imbalance and ulti-mately benefits the training of important people detection.

In addition, comparing the full model and our model us-ing both RankS and ISW, we clearly observe that the esti-mated effectiveness weight (EW, defined in Eq. 6) improvesthe performance (e.g., “Ours” improves the performance of“Oursw/o EW” from 86.34 % to 87.81 % on EMS if 33 % oflabelled images are used). This implies that our effective-ness weighting strategy is able to detect and neglect noisyunlabelled images with no important people, and this ben-efits important people detection. To further better under-stand how the effectiveness weight works, we visualize EWof several unlabelled images and present them in Figure 6.We clearly observe that if there are no important people inthe unlabelled image, EW is small (as shown in the secondrow in Figure 6), while if important people are present, EWis nearly 1 (as shown in the first row in Figure 6). This re-sult again clearly demonstrates the efficacy of our proposedEW on detecting noisy images and neglecting noisy sam-ples during training.

Additionally, we also evaluate the effect of differenttechniques (i.e., LP and MT) used to estimate importance

! = 0.98 ! = 0.96

! = 0.13 ! = 0.34Figure 6. Examples of unlabelled images and their effectivenessweights estimated automatically by our method.

score in our method during pseudo-labelling in Table 3,where “OursLP” implys our method using Label Propa-gation technique for importance score estimation duringpseudo-labelling. It is clearly shown in Table 3 that the vari-ants of method using different techniques for importancescore estimation yield similar results, which demonstratesthe stableness of our methods.

5. Conclusion

In this work, we study semi-supervised learning in thecontext of important people detection and propose a semi-supervised learning method for this task. Compared withrecent semi-supervised learning approaches, our method isshown to be able to effectively leverage the informationof unlabelled images to assist in model training. We alsoconduct extensive experiments on important people detec-tion by a semi-supervised learning method, and the re-sults confirm that 1) the pseudo-labels of individuals in agiven unlabelled image should have the special pattern inimportant people detection (i.e., the relatively high scoreshould be labelled as “important” and the rest remains “non-important”), and our proposed ranking-based sampling isable to achieve this; 2) our importance score weighting canalleviate the imbalance problem and boost performance;and 3) enabling the model to neglect the noisy unlabelledimages with no important people is important during semi-supervised learning. By our learning, we are able to avoidcostly labelling on important people detection and achievesatisfactory performance.

6. Acknowledgement

This work was supported partially by the Na-tional Key Research and Development Program ofChina (2018YFB1004903), NSFC(U1911401,U1811461),Guangdong Province Science and Technology Innova-tion Leading Talents (2016TX03X157), Guangdong NSFProject (No. 2018B030312002), Guangzhou ResearchProject (201902010037), and Research Projects of ZhejiangLab (No. 2019KD0AB03).

(") ($)

Fgiure (a) is the distribution of top 8 importance score in testing set in EMS datasets and Figure (b) is the statistics of unlabelled data’s pseudo-labels on EMS dataset. Better view in color.

Match

ingRa

te(%)

Rank

Match

ingRa

te(%)

Rank(a) CMC curve on EMS dataset (b) CMC curve on ENCAA dataset

0.890

0.339

0.044

0.001

(a) Importance scores estimated by our method.

0.082

0.034

0.021

0.005

(c) Pseudo-labels predicted by Mean Teacher.

1 00 0

(b) Pseudo-labels generated by our method.

0.278

0.250

0.035

0.098

(d) Pseudo-labels estimated by Label Propagation.

! = 0.98 ! = 0.96

! = 0.13 ! = 0.34Examples of unlabelled images and their effectiveness weights estimated automatically by our method.

Examples of our method’s results and pseudo-labels estimated by different methods during training