1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost...

14
1 Privacy-Preserving Deep Visual Recognition: An Adversarial Learning Framework and A New Dataset Haotao Wang*, Zhenyu Wu*, Zhangyang Wang, Zhaowen Wang, and Hailin Jin Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera applications, using deep learning. We formulate a unique adversarial training framework, that learns a degradation transform for the original video inputs, in order to explicitly optimize the trade-off between target task performance and the associated privacy budgets on the degraded video. We carefully analyze and benchmark three different optimization strategies to train the resulting model. Notably, the privacy budget, often defined and measured in task-driven contexts, cannot be reliably indicated using any single model performance, because a strong protection of privacy has to sustain against any possible model that tries to hack privacy information. In order to tackle this problem, we propose two strategies: model restarting and model ensemble, which can be easily plug-and-play into our training algorithms and further improve the performance. Extensive experiments have been carried out and analyzed. On the other hand, few public datasets are available with both utility and privacy labels provided, making the power of data-driven (supervised) learning not yet fully unleashed on this task. We first discuss an innovative heuristic of cross-dataset training and evaluation, that jointly utilizes two datasets with target task and privacy labels respectively, for adversarial training. To further alleviate this challenge, we have constructed a new dataset, termed PA-HMDB51, with both target task (action) and selected privacy attributes (gender, age, race, nudity, and relationship) labeled on a frame-wise basis. This first-of-its-kind video dataset further validates the effectiveness of our proposed framework, and opens up new opportunities for the research community. Our codes, models, and the PA-HMDB51 dataset will be made all available at: https://github.com/htwang14/PA-HMDB51. Index Terms—Visual privacy, action recognition, privacy-preserving learning, adversarial learning. 1 I NTRODUCTION S MART surveillance or smart home cameras, such as Amazon Echo and Nest Cam, are now found in millions of locations to remotely link users to their homes or offices, providing monitoring services to enhance security and/or notify environment changes, as well as lifelogging and intelligent services. The convenience and benefits, however, come at the heavy price of privacy intrusion from time to time. Due to their computationally demanding nature, not all visual recognition tasks can be run the resource- limited local device end, making transmitting (part of) data to the cloud indispensable. However, users have expressed growing concerns towards the abuse of their uploaded data, by the cloud service provider - a potentially malicious au- thorized party. That is different from the traditional privacy concerns, which mostly arise from the unsecured channel between cloud and device (e.g, malicious third-party eaves- dropping), therefore demanding new solutions. Is it at all possible to alleviate the privacy concerns, without compromising on user convenience? At the first glance, the question itself is posed as a dilemma: we would like a camera system to recognize important events and Haotao Wang, Zhenyu Wu, and Zhangyang Wang are with the Depart- ment of Computer Science and Engineering, Texas A&M University, College Station, TX, 77840. E-mail: {htwang,wuzhenyu sjtu,atlaswang}@tamu.edu The first two authors contributed equally to this work. Zhaowen Wang and Hailin Jin are with Adobe Research, San Jose, CA, 95110. E-mail: {zhawang,hljin}@adobe.com Correspondence to: Zhangyang Wang ([email protected]). assist human daily life by understanding its videos, while preventing it from obtaining sensitive visual information (such as faces, gender, skin color, etc.) that can intrude individual privacy. It thus becomes a new and appealing problem, to find an appropriate transform to “sanitize” the collected raw visual data at the local end, so that the transformed data itself will only enable certain target tasks while obstructing undesired privacy-related tasks. Recently, some new video acquisition approaches [1]–[3] proposed to intentionally capture or process videos in extremely low-resolution to create privacy-preserving ”anonymized videos”, and showed promising empirical results. This paper seeks to take a first step towards address- ing this brand new challenge of privacy-preserving visual recognition, via the following multi-fold contributions: A General Adversarial Training and Evaluation Frame- work. We formulate the privacy-preserving visual recog- nition in a unique adversarial training framework. The framework explicitly optimizes the trade-off between tar- get task performance and associated privacy budgets, by learning a transform from the original videos to anonymized videos. To mitigate the training instability, we then carefully design and experimentally compare three different optimization strategies. We eventually de- fine a novel two-fold protocol, to evaluate the trained models in terms of not only the utility, but also the privacy protection generalizability against unseen privacy hackers. Practical Approximations of “Universal” Privacy Pro- arXiv:1906.05675v2 [cs.CV] 29 Jul 2019

Transcript of 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost...

Page 1: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

1

Privacy-Preserving Deep Visual Recognition:An Adversarial Learning Framework

and A New DatasetHaotao Wang*, Zhenyu Wu*, Zhangyang Wang, Zhaowen Wang, and Hailin Jin

Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart cameraapplications, using deep learning. We formulate a unique adversarial training framework, that learns a degradation transform for theoriginal video inputs, in order to explicitly optimize the trade-off between target task performance and the associated privacy budgetson the degraded video. We carefully analyze and benchmark three different optimization strategies to train the resulting model. Notably,the privacy budget, often defined and measured in task-driven contexts, cannot be reliably indicated using any single modelperformance, because a strong protection of privacy has to sustain against any possible model that tries to hack privacy information. Inorder to tackle this problem, we propose two strategies: model restarting and model ensemble, which can be easily plug-and-play intoour training algorithms and further improve the performance. Extensive experiments have been carried out and analyzed.

On the other hand, few public datasets are available with both utility and privacy labels provided, making the power of data-driven(supervised) learning not yet fully unleashed on this task. We first discuss an innovative heuristic of cross-dataset training andevaluation, that jointly utilizes two datasets with target task and privacy labels respectively, for adversarial training. To further alleviatethis challenge, we have constructed a new dataset, termed PA-HMDB51, with both target task (action) and selected privacy attributes(gender, age, race, nudity, and relationship) labeled on a frame-wise basis. This first-of-its-kind video dataset further validates theeffectiveness of our proposed framework, and opens up new opportunities for the research community. Our codes, models, and thePA-HMDB51 dataset will be made all available at: https://github.com/htwang14/PA-HMDB51.

Index Terms—Visual privacy, action recognition, privacy-preserving learning, adversarial learning.

F

1 INTRODUCTION

SMART surveillance or smart home cameras, such asAmazon Echo and Nest Cam, are now found in millions

of locations to remotely link users to their homes or offices,providing monitoring services to enhance security and/ornotify environment changes, as well as lifelogging andintelligent services. The convenience and benefits, however,come at the heavy price of privacy intrusion from timeto time. Due to their computationally demanding nature,not all visual recognition tasks can be run the resource-limited local device end, making transmitting (part of) datato the cloud indispensable. However, users have expressedgrowing concerns towards the abuse of their uploaded data,by the cloud service provider - a potentially malicious au-thorized party. That is different from the traditional privacyconcerns, which mostly arise from the unsecured channelbetween cloud and device (e.g, malicious third-party eaves-dropping), therefore demanding new solutions.

Is it at all possible to alleviate the privacy concerns,without compromising on user convenience? At the firstglance, the question itself is posed as a dilemma: we wouldlike a camera system to recognize important events and

• Haotao Wang, Zhenyu Wu, and Zhangyang Wang are with the Depart-ment of Computer Science and Engineering, Texas A&M University,College Station, TX, 77840.E-mail: {htwang,wuzhenyu sjtu,atlaswang}@tamu.eduThe first two authors contributed equally to this work.

• Zhaowen Wang and Hailin Jin are with Adobe Research, San Jose, CA,95110. E-mail: {zhawang,hljin}@adobe.com

• Correspondence to: Zhangyang Wang ([email protected]).

assist human daily life by understanding its videos, whilepreventing it from obtaining sensitive visual information(such as faces, gender, skin color, etc.) that can intrudeindividual privacy. It thus becomes a new and appealingproblem, to find an appropriate transform to “sanitize”the collected raw visual data at the local end, so that thetransformed data itself will only enable certain target taskswhile obstructing undesired privacy-related tasks. Recently,some new video acquisition approaches [1]–[3] proposedto intentionally capture or process videos in extremelylow-resolution to create privacy-preserving ”anonymizedvideos”, and showed promising empirical results.

This paper seeks to take a first step towards address-ing this brand new challenge of privacy-preserving visualrecognition, via the following multi-fold contributions:

• A General Adversarial Training and Evaluation Frame-work. We formulate the privacy-preserving visual recog-nition in a unique adversarial training framework. Theframework explicitly optimizes the trade-off between tar-get task performance and associated privacy budgets,by learning a transform from the original videos toanonymized videos. To mitigate the training instability,we then carefully design and experimentally comparethree different optimization strategies. We eventually de-fine a novel two-fold protocol, to evaluate the trainedmodels in terms of not only the utility, but also the privacyprotection generalizability against unseen privacy hackers.

• Practical Approximations of “Universal” Privacy Pro-

arX

iv:1

906.

0567

5v2

[cs

.CV

] 2

9 Ju

l 201

9

Page 2: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

2

tection. The privacy budget in our framework cannot besimply defined w.r.t. one single privacy attribute predic-tion model, as the ideal protection of privacy has to beuniversal and model-agnostic, i.e., obstructing every pos-sible attacker model from predicting privacy information.To resolve the so-called “∀ challenge”, we propose twostrategies, i.e., restarting and ensembling, to enhance thegeneralization capability of the learned degradation todefend against unseen models.

• A New Dataset for this New Problem. There are fewoff-the-shelf datasets that have both utility and privacyattribute annotations. To alleviate the lack of appropriatetraining data, we (in our previous work [4]) introduce a“cross-dataset training” alternative, based on the hypoth-esis that privacy attributes have good “transferability”.In this paper, we construct the very first dataset namedPA-HMDB51 (Privacy Attribute HMDB51), for the taskof privacy-preserving action recognition from video. Thedataset consists of 580 videos originally from HMDB51.For each video, we have annotated both utility (action)and privacy (five attributes including skin color, face,gender, nudity and relationship) labels on a frame-wisebasis. We benchmark our proposed framework on the newPA-HMDB51 and validate its effectiveness.

The paper is a significant extension from the previous con-ference version [4]. We have included: (1) a detailed discus-sion and comparison on three optimization strategies for theproposed framework (only one heuristic was presented in[4]); (2) a much expanded experimental and analysis section;and most importantly (3) the construction of the new PA-HMDB51 dataset, and the associated benchmarking efforts.

2 RELATED WORK

2.1 Privacy Protection in Computer VisionWith pervasive camera for surveillance or smart home de-vices, privacy-preserving visual recognition has draw in-creasing interests from both industry and academia. Mostclassical cryptographic solutions secure the communicationagainst unauthorized access from attackers. However, theyare not immediately applicable to preventing authorizedagents (such as the backend analytics) from the unau-thorized abuse of information, that causes privacy breachconcerns. A few encryption based solutions, such as Ho-momorphic Encryption (HE) [5], [6], were developed tolocally encrypt visual information. The server can only getaccess to the enciphered data and conduct utility task onit. However, many encryption based solutions will incurhigh computational costs at the local platforms. It is alsochallenging to generalize the cryptosystems to more com-plicated classifiers. [7] combined the detection of regionsof interest and the real encryption techniques to improveprivacy while allowing general surveillance to continue. Aseemingly reasonable, and computationally cheaper optionis to extract feature descriptors from raw images, and trans-mit those features only. Unfortunately, a previous study [8]revealed that considerable details of original images couldstill be recovered from standard HOG or SIFT features (eventhey look visually distinct from natural images).

An alternative toward a privacy-preserving vision sys-tem concerns the concept of anonymized videos. Such

videos are intentionally captured or processed to be inspecial low quality conditions, that only allow for the recog-nition of some target events or activities, while avoiding theunwanted leak of the identity information for the humansubjects in the video [1]–[3]. [1] showed that even at theextreme low resolutions, reliable action recognition couldbe achieved by learning appropriate downsampling trans-forms, with neither unrealistic activity-location assump-tions nor extra specific hardware resources. The authorsempirically verified that conventional face recognition eas-ily failed on the generated low-resolution videos. [2] usesimage operations like blurring and superpixel clusteringto get anonymized videos while [3] uses low resolutioncamera hardware to get extreme low resolution (e.g., 16∗12)videos as anonymized videos. [9] uses cartoon-like effectswith a customized version of mean shift filtering. [10], [11]proposed to use privacy preserving optics to filter sensi-tive information from the incident light-field before sensormeasurements are made, by k-anonymity and defocus blur.Earlier work [12] explored privacy-preserving tracking andcoarse pose estimation using a network of ceiling-mountedtime-of-flight low-resolution sensors. [13] adopted a net-work of ceiling-mounted binary passive infrared sensors.However, both works [12], [13] handled only a limited setof activities performed at specific constrained areas in theroom. The usage of low-resolution anonymized videos [1],[3] is computationally cheaper, and is also compatible withsensor and bandwidth constraints. However, [1], [3] remainempirical in protecting privacy. In particular, neither weretheir models learned towards protecting any visual privacy,nor were the privacy-preserving effects carefully analyzedand evaluated. In other words, privacy protection in [1],[3] came as a ”side product” of down-sampling, and wasnot a result of any optimization. The authors of [1], [3] alsodid not extend their efforts to studying deep learning-basedrecognition, making their task performance less competitive.The recent progress of low-resolution object recognition [14],[15] also put their privacy protection effects in jeopardy.

2.2 Privacy Protection in Social Media/Photo Sharing

User privacy protection is also a topic of extensive interestsin the social media field, especially for photo sharing. Themost common means to protect user privacy in a uploadedphoto is to add empirical obfuscations, such as blurring, mo-saicing or cropping out certain regions (usually faces) [16].However, extensive research showed that such an empiricalmeans can be easily hacked too [17], [18]. A latest work[19] described a game-theoretical system in which the photoowner and the recognition model strive for antagonisticgoals of dis-enabling recognition, and better obfuscationways could be learned from their competition. However,it was only designed to confuse one specific recognitionmodel, via finding its adversarial perturbations. That cancaused obvious overfitting as simply changing to anotherrecognition model will likely put the learning efforts in vain:such perturbations even cannot protect privacy from humaneyes. Their problem setting thus deviated far away fromour target problem. Another notable difference is that insocial photo sharing, we usually hope to cause minimumperceptual quality loss to those photos, after applying any

Page 3: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

3

privacy-preserving transform to them. The same concerndoes not exist in our scenario, allowing us to explore muchmore free, even aggressive image distortions.

A useful resource to us was found in [20], which definedconcrete privacy attributes and correlated them to imagecontent. The authors categorized possible private informa-tion in images, and then run a user study to understand theprivacy preferences. They then provided a sizable set of 22kimages annotated with 68 privacy attributes, on which theytrained privacy attribute predictors.

3 METHOD

3.1 Problem definition

Assume our training data X (raw visual data captured bycamera) are associated with a target task T and a privacybudget B. Since T is usually a supervised task, e.g. actionrecognition or visual tracking, a label set YT is provided onX , and a standard cost function LT (e.g. cross-entropy) isdefined to evaluate the task performance on T . And usuallythere is a state-of-the-art deep neural network fT whichtakes X as input and predicts the target labels. On theother hand, we need to define a budget cost function JBto evaluate the privacy leakage of its input data: the smallerJB(·) is, the less privacy information its input contains.

We seek an optimal degradation function f∗D to trans-form the original X to f∗D(X) as the common input for bothfT and fB , and an optimal target model f∗T such that:• f∗D has filtered out the privacy information contained

in X , i.e.

JB(f∗D(X))� JB(X)

• the performance of fT is minimally affected whenusing the degraded visual data f∗D(X) compared towhen using the original data X , i.e.

LT (f∗T (f

∗D(X)), YT ) ' min

fTLT (fT (X), YT )

To achieve these two goals, we mathematically formulatethe problem as solving the following optimization problem:

f∗T , f∗D = argmin

fT ,fD

[LT (fT (fD(X)), YT )+

γJB(fD(X))](1)

The definition of the privacy budget cost JB is notstraightforward. Practically, it needs to be placed in con-crete application contexts, often in a task-driven way. Forexample, in smart workplaces or smart homes with videosurveillance, one might often want to avoid a disclosure ofthe face or identity of persons. Therefore, to reduce JB couldbe interpreted as to suppress the success rate of identityrecognition or verification. Other privacy-related attributes,such as race, gender, or age, can be similarly defined too.We denote the privacy-related annotations (such as identitylabel) as YB , and rewrite JB(·) as JB(fB(·), YB), where fBdenotes the budget model who takes (degraded or original)visual data as input and predicts the corresponding privacyinformation. Different from LT , minimizing JB will encour-age fB(·) to diverge from YB .

Such a supervised, task-driven definition of JB posesat least two challenges: (1) Dataset challenge: The privacybudget-related annotations, denoted as YB , often have lessavailability than target task labels. Specifically, it is oftenchallenging to have both YT and YB ready on the same X ;(2) ∀ challenge: Considering the nature of privacy protec-tion, it is not sufficient to merely suppress the success rate ofone fB model. Instead, define a privacy prediction functionfamily P : fD(X) 7→ YB , the ideal privacy protection offD should be reflected as suppressing every possible modelfD from P . That diverts from the common supervisedtraining goal, where only one model needs to be found tosuccessfully fulfill the target task.

We address the Dataset challenge by two ways: (1)cross dataset training and validation: see section 3.6; andmore importantly (2) building a new dataset containingboth utility and privacy labels: see section 5. We defer theirdiscussion to respective experimental paragraphs.

Handling the ∀ challenge is more challenging. First were-write the general form 1 with the task-driven definitionof LB as follows:

f∗T , f∗D = argmin

fT ,fD

[LT (fT (fD(X)), YT )+

γ supfB∈P

JB(fB(fD(X)), YB)](2)

The ∀ challenge in essence is the infeasibility to directlysolve 2, due to the existence of the superior operation. Sowe propose to solve the following problem instead:

θ∗T , θ∗D = argmin

θT ,θD

[LT (fT (fD(X)), YT )+

γmaxθB

JB(fB(fD(X)), YB)](3)

where fB has a fixed form and is parameterized by θB .Similarly, fD and fT are parameterized by θD and θTcorrespondingly. Note that problem 3 is a pretty wild ap-proximation to problem 2. But experimental results demon-strate that optimizing equation 3 can already achieve resultsmuch better than the baseline methods. We further propose”model ensemble” and ”model restarting” (see section 3.4)to handle the ∀ challenge better and further boost theexperimental results.

Without loss of generality, we assume both fT and fB tobe classification models and output class labels. To optimizethe target task performance, LT could be simply chosen asthe KL divergence: KL(fT (fd(X))||YT ). Definition of JB ,however, is not as trivial. We discuss different forms ofJB , and correspondingly different optimization strategies,in sections 3.3.1-3.3.3.

3.2 Basic framework

Figure 1 depicts the basic framework implementing theproposed formulation 2. The framework consists three parts:the degradation model fD, the target model fT and thebudget model fB . fD takes raw video X as input, filtersout privacy information contained in X and outputs theanonymized video fD(X). fT takes fD(X) as input andcarries out the target task. fB also take fD(X) as inputand try to predict the privacy information from fD(X). Allthree models are implemented with deep neural networks

Page 4: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

4

min(𝐿𝑇−𝛾𝐿𝐵)

𝐿𝑇 𝐿𝐵

𝑓𝐷(𝑋)TargetTask

Model 𝑓𝑇

PrivacyPredictionModel 𝑓𝐵

ActiveDegradation

Model 𝑓𝐷

X

Anonymized Video

𝑓𝑇 𝑓𝐷 𝑋 𝑓𝐵 𝑓𝐷 𝑋

𝑌𝑇 𝑌𝐵

Raw Video

Fig. 1. Basic adversarial training framework for privacy-preserving visualrecognition.

and their parameters are learnable during the training pro-cedure. The entire pipeline is trained under the guidance ofthe hybrid loss of LT and JB . The goal of the training proce-dure is to find a degradation model which can filter out theprivacy information contained in the original video whilekeeping useful information for the utility task, and to find atarget model which can achieve good performance on targettask using degraded videos fD(X). Similar frameworkshave been used in feature disentanglement [21] [22] [23][24]. After training, the learned degradation model can beapplied on local device (e.g. smart camera). We can convertraw video to degraded video locally and only transfer thedegraded video through Internet to the backend (e.g. cloud)for target task analysis, so that the privacy informationcontained in the raw videos will be invisible on the backend.

Specifically, fD is implemented using the model in [25],which can be taken as a 2D convolution based frame-levelfilter. In other words, fD converts each frame in X into afeature map of the same shape as the original frame. Weuse state-of-the-art human action recognition model C3D[26] as fT and state-of-the-art image classification models,such as ResNet [27] and MobileNet [28], as fB . Since theaction recognition model we use is C3D, we need to splitthe videos into clips with fixed frame number. Each clip is4D tensor of shape [T,W,H,C], where T is the number offrames in each clip and W , H , C are the width, height andcolor channel number of each frame correspondingly. UnlikefT who takes a 4D tensor as an input data sample, fB takesa 3D tensor (i.e. a frame) as input. We average the logits overthe temporal dimension of each video clip to calculate JBand predict the budget task label.

3.3 Optimization StrategiesSimilar to GANs [29], our model is prone to collapse andget stuck at bad local minimums during training. We thuscarefully designed three different optimization schemes tosolve this hard optimization problem.

3.3.1 Gradient reverse layer (GRL)We can consider problem 3 as a saddle point problem:

θ∗D, θ∗T = argmin

θD,θT

L(θD, θT , θ∗B)

θ∗B = argmaxθB

L(θ∗D, θ∗T , θB)

where L(θD, θT , θB) = LT − γLB and LB = −JB =KL(fB(fD(X))||YB) is KL divergence loss function.

GRL [30] is a state-of-the-art algorithm to solve such asaddle point problem. The underlying mathematical gist issimply solving the problem by following these update rules,

θD ← θD − αD ∗ ∇θD (LT − γLB) (4a)θT ← θT − αT ∗ ∇θTLT (4b)θB ← θB − αB ∗ ∇θBLB (4c)

We denote this method as Ours-KL in the followingparts and the formal description is in algorithm 1

Algorithm 1: Ours-KL algorithmInput: Target labels YT , Budget labels YB , visual data

X , step size αT , αB , αD, accuracy threshholdth1, th2, maximum iteration max iter

Output: Degradation model parameter θD , targetmodel parameter θT and budget modelparameter θB

1 initialize θD , θT and θB ;2 for i← 1 to max iter do3 Update θD using equation 4a4 while target task validation accuracy ≤ th1 do5 Update θT using equation 4b6 end7 while budget task training accuracy ≤ th2 do8 Update θB using equation 4c9 end

10 end

3.3.2 Maximize entropyUpdating θD to maximize cross-entropy loss LB mightnot be the best choice in our setting. We don’t need thebudget model to misclassify data to a false class with highconfidence. For example, if a sample has ground truth label[1, 0, 0, 0], one of the global optimum for maximizing LB isfB(fD(X)) = [0, 1, 0, 0]. We don’t need fD to go that farto this point so that fB will misclassify this training samplewith high confidence. Instead, a more reasonable output forfB(fD(X)) is somewhere near [0.25, 0.25, 0.25, 0.25], whichmeans fD has filtered most information that is necessary forthe budget task. Based on these intuitions, we formulate anew optimization scheme:

θD ← θD − αD ∗ ∇θD (LT − γHB)

θT ← θT − αT ∗ ∇θTLTθB ← θB − αB ∗ ∇θBLB

where LT and LB are still KL divergence loss functions asin equation 4, and HB = H(fB(fD(X))) is the entropy offB(fD(X)).

In practice, we update θT and θD in an end-to-end waywhen minimizingLT . That’s to say, the optimization schemeis actually as follows:

θD ← θD − αD ∗ ∇θD (LT − γHB) (6a)θT , θD ← θT , θD − αT ∗ ∇θT ,θDLT (6b)

θB ← θB − αB ∗ ∇θBLB (6c)

Page 5: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

5

We denote this method as Ours-Entropy in the followingparts and the formal description is in algorithm 2.

Algorithm 2: Ours-Entropy algorithmInput: Target labels YT , Budget labels YB , visual data

X , step size αT , αB , αD, accuracy threshholdth1, th2, maximum iteration max iter

Output: Degradation model parameter θD , targetmodel parameter θT and budget modelparameter θB

1 initialize θD , θT and θB ;2 for i← 1 to max iter do3 Update θD using equation 6a4 while target task validation accuracy ≤ th1 do5 Update θT , θD using equation 6b6 end7 while budget task training accuracy ≤ th2 do8 Update θB using equation 6c9 end

10 end

3.3.3 Alternative optimization of two loss functionsThe goal in equation 3 can also be formulated as alterna-tively solving the following two optimization problems:

minθD,θT

LT (fT (fD(X)), YT )

minθB

maxθD

LB(fB(fD(X)), YB)

With a little abuse of notations, we rewrite these two lossfunctions in terms of neural network parameters:

minθD,θT

LT (θD, θT ) (8a)

minθB

maxθD

LB(θD, θB) (8b)

Equation 8a is an ordinary minimization problem which canbe solved by training fD and fT in an end-to-end fashion.Equation 8b is a minimax problem which we solve bythe recent state-of-the-art minimax algorithm K-beam [31].K-beam keeps tracking K different sets of budget modelparameters, denoted as {θB , i}Ki=1, during training time andalternatively updates θD and {θB,i}Ki=1. More specifically,each iteration of the training procedure can be divided intotwo phases: Min step and Max step. Suppose at the t-thiteration, the parameter of degradation and budget modelsare θtD and {θtB,i}Ki=1 correspondingly. During Max step,it first select θB,j where j = argmini∈1,...,M LB(θ

tD, θ

tB,i)

and update θD using gradient descend on LB(θD, θtB,i) to

get θt+1D . During Min step, it updates all K budget model

parameters separately by gradient descend on LD(θt+1D , θB)

to get {θt+1B,i }Ki=1. We suggest the readers to refer to the

original paper for more details.Based on K-beam algorithm, we design algorithm 3 to

alternatively solve the two loss functions in equation 8: Wedenote this method as Ours-K-beam in the following parts.

3.4 Addressing the ∀ ChallengeTo improve the generalization ability of learned fD over allpossible fB ∈ P (i.e, privacy cannot be reliably predicted

Algorithm 3: Ours-K-beam algorithmInput: Target labels YT , Budget labels YB , visual data

X , step size αT , αB , αD, budget model beamnumber K , accuracy thresholds th1, th2,number of iteration max iter, d iter

Output: Degradation model parameter θD , targetmodel parameter θT and budget modelparameter θB

1 initialize θD , θT and {θB,i}Ki=1;2 for t← 1 to max iter do3 LB Max step:4 j = argmini∈1,...,M LB(θD, θB,i)5 for i← 1 to d iter do6 θD ← θD + αD ∗ ∇θDLB(θD, θB,j)7 end8 LB Min step:9 for k ← 1 to K do

10 while budget task training accuracy ≤ th2 do11 θB,i ← θB,i − αB ∗ ∇θBLB(θD, θB,i)12 end13 end14 LT step:15 while target task validation accuracy ≤ th1 do16 θT , θD ← θT , θD − αT ∗ ∇θT ,θDLT (θD, θT )17 end18 end

by any model), we hereby discuss two simple and easy-to-implement options. Other more sophisticated model re-sampling or model search approaches, such as [32], will beexplored in future work.

3.4.1 Budget Model Restarting

At certain point of training (e.g., when the privacy budgetLB(fB(fD(X))) stops decreasing any further), we replacethe current weights in fB with random weights. Such a ran-dom re-starting aims to avoid trivial overfitting between fBand fD (i.e., fD is only specialized at confusing the currentfB), without incurring more parameters. We then start totrain the new model fB to be a strong competitor, w.r.t.the current fD(X): specifically, we freeze the training of fDand fT , and change to minimizing KL(fB(fD(X))||YB),until the new fB has been trained from scratch to becomea strong privacy prediction model over current fD(X). Wethen resume adversarial training by unfreezing fD and fT ,as well as replacing the loss for fB back to the negativeentropy. It can repeat several times.

3.4.2 Budget Model Ensemble

The other strategy proposes to approximate the continuousP with a discrete set of M sample functions. Assuming thebudget model ensemble {f ib}Mi=1, we turn to minimizing thefollowing discretized surrogate of 2:

θ∗T , θ∗D = argmin

θT ,θD

[LT (fT (fD(X)), YT )+

γ maxi=1,...,M

maxθiB

JB(fiB(fD(X)), YB)]

(9)

Page 6: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

6

The previous basic framework is a special case of equation 9with M = 1. The ensemble strategy can be easily combinedwith re-starting.

3.4.3 Combine Budget Model Restarting and BudgetModel Ensemble with Ours-EntropyBudget Model Restarting and Budget Model Ensemble canbe easily combined with all three optimization schemesdescribed in section 3.3.1-3.3.3. We take Ours-Entropy as anexample here.

When model ensemble is used, we denote LB,i :=KL(fB,i(fD(X))||YB), HB,i := H(fB,i(fD(X))) and takeHB := mini∈{1,...,M}HB,i in equation 6a. That’s to saywe only suppress the model fB,i with the largest privacyleakage JB , e.g. the ”most confident” one about its currentprivacy prediction, when updating degradation model. Butwe still update all M budget models in equation 6c. So theparameter updating scheme is:

θD ← θD − αD ∗ ∇θD (LT − γ mini∈{1,...,M}

HB,i) (10a)

θT , θD ← θT , θD − αT ∗ ∇θT ,θDLT (10b)θB,i ← θB,i − αB ∗ ∇θB,i

LB,i, i = 1, . . . ,M (10c)

The formal description of Ours-Entropy algorithm withmodel restarting/ensemble is given in algorithm 4.

Algorithm 4: Ours-Entropy algorithm (with modelrestarting and model ensemble)

Input: Target labels YT , Budget labels YB , visual dataX , step size αT , αB , αD, model ensemblenumber M , accuracy threshold th1, th2,maximum iteration max iter, restartingiteration restarting iter.

Output: Degradation model parameter θD , targetmodel parameter θT and budget modelparameter θB

1 initialize θD , θT and θB ;2 for i← 1 to max iter do3 if i ≡ 0 (mod restarting iter) then4 Reinitialize {θB,i}Mi=1

5 end6 Update θD using equation 10a7 while target task validation accuracy ≤ th1 do8 Update θT , θD using equation 10b9 end

10 for j ← 1 to M do11 while budget task training accuracy ≤ th2 do12 Update θB,i using equation 10c13 end14 end15 end

3.5 Two-Fold Evaluation ProtocolAs a balance between two task models, the evaluation pro-tocol for privacy-preserving visual recognition is inevitablymore intricate than classical visual recognition tasks.

After we get fD , fT and fD by solving problem 3, weneed to evaluate the performance in two folds: (1) whether

the learned target task model maintains satisfactory perfor-mance on degraded videos; (2) whether the performance ofan arbitrary privacy prediction model on degraded videoswill deteriorate. Suppose we have training dataset Xt withtarget and budget task ground truth labels Y tT and Y tBand evaluation dataset Xe with target and budget taskground truth labels Y eT and Y eB . The first fold can follow thetraditional evaluation routine: compare fT (fD(X)) with Y eTto get the evaluation accuracy on target task, denoted asAT ,which we expect to be as high as possible.

For the second fold, it is apparently insufficient if weonly observe that the learned fD and fB lead to poorclassification accuracy onXe, because of the ∀ challenge: theattacker can select any budget model to try to steal privacyinformation from degraded videos fD(X). To empiricallyverify that fD prohibits reliable privacy prediction for otherpossible budget models, we propose a novel procedure: werandomly re-sample N models {f jB}Nj=1 from P . Then wetrain these N models on degraded training videos fD(Xt)to make correct predictions on privacy information, i.e.minfj

BLB(f

jB(fD(X

t)), Y tB) for j ∈ 1, . . . , N . (Note thatfD is fixed during this training procedure.) After that, weapply each f jB on degraded evaluation videos fD(Xe) andcompare the outputs f jB(fD(X

e)) with Y eT to get evaluationbudget accuracy of the j-th budget model. We select thehighest accuracy among all N budget models and use it asthe final budget accuracy ANB , which we expect to be as lowas possible.

3.6 Cross-Dataset Training and Evaluation: An initialstep towards alleviating the dataset challengeAn ideal dataset to train our framework would be, forexample, a set of human action videos with both action classand privacy attributes labeled. However, to the best of ourknowledge, no public dataset well satisfies this condition.We propose to use cross-dataset training and evaluation asa workaround method, details of which can be found in sec-tion 3.6. In brief, we train action recognition (target task) onhuman action datasets, such as UCF101 [33] and HMDB51[34], and train privacy protection (budget task) on visualprivacy dataset VISPR [20], while letting the two interact viatheir shared component - the learned degradation model.

More specifically, during training, we have twopipelines: one is fD + fT trained on UCF101 or HMDB51 foraction recognition; the other is fD + fB trained on VISPRto suppress multiple privacy attribute prediction. The twopipelines share the same parameters for fD. The initial-ization and alternating training strategy remain unchangedfrom SBU experiments, as shown in algorithm 2. Duringevaluation, we perform the first part of two-fold evaluation,i.e. action recognition, on UCF101 or HMDB51 testing set.We then evaluate privacy protection performance of Nbudget models using the VISPR testing sets. Such cross-dataset training and evaluation sheds new possibilities ontraining privacy-preserving recognition models, even underthe practical shortages of datasets that have been annotatedfor both tasks.

Beyond this initial step forward, we further constructa new datatset dedicated to the privacy-preserving visualrecognition task. It will be presented in section 5.

Page 7: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

7

4 SIMULATION EXPERIMENTS

We show the effectiveness of our framework on privacy-preserving action recognition on existing datasets. The targettask is human action recognition since it is a highly de-manded feature in smart home and smart workplace ap-plication. Experiments are carried out on three widely usedhuman action recognition datasets: SBU Kinect InteractionDataset [35], UCF101 [33] and HMDB51 [34]. The budgettask varies in different settings. In the experiments on SBUdataset, the budget task is to prevent the videos leakinghuman identity information. In the experiments on UCF101and HMDB51, the budget task is to protect visual privacyattributes as defined in [20]. We emphasize that generalframework proposed in section 3.2 can be used for a largevariety of target task and budget task combinations, not onlylimited to these two settings mentioned above.

4.1 Identity-Preserving Action Recognition on SBU:Single-Dataset Training

We compare our framework with three baseline methodsproposed in [1] and [16] to show our methods’ signifi-cant superiority in balancing privacy protection and modelutility. We use three different optimization schemes as de-scribed in section 3.3.1-3.3.3 on our framework and em-pirically shows all three largely outperform the baselinemethods. We also show that adding the model ensembleand model restarting, as described in section 3.4, to the op-timization procedure can further improve the performanceof our method.

4.1.1 Dataset and Problem SettingSBU Kinect Interaction Dataset [35] is a two-person in-teraction dataset for video-based action recognition. Sevenparticipants performed actions and the dataset is composedof 21 sets, each containing videos of a pair of differentpersons performing all 8 interactions. However, some setscontain the same two actors but with different person actingand reacting. For example, in set 1, actor 1 is acting and actor2 is reacting; in set 4, actor 2 is acting and actor 1 is reacting.These two sets have the same actors, so we combine them asone single class to fit with our experimental setting better.In this way, we combine all sets with the same actors andfinally get 13 different actor pairs. The target task on thisdataset is action recognition, which could be taken as aclassification task with 8 different classes. The budget taskis to recognize the actor pairs of the videos, which couldbe taken as a classification task with 13 different classes.Following the notations in section 3.2, in this dataset, wehave W = 112, H = 112, C = 3, and we set T = 16. Notethat the original resolution for SBU is 640 ∗ 480. We firstdownsample video frames to resolution 160 ∗ 120 and thencrop each frame to 112 ∗ 112.

4.1.2 Implementation DetailsWe compare the following six methods:• Naive Downsample: using raw RGB frames under

different down-sampling rates, following [1].• Cropping-Face: detecting and cropping out faces from

RGB frames, following [16].

0 20 40 60 80 100Actor Pair Accuracy AN

B (%)0

20

40

60

80

100

Actio

n ac

cura

cy A

T (%

)

Naive DownsampleCrop-FaceCrop-WholeOurs-KLOurs-EntropyOurs-Kbeam

Fig. 2. Target and Budget Task Performance Trade-off on SBU Dataset.For Naive Downsample method, a larger marker means a larger downsampling rate is adopted. For Ours-Kbeam method, a larger markermeans a larger K in algorithm 3. Neither model restarting nor modelensemble is used.

• Cropping-Whole: detecting and cropping out wholeactor bodies from RGB frames, following [16].

• Ours-KL: as described in section 3.3.1.• Ours-Entropy: as described in section 3.3.2.• Ours-Kbeam: as described in section 3.3.3.

In all three algorithms 1-3, we set step sizes αT = 10−5,αB = 10−2, αD = 10−4, accuracy thresholds th1 = 85%,th2 = 99% and max iter = 67. In algorithm 3, we setd iter to be 30 and we tried K = 1, 2, 4, 8. In algorithm4, we set restarting iter to be 100 and experimented withM = 1, 2, 4, 8. MobileNet [28] with different depth multipli-ers ranging from 0.6 to 0.44 are used as ensembled models.Other hyper-parameters of algorithm 4 are identical withthose in algorithm 2. We set γ=2 in loss function 3 and useAdam optimizer [36] to update all parameters. For NaiveDownsample method, we use down-sampling rate rangingfrom 1 (i.e. no down-sampling) to 56.

4.1.3 Result Comparison and Discussion

We present the experimental results in figure 2, which dis-plays the trade-off between the action recognition accuracyAT and the actor pair recognition accuracy ANB . In order tointerpret this figure, we should note that a desirable trade-off should incur minimal target accuracy AT (y-axis) whilereducing budget accuracy ANB (x-axis). Therefore, a pointcloser to the top-left corner represents a degrade model withmore desirable performance. The magenta dotted line sug-gests the target accuracy AT on original unprotected videos.This can be roughly considered as the AT upper bound forall privacy protection methods, under the assumption thatfD will unavoidably filter out some useful information forthe target task.

Page 8: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

8

0 20 40 60 80 100Actor Pair Accuracy AN

B (%)0

20

40

60

80

100Ac

tion

accu

racy

AT (

%)

Naive DownsampleCrop-FaceCrop-WholeOurs-Entropy(no restarting)Ours-Entropy(restarting)

Fig. 3. Target and Budget Task Performance Trade-off on SBU Dataset.For Naive Downsample method, a larger marker means a larger downsampling rate is adopted.

As we can see, all three of our methods largely outper-form the baselines. Crop-Face and Naive Downsample withlow downsample rate can lead to decent action accuracy,but the budget accuracy ANB is still very high, meaningthese methods fail to protect privacy. On the other hand,Crop-Whole and Naive Downsample with downsamplerate as high as 56 can effectively suppress ANB to a lowlevel, but AT also suffers a huge negative impact, whichmeans the degraded videos are of little practical utility. Ourmethods, in constrast, achieves a great balance betweenutility and privacy protection. Ours-Entropy and Ours-Kbeam with K = 2 can both decrease ANB by around 30%with nearly no harm on AT . Ours-KL and Ours-Kbeamwith K = 1, 4 achieve slightly worse trade-off comparedwith Ours-Entropy, but they still largely outperforms thebaseline methods. Due to its ease of implementation andlow complexity, we use Ours-Entropy as the default optionin our framework, unless otherwise noted.

4.1.4 Effectiveness of Model Restarting and Ensemble

In this section, we add model restarting and model ensem-ble to Ours-Entropy, as shown in algorithm 4, to furtherimprove the performance. Note that model restarting andmodel ensemble can be easily combined with all three of ourmethods, and we just pick Ours-Entropy here to show theireffectiveness. The results are shown in figure 3. As we cansee, using model restarting can suppress ANB much further,with no additional harm on AT . Model ensemble also helpsto improve the trade-off.

4.2 Action Recognition on UCF101 with Multiple Pri-vacy Attributes Protected: Cross-Dataset Training4.2.1 Dataset and Problem SettingUCF101 is an action recognition dataset with 13,320 real-lifehuman action videos collected from YouTube. It containsvideos of 101 different actions. We use official train-testsplit for this dataset. Similar to what we did with SBU, wespatially downsample and crop videos frames to 112 ∗ 112resolution, and then we temporally clip the videos to videoclips with shape [T,W,H,C] = [16, 112, 112, 3]. The targettask T is to do human action recognition on UCF101, whichcan be taken as a video classification task with 101 classes.

VISPR is a dataset with 22,167 images annotated with68 privacy attributes, e.g. semi-nudity, hobbies, face, race,gender, skin color and so on. Each attribute of an image islabeled as ’present’ and ’non-present’ depending whetherthe certain privacy attribute information is contained in theimage. We select 7 out of 68 attributes, which frequentlyappear in UCF101 datasets and well fit the smart homeapplication settings, to protect in our experiment. That’s tosay, we only consider these 7 attributes during training andevaluation. These 7 attributes are: semi-nudity, occupation,hobbies, sports, personal relationship, social relationshipand safe. The budget task B is to predict privacy attributeson VISPR dataset, which can be taken as a multi-label imageclassification task (7 labels, each is a binary classificationtask). We use class-based mean average precision (cMAP),whose definition can be referred to [20], to measure theperformance of this budget task. Official train-test split isused on VISPR dataset.

Fig. 4. Performance Trade-off on UCF-101/VISPR dataset.

4.2.2 Implementation DetailsExperiments on SBU has shown Ours-Entropy achievesslightly superior performance than Ours-KL and Ours-kbeam, so we only compare Ours-Entropy with baseline

Page 9: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

9

methods on this dataset. In algorithm 4, we set stepsizes αT = 10−5, αB = 10−2, αD = 10−4, accuracythresholds th1 = 70%, th2 = 99% and max iter =800, restarting iter = 100 and experimented withM = 1, 2, 4, 8. Models from {ResNet-V1-50, ResNet-V2-50,MobileNet-1.0, MobileNet-0.75} are used as {f iB}Mi=1. We setγ = 0.5 in loss function 3 and use Adam optimizer to updateall parameters. For Naive Downsample method, we use fourdifferent down-sampling rates: 1, 2, 4, 6.

4.2.3 Results Analyses

We present the experimental results in figure 4. NaiveDownsampling causes AT to drop dramatically while ANBonly drops a little bit, which means the utility of videos isgreatly damaged while the privacy information is hardly fil-tered out. On the contrary, with the help of model restartingand model ensemble, Ours-Entropy can decrease ANB by 7%while keeping AT as high as we can get on the original un-degraded videos, meaning the privacy is protected at almostno cost on the utility. Hence, Ours-Entropy outperformsNaive Downsampling in this experiment.

5 PA-HMDB51: A NEW BENCHMARK

5.1 Motivation

To the best of our knowledge, there is no public datasetcontaining both human action and privacy attribute labelson the same videos. The lack of available datasets hasnot only made it difficult to employ a data driven jointtraining method, but more importantly making it impossibleto directly evaluate the performance of a learned model tokeep privacy while not harming utility on a single dataset.To solve this problem, we annotate and present the veryfirst human action video dataset with privacy attributeslabeled, named PA-HMDB51 (Privacy Attribute HMDB51).We evaluate our method on this newly built dataset andfurther demonstrate our method’s effectiveness.

5.2 Selecting and Annoating Privacy Attributes

A recent work [20] has defined 68 privacy attributes whichcould be disclosed by images. However, most of themseldom make occurrence in public human action datasets.We carefully selected 7 privacy attributes which are mostrelevant to our smart home settings, out of 68 attributesfrom [20]. These seven attributes are: skin color, gender,face (partial), face (complete), nudity, personal relationshipand social circle. We further combined those 7 attributesinto 5 to better fit the human action videos: combine ”face(partial)” and ”face (complete)” into one attribute ”face” andcombine ”personal relationship” and ”social circle” into ”re-lationship”. To this end, we have got five privacy attributeswhich widely appear in public human action datasets andare closely relevant to our smart home setting. The detaileddescription of each attribute and their possible ground truthvalues and their corresponding meaning are listed in table1.

Privacy attributes may vary during the video clip. Forexample, in some frames we may see a person’s full face,while in next frames the person may turn around and the

face is no longer visible. We therefore decide to label allprivacy attributes for each frame.

The annotation of privacy labels was manually per-formed by a group of students at the CSE department ofTexas A&M University. Each video was at least annotatedby three individuals, and then cross-checked.

5.3 HMBD51 as the Data Source

Now that we have defined the five privacy attributes, weneed to identify a source of human action videos to label.There are a number of choices available, such as [33],[34], [37]–[39]. We choose HMDB51 [34] to label privacyattributes, since it consists of more diverse privacy infor-mation, especially nudity/semi-nudity.

We provide frame-level annotation of the selected 5privacy attributes on 592 videos selected from HMDB51.In this paper, we treat all 592 videos as testing samples;however, we do not exclude the future possibility to usethem for training.

5.4 Dataset Statistics

5.4.1 Action Distribution

When selecting videos from HMDB51 dataset, we considertwo criteria on action labels. First, the action labels shouldbe as balanced. Second (and more implicitly), we select morevideos with non-trivial privacy labels. For example, ”brushhair” action contains many videos with ”semi-nudity” at-tribute; ”drink” action contains many videos with ”can tellrelationship” attribute. Despite their practical importance,the two privacy attributes are relatively less seen in theentire HMDB51 dataset, so we tend to select more videoswith the two attributes, regardless of their action classes.

The resultant distribution of action labels are depicted inFigure 5, showing a relative class balanace.

5.4.2 Privacy Attribute Distribution

We try to make the label distribution for each attribute asbalanced as possible by manually selecting those videoscontaining uncommon privacy attribute values in originalHMDB51 to label. For instance, videos with semi-nudity areoverall uncommon, so we deliberately select those videoscontaining semi-nudity into our PA-HMDB51 dataset. Peo-ple are sensitive about releasing their privacy to the public,so the privacy attributes are highly unbalanced in anypublic video datasets by natural. Although we have usedthis method to try to relief this problem, the PA-HMDB51is still unbalanced. Figure 6 shows the frame-level labeldistribution of all five privacy attributes. We can see thevideos with ”cannot tell gender” and ”cannot tell skin color”still take a very small portion in the whole dataset.

Note that the total frame number of each privacy at-tribute is different since some vague frames may not belabeled with some certain attribute(s). For example, in someframe, it is hard to tell whether the face is complete or partial(it is somewhere in between), but you can tell the skin colorof the actor in all those frames. In this situation, we wouldnot label ’face’ attribute on those vague frames but wouldstill label ’skin color’ attribute.

Page 10: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

10

brush

hair

cartw

heelcat

chche

wcla

pclim

b

climb s

tairsdiv

e

draw sw

orddri

bbledri

nk eat

fall fl

oor

fencin

gflic

flacgo

lf

hand

stand hit hu

gjum

pkic

k

kick b

all kisslau

ghpickpo

urpu

lluppu

nchpush

pushu

p

ride b

ike

ride h

orse run

shake

hand

s

shoot

ball

shoot

bow

shoot

gun sit

situpsm

ilesm

oke

somers

aultsta

nd

swing

baseb

allsw

ord

sword

exerc

ise talkthr

owturnwalkwav

e

Actions

0

5

10

15

20

25

30Vi

deo

Num

ber

Fig. 5. Action distribution of PA-HMDB51. Each column shows the number of videos with a certain action. For example, the first column showsthere are 25 ”brush hair” videos in PA-HMDB51 dataset.

skin_color relationship face nudity genderPrivacy Attribute

0

10000

20000

30000

40000

50000

Fram

e N

umbe

r

0123

Fig. 6. Label distribution per privacy attribute in PA-HMDB51. Definitionsof label values (0,1,2,3) for each attribute are described in Table 1.

5.4.3 Action-Attribute CorrelationIf there is a strong correlation between a privacy attributeand an action, it would be harder to remove the privacyinformation from the videos without much harm of theaction recognition task. For example, we would expect highcorrelation between ’gender’ and the action ’brush hair’,since this action is carried our much more often by femalethan by male. Figure 7 shows the correlation between pri-vacy attributes and actions.

5.5 Benchmark Results on PA-HMDB515.5.1 Dataset and Problem SettingWe train our models using cross-dataset training onHMDB51 and VISPR datasets as we did in section 4.2, exceptthat we use the five attributes defined in Table 1 on VISPRinstead of the seven used in section 4.2. The trained modelsare directly evaluated on PA-HMDB51 dataset for both

target task T and budget task B,without any re-training oradaptation. We use the rest videos in HMDB51 not includedin PA-HMDB51 as training set. Similar with the UCF101experiments, the target task T (i.e. action recognition) canbe taken as a video classification problem with 51 classes,and the budget task B (i.e. privacy attribute prediction) canbe taken as a multi-label image classification task with twoclasses for each privacy attribute label. Notably, althoughPA-HMDB51 has provided concrete multi-class labels withspecific privacy attribute classes, we convert them intobinary labels during testing. For example, for “gender”attribute, we have provided ground truth labels “male”,“female” and “cannot tell”, but we only use “can tell” and“cannot tell” in our experiments, via combining “male” and“female” into the one class of “can tell”. This is becausewe have to keep the testing protocol on PA-HMDB51 inconsistency with the training protocol on VISPR (a multi-label, “either-or” type binary classification task, so that ourmodels cross-trained on UCF101-VISPR can be evaluateddirectly. Meanwhile, We hope to extend training to PA-HMDB51 in the future, so that the budget task as a multi-label can be formulated and evaluated as a multi-label,multi-classification problem.

The inputs X to our framework are clips of shape[T,W,H,C] = [16, 112, 112, 3], just the same as in theSBU and UCF101 experiments. All implementation detailsare identical with the UCF101 case, except that we adjustth1 = 0.7 and th2 = 0.95.

5.5.2 Results and Analysis

The results of Ours-KL method with M = 1, 4 (with andwithout restarting) and naive downsampling method withdownsample rates 1, 4, 8, 28 are shown in figure 8. Our

Page 11: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

11

brus

h hair

cartw

heel

catch

chew cla

pcli

mbcli

mb sta

irs dive

draw

swor

ddr

ibble

drink ea

tfal

l floo

rfen

cing

flic fla

cgo

lfha

ndsta

nd hit hug

jump

kick

kick b

all kiss

laugh pick

pour

pullu

ppu

nch

push

push

uprid

e bike

ride h

orse run

shak

e han

dssh

oot b

allsh

oot b

owsh

oot g

un sitsit

upsm

ilesm

oke

somer

sault

stand

swing

base

ball

swor

d

swor

d exe

rcise talk

thro

wtu

rnwa

lkwa

ve

NoneWhite

Brown/YellowBlack

skin color

0100020003000400050006000

brus

h hair

cartw

heel

catch

chew cla

pcli

mbcli

mb sta

irs dive

draw

swor

ddr

ibble

drink ea

tfal

l floo

rfen

cing

flic fla

cgo

lfha

ndsta

nd hit hug

jump

kick

kick b

all kiss

laugh pick

pour

pullu

ppu

nch

push

push

uprid

e bike

ride h

orse run

shak

e han

dssh

oot b

allsh

oot b

owsh

oot g

un sitsit

upsm

ilesm

oke

somer

sault

stand

swing

base

ball

swor

d

swor

d exe

rcise talk

thro

wtu

rnwa

lkwa

ve

NoneExist

relationship

01000200030004000

brus

h hair

cartw

heel

catch

chew cla

pcli

mbcli

mb sta

irs dive

draw

swor

ddr

ibble

drink ea

tfal

l floo

rfen

cing

flic fla

cgo

lfha

ndsta

nd hit hug

jump

kick

kick b

all kiss

laugh pick

pour

pullu

ppu

nch

push

push

uprid

e bike

ride h

orse run

shak

e han

dssh

oot b

allsh

oot b

owsh

oot g

un sitsit

upsm

ilesm

oke

somer

sault

stand

swing

base

ball

swor

d

swor

d exe

rcise talk

thro

wtu

rnwa

lkwa

ve

NonePartial

Complete

face

05001000150020002500

brus

h hair

cartw

heel

catch

chew cla

pcli

mbcli

mb sta

irs dive

draw

swor

ddr

ibble

drink ea

tfal

l floo

rfen

cing

flic fla

cgo

lfha

ndsta

nd hit hug

jump

kick

kick b

all kiss

laugh pick

pour

pullu

ppu

nch

push

push

uprid

e bike

ride h

orse run

shak

e han

dssh

oot b

allsh

oot b

owsh

oot g

un sitsit

upsm

ilesm

oke

somer

sault

stand

swing

base

ball

swor

d

swor

d exe

rcise talk

thro

wtu

rnwa

lkwa

ve

Strictly noneNone

Semi-nudity

nudity

05001000150020002500

brus

h hair

cartw

heel

catch

chew cla

pcli

mbcli

mb sta

irs dive

draw

swor

ddr

ibble

drink ea

tfal

l floo

rfen

cing

flic fla

cgo

lfha

ndsta

nd hit hug

jump

kick

kick b

all kiss

laugh pick

pour

pullu

ppu

nch

push

push

uprid

e bike

ride h

orse run

shak

e han

dssh

oot b

allsh

oot b

owsh

oot g

un sitsit

upsm

ilesm

oke

somer

sault

stand

swing

base

ball

swor

d

swor

d exe

rcise talk

thro

wtu

rnwa

lkwa

ve

Cannot tellMale

Female

gender

010002000300040005000

Fig. 7. Action-PA correlation in PA-HMDB51 dataset. The color represents the number of frames of each action containing a specific privacy attributevalue. For example, in the ”relationship” subplot, the intersection block of row ”Exist” and column ”kiss” shows the number of frames with ”relationshipexist” label in all kiss videos.

methods achieve much better privacy-utility trade-off com-pared with baseline methods. When M = 4, our methodsare able to decrease privacy cMAP by around 8% with littleharm on utility accuracy. Overall, the privacy gains are morelimited compared to the previous two experiments, becauseno (re-)training is performed; but the overall comparisontrends show the same consistency.

6 CONCLUSION

We proposed an innovative framework to solve the newly-established task of privacy-preserving visual recognition.To tackle the challenging adversarial learning process, weinvestigate three different optimization schemes. To fur-ther tackle the ∀ challenge of universal privacy protection,

Page 12: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

12

TABLE 1Attribute Definition on PA-HMDB51 Dataset

Attribute Possible Values Meaning

Skin Color

0 invisible Skin color of the actor is invisible.1 white Skin color of the actor is white.

2 brown/yellow Skin color of the actor is brown/yellow.3 black Skin color of the actor is black.

Face0 No face Less than 10% of the actor’s face is visible.

1 Partial face Less than 70% but more than 10% of the actor’s face is visible.2 Whole face More than 70% of the actor’s face is visible.

Gender0 Cannot tell Cannot tell the person’s gender.

1 Male It’s an actor.2 Female It’s an actress.

Nudity0 The actor/actress is wearing long sleeves and pants.1 The actor/actress is wearing short sleeves or shorts/short skirts.2 The actor/actress is of semi-nudity.

Relationship0 Cannot tell Relationships (such as friends, couples, etc.) between the actors/actress cannot be told from the video.

1 Can tell Relationships between the actors/actress can be told from the video.

TABLE 2Attribute Definition on PA-HMDB51 Dataset

Frame Action Privacy Attributes

Brush hair

• skin color: white• face: no• gender: female• nudity: level 2• relationship: no

Pullup

• skin color: white• face: whole• gender: male• nudity: level 1• relationship: no

we proposed model restarting and ensemble, which areshown to further improve the obtained trade-off. Varioussimulations confirmed the effectiveness of this proposalframework. Last but not least, we establish the very firstdataset for privacy-preserving video action recognition, an

effort that we hope could engage a broader community intothis research direction.

We note there being a large room to improve the pro-posed framework before it can achieve practical usefulness.For example, the definition of privacy leakage risk is core to

Page 13: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

13

Fig. 8. Performance trade-off on PA-HMDB51 dataset.

the framework. Considering the ∀ challenge, current LB de-fined with any specific fB is insufficient; the budget modelensemble could be only viewed as a rough discretizedapproximation of P . More elegant ways to model this ∀ op-timization may lead to further performance breakthroughs.

REFERENCES

[1] M. S. Ryoo, B. Rothrock, C. Fleming, and H. J. Yang, “Privacy-preserving human activity recognition from extreme low resolu-tion,” in Thirty-First AAAI Conference on Artificial Intelligence, 2017.

[2] D. J. Butler, J. Huang, F. Roesner, and M. Cakmak, “Theprivacy-utility tradeoff for remotely teleoperated robots,” in 10thACM/IEEE International Conference on Human-Robot Interaction,2015, pp. 27–34.

[3] J. Dai, B. Saghafi, J. Wu, J. Konrad, and P. Ishwar, “Towardsprivacy-preserving recognition of human activities,” in IEEE Inter-national Conference on Image Processing (ICIP), 2015, pp. 4238–4242.

[4] Z. Wu, Z. Wang, Z. Wang, and H. Jin, “Towards privacy-preservingvisual recognition via adversarial training: A pilot study,” inProceedings of the European Conference on Computer Vision (ECCV),2018, pp. 606–624.

[5] C. Gentry et al., “Fully homomorphic encryption using ideallattices.” in Annual ACM Symposium on the Theory of Computing(STOC), 2009, pp. 169–178.

[6] P. Xie, M. Bilenko, T. Finley, R. Gilad-Bachrach, K. Lauter, andM. Naehrig, “Crypto-nets: Neural networks over encrypted data,”arXiv preprint arXiv:1412.6181, 2014.

[7] A. Chattopadhyay and T. E. Boult, “Privacycam: a privacy preserv-ing camera using uclinux on the blackfin dsp,” in IEEE Conferenceon Computer Vision and Pattern Recognition (CVPR), 2007, pp. 1–8.

[8] A. Mahendran and A. Vedaldi, “Visualizing deep convolutionalneural networks using natural pre-images,” International Journal ofComputer Vision (IJCV), vol. 120, no. 3, pp. 233–255, May 2016.

[9] T. Winkler, A. Erdelyi, and B. Rinner, “Trusteye. m4: protecting thesensornot the camera,” in 2014 11th IEEE International Conferenceon Advanced Video and Signal Based Surveillance (AVSS). IEEE, 2014,pp. 159–164.

[10] F. Pittaluga and S. J. Koppal, “Privacy preserving optics forminiature vision sensors,” in Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition (CVPR), 2015, pp. 314–324.

[11] ——, “Pre-capture privacy for small vision sensors,” IEEE Transac-tions on Pattern Analysis and Machine Intelligence (TPAMI), vol. 39,no. 11, pp. 2215–2226, 2017.

[12] L. Jia and R. J. Radke, “Using time-of-flight measurements forprivacy-preserving tracking in a smart room,” IEEE Transactionson Industrial Informatics, vol. 10, no. 1, pp. 689–696, 2014.

[13] S. Tao, M. Kudo, and H. Nonaka, “Privacy-preserved behavioranalysis and fall detection by an infrared ceiling sensor network,”Sensors, vol. 12, no. 12, pp. 16 920–16 936, 2012.

[14] Z. Wang, S. Chang, Y. Yang, D. Liu, and T. S. Huang, “Studyingvery low resolution recognition using deep networks,” 2016 IEEEConference on Computer Vision and Pattern Recognition (CVPR), pp.4792–4800, 2016.

[15] B. Cheng, Z. Wang, Z. Zhang, Z. Li, D. Liu, J. Yang, S. Huang,and T. S. Huang, “Robust emotion recognition from low qualityand low bit rate video: A deep learning approach,” in 2017 Sev-enth International Conference on Affective Computing and IntelligentInteraction (ACII). IEEE, 2017, pp. 65–70.

[16] Y. Li, N. Vishwamitra, B. P. Knijnenburg, H. Hu, and K. Caine,“Blur vs. block: Investigating the effectiveness of privacy-enhancing obfuscation for images,” in IEEE Conference on ComputerVision and Pattern Recognition Workshops (CVPRW), 2017, pp. 1343–1351.

[17] S. J. Oh, R. Benenson, M. Fritz, and B. Schiele, “Faceless personrecognition: Privacy implications in social media,” in EuropeanConference on Computer Vision. Springer, 2016, pp. 19–35.

[18] R. McPherson, R. Shokri, and V. Shmatikov, “Defeating imageobfuscation with deep learning,” arXiv preprint arXiv:1609.00408,2016.

[19] S. J. Oh, M. Fritz, and B. Schiele, “Adversarial image perturbationfor privacy protection a game theory perspective,” in 2017 IEEEInternational Conference on Computer Vision (ICCV). IEEE, 2017,pp. 1491–1500.

[20] T. Orekondy, B. Schiele, and M. Fritz, “Towards a visual privacyadvisor: Understanding and predicting privacy risks in images,”in Proceedings of the IEEE International Conference on Computer Vision(ICCV), 2017, pp. 3686–3695.

[21] X. Xiang and T. D. Tran, “Linear disentangled representation learn-ing for facial actions,” IEEE Transactions on Circuits and Systems forVideo Technology, vol. 28, no. 12, pp. 3539–3544, 2018.

[22] G. Desjardins, A. Courville, and Y. Bengio, “Disentanglingfactors of variation via generative entangling,” arXiv preprintarXiv:1210.5474, 2012.

[23] A. Gonzalez-Garcia, J. van de Weijer, and Y. Bengio, “Image-to-image translation for cross-domain disentanglement,” arXivpreprint arXiv:1805.09730, 2018.

[24] N. Siddharth, B. Paige, A. Desmaison, J.-W. van de Meent, F. Wood,N. D. Goodman, P. Kohli, and P. H. Torr, “Learning disentangledrepresentations in deep generative models,” 2016.

[25] J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-timestyle transfer and super-resolution,” in Proceedings of the EuropeanConference on Computer Vision (ECCV), 2016.

[26] D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri, “Learn-ing spatiotemporal features with 3d convolutional networks,” inProceedings of the IEEE International Conference on Computer Vision(ICCV), 2015, pp. 4489–4497.

[27] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning forimage recognition,” 2016 IEEE Conference on Computer Vision andPattern Recognition (CVPR), 2016.

[28] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang,T. Weyand, M. Andreetto, and H. Adam, “Mobilenets: Efficientconvolutional neural networks for mobile vision applications,”arXiv preprint arXiv:1704.04861, 2017.

[29] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley,S. Ozair, A. Courville, and Y. Bengio, “Generative adversarialnets,” in Advances in Neural Information Processing Systems (NIPS),2014, pp. 2672–2680.

[30] Y. Ganin and V. Lempitsky, “Unsupervised domain adaptation bybackpropagation,” in Proceedings of the 32nd International Conferenceon Machine Learning (ICML), 2015, pp. 1180–1189.

[31] J. Hamm and Y.-K. Noh, “K-beam minimax: Efficient optimizationfor deep adversarial learning,” in Proceedings of the 35th Interna-tional Conference on Machine Learning (ICML), 2018, pp. 1881–1889.

[32] B. Zoph, V. Vasudevan, J. Shlens, and Q. V. Le, “Learning transfer-able architectures for scalable image recognition,” in IEEE Confer-ence on Computer Vision and Pattern Recognition (CVPR), 2018, pp.8697–8710.

[33] K. Soomro, A. R. Zamir, and M. Shah, “Ucf101: A dataset of 101human actions classes from videos in the wild,” arXiv preprintarXiv:1212.0402, 2012.

[34] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre, “Hmdb:a large video database for human motion recognition,” in Proceed-

Page 14: 1 Privacy-Preserving Deep Visual Recognition: An ... · Abstract—This paper aims to boost privacy-preserving visual recognition, an increasingly demanded feature in smart camera

14

ings of the IEEE International Conference on Computer Vision (ICCV),2011, pp. 2556–2563.

[35] K. Yun, J. Honorio, D. Chattopadhyay, T. L. Berg, and D. Samaras,“Two-person interaction detection using body-pose features andmultiple instance learning,” in IEEE Computer Society Conference onComputer Vision and Pattern Recognition Workshops (CVPRW), 2012,pp. 28–35.

[36] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimiza-tion,” in International Conference for Learning Representations (ICLR),2015.

[37] B. G. Fabian Caba Heilbron, Victor Escorcia and J. C. Niebles,“Activitynet: A large-scale video benchmark for human activityunderstanding,” in Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition, 2015, pp. 961–970.

[38] C. Gu, C. Sun, D. A. Ross, C. Vondrick, C. Pantofaru, Y. Li,S. Vijayanarasimhan, G. Toderici, S. Ricco, R. Sukthankar, andet al., “Ava: A video dataset of spatio-temporally localizedatomic visual actions,” 2018 IEEE/CVF Conference on ComputerVision and Pattern Recognition, Jun 2018. [Online]. Available:http://dx.doi.org/10.1109/cvpr.2018.00633

[39] W. Kay, J. Carreira, K. Simonyan, B. Zhang, C. Hillier, S. Vijaya-narasimhan, F. Viola, T. Green, T. Back, P. Natsev, M. Suleyman,and A. Zisserman, “The kinetics human action video dataset,”2017.

Haotao Wang received the B.E. degree in elec-tronics engineering from Tsinghua University,China, in 2018. He is working toward the PhDdegree in Texas A&M University, under the su-pervision of Zhangyang Wang. His research in-terests lie in computer vision and machine learn-ing, especially in privacy and adversarial attackand defense.

Zhenyu Wu received the M.S. and B.E. degreesfrom the Ohio State University and ShanghaiJiao Tong University respectively. He is currentlya Ph.D. student in Texas A&M University, ad-vised by Zhangyang Wang. His research inter-ests include visual privacy protection, neural net-work compression, object detection, fairness ingenerative models and hand pose estimation.

Zhangyang Wang is an Assistant Professor ofComputer Science and Engineering (CSE), atthe Texas A&M University (TAMU). During 2012-2016, he was a Ph.D. student in the Electricaland Computer Engineering (ECE) Department,at the University of Illinois at Urbana-Champaign(UIUC), working with Professor Thomas S.Huang. Dr. Wangs research has been address-ing machine learning, computer vision and op-timization problems, as well as their interdisci-plinary applications. He has co-authored over 80

papers, and published 2 books and 1 chapter. He has been granted 3patents, and has received around 20 research awards and scholarships.

Zhaowen Wang received the B.E. and M.S.degrees from Shanghai Jiao Tong University,China, in 2006 and 2009, respectively, and thePh.D. degree in electrical and computer engi-neering from the University of Illinois at Urbana-Champaign, in 2014. He is currently a SeniorResearch Scientist with the Creative IntelligenceLab, Adobe Inc. His research has been focusedon understanding and enhancing images, videosand graphics via machine learning algorithms,with a particular interest in sparse coding and

deep learning.

Hailin Jin is a Senior Principal Scientist atAdobe Research. He received his Master of Sci-ence and Doctor of Science degrees in ElectricalEngineering from Washington University in SaintLouis in 2000 and 2003, respectively. His advi-sor was Professor Stefano Soatto. Between fall2003 and fall 2004, he was a postdoctoral re-searcher at the Computer Science Department,University of California at Los Angeles. His cur-rent research interests include: deep learning,natural language processing, computer vision,

video, image search, 3D reconstruction, structure and motion estima-tion, optical flow, stereo, and image-based modeling and rendering. Hiswork can be found in many Adobe products including Photoshop, AfterEffects, Premiere Pro, and Photoshop Lightroom.