Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to...

11
1 Color Face Recognition using High-Dimension Quaternion-based Adaptive Representation Qingxiang Feng, Yicong Zhou, Senior Member, IEEE, Abstract—Recently, quaternion collaborative representation- based classification (QCRC) and quaternion sparse representation-based classification (QSRC) have been proposed for color face recognition. They can obtain correlation information among different color channels. However, their performance is unstable in different conditions. For example, QSRC performs better than than QCRC on some situations but worse on other situations. To benefit from quaternion-based e2-norm minimization in QCRC and quaternion-based e1-norm minimization in QSRC, we propose the quaternion-based adaptive representation (QAR) that uses a quaternion-based ep-norm minimization (1 p 2) for color face recognition. To obtain the high dimension correlation information among different color channels, we further propose the high-dimension quaternion-based adaptive representation (HD-QAR). The experimental results demonstrate that the proposed QAR and HD-QAR achieve better recognition rates than QCRC, QSRC and several state-of-the-art methods. Index Terms—Sparse Representation, Collaborative Represen- tation, Quaternion, Color Face Recognition. I. I NTRODUCTION F ACE recognition systems are relied on classification methods. Recently, representation-based classification methods [1]–[7] have attracted more attention and ob- tained good performance in face recognition. Two typi- cal representation-based classification methods are sparse representation-based classification (SRC) [1] [8] and collabo- rative representation based classification (CRC) [9]. For color face recognition, SRC and CRC independently use color channels of color face images. As a result, they fails to consider the structural correlation information among differ- ent color channels. The previous research works [10]–[15] proved that quaternion can obtain the information among different channels of the color images by representing three color channels as three imaginary parts. For example, [14] proposed a quaternion-based structural similarity index and used it to assess the quality of color images. [16] pro- posed a vector sparse representation using quaternion matrix analysis and successfully applied it to various color image processing. [17] proposed two methods to obtain correlation information among red, green and blue channels for color face recognition. These two methods are named quaternion collaborative representation-based classification (QCRC) and This work was supported in part by the Macau Science and Technology Development Fund under Grant FDCT/016/2015/A1 and by the Research Committee at University of Macau under Grants MYRG2014-00003-FST and MYRG2016-00123-FST. (Corresponding author is Yicong Zhou.) All authors are with the Department of Computer and Information Science, University of Macau, Macau 999078, China (e-mail: [email protected]; [email protected]). quaternion sparse representation-based classification (QSRC). QCRC obtains the sparse coefficient by solving quaternion- based e 2 -norm minimization problem while QSRC solves quaternion e 1 -norm minimization problem to compute the sparse coefficient. From the results in [17], we find an in- teresting phenomenon that QSRC obtains better performance than QCRC on some situations while QCRC performs better than QSRC on other situations. Table II proves this. That is, quaternion-based e 2 -norm minimization and quaternion-based e 1 -norm minimization both have their advantages. Motivated by this, we want to design a quaternion-based e p - norm minimization (1 p 2) to solve the sparse optimiza- tion problem such that it can benefit both from quaternion- based e 2 -norm minimization in QCRC and quaternion-based e 1 -norm minimization in QSRC. Moreover, [18] shows that high dimension information in kernel space is helpful for sparse. We also want to design a kernel-quaternion-based transformation that can obtain the high dimension correlation information among different color channels. Our main contri- butions are described as follows. To benefit both from quaternion-based e 2 -norm mini- mization in QCRC and quaternion-based e 1 -norm mini- mization in QSRC, we propose quaternion-based adaptive representation (QAR) method. QAR uses a quaternion- based e p -norm minimization (1 p 2) that can adap- tively combine quaternion-based e 2 -norm minimization and quaternion-based e 1 -norm minimization. To our best knowledge, the high-dimension quaternion has not been applied. Thus, we propose the high- dimension quaternion-based adaptive representation (HD- QAR). It is able to obtain high-dimension correlation information among different color channels. Because the product of high-dimension quaternion is non- communicative, it is quite difficult to solve the high- dimension quaternion-based optimization problem in HD- QAR. Thus, we propose the quaternion-based kernel matrix and two high-dimension-quaternion operators to solve optimization problem in HD-QAR be easy. Extensive experiments on six well-known color face databases show that the proposed methods have better performance than several state-of-the-art methods. II. RELATED WORK This section reviews the related works including quaternion, kernel and representation-based methods. arXiv:1712.01642v1 [cs.CV] 19 Nov 2017

Transcript of Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to...

Page 1: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

1

Color Face Recognition using High-DimensionQuaternion-based Adaptive Representation

Qingxiang Feng, Yicong Zhou, Senior Member, IEEE,

Abstract—Recently, quaternion collaborative representation-based classification (QCRC) and quaternion sparserepresentation-based classification (QSRC) have been proposedfor color face recognition. They can obtain correlationinformation among different color channels. However, theirperformance is unstable in different conditions. For example,QSRC performs better than than QCRC on some situationsbut worse on other situations. To benefit from quaternion-basede2-norm minimization in QCRC and quaternion-based e1-normminimization in QSRC, we propose the quaternion-basedadaptive representation (QAR) that uses a quaternion-basedep-norm minimization (1 ≤ p ≤ 2) for color face recognition.To obtain the high dimension correlation information amongdifferent color channels, we further propose the high-dimensionquaternion-based adaptive representation (HD-QAR). Theexperimental results demonstrate that the proposed QAR andHD-QAR achieve better recognition rates than QCRC, QSRCand several state-of-the-art methods.

Index Terms—Sparse Representation, Collaborative Represen-tation, Quaternion, Color Face Recognition.

I. INTRODUCTION

FACE recognition systems are relied on classificationmethods. Recently, representation-based classification

methods [1]–[7] have attracted more attention and ob-tained good performance in face recognition. Two typi-cal representation-based classification methods are sparserepresentation-based classification (SRC) [1] [8] and collabo-rative representation based classification (CRC) [9]. For colorface recognition, SRC and CRC independently use colorchannels of color face images. As a result, they fails toconsider the structural correlation information among differ-ent color channels. The previous research works [10]–[15]proved that quaternion can obtain the information amongdifferent channels of the color images by representing threecolor channels as three imaginary parts. For example, [14]proposed a quaternion-based structural similarity index andused it to assess the quality of color images. [16] pro-posed a vector sparse representation using quaternion matrixanalysis and successfully applied it to various color imageprocessing. [17] proposed two methods to obtain correlationinformation among red, green and blue channels for colorface recognition. These two methods are named quaternioncollaborative representation-based classification (QCRC) and

This work was supported in part by the Macau Science and TechnologyDevelopment Fund under Grant FDCT/016/2015/A1 and by the ResearchCommittee at University of Macau under Grants MYRG2014-00003-FST andMYRG2016-00123-FST. (Corresponding author is Yicong Zhou.)

All authors are with the Department of Computer and Information Science,University of Macau, Macau 999078, China (e-mail: [email protected];[email protected]).

quaternion sparse representation-based classification (QSRC).QCRC obtains the sparse coefficient by solving quaternion-based e2-norm minimization problem while QSRC solvesquaternion e1-norm minimization problem to compute thesparse coefficient. From the results in [17], we find an in-teresting phenomenon that QSRC obtains better performancethan QCRC on some situations while QCRC performs betterthan QSRC on other situations. Table II proves this. That is,quaternion-based e2-norm minimization and quaternion-basede1-norm minimization both have their advantages.

Motivated by this, we want to design a quaternion-based ep-norm minimization (1 ≤ p ≤ 2) to solve the sparse optimiza-tion problem such that it can benefit both from quaternion-based e2-norm minimization in QCRC and quaternion-basede1-norm minimization in QSRC. Moreover, [18] shows thathigh dimension information in kernel space is helpful forsparse. We also want to design a kernel-quaternion-basedtransformation that can obtain the high dimension correlationinformation among different color channels. Our main contri-butions are described as follows.

• To benefit both from quaternion-based e2-norm mini-mization in QCRC and quaternion-based e1-norm mini-mization in QSRC, we propose quaternion-based adaptiverepresentation (QAR) method. QAR uses a quaternion-based ep-norm minimization (1 ≤ p ≤ 2) that can adap-tively combine quaternion-based e2-norm minimizationand quaternion-based e1-norm minimization.

• To our best knowledge, the high-dimension quaternionhas not been applied. Thus, we propose the high-dimension quaternion-based adaptive representation (HD-QAR). It is able to obtain high-dimension correlationinformation among different color channels.

• Because the product of high-dimension quaternion is non-communicative, it is quite difficult to solve the high-dimension quaternion-based optimization problem in HD-QAR. Thus, we propose the quaternion-based kernelmatrix and two high-dimension-quaternion operators tosolve optimization problem in HD-QAR be easy.

• Extensive experiments on six well-known color facedatabases show that the proposed methods have betterperformance than several state-of-the-art methods.

II. RELATED WORK

This section reviews the related works including quaternion,kernel and representation-based methods.

arX

iv:1

712.

0164

2v1

[cs

.CV

] 1

9 N

ov 2

017

Page 2: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

2

TABLE INOTATION SUMMARY

Notation DescriptionX Entire training setXc All samples of the cth classxi The ith sample of Xq Dimension of a sampleNc Number of samples of the cth classM Number of classesL Number of samples of Xx A testing sampleX Matrix-based quaternionx Vector-based quaternion

A. Quaternion Algebra

Hamilton proposed the quaternion space that has threeimagery units i, j and k [19]. These three imagery units aredefined as

i2 = j2 = k2 = ijk = −1 (1)

A quaternion q is described by

q = q0 + q1i+ q2j + q3k (2)

where q0, q1, q2, q3 are real numbers. The conjugatequaternion of q is

¯q = q0 − q1i− q2j − q3k (3)

The modulus of quaternion q is

|q| =√qq =

√q20 + q21 + q22 + q23 (4)

The addition operation of two quaternions p and q is

q+ p = (q0 + p0) + (q1 + p1)i+ (q2 + p2)j+ (q3 + p3)k (5)

The multiplication operation of two quaternions p and q is

qp = (q0p0 − q1p1 − q2p2 − q3p3)+(q0p1 + q1p0 + q2p3 − q3p2)i+(q0p2 − q1p3 + q2p0 + q3p1)j+(q0p3 + q1p2 − q2p3 + q3p0)k

(6)

B. Kernel trick

The kernel techniques can map the data to a high-dimensionspace. Using the kernel trick, we can obtain the high dimensioninformation without computing the mapping explicitly. In thispaper, we use a well-known kernel called the Gaussian radialbasis function (RBF) kernel. It can be represented as

k(a, b) = φ(a)Tφ(b) = exp(−||a− b||2

δ) (7)

where a and b are any two original samples, δ is a parameter.In the kernel trick, φ(∗) is unknown. We can access the featurespace only via k(∗, ∗) .

TABLE IIMEAN RESULTS OF TABLE III IN [17]

Method QCRC QSRC1 48.70 51.572 67.94 68.673 75.80 75.414 78.99 78.41

C. Representation-based Methods

Wright et al. proposed sparse representation classification(SRC) in 2009. It represents the testing sample by the linearcombination of training samples of all classes. SRC solvese1-minimization optimization problem as

min ||α||1 s.t. x = Xα (8)

To reduce the computation cost and obtain the better represen-tation, Zhang et al. proposed the collaborative representationbased classification (CRC). In CRC, the authors argued thatthe collaborative representation should be better than thelinear combination of training samples. CRC solves the e2-minimization optimization problem as

min ||α||22 s.t. x = Xα (9)

They are two typical sparse-based methods. A lot modifi-cations of SRC/CRC have been proposed for various visualrecognition tasks [20]–[24]. For example, Gao et al. denotedthat the nonlinear high dimensional features are useful forsparse representation, and then proposed kernel-based sparserepresentation classification (KSRC) [18], [25],

min ||α||1 s.t. Φ(x) = Φ(X)α (10)

Wang et al. proposed kernel-based collaborative representationclassification (KCRC) [26], which is described as

min ||α||2 s.t. Φ(x) = Φ(X)α (11)

Zou et al. denoted that quaternion can help sparse representa-tion to obtain the correlation information among different colorchannels. Then, they proposed quaternion sparse representa-tion classification (QSRC) and quaternion collaborative rep-resentation based classification (QCRC). QSRC is describedas

min ||α||1 s.t. x = Xα (12)

QCRC is described as

min ||α||2 s.t. x = Xα (13)

III. PROPOSED QAR

In this section, we propose a new method, called quaternion-based adaptive representation (QAR). We first introduce itsmotivation. Then, we define and analyze its objective function.Afterwards, the solution of this objective function is given.Last, we describe the classification rule and summary thismethod.

Page 3: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

3

Fig. 1. Motivation: Because quaternion-based e2-norm and e1-norm bothhave their own advantages, we want to design a quaternion-based ep-norm(1 ≤ p ≤ 2) that can adaptively combine quaternion-based e2-norm ande1-norm and benefit from both. The above is the geometric interpretation ofep-norm, e2-norm and e1-norm.

A. Motivation

To consider the structural correlation information amongdifferent color channels, [17] proposed two methods forcolor face recognition. They are quaternion collaborativerepresentation-based classification (QCRC) using e1-normminimization and quaternion sparse representation-based clas-sification (QSRC) using e2-norm minimization. Table II listssome results copied from [17]. From Table II and [17], we findthat QSRC obtains the better performance than QCRC on somesituations while QCRC obtains the better performance thanQSRC on other situations. This means that quaternion-basede2-norm minimization and quaternion-based e1-norm mini-mization both have their own advantages. Motivated by this,this paper designs a quaternion-based ep-norm minimization(1 ≤ p ≤ 2) to solve the sparse optimization problem such thatit can benefit from quaternion-based e2-norm minimization inQCRC and quaternion-based e1-norm minimization in QSRC.We give an example in Fig. 1 to describe the geometricinterpretation of ep-norm, e2-norm and e1-norm.

B. Objective function of QAR

This paper tries to propose a quaternion-based ep-norm thatcan adaptively combine quaternion-based e2-norm and e1-norm and benefit from both. Our objective function can bedescribed as follows.

min ||α||p s.t. x = Xα (14)

where 1 ≤ p ≤ 2. Now, we face a problem how to obtain abetter ep-norm. To solve this, we give a new objective functionthat can be treated as ep-norm as

min ||XDiag(α)||∗s.t. x = Xα

(15)

where || • ||∗ means the nuclear or trace norm.Because the product of quaternions are non-communicative,it is quite difficult to solve the quaternion-based optimization

problem in QAR. Following [17], we give two quaternion-based operations <(∗) and ℵ(∗) . They are defined as follows.Given a matrix-based quaternion V = V0 + V1i+ V2j + V3k. <(V ) is equal to

<(V ) :=

V0 −V1 −V2 −V3V1 V0 −V3 V2V2 V3 V0 −V1V3 −V2 V1 V0

(16)

Given a vector-based quaternion v = v0 + v1i + v2j + v3k.ℵ(v) is equal to

ℵ (v) =[vT0 vT1 vT2 vT3

]T(17)

Using the two quaternion-based operations <(∗) and ℵ(∗), theobjective function in Eq. (14) can be rewritten as

min ||ℵ(α)||p s.t. ℵ(x) = <(X)ℵ(α) (18)

and the objective function in Eq. (15) can be rewritten as

min ||<(X)Diag(ℵ(α))||∗s.t. ℵ(x) = <(X)ℵ(α)

(19)

Next, we will explain why Eq. (19) can be treated as ep-norm(1 ≤ p ≤ 2).

C. Analysis of objective function

This section analyzes the objective function and explainwhy min ||<(X)Diag(ℵ(α))||∗ in Eq. (19) is equivalent tomin ||ℵ(α)||p in Eq (18). In order to explain the proposedobjective function can be treated as ep-norm minimization(1 ≤ p ≤ 2), we give two examples. Suppose that the subjects(columns) of <(X) are orthogonal and different from eachother, that is, <(X)T<(X) = I , where I is an identity matrix.We define Ω = ||<(X)Diag(ℵ(α))||∗ and its decompositionis

Ω = Tr

[(<(X)Diag(ℵ(α))

)T (<(X)Diag(ℵ(α))

)]1/2= Tr

[(Diag(ℵ(α)))

T(Diag(ℵ(α)))

]1/2= ||ℵ(α)||1

(20)Based on Eq. (20), we know that the objective function in Eq.(19) is equivalent to e1-norm minimization problem.On the contrary, we suppose that the subjects (columns)of <(X) are same to the first column of <(X) , that is<(X)T<(X) = 11T ( 1 is a vector that all elements areone). We define the first column of <(X) as <(X)1 and Thedecomposition is

Ω = ||<(X)11TDiag(ℵ(α))||∗ (21)

Suppose that the rank of a matrix A ∈ Rm×n is r. We canobtain the inequalities [27] [28] as follows. 1

||A||F ≤ ||A||∗ ≤√r||A||F (22)

1https://en.wikipedia.org/wiki/Matrix norm

Page 4: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

4

where || • ||F is the Frobenius norm. Because therank of <(X)11TDiag(ℵ(α)) is one (r = 1), Ω =||<(X)Diag(ℵ(α))||∗ can be re-wrote as

Ω = ||<(X)11TDiag(ℵ(α))||∗= ||<(X)11TDiag(ℵ(α))||F= ||<(X)1||2||1T ||2||Diag(ℵ(α))||2= ||ℵ(α)||2

(23)

In the real applications, the subjects (columns) of data set<(X) are not different each other or same to each other.Consider the Eqs. (20) and (23), our objective function couldbe treated as a combination of the quaternion-based e1-norm minimization in QSRC and quaternion-based e2-normminimization in QCRC. That is, Eq. (19) can be treatedas quaternion-based ep-norm. Our objective function couldbenefit from both quaternion-based e1-norm minimization andquaternion-based e2-norm minimization based on the correc-tion of the subjects (columns) in the data set <(X) .

D. Solution of objective function

Based on the optimization methods in [29] and [30], weuse the inexact augmented Lagrange multipliers (IALM) [40]method to solve the proposed ep-norm minimization. Weconvert the minimization problem in Eq.(19) to the followingoptimization problem as

minZ,z,ℵ(α)

||Z||∗ + ||z||1

s.t. z = ℵ(x)−<(X)ℵ(α)

Z = <(X)Diag(ℵ(α))

(24)

We further convert the optimization problem in Eq. (24) to aaugmented Lagrange multiplier problem as follows.

L(Z, z,ℵ(α)) = λ||Z||∗ +mT1 (ℵ(x)−<(X)ℵ(α)− z)

+ ||z||1 + Tr[MT2 (Z −<(X)Diag(ℵ(α)))]

+ u2 (||a−<(X)ℵ(α)− z||22 + ||Z −<(X)Diag(ℵ(α))||2F )

(25)where u > 0 is a parameter, m1 and M2 are Lagrangemultipliers. We can alternatively optimize the variates Z, z andℵ(α) by fixing others. The detailed optimization processes ofZ, z and ℵ(α) are described as follows.Update Z by fixing z and ℵ(α) . It is to solve the followingminimization problem as

Z∗ = arg minZL(Z, z,ℵ(α))

= arg minZλ||Z||∗ + Tr(MT

2 Z)

+ u2 ||Z −<(X)Diag(ℵ(α))||2F

= arg minZ

λu ||Z||∗ + 1

2 ||Z − (<(X)Diag(ℵ(α))− 1uM2)||2F

(26)We can approximately solve the above minimization problemin Eq. (26) using the singular value thresholding (SVT)operator [31].

Then, update ℵ(α) by fixing z and Z. It is to solve thefollowing minimization problem as.

ℵ(α)∗ = arg minℵ(α)

L(Z, z,ℵ(α))

= arg minℵ(α)−mT

1 Jℵ(α)− Tr(MT2 JDiag(ℵ(α)))

+ u2 (ℵ(α)TJTJℵ(α)− 2(ℵ(x)− z)TJℵ(α)

− 2Tr(ZTJDiag(ℵ(α)))+ Tr((JDiag(ℵ(α)))TJDiag(ℵ(α))))

= arg minℵ(α)

u2ℵ(α)T (JTJ +Diag(diag(JTJ)))ℵ(α)−

(JTm1 + uJTJ(ℵ(x)− z) + diag(MT2 J + uZTJ))Tℵ(α)

(27)where J = <(X). The above minimization problem in Eq.(27) can be solved by

ℵ(α)∗ =(JTJ +Diag(diag(JTJ)))−1JT ( 1

um1 + J(ℵ(α)− z))+(JTJ +Diag(diag(JTJ)))−1diag(JT ( 1

uM2 + Z))(28)

Next, update z by fixing Z and ℵ(α). It is to solve thefollowing minimization problem as

z∗ = arg minzL(Z, z,ℵ(α))

= arg minz||z||1 −mT

1 z + u2 ||a−<(X)ℵ(α)− z||2F

= arg minz

1u ||z||1 + 1

2 ||z − (a−<(X)ℵ(α) + 1um1)||22

(29)We can solve the minimization problem in Eq. (29) by usingthe soft thresholding (shrinkage) operator [32]. After updatingZ, z and ℵ(α), the multipliers m1,M2 can be updated by

m1 = m1 + u(ℵ(x)−<(X)ℵ(α)− z)M2 = M2 + u(Z −<(X)Diag(ℵ(α)))

(30)

We update the parameter u by u = min(ρu, umax), whereρ, umax are two constant values.

E. Classification

After obtaining the results of the optimization problemof the objective function, we calculate the distance betweentesting sample x and each class by

dc = ||ℵ(x)−<(Xc)ℵ(αc)|| (31)

Finally, the testing sample will be classified into the class withthe minimization distance by

c∗ = arg min(dc) (32)

The detailed classification procedures of QAR are summarizedin Algorithm 1.

IV. PROPOSED HD-QAR

This section proposes high-dimension quaternion basedadaptive representation (HD-QAR). We first introduce themotivation and its objective function. The solution of this ob-jective function and the classification rule are then presented.Finally, we compare the proposed methods with several state-of-the-art methods.

Page 5: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

5

Algorithm 1 Proposed Quaternion-based Adaptive Representation (QAR) MethodRequire: The original training set X with L samples and a testing sample x. Initialize the value of Z, z, α,m1,M2, u, ρ, ε

and umaxEnsure: Label of the testing sample x.

1: Constitute the objective function using the original training set X and testing sample x as

ℵ(α)∗ = arg min ||ℵ(x)−<(X)ℵ(α)||1 + λ||<(X)Diag(ℵ(α))||∗

2: Transform the objective function to a Lagrange multiplier problem as

L(Z, z,ℵ(α)) = λ||Z||∗ + ||z||1 +mT1 (a−<(X)ℵ(α)− z) + Tr[MT

2 (Z −<(X)Diag(ℵ(α)))]

+ u2 (||ℵ(x)−<(X)ℵ(α)− z||22 + ||Z −<(X)Diag(ℵ(α))||2F )

3: while until convergence do4: Update Z by fixing the others as Z∗ = arg min

Z

λu ||Z||∗ + 1

2 ||Z − (<(X)Diag(ℵ(α))− 1uM2)||2F

5: Update ℵ(α) by fixing the others as

ℵ(α)∗ = (<(X)T<(X) +Diag(diag(<(X)T<(X))))−1<(X)T ( 1um1 + <(X)(ℵ(x)− z))

+(<(X)T<(X) +Diag(diag(<(X)T<(X))))−1diag(<(X)T ( 1uM2 + Z))

6: Update z by fixing the others as z∗ = arg minz

1u ||z||1 + 1

2 ||z − (ℵ(x)−<(X)ℵ(α) + 1um1)||2

7: The multipliers m1,M2 are updated by m1 = m1+u(ℵ(x)−<(X)ℵ(α)−z) and M2 = M2+u(Z−<(X)Diag(ℵ(α)))

8: Update the parameter u by u = min(ρu, umax)

9: Check convergence by ||(ℵ(x)−<(X)ℵ(α)− z||∞ ≤ ε and ||E −ADiag(β)||∞ ≤ ε10: end while11: Compute the distance dc = ||ℵ(x)−<(Xc)ℵ(αc)|| and classify the testing sample x by c∗ = arg min(dc)

Fig. 2. Motivation of HD-AQR. Given a color face image, it has red, greenand blue channel. Quaternion transformation can obtain the correlation infor-mation among different color channels. The kernel quaternion transformationcan obtain the high dimension correlation information among different colorchannels.

A. Motivation and Objective Function

To better describe the motivation, we give an exampleof color face images shown in Fig. 2. A color face imagehas red, green and blue channels. The quaternion can obtainstructural correlation information among different color chan-nels. [18], [25], [33] presented high dimension informationin kernel space are helpful for sparse. Thus, we want todesign a kernel quaternion transformation that can obtainthe high dimension correlation information among differentcolor channels. Suppose that there is a mapping function,Φ(.) : Rq → RQ(q << Q) , which maps a vector-basedquaternion x and a matrix-based quaternion X to the high

dimensional feature space as x→ Φ(x) X → Φ(X) . Usingthis mapping function, the objective function of HD-QAR isas follows.

min ||α||p s.t. Φ(x) = Φ(X)α (33)

where 1 ≤ p ≤ 2.

B. Two kernel-based quaternion operations

Given a matrix-based quaternion V = V0 +V1i+V2j+V3k, the high dimension matrix quaternion is φ(V ) = φ(V0) +φ(V1)i+ φ(V2)j + φ(V3)k . Given a vector-based quaternionv = v0+v1i+v2j+v3k , the high dimension matrix quaternionis φ(v) = φ(v0) + φ(v1)i+ φ(v2)j + φ(v3)k . Because φ(V )and φ(v) contain the complex number, we need to design twooperations to transform φ(V ) and φ(v) to the real numbersuch that the optimization problem in Eq.(33) can be computedeasily. These operations are

<φ(V ) :=

φ(V0) −φ(V1) −φ(V2) −φ(V3)φ(V1) φ(V0) −φ(V3) φ(V2)φ(V2) φ(V3) φ(V0) −φ(V1)φ(V3) −φ(V2) φ(V1) φ(V0)

(34)

and

ℵφ (v) =[φ(vT0 ) φ(vT1 ) φ(vT2 ) φ(vT3 )

]T(35)

Page 6: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

6

Algorithm 2 Proposed High-Dimension Quaternion-based Adaptive Representation (HD-QAR) MethodRequire: The original training set X with L samples and a testing sample x. Initialize the value of S, s, α,m1,M2, u, ρ, ε

and umaxEnsure: Label of the testing sample x.

1: Use X and x to constitute quaternion-based kernel matrix and kernel vector as

K =

k(ℵφ(x1),ℵφ(x1)

)... k

(ℵφ(x1),ℵφ(xL)

)... ... ...

k(ℵφ(xL),ℵφ(x1)

)... k

(ℵφ(xL),ℵφ(xL)

) k =

[k(ℵφ(x1),ℵφ(x)

)... k

(ℵφ(xL),ℵφ(x)

) ]2: Use K and k to constitute the objective function as

minS,s,γ||S||∗ + ||s||1 s.t. s = k− Kγ S = KDiag(γ)

3: Transform the objective function to a Lagrange multiplier problem as

L(S, s, γ) = λ||S||∗ + ||s||1 + yT1 (k −Kγ − s) + Tr[Y T2 (S −KDiag(γ))] +u

2(||k −Kγ − s||22 + ||S −KDiag(γ)||2F )

4: while until convergence do5: Update S by fixing the others with S∗ = arg min

S

λu ||S||∗ + 1

2 ||S − (KDiag(γ)− 1uM2)||2F

6: Update γ by fixing the others with

γ∗ = (KTK +Diag(diag(KTK)))−1KT (1

um1+K(k−s))+(KTK +Diag(diag(KTK)))−1diag(KT (

1

uM2+S))

7: Update s by fixing the others with s∗ = arg mins

1u ||s||1 + 1

2 ||s− (x−Kγ + 1um1)||2

8: The multipliers m1,M2 are updated by m1 = m1 + u(k −Kγ − s) and M2 = M2 + u(S −KDiag(γ))

9: Update the parameter u by u = min(ρu, umax)

10: Check convergence by ||k −Kγ − s||∞ ≤ ε and ||S −KDiag(γ)||∞ ≤ ε11: end while12: Compute the distance dc = ||k−Kcαc||

||αc|| and classify the testing sample x by c∗ = arg min(dc)

C. Solution of objective function

Using the proposed two kernel-based quaternion operations,the proposed objective function of HD-QAR can be repre-sented as.

min ||<φ(X)Diag(γ)||∗s.t. ℵφ (x) = <φ(X)γ

(36)

Because the product of high-dimension quaternion are non-communicative, solving the high-dimension quaternion-basedoptimization problem in HD-QAR is quite challenging. Thus,we propose quaternion-based kernel matrix K and quaternion-based kernel vector k as follows. Suppose that K =<φ(X)T<φ(X) and k = <φ(X)Tℵφ(x), which can becomputed as follows:

K = <φ(X)T<φ(X)

=

k(ℵφ(x1),ℵφ(x1)

)... k

(ℵφ(x1),ℵφ(xL)

)... ... ...

k(ℵφ(xL),ℵφ(x1)

)... k

(ℵφ(xL),ℵφ(xL)

)(37)

and

k = <φ(X)Tℵφ(x)=[k(ℵφ(x1),ℵφ(x)

)... k

(ℵφ(xL),ℵφ(x)

) ] (38)

Using the Eqs.(37) and (38), the Eq.(36) is rewritten as

minS,s,γ||S||∗ + ||s||1

s.t. s = k− KγS = KDiag(γ)

(39)

The minimization problem in Eq.(39) is then transformed intothe following augmented Lagrange multiplier problem as

L(S, s, γ) = λ||S||∗ + ||s||1 + yT1 (k −Kγ − s)+ Tr[Y T2 (S −KDiag(γ))]+ u

2 (||k −Kγ − s||22 + ||S −KDiag(γ)||2F )(40)

where u > 0 is a parameter, m1 and M2 are Lagrangemultipliers. Similar to the proposed QAR, variables S, s and γin Eq.(40) can be optimized alternatively when the other twovariables are fixed.Update S when s and γ are fixed, which is equivalent to solvethe minimization problem

S∗ = arg minE

L(S, s, γ)

= arg minSλ||S||∗ + Tr(MT

2 E) + u2 ||S −KDiag(γ)||2F

= arg minS

λu ||S||∗ + 1

2 ||S − (KDiag(γ)− 1uM2)||2F

(41)

Page 7: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

7

(a) (b)

(c) (d)

(e) (f)

Fig. 3. Several color images from the (a) AR database, (b) Caltech database, (c) colorFERET database, (d) SCface database, (e) CMU-PIE database and (f)LFW database.

TABLE IIITHE PROPOSE METHODS VS WELL-KNOWN SPARSE METHODS.

Methods Objective Function Solved minimization problemSRC min ||α||1 s.t. x = Xα e1-norm minimizationCRC min ||α||2 s.t. x = Xα e2-norm minimizationKSRC min ||α||1 s.t. φ(x) = φ(X)α high-dimension-based e1-norm minimizationKCRC min ||α||2 s.t. φ(x) = φ(X)α high-dimension-based e2-norm minimizationQSRC min ||α||1 s.t. x = Xα quaternion-based e1-norm minimizationQCRC min ||α||2 s.t. x = Xα quaternion-based e2-norm minimizationQAR min ||Xα||∗ s.t. x = Xα quaternion-based ep-norm (1 < p < 2) minimizationHD-QAR min ||φ(X)α||∗ s.t. φ(x) = φ(X)α high-dimension-quaternion-based ep-norm (1 < p < 2) minimization

We use the singular value thresholding (SVT) operator [31]to approximately solve the minimization problem in Eq.(41).Afterwards, update γ when s and S are fixed, which can besolved by the following minimization problem

γ∗ = arg minγL(S, s, γ)

= arg minγ−mT

1Kγ − Tr(MT2 KDiag(γ))

+u2 (γTKTKγ − 2(k − s)TKγ − 2Tr(STKDiag(γ))

+Tr((KDiag(γ))TKDiag(γ)))= arg min

γ

u2 γ

T (KTK +Diag(diag(KTK)))γ−(KTm1 + uKTK(k − s) + diag(MT

2 K + uSTK))T γ(42)

The above minimization problem can be easily solved by

γ∗ = (KTK +Diag(diag(KTK)))−1KT ( 1um1 +K(k − s))

+(KTK +Diag(diag(KTK)))−1diag(KT ( 1uM2 + S))

(43)Next, update s when S and γ are fixed, which can be solvedas follows.

s∗ = arg minsL(S, s, γ)

= arg mins||s||1 −mT

1 s+ u2 ||k −Kγ − s||

2F

= arg mins

1u ||s||1 + 1

2 ||s− (k −Kγ + 1um1)||22

(44)

We use the soft thresholding (shrinkage) operator [32] to solvethe solution of the minimization problem in Eq.(44). After

updating S, s and γ, the multipliers m1,M2 are updated by

m1 = m1 + u(k −Kγ − s)M2 = M2 + u(S −KDiag(γ))

(45)

Then, update u by u = min(ρu, umax), where ρ, umax are twoconstants.

D. Classification

Using the sparse coefficient γ, we compute the distancebetween testing sample and the cth class as

dc(x) =||k −Kcαc||||αc||

(46)

The rule of HD-QAR is to select the class with the minimumone as

minc∗

dc(x) , c = 1, 2, ...,M (47)

The complete classification procedure of HD-QAR is shownin Algorithm 2.

E. Proposed methods vs Well-known methods

The main difference between the proposed methods andthe well-known methods is the way how to obtain the sparsecoefficients. Table III summarizes and compares the proposedQAR and HD-QAR with these methods.. SRC [1] solvese1-norm minimization problem. CRC [9] solves e2-norm

Page 8: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

8

minimization problem. KSRC [18], [25], [33] solve high-dimension-based e1-norm minimization problem. KCRC [26]solves high-dimension-based e2-norm minimization problem.QSRC [17] solves quaternion-based e1-norm minimizationproblem. QCRC [17] solves quaternion-based e2-norm mini-mization problem. The proposed QAR solves quaternion-basedep-norm (1 < p < 2) minimization problem while the pro-posed HD-QAR solves high-dimension-quaternion-based ep-norm (1 < p < 2) minimization problem. [17] also comparesdifference between the quaternion-based sparse methods withthe group-based sparse representation classification (GSRC)methods [34]–[36], joint-based sparse representation classifi-cation (JSRC) methods [37], [38].

V. EXPERIMENT RESULTS

This section assesses the effectiveness of the proposed meth-ods. We compare the proposed methods with CRC, ECRC-M,ECRC-A, QCRC, KCRC, SRC, ESRC-M, ESRC-A, GSRC,JSRC, QSRC and KSRC on several well-known color facedatabases.

A. Face Recognition with pose, lighting and expressions vari-ations

AR face database includes 126 people (70 men and 56women) with more than 4,000 color face images. These faceimages have different illumination conditions, expressionsand occlusions (sun glasses and scarf). Each person has twosessions (2 different days). The same face images were takenin both sessions. The face images were taken at the CVCunder strictly controlled conditions. Participants may havedifferent wear (clothes, glasses, etc.), make-up, hair style, etc.Following [17], we select a subset including the images withillumination and expression variations for this experiment.For each person, we randomly choose n (1, 2, 3, 4, 5) faceimages from session 1 as training set and the face images insession 2 are utilized as testing set. To obtain the convincingresults, we randomly select the images and run this processten times. The average recognition rates of this experiment isdescribed in Fig. 4 (a), and the mean recognition rates of alln (1, 2, 3, 4, 5) are listed in Table IV.Caltech Faces Database has 27 subjects with 450 colorface images under various illumination, expression andbackgrounds. Following [17], we use the Viola-Jones facedetector [39] to detect faces. Then, n (1, 2, 3, 4, 5) faceimages of each person are selected as training set and therest face images are used as testing set. To get the convincingresults, we also randomly select the images and run thisprocess ten times. Fig. 4 (b) shows the average recognitionrates of this experiment. Table V lists mean recognition ratesof all n (1, 2, 3, 4, 5).Color FERET (Facial Recognition Technology) Databasehas 119 subjects with 14, 126 face images. Under a semi-controlled environment, these images are collected in 15sessions. Following [17], the frontal face images of the setswith the letter code fa and fb are utilized for experiments.The Viola-Jones face detector [39] is used to detect faces. Inthe experiment, we select p (20, 30, 40, 50, 60) percent of

TABLE IVTHE RECOGNITION RATES OF SEVERAL METHODS ON AR FACE DATABASE

Methods Mean Recognition RateCRC 63.25KCRC 63.81ECRC-M 63.81ECRC-A 68.74QCRC 70.53SRC 58.21KSRC 58.84ESRC-M 66.95ESRC-A 67.65GSRC 69.89JSRC 69.56QSRC 71.03QAR 72.42HD-QAR 72.87

TABLE VTHE RECOGNITION RATES OF SEVERAL METHODS ON CALTECH FACE

DATABASE

Methods Mean Recognition RateCRC 83.60KCRC 84.09ECRC-M 85.40ECRC-A 86.00QCRC 88.87SRC 83.07KSRC 83.47ESRC-M 86.77ESRC-A 86.20GSRC 86.68JSRC 86.95QSRC 88.52QAR 90.29HD-QAR 90.70

face images of each person as training set, and the rest faceimages are used as testing set. To get the convincing results,we also randomly select the images and run this process tentimes. The average recognition rates of this experiment areshown in Fig. 4 (c). The mean recognition rates of all p (20,30, 40, 50, 60) are listed in Table VI.Considering the Tables IV, V and VI, we conclude threepoints as follows.

• Due to the fact that the quaternion can obtain the cor-relation information among different channels of colorimages, the quaternion-based sparse methods obtain thebetter performance compared to the non-quaternion-basedsparse methods.

• The high dimension feature in kernel space can help thesparse-based methods obtain better performance. Thus,the kernel-based sparse methods outperform the corre-sponding linear-based sparse methods.

• The adaptive ep-norm (1 ≤ p ≤ 2) minimization canbenefit from both e1-norm minimization and e2-normminimization according to the correction of the samplesin the prototype set X . Therefore, the proposed methodsobtain better performance than QCRC and QSRC in [17].

Page 9: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

9

(a) (b)

(c) (d)

Fig. 4. The recognition rates of several methods on the (a) AR database, (b) Caltech database, (c) colorFERET database and (d) SCface database.

TABLE VITHE RECOGNITION RATES OF SEVERAL METHODS ON COLORFERET FACE

DATABASE

Methods Mean Recognition RateCRC 70.24KCRC 70.76ECRC-M 72.17ECRC-A 72.49QCRC 74.82SRC 70.73KSRC 71.27ESRC-M 72.92ESRC-A 74.01GSRC 73.77JSRC 73.43QSRC 75.14QAR 76.97HD-QAR 77.39

B. Face Recognition with low resolution

In this section, we use the SCface (Surveillance CamerasFace) to assess the effectiveness of the proposed methods forface recognition with low resolution. The SCface database isa dataset of static face images of human. Face images werecollected in uncontrolled indoor environment with 5 videosurveillance cameras of various qualities. This dataset includes130 persons with 4160 face images (in visible and infraredspectrum) from different quality cameras mimic the real-world

conditions. The sizes of face images are various. Due to someof the surveillance face images are of quite low resolutionand quality, this dataset is quite challenging. Following [17],we randomly choose 2080 face images of 130 persons in thisexperiment. Then, p (10, 20, 30, 40, 50) percent of face imagesof each person are used as training set, and the rest face imagesare used as testing set. To achieve the convincing results, wealso randomly select the images and run this process ten times.The average recognition rates of this experiment are plottedin Fig. 4 (d) and the mean recognition rates of all p (10, 20,30, 40, 50) are listed in Table VII. From this table, we can getthe following conclusions. Quaternion-based sparse methodsoutperform the non-quaternion-based sparse methods. Kernel-based sparse methods have better performance than linear-based sparse methods. The adaptive ep-norm (1 < p < 2)minimization can benefit from both e1-norm minimizationand e2-norm minimization according to the correction of thesamples in the prototype set X . Therefore, the proposedmethods obtain the better performance than QCRC and QSRC

C. Face Recognition with complicated background

This section uses the CMU PIE face database to evaluatethe effectiveness of the proposed methods for face recognitionwith complicated background. CMU PIE database includesmore than 40,000 facial images of 68 people. Using theCMU (Carnegie Mellon University) 3D Room, this dataset is

Page 10: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

10

(a) (b)

Fig. 5. The recognition rates of several methods on the (a) LFW face database with normal image and (b) LFW face database with noise image.

TABLE VIITHE RECOGNITION RATES OF SEVERAL METHODS ON SCFACE FACE

DATABASE

Methods Mean Recognition RateCRC 62.46KCRC 63.16ECRC-M 63.18ECRC-A 64.47QCRC 65.53SRC 60.96KSRC 61.32ESRC-M 62.02ESRC-A 62.48GSRC 63.50JSRC 63.53QSRC 64.16QAR 67.15HD-QAR 67.72

collected by imaging each person across 13 different poses,under 43 different illumination conditions, and with fourdifferent expressions. This dataset is quite challenging becauseall face images have the various and complicated background.The experiment is implemented as follows: one-third imagesof expressions of each person are selected to form training set,while the rest two third images are considered as test images.The experiment results of comparison methods are listed inTable VIII. We can observe that the proposed methods obtainthe better performance than other state-of-the-art methods.

D. Face Recognition in the Wild

This section uses the LFW (Labeled Faces in the Wild)to assess the performance of the proposed methods forface recognition against environment in the wild. LFW facedatabase has 5749 persons with 13233 color face images takenin real and unconstrained environments. For each person, theface images have large visual variations in illumination, pose,occlusion, etc. This database is quite challenging becausethe face images of each person have the large variations inunconstrained environments. Following [17], we choose thepersons with over n (40,60, 80) face images in this experiment.

TABLE VIIITHE RECOGNITION RATES OF SEVERAL METHODS ON CMUPIE FACE

DATABASE

Methods Mean Recognition RateCRC 75.12KCRC 76.10ECRC-M 79.42ECRC-A 78.10QCRC 83.84SRC 77.27KSRC 78.04ESRC-M 84.12ESRC-A 83.79QSRC 85.04QAR 86.27HD-QAR 86.79

Then we randomly select 10 percent of the face imagesper person as training set and the rest images are used astesting set. To get the convincing results, we also randomlyselect the images and run this process ten times. The averagerecognition rates of the comparison methods are listed inFig. 5 (a). Observing this figure, we know that the proposedmethods could improve the performance of other comparisonmethods and obtain the higher recognition rate, in spite ofthe face images having the large variations in unconstrainedenvironments.

E. Face recognition with noise image

In this subsection, we use the LFW (Labeled Faces in theWild) face database to test the robustness of the proposedmethods against noise. The training sets are same as thoseof the above section. The testing set is obtained by Matlabfunction ”imnoise” to insert the Gaussian white noise. Noticethat only the testing face images are inserted with the Gaussianwhite noise, and the training face images are utilized theoriginal images of the database. To get the convincing results,we also randomly select the images and run this processten times. The average recognition rates of the comparisonmethods are listed in Fig. 5 (b). Observing this figure, we know

Page 11: Color Face Recognition using High-Dimension Quaternion-based … · 2017. 12. 6. · used it to assess the quality of color images. [16] pro-posed a vector sparse representation using

11

that the proposed methods could outperform other comparisonmethods and obtain the higher recognition rates for faceimages with the noise and large variations in unconstrainedenvironments.

VI. CONCLUSION

In this paper, quaternion-based adaptive representation(QAR) was proposed for color face recognition. QAR canadaptively combine the e1-norm and e2-norm and benefit fromboth. Moreover, we proposed the high-dimension quaternion-based adaptive representation (HD-QAR), which can obtainthe high dimension feature information in the kernel space,Thus, HD-QAR can obtain the correlation information amongthe different channels. Experimental results showed that theproposed methods achieve better performance compared toseveral state-of-the-art methods. The analyses and the experi-mental results on three databases confirmed the effectivenessof the proposed methods for color face recognition.

REFERENCES

[1] J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan, “Sparserepresentation for computer vision and pattern recognition,” Proceedingsof the IEEE, vol. 98, no. 6, pp. 1031–1044, 2010.

[2] Q. Feng, C. Yuan, J.-S. Pan, J.-F. Yang, Y.-T. Chou, Y. Zhou, and W. Li,“Superimposed sparse parameter classifiers for face recognition,” 2016.

[3] Z. Lai, W. K. Wong, Y. Xu, J. Yang, and D. Zhang, “Approximate orthog-onal sparse embedding for dimensionality reduction,” IEEE transactionson neural networks and learning systems, vol. 27, no. 4, pp. 723–735,2016.

[4] Q. Feng, C. Yuan, J. Huang, and W. Li, “Center-based weighted kernellinear regression for image classification,” in Image Processing (ICIP),2015 IEEE International Conference on. IEEE, 2015, pp. 3630–3634.

[5] J.-S. Pan, Q. Feng, L. Yan, and J.-F. Yang, “Neighborhood feature linesegment for image classification,” T-CSVT, vol. 25, no. 3, pp. 387–398,2015.

[6] Y. Xu, X. Fang, J. Wu, X. Li, and D. Zhang, “Discriminative transfersubspace learning via low-rank and sparse representation,” IEEE Trans-actions on Image Processing, vol. 25, no. 2, pp. 850–863, 2016.

[7] Q. Feng, J.-S. Pan, and L. Yan, “Two classifiers based on nearest featureplane for recognition,” in 2013 IEEE International Conference on ImageProcessing. IEEE, 2013, pp. 3216–3219.

[8] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robustface recognition via sparse representation,” T-PAMI, vol. 31, no. 2, pp.210–227, 2009.

[9] L. Zhang, M. Yang, and X. Feng, “Sparse representation or collaborativerepresentation: Which helps face recognition?” in ICCV. IEEE, 2011,pp. 471–478.

[10] S.-C. Pei, J.-J. Ding, and J.-H. Chang, “Efficient implementation ofquaternion fourier transform, convolution, and correlation by 2-d com-plex fft,” IEEE Transactions on Signal Processing, vol. 49, no. 11, pp.2783–2797, 2001.

[11] T. A. Ell and S. J. Sangwine, “Hypercomplex fourier transforms of colorimages,” IEEE Transactions on image processing, vol. 16, no. 1, pp. 22–35, 2007.

[12] D. S. Alexiadis and G. D. Sergiadis, “Estimation of motions in colorimage sequences using hypercomplex fourier transforms,” IEEE Trans-actions on Image Processing, vol. 18, no. 1, pp. 168–187, 2009.

[13] L. Jin, H. Liu, X. Xu, and E. Song, “Quaternion-based impulse noiseremoval from color video sequences,” IEEE Transactions on Circuitsand Systems for Video Technology, vol. 23, no. 5, pp. 741–755, 2013.

[14] A. Kolaman and O. Yadid-Pecht, “Quaternion structural similarity: a newquality index for color images,” IEEE Transactions on Image Processing,vol. 21, no. 4, pp. 1526–1536, 2012.

[15] Q. Barthelemy, A. Larue, and J. I. Mars, “Color sparse representationsfor image processing: review, models, and prospects,” IEEE Transactionson Image Processing, vol. 24, no. 11, pp. 3978–3989, 2015.

[16] Y. Xu, L. Yu, H. Xu, H. Zhang, and T. Nguyen, “Vector sparserepresentation of color image using quaternion matrix analysis,” IEEETransactions on Image Processing, vol. 24, no. 4, pp. 1315–1329, 2015.

[17] C. Zou, K. I. Kou, and Y. Wang, “Quaternion collaborative and sparserepresentation with application to color face recognition,” IEEE Trans-actions on Image Processing, vol. 25, no. 7, pp. 3287–3302, 2016.

[18] S. Gao, I. W.-H. Tsang, and L.-T. Chia, “Kernel sparse representation forimage classification and face recognition,” in ECCV. Springer, 2010,pp. 1–14.

[19] W. R. Hamilton, “Ii. on quaternions; or on a new system of imaginariesin algebra,” Philosophical Magazine Series 3, vol. 25, no. 163, pp. 10–13, 1844.

[20] Y. Xu, D. Zhang, J. Yang, and J.-Y. Yang, “A two-phase test samplesparse representation method for use with face recognition,” T-CSVT,vol. 21, no. 9, pp. 1255–1262, 2011.

[21] J. Wang, C. Lu, M. Wang, P. Li, S. Yan, and X. Hu, “Robust facerecognition via adaptive sparse representation,” IEEE transactions oncybernetics, vol. 44, no. 12, pp. 2368–2378, 2014.

[22] E. Grave, G. R. Obozinski, and F. R. Bach, “Trace lasso: a trace normregularization for correlated designs,” in Advances in Neural InformationProcessing Systems, 2011, pp. 2187–2195.

[23] Q. Feng and Y. Zhou, “Kernel regularized data uncertainty for actionrecognition,” IEEE Transactions on Circuits and Systems for VideoTechnology, vol. 27, no. 3, pp. 577–588, 2017.

[24] ——, “Kernel combined sparse representation for disease recognition,”IEEE Transactions on Multimedia, vol. 18, no. 10, pp. 1956–1968, 2016.

[25] S. Gao, I. W. Tsang, and L.-T. Chia, “Sparse representation withkernels,” T-IP, vol. 22, no. 2, pp. 423–434, 2013.

[26] B. Wang, W. Li, N. Poh, and Q. Liao, “Kernel collaborativerepresentation-based classifier for face recognition,” in ICASSP. IEEE,2013, pp. 2877–2881.

[27] G. H. Golub and C. F. Van Loan, Matrix computations. JHU Press,2012, vol. 3.

[28] R. A. Horn and C. R. Johnson, Matrix analysis. Cambridge universitypress, 2012.

[29] E. J. Candes, X. Li, Y. Ma, and J. Wright, “Robust principal componentanalysis?” Journal of the ACM (JACM), vol. 58, no. 3, p. 11, 2011.

[30] G. Liu, Z. Lin, and Y. Yu, “Robust subspace segmentation by low-rankrepresentation,” in Proceedings of the 27th international conference onmachine learning (ICML-10), 2010, pp. 663–670.

[31] J.-F. Cai, E. J. Candes, and Z. Shen, “A singular value thresholding al-gorithm for matrix completion,” SIAM Journal on Optimization, vol. 20,no. 4, pp. 1956–1982, 2010.

[32] E. T. Hale, W. Yin, and Y. Zhang, “Fixed-point continuation for\ell 1-minimization: Methodology and convergence,” SIAM Journal onOptimization, vol. 19, no. 3, pp. 1107–1130, 2008.

[33] L. Zhang, W.-D. Zhou, P.-C. Chang, J. Liu, Z. Yan, T. Wang, and F.-Z.Li, “Kernel sparse representation-based classifier,” T-SP, vol. 60, no. 4,pp. 1684–1695, 2012.

[34] J. Huang, T. Zhang et al., “The benefit of group sparsity,” The Annalsof Statistics, vol. 38, no. 4, pp. 1978–2004, 2010.

[35] J. Huang, T. Zhang, and D. Metaxas, “Learning with structured sparsity,”Journal of Machine Learning Research, vol. 12, no. Nov, pp. 3371–3412,2011.

[36] X. Lv, G. Bi, and C. Wan, “The group lasso for stable recovery of block-sparse signal representations,” IEEE Transactions on Signal Processing,vol. 59, no. 4, pp. 1371–1382, 2011.

[37] O. Lee, J. M. Kim, Y. Bresler, and J. C. Ye, “Compressive diffuse opticaltomography: noniterative exact reconstruction using joint sparsity,” IEEEtransactions on medical imaging, vol. 30, no. 5, pp. 1129–1142, 2011.

[38] J. M. Kim, O. K. Lee, and J. C. Ye, “Compressive music: Revisiting thelink between compressive sensing and array signal processing,” IEEETransactions on Information Theory, vol. 58, no. 1, pp. 278–301, 2012.

[39] P. Viola and M. J. Jones, “Robust real-time face detection,” Internationaljournal of computer vision, vol. 57, no. 2, pp. 137–154, 2004.