Dimension Reduction for the Conditional Mean in Regression

29
DIMENSION REDUCTION FOR CONDITIONAL MEAN IN REGRESSION BY R. DENNIS COOK AND BING LI University of Minnesota and Pennsylvania State University In many situations regression analysis is mostly concerned with inferring about the conditional mean of the response given the predictors, and less concerned with the other aspects of the conditional distribution. In this paper we develop dimension reduction methods that incorporate this consideration. We introduce the notion of the Central Mean Subspace (CMS), a natural inferential object for dimension reduction when the mean function is of interest. We study properties of the CMS, and develop methods to estimate it. These methods include a new class of estimators which requires fewer conditions than pHd, and which displays a clear advantage when one of the conditions for pHd is violated. CMS also reveals a transparent distinction among the existing methods for dimension reduction: OLS, pHd, SIR, and SAVE. We apply the new methods to a data set involving recumbent cows. 1. Introduction. Empirical evidence accumulated during the past few years indi- cates that recently developed dimension-reduction methods can be quite effective for constructing regression summary plots in a largely non-parametric context. In full gener- ality, the goal of a regression is to infer about the conditional distribution of the univariate response variable given the vector of predictors : How does the conditional distribution of change with the value assumed by ? The dimension-reduction methods approach this question through a population meta-parameter called the central Supported in part by National Science Foundation grants DMS-9703777 and DMS-0103983. Supported in part by National Science Foundation grant DMS-9626249. AMS 1991 subject classifications. Primary Key words and phrases. Central subspace, Graphics, Regression, pHd, SAVE, SIR, Visualization. 1

Transcript of Dimension Reduction for the Conditional Mean in Regression

Page 1: Dimension Reduction for the Conditional Mean in Regression

DIMENSION REDUCTION FOR CONDITIONAL MEAN IN REGRESSION

BY R. DENNIS COOK�

AND BING LI�

University of Minnesota and Pennsylvania State University

In many situations regression analysis is mostly concerned with inferring

about the conditional mean of the response given the predictors, and less concerned

with the other aspects of the conditional distribution. In this paper we develop

dimension reduction methods that incorporate this consideration. We introduce

the notion of the Central Mean Subspace (CMS), a natural inferential object for

dimension reduction when the mean function is of interest. We study properties of

the CMS, and develop methods to estimate it. These methods include a new class

of estimators which requires fewer conditions than pHd, and which displays a clear

advantage when one of the conditions for pHd is violated. CMS also reveals a

transparent distinction among the existing methods for dimension reduction: OLS,

pHd, SIR, and SAVE. We apply the new methods to a data set involving recumbent

cows.

1. Introduction. Empirical evidence accumulated during the past few years indi-

cates that recently developed dimension-reduction methods can be quite effective for

constructing regression summary plots in a largely non-parametric context. In full gener-

ality, the goal of a regression is to infer about the conditional distribution of the univariate

response variable�

given the ����� vector of predictors � : How does the conditional

distribution of� � change with the value assumed by � ? The dimension-reduction

methods approach this question through a population meta-parameter called the central

�Supported in part by National Science Foundation grants DMS-9703777 and DMS-0103983.�Supported in part by National Science Foundation grant DMS-9626249.

AMS 1991 subject classifications. Primary

Key words and phrases. Central subspace, Graphics, Regression, pHd, SAVE, SIR, Visualization.

1

Page 2: Dimension Reduction for the Conditional Mean in Regression

subspace which is denoted by ����� � . The central subspace is the smallest subspace of

� ��such that is independent of � given �� �� , where the columns of the matrix �

form any basis for the subspace. Knowledge of the central subspace is useful for parsi-

moniously characterizing how the distribution of �� � changes with the value of � . If

����� � is known, the minimal sufficient summary plot of versus �� �� can be used to

guide subsequent analysis. If an estimated basis �� of ����� � is available then the summary

plot of versus �� � can be used similarly.

Summary plots based on estimates of ����� � can be of significant value in many phases

of a regression analysis, particularly during the initial phases when an adequate parsi-

moniously parameterized model is not yet available. Methods of estimating the central

subspace or portions thereof include ordinary least squares (OLS), graphical regression

(Cook 1994a), principal Hessian directions (pHd, Li 1992), sliced average variance esti-

mation (SAVE, Cook and Weisberg 1991), sliced inverse regression (SIR, Li 1991) and

parametric inverse regression (PIR, Bura and Cook 2001). Cook and Weisberg (1999)

gave an introductory account of studying regressions via central subspaces. A compre-

hensive discussion is available in Cook (1998a).

While the central subspace is designed to give a complete picture of the dependence of

on � , certain characteristics of �� � may often be of special interest. Indeed, regression

is understood by some to imply a study of the mean function ������ ��� . Pursuing the mean

function through the central subspace can be inefficient because the scope of the statistical

inquiry may be much larger than necessary.

In Section 2 we introduce the central mean subspace (CMS) and study its properties.

The construction of a CMS is similar in spirit to that for a central subspace, but dimension

reduction is aimed at reducing the mean function alone, leaving the rest of �� � as the

“nuisance parameter.”

2

Page 3: Dimension Reduction for the Conditional Mean in Regression

In Section 3 we connect the central mean subspace to literature on the central subspace.

In particular, we show that, in the population, many known techniques for constructing

vectors in the central subspace in fact produce vectors in the central mean subspace.

These results provide an important distinction between known methods – OLS, pHd, SIR

and SAVE – for estimating vectors in the central subspace ����� � . And they imply that in

the first instance we can use two known methods – OLS and pHd – to estimate vectors

in the central mean subspace.

In Section 4 we describe a new class of methods for estimating vectors in the CMS.

This class may be useful because it requires fewer constraints than pHd on the distribution

of the predictors. The new method is applied in Section 6. To avoid interrupting the

development, we have placed most of the proofs in the appendix.

2. Mean dimension-reduction subspaces and their properties. Consider a regres-

sion consisting of a univariate response variable � and a ��� vector of random predictors

�. We assume throughout this article that the data ������������ , �������������� � , are iid ob-

servations on !"�#� �%$with finite moments. We use �&!"' $

to denote a general subspace

of ( )+* , where ' is a �,�.- matrix whose columns form a basis in �&!"' $. We use �0/21 3

to denote a dimension-reduction subspace, where the subscript indicates what response

and predictors are involved, and whether the whole conditional distribution or only the

conditional mean is of interest. In some instances we use �&!24���' $, where 4 is a �,�5

vector, to denote the subspace of ( )&* spanned by 4 and the columns of ' .

2.1. Overview of central subspaces. A dimension-reduction subspace (Li 1991) for

the regression of � on�

is any subspace �6!27 $such that

� �98 7;: � �(2.1)

3

Page 4: Dimension Reduction for the Conditional Mean in Regression

where denotes independence and � denotes a ����� matrix with ����� . The statement is

thus that is independent of given ��� . It is equivalent to saying that the distribution

of �� is the same as that of �� ���� for all values of . It implies that the ����� predictor

vector can be replaced by the ����� predictor vector ���� without loss of regression

information, and thus represents a potentially useful reduction in the dimension of the

predictor vector. When the intersection of all dimension-reduction subspaces is itself a

dimension-reduction subspace it is called the central subspace (Cook 1994b,1998a) and

denoted by ����� � . The central subspace is assumed to exist throughout this article.

As mentioned in the Introduction, the central subspace is designed to capture the

complete conditional distribution of �� and thereby give a full picture of the dependence

of on . On the other hand, when the conditional mean �� !�� �" is of special interest,

it may be useful to adapt our inquiry to fit that more specific objective.

2.2. Central mean subspace. When focusing on the conditional mean, dimension

reduction hinges on finding a ����# matrix $ , #��%� , so that the #&�'� random vector

$(�� contains all the information about that is available from �( )�� �" . This is less

restrictive than requiring that $*�� contain all the information about that is available

from as in the current literature associated with the central subspace. The following

definition formalizes this idea.

DEFINITION 1. If

�� !�� �"+� $ �

then �� )$," is a mean dimension-reduction subspace for the regression of on .

It follows from this definition that a dimension-reduction subspace is necessarily a

mean dimension-reduction subspace, because -� $.� implies �� !�� �"/� $*�� .

4

Page 5: Dimension Reduction for the Conditional Mean in Regression

The next proposition gives equivalent conditions for the conditional independence used

in Definition 1.

PROPOSITION 1. The following statements are equivalent:

(i) � ������� ��� � ��� ,

(ii) �������������������� ������ � � ������� ,

(iii) ��� �� �� is a function of �!��� .

The first condition is the same as Definition 1. The second condition is that, given �"��� ,

� and ��� �� �� must be uncorrelated. The final condition is what might be suggested by

intuition, ��� �� �� �#��� �� �!�$�� . Any of these three conditions could be taken as the

definition of a mean dimension-reduction subspace.

Paralleling the development of central subspaces, we would like the smallest mean

dimension-reduction subspace, as formalized in the next definition.

DEFINITION 2. Let %'&)(�*�+ ,.-/�10�%$2 where intersection is over all mean dimension-

reduction subspaces % 2 . If %'&)(3*4+ ,.- is itself a mean dimension-reduction subspace, it

is called the central mean dimension-reduction subspace, or simply the central mean

subspace (CMS).

The CMS does not always exist, because the intersection of two mean dimension-

reduction subspaces is not necessarily a mean dimension-reduction subspace. However,

when it does exist, %$&)(�*4+ ,.- 56%�*7+ , because the former is the intersection of a larger

collection of subspaces. Hence it may be possible to reduce the dimension from that of

%�*�+ , if ������ �� alone is concerned. Under mild conditions, the existence and uniqueness

of the CMS can be guaranteed in the same way that the existence of the central subspace

is guaranteed. For instance, if the domain of � is open and convex, then the CMS exists

5

Page 6: Dimension Reduction for the Conditional Mean in Regression

and is unique. For location regressions where � ��� ������� �� , the central subspace and

the CMS are the same, ��� ���� ���������� � . Additional existence results are available in

such cases (Cook 1994a; 1998a, p. 111). We assume in the remainder of this article that

the CMS always exists.

The CMS is intended to play the same role when considering the conditional mean as

the central subspace does when inquiring about the full conditional distribution of ��� � .

If an estimated basis �� of ��� ���� ��� is available then the summary plot of � versus ��� �can provide a low-dimensional visualization of the mean function. Methods of estimating

��� ���� ��� are discussed later in this article.

The central subspace is invariant under one-to-one transformations ! of the response,

��"�#� �$�%���� � . This property does not extend to the CMS because ��� ���� ��� does not

in general equal ��� ��"�#� ��� , although the central subspace is an invariant upper bound on

the CMS, ��� ��"�#� ����&'���� � . Standard response transformation methodology in linear

regression exploits this flexibility by attempting to find a ! so that (*)�+,�-"�� ��"�#� ���.��0/and ���1!2�1�23� �4 is linear.

It is sometimes helpful to transform � linearly so that the transformed predictors

are uncorrelated, and then study the relation between � and the transformed predictors.

In general, if 56�87 �89;: for some invertible matrix 7 and some vector : . Then

��� ���� <=�>�?7A@CBD��� ���� ��� is the CMS for the regression of � on 5 .

In most of the subsequent developments, we work in terms of the standardized pre-

dictor

5E�GF @�BIHKJLML ���ONP���1�4IRQ

where F LML �6SUTWVX�1�� is assumed to be positive definite. In terms of this standardized

predictor, the CMS is ��� ���� �����YF @CBZHZJL3L ��� ���� <[� . The corresponding sample version �56

Page 7: Dimension Reduction for the Conditional Mean in Regression

is obtained by replacing ����� and ����� with their usual moment estimates, ������ and

�� ����� .

3. Vectors in the central mean subspace. Having established some basic properties

of the CMS, we now turn our attention to finding population vectors in that subspace.

We will survey available methods for constructing vectors in the central subspace and

demonstrate that some of them in fact produces vectors in CMS. By categorizing and

assessing these methods in relation to CMS, we set the stage for a new estimation method

introduced in Section 4.

3.1. Objective functions and OLS. In their pioneering article Li and Duan (1989)

demonstrated that a class of estimators, which includes OLS, correctly estimate the di-

rection of the regression parameter regardless of the shape of the regression function,

provided that the predictor satisfies a linearity condition. They considered an objective

function of the form ���������������� ����� �!��"$#%�'&�)( where �+*-, .0/ and �1*, .%2 . Here, the

expectation is with respect to the joint distribution of & and # , and ����34�'&0 is strictly

convex in 3 . This use of an objective function is not meant to imply that any associated

model is true or even provides an adequate fit of the data. Nevertheless, there is a useful

connection between 57698;:=< >@? and the vectors derived from these objective functions.

Let

��A=�CB���EDGF'H�I�J;KLNM O ����������(3.2)

denote the population minimizers, and let P be a basis matrix for 5�:=< > . Li and Duan

(1989) showed, in effect, that if ���Q#�R P�"S#= is linear in # and TUJVIW�X5$:Y< >S+�[Z , then

B�*\5$:Y< > . From this it is straightforward to relax the dimension restriction: if ���Q#�R P�"S#=is linear in # then B]*\5S:Y< > (Cook 1998a, p. 143–147).

7

Page 8: Dimension Reduction for the Conditional Mean in Regression

Where � falls in relation to �������� �� depends on the choice of : for some choices

� belongs to ��������� �� while for others it may belong to ���� ����������� �� . However, if we

restrict attention to objective functions

������������ �"!$#&%(' !)�*�+�������,#-�/.����0�����1�,#(3.3)

based on the natural exponential family for some strictly convex function . , then �always belongs to �����2�,� �� :

THEOREM 1. Let 3 be a basis matrix for �������� �� , assume that 45�6�57 3 � �# is a

linear function of � and let � be as defined in (3.2) using an exponential family objective

function (3.3). Then �98:�����2�,� �� .

The exponential family objective function (3.3) covers many estimation methods used

in practice. In particular, OLS is obtained by setting .��*;<#&%=;?>A@CB . For future reference

we denote the OLS coefficient vector 45�*!D�# by � EGF .

3.2. SIR and SAVE. It is easy to see that SIR (Li 1991) and SAVE (Cook and

Weisberg 1991) can find vectors in ���,� H�I�������� �� : Suppose that ���� is spanned by

the columns of the matrix J and K�L is the projection onto ���� . Then, 45�6�57 !M#H%45N 45�O��7 !,�"J � �,#P7 !RQ�%S45N 45�O��7 J � �,#P7 !MQ-%TK L 4��O��7 !D#U� so that 45�O��7 !$# belongs to ����� .

However, if we replace J by 3 , the basis for �-������ �� , then the second equality need

not hold because we do not necessarily have the conditional independence ! ��7 3 � � .

Hence, in general, 45�O��7 !$# is in ���� but not necessarily in �����2�,� �� . The same is true

of the vectors derived from SAVE.

3.3. V -based pHd. Li’s (1992) V -based method of principal Hessian directions (pHd)

is based on using the third moment matrix W EGFXF %=4 YZ�*!T'?4���!$#"#"��� �,[ to infer about

8

Page 9: Dimension Reduction for the Conditional Mean in Regression

����� �, assuming normally distributed predictors. However, as described in the next theo-

rem, pHd in fact targets the CMS.

THEOREM 2. Let � be a basis matrix for������� ��

. If ������� ������� is a linear function

of � and if ����������� ������� is uncorrelated with , then� �"!$#&%('*) #&%+% �-, �.�/��� ��

.

The requirement that ������� ���0��� be linear implies !21 ���/�-� �3 as an application of

Theorem 1. Both conditions stated in this theorem are used to obtain the conclusion that

� ��) #4%5% �-, �.����� �� .

Cook (1998b, Theorem 1) showed that the inference procedure proposed by Li (1992)

is not as straightforward as originally claimed. However, the inference procedure simpli-

fies greatly if 6�798:�; �'5���=<2> . This condition will not normally hold, but is guaranteed

to hold if we replace with population OLS residuals ?@<A CBD���; E�.BF! �#4%9� . In short,

we can use the relatively straightforward inference procedure suggested by Cook (1998b)

for inferring about the related subspace� ��)HG %5% � , where )IG %5% is constructed as ) #4%5% ,

except that is replaced with the residual ? :

) G %5% <J���"?(�-� � �4K(3.4)

For this to be useful, we must understand the relationship between���/�-� ��

, the subspace

we would like to know, and��� G � �� , the CMS that we can infer about easily. This is the

topic of the next proposition.

Recall that � is the basis matrix for���/��� ��

, and that !L#4%M<C���; N��� . Define a residual

?@<C?O�P �'Q! �#4%9����<JR�"! �#4%9����SUT��V! �#&%����W (3.5)

to be a function of and ! �#&% � that is linear in . For example, ?X<A YBZ���; E��B[! �#4% �satisfies these conditions as does the Pearson residual computed from a logistic regression

model.

9

Page 10: Dimension Reduction for the Conditional Mean in Regression

PROPOSITION 2. Assume that ������� ����� is linear. Then

� ������� ������ ������� ������� ���! #"$&%(3.6)

where the summation on the right means the collection of vectors of the form � � ��'with � in

� �����(� ���and � ' in

� ���) #"$ .

Proposition 2 is useful because it allows us to infer about�*�����(� ���

by inferring

about� ���) +"� and

�,�����-� ���using relatively straightforward procedures in the literature.

In particular, it follows from Theorem 2 applied to the regression of . on � that

� ��/ � "0" (1 � ������� �2�, and we can infer about

� �3/ � "0" by using the inference procedures

discussed by Cook (1998b).

Finally, assuming that �4����� � � #"5�� is linear and 687:9$����� � � #"5�� is constant, it can be

verified that

/ � "0" � / +";"4<>=@?$ACB ����� � #"$��ED5FHG&�� #"IG#JLKThus,

� �M�) #"N%0/ � "0" �O� �M�) #"L%&/ #"0" , which returns us to the regression of P on � .

4. Vectors in the CMS that require only the linear conditional means.

4.1. Population structure. The vectors in the CMS described in the last section re-

quire two essential conditions:

C.1: the conditional mean ������� �8�Q�� is linear in � ,

C.2: the conditional variance Var ����� ���Q�� is uncorrelated with P .

Predictor linearity conditions such as C.1 used in Theorem 1 are common in dimension

reduction. In general, the requirement that �4����� RS�Q�� be linear in � is equivalent to

requiring that �4����� = T �� � =,T � where =,T is the projection operator for� �URS with

respect to the standard inner product (Cook 1998a, p. 57). If � follows an elliptically

contoured distribution then �4�3�4� RS��� is linear for all conforming R ’s. Hall and Li

10

Page 11: Dimension Reduction for the Conditional Mean in Regression

(1993) show that such linearity will hold to a reasonable approximation in many problems.

The intuition here is that conditional expectations of the form ������� ����� become more

linear as � increases with �� �������������� ���� fixed. This is related to the work of Diaconis

and Freedman (1984) who argue that almost all low-dimensional projections of high-

dimensional data sets are nearly normal. In addition, these conditions might be induced

by using predictor transformations and predictor weighting (Cook and Nachtsheim 1994).

With condition C.1 we can use known methods based on exponential family objective

functions to estimate a vector in ��������� ��� (Theorem 1). Such methods may be sufficient

when �� �� ���!���"��� ����$#&% , but other methods are needed when �� �������������� ����(')% . To

obtain multiple vectors in �������*� ��� using + -based pHd we had to require condition C.2

(Theorem 2) which is implied when Var ����� ���,�� is constant. Both conditions are implied

when � is normally distributed, although normality is not a necessary condition. In this

section we introduce a new class of vectors in the CMS that requires only C.1. The next

theorem lies at the core of this method.

THEOREM 3. Suppose that - and . are measurable functions of ���,� and that

���/�0� �1�,�� is linear in � . Then �324��-65879.:;�3<:=>��������� ��� , provided that �/-65?79.:@�is integrable.

This theorem provides a method of forming vectors in the CMS, which in turn provides

a basis for constructing estimates of at least a portion of the CMS. We refer to these

as COZY vectors since they come from COvariances between � and and constructed

responses 5BAC#D-65879. .

To see how Theorem 3 might be used to construct vectors in the CMS, suppose we

know one vector EGF(=H�!���"�*� ��� . We can find another vector E�IJ=K�!������� ��� by choosing

11

Page 12: Dimension Reduction for the Conditional Mean in Regression

appropriate functions ��� ������and �� �����

, and then forming the covariance �� ��������� ���� between

��� ��� � ��� ��� �"! � � �#� ��� (which is an instance of the constructed

response�$�

) and the standardized predictor vector � . This process can then be iterated

in the hope of finding additional vectors in the CMS:

�&%'� ��(&) � � �*�%,+� ��� ��! � � �#�%,+- ���/.��10� ��(2� �%,+� �10&354 �"6 387#3:9;9<9(4.7)

where���%,+� �=� � ���%,+- ��� �>! � � �#�%,+- ��� . Of course, we would need at most ?@�

A�B<C �ED-FHG;IKJ LNM � vectors.

Two general questions remain regarding the construction of COZY vectors: How can

we find the first vector � that is needed to prime the process? And how should we

choose the functions � and � used in forming the constructed responses�O�

? We address

these issues in the next section when developing a new method of estimating the CMS

based on the COZY class.

4.2. Iterative Hessian Transformation. A first application of Theorem 3 is straight-

forward: Setting P�Q6 and R"�TS with probability 1,�U(&� P �V! R �W��0 �TX�Y,Z which we

know belongs toD[FHG;I�J L\M

under condition C.1 (Theorem 1). Since any function of X �Y,Z �is measurable with respect to ] � � , we can use this first result to construct additional

vectors in the CMS:

COROLLARY 1. Let ��� �^��_�and �Q� �`��a�

be any functions such that

( � � X �Y:Z ��� �! � � X �Y,Z ���b0c� is integrable. Assume condition C.1. Then any vector of the

form�')d( � � X �Y,Z ��� �e! � � X �Y:Z ���b0f�g. belongs to

D[FHG;I�J LNM.

To apply Corollary 1 we must choose the functions � and � . The next corollary gives

two specific COZY sets obtained by choosing (i) � �Eh � � h and � ��h � �Qi hkj�l� � , and (ii)

12

Page 13: Dimension Reduction for the Conditional Mean in Regression

���������� and ������� ����������������� . Starting with ��� ������ , we then use (4.7) to form

subsequent COZY vectors.

COROLLARY 2. Assume condition C.1. Then

(i) �! #"%$'&)(+*���,�'�����.-./��10�243%265757598.:<;>=@?BADC EGF , and

(ii) �H #"I$'&I(J*K �,� �����.-./��10�243%265757598.:<;>=@?BADC EGF ,

where ( K �,� is as defined in (3.4).

We restrict most of the following discussion to case (i); similar comments hold for case

(ii).

It follows from Corollary 2 that, unless � ��� is an eigenvector of ( ���,� , the sequence of

COZY vectors ����� , ( ���,� �L��� , ( ��6�M�>����� , ... provides a set of different vectors in ;N=@?7ADC EGF .One question that remains is how large / must be in order for the first / vectors, �O��� , ...,

( *�P>Q�6�M� � ��� , to exhaust all possible vectors in the sequence. This is important because in

practice we can compute only a finite number of these vectors. This question is answered

by the next proposition; its proof is straightforward and omitted.

PROPOSITION 3. Let R be a S�T�S matrix and � be a S -dimensional vector. If R * �belongs to the subspace spanned by � , ..., R *�P'Q � , then so does RVU � for any W.X / .

Since ( * �6�M� ����� belongs to ;Y=@?7A�C EGF , which has dimension Z , Proposition 3 implies that

there is an integer W\[]Z such that the first W vectors in the sequence, � ��� , ..., ( U P>Q���,� � �6� ,are linearly independent, and all the subsequent vectors are linearly dependent on them.

This suggests the following estimation scheme (at the population level). First, compute

the S COZY vectors � �6� , ( ���,� � ��� , ..., (.^ P'Q���,� � ��� . Let _ be the S by S matrix with

these vectors as columns and form the matrix _�_a` . If b is an eigenvector of _c_d` ,

then by Corollary 2 it is also the eigenvector of e\fg_c_c`Ye+f , where e.f is the projection

13

Page 14: Dimension Reduction for the Conditional Mean in Regression

matrix onto ��������� � , and hence �� ���������� � . Now let ����������������� be all the eigenvectors

of ����� , so ordered that their corresponding eigenvalues form a descending sequence

� �! "�����# �%$ '&)(*�����+(,&-� Then ��� is the first eigenvector in ��������� � , ��. is the

second eigenvector, and so on.

Because this method is based on the iterative transformation of /1032 by the Hessian

matrix 4 03252 , we call it the Iterative Hessian Transformation method, or IHT. Note,

however, that when 6 is non-Gaussian 4 03252 ceases to have the interpretation of the

Hessian matrix of the regression function (Cook 1998b).

The IHT directions ��7 can be back-transformed to the original scale: 8�79(:4<; �>=?.@�@ ��7 ,A (CBD�E���E�3�GF . Sample IHT directions can be constructed by replacing 4 @H@ , 4 03252 and

4JI 252 by their sample estimates, and /1032 by the vector of least squares coefficients. The

linear combinations ���7 6K(L8M�71NPO'QSRTNGO�U�U will be called the IHT predictors.

Let the columns of the matrix V form a basis for �W��� . Li (1992, theorem 6.1) showed

in effect that if RTN 6TX VW�Y6 U is linear and /Z [���\� then � N 4 0E2?2 / U ]����� . While Li’s

result is related to that of Corollary 2, there are two key differences. First, Li’s result is

in terms of the central subspace and not the more specific CMS as defined here. Second,

Li used his result to address the robustness of pHd rather than to suggest the possibility

of iterating to find additional vectors in the CMS.

4.3. Illustrations. The results of the previous section show that we may need all FCOZYvectors to be assured of exhausting �������#� � , but in practice we may not need more

than the first ^<( dim N �������#� � U of them. For instance, if ^_(a` , /b0E2 and 4 03252 /b032 could

be sufficient, as illustrated in the following example.

To provide some insight into the IHT method, consider the regression model

c (ed � � 6[fhg N d �. 6 U f�i14

Page 15: Dimension Reduction for the Conditional Mean in Regression

where � �and without loss of generality ��������� . We assume that ��� �� ������ � ,

��� �� ������ � and ��� �� ���� ���������� � are linear. Also, without loss of generality, we constrain

������and ��� ������ � to be uncorrelated.

The CMS is spanned by � � � ��� � � , and

� �"! $#&%(')�*� �+� � � �-,

���� ��. ���� � ��� ���� � �0/1 � � 1 �

� �

Next,

2 �"!3! �4�"! 5����� �6� � � � �

5��.7� � � �8� �� � / , ��. ��� � �� � � �6� � � � /

� �1 � � 1 � ��.7� � � �9� �;:+/ , �*<>=@? , <�A8B ? =DCE�;��. ��� � �� � � �6� � � � /

� �1 � � 1 � ��.7� � � � � �;:+/ ,

� �1 � � 1 � ��.7� � � � � �

���� � ���� �F/

, G =H? � �1 G = ? � � 1� ��. ��� � �� � �E� � �� G =@? � �E� � � ��� �0/(4.8)

The first two terms on the right of (4.8) are in I�� ���"! � . The first term will be 0 if

���� �is symmetric. The second term will generally be nonzero. The third term is in the

subspace spanned by the part of� �

that is orthogonal to� �

and that is what we want.

It generally depends on higher order moments and we would expect it to be nonzero,

although it can be zero in carefully constructed cases. For example, it equals zero if� �

is orthogonal to� �

and���� � ����9�

.

Using2�J !3!

we obtain similar results,

2 J !3! �4�"! K��. ��� � �� � � �6� � � � /

� �1 � � 1 � ��.L� � � �8� �

���� � �� � �0/

15

Page 16: Dimension Reduction for the Conditional Mean in Regression

� �������� � � ��� � ���� ��� �������� � ���� ����� ��� � ���� �����(4.9)

This result differs from (4.8) by the term � � � �� � � � �������! � � � which belongs to " �$#�%�& � ,but otherwise the results are essentially the same. In particular,

" �$# %�&('�) %�&�& # %�& ��* " �$# %�&('�),+ &�& # %�& ��* ".-0/21�3 465provided none of the key terms vanish.

For a numerical illustration, we generated 200 observations on 5 predictors and a

response as follows:

7 � *98 �7 �;: 7 � * 7 � � 8 �

7� *98 �

7=< : 7 � * �?> � 7 � A@(��8 <7CB *98 BD * 7 � � 7 �� A@

All errors 8(E are independent standard normal random variables. The responseD

was

generated without error to emphasize the qualitative nature of the results. The CMS is

spanned by �?> 'GF�'HF�'HF�'GF � � and � FI' > 'HF�'HF�'GF � � . Condition C.1 holds, but C.2 does not hold

because JLKNM � 7 < : 7 � �L* �?> � 7 � A@(� � which is correlated withD

. Table 1 gives the first

two pHd directions, OP � and OP � , and the first two sample IHT directions, OQ � and OQ � . pHd

found7 � and

7 <to be the important predictors, while IHT correctly picked

7 � and7 � .

In effect, pHd missed the linear component in favor of the quadratic component and7R<

.

Figure 1 gives a visual representation of these results. The response surface from IHT

gives a very good representation of the true surface, while the surface for pHd shows

only a relatively rough quadratic.

16

Page 17: Dimension Reduction for the Conditional Mean in Regression

TABLE 1

Sample pHd directions � ��� and ���� and IHT directions �� � and �� � from the simulated data.

�� � �� � � � � � � � ��� �����

0.166 0.022� ��� �����

� � ��� ����� � ��� �����0.999 0.024

��0.050 0.151 0.001 0.062

�� � ��������� � ��� �� ��0.032

� ��� �����"! � ��� ���#

0.035 0.016 0.017

The fact that pHd missed the linear trend in this example may not be surprising because

it is not very effective at finding linear trends in the first place (Cook 1998b). In addition,

as illustrated by this example, failure of condition C.2 can influence the behavior of pHd.

In contrast, IHT can find linear trends and does not require C.2 to be effective.

In a variety of similar simulations, IHT was found to be superior to pHd whenever there

is a clear linear trend in the data, even if condition C.2 holds. The linear trend is necessary

to provide a reasonable estimate of $&%�' , which is needed to prime the procedure. If there

is not a linear trend, IHT failed as expected. Basing IHT on the residuals ( (that is, using

)"* + '�',$&%,' ) instead of the response worked well in many situations.

5. Asymptotic behavior. In the previous sections we focused on the characterization

and estimation of the CMS at the population level. A full-fledged asymptotic analysis

of the sample properties exceeds the scope of this paper, and will be developed in the

authors’ future work. In this section we provide a few basic ideas and results. Three

types of asymptotic problems will concern us: the - . -consistency of the estimators, the

17

Page 18: Dimension Reduction for the Conditional Mean in Regression

�� �

a. IHT, View 1 b. IHT, View 2

c. pHd, View 1 d. pHd, View 2

FIG. 1. Summary views from IHT and pHd for the simulated data. View 1 shows the quadratic component

while View 2 is provided for contrast. For IHT, �� �� and ���������� ������� �! �"����#$ � . For pHd, �� ��and �%�&���'�( )� �* � �( � �* � #+ � .

asymptotic effects of standardization, and the test statistic for determining the dimension

of the CMS. We will now outline the asymptotic development for IHT.

The , - -consistency is easy to demonstrate. Let ./ be the sample version of/

; that

is

./1032 .46587:9 .; 5<7=7 .4>587?9A@B@C@B9 .;EDGF$H587I7 .46587KJL918

Page 19: Dimension Reduction for the Conditional Mean in Regression

where ���������� �������, �� ����� ��� �����������

. By the central limit theorem ������ and �� �����are � � -consistent. Hence �� �� � , being a (matrix-valued) smooth function of ������ and

�� ����� , is a � � -consistent estimator of��� �

. Consequently the eigenvectors of former

are � � -consistent estimators of the eigenvectors of the latter.

We have described the estimators for CMS in terms of the standardized explanatory

vectors�

to simplify the presentation. In practice, we first transform the observed into �� by ��!� ��#" $&%�'()( � +* �� � �&� , and then apply IHT to �� to obtain the estimated

vectors in ,.-0/2143 576 , say �8 $:9);<;2;29 �8>= . They are then multiplied by �� ".$?%�'(@( to become the

estimated vectors in ,>-0/25�3 AB6 , say �C $D9�;2;2;<9 �C = . Since both transformations involve only

� � -consistent estimators such as �� ()( and� � � , they do not affect the � � -consistency

of the estimators of the CMS. In other words �8 $ 9�;2;<;29 �8 = are � � -consistent estimators of

the vectors in ,>-0/25�3 1E6 , and �C $ 9�;2;<;29 �C = are � � -consistent estimators of the vectors in

,.-0/25�3 AB6 .An asymptotic test for dimension can be developed along the following lines. If the

rank of�

is F , then the smallest GH*IF eigenvalues�J� �

are 0. Hence the corresponding

eigenvalues of �� �� � behave like noise and their sum should converge to a definite

distribution. This sum, together with its asymptotic distribution, can then be used to test

the hypothesis KJLNM rank� �J� �4��� F . The largest F for which this hypothesis is rejected

is an estimator of the dimension of the CMS. To see how this asymptotic distribution

can be derived, let vec be the transformation that maps a matrix O �P��Q $R9);<;2;<9 QTSU� to the

column vector��QV�$ 9�;2;<;29 QV�S �?� . For any G -dimensional vector

�and GXWJG matrix

�, let

YZ�!Y7�[� 9 � ���\�[� � 9 vec� � � � � �

and ] �^Y ��� vec�[� 9);<;2;<9 �

S " $ ��� 9and denote

Y4� �� ��� 9 �� ����� � by � Y andY4�_� ��� 9 � ����� � by

Y L . Because the components of � Yare moment estimators, by the central limit theorem � � � � Y * Y L � converges in distribution

to a normal random vector with mean ` and covariance matrix a �bY L � . By construction,

19

Page 20: Dimension Reduction for the Conditional Mean in Regression

�������is a vector-valued smooth function of

�and

���������vec

�� ��. Hence, applying the

delta method, we have the convergence� � �vec

��� ����vec

�� ������������� �"!�# ����� �# �%$ & ���%'��# � $ ��� �# � (*)From here we can apply the Eaton-Tyler factorization (Eaton and Tyler, 1994) to derive

the desired asymptotic distribution.

6. Recumbent cows. For unknown reasons, pregnant dairy cows can become re-

cumbent — they lay down — either shortly before or after calving. This condition can

be serious, and frequently leads to death of the cow. Clark, Henderson, Hoggard, Ellison

and Young (1987) analyze data collected at the Ruakura (N.Z.) Animal Health Laboratory

on a sample of recumbent cows. We use 254 cases with complete records to illustrate

the dimension-reduction methods described in this article. The response is binary + �-,for surviving cows, + �/.

otherwise.

The 0 �21predictors are 35476 ��8:9<;:� , logarithm of serum asparate amino transferase

(U/l at 30C); 3=4>6 ��?�@A� , logarithm of serum creatine phosphokinase (U/l at 30C); and3=4>6 ��B:CEDF8G� , logarithm of serum urea (mmol/l). These predictors were used in one of the

candidate logistic models studied by Clark, et al. We deleted two anomalous cases at the

outset, as did Clark, et al.

The transformations of the blood measurements to logarithms effectively induced the

linear conditional predictor expectations of the kind discussed in this article (condition

C.1), although this rationale was not stated in the report by Clark, et al. However, an

analysis of the predictors indicated that condition C.2 might not be reasonable. For

example, the score test for heteroscedasticity (Cook and Weisberg 1983) in the linear

regression of 3=4>6 ��B:CEDF8G� on� 35476 ��8:9<;:�H! 35476 ��?�@A�I� has a p-value of 0.001. The lack

of constant variances J�KML ��N:O P $ NF� may cause problems for pHd because this allowsJ�KML �QNRO P $ NF� and + to be correlated. To avoid this issue we would like to use a dimension

20

Page 21: Dimension Reduction for the Conditional Mean in Regression

TABLE 2

Sample IHT directions ������ ������� ��� ���� from the recumbent cow data. Predictors have been standardized

marginally to have sample standard deviation equal to one.

�� �� ����

��������� �"!$#0.761 0.041 0.716

�������&%(')#0.387 *,+.- /102/ *,+3- 4�0�/

�������6587 9:�8#0.521 0.803 *,+.- +�43;

reduction method that requires only condition C.1. SIR is one possibility, but because the

response is binary it can find at most one direction in <>=�?A@8B CD , a limitation that is shared

by all methods covered by Theorem 1. In short, IHT introduced in Section 4.2 seems to

be the only known method that can find multiple directions in <(=�?A@8B CD and requires only

condition C.1.

Table 2 shows the sample IHT directions from the regression of the binary response

on the three predictors using a sample version of the population calculations described in

Section 4.2. The directions were computed in the scale of the original predictors, except

the predictors have been standardized marginally to have sample standard deviation equal

to one. All three sample IHT vectors EFHGIKJLJ EM IKJ were used to form the sample version of

N.

Applying the graphical regression methods proposed by Cook (1996, 1988a) to the

regression of O on the three IHT predictors, we inferred that a good summary is provided

by the binary response plot for the first two IHT predictors shown in Figure 2. The

correlation between the first IHT predictor and the linear combination of the predictors

from a logistic fit is PRQ S3S"S , so this discussion covers what might be a first step for

21

Page 22: Dimension Reduction for the Conditional Mean in Regression

��� � ����� � ��������� �������

� ������� !"��# �$ � "

%�& %('*) +*, %�-.) , -.) +�, /01

02435

6

2 35

1

FIG. 2. Binary response plot for the first two IHT predictors in the recumbent cow data. Filled circles

correspond to survivors.

many. Our interpretation of the plot in Figure 2 is based on visual comparisons of the

empirical conditional distributions of the IHT predictors given the response. We see from

the figure that the conditional distributions for the first two IHT predictors differ primarily

in location and scale; difference in orientation seem less important. Additionally, both

conditional distributions seem consistent with samples from bivariate normal distributions.

Using this as a guide, we concluded that a first logistic model should include (a) linear

and quadratic terms in the first IHT predictor since the conditional distributions for that

predictor differ in both location and scale, and (b) a quadratic term in the second IHT

predictor since the conditional distributions for that predictor differ primarily in scale.

22

Page 23: Dimension Reduction for the Conditional Mean in Regression

The fit of a logistic model with the three IHT predictors, their squares, and a cross product

between the first two predictors sustains these conclusions, as the ratios of the estimated

coefficients to their standard errors are less than one for all terms except the linear and

quadratic terms in the first predictor and the quadratic in the second. And the absolute

coefficients for these terms are all larger than about 2 standard errors.

The goal of this example is to illustrate that IHT summary plots can be informative

and can capture the primary dependence in the data. Subsequent use of the summary plot

must depend strongly on the applications context.

7. Discussion.

7.1. Overview. We addressed the problem of how to do dimension reduction in

regression when the mean function alone, rather than the whole conditional distribution,

is of interest. From this consideration the Central Mean Subspace arises naturally as

the inferential object. The CMS is that part of the central subspace that captures all the

information about the mean function available in the predictors, and shares many parallel

properties with the latter. Moreover, the CMS categorizes transparently the previous

methods into two kinds: those, such as OLS and pHd, that estimate vectors in the CMS

and those, such as SIR and SAVE, that can pick up vectors outside the CMS. This gives

us a clear idea of what these estimators do, and thereby provides guidance on their use.

The new IHT method, as well as other methods that might be developed from the class

of COZY vectors, have two potential advantages: they allows the investigator to focus

on the mean function and they do not require constraints on the conditional variances

variances��������� ��� ���

, although linear conditional expectations � ��� ������� are still re-

quired. Our simulations show that IHT performs better than pHd when the conditional

variance��������� �������

is not a constant. We make no claim that IHT method is more

“powerful” than pHd when both conditions C.1 and C.2 hold. The IHT predictors, like

23

Page 24: Dimension Reduction for the Conditional Mean in Regression

the pHd predictors, are partially ordered on their likely relation to the CMS. Plots of the

response versus the first few IHT predictors and, in particular, graphical regression can

be used to aid in determining the number of “significant” directions just as they are used

with other dimension reduction methods in regression (see Cook 1998a for an overview.)

However, formal testing methods are required for full effectiveness. Such methods are

under consideration, as outlined in Section 5.

7.2. Comparing ����� � and ������� � � . A comparison of ����� � and �������� ��� may be of

general interest in all regressions and of particular interest in some regressions. Because

�������� � ��������� � , we can use their dimensions as a rough indicator of the difference

between the two subspaces. If ���������������� ������� �����!������� ��� then all of the regression

structure is captured by the mean, while there may be much more to the regression

structure otherwise.

One method for inferring about aspects of "#�$�%��� �'&(������)� � � can be developed from

recent results on partial dimension reduction by Chiaromonte, Cook and Li (2001). Let

* be an orthonormal basis for �+������ � � and let � *-,.*%/ � be an orthonormal basis for 0 132 .Define 4 � *)5�6 and 78� *)5/ 6 . Then, assuming it exists, the central partial subspace

� �:9;���� < for =?>@�A7 , 4B� is the intersection of all subspaces �C�ED?� such that

= 7;>@�ED 5 7 , 4B�(7.10)

It can be shown that ����� �F�G������� � ��H *%/ � �:9I���� < and consequently we can infer about

" via � �@9I���� < .

For example, suppose 6 is normally distributed so that 4 7 . Then (7.10) holds if

and only if �E= , 4B� 7;> D 5 7 . Thus, we can infer about � �:9;���� < by applying SIR/SAVE to

the multivariate regression of �E= , 4B� on 7 . Because * will usually be estimated rather

than known, new asymptotic tests need to be developed for careful application.

24

Page 25: Dimension Reduction for the Conditional Mean in Regression

Appendix: proofs

PROOF OF PROPOSITION 1. That (i) implies (ii) is immediate. That (iii) implies (i)

is also immediate, because, if ������� �� is a function of �� � , then, given ���� , ������� ��is a constant, and hence independent of any other random variable. Now let’s prove that

(ii) implies (iii). By (ii),

������������� ����� � ������������ � �������������� ���� � ���!

The left hand side is

��" ������������� ����� ���#� � �%$&�'���(" �����)� ���*$,+(� � �-�/.

and the right hand side is ����" ������� ���� 01�1�%$2� + . Therefore Var " ������ ���� ��1�%$&�'34 Thus,

given ���� , ������ �� is a constant.

PROOF OF THEOREM 1. We first rewrite 56��78.:9;� making use of the fact that < is

a basis for the central mean subspace:

56��78.=9>�?�@��"BA��)��7�CD9 �>E ��CGF1��7�CH9 �;E �I$

�@��"BA�������� < � E ����7�CH9 � E ��CGF1��7�CH9 � E �*$J ��"BA�������� < � E ����7�CH9 � ��� E � < � E �=��CGF ��7KCD9 � ��� E � < � E �:�I$

�@��"BA��)��7�CD9 � LNM#E ��CGF1��7�CH9 � LOM#E �*$*

The second equality follows because < is a basis for P1Q&RTS?U V8W . The inequality follows

because of convexity. The next equality stems from the linearity of ��� E � <X� E � which is

equivalent to requiring that ��� E � <?� E �?� LOM#E , where LNM is the projection onto P�Q&RTSYU VZWwith respect to the usual inner product.

Thus,

5[��78.:9;� J 56��78. LNM 9;�25

Page 26: Dimension Reduction for the Conditional Mean in Regression

and the conclusion now follows because � is unique.

PROOF OF THEOREM 2. It is sufficient to show that if ��������� , then �� ����������������������� �"! #%$ . Assume then �����&�"�� . Now

��������� � �"�'�( ������)��� �+* ,-� �"�.�

�'�( ��/��0213��� � 02154768:9;�<� * , � ���=�.�

�'021>��������� � �?0�1A@

The conclusion follows since �� �0B1C��������� � �=021D�5�E����� ��! #A$ .

PROOF OF PROPOSITION 2. By definition of �B��� ��! #%$ , � ����� * F � * , � F . Because

of the linearity of ���<� * , �G��� , ��HJI5KL����� �"! #%$ and hence

M ����N?� �HOIP��� M ������� * ���.N?� �HOIQ�"� * , � �R@

Because M is linear in � ,

��S M ����N?� �HOI �"� * �-T2� M ������� * ���.N=� �HOI ���.@

Thus M ��� M * ��� * , �-� , so that �B��� ��! #%$ is a mean dimension-reduction subspace for the

regression of M on � , and

�����VU�! #A$��W�����X��! #%$(7.11)

Next, let Y be the matrix whose columns form a basis of �Z��� U;! #%$ ; that is, �B���VU�! #A$"�����YG� . Then M ��� M * �"� * YZ�G� . But by Proposition 1, given YZ�G� , ��� M * �"� is constant.

Hence,

M ��� M * ��� * ��Y � �(N?� �HJIP���

and therefore

� M N?� �HOI �"� ����� M * �"�ON?� �HOI �"� * ��Y � �(N?� �HOI �"�O@26

Page 27: Dimension Reduction for the Conditional Mean in Regression

Since ����������� ������������������ ���������� , it follows that � �����! ���" #�%$ � �'&(���������� and thus

)+*�,.-"/ 021435) � �����6� is a mean dimension-reduction subspace for the regression of � on �. Combining this with (7.11) gives the desired conclusion

)+*�,879/ 021�:;)+*�,.-"/ 021<3=) � ������� :;)+*�,87�/ 021+3) �>���?�6�@� )+*�,87�/ 0A1 &where the last equality follows because the linearity of �B�C�� D � ��� implies �9�?�FE )+*�,87�/ 021 .

PROOF OF THEOREM 3. Let G �IHJ�����! ��� 3LKand, as before, let MFN be the

projection matrix onto) ��D�� . Then

��OP�QHJ� 3;K ����R��S�'OP�QHJTU���V ��� 3WK ����R��X�B�%GY���[ZSince H and

Kare measurable with respect to D � � , and since �B�%�V �����\�B���] D � ���

by Proposition 1, the random variable G is measurable with respect to D � � . Hence

�B�CGY�����^��O�G_�B�Q�B D � ���`R��XMJNP���%GY���[&where the right hand side belong to

) �>D�� .

PROOF OF COROLLARY 2. We only prove part (i); part (ii) can be proved similarly.

In Corollary 1, take ab�>cd�@�Sc and e��>cd���Y��c(�B�%�f� . Then

��ghO6ai�>������ ���d� 3 e2�>������ ���`R��kj<�ml ���`� �����belongs to

) �>D�� . Now suppose that nY�ol-?p<q�?�r� � ��� belongs to

) �>D�� . Then n � � is

measurable with respect to D � � . By Theorem 3, �'OP�%n � ���d�s��Rt�Yl ���`� nu�Yl-���`�<�����

also belongs to) ��D�� . Hence, by induction, all vectors of the form l

-���`�<�9�?� belong to

) �>D�� .

Acknowledgments. We would like to thank two referees and an Associate Editor for

their suggestions that led to significant improvement of this article.

27

Page 28: Dimension Reduction for the Conditional Mean in Regression

REFERENCES

Bura, E. and Cook, R. D. (2001). Estimating the structural dimension of regressions via parametric inverse

regression. Journal of the Royal Statistical Society, Ser. B 63, 393-410.

Chiaromonte, F., Cook, R.D. and Li, B. (2001). Sufficient dimension reduction in regressions with categorical

predictors. Submitted.

Clark, R. G., Henderson, H. V., Hoggard, G. K. Ellison, R. S. and Young, B. J. (1987). The ability of

biochemical and haematolgical tests to predict recovery in periparturient recumbent cows. New

Zealand Veterinary Journal 35, 126–133.

Cook, R.D. (1992). Regression plotting based on quadratic predictors. In Y. Dodge (ed.), L1-Statistical

Analysis and Related Methods, pp. 1156–128. Amsterdam: North Holland.

Cook, R.D. (1994a). On the interpretation of regression plots. Journal of the American Statistical Association

89, 177–189.

Cook, R.D. (1994b). Using dimension-reduction subspaces to identify important inputs in models of physical

systems. In 1994 Proceedings of the Section on Physical and Engineering Sciences, Alexandria,

VA: American Statistical Association, 18–25.

Cook, R.D. (1996). Graphics for regressions with a binary response. Journal of the American Statistical

Association 91, 983–992.

Cook, R.D. (1998a). Regression Graphics. New York, Wiley.

Cook, R.D. (1998b). Principal Hessian directions revisited. Journal of the American Statistical Association

93, 84–100.

Cook, R.D. and Nachtsheim, C.J, (1994). Reweighting to achieve elliptically contoured covariates in regres-

sion. Journal of the American Statistical Association 89, 592–599.

Cook, R.D. and Weisberg, S. (1983). Diagnostics for heteroscedasticity in regression. Biometrika 70, 1–10.

Cook, R.D. and Weisberg, S. (1991). Discussion of “Sliced inverse regression for dimension reduction”.

Journal of the American Statistical Association 86, 316–342.

Cook, R.D. and Weisberg, S. (1999). Graphics in statistical analysis: Is the medium the message? The

American Statistician 53, 29–37.

Diaconis, P. and Freedman, D. (1984). Asymptotics of Graphical Projection Pursuit. The Annals of Statistics

13, 496–499.

28

Page 29: Dimension Reduction for the Conditional Mean in Regression

Eaton, M.L. and Tyler, D. (1994). The asymptotic distribution of singular values with application to canonical

correlations and correspondence analysis. Journal of Multivariate Analysis 50, 238-264.

Hall, P. and Li, K.C. (1992). On almost linearity of low dimensional projections from high dimensional

data. The Annals of Statistics 21, 867–889.

Li, K.C. (1991). Sliced inverse regression for dimension reduction (with discussion). Journal of the American

Statistical Association 86, 316–342.

Li, K.C. (1992). On principal Hessian directions for data visualization and dimension reduction: Another

application of Stein’s lemma. Journal of the American Statistical Association 87, 1025–1039.

Li, K.C. and Duan, N. (1989). Regression analysis under link violation. The Annals of Statistics 17,

1009–1952.

SCHOOL OF STATISTICS

1994 BUFORD AVE.

UNIVERSITY OF MINNESOTA

ST. PAUL, MINNESOTA 55108

DEPARTMENT OF STATISTICS

THE PENNSYLVANIA STATE UNIVERSITY

326 THOMAS BUILDING

UNIVERSITY PARK, PENNSYLVANIA 16802

29