Robust dimension reduction

A. Bergesio1 M.E. Szretter2 V.J. Yohai3
  • 1

    Department of Mathematics, Universidad del Litoral, Santa Fe, Argentina [anbergesio@gmail.com]

  • 2

    Department of Mathematics, Universidad de Buenos Aires, Buenos Aires, Argentina [meszre@dm.uba.ar ]

  • 3

    Department of Mathematics, Universidad de Buenos Aires, Buenos Aires, Argentina and CONICET, Argentina [victoryohai@gmail.com]

Keywords: Sufficient reduction, Principal Fitted Components and Reduced Rank regressions

Abstract

Non-parametric regression is a very flexible procedure to establish the relationship between a response y variable and several covariables 𝐱=(x1,,xp). Non-parametric regression techniques deal with fitting a model of the form y=g(x1,,xp)+δ without assuming any predetermined parametric model for g. However the number of observations required for non-parametric regression increases exponentially with the number of covariables, and this number may be larger of what is generally available. This is usually known as the dimensionality curse problem.

Cook [2007] defined the concept of sufficient reduction statistics to overcome this problem. Suppose that y is the response and 𝐱Rp is the vector of covariables. Let 𝐳=g(𝐱), g:RpRd. Then 𝐳 is a sufficient reduction statistics of size d<p for y if, 𝒟(y|𝐳)=𝒟(y|𝐱) 𝒟(y|𝐳). Therefore the vector 𝐳 contains all the information that 𝐳 has about y. This imply that 𝐱 can be replaced by 𝐳 in the non-parametric regression. To obtain a sufficient reduction statistics, Cook [2007] introduced the principal fitted components (PFC) model. The PFC model assumes that

𝐱=Γ0B0𝐯(𝐲)+Δ01/2𝐮

where Γ0 is a p×d matrix, d<p, B0 is a d×r matrix where rd and v:RRr, Γ0TΔ-1Γ0=Id, E(𝐮T𝐮)=Ip. Then, E(𝐱)𝒮d, where 𝒮d is the subspace of dimension d generated by the columns of Γ0. For example v(y) =(1,y,y2,,yd-1 ). Calling F0=Γ0B0, we can also write 𝐱=F0𝐯(y)+𝜺, where rank(F0)=d and E(𝜺T𝜺)=Δ0. This imply that 𝐱 follows a reduced rank multiple regression model of rank d with regressor equal to v(𝐲). Therefore the proposed estimators can also be used to estimate these models.

Cook [2007] and Cook and Forzani [2008] proved that under the PFC model where 𝜺 is Np(𝟎,Δ0), a sufficient reduction statistics of dimension d for y is given by 𝐳=Γ0TΔ0-1𝐱.

Let (𝐱1,,y1),,(𝐱n,yn) be a random sample of the PFC model and consider the Mahalanobis distances between the predicted and observed values of 𝐱i

MDi(F,Δ)2=(𝐱i-F𝐯(yi))TΔ-1(𝐱i-F𝐯(yi))

Then the maximum likelihood estimator Θ0=(Γ0,B0,Δ0) is given by

Θ^=argminΘdet(Δ)

subject to

1ni=1nMDi(F,Δ)2=p,rank(F)=d

As most of the ML estimators that assume normal errors, the ML estimator of the PFC model is very sensitive to a few outliers, even a single outlier may take the ML estimator beyond any limit. To overcome this problem a class of robust estimators for the PFC model based on a τ-scale estimator, see Yohai and Zamar [1988], were proposed. The τ-scale estimators are highly robust and may have breakdown point 0.5. Let sτ be a τ-scale estimator, and let κ its asymptotic value for a sample of the χ2 distribution with p degrees of freedom. Then a class of robust estimators for the PFC model (called τ-PFC estimators) is given by

Θ^τ=argminΘdet(Δ)

subject to

sτ2(MD1(𝚯),,MDn(𝚯))=κ,rank(F)=d.

These estimators are strongly consistent under general conditions. A computational procedure based on iterative weighted maximum likelihood estimators, where the weights penalize outlier observations is given. A procedure based on cross validation to determine d, that is the dimension of the sufficient reduction statistics, is also proposed. A Monte Carlo study shows that the τ-PFC estimators are simultaneously highly efficient and robust. We also give an example with real data where the non-parametric model using the reduction sufficient statistics obtained with the proposed robust sufficient reduction statistics works better than using the one obtained with the maximum likelihood estimator.

References

  • Cook [2007] R. D. Cook. Fisher lecture: Dimension reduction in regressionl. Statistical Science, 22(2):1–26, 2007.
  • Cook and Forzani [2008] R. D. Cook and L. Forzani. Principal fitted components fordimension reduction in regression. Statistical Science, 23(4):485–501, 2008.
  • Yohai and Zamar [1988] R. D. Yohai and R. H. Zamar. High breakdown-point estimates of regression by means of the minimization of an efficient scale. Journal of the American Statistical Association, 83(402):406–413, 1988.