Sign and Rank Covariance Matrices - CiteSeerX

200 downloads 0 Views 201KB Size Report
erties of the sign and rank covariance matrices are described. ... estimates and tests in the one sample, two-sample and several sample cases and also.
Sign and Rank Covariance Matrices: Statistical Properties and Application to Principal Components Analysis Christophe Croux, Esa Ollila, and Hannu Oja Abstract. In this paper, the estimation of covariance matrices based on multivariate sign and rank vectors is discussed. Equivariance and robustness properties of the sign and rank covariance matrices are described. We show their use for the principal components analysis (PCA) problem. Limiting efficiencies of the estimation procedures for PCA are compared.

1. Introduction Let X = {x1 , ..., xn } be a univariate data set. The univariate signs and centered ranks Si and Ri can be defined implicitly by (1.1)

ave{|xi |} = ave{Si xi } and

ave{|xi − xj |} = ave{Ri xi }.

If the sign function S is defined as S(x) = −1, 0, 1 for x < 0, = 0, > 0, then Si = S(xi ) and

Ri = avej {S(xi − xj )}

In the univariate case, signs and ranks have been extensively used to construct estimates and tests in the one sample, two-sample and several sample cases and also in univariate regression problems. The concepts of sign and rank can be extended to the multivariate case by considering extensions of the L1 type objective functions in (1.1). Let X = {x1 , ..., xn } be a k-variate data set. We first generalize the concept of sign and ranks utilizing the Euclidean distance || · ||, resulting into the spatial signs and ranks. These multivariate vector valued spatial signs and ranks S i and Ri verify, analogous to (1.1), ave{||xi ||} = ave{S Ti xi } and ave{||xi − xj ||} = ave{RTi xi }. If

½ S(x) =

kxk−1 x 0

if x 6= 0 if x = 0

1991 Mathematics Subject Classification. Primary 62H99; Secondary 62G35. Key words and phrases. Covariance and correlation matrices, Eigenvectors and eigenvalues, Multivariate median, Multivariate Sign and Rank, Robustness.

2

Croux, Ollila, and Oja

then signs and centered ranks are given by S i = S(xi ) and Ri = avej {S(xi − xj )}. The spatial signs and ranks are rotation equivariant but, unfortunately, not affine equivariant. Therefore we need to introduce the Oja objective function in (1.1). Every subset of (k + 1) points xi1 , ..., xik+1 of X defines a k-variate simplex with volume V (xi1 , ..., xik+1 ). The affine equivariant k-variate signs S i and ranks Ri , i = 1, ..., n are implicitly defined by n o © ª k! ave V (0, xi1 , ..., xik−1 , xi ) = ave S Ti xi and

n o © ª k! ave V (xi1 , ..., xik , xi ) = ave RTi xi ,

respectively. For a review on multivariate signs and ranks we refer to [6], where also explicit expressions for the affine equivariant signs and ranks are given. Contrary to the univariate case, the multivariate signs and ranks can be used to estimate the covariance structure also. In Section 2 we review the most important properties of the covariance matrices based on multivariate signs and ranks. Since covariance matrices play a key-role in multivariate analysis, it is natural to perform well-known multivariate techniques as principal component analysis, factor analysis, canonical correlation analysis, ... using the sign and rank based covariance matrices. This yields a class of methods of a non-parametric nature, more robust than the classical procedures, while still being very efficient. Efficiency properties of the different methods are presented in the context of principal component analysis. In Section 3 we treat the affine equivariant procedures, while Section 4 discusses the approach based on the spatial signs. The latter approach to principal component analysis is appealing since is highly robust, and fast to carry out. Its efficiencies depend on the (standardized) eigenvalues and need to be computed numerically. A small simulation study will be presented to compare its finite sample efficiencies with the asymptotic ones. While this paper puts focus on principal component analysis, other applications of sign and rank methods do exist. A first overview has been made by [15], and more detailed studies in the context of canonical correlation analysis, multivariate regression, independent component analysis and factor analysis are currently being carried out. Let us also mention a nice application to wireless communications [16] using robust principal component analysis.

2. Sign and rank covariance matrices 2.1. Definitions For constructing the sign covariance matrix, first consider the centered data set ˆ ..., xn − µ}. ˆ Y = {y 1 , ..., y n } = {x1 − µ,

Sign and rank covariance matrices

3

ˆ is chosen so that the sum of sign vectors based on Y is a The location estimator µ ˆ becomes the spatial median or L1 -median, zero vector. When using spatial signs, µ ˆ the Oja median [7]. Let and when using the affine equivariant signs we obtain for µ Sˆi be the signs of the centered observations. The sign covariance matrix (SCM) is now nothing else but the usual covariance matrix computed from the multivariate signs: (2.1)

T SCM = ave{Sˆi Sˆi }.

In the same way, the rank covariance matrix (RCM) is defined as (2.2)

RCM = ave{Ri Ri T },

where the rank vectors are computed from the uncentered data, so no initial location estimate is necessary here. Note that once the signs and ranks are obtained, computation of the SCM and RCM is immediate. Depending on whether spatial or affine equivariant signs and ranks are used, covariance matrices with different equivariance, efficiency and robustness properties will be obtained. In all cases, the sign and rank covariance matrices are asymptotically equivalent with U-statistics and the population counterparts, population sign and rank covariance matrices, are then naturally defined as the expected values of the related kernels. It also follows that the SCM and RCM are consistent estimators for their population counterparts and have limiting multinormal distributions (see [15]). We will often refer to an elliptically symmetric distribution F . This means that there exists a positive definite scatter matrix Σ and a location parameter µ such that z i = Σ−1/2 (xi − µ) has a spherically symmetric distribution F0 . If the population covariance matrix of F0 exists, we will take it without lost of generality equal to the identity matrix, and then Σ becomes the covariance matrix of F . We can decompose Σ as Σ = λ1/k P LP T where L is a diagonal matrix with Det(L) = 1 and descending diagonal elements, P is an orthogonal matrix containing the eigenvectors of Σ in its columns, and λ = Det(Σ) is the Wilks’ generalized variance. 2.2. Spatial sign and rank covariance matrices The covariance matrices based on spatial signs and ranks are called the spatial sign covariance matrix and the spatial rank covariance matrix. In the following we list some of their properties, see e.g. [5] and [13]. • The spatial SCM and RCM are rotation equivariant but not affine equivariant. If the SCM∗ and RCM∗ are calculated from the transformed observations x∗i = Γxi + b, with Γ orthogonal, then SCM ∗ = Γ SCM ΓT and RCM ∗ = Γ RCM ΓT .

4

Croux, Ollila, and Oja

• At an elliptical model distribution F , the population sign and rank covariance matrices are given by SCM (F ) = P LS P T and RCM (F ) = P LR P T , where LS and LR are diagonal matrices with positive descending elements on the diagonal. The eigenvectors of the spatial SCM and RCM are thus the same as those of Σ. Therefore spatial SCM and RCM may be used in the principal component analysis (PCA) and subspace estimation. • The influence functions of SCM and RCM are uniformly bounded, indicating their robustness. The influence function of the spatial SCM at a distribution F symmetric around zero has the simple form (2.3)

IF (x, SCM, F ) = S(x)S(x)T − SCM (F )

and is seen to be constant in the radius kxk of the contamination point x. 2.3. Affine equivariant sign and rank covariance matrices If affine equivariant signs and ranks were used, the matrices are similarly called the affine equivariant sign covariance matrix and the affine equivariant rank covariance matrix. Their most important properties are listed below. Full detail can be found in [9] and [10]. • The SCM and RCM are affine equivariant in the sense that if SCM∗ and RCM∗ are calculated from the transformed observations x∗i = Axi + b, with nonsingular A, then SCM ∗ = det(A)2 (A−1 )T SCM A−1 and RCM ∗ = det(A)2 (A−1 )T RCM A−1 . • At an elliptical model distribution F , the population sign and rank covariance matrices are SCM (F ) = λS (F )P L−1 P T and RCM (F ) = λR (F )P L−1 P T . We see that SCM (F ) and RCM (F ) are proportional to the inverse of the covariance matrix. The eigenvectors of Σ can be estimated using the affine equivariant SCM and RCM , but also the eigenvalues. Since the correlation structure of F can be retrieved, applications are possible in multivariate regression analysis [11], canonical correlations analysis (CCA), etc. • The influence functions of SCM and RCM at a spherical distribution F0 can be written in the form (2.4)

IF (x, C, F0 ) = αC (r)uuT − βC (r)Ik

where the vector x = ru is decomposed in its radius r = kxk and its direction vector u = u/kuk. At the spherical distribution F0 , the variables r and u are independent. In fact, it can be shown that the influence function of any affine equivariant scatter matrix functional C possessing an influence function can be written as in (2.4) with αC (r) and βC (r) for two real valued functions [2]. When C stands for the SCM, αC (r) is a linear function in r, confirming its

Sign and rank covariance matrices

5

L1 -type character. For the RCM, the influence function is approximately linear in the radius of the contamination point. For the sample covariance matrix, we have a quadratic influence function. So although the influence functions of the affine equivariant SCM and RCM are not bounded, they indicate that the sign and rank procedures are more robust than the classical ones. • Asymptotic efficiencies have been computed. The resulting estimates (in PCA, CCA, multivariate regression, etc.) are highly efficient in the multivariate normal case and perform better than the estimates based on the regular sample covariance matrix for heavy tailed distributions. 2.4. Tyler’s scatter matrix estimate Hettmansperger and Randles [3] discussed location and scatter estimate based on spatial signs, but now resulting in an affine equivariant procedure. The idea is as follows: choose a positive definite symmetric k × k matrix V with T race(V ) = k and a k-vector µ, such that 1 Ik . k where the spatial signs S i = S(z i ) are being computed from the standardized observations zi = V −1/2 (xi − µ). The resulting estimate Vˆ equals Tyler’s Mestimate [12]. Its properties (see [12, 3]) include: ave{S(z i )} = 0 and ave{S(z i )S(z i )T } =

• It is affine equivariant in the sense that if SCM∗ and RCM∗ are calculated from the transformed observations x∗i = Axi + b, with nonsingular A, then Vˆ ∗ =

k T race(A Vˆ AT )

A Vˆ AT .

Tyler’s estimator can be seen as a shape matrix estimator, rather then a scatter matrix estimator, since it estimates Σ only upto a constant [8]. • The influence functions of Vˆ at a spherical distribution F0 has the same form as (2.3) and is bounded and even constant in the radius of the contamination point. Although the formal breakdown point of Vˆ is not very high, it turns out that for many data configurations Vˆ is very robust against outliers. • Tyler’s estimate is consistent and has a limiting multivariate normal distribution. The limiting variance is independent of F0 , which is an appealing nonparametric property.

3. Principal component analysis using affine equivariant signs and ranks We first state the problem. Let X = {x1 , ..., xn } be a random sample from an elliptical distribution F with covariance matrix Σ = P ΛP T where Λ = diag(λ1 , ..., λk ) with λ1 > ... > λk and P = (p1 , ..., pk ) . As before, write F0 for the spherical distribution of z ∼ Σ−1/2 (xi − µ) = Λ−1/2 P T (xi − µ).

6

Croux, Ollila, and Oja

The problem is to estimate the eigenvectors pj , j = 1, ..., k. In [2] we find ˆ C,j computed from a regular the following general result for eigenvector estimate p affine equivariant scatter matrix estimator C. √ Proposition 3.1. Under general assumptions, the limiting distribution of n(ˆ pC,j − pj ) is a multivariate normal distribution with limiting covariance matrix · ¸ 2 λj λl δC (F0 ) X T ASV(pC,j ; F ) = pp k(k + 2) (λj − λl )2 l l l6=j

2 (F0 ) δC

where the constant αC is given by (2.4).

=

2 EF0 [αC (r)]

With (3.1)

T race(ASV(pC,j ; F )) =

depends on the used scatter matrix C and

· ¸ 2 δC (F0 ) X λj λl k(k + 2) (λj − λl )2 l6=j

a natural measure of efficiency when comparing competing estimates based on affine equivariant scatter matrices C1 and C2 is given by (3.2)

ARE(C2 , C1 ; F ) =

T race(ASV(pC1 ,j ; F )) δ 2 1 (F0 ) = C 2 (F ) . T race(ASV(pC2 ,j ; F )) δC 0 2

2 Note that (3.2) only depends on F0 and that only the constants δC (F0 ) are needed in efficiency comparisons. With r = kzk the radius of z ∼ F0 , the constant is 2 δC (F0 ) =

k 2 E(r4 ) E 2 (r2 )

for the regular sample covariance matrix. For the affine equivariant sign covariance matrix and for Tyler’s estimate, one gets · ¸ 4E(r2 ) 2 2 δC (F0 ) = k 2 − 3 and δC (F0 ) = (k + 2)2 , E 2 (r) respectively. Now we are ready to compare the efficiency of the estimates of pj based on the different affine equivariant scatter matrices. In the table below we give asymptotic relative efficiencies of the eigenvector estimates based on SCM, RCM, and Tyler’s estimator relative to eigenvector estimates based on the sample covariance matrix. We took for F0 once the normal distribution and once a multivariate student t-distribution with 5 degrees of freedom, t(5), and considered dimensions k = 2, 5, 10 : It is confirmed that at normal distributions the efficiencies are very high for the affine equivariant SCM and RCM. Also Tyler’s estimator is quite efficient in the higher dimensions. At the heavy tailed t(5)-distribution all estimators clearly dominate the classical procedure.

Sign and rank covariance matrices

SCM RCM Tyler’s M

k=2 0.96 0.99 0.50

Normal k=5 0.99 0.97 0.71

k = 10 0.99 0.97 0.83

k=2 2.00 2.05 1.50

t(5) k=5 1.91 1.92 2.14

7

k = 10 1.84 1.86 2.50

4. Principal component analysis using spatial signs Besides its high robustness, an important virtue of the spatial SCM is its simplicity and ease of implementation. In contrast with many other highly robust multivariate estimators, the required computation time is extremely limited. For centering the data, we need to compute L1 -median, which comes down to solving a convex optimization problem and can be done very fast and accurately. Then, the computation of the actual SCM is only an explicit O(n) operation. A nice feature of PCA based on the spatial SCM is that it can also be applied in case the number of variables k is higher than the number of observations n, as was done by [4] for a functional data analysis problem. Indeed, the definition of the spatial median and the spatial SCM rely on Euclidean distances, which can be computed without any problem in any dimension. Of course, the spatial SCM will not have full rank in this case. Note that the affine equivariant SCM and RCM are not well defined for k > n, since the volumes of the simplices appearing in their definition will all be equal to zero. Experience shows us that even for high dimensional data sets, the spatial SCM can be computed without any difficulty. But the spatial SCM has one major drawback: it is only orthogonal equivariant, and not affine equivariant. In the context of principal component analysis, however, orthogonal equivariance is sufficient: let p∗j be calculated from the transformed observations x∗i = αΓxi + b, with Γ orthogonal, b a vector and α a scalar, then we have p∗j = Γpj for j = 1, . . . , k. We investigate the efficiency properties of the eigenvector estimates based on the spatial SCM, theoretically and by means of a simulation experiment. We do not considered yet the spatial RCM: we infer that its efficiencies will be somewhat higher. But we should not forget that the computational complexity of the RCM is is O(n2 ), making it less attractive as a competitor for the spatial SCM in the context of PCA. 4.1. Efficiency for PCA based on the spatial SCM As in Section 3, suppose that the data are generated from an elliptically symmetric distribution with covariance matrix Σ = P ΛP T . Now we cannot reduce the inference problem to the spherically symmetric model, since this would require a non-orthogonal transformation of the data. But by means of a rotation, we obtain z i = P T (xi − µ) ∼ F1 , a centered distribution having diagonal covariance matrix diag(λ1 , . . . , λk ). Note that we can decompose z i as z i = ri Λ1/2 ui , with ui uniformly distributed on the periphery of the k-variate unit sphere and independent of ri = kΛ−1/2 z i k.

8

Croux, Ollila, and Oja

As mentioned in Section 2, the population SCM is of the form P ΛS P T where we can write now · 1/2 T 1/2 ¸ Λ uu Λ zz T ]=E = diag(λS,1 , ..., λS,k ) with λS,1 > ... > λS,k . ΛS = EF1 [ kzk2 uT Λu The eigenvalues λS,j of the spatial SCM depend in a non-linear way on the eigenˆ S,j of the spatial values of Σ. The limiting distribution of the jth eigenvector p SCM can be derived via similar arguments as in [1] and [2]. √ Proposition 4.1. Under general assumptions, the limiting distribution of n(ˆ pS,j − pj ) is a multivariate normal distribution with limiting covariance matrix " # X λl λj bjl (Λ) T ASV(pS,j ; F ) = , 2 pl pl (c (Λ)λ − c (Λ)λ ) j j l l l6=j where cl (Λ) = E[u2j /(γ1 u21 + . . . + γk u2k )] for l = 1, . . . , k, and for 1 ≤ j, l ≤ k bjl (Λ) = E[u2j u2l /(γ1 u21 + . . . + γk u2k )]2 , with (u1 , . . . , uk ) the components of a random variable u, uniformly distributed on the periphery of a unit sphere, and γ1 , . . . , γp the standardized eigenvalues, that is γj = λj /(λ1 + . . . , λk ). It is important to note that this limit distribution only depends on the standardized eigenvalues, but not on the underlying elliptical model (i.e. it does not depend on the distribution of the radius r = kzk for z ∼ F0 .) In this sense, and similar to Tyler’s estimator, the spatial SCM is of a non-parametric nature. Efficiency comparisons can be based again on " # X λi λj bij (Λ) (4.1) T race(ASV(pS,j ; F )) = 2 . (λi ci (Λ) − λj cj (Λ)) j6=i We emphasize that asymptotic relative efficiencies, as defined in (3.2), will now depend on the standardized eigenvalues (in contrast with the affine equivariant procedure). When we are tending to a spherically symmetric distribution, we have that cj (Λ) = 1/k and bij (Λ) = 1/(k(k + 2)), yielding a limit value for the trace of the asymptotic variance · ¸ k+2X λj λl , k (λj − λl )2 l6=j

the same result as for the affine equivariant Tyler’s estimator. The corresponding relative efficiency with respect to the classical estimator is then, for this limit case, given by k/(k + 2). In the bivariate case, it is possible to compute the asymptotic variances explicitly. Let γ = λ2 /λ1 . After some calculus we obtain from (4.1) and for j = 1 an

9

0.5

Sign and rank covariance matrices

0.0

0.1

0.2

0.3

0.4

Efficiency

0.0

0.2

0.4

0.6

0.8

1.0

gamma

Figure 1. Asymptotic relative efficiency of the first eigenvector estimate based on the spatial SCM w.r.t. the estimate based on the sample covariance matrix for k = 2 at a normal distribution N (0, diag(λ1 , λ2 )) as a function of γ = λ2 /λ1 .

√ γ

γ asymptotic variance of 2(1−√γ)2 , while for the classical estimator we get (γ−1) 2. Figure 1 pictures the relative asymptotic efficiency of the first eigenvector estimate based on the spatial SCM with respect to the classical estimate as a function of γ. We see that the efficiency increases in γ, and tends fast to its maximal value of k/(k + 2) = 0.5 when γ tends to one. So the efficiencies get higher when we are closer to a spherically symmetric distribution. Indeed, in the later case the information lost when projecting the data on the unit sphere, i.e. when computing the spatial signs, is minimal. On the other hand, note that already for γ >= 0.2, the efficiency loss w.r.t the maximal value (which equals the efficiency of Tyler’s estimate) remains reasonable.

Using numerical techniques, we also computed for the three-variate case relative asymptotic efficiencies of the first eigenvector. In Figure 2 we see how these efficiencies depend on γ = λ2 /λ1 and δ = λ3 /λ2 , the relative magnitudes of the different eigenvalues. Again we see that maximal efficiency of k/(k + 2) = 0.6 is attained when we tend to the spherical case. Efficiencies tend to zero when the last two eigenvalues become negligible, but not when only δ tends to zero. Except for the extreme cases, where there is a huge disproportion between the first and the other two eigenvalues, we can say that the efficiencies are fairly close to their maximal value. The plot for the efficiency of the second eigenvector estimate is very similar and not represented here.

10

Croux, Ollila, and Oja

0.7 0.6

Efficiency

0.5 0.4 0.3 0.2 0.1 0 1 0.8

1 0.6

0.8 0.6

0.4 0.4

0.2

0.2 0

delta

0

gamma

Figure 2. Asymptotic relative efficiency of the first eigenvector estimate based on the spatial SCM w.r.t. the estimate based on the sample covariance matrix for k = 3 at a normal distribution N (0, diag(λ1 , λ2 , λ3 )) as a function of γ = λ2 /λ1 and δ = λ3 /λ2 4.2. A remark on the limit distribution of squared angular distances Until now, we worked with the trace of the covariance matrix to measure efficiencies. Another natural way to measure the accuracy of the estimate is to use the ˆ j and the true value pj . Indeed, the eigenangular distance between the estimate p vectors are giving us interesting directions to project the data on, and we are often more interested in the direction than in the components of the eigenvector. Below we give the limiting distribution of the squared angular distance of an eigenvector estimator based on regular affine equivariant scatter matrix estimators or on the spatial SCM. ³ ´2 Proposition 4.2. Under general assumptions, n cos−1 (|ˆ pTj pj | has a limiting distribution given by X¡ ¢ pTl ASV(pj ; F )pl χ2l (1), l6=j

where χ21 (1), ..., χ2k (1) are independent chi squared variables with one degree of freedom. (The above result follows from propositions 3.1 and 4.1, combined with √ P (pˆj T pj )2 = 1 − l6=j (ˆ pTj pl )2 , (cos−1 1 − u)2 ≈ u for u small, and standard

Sign and rank covariance matrices

11

asymptotics). It is now nice to see that the limiting expectations of the squared angular distances converge to the trace of the limiting covariance matrix of the eigenvector estimate: ³ ´2 lim nE[ cos−1 |ˆ pTj pj | ] = T race(ASV(pj ; F )).

(4.2)

n

4.3. Simulation study In the preceding section large-sample efficiencies were obtained for the spatial SCM eigenvector estimates relative to corresponding estimates based on the sample covariance matrix. In this section, finite-sample efficiencies are obtained by means of a modest simulation study. We take an almost identical simulation setup as in [9], were simulations have been performed for the affine equivariant SCM. For m = 10000 samples of sizes n = 20, 50, 100, 300, 1000, observations were generated from a k-variate elliptical t-distribution with ν degrees of freedom and covariance matrix Σ = diag(1, . . . , k). Our choices are k = 2, 3 and ν = 1, 5, 6, 8, 15, ∞. Please note that ν = ∞ corresponds to multivariate normal samples. The estimated quantity is the direction of the first eigenvector. The error in ˆ 1 |). The mean squared angular distance direction is given by the angle cos−1 (|pT1 p over all simulations is then m

MSE(ˆ p1 ) =

1 X (i) ˆ 1 |)2 cos−1 (|pT1 p m i=1

(i)

ˆ 1 is the estimate computed from the ith generated sample. The relative where p efficiencies reported in Table 1 are the ratios of these MSEs. It follows from (4.2) that these relative efficiencies tend asymptotically to (3.2). First of all, note that the finite sample efficiencies converge well to the asymptotic ones listed under n = ∞. An exception is of course the case ν = 1, where the population covariance matrix is not existing. However, the scatter matrix of this multivariate t(1) distribution can still be estimated with the spatial SCM. Existence of the population spatial SCM does indeed not require any moment conditions. Somewhat slower convergence is seen at ν = 5, and the efficiencies for k = 3 are significantly larger than for k = 2. Note that for heavy tailed distributions, the spatial SCM outperforms the classical estimator. It turns out that although the efficiencies are not so high as for the affine equivariant SCM, they are still fairly large, certainly in higher dimensions. Taking into account the good robustness properties, and the ease of computation of the spatial SCM, we may conclude that this approach is an excellent competitor of other existing procedures for robust principal components analysis.

12

Croux, Ollila, and Oja Table 1. Simulated finite sample efficiencies of the first eigenvector estimates based on the spatial SCM relative to eigenvector estimates based on the sample covariance matrix. Samples are generated from a k-variate t-distribution with ν degrees of freedom and Σ = diag(1, . . . , k).

k=2

k=3

n = 20 n = 50 n = 100 n = 300 n = 1000 n=∞ n = 20 n = 50 n = 100 n = 300 n = 1000 n=∞

1 1.679 3.738 8.472 32.52 120.5 ∞ 1.523 2.563 4.312 15.92 64.69 ∞

Degrees of freedom 5 6 8 15 0.798 0.757 0.701 0.634 0.890 0.756 0.634 0.508 0.979 0.816 0.653 0.495 1.193 0.926 0.679 0.551 1.215 0.942 0.724 0.564 1.456 0.971 0.728 0.574 0.948 0.915 0.870 0.816 1.049 0.952 0.836 0.732 1.124 0.963 0.839 0.651 1.496 1.096 0.849 0.611 1.637 1.174 0.908 0.669 1.750 1.167 0.875 0.689

∞ 0.566 0.425 0.388 0.475 0.477 0.485 0.745 0.646 0.547 0.536 0.577 0.583

References [1] F. Critchley, Influence in principal components analysis. Biometrika 72 (1985), 627– 636. [2] C. Croux, and G. Haesbroeck, Principal component analysis based on robust estimators of the covariance and correlation matrix: influence functions and efficiencies. Biometrika 87 (2000), 603–618. [3] T.P. Hettmansperger, and R. Randles, A practical affine equivariant multivariate median. Biometrika (2002), to appear. [4] N. Locantore, J.S. Marron, D.G. Simpson, N. Tripoli, J.T. Zhang, and K.L. Cohen, Robust principal components for functional data. Test 8 (1999), 1–73. [5] J. I. Marden, Some robust estimates of principal components. Statistics and Probability Letters 43 (1999), 349–359. [6] H. Oja, Affine invariant multivariate sign and rank tests and corresponding estimates: a review. Scandanavian Journal of Statistics 26 (1999), 319-343. [7] H. Oja, Descriptive statistics for multivariate distributions. Statistics and Probability Letters 1 (1983), 327–332. [8] E. Ollila, T.P. Hettmansperger, and H. Oja, Affine equivariant multivariate sign methods. Submitted (2001a). [9] E. Ollila, H. Oja, and C. Croux, The affine equivariant sign covariance matrix: asymptotic behaviour and efficiencies. Submitted (2001b). [10] E. Ollila, C. Croux, and H. Oja, Influence functions and asymptotic efficiency of the affine equivariant rank covariance matrix. Submitted (2002a).

Sign and rank covariance matrices

13

[11] E. Ollila, T.P. Hettmansperger, and H. Oja, Estimates of regression coefficients based on the sign covariance matrix. Journal of the Royal Statistical Society, Series B (2002b), to appear. [12] D.E. Tyler, A distribution-free M-estimator of multivariate scatter. The Annals of Statistics 15 (1987), 234-251. [13] S. Visuri, V. Koivunen, and H. Oja , Sign and rank covariance matrices. Journal of Statistical Planning and Inference 91 (2000) , 557-575. [14] S. Visuri, E. Ollila, V. Koivunen, J. M¨ ott¨ onen, and H. Oja, Affine equivariant multivariate rank methods. Journal of Statistical Planning and Inference (2001), to appear. [15] S. Visuri, E. Ollila, V. Koivunen, J. M¨ ott¨ onen, and H. Oja, Affine equivariant multivariate rank methods. Journal of Statistical Planning and Inference (2001), to appear. [16] S. Visuri, H. Oja, and V. Koivunen, Robust subspace DoA estimation for wireless communications. Proceedings of the IEEE 51st Vehicular Technology Conference, Tokyo, Japan, (2000), 2511-2555.

Department of Applied Economics, University of Leuven, Naamsestraat 69, B-3000 Leuven, Belgium E-mail address: [email protected] Department of Mathematics and Statistics, University of Jyv¨askyl¨a, P.O. Box 35, FIN-40351 Jyv¨askyl¨a, Finland E-mail address: [email protected] Department of Mathematics and Statistics, University of Jyv¨askyl¨a, P.O. Box 35, FIN-40351 Jyv¨askyl¨a, Finland E-mail address: [email protected]