کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
5045595 | 1475849 | 2018 | 7 صفحه PDF | دانلود رایگان |
- Psychologists do not pay much attention to multivariate outliers.
- When they do they do not report the method used to detect these outliers or use questionable methods.
- We highlight the shortcomings of the currently used methods.
- We propose a robust alternative, namely the Minimum Covariance Determinant.
- We explain its computation in R and SPSS softwares.
A look at the psychology literature reveals that researchers still seem to encounter difficulties in coping with multivariate outliers. Multivariate outliers can severely distort the estimation of population parameters. Detecting multivariate outliers is mainly disregarded or done by using the basic Mahalanobis distance. However, that indicator uses the multivariate sample mean and covariance matrix that are particularly sensitive to outliers. Hence, this method is problematic. We highlight the disadvantages of the basic Mahalanobis distance and argue instead in favor of a robust Mahalanobis distance. In particular, we present a variant based on the Minimum Covariance Determinant, a more robust procedure that is easy to implement. Using Monte Carlo simulations of bivariate sample distributions varying in size (ns = 20, 100, 500) and population correlation coefficient (Ï = .10, .30, .50), we demonstrate the detrimental impact of outliers on parameter estimation and show the superiority of the MCD over the Mahalanobis distance. We also make recommendations for deciding whether to include vs. exclude outliers. Finally, we provide the procedures for calculating this indicator in R and SPSS software.
Journal: Journal of Experimental Social Psychology - Volume 74, January 2018, Pages 150-156