Article ID Journal Published Year Pages File Type
10525246 Journal of Statistical Planning and Inference 2005 10 Pages PDF
Abstract
We consider the problem of model (or variable) selection in the classical regression model based on cross-validation with an added penalty term for penalizing overfitting. Under some weak conditions, the new criterion is shown to be strongly consistent in the sense that with probability one, for all large n, the criterion chooses the smallest true model. The penalty function denoted by Cn depends on the sample size n and is chosen to ensure the consistency in the selection of true model. There are various choices of Cn suggested in the literature on model selection. In this paper we show that a particular choice of Cn based on observed data, which makes it random, preserves the consistency property and provides improved performance over a fixed choice of Cn.
Related Topics
Physical Sciences and Engineering Mathematics Applied Mathematics
Authors
, ,