Article ID Journal Published Year Pages File Type
5773033 Linear Algebra and its Applications 2018 22 Pages PDF
Abstract
We consider a general binary model for which conditional probability of success given vector of predictors X equals q(β1TX,…,βkTX) and a family of possibly misspecified logistic regressions fitted to it. In the case when X satisfies linearity condition we show that their algebraic structure is uniquely determined and that the vector β⁎ corresponding to Kullback-Leibler projection on this family is a linear combination of β1,…,βk. This generalizes the known result proved by P. Ruud for k=1 which says that the true and projected vectors are collinear. It also follows that the projected vector has the same direction as the first canonical vector which justifies frequent observations that logistic fit yields well performing classifiers even if misspecification is expected. In the special case of additive binary model with multivariate normal predictors and when response function q is a convex combination of univariate responses we show that the variance of β⁎TX is not larger than the maximal variance of the projected linear combinations for the corresponding univariate problems. In the case of balanced additive logistic model it follows that the contribution of βi to β⁎ is bounded by the corresponding coefficient in the convex representation of response function q.
Related Topics
Physical Sciences and Engineering Mathematics Algebra and Number Theory
Authors
, ,