Table of Contents
Fetching ...

The exact amount of t-ness that the normal model can tolerate

Nils Lid Hjort

Abstract

Suppose that the normal model is used for data $Y_1,\ldots,Y_n$, but that the true distribution is a t-distribution with location and scale parameters $ξ$ and $σ$ and $m$ degrees of freedom. The normal model corresponds to $m=\infty$. Using a local asymptotic framework where $m$ is allowed to increase with $n$ two classes of estimands are identified. One small class, which in particular contains the functions of $ξ$ alone, is only affected by t-ness to the second order, and maximum likelihood estimation in the two- or three-parameter models become equivalent. For all other estimands it is shown that if $m\ge1.458\sqrt{n}$, then maximum likelihood estimation using the incorrect normal model is still more precise than using the correct three-parameter model. This is furthermore shown to be true in regression models with t-distributed residuals. We also propose and analyse compromise estimators that in various ways interpolate between the normal and the nonnormal models. A separate section extends the t-ness results to general normal scale mixtures, in which case the tolerance radius around the normal error distribution takes the form of an upper bound $0.3429/\sqrt{n}$ for the variance of the scale mixture distribution. Proving our results requires somewhat nonstandard `corner asymptotics' since behaviour of estimators must be studied when the crucial parameter $γ=1/m$ is close to zero, which is not an inner point of the parameter space, and the maximum likelihood estimator of $m$ is equal to $\infty$ with positive probability.

The exact amount of t-ness that the normal model can tolerate

Abstract

Suppose that the normal model is used for data , but that the true distribution is a t-distribution with location and scale parameters and and degrees of freedom. The normal model corresponds to . Using a local asymptotic framework where is allowed to increase with two classes of estimands are identified. One small class, which in particular contains the functions of alone, is only affected by t-ness to the second order, and maximum likelihood estimation in the two- or three-parameter models become equivalent. For all other estimands it is shown that if , then maximum likelihood estimation using the incorrect normal model is still more precise than using the correct three-parameter model. This is furthermore shown to be true in regression models with t-distributed residuals. We also propose and analyse compromise estimators that in various ways interpolate between the normal and the nonnormal models. A separate section extends the t-ness results to general normal scale mixtures, in which case the tolerance radius around the normal error distribution takes the form of an upper bound for the variance of the scale mixture distribution. Proving our results requires somewhat nonstandard `corner asymptotics' since behaviour of estimators must be studied when the crucial parameter is close to zero, which is not an inner point of the parameter space, and the maximum likelihood estimator of is equal to with positive probability.

Paper Structure

This paper contains 5 sections, 66 equations.