next up previous


Postscript version of these notes

STAT 350: Lecture 35

Estimating equations: an introduction via glm

Estimating Equations: refers to equations of the form

\begin{displaymath}h(X,\theta) = 0
\end{displaymath}

which are solved for $\theta$ to get estimates $\hat\theta$. Examples:

1.
The normal equations in linear regression:

\begin{displaymath}X^TY-X^TX\beta=0
\end{displaymath}

2.
The likelihood equations:

\begin{displaymath}\frac{\partial\ell}{\partial\theta} = 0
\end{displaymath}

where $\ell(\theta)$ is the log-likelihood.

3.
The equation which must be solved to do non-linear least squares:

\begin{displaymath}\sum (Y_i-\mu_i)\frac{\partial\mu_i}{\partial\theta}=0
\end{displaymath}

4.
The iteratively reweighted least squares estimating equation:

\begin{displaymath}\sum \frac{Y_i-\mu_i}{\sigma_i^2}\frac{\partial\mu_i}{\partial\theta}=0
\end{displaymath}

where, in a generalized linear model the variance $\sigma_i^2$ is a known (except possibly for a multiplicative constant) function of $\mu_i$.

Only the first of these equations can usually be solved analytically. In Lecture 34 I showed you an example of an iterative technique of solving such equations.

Theory of Generalized Linear Models

The likelihood function for a Poisson regression model is:

\begin{displaymath}L(\beta) = \prod \frac{\mu_i^{y_i}}{y_i!} \exp(-\sum\mu_i)
\end{displaymath}

and the log-likelihood is

\begin{displaymath}\sum y_i \log\mu_i -\sum \mu_i - \sum \log(y_i!)
\end{displaymath}

A typical glm model is

\begin{displaymath}\mu_i = \exp(x_i^T\beta)
\end{displaymath}

where the xi are covariate values for the ith observation (often including an intercept term just as in standard linear regression).

In this case the log-likelihood is

\begin{displaymath}\sum y_i x_i^T\beta- \sum \exp(x_i^T\beta) - \sum\log(y_i!)
\end{displaymath}

which should be treated as a function of $\beta$ and maximized.

The derivative of this log-likelihood with respect to $\beta_k$ is

\begin{displaymath}\sum y_i x_{ik} - \sum \exp(x_i^T\beta) x_{i,k} = \sum (y_i -
\mu_i) x_{i,k}
\end{displaymath}

If $\beta$ has p components then setting these p derivatives equal to 0 gives the likelihood equations.

For a Poisson model the variance is given by

\begin{displaymath}\sigma_i^2 = \mu_i = \exp(x_i^T\beta)
\end{displaymath}

so the likelihood equations can be written as

\begin{displaymath}\sum \frac{(y_i -\mu_i)x_{i,k}\mu_i}{\mu_i} = \sum \frac{(y_i -\mu_i)}{\sigma_i^2}
\frac{\partial\mu_i}{\partial\beta_k} = 0
\end{displaymath}

which is the fourth equation above.

These equations are solved iteratively, as in non-linear regression, but with the iteration now involving weighted least squares. The resulting scheme is called iteratively reweighted least squares.

1.
Begin with a guess for the standard deviations $\sigma_i$ (taking them all equal to 1 is simple).

2.
Do (non-linear) weighted least squares using the guessed weights. Get estimated regression parameters $\hat\beta^{(0)}$.

3.
Use these to compute estimated variances $\hat\sigma_i^2$. Go back to do weighted least squares with these weights and get $\hat\beta^{(1)}$.

4.
Iterate (repeat over and over) until estimates not really changing.

If the $\hat\beta^{(k)}$ converge as $k\to\infty$ to something, say, $\hat\beta$ then since

\begin{displaymath}\sum \left[\frac{y_i -\mu_i(\hat\beta^{(k+1)})}{\sigma_i^2(\h...
...partial\mu_i(\hat\beta^{(k+1)})}{\partial\hat\beta^{(k+1)}}= 0
\end{displaymath}

we learn that $\hat\beta$ must be a root of the equation

\begin{displaymath}\sum
\left[\frac{y_i -\mu_i(\hat\beta)}{\sigma_i^2(\hat\beta)}\right]
\frac{\partial\mu_i(\hat\beta)}{\partial\hat\beta}
= 0
\end{displaymath}

which is the last of our example estimating equations.

Distribution of Estimators

Distribution Theory is the subject of computing the distribution of statistics, estimators and pivots. Examples in this course are the Multivariate Normal Distribution, the theorems about the chi-squared distribution of quadratic forms, the theorems that F statistics have F distributions when the null hypothesis is true, the theorems that show a t pivot has a t distribution.

Exact Distribution Theory: name applied to exact results such as those in previous example when the errors are assumed to have exactly normal distributions.

Asymptotic or Large Sample Distribution Theory: same sort of conclusions but only approximately true and assuming n is large. Theorems of the form:

\begin{displaymath}\lim_{n\to\infty} P(T_n \le t) = F(t)
\end{displaymath}

Sketch of reasoning in special case

POISSON EXAMPLE: p=1

Assume Yi has a Poisson distribution with mean $\mu_i=e^{x_i\beta}$ where now $\beta$ is a scalar.

The estimating equation (the likelihood equation) is

\begin{displaymath}U(\beta)=h(Y_1,\ldots,Y_n,\beta)= \sum (Y_i-e^{x_i\beta})x_i = 0
\end{displaymath}

It is now important to distinguish between a value of $\beta$ which we are trying out in the estimating equation and the true value of $\beta$ which I will call $\beta_0$. If we happen to try out the true value of $\beta$ in U then we find

\begin{displaymath}E_{\beta_0}(U(\beta_0)) = \sum x_iE_{\beta_0}(Y_i-\mu_i) = 0
\end{displaymath}

On the other hand if we try out a value of $\beta$ other than the correct one we find

\begin{displaymath}E_{\beta_0}(U(\beta)) = \sum x_i(e^{x_i\beta} - e^{x_i\beta_0}) \neq 0 \, .
\end{displaymath}

But $U(\beta)$ is a sum of independent random variables so by the law of large numbers (law of averages) must be close to its expected value. This means: if we stick in a value of $\beta$ far from the right value we will not get 0 while if we stick in a value of $\beta$ close to the right answer we will get something close to 0. This can sometimes be turned in to the assertion:

The glm estimate of $\beta$ is consistent, that is, it converges to the correct answer as the sample size goes to $\infty$.

The next theoretical step is another linearization. If $\hat\beta$is the root of the equation, that is, $U(\hat\beta)=0$, then

\begin{displaymath}0 = U(\hat\beta) \approx U(\beta_0) + (\hat\beta-\beta_0) U^\prime
(\beta_0)
\end{displaymath}

This is a Taylor's expansion. In our case the derivative $U^\prime$is

\begin{displaymath}U^\prime(\beta)= -\sum x_i^2 e^{x_i\beta}
\end{displaymath}

so that approximately

\begin{displaymath}\hat\beta = \frac{\sum(Y_i-\mu_i)x_i}{\sum x_i^2 e^{x_i\beta_0}}
\end{displaymath}

The right hand side of this formula has expected value 0, variance

\begin{displaymath}\frac{\sum x_i^2 Var(Y_i)}{\left(\sum x_i^2 e^{x_i\beta_0}\right)^2}
\end{displaymath}

which simplifies to

\begin{displaymath}\frac{1}{\sum x_i^2 e^{x_i\beta_0}}
\end{displaymath}

This means that an approximate standard error of $\hat\beta$ is

\begin{displaymath}\frac{1}{\sqrt{\sum x_i^2 e^{x_i\beta_0}}}
\end{displaymath}

that an estimated approximate standard error is

\begin{displaymath}\frac{1}{\sqrt{\sum x_i^2 e^{x_i\hat\beta}}}
\end{displaymath}

Finally, since the formula shows that $\hat\beta-\beta_0$ is a sum of independent terms the central limit theorem suggests that $\hat\beta$has an approximate normal distribution and that

\begin{displaymath}\sqrt{\sum x_i^2 e^{x_i\hat\beta}}(\hat\beta-\beta_0)
\end{displaymath}

is an approximate pivot with approximately a N(0,1) distribution. You should be able to turn this assertion into a 95% (approximate) confidence interval for $\beta_0$.

Scope of these ideas

The ideas in the above calculation can be used in many contexts.

Further exploration of the ideas in this course


next up previous



Richard Lockhart
1999-03-24