ABSTRACT

So far we have approached the linear model mostly as a method of mathematical approximation. In this chapter, we pose the Gauss-Markov model, which embodies the most common assumptions for the statistical approach to the linear model, leading to the Gauss-Markov Theorem. The Gauss-Markov model takes the form

y = Xb+ e (4.1) where y is the (N × 1) vector of observed responses and X is the (N × p) known design matrix. As before, the coefficient vector b is unknown and to be determined or estimated. The main features of the Gauss-Markov model are the assumptions on the error e:

E(e) = 0 and Cov (e) = σ 2IN . (4.2) The notation for expectation and covariances above can be rewritten component by component:

(E(e))i = i th component of E(e) = E(ei ) (Cov (e))i j = i, j th element of covariance matrix = Cov (ei , e j )

so that the Gauss-Markov assumptions can be rewritten as

E(ei ) = 0, i = 1, . . . , N

Cov(ei , e j ) = {

σ 2 for i = j 0 for i = j

,

that is, the errors in the model have a zero mean, constant variance, and are uncorrelated. An alternative view of the Gauss-Markov model does not employ the error vector e:

E(y) = Xb,Cov(y) = σ 2IN . The assumptions in the Gauss-Markovmodel are easily acceptable for many practical problems, and deviations from these assumptions will be considered in more detail later.