Estimation

이하와 같은 linear model 고려. 이때 의 i번째 row vector이며, .

Identifiability and Estimability

Identifiable

모델에서의 무한한 갯수의 관측치를 보유한다면, 모델의 underlying 패러미터의 참값을 획득하는 것이 가능한 성질.

A general linear model is a parameterization

The parameter is identifiable if for any and implies . If is identifiable, we say that the parameterization is identifiable. (패러미터 가 identifiable하다면, 우리는 해당 패러미터의 parameterization 또한 identifiable 하다) Moreover, a vector-valued function is identifiable if implies .

For regression models for which , the parameters are identifiable: is nonsingular, so if , then

A function is identifiable is a function of .

Estimable

The results in the last section suggest that some linear combinations of in the less than full rank case will not be estimable.

The linear parametric function is an estimable function if there exists a vector such that .

A vector-valued linear function of , is estimable if for some matrix P; In other words, is estimable if .

Clearly, if is estimable, it is identifiable and therefore it is a reasonable thing to estimate.

  • estimable identifiable

For estimable functions , although need not be unique, its perpendicular projection (columnwise) onto is unique: let be matrices with , then

  • Example 2.1.4 and 2.1.5

‘s estimate, , is unbiased if .

if for some scalar and vector , is a linear estimate of .

if and ; say, , then a linear estimate is unbiased

is estimable there exists such that for any .


Estimation: Least Squares

Estimating is to take a vector in closest to ;

for any Least Squares Estimate , LSE of , e.g., .

  • Theorem 2.2.1

where is the perpendicular projection operator onto , then

is a LSE of

  • Corollary 2.2.2

  • Corollary 2.2.3

The unique LSE of .

※ Note: the unique LSE of .

  • Theorem 2.2.4

the LSE of is unique only if is estimable: if , so that .

※ Note: When is not identifiable, we need side conditions imposed on the parameters to estimate nonidentifiable parameters.

※ Note: With (overparameterized model), we need individual side conditions to identify and estimate the parameters.

  • Proposition 2.2.5

If , then .

let’s decompose

이때

  • Theorem 2.2.6

Let and . At below formula, denominator is degrees of freedom for error.

Then an UE of , MSE, is as below.


Estimation: Best Linear Unbiased

  • Definition 2.3.1

is a Best Linear Unbiased Estimate(BLUE) of if is unbiased.

e.g., and if for any other linear unbiased estimate , .

  • Theorem 2.3.2: Gauss-Markov thm

Consider with , . Let be estimable.

Then LSE of BLUE of .

  • Corollary 2.3.3

Let . Then there exists a unique BLUE for any estimable function .


Estimation: Maximum Likelihood

Assume that . Then the Maximum Likelihood Estimates (MLEs) of and are obtained by maximizing the log of the likelihood so that


Estimation: Minimum Variance Unbiased

Assume that with .

if implies that , A vector-valued sufficient statistic is said to be complete

If is a complete sufficient statistic, then is a Minimum Variance Unbiased Estimate (MVUE) of .

  • Theorem 2.5.3

let and let be a rvec with pdf as below. then is a complete sufficient statistics provided that neither nor satisfies any linear constraints.

  • Theorem 2.5.4

MSE is a , and whenever .


Sampling Distributions of Estimates

Assume that with . Then . then

\begin{alignat}{4} \Lambda ' \hat \beta &= P' M Y &&\sim N(\Lambda ' \beta , \; &&\sigma^2 P'MP&&\; \; \; ) && \; \; \; \; \; \; \; \; \; \;&& && && \\ & &&\sim N(\Lambda ' \beta , \; &&\sigma^2 \Lambda ' (X'X)^{-} \Lambda&&\; \; \; ) && && \because && \;M && =X(X'X)^- X' \\ & && && && && && && \; \hat Y && = MY &&\sim N(X\beta, \sigma^2 M) \\ \hat \beta &= (X'X)^- X'Y &&\sim N(\beta , \; &&\sigma^2 (X'X)^{-1}) && && && && && && (\text{if X is of full rank}) \end{alignat}

Do Exercise 2.1. Show that


Generalized Least Squares(GLS)

Assume that for some known positive definite ,

\begin{alignat}{3} Y &= X \beta &&+ \epsilon && \; \; \; \; \; \; \; \; \; \; && E(\epsilon)&&=0, \; \; &&\; Cov(\epsilon) &&= \sigma^2 \Sigma \tag{1} \\ \Sigma^{-\tfrac{1}{2}}Y &= \Sigma^{-\tfrac{1}{2}} X \beta &&+ \Sigma^{-\tfrac{1}{2}} \epsilon && \; \; \; \; \; \; \; \; \; \; && E(\Sigma^{-\tfrac{1}{2}} \epsilon)&&=0, &&\; Cov(\Sigma^{-\tfrac{1}{2}} \epsilon) &&= \sigma^2 I \tag{2, by SVD} \\ Y_\ast &= X_\ast \beta &&+ \epsilon_\ast && \; \; \; \; \; \; \; \; \; \; && E( \epsilon_\ast)&&=0, &&\; Cov( \epsilon_\ast) &&= \sigma^2 I \end{alignat}
  • Theorem 2.7.1
  1. estimable in model (1) if is estimable in model (2).
  2. is GLSE of , which is Normal Equation of GLS.
  • For any estimable function, there exists a unique GLSE.
  1. GLSE estimate of estimable , is BLUE of .
  2. let . then, GLSE of estimable , is MVUE.
  3. let . then, .

Normal Equation of GLS can be rewritten as

is a projection operator onto .

Let be estimable. Then .

  • Note: is residual vector of GLSE.

denominator는 .

Let be nonsingular and . Then least squares estimates are BLUEs.

  • Note: for diagonal , GLS is referred to as Weighted Least Squares (WLS).

  • Exercise 2.5.

Show that is the perpendicular projection operator onto when the inner product between two vectors and is defined as .