• 検索結果がありません。

When these conditions hold, we can rewrite e β

N/A
N/A
Protected

Academic year: 2021

シェア "When these conditions hold, we can rewrite e β"

Copied!
4
0
0

読み込み中.... (全文を見る)

全文

(1)

When these conditions hold, we can rewrite e β

2

as:

e β

2

= β

2

+

X

n i=1

i

+ d

i

)u

i

. The variance of e β

2

is derived as:

V(e β

2

) = V β

2

+

X

n i=1

i

+ d

i

)u

i

= V X

n

i=1

i

+ d

i

)u

i

= X

n

i=1

V

i

+ d

i

)u

i

= X

n

i=1

i

+ d

i

)

2

V(u

i

) = σ

2

( X

n

i=1

ω

2i

+ 2 X

n

i=1

ω

i

d

i

+ X

n

i=1

d

2i

)

= σ

2

( X

n

i=1

ω

2i

+ X

n

i=1

d

2i

).

24

From unbiasedness ofe β

2

, using P

n

i=1

d

i

= 0 and P

n

i=1

d

i

x

i

= 0, we obtain:

X

n i=1

ω

i

d

i

= P

n

i=1

(x

i

− x)d

i

P

n

i=1

(x

i

− x)

2

= P

n

i=1

x

i

d

i

− X P

n i=1

d

i

P

n

i=1

(x

i

− x)

2

= 0,

which is utilized to obtain the variance of e β

2

in the third line of the above equation.

From (15), the variance of ˆ β

2

is given by: V( ˆ β

2

) = σ

2

P

n i=1

ω

2i

. Therefore, we have:

V(e β

2

) ≥ V( ˆ β

2

), because of P

n

i=1

d

2i

≥ 0.

25

When P

n

i=1

d

i2

= 0, i.e., when d

1

= d

2

= · · · = d

n

= 0, we have the equality: V(e β

2

)

= V( ˆ β

2

).

Thus, in the case of d

1

= d

2

= · · · = d

n

= 0, ˆ β

2

is equivalent to e β

2

.

As shown above, the least squares estimator ˆ β

2

gives us the minimum variance linear unbiased estimator ( ࠷খ෼ࢄઢܗෆภਪఆྔ ), or equivalently the best linear unbiased estimator ( ࠷ྑઢܗෆภਪఆྔɼ BLUE), which is called the Gauss-Markov theorem ( Ψ΢εɾϚϧίϑఆཧ ).

26

Asymptotic Properties of ˆ β

2

: We assume that as n goes to infinity we have the following:

1 n

X

n i=1

(x

i

− x)

2

−→ m < ∞ , where m is a constant value. From (12), we obtain:

n X

n

i=1

ω

2i

= 1 (1/n) P

n

i=1

(x

i

− x) −→ 1

m .

Note that f(x

n

) −→ f(m) when x

n

−→ m, called Slutsky’s theorem ( εϧπΩʔ ఆཧ ), where m is a constant value and f( · ) is a function.

27

We show both consistency of ˆ β

2

and asymptotic normality of √ n( ˆ β

2

− β

2

).

˔ First, we prove that ˆ β

2

is a consistent estimator of β

2

. Chebyshev’s inequality is given by:

P( | X − µ | > ) ≤ σ

2

2

, where µ = E(X) and σ

2

= V(X).

Replace X, E(X) and V(X) by:

β ˆ

2

, E( ˆ β

2

) = β

2

, and V( ˆ β

2

) = σ

2

X

n

i=1

ω

2i

= σ

2

P

n

i=1

(x

i

− x) , respectively.

28

Then, when n −→ ∞ , we obtain the following result:

P( | β ˆ

2

− β

2

| > ) ≤ σ

2

P

n i=1

ω

2i

2

= σ

2

n P

n

i=1

ω

2i

n

2

−→ 0, where P

n

i=1

ω

2i

−→ 0 because n P

n

i=1

ω

2i

−→ 1

m from the assumption.

Thus, we obtain the result that ˆ β

2

−→ β

2

as n −→ ∞ .

Therefore, we can conclude that ˆ β

2

is a consistent estimator of β

2

.

29

(2)

˔ Next, we want to show that √

n( ˆ β

2

− β

2

) is asymptotically normal.

Note that ˆ β

2

= β

2

+ P

n

i=1

ω

i

u

i

as in (13).

From the central limit theorem, asymptotic normality is shown as follows:

P

n

i=1

ω

i

u

i

− E( P

n i=1

ω

i

u

i

) p V( P

n

i=1

ω

i

u

i

) =

P

n i=1

ω

i

u

i

σ qP

n i=1

ω

2i

= β ˆ

2

− β

2

σ/ pP

n

i=1

(x

i

− x)

2

−→ N(0,1), where E( P

n

i=1

ω

i

u

i

) = 0, V( P

n

i=1

ω

i

u

i

) = σ

2

P

n

i=1

ω

2i

, and P

n

i=1

ω

i

u

i

= β ˆ

2

− β

2

are substituted in the first and second equalities.

30

Moreover, we can rewrite as follows:

β ˆ

2

− β

2

σ/ pP

n

i=1

(x

i

− x)

2

=

√ n( ˆ β

2

− β

2

) σ/ p

(1/n) P

n

i=1

(x

i

− x)

2

−→

√ n( ˆ β

2

− β

2

) σ/ √

m −→ N(0, 1), or equivalently,

√ n( ˆ β

2

− β

2

) −→ N(0, σ

2

m ).

Thus, the asymptotic normality of √

n( ˆ β

2

− β

2

) is shown.

31

Finally, replacing σ

2

by its consistent estimator s

2

, it is known as follows:

β ˆ

2

− β

2

s/ pP

n

i=1

(x

i

− x)

2

−→ N(0,1), (16) where s

2

is defined as:

s

2

= 1 n − 2

X

n i=1

e

2i

= 1 n − 2

X

n i=1

(y

i

− β ˆ

1

− β ˆ

2

x

i

)

2

, (17) which is a consistent and unbiased estimator of σ

2

. −→ Proved later.

Thus, using (16), in large sample we can construct the confidence interval and test the hypothesis.

32

Exact Distribution of ˆ β

2

: We have shown asymptotic normality of √ n( ˆ β

2

− β

2

), which is one of the large sample properties.

Now, we discuss the small sample properties of ˆ β

2

.

In order to obtain the distribution of ˆ β

2

in small sample, the distribution of the error term has to be assumed.

Therefore, the extra assumption is that u

i

∼ N(0, σ

2

).

Writing (13), again, ˆ β

2

is represented as:

β ˆ

2

= β

2

+ X

n

i=1

ω

i

u

i

.

First, we obtain the distribution of the second term in the above equation.

33

Using the moment-generating function, P

n

i=1

ω

i

u

i

is distributed as:

X

n i=1

ω

i

u

i

∼ N (0, σ

2

X

n

i=1

ω

2i

).

Therefore, ˆ β

2

is distributed as:

β ˆ

2

= β

2

+ X

n

i=1

ω

i

u

i

∼ N(β

2

, σ

2

X

n

i=1

ω

2i

), or equivalently,

β ˆ

2

− β

2

σ qP

n i=1

ω

2i

= β ˆ

2

− β

2

σ/ pP

n

i=1

(x

i

− x)

2

∼ N(0,1), for any n.

34

Moreover, replacing σ

2

by its estimator s

2

defined in (17), it is known that we have:

β ˆ

2

− β

2

s/ pP

n

i=1

(x

i

− x)

2

∼ t(n − 2),

where t(n − 2) denotes t distribution with n − 2 degrees of freedom.

Thus, under normality assumption on the error term u

i

, the t(n − 2) distribution is used for the confidence interval and the testing hypothesis in small sample.

Or, taking the square on both sides, β ˆ

2

− β

2

s/ pP

n i=1

(x

i

− x)

2

2

∼ F(1,n − 2), which will be proved later.

35

(3)

Before going to multiple regression model ( ॏճؼϞσϧ ),

2 Some Formulas of Matrix Algebra

1. Let A =

 

 

 

a

11

a

12

· · · a

1k

a

21

a

22

· · · a

2k

.. . .. . . .. .. . a

l1

a

l2

· · · a

lk

 

 

 

= [a

i j

],

which is a l × k matrix, where a

i j

denotes ith row and jth column of A.

36

The transposed matrix ( సஔߦྻ ) of A, denoted by A

0

, is defined as:

A

0

=

 

 

 

a

11

a

21

· · · a

l1

a

12

a

22

· · · a

l2

.. . .. . . .. .. . a

1k

a

2k

· · · a

lk

 

 

 

= [a

ji

],

where the ith row of A

0

is the ith column of A.

2. (Ax)

0

= x

0

A

0

,

where A and x are a l × k matrix and a k × 1 vector, respectively.

37

3. a

0

= a,

where a denotes a scalar.

4. ∂a

0

x

∂x = a,

where a and x are k × 1 vectors.

5. ∂x

0

Ax

∂x = (A + A

0

)x,

where A and x are a k × k matrix and a k × 1 vector, respectively.

38

Especially, when A is symmetric,

∂x

0

Ax

∂x = 2Ax.

6. Let A and B be k × k matrices, and I

k

be a k × k identity matrix ( ୯Ґߦྻ ) (one in the diagonal elements and zero in the other elements).

When AB = I

k

, B is called the inverse matrix ( ٯߦྻ ) of A, denoted by B = A

−1

.

That is, AA

−1

= A

−1

A = I

k

. 39

7. Let A be a k × k matrix and x be a k × 1 vector.

If A is a positive definite matrix ( ਖ਼஋ఆූ߸ߦྻ ), for any x except for x = 0 we have:

x

0

Ax > 0.

If A is a positive semidefinite matrix ( ඇෛ஋ఆූ߸ߦྻ ), for any x except for x = 0 we have:

x

0

Ax ≥ 0.

40

If A is a negative definite matrix ( ෛ஋ఆූ߸ߦྻ ), for any x except for x = 0 we have:

x

0

Ax < 0.

If A is a negative semidefinite matrix ( ඇਖ਼஋ఆූ߸ߦྻ ), for any x except for x = 0 we have:

x

0

Ax ≤ 0.

41

(4)

Trace, Rank and etc.: A : k × k, B : n × k, C : k × n.

1. The trace ( τϨʔε ) of A is: tr(A) = X

k

i=1

a

ii

, where A = [a

i j

] .

2. The rank ( ϥϯΫɼ֊਺ ) of A is the maximum number of linearly indepen- dent column (or row) vectors of A, which is denoted by rank(A).

3. If A is an idempotent matrix ( ΂͖౳ߦྻ ), A = A

2

.

42

4. If A is an idempotent and symmetric matrix, A = A

2

= A

0

A .

5. A is idempotent if and only if the eigen values of A consist of 1 and 0.

6. If A is idempotent, rank(A) =tr(A) .

7. tr(BC) =tr(CB)

43

Distributions in Matrix Form:

1. Let X, µ and Σ be k × 1, k × 1 and k × k matrices.

When X ∼ N (µ, Σ), the density function of X is given by:

f (x) = 1

(2π)

k/2

|Σ| exp − 1

2 (x − µ)

0

Σ

−1

(x − µ)

! .

E(X) = µ and V(X) = E

(X − µ)(X − µ)

0

= Σ The moment-generating function: φ(θ) = E

exp(θ

0

X)

= exp(θ

0

µ +

12

θ

0

Σθ)

44

参照

関連したドキュメント

In this paper we give the Nim value analysis of this game and show its relationship with Beatty’s Theorem.. The game is a one-pile counter pickup game for which the maximum number

The problem is modelled by the Stefan problem with a modified Gibbs-Thomson law, which includes the anisotropic mean curvature corresponding to a surface energy that depends on

In this work we study spacelike and timelike surfaces of revolution in Minkowski space E 3 1 that satisfy the linear Weingarten relation aH + bK = c, where H and K denote the

so.. Then the point [−4, 2t] can always be in a system of generators.. Let us now concentrate on integral points. When the rank is one, the structure of the Mordell-Weil group is

Using general ideas from Theorem 4 of [3] and the Schwarz symmetrization, we obtain the following theorem on radial symmetry in the case of p &gt; 1..

Kilbas; Conditions of the existence of a classical solution of a Cauchy type problem for the diffusion equation with the Riemann-Liouville partial derivative, Differential Equations,

Here we do not consider the case where the discontinuity curve is the conic (DL), because first in [11, 13] it was proved that discontinuous piecewise linear differential

It turns out that the symbol which is defined in a probabilistic way coincides with the analytic (in the sense of pseudo-differential operators) symbol for the class of Feller