Documente Academic
Documente Profesional
Documente Cultură
Continuous
Review of Probability
Review of Statistics
F = P r(X x)
F =
p(x)dx
Types of Data
Cross-Sectional
f=
Panel Data
Multiple entities, where each entity is observed at multiple time periods
Probability Distribution
Cummulative Probability Distribution
0.8
0.8
0.1
0.9
0.05
0.95
0.05
1.00
d
F = P r(X = x)
dx
2
Best Reference: pg. 35 (Key Concept 2.3) of S&W!
2.2 Expectation
Sum of all possible values a random variable X can take,
weighted by the probability of each outcome.
E[X] =
Notation:
E[ ] =
Consistency: as we approach infinite data, our estimator
gets close to the true parameter i.e.
P r[|
xi P (X = xi )
| < "] ! 1 as N ! 1.
Convergence in probability:
!p
= E[X]
Intuition: when the sample size is large, our estimate is approximately equal to the true parameter
Expectation Property:
Eciency:
- is more ecient than e if it has a lower variance
V ar( ) V ar( )
2.3 Variance
V ar[X] = E[(X
x ) 2 ] =
2
X
)2 P (X = xi )
(xi
V ar( ) V ar( ) 8 e E[ e] =
= V ar[X]
V ar[aX + bY + c] = a2
2
x
+ b2
2
y
+ 2ab
xy
XY
= E[(X
x )(Y
y )]
XY
Corr(X, Y ) =
X Y
Kurtosis: E[(X
1X
xi = x
n i
X
X
2
X)
N (X ,
2
X)
OLS Assumptions
1.is,Conditional
That
we
E[ui |xi ] = 0
2. i.i.d Sampling
he Ordinary Least Squares Estimator (OLS) chooses the regression coefficients (b0 , b1 ) so that the sample regresXi 6= Xj
Theninthe
sample
parameter
is denoted:
on line as depicted
equation
(2) is estimated
as close as possible
to the observed
data collected.
ollectively,
and
rom here on, instead of using the notation (b0 , b1 ), we will use ( 0 , 1 ) to denote the OLS estimators3.
for Large
the
use ui instead of i to denote the estimate of the error
opulation regression parameters ( 0 , 1 ). We will also
erm (or also known as fitted residual). In other words, re-writing our sample regression line, we obtain:
Yi = 0 + 1 Xi + ui
SLR Model
Notice that OLS estimators regression coefficients and the error term for our sample regression line have a hat
Y = 0 + between
uiactual population parameters (without the
1 Xi +the
n top of them. This notation enables us toidifferentiate
Coecient
hats) and the estimated parameters (with the hats) and is also supposed to make it more intuitive
for you
We use some kind of method to get estimators of coecients.
a.k.a 0 is the OLS estimator for 0 ).
model
o formalize what is meant as as close as possible to the observed data collected, the regression coefficients ( 0 ,
1 ) are chosen to minimize the Sum of Squared Residuals (SSR), which is the sum of the differences in the
alues predicted in our estimated linear regression
the
actual values.
Yi = from
i
0 + 1 Xi + u
Mathematically, the minimization of the Sum of Squared Residuals (SSR) is written as:
min
N
X
ui 2 = min
i=1
N
X
(Yi
Y = 0 + 1 X + u
Yi )2
i=1
here the sample regression is denoted in equation (3) and the corresponding predicted value of Yi based
n the sample regression line denoted as:
Yi = 0 + 1 Xi
i model
(Xby
X)(Y
Y ) is known
xy as the fitted
he part of the outcome that is not
explained
=
Pi the estimated
= 2
mathematically as:
2
(Xi X)
x
0 :
residual, written
Yi
ui = Yi
Measures of Fit
Explained Sum of Squares (ESS): the amount of deviation of the predicted values of Y , Y from the sample
averages Y
ESS =
n
X
3
(Yi
(Yi
the
interpretation.
Statistics Review
in X on Y .
, measures
of thechange
tail.
is to say
thattheathickness
one unit
in X should have
Sample Average
a 1 unit change in Y .
3
Kurtosis:
That
Y )2
10/3 andconstant
10/5, 2016 0 may or may not have a useful
E[(X x )4 ]
n
X
SLC Econ
MT#1
Session
the 140
estimator
thatReview
captures the
expected eect on
Y )2
=
For a sample of X1 , , Xn , the sample average is: X
1
n
n
X
Xi
n
n error, also
n known as the dierence be ui : the X
prediction
1 X
1 X
1
= E( 1
E(X)
Xi ) =
E(Xi ) =
x = n x = x
tweenn the
observed
values
of
Y
and
the predicted value.
n
n
n
i=1
i=1
i=1
n
n
n
2
2
X
X
X
If
u
=
0,
it
means
that
our
sample
perfectly
1
1
1
x
2regression
2
= V ar( 1
V ar(X)
Xi ) = 2
V ar(Xi ) = 2
, lim x = 0
x = 2 n x =
n
n!1 n
n
n
n
n
predicts the
regression.
i=1 population
i=1
i=1
i=1
Statistic Theorems
will converge in p
of Large Number: As the sample size n approaches infinity, the sample mean X
Sum of Squared Residuals (SSR): the sum of the dif- Law
bility to the population mean x .
Hypothesis
Tesing
General
form
of
a
t-statistics:
ferences in the values predicted in our estimated linear lim X = x
n!1
regression Y from the actual values Y
Central Limit Theorem: As the number of samples taken approches infinity, the distribution of sample m
approaches normal.
X
estimator hypothesized
n
N (x , 2 )
X
X
t=
x
2
X
standard
error(estimator)
SSR =
(Yi Yi )
Zn =
N (0, 1)
x
Hypothesis Testing
T SS = ESS + SSR
T-statistic: the distance of your observed value from your null hypothesis.
SSR
T SS
P-value: smallest level of significance at which your null hypothesis will be rejected.
Hypothesis testing conditions and conclusions:
If t-stat > critical value: reject the null hypothesis
If p-value < significance value: reject
Calculating Confidence Interval: H0 SE
Critical V alue
Calculating t-statistic:
For a single variable: t =
Observed V alue
Std.Dev
H0
P-value: smallest level of significance at which your null hypothesis will be rejected.
Hypothesis testing conditions and conclusions:
If t-stat > critical value: reject the null hypothesis
Critical V alue
Calculating t-statistic:
For a single variable: t =
Observed V alue
Std.Dev
H0
1
x
Sx 2
Nx
y
+
dent from
Y and Cov(X,
Z)variance]
= 3 while
Z) Z
= be7.random va
1. [Properties
of expected
values and
LetCov(Y,
X, Y and
N (3,Using
25), Y the
N
(1, 4) and Z given
N (16,
1). X
is independent from Y and C
information
above,
compute
Cov(Y, Z) = 7. Using the information given above, compute
null
SE
Sy 2
Ny
1
N
1
N
N
P
N
P
Xi
Y =
Yi
i=1
Xi
i=1
1
N
N
P
i=1
Xi2
1
N
N
P
Xi Y i
Y
X
i=1
X2
3. [Probability distribution] Let Y denote the number of heads that occur when t
2.
of Estimators
(a)Properties
Derive the probability
distribution of Y .
(b)
Derive
Let the cumulative probability distribution of Y .
(c) Compute the mean and
of1 X
Y .i 140
Yi variance
= 0 -+Econ
+ ui
Review
Exercises
bwght = 119.77
0.514 cigs
1. [Unbiasedness] Assume Xi , i = 1, ..., N , iid random variables. Are the following estimators biased? If
so, calculate the bias.
(a) M1 =
2X1 +3X2
(a) What 5isN the predicted birth weight when cigs = 0? What about when ci
P
(b) M2day)?
= N1+1 Comment
Xi
on the dierence.
i=1
=
(c) M3 = X
Thanks
for previous
sharing
his section
2.
[Eciency]
Consider GSI
Xi , iEdson
= 1, ...,Severnini
N , iid random
variables.
The notes.
estimator M1 =
= X1 +X2 ? Justify your answer.
ecient that X
2
2X1 +3X2
5
is more
20
30
60
70
55
65
90
95
5
Chapter 3: Statistics
6
Chapter 5: Hypothesis Tesing