Sunteți pe pagina 1din 32

Parameter Estimation, Sampling

Distributions & Hypothesis Testing


Estimation and CLT

Parameter Estimation & Hypothesis Testing


• In doing research, we are usually interested in some feature of a
population distribution (which can be described using population
parameters)

• Since populations are difficult (or impossible) to collect data on, we


estimate population parameters using point estimates based on
sample statistics

• Sample statistics vary from sample to sample, making point estimates


variable and unreliable

• The distribution of a statistic (estimate) computed across many different


samples is called the sampling distribution of that statistic (estimate)

• We can use the sampling distribution to estimate the likelihood


associated with a hypothesized population parameter or the margin of
error (or confidence interval) associated with a point estimate

01:830:400 Spring 2019 2


Estimation and CLT

characterized by
population parameters

characterized by
sample statistics

01:830:400 Spring 2019 3


Estimation and CLT

Law of Large Numbers

Mean Sample Age vs. Size


1 n
Let X n   Xi,
n i

and

k
  E[ X ]   pi X i ,
i

then

n    
lim  P X n       1

01:830:400 Spring 2019 4


Estimation and CLT

Sampling Distributions
• How reliable are sample statistics (as estimators) for a finite
sample size?

01:830:400 Spring 2019 5


Estimation and CLT

Central Limit Theorem


Thanks to the central limit theorem we can compute the
sampling distribution of the mean without having to actually draw
samples and compute sample means.
• Central limit theorem:

– Given a population with mean µ and standard deviation σ, the sampling


distribution of the mean (i.e., the distribution of sample means) will itself
have a mean of µ and a standard deviation (standard error) of  / n

– Furthermore, whatever the distribution of the parent population, this


sampling distribution will approach the normal distribution as the sample
size (n) increases.

01:830:400 Spring 2019 6


Estimation and CLT

Standard Error
• Just as the standard deviation (σ)of a population of scores
provides a measure of the average distance between an
individual score (x) and the population mean (µ), the standard
error (𝜎𝑋ത ) provides a measure of the average distance
between the sample mean (𝑋) ത and the population mean (µ).


X 
n

01:830:400 Spring 2019 7


Estimation and CLT

Hypothesis Testing
• Procedure for traditional (NHST) hypothesis testing

• Roots
– Significance testing: (Karl) Pearson & Fisher
– Decision-theoretic hypothesis testing: Neyman & (Egon) Pearson

• Logic of the individual and combined approaches

01:830:400 Spring 2019 8


Estimation and CLT

Traditional (NHST) Hypothesis Testing


1. Begin with a research hypothesis H1 (defined in terms of population
parameters)

2. Set up the null hypothesis H0

3. Construct the sampling distribution of a particular statistic under the


assumption that the null hypothesis is true

4. Collect some data and use it to compute a sample statistic

5. Compare the sample statistic to the distribution constructed in step (3)

6. Reject or retain H0 depending on the probability, under H0, of obtaining a


sample statistic as extreme as the one we observed

01:830:400 Spring 2019 9


Estimation and CLT

Roots: Inferential Significance Testing


• Significance ‘Testing’, as conceived by Fisher (and Karl
Pearson) was conceived as a heuristic for building an
inductive case for or against a particular model

– Pearson (1900) conceived of p (essentially equivalent to a modern two-


tailed p-value) as an index of the validity of a hypothesis. He later
(1914) popularized this index by publishing tables of this value for a
number of standard distributions

– Fisher (1925) suggests using p = 0.05 (or some smaller value) as a


heuristic to determine whether to further consider the results of an
experiment

• The ideas of the p-value, of the null hypothesis, and of


significance come from this approach

01:830:400 Spring 2019 10


Estimation and CLT

Roots: Decision-Theoretic Hypothesis Testing


• Hypothesis testing was conceived by Jerzy Neyman and
Egon Pearson (Karl’s son) as an efficient and objective
alternative to significance testing

– Neyman & Pearson (1933) write an abstract paper investigating an


optimal long-run strategy for testing pairs of hypotheses. They suggest
comparing the log likelihood ratio of each hypothesis to a criterion
computed from a fixed tail probability of incorrectly classifying one of the
two hypotheses

• The concepts of Type I and Type II errors, α, β, power, critical


regions, and fixed-criterion hypothesis testing all come from
this approach

01:830:400 Spring 2019 11


Estimation and CLT

Differences Between the Approaches


Fisher Neyman—Pearson
• Set up a statistical null • Set up two statistical hypotheses
hypothesis (must be exact) (H0 & H1), both of which must be
exact
• Report the exact level of
significance (p) • Decide on α, β, and sample size
before the experiment, these will
• If the result is not significant, define a rejection region
draw no conclusions.
• If the data fall into the rejection
• Only use this procedure to draw region of H0, accept H1,
provisional conclusions otherwise accept H0.

• Always make a decision based


on the available information

01:830:400 Spring 2019 12


Estimation and CLT

Hypothesis Testing & The Null Hypothesis


Why do we test the null hypothesis H0?

Philosophical arguments:

• Finite observations cannot prove categorical propositions, only


disprove them

• Puts the burden on the researcher


– Anyone can create an apparent difference between conditions by using very
small sample sizes
– Assume no effect (or standard effect) until given sufficient evidence

Practical argument:

• The null hypothesis is specific and well-defined, making it easy to


predict a sampling distribution

01:830:400 Spring 2019 13


Estimation and CLT

Rejection Regions
α= 0.05; 1-tailed test α= 0.05; 2-tailed test
(test that µ1 > µ0) (test that µ1 ≠ µ0)

p X 

X X

Why 0.05?
01:830:400 Spring 2019 14
Estimation and CLT

Errors in Hypothesis Testing


• Because the hypothesis test relies on sample data, and
sample data are variable, there is always a risk that the
hypothesis test will lead to the wrong conclusion.

• Two types of errors are possible:


– Type I errors (false positives)
– Type II errors (false negatives)

01:830:400 Spring 2019 15


Estimation and CLT

Errors in Hypothesis Testing

01:830:400 Spring 2019 16


Estimation and CLT

Errors in Hypothesis Testing

01:830:400 Spring 2019 17


Estimation and CLT

 0  1  

Population distributions

Raw scores (x)

n4
 
M  
n 2

Sampling distributions


Sample means (𝑋)
01:830:400 Spring 2019 18
Estimation and CLT

Errors in Hypothesis Testing

P=α P = 1-β

P = 1-α P=β

01:830:400 Spring 2019 19


Estimation and CLT

Power
• The statistical power of a test is simply the probability of
correctly rejecting the null hypothesis when it is false

• For our purposes, you can think of this as the probability that
the test will classify an actual difference in population means
as significant.

01:830:400 Spring 2019 20


Estimation and CLT

 0  1  

Population distributions

Raw scores (x)

n4
 
X  
n 2

Sampling distributions
power  1  


Sample means (𝑋)
01:830:400 Spring 2019 21
Estimation and CLT

Factors that Affect the Power of a Test


1. The probability of a Type I error (α), or the level of
significance, and the criterion for rejecting H0, which are
directly related to each other.

2. The true difference between the underlying population


means under the alternative hypothesis (μ1 - μ0).

3. The standard error(s) of the mean(s), which is a function of


the sample size n and the population variance σ2.

4. The particular research design and test used and whether


the test is one or two-tailed.

01:830:400 Spring 2019 22


Estimation and CLT

Power as a Function of α

Population
distributions

  0.05
  0.73
power  0.27

Sampling 1    power
distributions 


01:830:400 Spring 2019 X 23
Estimation and CLT

Power as a Function of α

Population
distributions

  0.10
  0.62
power  0.38

Sampling 1    power
distributions 

01:830:400 Spring 2019 X 24


Estimation and CLT

Power as a Function of α

Population
distributions

  0.20
  0.48
power  0.52

Sampling 1    power
distributions 

01:830:400 Spring 2019 X 25


Estimation and CLT

Power as a Function of (μ1 - μ0)


1  0  0.5

Population
distributions

1  0  0.5
  0.84
power  0.16

Sampling 1    power
distributions 

01:830:400 Spring 2019 X 26


Estimation and CLT

Power as a Function of (μ1 - μ0)


1  0  1.0

Population
distributions

1  0  1.0
  0.62
power  0.38

Sampling 1    power
distributions 

01:830:400 Spring 2019 X 27


Estimation and CLT

Power as a Function of (μ1 - μ0)


1  0  2.0

Population
distributions

1  0  2.0
  0.36
power  0.84

Sampling
1    power
distributions

01:830:400 Spring 2019 X 28


Estimation and CLT

Power as a Function of n and σ


  1.5

Population
distributions

n4

X   0.75
4
  0.81
power  0.19
Sampling 1    power
distributions 

01:830:400 Spring 2019 X 29


Estimation and CLT

Power as a Function of n and σ


  0.75

Population
distributions

n4

X   0.375
4
  0.15
power  0.85
Sampling 1    power
distributions
 

01:830:400 Spring 2019 X 30


Estimation and CLT

Power as a Function of n and σ


  1.5

Population
distributions

n  16

X   0.375
16
  0.15
power  0.85
Sampling 1    power
distributions
 

01:830:400 Spring 2019 X 31


Estimation and CLT

Some Pros & Cons of Hypothesis Testing


Pros Cons

• Objective method for • Rigid, 1-bit decision making


making decisions regarding
data • Absolves scientists from
thinking carefully about
• Simple rules, do not require analysis
statistics expertise
• Long-run guarantees rely
• In the absence of auxiliary on replication and unbiased
biases (and in scrupulous reporting & publication
hands), guarantees correct
decisions in the long run • p-values & significance level
not useful for meta-analysis

01:830:400 Spring 2019 32

S-ar putea să vă placă și