Documente Academic
Documente Profesional
Documente Cultură
Agenda
1 Session 12
Point Estimation of Parameters
2 References
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The field of statistical inference consists of those methods
used to make decisions or to draw conclusions about a
population.
These methods utilize the information contained in a sample
from the population in drawing conclusions.
Statistical inference may be divided into two major areas:
Parameters estimation.
Hypothesis testing.
We will see that it is possible to establish the precision of the
estimate.
Since a statistic is a random variable, it has a probability
distribution. We call the probability distribution of a statistic
a sampling distribution.
Session 12 References
Introduction
Statistical Inference
The parameters will be represented by .
The objective of point estimation is to select a single number,
based on sample data, that is the most plausible value for .
If X is a random variable with probability distribution f (x),
characterized by the unknown parameter , and if
X1 , X2 , .., Xn is a random sample of size n from X , the
statistics = h(X1 , X2 , ..., Xn ) is called a point estimator of .
Note that is a random variable because it is a function of
random variables.
After the sample has been selected, takes on a particular
numerical value called the point estimate of .
Session 12 References
Introduction
Statistical Inference
The parameters will be represented by .
The objective of point estimation is to select a single number,
based on sample data, that is the most plausible value for .
If X is a random variable with probability distribution f (x),
characterized by the unknown parameter , and if
X1 , X2 , .., Xn is a random sample of size n from X , the
statistics = h(X1 , X2 , ..., Xn ) is called a point estimator of .
Note that is a random variable because it is a function of
random variables.
After the sample has been selected, takes on a particular
numerical value called the point estimate of .
Session 12 References
Introduction
Statistical Inference
The parameters will be represented by .
The objective of point estimation is to select a single number,
based on sample data, that is the most plausible value for .
If X is a random variable with probability distribution f (x),
characterized by the unknown parameter , and if
X1 , X2 , .., Xn is a random sample of size n from X , the
statistics = h(X1 , X2 , ..., Xn ) is called a point estimator of .
Note that is a random variable because it is a function of
random variables.
After the sample has been selected, takes on a particular
numerical value called the point estimate of .
Session 12 References
Introduction
Statistical Inference
The parameters will be represented by .
The objective of point estimation is to select a single number,
based on sample data, that is the most plausible value for .
If X is a random variable with probability distribution f (x),
characterized by the unknown parameter , and if
X1 , X2 , .., Xn is a random sample of size n from X , the
statistics = h(X1 , X2 , ..., Xn ) is called a point estimator of .
Note that is a random variable because it is a function of
random variables.
After the sample has been selected, takes on a particular
numerical value called the point estimate of .
Session 12 References
Introduction
Statistical Inference
The parameters will be represented by .
The objective of point estimation is to select a single number,
based on sample data, that is the most plausible value for .
If X is a random variable with probability distribution f (x),
characterized by the unknown parameter , and if
X1 , X2 , .., Xn is a random sample of size n from X , the
statistics = h(X1 , X2 , ..., Xn ) is called a point estimator of .
Note that is a random variable because it is a function of
random variables.
After the sample has been selected, takes on a particular
numerical value called the point estimate of .
Session 12 References
Point estimate
Definition
A point estimate of some population parameter is a single
numerical value of a statistics . The statistics is called the
point estimator. For example:
Some population parameter e.g. or 2
The statistics e.g. X or S 2
Single numerical value e.g. The computed values
Session 12 References
Point estimate
Definition
A point estimate of some population parameter is a single
numerical value of a statistics . The statistics is called the
point estimator. For example:
Some population parameter e.g. or 2
The statistics e.g. X or S 2
Single numerical value e.g. The computed values
Session 12 References
Point estimate
Definition
A point estimate of some population parameter is a single
numerical value of a statistics . The statistics is called the
point estimator. For example:
Some population parameter e.g. or 2
The statistics e.g. X or S 2
Single numerical value e.g. The computed values
Session 12 References
Point estimate
Choices
We may have several different choices for the point estimator
of a parameter.
If we wish to estimate the mean of a population, we might
consider the sample mean, the sample median, or perhaps the
average of the smallest and largest observations in the sample
as point estimators.
In order to decide which point estimator of a particular
parameter is the best one to use, we need to examine their
statistical properties and develop some criteria for comparing
estimators.
Session 12 References
Point estimate
Choices
We may have several different choices for the point estimator
of a parameter.
If we wish to estimate the mean of a population, we might
consider the sample mean, the sample median, or perhaps the
average of the smallest and largest observations in the sample
as point estimators.
In order to decide which point estimator of a particular
parameter is the best one to use, we need to examine their
statistical properties and develop some criteria for comparing
estimators.
Session 12 References
Point estimate
Choices
We may have several different choices for the point estimator
of a parameter.
If we wish to estimate the mean of a population, we might
consider the sample mean, the sample median, or perhaps the
average of the smallest and largest observations in the sample
as point estimators.
In order to decide which point estimator of a particular
parameter is the best one to use, we need to examine their
statistical properties and develop some criteria for comparing
estimators.
Session 12 References
Estimators
An estimator should be close in some sense to the true
value of the unknown parameter.
Formally, we say that is an unbiased estimator of if the
expected value of is equal to . i.e. E () = .
If the estimator is not unbiased, then the difference E ()
is called the bias of the estimator .
In the book is demostrated that E (X ) = and E (S 2 ) = 2
but E (S) 6= . For large samples, the bias is very small.
However, there are good reasons for using S as an estimator
of in samples from normal distributions.
Session 12 References
Estimators
An estimator should be close in some sense to the true
value of the unknown parameter.
Formally, we say that is an unbiased estimator of if the
expected value of is equal to . i.e. E () = .
If the estimator is not unbiased, then the difference E ()
is called the bias of the estimator .
In the book is demostrated that E (X ) = and E (S 2 ) = 2
but E (S) 6= . For large samples, the bias is very small.
However, there are good reasons for using S as an estimator
of in samples from normal distributions.
Session 12 References
Estimators
An estimator should be close in some sense to the true
value of the unknown parameter.
Formally, we say that is an unbiased estimator of if the
expected value of is equal to . i.e. E () = .
If the estimator is not unbiased, then the difference E ()
is called the bias of the estimator .
In the book is demostrated that E (X ) = and E (S 2 ) = 2
but E (S) 6= . For large samples, the bias is very small.
However, there are good reasons for using S as an estimator
of in samples from normal distributions.
Session 12 References
Estimators
An estimator should be close in some sense to the true
value of the unknown parameter.
Formally, we say that is an unbiased estimator of if the
expected value of is equal to . i.e. E () = .
If the estimator is not unbiased, then the difference E ()
is called the bias of the estimator .
In the book is demostrated that E (X ) = and E (S 2 ) = 2
but E (S) 6= . For large samples, the bias is very small.
However, there are good reasons for using S as an estimator
of in samples from normal distributions.
Session 12 References
Estimators
An estimator should be close in some sense to the true
value of the unknown parameter.
Formally, we say that is an unbiased estimator of if the
expected value of is equal to . i.e. E () = .
If the estimator is not unbiased, then the difference E ()
is called the bias of the estimator .
In the book is demostrated that E (X ) = and E (S 2 ) = 2
but E (S) 6= . For large samples, the bias is very small.
However, there are good reasons for using S as an estimator
of in samples from normal distributions.
Session 12 References
Definition
Since 1 has a smaller variance than 2 , the estimator 1 is
more likely to produce an estimate close to the true value .
A logical principle of estimation, when selecting among several
estimators, is to choose the estimator that has minimum
variance.
If we consider all unbiased estimators of , the one with
smallest variance is called the minimum variance unbiased
estimator (MVUE).
If X1 , X2 , ..., Xn is a random sample of size n from a normal
distribution with mean and variance 2 , the sample mean X
is the MVUE for .
Session 12 References
Definition
Since 1 has a smaller variance than 2 , the estimator 1 is
more likely to produce an estimate close to the true value .
A logical principle of estimation, when selecting among several
estimators, is to choose the estimator that has minimum
variance.
If we consider all unbiased estimators of , the one with
smallest variance is called the minimum variance unbiased
estimator (MVUE).
If X1 , X2 , ..., Xn is a random sample of size n from a normal
distribution with mean and variance 2 , the sample mean X
is the MVUE for .
Session 12 References
Definition
Since 1 has a smaller variance than 2 , the estimator 1 is
more likely to produce an estimate close to the true value .
A logical principle of estimation, when selecting among several
estimators, is to choose the estimator that has minimum
variance.
If we consider all unbiased estimators of , the one with
smallest variance is called the minimum variance unbiased
estimator (MVUE).
If X1 , X2 , ..., Xn is a random sample of size n from a normal
distribution with mean and variance 2 , the sample mean X
is the MVUE for .
Session 12 References
Definition
Since 1 has a smaller variance than 2 , the estimator 1 is
more likely to produce an estimate close to the true value .
A logical principle of estimation, when selecting among several
estimators, is to choose the estimator that has minimum
variance.
If we consider all unbiased estimators of , the one with
smallest variance is called the minimum variance unbiased
estimator (MVUE).
If X1 , X2 , ..., Xn is a random sample of size n from a normal
distribution with mean and variance 2 , the sample mean X
is the MVUE for .
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Definition
When the numerical value or point estimate of a parameter is
reported, it is usually desirable to give some idea of the
precision of estimation.
The measure of precision usually employed is the standard
error of the estimator that has been used.
The standard error
q of an estimator is its standard deviation,
given by = V ().
Suppose we are sampling from a normal distribution with
mean and variance 2 .
The distribution of X is normal with mean and variance
2 /n, so the standard error of X is X = / n.
If we did not know but substituted the sample standard
deviation S into the above equation, X = S/ n.
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Standard Error
Example
Generates a population of 250000 elements normally
distributed (use randn).
Choose a sample of 50000. Use datasample(data,k)
Compute the sample mean X .
Compute the sample standard deviation S.
Compute the standard error of X .
Compute the standard error percentaje with respect to X .
What does it means?
Session 12 References
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Suppose we have a random sample of size 2n from a
population denoted by X , and E (X ) = and V (X ) = 2 .
2n n
1 X 1X
Let X1 = Xi and X2 = Xi be two estimators of .
2n n
i=1 i=1
Which is the better estimator of ? Explain your choice.
Both estimators are unbiased because E (X ) = .
However in the first case the variance is V (X1 ) = 2 /2n
While in the second case the variance is V (X2 ) = 2 /n.
MSE (1 ) 2 /2n n 1
Which means that MSE (2 )
= 2 /n
= 2n = 2
Exercise
Let X1 , X2 , ..., X7 denote a random sample from a population
having mean and variance 2 . Consider the following estimators
of :
X1 + X2 + ... + X7
1 =
7
2X1 X6 + X4
2 =
2
Exercise
1 1
E (1 ) = [E (X1 ) + E (X2 ) + ... + E (X7 )] = (7) =
7 7
1 1
E (2 ) = [E (2X1 ) + E (X6 ) + E (X7 )] = (2 + ) =
2 2
Both 1 and 2 are unbiased estimates of since the expected
values of these statistics are equivalent to the true mean, .
X1 + X2 + .. + X7 1
V (1 ) = V [ ] = (V (X1 )+V (X2 )+...+V (X7 ))
7 49
1 1
= (7 2 ) = 2
49 7
Session 12 References
Exercise
2X1 X6 + X4 1
V (2 ) = V [ ] = (V (2X1 ) + V (X6 ) + V (X4 ))
2 4
1 1
= (4V (X1 ) + V (X6 ) + V (X4 )) = (4 2 + 2 + 2 )
4 4
1 3
= (6 2 ) = 2
4 2
Since both estimators are unbiased, the variances can be compared
to decide which is the better estimator. The variance of 1 is
smaller than that of 2 , 1 is the better estimator.
Session 12 References
References I