Sunteți pe pagina 1din 21

Chapter 10

Two Stage Sampling (Subsampling)

In cluster sampling, all the elements in the selected clusters are surveyed. Moreover, the efficiency in
cluster sampling depends on size of the cluster. As the size increases, the efficiency decreases. It
suggests that higher precision can be attained by distributing a given number of elements over a large
number of clusters and then by taking a small number of clusters and enumerating all elements within
them. This is achieved in subsampling.

In subsampling
- divide the population into clusters.
- Select a sample of clusters [first stage}
- From each of the selected cluster, select a sample of specified number of elements [second stage]

The clusters which form the units of sampling at the first stage are called the first stage units and the
units or group of units within clusters which form the unit of clusters are called the second stage units or
subunits.

The procedure is generalized to three or more stages and is then termed as multistage sampling.

For example, in a crop survey


- villages are the first stage units,
- fields within the villages are the second stage units and
- plots within the fields are the third stage units.

In another example, to obtain a sample of fishes from a commercial fishery


- first take a sample of boats and
- then take a sample of fishes from each selected boat.

Two stage sampling with equal first stage units:


Assume that
- population consists of NM elements.
- NM elements are grouped into N first stage units of M second stage units each, (i.e., N
clusters, each cluster is of size M )
- Sample of n first stage units is selected (i.e., choose n clusters)
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 1
- Sample of m second stage units is selected from each selected first stage unit (i.e., choose m
units from each cluster).
- Units at each stage are selected with SRSWOR.

Cluster sampling is a special case of two stage sampling in the sense that from a population of N
clusters of equal size m = M , a sample of n clusters are chosen.
If further M= m= 1, we get SRSWOR.
If n = N , we have the case of stratified sampling.

yij : Value of the characteristic under study for the j th second stage units of the i th first stage

=
unit; i 1,=
2,..., N ; j 1, 2,.., m.
M
1
Yi =
M
∑y
j =1
ij : mean per 2nd stage unit of i th 1st stage units in the population.

1 N M 1 N
=Y ∑∑ ij N=
=
MN =i 1 =j 1
y ∑ yi YMN : mean per second stage unit in the population
=i 1

1 m
yi = ∑
m j =1
yij : mean per second stage unit in the i th first stage unit in the sample.

1 n m 1 n
=y ∑∑ ij n=
=
mn=i 1 =j 1
y ∑ yi ymn : mean per second stage in the sample.
=i 1

Advantages:

The principle advantage of two stage sampling is that it is more flexible than the one stage sampling. It
reduces to one stage sampling when m = M but unless this is the best choice of m , we have the
opportunity of taking some smaller value that appears more efficient. As usual, this choice reduces to a
balance between statistical precision and cost. When units of the first stage agree very closely, then
consideration of precision suggests a small value of m . On the other hand, it is sometimes as cheap to
measure the whole of a unit as to a sample. For example, when the unit is a household and a single
respondent can give as accurate data as all the members of the household.

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 2
A pictorial scheme of two stage sampling scheme is as follows:

Population (MN units)

Cluster Cluster Cluster Population


M units M units … … … M units N clusters
(large in
number)
N clusters

Cluster Cluster First stage


Cluster sample
M units M units … … …
M units n clusters
(small in
n clusters number)

Second stage
sample
m units
Cluster Cluster Cluster n clusters
… … …
m units m units m units (large number
of elements
mn units from each
cluster)

Note: The expectations under two stage sampling scheme depend on the stages. For example, the
expectation at second stage unit will be dependent on first stage unit in the sense that second stage unit
will be in the sample provided it was selected in the first stage.

To calculate the average


- First average the estimator over all the second stage selections that can be drawn from a fixed
set of n units that the plan selects.
- Then average over all the possible selections of n units by the plan.

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 3
In case of two stage sampling,

E (θˆ) = E1[ E2 (θˆ)]


↓ ↓ 
average average average over
over over all all possible 2nd stage
all 1st stage selections from
samples samples a fixed set of units

In case of three stage sampling,

{
E (θˆ) = E1  E2 E3 (θˆ)  .
  }

To calculate the variance, we proceed as follows:


In case of two stage sampling,
θˆ) E (θˆ − θ ) 2
Var (=
= E E (θˆ − θ ) 2
1 2

Consider

E2 (θˆ − θ )=
2
E2 (θˆ 2 ) − 2θ E2 (θˆ) + θ 2

{ }
.
=  E2 (θˆ + V2 (θˆ)  − 2θ E2 (θˆ) + θ 2
2

 

Now average over first stage selection as


2
ˆ − θ ) 2 E  E (θˆ)  + E V (θˆ)  − 2θ E E (θˆ) + E (θ 2 )
E1 E2 (θ= 1 2  1 2  1 2 1

{ }
= E1  E2 (θˆ) − θ 2  + E1 V2 (θˆ) 
2

 
= Var (θˆ) V1  E2 (θˆ)  + E1 V2 (θˆ)  .

In case of three stage sampling,

{ } { } { }
Var (θˆ) = V1  E E3 (θˆ)  + E1 V2 E3 (θˆ)  + E1  E2 V3 (θˆ)  .
 2     

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 4
Estimation of population mean:
Consider y = ymn as an estimator of the population mean Y .

Bias:
Consider
E ( y ) = E1 [ E2 ( ymn ) ]
=
E1  E2 ( yim i )  (as 2nd stage is dependent on 1st stage)

=
E1  E2 ( yim i )  (as yi is unbiased for Yi due to SRSWOR)

1 n 
= E1  ∑ Yi 
 n i =1 
1 N
= ∑ Yi
N i =1
=Y .

Thus ymn is an unbiased estimator of the population mean.

Variance
) E1 V2 ( y i )  + V1 [ E2 ( y / i ) ]
Var ( y =
 1 n   1 n 
E1 V2  ∑ yi i  + V1  E2  ∑ yi / i 
=   n i 1=   n i 1 
1 n  1 n 
E1  2 ∑ V ( yi i )  + V1  ∑ E2 ( yi / i ) 
=  n i 1=  n i 1 
1 n
1 1   1 n

= E1  2 ∑  −  Si2  + V1  ∑ Yi 
= m M  
 n i 1= n i 1 
1 n 1 1 
2 ∑
= − E1 ( Si2 ) + V1 ( yc )
n i =1  m M 
                                       (where yc is based on cluster means as in cluster sampling)
1 1 
1 N −n 2
= n  −  S w2 + Sb
m M 
n2 Nn
1 1 1  2 1 1  2
=  −  S w +  −  Sb
nm M  n N 
2
1 N 2
(Yij − Yi )
N M
1
=
where S w2 = ∑ i N (M − 1)=∑∑
N =i 1
S
i 1 =j 1

1 N
=S ∑
N − 1 i =1
b(Yi − Y ) 2
2

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 5
Estimate of variance
An unbiased estimator of variance of y can be obtained by replacing Sb2 and S w2 by their unbiased

estimators in the expression of variance of y .

Consider an estimator of
N
1
S w2 =
N
∑S
i =1
i
2

∑ ( yij − Yi )
1 M
=
where Si2
M − 1 j =1
1 n 2
as ∑ si
sw2 =
n i =1
1 m
=
where si2 ∑ ( yij − yi )2 .
m − 1 j =1
So

E ( sw2 ) = E1 E2 ( sw2 i )

1 n 
= E1 E2  ∑ si2 i 
 n i =1 
n
1
= E1 ∑  E2 ( si2 i ) 
n i =1
1 n 2
= E1 ∑ Si
n i =1
(as SRSWOR is used)

1 n
= ∑
n i =1
E1 ( Si2 )

1 N  1 N 2
= ∑  N ∑
N i 1=
= i 1
Si 

N
1
=
N
∑S
i =1
i
2

=S 2
w

so sw2 is an unbiased estimator of S w2 .

Consider
1 n
=sb2 ∑
n − 1 i =1
( yi − y ) 2

as an estimator of
1 N
=Sb2 ∑
N − 1 i =1
(Yi − Y ) 2 .

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 6
So
1  n 
=E ( sb2 ) E  ∑ ( yi − y ) 2 
n − 1  i =1 
 n 
(n − 1) E ( sb2 ) = E  ∑ yi2 − ny 2 
 i =1 
 n

= E  ∑ yi2  − nE ( y 2 )
 i =1 
  n 
= E1  E2  ∑ yi2   − n Var ( y ) + { E ( y )} 
2

  i =1    

 n   1 1   1 1 1 
= E1  ∑ E2 ( yi2 ) i )  − n  −  Sb2 +  −  S w2 + Y 2 
 i =1   n N  m M n 
 n
{
 i =1
}
2 


 1 1   1 1 1 
= E1  ∑ Var ( yi ) + ( E ( yi )  − n  −  Sb2 +  −  S w2 + Y 2 
 n N  m M n 
 n  1 1  2 2   1 1  2  1 1 1 2 2
= E1  ∑  −  Si + Yi  − n  −  Sb +  −  Sw + Y 
 i =1  m M    n N  m M n 
1  n  1 1  2 2   1 1  2  1 1 1 2 2
= nE1  ∑  − Si + Yi  − n  −  Sb +  −  Sw + Y 
 n  i =1  m M    n N  m M n 
 1 1  1 N 1 N   1 1   1 1 1 
= n  −  ∑ Si2 + ∑ Yi 2  − n  −  Sb2 +  −  S w2 + Y 2 
 m =
M  N i 1= N i1   n N  m M n 
 1 1  1 N
  1 1   1 1 1 
= n  −  S w2 + ∑ Yi 2  − n  −  Sb2 +  −  S w2 + Y 2 
 m M  N i =1   n N  m M n 
1 1  2 n N 2 1 1  2
=(n − 1)  −  S w + ∑ Yi − nY − n  −  Sb
2

m M  N i =1 n N 
1 1  2 nN 2 2 1 1  2
=(n − 1)  −  S w +  ∑ Yi − NY  − n  −  Sb
m M  N  i =1  n N 
1 1  n 1 1 
= (n − 1)  −  S w2 + ( N − 1) Sb2 − n  −  Sb2
m M  N n N 
1 1 
= (n − 1)  −  S w2 + (n − 1) Sb2 .
m M 
1 1  2
⇒ E ( sb2 ) =  −  S w + Sb
2

m M 
 1 1  
or E  sb2 −  −  sw2  = Sb2 .
  m M  
Thus

 ( y )= 1  1 − 1  Sˆ 2 +  1 − 1  Sˆ 2
Var   ω   b
nm M  n N 
1  1 1  2  1 1  2  1 1  2
=  −  sw +  −   sb −  −  sw 
nm M   n N  m M  
11 1  2 1 1  2
=  −  sw +  −  sb .
N m M  n N 
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 7
Allocation of sample to the two stages: Equal first stage units:
The variance of sample mean in the case of two stage sampling is

 ( y )= 1  1 − 1  S 2 +  1 − 1  S 2 .
Var   w   b
nm M  n N 
It depends on Sb2 , S w2 , n and m. So the cost of survey of units in the two stage sample depends on
n and m.

Case 1. When cost is fixed


We find the values of n and m so that the variance is minimum for given cost.

(I) When cost function is C = kmn


Let the cost of survey be proportional to sample size as
C = knm
where C is the total cost and k is constant.
C0
When cost is fixed as C = C0 . Substituting m = in Var ( y ), we get
kn
1  2 S w2  Sb2 1 kn 2
Var ( y =
)  Sb −  − + Sw
n M  N n C0
1  2 S w2   Sb2 kS w2 
=  Sb − − − .
n M   N C0 

 S2 
This variance is monotonic decreasing function of n if  Sb2 − w  > 0. The variance is minimum
 M 

when n assumes maximum value, i.e.,


C0
nˆ = corresponding to m 1.
k
 S2 
If  Sb2 − w  < 0 (i.e., intraclass correlation is negative for large N ) , then the variance is a monotonic
 M 

C0
increasing function of n , It reaches minimum when n assumes the minimum value, i.e., nˆ =
kM
(i.e., no subsampling).

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 8
(II) When cost function is =
C k1n + k2 mn

Let cost C be fixed as C=


0 k1n + k2 mn where k1 and k2 are positive constants. The terms k1 and k2
denote the costs of per unit observations in first and second stages respectively. Minimize the variance of
sample mean under the two stage with respect to m subject to the restriction C=
0 k1n + k2 mn .

We have
 S2   S2   S2  k S2
C0 Var ( y ) + b  = k1  Sb2 − w  + k2 S w2 + mk2  Sb2 − w  + 1 w .
 N  M  M m
 S2 
When  Sb2 − w  > 0, then
 M
2 2
 S2    S2     S2  k S2 
C0 Var ( y ) + b=  k1  Sb2 − w  + k2 S w2  +  mk2  Sb2 − w  − 1 w 
 N   M    M m 

which is minimum when the second term of right hand side is zero. So we obtain

k1 S w2
mˆ = .
k2  2 S w2 
 Sb − 
 M

The optimum n follows from C=


0 k1n + k2 mn as

C0
nˆ = .
k1 + k2 mˆ

 S2 
When  Sb2 − w  ≤ 0 then
 M

 S2   S2   S2  k S2
C0 Var ( y ) + b  = k1  Sb2 − w  + k2 S w2 + mk2  Sb2 − w  + 1 w
 N  M  M m

is minimum if m is the greatest attainable integer. Hence in this case, when


C0
C0 ≥ k1 + k2=
M ; mˆ M =
and nˆ .
k1 + k2 M
C0 − k1
If C0 ≥ k1 + k= ˆ
2 M ; then m = and nˆ 1.
k2

If N is large, then S w2 ≈ S 2 (1 − ρ )

S w2
S w2 − ≈ ρS2
M

k1  1 
mˆ ≈  − 1 .
k2  ρ 

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 9
Case 2: When variance is fixed
Now we find the sample sizes when variance is fixed, say as V0 .

1 1 1  2 1 1  2
V0= −  S w +  −  Sb
nm M  n N 
1 1 
Sb2 +  −  S w2
⇒n= 
m M
Sb2
V0 +
N
So
 2 S w2 
 Sb −  kS w2
= = km 
C kmn M + .
2 2
 V + Sb  V + Sb
 0  0
 N  N
 2 S w2 
If  Sb −  > 0, C attains minimum when m assumes the smallest integral value, i.e., 1.
 M

 S2 
If  Sb2 − w  < 0 , C attains minimum when mˆ = M .
 M

Comparison of two stage sampling with one stage sampling


One stage sampling procedures are comparable with two stage sampling procedures when either
(i) sampling mn elements in one single stage or
mn
(ii) sampling first stage units as cluster without sub-sampling.
M
We consider both the cases.

Case 1: Sampling mn elements in one single stage


The variance of sample mean based on
- mn elements selected by SRSWOR (one stage) is given by
 1 1  2
V ( y=
SRS )  − S
 mn MN 
- two stage sampling is given by
1 1 1  2 1 1  2
V ( yTS )=  −  S w +  −  Sb .
nm M  n N 

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 10
The intraclass correlation coefficient is

 N − 1  2 Sw
2

  Sb −
M M ( N − 1) Sb − NS w
2 2
=ρ  N 
= ;−
1
≤ ρ ≤1 (1)
 NM − 1  2 ( MN − 1) S 2
M −1
 S
 NM 
and using the identity
N M N M N M

∑∑ ( yij − Y=
=i 1 =j 1
)2 ∑∑ ( yij − Yi )2 + ∑∑ (Yi − Y )2
=i 1 =j 1 =i 1 =j 1

( NM − 1) S =( N − 1) MS + N ( M − 1) S w2
2 2
b (2)

1 N M 1 M
=
where Y = ∑∑ ij i M =∑
MN =i 1 =j 1
y , Y
j 1
yij .

Now we need to find Sb2 and S w2 from (1) and (2) in terms of S 2 . From (1), we have

 MN − 1   N −1 
S w2 =
−  MS ρ + 
2 2
 MSb . (3)
 N   N 

Substituting it in (2) gives


 N − 1   MN − 1  2 
( NM − 1) S 2 =( N − 1) MSb2 + N ( M − 1)   MSb −   MS ρ 
2

 N   N  
= ( N − 1) MSb + ( M − 1)( N − 1) Sb − ρ M ( M − 1)( MN − 1) S
2 2 2

= ( N − 1) MSb2 [1 + ( M − 1)] − ρ M ( M − 1)( MN − 1) S 2


= ( N − 1) MSb2 − ρ M ( M − 1)( MN − 1) S 2
( MN − 1) S 2
=
⇒ Sb2 [1 + ( M − 1) ρ ]
M 2 ( N − 1)

Substituting it in (3) gives

N ( M − 1) S w2= ( NM − 1) S 2 − ( N − 1) MSb2
 ( MN − 1) S 2 
= ( NM − 1) S 2 − ( N − 1) M  2 [1 + ( M − 1) ρ ] 
 M ( N − 1) 
 M − 1 − ( M − 1) ρ 
= ( NM − 1) S 2  
 M
= ( NM − 1) S 2 ( M − 1)(1 − ρ )
 MN − 1  2
⇒=S w2   S (1 − ρ ).
 MN 

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 11
Substituting Sb2 and S w2 in Var ( yTS )

 MN − 1  S  m(n − 1) N −n m M − m 
2
=
V ( yTS )   1− +ρ ( M − 1) −  .
 MN  mn  M ( N − 1)  N −1 M M 
m
When subsampling rate is small, MN − 1 ≈ MN and M − 1 ≈ M , then
M
S2
V ( ySRS ) =
mn
S2   N −n 
V ( yTS ) =1 + ρ  m − 1  .
mn   N −1 
The relative efficiency of the two stage in relation to one stage sampling of SRSWOR is
Var ( yTS )  N −n 
RE = 1+ ρ 
= m − 1 .
Var ( ySRS )  N −1 
N −n N −n
If N − 1 ≈ N and finite population correction is ignorable, then ≈ ≈ 1, then
N −1 N
1 + ρ (m − 1).
RE =

Case 2: Comparison with cluster sampling


mn
Suppose a random sample of clusters, without further subsampling is selected.
M
The variance of the sample mean of equivalent mn / M clusters is
M 1 2
Var (=
ycl )  −  Sb .
 mn N 
The variance of sample mean under the two stage sampling is
1 1 1  2 1 1  2
Var ( yTS )=  −  S w +  −  Sb .
nm M  n N 
So Var ( ycl ) exceedes Var ( yTS ) by

1M  2 1 2 
 − 1   Sb − S w 
n m  M 
which is approximately

1M  2  2 S w2 
 − 1  ρ S for large N and  Sb −  > 0.
n m   M

MN − 1 S 2
=
where Sb2 [1 + ρ ( M − 1)]
M ( N − 1) M
MN − 1 2
=S w2 S (1 − ρ )
MN

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 12
So smaller the m / M , larger the reduction in the variance of two stage sample over a cluster sample.
 S2 
When  Sb2 − w  < 0 then the subsampling will lead to loss in precision.
 M

Two stage sampling with unequal first stage units:


Consider two stage sampling when the first stage units are of unequal size and SRSWOR is employed at
each stage.
Let
yij : value of j th second stage unit of the i th first stage unit.

M i : number of second stage units in i th first stage units (i = 1, 2,..., N ) .


N
M 0 = ∑ M i : total number of second stage units in the population.
i =1

mi : number of second stage units to be selected from i th first stage unit, if it is in the sample.
n
m0 = ∑ mi : total number of second stage units in the sample.
i =1

mi
1
yi ( mi ) =
mi
∑yj =1
ij

Mi
1
Yi =
Mi
∑y j =1
ij

1 N
=Y = ∑ yi
N i =1
YN
N Mi N
∑∑ yij ∑M Y
1 i i N
=Y =i 1 =j 1
=N
=
i =1
∑u Y i i

∑ Mi
MN N i =1

i =1

Mi
ui =
M
1 N
M = ∑ Mi
N i =1

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 13
The pictorial scheme of two stage sampling with unequal first stage units case is as follows:

Population (MN units)

Cluster Cluster Cluster Population


M1 M2 … … … MN N clusters
units units units

N clusters

Cluster Cluster Cluster First stage


M1 M2 … … … Mn sample
units units n clusters n clusters
units
(small)

Second stage
Cluster Cluster Cluster sample
m1 units m2 units … … … mn units n clusters
(small)

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 14
Now we consider different estimators for the estimation of population mean.
1. Estimator based on the first stage unit means in the sample:
1 n

Y y=
S2 ∑ yi ( mi )
n i =1

Bias:
1 n 
E ( yS 2 ) = E  ∑ yi ( mi ) 
 n i =1 
1 n

= E1  ∑ E2 ( yi ( mi ) ) 
 n i =1 
1 n 
= E1  ∑ Yi  [Since a sample of size mi is selected out of M i units by SRSWOR]
 n i =1 
1 N
= ∑ Yi
N i =1
=YN
≠ Y.

So yS 2 is a biased estimator of Y and its bias is given by

Bias=
( yS 2 ) E ( yS 2 ) − Y
1 N 1 N
=
=
∑ i NM ∑
Y
N i 1=

i 1
M iYi

1 N 1  N  N 
=
−  ∑ M iYi −  ∑ Yi  ∑ M i  
NM
= i 1 N  i 1=
=  i 1 
N
1
= ∑ (M i − M )(Yi − YN ).
NM i =1
This bias can be estimated by

(y ) = N −1 n
Bias S2 − ∑ (M i − m)( yi ( mi ) − yS 2 )
NM (n − 1) i =1
which can be seen as follows:
N −1  1 n 
 ( y ) =
E  Bias − E1  ∑ E2 {( M i − m)( yi ( mi ) − yS 2 ) / n}
 S 2  NM  n − 1 i =1 
N −1  1 n 
=
− E ∑
NM  n − 1 i =1
( M i − m)(Yi − yn ) 

N
1
=

NM
∑ (M
i =1
i − M )(Yi − YN )

= YN − Y

1 n
where yn = ∑ Yi .
n i =1

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 15
An unbiased estimator of the population mean Y is thus obtained as
N −1 1 n
yS 2 + ∑ (M i − m)( yi ( mi ) − yS 2 ) .
NM n − 1 i =1
Note that the bias arises due to the inequality of sizes of the first stage units and probability of selection
of second stage units varies from one first stage to another.

Variance:
Var ( yS=
2) Var  E ( yS 2 n)  + E Var ( yS 2 n) 
1 n  1 n 
= Var  ∑ yi  + E  2 ∑ Var ( yi ( mi ) i ) 
=  n i 1=  n i 1 
1 1  2 1 n
 1 1  2
=
 −  Sb + E  2
n N 
∑ m −  Si 
n i =1  i Mi  
1 1  2 1 N  1 1  2
=
 −  Sb +
n N 
∑  −
Nn i =1  mi M i
 Si

( )
2
1 N
=
where S 2
∑ Yi − YN
N − 1 i =1
b

2
1 Mi
=S ∑ ( yij − Yi ) .
2

M i − 1 j =1
i

The MSE can be obtained as

= ( yS 2 ) Var ( yS 2 ) + [ Bias ( yS 2 ) ] .
2
MSE

Estimation of variance:
Consider mean square between cluster means in the sample
1 n
∑ ( )
2
=sb2 yi ( mi ) − yS 2 .
n − 1 i =1
It can be shown that

1 N
 1 1  2
E ( sb2 ) =
Sb2 + ∑ m − Si .
N i =1  i Mi 
1 mi
=
Also si2 ∑
mi − 1 j =1
( yij − yi ( mi ) ) 2

1 Mi
E( s =
2
i) S= 2

M i − 1 j =1
i ( yij − Yi ) 2

1 n  1 1  2 1 N  1 1  2
So E  ∑  −  si  = ∑  −  Si .
 n i 1  i
=m M i   N i 1  i
m M i 

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 16
Thus
1 n  1 1  2
Sb2 + E  ∑  −
E ( sb2 ) =  si 
 n i =1  i
m M i  
and an unbiased estimator of Sb2 is

1 n  1 1  2
Sb =
ˆ 2
sb − ∑  −
2
 si .
n i =1  mi M i 

So an estimator of the variance can be obtained by replacing Sb2 and Si2 by their unbiased estimators as

(y ) = 1 1  ˆ2 1 N  1 1  ˆ2
Var S2  −  Sb +
n N 
∑  −
Nn i =1  mi M i
Si .

2. Estimation based on first stage unit totals:


1 n M i yi ( mi )
Yˆ y=
= *
S2 ∑
n i =1 M
1 n
= ∑ ui yi ( mi )
n i =1
Mi
where ui = .
M
Bias
1 n 
E ( yS* 2 ) = E  ∑ ui yi ( mi ) 
 n i =1 
1 n

= E  ∑ ui E2 ( yi ( mi ) i ) 
 n i =1 
1 n

= E  ∑ uiYi 
 n i =1 
N
1
=
N
∑u Y
i =1
i i

=Y.

Thus yS* 2 is an unbiased estimator of Y .

Variance:
Var ( yS* 2 ) Var  E ( yS* 2 n)  + E Var ( yS* 2 n) 
=
1 n  1 n 
= Var  ∑ uiYi  + E  2 ∑ ui2Var ( yi ( mi ) i ) 
=  n i 1=  n i 1 
 1 1  *2 1 N
 1 1  2
=−

n N 
 Sb + ∑u 2
i  −  Si
nN i =1  mi M i 
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 17
1 Mi
=
where Si2 ∑
M i − 1 j =1
( yij − Yi ) 2

1 N
=Sb*2 ∑
N − 1 j =1
(uiYi − Y ) 2 .

3. Estimator based on ratio estimator:


n

∑M y i i ( mi )
Yˆ y=
= **
S2
i =1
n

∑M
i =1
i

∑u y i i ( mi )
= i =1
n

∑ui =1
i

yS* 2
=
un

Mi 1 n
=
where ui =
M
, un ∑ ui .
n i =1
This estimator can be seen as if arising by the ratio method of estimation as follows:

Let yi* = ui yi ( mi )
Mi
=xi* = , i 1, 2,..., N
M
be the values of study variable and auxiliary variable in reference to the ratio method of estimation. Then
1 n *
=y* =∑
n i =1
yi yS* 2

1 n *
=x* = ∑ xi un
n i =1
1 N *
=X* = ∑ X i 1.
N i =1

The corresponding ratio estimator of Y is


y* yS* 2
YˆR
= =X* = 1 yS**2 .
x* un

So the bias and mean squared error of yS**2 can be obtained directly from the results of ratio estimator.
Recall that in ratio method of estimation, the bias and MSE of the ratio estimator upto second order of
approximation is

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 18
N −n
Bias ( yˆ R ) ≈ Y (C x2 − 2 ρ C xC y )
Nn
Var ( x ) Cov( x , y ) 
= Y − 
 X
2
XY
MSE (YˆR ) ≈ Var ( y ) + R 2Var ( x ) − 2 RCov( x , y ) 

Y
where R = .
X
Bias:
The bias of yS**2 up to second order of approximation is

Var ( xS*2 ) Cov( xS*2 , yS* 2 ) 


=
Bias ( yS**2 ) Y  2
− 
 X XY 
1 n
where xS*2 is the mean of auxiliary variable similar to yS* 2 as xS*2 = ∑ xi ( mi ) .
n i =1

Now we find Cov( xS*2 , yS* 2 ).

 1 n 1 n   1 n 1 n 
Cov( xS*2 , yS* 2 ) Cov  E  ∑ ui xi ( mi ) , ∑ ui yi ( mi )   + E Cov  ∑ ui xi ( mi ) , ∑ ui yi ( mi )  
=   n i 1= ni1  =   n i 1= ni1 
1 n 1 n  1 n 
Cov  ∑ ui E ( xi ( mi ) ), ∑ ui E ( yi ( mi ) )  + E  2 ∑ ui2Cov( xi ( mi ) , yi ( mi ) ) i 
=  n i 1= ni1 =  n i 1 
1 n 1 n  1 n  1 1  
= Cov  ∑ ui X i , ∑ uiYi  + E  2 ∑ ui2  − Sixy 
=  n i 1= ni1  =  n i 1  mi M i  
1 1  * 1 N
 1 1 
=
 −  Sbxy +
n N 
∑u 2
i  − Sixy
nN i =1  mi M i 
where
1 N
=*
Sbxy ∑ (ui X i − X )(uiYi − Y )
N − 1 i =1
1 Mi
=
Sixy ∑ ( xij − X i )( yij − Yi ).
M i − 1 j =1

Similarly, Var ( xS*2 ) can be obtained by replacing x in place of y in Cov( xS*2 , yS* 2 ) as

 1 1  *2 1 N
 1 1  2
Var ( xS*2 ) =−

n N 
 Sbx + ∑u 2
i  − Six
nN i =1  mi M i 
1 N
=
where Sbx*2 ∑ (ui X i − X )2
N − 1 i =1
1 Mi
=Six*2 ∑ ( xij − X i )2 .
M i − 1 i =1

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 19
Substituting Cov( xS*2 , yS* 2 ) and Var ( xS*2 ) in Bias ( yS**2 ), we obtain the approximate bias as

 1 1   Sbx*2 Sbxy
*
 1 N
  1 1   Six2 Sixy  
Bias ( y ) ≈ Y  −   2 −
**
S2  +
XY  nN
∑ u 2
i  −  2 −   .
 n N   X i =1   mi M i   X XY  

Mean squared error


MSE ( yS**2 ) ≈ Var ( yS* 2 ) − 2 R*Cov( xS*2 , yS* 2 ) + R*2Var ( xS*2 )
 1 1  *2 1 N
 1 1  2
Var ( yS**2 ) =−

n N 
 Sby + ∑u 2
i  −  Siy
nN i =1  mi M i 
 1 1  *2 1  1 N
1  2
Var ( xS**2 ) =−

n N 
 Sbx +  − ∑u 2
 Six
i
nN i =1  mi M i 
1 1  * 1 N 2 1 1 
Cov( xS*2 , yS**2 ) =
 −  bxy
n N 
S + ∑ ui  −
nN i =1  mi M i
 Sixy

where
1 N
=Sby*2 ∑ (uiYi − Y )2
N − 1 i =1
1 Mi
=Siy*2 ∑ ( yij − Yi )2
M i − 1 j =1
Y
R=
*
= Y.
X
Thus

1 1    1 1  2 
MSE ( yS**2 ) ≈  −  ( Sby*2 − 2 R* Sbxy + R*2 Sbx*2 ) +  ( Siy − 2 R Sixy + R Six ) .
N
1
n N 
*
∑ u 2
i  −
* *2 2

nN i =1   mi M i  
Also

1 1  1 N 2 1 N  2 1 1  2 
∑ ui (Yi − R* X i ) + ∑  ( Siy − 2 R Sixy + R Six ) .
2
MSE ( yS**2 ) ≈  −  ui  −
* *2 2

=  n N  N −1 i 1= nN i 1   mi M i  

Estimate of variance
Consider

( ui yi ( mi ) − yS* 2 )( ui xi ( mi ) − xS*2 ) 
1 n
s=
*
bxy ∑
n − 1 i =1  

∑ ( xij − xi ( mi ) )( yij − yi ( mi ) ).


1 n
s=
mi − 1 j =1 
ixy

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 20
It can be shown that
 1 1 
E ( sbxy )= 1 N
* *
Sbxy + ∑ ui2  − Sixy
N i =1  mi M i 
E ( sixy ) = Sixy .
So
1 n  1 1   1 N  2 1 1  
E  ∑ ui2  −  sixy=
 ∑ ui  −  Sixy .
 n i 1= mi M i   N i 1   mi M i  
Thus

1 n  1 1 
Sˆbxy
*
= *
sbxy − ∑ ui2  −  sixy
n i =1  mi M i 
1 n  1 1  2
Sˆbx*2 = *2
sbx − ∑ ui2  −  six
n i =1  mi M i 
1 n  1 1  2
Sˆby*2 =*2
sby − ∑ ui2  −  siy .
n i =1  mi M i 

Also
 1 n   1 1  2  1 N  2  1 1  2
E  ∑ ui2  −  six =
 ∑ ui  −  Six 
 n i 1 =  mi M i   N i 1   mi M i  
 1 n   1 1  2  1 N  2  1 1  2
E  ∑ ui2  −  siy =
 ∑ ui  −  Siy .
 n i 1 =  mi M i   N i 1   mi M i  

A consistent estimator of MSE of yS**2 can be obtained by substituting the unbiased estimators of

respective statistics in MSE ( yS**2 ) as

 ( y ** ) ≈  1 − 1  ( s*2 − 2r * s* + r *2 s*2 )
MSE S2   by bxy bx
n N 
1 n 2 1 1  2
+ ∑ ui  −  ( siy − 2r sixy + r six )
* *2 2

nN i =1  mi M i 
1 1  1 n
∑ ( yi ( mi ) − r * xi ( mi ) )
2
≈ − 
 n N  n − 1 i =1
1 n  2 1 1  2  yS* 2
+ ∑ ui  −  ( siy − 2r sixy + r six )  where r * =
* *2 2
.
nN i =1   mi M i   xS*2

Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 21

S-ar putea să vă placă și