Documente Academic
Documente Profesional
Documente Cultură
Advisors:
George Casella Stephen Fienberg Ingram Olkin
Probability Theory
Independence, Interchangeability,
Martingales
Third Edition
, Springer
Yuan Shih Chow Henry Teicher
Department of Statistics Department of Statistics
Columbia University Rutgers University
New York, NY 10027 New Brunswick, NJ 08903
USA USA
Editorial Board
Ingram Olkin
Stephen Fienberg Department of Statistics
Department of Statistics Stanford University
Carnegie Mellon University Stanford, CA 94305
Pittsburgh, PA 15213-3890 USA
USA
George Casella
Biometrics Unit
Cornell University
Ithaca, NY 14853-7081
USA
987654321
ISBN 978-0-387-40607-7
To our teachers
J. L. Doob and 1. Wolfowitz
Preface to the Third Edition
vii
Preface to the Second Edition
Apart from new examples and exercises, some simplifications of proofs, minor
improvements, and correction of typographical errors, the principal change
from the first edition is the addition of section 9.5, dealing with the central
limit theorem for martingales and more general stochastic arrays.
ix
Preface to the First Edition
References
J. M. Keynes, A Treatise on Probability, 1921; MacMillan, London, 1943.
A. Kolmogorov, Foundations ot'the Theory of Probability, 1933; Chelsea, New York,
1950.
R. von Mises, Probability, Statistics and Truth, 1928; Wm. Hodge, London, 1939.
R. von Mises, Mathematical Theory of Probability and Statistics, 1931 (H. Geiringer,
editor), Academic Press, N.Y., 1964.
O. Ore, .. Appendix," Cardano, The Gambling Scholar, Princeton University Press,
1953; Holt, New York, 1961.
I. Todhunter, A History of the Mathematical Theory of Probability, 1865; Chelsea,
New York, 1949.
1 In the same notational vein, Theorem 3.4.2 signifies Theorem 2 or Section 4 of Chapter 3.
Contents
3 Independence 54
3.1 Independence, random allocation of balls into cells 54
3.2 Borel-Cantelli theorem, characterization of
independence, Kolmogorov zero-one law 61
xv
xvi Contents
11 Martingales 404
11.1 Upcrossing inequality and convergence 404
11.2 Martingale extension of Marcinkiewicz-
Zygmund inequalities 412
11.3 Convex function inequalities for martingales 421
11.4 Stochastic inequalities 432
Index 479
List of Abbreviations
xix
List of Symbols and Conventions
xxi
xxii List of Symbols and Conventions
UA* (U
AeA
A A =
AeA
A A) * A, A U AA =
* AeA A U AA)'
* ( AeA
n * (n
Ae A
A;. A =
.I.e A
Ai.) * A, A * n
AeA
AA = A* (n
AeA
A;.).
Then
A - U A A=
;'eA
n(A -
AeA
A A)' A - nA;. = U(A -
AeA AeA
AA)'
n UA
00 00 00 00
lim An
"-00
= U nAk,
n= 1 k=n
lim An
"-00
=
n= 1 k=n
k
and note that, employing the abbreviation i.o. to designate" infinitely often,"
lim An = {w: wEAn for infinitely many n} = {w: wEAn' i.o.}
(1)
lim An = {w: wEAn for all but a finite number of indices n}.
1.1 Sets and Set Operations 3
To prove, for example, the first relation, let A = {W: wEAn' i.o.}. Then
WE A iff for every positive integer m, there exists n ~ m such that wEAn,
that is, iff for every positive integer m, W E U:'=m
An i.e., iff WE t An· n:= U:'=m
In view of (1), lim An C lim An' but these two sets need not be equal
(Exercise 3). If lim An = ITii1 An = A (say), A is called the limit of the sequence
An; this situation is depicted by writing lim An = A or An -+ A. If Ate
A 2 C ... (resp. AI=' A 2 =' ...) the sequence An is said to be increasing
(resp. decreasing). In either case, {An' n ~ l} is called monotone.
Palpably, for every monotone sequence An, limn~oo An exists; in fact, if
{An} is increasing, limn~oo An = U:,=
1 An, while if {An} is decreasing,
limn~oo An = n:,=
1 An· Consequently, for any sequence of sets An,
00
lim An = lim UA k,
n-oo "-00 k=n
limA n = lim
n-+oo n-oo k=n
n
00
Ak •
EXERCISES I. 1
I. Prove (i) if An is countable, n ~ 1, so is U:= 1 An; (ii) if A is uncountable and B => A,
then B is uncountable.
2. Show that U:= 1 [0, n/(n + I» = [0, I), n:= I (0, I/n) = 0·
3. Prove that lim n-<Xl An C ITm n-<Xl An. Specify lim An and lim An when A 2j = B,
A 2i - 1 =C,j= 1,2, ....
4. Verify that U:'= 1 An = Iim n_<Xl U~ Aj and n:= I An = Iim n_<Xl nj= I A j . Moreover,
if {An' n ~ I} is a sequence of disjoint sets, i.e., Aj Aj = 0, i # j, then
<Xl
lim U A j = 0·
,,""cc j=n
5. Prove that ITmn(A n U Bn) = ITm n An U ITm n Bn and limn An' Bn = limn An' limn Bn.
Moreover, lim An = A and lim Bn = B imply limn (An U Bn) = A u Band
lim AnBn = AB.
6. Demonstrate that if B is a countable set and Bn = {(b I' ... , hn): hi E B for 1 ::; i ::; n l.
then Bn is countable, n ~ I.
7. Prove that the set S consisting of all infinite sequences with entries 0 or 1 is non-
denumerable and conclude that the set of real numbers in [0, I] or any nondegenerate
interval is nondenumerable. Hint: If S were countable, i.e., S = {sn, n ~ I} where
Sn = (x nl . Xn2 ,·· .), then (J - XII' I - Xn, ... , I - x nn ,.·.) would be an infinite
sequence of zeros and ones not in S.
8. If an is a sequence of real numbers, 0 ::; an ::; 00, prove that
U[0, an)
n=1
= [0, supan),
n21
U[0. (~)n]
n=1 n
# [0, sup (n_",=-~)n].
n n~l
9. For any sequence of sets {An, n ~ I}, define B I = AI' Bn+ 1 = Bn L\ An+!, n ~ 1.
Prove that limn Bn exists iff lim An exists and is empty.
4 I Classes of Sets, Measures, and Probability Spaces
Au A C = 0,
A - B = ABC, IA ~ IB iff A c B,
with the last inequality becoming an equality for all W iff AB = 0. Let A be an
arbitrary set and {A;., ,l E A} a class of subsets ofO. It is convenient to adopt
the conventions
U A). = 0,
;'e0
n
;'e0
A;. =0.
Moreover,
If A;. . A;.' = 0 forA., ,l' E A and,l '# ,l', the sets A;. are called disjoint. A class of
disjoint sets will be referred to as a disjoint class.
If {An' n ~ I} is a sequence of subsets of 0, then {I An' n ~ I} is a sequence
of functions on 0 with values 0 or 1 and
Moreover,
(1)
1.2 Spaces and Indicators 5
Equality holds in (I) iff {An' n ::?: I} is a disjoint class. The following identity
(2) is a refinement of the finite counterpart of (I) : For Ai c n, 1 :::;; i :::;; n, set
Then
(2)
In proof of (2), if for some WEn, I U7 Aj( w) = 0, clearly Sk( w) = 0, 1 :::;; k :::;; n,
whence (2) obtains. On the other hand, if I u74w) = 1, then W E A j for at least
one j, 1 :::;; j :::;; n. Suppose that W belongs to exactly m of the sets AI' ... , An.
Then St(w) = m, siw) = (D,···, sm(w) = 1, Sm+ t(w) = ... = sn(w) = 0,
whence
EXERCISES 1.2
1. Verify that
A~B = Ae ~ Be, C =A~B iff A = B ~ C,
00 00 00
U An ~ U Bn U (An ~ Bn),
, ,
C
1
n An ~ n B
00 00
n C U (An ~ Bn )·
1 I 1
2. Prove that (lim._ oo A.t = lim._ oo A~ and (lim n _ oo A.)e = Iim._ oo A~; also that
lim B. = B implies lim B~ = Be and lim A ~ B. = A ~ B.
3. Prove that IlimA n = ITiTI IAn and that IlimA n = lim IAn whenever either side exists.
4. If A. c n, n 2': 1, show that
6. Apropos of(2), prove that if B., is the set of points belonging to exactly m(1 ::; m ::; n)
of A" ... ,A., then
thenf~ i fo·
9. If I, and 12 are real functions on Q, prove that for all real x and rational r
(w: 1,(w) + f2(W) < x} = U {w: f.(w) < r}· (w: f2(W) < x - r}.
all ,.
Evidently, (ii) implies that for every positive integer n, U~ A jEd when-
ever AjEd, 1 ~j ~ n, while both (i) and (ii) entail A t A 2 Ed, if AjEd,
j = 1, 2,; also, since d is nonempty, 0 E d, 0 E d. Clearly, (iii) implies (ii)
by taking An = A 2 , n ~ 2. Note that a a-algebra is closed under countable
intersections.
Definition. The minimal algebra Iff' (resp. a-algebra, monotone class) contain-
ing a nonempty class Iff ofsubsets of0, is an algebra (resp. a-algebra, monotone
class) such that
i. Iff'::::> Iff,
1.3 a-Algebras. Measurable Spaces. and Product Spaces 7
ii. tff" :::> tff' whenever tff" :::> tff and tff" is an algebra (resp. a-algebra, monotone
class).
Such a minimal algebra tff' (resp. a-algebra, monotone class) containing tff
is also called the algebra (resp. a-algebra, monotone class) generated by tff and
is denoted by d(tff) (resp. a(tff), m(tff».
To demonstrate the existence of d(tff) (resp. a(tff), m(Iff)), let
Qs = {~: ~ :::> tff, ~ is an algebra (resp. a-algebra, monotone class)}
Then Qs is nonempty since So E Qs' Since an arbitrary intersection ofalgebras
(resp. a-algebras, monotone classes) is itself an algebra (resp. a-algebra,
monotone class), if £:0 = n£lteQS~' then £:0 is an algebra (resp. a-algebra,
monotone class) and £:0 :::> tff. Obviously, £:0 is minimal.
For an arbitrary class tff of sets, it is impossible to give a constructive
procedure for obtaining a(tff) (see Hausdorff (1957»). However, some in-
dication of its structure is given by
Theorem 2. If a A-class d contains an-class £:0, then d :::> a(£:0), the a-algebra
generated by £:0. In particular, if £:0 is an-class, A(£:0) = a(£:0).
8 I Classes of Sets, Measures, and Probability Spaces
X A j = {(w
j= I
t, W z ,· .. , w n): wjEA j C OJ, 1 S j S n},
n
AI X A z x ... x An = XA j ,
j= I
.~ d
I-I
j = a({.~
I-I
Aj : Aj E d j , I s js n}),
j~I(Oj,dJ = (~Oj' j~d}
Then X?=I OJ is called the product space (with components OJ, 1 sis n).
Sets of the form X7= I Ai with Ai C 0i, 1 sis n, are called (n-dimen-
sional) rectangles of the product space X7= I OJ, and, moreover, if A j E d j ,
1 sis n, they are dubbed measurable rectangles or rectangles with measur-
able sides. Clearly, the intersection (but not the union) of any two measurable
rectangles of a given product space is a measurable rectangle in that space. In
other words, the class of measurable rectangles of X7 = I OJ is an-class.
Theorem 4. If (OJ, d), 1 sis n are measurable spaces, the class .91 of all
finite unions of disjoint rectangles X7= I Ai with A j E d j , l s i s n, is the
algebra generated by the class of all measurable rectangles of the product
space X7= I OJ.
PROOF. Let "§ denote the n-c1ass of measurable rectangles of X7= I OJ' Now
d is also a n-c1ass since if Ai = Uj:: 1 Eij E d, i = 1,2, with Eij E '§, then
AI' A 2 = Uh.k
E lh E 2k E d. Moreover, if E = E I X ... x En E '§,
E C
= EI X ... X En _ I X E~ u EI X ... X En _ 2 X E~ _ I X On
n
U ... u E~ X O2 X ... X On = UD j (say) E d
I
12 I,. .
° °
able sides, is also the a-algebra generated by the algebra d of Theorem 4.
Xt
The points of = I OJ, (0 I X 02) X 3, and I X (0 2 X 03) are respec-
tively of the form (WI, W2, W3), «WI' W2), W3), and (WI, (W2' w 3 )), where
W j E OJ for i = 1,2, 3 and formally they are different. However, between any
two of the three points there is a natural one-to-one correspondence and so
each one ofthe points will be identified with (WI' W2' W3)' Similarly, «WI' w 2 ),
(W3' w 4 )) will be identified with (WI, W 2 , W 3 , w 4 ), etc. Under this convention
it is easily seen that for 1 S m < n
" "
ff = X
j=
d·
I"
Yf = X
j=m+ 1
d;,
ff' = {.X
1= 1
Aj : Aj E d j , 1 $; i $; n}, <;1' = {.X.= 1
A j : A j E d j, 1 $; i $; m},
Yf' ={ X A j: A j
i::=m+ 1
E d j, m + 1 $; i $; n}.
By definition, ff = ax(ff'), <;1 = ay{<;1'), Yf = az{Yf'), and <;1 x Yf =
0'( {A x B: A E <;1, BE Yf}), where aQ is the a-algebra relative to the space Q.
Now if X?=I AjEff', then X:"=I A j E<;1' and X?=m+1 AjEYf', whence
X?=I A j = (X:"=I A j ) x (X?=m+1 AJE<;1 x Yf,implyingff c <;1 x Yf,On
theotherhand,ifAE<;1andBEYf,thenA x B = (A x 0m+1 X .,. X 0") (\
(0 1 X ... x Om X B) E ff and so <;1 x Yf c ff. 0
In the product measurable space X?= I(Oj, d j ), n = 1,2, ... , sets of the
form A x (X?=m+ 1 OJ with A E X:,,= 1 d j for 1 $; m < n are called cylinders
with m-dimensional base A. The quintessence ofa cylinder set is that all but a
finite number of coordinates are unrestricted. The notion of a cylinder is
important in the case of an infinite-dimensional product space.
Let (OJ, dJ, i = 1,2, ... , be measurable spaces and define
XA
OC>
Xd
j= 1
j = 0'(<;1),
"
d" = X" d;, (0", d") = X (OJ, d j ), 1 $; n $; 00.
j= I j= 1
1.3 a-Algebras, Measurable Spaces, and Product Spaces II
x + (- (0) = - 00 +x = - 00,
~=_x_=o,
00 -00
x·oo=oo·x=(-x)·(-oo)=(-oo)·(-x)=oo or ° or -00,
accordingasx > O,X = O,x < 0. Note that oojooand 00 - 00 are not defined.
For x E R, the sets [ - 00, x], (- 00, x), (- 00, x], [ - 00, x), [x, 00], (x, (0),
[x, (0), (x, 00] are termed infinite intervals.
The elements of the a-algebra !!d generated by the class of infinite intervals
of the form [ - 00, x), - 00 < x < 00, are known as the Borel sets (of the line)
or linear Borel sets or Borel sets in R. The measurable space (R,!!d) is called the
Borel line or I-dimensional Borel space.
The product measurable spaces (R", .j$") and (ROO, !!d OO ), emanating from
the Borel line (R, !!d), are called the n-dimensional Borel space and infinite-
dimensional Borel space respectively. Moreover, the sets of !!d" are termed
n-dimensional Borel sets. 1 ~ n ~ 00. Since every point WE ROO is an infinite
sequence of real numbers, the infinite dimensional Borel space (ROO, !!d OO ) is
also alluded to as a sequence space.
For any interval J c R the a-algebra generated by the class of all subin-
tervals of J coincides with f!4 n J according to Theorem 3. Thus, it is natural
to describe the a-algebra in question as the class of Borel subsets of J.
Similar comments apply to any n-dimensional interval (rectangle) J" c R".
EXERCISES 1.3
J. Prove that .fi1 is a a-algebra if and only if it is both a A.-class and a n-class. If a class
,91 c f», then a(.fi1) c a(£il).
2. If;f' is an algebra such that for every disjoint sequence {F., n ~ J} of sets in .?
U;,,;I F. E Y;, then Y; is a a-algebra.
3. Let (Qi' ,fi1,) be measurable spaces and .9I i = IT(~,) with Q i E ~i' I :s:; i :s:; n. If
f§ = {A I x A 2 X ... x A.: Ai E ~i' 1 :s:; i :s:; n}, then IT(~) = .fi1 I x .91 2 X ..• x .91.
011 Q1 X Q2 X ... x Q •.
12 I Classes of Sets. Measures. and Probability Spaces
6. If (Qi' .cY',), i ~ I, are measurable spaces, the class Iff of all cylinder sets of Qi Xf
with bases in X';' ·cY'i for some In ~ I is an algebra, but not a a-algebra. Moreover,
setting!!iJ = {XT= I Ai: Ai E ,Wi} verify that X,oc;, I .cY'j = a(fl1) = a(6').
7. Let!!iJ be a n-class of subsets of Q and '§ the class of all finite unions of disjoint
sets of!!iJ with 0 E!!iJ. If DC E '§ for every D E!!iJ, prove that '§ is the algebra generated
by!!iJ.
8. Show that the class of Borel sets :fI may be generated by {(x, N], - x., < x < x.,}
or by.'I' = {{ + Xl}. [ - Xl, Xl], [a, b), - N :-0; a :-0; b :-0; X }.
10. If.<I is the class of open sets of Qn = ( -XJ, XJ t, then a(.<I) = .fln n Qn.
° °
Let 0 1 and O 2 be reference sets and X a function defined on 01' with values in
2, the latter being denoted by X: 0 1 --+ 2, For every subset A of O 2 and
class '§ of subsets of il 2, define
The set X - I(A) and the class X - I(,§) are called respectively the inverse images
of the set A and of the class '§. Clearly, if for each AE A the set A;. C 2, then °
X-I(UA;.) = UX-I(A;.),
;'EA ;'EA
°
and hence the inverse image X - I(,§) of a a-algebra '§ on O 2 is a a-algebra on
OJ and the class {B: B c 2 , X - I(B) E ff} is a a-algebra on O 2 if .11' is a
a-algebra on 1 , °
Lemma I. For any mapping X: 0 1 --+ O 2 and class d of subsets ofil 2 ,
X-I(an,(.w'» = an,(X-1(d».
PROOF. Since X- l (an 2 (d» is a a-algebra containing X-I (d), it must be shown
for any a-algebra .@ of subsets of ill with'@ ~ X-I (d) that.@ ~ X-I (an 2 (d».
Since.@ is a a-algebra, the class fJl = {B: B c 02,X- l (B)E.@} is likewise a
1.4 Measurable Transformations 13
<T-algebra and this together with the relation fJl ::J d implies fJl ::J <Tn,(d).
Thus
o
Suppose that X: Q t ~ Q 2 and Y: Q 2 ~ Q3' If Y(X) is defined by the usual
composition, that is,
Y(X)(W) = Y(X(w», WE Q,
~ = {.X
,=
A j : AjEdj , I ~ i ~ n},
1
~= U{X
m;;:; 1 i=l
A j : A j E d j, 1 ~ i ~ m, and Ai = Qi' i > m}.
Then <T(~) = X7= I d j and by the prior lemma
°
by (x, 00], - 00 < x < 00, it follows from Lemma I that a real-valued func-
tion X on is measurable iff for every finite x, {X(w) < x} Ed or
{X(w) > x} Ed, XE( -00, (0).
In the special case of a measurable transformation .r from the Borel line
(R, f!J) to itself, f is termed a Borel or Borel-measurable function. Since every
open set in ( - 00, (0) is a countable union of open intervals and every open
interval is a Borel set, every continuous function on (- 00, (0) is a Borel
function. Hence a real function on [ - 00, 00] with range [ - 00, 00] which is
continuous on ( - 00, (0) is always a Borel function.
Similarly, for any finite or infinite interval J c R measurable transforma-
tions from (J, f!J. J) to (R, f!J) are termed Borel functions or Borel functions
on J. Analogously, for any finite or infinite rectangle r eRn, measurable
functions from (In, &In . r) to (R, f!J) are called Borel functions of n variables
or Borel functions on r. Since every open set of ( - 00, 00 t is a countable
union of open rectangles, continuous functions on ( - 00, 00 t are likewise
Borel functions.
1.4 Measurable Transformations 15
The a-algebra n:= 1 a( X n' X n + I' ...), which is called the tail a-algebra of
{X n , n ~ I}, wi11 be encountered in Chapter 3.
1.4 Measurable Transformations 17
Theorem 4. Let X I' ... , X n be real measurable functions on (n, d). A (real)
function on n is O"(X j, . . . , X n)-measurable iff it has the form f(X (, ... , X n),
where f is a Borelfunction on W.
PROOF. Let C§ denote the class of Borel functions on Wand define
.Yf' = U(X t, X n): f E C§, f(X I ' " ' ' X n) 2 O},
... ,
~ = {{(w: Xlw):s; Xi' 1 :s; i:S; n}: -00 < Xi < 00,1 :s; i :s; n}
Then ~ is a n-class of subsets of n, O"(~) = O"(X I' . . . , X n), and IDE YI' for
D E~. Moreover, YI' is a A.-system. In fact, 1 E..if, and if /;(X I' . . . , X n) E..if,
C i is a finite real number for i = 1,2, and
Now let fm(X 1 , ... ,Xn)E..if,m21, and fm(Xj, ... ,Xn)i f(XI, ,X n).
Set gm = max(.fl' ... ,fm)' Then gm i g E C§ and O:s; fm( X I' , X n) =
gm(X I"'" X n) :s; fm+ I(X I"'" X n)· Therefore,
f (X I' . . . , X n) = g( Xl' ... , X n) E YI'
and YI' is a A.-system. By Theorem 3, .Yf' contains all nonnegative O"(~)
measurable functions. Hence, if Y is 0"( X I' . . . , X n)-measurable, both Y +
and Y- are in Yl'sothat Y = y+ - Y- =f(X1, ... ,Xn)forsomeBorel
function f on Rn.
Conversely, let f be a Borel function on Rn and
X(w) = (X I(W), ... , Xn(w», WEn.
Since Xi is measurable from (n, O"(X (, ... , X.» to (R, :JI) for 1 :s; i :s; n,
X is measurable from (n, 0"( X I. X z, ... , X n» to (W, :JIn) by Theorem 1. By
hypothesis,f is measurable from (W, :JI') to (R, :JI), whence f(X (, ... , X n) =
f(X) is measurable from (n, O"(X (, ... , X.» to (R, 86), that is,.f(X I' . . . , X n)
is O"(X" ... , Xn)-measurable. 0
In similar fashion one can prove
EXERCISES 1.4
1. Let T: 0 1 --+ O2 and A.l C 0 1 for AEA. Show that T(U.l A.l) = U.l T(A.l) but that
T(A.l - A.l') need not equal T(A.l) - T(A.l')' where T(A) = {T(wd: WI EA},
T(d) = {T(A): AEd}. If T is-one-to-one and onto, then d a O'-algebra on 0 1
entails T(d) is a O'-algebra on 02'
2. In the notation of Exercise 1, prove that A c T-l(T(A») with set equality holding
if T is one-to-one. Also, T(T-l(B» c B c O 2 with equality if T is onto.
3. A function X on (0, §, P) is called simple if for some finite integer n and real
numbers Xl' ... , X n it is representable as X = I7=lx;IA, for {A i ,1 $ i $ n} a
disjoint subclass of §. Prove that any non-negative measurable function is an
increasing limit of non-negative simple functions.
4. Prove that (X + Y)+ $ X+ + Y+, (X + Yf $ X- + Y-, X+ $ (X + Y)+ + Y-.
Also if Bnt B, then X-I(BnH X-I (B).
5. Let X and Ybe real measurable functions on (0, d) and c any real constant. Show
that {w: X(w) < Y(w) + c}, {w: X(w) $ Y(w) + c}, {w: X(w) = Y(w)} Ed.
6. Prove that if X is a real measurable function on (0, d), so is IXI. Is the converse
true?
7. If X and Yare real functions on (0, d) then X + iY is measurable iff (X, Y) is
measurable from (0, d) to (R 2 ,gJ2).
8. If (OJ, d;), i = 1,2 are measurable spaces and X: 0 1 --+ O 2 then X is a measurable
transformation from (Ol,dl ) to (02,d2) provided X- l (£0) c d l for some class
£0 of sets for which 0'(£0) = d 2 •
9. Prove that any real monotone function on R is Borel measurable and has at most
a countable number of discontinuities.
to. For any linear Borel set B and T.v. X prove that O'(XI[XeBI) = X-I (B) n O'(X).
Note in the u-additive case that the definition precludes conditional con-
vergence of the series. Clearly, if 0 E d, countable additivity implies finite
additivity.
From a purely formal vantage point the prior definition relegates prob-
ability theory to a special case of measure theory. However, such important
notions as independence (Chapter 3) and other links between probability and
the real world have nurtured probability theory and given it a life and direc-
tion of its own.
In addition to the basic property of u-additivity, a measure p, on an algebra
d is monotone, that is, p,{A.} ~ p,{A 2 } whenever Al c A 2 , AjEd, i = 1,
2 (Exercise l),and, moreover,subadditive, that is, ifUI" A j E d, A j E d,j ~ 1,
then
(1)
PROOF. Since s1 is the class of all finite unions of disjoint sets in Y', every
A E s1 has a finite partition {A., 1 ~ n ~ m} in .C/'. For such an A, define
1.5 Additive Set Functions, Measures, and Probability Spaces 21
= L L Jl(CnB j ) = L L Jl(CnB)
j;1 n;1 n;1 j;1
m
= L Jl(Cn) = v(A). o
1
In this case Ji. is subadditive on.xl, i,e" ifUf AjE.xI, AjE.xI,j ~ 1, then
Ji.(Q A j) ~ ~ Ji.(A).
iv. ifJi. is a-additive, then for every decreasing sequence An E .xl with Ji.(A t} < 00
and limn An = A E .xl,
(4)
The second part of (iii) follows from the first part and (ii).
(iv) If J.l is cr-additive, and An' n ~ 1, is a decreasing sequence of.91 sets with
J.l(A I) < 00 and limn An = A Ed, then Al - An is an increasing sequence in
.91, whence by (iii)
Since J.l(A 1 ) < 00, (4) holds. Conversely, if {An} is any cr-partition in .91 of
some set A Ed, then Ui'=nAjE.9I, n~ 1, and Ui'=nAj!fIn1An=0,
whence by hypothesis lim J.l{Ui'=n A j} = O. By finite additivity,
A finite measure space is a measure space (0, ff, J.l) with J.l finite. In such a
case, the finiteness proviso in Theorem 2 (iv) is superfluous.
Corollary 2. If (0, ff, P) is a probability space, P{lim n An} = limn P{A n} for
every sequence of sets An E:F whose limit exists.
EXERCISES 1.5
1. Show that a finite measure II on a class .'# is (i) subtractive, i.e.. A; E .91, i = 1, 2.
AI C A 2 and A z - AI E.W implies/l(Az - AJl = /l(A z ) - /l(A I ), and (ii) monotone,
i.e., /l(A I) =:; /l(A 2 ) with A; as in (i). In view of (i), if there is one set in d wIth finite
measure, the proviso 11(0) = 0 is automatically satisfied.
2. If (Q, sd, /l) is a measure space, prove that /l(lim An) :5 lim /l(A n) for An E d.
Analogously, if /l(U~n A;) < 00, some n ~ I, then /l(rrm An) ~ rrm /l(A n).
p{lim An}
11-+00
= lim p{
n ..... oo
n
J=n
A j }, p{lim An}
n-oo
= lim p{
n-oo
UA
J=n
j}
. N.(A) }
£t = { A Ed: /l{A} = 11m - - exists.
"-+00 n
Prove that £t is closed under complementation and finite disjoint unions but that
nonetheless £t is not an algebra. Also /l {A}, called the asymptotic density of A, is
additive but not a-additive. Hint: Let Bk = {odd integers in [2 2\ 22k + I)} and
B~ = {even integers in [2 2k - 1, 2 Zk )}. If B = Ui"= 1(B k U B~) and A = {odd integers
ofQ}, then A E~, BE~, but AB¢~.
8. Letfbe a monotone increasing function on Q = [0, I) such that 0 =:; f =:; I, and ~
the class of finite unions of disjoint intervals [a, b) c Q.
i. ~ is an algebra.
1.6 Induced Measures and Distribution Functions 25
ii. Put Il(A) = D~ \ (f(b j) - f(a}) for A = Uj~ I [aj' b} E <§, where 0::0; a l <
b l ::0; a2 < b 2 ::0; ... ::0; a. < b. ::0; I. Then Il is additive on <§.
iii. If Il is u-additive on <§, thenfis left continuous, i.e.,«£) = limo <';-0 f(t - (i) for
every t E Q.
iv. Iffis left continuous, then Il is u-additive on r.§. Hint: For part (iv), see the proof
of Lemma 6.1.1.
9. Let (Q, ff, Il) be a measure space and r.§ a sub-u-algebra of .~. Prove that
<§* = {G fj, N: G E <§, N E .~, Il{N} = o}
is a sub-u-algebra of ff. Then <§* is called the completion of <§. When <§ = ff then
<§* = ff and so the completion of ff is rather defined by ff* = {A fj, M: A E .~,
MeN E .~,Il{N} = O}. If Il* on ff* is defined by 11*{A 11 M} = I1{A}, check that
(Q, .~*, Il*) is a measure space. It is called the completion of (Q, .r;;, 11).
10. (Cantor ternary set.) Let II = III = (t,~) be the "open middle third" of I = [0, I].
Analogously, for j ~ 2, let I j • b k = I, ... , 2j -\, be the open middle third of each of
the 2j - 1 intervals of I - UI=! Ij, where I j = Uf~-" I j . k , j ~ I. Show that the
Cantor set C = I - U~ \ I j satisfies: (i) x E C iff the digit I is not needed in the
triadic expansion x = L~ I xjr\ where, in general, X j = 0, 1,2; (ii) C is un-
countable; (iii) C is perfect (all its points are limit points); (iv) C is nowhere dense
(the closure of C contains no nondegenerate interval); (v) if Il is a measure on the
Borel sets of [0, I] whose restriction to intervals coincides with "length" (such a
" Lebesgue measure" appears in Section 6.1), then Il {C} = O.
It is readily verified thatJ-lx isa measureond and consequently that (n, d, J-lx)
is a measure space.
In particular, a random variable X on a probability space (0, ff, P)
engenders a new (Borel) probability space (R, 14, P x), where 14 is the class of
Borel sets of R = [ - 00, 00]. To focus attention on the latter, i.e., to consider
X of primary interest, is tantamount to restricting the former to (0, a(X), P).
Associated with any random variable X on a probability space (0, ff, P) is a
real function F x called the distribution function of the random variable X and
defined for all x E[ -00,00] by
Fx(x) = P{w: X(w) < x} = P{X- 1(( -00, x»)} (1)
26 I Classes of Sets, Measures. and Probability Spaces
It follows from Theorem 1.5.2 and the fact that a r.v. is finite a.s., that the
"distribution function of any r.v." is a distribution function in the sense ofthe
prior definition. Conversely, for every distribution function G there exists a
r.v. X on some probability space (n, §', P) such that Fx = G. The prooffor a
special case is given in this section but a general treatment will be deferred to
Section 6.3.
Analogously to tbe preceding, n random variables X I' ... , X n on a
probability space (n, §', P) induce via the map X = (X I"'" X n ), the new
(Borel) probability space (R n, 14n, P x), where 14n is the class of n-dimensional
Borel sets of R n and P x is the induced measure. Rather than shift from the basic
probability space, one may restrict attention to (n, a(X I" .. , X n ), P).
Associated with the random vector X i.e., with the n-tuple of random variables
(X I> ••• , X n), is a real function on Rn called the (joint) distribution function of
(X I' ... , X n ), denoted by F XI, ...,X n and defined by
The joint dJ. F XI, .... Xnsatisfies properties akin to (i), (ii), (iii) (see Section 6.3),
but monotonicity in each argument is insufficient and the analogue of (i)
involves a mixed difference. Nonetheless, if a dJ. on Rn is properly defined, it is
again true that the "dJ. of (X I' ... , X n )" is a dJ. on R" and that, conversely,
givenadJ.G(x l , ... , xn)onRn,therealwaysexistrandomvariablesX t ,··" X n
on some probability space whose joint dJ. F XI, .... x n = G.
In the same vein, a sequence {X n , n 2: I} of r.v.s on a probability space
(n, §', P) induces a Borel probability space (ROO, 14 00 , P x) via the map
X = (X I' X 2' ...) or alternatively places in relief (n, a(X n , n 2: 1), P). Note
that the dJ.s of all finite n-tuples ofr.v.s (X it ..... Xi) are determined via
It will be proved in Section 6.4 that if the dJ.s Git in are prescribed in a
consistent manner for all choices of indices I .::;; it < < in and all n 2: 1,
there exists a sequence {Xn,n 2: 1} ofr.v.s on a probability space such that
1.6 Induced Measures and Distribution Functions 2.7
G(x) = L
j:Xj<x
Pj' XER, (3)
where Pj > 0 for allj, LalljPj = 1, and S = {Xj: 1 5;j < n + 1 5; oo} is a subset
of( -<X), (0). The associated function
Theorem 1. If(X I " ' " X n ) and (YI , ••• , y,,) are random vectors with identical
distribution functions, that is,
then g( X I' ... , X n) and g( YI , ... , y,,) have identical dJ.s for any finite Borel
function 9 on W.
PROOF. It follows from Theorem 1.4.4 that g(X t, ... , X n) and g( YI , ... , Yn)
are r.v.s. Set
rg = {B: BE BIn, P{(X I> ... , X n) E B} = P{(YI , ... , Yn) E Bn,
!0 = {D: D = X[-
J= I
00, c), cj real}.
Corollary 1. If(X I"'" X n ) and (YI , ••• , y") are random vectors with identical
distribution functions, thenfor any Borelfunction 9 on Wand linear Borel set B
EXERCISES 1.6
1. Prove that G as defined in (3) is adJ.; verify that (n, .~, P) as defined thereafter is a
probability space and that F x = G.
2. If P,>. I ~ 0, PES, t E T, and Lie
T Lpes PI'. I = I, where Sand T are countable subsets
of (- 00, Xi), define a probability space and random variables X, Yon it with
Fx(x) = L LP
p<xteT
p ." Fy(Y)= L LP
t<ypeS
p ."
Hint: Take n = S x T.
3. Prove that if H(x) = P {X ~ x}, where X is a r. v. on (n, .~, P), then H is a nondecreas-
ing right-continuous function with H( - 00) = 0, H( + 00) = 1. Some authors define
such functions to be d.f.s.
4. For adJ. F, the set
Comments
The notion of a n-class and A-class (Section 3) seems to have originated with
Dynkin (1961).
References
J. L. Doob, "Supplement," Stochastic Processes, Wiley, New York, 1953.
E. B. Dynkin, Theory of Markol' Processes (D. E. Brown, translator), Prentice-Hall,
Englewood Cliffs, New Jersey, 1961.
Paul R. Halmos, Measure Theory, Van Nostrand, Princeton, 1950; Springer-Verlag,
Berlin and New York, 1974.
Felix Hausdorff, Set Theory (J. Aumman et al., translators), Chelsea, New York, 1957.
Stanislaw Saks, Theory of the Integral, (L. C. Young, translator), Stechert-Hafner,
New York, 1937.
2
Binomial Random Variables
30
2.1 Poisson Theorem, Interchangeable Events, and Their Limiting Probabilities 31
and in part to its leading (in the limit) to even more important distributions,
such as the Poisson and the normal.
As is customary, the combinatorial symbol (j) abbreviates n !/j !(n - j)! for
integers n ~ j ~ 0 and is defined as zero when j > n or j < O.
The binomial d.f. is a discrete dJ. with jumps of sizes (j)pi( 1 - p)" - i at the
points j = 0, 1, ... , n. In other words, its p.dJ. is completely characterized
(Section 1.6) by S = {O, 1,2, ... , n} and {(j)pi(1 - p)"- i, ~ j ~ n}. Here,
p E (0, 1) and n is a positive integer.
°
The construction of Section 1.6 shows that it is a simple matter to define
a probability space and a r.v. X on this space whose associated dJ. is binomial.
Thus, a r.v. X will be alluded to as a binomial r.v. with p.dJ. b(k; n, p) if for
some positive integer nand p E (0, 1)
The Poisson d.f. is a discrete dJ. with jumps of sizes )./e -)./j! at the points
j = 0, 1,2, ... , and so X will be called a Poisson r.v. with p.dJ. p(k;)") if for
some)., E (0, (0)
).,ke -)'
P{X = k} = k! == p(k; ).,), k = 0, 1,2,.... (2)
x 1
= - - e _(u-II)2/2.,.2 du,
G( x )
J- 00 afiJc
XER = [- 00,00]. (3)
e
The parameters of the distribution are E ( - 00, (0) and a 2 E (0, (0). Here,
a 2 = 0 may be regarded as a degenerate distribution. It is customary to
denote the standard normal distribution corresponding to the case = 0, e
a 2 = 1 by
1
J
x
<l>(x) = - - e- u2 2
du, q>(x) = -1- e- x 2 / 2 . (4)
fiJc
/
-00
fiJc
Here q>(x) is called the (standard) normal density function. A r.v. X is
e
normally distributed with parameters and a 2 , written X is N(9, ( 2 ) if
P{X < x} X-
= G(x) = <I> ( -a- e) , XER. (5)
32 2 Binomial Random Variables
Theorem I (Poisson, 1832). IfSn is a binomial LV. with p.d.f. b(k; n, Pn), n ~ I,
and as n -+ 00, npn = A. + o(1)for some A.E [0, (0), then/or k = 0, 1,2, ...
A.k -).
lim P{Sn = k} = -kef .
n-oo .
P{Sn = k} = (~)p~q~-k
-_ n(n - 1)··· (n - k + 1) [A.
-+0 (1)]k[
- A. (1)]n-k
1--+0-
k! n n n n
n n
1)
= -n (n-- - ... (n - k +
n
1) -k!1[A. + 0(1)] k
o
By refining the arguments of the preceding proof, a more general result is
obtainable. To pave the way for the introduction of a new p.dJ. subsuming the
binomial p.dJ., two lemmas which are probabilistic or measure analogues of
(2) and (3) of Section 1.2 will be proved.
q> (0 A
1
j) = I q>(A) - l$h<j,$n
j=l
L q>(AhAj,) + ... + (_l)n-lq>(A1A z ··· An)'
(6)
2.1 Poisson Theorem, Interchangeable Events, and Their Limiting Probabilities 33
PROOF. Set CPj = cp(A), CPh.j, ... j. = cp(AhAj, ... A j), r ~ 2, Clearly, (6) holds
for n = 1. Suppose, inductively, that it holds for a fixed but arbitrary positive
integer m in [1, n), whence Icp(U~+ 1 A)I < 00. Now,
(7)
cp(A m+l ) = cp(y AjA m+l ) + cp(Am+1 - y A j).
where Icp(U~ Aj)1 < 00, Icp(Am+I)1 < 00, Icp(U~ AjAm)1 < 00. By the
induction hypothesis
(9)
Corollary 2. If {A I' " " An} are interchangeable events ofa probability space
(Q, ~, P) with Pj defined by (10), then
The novel part of the next lemma is (13), since for A = Q, (12) is merely the
complementary version of the Poincare formula (9),
34 2 Binomial Random Variables
PROOF. By Lemma 1
represents Bmas a disjoint union, (13) follows by summing over il"" ,im in
(14). 0
Corollary 3. If AI> ... , An are events of a probability space (0, iF, P) and
Tk = II sj, < ... <jkSn P{Aj,A h AjJ for k = 1,2, ... , n, then
P{exactly m of AI' , An occur}
m 2 0.
Am+i
- L
00
(-l)i~ = It + 12 + 13 (say).
i=no m.J.
36 2 Binomial Random Variables
Since A. n + o( I) = np\n) ~ ..1.0 , there exists A. > such that A.n < A., all n. For
°
..1. 0
any I> > and fixed m, choose no to satisfy
00 A.m+j
L -,-.,<1>,
j=nom.J.
whence 11 3 1 < I> and via (18)
n-m n(n - 1)··· (n - m - j + 1) n-m lm+ j
II 21 ~ L
j=no
,.,
m.J.
pl::~ j::;; L ~<
j=nom.J.
1>.
= l~ + j + 0(1) - l~ + j = 0(1),
n.
, p(n) . _ lm+ j > (n _ m _ J')m+jp(n) . _ lm+j
.),m+J n - m+] n
( n-m-J.
= I_ m + J.)m+ j nm+jp(n)._lm+ j
( n m+J n
L Ib(k; n, p) -
k=O
p(k; l)l-+ ° (21)
as n -+ 00, np -+ l E (0, (0), where b(k; n, p) is defined as zero for k > n. This is
an instance of the more general phenomenon of the following
Then, as n --+ 00
00
L Pj <
j>N
B. (24)
N
L IPn,j - pjl --+ 0, (25)
j= 1
whence as n --+ 00
(26)
IIri1 L IPn,j -
n-oo j= 1
pjl = Iiffi
n-oo j>N
L IPn,j - pjl :::;; Iiffi
n-oo j>N
L (Pn,j + p) :::;; 21';,
which, in view of the arbitrariness of 1';, is tantamount to (23). o
EXERCISES 2.1
1. Prove that for n = 1,2, ...
2. Verify that
L kp(k; 1) = L
00 00
1= Pp(k; 1) - 12,
k=O k=O
L• p(k; 1) = 1
-
foo e-Xx' dx,
k=O n! A
L k p(k; 1) = 1
00
3 3
+ 31 2 + 1,
.
k=O
3. Verify for all positive integers nl> n2 and nonnegative integers k that
I ( ( - I)N - m)
pIN) = _ 1_ I + -I - -I + ... + - -- (i)
m m! 2! 3! (N - m)! '
(N) e-
I
I I (ii)
IPm --;T <m!(N-m+ I)!'
then
2m:o=::; 11.
8. Find the formula analogous to (15) for the probability that at least m of the events
AI>"" An occur.
9. Let Po = I and Pm' I :0=::; m :0=::; 11, be real numbers such that
m
qm = (n)m ni: (_IY(" -: m)Pm+
)=0 ]
j ~0
for 0 :0=::; m :0=::; 11. Then L::,=o qm = Po = I and there exists a probability space
(Q, .¥, P) and interchangeable events A I' ..• , An with Pi = P{A I A 2 ... A j},
I :O=::;j:O=::; 11. Hint: Take Q = {w = (WI>' .. , wn):w j = Oor I, I :O=::;j:O=::; Il}, Y; = class
of all subsets ofQ, and P({w}) = qml('~,) whenever D= I w j = m. Set
Aj = {w = (wl, ... ,wn):wj = I},I :O=::;j:O=::; 11.
2.2 Bernoulli, Borel Theorems 39
10. Verify that if Pi = pi in (16) i.e., in Exercise 9, the probabilities qm coalesce to the
binomial and P{A i, A i2 ··· Ai.} = n~=, P{A i), 1 ~ i, < ... < ik ~ n.
12. Prove that if r = r. = A. N In -+ CX) and A. -+ A E [0, CX) in the hypergeometric case
of Exercise II, then Pi = P}·' satisfies (18) and (19) of Theorem 2, so that the hyper-
geometric p.dJ. tends to the Poisson p.dJ. under the stated condition.
Prove that
PROOF.
p{ISn_pl~e}=p{ISn-npl~ne}=
n
L
Ik-npl2:ne
P{Sn=k}
~ L (k -2 ;)2 P{Sn = k}
Ik-npl2:nt ne
1
L (k -
n
~ 22 np)2 P{Sn = k}.
ne k=O
40 2 Binomial Random Variables
The prior inequality is a special case of the simple but very useful Tchebychev
inequality discussed in Chapter 4. Moreover, the last sum which will be
identified in Chapter 4 as the variance of Sn (or its dJ.) equals
2~ (n-2)! k-2n-k
= n(n - 1)p /;'2 (k - 2)!(n - k)! P q
I) ~ (n-1)! k-ln-k 22
- (2 np - np/;'l (k _ 1)!(n _ k)! P q +n p
Therefore,
o (2)
Theorem 2. Let Sn, n :2: 1, be binomial LV.S with p.dJ.s b(k; n, p), 0 < p < 1,
n :2: 1. Then for every t; > 0
Since
o ~j ~ n,
2.2 Bernoulli, Borel Theorems 41
it follows that
f (k - np)4(n)pkqn-k
k=O k
p{ I Sn - p
n
I ~ t;} = Ik-npiL ~n£P{Sn = k} ~ n; t; 4 f (k -
k=O
n
np)4(k )pkqn-\
and so
00
p{lim Sn = p} = 1.
"-00 n
PROOF. According to Theorem 2, for every e > 0
e > O.
Thus, by Lemma 1
then
lim Bn(P) = f(P) uniformly for p E [0, 1]. (4)
PROOF. Let 8 n be a binomial r.v. with p.d.f. b(k; n, p). Since every continuous
function on [0, 1] is bounded and uniformly continuous thereon, I f(P)1 ~
M < 00 for p E [0, 1], and for every e > 0 there exists () > 0 such that
If(P) - f(P') I < eiflp - p'l < () and 0 ~ p,p' ~ 1. Then,settingq = 1 - P
and An = {j: Ij/n - pi < ()},
~ e L bU; n, p) + 2M L
An A:i
bU; n, p)
By (2),
P {I Sn I } P(1-p)
--;; - p ~ () ~ n{)2 ~
1
4n{)2'
and so if n ~ M(e{)2)- 1,
PROOF. Set A = {oo: lim n _ oo y" = Yo} and suppose that PtA} < 1 or equiv-
alently PtA"} > O. Since A" is countable and, moreover, a countable union of
null events is a null event, there exists 00 0 E A" with P{ oo o} = b > O. Moreover,
000 E A" implies that for some t:: > 0 and subsequence nj of the positive
integers, I y"/ooo) - Yo(oo o)! > t::,j ~ 1, whence
EXERCISES 2.2
1. Verify that if np ...... ). E (0, (0), then D kib(k; n, p) ...... LQ' kip(k; ).), j = 1,2. Hint:
Recall Exercise 2.1.2.
2. If gm().) = 2:;;"=0 (k - :A.ymp(k; A), show that gd:A.) = 0, gz(:A.) = g3(:A.) = :A., g4(:A.) =
3:A. z + A, g6().) = 15).3 + 25A z + A.
3. (i) Prove a weak law of large numbers where Sn has p.d.f. p(k; n).), n 2: I, that is,
P{I(Sjn) - AI > e} = 0(1), e > O. (ii) If {X, X n, n 2: I} are LV.S with a common
d.f. and n P{ I X I > n} = 0(1), then (I/n)max, ';i,;nl Xii .f, o.
4. Prove a strong law of large numbers where Sn has p.d.f. p(k; n:A.), n 2: 1, that is,
lim n_ oo (Sn/n) = :A., a.c. Hint: Consider P{lSn - n).1 4 > n4 e4 }.
5. Show that the Borel-Cantelli lemma is valid on any measure space (n, .91, /1), that is,
Do
/1{A n} < 00 implies /1{1lm n An} = 0 for An E .e/, n 2: 1.
6. Let {Xn.i , j 2: I, n 2: I} be a sequence ofr.v.s such that 2:J=' P{IXn) > e} ~O,
all e > O. Prove that sUPh I IXn.jl !. O.
lim In (k -c::np)4b(k; n, p) = 1
r::;::"
foo
t 4 e- t '/2 dt = lim I00 (k _!1).)4 p(k; A).
n-oo k=O v' npq v' 2n - 00 .1-00 k=O v' A
9. (i) For any LV. X and constants an ...... 0, verify that anX ~ O. Hint: It suffices to prove
c.1 XI~ 0, where Cn = supj"nlajl. (ii) If {X n , n 2: I} are r.v.s with identical d.f.s and
{a.} are as in (i), then an X• .f, O.
2.3 Central Limit Theorem for Binomial Random Variables, Large Deviations 45
Then
1 n + I 1 1+ b b; b~
logan = (n + 2)log-- = - I o g - -n = 1 + - + - + ...
n 2b n I - bn 3 5
= 1 + bn (say), where
1 b; ~ b;
< -3 < Un < -3 (I + bn + bn + .,.)
2 4
12n + 1 12(n + 1) + 1
(2)
12n(n + I)'
so that
C()
Therefore,
(n + l)n+O/2l n n
L log a + L bj = n + C - L b
C()
log I = j = n j
n. 1 j;l j;n+t
Set K = e 1 - C
> O. Then, replacing n by n - 1 in (3),
n ~ 2. (4)
e' (k
= __ _ k)-n+k-(1/2 1(npq)-1/2
_ )-k-(1/21(n__
J2it np nq
where e = en - ek - en-ko Since kin -+ p, e = 0(n- 1). Now
k n-k
log{(2nnpq)1/2b(k; n, p)} = e - (k + !)log - - (n - k + !)log--
np nq
I3
(np + x~ +!) [\jrq
2
= e-
x q + 0 (IX
-;p - 2np n 2)] 3/
r,;pq +!)
- (nq - x V r1Plf [ !£ X2p
-x -nq - -2nq + 0 (IX
I3
-n3 / 2)]
I3 2
= r,;pq + x 2q - -x2q
e - [ x v"Plf 2
- + 0 (IX
+ -xj£
2 np
-n 1/2) + 0 (x-n )]
r,;pq
- [ -x v"Plf + x 2 p - -X2p
2
- + 0 - 1 2) + 0 (X- )]
- -x!£
2 nq n / n
(IX I3 2
whence
Note that the second statement in (11) of Corollary 1 asserts that the dJ
of the normalized binomial r.v. S: as defined in (10) tends to the normal dJ.
Such a tendency toward normality is a widespread phenomenon and ex-
emplifies a class of theorems known as central limit theorems (Chapter 9).
Corollary 1. If {Sn, n ;::: I} is a sequence ofbinomial r. v.s with p.dJ.s b(k; n, p),
0< p < 1, and
S* = Sn - np (10)
n ;;;M'
thenfor every pair affinite constants a < b and all x
lim P{a ~ S: ~ b} = <I>(b) - <I>(a), lim P{S: < x} = <I>(x). (11)
"" <I>(M
n
~+ 1) _ <I>(m
n
~-1} (12)
Since the right side of (12) approaches <I>(b) - <I>(a) as n -> 00 via the con-
tinuity of <1>( . ), the first part of (11) follows. As for the second, for
-oo<a<b<oo
lim P{S: ~ b} ;::: lim P{a ~ S: ~ b} = <I>(b) - <I>(a),
yielding as a -> b
lim P{S: < b} ;::: et>(b).
Similarly,
1- IliTi P{S: < a} = lim P{S: ;::: a} ;::: lim P{a ~ S: ~ b} = <I>(b) - <I>(a),
yielding as b -> 00
Il-~
------> -
I fb e -x'j2 dx + -12 h-", j
(21l)t 2
------> - -
K -b b K
yielding K ~ (21l)lj2. Consequently, K = (21l)tj2. o
As will become increasingly apparent, an important characteristic of any
dJ. F(x) is the manner in which it increases to one and decreases to zero-in
other words, the order of magnitude of its tails I - F(x) and F( - x) as
x -+ + 00. In the case of a normal LV., symmetry of the distribution entails
identical behavior of the upper and lower tails
x
-1-- - x 2{2
< foo e - .2j2 du < -1 e - x j2 .
2
2 e
+ x x X
whence
-I e -x
2j2 _
- JOO ( 1 + "21 ) e -.2j2 du> JOO e -.2j2 du > X
-1-- e -x
2j2
. o
x x U x +x 2
1 tf\{) 1 -x 2 j2
as x -+ 00,
- '¥\X '" X(21l)lj2 e ,
50 2 Binomial Random Variables
Theorem 2.lfSn is a sequence ofbinomial LV.S with p.dJ.s b(k; n, p), n ~ 1, and
{an, n ~ 1) is a sequence ofreal numbers with an -+ 00, an = o(n 1/6 ) as n -+ 00,
then setting S:
= (Sn - np)j(npq)1/2 as earlier,
(14)
_ Mn - np _ -1/2
XM n - (npq)1/2 - an + log an + O(n ),
Hence,
P{Sn = k}
-----'--"---'-- =n - k + 1 -p <n - Mn p
-
P{Sn = k - I} k q - Mn + 1 q'
2.3 Central Limit Theorem for Binomial Random Variables. Large Deviations 51
whence
P{S. ~ M.} =
.-Mn
L P{S. = M. + j}:$; j=O
L
.-M n (nM- M i-p)j P{S. = M.}
j=O .+ q
(M. + I)q n
:$; P{S. = M.} :$; P{S. = M.}.
M. +q - np M. - np
Since M. - np = xMJnpq)t /2, by Lemma 2
and
P{Sn ~ mn} = P{m n :$; Sn :$; M n} + P{Sn > M n}
EXERCISES 2.3
1. Show that the Bernoulli weak law of large numbers is consequence of the
DeMoivre-Laplace theorem.
2. Prove that as x
f'
-+ 00.
1 - <I>(x + (a/x)) _
- - - - - - + e a.
I - <I>(x)
hm. P[ a < ~
X;, -.l. < b = l1>(b) - l1>(a) J
;,-"" v.l.
II
by first showing for
z=
k-.l.
-- = O(.l.1/6) that
p(k; .l.) _ < i 81z1 3 C1zl
.ji I.l. 1/2¢J(Z) - .l. + .ji + .ji'
Obtain a large deviation result akin to Theorem 2.
5. Let Sn be binomial LV.S with p.d.f. b(k; n, p), 0 < p < 1, and S: = (Sn - np)jfiM.
Prove that for every rx > 1,
P{IS:I > (2rxlogn)I/2, i.o.} = O.
6. Utilize Exercise 5 to prove that for Sn as defined there and for every p > 1.
. Sn - np
hm - - p - = 0 a.c.
n--+co n
(If P = I, this yields the Borel strong law of large numbers.)
7. Prove for X;, as in Exercise 4 that P {X;, :.,; .l.} > c;" .l. > 0, where c;, = ! for integer
=
.l. and c;, = e- I otherwise. Hint: An,;, Ii=o(.l.iJjl)e-;' = P{X;,:";.l.} for n:";.l. <
n + 1, and A n.n - A n.n+! = J:+
I (.l. ne-;'jn!)d.l. > A +l,n+1 - A . +!, implying A . >
n nn nn
A n+, .n+!. Also A n.n - An,n+1 < A n.n - A n- I • n and by Exercise 4 P{X;, :.,; .l.} -+! as
.l. -+ 00.
8. For q > 0 and k = 0, 1, ... , n, let
N = [nq ++ IJ,
1
h =k- N.
where '1 < 2rx - 1, and that for Ihl :.,; nO,
(ii)
(Hint: For (i) apply Theorem 2, and for (ii) apply Lemma 2.
9. (Renyi) Let Sn be a binomial r.v. with p.d.f. b(k; n, p), where 0 < p < 1. If [np] is
the largest integer:.,; np and In, k n are integers such that In = O(n"), kn = O(n") for
some rx in (0, i), and k; - J;
= o(n), prove that
10. If 0 < a ~ 1 and -00 < a < 00 show that sup",Ict>(xa- 1 - a) - ct>(x)1 ~ (h)-1/2.
[a-I - 1 + lal]. Hint: Add and subtractct>(xa-I),and use the mean value theorem.
References
J. Bernoulli, Ars Conjectandi, Basel, 1713.
S. Bernstein, "Demonstration du theoreme de Weierstrass fondee sur Ie calcul des
probabilites," Soob. Charkov. Mat. Obs. 13 (1912), 1-2.
E. Borel, "Sur les probabilites demombrables et leurs applications arithmetiques,"
Rend. Circ. Mat. Palermo 27 (1909), 247-271.
W. Feller, An Introduction to Probability Theory and Its Applications, Vol. I, 3rd ed.,
Wiley, New York, 1950.
B. de Finetti, "La prevision, ses lois logiques, ses sources subjectives," Annales de
I'lnstitut Henri Poincare 7 (1937), 1-68.
J. Haag, "Sur un probleme general de probabilites et ses diverses applications," Proc.
Inst. Congr. Math., Toronto, 1924, 1928,629-674.
G. H. Hardy, Divergent Series, Clarendon Press, Oxford, 1949.
P. L. Hsu and H. Robbins, "Complete convergence and the law of large numbers,"
Proc. Nat. Acad. Sci. U.S.A. 33 (1947),25-31.
P. S. Laplace, Theorie analytique de probabiliu!s, 1812 (Vol. 7 in Oeuvres completes de
Laplace, Gauthier-Villars, Paris, 1886].
A. de Moivre, The Doctrine of Chances, 1718; 3rd ed., London, 1756.
S. D. Poisson, Recherches sur la probabilite des judgements, Paris, 1837.
A. Renyi, Foundations of Probability, Holden-Day, San Francisco, 1970.
H. Robbins, "A remark on Stirling's formula," Amer. Math. Mont/hy 62 (1955),
26-29.
J. Stirling, M ethodus Differentia/is, London, 1730.
H. Teicher, "An inequality on Poisson probabilities," Ann. Math. Stat. 26 (1955),
147-149.
3
Independence
54
3.1 Independence, Random Allocation of Balls into Cells 55
On the other hand, it is easily verified via (14) of Lemma 2.1.2 that At, lET,
are independent events iff the classes t:§t = {0, n, A" A~} are independent.
Then PI ..... ia.l, ... , an) is called the joint probability density function of
X I> .•. , X n and the latter are termed discrete LV.S. It is not difficult to ascer-
tain that the discrete random variables X I> •.• , X n are independent iff for
every choice of (ai' ... , an) in A x A x ... x A
PI, .... n(al'···' an) = PI(al)'" pian), (3)
where Pj is the (one-dimensional) probability density function of Xj' 1 ~
j ~ n. It follows from (3) that if the discrete random variables X I' ... , X n are
independent then for all real Al> ... , An
and the converse is also true. Even without the proviso of discreteness (4) is
still equivalent to independence, but the proof is more involved and hence
deferred to Section 2. This condition may be rephrased in terms of the joint
and one-dimensional d.f.s, namely, for all real Ai> 1 ~ i ~ n,
n
Unless the contrary is stated, any subsequent relationship among LV.S will be
interpreted in the a.c. sense. Since f(tl> t 2 ) = t 1 + t 2 is a Borel function on
(R 2, 91 2 ) (with the convention that 00 +(- (0) = ( - (0) + 00 = 0), the sum
Xl + X 2 is a LV. by Theorem 1.4.4. By induction, the sum L'i
Xi of n LV.S is a
r.v. for every n = 1,2, ....
. {Sn - np
lIm P ( )1/2 < X
} _
-
1
Fe.
IX e
-u2/2
du, X E ( - 00, (0), (6)
n-oo npq y2n -00
(7)
PROOF. Clearly, {Sn' n ~ 1} are LV.S, and since P{X n = x} = pXq1-X for
x = 0 or 1, by independence for k = 0, 1, ... , n
To say that r balls have been cast at random into n cells will mean that
every possible arrangement of these r balls into the n cells has the same
3.\ Independence, Random Allocation of Balls into Cells 57
According to (3), XI' ... , X, are independent and, taking cognizance of (8),
X j ' 1 ~ j ~ r, are i.i.d. random variables. Thus, a random allocation of r balls
into ncells is tantamount to considering i.i.d. LV.S X 1> .•• , X, with P{X 1 = k}
= lin, k = 1, ... , n. Let
P{AJ = n, P{X
j= 1
j ::1= i} = (1)'
1- -
n
;
Pm(r, n) = j~m(-ly-m
n .
j ( 1 - ~j)r,
m (n) (j) o :5; m :5; n. (9)
Theorem 2 (von Mises). Ifrn balls are cast at random into n cells so that each
arrangement has probability n- rn , where
n 2:: 1, (10)
Pm(rn,n) = ni:m(_lY( n+
)=0 m
.)(m+j)pl::~j'
] m
In Chapter 9 it will be shown that the dJ. ofthe number ofempty cells tends
to the normal distribution when r In ~ IX > 0 and even in certain cases if
rJn -+ 0 or 00.
3.1 Independence. Random Allocation of Balls into Cells 59
°
PROOF. Evidently, P{An} > for infinitely many values of n, and it may be
supposed without loss of generality that P {An} > 0, n ~ 1. Since
1 ~ pLQ An} ~ n~l P{A n)~t Ai} = n~l P{An{l- pt=Ql AjIAn}].
divergence of the series requires
lim inf[l -
n-+oo
p{. 0 AjIAn}] =
]=n+l
0,
EXERCISES 3.1
I. Events An. n ~ I, are independent iff their indicator functions I An' n ~ I, are
independent LV.S iff~n = {0. n, An' A~}. n ~ I. are independent classes.
2. If the classes of events d nand !2 are independent, n ~ I. so are U:,=. d nand !2.
3. (i) Any r.v. X is independent of a degenerate r.v. Y. (ii) Two disjoint events are
independent iff one of them has probability zero. (iii) If P {X = ± 1, Y = ± I} =
for all four pairs of signs, then X and Yare independent r.v.s.
*
4. Let n = {wo , w., w 2, w3}, §' = {A: A en}, P{wJ = to::; i ::; 3, and otherwise
Pisdefined by additivity. If A j = {wo , wJ, I ::; i ::; 3, then each of {A., A 2 }, {A l' A 3 },
{A 2' A 3} is a pair of independent events but the events A I' A 2, A 3 are not inde-
pendent. On the other hand, if B. = {wo}, B 2 = A 2 , B 3 = 0, then (I) obtains for
m = 3 but the events BI> B 2 are not independent.
5. (i) If X 1>"" X n are independent LV.S and 9j, I ::; i ::; n, are finite Borel functions on
( - 00, 00), then 1'; = 9j(Xj), 1 ::; i ::; n are independent LV.S. In particular,
- X., ... , - X n are independent LV.S. (ii) If Bi> I ::; i ::; n, are linear Borel sets and
1'; = XJrx,EB;J' where I A isthe indicator function oftheset A,show that {1';, I ::; i ::;.Il}
are independent LV.S.
8. If X and Yare independent LV.S and X + Y is degenerate, then both X and Yare
degenerate.
9. Let {X., n ~ I} be i.i.d. LV.S with P{X j = Xj} = 0 for i #j. If 1'; = D=I
J1Xj$x,J'
i ~ I, prove that {1';, i ~ I} are independent LV.S with P{ 1'; = }} = Ifi,} = 1,2, ... , i,
and i ~ I.
10. In Bernoulli trials with success probability p, let Y be the waiting time beyond the 1st
trial until the first success occurs, i.e., Y = } iff X j + 1 = I, X j = 0, i :$; }, where
{X J , . . . , Xj} are i.i.d. with P{X j = I} = p = I - P{X j = OJ. Find P{ Y = }},
} = 0, I, ... , which for obvious reasons is called the geometric distribution. If
YI ,· .• , Y,. are i.i.d. LV.S with a geometric distribution, find the p.d.f. of S, = D=
I 1';,
known as the negative binomial distribution (with parameter r). Hint: S, may be
envisaged as the waiting time beyond the rth trial until the rth success.
II. If Y and Z are independent Binomial (resp. Poisson, negative binomial) r.v.s with
parameters (n, p) and (m, p) (resp. A. 1 and A. 2, r. and r2), then Y + Z is a binomial
(resp. Poisson, negative binomial) LV. with parameter (m + n, p) (resp. A. I + A. 2 ,
r J + r2)' Thus, if {X., n ~ I} are i.i.d. Poisson LV.S with parameter A., the sum
S" = I? X j is a Poisson LV. with parameter nA..
12. In the random castingofr balls into ncells, let 1'; = I iftheithcellisemptyand =0
otherwise. For any k < n, show that the p.d.f. of 1';" ... , 1';k depends only on k and
not on i l , ... , ik • Hint: It suffices to consider P{ 1';, = I, ... , 1';k = I} for all k :$; n.
13. If{A.} is a sequence of independent events with P{A.} < I,n ~ I,andP{Uf A.}=I,
then P{A., i.o.} = I.
14. Let Q = [0, I] and .s;1 = Borel subsets of Q, and let P be a probability measure on
.s;1 such that P{[a, b)} = b = a for 0 :$; a :$; b :$; I. Such a "Lebesgue measure"
exists and is unique (Section 6.1). For WE Q, let W = W I W 2 ... be the decimal
expansion of W (for definiteness, no "finite expansion" is permitted). Prove that
{w., n ~ l} are independent LV.S with P{w. =}} = to,} = 0, 1, ... ,9.
15. If Np is a geometric LV. with parameter p, that is, P{ Np = k} = pqk, k ~ 0, prove that
limp~o P{pNp < x} = I - e-X, x > 0, and check that F(x; A.) = (I - e-AX)I[X>OI is
a dJ. for any A. > O. A r.v. with d.f. F(x; A.) is said to have an exponential distribution
with parameter A..
16. If {A., n ~ I} is a sequence of events with L~=I P {A.} = 00, show that
lim sup.~<Xl P{Uj~~AjIA.} = I for all m ~ 1.
3.2 Borel-Cantelli Theorem, Characterization of Independence 61
1 = lim P
n-oo
{Uj=n
Ai} = P{A n, i.o.}. o
p{rrm L ogn
n~oo
Nn
= I} = 1. (1)
p{rrm L ogn
n~oo
Nn
~ I} = 1. (2)
implying L::'=no P{A n} = co. Thus, by the Borel-Cantelli theorem, for a E(O, 1)
P{N kn ~ a Log k n , i.o.} ~ P{N kn ~ [a Log k n ] + 1, i.o.} = I,
yielding
p{rrm L
n-oo
Nn
og n
~ I} ~ p{lim
n-oo
N knk
Log n
~ I} = I,
Lemma 1. If~ and ~ are independent classes ofevents and ~ is a x-class, then ~
and (1(~) are independent.
PROOF. For any B E~, define
D~ = {X'li ) < ..1. 1 , .•. , X,l:,) < ..1.m }, where m is a positive integer, ..1.j a real
number, and tYI E 1;, 1 ~ j ~ m. If £i) I and £i)2 are independent classes, then so
are O"(X t E TI ) and O"(X t E T2 ).
" "
PROOF. Since £i) I and £i) 2 are n-classes and £i) I is independent of £i) 2, Lemma 1
ensures that £i)1 and 0"(£i)2) are independent. A second application of Lemma 1
yields independence of 0"(£i)1) and 0"(£i)2)' Since O"(£i)) = O"(X t E 1;), i = 1,2,
Lemma 2 is proved. "
Theorem 2. Random variables X I' ... , X n on a probability space (n, Y', P) are
independent if and only if
n
F XI ..... Xn = fl F Xi'
j= I
(4)
PROOF. Necessity is trivial since {Xj < Xj} is an event in O"(X j), 1 ~ j~ n. To
prove sufficiency, let £i) I be the class of sets of the type {X n < ..1.n} while £i) 2 is
the class of sets of the form {X I < ..1. 1, ... , X n- I < ..1.n- d. Then £i)1 and £i)2
are n-classes and independent. It follows from Lemma 2 that O"(Xn) and
O"(X h ... , X n - I ) are independent. Consequently, if An E O"(X n) and Ai E O"(X)
C O"(X I , . . . , X n - 1 ), 1 :::;; j < n,
(5)
for all A j E O"(X i ), 1 ~ i ~ n. Thus, the classes O"(X j ), 1 ~ i ~ n, and hence also
the r.v.s Xi' 1 ~ i ~ n, are independent. 0
co co
{An,i.o.} = n U Aj E<I(Xm,Xm+ 1 ,···), allm2:: 1.
n=m j=n
The Kolmogorov zero-one law (below) confines the value of P{ An, i.o.} to
zero or one while the Borel zero-one criterion provides a touchstone. The
zero-one law, however, applies to independent r.v.s other than indicator
functions.
EXERCISES 3.2
1. Find a trivial example of dependent events A" with I P{A"} divergent but
P{A", i.o.} < 1.
2. If X", n :?: I, are independent LV.S, prove that P{lim"~OC' X" = O} = I iff
<Xl
I P{IX"' :?: r.} < 00,
n= 1
all r. > O.
3. Prove that if(i) {X", n:?: I} are i.i.d. N(O, (12) random variables, then
P { r=
urn X" = (J} = 1.
"~<Xl J210g n
If, rather, {X", n :?: I} are i.i.d. exponential LV.S with parameter ;. (Exercise 3.1.15)
then P{limn~<Xl Xn/log n = r J } = 1. (ii) If {X", n ~ I} are i.i.d. Poisson LV.S with
parameter A, then P{limn~oo Xn(log log n)/log n = 1} = 1 irrespective of A. Hint:
Recall Exercise 2.1.4.
4. Show that random vectors X = (X" ... , X m) and Y = (YI , . . . , yo) are independent
(of one another) iff the joint dJ. of X and Y is the product of the dJ.s of X and of Y.
5. If {Xn , n :?: 1} is a sequence of finite-valued Lv.s and Sn = I1
Xi' determine which
of the following are tail events of {Xn , n:?: I}: {Sn converges}; {limSn > IimSn};
{lim Sn = oo}; {X n > C, i.o.}; {limX n = O}; {Sn > Cn' i.o.}.
6. If {Xn, n :?: I} is a sequence of independent finite-valued LV.S and {an, n :?: I} is a
sequence of finite constants with 0 < an --+ 00, show that lim Snlan and lim Snla" are
degenerate.
7. (Barndorff-Nielsen,) If {A", n :?: I} are events satisfying I:= 1 P{A"A~+ I} < 00
and P{A"} = 0(1), then P{A", i.o.} = O.
8. Let {Xn,n:?: I} be i.i.d. LV.S with P{X 1 = I} = P{X I = -I} = 1 and set S" =
Ii= 1 X;, Prove that
i. Iim"_<Xl P{SJ';;; < x} = <l>(x), - YJ < X < x,
ii. P{inf";" S" = -oo} = P{SUPn;'l Sn = oc} = 1.
9. If X", n :?: I, are i.i.d. demonstrate equivalence of the following relations:
10. Verify for i.i.d. r.v.s X n , n ~ I, that p{rrmn~oo X n = oo} = I if and only if XI is
unbounded above, i.e., P{X 1 < C} < I, all C < 00.
66 3 Independence
11. For each n ~ 1, let {X•. j,j ~ l} be a sequence of independent LV.S. Then
sup/X.)!. 0 iffIi'=1 P{IX.) > e} = 0(1), all e > O.
12. If {X., n 2:: I} are independent LV.S with E X. = 0, E X; = 1 which obey a central
limit theorem, i.e., lim._., P{S. 2:: xn 1/2} = I - <1>(x), x E ( - OC, (0), where S. =
2:7 Xi' n ~ 1, prove that fiffi._., S.Jn l / 2 = 00, a.c. Hinr: Utilize the zero-one law.
C = tOI .VI DI
co co co {
IX.+i(w) - X.(w)l <
I}
k.
Moreover, if P{C} = I, there exists a T.v. X on (n,~, P) such that
.-co X. = X} = I.
p{lim (I)
It suffices to define
X(w) = lim X.(w), WEC,
and clearly X is a r.v. In such a situation, the r.v.s X., n ~ I, are said to
converge almost certainly (or almost surely) to the r.v. X, denoted symbolically
by X. ~X. Even if X(w) is not a r.v. it is still possible that lim._ co X.(w) =
lim._ oo X.(w) = X(w) with probability one, likewise denoted by X.~X.
The former will be distinguished from the latter on occasion by writing
X.~ a T.v. X or X. ~X, finite.
that is, iffsuPj~nlXj - XI-f. O. The final condition of the lemma is simply a
restatement in finite terms of the following alternative form of (2):
I:: > O.
o
If X, X n, n ~ 1 are r.v.s with L:'=l P{IXn - XI> I::} < 00, all I:: > 0 then the
Borel-Cantelli lemma ensures that X n ~ X.
Strictly speaking, the latter is not a corollary of the lemma but rather an
immediate consequence of the definition and continuity.
PROOF. If X n ~ X, then for aliI:: > 0, n ~ no(l::) implies P{ IX n - X I > I::} < 1::.
Hence, for m > n ~ no(I::/2)
68 3 Independence
which implies (i). Conversely, if (i) obtains, then for any integer k ~ 1,
P{IX n - Xml > 2- k} < 2- k provided n> m ~ mk' Set n l = m\,ni+1 =
max(nj + 1, mi+I), and X~ = X nk and Ak = {IX~+ 1 - X~I > 2- k}. Then
:D"'= 1 P{A k } < 00, whence the Borel-Cantelli theorem ensures that, apart
from an w-set A of measure zero, IX~ + 1(w) - X~(w) I :$ 2- k provided k ~
some integer ko(w). Hence, for WE AC, as n -+ 00
00 00 1
~~~IX;" - X~I:$ k~nIX~+1 - X~I:$ k~nrk = 2n - 1 = 0(1),
and so
as k -+ 00, X k ~ X.
Next, if X n ~ X, any subsequence of {X n}, say X; -f. X, whence, as already
shown, there exists a subsequence of {X;}, say X~ ~ some LV. Y. Then
X~ -f. Y but also X~ -f. X, necessitating X = Y, a.c. (Exercise 1). Thus, the
subsequence X; has a further subsequence X~ ~ X.
Finally, if X n does not converge in probability to X, there exists an e > 0
and a subsequence Xnkwith P{IX nk - XI > e} > e.Butthennosubsequence
of X nk converges in probability to X (a fortiori almost certainly to X) in
violation of (ii).
Corollary 3. If random variables X n -f. X, then g(X n) -f. g(X) for every
continuous function g.
PROOF. Every subsequence of Yn = g(X n ) has a further subsequence y"k =
g(XnJ with X nk ~ X. By Corollary 2, Ynk = g(XnJ ~ g(X), whence, by
Lemma 2, g(X n) = Yn ~ g(X).
Conversely, (iii) entails sUPm>n P{IX m - Xnl > t:} = 0(1), all e > 0, and so
by Lemma 2 there exists a LV. X with X n E. X. Thus, for all e > 0
P{~~~IXm - e} ~ P{~~~IXm - n
n
XI > Xnl >
Suppose that for some constants 0 < bn i 00, bn+ Ilb n -+ 1, i.i.d. r.v.s
{X n , n 2:: I} and partial sums Sn = L:~ Xi, n 2:: 1,
Sn ---+
-
•.c. S finite. (4)
bn '
(5)
This is a restriction on the dJ. of X I (see Corollary 4.1.3) and thereby pro-
vides a necessary condition for (4). On the other hand, if (4) is replaced by
(7)
70 3 Independence
the dJ. of X 1 should likewise be constrained, but the simple subtraction of (5)
leads nowhere. However, (6) is still necessary for (7) if the r.v.s {X n' n ~ I}
are i.i.d. as follows from the second of the ensuing lemmas.
~ n~1P{BnAn~O>j} = J1 P{Bn}P{An:O~Aj}
~ (~~~ P{Bn}) pL91 An},
and in case (ii)
= ±P{Bj}P{Aj nA~} ~
j=1 j+1
inf P{BJP{U Aj},
Isjsn I
p{nm (Zn + y") ~ e} ~ p{lim Zn > e+ b}. lim P{Yn ~ -b}. (9)
n-oo "-00 "-00
3.3 Convergence in Probability 71
By hypothesis and Lemma 1 the left side of (10) is 0(1) as m -+ 00, and
moreover,
P{Bk} ::;; P{IS2"(k,j > 2n(k)-l e} + P{lSkl > 2n(k)-l e} = 0(1)
as k -+ 00. Consequently, P{Ui... Ad = 0(1) as m -+ 00, and so SJn ~ O.
o
Although, in general, a.c. convergence is much stronger than convergence
in probability, in the special case of sums Sn of independent random variables
the two are equivalent. A basic tool in demonstrating this equivalence for
sums of independent random variables is an inequality due to P. Levy. This
in turn necessitates the
for all e > 0, and so letting e -+ 0, P{X < a} ::; ! or equivalently P{X ~ a}
~ !. Thus, a is a median of X.
A pertinent observation concerning medians is that if for some constant c
P{IXI ~ c} < e::;!, then Im(X)1 ::; c. Moreover, if c is any finite constant,
cm(X) and m(X) + c are medians of cX and X + c respectively.
PROOF. Set So = 0 and define T to be the smallest integer j in [1, n] for which
Sj - m(Sj - Sn) ~ e (if such an integer exists) and T = n + 1 otherwise. If
B j = {m(Sj - Sn) ~ Sj - Sn}, 1 ::; j ::; n,
then P{B j };:::!. Since {w: T = j} E a(X I, ... , X), B j E a(X j + 1>'''' X n),
and {Sn ;::: e} ::> Uj= I Bj{T = j},
n n
It is easy to verify that X is symmetric iff P{X < x} = P{X > -x} for
every real x and also that zero is a median of a symmetric r.v. It follows from
Corollary 3.2.3 that sums of independent symmetric r.v.s are themselves
symmetric r.v.s. This leads directly to
(13)
(14)
PROOF. It suffices to verify sufficiency. By Lemma 2, for any e in (0, ;}), there
exists an integer ho such that n > h ~ ho implies, setting Sh,. = S. - Sh'
that P{lSh,.1 > e} < e. In view of an earlier comment, this entails Im(Sh,.)1
:5; e for n > h ~ ho . By Levy's inequality (12), for k > h ~ ho
EXERCISES 3.3
I. i. If X • .!. X and X • .!. Y,thenP{X = Y} = I.
ii. X • .!. X and Y• .!. Y imply X. + Y• .!. X + Y.
iii. X. ~ 0 implies m(X.) --+ O.
f
P lim X. ~ X ~ fiffi
0 X.} = 1.
) " ..... x "-00
Conversely, if fiffi.~ 00 X. = X 0 (resp.li.m..~ 00 X. = X 0), a.c., then for any € > 0
PiX. > X o + €} = 0(1) (resp. PiX. < X o - €} = 0(1»).
6. If {X., n ~ I} are independent, symmetric LV.S such that (lIb.) Ii Xi.!. 0 for some
positive constants b., then (llb.)max l ";,,. Xi!. O. Hint:
7. If the LV.S X.lb• ..!'. 0, where the constants b. satisfy 0 < b.l 00, then
max Im(X j - X.)I = o(b.).
1 SjS"
8. Prove for any LV.S {X.} and constants {b.} with 0 < b.l 00 that (i) (lIb.) I1 Xi ±
o implies X.lb. 2...
a.c.
0, (ii) if for identically
p
distributed {X.}, some nonzero
constant c and 0 < b. --+ 00, (lib.) L~ Xi --+ c, then b. - b.-I'
10. If r.v.s X. ~ X as n --+ 00 and {N., n ~ I} are positive-integer valued L v.s with
(i) N. ~ Xl, then X N " ~ X. If, rather, (ii) N• .!. Xl, that is, PiN. < C} = 0(1) all
C > 0 and X is a LV., then X N ".!. X.
II. If the w" on (O,:J', P) is .~.-measurable, n ~ 1, where .¥. is a decreasing
n:=
LV.
14. Let {X., n ~ I} be LV.S such that P{IX.I ~ c > O} ~ b > 0, n ~ I. If {a.} are
finite constants for which a. X. !. 0, then a. --+ O.
3.4 Bernoulli Trials 75
15. IfLv.s X. ~ X, finite, prove thatforevery t: > Othere isaset A, with PIA,} < t:such
that lim X.(w) = X(w) uniformly on A~. This is known as Egorov's theorem.
(Hint: Verify that if A•. k = n1=' {IX j - XI < 2- k } and A = {tim X. = X}, then
lim._., P{A•. d = P{tim._., A•. d ~ PIA} = I, whence for some integers nk
P{ A~ •. k} < t:/2 k. Take A, = Uk'= I A~ •. k)' State and prove the converse.
16. If {X., n ~ I} are independent LV.S, Sm.• = Li=m+ 1 X j , S. = SO.• , then for any
f.> 0
.
This is Ottaviani's inequality. Hint: If T = inf{j
U
~ I: ISjl > 2t:}. then
17. If {X, X., n ~ 1} are i.i.d. symmetric r.v.s and S. = L~ Xi' then (i)
n
P{S. > x} ~ ZP{X > 2X}p·-1 {X ~ 2x}
for x> 0, and (ii) P{S. > x} ~ (n/2)P{X > x} [1 - (n - 1)P{X > x}]. Part (i)
is untrue if aU "twos" are deleted (Take n = 2, x = !, and P{ X = ± I} = t). Hint:
Apropos of (ii), define T = inf{j ~ I: Xi> x}.
18. Let S. = I~ Xi where {X., n ~ I} are independent LV.S and suppose that
lim._., P{S._I ~ -bn} > 0, all {j> O. Then rrm
Sjn ~ C < <x, a.c. implies
I:,=I PIX. > t:n} < x., all t: > C.
p{ max Sj
1 SJsn
= N} = P{Sn = N} + P{Sn = N + I}
=r
n
([(n + : + 1)/2]). (3)
p{ maxSj ~ N,Sn < N} =P{T ~ n,Sn < N} = P{T < n,Sn < N}
Is,JS,n
= nfp{T = k}P{
k=1
i
i=k+1
Xi <o}
= nil P{T = k}P{ i Xi> o}
k=1 i=k+1
n-I
= L P{T =
k=1
k, Sn > N}
p{ m~xSj = N} = p{ ISjsn
ISjSn
m~xSj ~ N} - p{ m~xSj ~ N + I}
ISjsn
= 2 P{Sn ~ N} - 2 P{Sn ~ N + I} - P{Sn = N}
+P{Sn=N+l}
= P{Sn = N} + P{Sn = N + I}.
In proving the second equality of (3), if n + N = 2m for some m = 0, 1, ... ,
= p{ max Sj
ISjsn-1
~ oJ,
which is tantamount to the first equality of (4). To obtain the second, note
that via (2)
78 3 Independence
p{ max Sj:::;
ISjsn
o} = 1 - p{ ISjsn
max Sj ~ I} = 1 - 2 P{Sn ~ I} + P{Sn = I}
P{S2m = O} = C:)2- 2m
, P{S2m+1 = I} = Cmm+ 1)2- 2m - l
.
= P{ max Sj :::; O}
I sjSn- I
- P { max Sj :::; O}
I Sjsn
Next, a local limit theorem (7) and global limit (8) will be obtained for
- = maxI sjsn Sj' Accordmg
Sn . to (8), the dJ. of n -1/2-Sn tends as n - 00 to the
positive normal dJ. The name arises from the fact that if X is N(O, I) then IX I
has the dJ. 2C1l(x) - I, x > O.
(7)
x> 0, (8)
and since
p{lim S.
n-oo
= oo} = 1 = p{lim S. = -oo},
n-C('I
(11)
=t P{~~f J2 x ~ j - I} + t P{~~~ J2 X~ j + I}
i i
= tP{sUPS n ~j -
n~O
I} + tP{SUPSn
n~l
~j + I} = ¥qj-l + qj+l).
p{SUP Sn =
n~l
oo} = p{rrm Sn = oo} = l.
n-oo
They Yo = 1, Yr = 0, and (14) obtains but with p "# t. Hence, for 0 < i < r
P(Yi - Yi-I) = q(Yi+ 1 - Yi) or Yi+ I - Yi = S(Yi - Yi-l)·
Thus, for 0 < i < r
Yi+l - Yi = S2(Yi_1 - Yi-2) = ... = Si(Yl - Yo), (16)
3.4 Bernoulli Trials 81
If it were known that the left side of (15) was a continuous function of p,
then (15) would imply (13).
if s = r. i: 1. (18)
q
Interchanging p with q and j with k,
EXAMPLE 2 (S. Samuels). Let Sn = IJ:1 X j where {Xj , j ~ I} are Bernoulli trials
with parameter pE(O, l)and define T = inf{n ~ 1: ISnl = r}, r > oand T = 00
otherwise. If V = r + ST where ST = I::':1 Sn' lIT:n), then V and T are inde-
pendent random variables.
PROOF. Note that according to Example 1, P{T < a)} = 1. Let C(n, r) =
number of n-tuplets (X l ' ... , X n ) with Sn = r, ISjl < r, 1 ~ j < n. Then
P{T = n, V = 2r} = P{T = n, Sn = r} = C(n, r)p(n+r l/2 q(n-r l/2, n ~ r > O.
and by symmetry
P{T = n, V = O} = P{T = n, Sn = -r} = C(n, r)p(n-r l/2 q(n+r l/2.
Hence,
P{T= n} = [1 + (p/q)'J'P{T= n, V = O}, n~r > 0 (20)
and so
L P{T = n, U = O} = [1 + (pjq)'r
00
P{U = O} = 1
.
n=r
EXERCISES 3.4
I. Let Sn = D Xi> n ~ I, where {X n' n ~ I} are i.i.d. with P{X 1 = I} = P{X 1 = -I}
= 1. If Ii = inf {n ~ I: Sn = i}, i = I, 2, then
a. P{T1 < Xl} = I, I:,=
1 P{T1 > n} = 00,
b. T1 and T2 - TI are i.i.d.
References
O. Barndorff-Nielsen, "On the rate of growth of the partial maxima of a sequence of
independent, identically distributed random variables," Math. Scand. 9 (1961),
383-394.
L. Baum, M. Katz, and H. Stratton, "Strong Laws for Ruled Sums," Ann. Math. Stat.
42 (197\), 625-629.
F. Cantelli, "Su due applicazioni di un teorema di G. Boole," Rend. Accad. Naz.
Lincei 26 (1917).
K. L. Chung,"The strong law of large numbers," Proc. 2nd Berkeley Symp. Stat. and
Prob. (1951), 341-352.
K. L. Chung, Elementary Probability Theory with Stochastic Processes, Springer-Verlag,
Berlin, New York, 1974.
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
W. Feller, An Introduction to Probability Theory and Its Applications, Vol. I, 3rd ed.,
Wiley, New York, 1950.
A. Kolmogorov. Foundatiolls of Probability, (Nathan Morrison, translator), Chelsea,
New York, 1950.
r
P. Levy, Theorie de addition des variables aleatoires, Gauthier-Villars, Paris, 1937;
2nd ed., 1954.
M. Loeve, Probability Theory, 3rd ed. Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
R. VOn Mises," Uber aufteilungs und Besetzungs Wahrscheinlichkeiten," Revue de la
Faculte des Sciences de r Universite d'lstanbul, N.S. 4 (1939), 145-163.
A. Renyi, Foundations of Probability, Holden-Day, San Francisco, 1970.
4
Integration in a Probability Space
E X = hm
•
.L
00 i {i i+ I}
2" P 2ft < X ~ - 2
ft • (I)
n-oo 1=1
(2)
84
4.1 Definition, Properties of the Integral, Monotone Convergence Theorem 85
It is not difficult to see that the limit in (1) always exists, since setting
i i + I} <Xl i
Pn,i = P { 2n < X ~ -2" ' Sn = L fnPn,i' (3)
i= 1
additivity of P guarantees Pn,; = Pn+ I, 2; + Pn+ I, 2;+ I' whence
<Xl 2i
Sn = i~1 2n+1 (Pn+I,2i + Pn+I,2;+I) ~ Sn+I'
and so limn~ <Xl Sn exists, Furthermore, E X ~ Sn' n ~ 1, for X ~ O.
It is trivial to check that
E[I] = 1, 0 ~ E X ~ 00 when X ~ 0,
X ac. 0 ifEIXI = 0, P{IXI < oo} = 1 if EIXI < 00, (4)
EX = E YifP{X = y} = 1, E[-X] = -EX iflEXI ~ 00,
. L
EX = lIm i +-1 P {i
<Xl - i +-I}
-n< X < -
n~<Xl i=O 2
n 2 - 2n
-_ I'1m ~
L.
n~<Xl i=O
i
-n-
+
2
1 p{n 2
i ~
in
X < - + -I}
2
lim[
n~<Xl;=12
L<Xl'-'n P{'2~ < .+
X < -'-n- + L '-n P X
2
I} <Xl' 1 {
;=12
= ~. }]
2
(6')
86 4 Integration in a Probability Space
(7)
PROOF. (i) If P{X = oo} > 0, then X n = 00 for some n for which P{A n } > 0
so that (7) obtains. If, rather P{X = oo} = 0, then X m = 00 for some m
requires that P{A m } = 0, whence X m P{A m } = O. By setting x~ = 0 on any
An C {X = oo} and x~ = X n otherwise, and defining X' == X· lrx<oo) =
Lf x~IAn' it is evident that P{X' = X} = 1, EX' = E X, and X'(w) < 00
for all w. Let
J.x j
ooi 00 00
L xj P{A j} = L xjP{Aj}.
OC! OC!
EX = E X' =
1 1
Then X~l) i X, X~2) i X', and X~) ~ x~t I> j = 1,2; furthermore, via (i)
E X~2) = Sn ~ Sn+ 1 = E X~221' Consequently, if P{X = oo} = 0, E X~2) i E X
according to the definition (1). Moreover, under the same proviso P{X = oo}
= 0, subadditivity guarantees
X = L XJA;Bj' Y=LYjIAiBj'
i.i i.j
For the general case it may be supposed that P{X = oo} = 0, whence also
P{ Y = oo} = O. If X n = X~2) as in (14) and Yn is defined analogously via Y,
then these elementary functions satisfy P{X n 2 Yn 2 O} = 1, so that by the
part already proved E X n 2 E Yn • Hence, by (ii), E X 2 E Y.
Consequently, the Markov inequality follows directly from X 2 a1lx,,- al
and (i).
To prove the monotone convergence theorem, note first in case
P{ X = oo} > 0 that by the Markov inequality and Corollary 1.5.2
p{~
2"
< X <~}
m-2"
= p{xm >~}2" - p{xm>~2"_~}
2"
"~p
k
i~' 2"
. {'
2"
• +
~ <X<~---
- 2"
I} >a'
whence, via (15), for all large m
E Xm 2 J, 2n p
k i { i
2" < i
X m ~ ~ > a.
+ I} (16)
E [X + Y] =L (Xi + Yj)P{AiB j}
i,j
Thus, from
X+ + y+ = (X + Y)+ + [X- + Y- - (X + Y)-],
by the part already proved
E X+ +E y+ = E[X+ + Y+] = E(X + Y)+ + E[X- + Y- -(X + Y)-],
yielding again via the additivity or subtractivity already proved
EX + E Y = E(X + Y).
To dispatch (vi) in the nontrivial case EX < 00, E Y > - 00, note that
via (v) and (iii)
E X = E Y + E[X - Y] ~ E y.
Finally, apropos of (vii), let 0 < a < 00 and X ~ O. If P{X = oo} > 0,
then E[aX] = 00 = a E X. If P{X = oo} = 0, then, recalling (ii), X~2) =
I~ t CfJni i X, a.c., whence 0 ~ aX~2) i aX, a.c., and by (i) and (iv)
1
P{IXI ~ a} ::;; - E lXI, a>O (Markov inequality) (19)
a
and
IEXI ~ EIXI. (20)
If X is a discrete r.v. with values {xi' 1 ~ j ~ n} and p.dJ. {Pi' 1 ~j ~ n},
where 1 ~ n ~ 00, Theorem l(i) ensures that
n
EX = LXiPi (21)
i= 1
when xi ~ O. Consequently, (21) holds for any discrete T.v. provided that
I.i= 1 xt Pi or I.i= 1 xjpi is finite when n = 00. On the other hand, if X is an
90 4 Integration in a Probability Space
EI X. = I EX•. (23)
.=1 .=1
Without the nonnegativity proviso, (23) is, in general, false even when
Lr;.1 Xi converges, a.c.
L P{X ~ b
1
n} = L P{cp(X) ~ n} = L L P{j ~ cp(X) < j
1 n=1j=n
+ 1}
co
= L j P{j ~ cp(X) < j + I} = E Y
j= 1
and
co co co
Lo P{X > b n} = L L Pfj < cp(X) ~j + I}
n=O j=n
co
= L(J+ I)P{j < cp(X) ~j + I} = E Z,
j=O
Clearly, (25) follows from (24) with b(x) = x', r > 0. Finally, it suffices to
verify (26) when r = 1 since the statement for r > then follows by a change °
of variable. Now, for any a > 0, (25) ensures
and, similarly,
j=1
f J
Aj
X- dP = r
JU 1"Aj
X- dP.
= r
JUAj
X+ dP - r
JUAj
X- dP = r
J U 1"Aj
X dP. D
lim sup
a-tO n~ t
r
JIIXnl >a]
IXnl dP = O. (5)
PROOF. If {X n } is u.i., then sup EIXnl ~ C < 00, whence for any € > 0,
choosing E> as in (3), the Markov inequality ensures that
C
P{ IX n I > a} ~ a - tEl X n I ~ - < E>, n ~ 1,
a
provided a > C/E>. Consequently, from (3) for a > C/b,
sup
n~ 1
r
JIIxnl>a]
IXnl dP < €, (6)
E IX n I ~ a + r
JIIXnl >a)
IX n I dP ~ a + €, n ~ 1,
yielding (4). Moreover, selecting E> = €/a, for any A with P{A} < E> and all
n~ 1
f IXnl
A
dP = f
AIIXn!,;a)
IXnl dP + f AIIXnl>a)
IXnl dP
~ a P {A} + r
JIIXnl >a)
IXnI dP ~ € + € = 2€,
Associated with any probability space (O,~, P) are the ;R" spaces of
all measurable functions X (necessarily LV.S) for which E I X IP < 00, denoted
by ffl P or ffl p(O, ~, P), P > O. Random variables in ffl P will be dubbed
!E" r.v.s.
The inequalities of Section 1 show that a LV. X E ffl P iff the series
Lr' P{IXI > n l/P } converges.
PROOF. For a > 0, b > 0, (a + by ~ [2 max(a, b)]P ~ 2P[a P + bP] and (7)
follows.
In particular, X E ffl P' Y E ffl P imply X + Y E ffl P' 0
Among the most important and frequently used results of integration are
Lebesgue's dominated convergence theorem (Corollary 3) and monotone
convergence theorem, and Fatou's lemma. The second will be obtained by first
verifying the latteL
Hence, setting y" = infi~. X;, it follows from - K =:;; Y" i fu!!...-. 00 X~ and (10)
that
E lim X~ = lim E Y" =:;; lim E X~. (11)
~ E lim X~ - 8 ~ E lim X. - E,
and (8) obtains as 8 -+ O. 0
The next lemma, especially when A. = B., may be regarded as an extension
of the Borel-Cantelli theorem which relaxes the independence requirement
therem.
(12)
then
(14)
Moreover, if(12) holdsfor infinitely many integers k and (13) obtains, then
P{B., i.a.} = 1. (15)
p{O B ~ p{.O
J=k
i}
J=l
Ai+(II+ilkIAi+lIk} = p{ 0 h=.+l
AtI A :}
4.2 Indefinite Integrals. Uniform Integrability. Mean Convergence 9'/
for all large n, whence (14) follows from Theorem 3.1.3, noting that
k 00 00
L L P{A + = 00 = n=1
;=1 n=1
L P{A:.;}
j nk }
L P{ISnl :::;; e}
n=1
diverges or converges.
EXAMPLE 2 (Polya). Let {Sn, n ;::: O} be a simple random walk on (- 00, oor
whose initial position is at the origin, that is, Sn = D
Xj, where {X n , n ;::: l}
are i.i.d. random vectors with
m
and Ie;
i;;]
= 1.
P{A n} = 4- 2n ± .,
j=O
(2n)!. ,
[j.(n - }).]
2= (2n)4-
n
2n ±(~)2 = [(2n)2_2n]2 "'_,
j=O } n 1W
and so If P{A n} = 00. Moreover, AjA j = AjCjj for i < j where Cij =
{S2j - S2j = (O,O)}, implying
Lemma 6. Let {X n, n ;::: I} be independent, symmetric LV.S and {an' n ;::: I},
{en, n ;::: I} sequences of positive numbers with an ...... 00. If X~ = XnIUX"1 $c"I'
S~ = D= I Xj, Sn = D=
I Xj' then
then
4.2 Indefinite Integrals. Uniform Integrability. Mean Convergence 99
00
= L P{Sn ~ S~, N = n} ~ t.
n::;:m
m
implying
and hence also the conclusion of (17) by the Kolmogorov zero-one law.
To verify (18), set Xj = XjIlIXjl:scj) - XjI IIXj ! >'j] and note that by
symmetry and independence the joint distributions of (X t, ... , X n) and
(X!, ... , X:) are identical for all n. Hence, if n > m,
P{Sn ~ S~, N m = n}
= p{~ XjIUXjl>cj] ~ 0, ~ XjIlIXjl:scj) > an,S;:::;; ai' m:::;; i < n}
The next result reveals the extent to which mean convergence is stronger
than convergence in probability and provides a Cauchy criterion for the
former.
°
k-oo k-(JJ m~l
since {I X niP} is u.i. Again employing the latter, for any t: > 0, () > may be
chosen such that P{A} < () implies
sup E{IXnIPI A ] < t:, (19)
.. ~1
100 4 Integration in a Probability Space
PROOF.p The hypothesis implies that X • .!. X and Corollary 3.3.3 ensures that
IX.IP -+ IXI P, whence EIX.,p -+ EIXIP by Corollary 4. 0
In the proof of Theorem 2.2.2 it was shown for binomial r.v.s S. that
EIS. - np!4 = 0(n 2) and so EI(S.ln) - pl4 = 0(1). Thus, it follows directly
from Corollary 5 that E(S.ln)4 -+ p4. More generally, since S.ln ~ p and
IS.lnl :::;; 1 Corollaries 3 and 5 ensure that EIS.lnl/i ~ p/i for every P > O.
EXERCISES 4.2
I. Improve the inequality of Lemma 3 by showing that
P{lX11 > A.} S; 2 P{IXd > A., IXll <~} ~ 2 P{IX 1 + Xli >~}
5. Let PIX. = a. > O} = I/n = I - PIX. = O}, n ~ I. Is {X.,n ~ I} uj. if
i. a. = o(n),
ii. a. = en > O?
102 4 Integration in a Probability Space
6. If {X., n :2: I} are LV.S with sup. 2. E I X. IP < IX) for some P > 0, then {I X. I', n :2: I}
is u.i. for 0 < IX < p.
7. If the r.v.s X., n :2: I, are u.i., so are S./n, n :2: 1, where S. = I7= I Xi; in particular,
if X., n:2: 1, are identically distributed !f'. random variables, then {S./n, n :2: I}
is u.i.
8. Show that (i) if the Poisson r.v.s S. have p.d.f. p(k; nil), n :2: 1, then EI(S./n) - ill-+ O.
Hint: Recall Exercise 2.2.2. (ii) Any sequence ofr.v.s y".!. 0 iff E(I y"1/(1 + Iy"1)) =
0(1).
9. (i) Construct a sequence of LV.S that is u.i. from below and for which the expectation
of lim X. does not exist. (ii) Show that the hypothesis sup.2.1 X. IE!f' I of Corollary
3 is equivalent to IX.I $ Y E !f'., n :2: 1.
10. Let {X., n :2: I} be a simple symmetric random walk in R\ that is, {X.} are i.i.d.
random vectors such that P{X. = (e., ... , ek )} = 1/2k, where ej = 0, 1, or -I and
D eJ = 1. Prove that
II. If {X. X., n :2: I} are r. v.s on (n, ff, P), show that the indefinite integrals I' A X. dP-+
J
A X dP, finite, uniformly for all A E ff iff X. converges to X in mean.'
12. If the two sequences of integrable LV.S {X.}, {Y.} satisfy P{X.:2: Y.:2: O} = 1,
X. ~ X, Y. ~ Y, and EX. -+ E X, finite, then EI Y" - YI-+ O.
13. Let {X., n :2: I} be a sequence of !f'p LV.S, P > 0 with SUPUA IX.I P dP: n :2: 1 and
P{A} < 0] = o(l)aso -+ O. Then X. ~ X iff X• .!!4 X.
14. If X E!f'. (n, ff, P) and u(X), t§ are independent classes of events, prove that E XI A
= EX· P{A}, all A E t§.
15. (Kochen-Stone) If {Z., n :2: I} is a sequence of r.v.s with 0 < E Z; < IX), E Z. #- 0,
and ITm._oo(E Zn)2/E Z; > 0, then P{ITmn_ oo Zn/E Zn:2: I} > O. Hint: If Yn =
Zn/E Zn, there is a subsequence {n'} with E Y;' < K < IX). Replacing {n'} by {n} for
notational simplicity, E lim Y; $ K by Corollary 2. Since - Yn $ I + Y;, neces-
sarily E Iim( - Yn) exists. Then Theorem 2(ii) ensures (since {y"} is u.i.) that E r.m Y.
:2:r.mEY.=1.
16. (Kochen-Stone) If {An' n :2: I} is a sequence of events such that for some c > 0
i. P{AA} $ c P{A;} [P{A j _;} + P{Aj }], i < j
ii. I:'=I P{A n } = IX),
then P {A., i.o.} > O.
Hint: If Z. = L7=.I Aj , note that E Z; $ E Z. + 4c(I'i P{ A i })2 ~ (1 + 4c)(E Z.)2
for all large n since E Zn -+ IX) and, via Exercise 15,
P{A n, i.o.} :2: p{rrm ZJE Z. ~ I} > O.
(1)
Geometrically speaking, the value of a convex function at any point on the
line segment joining XI to X2 lies on or below the line segment joining g(x l )
and g(X2)' Since t = u«t - s)/(u - s» + s«u - t)/(u - s»,
t-s u-t
g(t) ~ - - g(u) + - - g(s), s < t < u, (2)
u-s u-s
or equivalently
g(t) - g(s) g(u) - g(t)
----~ , s < t < u. (3)
t-s u-t
If g is convex on an open interval J 0' it follows from (2) that lims<t_s g(t) ~
g(s), limr>s_t g(s) ~ g(t), limt<u_t g(u) ~ g(t), and limu>t_u g(t) ~ g(u),
whence g is continuous on J o. Furthermore, as a consequence of (3), a
differentiable function g is convex on J 0 iff g' is nondecreasing on J o. Thus,
if g is convex on J 0 and twice differentiable, g" ~ O. Conversely, if g" ~ 0 on
J o , a two-term Taylor expansion yields, setting ql = A., q2 = 1 - A.,
qjg(X j) ~ qj[g(qlx l + q2X2) + q3-j(Xj - X3_j)g'(q I XI + q2X2)], i = 1,2,
and summing, (I) holds, that is, g is convex on J o . Moreover, it is shown in
Hardy et al. (1934, pp. 91-95) that
i. If g is convex on an open interval J 0' it has left and right derivatives g; and
g~ at every point of J 0' with g; ~ g~, each derivative being nondecreasing,
n. If g is convex on an interval J, at each interior point ~ e J,
teJ. (4)
PROOF. By (4), g(t) ~ g(O) + tg~(O) for t e (- 00, (0), whence (5) holds. Since
monotonicity ensures t+ + cx(O) ~ cx(t+) ~ cx(t) ~ cx( - C) ~ - C + cx(O), the
hypothesis implies Icx(X)1 ~ IXI + Icx(O) I and so cx(X) is integrable. Con-
sequently, (4) yields E g-(cx(X) - E cx(X) + c) < 00. Set
fJ(t) = t - cx(t) - E X + E cx(X), t e (- 00, (0).
Then ElfJ(X)1 < 00 and E fJ(X) = O. If P{fJ(X) = O} = 1, (6) holds trivially.
Otherwise, fJ(t l ) < 0, fJ(t 2) > 0 for some tl> t 2 e( -00, (0). If to =
inf{t: (J(t) > O}, then t l ~ to ~ t 2 by monotonicity of t - cx(t), and
t ~ to if (J(t) > 0, t ~ to if fJ(t) < O. (7)
104 4 Integration in a Probability Space
Corollary I. If9 is a convex function on ( - 00, 00 ),for any !i't r.v. X and any
finite constant c
E g(X - E X + c) ~ g(c) (10)
and, in particular,
E g(X) ~ g(E X) (Jensen's inequality). (11)
(16)
4.3 Jensen, Holder. Schwarz Inequalities 105
PROOF. In proving (16), it may be supposed that 0 < IIxlip II YII p' < 00 since
(16) is trivial otherwise. Set
IXI
U = TIXII~'
entailing IIUll p = 1 = 1IVII p " Now, -log t is a convex function on (0, (0),
whence, via (1), for a, b > 0
p P
- log ( -a + -b ')
~ - -1 log a p
- -Ilog
pb' = -log ab,
P P' P p'
or equivalently
o ~ a, b ~ 00.
Thus,
1 1, 1 1
E UV < - E UP
- P
+ -p' E VP = -P + -p' = 1
,
(17)
PROOF, For 0 ~ Pt, P2 < 00 and qt, q2 > 0 with ql + q2 = 1, noting that
l/qj > 1, i = 1,2, Holder's inequality yields
(19)
106 4 Integration in a Probability Space
while its positive square root is the standard deviation of X. Clearly, for every
finite constant c, a 2 (X + c) = a 2 (X) and a 2 (cX) = c2 a 2 (X). The variance or
standard deviation of X provides information about the extent to which the
distribution of X clusters about its mean and this is reflected in the simple but
extremely useful Tchebychev inequality
a> 0, (20)
which follows from a direct application of the Markov inequality to the LV.
(X - E X)2.
A basic tool of probability theory is truncation. Two alternative methods
of truncating a LV. X are
(i)
(ii)
where a, c are constants such that - 00 ~ a < c ~ 00. One or both equality
signs in the set of the indicator function of (i) may be deleted. Whenever both a
and c are finite, Yand Y' are bounded r.v.s and hence have moments of all
orders. For X E ff l' Corollary 2 reveals that a~ ~ ai, whereas no comparable
inequality between a~. and ai exists (Exercise I).
If X, Yare LV.S with 0 < a(X), a(Y) < 00, the correlation coefficient
between X and Y or simply the correlation of X and Y is given by
E(X - EX)(Y - E Y)
Px. y = p(X, Y) = a(X). a(Y) . (21 )
E XY = EX· E Y. (22)
00
00
= P{mj < Y S m j + t} . E X.
By the monotone convergence theorem
00 00
00
0'2 (.±
)=1
Xj) = E (.± )=1
X j)2 = .± E Xf + .I.E
)=1 '''')
XjX j = .±
)=1
(12(X j). 0
and so according to Exercise 4.2.6 {I Sn/n la, n ~ I} is u.i. for < IX < p. If, °
moreover, {I X nIII, n ~ I} is u.i., so is {ISn/nl ll , n ~ I}, generalizing Exercise
4.2.7.
PROOF. Since the case p = 1 is obvious (Exercise 4.2.7), suppose that p > 1.
By Holder's inequality, ISnlnl1l ~ ~ I7= 1 IXl, whence uniform integrability
~M~~ 0
EXAMPLE 2. Let Sn = Ii=1 Xj' n ~ 1 where {Xn , n ~ I} are independent r.v.s
with EX. = 0, E X n2 = 1. If {Xn2 , n ~ I} is u.i., then {S;/n, n ~ I} is u.i.
PROOF. Define y" = XnIlIXnl~KI - E XnIlIXnl~KI' K > 0 and Z. = X. - y",
n ~ 1. Then {y", n ~ I} are independent r.v.s with E y" = and E Y,,2 ~ 1. °
The same statement likewise holds for {Z., n ~ I}. Hence, if T" = Ii=l lj,
J¥" = Ii=l Zj'
~ :2 [nK 4
+ 2 (;) ] ~ K 4+ 1
and so {(T,,/n l/2 )2, n ~ I} is u.i. by Exercise 4.2.6. Hence, for any I: > 0, there
is a () > 0 such that P {A} < () entails
E ( S2)
: IA ~ 2 E (T.•2 +n W 2 I ) < 1:.
n A
PROOF. Let r = min(l, p/2). If p < 2, then r < 1 < p and so by Exercise 4.3.8
(P-1l/(p-rl E y.
( lim E
"-CO
Y: ) ~ lim (E p)(l--"rl/(p-rl = 1
n-oo n
implying
(24)
EXERCISFS 4.3
1. (i) Show for any fill r.v. X that ai = E X 2 - (E X)2. (ii) If P{X = I} = pe(O, I)
andP{X = O} = P{X = 2} = (1- p)/2,then,settingZ = XlIX>I)' W = XlIXSII'
necessarily (1ij. < (1i < (1~.
2. Calculate the mean and variance of U., the number of empty cells in the random
assignment of r balls into n cells.
3. Prove (i) for any r.v.s {S., n ~ I} with (12(S.) = o(n 2 ) that n -1 (S. - E S.) .!. 0,
(ii) for any r.v. X and positive numbers a, t that necessarily P{X ~ a} ~ e- al E elx
(iii) ifm is a median of X e fill' then Im-E XI~EIX -E XI and hence Im:'E XI
~ (1 for X e fil2 with variance (12.
(iv) for fill r.v.S. X and Y with E XY finite, their cOl'ariance, denoted Cov[X, Y],
is defined by Cov[X, Y] = E[X - E X] [Y - E Y]. Prove that if f and g are
non-decreasing functions and X is an r.v. with E f(X), E g(X), E f(X)' g(X) finite,
then Cov[f(X), g(X)] ~ O.
(v) for any fil2 r.v.s. Xj' 1 ~ j ~ n, check that
(12 (t
j=l
Xj) = t
j=1
(12(X) +2 L
l$i<j$.
Cov(X;, Xj)'
*
4. For arbitrary real numbers ai' bi, 1 ~ i ~ n, prove that
8. Verify for 0 < a < b < d and any nonnegative r.v. Y that
E y b :s; (E P)ld-bl/1d-al(E yd)(b-al/1d-al.
Utilize this to conclude for any positive r.v.s {Y", n ~ I} and positive constants
{c., n ~ I} that ifI.""= I e: E Y:
< 00 for IX = IX I and 1X 2, where IX j > 0, i = 1,2, then
the series converges for all IX in [lXI' 1X2].
9. The moment-generating function (m.gJ.) of a r.v. X is the function rp(h) = E e hX •
If rp(h) is finite for h = ho > 0 and h = -h o, verify that (i) E e h1x1 :s; rp(h) +
rp( - h) < 00, 0 < h :s; ho (ii) log rp(h) is convex in [ - ho, ho] and strictly convex if
X is non-degenerate (iii) ifE X = 0, then rp(h) ~ 1 for Ihl :s; ho. (iv) The kth moment
of X is finite for every positive integer k and equals rplkl(O), (v) if X I and X 2 are
independent r.v.s with (finite) m.gJ.s rpt (h), rp2(h) (say) for Ihl < ho, then the m.gJ.
of X I + X 2 is rpl (h)' rp2(h) in this interval, (vi) if rp(h) is finite in [0, ho], all moments
of X+ are finite.
2
P2(n) = L P{AjAd, d. = P2(n) - p~(n).
n(n - 1) t $j<k".
Prove that (i) E Y; = dn + n-I[pI(n) - P2(n)], (ii) y" f.. 0 iff E Y; -+ 0 iff d. -+ 0,
(iii) Y" ~ 0 if d. = O(n-O) for some IX > O. Hint: Utilize Exercise 11.
13. For events {A., n ~ I} and e > 0, there exist distinct indices j, k in [1, n], where
n> l/e, such that P{AjA k } ~ piCn) - e, where PI(n) is as in Exercise 12. Hint:
P2(n) ~ piCn) - l/n.
4.3 Jensen, Holder, Schwarz Inequalities 111
14. Bernstein's Inequality. If S. = Ii=1 X j where {Xj' 1 5,j 5, n} are independent r.v.s
with E X j = 0, E Xl = ul, s; = Ii=1 ul > 0 which satisfy (i) EIXl5, (k!j2)ulc k- 2
for k > 2, 15, j 5, n,O < c < 00 (a fortiori if(i)' P{IXjl5, c} = 1,15, j 5, n), then
Hint: exp{tXj } 5, 1 + tXj + I~=2tkIXl/k! valid for t > 0 implies E elKj 5, expo
{ult 2/2(1 - tc)},O < tc < 1and choosing t = x/(s; + cx)yields Eexp{xS./(s; + cx)}
5, exp{x 2 /2(s; + cx)}, x> O. Now apply Exercise 4.3.3(ii).
15. If S. = Ii=1 X j where {Xj' 1 5, j 5, n} are independent LV.S with E X j = 0, E Xl =
ul, s; = Ii=1 ul > 0, then for A> Y > 1,
p{ max ISjl
15j5.
~ AS.} 5, Y~P{lS.1
- 1
~ (A - y)s.}.
Hint: 1fT = inf{1 5, j 5, n: ISjl ~ AS.} and T = n + 1otherwise, then for 1 < y < A
Hint. Use the Feller-Chung lemma with A j = {Sj > e},B j = {So - Sj > -EISol}.
References
Y. S. Chow and W. J. Studden, .. Monotonicity of the variance under truncation and
variations of Jensen's inequality," Ann. Math. Stat. 40 (1969), 1106-1108.
K. L. Chung and W. H. J. Fuchs, "On the distribution of values of sums of random
variables," Mem. Amer. Soc. 6 (\951).
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
P. R. Halmos, Measure Theory, Van Nostrand, Princeton, 1950; Springer-Verlag,
Berlin and New York, 1974.
P. Hall, "On the 2"p convergence of random variables," Proc. Cambridge Phi/os. Soc.
82 (1977). 439-446.
G. H. Hardy, J. E. Littlewood, and G. Polya, Inequalities, Cambridge Univ. Press,
London, 1934.
112 4 Integration in a Probability Space
S. B. Kochen and C. J. Stone, "A note on the Borel-CanteIli lemma," /II. Jour. Math. 8
(1964),248-251.
A. Liapounov, "Nouvelle forme du theoreme sur la limite de probabilite," Mem. Acad.
Sc. St. Petersbourg 12 (1905), No.5.
M. Loeve, Probability Theory, 3rd ed., Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
G. Polya, "Uber eine Aufgabe der Wahrscheinlichkeitsrechnung betreffend die
Irrfahrt im Strassennetz," Math. Ann. 84 (1921), 149-160.
S. Saks, Theory ofthe Integral (L. C. Young, translation), Stechert-Hafner, New York,
1937.
H. Teicher, "On the law of the iterated logarithm," Ann. Proh. 2 (1974),714-728.
5
Sums of Independent Random
Variables
113
114 5 Sums ofIndependent Random Variables
Lemma 1 (Abel). If {an}' {b n} are sequences ofreal numbers and An = D=o aj,
n ~ 0, then for n ~ 1
n n- 1
1
- I
n
aj --+ O. (6)
bn j= I
PROOF. This will be demonstrated in the alternative equivalent form that
convergence of If
aj entails II
ajb j = o(bn). By (4),
1 bl 1 n- I
-A: + A6 -b + -b I
n
For any I: > 0, choose the integer m so that I A! I < I: for j ;;:: m. Then
1 n- 1 _I n - 1
-I::::; limb L A!(bj + 1 - bj):::; limb L Aj(bj + 1 - bj):::; 1:,
n n m n n m
(8)
PROOF. It suffices to prove (8) since Theorem 1 guarantees the rest. To this
end, set Z2 = sup X~ and choose K > 0 sufficiently large so that
Vn = SminlT. n)'
(10)
n n
E Vn = E I
j= 1
XjI(T""j] = I
1
P{T ~ j}E Xj'
whence
PROOF. It suffices to consider separately the cases 1 :s; IXn :s; 2, n ~ 1, and
0< IXn < 1, n ~ 1. In the former instance, X;IIIX nl,;l) + IXnIIIIXnl>l):S;
IX nlan, whence (12) obtains. In the latter,
ro ro
2: E(X; + IXnI)IIIXnl,; 1) :s; 22: EIXnlan < 00,
1 1
Furthermore, if either
If Xnln llp converges a.c.
(i) °< p < 1 or (ii) 1 < P < 2 and E X I = 0, then
PROOF. Set Aj = {(j - WIP < IXII ~/IP},j ~ 1. Then for rx > p > °
~EIy"la = JI Jl n -
alP
{jlXlla = JI Jt- aiP
{jlXlla
~ .I: (r alP + ~
J; I rx P
p-a)IP) f IX t1
Aj
a
~ I: (; +~)
j; I rx ]
f IXti ~ ~EIX,IP
rxP
<
Aj
P
P
00, (13)
whence (rx = 2) If
(Y" - E Yn) converges a.c. by Theorem 1. Since, re-
calling Corollary 4.1.3,
I'"
I
P{X-1% 1= n}= I'" P{IXII > nllp }
n
Y
I
~ EIXIIP < 00,
the sequences {X nln IIP }, {y"} are equivalent, and so (X nln l/P - E y") Ir'
converges a.c.
°
In case (i), where < P < 1, In"';
I IEy"1 < 00 via (13) with rx = 1, and
this same series converges in case (ii), 1 < P < 2 and E X I = 0, since
I:IEy"I~In-'IPi
I I IIXnl>nl/p)
IXnl= fIn-liP
n;lj;n+1
f Aj
IX,I
.I ji,ln-IIPf 'XI,~~.I(j-l)(P-')IPf
J;2n;1 Aj P I J ;I Aj
IXII
EXAMPLE 1. Let {X n, n ;;::: 1} be i.i.d. r.v.s with EIX 11 < 00. If {an, n ;;::: l} are
real numbers such that an = O(1/n) and If
an converges, then 1 anX n L:'=
converges a.c.
PROOF. By considering X;; and X;; separately, it may and will be supposed
that XI;;::: O. Set
Y" = X~ - E X~.
00
:::; An~l j~ln-2
n f AjXi:::; 2A E X I'
where the last inequality follows via (13) with IX = 2, p = 1. Hence, Lan Yn
converges a.c. by the Khintchine-Kolmogorov convergence theorem. 0
b; Lb
00
2
j- = O(n) (14)
j=n
and {X, X n , n;;::: I} are i.i.d. r.v.s with I:'=l P{!XI > bn } < 00, then
00
L b;;l(Xn -
n= 1
E XIUX1s;b"l) (15)
then
(17)
converges a.c.
Remark. If bn/n l or bibn ;;::: AWn)6 for j ;;::: n, where b > t, A > 0, then
(14) holds; if bibn ;;::: AWn)lJ for j :::; n, where 0 < b < 1, A > 0, then (16)
obtains.
120 5 Sums oflndependent Random Variables
00 ex> co a:>
~ C L n P{A n} < 00
I
via (14) and (18). By Theorem 5.1.1, Lf bjl(lj - E lj) converges a.c., and
so, again employing (18), Lf bj-I(X j - E lj) converges a.c., yielding (15).
Moreover, if E X = 0,
<Xl
(21)
In case (i), a~" = E(X I I!la"Xd:S; I] - E Y,Y. If aL --+ for some subsequence °
nj , then XI Illa".xd:S; I] - E Y. .!. 0, implying Xl degenerate, contrary to the
j
J
hypothesis of (i). Thus lim. a~" > 0, whence (21) implies (20).
5.1 Three Series Theorem 121
IE Y"I = IJUX"I
f
> la"l- 1)
Xn I~ JIIX"I
f
> la"l- 1)
IXnl = 0(1)
lim at = lim(E Y; - E 2
y") = lim E Y; ~ lim E 2
1Y" I
n
E(S;" - Sm)2 = I E
keQ
xr
Now, if {nl, ... , nm } ~ {l, 2, ... ,j}, then
00
and if so, If
X nj = If
X j a.c.for every rearrangement {n j} of {j}. Moreover,
a series ofindependent r.v.s {X n } converges absolutely a.c. iff (i'), (iii'), and
ii". If EI Ynl < 00
hold.
PROOF. Since the series appearing in (i'), (ii'), (iii') are independent of the
order of summation, the three series theorem (Theorem 2) guarantees that
If X n converges a.c. unconditionally, and by Lemma 5
EXERCISES 5.1
1. Let S. = D X;, where (X", n ~ I} are independent LV.S. (i) IfI:'=1 P(IX.I > c}
= 00, all c> 0, then lim"_oo IS.I = 00, a.c. (ii) If (b", n ~ I} are positive constants for
which li.m.-oo P(S"-l > -bb"} > 0 for all b > 0, then fIn1 S"/b" s: C < 00, a.c.,
implies I:'= I PIX" > eb"} < 00 for all e > C. Hint: Recall Lemma 3.3.4.
2. Any series I~1 Xi of r.v.s converges absolutely a.c. if I;;'=1 EIX.I'" < 00 for some
sequence r" in (0, I].
5.1 Three Series Theorem 123
13. (Chung) Let '¥ be a positive, even function with x-Z'¥(x) 1, Ixrl,¥(x) l as Ixll. If
0< bn l 00, {X.} are independent with EX. = 0, I (E 'I'(X.)/'¥(b.» < 00, then
I (X.lb.) converges a.c. Hint: Apply Corollary 2 with X n replaced by Xnlb n.
15. If {b., n ~ I} are finite constants with 0 < b.l 00 and (i) b; IJ=. bj- Z = O(d(b.»,
where d is a nondecreasing mapping of [0, (0) into [0, (0) for which (ii) d(b.) ~
en> 0, n ~ I, and (iii) xZld(lxi) l as Ixll, then for any i.i.d. LV.S {X, X n , n ~ I}
with (iv) E d( IX I) < 00, the conclusion (15) obtains. Moreover, if E X = 0 and
(v) Ixl/d(lxl) 1 as Ixll and (vi) bn I:i= 1 b; 1 = O(d(b.», then (17) holds.
16. If {X, X., n ~ I} are i.i.d. LV.S with E X = 0, E XZ(l + log+ IXI)-26 < 00 for some
(j > 0, then I n- I/Z (log nr(l/Z)-~X. converges a.c.
124 5 Sums of Independent Random Variables
S -np 1
= _ L (Xi
n
n - E Xi) ~ O.
n n i~l
(a)
If, merely,
(b)
the sequence {X n} satisfies the classical weak law of large numbers (WLLN).
From a wider vista, n may not reflect the real magnitude and the expecta-
tions need not exist. Thus, there is occasion to consider the more general
strong and weak laws of large numbers
~ ~(X.
"-- I
- b.)~O
I P ,
an i~ I
where 0 < an t 00. Here, the smaller the order of magnitude of an, the more
precise the SLLN becomes; the fuzzy notion of an optimal choice of an
impinges on the law of the iterated logarithm (Chapter 10). Note, in this
context, Exercise 5.1.12.
The first SLLN may be reaped as a direct application of Kronecker's
lemma to Corollary 3 of the three series theorem (Theorem 5.1.2), thereby
obtaining Loeve's generalization of a result of Kolmogorov (IXn == 2).
for some choice of IXn in (0, 2], where E X n = 0 whenever 1 :::; IXn :::; 2, then
1 n
- LXj~O. (2)
n j~ 1
5.2 Laws of Large Numbers 125
(3)
for some finite constant c iff E IX I IP < 00, and if so, c = E X I when 1 S; P < 2
while c is arbitrary (and hence may be taken as zero)for 0 < p < 1.
PROOF. If (3) holds, then
X n _ Sn - nc _ (n - 1)I/P Sn-I - nc ~O
n l/p - n l/p n (n - l)llp ,
1 < P < 2;
p = 1; (4)
00 X
O<p<1.
~ nl;P'
a.c. o
EXAMPLE 1. If {X n' n ~ 1} are i.i.d. ffp random varibles for some p in (0, 2),
then (i) L::"=I n- 2jpX; < 00, a.c. Moreover, if E X = 0 whenever 1 ~ p < 2,
•. )
( It n -2jp,,\,~ X.S. ~O.
L.J = 2 J r I
PROOF. According to the proof of Theorem 5.1.3, y" = n - ljp X nlllx"1 sn1/PI'
n~ 1 and n-IIPX n , n ~ 1 are equivalent sequences and L::"=l E Y,,2 < 00
implying (i). Then, by Kronecker's lemma, n- 2IP Lj=1 XJ~ 0, and so
Theorem 2 ensures that
n- 2lp L XjS j _
n
j=2
1
1 2Ip [ S; -
= _n-
2
L XJ
n
j=l
]
~O. o
If {X n , n ~ 1} are i.i.d. LV.S, Sn = L~ Xi and {b n , n ~ 1} are constants
with 0 < bn t 00, the relationship between X ..Ibn -!:4 0 and (Sn - Cn)/bn ~ 0
for some sequence of constants {Cn , n ~ 1} is described by
Theorem 3. Let {X, X n, n ~ 1} be i.i.d. r.v.s with partial sums {Sn, n ~ I} and
{b n, n ~ I} constants such that 0 < bn t 00 and
00
(i) (Feller) If(rx.) b,jn t 00 or (P) bn/n!, the first halfof(6) holds and E X = 0
or (y) E X = 0 and
00 n
b; Lbj2 = O(n), bn L bj- I = O(n), (6)
j=n j= I
then
S
~~O
b . (7)
n
(8)
and, moreover, if E X = 0,
1 n
-
A"L. a j X j a.c.
---+
0. (9)
n I
PROOF. According to Example 5.1.2, condition (5), the initial portion of (6)
and Kronecker's lemma ensure
(10)
Lj P{b
00
by Example 5.1.2 (18), and so (7) follows from (10), thereby proving (0:) of
(i). In case ({3) since E X = 0,
128 5 Sums ofIndependent Random Variables
= 0(1)
(11)
For any i.i.d. LV.S {X n, n ~ I}, there exist constants Cnfor which
1
-An L a ~X. -
n
C-) ~ 0 (12)
1 J J J
iff
anXn
- - ----+
a.c. 0 (13)
An .
an(X n - Cn)/A n ~ O.
Although necessary and sufficient conditions for the WLLN are available
in the case of independen( r.v.s, a complete discussion (see Section 10.1)
requires the notion of symmetrization. The i.i.d. situation, however, is
amenable to methods akin to those already employed.
n P {I X I I> n} - 0, (15)
S~ - E S~ S~ p
n =-; - EXIII/Xdsnl- O. (16)
f
I I
= nLn Xi
j=1 [j-I<IXdsjl
n
Moreover, for max, $j$nICjl < ne, and hence for all large n,
~ p{ ,max
$)$n
IXjl < 3ne},
LP{ISnl<e}~-2Iim-k
00
n=O
1 IP
km
k-oo m n=O
m {I S
-!'.
n
Ie} m
<- =-2'
m
Since m is arbitrary, the series on the left diverges for all e > 0 and the con-
clusion follows from Example 4.2.1. The final remark stems from the strong
law of large numbers. 0
5.2 Laws of Large Numbers 131
(23)
L ai/k;
00 00 00 k
2
e L P{ISk n - E SkJ > ek n} ~ L a~k /p;. ~ L
n=1 n=1 n n=1 j=1
-I' I
Imk--+oo Sk - kE Sk I (-
a ~ - +0,
I)B -"h a.s.
yielding (25). D
Corollary 4. If Sn = Lj= 1 Xj' n 2:: 1, where {X n' n 2:: I} are non-negative, pair-
wise independent random variables such that
00
L n- 2 E X;IIXnSnl < 00
n=1
(27)
132 5 Sums of Independent Random Variables
(28)
n=l
then
(29)
-n1 ( i=l
Ln n
Y; - LEY;
i=l
)
~ o. (30)
implying as n -+ 00 that
-
1(
n
ES n -
n
LEY; =-
i=l ni=1
) 1L
EXJ[X,>i)-+O.
n
(31)
Since (28) ensures that {Xn' n ~ I} and {Y,., n ~ I} are equivalent sequences,
(30) holds with L~= 1 Y; replaced by Sn, and this together with (31) yields (29).
D
L L n- 2E(X+)2Iu_ <x+
00 00
= 1 S;j)
j=1 n=j
=CEX+<oo,
whence Corollary 4 guarantees
1 n
Sn/n=-
n j= 1
L
(xt -Xi-)~EXt -EXt =EX 1 • D
5.2 Laws of Large Numbers 133
PROOF. Define T = smallest integer j in [1, n] for which ISjl 2:: e if such
exists and T = n + 1 otherwise. Then {T 2:: j} E a(X I' ... , Xj-I) for
1 ~ j ~ n + 1, where the a-algebra in question is {0, O} whenj = 1. Thus,
X j and IIT"j) are independent r.v.s for 1 ~ j ~ n. Moreover, since ST = Sj
on {T = j}, I ~ j ~ n, and s:..inIT.n) = LJ=
I XjIIT"j]'
I~i
~ 2' L..-
e j= I [T=Jl
Sj2 = Ii
2'
e ITSn]
2
SminlT.n)
T,. p
- --+ c, where 0 < C < 00, (34)
n
Thus,
P{BrJ ~ P{T" ~ 2en , BrJ + P{T" > 2en }
~ P{Dn } + P{T" > 2en } = 0(1),
so that
S~. - M~ • .f. O.
n
o
For any sequence of LV.S {X n , n ~ I} and nonnegative numbers u, v
define
Su.v = L
t sjsv
X 1ul + j, Sv = So,v, So = O. (37)
5.2 Laws of Large Numbers 135
The r.v.s Su.v are called delayed sums (of the sequence X n). If I,.<Xl= I P{ ISnl
> m:} < 00 for every I: > 0, then by the Borel-Cantelli lemma, Sn/n ~ O.
The converse is false even when the {X n } are i.i.d. since then by the classical
SLLN (Theorem 5.2.2, p = 1), Snln ~ 0 iff E XI = 0, whereas according
to a theorem of Hsu-Robbins-Erdos (Chapter to) I,.<Xl= I P{lSnl > nl:} < 00,
aliI: > 0 iff E X I = 0, E xi < 00. However, in the i.i.d. case Snln ~ 0 does
imply that I,.<Xl= 1 (l/n)P{ ISnl > nl:} < 00 for every I: > 0 by a result of
Spitzer, and the next theorem due to Baum-Katz extends this.
For any sequence of r.v.s {X n' n ~ I} define via (37).
(40)
Suppose first that rxp = 1. Since {S!n.2n, n ~ I} are independent r.v.s, (40)
and the Borel-Cantelli theorem (Theorem 3.2.1) imply that for every I: > 0
and c = (log 2)-1
<Xl <Xl
00 > L P{S!n.2n ~ 2 l:} = L P{S!n ~ 2 l:}
an an
n=1 n=1
.
m-a/(ap-I)S:~P/(~P _I) ml/(~p _ I) ~ m-a/(ap- I) S:~P/(~P- I).m~p/(~p-') ~ 0
136 5 Sums of Independent Random Variables
EXERCISES 5.2
I. Prove for i.i.d. LV.S (X.} with S. = D Xi that (S. - C.)fn ~O for some sequence
ofconstantsC.iffEIXII < 00.
2. If {X.} are i.i.d. with EIXIIP = 00 forsomepE(O, 00), then P{lim.IS.lln l / P = oo} = 1.
3. Demonstratefor i.i.d. r.v.s {X.} that E sup. I X .Inl < 00 iff EI X Illog+ IX II < 00.
4. If S. = D
Xi' where {X.} are i.i.d. ft'p r.v.s for some p;<: 1, then EIS'/nI P -+
P
IE X I . Hint: Recall Example 4.3.1.
5. If {X.} are i.i.d. LV.S with E XI = I and {a.} are bounded real numbers, then
(lIn) D a j -+ I iff (lIn) D= I ajX j ~ I.
6. Let {X.} be i.i.d. LV.S with (S.ln) - C• .!. 0 for some sequence of constants {C.}.
Prove that
1 •
• )I/P ~ ajXj~O.
I at
( j=1
J=I
Conversely, ifa:/Il=1 at -
Cn, then (*)ensures X E 2'p(and also EX = Oifp = I).
Hint: if y,. = a.(IJ=1 af}-I/PX.· Inx.I".IIP], then EI y"1" :::; Cn-"/PEIXI"IIIXI""IPI
whence I~=I EI y"1" < 00 for (X> P and I~=I IE Y,.I < 00 for I < p < 2 as in
Theorem 5.1.3. N.B. Exercise 7 with p = I and Theorem 5.2.3(ii) are related to
summability.
8. For r > 0 and any LV. X, prove that X E .!l;. iff
00
~
b.
fX
I
j - ~ E XII!XI:<b"I'!' 0 iffn P{lXI
b.
> b.} = 0(1).
10. Prove that if {X., n ;;:: I} are independent r.v.s with EX. = 0, EX; = a;, s; =
D a; -+ 00, then s;; l(log s;)-" D Xi ~ 0 for (X > t-
Il. (Feller-Chung) Let P(x) be nonnegative and nonincreasing on (0, 00) and suppose
that the positive, nondecreasing sequence {b., n ~ I} satisfies (*) lim.- 00 b.,lb. >
c > I for some integer r or a fortiori either (**) b.lnP t for some fJ > 0 or (***)
b; If=. bj- 2 = O(n). Then Loo= I P(xb.) either diverges for all x > 0 or converges
for all x > O. (Hint: Any integer m satisfies'" :::; m < ,M I, where k = k(m) -+ 00 as
m -+ 00, whenceb. m ~ ckb.,alliargen). Thus, if {X, X.} are i.i.d. and Loo=l P{IXI >
Cb.} = 00 for some C > 0 and {b.} as stipulated, 1IiTI._ 00 I X.llb. 'c 00 = 1IiTI._ oo
IS. lib•.
12. (Heyde-Rogozin) If {X, X., n;;:: 1} are i.i.d. with (*) lim x _ oo x 2 P{lXI > x}1
E X 2 I lIxl :<x) > (X> 0, then for every sequence {b.} satisfying 0 < b. too either
1IiTI._ooIS.llb. a.C. CD or (Ilb.XS. - L~ E X/IIXI';bJJ~O. In particular, for sym-
metric i.i.d. LV.S satisfying (*), Sjb. - + 0 or IIiTI S./b. a.c. w. Hint: P{I X I > x} :::;
(I + b 2(X-l)P{lXI > bx}, b > I, x > x o , whence the series of Exercise II either
converges for all C > 0 or diverges for all C > O.
13. (Komlos-Revesz) Let {X.,n ~ I} be independent r.v.s with means E X.and positive
variances a; c
satisfying lim EX. = and I~= 1 (j;; 2 = 00. Then
15. (Derman-Robbins) {X.} i.i.d. with E X I nonexistent does not preclude P{S.Jn
--.. oo} = I. Hint: Let Iim x _..,x· PIX > x} > 0 and E(X-)P < 00 for I> f3 >
:x > O. The latter implies n- IIP I7 X i- ~ 0, while the former entails P{I7 xt =:;
Cn 1IP } =:; P{max 1 "i". xt =:; Cn l/P } =:; exp( -en Y ), y = I - rxfJ-l > 0, C > O.
16. (Klass) Let {X.} be i.i.d. with E XI = 0 and I:'=I P{IX.I > b.} < 00, where
b. > 0, n- b; L n-Ib; i
2
00.
Then EIS.I = o(b.). Hint: X j = y".j + Z•. j, where y".j = X/UXj!';b n ]' In particular,
X. i.i.d. with E X I = 0, EI X liP < 0:, P E [1,2), implies EIS.I = o(n 1IP ).
17. Strong Laws for Arrays. Let {X. i , 1 =:;; i =:;; n} bean arrayofr.v.'s that are identically
distributed and rowwise independent, i.e., {Xnl , ... , X •• } are independent for n ;;::: 2.
If Elxlllq<oo, 0<q<2, and EXll=O whenever l=:;;q<2, then
n- 2 /q I7=1 X'i~ O. (For an extension to 2 =:; q < 4, see Example 10.4.1.)
Clearly, for any sequence of r.v.s {X n}, {X n, u(X t, ... , X n), n;::: I} is a
stochastic sequence.
For any stochastic sequence {X n , fF n , n ;::: I} and fFn-timc T, define
X T = XT(W)(w), where X",(w) = rrm Xn(w), (2)
(3)
fF '" = u(9 fFn).
T is an {fF n, n ~ 1}-time. To prove fF T = t:§ T2' let A e t:§ T2' Then for m =
1, 2, ... and j = 1, ... , m - 1,
which implies that AefFT. Conversely, let AefF T. Then for r = 1,2, ... ,
and m = 0, I, ... ,
A{T = r + m} efFr+m,
A{T2 = m}· {T1 + m = r + m} = A{T = r + m}{Tl = r} e fF,+m,
A{T2 = m} e fF T, +m = t:§m,
P{A· [T
n
= j]}iG P{X j +i < AJ
<Xl n
= I
j= I
P{A . [T = j]} TI P{X i < AJ
i= I
= P{A} TI P{X
i= I
i < AJ. (7)
Hence, taking A = n,
n
and, in particular, P{X T+j < Aj } = P{X j < Aj } = P{X I < Aj }, 1 :S;j:S; n.
Thus, since n is an arbitrary positive integer, {X T + n' n 2:: I} are i.i.d. with the
same distribution as X I' Consequently, from (7)
Then, as noted earlier for j = 1, T(j + I) is a finite {X T J + n' n 2:: 1}-time, j 2:: 1.
The stopping variables {T(j+ I), j 2:: I} or {TW, j 2:: l} will be called copies
of T. Moreover, as folIows from earlier discussion, Tm = Lj= I TU) is a
(finite) {Xn}-time and ff'T~ c ff'T~+I' m 2:: 1.
In a sequence of Bernoulli trials with parameter p = t, i.e., Sn = Xi' I1
where {X n} are i.i.d. r.v.s with P{X j = ± I} = t, let T = TI = T(I) =
inf{n 2:: I:S n = O}and T(j+I) = inf{n 2:: I:S Tj +n = O}.Then ~ = If=1 Tli),
j 2:: I, are the return times to the origin and T(j) is the time between the
(j - l)st and jth return. If, rather, T = inf{n 2:: 1: Sn = l}, then ~ is the
first passage time through the barrier at j, whereas TW is the amount of time
required to pass from j - 1 to j. Either choice of T yields a finite {X n}-time
with infinite expectation (Exercises 3.4.1, 3.4.2).
142 5 Sums of Independent Random Variables
= p{T(m) = n,X T _
1
+1 < AI, ... ,X Tm _ +. < A.} I
EIXTI = r
J\T<",j
IXTI + r
J{T="'}
ITmi X.I
~
j=. J{T=j}
r IXjl + r
J{T="'1
Ilim X.I. (8)
EST = EX 1 • E T. (10)
PROOF. Suppose first that E I X I I < 00. Then by Theorem 4.3.3
00 00 00
If, rather, E X I = 00, then E Xl < 00, EX: = 00, whence by what has
just been proved
00
Theorem 2. Let {X n} be i.i.d. r.v.s, let Sn = D Xi' and let T be a finite {X n}-
time for which E ST exists. (i) IfE X I exists and either E X I#-O or E T < 00,
then (10) holds. (ii) IfP{ IX II> n} = o(n- I) and E T < 00, then
EST rEI
ET = X I lIX Ii sn)'
1m (II)
n-oo
*
PROOF. If T(n) = min[T, n], then T(n) is a finite stopping time and
ES?(n) E( XjIIT~j))
= 2
=
5.3 Stopping Times, Copies of Stopping Times, Wald's Equation 145
fi'n-t,
n- I n- I
whence
Since T(n) t T,
lim E S}(n) = (12 E T < 00. (17)
n~'"
E(ST(n) - ST(m»2 = E( ±
m+1
Xjllnjl)2 = (12[E T(n) - E T(m)]
E aTISTl' = JI LT=jlajlSjl'
146 5 Sums ofIndependent Random Variables
00 j
= M L L a" P {T =
j= ,,=
1 1
j} = MEAT·
Hence, for any finite stopping time T, (18) holds for T(n) = min[T, n],
yielding (18) as n -+ 00 by Fatou's lemma and monotone convergence. To
prove (19), note that by Holder's inequality
EIS Ir«
T
= E T«(l-«)
ISTlr« . T«(l-«) <
-
(E ISTI')«(E P)l-«
T1-« '
and so, employing (18) with all = l/n 1-«, All ~ n«/a., (19) follows. 0
(20)
then E T~ = 00 for c 2 > 1/a., 0 < a. < I, and E U~ = 00 for m > m«, all
a. > O. The latter statement follows from the former which, in turn, results
Gr
from (19) via
2
c «E T~ ~ EISTf« ~ E T~.
The same conclusion holds if {XII' n ~ I} are independent r.v.s with mean
zero and variance one which obey the central limit theorem (Chapter 9),
since r; is a finite stopping time (Exercise 3.2.12).
since the joint distributions of (Y", Zn) and (Yl' Z d are identical. The conclu-
sion then follows by induction. 0
If {X, X n' n 2:: I} are i.i.d. random variables with E X = 0, E X 2 < 00, then
limn .... 00 Sn/n 1/2 = 00, a.c., as noted in exercise 3.2.12. This remains true even
when E X 2 = 00, according to
Theorem 4 (Stone). If Sn = Ij= 1Xj' n 2:: 1, where {Xn' n 2:: 1} are i.i.d. random
variables with E Xl = 0, E IX11 > 0, then
lim Sn/n 1/2 = 00 = - lim Sn/n 1/2 , a.s. (22)
n.... oo
PROOF. It suffices to prove the first half of (22). To this end, choose finite
°
constants a, b such that a < < band P{a < Xl < O}P{O < Xl < b} > 0.
Then T = inf{n 2:: 1: a < X n < b} is an {Xn}-time with finite expectation.
Since, as noted earlier, the theorem holds when E xi < 00, it may be supposed
that E Xf = 00, whence 0< P{a < Xl < b} < 1. Set To = 0, T" = Lj= 1 TW,
n 2:: 1, where {TW,j 2:: 1} are copies of T and define
(23)
According to Lemma 3, { ~,j 2:: 1} and Z j' j 2:: 1} are each i.i.d. sequences and,
moreover, via Lemma 4, {~,j 2:: 1} is independent of {Zj,j 2:: I}. By Wald's
equation, E Y1 + E Zl = E If X j = 0, whence
n n n
ST n = I (~+ Zj) = I (~- E ~) + I (Zj - E Zj)' (24)
j=l j=l j=l
M + ST tm
2:: ~ (~ -
jf-1
E ~) --[1 (J't:.!2
~ (~ -
jf-1
E~) ] 1/2
(J'tlft (25)
148 5 Sums of Independent Random Variables
p{-k
arm
.I (lj -
)=1
E lj) > x}
= f p{ak11 /
k=m
2 .t
)=1
(lj - E lj) > x}p{r m = k}~ 1 - $(x)
for all x via the Central Limit Theorem for i.i.d. random variables (Corollary
9.1.2).
Consequently, for any x > 0, as n ~ 00,
Now, T.jr m = : Ii':: 1 TUl ~ E T < 00, implying via (26) that, for x > 0,
m
as m --+ 00,
and so
p{s T tm
/T.t 1 / 2 > xa
rn 3
E- 1 / 2 T. i.o.}
'
= lim
m-oo
p{.O [ST /Tr~/2 > xa3 E- T]}
)=m
t
j
12
/
~ 1 - $(x)
Thus, Qx = 1 by the Kolmogorov zero-one law and (22) follows as x --+ 00.
D
EXERCISES 5.3
1. Verify that if T is an {~}-time, X Tm +j is §Tm'j-measurable,j ~ 1, m ~ 1.
2. If T I and T2 are ~-times, so are T I T2, max(TI , T2), min(Tl , T2), and kTl , where k
is a positive integer.
3. If T is an integrable {X.}-time, where {X.} are independent LV.S with E X. = 0,
EIX.I ::; C < 00, n ~ 1, then Wald's equation holds.
4. If (Xi' ~), i ~ 1, are i.i.d. 2 2 random vectors with E Xl =E Yl = 0, and ~ =
a(X I , YI ,···, X n, Y,,), S. = I~ Xi' Un = I~ ~, then for any integrable ~-time T,
the identity E STUT = E T· EX l YI holds.
5. Let S. = I7=l Xi where {X., n ~ 1} are i.i.d. r.v.s with E XI = Jl > 0 and N = Np
an {X. }-time (or a LV. independent of {X., n ~ I}) having the geometric distribu-
tion. Then Iimp~o P{SN/E SN < x} = 1 - e-X, x> O.
6. Show that the condition E T < 00 cannot be dropped in Corollary 3. Hint: Con-
sider P{X. = 1} = P{X. = -I} = tand T= inf{n ~ 1: S. > O}.
7. If {X., n ~ I} are independent random variables with E X. = 0, E X; = 1 and T*
(resp. T*) = inf{n ~ 1: IS.I > (resp. <)en l /2 }, where inf 0 = 00, then E T* = 00,
e ~ 1 and E T* = 00, e ::; 1.
8. If {X., n ~ I} are independent LV.S with EX. = 0, E X; = a;, n ~ I and T is an
§ .-time for which E Ii aJ < 00, where (i) §. ~ a(X I' ... , X.), (ii) § nand
a(X.+ I) are independent, n ~ 1, then E(Ii X)2 = E Ii aJ.
13. Show that lim sup.~oo n -1/2IIi'= I Xii = 00, a.c. for any sequence of Li.d. random
variables {X., n ~ I} with EIXti > O.
14. Let S. = Ii'=1 Xi' n ~ 1, where {X, X., n ~ I} are i.i.d. random variables with
E X = 0, E IX I > o. For x > 0, p > 0, and k = 1, 2, ... , show that
qk == p{ max Sj//
I';j';.
~ 2kx}:S; p{ m~x X)/ ~ x} +
I';j';'
pk{ m~x
I';j';'
Sj// ~ x},
and if 1 :s; r :s; 2 and P< l/r, there exists a universal constant C',I1 such that
Theorem 1. Let {X.} be i.i.d. r.v.s with EIXII > 0, S. = D Xi' and define
T+ = inf{n ~ 1: Sn ~ O}, T~ = inf{n ~ 1: S. > c > O}. (1)
Then:
i. P{T+ < oo} = 1 iff rrm S. ~. 00, in which case E ST + > 0;
ii. E T+ < 00 iff E T~ < 00 for all c in (0, (0).
Hence, rrm
s. ~ rrm
n EXt ~ 0, a.c., and, moreover, s. = + 00, a.c., rrm
since EXt > 0 in view of the fact that the only permissible alternative,
namely, E Xl < 0, would imply (same corollary) that S.ln ~ E Xl < 0,
contradicting rrm
S. ~ 0, a.c. The converse is trivial.
Apropos of (ii), since via (i)
for any c > 0 there exists an integer k such that p{I~ SUI > c} ~ t. Setting
Zn = I(':- t)k+ t SUI, the r.v.s {Zn, n ~ l} are i.i.d. by Lemma 5.3.3 and
clearly Zn ~ 0, a.c., with P{ZI > c} ~ t. Define r = inf{n ~ 1: Zn > c}.
Then
P{r ~ n} = P{ZI ~ c"",Zn_1 ~ c} = pn-I{ZI ~ c},
whence
00 00 I
Er= IP{r~n}= Ipn-I{ZI~C}=p{Z }~2.
n=1 n=1 I>C
Moreover, since
t kt kt Tj Tk'
C < I Zj =
1
L SU) = j=1
1
I I
Tj_I+1
Xi = I Xj =
1
STk<'
The stopping rule T+ and its partial analogue T_ , defined in (2) below, are
interconnected as in
yielding
1 ~ P{T'_ = oo}E T+, (7)
whence P{ T'_ = oo} > 0 implies E T+ < 00. Conversely, if E T+ < 00,
via (6)
j k
1~ L P{T+ ~ n}P{T'_ > k - j} + L P{T+ ~ n};
n= 1 n=j
letting k -+ 00,
j 00
1
E T+ = P{T'_ = oo}
Similarly,
00
yielding
(I - ~)P{T'_ = oo} = P{L = oo},
(ii) Suppose that E I X 1 I > o. If ~ = I, then T_ < 00, a.c., and, replacing
{X.} by {-X.} in Theorem I(i), E Sr- < O. On the other hand, ~ = I
entails ST _ = 0, a.c., via (5), a flagrant contradiction.
(iii) If L and hence also T'_ is defective, E T+ < 00 by (3) and afortiori
T+ < 00, a.c., so that according to Theorem I(i)
(8)
Thus, T = inf{ n ~ 1: S. > K} < 00, a.c. for all K > 0 and, via Lemma 5.3.2,
P{fu!!,,~oo (ST+. - ST) ~ O} = P{fu!!,,~oo S. ~ O} = e> o.
Consequently, with probability at least e,
fu!!,,-oo S. = lim._ oo ST+. = ST + fu!!,,~oo (ST+. - ST) > K,
implying that P{fu!!,,-oo S. = oo} ~ e > 0 whence, by the Kolmogorov zero-
one law, fu!!,,- 00 S. = 00, a.c., that is, S. ~ 00.
Conversely, if lim._ oo S. a.c. 00, T cannot be finite since Theorem 1 (with
X -+ - X) then would entail lim S. a.c. -00. 0
.-00
ii. T+ is defective, lim Sn ac. - 00, E T_ < 00, E T'-c < 00, C > 0;
iii. T+ and L are finite, rrm Sn a.c. 00,lim Sn a.c. - 00, E T+ = E L = 00.
The next theorem, which extends Corollary 2, asserts that the same
conclusion holds if SJn !. 0 and E I X • I > O.
Corollary 3. Let {X n} be i.i.d. r.v.s with EIX II > 0, Sn= Li Xi' and n P{ IX t I
> n} O. (i) Ifeitherlim EX IIIIXd,;n) does not exist orlim E X IIIIXd,;n) ~ 0,
-+
then T+ < 00, a.c. (ii) If, moreover, EX: = E Xl = 00, and if lim E X I
X IUXd,;n) = c finite or EX IIUXtl,;n) has no limit, then p{nm Sn = oo} =
P{lim Sn = -oo} = 1.
Corollary 1 implies that apart from degeneracy only three possible modes
of behavior exist for sums Sn of i.i.d. r.v.s {X n}. If, moreover, E X: = E Xl
= 00, the same trichotomy will be shown to exist for the averages Snln.
Corollary 4. If {X n} are i.i.d. random variables, then, for every a in (0, t], one
of the following holds:
(i) Snln"~ 00.
(ii) Snln"~ -00.
(iii) lim Snln" •. c. lim Snln" •. c. -00.
Nc = max{j: Sj = *
Xi :5: c}, c > 0,
the r. v.s {X n' n ~ I} being i.i.d. with p. = E XI E (0, 00]. Although, N c is not
a stopping time, when X 1 ~ 0, a.c. Nc + 1 = Tc'. Thus, when X 1 ~ 0, a.c.,
(13) and (14) below hold with N c replacing T~ and the former is known as the
elementary renewal theorem. A stronger result (due to Blackwell) asserts
that E N c +" - E N c - alp. as c - 00 (modification being necessary when
{X n } are lattice r.v.s).
References 157
L
j= 1
n 1[Xj >£jj
Xj = o(n), I: > 0,
PROOF. For I: > 0, choose N ;;::: 1 such that Lj= 1 S[xj>£j) X j < nl: for n ;;::: N.
Then, if n ;;::: N,
N-l co
~ I: E 1'" + LE
j= 1
X j + LE
j=N
XjI[xj>£j. Tn~Jl
co
= I: E 1'" + 0(1) + L P{1'" ;;::: j}E XjI[xj>£j)
j=N
co k
= I: E 1'" + 0(1) + L LE
k=N j=N
X j I[Xj>£Jl P{1'" = k}
co
~ I: E 1'" + O( 1) + L I:k P{1'" = k} ~ 21: E 1'" + O( 1),
k=N
(14)
For any m in (0, J1.), choose the positive constant K sufficiently large to
ensure E X 1I IX I :$ K) > m and define
n
Then {X~} are i.i.d. and, as earlier, E V < 00. By Wald's equation
K +c ~ E S~ = E V· E X'I'
and so
E T~ EVII
--<--<--<-
K + c - K + c - E X'I - m'
whence
~I E T~ 1
1m --~-
c-ex> C m
and (m --> J1.)
~I ET~ I
lm--~-. (15)
C-ex> c J1.
If J1. = 00, (15) is tantamount to (13). If, rather, J1. < 00, then by Wald's
equation J1. E T~ = E Sf~ > c, implying
. E T~ 1
I1m - - > -
c -00 C - J1.'
(16)
Hence,
E(ST - C)2 ~ E(X;)2 ~ E X} = o(c) as c --> 00, (17)
so that
2
u2
uf = E(T - E T)2 = 2 E T
U
+ o(e) = 3" e + o(e),
Jl. Jl.
which is (14). o
When X n , n ~ 1 are i.i.d. with EX 1 = Jl. > 0, u 2 = E Xr E (0, 00), (Jl.3leu 2)1 /2
(1;, - (clJl.) has a limiting normal distribution as e -+ 00 according to Theorem
9.4.2.
EXAMPLE 1. If {X, X n , n ~ I} are nonnegative i.i.d. r.v.s with P{X > O} > 0
and Tx = inf{n ~ 1: Sn = I~ Xi > X},
X
! E Tx ~ E mm
. (X ,X ) ~ E Tx ' X> O. (20)
S~ = IX;,
1
yielding (20). o
Consider the plight of a gambler who makes a series of plays with out-
comes X n , n ~ 1, where the X n , n ~ 1, are i.i.d. r.v.s, and who has the option
of stopping at any time n ~ 1 with the fortune y" = maxI ,;;i';;n Xi - en,
e > O. Since the gambler is not clairvoyant his choice of a rule for cessation of
play is a stopping time, i.e., his decision to stop at any specific time n must
be based only on X I' ... , Xn (and not on X j with j > n). Does there exist an
optimal stopping rule, i.e., one which maximizes E YT over the class of all
finite stopping times?
(25)
(26)
co
= L E XjI[T~J1I[xj~PI
!
co
= L P{T ~j}E X!I[Xl~Pl
I
co
=I [E(X I - f3)+ + f3p]P{T ~ j}
I
e
= (e + f3p)E T = - + f3.
p
Moreover,
yielding the first part of (26). Moreover, if T' is any finite {Xn}-time for which
E Yr exists, it may be supposed without loss of generality that E Yr > - 00.
Since for any b > f3
n
Yn = max Xi - en ~ b + L [(Xi - b)+ - e]
ISisn 1
In
j=!
i [x]>.jj
Xl = o(n) (27)
162 5 Sums ofIndependent Random Variables
and define
1;, = 1;,(1), 1;,(m) = inf{n ~ m: IS.I > cn l/2 }, c > 0, m = 1,2,.... (28)
Then for all m ~ 1, E 1;,(m) < 00 for all c in (0, 1).
PROOF. The argument for m > 1 requires only minor modifications from that
for m = 1 which will therefore be supposed. Let c E (0, I) and assume that
E 1;, = 00. If V = min(1;" rt), clearly E V < 00 and E V - 00 as n - 00. Thus,
E xt = o(E V) by Lemma 2 and via Theorem 5.3.3
EXERCISES 5.4
I. Verify the second equality of (3).
2. Prove that the stopping rule T of (24) remains optimal when Y. = X. - en, e > O.
3. Prove Corollary 3.
4. Prove that if {X.} are i.i.d. with E X I = J.I E (0,00] and N, = sup{n ;::: I: S. $ e},
then Ncle ~ I/Jl..
5. Let {X.} be i.i.d. r.v.s with E X I = J.I > 0 and
T= T; = inf{n;::: I:S. > e} fore> O.
Prove that Tic ~ IIJl., a.c. If, moreover, E Ki < 00, then
ST - Jl. E T
ft .!. 0 as e -+ 00.
6. Prove that if {X.} are i.i.d. with E X I = J.I E (0, 00] and 1; = inf{n ;::: I: S. > en·},
e > 0,0 < IX < I, then
7 1 -. I
-~e-- ~ ~ as e -+ 00 and E 1; -
(e)I/(1-.'
~ .
7. (Chow-Robbins) Let {Y", n ;::: I} be positive LV.S with lim Y" = I, a.c., and {a.,
n ;::: I} positive constants with a. -+ 00, aJa._ I -+ I. For e > 0 define N = N c =
inf{n ;::: I: Y" $ a.le}. Prove that P{N < 00, lim, aNle = I} = I and, if E sup.;> I Y"
< 00, then lim,~", E aNle = I.
8. If S. = L'iXi' where {X., n ;::: I} are nondegenerate i.i.d. r.v.s, T = inf{j ;::: 1: Sj <
- a < 0 or S j > b > O} is of interest in sequential analysis (Wald). Prove that T is a
stopping variable with finite moments of all orders. Hint: P {T > rn} $ P { IS jr -
SrU- \)1 < a + b, I $ j $ n} for all integers r > O.
5.4 Chung-Fuchs Theorem, Elementary Renewal Theorem, Optimal Stopping 163
9. (Alternative proof that T'+ < 00, a.c., implies E L = 00.) Let {T (J), j ~ I} be
copies of T'+ and set T;, = I~ TW. Then Z = If
IITn<oo) = number of times S.
exceeds Si' i < n (take So = 0) and P{T_ ~ n + I} = P{S. > Si' 0 ~ i < n} =
P{U~ [1] = n]}, whence E T_ - I = E Z = 00.
10. If {X., n ~ I} are i.i.d. with S. = D Xi' then (i) Jmi S. = 00, a.c., iff there is a finite
{X.}-time T with EST> O. (ii) Moreover, when EXt = 00 Jmi S"jn = 00, a.c.,
iff there exists a finite {X .}-time T with E ST > - 00.
I I. If {X., n ~ I}arei.i.d. r.v.swithE XI = Jl > - 00 and S. = D Xi' 1; = inf{n ~ I:
S. > cn}, then E 1; < 00 for c < Jl and E 1; = 00 for c ~ Jl.
12. Let {So = I~ Xi' n ~ I, So = O} be a random walk on the line. A real number c
is said to be recurrent if PUS. - cl < e, i.o.} = I for all e > 0 and is called possible
if for every e > 0 there is an integer n such that P {IS. - c 1 < e} > O. Prove that if
c is possible and b is recurrent, then b - c is recurrent. Since every recurrent value
is clearly possible, the set Q of recurrent values is an additive group. Show that Q is
closed. Thus, if Q is nonempty, Q = (- 00, 00) or Q = Inc: c ¥- 0, n = 0, ± I,
± 2, ...} or Q = {O} (the latter only if X is degenerate at 0).
References
J. H. Abbott and Y. S. Chow, "Some necessary conditions for a.s. convergence of
sums of independent r. v.s.," Bull. Institute Math. Academia Sinica 1 (1973), 1-7.
L. E. Baum and M. Katz, "Convergence rates in the law of large numbers," Trans.
Amer. Math. Soc. 120 (1965), 108-123.
B. Brown, "Moments of a stopping rule related to the central limit theorem," Ann.
Math. Stat. 40 (1969), 1236-1249.
D. L. Burkholder, "Independent sequences with the Stein property," Ann. Math.
Stat. 39 (1968),1282-1288.
Y. S. Chow, "Local convergence of martingales and the law of large numbers," Ann.
Math. Stat. 36 (1965),552-558.
Y. S. Chow, "Delayed sums and Borel summability of independent, identically distrib-
uted random variables," Bull. Inst. Math., Academia Sinica 1 (1973), 207-220.
Y. S. Chow and H. Robbins, "On the asymptotic theory of fixed-width sequential
confidence intervals for the mean." Ann. Math. Stat. 36 (1965),457-462.
Y. S. Chow and H. Teicher, "Almost certain summability of i.i.d. random variables,"
Ann. Math. Stat. 42 (1971), 401-404.
Y. S. Chow, H. Robbins, and D. Siegmund, Great Expectations: The Theory ofOptimal
Stopping, Houghton Mifflin, Boston, 1972.
Y. S. Chow, H. Robbins, and H. Teicher, "Moments of randomly stopped sums,"
Ann. Math. Stat. 36 (1965),789-799.
K. L. Chung, "Note on some strong laws of large numbers," Amer. Jour. Math. 69
(1947),189-192.
K. L. Chung, A Course in Probability Theory, Harcourt Brace, New York, 1968; 2nded.,
Academic Press, New York, 1974.
K. L. Chung and W. H. J. Fuchs, "On the distribution of values of sums of random
variables," Mem. Amer. Math. Soc. 6 (1951).
K. L. Chung and D. Ornstein, "On the recurrence of sums of random variables," Bull.
Amer. Math. Soc. 68 (1962),30-32.
164 5 Sums oflndependent Random Variables
C. Derman and H. Robbins, "The SLLN when the first moment does not exist,"
Proc. Nat. Acad. Sci. U.S.A. 41 (1955),586-587.
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
K. B. Erickson, "The SLLN when the mean is undefined," Trans. Amer. Math. Soc.
185 (1973),371-381.
W. Feller, "Uber das Gesetz der grossen Zahlen," Acta Univ. Szeged., Sect. Sci. Math.
8 (1937),191-201.
W. Feller, "A limit theorem for random variables with infinite moments," Amer. Jour.
Math. 68 (1946),257-262.
W. Feller, An Introduction to Probability Theory and its applications, Vol. 2, Wiley,
New York, 1966.
R. Gundy and D. Siegmund, "On a stopping rule and the central limit theorem," Ann.
Math. Stat. 38 (1967),1915-1917.
C. C. Heyde, "Some renewal theorems with applications to a first passage problem,"
Ann. Math. Stat. 37 (1966), 699-710.
T. Kawata, Fourier Analysis in Probability Theory, Academic Press, New York, 1972.
H. Kesten, "The limit points ofa random walk," Ann. Math. Stat. 41 (1970),1173-1205.
A. Khintchine and A. Kolmogorov, "Uber Konvergenz von Reichen, derem Glieder
durch den Zufall bestimmt werden," Rec. Math. (Mat. Sbornik) 32 (1924),668-677.
A. Kolmogorov, "Uber die Summen durch den Zufall bestimmer unabbiingiger
Grossen," Math. Ann. 99 (1928),309-319; 102 (1930), 484-488.
M. J. Klass, "Properties of optimal extended-valued stopping rules," Ann. Prob. I
(1973),719-757.
M. Klass and H. Teicher, "Iterated logarithm laws for random variables barely with or
without finite mean," Ann. Prob. 5 (1977), 861-874.
K. Knopp, Theory and Application of Infinite Series, Stechert-Hafner, New York, 1928.
P. Levy, Theorie de faddition des variables aleatoires, Gauthier-Villars, Paris, 1937;
2nd ed., 1954.
M. Loeve, "On almost sure convergence," Proc. Second Berkeley Symp. Math. Stat.
Prob., pp. 279-303, Univ. of California Press, 1951.
M. Loeve, Probability Theory, 3rd ed., Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
J. Marcinkiewici! and A. Zygmund, "Sur les fonctions independantes," Fund. Math.
29 (1937),60-90.
P. Revesz, The Laws of Large Numbers, Academic Press, New York, 1968.
H. Robbins and E. Samuel, "An extension of a lemma of Wald," J. Appl. Prob. 3
(1966),272-273.
F. Spitzer, "A combinatorial lemma and its applications to probability theory," Trans.
Amer. Math. Soc. 82 (1956),323-339.
C. J. Stone, "The growth of a recurrent random walk," Ann. Math. Stat. 37 (1966),
1040-1041.
H. Teicher, "Almost certain convergence in double arrays," Z. Wahr. verw. Gebiete 69
(1985),331-345.
A. Wald, "On cumulative sums of random variables," Ann. Math. Stat. 15 (1944),
283-296.
6
Measure Extensions,
Lebesgue-Stieltjes Measure,
Kolmogorov Consistency Theorem
165
166 6 Measure Extensions
If S E Y, then by (1), v{S} :5: J.l{S} + J.l{0} + J.l{0} + ... = J.l{S}, while
if Uf Sn ::J S, where Sn E Y, n ~ I, then, since J.l is a measure on Y, by
Corollary 1.5.1 J.l{S} = J.l{Uf SnS} :5: If
J.l{Sn}, and so via (1), J.l{S} :5: v{S}.
(ii) v is subadditive, that is, for A o c Uf An C n
00
In proving (3) it may be supposed that v{A n } < 00, n ~ I. For any
e > 0 and n ~ 1, choose Sn.m E Y, m ~ I, with
00
(iv) vii is an algebra and for every DEvil and A c:: n and finite partition
{D n , n = 1,2, ... ,m} of D in vii,
m
v{A . D} = L v{A . Dn }. (8)
j
Hence, by (3)
when these limits exist. Iff(t) = f(t-), thenfis said to be continuous from
the left or left continuous at t E R. A function which is left continuous at all
points of some set T c R is called left continuous on T and when T = R,
simply left continuous. Similarly, f(t) = f(t+) defines right continuity at
t and the analogous terms are employed.
If f is a function with f(t - ) existing for each t E R, then g(t) = f(t - ) is
left continuous (Exercise 3). In particular, iff is a monotone function on R,
f(t-) exists and g(t) = f(t-) is left continuous. Since the set of discon-
tinuities of a monotone function is countable and two left-continuous
functions are identical if they coincide except for a countable set, every
monotone function m on R defines a unique left-continuous function F = F m
via F(t) = m(t - ).
Lemma 1. Let F be a nondecreasing, left continuous function on R = [- 00, 00]
withF(-oo) = F(-oo+)andIF(t)1 < oo,ltl < 00. If
J1.{[a, b)} = F(b) - F(a), - 00 :5 a :5 b :5 00,
J1.({00}) = 0 = J1.{0}, J1.{R} = F(oo) - F(-oo), (13)
then J1. is a measure on the semi-algebra Y.
6.1 Measure Extensions, Lebesgue-Stieltjes Measure 169
(14)
In proving the reverse inequality, since jJ.{R} = jJ.{[ -00, oo)}, it may be
supposed that S = [a, b), -00 ~ a < b ~ 00. Moreover, in view of the
hypothesized equality F( -00) = F( -00+) and left continuity, it suffices to
establish that
Since 00 ¢ [a, b), necessarily Sn #- {oo}, n ~ 1 whence Sn' [c, d) = [an, bn),
where c ~ an ~ ba ~ d, n ~ 1.
For any e > 0, set I n = (an - (jn, bn) where (jn > 0 satisfies F(a n) - F(a n -
n
(jn) < e/2 , n ~ 1 via left continuity. Then [c, d] =
whence by the Heine-Borel theorem [c, d) c J nk c U7
[an, bn) Cur
U7
In
[a nk - (jnk' bn) for
ur
some finite set of integers n1 , ••• , nm• By Corollary 1.5.1
m m
jJ.{[c, d)} ~ L (F(bn) - F(a nk - (jn) ~ L (F(bn) - F(a n) + e2- nk )
1 1
m
Lt (F(bn.) - + e ~ L jJ.{Sd + e.
00
~ F(a n.)
1
Moreover, .IIm =:> the a-algebra f!J ofall Borel sets of Rand Vm is unique on f!J.
(16). Then
= v*{{ -oo}} = v.. {{oo}} = v.. {{ -oo}}
v*{{oo}}
v*{[a, b)} = m(b-) - m(a-) = v..{[a, b)}, -00 < a:S; b < 00,
v*{[ -00, b)} = v*{( -00, b)} = lim v*{[a, b)} = lim v.. {[a, b)}
Q--oo
EXERCISES 6.1
1. Let J.l be a measure on a a-algebra .91 and defined = {A AN: A E.9I, NcB E.9I,
J.l{B} = O}. Then .91 is a a-algebra, .fj => .91, and ji is a complete measure on .91,
where ji{A AN} = J.l{A} for A A NEd.
2. Prove that the extension v (as defined in Theorem I) of a measure J.l on .Y' to the
a-algebra .I( of all v-measurable sets is complete.
3. Let f map R = [ - 00, 00] into R with f(t - ) existing for every t E R. Then g(t) =
f(t - ), t E R, is left continuous, i.e., g(t) = g(t - ), t E R.
5. If m(t) is as in Exercise 4 and G(t) = m(t +), - 00 :;;; t :;;; 00, then the nondecreasing,
right continuous function G determines a measure ji on the semi-algebra ii' of all
6.2 Integration in a Measure Space 171
finite or infinite intervals ofthe form (a, b], a < b, and also 0, R = [ - 00, 00], { - oo}
via jl{(a, b]} = G(b) - G(a), jl{{ -oon = Jl{0} = 0, jl{R} = G(oo) - G( -(0).
6. There is a I-I correspondence between dJ.s and probability measures on the Borel
sets of the line.
7. If (R,.n, v) is the Lebesgue measure space (of R), n = [0, I], .91 = n·.n, and
P = vi...., then (n, .91, P) is a probability space.
8. If X is a T.v. with dJ. F, then P{X+ = O} = F(O+), P{X+ < x} = F(x), x> 0, and
P{X- = O} = I - F(O),P{X- < x} = I - F(-x+),x > O.FindthedJ.ofIXI.
9. Give an example to show that the uniqueness assertion of Theorem I is not true
withouttherestriction ofa-finiteness on Y. Hint: Taken = {r: rrational,O::; r < I},
Y = {n· [a, b): 0 ::; a ::; b ::; I}, Jl(0) = 0 and Jl(A) = 00 if 0 ¥ A E Y, v(A) =
number of elements in A for A E a(Y), v*(A) = 2v(A) for A E a(Y).
10. If (R, .n, v) is the Lebesgue measure space of the real line and E is a Lebesgue
measurable set, so is E + x = {y + x: y E E} and, moreover, v{E + x} = v{E} for
every x E ( - 00, (0).
II. For any real x, y consider the equivalence relation x - y if x - y = r = rational.
Let the subset E of [0, I) contain exactly one point of each equivalence class. Then E
is a non-Lebesgue-measurable set. Hint: (i) if x E (0, I), thenx E E + r for some r in
(-I, 1). (ii)(E + r) n (E + s) = 0 for distinct rationals r, s. Thus, if E is Lebesgue
measurable so is S = U,eF(E + r), where F is the set of rationals in (-1, 1) via
Exercise 10. Then S c ( -1,2), whence 3 ~ v{S} = If
v{E}, implying v{S} =
I v{E} = O. However, by (i), [0, I) c U(E + r) = S, so that v{S} ~ 1, a contradic-
tion.
EX = Lx dJl.
E X = hm
•
L ni JI. {in2
00
n-ooi=12
< X ::;
i+ I}
-n-
2
. (I)
172 6 Measure Extensions
As in Chapter 4, it is easy to verify that the limit in (1) always exists and
that:
o :s;; E X :s;; 00 if X ~ 0, a.e., E I = fJ.(S);
IXI < oo,a.e., ifEIXI < 00; X=O,a.e., iffEIXI=O;
EX = E·Y if X = Y, a.e., and IE X I :s;; 00.
From (I) and (2) it is readily seen that if (S, L i , J1J, i = 1,2, are measure
spaces with fJ.l = fJ.2 b:. and X is a L1-measurable function, then
in the sense that if one of the integrals exists, so does the other and the two
are equal.
Associated with any measure space (S, L, fJ.) are the spaces !l' p =
!l'iS, L, fJ.), P > 0, of all measurable functions X for which E I X IP < 00.
For any measurable function X, and especially when X E !l' P' the !l'p-norm
of X is defined by
(4)
Let {X n , n ~ I} be a sequence of measurable functions on the measure
space (S, L, fJ.). Jf fJ.{lim X n -:f. limn X} = 0, set X = lim X n whence lim X n =
X, a.e., denoted by X n ~ X. If X is finite a.e., X n will be said to converge a.e.,
denoted by X n ~ X, finite. Alternatively, if X n is finite a.e., n ~ 1, then X n
converges in measure to X, written X n -4 X, iflimfJ.{IX n - XI > t:} = o for
all t: > O. These are obvious analogues of a.c. convergence and convergence
in probability on a probability space, but the correspondence is not without
limitation (see Exercise 6.2.4). In case IIX n - Xll p = 0(1), the earlier notation
:I!
X n ~ X will be employed.
In dealing with the basic properties of the integral in Chapter 4, the proof
of Theorem 4.1.1(i) utilized the fact that for any nonnegative integrable LV.
X on a probability space (0, JF, P)
EX = lIm
n-oo
•
;=0
L -+-
00 1
n P
i 2
{i
-n < X < --
2 - 2
n .
i+ I} (5)
6.2 Integration in a Measure Space 173
whence
2-(2n-l)p.{X > r(2n-1)} ~ 2. 2- 2n p.{X > r 2n } = 0(1).
Consequently, the difference of the right and left sides of (6), viz.,
limn_ex> 2- np.{X > r n}, is zero.
It may be noted via (6) or (1) that if X is a nonnegative, integrable function
on (S,~, p.), then p. is IT-finite on S{X > OJ.
E(X + Y) = EX + E Y.
vi. If X ~ Y, a.e., IE X I ~ 00, lEY I ~ 00, then E X
~ E y.
vii. IfE X, E Y, a EX + bEY are defined for finite a, b, then
E(aX + bY) = aEX + bE Y. (9)
viii. X is integrable iff IX I is integrable and if X ~ 0, Y ~ 0, a.e., then
E(X + Y)p ~ 2P(EXP + E ¥P),p > O.
174 6 Measure Extensions
PROOF. The argument follows that of Theorem 4.1.1 with two modifications.
Firstly, in proving (i) write Ini = (iI2 n, (i + 1)/2n], P.ni = p.{ X' E Ini }, and replace
(13) of Theorem 4.1.1 by
> a,
implying EX :s; Iim,,-oo E X n :s; ITiii n _ oo E X n :s; E X, that is, IXnl :s; Y E!l' I
and X n ~X imply
EX = limE X n • (13)
r
J[-CO,CO]
X dm = f
(-co,co)
X dm = r
J[-CO,CO)
X dm = f(-co,co)
X dm.
However, if - 00 < a < b< 00, S: X dm is not, in general, well defined since
r
J[a,bl
X dm = [m(a+) - m(a- )]X(a)
via additivity ofthe indefinite integral and Exercise 6.1.4. On the other hand, if
a and b are continuity points of m, i.e., m(b +) = m(b - ) and m(a +) = m(a - ),
S:
then X dm may be interpreted as the common value
r
~~~
X dm = f~~ X dm = f~~ X dm = r
~~~
X dm.
Thus, only when a and b are continuity points (including ± (0) of m will
the notationf: X dm be utilized for a Lebesgue-Stieltjes integral.
176 6 Measure Extensions
fX dm = - fx dm
(15)
in the sense that if either integral exists, so does the other and the two are equal.
PROOF. Note that g(X) is a real L-measurable function. Since g = g+ - g-
and g± are measurable functions on (T, d), it suffices to prove that every
nonnegative d-measurable function is in~, where
~ = {g: g is a nonnegative d-measurable function for which (15) holds}.
Now by monotone convergence and linearity, ~ is a monotone system (see
Section 1.4), and if g = I A for any A Ed, then
1
f/ dv = {dv = J1.{X- (A)} = f/A(X)dJ1..
in the sense that if either of the integrals exists, so does the other and the two
are equal.
PROOF. Let Vx be defined on !!4' by (16). Then, as noted earlier, Vx is the re-
striction to !!4 of the Lebesgue-Stieltjes measure determined by F x. By
Theorem 3
again in the aforementioned sense. Thus (17) and hence the corollary is
~~. 0
f
b n
VI.I f f(t)dm(t) I :s; fI f (t) Idm(t) :s; [m(b) - m(a)] a~~~xb I f(t) I,
Vl1. f f(t)dm(t) + f m(t)df(t) = m(b)f(b) - m(a)f(a),
V111. f m(t)df(t) = m(a) f df(t) + m(b) f df(t) for some ~ in [a, b],
IX. f f(t)m(t)dt = m(a) ff(t)dt + m(b) f f(t)dt for some ~ in [a, b],
J:
The Riemann-Stieltjes integral f(t)dm(t) has been defined for finite
closed intervals [a, b] and exists iff is continuous and m is finite and non-
decreasing on [a, b]. Moreover, iff is continuous and m is finite and non-
decreasing on [a, (0) for some finite constant a, the definition may be extended
via
provided the limit on the right exists. Analogously, for any finite constant b,
if f(t) is continuous and m(t) is finite and nondecreasing on ( - 00, b] (resp.
on ( - 00, (0» define
J
b
_ 00 f(t)dm(t) = a~i~oo L b
f(t)dm(t),
!~~ f
(23)
{oooo f(t)dm(t) = f(t)dm(t)
Q-+ - 00
provided the limits on the right exist (independently of the manner in which
b ---+ 00, a ---+ - (0). If a = - 00 or b = 00 or both, f(t)dm(t) is frequently J:
alluded to as an improper Riemann-Stieltjes integral.
The relationship between the Riemann-Stieltjes integral f(t)dm(t) J:
and the Lebesgue-Stieltjes integral Jla,b)f(t)dm(t) is embodied in
ila,b)
f(t)dm(t) = fb f(t)dm(t),
a
-oo<a<b<oo; (24)
moreover, if I J(- 00. oo)f(t)dm(t) I ::; 00, then the Riemann-Stieltjes integral
J~ 00 f(t)dm(t) exists and
t, oo/(t)dm(t) = 1 00
2(b - a)
a = t\» < t\n) < ... < t~n) = b, max (t~n) - t~n~ 1) < .
1 skSn n
180 6 Measure Extensions
On [a, b) define
n
fn(t) = L f(tL~dI[I~"2"I~n)(t),
k=!
i[a. b)
f(t)dm(t) = i [a. b)
f dv = lim
n
i[a, b)
fn dv
= lim E L f(tL~dI[I~n~l,Il:'»
f
n k=!
i eo)
[a.
f±(t)dm(t) =
b~eo
lim
m(b)=m(b-)
i [a,b)
f±(t)dm(t) = lim
b-eo
m(b)=m(b-)
fb f±(t)dm(t)
a
= leo f±(t)dm(t),
and the comparable statement with [a, (0) replaced by (- 00, b) follows
analogously. The remaining portion of (25) obtains by letting a -+ - 00
subject to m(a) = m(a - ). 0
whenever the integrals on the right are defined. In the special case where
m = I - F with F a d.f., the latter may also be expressed as
It may be noted via the definition that whenever J~f(t)dm(t) exists for
- 00 < a < b < 00, so does J=b f( -t)dm( -t) and the two are equal.
The absolute moment EIXI', r > 0 always exists for any r.v. X, and its
6.2 Integration in a Measure Space 181
Thus, in proving (i) it may be supposed that Elm(X)1 < 00, and so, since
= i [Itl2C)
Im(t)ldF(t) = 0(1)
PROOF. The first statement is an immediate consequence of(iii) and (x) of(21),
while the second follows similarly from (ii). 0
EXERCISES 6.2
l. If 111,112 are measures on (n, §) and S X d(1l1 + 112) exists so does S X dill' i = 1,2,
and S X d(llt + J,l2) = S X dJ,lI + S X d1l 2 ·
2. The integral of a nonnegative measurable function over a set of measure zero has the
value 0; also, if SA 9 dJ,l = 0 for every measurable set A, then 9 = 0, a.e.
6.2 Integration in a Measure Space 183
3. If S is the set of positive integers, ~ is the class of all subsets of S, and J.l(A) = number
of integers in A E ~, then (S, ~, J.l) is a non finite measure space and convergence in
measure is equivalent to uniform convergence everywhere.
4. Demonstrate in a nonfinite measure space (S, ~, J.l) that X. ~ X does not neces-
sarily imply X. 4 X. Hint: Utilize Exercise 3. If X. 4 X, does X. Y 4 XY?
~
5. If X. E .PiS, ~, J.l) and X. -4 X for some p > 0, then X E.P p and EI X.I P --+ EIXI P•
6. Iff is a finite, nondecreasing function and m is a continuous nondecreasing function
on [a, b], where - 00 < a < b < 00, then
f't
e
dF,xl(t) = c P{IXI ~ c} + f'e
P{lXI ~ t}dt.
(i) Show that a sequence ofr.v.'s {X., n ~ I} is u.i. iffsup.~ 1 S:'
P{IX.' ~ t} dt --+ 0
as c --+ 00.
ALV. X is said to be stochastically larger thanaLv. YifP{X ~ x} ~ P{Y ~ x} for
all x. (ii) If the LV.S X., n ~ I, are u.i. and IX.I is stochastically larger than IY.1.
n ~ I, then {y", n ~ l} is u.i.
9. If {X.,n ~ I} are.P I LV.S with a common distribution, then Emax1,;i,;.IXd =
o(n). Hint: Use Exercise 8 to establish u.i.
10. Show that the analogue of (2IXiv) for Lebesgue-Stiettjes integrals is not true in
general. Construct an example for which the Riemann-Stieltjes integral over a
finite interval [a, b] fails to exist.
II. Establish that g(t) = (sin t)/t is Riemann but not Lebesgue integrable over ( - 00, 00)
and find a function g(t) which is Lebesgue integrable but not Riemann integrable.
12. Let S= {t,2}, ~ = {{I}, {2},0,S} and J.l = counting measure. If Xes) =1,
Jensen's inequality fails for the convex function XZ.
13. Let S = {I, 2, ... }, ~ = {A: A c S}, J.l = counting measure on ~. If X.(s) =
n- 11(1 ,;s,;.I' then X. ~ 0 and EX. =
I despite the fact that for J.l{A} < l, E X.I A
= O. Thus, Theorem 4.2.3(i) may fail in a a-finite measure space.
14. If f, m are finite and nondecreasing on (- 00, 00) with f continuous, prove that
SID.b1f(t)dm(t) + SID.b] m(t)df(t) = f(b)m(b+) - f(a)m(a-).
15. Let P. E (0, I), q. = I - P., where np. --+ A. E (0, 00). Let J.l be counting measure on
the class of all subsets of!l = {t, 2, ...}. If XU) = ;Je-),/j! and X.U) = (j)~q:- i,
IJ !£p
provethatX. -+ XandX.~ X,p ~ 1. Hint: Apply Example 1 or Example 2.1.1.
16. (Erickson) In Example 5.4.1 thefunction a(x) = x/J~ [I - F(y)]dy was encountered,
where F is the dJ. of a nonnegative LV. X. (i) Show that a(x) is nondecreasing. (ii)
Prove that EX = 00 implies E a(X) = 00. Hint: E a(X) < 00 entails a(x) =
0«(1 - F(xW 1) and hence E X/g y dF(y) < 00, contradicting the Abel-Dini
theorem.
184 6 Measure Extensions
are $' 2- and $' I-measurable respectively according to Theorem 1.4.2. Thus,
JJ.z{A(l)(wI)} and JJ.l {A(2)(wz)} are well-defined real functions, the first on n l
and the second on nz . For notational simplicity these will be denoted by
JJ.Z{A(l)} and JJ.l{A(Z)} respectively.
Now if JJ.I and JJ.z are finite measures,
sf = {A E $': JJ.3-j{A(j)} is $'i-measurable, i = 1,2}
is a A.-class containing all rectangles with measurable sides, whence sf :::l $',
that is, JJ.3-d A(i)} is ~-measurable i = 1, 2; this carries over to the case
where JJ.i is a-finite, i = 1,2, since if n j = U;;"
I Bi,i' where Bij are disjoint
(2)
6.3 Product Measure, Fubini's Theorem, n-Dimensional Lebesgue-Stieltjes Measure 185
PROOF. For A E fF, define J.l by the first equality in (2). Then J.l ~ 0 and
J.l{0} = O. Since (Uf An)(Z) = Uf A~) for every sequence {An} of fF sets,
and disjointness of {An} entails that of {A~Z)}, the set function 11 so defined
is a measure on fF, whence (il, fF, J.l) is a measure space. Moreover, if
A = B I X B z , where B i E .-7;, i = 1,2, then
1l*{A} = i
0,
Ilz{A(l)}dll j, A E g;,
The measure J.l defined in (2) is called the product measure of the a-finite
measures J.lj and J.lz and is denoted by J.l1 x 117.' The a-finite measure space
(il j x il z , fF j X fFz, J.lI x J.lz) is referred to as the product measure space
(of the a-finite measure spaces (ili' fF i , J.la, i = 1,2).
In situations where more than one measure is floating around, almost
everywhere statements must be qualified. Thus, a.e. [J.lj] abbreviates the
statement" except for a set of J.ll-measure zero."
Corollary I. If (il, fF, J.l) is the product measure space of the a-finite measure
spaces (ili' fF i , J.la, i = 1,2, and J.l{A} = Ofor some A E fF, then J.lj {A(Z)(wz)}
= 0, a.e. [J.lz], and J.lz{A(l)(w j)} = 0, a.e. [J.lI].
Theorem 2 (Fubini). If (0, :IF, J.L) is the product measure space of the a-finite
measure spaces (0;, :IF i, J.LJ, i = 1, 2, and X is an :IF-measurablefunction whose
integral exists, then
°
of:IF by Theorem 1. Since ytJ is a monotone system, ytJ contains all nonnegative
:IF -measurable functions by Theorem 1.4.3. Moreover, if X 2: is integrable,
it is clear from (4) that so are S02 X(Wl> w 2)dJ.L2 and SOl X(Wl, w2)dJ.Ll for
almost all WI and W2 respectively. 0
n J.Lj{BJ.
n
Then J.L is called the product measure and denoted by J.Ll x ... x J.Ln or
X7; I J.Li· Associativity of ordinary multiplication together with Theorem 6.1.1
guarantees that X7; I J.Lj = (X~ 1 J.LJ x (X~+ I J.LJ for any integer m in [1, n).
The measure space (0, :IF, J.L) = (X~ OJ, X7; X7;
I :lF j , I J.LJ is alluded to as
the (a-finite) product measure space of (OJ, :lFj, J.Li), 1 ~ i ~ n, and is denoted
by X7; 1 (Oi, :lF j , J.LJ The extension of Fubini's theorem is immediate, and
(5, i)
lim F(x ... , x j _" Yj, x j + I,· .. , x n) = F(x . .. , x j , . .. , x n),
Xj>Yj-Xj " "
1 :s; j :s; n, (5, ii)
n
and
n
P{S} = (6)
n
o if S = X Sj
j; 1
with some Sj =0 or {oo}.
188 6 Measure Extensions
where F satisfies (5, i)-(5, iii) and is finite on (-co, co)". As in Lemma 6.1.1,
P is a measure on the semi-algebra Y", and so by Theorem 6.1.1, P can be
extended to .A", the class of all P-measurable sets, the extension being unique
on the class (fI" of n-dimensional Borel sets. The measure P and the measure
space (R", --It", P) are called respectively the n-dimensional Lebesgue-
Stieltjes measure and the n-dimensional Lebesgue-Stieltjes measure space
determined by F. The same terms respectively are used for the restriction of P
to f!J" and (R", fJI", P). If F is a dJ., that is, satisfies (5, iv) also, then (R", fJ4", P)
is a probability space.
If X" ... , X" are LV.S on a probability space (0, !IF, P), their joint dJ.
F x, ..... xn(Xt, ... , x") = P{X I < XI' ... , X" < x"}
is readily verified to be a dJ. on R" in the sense of (5). Conversely, given any
n-dimensional dJ. F as in (5, i)-(5, iv), there always exist n r.v.s X I"'" X"
on some probability space (0, !IF, P) whose joint dJ. is the preordained F.
It suffices to choose 0 = R",!IF = (fI", Xi(w) = ith coordinate of w, I ~ i ~ n,
and to define P on Y" via (6). According to the prior discussion, P is uniquely
determined on the Lebesgue-Stieltjes measure space (R", (fI", P). Moreover,
(8)
6.3 Product Measure, Fubini's Theorem, n-Dimensional Lebesgue-Stieltjes Measure 189
(9)
P{X I + X 2 < x} =
Jor I[x,+x,<X](w) dP(w) = J[z+y<X]
r d(F1 x F2 )(z, y)
= f:ooFI(X - y)dFiy),
and so the dJ. of XI + X 2 is F I * F2 • 0
EXAMPLE 1. Let (r, §) and (A, <§) be measurable spaces with v a measure on
<§ and JL" {A} a function on § x A such that JL" {.} is a measure on § for
t
almost all A E A and moreover JL" {A} is <§-measurable for every A E §. Define
PROOF. Let A' = {k JL" is a measure on §}. Then JL{A - A'} = 0 whence
JL{A} = fA'JL,,{A}dv(A). Via the monotone convergence theorem, JL is a mea-
sure on §. As usual, in verifying (12), it suffices to consider non-negative f.
fr
Let Yl' = {f ~ 0: f is §-measurable, f(y)dJL,,(Y) is <§-measurable and (12)
holds}. Then IA E Yl' for all A E § by definition and moreover by Theorem
6.2.1 Yl' is a monotone system. Thus, by Theorem 1.4.3, Yl' contains all
non-negative § -measurable functions. D
EXERCISES 6.3
1. Let (OJ, ff j , P j ) be a probability space, where OJ = (- 00, (0) and ff j = {B: B is a
Borel subset of(-oo,oo)},i= 1,2.U(Q,ff,P) = (0. x 0z,ff. x ffz,P. x P z)
and if Xlw) = Wj, i = 1,2, for W = (w., wz)e 0, then X. and X z are independent
LV.S on (n, ff, P).
2. un l= Oz = [0, 1], ff. = ff z = [0, 1] .~, fl. = Lebesgue measure, and fl2{A} =
number of points in A, then (OJ, ffi> flJ, i = 1,2, are measure spaces. The set
A = {(w.,wz):w. = wz}eff l x ffz,but
whence Theorem 1 as well as the uniqueness part of Theorem 6.1.1 is invalid if (1-
finiteness is omitted.
3. For any (not necessarily (1-finite) measure spaces (OJ, ff j , flj), i = 1,2, and any
integrable function X on the product measure space (n, ff, fl),
Hint: Consider X ~ 0 and utilize the remark just prior to Theorem 6.2.1.
6.4 Infinite-Dimensional Product Measure Space, Kolmogorov Consistency Theorem 191
sums of the double series Lj au are simply the elements of the first column. What are
the column sums? Taking III and 112 as counting measure, does this example con-
tradict Fubini's theorem?
5. Show that g(wl' W2) = e-w,w, - 2e- 2w ,w, is Lebesgue integrable (i) over Q 1 =
[I, 00) for each W 2 and (ii) over Q 2 = (0, IJ for each WI' but that Fubini's theorem
fails. Why?
6. An alternative construction of a LV. X on a probability space with a preassigned dJ.
F is to take Q = [0, I J, ff = Borel subsets of [0, I J, P = Lebesgue measure on
[0, IJ, and X(w) = F-I(w), where F-I(w) = sup{x: F(x) < w}.
7. Prove that the random vectors X = (X I' ... , Xm)and f = (fl , ... , fn)on (Q, ff, P)
are independent of one another iff their joint dJ. F x, y = F x . F y and conclude that X
and f independent entails (R m+n, gr+n, vx, y) = (R m, [JIm, Vx) X (Rn, [JIn, Vy), where
vx, Vy, and VX,y are the Lebesgue-Stieltjes measures determined by F x , F y, F x . y
respecti vely,
8. Random variables XI' X 2 have a bivariate Poisson distribution if
min(j. k) ai aj - iak - i
P{X = ' X = k} = e-(a,+a,+a12) " 12 I 2
I j, 2 L.
i=O I. ")'(k
"("j _ I . _ I')'.
for any pair of nonnegative integers (j, k), where aI' a2' al2 are nonnegative param-
eters. Define a probability space and LV.S X I' X 2 on it whose joint distribution is
bivariate Poisson and show that X j is a Poisson LV. with mean aj + a 12 . Prove that
the correlation p(X I> X 2) ~ 0 and that X I and X 2 are independent iff a l 2 = O.
9. Random variables XI" .. , X k have a multinomial distribution if P{X i = Xi' 1 ::;;
i ::;; k} = n! n~=1 (PN(Xi!» for any nonnegative integers Xi' I ::;; i ::;; k, with
L~ Xi = n and zero otherwise. Here, n is a positive integer and L~ pj = I, Pi> 0,
I ::;; i ::;; k. Prove that if {A 1 ::;; i ::;; k} is a partition orQ in ff with Pi = P{A;} and
j ,
a~ = Xaj,
j= I
00 n 00
i=n+ I
= r (P
Jn~
Pm){Bm(w~)}d(PI x ... x P
n + 1 X ... X n)
= r ... r p(n){A(w~)}dPl(wd"'dPn(wn)'
Jnn Jn,
(2)
6.4 Infinite-Dimensional Product Measure Space, KolmogoroY Consistency Theorem 193
p(m){An(wT, ... , w:)} ~ 2Em ' m = 1,2, ... ;'n = 1,2,.... (4)
.. 0, W m Of *
1 (w t , 000'
*) J
W m 'F
Bmo
The proof is based on the fact that if t§n = {A: A = Bn x Q~} is the class of
cylinders with n-dimensional bases B n E Xi= I fF j and
t\{A} = (PI x 000 x Pn){B n},
194 6 Measure Extensions
then Pn is a probability measure on the a-algebra t§n with Pn = Pn+ II~. and,
moreover, if P{A} = limn Pn{A} for A E t§ = Uf t§n, then P is a-additive
on the algebra t§, whence there is a unique extension to a(t§) = ~i' Xf
The following question then poses itself: If (O,~, PJ, i ~ 1, is a sequence
of probability spaces with ~ c ~+1 and Pi = Pi+d~;, is P{A} = limnPn{A},
A E t§ = Uf~, necessarily a-additive on the algebra '§? The answer is,
in general, negative; see Example 3. However, if 0 = ROO and g;, is the
class of cylinders of (ROO, &1 00 ) with n-dimensional bases, the answer becomes
affirmative.
and let ijn he the a-algebra ofcylinders in (ROO, &1 00 ) with n-dimensional Borel
bases. Ift§ = Uf ijn andfor each A = An X X:'+ I R with An E :!in,
PROOF. In view of Theorem 6.1.1, it suffices to prove that P is well defined and
a probability measure on t§. The former is an immediate consequence of (5)
and, clearly, P is nonnegative and additive on t§ with P{R oo } = l.
Let ~ n and ~ n denote the classes of all sets of the form J I X •.. x J nand
J I X ... x I n x R x R x ... respectively, where J i signifies an interval
of R, 1 :5: i :5: n, i.e., J i = [ai' bJ, [ai' bi), (ai' bi ), or (ai' bJ for - 00 :5: aj :5:
bi :5: 00. Then the classes J(' nand Ji" nof all finite unions of sets of ~ n and ~ n
respectively are algebras. If Ji" = Uf Ji"n' then Ji" is a subalgebra of'S,
whence P is defined and additive on Ji".
To check a-additivity on Ji", let {An, n ~ I} be a decreasing sequence of
sets of Ji" with infn P{An} = E> 0, whence An = Am X R x R x ... for .
some Am. E J(' m.' and mn + 1 > mn , n ~ l. Since P n is a probability measure on
J('n, every interval J I X ... x I n of ~n contains a closed subinterval whose
Pn-measure is arbitrarily close to that of J I X ... x J n' Thus, there is a
closed set Bn (which is a finite union of closed intervals) with
(7)
Let En = Bn X R x R···, whence (7) holds with An, En' P replacing Am.,
Bn , Pm. respectively. Consequently, if en = 8 1 . E2 ••• 8n ,
6.4 Infinite-Dimensional Product Measure Space, Kolmogorov Consistency Theorem 195
whence
l: l:
7<
P{L n } > P{A n }
-
- 2 ~ 2
and Cn -# 0. Let wIn) = (w\n), wT 1, ••• ) E En. Since En!' necessarily w(n+ p) E
Cnc fin' implying (w\n+p), ... ,w~.7P)EBn, p = 0, 1, .... Choose a sub-
sequence {n a } of the positive integers for which w\n 1k ) - . a limit W\OI (finite or
infinite) as k -. 00. Likewise, there is a subsequence {n2k} of {n1k} with
WT2k) -. W~OI as k -. 00, etc. Then
w(n kk ) = (w\n kk ), w~nkk), ...) -. (w\O), w~o), ..•) == w(O)
and so as k -. 00
n ~ 1.
Fl. .... n(x\' ... ,xn)= lim Fl. .... n+\(x\' ... ,xn+\). (8)
X .. + I - +dJ
Pn+\{~[ai,bJ R} = an+l,i~_ooPn+\{~:[ai,bJ}
X
bn + 1-00
tin
lim
+ 1- - 00
~:·:\F\ ..... n+\ = ~:.bF\, .... n = Pn{.X[aj,b J }
I = 1
bn + t - OO
PROOF. For w = (w\, w 2 , ...) E ROO, define ~{w) = Wj, j ~ I, and Fn , j> by
F2n = P{Y\ < y\, ... , Y2n < Y2n} = F X'.X2 Xn(Y\' Y3,"" Y2n-\)
. FX'.X2 XJY2' Y4,"" Y2n),
F2n - 1 = P{Y1 <YI"'" Y2n - t <Y2n-.}
= F X'.Xl Xn(Y\' Y3,"" Y2n-\)
. FX'.X2 Xn_'(Y2, Y4,"" Y2n-2)
are called the symmetrized XII. Given any sequence of LV.S {X"' n ~ l}, in
subsequent allusion to the symmetrized X"' namely X:, the distinction be-
tween X" and X~ (which are probabilistically indistinguishable) will be
glossed over and X:
will be written
n ~ 1,
where {X~, n ~ I} is independent of {X"' n ~ l} and possesses the same
joint distributions.
Since the {X:, n ~ I} are symmetric LV.S, it is frequently easier (especially
for independent {X"' n ~ I}) to prove a result for {X:} rather than attempt a
direct argument with {X,,} (see Chapter 10).
The following examples complement the three series theorem, the second
furnishing a typical exploitation of symmetrization. For any r.v.s {X"' n ~ I}
and positive constant c, define
" "
s;(c) = L a2(XjIlIXk;c]) = L (E XIIllxjl"C] 2
- E XjIuxjl"c)· (14)
j= 1 j= 1
If s;(c) = 0(1), all c > 0 and Loo= 1 P{lX,,1 > c} < 00, some c > 0 then
S" converges a.c. (17)
PROOF. Define
(18)
P ~ {
hm - Sn > - = I b} (19)
n-oo sn(c) - 2 '
and so, in view of the arbitrariness of b, (19) obtains with b = 00. The an-
alogous statement for the lower limit follows by symmetry.
Apropos of (16), via IXnI ::;; ISn I + ISn _ I I and the Borel-Cantelli theorem
and by symmetry wnn_ oo Sn = 00 = -lim Sn, a.c. The final statement, (17),
follows immediately from the three series theorem (Theorem 5.1.2). 0
If for some c > 0, Loo= I P{ IX nI > c} < 00, and s;(c) --+ 00, then
= ISnl ac.
JIm - - = 0 0 . (21)
n-oo sn(c)
Suppose that for some c > 0,
00
necessarily
2ah ~ afX"_X~)2C + 2e P{lXnl > e},
2
while the prior ones follow upon replacing n in this identity by suitable
subsequences {nJ. 0
Since Levy's inequality acquired a very simple form for symmetric r.v.s.
(Corollary 3.35), symmetrization is especially useful in proving the following
converse to Theorem 5.2.8.
I
00
naP-2P{ISnl ~ ena } < 00, (23)
n=1
then IX > 0 and EIXIP < 00. Moreover, EX = 0 if either IX < 1 or IX = 1 and
(23) holds for all 8 > O.
PROOF. By Theorem 4, there exists a sequence {X', X~, n ~ 1} such that
{X, X n, n ~ 1} and {X', X~, n ~ 1} are i.i.d. Set Y = X - X', Y" = X n - X~,
and 1'" = Ii=1 1], n ~ 1. Then, for some 6 > 0,
00
I nap - 2
P{IT"I ~ 2en
a
} < 00, (24)
n=1
so that, as m ~ 00,
2m 2m
I nap - 2
P{IT"I ~ 28(1 + 2a )ma } ~ I naP- 2 P{IT"I ~ 2ena } = 0(1). (25)
n=m n=m
Now, if 1',,* = maxI s,j S,n 11j1 and Y,,* = max 1 s,j S,n 11]1, for any tJ > 0, via Levy's
inequality,
.-1
= P{I YI ~ 2b} L pj{1 YI < 2b}
j=O
.-1
~ P{IYI ~ 2b} L
j=O
[1- 2P{I1j1 ~ b}J. (26)
while (25) and the initial inequality of (26) ensure that, as m -+ 00,
2m
map-tP{T",* ~ 1'/(1 + 2a)ma}::; L nap - 2 P{1;'* ~ 1'/(1 + 2a)ma} = 0(1).
n=m
for some positive constant C, yielding EI YIP < 00. Hence, via independence
of X and X',
L
00
Thus, E X = 0 if either r.x < 1 or r.x = 1 and (23) holds for all e > O. 0
6.4 Infinite-Dimensional Product Measure Space, Kolmogorov Consistency Theorem 201
whereas
00
EXERCISES 6.4
l. Verify that it is possible to define a sequence of independent LV.S {Xn } with specified
dJ.s F. on a probability space.
2. Let (n, !fI, v) be the probability space consisting of Lebesgue measure on the Borel
subsets of [0, 1]. Each w in [0, I] may be written in binary expansion as w = X 1 X 2 ...
= If 2-·X., where X. = X.(w) = 0 or I and this expansion is unique except for a
set of w of the form m/2n which is countable and hence has probability (Lebesgue
measure) zero. For definiteness regarding such w, only the expansion consisting of
infinitely many ones will be used. Show that for all n ~ I, {w: X .(w) = I} E (JI and that
{X., n ~ I} is a sequence of independent r.v.s. Describe the LV.S Y. where w =
In"":1 ,-'y" and, is an integer> 2.
d T of Borel cylinder sets is an algebra. Define f1IT = u(d T) and let T = [a, b],
-00 ~ a < b ~ 00. Do the sets {m(t): m(t) is bounded on T}, {m(t): m(t) is continu-
ous on T} belong to f1IT? If A* = {m(t): m(ti)e Ri , i = 1,2, ... }, and .91* is the class
of all such set A * (as t j and Rj vary), set f1I* = u(d*). Is f1IT = f1I*?
6. If S. =Li Xi where {X, X., n ;::: I} are i.i.d. then LI n-1P{IS.1 > en} < 00 all e > 0
iff E X = O. Hint: Sufficiency is contained in Theorem 5.2.7. For necessity, define
S: = Li Xr where {X:, n;::: I} is the symmetrized sequence. The hypothesis ensures
convergence ofLI n- I P{IS:I >ne} and hence also LI n-1P{max 1"j".IXrl >ne},
e > O.
Lemma 1. Let (n, ff, /l) be a a-finite measure space and {g" t E T} a nonempty
family of real, measurable functions. Then there exists a countable subset
To c: T such that
sup gl = esup gl'
leTo leT
PROOF. Since /l is a-finite, it suffices to prove the theorem when /l is finite;
moreover, by considering tan - 1 gl if necessary, it may be supposed that
IgIl ~ C < 00 for all t E T. Let 5 signify the class of all countable subsets
J c: T and set
whence (1. is finite. Choose In E 5, n 2 1 for which (1. = sUPn~ I E(SUPleI. gl)
and let To = Uf In· Then To is a countable subset of T and clearly (1. =
E[suPte To gIl The measurable function g = sUPle To gl satisfies (ii) since
otherwise for some t E T necessarily (1. < E max(g, gl) ~ (1.. Obviously, (iii)
holds, and so g = esuPleT gl' 0
6.5 Absolute Continuity of Measures, Distribution Functions 203
Definition. If (n,~, JlJ, i = 1,2, are two measure spaces and JlI {A} = 0
whenever Jl2{A} = 0, then JlI is said to be absolutely continuous with respect to
pz or simply pz-continuous. If, rather JlI {N C } = 0 for some set N E ~ with
Jl2 {N} = 0, then JlI is called pz-singular (or the pair JlI' Jl2 is dubbed singular).
If g is a nonnegative integrable function on (n,~, Jl), the indefinite
integral vg{A} = fA g dJl is absolutely continuous relative to Jl. This is
tantamount to saying that the integral of a nonnegative function g over a set
A of measure zero has the value zero (Exercise 6.2.2). The Radon- Nikodym
theorem asserts under modest assumptions that the indefinite integral v9 is the
prototype of a measure absolutely continuous with respect to Jl. The crucial
step in proving this is
Lemma 2. Let (n, ~,Jl) be a a-finite measure space and va a-finite measure on
~, and let ;t(' denote the family of all measurable functions h ~ 0 satisfying
fA h dJl ~ v{A}, A E~. Then
v{A} = t/J{A} + 19dJl' AE~ (I)
fA
h dJl = f A[hl ~ h21
hi dlJ, + fA[hl < h21
h 2 dlJ, ~ v{A}, A E 17,
and so it may be supposed that hn ~ hn+ I' n ~ 1. Then g = limn hn, whence by
the monotone convergence theorem
C0
1
Corollary 2. Iv{A} I < 00 for all A E .iF iff 9 is p,-integrable and v is a measure iff
9 ~ Oa.e. [jl].
LX dv = Lx ~: djl (4)
206 6 Measure Extensions
where dt signifies Lebesgue measure on (R, Bl), then, as noted in Section 1.6, F
is said to be absolutely continuous and f is called the density function of F. In
particular, if F is the dJ. attached to some r.v. X on a probability space, thenf
is also called the density of X. Clearly, when the f of (5) exists, it is unique to
within sets of Lebesgue measure zero.
i[XEB)
g(X) dP = [g(t)f(t)dt
JB
(6)
and, in particular,
i
[a, b)
f(t)dt = F(b) - F(a) = F(b - ) - F(a - ),
6.5 Absolute Continuity of Measures, Distribution Functions 207
= E /B(X)g(X) = [ g(X)dP,
J[XEBI
and, in particular,
EXERCISES 6.5
I. If t/J and Il are measures such that t/J is both Jl-continuous and Wsingular, then t/J == O.
2. Two measures Jl, v are called equivalent, denoted Jl == v, if each is absolutely con-
tinuous with respect to the other. Verify that this is an equivalence relation. If
(n,~, Jl) is a probability space, Xi is a nonnegative!f I random variable and Jli is the
indefinite integral of Xi> i = 1,2, then, if Jl{[X 1 = 0] ~ [X 2 = OJ} = 0, the two
indefinite integrals are equivalent measures.
3. If (O,~, Jli) is a measure space, i = 1,2, then JlI is absolutely continuous relative to
III +Jl2·
4. If F(x; a, b) = (b - a)-I(x - a), a :'S: x :'S: b, the corresponding measure F{ . } is
absolutely continuous relative to Lebesgue measure with Radon-Nikodym
derivativef(x) = (b - a)-l/la,;x';bl and F is called the uniform distribution on
208 6 Measure Extensions
10. Set functions {v., n ;;:: I} on (n, .'F, Jl) are uniformly absolutely continuous relative
to Jl if for all e > 0, Jl{A} < 0, implies Iv.{A}1 < e for all n;;:: I. The sequence
{v., n ;;:: I} is equicontinuous from above at 0 iffor all e > 0 and Am! 0, Iv.{ Am} I
< doralln;;:: Iwhereverm;;:: m,.Provethatifmeasures{v.,n;;:: I}areequicon-
tinuous from above at 0 and also absolutely continuous relative to Jl, then {v.,n ;;:: I}
are uniformly absolutely continuous relative to Jl.
11. If f. e If p(n, §, Jl), n ;;:: I, then Ilf. - fmllp = o(l) as n, m ..... 00 iff (i) f. - fm 4 0
as, n, m ..... 00 and (ii) JA 1f.IP dJl, n ;;:: I, are equicontinuous from above at 0·
12. Random variables X I' . . . , X. have a (nonsingular) joint normal distribution if their
dJ. is absolutely continuous with density defined by f(x l , •.• , x.) = (2n)-'/2
6.5 Absolute Continuity of Measures, Distribution Functions 209
jlAl exp{ -! Li.j= 1 ai/Xi - ei)(Xj - ej)}, where A = {aiJ is a positive definite
matrix of order n and IA I signifies the determinant of A. Here, e = (e l , ... , e.) is a
e
real vector. Verify that this yields a bona fide probability measure and that E Xi = i ,
p(X i , Xi) = au!aiaj' where {aij} is the inverse matrix of A.
References
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
P. R. Halmos, Measure Theory, Van Nostrand, Princeton, 1950; Springer-Verlag,
Berlin and New York, 1974.
G. H. Hardy, J. E. Littlewood, and G. Polya, Inequalities, Cambridge Univ. Press,
London. 1934.
A. N. Kolmogorov, Foundations of Probability (Nathan Morrison, translator),
Chelsea, New York, 1950.
M. Loeve, Probability Theory, 3rd ed., Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
E. J. McShane, Integration, Princeton Univ. Press, Princeton, 1944.
M. E. Monroe, Introduction to Measure and Integration, Addison-Wesley, Cambridge,
Mass., 1953.
H. Robbins, "Mixture of Distributions," Ann. Math. Statist. 19 (1948),360-369.
S. Saks, Theory of the Integral (L. C. Young, translator), Stechert-Hafner, New York,
1937.
J. L. Snell, "Applications of martingale system theorems," Trans. Amer. Math. Soc.
73 (1952), 293-312.
D. V. Widder, Advanced Calculus, 2nd ed., Prentice-Hall, Englewood Cliffs, New
Jersey, 1961.
7
Conditional Expectation,
Conditional Independence,
Introduction to Martingales
1. Y is '§-measurable,
11. SA Y dP = SA X dP, all A E '§.
A E fi'. (1)
210
7.1 Conditional Expectations 211
J Y dP = A.{A} = f X dP,
·A A
A Ef"S.
PROOF. Set
either v{D} < 00, implying DuE E fi} and hence a ~ P{D u E} = P{D} + a,
that is, P{D} = 0 = v{D}, or alternatively v{D} = 00, whence P{D} > 0 by
the P-continuity of v. 0
f
dv
v{A} = A dP dP, AEfF. (4)
dV'
on E
:;= :' {
on P.
212 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
Then dv/dP is g;--measurable and (4) holds for A E g;- n E. If, rather, A E
g;- n EC, then (3) ensures that (4) still obtains. Finally, if dv/dP were infinite
on a set B of positive P-measure, then for every measurable subset A c B
v{A} = {:;dP = 00 or 0
Corollary 2. Let the random vectors X = (X I' ... , X m ) and f = (fl , ... , y,,)
on (n, /F, P) be independent of one another and let f be a Borel function on
R m x WwithIEf(X, f)l::; oo·/f,forxERm ,
PROOF. Let F x' F y, and F x. y be the joint distribution functions of the random
vectors, X, f and (X, f) respectively, and denote the corresponding
Lebesgue-Stieltjes measures by vx, Vy, and vx . y. Since f is a Borel function
on R m + n , Theorem 1.4.2 ensures that f(x, y) is a Borel function on R n for
each fixed x E R m , and so by Theorem 6.3.3 and Fubini's theorem
is a Borel set and g(x) = [g + (x) - g _(x)]/ DC(X) is a Borel function whence
by Theorem 1.4.4, g(X) is u(X)-measurable.
By independence (R m +n, fJm+n, vx . y) = (R m, 81 m, Vx ) x (R n, fJn, Vy). If
A E u(X), Theorem 1.4.4 guarantees the existence of B E!JI m such that A =
{X E B}, and once more via Theorem 6.3.3 and Fubini's theorem
214 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
= r r f±(x, y)dvy(y)dvx(x)
J JR"
B
= r
JBXR"
f± dvx,y
= f,/±(X, Y)dP.
EXAMPLE 1 (Kesten). If {S" = L~ Xi' n ~ I}, where {X, X" ~ I} are non-
negative i.i.d. r.v.s with E X = 00, then fiiii"_ex> XJS"-l = 00, a.c.
PROOF. Set A j = {Sj-l ~ I:Xil, I: > 0, where So = 0, and note that AjA j C
Aj{2J=: +1 X h ~ I:X j} for i < j, implying for k ~ 1 via independence and
identical distributions
P{A i , UA ~ P{Aj}P{O A
]=l+k
j}
]=k
j }'
then, clearly, a(x)/x ! and (Exercise 6.2.16) a(x) i and E a(X) = 00. It follows
that E a(I:X) = 00, all I: > O. However, via Corollary 2 and Examples
5.4.1,6.2.2
and so Lemma 4.2.4 ensures P{A n, i.o.} = 1, all e > 0, which is tantamount
to the conclusion of Example l.
It follows immediately from this example that for any i.i.d. LV.S {X, X n ,
n ~ I} with EIXI = 00,
a.c. o
P/A} Lx dP if WE A
where either of the constants on the right can be construed as any number in
[ - 00, 00] when the corresponding set A or A C has probability zero.
More generally, if {An, n ~ 1} is a O'-partition of 0 in ~ with P{A n} > 0,
n ~ 1, and ~ = O'(A n , n ~ 1), then for any measurable function X with
IEXI ~ 00
a.c. (13)
216 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
°
Moreover, this remains valid even if PrAm} = for certain indices m, the
quantity in the parenthesis being then interpreted as an arbitrary constant.
Some simple consequences of the definition of conditional expectation are
If ':9'1' ':9'2 are a-algebras with ':9'1 c ':9'2 c !F and IE XI::; 00, then
E{E{XI':9'2} l':9'd = E{XI':9'd = E{E{XI':9'd 1':9'2}, a.c. (14, v)
the first equality of (14, v) follows. Since E{X I':9' d is ':9'j-measurable for
i = 1 and hence i = 2, the second equality of (14, v) follows from (14, iv).
D
Theorem 2. Let {X n' n ;::: I} and Y be random variables with ElY I < 00 and
':9' a a-algebra ofevents.
i. (Monotone convergence theorem for conditional expectations). If Y ::;
X n iX, a.c., then E{X nl':9'} i E{XI':9'}, a.c.
ii. (Farou lemma for conditional expectations). If Y ::; X n' n ;::: 1, a.c., then
E{lim X nl':9'} ::; lim E{X nl':9'}, a.c.
iii. (Lebesgue dominated convergence theorem for conditional expectations).
If X n ~ X and IXnl ::; I YI, n ;::: 1, a.c., then E{X nl':9'} ~ E{XI':9'}.
PROOF. (i) By the monotone property of conditional expectations, E{ Y I':9'}
::; E{X nl':9'} i some function Z, a.c. For A E ':9', by ordinary monotone
convergence
fA
Z dP = lim n f
n A n
E{X l':9'}dP = lim f A
X n dP = f
A
X dP,
L Y d/i(9 = I :~:
Y dP(9 = Iy E{XI~} dP(9,
Consequently,
PROOF. Since (ii) follows directly from (i), it suffices to prove the latter. To
this end, define
*( )
gt=lm
I' g(s) - g(t) ,
5-1- s-t
whence g* is a finite, nondecreasing function on ( - 00, (0). Now the secant
line of a convex function is always above the one-sided tangent line, that is,
g(t) ~ g(s) + (t - s)g*(s), -00 < S, t < 00, (17)
whence if A = {IXI ::;; M, g*(X) ~ a}, 0< M < 00 and B = {IXI < 00,
g*(X) ~ o} both g(X) and g*(X) are bounded on A E <;#, so by (17)
IAg(Y) ~ IAg(X) + IA(Y - X)g*(X), a.c.
Since IE IA(Y - X)g*(X)1 ::;; 00, by Theorem 3,
IA E{g(Y)I<;#}'= E{IAg(Y)I<;#} ~ IAg(X), a.c.
As M -+ 00, IA ~ IB , so (16) holds on B. Similarly, for {g*(X) ::;; 0, IXI < oo}.
Consider next D = {X = oo}. If g*(s) > 0, some s in (-00, (0), then (17)
ensures
I D E{g(Y)I<;#} ~ g(s)ID + I D E{Y - sl<;#}g*(s) = 00 = g(X)' ID
If g*(s)::;; 0, all s in( -00, (0), theng! whence g(oo) ::;; g(X), a.c. and (16) holds
on D since
I D E{g(Y)I<;#} ~ IDg(oo) = IDg(X).
Let D' = {X = -oo}. If g*(s) > 0, all s in (-00, (0), then 9 i whence
I D· E{g(Y)I<;#} ~ ID,g( -(0) = g(X)I D,
(Y - s)g*(s) implying
°
whereas if g*(s) < for some s in (-00, (0), then via (17) g(Y) ~ g(s) +
PROOF. If lk = SUPn~ 1 E{X; I[X;; >kll ~}, k > 0, then lk~ by hypothesis.
Since with probability one
°
E{Xnl~} = E{Xn(J[X,;-:<>kl + I[x,;->kj) I~} 2:: E{XnI[x,;- :<>kll~} - lk,
it follows via Theorem 2(ii) that for all k > °
lim E{X n I~} 2:: lim E{XnI[x,;- :<>kll~} - lk
n n
EXAMPLE 2. Let Sn = LJ= 1 X j ' where {X n' n 2:: I} are independent LV.S and
let {lX n, n 2:: I} be constants such thatP{Sn < IXn} > 0, n 2:: 1. Then
(18)
entails
PROOF. Set AN = U.~)=N [Sn 2:: IX n], N 2:: 1, and suppose that P{A N} = I for
220 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
a.c. (21)
f(-<Xl.aN)
[1 - h(x)] dP{SN < x} = P{SN < aN} - i [SN<aN]
h(SN) dP = 0,
and so h(x) = 1 for every x < an that is a point of increase of P {SN < x}. Since
P{SN < aN} > 0, there must be at least one such x. Then monotonicity of h
guarantees h(x) = 1 for all x ~ aN and (21) ensures
P{A N+ 1 I SN} = 1, a.c.
on [SN ~ aN]. Consequently, recalling (20), P{A N+ I I SN} = 1, a.c., and so
P{AN+d = 1. Since P{Ad = 1 by hypothesis, it follows inductively that
P{A N } = 1 for all N ~ 1, which, in turn, yields (19). 0
Corollary 6. If Sn = 2:7
Xi' where {X n, n ~ I} are independent LV.S with
P{Xn < O} > 0, n ~ 1, then for nonnegative, constants {an, n ~ I}, (18) implies
(19).
Remark. Clearly, both equality signs can be deleted from (18) and (19)
provided P{Sn < an} > 0, n ~ 1, is modified to P{Sn ::; an} > 0, n ~ 1.
EXERCISES 7.1
I. If X is a r.V. with IE X I ~ 00 and ':§ a a-algebra of events with a(X) and '!J indepen-
dent classes, then EIX I ':§} = EX. a.c. In particular. if IX n} are independent LV.S
with IE Xnl ~ 'X), n:2: I, then E{Xnl X to . . . , X n- I } = EIXnl X n+ l • X"+2' ... : =
E X n • a.c.
2. Let I An. n :2: I} be a-partition of n in .'F. Verify that if t§ = a(A n. n :2: I) and
IE XI ~ x. then
4. Let ~ be a semi-algebra of events and X, Y r.v.s with IE XI < 00, IE YI < 00. If
SoX dP :::;; So
Y dP, DE C/, then E{X IO'(.@)} :::;; E{ Y IO'(.@)}, a.c.
5. Let (X I' X 2) be jointly normally distributed (Exercise 6.5.12) with p.d.f.
2 1/2
[27t0'10'2(1 - p) r I exp { 1
2(1 _
[xi 2px IX2
2) "2 - - - +"2
x~J} .
P 0'1 0'10'2 0'2
6. Prove that
i. if X isan 9"2 r.v. and Yisar.v. such that E{X I Y} = Y,a.c.and E{ Y IX} = X,a.c.,
then X = Y, a.c.
ii. Let ~I' ~2 be a-algebras of events and X an 9"1 r.v. If Xl = E{XI~d, X 2 =
E(XII~2)' and X = X 2 , a.c., then Xl = X 2 , a.c.
iii. If X, Yare 9"1 r.v.s with E{XI Y} = Y, a.c., and E{YIX} = X, a.c., then
X = Y, a.c.
7. Prove that the set E of lemma I and the function dvldP are unique to within an
equivalence.
8. Let X be an!E 2 r.v. and <§ a a-algebra ofevents. Prove that (i) 0'2(E{ X I <§}) :::;; 0'2(X),
(ii) iffor any oc in ( - 00, 00), Y = min(X, oc),
E{[X - E{XI~}J21~} :?: E{[Y - E{YI':§}J21<;§}, a.c.
(iii) E(X - E{XI~})2 :::;; E(X - y)2 for any ~-measurable r.v. Y.
:/
9. Let <§ be an a-algebra of events and (X., n :?: I} r. v.s. If for some p :?: I, X. ---4 X,
then E[X. I~} ~E[X I ~}.
10. Let (g be a a-algebra of events and X a nonnegative r.v. Prove that E[X I~} =
esup{h: h is ~-measurable, h :?: 0, a.c., and SA h dP :::;; .fAX dP, all A E~}.
12. If [ X., n :?: I} are 9" I interchangeable r.v.s (Exercise 6.3.10) and fi'. = O'(L!= I X j,
) :?: n), prove that E{Xd fi'.} = (lin) D X j , a.c., I :::;; i :::;; n. More generally, if
[X., n:?: I} are interchangeable r.v.s, cp is a symmetric Borel function on R m
with E Icp(X ..... , X m)1 < 00, and fi'. = O'(U m.j,) :?: n), where
(mn)U m .• = L:
IS; I < ... < i m S n
cp(X i ,,···, Xi,.), n:?: m,
The latter, P{A I B}, is called the conditional probability of the event A given
the event B and according to 7.1(12), if P{B} > 0,
WEB. (3)
°
yield:
~ P{A I '1} ~ 1, a.c,; (4, i)
P{A 1'1} = 0, a.c" iff P{A} = 0, P{A I '1} = 1, a.c., iff P{A} = 1; (4, ii)
Property (iii) asserts that for each sequence {An} of disjoint events
except for w in a null set which may well depend on the particular sequence
{An}. It does not stipulate that there exists a fixed null set N such that
for every sequence {An} of disjoint events. In fact, the later is false (Halmos,
1950, p. 210).
7.2 Conditional Probabilities, Conditional Probability Measures 223
Theorem 1. Iffor some pair ff l' <§ ofa-algebras ofevents, P",{A} == P(A, w) is
a regular conditional probability on ff 1 given <§ and X is an ff i-measurable
function with IE X I ::; 00, then
P(A, W) = P{A I<§} a.c. (as in Exercise 7.1.2), and so P(A, w) is a regular
conditional probability relative to ff given cr;.
(6)
224 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
Then for each X E R 2, P(B, x) is a probability measure on .JJ2, and for each
B E .~2, P(B, x) is a Borel function in X2 and hence u(X 2)-measurable. More-
over, for BE fJl2 and A 2 = R X B 2 E u(X 2)
=f 1 fl(U I t)fit)du dt
B2 [u: (u. I) e BI
and so according to (2), P(B, x) = P{BIX 2}(x), a.c., for each BE .:fd 2. Con-
sequently, P(B, x) is a regular conditional probability measure on fJl2 given
u(X 2)' Hence, by Theorem 1 for any Borel function h on R 2 with
IE h(X l , X 2)1 :::;; 00
a.c. (10)
Moreover,
Can a version of Px{B I~} be chosen for each BE flI" so that on the comple-
ment of a single null set N E~, (11) holds for all disjoint sequences
{B m , m 2': I} c .9B"? An affirmative answer to this case has been given by
Doob in Theorem 2 (below)
By the properties enumerated in (4) there is a null set N E ~ such that for
WENC and all rational numbers Ai, Ai, rim
F:;'(A 1, , A") 2': F:;'(A.'I' ... ,A~) if Ai > Ai, I ~ i ~ n, (i)
F:;'(A 1, , A") = lim F:;'(r 1m , ... , r"m), (ii)
rim t Ai
1 SiSn
226 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
lim F:;'(A I , ... , An) = 0, 1 :$; i :$; n, lim F:;'(A b ... , An) = I, (iii)
Aj - - 00 Aj - 00
1 :5i$n
(iv)
where the notation in (iv) is that of (5) of Section 6.3; also, A:$; A' signifies
that the inequality holds for each coordinate. Define for any real numbers
pj and WE N C
F::'(PI"'" Pn) = lim Fn"'(A I ,···, An), Aj rational, 1 :$; i :$; n, (14)
Ai-+Jli-
iSiS"
(15)
Then for each WEn, F:;'(PI' ... , Pn) is an n-dimensional d.f. and hence
determines a Lebesgue-Stieltjes measure P,.} on f!Jn with p",{R n} '= I. For
B E f!Jn and WEn define
(16)
If
:If = {B: BE f!Jn, Px(B. w) is <§'-measurable, Px(B, w) = P{X-I(B)I<§'}, a.c.},
g = {B: B = ,X[-
.= I
00, A rational AjE( -00,(0),1
j ), :$; i :$; n}
then :If is a A-class, g is a n-c1ass, and by (13) through (16):If :=> g. Hence,
by Theorem 1.3.2,:If :=> a(g) = f!Jn, that is, Px(B, w) is a regular conditional
distribution for X given <§'. 0
Corollary 1. If X = (XI' X 2 , ... ) is a (countable) stochastic process on
(n, ~, P) and <§' is a a-algebra of events, there exists a regular conditional
distribution for X given <§'.
PROOF. For all n ~ I and rational Ai> I :$; i :$; n, choose F:;'(A b . , . , An) as in
(13). Select the null set N E <§' such that in addition to properties (i)-(iv), for
WEN C
lim F:;'+I(A I , ... , An + l ) = F:;'(A I , ... , An), n~1. (v)
A. n + I-a::>
Define F:;'(J.lI' ... , J.ln) as in (14), (15). Then for each WEn, {F:;', n ~ I} is a
consistent family of dJ.s and hence by Theorem 6.4.3 determines a measure
J.l", on (ROO, f!Joo). Define Px(B, w) = p",{B}, BE f!Joo, and
:If = {B: BE f!Joo, P x(B, w) is <§'-measurable,
Px(B, w) = P{X-I(B)\<§'}, a.c.},
PROOF. For BE tJ42 and WEn, let Pw{B} = P(B, w) be a regular conditional
distribution for (X, Y) given r§. By Corollary 2
E{IXYIIr§}(w) =
Jr !x x21 dPw ,
R2
1 a.c.,
Since for each WEn, Pw is a probability measure on tJ42, (17) follows from the
ordinary Holder inequality. 0
EXERCISES 7.2
2. If in Example 2, Flx') = Iim x,_,_oo F(x l , X2), i = 1,2, prove that the dJ. F I is a
mixture of the family F(x 1 I X2) in the sense of Exercise 8.2.3.
4. If LV.S X and Y have a joint normal distribution (Exercise 6.5.12), show that the
conditional distributions of X given Y and Y given X are both normal. Is the con-
verse true?
5. If XI' ... , X k have a multinomial distribution (Exercise 6.3.9) find the conditional
distribution of X I given X 2, the conditional mean E( X 1 I X 2}. and the conditional
variance a 2(X 11 X 2)'
8. Prove the conditional Minkowski inequality (Exercise 6.2.7) and also the condi-
tional Markov inequality E{IYIIr§} ~ cP{1 Y ~ clr§} for any c > 0, Ye 2 1 ,
and sub-a-algebra r§.
7.3 Conditional Independence, Interchangeable Random Variables 229
a.c. (1)
If the subscript n of Sn is interpreted as "time" and the r.v.s SI, Sz, S3 are
envisaged as chance occurrences of the past, present and future times
respectively, then the prior relationship may be picturesquely stated as, "The
past and future are conditionally independent given the present." In fact,
designating the LV. Sn as "the present," the r.v.s SI"'" Sn-l as "the past,"
and the r.v.s Sn+ 1 " ' " Sn+m as "the future," it may be verified for any n 2:: 2,
m > 0, that the past and the future are conditionally independent given the
present. This property holds not only for sums Sn of independent r.v.s but
more generally when the r.v.s {Sn' n 2:: I} constitute a Markov chain (Exercise
7.3.1).
a.c. (6)
PROOF.(i) (1) => (3): Let Ai E r§j, i = 1,2, 3. By the definition of conditional
expectation, (1), and Theorem 7.1.3,
fA2A3
IA, = f
A3
IA,A2 = f
A3
P{A I A z l r§3} = fA3
P{A 1 I r§3} . P{Azl r§3}
= f
A3
IA2 P{A 1 I r§3} = fA2A3
P{A 1 I r§3},
Apropos of (ii), since (3) =;> (4) trivially, it suffices to prove the reverse
implication. If
then.s;1 is a A.-class ~ fi) via (4), and so by Theorem 1.3.2,.s;1 ~ a(fi) = ~t,
which is tantamount to (3).
Concerning (iii), it suffices to prove (5) ==- (1), the converse being trivial.
Since
.s;1 = {A 2 E fF: P{D t A 2 ~3} = P{D I I ~3} . P{A 2 ~3}, a.c., all D 1 E fi)tl,
1 1
p{Q I
[Xi < xJ <#} = it P{X i < Xi I <#}, a.c.
n P{X I <
m
Then
E[~.(x) - ~m(x)]Z = 1m - nl (P{X I < x} - P{X I < X, X z < x}) -+ 0
m'n
as m, n -+ 00, whence ~.(x)!. some ff-measurable LV. ~(x), implying
(Exercise 3.3.1(iv))
n
m
j= I
~.(Xj)!. n
m
j= I
~(x), m~1. (10)
j=1
~.(Xj) = m L IIX;<Xj)
n j=1 i=1
= m L I(IX)
n HM
= m1 ( L + L ) I(IX). (11)
n ~eM-N ~eN
7.3 Conditional Independence, Interchangeable Random Variables 233
Since as n --+ 00
1 1 nm - n(n - I)· .. (n - m + 1)
Iii
n aeM-N
L l(a)::; Iii
n
L
aeM-N
1=
n
m --+ 0
f Ii~n(Xj)dP=n-mI f
Aj=1 aeN X-I(B)
l(a)dP+o(1). (12)
fX-I (B)
l(a)dP = f
(nX)-I(B)
1IX,,<Xl ..... X'.,,<XmldP
= f
X-I(B)
1[Xl <XI.·· .• Xm<xm] dP .
x f
X-I(B)
1 1XI <XI, .... X'!'<Xm) dP + 0(1)
On the other hand, the left side of(13) tends to SA nj= I ~(x)dP by (10) and
dominated convergence. Hence for any A E Y, all real x I' ... ,Xm , and positive
integers m,
f Ii~(x)dP= f
A j=1 A
1 I X I<Xl, ... ,Xm<xml dP ,
and therefore
n ~(x),
m
P{X I < XI"'" X m < XmI ff} = E{ I) ~(x) Iff} = I) ~(Xj), a.c. (16)
In particular, for m = I, ~(XI) = P{X I < xIIY} = P{X I < XI Iff}, a,c.,
and by (15), (16)
234 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
I
P{X I < x;I9'} = n
m
I
P{X I < xilff}
PROOF. For any rational A., define Fi(A.) = P{X; < A. I I'§}(w), i ~ 1. There
exists a null set N E I'§ such that for WE NO, all positive integers n, and all
rational A., A.', r, r l , . . . , rn :
nFi(r;).
n
Then properties (i)-(v) of Theorem 7.2.2 and Corollary 7.2.1 hold, so that
defining Fn,ro = nJ= I F Xj for WEN and continuing as in the proof of
Corollary 7.2.1, existence is ensured of a regular conditional distribution
pro for X given I'§.
7.3 Conditional Independence, Interchangeable Random Variables 235
Let {¢n, n ~ I} be coordinate LV.S on (ROO, fJloo, P"'). Since for all n ~ 2
and real XI"'" Xn
P"'{¢I < XI"'" ¢n < x n} = nn
i= 1
Ff'(xj-) = nP"'{¢j < xJ
n
i= 1
for WE N C , while
n P{X j < xJ
n
L
Roo
e
Apropos of (ii), set S~ = IJ=I Xi, T: = IJ=I j , where Xi = XAIXil~n)'
¢j = eill~jl,;;n), 1 ~j ~ n. Then P{Sn -:F S~} ~ IJ=I P{IXjl > n} = 0(1) and
In view of Corollaries 4 and 5 it seems plausible that many results valid for
sequences of i.i.d. random variables carryover to sequences of interchange-
able r.v.s (see Chapter 9).
Lemma 1. Let <§ be a a-algebra of events and f§* its completion. If X* is any
<§*-measurable function, there exists a <§-measurable function X such that
X = X*, a.c.
PROOF. Let
£ = {X*: X* ;;:: 0, X* is r1*-measurable, and X* = X, a.c.,
for some <§-measurable function X}.
Then £ is a monotone system and I G E £ for all G E <§*. Hence, £ con-
tains all nonnegative <§*-measurable functions. The general case now follows
from X* = X*+ - X*-. 0
7.3 Conditional Independence, Interchangeable Random Variables 237
Lemma 2. Let ~* be the completion ofa a-algebra ~ ofevents. (i) Ifl E X I ~ 00,
then E{X I ~*} = E{X I ~}, a.c. (ii) LV.S {X n , n ~ I} are conditionally
independent (resp. conditionally i.i.d.) given ~ iff {X n , n ~ I} are condition-
ally independent (resp. conditionally i.i.d.) given ~*.
PROOF. (i) Since by Lemma 1, E{X I ~*} = Y, a.c., where Y is ~-measurable,
E{X I ~*} = E{X I ~}, a.c., whence (ii) follows immediately from (i). 0
~n(x) = ~
n
f I IXj <x)!!2. P{X
j=1
1 < x I f7},
J
= E[IAjQ P{X < Xi I ~*} If7 = E[IAiQ P{X j < Xi I f7} If7
j J
= P{A I f7} . p{0[Xj < xiJ If7}. a.c.
EXERCISES 7.3
I. Let S denote the set of positive or nonnegative or all integers and let (P•• n E S} be a
probability density function. Define P = {Pi}} to be a stochastic matrix. that is,
Pi} ~ 0 for all i, j E S, and Iies Pi} = I for all i E S. (i) Verify that p{nj=o [Xi =
iJ} = Pio nj;;J Pij.ij+1 satisfies the consistency requirement of Theorem 6.4.3. The
LV.S (X., n ~ O} are called a temporally homogeneous Markov chain with state
space S. (ii) Check that Pi} = P{X.+ 1 = j IX. = i}, n ~ 0 for all i, j E S, and verify
for n ~ I, m > 0, that "the past" X 0' ... , X .-1 and "the future" X.+ I> . . . , X .+m
are conditionally independent given "the present" X•. (iii) Verify that the product
of two stochastic matrices is a stochastic matrix and interpret the equation pm+. =
pm. p. probabilistically.
2. Verify that the class Y of permutable events as defined in (7) is a a-algebra and prove
Corollary I.
3. If S. = Ii Xi where {X., n ~ I} are independent r.v.s then 8 1 and 8 3 are condi-
tionally independent given 8 2 , Hint: Via Corollary 7.1.2 if F3 is the dJ. of X 3
6. If a r.v. X is independent of the i.i.d. r.v.s {X., n ~ I}, show that y" = X. + X,
n ~ 1, are interchangeable r.v.s. Express that joint dJ. of (YI , ... , y,,) in terms of
the dJ.s of X and X I .
7. For any n ~ 2, find n interchangeable r.v.s {X 1, ... , X.} which cannot be embedded
in a collection of n + I interchangeable LV.S {XI"'" X., X.+ I }. Hint: Recall
Exercise 6.3.10.
8. If {X., n ~ I} are .Ii'l interchangeable LV.S with p = E(X 1 - EX I)(X 2 - E X I),
then E X I X 2 = 0 iff E X I = 0 and p = O.
9. If {X., n ~ I} are .Ii'2 interchangeable LV.S and {Y", n ~ I} is defined by X. =
y" + E{XdY'}, then {y", n ~ I} constitute uncorrelated interchangeable LV.S.
I' I I I I
I
E ~ ~ Xj ~ E n _ I ~ Xj n 2.
n- 1
,
~
11. Prove that if the r.v.'s {X n, n ~ I} of Corollary 5 are conditionally i.i.d. given rJ,
then the coordinate LV.'S {en, n ~ I} are i.i.d.
When N = {I, 2, ... , oo}, the preceding coincides with the definition of
Section 5.3 and T is a finite stopping time or stopping variable if
P{T = oo} = O.
If {Sn, ffn' n E N} is a submartingale and there exists a measurable
function R with IE R I ::s; 00 (resp. an ()nEN ffn-measurable function L with
IELI ::s; (0) such that for each n E N
then {Sn, ffn, N} is said to be closed on the right (resp. left). A closed sub-
martingale {Sn, ffn, N} is one which is closed both on the right and left. A
submartingale {Sn' ffn, N} is declared to have a last (resp., first) element if
N has a maximum (resp. minimum). Obviously, a submartingale with a last
(resp. first) element is closed on the right (resp.left). A martingale {Sn, ffn' N}
is said to be closed on the right by R (resp. on left by L) if (2) holds with
equality.
The submartingale {Sn, ff n' n :2: I} will be closed iff it is closed on the right
since it is automatically closed on the left by S I' An analogous statement
holds for {Sn, ffn, n ::s; -I}.
The important special case ofa submartingale or martingale {Sn, ffn, N}
with ffn = u(Sm, m ::s; n, mEN), n E N, will be denoted by {Sn, N} or {Sn,
nEN}.
Simple properties of conditional expectations readily entail the following:
i. The stochastic sequence {Sn, ffn, - 00 < n < oo} is a submartingale
iff for every finite integer n, IE Snl ::s; 00 and E{Sn+ II ffn} ~ Sn, a.c.
ii. If {Sn, ffn, N} is a submartingale and {Sn, <§n, n E N} is a stochastic
sequence with <§n c ffn, n EN, then {Sn, <§n, N} is a submartingale.
iii. Ifboth {Sn' ffn, N} and {S~, '~n, N} are submartingales and E Sn + E S~
exists for each n E N, then {Sn + S~, ffn' N} is a submartingale.
iv. An IE I stochastic sequence {Sn = I~ Xi, ffn, - 00 < n < oo} is a
submartingale (resp. martingale) iff for -00 < n < 00
E{X n+ 1 I ffn} :2: 0 (resp. E{X n+ 1 I ffn} = 0), a.c.;
n
E S; = I EX;, n ~ 1.
j= I
7.4 Introduction to Martingales 241
Um. n = (n)-I
m
L
1 ;s;i l < ... <i m :5n
cp(X i " ... , X im ), n ;?: m.
If iF n = a(Um.j,j ~ n), then for 1 ::s;; i l < ... < im ::s;; n + I and BEiF n+t
Consequently, if U: = Um. -n and iF: = iF -n' n::S;; -m, then {U:, iF:,
n ::s;; -m} is a martingale closed on the right. The important special cases
m = I, cp(x) = x and m = 2, cp(x 1, X2) = (Xl - X2)2j2 yield the arithmetic
mean U l.n = (ljn) L7= 1 Xi = Xn (say) and the sample variance U 2.n =
(n - 1)-1 L7=1 (Xi - Xn )2 respectively.
242 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
{Vn' ·'Y'n, n ~ I} is a martingale. Note that for N = {I, 2, ...}, EI Vnl < ro,
n ~ I, and (3) constitute an alternative definition of an ff j martingale
{Vn,fl'n,n ~ 1}.
n
EIS~I ~ EISnl + L E Xj = EISnl + E Sn -
j:2
EX) ~ 3 sup EISnl·
n~1
J [T< 00]
ST dP :F 00,
has positive probability, say P{ V} > lJ > O. It may and will be supposed,
replacing Sn by (IX - [J)- )(Sn - (J), that IX = 1, [J = O.
Set mo = 1, B o = Q, Vo = B o V and C~ = Bo{Sn > 1, Sj ~ 1, for mo ~
j < n}. Define A) = U~~ C~, where nl is large enough to ensure P{Vo - Ad
< lJ/4.
Next, define D~ = Al {Sn < 0, Sj 2 0, n) ~ j < n}, B) = U::'i D~, where
ml is large enough to guarantee P{Vo - Bd < lJ/4 and note that
m,
Pm,{A 1 - Bd 2 Pml{A1 - Bd + L pj{DJ}
j=nl
nl
2 Pn,{Ad = L Pnl{C~}
n=mo
244 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
"2 m2
Lemma 1. If {Sn, ~n, n EN} is a martingale and ({) is any real convex function
with IE({)(Sn) I :::; 00, then {({)(Sn)' ~n' nEN} is a submartingale. If (Sn' ~n'
n E N} is merely a submartingale but ({) is, in addition, nondecreasing, then
{({)(Sn)' ~ n' n E N} is likewise a submartingale.
By the conditional Jensen inequality, Theorem 7.1.4, for m < nand
PROOF.
m,nEN,
E{({)(Sn) I·~ m} ;;::: ({)(E{Sn I ~ m}) = ({)(Sm)' a.c.
yielding the first statement. The submartingale hypothesis combined with ({)
nondecreasing converts the prior equality to ;;:::, thereby preserving the
oo~~oo. 0
r
J[T<oo]
S;dP-:s;,M, r
J[T<oo)
ISTldP-:s;,2M-ES I -:s;,3supEI Snl.
n"21
(5)
PROOF.Set T' = min(T, n), n :2: I. Since {S;, g;n, n :2: I} is a submartingale
by Lemma 1,
E s;, = L n
j= I
I
[T=j)
st + I [T>n)
s; -:s;, Ln
j= I
I
[T=j)
s; + I
[T>n]
s; = E Sn+,
so that E S;, ~ M, whence the first part of (5) follows from Fatou's lemma,
Next,
E SI = i[T= I]
SI + i [T> I)
SI -:s;, i
[T= I)
SI + i[T> I)
S2
-:s;, i[T=I]
SI + i [T=2)
S2 + i
[T>2]
S2 -:s;, ...
-:s;, r
J[I :5T:5n)
ST + r
J[T>n)
Sn = EST"
implying
E 1ST' I = 2 E S;, - EST' -:s;, 2M - E SI'
and the remainder of (5) follows once more by Fatou. o
The special case X n == X o of the next corollary, due to Doob (1940), may
be considered a milestone in the development of martingales, When X 0 is
bounded, this had been obtained by Levy (1937).
Corollary 1. Let {g; n' n :2: 1} be a stochastic basis and {X n' n :2: O} a sequence
of.If I LV.S with X n ~ X 0 and E sUPn"211 Xnl < 00. Then, if
g; 00 = a( Ug;n),
n=1
PROOF. (i) It will first be demonstrated for any integrable r,v, X 0 that
E{X o I g;n} ~ E{X o I g; oo}·
r
JISn>C)
Sn= r
JISn>C)
E{Xolg;n}= r
JISn>C)
Xo--+O
as C --+ 00, and so {Sn' n 2 I} is u.i., whence Sn~ Sro by Corollary 4.2.4.
Thus, for all n 2 I and A E g;n, if m > n,
by the part (i) already proved. Since Ym ~ 0 and I Ym I ~ 2 sup IXnIE !l\,
it follows that E{ Ym I g; ao} ~ 0 as m --+ 00 by Theorem 7.1.2, whence
~~Q 0
Apropos of (ii), the hypothesis implies Sn ~ Soo via (i). Moreover, for
A E fF mand n 2= m, applying Theorem 4.2.2(ii) to - Sn 1A'
f f
A
Sm :5;
A
Sn :5; TIm
n
f f
A
Sn :5;
A
Soo,
i [Sn>k)
S+ <
n -
i[Sn>k)
S+
.
lim i
k-oo [Sn>k)
S: :5; lim
k-Q
i [Sn>k]
S+ = 0,
uniformly in n as k -+ 00. Now each S: is integrable and so {Sn+, n 2= I} are
u.i. By (ii) {Sn, fF n' 1 :5; n :5; oo} is a submartingale. To verify that it is closed
by S, define S~k) = max(Sn' -k), 1 :5; n :5; 00, and S(k) = max(S, -k), where
k > O. Then {S~k), n 2= I} are u.i., S~k)~S~), and by Lemma 1 {S~k), fF n,
n 2= I} is a submartingale closed by S(k). Hence, for A E fF n , n 2= 1,
L LS~k) ~ LS~) L
S(k) 2= 2= S 00'
f A
S = lim
k-oo
f A
S(k) 2= f A
S 00 ,
Theorem 4. If
is a submartingale with E sUPn;, I X: < 00, then Sn converges a.c. on the set
{XI> -00, sUPn;,1 Sn < oo}.
PROOF. For any c> 0, define T = T;, = inf{n ~ 1: Sn > c}. Then T is a
stopping time and {T;, = oo} = {suPn;, I Sn ~ c} --+ {suPn;, I Sn < oo} as
c --+ 00. As seen in Example 5, {Un = LJ=I XjI[T;,j), fF n, n ~ I} is a
submartingale and
E U: = E(.f. XjIIT;,jJ) +
J= 1
~ E(.f. X j I IT >i1) + + E(f. XjIIT=jl)+
J= 1 I
i IT>n)
s: = 0, (6)
then for n ~ 1
f
A[T:2:nl
ST> f
A[T:2:nl
Sn· (9)
Now
f A[T:2:nl
Sn = f A[T=nl
Sn + f A[T>nl
Sn
= f A[T=nl
ST + f A[T:2:n+ II
Sn+ I·
The last term on the right is the first term on the left with n replaced by n + 1,
so, repeating the argument m - n - I times,
fA[T:2:nl
Sn:$ f
A[nST<ml
ST + f A[T:2:ml
Sm = f
A[nSTSml
ST + fA[T>ml
Sm· (10)
Let BE!F Tn and Bm = B[T" = m]. Then Bm E j" m and T,,+ t ~ m a.c. on
Bm • By (i)
whence, summing on m,
implying (11). o
250 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
lim
n-",
r
JIT>nl
ISnl = 0, (12)
thenfor n ~ 1
T = inf{j: -n ~j ~ -1, Yj ~ I}
7.4 Introduction to Martingales 251
00 T
= EI I[T~nl E{X: I§"n-I} = E I E{X: I§"n- d < 00.
1 1
I
n
I[T>n)
s+n <-
I
[T>n)
L
n
j=1
x:t
'\' . J -
< '\'
T
L.J
[T>n)j=1
x:t = 0(1)
as n --+ 00. Thus, (6) and consequently the conclusion (7) of Theorem 5 hold.
Under (ii), sUPn2: 1 E S: < 00 whence E S; < 00 by Lemma 2. Since
P{T> n} = 0(1), the remainder of(6) and consequently (15) follow from u.i.
o
Corollary 5. If {Sn = I7
Xi' §"n' n ?: 1} is an It'l martingale and T is a
finite {§"n} -time satisfying
T
ELE{lXnll§"n-d < 00 (16)
n=1
(in particular, if T is a bounded r.v.) or if {Sn' §" n' n ?: 1} is u.i., then for any
252 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
n ~ 1
(20)
Thus, if E S} = 00, equality holds in (20). In the contrary case E S} < 00, in
view of (20), equality will hold therein recalling (19) if,
(21 )
lim
n-""
i
IT>n]
S; < 00 (22)
or
lim i
n-"" [T>n]
ISnl =0 (23)
lim
n~co
r
J[T>n)
S; ~ k lim
n~co
r
JIT>n,ISnl>k)
ISn I = k· c --+ 00
whence on {T ~ n}
Since E{S} I ~n} = E{S}(n) I ~n} = Shn) on {T < n}, (21) follows. 0
lim r
n~co J[T>n)
ISnl = 0, lim r
n~co J[T>n)
S; < 00,
T (24)
EI
n=1
x; < 00,
Either of the last two conditions of(24) implies E ST = EX l ' If {Sn = D Xi'
~ n' n ~ I} is merely an .se 1 martingale, the last condition of (24) entails
EST=EX t ·
i [T>n)
S; ::; E sLn = E
TAn
I XJ ::; E I XI.
1
T
1
(26)
and so in view of Lemma 3 and Corollary 4 it suffices to note that the final
condition of (24) ensures that
T
EISTI ::; E I IXnl < 00
I
254 7 Conditional Expectation. Conditional Independence, Introduction to Martingales
and
n T
EISnIIIT>n] ~ E L IXjIIIT>n) ~ E I
1 t
IXjIIIT>n] = 0(1). 0
(27)
implies
E Si- = ELf Xl = ELf af. (29)
r 2
(.,---------,-,.-a_+_S--'-T_-,.--) 2 < a + Y1 + _1_ (32)
JIT<OO) b + Y1 + ... + YT - (b + y1 )2 b + Y1 '
f (+
ITSn)
b Ya + ST
I + ... + T
~ E Vj E{
Y )2 ~ L.
j~ I
(a +2
Sj) - (a + Sj_l) 2 IY'
tZ
j- I }
n
= vI(a 2 + YJ + E L: Vj lj.
j~ 2
Since
Y 1
v· Y. = ) < ,..--------
)) (b + YI + ... + lj)2 - b + YI + ... + lj_1 b + YI + ... + lj'
the conclusion (32) follows as n -+ 00. o
Theorem 8. If {Sn = L:~ Xj' Ji'n' n ~ I} is a nonnegative 2 1 submartingale
and {v n, Ji' n - 1> n ~ I} Q stochastic sequence of 2 00 r. v.s with Vn ~ Vn+ I ~ 0,
a.c., then for any A> 0, (i)
A p{ I sjSnVjSj ~ A} + r
max vnS n ~ j~i I E VjX j (33)
J[maxvjsj<).]
P { max Sj
I sjSn
~ A} ~ -1
A
f
[maxSj2:).)
Sn, (34)
whence
IISnllp ~ II max Sj
I I <)Sn
II
IP
~ P ~ 1 IISnllp, p> 1,
°
martingale and {b n, n ~ I} is a positive, nondecreasing real sequence, then for
any A>
P { V.I
max -.!... > A < -12 } L --)
Eu
n
2
(36)
I~j~n bj - I - ..1. j~1 b} .
Ai p{ max VjSj
1 :5j:5:n
~ Ai} + fL ~lf~"VjSj<).r]
VnS n
~ i [T;Sn)
VT;ST; + i [T;>n)
VnSn = E VTiST,
Ti n
E S: - 1 s; E(S: - 1)+ = L oo
P{S: - 1 2: )'}d)'
1 1
00
s; -1- s. dP d)'
o A. + 1 [S~ ~ H I)
r(S~-W d)'
=ES. A+l=ES.log+S:.
Jo
Since for constants a ;::: 0, b > 0 necessarily a log b :$: a log+ a + be-I,
E S: - 1 s; E S. log+ S. + e- I E S:,
from which the second portion of (35) is immediate. o
EXAMPLE 9. If {S., $i., n ;::: I} is a submartingale and h is any nonnegative,
increasing, convex function, then for any positive t and real x
E h(tS.)
P { max Sj 2: x } s; h( ) (38)
I ~)~. tx
and, in particular,
p{ max Sj 2: x}
I~)~.
S; p{h(m~x tSj) 2: h(tX)} =
I$)$n
p{ m~x h(tS) 2: h(tX)}
15)$.
< E h(tS.). 0
- h(tx)
7.4 Introduction to Martingales 257
EXAMPLE 10. If Sn = Lj=1 Xj' where {X n, n ~ 1} are i.i.d. fiJ p r.v.'s for some
= 0 whenever 1 :5: P < 2, then
p in (0, 2) with E XI
00
EXERCISES 7.4
I. If {S., n :2: I} is a martingale satisfying (i) I1'=1
E X/ < 00 or (ii) If
E(IXjl/uxjl>l]
1])
+ X/IUXjl:S < 00 or (iii) I1'=1
EIXjlP < 00 for some p in [1, 2], then converges S.
a.c. Hint: For (ii) consider X; = X j l[I Xk;l] - E{X/[IXj l:S1)IX 1 , ••• , Xj-d and
Xi' = Xj - X;.
2. In statistical problems, likelihood ratios U. = g.(X l ' ... , X.)/f,,(X I ' ... , X.) are
encountered. where!., g. are densities, each being a candidate for the ac[ual density
ofLv.s XI.' .. , X •. If {X., n :2: I} are coordinate LV.S on (ROO, atoo, P) and g. vanishes
whenever f" does, show that {U., n ~ I} is a martingale whenf. is [he true density.
3. There exist martingales {S., ff., n :2: I} and stopping variables T for which (13)
fails. Let {X n , n ~ I} be i.i.d. with E XI = 0 and set T = inf{n:2: I:S. = D Xi
> O}. Then for n = I, E{ST - S.lff.} > 0 on [T > I].
4. If {S., ff., n :2: I} is a martingale or positive submartingale. then for any stopping
time T, EISTI ~ lim._oo EIS.I. In particular, if {S.,ff.,n:2: I} is!f'. bounded,
EISTI < 00 for every stopping time T.
5. If E 1STI < 00, it is spurious generality to have !im,,-oo rather than lim._oo in (6) or
(12). Hint: If v" = S;; in the first case and IS. I in the second, then, as in the proof of
Theorem 5,
6. (i) Find a positive !f'1 martingale which is not u.i. (ii) If Y", n :2: 1, are r.v.s on
(!l, ff, P) and A E u(Y1 , Y2 , •• • ), then P{AI Y1 , •.• , Y,,} ~ I A • Hint: Apply Corol-
lary 1. (iii) If {ff., n:2: I} is a stochastic basis and {X., n:2: I} are r.v.'s with
258 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
E sUP.«lIX.I<co, then
E{limn_co X.lffco } ::;; limn-co E{X.lffco }·
8. If {So = II=l Xi' n ~ I} is an Ifp martingale for some p in [1,2] and 0 < b. i co
then lim S./b. = 0, a.c. on the set A where I:'=2 b;PE{lX.IPIff.-d < 00. In par-
ticular, S./b. ~ 0 if P{A} = 1. Hint: Set T" = T 1\ n where T = inf{n ~ 1:
D~: bj-PE{ IXjIPIj"j-l} > K > 0 and apply Exercise 1 to Ii" X)bj.
9. Let Y be an 2. LV. and {'§., -00 < n < oo} a stochastic basis. If '§co =
a(U,:,co '§.), '§-co = n,:,co '§., and U. = E{YI'§.}, -co::;; n::;; 00, then {U.,
'§., - 00 ::;; n ::;; oo} is a martingale.
11. Show, if (S., ff., n ~ I} is a nonnegative martingale with E S. = 1, that PIS. ~ A.,
some n ~ I} ::;; I/X
12. Prove that an If. stochastic sequence {II= t Xi' a(X., ... , X.), n ~ I} is a martin-
gale iff EX. + 1 cp(X I' ... , X.) = 0 for all n ~ I and all bounded Borel functions cp,
whereas the 2. LV.S {X., n ~ I} are independent iff for all n ~ 1
prove that P{X.E A, i.o.} ::;; P{X.EB, i.o.}. Hint: If FN= Ui=N [XjEB],
limn_co P{FNIX 1 ••• X n } = IF N '
15. (Doob) If (X.n ~ I} are independent LV.S with E X n = 0, n ~ 1, and S. = I~ Xi,
S: = maxI sis.ISd, then E S: ::;; 8 EIS.I, thereby improving (33) for p = 1. Hint:
Via Ottaviani's inequality (Exercise 3.3.16)
fo
co PIS: > 2y}dy::;; 2 EIS.I + I co
2EIS,,1
2 P{lS.1 > y}dy.
7.5 U-Statistics 259
7.5. U -Statistics
Let h be a measurable symmetric function on R\ k ~ 1 (i.e., invariant under
the permutations of its arguments for k ~ 2) and {X.., n ~ 1} a sequence of
i.i.d. random variables. Then (Example 7.4.3) a sequence of V-statistics Vk...(h)
and their corresponding sums Sk... (h) are defined by
n
(k ) Vk...(h) = Sk...(h) = . L.
1:5:11<"'<lk:5:n
h(X il , ... , X ik ), n ~ k~ 1. (1)
For hE .!f 1 , that is, Elhl = Elh(X 1 , ... , Xk)1 < 00, the V-statistic Vk...(h)
and its "kernel" h are said to be degenerate of order i - 1 where 2 ~ i ~ k if
E{h(X 1, ... , XdlX 1 ' ' ' ' , Xj} = O,a.s. forj = i - 1but not forj = i; otherwise
(i.e., if E{h(X 1"'" Xk)IX d is not a.s. zero), they are non-degenerate. In the
particular case i = k, they are called completely degenerate. It is sometimes
convenient to express non-degeneracy as "degeneracy of order zero."
Let I denote the identity operator, that is, If = f, and define ~ and opera-
tors Qj' 1 ~ j ~ k by
~ = QJ = E{f(X 1 , ... , Xk)IX«, 1 ~ IX ~ k, IX #j}
for any function f on R k with E If I < 00.
and, for r ~ 1,
ft(X,) = E{J(X" X,+t, ... , X'+k-dIX,} - Ef
are i.i.d. random variables.
Lemma 2. The functions Jj = Jj(X t, ... , X) defined in (2) are completely degen-
erate for 2 :S::j:S:: k and
k
f(Xt, .. ·,Xk)=Ef+ Lt ft(X;)
i=
+ t ';;i,L< i ,;;k
fz(X i , Xi,) + ...
2
+ L
1 $;;i t <···<ik-t::::;k
h-t(Xi,,··.,X ik _,)
n (I -
j
E{JjIX t,· .. , Xj-d = Qj'" QkJj = Qj'" Qk Q;)Jj* = 0, a.c.
i= t
whence Jj(X t, ... , X) is completely degenerate, 2 :s:: j :s:: k. The final statement
follows from the representation
n [(I -
k
f = Qi) + Q;]f,
i= 1
<;§t and <;§z are conditionally independent given <;§3 by Corollary 7.3.3. Since
f(X i " .•. , Xi) is a(<;§t U <;§3)-measurable, (i) follows from Theorem 7.3.1 (iv).
Apropos of (ii), if A is the set of distinct integers among Pt, ... , 13m' it, ... ,
ik- t , then {it, ... , ik- d = A n {it, ... , ik}, so that(ii) follows from (i). 0
7.5 V-Statistics 261
{.f. aj
J=k
. L.
1:S;I,<'''<lk-,<J
.h(Xi",,,,Xik_,,Xj),ofFn,n~k}
is a martingale and, in particular (a j = l,j ~ 1), so is {Sk,n(h), ofFn, n ~ k}.
Vk,n(h) = ±(~)
j= 1 ]
Vj,n(h j), (3)
n)
( k U.,,(h) ~ S.,,(h.) + k-1
j~'
C)G)
C) (n) k (k)
SJ,.(h j ) ~ k J~' j UJ,.(hj >
Lemma 4. Let 1 ~ i 1 < ... < ik ~ nand 1 ~ 131 < ... < 13m ~ n be two sets of
integers with {i 1, ... , ik } n {131"'" 13m} = {i 1, ... , ij }, where 1 ~j < k. (i) Iff
is a completely degenerate function on Rk, k ;::: 2, and 9 is a function on R m with
IE gl ~ 00, IEfgl ~ 00, then
Ef(X i" ... , XiJg(X p" ... , XpJ = O.
(ii) If m = k and h is a symmetric function on Rk, k ;::: 2, with E h 2 < 00, then
E h(X i" ... , XiJh(X p" ... , XpJ = E[E{h(X 1 , ... , X k)!X 1 , ... , XJ2,
and the right side is an increasing function of j.
PROOF. Employing Lemma 3(i),
E{f(X j 1 , ••• , Xi k p , ...
)g(XI m
, X p )IX j 1 , ... , Xj}
)
and so
7.5 U-Statistics 263
PROOF. Set qj = E[E 2 {h(X 1 , .•. , Xk)IX 1 " ' " Xj}]. Since the number of pairs
(J~k.n(hl=(kn)-2 L L
l~il<"'<ik~n l~PI<"'<Pk~n
Eh(Xi"···,XiJh(Xp,, ... ,XpJ
Then Pk = P and for P in (1, 2k/(2k - j)), the sequence {Pj' 1 ~ j ~ k} is de-
creasing, and Pj E (1, 2).
provided
1ic.n(h) = L: r
j=k
a
L:
lsi,<"'<ik _,<j
giXi" ... , Xik-l' Xj}'
If f7n = a(X l ' ... , X n), Corollary 1 asserts that {Sk.n(h), f7n, n ~ k} is a mar-
tingale and the same is true of {lic.n(h), f7n, n ~ k} since, via Lemma 3(i),
E{gn(Xi " ... , Xik-l' X n)lf7n-d
= E{gn(Xi " ... , X ik _" Xn)IX i " ... , X ik _,} a.s. O.
Hence, employing Lemma 4(i) and setting An = {(n - 1) < Ihl 1/a ~ n},n ~ 1,
Thus, in view of
sup EISk.ih)1 ~ sup EISk.n(h) -lic.ih)1 + sup Ellic.n(h)!,
n;;,k n;;,k n;;,k
n- k/PSk.n_1(h) = n- a t
j=k
L
l:S;i,<'''<i k _,<j
h(X i" ... , X ik _" X)~O. (8)
to .PP' 1 < P < 2 and k is any integer ~ 2. Clearly, (8) is tantamount to (6)
since E h = O.
Suppose next that 2 :::; i < k. Then, via Hoeffding's decomposition (Theo-
rem 1) and since again E h = 0,
k p
k 1 (n - j)!
:::; n- / ~. (k _ ')' (n _ k)'. ISj,.(h)1
)-1 ] •
k 1
" k-j-k/PIS (h)1
/..-. (k _ ] ')'• n
:::; )=1 j,. j
k 1 j
= ~. (k _ .),ISj,.(h)l/n /PJ , (9)
)=1 ] •
D k- 1,j(h) = L
1 :::;i 1 < ... <i k - 1 <j
h(X i " ... , Xik-l' X).
266 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
Now,
:::; j=k
for some choice of 1 :::; i1 < ... < i k - 1 < j}
:::; L L
00 00
j=k
j
1 E h·} -< "L.. J·k-l-. "L.. E hIAn
n=1
Thus, the series L'f'=kF r'D k- 1 ,j(h) and hence also LJ=kF·Dk- 1•j (h)converges
a.s. whence n-k/PSk.n(h)~ O. 0
n 1/2 Vk,n(h) _
-172
n
k ~ E{h(X" ... , X r+k-l)!Xr}+
L,
r= 1
n 1/2 ~
L,
j= 2
(k) Vj,n()'
}
.
h
As in the proof of Theorem 2, (11) implies that hj E 2 2j/(2j-l), 2 5, j 5, k
whence (12) follows from Corollary 4. Clearly, Elh1 4 / 3 < 00 ensures (11).
o
A Central Limit Theorem and Law of the Iterated Logarithm for non-
degenerate V-statistics will be proved in Sections 9.1 and 10.2.
EXERCISE 7.5
k
(k + l)Sk+
-
1 •• =
" .-
'-- (-lYSk_ ).•S.U+l) , k ~ 1,
)=0
and (ii) if {X, X., n ~ 1} are i.i.d. with E X = 0, then Sk.• /(k) is a completely degener-
ate U-statisticfor k ~ 2.
2. Let Sk.•, n ~ k and S~\j ~ 1 be as in Exercise 1, where {X, X., n ~ 1} are i.i.d., and
let {b n , n ~ 1} be constants such that 0 < b./n 1/ 2 t 00. If (i)(l/b.) D XI~ 0 and
(ii) I~=l P{lXI > bn } < 00, prove that Sk,./b:~~,k ~ 2. In particular, if X E 2 p ,
o< p < 2 and E X = 0 whenever p ~ 1, then Sk.•/nk/P ~ O. Hint: Employ the
Newton identities
where the sum is over all non-negative integers m), 1 :$; j :$; ksatisfyingI'= 1jm) = k.
3. Find a non-degenerate kernel h such that Elhl P = 00, some p > 1, but E{h(X l' ... ,
Xk)IX dE !l'q, all q > O.
4. Show that if E h(X 1- ... _ X k) = 0, for 2 :$; j :$; k
)
E{h(X l ,· .. ,Xk)IX 1 ,···,X)} = hiX 1""'X) + I
1=1
hl (X;)
+ I
1 $1, <1,$)
h2 (X I" XI,) + ...
5. (Kemperman) For any countable partition of (-00, (0) into Borel sets B),
j = 1, 2 ... , define the non-negative, symmetric function h by h(X l ' ... , Xk) =
Ij=1 Ujn~=1 [[X/EBil' k ~ 2 and let {X, X n , n ~ 1} be i.i.d. random variables with
P{X E B)} = 'Ttj > O,j ~ 1, where I f 'Tt) = 1. Set 'Tt) = cfj", u) = jP with IX> 1,p > 1.
Prove that if klX - fJ > 1 ~ klX - pfJ and 1X[1 + (k - l)p] > 1 + pfJ, then hE 2 1 ,
268 7 Conditional Expectation, Conditional Independence, Introduction to Martingales
h ¢!I! P' E{h(X 1 " ' " Xk)IX dE !l!p. These inequalities hold, for example, if p=
k - 1 and 1 < IX < [1 + (k - l)p]/k.
6. If 1 < P < (2k/2k - i), there exist degenerate kernels of order i = 1, 2::;; i ::;; k
such that nk(P-l)/PVk.• ~ 0 despite EIE{h(X 1"'" Xk)IX 1"", Xi}IPi = 00, where
Pi is as in (5). Hint: Let {X., n ~ I} be symmetric i.i.d. random variables with
P{IXI > t} = ct- P(1og t)-I, t ~ 2, where 1 < P < 2. Then
is degenerate of order i-I, 2 ::;; i ::;; k. Hint: Set X; = X,IlIX,1 $v'2v1p,], I ~ 1, IX > O.
References
Y. V. Borovskich and V. S. Korolyuk, Theory of V-Statistics, Kluwer Academic
Publishers, Boston, 1994.
L. Breiman, Probability, Addison-Wesley, Reading, Mass., 1968.
H. Buhlman, "Austauschbare stochastiche Variabeln und ihre Grenzwartsatze,"
Univ. of California Publications in Statistics, 3 (1960), 1-36.
Y. S. Chow, "A martingale inequality and the law ofIargenumbers," Proc. Amer. Math.
Soc. 11 (1960),107-111.
Y. S. Chow, H. Robbins, and D. Siegmund, Great Expectations: The Theory ofOptimal
Stopping, Houghton Mifflin, Boston, 1972.
Y. S. Chow, H. Robbins, and H. Teicher, "Moments of randomly stopped sums," Ann.
Math. Stat. 36( 1965), 789-799.
K. L. Chung, A course in Probability Theory, Harcourt Brace, New York, 1968; 2nd ed.,
Academic Press, New York, 1974.
J. L. Doob, "Regular properties of certain families of chance variables," Trans. Amer.
Math. Soc. 47 (1940), 455-486.
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
L. E. Dubins and D. A. Freedman, "A sharper form of the Borel-Cantelli lemma and
the strong law," Ann. Math. Stat. 36 (1965), 800-807.
E. B. Dynkin and A. Mandelbaum, "Symmetric statistics, Poisson point processes and
multiple Wiener integrals," Ann. Statist. 11 (1983), 739-745.
B. de Finetti, La prevision; ses lois logiques, ses sources subjectives," Annales de
/'Institut Henri Poincare 7 (1937), 1-68.
E. Gine and 1. Zinn, "Marcinkiewicz-type laws of large numbers and convergence of
moments for V-statistics," Probability in Banach Spaces 8 (1992), 273-291, Birk-
hauser, Boston.
J. Hajek and A. Rcnyi, "Generalization of an inequality of Kolmogorov," Acta Math.
Acad. Sci. Hung. 6 (1955),281-283.
P. R. Halmos, Measure Theory, Van Nostrand, Princeton, N. J., 1950; Springer-Verlag,
Berlin and New York, 1974.
E. Hewitt and L. J. Savage, "Symmetric measures on Cartesian products," Trans. Amer.
Math. Soc. 80 (1955), 470-501.
W. Hoeffding, "The strong low of large numbers for V -statistics," Institute of Statistics
Mimeo Series 302 (1961), University of North Carolina, Chapel Hill, NC.
D. G. Kendall, "On finite and infinite sequences of exchangeable events," Studia Scient.
Math. Hung. 2 (1967),319-327.
7.5 U-Statistics 269
270
8.1 Convergence of Distribution Functions 271
1
Pd{x}} = -.P{{x}},
(XI
PI{B} = 0, BE 5~· Pi,
then the dJ. corresponding to PI' say F I' is discrete. Moreover, p* =
[1/(1 - (Xt)](P - (XIPI) is a probability measure vanishing on all one-point
sets, whence its corresponding d.f. F* = [1/(1 - (XI)](F - (XtFt) is con-
°°
tinuous. If p* is absolutely continuous (resp. singular) relative to Lebesgue
measure, its dJ. F* may be taken as F 2 (resp. F 3) and (X3 = (iesp. (X2 = 0).
Otherwise, by Corollary 6.5.1, F* = f3F 2 + (1 - f3)F 3' < f3 < 1, where F2
is absolutely continuous and, moreover, F 3 is singular. Thus, F - (XIF I =
(1 - (X t)F* = f3(1 - (X t)F 2 + (l - f3) (l - (X t)F 3 is the asserted decom-
position. 0
In the special case of dJ.s F., complete convergence of {F.} guarantees that
the "limit function" F, if left continuous (as may and will be supposed via
Lemma 8.2.1 even when merely F. ~ F), is a d.f.
272 8 Distribution Functions and Characteristic Functions
P{X n < x} = P {X n < x, IX n - Y"I < t:} + P{X n < x, IX n - Ynl ~ t:}
~Fdx+t:)+P{IXn- Ynl~d
and, analogously,
Thus,
Corollary 2. If {X n}, {Yn}, {Zn} are sequences of r.v.s on (0, iF, P) with
X n 2. X F, Yn X. a, Zn ~ b, where a, b are finite constants, then
X n Y" + Zn 2. aXF + b.
Note. Here, aX F + b is a fictitious r.v. whose distribution coincides with
that of aX + b when X is a bona fide LV. with dJ. F.
such that F(h) - F( - h) ~ 1 - fJ. Then, for all sufficiently large n and any
t:> 0,
P { I Un X nI > t:} ~ P { I Un X nI > t:, 0 < IXnI ~ h} + 2fJ
~ P{lUnl > t:lh} + 2fJ ~2fJ
and the result follows as fJ -+ O. D
j=O
'f
.
+ o(lx - elk),
valid as x -+ c under the hypothesis below (Cramer, 1946, p. 290) leads directly
to
f(X n) = ±(Xn.~
j=O J.
cy f(j)(c) + op(b:).
It will be advantageous to prove the ensuing for nondecreasing functions
Gn on ( - 00, (0). In the special case where the Gn are dJ.s, the condition which
conjoined with weak convergence yields complete convergence is (iii)(y) in
PROOF. Since G.( - 00) :::;; G.(x) :::;; G.( 00), takingx E C(Go) and lettingn - 00,
yielding (i) as x - ± 00. Then (ii) follows immediately from (i). That (a)
implies ({3) is trivial. For the reverse implication, let 6.G n - 6.G o < 00. Then
Go( ± 00) are finite and by (i)
= 6.G o + ITm Gn( - 00) :::;; Go( 00) :::;; lim Gn( 00),
whence lim n_ <Xl Gn( (0) = Go( 00), finite, and so lim Gn( - 00) = Go( - 00),
finite.
Under(y),forany£ > ochoose a = a(£) > Osuchthat6.Gn - 6.G n(a) < £,
n ~ 1, for a ~ a. Then if ±a E C(G o),
ensuring 6.G o < 00 by (ii), and since £ is arbitrary, ITm n 6.G n :::;; 6.G o . In
conjunction with (ii), this yields ({3). Conversely, under ({3), for any £ > 0
choose the integer nl such that n ~ nl entails 6.G n - 6.G o < £ and select
a > 0 with ±a E C(G o) such that 6.G o - 6.G o(a) < £. Then for n ~ some
integer n2, 6.G oUi) - 6.G n(a) < £, implying for n ~ no = max(nl, n2) that
Choose aj such that 6.G j - 6.G j(aj) < 3£, 1 :::;; j < no, whence for a ~ a' =
max (a, at> ... , ano-I)
sup[6.G n - 6.G.(a)] < 3£,
n~1
lim fb g dF n
= fb g dF. (1)
n- 00 a CJ
Then
Hn == fgdFn - f9dF
+ [S;"'9(Xi)dFn(X) - S;"'9(XMF(X)]
+ [fi"g(XMF(X) - fi"g(X)dF(X)]}
Gn(x) = fg dF n •
Clearly, (i) iffand g are u.i. relative to {F,,}, so aref+ and af + bg for any
finite real numbers a, b. (ii) iff, g are continuous, If I ~ Igl, and g is u.i.
relative to {F,,}, so is f.
Thus, in the case of dJ.s, Lemma l(iii) may be rephrased as follows:
If the dJ.s F" ~ F, then F,,':' F iff {F,,} is tight iff every bounded continuous
function g is u.i. relative to {F,,}.
(3)
(ii) If the dJ.s F" ~ F and g is continuous on ( - 00, (0) with lim y _ ± ao g(y)
= 0, then Eq. (4) holds.
8.1 Convergence of Distribution Functions 277
PROOF. This follows from Corollary 5 since for 0 ~ r < s and some C in
(0, (0)
[
J[lxl~al
lxi' dF.(x) ~
a
.t-, flx1s dF.(x) < a~" n~1. D
nr
PROOF. Throughout, in addition to any other requirements choose a > 0 so
that ±a E C(Fn ). If 9 is u.i. relative to {F n } and 8> 0, select a so that for
n ~ 1 it also satisfies the first inequality of
8> f ~al
Ilyl
Ig(y)ldF n = [
J UXnl ~a)
Ig(Xn)ldP ~ [
JUg(Xnll >b]
Ig(Xn)1 dP, (5)
whereb = max{ Ig(y)l: Iyl ~ a} and the equality holds via Theorem 6.2.4 and
Corollary 6.2.1. Thus {g(X n )} are u.i.
Conversely, in case (i), as a -+ 00 there exists K = K a -+ 00 such that
and then select a> 0 so that sUPn~1 P{IXnl ::2': a} < e/b. Then, for n::2': 1
f.
lIyl~al
Ig(Y)ldFn(Y) = i
[IXHI~al
Ig(Xn)1 dP ~ e + b P{lXnl ::2': a} < 2e,
Corollary 8. For some p > 0, let {X n, n::2': I} be Ifp r.v.s on (0, f/', P) with
X n ~ X F· Then EIXnlP -+ EIXFIP,jinite iff {IXnIP, n ::2': 1} is u.i.
If f/'* denotes the class of all d.f.s on R, many distances may be defined
on f/'*. One prominent choice is d*[F, G] = SUpxERIF(x) - G(x) I (see
Exercise 2). The Levy distance d[F, G] corresponds to the maximum distance
between F and G measured along lines of slope - 1 (in contemplating this,
draw vertical lines connecting F and also G at any discontinuities) multiplied
by the factor 1/.j2. Formally,
d[Fn, F] = inf{h > 0: F(x - h) - h ~ Fn(x) ~ F(x + h) + h, all x}. (6)
Theorem 3. Let {F, F., n ::2': 1} bed.f.s. Then (i) Fn '; F iff (ii) J9 dF n -+ J9 dF
for every bounded, continuous function 9 iff (iii) d[F., F] -+ 0 iff (iv) rrm Fn{C}
~ F{C}, lim F.{V} ~ F{V} for all closed sets C and open sets V, where
Fn{·}, F{·} are the probability measures determined by F., F respectively.
PROOF. That (i) implies (ii) is the Helly-Bray theorem, Corollary 6. To
show that (i) implies (iii), for any e > 0 choose a, bE C(F) such that e/2
exceeds both F(a) and 1 - F(b) and then select ajE C(F), 0 ~ j ~ m,
with ao = a < al < ... < am = band laj - aj-II < e, 1 ~ j ~ m. Deter-
mine N j , 0 ~ j ~ m, so that n ::2': N j entails IFn(a) - F(a) I < e/2 and set
N = maXO,;;j,;;m N j • Let n > N. If x ~ ao,
e
F.(x) ::2': 0 > F(x) - 2 ::2': F(x - e) - e,
e
Fn(x) ::2': F.(aj_,) > F(aj_ I) - 2 ::2': F(x - e) - e.
To verify that (iii) implies (i), for any X o E C(F) and e > 0 choose 1J > 0
such that Ix - Xo I :$ 1J entails IF(x) - F(xo) I < e. Set h = min(e, 1J) and
select N so that d[F., F] < h when n ~ N. Then, for n ~ N from (6),
Fn(xo) :$ F(xo + h) + h :$ F(xo + 1J) + e :$ F(xo) + 2e,
Fn(xo) ~ F(xo - h) - h ~ F(xo - 1J) - e ~ F(xo) - 2e,
and (i) follows.
To obtain (i) from (ii), retain the choice of xo, 1J, e; define
I
I, x :$ Xo - 1J
h(x) = x o ; x, Xo - 1J :$ x :$ Xo
0, x> xo,
= h(x), hix) = h(x
r:-\I r:
and set hl(x) - 1J). For any dJ. G
Hence,
280 8 Distribution Functions and Characteristic Functions
If X F is a fictitious LV. with dJ. F and 9 is a finite Borel function on (-00, 00),
it is natural to signify by g(X F) a fictitious r.v. with dJ. G(x) = F {g-l( -00, x)},
where, as earlier, F{·} represents the probability measure determined by the
dJ. F(·).
EXERCISES 8.1
I. Let F be the d.f. ofa LV. Y. where pry = I} = pry = O} = 1and define X n == Y,
X = I - Y. Verify that X n 2.. X but X n -¥. X. Also prove that if X n 2.. X F' where F is
degenerate at c (i.e.• F ( {c}} = I), then X n !'., C.
3. If X and Yare r.v.s with d.f.s F and G and P{IX - YI ~ £} < £, then the Levy
distance d[F, G] :::;; £.
4. If dJ.s F n .s. F 0 and mn is the unique median of F n' n ~ 0, prove that mn --> mo. Can
an analogous statement be made if the medians are not unique?
8.2 Weak Compactness, Frechet-Shohat, Glivenko-Cantelli Theorems 281
5. Define X = 2 If= 1 X/Y, where the {Xj} are i.i.d. r.v.s on some probability space
with P{X I = I} = P{X I = O} = 1. Then 0 s X s I, and if XI = I, then X ~~,
while if X I = 0, X s 2 If= 2 r j = t. Verify that the dJ. F of X satisfies F(x) =
rkFWx),O < x < r\ k ~ I, and that F is a singular dJ.
6. If {S., n ~ I} is a sequence of binomial LV.S with p.dJ. b(k; n, p), find the density and
dJ. F such that F. ~ F, where F. is the dJ. of(npq)-l(S. - np)2.
7. If g(x) = x", ex> 0, and P(X. = a.} = I/n = I - P(X. = OJ, is g u.i. relative to
F. = F X n if a. = a· nil", a > 0; if a. == a?
10. If LV.S X.!. X and IglP is u.i. relative to Fxn , then g(X.).::4 g(X).
II. If (F.(x) = J~ <X> f.(y)dy, n ~ O} are absolutely continuous d.f.s with f.(x) --+ fo(x),
a.e., then F.{B} --+ Fo{B} uniformly in all Borel sets B. Hint: J'.'.'<X>If,,(x) - fo(x)/dx
--+ O.
12. FinddJ.sF. ~ Fanda Borel set Bsuch that F{B} = O,F.{B} == 1. Hint: IfY.,n ~ I
is a sequence of binomial r.v.s with parameters nand p = t, let F n be the d.f. of
[Y., - (n 2 /2)]/(n/2).
13. If dJ.s F. ~ F, where F is continuous, then F. converges to F uniformly on ( - 00, 00).
14. Ifforn ~ 1 and lui < UoE(O, 00), them.gJ.s q>.(u) = J'.'.'<X>e"X dF.(x) < g(u) < 00 and
F.':' F, then q>.(u) --+ J'.'.'<x> e"X dF(x), finite for lui < u o'
15. Let S. = Ii Xi where {X., n ~ I} are i.i.d. r.v.s with E Xl = Jl. > O. If N = N p is
an {X.} - time (or a LV. independent of {X., n ~ I}) having the geometric distri-
bution then limp_o P{SN/E SN < x} = I - e- x , x> O. Hint: Recall Exercise 3.1.15.
16. Let H k ('), k ~ I, denote the Hermite polynomial of degree k, which satisfies
HHdx) = xH k(x)-kH k _ 1(X), k~l, with Ho(x) = I, Hdx)=x. If Sk,.=
II :5i, <"'<i.:5. X i,··· Xi., n ~ k ~ I, where {X, X., n ~ I} are i.i.d. with E X = 0,
-
E X 2 = 1, prove that k! Sk,./n kl 2 --+
d
Hk(Z), where Z is a standard normal random
variable. Hint: Use induction, Exercise 7.5.1, and Theorem 8.1.1.
G(x) ....... G(Yn) :::; F(Yn+ I) :::; F(x) :::; F(x n) :::; G(x n) -+ G(x), 0
yielding the final statement of the lemma.
Lemma 2. (i) Every sequence of dJ.s is weakly compact, that is, contains a
subsequence c;onverging weakly to a left-continuous function. (ii) A sequence of
dJ.s {F n} is completely compact, i.e., every subsequence contains a further
subsequence converging completely, iff {F n' n ~ I} is tight.
(iii) A sequence of d.f.s {F n} converges completely (resp. weakly) to F iff
every subsequence of {F n} has itselfa subsequence converging completely (resp.
weakly) to the same function F.
PROOF. Let D = {r j } be a countable dense set, say the rationals. Since
0:::; Fn(r l ) :::; I, there exists a convergent subsequence {Fn'i(rl),j ~ I}. Then
O:s; Fn'i(rZ) :s; I and there exists a convergent subsequence {Fn2irz)} of
{Fn,/rz)}. Continuing in this fashion, the diagonal sequence {Fnjj' j ~ I}
converges to a bounded nondecreasing function G on D, whence by Lemma 1
F(x) = lim G(y)
Apropos of (iii), let x E C(F) and suppose that Fix) does not tend toward
F(x) as n -> 00. Then there exists a subsequence F n· for which lim Fn·(x) exists
but differs from F(x), thereby precluding Fn • from having a subsequence
converging completely (or weakly) to F. The remaining portion of (iii) is
trivial. 0
Lemma 8.1.3 ensures the following
and since as n --+ 00 the right side --+ 11k, which is arbitrarily small for large k,
the left side --+ 0 as n --+ 00. 0
If X I' X Z' ... , are i.i.d. r.v.s, the sample or empirical d.f. F~ based on
Xl' ... , X n is defined by F:;>(x) = (lIn) D=
1 I[X j <x)(w). Note that for all n
and x, F~(x) is a r. v. and for every n and almost all w, F~(x) is a dJ. Therefore,
for almost all w, {F~(x), n ~ I} is a sequence of dJ.s.
Definition. Two d.f.s F and G are said to be of the same type if for some
positive a and real b
G(x) = F(ax + b). (I)
Theorem 3. If {F n} are dJ.s such that for some positive constants an and real
constants bn
where F and G are nondegeneratedJ.s, then F and G are ofthe same type, that is,
(1) holds, and an--+ a, bn --+ b.
8.2 Weak Compactness, Frechet-Shohat, Glivenko-Cantelli Theorems 285
PROOF. Note at the outset that if H is any nondegenerate dJ. with H(a l x + b l )
= H(a2x + b 2) for all x, where aj and bj are real,j = 1,2, and a l + a2 ¥- 0,
then al = a2, b l = b 2. For if alx + b l ¥- a 2x + b 2, then
x' = t [(al + a2)x + b l + b 2]
cannot be a point of increase of H(x). Thus, if x' is a point ofincrease,alx + b l
= a2 x + b 2 and, since there are two distinct points of increase, necessarily
al = a2,b l = b 2·
The hypothesis (and proof) is easily couched in terms of LV.S, namely,
X n ~ X F (nondegenerate) and y" = a;; I(X n - bn) ~ YG (nondegenerate),
where an > O. Let a be a finite or infinite limit point of {an} and {n'} a sub-
sequence of the positive integers {n} with an' --> a; let b be a finite or infinite
limit point of {b n.} and {nO} a subsequence of {n'} with bn·· --> b.
If a = 00, Xn·la n· ~ 0, whence by Theorem 8.1.1 -bn"la n" = Yn" -
(Xn"la n,,) ~ YG , which is impossible since YG is nondegenerate. Likewise,
a = 0 is precluded since this would entail bn" = X n" - an" Yn" ~ X F' Thus,
o < a < 00 and, since (an" - a) y"" ~ 0,
aYn" + bn" = X n" - (an" - a)Yn" ~ X F ,
which ensures b finite in view of a Yn" ~ a YG . Thus, by Corollary 8.1.3
YG ~ y"" = a,;;.I(X n" - bn,,) ~ a-I(X F - b)
or G(x) = F(ax + b). By the remark at the outset, no other subsequence of
{b n·} can have an alternative limit point b', that is, lim bn· = b. Analogously,
if {n*} is a subsequence of {n} with an' --> a*, bn• --+ b*, the prior reasoning
requires F(ax + b) = F(a*x + b*), which, by the initial comment, entails
a = a*, b = b*. Thus an --> a, bn --> b. D
Corollary 2. If Fn(anx + bn) -=. F(x), Fn(Ct.nx + Pn) -=. G(x), where F, G are
nondegenerate, an' Ct. n > 0, then F and G are related by (1) and Ct.nla n --> a,
(Pn - bn)lan --+ b. In particular, if G = F, Ct. n -- an' Pn - bn = o(a n)·
EXERCISES 8.2
I. Give an example showing that Theorem 3 is invalid without the nondegeneracy
hypothesis.
2. Verify that the class 9'* of all dJ.s on R is a complete metric space under the Levy
distance.
3. Let 1\ be a Borel subset of Rm where R = ( - co, co), m ;:::: 1, and for every AE 1\ let
F(x; A) be a dJ. on R such that F(x; A) is a Borel function on Rm + I. If G is any dJ.
whose support c 1\, show that H(x) = .fA F(x: A)dG(A) is a dJ on R. It is called a
286 8 Distribution Functions and Characteristic Functions
G-mixture ofthe family.F = {F(x; i.), i. E ;\} or simply a mixture. Convolution is the
special case F(x; A) = F(x - A), m = I. If EH and Ef ; denote integration relative to
Hand F(x; A) respectively, show for any Borel function cp with EH Icp I < 'X that
EH[cp] = JEdcp]dG(A).
4. Let Q be the planar region bounded by the quadrilateral with vertices (0,0), (I, I),
(0, 1), (1, I) and the triangle with vertices (1,0), (I, 0), (I, t). If X, Yare jointly uni-
formly distributed on Q, that is, P is Lebesgue measure on the Borel subsets of Q,
prove that F x + y = F x * F y despite the nonindependence of X and Y.
6. The convolution of two d.f.s is absolutely continuous ifat least one of the components
is absolutely continuous. The converse is false (Exercise 8.4.6).
7. If {X., n :;:.: I} are i.i.d. with a uniform distribution on [0, I], that is, F(x) = x for
0::; x ::; I, show that n(max i SjS. Xi - I) 2.. X G •
Demonstrate that if F = F 1 * F 2' Q I (a12) . Q2(a12) ::; Q(a) ::; minj = I. 2 Qj(a), and
deduce that F I * F 2 is continuous iff F I or F 2 is continuous.
9. Let F i be a discrete dJ. with maximum jump qj, i :;:.: I, and suppose that G. =
FI * F 2 * ... * F. ~ G. Prove that if G is continuous, Oi= I qi = 0(1).
10. If flJ2 are densities, their convolution is defined by
f = fl * f2 = J
fl(X - y)f2(y)dy·
Verify that if F j is an absolutely continuous dJ. with density Jj,j = 1,2, then F =
F I * F 2 is absolutely continuous with density f = II * I2'
11. (Chow-Robbins) Prove that if F is a dJ. with F(O) = 0, then G(x) = F(x + n) is Of
J
a dJ. iff x dF(x) <x. More generally, Ghas a finite kth moment iff F has a finite
(k + I)st moment.
12. If X and Yare independent r.v.s such that X + Yand X have identical dJ.s, then
Y = 0, a.c. Hint: Employ the Levy concentration function of Exercise 8.
lim -2
C-oc n
1
-c
I C
e- Ita -
it
e- Itb
</J(t)dt
_ P{
- a<
X
<
b}
--- = a} +
+ P{X 2
P{X = b}
. (3)
If a, bE C(F), where F = F x' the right side of (3) reduces to F(b) - F(a).
The integrand of (3) is defined at t = 0 by continuity and is not, in general,
absolutely integrable over ( - 00, 00).
PROOF. Set
I(C) = -2
1 fC e- ita -. e- itb </J(t)dt = -21 fe e- ita - e- irb E eUx. dt.
n -c It 1r -c it
288 8 Distribution Functions and Characteristic Functions
Since [(e- ita - e-ilb)/it]eiIX is bounded for all wand t, by Fubini's theorem
I fC eit(X - a) _ eit(X - b)
I(C) = -2 E dt
n -c it
=-
1sin t
-dt - E
[fC(X - a) fC(X - b) sin t
-dt
J= EJdX),
not 0 t.
Ifc(u-a) . {I,
where
a<u<b
- sm t d ('-00 1.
( )-
J CU - -- t ~ 2' u = a,b
n C(u-b) t 0
, u < a, u > b.
Since IJdu)1 ~ 2 for all - 00 < u, C < 00, by dominated convergence
lim I(C) = lim E JdX) = E lim JdX)
C-+C() C-oo C-oo
1 P{X = a} + P{X = b}
= E[2 I [x=aorbl + I[a<x<bJ] = 2 + P{a < X < b}.
o
Corollary 1. There is a one-to-one correspondence between dJ.s on Rand
their cJ.s.
PROOF. The very definition (1') shows that identical dJ.s have identical d.s,
while the converse follows from
I fC e- i1a - e- i1b
F(b) = lim lim -2 cp(t)dt, bE C(F). o
a--oo c-oo n -c it
Corollary 2. If J~ 00 Icp(t) Idt < 00, then for - 00 < a< b< 00
-2
I foo e- ita - e- itb
. cp(t)dt = P{a < X < b} +
P{X = a} + P{X = b}
2
n _00 It
= F(b) - F(a),
(3')
where the corresponding dJ. F is absolutely continuous with a bounded,
continuous density
f(x) = F(x) = -2I
e- Ilx cp(t)dt. foo (4) .
n _00
PROOF. In view of the hypothesis, (3) transcribes as the first equality in (3'),
whence, letting b -+ a +, Lebesgue's dominated convergence theorem ensures
P{X = a} = O. Hence, F is continuous, yielding the second equality. Since
for b approaching a the integrand of (3') divided by b - a tends to e-itacp(t),
(4) follows by dominated convergence. Boundedness of f is apparent from
8.3 Characteristic Functions, Inversion Formula, Uvy Continuity Theorem 289
f bf(x)dx =
a
1
-2
1t
fb dx fex>
a -ex>
.
e-llxcp(t)dt
= 2~ f~ex>(fe-iIX dx )cp(t)dt
I fex> e- il• - e- i1b
= -2 . cp(t)dt = F(b) - F(a)
1t _ ex> It
by (3'). Since f is continuous, f(x) ~ 0 on (- 00, (0) and, letting b -+ 00,
a-+ -OO,fE2'I(-OO, (0) and Ilfll = 1. 0
Theorem 2. If Fi is a dJ. with corresponding c.f. cpj, i = 0, I, 2, then F 0 =
F 1 * F 2 iff CPo = CPI . CP2·
ilZ
PROOF. Suppose F 0 = F t * F 2. Since e is bounded and continuous, by (II)
of Section 6.3
Conversely, if CPo = cP 1 • CP2 and F = F 1 * F 2 has c.f. cP, then by the portion of
the theorem just proved cP = CPl· CP2 = CPo, whence by Corollary 1, F o =
F = F 1 * F 2. 0
Corollary 3. If X I' X 2 are independent LV.S on some probability space with
cJ.s cP x I' cP X 2' then the cJ. of their sum is
(5)
(6)
t-: t-:
where
IIII = I e
i1x
dFn(X) - e
i1x
dF(x) I :$; F n( - M) + F( - M) < 3e,
M
= [Fn(x) - F(x)]e iIX 1 - it fM ei'X[Fn(x) - F(x)] dx,
-M -M
(7)
Since lim Fn(x) = F(x) except on a set of Lebesgue measure zero, the Lebesgue
dominated convergence theorem ensures that for n ;;:: N 2 the last term in (7)
is less than e. Hence, for n;;:: N = max(N1 , N2 ) and It I :$; T, II 3 1 < 7e, and
so from (6), qJn(t) -+ qJ(t) uniformly in It I :$; T.
Conversely, suppose that qJn(t) -+ g(t) where g is continuous at t = O.
By Lemma 8.2.2 there is a monotone, left-continuous function F and a
subsequence {F n,} of {F n} such that Fn; ~ F. Then for any (j > 0
(8)
I
2(j
fd
_/(t)dt =
foo -00
sin (jx
~ dF(x),
8.3 Characteristic Functions, Inversion Formula, Uvy Continuity Theorem 291
and so, as b --+ 0, by continuity of 9 at the origin and, once more, dominated
convergence,
F( 00) - F( - 00) = g(O) = lim lpn(O) = l.
Thus, F n; ~ F, whence by the portion of the theorem already proved lpn. --+ lpF'
implying 9 = lpF' Analogously, every subsequence of {F n} has itself a sub-
sequence converging completely to adJ. F* and lpF* = 9 = lpF, entailing
F* = F. Consequently, Lemma 8.2.2(iii) ensures F n -=. F. 0
Corollary 5. If {F, F n, n ~ I} are d.f.s with corresponding d.s {lp, lpn' n ~ I},
then Fn -=. F iff lim lpn = lp.
The few but powerful theorems concerning d.s already established permit
the extension of Theorem 3.3.1 via the ensuing
L[I -~{lp(t)}]dt
lp,
2)-2
f b
o (l - cos Cu)du ~ [)3
(
b +; (10)
>
-
i!IXI"Cj
[)3 ([) + ~) - 2 dP
IXI
~b 3
(b + ~r2p{IXI ~ C}.
To this end, choose the minimal [) 1 ~ [) such that C[) 1 = 2kn for some positive
integer k. Then [)l < [) + 2n/C and
fo
Cb
(l - cos u)du = L
k f
2j"b 1b
' (l - cos u)du.
j=l 2(j-l)"blbl
Now for 0 < b < 2n and any real a
f
O+b fbl2 fb l2
cos u du ~ cos u du =2 cos u du,
o -N2 0
Cb
fo (1 - cos u)du ~ 2k
f1tb/bl
0 (1 - cos u)du = 2k [(no)
~ - sin ~
(no)]
0)3 Co 3 Co 3
~ 2kn ( 0 1 = oi ~ (0 + 2njC?
and (10) follows. o
(O,~, P) andfor some T >
Corollary 6. If {X n} is a sequence ofr.v.s on the °
c.f.ofX m - X n,saYCfJm.n(t),convergesto 1 uniformly in It I ~ Tasm > n -+ 00,
then X n ~ some LV. X on (O,~, P).
that Sn~S,
PROOF. For n ~ 1, let CfJn' t/Jn, and CfJ be the d.s of Sn, X n and F respectively. By
°
hypothesis and Theorem 3, CfJn(t) -+ CfJ(t) uniformly in every bounded interval.
Hence, for any 6 in (0, !) there exists T > and an integer N > such that for
It I ~ T and n ~ N
°
ICfJ(t) I > !, ICfJn(t) - CfJ(t) I < 6, ICfJn+k(t) - CfJn(t) I < 26, k ~ 1, (11)
n+m I = ICfJn(t)!ICfJn(t)
1
CfJn+ m(t) I < 46, m~l.
I 1 - nn t/JJ{t) -
EXERCISES 8.3
I. If q> is a c.f. so are I q>(t) 12 and .J4l (q>(t)} where .J4l {z} denotes the real part of z. Find the
c.f. of the singular dJ. of Exercise 8.1.5.
2. If Fn, Gn are dJ.s with F n -"-> F, Gn ':" G, then F n * Gn -"-> F * G.
3. If H is a G-mixture of.:1' = (F(x; A), A E t\} as in Exercise 8.2.3 and q>(t, A) is the d.
corresponding to F(x; A), show that q>H(t) = SA q>(t; A)dG(A). Thus, if (q>n(t), n 2: I}
is a sequence of c.f.s and Lf Cn = I, Cn 2: 0, q>(t) = Lf Cjq>j is a c.f. Verify that
exp{A[q>(t) - I]}, A> 0, is a c.f. if q> is.
5. Prove that a real-valued c.f. cp satisfies cp(t + h) ~ cp(h) for t > 0 and sufficiently
small h > O.
6. Verify that the d. corresponding to a uniform distribution on (-IX,:X) is cp,(t) =
(sin IXt)/:Xt, IX > 0, and that lim cp.(t) exists but F" '!4 to adJ.
7. Find thed.cpx ifP{X = IX} = P{X = -IX} = -!-. Show by iteration of an elementary
trigonometric identity that
-
I ..
fn e-UJcpx(t)dl = P{X = j}.
2n -n
Give an analogous formula for S" = D= I Xi> where {X;} are i.i.d. integer-valued
LV.S.
f
10. Prove that
I e-i,xcp(t)
F(x + ) + F(x - ) = I - - . dt
n It
where
Ie) ,
f (f = lim
c ..... o -c
-<
+
£
for F continuous.
12. Prove for any dJ. with d. cp that for all real x and h > 0
-
I fX+ 2h F(y)dy - -
I fX F(y)dy = -
I foo (Sin
-
U)2 .
e-·uxtPcp(u/h)du.
2h x 2h x- 2h n - 00 u
Hint: Apply the Levy inversion formula to F * Gh , where Gh is the uniform dis-
tribution on [-h, h].
14. Let H,(x) = SA F(x; A)dGi(A) be a G,-mixture (Exercise 8.2.3) of the additively
closed family $' = {F(x; 1), lEA c R m } (Exercise 8.4.5). Verify that the convolu-
tion HI * H2 is a (G I * G2 )-mixture of $'. Hint: Utilize Exercise 3.
15. If, in Exercise 14, m = I and A = [0, (0) or [0, I, 2, ...] or (nonnegative rationals},
then the mapping of ~ = (G: G(A} = I} onto .Yt = (H: H(x) = SA F(x; l)dG(A.),
G E~} is one-to-one. Then.Yt is said to be identifiable. Hint: ljJ(z; G) = .fA ZA dG(l) is
analytic in 0 < Izl < I and tp(t; 1) = tpA(t), 1 E A where tp(t) = tp(t; I).
Symmetric I
(I, -I} cos t
Bernoulli 2
Binomial G)pX(1 - p r x xE(O,I, ... ,n} (pe i ' + 1 - p)",O < p < 1
Poisson lx~
-A
x!
XE (0, I, ...} exp(l(e i' - I)}, 1> 0
Normal I
--exp
(J$
{-(X-tW }
20'2
(- 00, (0) ei8 ,-a 2,2/ 2 , (J > 0
Symmetric
Uniform
-
I [ -rx, IX]
sin rxt
- - rx>O
2rx IXt '
~ (I - 1:1)
2(1 - cos rxt)
Triangular [-IX, IX] , rx > 0
IX 2 t 2
Inverse
Triangular
I- cos rxx
(- 00, (0) [I---;. ' It'r rx > 0
1trxx 2
A
Gamma
IX
[(1) x
),-1 -(U
e
(0,00) (I - ~rA, 1> 0, IX> 0
rx
Cauchy (- 00, (0) e-· III , rx > 0
1t(rx 2
+ x 2)
Ifl
00,
" (. \i (. )"+
ei. - " I t ) _ It
L- -.-, - - - , - -u du
ei'U(I)"
i=O J. n. 0
(1)
(')"+Ifl
_It__
, e
ilU(1 _
u
)" d
u
=~,
(.)" (.)" fl
+ ( _It I)' e
ilU(1 _
u
)"-1 d
u,
n. 0 n. n . 0
the first part of (1) follows by summing while the second is obtained in-
ductively.
To prove (2), let I" denote the left side of (I). Since Ieil - II = 21 sin t/21 ~
1 O
2 - lt10 for 0 ~ fJ ~ I, let n ~ I, whence from (1)
~2
1 lf 'l f'"+1 f'
-0 0 0 . .. 0
3
t~ dt 2 ... dt"+ I
2
1
-
= (I + fJ) ... (n + fJ)'
O
ltl"H
0
PROOF. (2) with {) = 1, n = 0 yields the bound 2(elzl - 1). However, directly
le
z
- II = I
I~ zi/j! ~ ~ Izli/j! = elzl - 1. 0
Theorem l. If X is a LV. with d. ({J and EIXI"H < 00 for some nonnegative
integer n and some {) in [0, I], then ({J has continuous derivatives q>(k l of orders
k :s; nand
I ~ k ~ n, 0)
q> ()
t = ~(it)iEXj
L. ., + 0 (II")
t as t -+ O. (5)
j=O J.
296 8 Distribution Functions and Characteristic Functions
i'i'k i'2.
PROOF.
e ilX - L -.-,-
k-t (itXY = (iX)k ... e"' X dt t ... dt k, (6)
j=O ). 0 0 0
so that
and
- +
IjJk(t
--, h)---IjJk(t)
- - -_
h
E('X)k
I -1
h
f'+h i'k
I 0
... i 0
'2
e i'IX d tt ... d t k.
Since
2
I~ f+h f~k ... { e i/ IXdt t ... dtkl :5: (It I + Ihl)k-t = 0(1) as h -+ 0,
(7)
and, in particular,
cp'(t) = ljJ't(t) = i E Xe itX .
Repeating the previous argument, it follows via (7) that for 2 :5: k :5: n
2
d 1jJ (t) = ik E X k
1jJ~2)(t) = --~-
dt 0 0
...
0
i'i'k-2 it2
eitlX dt t ... dt k- 2,
IW-1)(t) = jk E X k {e itlX dt t ,
and cp has a continuous nth derivative, whence (5) follows from Taylor's
theorem.
8.4 The Nature of Characteristic Functions 297
To prove the converse, define for any finite function g for - 00 < x, h < 00
t 2n
Ip(t) = 1p(0) + tlp'(O) + ... + (2n)! [1p(2n)(0) + 0(1)],
Proposition 1. A d. Ip(t) is real valued (for real t) iff its dJ. F is symmetric.
PROOF. The proof is most easily couched in terms of LV.S. If F is symmetric,
that is, X and - X are identically distributed,
Ip(t) = Ipx(t) = Ip-x(t) = Ipx( -t) = Ip(t),
so that Ip is real. Conversely, if Ip is real,
Ipx(t) = Ipx(t) = Ipx( - t) = Ip - x(t),
and so by the one-to-one correspondence between d.s and dJ.s, X and - X
are identically distributed, that is, F is symmetric. 0
PROOF. Since 1cp(t o) 1 = 1, cp(t o) = ei80 for some real (Jo, whence 1 = e - i8ocp(t O)
= J ei(l ox- 8o) dF(x) = Jcos(tox - (Jo)dF(x). In r.v. parlance,
E[1 - cos(t o X - (Jo)] = 0,
necessitating
P{toX - (Jo = 2jn,j = 0, ± 1, ...}
2jn + (J 0 .
=p{ X = to
}
,J = 0, ± 1,... = 1.
( kl) = -;(JI -
2n jl - -; (Jo = 2n (k 2
j2 - -; ) ,
then
PROOF. If an -r> 00, there is a subsequence {n'} of the positive integers {n}
(which, for notational simplicity, will be replaced by {n}) such that an -+ a,
finite. By Corollary 8.1.3, Sn - bn = an(Sn - bn)/an 2.. aSG' Thus, if t/J and cP
are d.s respectively of XI and aSG' by Theorem 8.3.3 e-ibn't/Jn(t) -+ cp(t).
Clearly, this entails cp(t) = 0 for all t such that It/J(t) I < 1. By Corollary 2,
cp(t) = 0 for all except perhaps countably many isolated values of t, whence
cp(t) == 0 by continuity, a contradiction. Thus, an -+ 00, whence X dan ~ 0
and likewise X Jan ~ O. Theorem 8.1.1 then ensures that (Sn- 1 - bn)/an 2.. SG'
which in conjunction with (Sn-I - bn-I)/an-I 2.. SG necessitates
by Corollary 8.2.2. D
The next proposition reveals inter alia that CPa(t) = e-/'I is a c.f. for
o < ex ~ 1 but fails to detect that CPa is likewise a d. for 1 < ex ~ 2, a fact
which will be established in Theorem 12.3.2.
() t/Jj-Itj - t/Jjt j - I
Yj\t = t/Jjt -- t/Jj-I
tj_ 1
t + -'---'-----"--''-----'---'~...e:-
t - t _
j j j 1
m
L Cj = Yj(O),
i=j
300 8 Distribution Functions and Characteristic Functions
Then
m C. m
Yj(t) = - t .L. -;f + .Lei>
l=) A. 1=)
./,
'l'j+1
./,
+ 'l'j-I ./, = (t
~ 2'1'j
j + 1t+ t 1) ./,
j -
'l'j
j
by convexity, implying y/O) ~ Yj+ 1(0). Thus, Cj = y/O) - Yj+ 1(0) ~ 0 and
Lj=1 Cj = YI(O) = 1. Furthermore,ljJm(t) == Lj=l cp - (ItI!Aj)t coincides
with Yj(ltl) for t j _ 1 S; It I S; t j , 1 S; j S; m, and ljJm<t) is a d., m ~ 1, via
Exercise 8.3.3 and the table of c.f.s. By hypothesis ljJ(O) = ljJ(O+ ), whence
ljJ(t) = Iim n .... co ljJn(t), and so ljJ is a d. by Theorem 8.3.3. 0
sin 2t sin t
- - = --cost.
2t t
Definition. A family g;' of dJ.s is factor closed if for all FE g;' the convolu-
tion relation F = G 1 * G 2 with G i a nondegenerate dJ. implies that G i E g;',
i = 1,2.
The theorem of Cramer-Levy may be rephrased by saying that the family
of normal d.f.s is factor closed. The only known method of proof is via d.s and
requires a discussion of analytic d.s.
8.4 The Nature of Characteristic Functions 301
whence
f
j=1
2j 1
P2j_ 1 t - <
(2j - I)! - j=1
f (l + P2)t
(2j - I)!
2j
-
1
< '+ d
- e dt
i (2j)!
I
2j
P2j t <
00
so that (iii) = (iv). Finally, if (iv) obtains, then, employing the language of
random variables, (9) implies that
E f I(itX)jl
j=O j!
= Ee 1tX1 < 00 for It I < r,
<p(t) = E eitX = f
j=O
E (it~)j =
).
f
j=o}'
IXJ~i,t)j. o
Corollary 5. A d. <p with dJ. F is entire iff J~ 00 er1xl dF(x) < 00 for all r > O.
The next proposition asserts that a d. q> is r-analytic iff q>(z) is defined and
analytic in the strip 1.1"(z) I = 1y 1 < r ofthe complex plane. Consequently, the
statement that <p is r-analytic may also be interpreted as asserting that <p(z)
as defined by (10) is analytic in 1.1"(z) I < r.
Proposition 5. A d. <p with dJ. F is r-analytic iff <p(z) (in (10» is an analytic
function in 1.1"(z) I < r, and in this case for k = 0, 1,2, ... ,
OO'j j-k
(k)( )
z
= "L. I ajz Izi < r, (1t)
q> j=k j - k) ,'.
(.
b. (izy fb .
= f e'zx dF(x) = L -.-,
00
g(z) Xl dF(x),
a j=O j. a
and since the latter series converges uniformly in 1z I ~ M for all 0 < M < 00,
g(z) is entire and for k = 0, 1, ...
j-k fb fb
g(k)(Z) = L 00
~z , x
j=k(j-k).
'j
a
j
dF(x) =
a
(ixte izx dF(x).
Then by the preceding, g.(z) is entire, n ;::: 1, and for 1.1"(z)\ ~ s < r
f:
.=m
19n(Z) \ S; [
Jllul 2:am_ d
e1u.1( zlI dF(u) S; [
Jllul 2:am_ II
~luldF(u) < 00
by Proposition 4. Thus, If
g.(z) converges uniformly in 1.1"(z) 1 S; s < r,
whence (Titchmarsh, 1932, p. 95) <p(z) = J~ 00 eizu dF(u) = Lf 9.(Z) is
analytic in 1.1"(z) 1 < r. Moreover, for 1.1"(z) I < rand k = 0, 1, ... , setting
ao = 0,
= f [
n=1 J[an-,,;lxl<anl
(ixte izx dF(x) = E (iX)keiZX
8.4 The Nature of Characteristic Functions 303
and so, via (and in the notation of) Proposition 4, for 1z 1:::;; s < r
f J=kf IUik·(i~k-~). i
n=1 [0.-1,,1"1<0.)
j
u dF(u) I
co ~-k dk co p.~
: :; j=k
L U _ k)I. P = dk Sj=O)'
L ~- < 00,
j
whence
(k) _ ik(izy-k..
.L (. _ k)IC1..
00
Proposition 6. (i) If F is a dJ. whose d. CfJ is r-analytic for some r > 0, then F
is uniquely determined by its moment sequence {IXj = xj dF(x),j ~ 1}. J
(ii) A d. CfJ with moment sequence {IX j , j ~ I} is r-analytic for some r > 0 iff
~ (IX 2n)I/2n
hm 2 < 00. (12)
n-+ 00 n
j
PROOF. (i) Denote F, CfJ by F I' CfJI and suppose for j ~ 1 that IXj = Jx dFk(x),
k = 1, 2, where F2 is a dJ. with d. CfJ2' By hypothesis and Proposition 4,
CfJ2 is r-analytic, whence by Proposition 5, CfJb) is analytic in 1..I(z)1 < r,
k = 1,2, and
IX~iz)j
L _J_.,- = CfJ2(Z),
co
CfJI(Z) = Izi < r.
j=O ).
iff x = x m , where
m
xm
= log X m
( 1) Xm
1 - log X m :-s; log X m.
As m -> 00, xm-> 00, and therefore m log xm/x m -> 1, whence (m/2)log m :-s;
Xm :-s; 2m log m for m ~ mo. Define
Now for m ~ mo
Thus qJ is not r-analytic for any r > 0 by Proposition 6, whereas the Carleman
criterion ensures that the distribution in question is uniquely determined by
its moments. 0
Proposition 7. If qJ, qJh qJ2 are d.s with qJ(t) = qJI (t)qJit)for all real t and qJ is
ro-analytic, then qJi is ro-analytic, i = 1,2, and qJ(z) = qJI(Z)' qJiz) for
Izi < roo
PROOF. If F, F h F 2 are the corresponding dJ.s, by Proposition 4
Since neither integral on the right vanishes, J~ <Xl e"" dF;(x) < 00, i = 1, 2,
for Irl < r o, so that for i = 1,2
whence J~ <Xl erlxl dFb) < 00 for 0 < r < ro and i = 1,2. Again invoking
Proposition 4, cP; is ro-analytic, i = 1,2. Finally, since cp(z) and qJI(Z)' cpiz)
are both analytic in IZI < r o and coincide on the real axis, they must coincide
in Izi < roo D
Corollary 6. If cP, CPI, CP2 are c.f.s with cp(t) = CPI(t)· CP2(t)for all real t and cP is
an entire c.f., so is cP;, i = 1,2, and qJ(z) = qJI(Z)' CP2(z)for all complex Z.
-1
n
1
211
0
u(r, O)cos nO dO = IXn rn, -I
n
1211
0
u(r, O)sin nO dO = - Pnrn
lanr"1 = I~ f 1lu
(r, (J)e-
i 1J
• dO 1:$ ~ f 1l
!u(r, 0)1 dO, n > O.
Now 9l{cp(0)} = 1X0 = (l/2n) g1l u(r, O)dO, implying for r> 0 that la.rnl +
21X0 :$ ljn S~1I [u(r, 0) + lu(r, 0)1] dO :$ 4A +(r), which is Lemma 3. 0
distribution is. Thus, without loss of generality, qJJ{t) = e icjl Li:j;o Pjkeikt
with Cj real, 0 ::; Pjk ::; I, PjO > O. Since qJ is entire, setting z = ir, r real, in
Corollary 6,
(pe- r + qt = e- r(C'+ C2I nL
2
j; 1 k;O
nj
pjke-rk,
4. Find the dJ. whose c.f. is C L~ 2 [cos jtjV logj)] and utilize this d. to show that the
converse of Theorem I is false for odd derivatives (e.g., cp').
5. A family of d.f.s ff 0 = (F(x; ),), ), E A} with A an Abelian semigroup under addition,
is called reproductive or additively closed if F(x: ),1) • F(x: ),2) = F(x:)" + ),2) for
all ),j E A, j = I, 2. If A = {O, 1,2, ...}, the corresponding family of d.s cp(t; ),) =
[cp(t)].l for some d. cp. Show that if cp(t) and Ijcp(t) are both d.s, then cp is degenerate.
Hence, if A c ( - 00, 'XJ), usually A c [0, 'XJ).
6. Prove that the convolution of 2 singular dJ.s may be absolutely continuous. Hint:
sin t O O t
-
t
= n
cos 2 2j -
J
1 . n00 cos 2t
1
2j
7. Show that
e il - D~A [(it)jjj!]
cp"(t) = (it)"jn!
is a d., n ~ I.
10. If Fj,j = 1,2,3, are dJ.s with F I * F 3 = F 2 * F 3 , it does not follow that F I = F 2.
Hint: If qJ is the d. of a LV. X with PIX = O} =.!, PIX = ± kn} = 2j(k 2n 2)},
k = 1,3,5, ... , then qJ is a periodic c.f. coinciding with qJl(t) = (I - Itl)/IIIISI1 in
[-I, I].
11. There exist uncountably many absolutely continuous dJ.s with support (0, ::t:J) and
identical moments Vb k = 1,2, .... Hint: For 0 < cx < 1, r > 0, and a nonnegative
integer k, set c = (k + l)/cx, b = r + is in the well-known formula b-cr(c) =
S~ i-Ie-by dy. Convert this to a similar example with support (- oc, ::t:J).
12. Prove the Raikov-Ottaviani theorem that the family of Poisson dJ.s is factor closed.
Hint: It may be supposed, without loss of generality, that the support Sj (of any
factor F j) C {O, 1, 2, ... }. Rather than d.s, utilize the corresponding probability
generating functions t/J( w) = e),(W- 11, t/J l v,) = E ",x j = Ii';:o P [X j = ilw i , j = I, 2.
13. If c.f.s qJ., qJ2 satisfy nJ: I [qJj(t)]'j = (pe i • + q)",
where cx j > 0, j = 1,2, then
qJj{t)= eifCj(peil + q)"j with n l ', n2 positive integers such that (Xln l + cx 2nz = n; also
CI(XI + C2CX2 = O.
14. If I - qJ(t) = 0(1 t I') as t -. 0 for some (X in (0,2], then PI I X I > c) = 0«'-') as
C -. Xj. Hint: Integrate.f IIxl > c] (I - cos tx)dF(x) ~ kt' over (0, C- I) or use Lemma
lj.3.1.
15. Prove that
2
-
foo I - .3f(<p(t)}
2 dt =
foo IxldF(x).
n -00 t -00
16. Show that a c.f. is positive definite i.e., D.k: I qJ(t j - tk)pi5k ~ 0 for any n ~ I.
real t I' . . . , t n and complex Plo ... , Pn' Conversely, a continuous, positive definite
function qJ on ( - rx:, ::t:J) with qJ(O) = I is a d. This is Bochner's theorem.
It is easy to verify that a joint dJ. F(x I' . . . , Xk) is continuous at a point
(XI, . .. , Xk) if its marginal d.C.
= E n[!I[xj=ajorbj) + I[a J
<Xj <bja
j= I
and the right side reduces to P{aj < X j < bj , 1 ~j ~ k}whena = (al, ... ,ak)
and b = (b l , ... , bk) are in C(F).
The reformulation and proof of Theorem 8.3.2 for R k are immediate. The
statement of the Levy continuity theorem (Theorem 8.3.3) carries over
verbatim to R k and the proof generalizes readily.
It is worthy of note that independence may be characterized in terms of
d.s as well as dJ.s.
n ({)Xj(t j).
k
via Fubini's theorem, if(2) holds, the d.s of F x, ..... Xk and the product measure
FIx ... x Fk are identical. But then by the uniqueness theorem for dJ.s
and d.s on R k so are the dJ.s, that is,
k
Fx, ..... Xk = nF x ,
j= 1
and so, setting u = 1, the family of univariate d.s on the right determines the
multivariate d. on the left.
EXERCISES 8.5
1. Let F(xl> X2) be the dJ. corresponding to a uniform density over the interval (0, 1) of
the X2 axis of R 2. Verify that (i) F is discontinuous at all points (0, X2) with X2 > 0,
°
(ii) the marginal dJ. F I (x I) is discontinuous I/r x I = 0, (iii) F 2(X2) is continuous. Note
that F { . } assigns probability to all points of R 2 • Construct an F for which C(F) '#
discontinuity set of F.
2. If q> is a c.f. on R I withdJ. F, what is the dJ. corresponding to '/J{tl'"'' tk) = q>{L~ t j )?
3. Prove the dJ. F(x) is continuous at x = (XI"'" X2) if X E C(F). Construct a dis-
continuous density I on R2 with continuous marginal densities II and 12'
References 311
4. If Fi , I :$ i :$ k, are dJ.s on RI, show that for any IX in [-I, I], F(Xh""
nw -
Xk) =
[I + IX Fi(xj)] n~= 1 Fj(Xj) is a dJ. with the given marginal dJ.s.
5. Prove that X = (X 1, ... , X k) has a multivariate normal distribution with mean
vector () = «()l, ... , ()k) and covariance matrix ~ = {aij} iff every linear combination
cX' = L~ CjX j has a normal distribution on R 1 with mean C/I' and variance c~c'.
12. Let the random vectors X. ~ X 0, where X 0 = (X 01' ... , X Ok) is a possibly fictitious
random vector with dJ. F. If {Y", n ~ O} are k-dimensional random vectors whose
ith component is gi(X•. 1" " , X•. k ), I :$ i :$ k, n ~ 0, where {gj, I :$ i :$ k} are
continuous functions on R\ then Y. ~ Yo.
References
H. E. Bray, "Elementary properties of the Stieltjes integral," Ann. Math. 20 (1919),
177-186.
F. P. Cantelli, "Una teoria astratta del calcola delle probabilitci," 1st. Ital. Attuari 3
(1932).
H. Chernoff. "Large sample theory: Parametric case, Ann. Math. Stat. 27 (1956),1-22.
Y. S. Chow and H. Robbins, "On optimal stopping rules:' Z. Wahr. 2 (1963), 33-49.
K. L. Chung, A Course in Probability Theory, Harcourt Brace, New York, 1968; 2nd ed.,
Academic Press, New York, 1974.
H. Cramer, "Ober eine Eigenschaft der normalen Verteilungsfunktion," Math. Z. 41
(1936),405-414.
H. Cramer, Mathematical Methods ofStatistics, Princeton Univ. Press, Princeton, 1946.
H. Cramer, Random Variables and Probability Distributions, Cambridge Tracts Math.
No. 36, Cambridge Univ. Press, London, 1937; 3rd ed., 1970.
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
W. Feller, An Introduction to Probability Theory and Its Applications, Vol. 2, Wiley,
New York, 1966.
M. Frechet and J. Shohat, "A proof of the generalized second limit theorem in the
theory of probability," Trans. Amer. Math. Soc. 33 (1931).
J. Glivenko, Stieltjes Integral, 1936 [in Russian].
312 8 Distribution Functions and Characteristic Functions
n
Definition. Random variables {X n , ~ I} with E X n = 0, E X; = < 00, a;
and dJ.s F n are said to obey the Lindeberg condition if s; = L~ af > 0 for
i
some nand
f
j= 1 lIxl> tS;)
2
x dFi x ) = o(s;), all E > O. (2)
313
314 9 Centra) Limit Theorems
Condition (2) requires that Sft -+ 00 and is equivalent to the classical form
of the Lindeberg condition, viz.,
±i
j= 1 IIxl >u.)
x 2 dFi x ) = o(s;) for all 8> O. (2')
Monotonicity of Sft yields (2) ~ (2'), while the reverse implication follows by
noting that for all 8 > 0 and arbitrarily small (j > 0
S;;2 Ln xi LL
2 dFi x ) = S;;2 + L ]
j= 1 IIxl> Uj) ":sjS6sn j:sJ>6s.
s; (j2 + S;;2 f
j =1
i IIxl > .6sn )
X2 dF j -+ (j2.
max
j
alS;; 2 s; max s;; 2 [8 2S; +
j
iIIxl>.s.)
X2dF j] = 8
2
+ 0(1).
By the Cramer-Levy theorem (Theorem 8.4.3) both of the c.f.s CfJl(t) and
fLoo
= 2 CfJJ{t) must be normal. Isolating CfJ2(t) analogously, it follows that X 2
and eventually all Xft are normally distributed.
(4)
imply (2).
9.1 Independent Components 315
PROOF. Let t, t: be fixed but arbitrary real numbers, the latter being positive.
Set
t2X 2
Y't)
~
=e
o
l
/
Xj - 1- itX 0
J
+ _-.!...
2'
and note that I ~{t)1 ::s; min[t 2 XJ, (l/6)ltXjI3] by Lemma 8.4.1 with () = I
and n = 1,2. Consequently, recalling that E X j = 0,
(5)
Thus, for 1 ::s; j ::s; n, setting So = 0 = So and utilizing independence and (5),
2
t }1
IE exp {it (
So) S~t2}
t
{(SO_l) S~-l
+ ~S; - E exp it ~n + 2s; J
2E exp {itX} 2
{U t }1
2
exp - ~S;
2
::s; e/ /
I sn j -
(6)
(7)
and, since t: is arbitrary, the conclusion follows from the hypothesis and
Theorem 8.3.3.
Conversely, (4) entails (3) via
Uo
max --!.::s; max --!.
Uo + max --!.
Uo --+ 0
t :s,j:s,n Sn 1 :s,j:s,m Sn m< j:s,n Sj
316 9 Central Limit Theorems
= O(1).f <Pj
)= 1
I (.!-) - 11 = 0(1)
Sn
(8)
max
1 $,) $,n
I<Pj (!.-) - 1 I= 0(1),
Sn
Since the integrand on the right :s; 2 :s; 2x 2 /fh;, while that on the left :s;
t 2 x 212s;, it follows that as n ~ 00
Corollary 2. If {X.} are i.i.d. with EX. = /1, E(X. - /1)2 = a 2 E (0, (0), then
I1m· p{S.afi
• -00
- n/1 < x } = -1- IX
j2;r. - 00
e -"2/2 du. (11)
(12)
n = 0, EX; = a;
Definition. Random variables {X., ~ I} with EX.
said to obey a Lindeberg condition of order r > if ° are
f r
j= I Jllxl>tSj]
Ixl'dF/x) = o(s~), all [; > 0. (13)
and also to (I 3") (defined as (13) with Sj replaced by s.). Clearly (13') ~ (13) ~
(13") and so to establish equivalence for r > 2 it suffices to verify that
(13") ~ (13'). The latter follows from the fact that for r > 2 and all [; > °
• •
L EIXX ~ L E{([;s.y-
j= I j= I
2
XJ I IIX jl,;;tS"] + IXXIllxjl>tS")}
~ [;'-2S~ + o(s~).
According to (10), a Lindeberg condition of order r > 2 implies that of order
q, for all q in [2, r]; in particular, a Lindeberg condition of order r > 2
implies the central limit theorem (9).
Then, Qj,n + X j = Qj+ I,n + ll+ I> 1 :$ j < n, and P')(t) = r!, whence
(14)
= O(1)E(,
Xj i
l ~ Illl)E[ I~: r- i + I YOI'-i]
Xj1i
= O(I)EC s7 Illl)
for i = 3, ... ,r, recalling that {I Sj la, 1 :$ j :$ n} is a submartingale, IX ~ 1.
The latter together with the fact that {X n } obeys Lindeberg conditions of
orders 3,4, ... , r ensures that the last expression in (14) is 0(1), noting that
for i ~ 3, Ellll i = Cia~ = CJE XJ)i /2 :$ C i EIXjl i for some constant Ci in
(0, (0). D
If
oo (sin X)2 Hc (2X)dX
-00 x T
I~ 2Mb[~2 - 3 foo sin: x dxJ
n/2 x
~ 2Mb[~ - ~J
2 Tb
(l5)
where 15 = (1j2M)supx IF(x) - G(x) I and Hc(x) = F(x + c) - G(x + c).
PROOF. Since G is necessarily bounded, 15 is finite and the integral on the left
of (15) exists. In the nontrivial case 15 > 0, there exists a real sequence {x n }
with F(x n ) - G(x n) -+ ±2Mb. Since F(x) - G(x) -+ 0 as x -+ ±oo, {xn }
has a finite limit point, say b, and the continuity of G ensures F(b) - G(b) :::;
-2Mb or F(b +) - G(b) ~ 2Mb. Suppose the former for specificity and set
c = b - b. Then if Ix I < 15, by the mean value theorem
= -2Mb fb 1 - c~s Tx dx
o x
1 llxl>b)
1 - c~s Tx Hc(x)dx :::; 2Mb
X
1 lIxj>b)
1 - c~s Tx dx
X
oo sin2 x
= 4MbT
i bT/2
-2-
X
dx. (l7)
Without loss of generality, suppose T large enough so that the middle of (15)
is positive and hence the sum ofthe right sides of (16) and (17) negative. Then
If-00 I (1
=-
oo
lIxlsb]
+ 1) >
llxj>b) -
2M bT - - 3
2
[n ibT/2
oo
-
2
sin -
2
x
dx x ] '
which is tantamount to the first inequality in (15), and the second follows
directly therefrom. 0
320 9 Central Limit Theorems
whence
Joo 2 (2X)d
-_ 2T sin x He --
--2- X
-00 x T
and so
whence by Lemma 1
yielding (18). D
9.1 Independent Components 321
whence, summing on j,
< 8- (rn
--l tl)2+0 exp - {_t 2
+ -2} < 3 (rn1tl)2+0
-- e
-/ /2 2
- 5 Sn 2 5 Sn
o
Lemma 4. Under the conditions of Lemma 3
I<P: (~) - e-
t2/2
1 :$; 16 (r~~tIYH e- /2/3 for
1 (s
It I :$; [ 36 r"n
)2+0J1 /O
e2r
J J
I
2H 2H
+2
2H
I ~ 1 EI X j l
2H
~ 1- + 81 Y:: 1
sup
- 00 <x< 00
IP{Sn < xs n } - ~ IX
Y 2n - 00
e-
y2 2
/ dy I~ cb(rn)2H. Sn
(20)
Remark. The thrust of the theorem is for rn/sn = 0(1), in which case
Esseen (1945) asserts that C b ~ 7.5 when b = 1.
PROOF. If F(x) = P{Sn < xsn} and cI>(x) = standard normal dJ., <D'(x) =
(1/.jbr)e- x2 / 2 ~ 1/.jbr = M and, since both F and <D have mean zero and
°
variance one, Tchebychev's inequality ensures F(x) ~ 1/x 2 , X < 0, and
I - F(x) ~ 1/x 2 , X > (similarly for <D). Thus, F - <D E!I' 10 whence by
Lemma 2, taking T d = (1/36) (sn/rn)2+b and recalling Lemma 4,
supl P{Sn < xsn } - cI>(x)1
x
9.1 Independent Components 323
Consequently, (20) is valid for rnls n ::; 1 and holds trivially with C" = I
when r nlsn > 1. 0
-n
Bn
iIIxl<B.)
xdF(x). (24)
PROOF, Note that either (22) or (23) implies that E xi > o. If J~ 00 x 2 dF(x) < 00,
2
thenc Jllxl>c) dF ::; Juxl>C) x 2 dF(x) = o(1)asc -+ 00,sothat(23)obtainsand
in this case (22) has already been proved with
An = jn foo xdF,
(1 - 00
324 9 Central Limit Theorems
which differ from the choices of (25), (24) only in the sense of Corollary 8.2.2.
Thus, it may and will be supposed that E X 2 = 00.
By the dominated convergence theorem
c- 2 f x 2 dF(x) = 0(1)
JUxl<C]
Bn = sup{c: c- 2 f
JUxl <cl
x 2 dF(x) ~ !},
n
(25)
;2 r 2
x dF(x) = 1. (26)
n JIlXI<Bnl
_; f
Bn J Uxl < cBnl
x 2 dF(x) = n
2
f
Bn JUxl < Bnl
x 2 dF(x) [1 _J[CBn<lxl<Bn~x2
JUxl < BnlX
dF(X)]
dF(x)
-+ 1.
(27)
Thus, via (27) and (23)
lim n P{IXII > eBn} = lim nJllxl>cBnldF
2 2 = 0 (28)
n~oo n~oo nBn J[lxl<cB nlx dF
for 0 < e ~ I and hence for all e > O. Likewise (27) holds for all e > O.
Define Xi = Xj[UXI<Bnl' 1 ~ j ~ n, and S~ = D Xi. Now (26) entails
n l / 2 = o(B n) so that if y" = (n l / 2jBn)X 1[[IXIi <cBnl' then Yn !. 0 and, moreover,
E Y; -+ I by (27). Thus, {y", n ~ I} is uniformly integrable, whence
(n tl2 j Bn ) JUxl < cBnl X dF = E Yn = 0(1) for all e > 0, and so (27) implies
--
n
B;
[i x 2 dF -
Uxl < cBn]
(i [Ixl < cBnl
x dF )2] -+ 1
'
e > O. (29)
Now for e = 1, (29) asserts that s; == (j2(S~) '" B; and, moreover, E Yn = 0(1)
implies E X'I = 0(B n n- I /2 ). Thus, for 0 < e ~ 1 and all large n
9.1 Independent Components 325
Conversely, if (22) prevails, then (28) and (29) hold for all E > 0 by
Corollary 12.2.3, whence
(EB n)2 Jllxl>Bn£)dF
2
E n P{IXti > BnE}
JIIxl<Bn 2
£)x
S 2
dF(x)
2 J 2 = 0(1)
nB n [lIxl<B n£]x dF - ([lxl<£Bn)xdF) ]
J
for all E > 0, which is tantamount to (23). 0
EXAMPLE 1 (Friedman, Katz, Koopmans). If {X, X n , n ;;:: 1} are i.i.d. LV.S with
EX = 0, E X 2 = u 2 E (0, 00), U; = U2(X~) where X~ = (X /\ n l /2 ) v (_n l /2 ),
Sn = LJ=I Xj' then
L n- I sup IP{Snjn l /2 < x} - <I> (xju ) I < 00.
00
(30) n
n=l x
PROOF. Since Un - U, it may and will be supposed that Un > 0, n ;;:: 1 and that
U = 1. Let S~ be the sum on n i.i.d. r.v.s having the common distribution of X~
and set Jln = E X~. In view of
IP {Sn < xn 1/2} - P {S~ < xn 1/2 }I s n P {IX I > n 1/2 }
it suffices to prove (30) with Sn replaced by Now
)J
S~.
l 2 l 2
Sf - nJl xn / - nJl } (x - n / Jl
P{S~ < xn l / 2 } - <I>(xjun) = [ P { n 1/2 n < 1/2 n _ <I> n
n ~ n ~ ~
L L L
00 00 00
f n-1B~::; f
n=1 n=1
n_llet>(-nl/2fln) - et>(0)! = C4
(Tn n=1
n- 1 f f0
n
I/
21
1'"1/11" e- t2 / 2 dt
n=1
L n- 1 IP {Sn < O} - tl < 00. (31)
°
Theorem 5 (CLT for non-degenerate U-statistics). Let Uk,n(h), n ~ k ~ 2 be a
sequence of U-statistics with E h = and (T2 = EIE{h(X 1'"'' Xk)IX d]2 E
(0,00). If
E{h(Xl, ... ,Xk)IX1, ... ,XJE22j/(2j-l), 2::;j::;k (32)
(a fortiori, if E Ih1 4 / 3 < (0), then for any real x
Since y,. = E{h(X r, ... , X r+k-l)IXr}, r ~ 1, are i.i.d. random variables with
E y,. = 0, E Y,.2 = (T2 E (0, 00), the conclusion (33) follows from Corollary 2 and
Theorem 8.1.1. 0
EXERCISFS 9.1
I. Show that if {X., n ~ I} are independent LV.S with IX.I ~ C., a.c., n ~ I, and
C. = o(s.), where s; = L~ E(X j - E X j)2 -+ 00, then (S. - E S.)/s. ~ No. I'
2. If {X. j , 1 ~j ~ kn -+ oo} are rowwise independent LV.S with S. = D',;,I X. j ,
EX.j = 0, E X;j = a;j' s; = D',;,
1 a;j -+ 00, then S./s. ~ No. I jf
k.
L E X;JlIX.il> £s.l = o(s;), e > O.
j=1
4. Failure of the Lindeberg condition does not preclude asymptotic normality. Let
{Y.} be i.i.d. with E Y" = 0, E Y; = I; let {Z.} be independent with P{Z. = ± n} =
1/2n 2, P{Z. = O} = I - (1/n 2) and {Z.} independent of {Y.}. Prove that if X. =
Y" + Zn, Sn = D Xi' then Sn/';;' ~ N O• 1 and the Lindeberg condition cannot
hold. Explain why this does not contravene Theorem I.
5. Let {Y", n ;::: I} be i.i.d. LV.S with finite variance a 2 (say a 2 = I) and let {a;, n ;::: I}
be nonzero constants with 5; = I~ af -> 00. Show that the weighted i.i.d. LV.S
{an Y", n ;::: I} obey the central limit theorem, i.e., (lis.) D
a j >j ~ N o. 1 if an = o(sn)
and E Y = O.
6. Let {X n} be independent with
Again, Sn/';;' has a limiting normal distribution despite the Lindeberg condition
being vitiated.
7. Prove that the Liapounov condition of Corollary 9.1.1 is more stringent the larger
the value of b.
8. (i) For what positive values of lX, ifany, does a CL.T. hold for i.i.d. symmetric random
variables with F(x) = I - 1/(2x"), x ;::: I, F(x) = to 0 ~ x ~ I.
(ii) Does a c.L.T. hold for independent {X.} with P{Xn = ± I} = t
P{Xn = ±n}
= 1/(2n 3 ), P{X. = O} = (1/2) - (lln 3 )? Hint: Apply Theorems 4 and I.
9. CL.T. for V-statistics: Let {X.} be i.i.d. and cp(x l , ... , x m ) a real, symmetric function
of its m arguments. If E cp2(X I' ... , X m) < 00 and E cp(X I' , X m) = 8, then
n l /2(V n - 8) ~ N o.n , where V n = (~)-I I I $;, < ... <im$n cp(X i " , XiJ and a =
2
m E[E {cp(X I"'" Xm)IX d - 8 ]. Hint: A CL.T. applies to
2 2 2
since the components of the sum are i.i.d. with mean 0 and variance a 21m 2 , while
V n - v".!. 0 via E(V n - v,,)2 = 0(1).
k-dimensional normal with mean vector zero and covariance matrix {aij}, where
aij = min(i, j). Conclude that
328 9 Central Limit Theorems
13. Let {X., n ;::: I} be i.i.d. with symmetric density f(x) = Ixl- 3 IlIxl> I)' Show that
(n log n)-1/2 Ii Xi 2.. NO,I'
and if Sn = L~ Xi' T,. = L~ ~i' then via Corollaries 7.2.2, 7.3.4, 7.3.5, and
9.2 Interchangeable Components 329
dominated convergence
so that
E e-[1-a~Wlt2/2 = 1, all t E (-00, (0),
(5)
Thus, recalling Exercises 7.3.6 and 6.3.10, {X. j , 1 :::; j :::; n} is not embeddable
in an infinite sequence of interchangeable r.v.s. Nonetheless, asymptotic
normality of the distribution of the number of empty cells, i.e., Ii= t X. j ,
can be prove~ by more ad hoc methods.
By way of preliminaries, set U = U. = Ii X. j and note that
EU = I• E X.
j=t
j = n (I)N
1- -
n
(6)
-(I - -(1 -
and from (5)
= n[(1 -
- ~r + (n - 1)(1 - ~r n(1 - ~rN]:::; n(1 - ~r· (7)
Let Sj = SJ"l denote the waiting time until the occupation of the jth new
cell by a ball. Set So = 0 and y".j = Sj - Sj-t, j ~ l. Clearly, {y".j, 1 :::; j :::; n}
are independent with Y,..t = SI = 1 and
Un - E Un d N
(J ---> 0 , I' (II)
(12)
aeO }
= ne- 20 eO - a - 1 -
[ 2n (1 + 0(1» + O(an- 1 ) + O(an- 2)
ae-o
= ne- 2°b 2 - -2-(1 + 0(1» + O(ae- 20 )
(14)
a
- (1 + 0(1» = 0(1) if a -+ 00
n
ae-
2a
a
2
ae
a
2 a
ae
a
ne- b = nb = n(e _ 1 _ a) =
2a(1 + 0(1» 0
na 2 = N If a
(1) . -+ °
if a -+ a. E (0, (0),
:;
(ii) From the definition of a and (14), if (12 -+ 00, then N -+ 00, and from (7)
implying a = o(n) when (12 -+ 00. On the other hand, if ne - 2ab2 -+ 00, then
N
00 +- ne- 2a b 2 = - e- 2a(e a - 1 - a) = O(N),
a
a ae a e 2a
- ::; -b2 ::; -b2 = 0(1),
n n n
and so again N -+ 00 and a = o(n). Hence, if one of (12 and ne - 2ab2 tends to 00,
then by (i), (10) holds and the other likewise approaches 00. Now assume that
2a 2
(12 = ne- b (1 + 0(1» -+ 00. (15)
From (15)
ea b a
by' n
; : -+ 0, .fit -+ 0,
.fit(e - 1)
b -+ 00.
(16)
In order to evaluate
P{ j;;nbe _ < x
V - ne- a
a
}
= P{U n < ne- a + x.fitbe- a} (17)
recalling (16).
From (9) and (17)
Now by (8)
n
E Y. . = - - - , - -
n.) n - j + l'
for j ~ 1, whence
k n 1
L n - J. + 1 = L+ -;-,J
n
E Sk = n
j= 1 n-k 1
2
a (Sk) = i [(
j= 1 n- J
~ + 1)2 - n - J~ + I] = f
n-k+ 1
(~:
J
- J~).
Since for m = 1,2
I
i dt 1 1 1
<
0_. -
tm - -.m-('_l)m
< --'m
;-1 J J J
it follows that
n nil 1 k
o <- log - - - L - < -- - - = -,---:-.,-
n - k n-k+ I j - n - k n n(n - k)'
1 1)
0:::;; ( n _ k - ~ - n-hlnp i l
: :; (n - k)2 -
1 2nk
n2 :::;; n2(n _ k)2'
whence, recalling (19) and (20),
N - E Sk = N - n log n : k + °C ~ k)
2
a (Sk) = n nk n (nk)
_ k - n log n _ k + 0 (n _ k)2
334 9 Central Limit Theorems
= ne- + 0(1)
O
(25)
To complete the proof, by Exercise 9.1.2 it remains to verify that {Yn,j - E y",j'
2 ~ j ~ k n } obey the Lindeberg condition.
Setting lj' = y",j+l - 1 and % = j/n, P{Y/ = h} = qjh(l - q), h = 0, I, ... ,
1 ~ j ~ k - 1. Recalling Exercise 4.1.1 and (23),
(1§k = I
j=l
[(-!L)2
1 - qj
+ -!LJ = nb 2 (1 + 0(1»
1- %
I
j=l
E lj'3 =6 I
j=l
[(-!L)3
1- %
+ (-!L)2J +
1 - qj
kf -!L.
j=l1 - qj
For 1 ~ j ~ k, noting (20) and (16),
q. k
__
J _ ~ __ = eO _ 1 + 0(be o/n 1/2 ) = 0(bn 1/ 2 ) = o«1s)
1 - qj n- k
whence
Clearly,
and so
k-l k-l
L
j=l
Ellj' - E lj'1 3 ~ 8 L
j=l
E lj'3 = o«1§J
9.2 Interchangeable Components 335
Thus, the Liapounov and hence also the Lindeberg condition of Exercise 9.1.2
holds. 0
°
(e' - 1 - 1X)1/2 if a -'IXE(O, (0), by a(nI2)1/2 if a -. 0, and by (ne- a )1/2 if
a -. 00. (ii) If aln > 6> 0, then (J2 ~ ne- a - . via (7); if N = na ~ C < 00,
then n l/2 a = 0(1) and (J2 = (na 2/2)(1 + 0(1» + O(a) -. via (14). (iii) If °
(J2 -. (J~ E (0, (0) and a -. 00, then via (10) a = log n - log (J~ + 0(1); if (J2 -.
(J~ E (0, (0) and a -. 0, then a 2 = (2(J5In)(1 + 0(1». Here, the possible
°
limiting distributions can be completely determined (see Theorem 4, Theorem
3.1.4, and Exercise 6). (iv) If (J2 -.0, either a -. or a -. 00 and the limit
distributions are degenerate (Exercise 6).
lim P{U
n
- (n _ N
n
) =j} = «(j2/2~~-62/2, j = 0, 1, .... (26)
n-oo J.
PROOF. According to (9) and (8) with N = Nn,j
P{U n ~ n - N + k} = P{SN-k ~ N} = P{SN-k - (N - k) ~ k},
whereS N _ k - (N - k) = 'L7=-Nlj -
I)and{lj - l,j;;::: 2}areindependent,
geometric r.v.s with success probabilities 1 - U - 1)ln. Hence, the char-
acteristic function of SN-k - (N - k), say q>(t), is given by
q>(t) n
= N-k( 1 -
)=1
'-1)[ '-1 J-I
J__
n
1 _ J_ _ e it
n
=N
n- exp k { • _
_J_ _1 +J_
. __1 eit + 0 ( J2'2)}
j= Inn n
= exp{~~ (e it
- 1) + O(~O}
-.exp{(j; (e it -l)},
and Corollary 8.3.5 guarantees that the dJ. of Un - (n - N n) tends to a
Poisson d.f. with parameter (j2/2, yielding (26). 0
EXERCISES 9.2
I. For an arbitrary dJ. G, concoct a sequence of interchangeable r.v.s for which
P{S. < x~} -+ J<Il(x/y)dG(y).
2. If {X.,n ~ I} are interchangeable r.v.s with E XI = 0, E xi
°
Cov(X., X 2) = = Cov(Xi, XD, then Corollary 9.1.4 holds.
= 1, EIXd 3 < 00,
when C§. is the a-algebra relative to which {X. i , i ~ I} are conditionally i.i.d. If
';;'m.(w)!. 0, a.(w)!. I, rY..(w)/';;'a;'(w)!. 0, and S. = D=I X.;, then S.;.;;, ~
N o. 1 ,
4. If for each n ~ I {X. i , i ~ I} are interchangeable LV.S with E X. I = 0, E X;I = I,
EIX. 113 < 00 and E X. I X. 2 = o(n- I ), E X;IX;2 --> I, EIX. 113 = o(n I/2 ), then
S./';;' ~ N o. 1,
5. If two of the last three conditions of Exercise 4 obtain but the third is violated,
construct a sequence {X .i, i ~ I} of interchangeable processes for which S./';;' f2-.
N o. 1,
6. Apropos of Theorems 3 and 4, prove that (i) if a. - log n --> 00, P{L'i X. i = O} --> I,
(ii) if N a. --> 0, P {I X. i = n - N} --> I. (iii) if a. - log n --> fJ finite and P A designates
a Poisson r.V. with mean A., I7=1 X. i ~ P. xp{-6} (see Theorem 3.1.2).
7. Let the i.i.d. LV.S {Z., n ~ I} be independent of Y, where Yis uniformly distributed on
(0, I) and P{Z. = ± I} = t. Then X. = y- 1/ 2 Z., n ~ I, are !f I interchangeable
LV.S with (lIn) L'i Xi ~ 0 but E X IX 2 does not exist.
•
I EIE{XJI~j-d - uJI = o(s;), (1)
j= I
then (l/sn)Ij= I X j ~ N o. 1 ,
PROOF. The proof of Theorem 9.1.1 may easily be adapted to handle the
current situation. In fact, let ~{t), ait) be exactly as defined there and set
2
bit) = ~ [E{XJI~j- d - uJJ.
Then 9.1(5) holds provided (i) all expectations E therein are changed to
E{ 'l~j_I}' (ii) the term bP/s.) is added to the expression within the absolute
value signs on the right, and (iii) the term Ibi t/ s.) I is added within the brackets
9.3 The Martingale Case 337
on the extreme right. Analogously, (6) obtains if the alteration (iii) is effected
and E is replaced by E{·lffj_d in the second expectation of the second
expression (where independence was used) and in all succeeding expressions
of (6). Finally, (7) holds if E :D=
I Ibit/sn) I is appended to the extreme
right side and so, recalling (l), the theorem follows. 0
A version of the Berry-Esseen theorem also carries over to martingales.
J/
E S~ = s~ = :L'i a;, n 2 I, there exists an absolute constant Co such that
PROOF. Set
where Yo is normal with mean 0, variance I, and IlfU)11 = inf[M: J.lU f(j)(x) I 2
M} = OJ, I ~ j ~ 3, with J.l denoting Lebesgue measure on the real line.
338 9 Central Limit Theorems
PROOF. Let {Y,., n ~ O} be independent normal LV.S with means zero and
variances a; (a5 = 1) with {Yn , n ~ O} independent of {X n , n ~ I}.
Set Qj,n = D=:
Xi + LJ+ I 1';, 1 ~ j ~ n, and note that Qj,n + Xj =
Qj+ I,n + lJ+ I' By Lemma 1
If(a + h) - f(a) - hf'(a) - (~)P2)(a) I ~ Iht IIf(3)1I,
whence for some LV.S ()j with I()jl ~ I,j = 1,2,
(5)
£
O<t<-
r
- -4
h(t) = 3£63
1
G- t -"2' (6)
1
2'
f(t) = h (t - x - n +~ (7)
9.3 The Martingale Case 339
p{:: < x} :$; Ef G:) :s; E f(Yo) + I :s; $(x + t:) + I, (8)
where
By the mean value theorem $(x + t:) - $(x) :s; t:/jb., whence from (8)
{Sn}
P - < x :s; $(x)
Sn
+ t:
M:
V 2n
+ I. (9)
Define
EXERCISES 9.3
1. Utilize Theorem 9.3.1 to prove sufficiency in Theorem 9.2.1.
2. If {X.} are independent with E X j = 0, E xl = (fl, EIXY H < 00, 0 < b ::; I,
modify the right side of (4) of Lemma 2 to (1I!(2)1I + IIP))II) D=
I EIX;i2+6. Hint:
and according to the hypothesis the first factor on the right side of (1) con-
verges in probability to one. Moreover, Skjk~/2 as a subsequence of Snln l ' 2
converges in distribution to No. I' Thus, to prove the theorem it suffices to
establish that
(2)
Since the event corresponding to the first term on the right implies A: u A;,
where
A; = { max
kn(l-6)~j~kn
ISj - Sk..l > ek~/2},
and since by Kolmogorov's inequality
P(A;)::; (e 2kn)-1 E(Sk n ±[6kn J - SkY::; (e 2kn)-I(jk n = e- 2(j,
it follows from (3) that
P{IS'n - Sk..l > ek~/2} ::; 2&-2 + 0(1). (4)
For arbitrary e > 0, the first term on the right in (4) tends to zero with (j and
the theorem follows. 0
9.4 Miscellaneous Central Limit Theorems 341
PROOF. For simplicity, take J-L = l. Since via the strong law of large numbers
1;,le l /(l-a) ~ 1 as e -+ 00 (Exercise 5.4.6), by Theorems 1 and 8.Ll
p{ max ~- Jln 1 - a
> - xan O / 2l - a }
O$j$nJ
= P {max
o$j$.J
~! > c} = P {1; ~ n}
~ I P{1;(O) < n ~ N c }
n=1
ex> n-l
= L I P{1;(O) = k, Sj ~ dor somej ~ n}
n= I k= 1
= Jl nJ+I {1;(O) =
P k}pt~~~/j < O}
JI P {1;(O) = k}JIP{~~~Sj < O}
f p{inf Sj < O} ~ E No <
n= 1 J~n
00
°
(I/crn I/2 )max 1 $;j$;n Sj converges to the positive normal distribution, i.e., for
x> 0, y >
2
lim P {1;,i > c : } = lim P {m~x Sj:5: xcrn
l/2 } = 2l1> ( ~/2) - 1. (10)
c-oo cr n-oo 1 ~J~ny y
Note. For y = I, the right side of (10) is a dJ. in x, namely, the positive
normal distribution, while for x = 1 it is one minus a dJ. in y; the latter,
2[1 - lI>(y- 1/ 2 )], Y > 0, is the so-called positive stable distribution of
characteristic exponent! (Chapter 12).
PROOF. Without loss of generality, take cr = I = y and let x > 0. In view of
the Lindeberg condition, for every () > °
n I n
j~IP{lXjl > ()~}::;; n{)2 j~IEXJI[IXjl>cSJlil = 0(1). (II)
then Sn = Snk = L~:J 1';. Moreover, Yo, ... , ¥,.-1 are independent LV.S and,
furthermore, for fixed j = 0, 1, ... , k - 1 as n - 00
P{TxJli :5: n} - P{Sn > x~} = P{Sn :5: x~, max Sm > x~}
t :::;;m<n
n-l n-l
(13)
: ; L P{Txyin =
i= 1
i, Ai} + L P{Sn::;; x~, Txyin =
i= 1
i, Ai}
.-1
D= L1 P{'Txvr.; =
j=
i}P{AJ
.- 1 1 1 2n 2
~ L -2(nm(i)-nm(j)-I)P{Txm=i}~-2-k
i= 1 ne ne
=-k2'
e
(14)
L [P{S. ~ (x - = i, An
.- I
B~ e)fi, Txm
i= 1
.- 1
~ L P{Txm = i, S'm(j) -
i= 1
S. > O} + P{(x - e)fi < S. ~ xfi}
.-1
= L P{TxJii = i}P{S'm(j) - S. > O} + O(e). (15)
j= 1
To obtain the reverse inequality for the lower limit, observe that A. = 0,
whence via (14)
P{Txv" ~ n} - P{S. > (x + 2e)fi}
= P{S. ~ (x + 2e)fi, max Sj > xfi}
•
2: L P{TxJii = i, S. < S'm,'" IXd ~ efi, An
i= 1
2: f P{TXJii
i= 1
= i, S. < S'm(,,' IXd ~ efi} -
i= 1
f P{TxJii = i, Ai}
9.5 Central Limit Theorems for Double Arrays 345
n n 2
~ i~IP{TXJiI = i,Sn < Sn m,,,} - i~IP{lX;i > ejn} - kt;2
I n 2
= "2 i~l P{TxJiI = i} + 0(1) - ke 2
in view of (11) and the last equality of (15) and its aftermath. Thus,
1 . 2
-2 lIm P{TxJiI :s; n} ~ 1 - Cl>(x + 2e) - -k2' (19)
ft-Q) e
Letting k - 00 and then e - 0 in (18) and (19), it follows that as n - 00
EXERCISES 9.4
1. Construct a sequence of independent r.v.s {X.} with identical means and variances
for which the Lindeberg condition fails.
2. Let{X., n ~ I} be independent LV.S obeying the central limit theorem with EX. = ~,
EX; == ~2 + (12 < 00. If Nc(lX) = sup{k ~ I: Sk :::::; ck"}, c > 0, prove if ~ > 0,
o : : :; IX < I that Ne<lX) is a bona fide LV.
3. Let {X., n ~ I} be independent LV.S with (l/fi) D (E X j - ~) -+ 0 for ~ E (0,00),
(lIn) D (12(X) -+ (12 E (0,00), and (l/fi(1)(s. - DE Xj) ~ No,l' Show that the
conclusions of Theorems I and 2 remain valid.
4. If {X•• n ~ I} are i.i.d. LV.S with J1. = EX> 0 and 0 < b. --+ 00, then, if
likewise (llb.)(max l S;js;.Sj - n~) ~ SF' This yields an alternative proof of Corollary
9.4.1 in the special case IX = O.
Theorem I (Hall, Heyde). For each n ~ 1, let {Sn,j = '~J=1 Xn,i, :F..,j, 1 ~ j ~
kn < oo} be an 2 2 stochastic sequence with X: = maxI $i$k n IXn,;!, Un~j =
I3=1 X;,i' 1 ~ j ~ k n such thatjor some :F..,I-measurable r.v. u;
U 2 - u2 !. 0
n n(where U 2 = U 2 ) (1)
"n.k,.
X:!. 0 (2)
Un2 ~ 1];' (3)
sup E
n2:1
X: 2
< 00 (4)
k ~
L E{Xn):F..,j-d !.O,
j=1
L E 2{Xn,jl:F..,j_d!.O
j=1
(5)
whence
~ 2 E(X: 2 + max Ej _ X: 2) ~ 10 E
jSk,.
1
2
< 00 X:
implying sUPn;" 1 E(X~*)2 < 00. Hence (1), (2), (3) and (4) hold for the primed
r.v.'s (X~j}'
F or any c > 0, Iet I] ,,2 = I] 2 1\ C, Un"2 = Un2 1\ C, X"nj = X'n,j I [~;"Jl W here IX =
min{1 ~ i ~ kn: U2i> c} or k n according as U~2 > c or not. Define U~',i' S;,j'
X;* in an obvious fashion. Since either U~2 < c whence U~'2 = U~2 or IX ~ kn
implying c < U~'2 ~ C + (X~*)2,
EI T,.1 2 = E nk
j=1
(1 + X;/) = E(l + X;,~) n (1 + X;/)
j<~
9.5 Central Limit Theorems for Double Arrays 347
eiS~ = 1'" W:' = (1'" - 1) J.v" + (1'" - I)(W:' - J.v,,) + W:' (7)
where
Since on {X"*
n I"~n r(X"n,}·)1
< I} 'L.Il=! <" ·IX" .1
~J n.J
3
< X"*U"2
- n n P
-+ 0,
(3)" guarantees
(9)
while (1)" and (8) ensure W:'/J.v,,!. 1. Thus, 0 :5: J.v" :5: 1 entails W:' - J.v,,!. 0
a~d so recalling that {T,,} is u.i. (and a fortiori, tight)
(1'" - I)(W:' - J.v,,)!. O. (10)
Next, since {X;.j' .fF".j, 1 :5: j :5: k"} are martingale differences,
j=1
while (9) and (10) ensure that the right and hence left side of (12).2. e-~"2/2.
Since the left side is u.i., Corollary 8.1.8 in conjunction with (11) guarantees
that
(13)
348 9 Central Limit Theorems
Consequently,
IE(eiS~ _ e-,,2/ 2 )! ~ IE(eis~ _ eiS;;)1 + IE e iS;; _ e,,"2/2 1+ \E(e-,,"2/ 2 _ e-,,2/ 2 )1
~ 2P{U~2 > c} + IE(eiS~ _ e-,,"2/ 2)1 + P{,,2 > c}
which, in conjunction with (13) implies that for any t: >0
lim IE(eiS~ - e-,,2/ 2 )1 ~ 3s
provided c :?: Ct' Replacing S~ by tS~ for t #- 0,
lim E ei'S~ = E e-,2,,2/2.
p{ )=1UA j} t
~ e+ p{ P{Ajl~-d > e},
J-I
e > O. (14)
pLQ AltLn jt
~ e]} ~ pLQ Aj[}J.j ~ e]} = E P{Ajl~_I}I(~j~t) ~ e l
so that
necessarily
2 2
Q(AG) = E ZIAG - E ZI[Y2+Z~ IIAG + E y I[y2+Z>IIAG - E y I[Y2>IIAG ~ O.
c 2
DnA , y + Z ~ 1 iff(2Y + W + 1)(2Y + W - 1) ~ iff -(1 + W)/2 ~ Y ~ o
(1 - W)/2. Hence A c[y 2 + Z ~ 1] C [y 2 ~ 1] so that A c[y 2 + Z ~ 1]·
[y2 > 1] = ,p. Furthermore, since Z < 0 iff 12 y + WI < IYI which, in turn,
implies I YI < IWI, necessarily [Z < O]A C C [y 2 < 1] so that Z ~ 0 on
A c [y 2 ~ 1]. Thus
Q(AcG) = E(y 2 + Z)I[Y2+Z>I2:y2WG - E y2I[Y2+Z~I<Y2WG
+ E ZI[y2+Z>I,Y2>IIAcG ~ 0
whence
Q(G) = Q(AG) + Q(AcG) ~ O. D
Sn,j = L
i=l
y",i'
If
kn
and
{/In, n ~ I} is tight (19)
then
p
max ISn,j - /In,jl-+ O. (20)
1 $j~k"
350 9 Central Limit Theorems
implying
lim P
"-00
{t I
1
Y".i - Y:) > o} : ; sup P{JLn > K} + lim P{ Y,,* > K-
n~ 1 n-oo
2
} ~0
whence defining S~.j' S~, JL~.j' JL~ analogously
Moreover,
P {max IJLn,j -
1 SjSk n
JL~) > fJ} ::; P {t E{Y",JIYn,j>K'211'?;'.j-d > fJ/2}
1
n~l n~l
(F /4) P {m~x IS~,j - JL~) > b} ::; ~ E m~x (S~.j - JL~,Y ::; E(S~ - JL~)2
k" kn II"
::; L E Y:j ::; K- 2E L Y:. j = K- 2 E L E {Y:).?;..i-d
i=l 1 1
9.5 Central Limit Theorems for Double Arrays 351
Finally,
max ISn,j - J.ln)
j
The latter ensures that V;2 is tight which, in conjunction with (22) for {X~.j}
guarantees that
and note that (2) and (5) hold trivially for {X;,j}' Likewise (1) and (3) obtain
since U~2 - U~'2 !. 0 in view of
U,2 - U"2 < U'21 k n '2
n n - n [Ln:, Ej-l{Xn.iIIX~.jl><l}>l)
and the conditional Lindeberg condition.
Moreover,
kn
E max
.
X"~
n,) -
< /;2 +E "X,2.I,
!-- n,) IIX'\iI>£,L~1
j '2
Ei-,{Xn.;lIIX· il>"}::S; 1)
}5':k n )=1 IJ.
kn
f
k
v,,2 = [E{X;,jlff",j-d - E 2{Xn •j lff".j_d].!.,,2 (23')
j=1
for some non-negative n:;'1 ff".I-measurable r.v. ,,2, then the conclusion of
Theorem 2 holds.
EXERCISES 9.5
1. (Helland) If for n;;:: 1, {Xnj , !F,.j' 1 :::;; j :::;; k n} is a stochastic sequence on (O,§', P)
with E Xn* = E maXj IXnjl = 0(1) as n .... 00, then L~~I Ej_1{IXnPllx nJ I>tl} !. 0, e > O.
In particular, ifevents A nj E!F,.j where!F,., c ... C !F,.k n C jO, then p{U~n A nj } =0(1)
as n .... 00 implies D~I P{Anjl!F,..j-l} !. O.
Hint: Let T" = inf{1 :::;; j :::;; k n : IX.jl > e} and T" = k. otherwise. Then
2. (Helland) (i) If for each n;;:: 1, {X.j ,9';,j' 1 :::;; j:::;; k.} is a stochastic sequence with
IXnjl :::;; c, 1 :::;; j :::;; k. and U; = Dn X~ !. 1, then {v,,2, n ;;:: 1} is tight.
Hint: If 1;, = inf{l :::;; j :::;; kn , Lt=1 X;i > 2} and T" = k. otherwise, then
Tn
P{v,,2 > a} ~ P{U; > 2} + a-I E I X;j ~ 0(1) + (c + 2)a- 1•
1
References
F. Anscombe, "Large sample theory of sequential estimation," Proc. Cambro Phi/os.
Soc. 48 (1952),.600-607.
S. Bernstein, "Several comments concerning the limit theorem of Liapounov," Dokl.
Akad. Nauk. SSSR 24(1939),3-7.
A. C. Berry, "The accuracy of the Gaussian approximation to the sum of independent
variates," Trans. Amer. Math. Soc. 49 (1941), 122-136.
J. Blum, D. Hanson, and 1. Rosenblatt, "On the CLT for the sum of a random number
of random variables," Z. Wahr. Verw. Geb. 1 (1962-1963), 389-393.
J. Blum, H. Chernoff, M. Rosenblatt, and H. Teicher, "Central limit theorems for
interchangeable processes," Can. Jour. Math. 10 (1958),222-229.
K. L. Chung, A Course in Probability Theory, Harcourt Brace, New York, 1968; 2nd ed.,
Academic Press, New York, 1974.
W. Doeblin, "Sur deux problemes de M. Kolmogorov concernant les chaines denom-
brables," Bull, Soc. Math. France 66 (1938),210-220.
J. L. Doob, Stochastic Processes, Wiley New York, 1953.
A. Dvoretzky, "Asymptotic normality for sums of dependent random variables,"
Proc. Sixth Berkeley Symp. on Stat. and Prob. 1970,513-535.
P. Erdos and M. Kac, "On certain limit theorems of the theory of probability," Bull.
Amer. Math. Soc. 52 (1946), 292-302.
C. Esseen, "Fourier analysis of distribution functions," Acta Math. 77 (1945), 1-125.
W. Feller, .. Ober den Zentralen Grenzwertsatz der wahrscheinlichkeitsrechnung," Math,
Zeit. 40 (1935),521-559.
N. Friedman, M. Katz, and L. Koopmans, "Convergence rates for the central limit
theorem," Proc. Nat. Acad. Sci. 56 (1966), 1062-1065.
P. Hall and C. C. Heyde, Martingale Limit Theory and its Application, Academic Press,
New York, 1980.
K. Knopp, Theory and Application of J"jinite Series, Stechert-Hafner, New York, 1928.
P. Levy, Theorie de faddition des variables aleatoiries, Gauthier-Villars, Paris, 1937;
2nd ed., 1954.
J. Lindeberg, "Eine neue Herleitung des Exponentialgesetzes in der Wahrschein-
lichkeitsrechnung," Math. Zeit. 15 (1922),211-225.
D. L. McLeish, "Dependent Central Limit Theorems and invariance principles," Ann.
Prob. 2 (1974), 620-628.
J. Mogyorodi, "A CLT for the sum of a random number of independent random
variables," Magyor. Tud. Akad. Mat. Kutato Int. Kozl. 7 (1962), 409--424.
A. Renyi, "Three new proofs and a generalization of a theorem of Irving Weiss,"
Magyor. Tud. Akad. Mat. Kutato Int. Kozl. 7 (1962),203-214.
A. Renyi, "On the CLT for the sum of a random number of independent random
variables, Acta Math. Acad. Sci. Hung. 11 (1960), 97-102.
B. Rosen, "On the asymptotic distribution of sums of independent, identically distri-
buted random variables," ArkivfOr Mat. 4 (1962),323-332.
D. Siegmund, "On the asymptotic normality of one-sided stopping rules," Ann. Math.
Stat. 39 (1968), 1493-1497.
H. Teicher, "On interchangeable random variables," Studi di Probabilita Statistica e
Ricerca Operativa in Onore di Giuseppe Pompilj, pp. 141-148, Gubbio, 1971.
H. Teicher, "A classical limit theorem without invariance or reflection, Ann. Math. Stat.
43 (1973), 702-704.
P. Van Beek, "An application of the Fourier method to the problem of sharpening the
Berry-Esseen inequality," Z. Wahr. 23 (1972), 187-197.
I. Weiss, "Limit distributions in some occupancy problems," Ann. Math. Stat. 29 (1958),
878-884.
V. Zolotarev, "An absolute estimate of the remainder term in the c.L.T.," Theor. Prob.
and its Appl. II (1966), 95-105.
10
Limit Theorems for
Independent Random Variables
n;::: I,
354
10.1 Laws of Large Numbers 355
t p{ max (X j
l~)~n
- m) ~ e} 5. p{ max X1 ~ e},
l~J~n
(I)
n n
5. 2 L
j=l
P{T=j, IXjl ~ e} =2 L
j=l
P{T=j, IX}I ~ e}
5. 2P{ m~x
15is n
IXjl ~ e}.
This yields the first inequality of (2) and the second follows from
kn
L (12(XnjIIIXnil< 11) -+ 0,
j= I
(7)
in which case
kn
An - L E X )II
j= I
n X ni!<II-+ O. (8)
and note that via independence (7) ensures Vn - E Vn .!:. O. Since by (6)
kn
n
whence
ex p {- II
P{IX:j ! ~ t:}} ~ J= I
P{IX:jl < t:} = p{maxlx:jl < t:} -+
J $k n
I
Since (5) ensures Imnjl < c;, 1 ~ j ~ k n for all large n, by Lemma 1
V: = L Y:
k"
j •
j= 1
By (6), (9), and (4), v" - An!. 0 entailing V: !. O. Hence, if V:k = D= 1 Y:j ,
1 ~ k ~ kn , by Levy's inequality for all c; > 0
(12)
kn
L E(X nj -
j; I
mnj)21Ilxnrmnj!<11 = 0(1), (14)
PROOF. Since Znj = X nj - m nj has zero as a median, (12) follows from (13)
and (14) by Theorem I, noting (9). Conversely, under (12), setting Un =
D~I Znj, B n = An - D~I mnj ,
Un - Bn ~ 0, (15)
and let (¥~ I> ... , ¥~kJ and (Y" I' ... , ¥nkJ be i.i.d. random vectors. It follows
from (15) and (13) that
and since I(¥nj - ¥~)/31 < I and (Exercise 3.3.3) m( Y,,) = 0, Theorem I and
Lemma I ensure that
kn
L P {Y"j ;;::: e} = 0(1), (16)
j; I
kn
L E ¥njl[Ynj< II --+ l.
j;1
(17)
Moreover, ifL~~ I E ¥nj = I, then Sn!. I and max l 5,j5,k m(¥n) --+ 0 iff
n
kn
j; I
L E ¥nJ[Ynj~'1 = 0(1), e> O. (18)
PROOF. Necessity of (16) and (17) is an immediate consequence of (6) and (8)
of Theorem 1 with An == l. Sufficiency likewise follows from Theorem I once
10.1 Laws of Large Numbers 359
o::; e L P{Ynj 2 e}
j; 1
kn kn
L P{lXjl
j; 1
2 bn } = 0(1), (19)
(20)
(21)
max X.)
m ( ---1
bn
I ::; - I max Im(X)1
bn 1 SjSno
+ max I (X ")
m---1
bj
1
1
1 sjsn I no<jsn
~ ~Up m(~-l)
b)
I~o,
)"no
360 10 Limit Theorems for Independent Random Variables
conditions (4) and (5) of Theorem 1 hold with An = 0, so that (6), (7), (8) are
tantamount to (19), (20), (21). D
xt
k-l n
+ (_l)HI L
11
= (k + l)Q
k+!,n.L..
+ ~I (_I)i+1 [S~-i ~ X!+I + CHI
+ Ck - i ] !--)
(k _ ')1
,=1 I . )=1
+ l)QHI,n + CHI'
= (k
Since (24) clearly holds for k = 1 and 2, the lemma follows. 0
min I C, whence by the induction hypothesis Lj,;l Yj :5; (2k - 3)A, implying
D= 1 Yj :5; (2k - 2)A. Consequently, noting that (26) entails LjelYj :5; A,
n m
LYj:5; L Yj + LYj :5; (2k - I)A,
1 1 je 1
LI = L b j-
2
aJ,
j= 1
Q) jk-1 j,-1
(27)
Lk = L b;;,2kaJk L aJ.-l··· L aJ" k ~ 2.
j.=k j.-,=k-l h=1
Corollary 5.2.1 states that for independent r.v.s with E X n = 0, EX; = a;,
the convergence ofLI is sufficient for the classical strong law (where bn = n).
The next theorem asserts that convergence of Lk for some k ~ 2 in con-
junction with the necessary condition X n = o(b n), a.c., ensures the generalized
strong law SJb n ~ O.
Then {ur,.,n, /Fn' n ~ k} is a martingale, where /F n = a(X j , 1 :5; j :5; n), and,
moreover, E Wrn is the series in (27) modified in that the first summation
only goes up to n rather than 00. The convergence of Lk thus ensures that
{ur,..n, /Fn, n ~ k} is an 2 2 bounded martingale, hence convergent to some
r.v. by Theorem 7.4.3. Consequently, by Kronecker's lemma
n
bn-kU k,n = b-
n "L, X j V k - l , j - l --+
k a.c. 0
. (29)
j=k
This proves the theorem for k = 1, in which case (28) is superfluous. Next, in
view of (27)
n ~-1 j,-l
Zk,n = L bj~2kXJk L XJ._,,,. L Xl., n ~ k,
jk =k j. - 1 =k - 1 h =1
362 10 Limit Theorems for Independent Random Variables
· I (. )h12
I ,I X~ ~ .L XJ = o(b~),
h 2. 2. (30)
J= I J= I
•
s;; 1(log s.)-a I Xj~O,
j= 1
(32)
hold.
PROOF. Setting s; = L~ aJ, if k~ > 1, the series Lk of (27) converges, being
dominated by
"'.
L.
[b;;2ka;(~1L. a 2)k-l]J
< C~
- .~l s;(log s;)U
a; < 00. o
10.1 Laws of Large Numbers 363
If higher-order moments are assumed finite, the next result asserts that
convergence of a single series suffices for the classical strong law.
and so, in view of Lemma 3.3.1 it suffices to show that the right side of (34) is
0(1). Now,
(35)
It follows via (33) and Kronecker's lemma that EIS.1 2 r = 0(n 2r ) and, moreover,
that the series on the right of (35) is bounded by
where 1 < tX ~ 2 and A n +l,JA n • a ~ Y< 00, all n ~ 1. If, for some P in
[0, l/tX) and positive b, c,
00
then
(38)
1 n
A (I
n og2
A)I
n
p.L (Vi -
1= I
E Vi) ~ O.
A (I
n og2
A)1
n
p.L (l'i -
1= I
E l'i) ~ O. (39)
: :; p{ max
"k-I <nSnk
U. > 2y2EA•• _,(I0g2 A•• _Y-//, i.O.}
(40)
•
t; = E U; : :; L E i= 1
XrIlIxil:scAi(IOg2Ai)-Pj
Consequently, setting
EA 2(log A)1 - 2//
A. =
• •
t;
2.
,
it follows that e.' x. = 2e. Since (41) ensures that 2eA.x. ~ 2e«-lE(log2 A.)l-//«
-+ 00 and h(x) ~ (x/2)log(1 + x) as x -+ 00, for all large n (see 10.2(1»
x;h(2eA.) ~ eA. x; log(1 + 2eA.) = E(log 2 A.) log(1 + 2eA.) > 8e 2 1og 2 A•.
1
:::;; exp{ -t x ;.h(2eA.,>/4e 2
} :::;;exp{ - 21og 2 A•.} :::;; (k logy)2
~ U.
11m 1 /I < 0, a.c. (42)
• -00 A.(I0g2 A.) - -
then
(45)
00
is necessary and sufficient for Sn/n ~ 0 when IXnI < Knjlog2 n, a.c. for
n ~ 1. Unrestricted necessary and sufficient conditions depending upon
solutions of equations involving truncated moment generating functions
have been given by Nagaev (1972).
EXERCISES 10.1
3. Let {X., n::e: I} be independent LV.S with EX. = 0, EX; = 0'; < 00, s; =
L~ af -> 00, and a. = o(s.). Prove the result of Raikov that
. I ~ 2 p
Iff 2 L. X j -> I.
sn j= t
5. Define a sequence of LV.S for which (lin) X;..:.=.. 0 but D L (X .In) diverges a.c.,
revealing limitations to the" Kronecker lemma approach."
8. Show that Theorem 2extends to the case where {X n' n ::e: I} are martingaledifferences
with constant conditional variances 0';.
9. If {X n } are independent LV.S with variances 0';, s; = L~ af = o(b;), and
D' (E X~/b~) < X!, where bn i 00, then (JIb;) I L7= xf ~ o.
10. Let X n = b· Yn , n ::e: I, b > I, where {Yn } are bounded i.i.d. random variables. Prove
that (llb n) L7= n
1 X; ~ 0 provided bnlb -> 00. Compare with Exercise 5.2.8.
Show that s; = D
af - (2),) - I exp{2).nllog n} and that Ilb n L7 = 1 X i ~ 0 when-
ever s; log log s; = o(b;).
368 10 Limit Theorems for Independent Random Variables
Lemma 1. Let 8n = IJ=l Xi where {Xi' 1 ::; j ::; n} are independent r.v.s with
s;
E Xi = 0, E Xl = ul, = IJ=l ul > o.
(i) If P {Xi::; CnSn} = 1, 1 ::; j ::; n, then
t > 0, (2)
and if, in addition, CnX n ::; an, thenfor all An and X n such that an· X n > 0, anAn > 0
t> 0, (4)
tX"J = 1 + t E --+
2
EelXj/Sn = I + E [elXj/Sn - I - _J X g (tX")2 _J . (6)
Sn Sn Sn
If, rather, (ii) obtains, then (6) in conjunction with the elementary in-
equality (l + u)e > e u > 0, yields for t > 0
Ul U
,
p{ m~x Sj ~ }X"S"} ::::;; e-J.txnsn E etSn ::::;; exp{ -AtX"S" + t s;g(c"s"t}}, (8)
I S)S"
2
(9)
(10)
PROOF. Let b" = s"(log2 s;r 1/ 2 and suppose d"/b" -+ a ~ O. Since 0 < b" i co, it
follows that b;l(max l 5;;iS" d;) -+ a and so d" may be taken to be increasing.
For any ex>1, define no=inf{n~1:s"~ex} and nk = inf{n> nk-I:
k
s" ~ exS"k_I}' k ~ 1. Then S"k_l ~ S"k- 1 < exS"k_l so that S"k- 1 ~ ex and
log2 S;k_l '" log2 S;k- I · Hence,
370 10 Limit Theorems for Independent Random Variables
P {Sn > Aa 2sn(log2 S;)1/2, i.o.} ::5: P { max Sn > Aa 2Snk _1 (log2 S;k_Y/2, i.O.}
nk-I ~n<nk
p{ I
max Sj > !h-l(y)Snk-I(lOg2S;k-dl/2}::5:
$;j<nk U
eXP{-~h(h-I(Y»IOg2S;k_l}
U
proving the second part of (i). If rather, a = 0 then (12) holds for arbitrarily
small a. Since h(a) '" a2/2 as a -+ 0, necessarily h-I(a) '" (2a)I/2 whence
(l/a)h- l (a 2 ) -+.j2 as a -+ 0 yielding the first portion of (i). Finally, under (ii),
(12) holds for both Sn and -Sn so that (10) obtains. 0
For any positive integer n, let {Xn, j, 1 ::5: j ::5: n} constitute independent LV.S
with dJ.s Fn,j and finite moment generating functions ((In.J{t) = exp{I/Injt)}
for 0 ::5: t < to' Suppose that Sn,n = 2:J=1 Xn,j has dJ. Fn. For any tin [0, to),
define associated d.C.s F~~~ by
F(t) ~x) I
= ---- IX e'Y dF ~y)
n,J ({Jnjt) -00 n,J
and let {Xnjt), I ::5: j ::5: n} be (fictitious) independent r.v.s with dJ.s {F~!j'
1 ::5: j ::5: n}. Since the d. of Xnjt) is ({Jnjt + iU)/({JnJt), setting I/In(t) =
L:J= I I/Injt), the c.f. of Sit) = L:J= 1 Xnjt) is given by
nn ({In J{t + iu) }
Ee
. S
1U
"
(I)
= . = exp{I/I.(t + iu) - I/In(t) .
j= I ((Jnjt)
Thus, the mean and variance of Sn(t) are I/I~(t) and I/I;(t) respectively and,
moreover, the dJ. of Sit) is
10.2 Law of the Iterated Logarithm 371
If 1/1. and its derivatives can be approximated with sufficient accuracy, (13)
holds forth the possibility of obtaining a lower bound for the probability that
a sum of independent LV.S with zero means exceeds a multiple of its standard
deviation.
PROOF. Let qJj(t) denote the m.gJ. of Xj and set S._. = S./s. = IJ=1 X)s. and
c. = d./s.,
°
Since, in the notation leading to (13), qJ.)t) = qJit/s.), 1 ~ j ::;; n,
and gl (x) = X-I (eX - 1)j, it follows for t > and 1 ::;; j ::;; n that
< a2
'1'._ J'
m' It) = -dtd E e,Xj/ = E --!.
X·
S
s
"(e,Xj!s" - 1) - tg (+tc)---l.
~ I - • S2'
• •
Thus, via (5), (15), (16) and g(O) = 1, gl(O) = 1, for any y in (0,1) and all
large n
2
tn
t/tn(tn) - tnt/t~(tn) ~ t;[g( -tncn) - t;c;g2( -tncn) - gt(tnCn)] ~ -2 (l + y),
= (1 - y)t n - t/t~(tn) = _ t (1 + (1» < -ytn
Vn - J t/t~(tn) y n 0 - 2 .
x r
P{Sn> (1 - y)sntn} ~ exp{t/tn(t n) - tnt/t~(tn)}
since
n
Sn(t n) - t/t~(tn)
LZnj~ No,t
Jt/t~(tn) j= 1
The strong law asserts under certain conditions that with probability one
sums Sn of independent r.v.s with zero means are o(n). In the symmetric
Bernoulli case, Hausdorff proved in 1913 that Sn ac O(n(l/2)+t), € > 0. The
order of magnitude was improved to O(Jn log n) by Hardy and Littlewood
in 1914 and to O(Jn log2 n) by Khintchine in 1923. (Here, as elsewhere,
log2 n denotes log log nand logk + 1 n = log logk n, k ~ 1). One year later
Khintchine obtained the iterated logarithm law for the special case in
question and in 1929 Kolmogorov proved
10.2 Law of the Iterated Logarithm 373
(18)
PROOF. Choose the integers nk , k ~ 1 such that SOk ~ a l < SOk+1 and note that
u;/s; = 0(1) whence SOk ' " a k • According to Corollary 1,
a.c. (19)
(20)
for all large k. Thus, taking X Ok = hk in Lemma 2, noting (20) and that
dokhk/g k = 0(1),
P{Ad ~ C y exp{ -hf(1 - yXI - yZ)/2} ~ C y exp{ -(1 - yZ)Zlog k}
Cy
=k~
To extend the law of the iterated logarithm (LIL) from bounded to un-
bounded LV.S without losing ground, a refined truncation is necessary. This
means that the truncation constants, far from being universal, should (as
first realized by Hartman and Wintner in the i.i.d. case) depend upon the
tails of the distributions of the r.v.s involved.
Let {X., n 2:: I} denote independent random variables with EX. = 0,
EX; = a;, s; = D=t a? -+00. Then {X., n 2:: I} obeys the LIL if (17)
obtains.
Theorem 2. If {X., n 2:: I} are independent LV.S with EX. = 0, EX; = a;,
s; = L~ a? --+ 00, in order that {X., n 2:: l} obey the LIL it is necessary that
00
L P{X. > <5S.(I0g2 S;)t/2} < 00, <5 > fi. (22)
.=1
PROOF. If b; = 2s; log2 s; and S. = L~ Xi' So = 0, then lim._", Snlb. 3':;,1. Now
°
S._llb. ~ since a 2 (S._d = S;_I = o(b;), and clearly S._I is independent of
~X., X.+1' ... ) for all n 2:: 1. Hence, by Lemma 3.3.4(ii) lim X.lb. ::;; 1 + e, a.c.
and (22) follows by the Borel-Cantelli theorem. 0
Corollary 3. Under the hypothesis of Theorem 2, in order that both {X.} and
{-X.} obey the LIL, it is necessary that
00
L P{ IX. I > <5S.(I0g2 S;)1/2} < 00, <5 > fi. (23)
.=1
The next result stipulates two conditions which, conjoined with (23) for a
fixed <5, are sufficient for the LIL. One of these, (25), clearly implies the
Lindeberg criterion and hence the asymptotic normality of L~ X )s•.
00
(25)
L
00
2
n= 1 Sn(\og2
1 2
i
Sn) [£Snllog2s~)-1/2<lxI56sn(log,s~)1/2)
X
2
dFn(x) < 00 for all e > 0,
(26)
then the law of the iterated logarithm (17) holdsfor {X n} and -{X n}. Alter-
natively, if(24) is valid for all e5 > 0, (25) obtains and (26) is replaced by
00 j. - 1 h - 1
L(S].log 2 S]k)-kyj• L Yj._,··· LYj,<oo, (27)
j. = k j. - I =k- 1 j, = 1
and hence permits the choice of integers nk+ 1 > nk such that <fJn(k - 2) < k - 2
for n ~ nk> k ~ 1. Define e~ = k - 2, nk ::s; n < nk+ l' k ~ 1. Then e~ l and °
for nk ::s; n < nk+ I
as n -> 00 provided en = e~. Proceeding in a similar spirit with the tail of the
series of (26), there is a sequence e~ = 0(1) such that
nk + 1
L L ... ::s; L k-
00 00
= 2
< 00, (29)
k= 1 n>n. k= 1
where en = e~.
Consequently, en = max(e~, e~) = 0(1) and both (25) and (26) hold with e
replaced by ej and en respectively.
Define truncation constants {b n , n ~ I} by
376 10 Limit Theorems for Independent Random Variables
and set
n n n
I I I
Now
0'; - at = E X;IlIx.l>b.) + E 2
XnIIIX.lsb.1 ::; 2 E X;IlIx.l>b.I'
recalling that E X n = 0, and so (28) ensures o'~~ - at. Thus Theorem
yields
S~ - E S~ a.c. I.
fIn1 2 1/2 (31)
n- 00 sn(2 log2 Sn)
(32)
Thirdly, (24) implies that S;' = 0(1) with probability one, and, further-
more,
IE S;/I::; ±i
i= I [Ixl >clsj(loK2S~)1/2)
IxldFi x )
+ itl LXI>s.(IOg2S~)_1/2)IXldFJ{X)
Now
The first corollary reduces the number of conditions of the theorem while
the second circumvents the unwieldy series of (27).
(35)
where
Theorem 4. Let {a. Y,,} E Q with a; = 0(s;/log 2 s;) and ai = I. If either (i) for
some (X in (0, 2]
a; 2),,/2 i
f (2s. log2 Iyl" dF(y) < co for all e> 0 (37)
n=I s. [y2;" esMC1~ log2 s~)
or (ii)
~ {2
.~I P Y I >
[)s; log2
a;
s;} < co fi II ~
or aU>
0 (38)
(39)
(40)
I·
2:
s.
La;
j= 1
i
[y2>es'/a2Iog2s'l
J J J
y2 dF(y) = 0(1), e> 0, (41)
and is automatic whenever the integral therein is o( I), that is, whenever
10.2 Law of the Iterated Logarithm 379
a; = o(s;/log 2 s;). The first part of the theorem thus follows from Corollary 4
since (37) is just a transcription of (34).
Likewise, (38) is a transliteration of (24), and so the second portion will
follow from Corollary 5 once (35) is established. To this end, note that
a; = 0(s;/log 2 s;) entails lim a;/s;_ 1 = 0, whence
and
log s; = (I + an)log s; - I' (42)
so that,
s;(Iog s; _ I)P log2 s;
= (s;_ I + a;)[(1 + an)-I log S;]P[lOg2 S;_I + log(I + an)]
= (I + ~)[I
S;_I
+ 0(1)
P log S;_I
][1 + 0(1)
(log s;_I)log s;
]
Yn = n2
a; , n~l. (44)
Sn
Note that YI = 1,0 < Yn < n, n > I, and
S2
~=n
n ( Y
1--.1
)-1 (45)
si j= 2 j
380 10 Limit Theorems for Independent Random Variables
then for every J-II > 0 and real J-I2 necessarily nl'l/(Iog2 s;t 2 i OC) (aI/large n)
and
(49)
implying
J-II + 0(1)
n
for J-I2 ~ 0; the same conclusion is obvious when J-I2 < 0, so that for all J-I2
nI'l (n - 1)1'1 J11(l + 0(1»
(50)
(I0g2 S;)1'2 (I0g2 S;_1)1'2 ~ n I'I(lOg2 s;t 2 '
l
10.2 Law of the Iterated Logarithm 381
Theorem 5. If {un Y,,} E Q, where u; = o(S;/lOg2 s;) and ')In = 0((lOg2 s;)P) for
some p < 1, then the LIL holds for {un Y", n 2: 1} provided E y2 < 00.
PROOF. According to Theorem 4 it suffices to verify (37) for some IX in (0, 2].
Now the hypotheses entail ')In = o(n), thus a fortiori (46) and also ')In ~
K(log2 s;l, whence Lemma 3 is applicable. Setting
ej ej
r·
J
== ')Ij log2 s} -> K(lOg2 SJ)1 +/1 == q.,J
this lemma guarantees qj i 00 all large j (and for convenience this will be
supposed for allj 2: 1); the lemma also certifies for any IX in [0, 2) that
~ K. f
n=1
(lOg2 S;)1 +p-a [
JIQn S y 2<qn+tl
y2 dF(y) < 00
Corollary 6. If s; -+ 00, ')In = 0(1), and {Y, Y", n 2: 1} are i.i.d. random
variables with E Y = 0, the LIL holds for {Un Y,,} and {- Un Y,,} iff E Y 2 < 00.
PROOF. The hypothesis implies (46), whence (47) ensures
~
L, p{ 2
Yt >
~ n log2 s;a~} <
U 00, J > 2a~. (51 )
n= I 1'n
p{lim ILJ=t
n~ 00 sn(log2
aj~~
sn)
= oo} = 1. (52)
PROOF. Let {Y:, n ~ I} denote the symmetrized {Yn } and for c > 0 set
- Y*I
Y 'n- n IIY~I:scl' X'n-an
- Y'n,an d ac2 -- E y'2
n' Thensn,2 =L,j=taXj-acsn,
- "n 2 _ 22
whence {X~} obey the conditions of Theorem I, implying
-
{n~
P lim
00
d- }
L~-I a·Y'·
Sn Og2 sn)
t~2 > ac
}
= 1.
By Lemma 4.2.6
(53)
and since ac --+ 00 as c --+ 00, (53) holds with ac replaced by + 00, which, in
turn, yields (52). 0
EXERCISES 10.2
I. Show under the conditions of Theorem 1 or Corollary 7 that
~ S. r:. a.c.
hm ". 2 ". 2 1/2 = V 2,
'-00 (L.,j=1 X j IOg2 L.,j=1 Xj)
2. If a.A.. ~ e - 1, the upper bound in (3) of Lemma 1 may be replaced by
exp{ -(A..' x;)fan(e - I)}.
3. Verify that the LIL holds for independent r.v.s X. distributed as N(O, a;) provided
5; = L~ a1 -+ 00, a. = 0(5.). Hint: Use a sharp estimate of the normal tail for the
dJ. of S./s•.
according as eJl > 0 oreJl ~ (e - 1)a2 /d•. Hint: Apply Lemma 1 with A.. = eJl/a 2 ,
x. = an l /2 .
5. When {X.} are independent with EX. = 0, EX; = a;, = D af, X. S;
5; c.s. i,
a.c., lim c.x. = 0, check via (2) that for all y > 0, r > 0, and all large n
p{ max Sj >
t SJ'5n
(I + y)'x.s.} s; exp{ -!x;(I + y)2.-I}.
6. Under the conditions of Theorem I, show that with probability one that every point
of [ -1,1] is a limit point of S./s.(2Iog 2 S.)1/2. Hint: For d # 0, 0 < d < I, y > 0,
setting ak = (I - y)dh k, bk = (I + y)2 dh k, and 7k = s •• - S•• _" for all large k
P{ghak < 7k < bkgd = P{7k > akgd - P{7k > bkgd
-(I + y)ai } {-(I + Y)bf} 1 {-(I - Y)d hi }
2
~ Cyexp { 2 - exp 2 > "2Cyexp 2
via Exercise 5.
9 Let S. = D'=I Xi where {X, X., n ~ I} are i.i.d. with E X = 0, E X 2 = 1 and let T
be an {X.}-time with E T < 00. If T,. = Ij=1 TW where TW, j ~ 1 are copies of T,
then
10. If {X., n ~ I} are i.i.d. with E ei'x, = e-I'I', 0 < a. ~ I, prove that
P{Ilill!n-I/'S.ll/lou. = ell'} = I,
that is, P{IS.I > n /'(Iogn)(1+t ll', i.o.} = Oor 1 according as/; > 0 or /; < O. Hint:
l
Show that P{n-I/'IS., > x} = P{lX II> x} and use the known (Chapter 12)
fact that PiX II > x} - Cx-' as x --+ 00.
11. If {X., n ~ I} are interchangeable r.v.s with E X I = 0, E xi = I, Cov(X I' X 2) =
0= Cov(Xi, X~), then 1lill(2n log2 n)-112 D
Xi = I, a.c.
12. For {X., n ~ I} as in Lemma 2 except that (*) lim d.x./s. = a > 0, x. --+ 00, prove
that for all y in (0, 1) and all u in (0, uo)
t
pin (0, (0) there exist positive,finite constants A p , Bp such that
where IX I , .•• , IX j are positive integers with L{= I IX i = 2k, A<lI •...• <lj
(IX I + + IXj)!/IX I !'" IX j !, and i l , ... , i j are distinct integers in [I, n]. Since
E Xf; Xf{ = I when IXI>' •• , IX j are all even and zero otherwise,
E S2k
n
= "L, A 2fJ,,···.2fJj.,
c 2fJ , ... c 'j2fJj ,
E Sn2k -- L AzfJl
A
..... zfJj . A fJ
I ... ·•
ZfJ, ... c·ZfJ )
fJ j c·" Ij
fJ, ... ·• fJj
~ B~:s;k,
where s; = L7= cf and I
{31!"'{3j!
BZk
Zk = sup AzfJl ..... ZfJj = sup (2k)!
-----
AfJ, ... fJj (2{31)!'" (2{3)' k!
2k(2k - I)···(k + I) 2k(2k - I)· .. (k + I)
< sup . < ------:,.......,:----,-...,...----~
- nl= I 2{3l2{3j - 1) ... ({3i + I) - 2fJ , + ... +fJj
Thus, when p = 2k the upper inequality of (I) holds with Blk ~ kllz. Since
IISnllp is increasing in p, IISnllp ~ IISnib ~ Blks n for p ~ 2k, whence the
upper inequality of (I) obtains with Bp ~ kllz, where k is the smallest integer
~ p12.
It suffices to establish the lower inequality for 0 < p < 2 since IISnllp ~
I Snllz = Sn for p ~ 2. Recalling the logarithmic convexity of the Y p norm
established in Section 4.3 and choosing r l , r z > 0 such that r l + r z = I,
pr l + 4r z = 2,
s; = IISnll~ ~ IISnl!:"IISnll:'2 ~ IISnll~"(21IZsn)4'Z,
whence
IISnll~" ~ 4-'2S;-4'2 = 4-'2S~",
IISnilp ~ 4-'2lp"sn
Hence, the lower inequality holds for 0 < p < 2 with Ap ~ 4 -'21p'l =
2-(Z-Pllp and for p ~ 2 with Ap ~ 1. 0
Ee =
,s2
~ E SZj ~ ~ V /2 s)Zj = L ~ (tsZY < (tsZey
j=O)' j=O}' j=O}' j=O
since /Ij! < L:= Z
0 j"In! = ei. Thus, E e,s2 < 00 for ts e < 1. Finally, since
386 10 Limit Theorems for Independent Random Variables
Sn -+ S, for any t > 0 the integer n may be chosen so that 2te(s2 - s;) < 1.
Then
E e,S2 = E e,(S-s.+s.)2 ~ E[e2IS~ . e21(S-S.)'] < 00
.
(3)
whence, recalling Lemma I0.1.1 (or repeating the earlier two-term martingale
10.3 Marcinkiewicz-Zygmund Inequality, Dominated Ergodic Theorems 387
argument),
and so (2) follows from (4) and (5), the upper and lower constants B p and A p
being twice and one half respectively those of the Khintchine inequality. 0
iff
EIXI' < 00, r> I, and EIXllog+ IXI < 00, r = I. (9)
PROOF. Since {X: , n 2 I} and {X;; , n 2 I} are each i.i.d. with moments of the
same order as {Xn}, Example 7.4.3 stipulates that {(Iln) L1 xt,.? n' n 2 I}
and {(Iln) D X j- , .?n, n 2 I} are positive (reversed) martingales, whence
(9) and Theorem 7.4.8 (35) ensure that (8) holds with X replaced by X + or
X -. The exact conclusion (8) then follows by Lemma 4.2.3.
Conversely, if(8) obtains forr 2 I,EIXII' ~ ESUPn~1 n-'IL1 Xii' < 00,
so that only the case r = 1 needs further attention. Now
Esupn-IIXnl=Esupn- 1 ~Xj-
n n- t n I I I
~Xi
I
~2Esupn-1 ~Xi <00,
388 10 Limit Theorems for Independent Random Variables
ro r
Xl Xl
f
00
OO
L P{ I X I 2
00
= foo L I1n,,;'-'IXIJ dt dP
00
00 >
fM n=)
nt}dt
lIXI;z,MJ M n=1
Lemma 1. If{ Y,., n 2 1} are independent, nonnegative r.v.s then E(Lf YS < 00
for some r 2 I provided
00 00
n=)
L E Y~ < 00, LEY: <
n= 1
00, (10)
(12)
\
E sup
IL?;\ XJ <
,/2 00 (13)
n;,e" (n log2 n)
iff
X 2 log IXI
EIXI' < 00, r> 2, and E log21XI IllXI>e"1 < 00, r = 2. (14)
n n
S~ = LX;, s~ = LX;"
i I
a:>
It follows from Example 7.4.9 and (1) of Lemma 10.2.1, noting g(l) < 1, that
for 0 < tb n ~ 1
for some positive constant a. Therefore, choosing Uo such that auo > 5,
it follows via (15) that
a:>
f "0 U,-I P{sup cnlS~1 2: u}du ~ 4J, fa:>
a:>
"0 U,-I exp{ -(au - 1)log(k + 1)}du
a:> 1
< c~ k 2 < 00,
For r > 2 the converse follows from the necessity part of Theorem 3.
When r = 2, Theorem 3 merely yields E xi < 00. However,
X; 2E S; + S;_I
E sup ~ sup < 00,
n~e' n log2 n n~e' n log2 n
10.3 Marcinkiewicz-Zygmund Inequality, Dominated Ergodic Theorems 391
and so, choosing AI > ee such that P{IX I < AI} > 0,
~ }J I
P{X 2 < Mj IOg2 j} f: ~ P{X; ~ tn IOg2 n}dt.
Now E X 2 < 00 entails positivity of the infinite product, whence for some
C>O
00 > (00 L P{X 2 ~ tn IOg2 n}dt = f (00 L I'nlo82n:sx2t-'j dt dP
JM n IIXI2:C] J M n
>
- i lIX!2:C] M
f X2(t IOg2X2X t2 I - ee ) dt dP
X2
i f
X2
> 2 dt dP + O( 1)
- lIX/2:C] M t IOg2 X
X 2 10glXI
=2E IOg21XI IlIX1>e-j+O(I). D
EXERCISES 10.3
I. Show via examples that Theorem 3 is false for 0 < r < I and Theorem 4 fails for
I ::; r < 2.
2. Verify under the hypothesis of Theorem 4 that for any finite {X n}-time T
E[(T log2 T)- l l2jSTIJ < oc.
3. If X(1), X(21 are LV.S. with respective symmetric dJ.s F and G, where F(x) ~ G(x) for
all x > 0, then F.(x) ~ Gn(x), x > 0, where F n (resp. Gn) is the n-fold convolution of
F ( resp. G) WI'th't If H ence, I'f SU)
I se. n -- "n
L"i = 1 XU)·
i ' ) -- I , 2,were
h i ' I _< I. <
XU) _ n, are
are i.i.d., then EIS~I)IP ::; EIS~2)IP, n ~ I, P > O.
4. Show that in Theorem 3 sufficiency and necessity for r > I extend to interchangeable
LV.S. Is (9) necessary for r = I?
S.1f {Xn,n~ I} are independent LV.S with EXn=O, n~ I, and pE(I,2], then
EmaxlsjsnII{=1 Xd P ::; ApD=1 EIXjlP for some finite constant Ap. Hint:
Use the Doob and Marcinkiewicz-Zygmund inequalities.
6. USn = I7=1 Xi where {X, X n, n ~ I} are i.i.d. r.v.s,prove that E exp{t sup cn'Snl} < 00
for some t > OiffE{exptlXI} < 00 for some t > 0 where Cn = (nlog 2 nf l /2, n ~ ee.
392 10 Limit Theorems for Independent Random Variables
PROOF. Suppose without loss of generality that E I X I > 0, E(X +)P < 00, and
EIXIY < 00. Clearly, for any k > 0, by Lemma 5.3.5,
P{Sn ~ na} :::;; nP{X > na/2k} + Pk{Sn ~ na/2k}. (3)
(4)
I :::;; C 1 2
Ekl XI Y:::;; C(EIXl y)(a p -I I!(aY-I). (6)
n: 1
10.4 Maxima of Random Walks 393
;a
Hence, from (2), (5), and (6)
co
L nP - 2 P{Sn ~ n} ~ CpEIXI P ,
n=l
co (7)
Ln P
-
2
P{S: ~ n} ~ 2C p EIXI P .
n=l
PROOF. (8) follows immediately from Corollary 1 applied to X/£ while the
converse is the special case IX = 1, p = 2 of Theorem 6.4.5. 0
00
J(e):$ max(2 a p-2, 1) L nap - 2
P{Y. ~ en a}, (14)
n= 1
=
f OO
o
V, p
-
2 P
{(M(e») 1/.
--
e
>
-
V
}
dv = -1- E [M(e)]I.p-I)/.
--
ap - 1 e '
so that
E[L(r.)]·P-1 = (ap - 1) L oo
v· p- 2 P{L(e) ~ v}dv ~ (ap - 1)I(e). 0
t
and EX °
Lemma 2. Let {X, X n, n ~ l} be i.i.d. r.v.s with EIXI I /. < 00, where a >
= if a ~ 1. If Sn = D Xi and
where e > °and sup 0 = 0, then E V(e) < 00 implies E LH2e) < 00, Y > 0.
PROOF. Set A j = {Xj ~ 2ej·}, B j = {ISj- II ~ ell. Now n-·Sn ~o by
Theorem 5.2.2, and so P{Bn} -+ 1 as n -+ 00. Since for n ~ 1 the classes {B n}
and {An' AnA~+ I""} are independent, by Lemma 3.3.3 for n ~ no
J I (e) = Loo
v· p - 2
P{X v ~ w·}dv. (19)
(21)
PROOF. The second inequality of (20) follows immediately from
E(X+Y = a.p 1 00
$ I + 2ap 1 00
Finally, if X' = max(X, c), C > 0, then (22) holds with X+ replaced by X'.
IfE(X+Y = 00, then, since a.p > 1, E(X')I/" = o(E(X'Y) as C -. 00, implying
J 1(1) = 00. Thus, (21) obtains. 0
Then, (i)
and J(£) :::;; 1(£) < 00, where the latter are as in (12) but with Sv replacing v".
(iii) Let lX> 1, EIXII/<x < 00, and EX = 0 if ex :::;; 1. If either of the
conditions of (23) holds or 1(£) < 00 or J(£) < 00, then E(X+Y < 00.
PROOF. By Lemma 3, E(X+Y < 00 iff J 1(1) < 00 iff J 1(£) < 00 for all £ > O.
Then Lemma 1 ensures E[M 1(£)]l<XP-ll/<x < 00,1(2£) < 00, £ > 0, and hence
E[LI(£)]<xP-I < 00 since L I (£) :::;; L I (£). Conversely, by Lemma 1
E[M 1(£)]l<X P- Il/<X < 00
implies 11(£) < 00, £ > 0, and hence J \(£) < 00, £ > O. Moreover, if
E[LI(£)]<xP-I < 00, £ > 0, then by Lemma 3, J I(£) < 00, £ > O.
Apropos of (ii), if 1 < lX < 1, the first half of (23) follows from Theorem 2.
Then by Lemma I, J(£) :::;; 1(£) < 00 and E[L(£)YP-I ::; E[[(£)]<X P- \ <: 00.
If, rather, lX 2 1, define X~ = XnI[Xn~ -C]' C > O. Then S~ = D Xj 2 Sn'
Since E(X+Y < 00 for some p 2 1 (p > 1 if lX = 1), necessarily y ='
min(p, 2) E [1, 2] and EIX'II Y < 00. Hence, by Theorem 2
n )](<X P - I)/<X
,L
:::;; E [ sup ( (Xj - E X'I) - w<X < 00
n~O )=1
whence E[L(£)YP-I < 00. Then by Lemma 2, E[LI(£)YP-I < 00, implying
E(X+Y < 00 by part (i). 0
Corollary 5. If{X, X n , n 2 1} are i.i.d. with E X = Jl. > 0, E(X-)2 < 00 and
No = sup{n 20: Sn :$ O}, where Sn = L~ Xi' then E No < 00 and
f p{inf Sj O} <
n=1 J",n
:$ 00.
:$ LP
<0 {
sup j (L li - Jl.)
- J1
> n- } < 00
n=1 j",1 i=1 2 2
by Corollary 3.. Since {No 2 n} = U~n {Sj :$ O} c {infj ",. Sj :$ O}, neces-
sarily E No < 00. 0
For any r.v.s Y, Z let the relation Y '" Z signify that Y and Z have identical
distributions.
n21. (24)
(26)
Since (5' + X')- :::; (X')- = X-, 5' :::; 5, and lime_oo 5' = S, by Lebesgue's
dominated convergence theorem E 5' ~ E Sand ais, + X') - --+ afs+ X) - and
(iii) follows from (26), 0
E(t;V _ - Su _) r
J[U+<ool
(Su + - t;V J ----> (J; ,
(J2
E( - ST - ). E ST + <
-
-
2'
III. J[U + < 00] (Su + - t;V +) < 00 iff E M(t;) < 00 iff E(X+)2 < 00.
PROOF. By Corollaries 5.4.1, 5.4.2, T+ and L are finite stopping times and,
clearly, lim,_o V ±(t;) = T±. The second equality of (i) follows immediately
from Theorem 5.4.2. Apropos of the first, define VO) = V +, V1i + J) = 00 on
{VIi) = oo} and otherwise V n = Ii=J VIi) where
V(i+J) = inf { n z I: t
U·+n
j=U;+l
(Xj -
}
t;) > 0 , i z l.
(30)
Now, setting
n z I, (31)
necessarily
00
Since
00
00
To prove (ii), note that via Wald's equation, the initial notation of (31),
and part (i),
E( - Wv J . E Wv • / [U • < 00 I = e E V - . E Wv • / [U • < 00 I
(72
= e E M(e) --+ -
2
Since both E( - Sr _) and E Sr. are positive, each is therefore finite and (ii)
is proved.
Finally, as already observed, E V _ = I/P{ V + = oo} < 00, whence (iii)
follows from (i) and Corollary 3.
References
L. E. Baum and M. Katz, "Convergence rates in the law of large numbers," Trans.
Amer. Math. Soc. 120 (1965), 108-123.
H. D. Brunk, "The strong law of large numbers," Duke Math. J. 15 (1948), 181-195.
Y. S. Chow, "A martingale inequality and the law oflarge numbers," Proc. Amer. Math.
Soc. II (1960), 107-111.
Y. S. Chow, "On a strong law of large numbers for martingales," Ann. Math. Stat. 38
(1967),610.
Y. S. Chow, .. Delayed sums and Borel summability of independent, identically dis-
tributed random variables, Bull. [nst. Math., Academia Sinica I (1973), 207-220.
402 10 Limit Theorems for Independent Random Variables
Y. S. Chow and T. L. Lai, "Some one-sided theorems on the tail distribution of sample
sums with applications to the last time and largest excess of boundary crossings,
Trans. Amer. Math. Soc. (1975).
Y. S. Chow, H. Robbins, and D. Siegmund, Great Expectations: The Theory ofOptimal
Stopping, Houghton Mifflin, Boston, 1972.
K. L. Chung, "Note on some strong laws of large numbers," Amer. J. Math. 69 (1947),
189-192.
K. L. Chung, "The strong law of large numbers," Proc. 2nd Berkeley Symp. Stat. and
Prob. (\951), 341-352.
J. L. Doob, Stochastic Processes, Wiley, New York, 1953.
V. A. Egorov. "On the strong law oflarge numbers and the law of the iterated logarithm
for sequences of independent random variables." Theor. Prob. Appl. 15 (1970),
509-514.
W. Feller, An Introduction to Probability Theory and Its Applications, Vol. 2, Wiley,
New York, 1966.
W. Feller, "An extension of the law of the iterated logarithm to variables without
variance," Joum. Math. and Mech. 18 (\968),343-356.
B. V. Gnedenko and A. N. Kolmogorov, Limit Distributionsfor Sums of Independent
Random Variables (K. L. Chung, translator), Addison-Wesley, Reading, Mass.,
1954.
P. Hartman and A. Wintner, "On the law of the iterated logarithm," Amer. Jour. Math.
63 (1941), 169-176.
P. L. Hsu and H. Robbins, "Complete convergence and the law of large numbers,
Pmc. Nat. Acad. Sci. U.S.A. 33 (\947),25-31.
A. Khintchine, "Uber dyadische Bruche," Math. Zeit. 18 (1923),109-116.
A. Khintchine. "Uber einen Satz der Wahrscheinlichkeitsrechnung," Fund. Math. 6
(1924),9-20.
J. Kiefer and J. Wolfowitz, "On the characteristics of the general queueing process with
applications to random walk," Ann. Math. Stat. 27 (1956),147-161.
J. F. C. Kingman, "Some inequalities for the queue GI/G/I," Biometrika 49 (1962),
315-324.
A. Kolmogorov, "Uber der Gesetz des Iterierten Logarithmus," Math. Annalen 101
(1929),126-135.
M. Loeve. Probability Theory, 3rd ed., Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
J. Marcinkievicz and A. Zygmund, "Sur les fonctions independantes," Fund. Math. 29
(1937). 60-90.
1. Marcinkiewicz and A. Zygmund, "Quelques theoremes sur les fonctions indepen-
dantes," Studia Math. 7 (1938),104-120.
S. V. Nagaev, "On necessary and sufficient conditions for the strong law of large
numbers," Theor. Proh. Appl. 17 (1972),573-581.
Y. V. Prohorov. "The strong law of large numbers," Izv. Akad. Nauk. Ser. Mat. 14
(1,950),523-536 [in Russian].
Y. V. Prohorov, "Some remarks on the strong law of large numbers," Theor. Prob.
Appl. 4 (\959), 201-208.
D. A. Raikov, "On a connection between the central limit theorem in the theory of
probability and the law of large numbers," /zv. Nauk USSR, Sov. Math. (1938),
323-338.
D. Siegmund, "On moments of the maximum of normed sums," Ann. Math. Stat. 40
(\969),527-531.
F. Spitzer, "A combinatorial lemma and its applications to probability theory,"
Trans. Amer. Math. Soc. 82 (\956), 323-339.
F. Spitzer, "The Wiener-Hopf equation whose kernal is a probability density," Duke
Math. Jour. 24 (\ 957). 327-343.
References 403
V. Strassen, "A converse to law of the iterated logarithm," Z. Wahr. 4 (1966), 265-268.
H. M. Taylor, "Bounds for stopped partial sums," Ann. Math. Stat. 43(1972), 733-747.
H. Teicher, "A dominated ergodic type theorem," Z. Wahr. 8 (1967), 113-116.
H. Teicher, "Some new conditions for the strong law," Proc. Nat. Acad. Sci. U.S.A. 59
(1968), 705-707.
H. Teicher, "Completion of a dominated ergodic theorem," Ann. Math. Stat. 42 (1971),
2156-2158.
H. Teicher, "On interchangeable random variables," Studi di Probabilita, Statistica e
Ricerca Operative in Onore di Giuseppe Pompilj, pp. 141-148, Gubbio, 1971.
H. Teicher, "On the law of the iterated logarithm, "Ann. Prob. 2 (1974), 714-728.
H. Teicher, "A necessary condition for the iterated logarithm law," Z. Wahr. Verw. Geb.
(1975),343-349.
H. Teicher, "Generalized exponential bounds, iterated logarithm and strong laws,"
Z. Wahr. Verw. Geb. (1979), 293-307.
N. Wiener, "The ergodic theorem," Duke Math. Jour. 5 (1938) 1-18.
A. Zygmund, Trigonometric Series, Vol. I, Cambridge, 1959.
11
Martingales
Definition. If r l < r 2 and a l , •.• , an are finite, real numbers, the number of
upcrossings of the interval [r I' r 2] by the sequence a I' a 2 , ••. ,an is defined
as the number of times the elements a j pass from "on or below r l " to "on or
above r2'" More precisely, let 0(1 be the smallest integer (if any) for which
aa, ~ r l , and in general for j ~ 2 let O(j be the smallest integer (if any) ex-
ceeding O(j_ I for which
404
ll.l Upcrossing Inequality and Convergence 405
If 2u is the largest even integer j for which (Xj is defined then u is called the
number of upcrossings (if (X2 is undefined, then u = 0).
+1 S2 ::s; E S2'
[S, >0]
Suppose inductively that (1) holds with n replaced by n - 1 for all sub-
martingales, and set
1::S;j::S;n-2
So ifO < S._I < r
T,,- 1 = { S._I otherwise.
For A E a(Tj , I ::s; j ::s; n - 2) = a(Sj' 1 ::s; j ::s; n - 2)
Hence, E{T,,- ,11j, 1 ::s; j ::s; n - 2} ~ T,,- 2' a.c. Clearly, for 2 ::s; m < n - 1
E{Tm l1j, l::S;j::S; m - I} ~ Tm - I, a.c., and so {1j'~j' l::S;j::S; n - I} is
a nonnegative submartingale.
Let r be the number of upcrossings of [0, r] by (TI , ... , T" _I)' Then
U = V + I[Sn_I=O.Sn~r],
whence by the induction hypothesis
ErU + ES I = ErV + ErI[Sn_I=O,Sn~rl + ES.
::s; E T,,-I + ErI[Sn_l=O.Sn~rl
= 1 [0 <Sn _, <rl
s. + 1 [Sn _ , ~ rl
S._I + ErI[Sn_,=O,Sn,U]
yielding the initial portion of (i). Moreover, iflim n__ 00 E Sn =K > - 00, for
-oo<nS;O
r
)lIs.l~aJ
ISn I = r
)[s.~aJ
Sn +(r
)[s.> -aJ
Sn) - E Sn
S; r
)[S.~al
Sm + (r )[s"> -aJ
Sm) - K
S; r
JIIS"I ~aJ
ISm I + E Sm - K < r
)IIS"I ~aJ
ISm I + 8 < 28 (2)
11.1 Upcrossing Inequality and Convergence 407
for large a since P{ ISn I ;;::: a} :5: a- I E ISn I -+ 0 as a -+ 00. The conclusion of
(2) clearly also holds for the finitely many integers n in Em, OJ, and so uniform
integrability is established.
Next, for A E :IF _ ro and every m :5: 0
Is - ro =
A
lim
n- -ro A
I
Sn
A
~ ISm,
whence E{Sm I:IF _ ro} ;;::: S _ ro' a.c. for all m, concluding the proof of (i).
Apropos of (ii), {S~, :lFn , - 00 < n ~ O} is a nonnegative submartingale
which is easily seen to be u.i., whence the conclusion of (ii) follows. 0
Since U-statistics
Um,n = (n)-I
m
I
1 Sil <O"<im$n
<p(X j " ... , XjJ, n;;::: m,
Corollary 3.If{Sn, :lFn, - 00 :5: n :5: O} is a submartingale with E Sri < 00 and
PROOF. By Theorem 1, Sn *
lim n__ ro E Sn = K > -00, then Sn* S and E{SI:IF -ro} ;;::: S-ro, a.c.
S as n -+ 00. For A E :IF - ro
(3)
n--oo
408 11 Martingales
n- - 00
PROOF. The initial portion of (3) is Corollary 7.4.1. Apropos of the second,
note that by Fatou, EI Lool < 00. If Dn = E{Y"Ig;n} - E{Loolg; -oo}, then
for n ~ m
lim IDnl
11--00
~ lim E{SUP1 }}- L0011g;n}
11--00 )5.m
Theorem 2 (Austin). If {Sn, g;n, -00 < n < oo} is a martingale with
sUPnEISnl = K < oo,thenI:~oo(Sn - Sn_t)2 < oo,a.c.
PROOF. By Theorem 1, Sn~ an g; _oo-measurable r.v. S-oo as n --+ -00.
Since S-oo E ft'l' it may and will be supposed that S-oo = 0, a.c. Set X n =
Sn - Sn-l' whence by Corollary 2, Ii; _00 X j = Sn, a.c. For any C >
define T = inf{n > - 00: ISnl > q, where inf{0} = 00. Then, noting that
°
ISjl ~ Con {T > j},
1 I XJ
[T;oo] -00
n
~ I
n
j;-oo
1 [T>jl
(Sj - Sj_ d 2
Now
In
-00
1
[T>Jl
In {(1
(SJ - SJ- 1) = -00 [T>Jl
SJ - 1
[T>j-ll
SJ- 1) +1
IT;jl
SJ- t }
=1 S;+1
IT>nl IT"nl
Sf-l~C2+1IT"nlSf-l'
11.1 Upcrossing Inequality and Convergence 409
i r
j=-ooJ[T>jl
Sj_I(Sj - Sj_l) = ±r
-ooJ[T=j)
(Sj_ISj - SJ-I)
:s;; r
J[T:5.)
(C1STI - S}_I)
:s;; CK - r
J[T:5.]
S}_I'
recalling Exercise 7.4.4. Consequently, (5) ensures that for all n > - 00
r
J[T=oo)
±
-00
XJ :s;; C 2 + 2CK.
Thus, L~oo XJ < 00, a.c. on the set {T = oo}. By Corollary 2, P{T = oo} =
P{sup.~ -00 IS.I :s;; C} --+ 1 as C --+ 00, and so L~ 00 XJ converges a.c. 0
with c, = 2', r #- 2, C z = 1,
00
E L IX~Z)I ~ 2 EIXTI[T<oo)1 ~ 2 E X*, (9)
1
where
00
~ 2'-1 L E(lXnI[T>nl + IE{XnI[T>n)l~n- d I')
1
00 T-I
~ 2' E DXnl'I[T>n) = 2' E L IXnl'
1 1
/
E(~ IX~3)l'r/' ~ E(~ IXnl'r ' = E v,.,
and, since {X(3)' > O} c {v,. > K},
P{X(3)' > O} ~ P{v,. > K} ~ K- 1 E v,.,
11.1 Upcrossing Inequality and Convergence 411
I x; Y;
00
implying EX" < 00. By (i), S~ converges a.c., whence If X n Y" converges
a.c. on {T = oo}. Since Sn converges a.c. and Y* < 00 a.c., P{T = oo} ~ 1
as K ~ 00, whence If X n Y" converges a.c. 0
EXERCISES 11.1
2. If {X n , n ;;:: I} are !L'I interchangeable r.V.S, prove that I/n L~ Xi~ some LV.
Y (also in !L'I)' Verify that Y = E{X11nf ~n}' where ~n = a(D X,,};;:: n). Hint:
n- 1 D Xi = E{n- I D Xd~n} = E{X II~n}'
f~ = max IJiI, n ~ 1,
I s;js;.
00 ) 1/2 • )1 /2
S(f) = Soo(f) = ~ dJ
(
' S.(f) = ( ~dJ .
and so it remains to prove the first inequality in (1). Setting T" = T /\ n for
n ~ 1,
.-1
S;-I(f) + f;-, = 2 L
ls;jS;kS;.-1
djd k =2 L dj(f.-,
j~1
- Ji-I)
Since IfTJTn-11 ::;; A.(A. + IfTI) and ElfTI < 00 by Exercise 7.4.4, the first
inequality of (1) follows as n -+ 00. 0
Lemma 3. Letf be a nonnegative submartingale, 0< e < 00, y" = Sn(ej) v f:,
n ~ 1. Then,for A. > 0, {3 = (1 + 2e 2)1/2, and p E (1, (0)
9p 3/2
IISn(f)ll p ::;; - 1 Ilfnllp, (5)
p-
9p 3/2
IIS(f)ll p ::;; - 1 Ilfll p. (6)
p-
PROOF. Define I j = I[sj(9f»,,) and gj = IJj,j ~ 1. Since Ij+1 ~ I j, neces-
sarily g = {gn, n ~ I} is a nonnegative submartingale. Let T = inf{n ~ I,:
Sn(ef) > A.}. On the set {Sn(ef) > {3A..f: ::;; A.}, note that T ::;; n, g: ::;; A., and
IdTI = 1fT - fT-t1::;; fT v fT-I ::;; f:::;; A., so that, recalling the definition
of {3,
(1 + 2e2)A 2 < S;(ef) = S}-I(ef) + e 2d} + e 2 L dJ
T<j'5,n
::;; A2 + e 2A2 + e 2 L (gj - gj_I)2
T<j'5,n
::;; (1 + e 2)A 2 + e 2S;(g),
implying that Sn(g) > A on the set in question. Hence, applying Lemma 2 to
{gt> ... , gn, gn," .},
A P{Sn(ej) > {3A, f: ::;; A} ::;; AP{Sig) > A.. g: ::;; A} ::;; 211gnll = 211 I nfnll·
On the other hand, by Theorem 7.4.8 (34)
:$ 3 i [Y n > Al
fn
:$ ~I
P-
Ilfnllpll Ynll~-I,
implying
Choose () = p- liZ, whence fJP = (I + (2jp»pIZ < 3 and (5) follows. Finally,
let n -+ 00 in (5) to obtain (6). 0
I :$j:$ n,
11.2 Martingale Extension of Marcinkiewicz-Zygmund Inequalities 415
Corollary 2. Iff = Un, ffn , n z I} is an !£I2r martingale such that for some r z 1
(II)
integrable. To this end, set m = [v'] for v> 0, whence, employing Theorem
7.4.8,
P{Z ~ v} ~ P{T ~ v'} + P{T ~ m, Z ~ v}
in view of (29) of Corollary 7.4.7 (which holds in general) and (12). Con-
sequently, from (13) and (14)
IIS.U)II = E S.U) = E ( JI
TA.
1
)1/2
= E(T /\ n)1/2 -+ 00,
11.2 Martingale Extension of Marcinkiewicz-Zygmund Inequalities 417
so that the first inequality of(7) fails for p = 1. However, the second inequality
does hold. More precisely, Corollary 1 obtains when p = 1, as will be shown
in the next theorem.
P{f* > fJA, S(f) v V* ~ tJA} ~ (fJ _ ~~ 1)2 P{f* > A}, (15)
9tJ 2
P{S(f) > fJA, f* v v* ~ tJA} ~ fJ2 _ tJ2 _ 1 P{S(f) > A}. (16)
hn = LdjI[IJ<j$vl\al
j= 1
:::;; [(/32 - 1 - J2)A 2 r I E (g*)2 :::;; /32 _ ~~2_ I P{S(f) > A}, 0
00
PROOF, The validity of(l9) is clear. On the set {jdjl > 2dj_ d, Idjl + 2dj_ t :::;;
21djl :::;; 2dj, implying
00 00
:::;; 2 E d* + 2 E d* = 4 E d*, o
Theorem 2 (Davis). There exist constants 0 < A < B < 00 such that, for any
fill martingale f = {f", n ~ I},
A E S(f) :::;; Ef* :::;; BE S(f), (22)
11.2 Martingale Extension of Marcinkiewicz-Zygmund Inequalities 419
9<5 2
P{S(g) > p..1., g* v 4d* ~ <5..1.} ~ p2 _ <5 2 _ 1 P{S(g) > ..1.}.
Hence,
2<5 2
P{g* > P..1.} ~ P{S(g) > <5..1.} + P{4d* > <5A.} + (P _ <5 _ 1)2 P{g* > ..1.},
9<5 2
P{S(g) > P..1.} ~ P{g* > <5..1.} + P{4d* > <5,1,} + p2 _ <5 2 _ 1 P{S(g) > ,1,},
implying
2<5 2
P- I Eg* -< <5- t ES(g) + 4<5- 1 Ed* + Eg* (25)
(P - <5 - 1)2 '
p- 1 E S(g) ~ <5 - 1 E g* +
9<5 2
4<5 - I E d* + p2 _ <5 2 _ 1 E S(g). (26)
In order to subtract in (25) and (26), rewrite these for the martingale
{gt> g2, ... , g., g., g., ...}, obtaining via (27) and (28)
2
2<5 ]
[ p-l - (P _ <5 _ 1)2 Eg: ~ <5- 1 ES.(f) + 8<5- 1 Ed: ~ 9<5- 1 ES(f),
(29)
(30)
420 11 Martingales
For small b, the coefficients on the left in (29), (30) are positive, and so, letting
n -+ 00 and recalling (23), (24), there exist constants B I , B, AI, A for which
E f* ~ BI E S(f) + 4 E d* ~ BE S(f),
E S(f) ~ A I E f* + 4 E d* ~ A - I E f*. 0
Set
Y~ = Y. - Y~.
•
~ (nM)p + LEI YjlP ~ (nM)p + nt:
I
and (31) follows. If, rather, 1 ~ p < 2, then Theorems 1 and 2 guarantee the
existence of a constant C with
EXERCISES 11.2
I. Let {X., n ~ I} be independent LV.S with EX. = 0, EX; = I for n ~ 1. If T =
inf{n ~ l:D X j > O}, then P{T < oo} = I and E T I /2 = 00.
3. If {X, X., n ~ l} are martingale differences with EX = 0, EIXI P < 00 for some
p ~ 2, and {S., n ~ I} are the partial sums, then {IS./n 1/ 2 IP, n ~ l} are u.i. Hint:
Consider EIS~IP+l and EIS;I Pwhere X~ = X.I[IXnl$MJ - E{X.I[IXnl$MJIX 1 ••• X.- 1 }
and X; = X n - X~.
4. If {X., n ~ I} are martingale differences and p E (I, 2J then E SUP.2 I ID X X :s;
A p L;'O EI X.I P for some finite constant A p .
(I)
where f* and S(f) are defined in Section 2. Here, it will be shown for any
convex, nondecreasing function <1> on [0, 00) with <1>(0) = 0 and
Lemma 1. Let <1> be a nondecreasing function on [0, 00], continuous on [0, 00)
with <1>(0) = 0, <1>( 00) = <1>(00 - ), and
<1>(2.1) :s;; c<1>(A) for all A E [0, 00) and some c > O. (3)
Iff and g are nonnegative, measurable functions on (0, ff, P) and b, e, f3 are
positive constants with f3 > 1 satisfying
eC I +Iogp < 1 where 210gp = f3, (4)
P{g> f3A,f :s;; bA} :s;; eP{g > A} for all A E (0,00), (5)
there exists A = A e• P. o. e E (0, 00) such that
E<1>(g):s;; A E<1>(f). (6)
PROOF. Suppose without loss of generality that E <1>(f) < 00. From (3) and
(4) there exists y = 'Ie. p E (0, 00) and Yf = Yfe.o E (0, 00) with 'Ie < 1 such that
for all A> 0
(7)
422 II Martingales
Let gn =9 /\ n for n ~ 1. Then gn satisfies (5), so that from (6) and (9)
implying
(1 - ')'t:)E <l>(gn) ~ ')'11 E <I>(f),
and hence (6) and A = ')'11/(1 - ')'t:). o
Lemma 2 (Zygmund). If<l> is a convex function on an interval [a, b), where a
is finite, there exists a nondecreasing, integrable cp on [a, c)for every c E (a, b)
such that
°
[0, 00], finite and convex on [0, 00) with <1>(0) = 0, <I>( 00) = <I>( 00 - ) = 00,
and <I>(2A.) ~ c<l>(A.) for all A. ~ and some c E (0, 00). Then there exists a
constant B = Be E (0, 00) such that for every sequence {Zn, n ~ I} of non-
negative measurable functions on (0, ~, P)
(11 )
11.3 Convex Function Inequalities for Martingales 423
• 00
In verifying (II) it may be supposed that E Z > 0 and E «I>(Z) < 00. Then
convexity and cI>(00-) = 00 entail E Z < 00, so that E W < 00. For A. ~ 0
define the S;;.-time T = inf{n ~ 0: JtY,,+ 1 > A.}, whence
i
[O,;T,;n]
(W - Z - A.) =
j=O
f i
[T=j)
(~ - A. + W - ~- Z)
~
j=O
f i
[T=j)
E{W - ~- ZIS;;j} ~0
or
i [O,;T,;.)
(W - A.) ~ i [O,;T,;.)
Z,
where the integral is in the Lebesgue-Stieltjes sense. From (14), recalling (12),
= 100
~ fOO
o
r
J[W >")
Zn dP dq>(u) + q>(O)E Zn
f
n
implying
'P(t) = LljJ(U)dU.
(20)
PROOF. When <I>(t) = t P, (13) yields d = p, whence (20) follows directly from
(19).
00 00
(27)
11.3 Convex Function Inequalities for Martingales 427
then for any A. > 0, f3 > 1, and b E (0, f3 - 1) there exists a finite, positive
constant Ba such that
B ba
PU* > f3)., s(f) v d* ::;; bA.} ::;; (f3 _ ~ _ b)a PU* > A.}. (28)
Theorem 2. Let «I> be a nondecreasing function on [0, 00], continuous on [0, (0)
with «1>(0) = 0, «1>(00) = «I>( <X) - ), and «I>(2A.) ::;; c«l>(A.) for all A. > and some
c E (0, (0). Then, for every IX in [1, 2] there exists a finite positive constant
°
B = Be• a such that for any martingale f
E «I>(f*) ::;; B E «I>(s(f» + B E «I>(d*), (29)
PROOF. Choose f3 = 3 and bE (0, 1) such that Ba baC 3 < 1, where B is the
constant in (28). Then by Lemmas 1 and 4 there exists B = Be. a E (0, (0)
such that
E «I>(f*) ::;; B E «I>(s(f) v d*) ::;; B E[«I>(s(f» + «I>(d*)]. 0
428 11 Martingales
PROOF. Set
X~ = XnIIIXnl9nJ - E XnIIIXnI9nJ'
X~ = XnIIIXnl>snJ - E XnIIIXnl>snJ'
and Sn = L~ Xj' S~ = D Xi, s; = D Xi· Then Sn = S~ + S; and, by
Theorem 2, for some constant C E (0, 00)
]('+ll/2
L E(Xi?
n
EIS~I'+ 1 ~ C [ + C E max IXil'+ 1
1 I$j$n
~ C(S~+I + 2r+ls~+I).
*
Again via Theorem 2, for some B in (0, 00)
since as noted in Section 9.1 (see, e.g., (10) therein), a Lindeberg condition of
order r > 2 ensures that of lower orders. Thus, E IS:/sn I' = 0(1), implying
{I S:/sn I', n ~ I} is uniformly integrable. Consequently, the same is true of
{ISn/snl', n ~ I}, and so, in view of the central limit theorem (Theorem 9.1.1)
and Corollary 8.1.8, the conclusion (31) follows.
for some finite nonnegative, nonincreasing function cp on [0, 00) and, moreover,
= - 100
°
Lemma 6. If <I> is a nondecreasing function on [0, 00], finite and concave on
(0, 00) with <1>(0) = and <I>( 00) = <1>(00 - ), thenfor every sequence {zn, n ~ I}
of nonnegative measurable functions on (n,:#', P)
°
PROOF. Define
for A >
Jv", Zn, W = Woo, and Z = Zoo as in the proofof Lemma 3, and
set T = inf{n ~ 0: Jv,,+ t > A}. Then WT ~ W A A and
00 00
= E WT ~ E(W A A),
whence
E(Z A A) ~ E(ZT + A.I IT < 00)
Corollary 3. If °
< p < IX and 1 ~ IX ~ 2, there exists a constant
A = A<I E (0, (0) such that, for any martingale f,
(37)
T
E L
i= I
E{IXil11~i_d < 00, (38)
(39)
PROOF. When y = 1, the first two lines of the proof of Theorem 7.4.6 (with X:
replaced by IXIII) yield (39). Fory > 1, set v",11 = Ll= I E{IXlllal~i_I}' IX> 0,
n ~ 1, and Yi = IXiI - E{IXill~i-d, i ~ 1, and consider the martingale
f.. = Ll=l YiIIT;;,il' n ~ 1. Since DO=1 E{I YiIIIT~i]I~i-d = L[E{I Yill~i-d
~ 2V1 ,T, it follows, taking IX = 1, et>(x) = Ixl 1 in Theorem 11.3.3, that
1 1
L L
E TAll Yi 1 ~ E sup I" YiIIT;;,i) 1
II II ~ I I
= B·lE(2V1,T)1 + E JI IIT;;,ilE{IYiI11~i-dJ
I
~ 21 + B[E Vl,T + E ~,T].
Consequently,
E ( TAll
~ IX;! )1 ~21 E TAli
~ [I Yi 11 +EVl,T] =0(1),
and so monotone convergence yields (39). o
Theorem 4. If {SII = LJ= I Xi' ~II' n ~ I} is a martingale and T is an ~II-time
with
T )112
E L E{IX;!11~i_d <
T
00, E ( ~ E{Xfl~i-1 < 00 (40)
I
EXERCISE 11.3
1. Prove Corollary 11.2.3 by applying Corollary 11.3.3.
E f3+V <_13_
!f 1 + af3 + V-I + ap'
PROOF.
1 1 aV
1 + af3 + V - 1 + ap = 1 + af3 + V 1 + ap
=
aV (1 1 av)
1 + ap 1 + ap + 1 + af3 + V 1 + ap ~
2 2
aV + a V
(1 + ap)2 ' (3)
and so via (2)
1 1 1 a E!f U
1 + af3 + E!f U - 1 + ap 1 + af3 + E!f U 1 + ap
< - a 2 ai( U)
(4)
- (I + af3 + E!f U)2'
11.4 Stochastic Inequalities 433
E 1 _ E 1
'§ 1 + 1X{3 + U - '§ 1 + 1X({3 + E'§ U) + (U - E'§ U)
1 1X2CJ~( U) 1
< + <--.
- 1 + 1X{3 + E'§ U (I + 1X{3 + E'§ U)2 - 1 + 1X/3
Consequently,
E f3 + U
'§ I + 1X{3 + U
= E ! [-I +
'§ IX
1 + IX ] <! [-I
1 + 1X{3 + U - IX
+~]
1 + 1X/3
f3
=1+1X/3"
o
E~
Sn+1
n 1 + IXSn+1
-E
- ~ Sn+Xn+1 Sn
<----"--,~
n 1 + IXS n + X n+1 - 1 + IXS n
o
According to the ensuing theorem, it is uncertain that a nonnegative super-
martingale Sn starting in (0, I) will ever exceed one if the conditional coef-
ficients of variation of the differences are bounded away from zero.
PROOF. Let T = inf {n ;;::: I: Sn ;;::: l} and set v" = L7= I lj, where
y" = XnI[T>nl + (l - Sn-I)I[T=nl' So = o. (8)
On [T = n], 0 < Yn = 1 - Sn-I ~ X n , and so Y; ~ X; on n, whence
E'F n_I Y; ~ E~n_ IX;, n > 1. On the set [T ;;::: n]
E~n_. Yn ~ E~n_,Xn ~ -IXCJ}n_,(X n),
so that E}n_. Y" ; : : E}n_ .Xn, implying CJ}n_ .(y") ~ lX}n __ ,(Xn) on [T ;;::: n].
Therefore, (6) obtains for {y"} on [T ;;::: n], and on the complement [T < n]
434 11 Martingales
it is trivially true since y" = O. Hence, Lemma 2 ensures that {v,,/( I + ex v,,),
~n' n ~ I} is a supermartingale. Consequently, noting that VT = I, setting
T(n) = min[T, n], and employing Theorem 7.4.5.
· E
< I1m VT(n)
. < E Y1
- n-oo I + exVT(n) I + exYI
Since Xl = Y1 , (7) follows. D
U = Sn + K - L:j;o E~j_' XJ n ~ 0,
n K +b '
whence Un = Un - U n- 1 = (K + b)-I[X n - E~n_. X;], n ~ 1, and Uo =
K(K + b)-l = Uo . Note that for n ~ I
-':1
E ~n-I U
n
- - E ~n-' X n2
=K+b
and
a}n_.(u n) = (K + b)-2 E~n_. X; = -(K + b)-l E~n_' Un' n ~ I,
so that for ex = (K + b)
(10)
Then, via Theorem 1
E~ e
YU
~ I + Y' E!J V Z ~ exp(Y' E!J V 2 ). (12)
.r.
{ ex p {- )=1 Uj(U+ ~j .)'}COSh A( Y+L~n
lUI U+
.), jOn,
lU)
n ~ I}
is a supermartingale.
PROOF. Designate the putative supermartingale by {v", jOn, n ~ 1} and set
Uo =u. By Lemma 3, for n ~ 2
~ exp { - n
.LUj
(A)' (A )'} A(yLn.
" j . + Un "n . cosh
+8 n - 1)
~ Vn -l·
)=1 L.o U • L.o U ) oU)
Setting Vo = cosh(Ayju), the same argument shows that irE Xl = 0, {v", jOn,
n ~ O} is likewise a supermartingale. 0
E cosh "T +E 8
A(y T)
XZ ~ e
u(A)' Ay
- cosh - . (13)
U + L.l 'j_1 j U U
436 11 Martingales
( A)j (A)j(- -
U )2
(U ~ a)' I1- -< I1-
00 00
j=2j! U + a- j=d! U U + a
= (u: ar(~r
implying
I u· (A)'
n
j=l
- .- < (A)'
IbUj -
J
- I u·
j=l U
n
J
( - .- -U)2
If=ou j
< u2
-
-
U
(1 1)
(A)' I-;=;-:--
n
I Ib I Uj D=ou j
(14)
recalling (14). o
~ eXP{ASn_ 1 - X ~u j + XU n}
= eXP{AS n_ 1 - X nt1Uj},
I - IX 8 I + IX -8 I I+IX
<--e +--e < for 0 :::; lJ :::; log I _ IX
- 2 2 -
and
P.+IC ;f)=PP{A.+I(~~f)IXI = I}
+ q P { A. + 1 C~ f) IXI = - 1}
= p + q P{A.(f)} = P + qP.(f)·
Hence, for 0 :::; f :::; 1 and n 2: 0
P.+ 1(f) = pp.(2f) + qp.(2f - 1), (20)
and obviously (20) holds for f :::; 0 or f 2: 1. Consequently, (20) obtains for all
f E (-00,00) and all n 2: O.
To prove (19), define for 0:::; f - s :::; f and n 2: 0
!l.(f, s) = P.+ I(f) - PP.(f + s) - qP.(f - s). (21)
If f + s < 1, then !lo(f, s) = PI(f) 2: 0, while if f + s 2: 1, monotonicity
of P. and (20) ensure !lo(f, s) 2: O. Hence, for 0 :::; f - s :::; f and n = 0
(22)
Now from (20) and (21)
!l.(f, s) = p(P.(2f) - P.(f + s)] + q(P.(2f - 1) - P.(f - s)],
and employing (20) once more
!l.+ l(f, s) = p 2[P.(4f) - p.(2f + 2s)] + q2(P.(4f - 3) - p.(2f - 2s - 1)]
+ pq(P.(4f - 1) - p.(2f + 2s - 1)
+ p.(4f - 2) - p.(2f - 2s)], (23)
and so
!l.+ l(f, s) = pA.(2f, 2s) + q{p(P.(4f - 2) - p.(2f + 2s - 1)]
+ q(P.(4f - 3) - p.(2f - 2s - 1m· (24)
To verify (22) for n 2: 1, suppose inductively that it holds for a fixed
but arbitrary integer n. If (i) f - s 2: t, then 2f - 1 2: 2s, whence by (24)
!l.+I(f, s) = p!l.(2f, 2s) + q!l.(2f - 1, 2s) 2: 0; if (ii) f + s :::; t, then 2f-
2s :::; 2f + 2s :::; 1 and again via (24), !l. + 1(f, s) 2: p!l.(2f, 2s) 2: 0; finally, if
t
(iii) 0 :::; f - s < < f + s, then 1 < f, and by (23)
!l.+ l(f, s) 2: pq[p.(4f - 1) - p.(2f + 2s - 1) + P.(4f - 2) - p.(2f - 2s)].
(25)
Since q 2: p, if s > 1, then from (25)
!l.+ I(f, s) 2: p{(P(P.(4f - 1) - p.(2f + 2s - 1)]
440 11 Martingales
Lemma 6. Let {X n, :l'n, I ::;; n ::;; N < oo} be an .ff I stochastic sequence, let
CN be the class of all stopping rules T with P{T ::;; N} = I and define
YN = yZ = XN,
(26)
Yn = Y: = max [X n, E{Yn+ II:l'n}], I::;; n < N.
Then, if a = inf{n ~ I: X n = Yn},
supEXT=EYI =EX u ' (27)
TeCN
PROOF. Clearly, {Yn, :l'n, I ::; n ::; N} is a supermartingale. Thus, for any
TEC N ,
Moreover,
Lemma 7. For any random variables Yt> ... , YN setting X n = llYn "2 I) and
:l'. = a(YI , ... , y"), I ::;; n ::;; N,
p{ max Yn
l:5,n:5,N
~ I} = EYI (28)
P{IsjsN
max lj ~ I} = P{IsjSN
max Xj = I}
= P{T < N} + P{T = N, X N = I}
=EXr=EYI
by Lemma 6. o
TheoremS (Dvoretzky). Let {X., n ~ l} be i.i.d. with P{X I = I} = P = 1 -
P{X I = -I} and q = 1 - P ~!. Set !!F. = <1(X I ,· .. ,X.), n ~ 1, and
!!F o = {0, Q}. For any constant 0 < f = fo ~ I, designate the fortunes
associated with the bold strategy by
f. = In-I + in-IX. where ii = min(a, 1 - a). (29)
If g., n ~ I, are any !!F.-measurable functions with 0 ~ g. ~ 1 and h., n ~ 1,
are the fortunes associated with this alternative betting strategy, that is,
ho = f and
n ~ 1, (30)
then for aU n = 0, 1, ... , and aUf E (0, 1)
pt~:SXN h ~ I} = Edo,
j (33)
d. = 1 = PN-.(h.) if h. ~ I,
while if h. < 1, via the induction hypothesis with probability one,
d. = E{d.+ I I!!F.} ~ E{PN-n-I(h.+ I)I!!F.}
= E{PN_._I(h. + h.g.Xn+I)ISiO.}
442 11 Martingales
p{ max hj
O~;,j:sN
~ t} = Edo ~ PN(ho) = PN(f). o
References
D. G. Austin, "A sample property of martingales," Ann. Math. Stat. 37 (1966), 1396-
1397.
D. Blackwell, "On optimal systems," Ann. Math. Stat. 25 (1954),394-397.
B. M. Brown, "A note on convergence of moments," Ann. Math. Stat. 42 (1971),
777-779.
D. L. Burkholder, "Martingale transforms," Ann. Math. Stat. 37 (1966),1494-1504.
D. L. Burkholder, "Distribution function inequalities for martingales," Ann. Prob-
ability 1 (1973), 19-42.
D. L. Burkholder and R. F. Gundy, "Extrapolation and interpolation of quasi-linear
operators on martingales," Acta Math. 124 (1970),249-304.
D. L. Burkholder, B. J. Davis, and R. F. Gundy, "Inequalities for convex functions of
operators on martingales," Proc. Sixth Berkeley Symp. Math. Stat. Prob. 2 (1972),
223-240.
Y.S. Chow, "On a strong law of large numbers for martingales," Ann. Math. Stat. 38
(\967),610.
Y. S. Chow, "Convergence of sums of squares of martingale differences," Ann. Math.
Stat. 39 (1968),123-133.
Y. S. Chow, "On the Lp-convergence for n-IIPS", 0< p < 2," Ann. Math. Stat. 42
(1971),393-394.
K. L. Chung, A Course in Probability Theory, Harcourt Brace, New York, 1968; 2nd ed.,
Academic Press, New York, 1974.
B. Davis, "A comparison test for martingale inequalities," Ann. Math. Stat. 40 (1969),
505-508.
1. L. Doob, Stochastic Processes, Wiley, New York, 1953.
L. E. Dubins and D. A. Freedman, "A sharper form of the Borel-Cantelli lemma and the
strong law," Ann. Math. Stat. 36 (1965),800-807.
L. E. Dubins and L. J. Savage, How to Gamble /frou Must, McGraw-Hill, New York,
1965.
A. M. Garsia, "On a convex function inequality for martingales," Ann. Probability 1
(1973),171-174.
R. F. Gundy, "A decomposition for g\-bounded martingales," Ann. Math. Stat. 39
(\968), 134-138.
J. Neveu, Martingales a temps discrets, Masson, Paris, 1972.
References 443
R. Panzone, "Alternative proofs for certain upcrossing inequalities," Ann. Math. Stat.
38 (1967),735-741.
E. M. Stein, Topics in Harmonic Analysis Related to the Littlewood-Paley Theory,
Princeton Univ. Press, Princeton, 1970.
H. Teicher, "Moments of randomly stopped sums-revisited," Jour. Theor. Prob. 9
(1995), 779-793.
A. Zygmund, Trigonometric Series, Vol. I, Cambridge, 1959.
12
Infinitely Divisible Laws
Row sums L~::'l X,,; of arrays of random variables {X,,;, 1 ;s; i ;s; k" -+ 00,
n ~ I} that are rowwise independent have been considered briefly with respect
to the Marcinkiewicz-Zygmund type strong laws of large numbers (Example
10.4.1). In this same context, non-Iterated Logarithm laws and generalizations
thereof have been dealt with by H. Cramer and C. Esseen (see references at
the end of this chapter). Here, limit distributions of row sums of the variables
in such an array will be treated.
It is a remarkable fact that the class oflimit distributions of normed sums of
i.i.d. random variables is severely circumscribed. If the underlying L v.s, say
{X"' n ~ I} have merely absolute moments of order r, then for r ~ 2 only
the normal distribution can arise as a limit, while if 0 < r ;S; 2, the limit law
belongs to a class called stable distributions. If the basic LV.S are merely
independent (and infinitesimal when normed cf. (l) of Section 2), a larger
class of limit laws, the so-called class !f' emerges. But even the class !f' does
not contain a distribution of such crucial importance as the Poisson. A
perusal of the derivation (Chapter 2) ofthe Poisson law as a limit of binomial
laws B" reveals that the success probability associated with B" is a function
of n. Thus, if B,,_ I is envisaged as the distribution of the sum of i.i.d. random
variables Yl , ... , Y,,-l, then B" must be the distribution of the sum of n
different i.i.d. random variables which, therefore, may as well be labeled
X". l' X". 2' ... , X".". In other words, to obtain the Poisson law as a limit of
distributions of sums of i.i.d. (or even independent) random variables, a
double sequence schema {X "j' j = 1, ... , k,,_ oo} must be employed (with
X".I"'" X".k n independent within rows for each n = 1,2, ...). Under one
further proviso, the class of limit laws of (row) sums of such r.v.s coincides
with the class of infinitely divisible laws.
444
12.1 Infinitely Divisible Characteristic Functions 445
Definition. AdJ. F is called infinitely divisible (i.d.) if for every integer n ;::: 1
there exists adJ. Fnsuch that F = Fn * Fn * ... * Fn = (F n)"' or equivalently if
its d. cP (also called i.d.) is the nth power of a c.f. CPn for every integer n ;::: 1.
(-Iy-I .
= L .
<Xl
Proposition 4. A finite product of i.d. d.s is i.d. Moreover, if i.d. d.s q>k - q>,
a d., then this limit d. q> is i.d.
PROOF. Clearly, if q> = q>~, t/J = t/J~, n ~ 1, then q>' t/J = [q>n . t/Jn]n, n ~ 1,
shows that a product of two and hence any finite number of Ld. d.s is i.d.
Suppose next that the i.d. d.s q>k - q>, a d. Then, the i.d. d.s t/Jk(t) =
lq>k(t)1 2 = q>k(t)· q>k( -t) - the d. t/J(t) = 1q>(tW. Consequently, t/J~/n as the
positive nth root of the positive function t/Jk tends as k - 00 to the nonnegative
nth root t/Jl/nofthe nonnegative function t/J. Since for n ~ 1, t/J~/n is a sequence
of d.s whose limit t/J l /n is continuous at t = 0, t/J l /n is a d. for n ~ 1. Thus, t/J
is i.d. and hence nonvanishing. Consequently, <p is nonvanishing, whence
Log q> is defined by Lemma 1. By Lemma 2,
Since c.f.s exp{A.(e iIU - 1) + itO}, A. > 0, of Poisson type are i.d., it follows
from Proposition 4 that exp{D= 1 [A.J{eiluj - 1) + itOj ]} and hence also
exp{itO + J~«) (e i1u - I)dG(u)} with G a bounded, increasing function is i.d.
The latter comes close but does not quite exhaust the class of i.d. d.s.
Proposition 5. The class of i.d. laws coincides with the class of distribution
limits offinite convolutions ofdistributions of Poisson type.
PROOF. That every such limit is i.d. follows directly from Proposition 4.
Conversely, if q> is i.d., so that q> = q>~, n ~ 1, then
that is
lim exp{n[lpn(t) - I]} = lp(t).
Now,
n ~ 1,
and a net -00 < - M n = Un,l < U n ,2 < ... < un,kn+1 = M n < 00 may be
chosen whose points are continuity points of F n and such that
1
max(u n j+1 - un) ~-3
j' '2n
and
Then for It I ~ n, choosing An, j = n[Fn< Un, j +I) - Fn( Un, j)],
1 f"n,i+ 1
~"2 dFn(u).
n "n,j
<Xl .~ An,){eil"n,i - I
If -<Xl
(eil" - l)n dFn(u) - 1)
)-1
f.
kn 1 f"n'i 1
~ 2n dFn(u) + L"2 + 1
dFn(u) ~ -.
lI"I~Mnl 1 n "n,i n
2onsequently, for It I ~ n and sufficiently large n
I
en('Pn(tj-Ij - .n
)=1
exp(Anjeit"n,i - 1»\
n-oo j= 1
1». 0
Theorem 1. <p(t) = exp{l{!(t; y, G)} as defined by (1) is an i.d. d.for every real
y and G as stipulated. Moreover, <p uniquely determines y and G.
PROOF. The integrand h(t, u) of l{! satisfies Ih(t, u)1 ::s; C < 00 for It I ::s; T
and all real u. Choose -M n = Un,l < Un ,2 < ... < un ,kn +l = M n to be
nonzero continuity points of G for which
y G(un,j+ I) - G(un,j)
an,j = kn - u .
n,)
'
n exp(itan,j + AnJeilUn.i -
kn
= lim 1»,
ft-(I) j=l
H(u) =
Theorem 2. Let {y, YN' n ~ I} be finite real numbers and {G, GN, n ~ I}
nondecreasing left-continuous functions of bounded variation which vanish at
- 00.1fYN -+ Yand GN~ G, then t/I(t, YN' G,,) -+ t/I(t; Y, G)for all real t, where t/I
is as in (1). Conversely, if t/I(t; Y", GN ) tends to a continuous function g(t)
as n -+ 00, then necessarily g(t) = t/I(t, Y, G) and YN -+ Y, GN~ G.
PROOF. °
If G(oo) = 0, then G,,(oo) -+ by complete convergence, whence
t/I(t; YN' G,,) -+ t/I(t; Y, G) = iyt, recalling that the integrand h(t, u) of (1) is
bounded in modulus for fixed t. If G( (0) > 0, then G,,( (0) > for all large n
by Lemma 8.1.1, whence (ljG,,( 00 »t/I(t; Y", G,,) -+ (ljG( (0» t/I(t; Y, G) by the
°
Helly-Bray theorem, and so t/I(t; y", GN ) -+ t/I(t; Y, G).
Apropos of the final assertion, Theorem 1 ensures that the i.d. c.f.s
= e"'(I; Yn, G n ) -+ e9(I), continuous. Thus, e9(I) is a c.f. and i.d. by Theorem 1
e"'n(l)
and Proposition 4. Define oc(t) = Log e9(I) and ocN(t) = Log e"'n(l) = t/lN(t). By
Theorem 8.3.3 e"'n(l) -+ e9(I) uniformly in It I ~ T for all T E (0, (0), whence
by Lemma 2, t/lN(t) -+ oc(t) uniformly in It I ~ T and oc(t) is continuous. Hence,
recalling the proof and notation of the last part of Theorem 1 and defining
'+ foo-00eily dHiy),
f1-1 t/lN(y)dy =
I
y"(t) = 2t/1,,(t) -
f
'+l
V(t) = 2oc(t) - 1_IOC(y)dy,
12.1 Infinitely Divisible Characteristic Functions 451
it follows that v,,(t) ..... V(t), continuous, and, in particular, Hft(oo) = v,,(0) .....
V(O), whence V(O) ;;::: O. If V(O) = 0, then
x) (1~
f
2
Hft(oo) = 2 oo ( 1 -
-00
sin
-~
+ dGft(x) ..... 0,
X )
implying Gft( (0) ..... 0, whence G(u) == 0 and necessarily ')1ft tends to a finite
limit ')I. If V(O) > 0, the dJ.s Hft(u)/Hft( (0) (whose d.s v,,(t)jVft(O) ..... V(t)/V(O»
converge to a limit dJ., say H(u)/V(O). Thus, Hft -4 H, and by the Helly-Bray
theorem for any continuity point u of H, recalling Theorem 6.5.2,
f
u ( sin y) - 1 y2
2Gft(U) = 1- - -1--2 dHft(Y)
-00 Y +Y
f
u ( sin y) - 1 y2
--+ 1- - --2 dH(y). (2)
-00 y 1+Y
Define G(u) to be the integral on the right side of (2). Since the continuity
points of G and H are identical, Gft ..:. G. Hence, ')1ft tends to a finite limit ')I.
Clearly, r/I(t, ')I, G) = g(t). 0
From the preceding, a canonical form for i.d. cJ.s known as the Levy-
Khintchine representation follows readily.
oo
f
U
= it --2 n dFft(U)
-00 1 +u
(1
f
2 2
+ . - 1 - -itu-) -+-
oo ( e"u u ) - -un d F (u).
_ 00 1 + u2 u2 I + u2 ft
Set
nu fU ny2
f
oo
')1ft = -1--2 dFft(u), Giu) = -1- 2 dFft(y),
-00 +U -00 +y
r/lft = r/I(t; ')1ft, Gft)·
452 12 Infinitely Divisible Laws
=- oo .
iIU
(eilU - e- )2 1 + u
2
- 2 - dG(u) = -
foo (sin tU)2
- - (1 + u2) dG(u),
f -00 21t u -00 tu
whence
-.p"(O) ~ lim
,-0 I (
lIul:s 1/1]
-
tu
tu) 2 (1
sin - + u2)dG(u) ~ (sin 1)2 fOO
- 00
(1 + u2)dG(u).
Thus, G*(u) = J~ 00 (1 + y2)dG(y) has the asserted properties i.e., G*( 00) < 00
and
J(
2
eilU - 1- ~)
1 + u2
(1 +u2u )dG(U) = JI~(ei'U - 1-~) ~ dG*(u)
1 + u2 u2
u12 (u - 1 : u 2) = 0(1),
necessarily r(t) = .p(t; y, G) + itc for some constant c, where
f
dG*(u)
-> iy* +i (e itu - I) - u - = f'(t) (5)
~
ei(t"-t)u - 1 - i(t' - t)u
,
I + lu(1 - e-itU)1 :s; u 2 (! + Itl),
I t - t
which is integrable with respect to u - 2 dG*(u). Analogously, as t' -> t
it u itu
f'(t'), - f'(t) =1. f e ' , - e dG*( U ) - > - f eitu dG*( u.)
t - t u(t - t)
Thus, f and hence qJ has a finite second derivative, whence the transform of qJ
has a finite second moment. Moreover, f" and hence qJ" uniquely determines
G* and therefore also y*. From (5), y* = - if'(O) is the mean ofthe underlying
distribution and it is readily verified that G*( (0) is the variance. 0
EXERCISES 12.1
3. Verify that the function G of (3) has a finite moment of even order 2k iff the same is
true for the underlying i.d. distribution.
4. Let cp be a non-i.d. d. having the representation (3) with G not nondecreasing (but
otherwise as in (1». Prove that y and G are still uniquely determined by cp.
5. If
1X(1 - fI) I - f3 k
PIX = -I} = - - - - , PIX = k} = - - ( 1 + 1Xf3)f3 ,k = 0, I, ... ,
I+IX I+IX
where 0 < IX < f3 < I, show that X has an i.d. c.f. cp(t) iff IX = 0 and, further, that
Icp(tW is i.d. even when IX -:f. O.
6. Show for an i.d. c.f. cp(t) = exp{ljJ(t; y, G)} that if the support of its d.f. F is bounded
from below, so is that of G. Is the converse true?
7. Prove that if CPo = cpZ" k ~ I, where CPk is a c.f. for k ~ 0 and nk is a sequence of
positive integers -+ 00, then cp is i.d.
8. Prove that if {X n' n ~ I} are i.i.d. r.v.s with d.f. G and N is a Poisson LV. (parameter A.)
J
independent of {X n' n ~ I}, then the c.f. of 1~ Xi is exp{A. (e itu - l)dG(u)}.
9. Show that an i.d. mixing G of an additively closed family ff = {F(x; A.), A.EA c: Rm}
yields an i.d. mixture H. Hint: Recall Exercise 8.3.14.
454 12 Infinitely Divisible Laws
with corresponding dJ.s {Fn.d and c.f.s {q>n.d such that within each row,
i.e., for each n ~ 1, the r.v.s X n. l , X n. 2 " " , X n. kn are independent. The r.v.s
of an array such as in (1) will be called infinitesimal if
that is, if the row elements become uniformly small in the sense of (i).
Exactly as in the proof of the weak law of large numbers (Section 10.1),
this implies
(ii) max Im(Xn.k)1 = 0(1),
I sksk n
(iii) max i
ISkSkn Ilxl<rl
Ixl'dFnk(x) = 0(1) for all r > 0, r > O.
(i')
or
(i") max II - q>n.k(t) I = 0(1) uniformly in It I =:;; T for all T > O.
1 SkSk n
e
-1--
2
max P{lXn.kl > e} =:;; max
+ e2 k k
x 2 dFnk(x)
-1--
IIxl ~ tJ + x
i 2
= 0(1),
12.2 Infinitely Divisible Laws as Limits 455
as n --+ 00, and then E; --+ 0, so that (i) implies (in). Finally, since 1 - 9l {cp} ::;;
11 - cpl, Lemma 8.3.1 stipulates a positive constant a(c, <5) such that
max P{lX.d > c}::;; a(c, <5) rlmaxll - CP•. k(t)ldt = 0(1)
k o k J
for all c > 0, whence (in) ensures (i). o
For fixed but arbitrary f, 0< f < 00, define
Lemma 2. If {X.,d are infinitesimal and {F.,k> cPn.d are defined by (2), then
for any f, T > 0 and n ~ NT' there exist positive constants c i = Ci(T, f),
i = 1,2, such thatfor 1 ::;; k ::;; k.
+ r it(x -
IJUxl<r] a)dF(x) I+ 2P{!XI ~ f}
::;; T
2
2
r
JUxl<r)
(x _ a)l dF + T r (x -
IJUxl<r) a)dF I
+ 2P{!XI ~ f}. (3)
456 12 Infinitely Divisible Laws
Ir (x - a) dFI = lal r dF
JlIxl<'J
S;
lal[1
JlIxl:;::,]
+ (r + lal)2]
(r - lal)
2
I IIxl:;::']
(x - a)2
d
( ) 2 F,
1+ x - a
(4)
sup II - CPn.k(t)1
~ S;
2 + lal
[I + (r + lal)] -2 + ( _I 1)2
2 (T 2
T) f-1--
x
dFn.
2
-
k
III ~ r r a + x2
where, noting that (i) entails max k Ian. k I < r/2 for all large n,
-< 1+-
4 r2 C '
1
yielding the lower bound. Next, if F* denotes the dJ. of a LV. X* with c.f.
Icp(tW, from the elementary inequality
sin
( I - -~
TX) (I~
+x
2
)
~ c(T) > 0,
oo
- dF*(x)
Tx
f
X2
~ Te(T) --2 dF*(x). (5)
_ool+x
For any LV. X with dJ. F and median m, let X* denote the symmetrized X
and define
Fm(x) = PIX - m < x}, qm(x) = P{IX - ml ~ x},
q*(x) = P{IX*I ~ x}.
12.2 Infinitely Divisible Laws as Limits 457
f:oo 1 :2X2 dF =
m
l°Otf"~dC :2 X2 )::; 21°O q *(X)dC :2 X2 )
oo
f
X2
=2 --2 dF*(x). (6)
_ool+x
Moreover, from the elementary inequality
2(m - a)(x - a) and the first equality in (4),
:; i [Ixl <r]
(x - m)2 dF + 2r(r + Iml) i
lIxl2: r]
dF,
whence
f~
+
dF = f
1 x 1
(x - a)2
+ (x - a)
2 dF ::; i
lIxl<rJ
(x - a? dF + i lIxj2:r]
dF
But
i IIxl<r]
(x - m)2 dF ::; [1 + (r + Im1)2] i IIxl<r)1
(x - m)2 2 dF
+ (x - m)
oo X2
::; [1 + (r + Im1)2]
f - - 2 dF m
_ool+x
and as in (4)
<
1 + (r + Iml)2 foo - -x 2d Fm
- (r - 1m 1)2 _ 1 + x2
00 '
so that, combining these with (7) and recalling (6) and (5),
2d 2
Jo (l
( 2
::; Tc(T) - 1<p(u)1 )du. (8)
458 12 Infinitely Divisible Laws
d 2 = [1 + (r + Im I)]
2 (
1+
1 + 2r(r + Im
(r _ 1m 1)2
I»)
exists and is continuous at t = 0, then for any r in (0, (0) there exists a constant
C depending on rand {lpn.k} such that
k
L
k=l
n
f+ x
2
-
2 dFn.k(x) :s;
-1--
X
c.
PROOF. Since n~':.l lq>n.k(tW -+ f2(t), continuous at zero,f2 is a d., whence
T may be chosen so thatf2 > i for It I :s; T. Then, by uniform convergence
n~nllpn.k(tW >! for n ~ NT and It I s; T and D':.lloglq>n.k(t)l-+ logf(t)
uniformly in ItiS; T. Hence from Lemma 2
L
k
n foo x
2
-1--2 dFn.k(x) :s; -C2 L
k
n iT logl q>n.k(t) Idt
k=l -00 +x k=10
-+ -c 2 f:IOgf(t)dt. 0
Lemma 4. If {Xn.d are infinitesimal LV.S such that for some r in (0,00) there
J
exists a C in (0, (0) with D':.l [x 2j(1 + x 2)]dFn. k(x) :s; C, n ~ 1, then
L~':.l [Log <Pn.k(t) - (<Pn.k(t) - 1)] = 0(1) for all real t. Moreover, for any
constants An and all real t
n q>n.k(t) -
kn
Log e- itAn I/I(t; Yn' Gn) = 0(1), (9)
k=l
where
Furthermore, infinitesimality implies (iv), whence Log qin,k is well defined for
It I ::;;
T and 1 ::;; k ::;; k n provided n ~ n(T, t), whence under these circum-
stances
ILog qin,k(t) - (qin,k(t) - 1)1 ::;; lqin,k(t) - 11 2 .
Thus, since T is arbitrary, for all real t if n ~ n (T, Itl),
C ~
::;; - max l4>n,k(t) - II = 0(1). (10)
CI I "k"k n
Next,
L I"
n 2
Gn(u) = k
-I-x- 2 dFn,k,
-
I -00 +X
(9) follows from (10). D
The connections between i.d. laws and the array of (1) is unfolded in
Theorem 1 below.
Theorem l. If {X n, b I ::;; k ::;; k n --+ 00, n ~ I} are infinitesimal, rowwise
independent LV.S, the class oflimit distributions ofcentered sums L~~ I X n,k - An
coincides with the class ofi.d.laws. Moreover, D~ I X n. k - An .!4 i,d. distribu-
tion characterized by (y, G) iff Yn --+ Y, Gn .s. G, where
Yn =- An + ~
k=1
(an'k X 2 dFn,k(X~,
+ f -1
+X ')
L
k
Gn(u) = x -
n
-1- - 2 dFn,k(X),
f" 2
k=1 -00 +X
and t is an arbitrary but fixl:d constant in (0, (0).
460 12 Infinitely Divisible Laws
An = I (an.k + f~
k= I +
dFn.k) - Y +
1 X
0(1)
k [1
L
n
- Fnk(u)] --+
foo -1 -+2-x 2 dG(x), 0< U E C(G) (11)
1 U X
lim lim
,-On-ook=1
I [r J1xl<,]
X2 dFnk(x) - (r
J1xl<'J
X dFnk(X») 2] = G(O+) - G(O-),
(12)
- An + Ir
1 J l1 xl<rJ
X dFnk(x) --+ y + r
J l1 xl<tl
X dG(x) - r
J l1 xl<:rJ X
~ dG(x) (13)
(14)
and
This will be effected by proving (14) <:> (14') <:> (II '), (12') <:> (I 1), (12), and,
moreover, that when (14) obtains (13)<:>(15). In this schema, (11') is (II)
12.2 Infinitely Divisible Laws as Limits 461
t
kn _
Fnk(x) =
kn
t IX_
I + UZ UZ
00 ~ I + UZ dFnk(u)
_
z z
= IX I +zu dGn(u)--+ IX I +zu dG(u), x < 0,
-00 u -00 U
L
k
n
[I
_
- Fnk(x)] =
Joo -I Z
+
- dGn(u)
Joo -I +z - dG(u),
U
Z
--+
U
Z
x> O.
I X u x u
Conversely, (11 '), (12') = (14') since for continuity points u, by the Helly-
Bray theorem,
"
-00 +x
-00 1
I +X
I
XZ Z
--+ - I - -Z --z-dG(x) = G(u), u < 0,
-00 +x x
Gn(oo) - Giu) = 1 00
--+ 1 00
u 2
+x
t
_1_2 ~ f x 2 dFnk(x) ~L f
~ dFnk(x) ~ L f x 2 dFnk>
I + e k=1 JUxl<t) JUxl<t)I + x k k Jllxl<tl
In(e) = I If
k=1 JUxl<tl
x 2 dFnk(x) - [f
JUxl<tl
x 2 dFnk(x) - (f
JUxl<tl
x dFnk(X») 2] I
= 0(1).
Recalling that Iank I ~ an = 0(1), for 0 < e < r (omitting subscripts
temporarily)
f x 2 dFnk - f (x - ank)2 dF nk I
IJUxl<tl JUxl<tl
= If (x - a)2 dF - f (x - a)2 dF I
JUx-al<t) JUxl<tl
= Iilx-al<t:Slxll - lX'<t:s,x-alll
~ e2 f dF + f (e + lal)2 dF
Jlt :slxl:s t+lall Jlt -Ial :slxl <t)
~ (e + an? r
Jlt - an :slxl:s t+anl
dF nk
12.2 Infinitely Divisible Laws as Limits 463
and
I
2
f (x - ank)2 dF nk - f x 2 dF nk + (f x dF nk )
I J11xl«1 JUxl«1 JUxl«J
=
(1
1 Uxl < <I
x dF - a
) 12 - a2
Uxl2!: <I
dF I
=
I(1 1<:5lxl<rl
x dF
) 1
2 - a2
Ilxl2!:<l
dF I
.::;; an r
J I <:5lxl < rl
r dF + a; r
J ux!2!: <I
dF
implying
since, choosing 11 in (0, E), where ±E, E ± 11 are continuity points of G, and
noting that an < 11 for n ~ N. (and recalling that (11) ¢ > (11'))
1 + x2
f
oo
+ - - 2- dG(x) < 00,
< x
= 0(1) as 11 -> o.
464 12 Infinitely Divisible Laws
To complete the proof of the theorem it remains to show that (13) ~ (15)
under (14). Since for all k
= IJrlIx-al<t)(x - a)dF - r
JlIxl < r)
(x - a)dF +a r
JlIxl;;, f]
dF I
< r
IJ[ix-al<f:<;lxll (x - a)dF - r
JlIxl<f:<;lx-all
(x - a)dF I
+Ial r
JlIxl;;'f)
dF
~, r dF
J[f:<;lxl<f+lall
+ (, + lal) r
J[f-Ial:<; Ixl < f]
dF + lal r
J lIxl ;;, f)
dF
via (16), and so, recalling that ±' are continuity points of G, it follows from
I
k=1
f-~
I+x
dFnk(x) = I (1
k=l [iXI<f)
x dFnk - 1Ilxl<f)l+x
~ dFnk(x)
+
I lIxl;;'r)
x2
--
1+ x
-
dFnk(x) )
= 0(1) - r
J[iXI<f]
x dGn(x) + r
JlIxl;;'f]
~ dGn(x)
x
that (13)~(15) under (14). 0
or equivalently
L
k= 1
k
n
I [lxl" £)
dF nk = 0(1), e> O. (18)
(ii) {Xn,d are infinitesimal and LZ~ 1 (X n. k - an,k) has a limiting normal
distribution necessarily N(O, (12) iffjor all e > 0, (18) holds and
lim ~
n k= 1 J[lxl<d
[r x 2dFnk(x) - (rJ[lxl<tl
x dF nk (X»)2] = (12. (19)
PROOF. (i) tfL X n. k - An has a limiting distribution, then (II), (12), and (13)
hold. Moreover, if the limit is normal, since exp{tjJ(t, y, G)} = e-(<1' r' /2)+it9 iff
y = 0, G(u) = 0 for u < 0 and G(u) = (12, U > 0, (11) of Theorem 2 requires
Dn Fnk ( -e) = 0(1) = Dn[1 -
Fnk(e)], e > 0, which is tantamount to (18)
and also to fl~n(l -
S[lxl>t)dFnk(x»-> I, f, > 0, in view of 1- LPi:-:;
fl (1 - Pi) :-:; e - L.p, :-:; 1 (where 0 :-:; Pi :-:; 1). It is therefore also equivalent to
(17)via P{maxJsksdXnkl;;::: e} = I - nZ~J(1 - P{IXn.d;;::: f.}).
Apropos of (ii), (18) and (19) ensure (II) and (12) with G as above and
(12 = G(O+) - G(O-), while (13) is satisfied with An = D~ 1 an.b Y = O.
Thus, D~ J (X n. k - an.d has a limiting normal dJ. N(O, (12) by Theorem 2.
As for the converse, (18) holds via (i), and therefore it suffices to verify that
(12) and (18) entail (19). Now, for 0 < [;' < f,
Lk{I
n
k= I [lxl<d
x 2 dF nk - (I [lxl<t]
)2}
x dF nk
L {I
k" x 2 dF nk - (I x dF nk )2}
k = I. [Ixl < t') [lxl < t']
+ L {I k" x 2 dF nk - (1 x dF nk )2}
k =1 It' slxl < t) [t' slxl < t)
0:-:; L {I
k" x 2 dF nk - (I x dF nk )2}
1
< Ixl < t) slxl < £)
I
k= 1 [t' It'
:-:; L L
n n
k x 2 dF nk :-:; e2 k dFnk = 0(1),
k= J It' slxl < t) k= J [Ixl ~ t']
o< 2 L
k
n
II [lxl <t')
x dF nk
III slxl <t)
x dF nk
I
I
k= 1 It'
~ 2£f,' L
k dF nk = 0(1).
k= J [lxl"')
466 12 Infinitely Divisible Laws
lim "
-
.~oo
L.
k {1 n
k= t Uxl<e)
x 2 dF. k - (1 Ilxl<e)
X dF. k )2}
= fiili L {1 (1 X dF k)2} .
n
k x 2 dF. k -
• ~ 00 k= 1 Uxl < e') llxl < e')
Thus, the prior upper and lower limits are independent of e, whence (12)
ensures (l9). 0
PROOF. In view of
L
k= 1
k
n
1
llxl ;,e)
x 2 dF.k(x)
~maX[I(r
k= 1 JUxl;, e)
XdF. k(x))2,e I
k= 1
r
JUxl;, e)
I XldF. k(X),[;2I
k= 1
r
J uxl ;, e)
dF.k(X)] ,
(22)
(21) implies (18) and also D=
1 (Jlxl <e X dF. k )2 = o(1)for all [; > 0, noting that
E X. k = O. Consequently (21), which is equivalent to D~ 1 JUxl<e) x 2 dF.k(x)
--+ 1 via L~~ 1 E X;k = 1, also implies (19). Then, Theorem 3 guarantees a
limiting standard normal dJ. for D~ 1 X n. k since D~ 1 a.k = 0(1) via (22),
Conversely, if Ihl X •. k is asymptotically N(O, 1), then taking y = 0 = A. in
Theorem 2, (11) and (13) ensure I ~n a. k = o( 1) so that for all [; > 0,
via (19), implying D~ 1 Jllxl <e) X2 dF.k(x) --+ 1 for all [; > 0, and, as already
noted, this is tantamount to (21). 0
Corollary 3. If {X., n ~ I} are independent LV.S with d.f.s {F., n ~ I}, then
(lIB.) L~ Xi - A. has a limiting standard normal dJ. and {XkIB., 1 :$; k :$; n}
are infinitesimal for some sequence {B.} of positive constants tending to 00 iff
for all [; > 0
(23)
12.2 Infinitely Divisible Laws as Limits 467
(24)
where
EXERCISES 12.2
1. If {X n . k , I ~ k ~ kn ---+X.J,n:2: I} are rowwise independent LV.S, prove that
L~n X nk -.':. some constant )I and that {Xn.k} are infinitesimal iff for every e > 0
(i) If I !lxl2:c)
dFnk(x) = 0(1), (ii) If I Ilxl<t)
x dF.k(x) ---+ )I,
(iii) I {f
k= I !lxl<')
x 2 dFnk(x) - (f
. Ixl<'
x dF•. b))2} = 0(1).
(iv) k~l
k
n
SOC:
, dFnjx) = 0(1), (v) I
k= I
f' x dF.ix) ---+ 1.
Jo
3. If {X•. k> I~ k ~ k....... ,Xl} are rowwise independent positive LV.S with finite
expectations satisfying D~ 1 E X. k = I, then D~ 1 X •. k -.':. I and {X •. k} are
infinitesimal iff for every c > 0, (v) holds.
7. Give necessary and sufficient conditions for sums L~n X nk of rowwise independent,
infinitesimal LV.S {Xn.d to have limiting Poisson distributions.
8. Prove that if {X., n :2: I} are independent LV.S, there exist constants An' Bn > 0
such that B; I D
Xi - Anhas a limiting standard normal d.f. and {Xk/B n, 1 ~ k ~ Il}
are infinitesimal iff there exists constants Cn ---+ OCJ with
(vi) if
1 !lxl>Cn)
dF k = 0(1), (vii) -~
C.
i 1
{f !lxl <Cn)
2
x dF k - (f IIxl <Cn!
XdFk)2}""".N.
468 12 Infinitely Divisible Laws
Hint: Under (23) and (24), choose Gn --+ 0 such that Gn Bn --+ CIJ and then determine
nj such that for n 2 nj the left side of (23) (resp. (24» with G = Gj is < I/j (resp.
> I - I/j). Then take C n = GjBn for nj ~ n < nj + I' Conversely, under (vi), (vii)
choose B; to be C; multiplied by the left side of (vii), whence C n = o(B n ) and (23),
(24) hold.
As indicated at the outset of this chapter, the class of limit laws of normed
sums of i.i.d. random variables is a narrow subclass of the infinitely divisible
laws consisting of stable distributions.
Definition. AdJ. F or its d. q> is called stable if for every pair of positive
constants b l , b 2 and real constants ai' a2 there exists b > Oand real a such that
(I)
Clearly, if F(x) is stable, so is F(cx + d), c > 0, so that one may speak of
"stable types." Patently, degenerate distributions and normal distributions
are stable, and in fact these are the only stable dJ.s with finite variance. The
class of stable d.s will be completely characterized but, unfortunately,
explicit expressions for stable dJ.s are known in only a handful of cases.
p {~
B
n .=1
t Xi - An < x} = P {f Xi < Bnx + AnBn}
1
By hypothesis, the left and hence right side of (2) converges in distribution to
F(b.tx + at)*F(bi·x + a2)·Ontheotherhand,(lIBm + n ) L7+ n Xi - A m+n
converges in distribution. According to Corollary 8.2.2, the two limit
distributions must be of the same type, that is, (1) obtains for some b > 0
anda. 0
It follows immediately from Theorem 1 that the stable distributions
form a subclass of the infinitely divisible laws and hence (l) may be used in
conjunction with the representation of i.d. c.f.s to glean further information.
z
l/t(bt) = itby + foo (e i1bX
-00
- I - I itbx Z) I
+x
\X
X
dG(x)
y )b + i (y)b
f
Z
oo ( . it
= itby + _ 00 e"Y - I - I + i Ib z i dG
= it [bY + 0 - b )
Z
f~oo I : yZ dG(~) ]
Z
+ foo (eilY _ 1 _~) b + yZ dG(~)
_ 00 I + yZ yZ b
= itb' + foo
-00
(eilY _ 1_~)(1 + +
+i i i +
1
i) b
Z
),Z
i dG(~).
b
(4)
f x
-00
bi + i dG
---=---'--;;-z
I+y
(1-) + fX
bl -00
b~ + i Z dG
I +y
(1-) = fX
bz -00
Z
b + i Z dG (~) ,
1+y b
-ex bZ + u
r(x+h)= -00 UZ
f (u)
dG b'
Thus, from (5), for all x and arbitrary hI' h z there exists h such that
+ hi) + J(x + h z ) = J(x + h),
J(x
r(x + hi) + r(x + h z ) = r(x + h). (7)
l()
X = leO) ·e -ax =-e
C I -ax
,
IX
+ y2
f
oo I C
--2- dG(y) = l(log x) = ~ X-a, x> 0,
x y IX
or
x 1- a
dG(x) = C I -1--2 dx, x>O CI = al(O), (8)
+X
472 12 Infinitely Divisible Laws
Since G(oo) - G(O+) < 00, necessarily 0 < a < 2 and moreover, from (7)
e-a(x+h d + e- a (x+h = 2) e-a(x+h) or
(9)
Proceeding in similar fashion with r(x) if G(O-) > 0, it follows that
1
00
ljJ(t) = iyt + C (
e,tx
-
- 1 - -itx-2) -
dx-
1 0 + l+x x l +a
+ C2 f O- (
_ 00
itx- ) -
e itx - 1 - - dx-
1 + x 2 IX II + a .
(12)
fO- (e
1
00
itx dX itx dX (13)
+C I (e -l)~+C2 -l)-I-II+a-
0+ X -00 X
By contour integration
0= -
(e lZ
dz
1)- =
f R(e'" -
2 -
1)
dv
-+
fR' (e- U
1)-
du
-
v l +a fu 1 +a
- -
1Q zl+a R, R2
2 _ 11</2 _ R _eilJide
+ I(-I)J (exp(iR j e '8 ) - I) l;a i8(1+a)-
j=l 0 Rj e
Now, since v == lexp(iRe i8 ) - 11:5: 2R for 0 < R < 1 and v:5: e-Rsin8 +
e
1 :5: 2 for 0 :5: :5: n12,
1</2
IJ[
1</2 _
(exp(iRe'8) - I)
ide
:5:
2
10
--81
R I - a de = 0(1) as R --+ 0
[1</2 de
1J
o Rae' a
it follows that
via integration by parts and the recursion formula for the r function. Thus,
if t > 0,
Since
setting
and
an
c = - (c\ + c 2 )· r( -a)cos 2 ~ 0,
[00 . dx
l{J(t) = ity" + C 1 J (e"X - I - itx) Xl+a
o
+ C 2 f o (itX
e . ) -dx-
- I - Itx l a (5)
-00 Ixl + '
= i-aM(ex) say, where 0< M(ex) < 00, and so, setting
1r
C = - M(ex)(c 1 + c 2 )coS"2 ex ~ 0,
and exactly as in case (i) the above also holds for I < 0.
(iii) ex = I. Since
if t > 0,
12.3 Stable Laws 475
r'( .,
Jo e' x -
itx ) dx
I - 1 + X 2 ~2
=
1
o
00
cos tx - I d X
x2
+I .1 00
0
(
sm
•
tx - -tx-2 ) -dx2
I +x x
= - -n t + i lim -
2 <-0 +
[f<'
t sin v dv + t foo(sin v
<
-2-
V <
-2- -
V v(l
I) J
+ v )
2 dv
n
= -"2 t - it log t + itYo ,
noting
3
sin v I = v + O(v ) _ ~ + _v_ = O(v) as v -+ o.
7 - v(1 +v 2
) v2 V I + v2
which coincides with (3) for t > 0 and also for t < O. Clearly, I fJ I ::;; I
from (14).
Conversely, suppose that rp(t) is defined by (3). If IX = 2, then rp(t) =
exp{ iyt - ct 2 /2}, so that rp is a normal d. and hence a stable d. If, rather,
o < IX < 2, let J - (0) and J + (0) be determined by
r(o) - J(O)
(16)
J (0) + J(O) = fJ,
I
x dy
IXr(o) _oolyla-I(I +y2)' x<O
=
G(x)
r
1 IXJ(O) Jo ya
dy
1( 1 + i) + G(O), x>O (18)
476 12 Infinitely Divisible Laws
Then
+ IXr(O)
f o ( e'., x_I - -itx- ) ----
-00
dx
1 + x 2 Ixl l +~
and, from the computations following (12), for some y" E ( - 00,00)
Hence, from (3), (19), and (20), choosing y' =y- y",
qJ(t) = exp{l/!(t)},
I. Show that the mass of a proper stable dJ. is confined to (0, (0) (resp. ( - 00, 0» iff
P = -I (resp. P = I).
2. Prove that if X has a stable distribution with characteristic exponent IX E (0, 2), then
E I X IP < 00 for P < IX and E IX IP = 00 for P > IX. Hint: If X = X I + X 2' where
X l' X 2 are i.i.d., then EIX IP < 00 for P < IX by Exercise 8.4.11. If P > IX, then EIX IP
= 00 by Theorem 8.4.1 (4).
References 477
3. Prove that all proper stable distributions are continuous and infinitely differentiable.
4. A d.r. F(with c.r. cp) is said to be in the domain ofattraction (resp. domain of normal
attraction) of a (stable) distribution G (with d. ljJ) if for suitable constants A., B.
(resp. for some An and B. = bn l /), lim e;'A·cp'(t/Bn) = ljJ(t). Show that every stable
distribution belongs to its own domain of (normal) attraction.
5. In coin tossing, prove that the probability that the mth return to equilibrium (equal
number of heads and tails) occurs before m2 x tends to 2[1 - <I>(X- 1/2)] as m -+ 00. As
usual, <I> is the standard normal d.f.
6. The limit distribution of Exercise 5 has density f(x) = (2nx 3 )-1/2 e -lflx, x> 0.
This is actually the stable density function corresponding to IX = t, fJ = -I, y = 0,
c = l.
7. If Sn = I~ Xi> n ;;:: 1 is a random walk with X I having a symmetric stable distribu-
tion of characteristic exponent IX, show that Sn/n 1/. also has this distribution. Hence, if
I $ r < IX < 2, E Is. I = Cn 1/. for some C in (0, (0) whence the conclusion ofTheorem
10.3.4 fails for I $ r < 2.
References
K. L. Chung, A Course in Probability Theory, Harcourt Brace, New York, 1968; 2nded.,
Academic Press, New York, 1974.
H. Cramer, "Su un teorema relativo alia legge uniforme dei grande numeri," Giornale
del/' lstituto degli Attuari 5 (1934), 1-13.
e. Esseen, "Fourier analysis of distribution functions," Acta Math 77 (1945), 79.
B. V. Gnedenko and A. N. Kolmogorov, Limit Distributions for Sums of Independent
Random Variables, Addison-Wesley, Reading, Mass., 1954.
P. Levy, Theorie de faddition des variables aleatoires, Garthier-Villars, Paris, 1937;
2nd ed., 1954.
M. Loeve, Probability Theory, 3rd ed., Van Nostrand, Princeton, 1963; 4th ed., Springer-
Verlag, Berlin and New York, 1977-1978.
Index
479
480 Index
Event, 19 elementary, 85
Expectation (mean), 84 finite set, 18
existence of, 84 integrable, 84, 89, 172
Exponential joint distribution, 26, 187
distribution, 60 left continuous, 25, 26
random variable, 60, 65 moment generating, 110
Extension monotone set, 19
of a sequence, 90 probability density, 27
of a set function, 20, 165 set, 18
(i-additive (countably additive)
Factor closed, 300 set, 19
Fatou lemma, 95, 174 (i-finite set, 18
for conditional expectations, 216 simple, 18
extended, 218 subadditive set, 19
Feller, 53, 61, 70, 83, 126, 128, 137, subtractive set, 24
164,311,313,314,353,402 tail,64
-Chung lemma, 70
weak law of large numbers, 128 Gambler's ruin problem, 81, 250
Fictitious random variable, 272 Gamma distribution, 294
Finite Garsia, 425, 442
measure space, 23 Geometric distribution, 60, 91
partition, 18 Glivenko, 284, 311
permutation, 232 -Cantelli theorem, 284, 311
real line, 11 Gnedenko, 312,402,477
set, 2 Gundy, 157, 164,422,425,442
set function, 18
stopping time (rule, variable), 138, Haag, 33, 53
240 Hajek, 255, 268
Finitely additive set function, 18 -Renyi inequality, 255
First passage time, 141 Halmos,29, 111, 177,209,222,268
Frechet, 283, 311 Hanson, 353
-Shohat theorem, 283 Hardy, 53, 103, 111,209,297,312,
Freedman, 254,268,442 368,372
Friedman, 325 Harris, 258
Fubini theorem, 186, 191,213 Hartman, 374, 378,402
Fuchs, 98, 111, 154, 163 - Wintner law of the iterated
Function logarithm, 382
d-measurable,15 Hausdorff, 29, 368, 372
absolutely continuous, 27 Helly, 274, 312
additive, 18 -Bray lemma, 274
Borel (measurable), 14 -Bray theorem, 276
convex, 102-104 Hewitt, 238, 268
density, 27, 31,46,206 Heyde, 137,164, 346, 353
discrete distribution, 27 Hoeffding, 268
distribution, 26 decomposition, 261
Index 483
Teicher, 53, 112, 137, 163, 164, 268, Wald, 158, 162, 164
269,312,353,403,443 equation, 143, 144,253, 254,415
Three series theorem, 117, 123 Weak
Tight, 276 compactness (sequential), 282
Titchmarsh, 302, 303, 306, 312 law of large numbers, 124, 128,
Total variation, 208 235,356-359
Triangular distribution, 294 Weierstrass, 121
Truncation, 106, 110, 182,213 approximation theorem, 42
Weiss, I., 353
Vncorrelated, 106 Widder, 178, 209
Vniform distribution, 39, 293 Wiener dominated ergodic theorem,
Uniformly 387,403
absolutely continuous, 208 Wintner, 374, 378, 402
bounded random variables, 116 Wolfowitz, 347,402
integrable (u.i.) random variables,
93,94 Yadrenko, 149
integrable relative to distribution Young, L. C, 112
functions, 276, 277 Young, W. H., inequality, 184,424
V-statistics, 241, 259-268
Central limit theorem, 326 Zero-one law, 97
degenerate, 259 Hewitt-Savage, 238
decomposition of, 261 Kolmogorov,64
Strong law of large numbers, 263 for interchangeable random
Law of the iterated logarithm, 382 variables, 238
see also Borel zero-one criterion
Van Beek, 323, 353 Zolotarev, 353
Variance, 105 Zygmund, 118, 121, 125, 164,386,
Von Mises, 58, 83 387,402,422,443
Springer Texts in Statistics (continued from page ii)