Documente Academic
Documente Profesional
Documente Cultură
Appendix A
Version 2.3
© Dr. Michael P. McLaughlin
1993-2001
Third printing
This software and its documentation are distributed free of charge and may neither be sold
nor repackaged for sale in whole or in part.
A-2
PREFACE
All distributions are shown in their parameterized, not standard forms. In some cases, the
definition of a distribution may vary slightly from a definition given in the literature. This
happens either because there is more than one definition or, in the case of parameters,
because Regress+ requires a parameter to be constrained, usually to guarantee convergence.
There are a few well-known distributions that are not included here, either because they
are seldom used to model empirical data or they lack a convenient analytical form for the
CDF. Conversely, many of the distributions that are included are rarely discussed yet are
very useful for describing real-world datasets.
mpmcl@mitre.org
Michael P. McLaughlin
McLean, VA
September, 1999
A-3
A-4
Table of Contents
Distributions shown in bold are those which appear in the Regress+ menu. Distributions
shown in plain text are aliases and/or special cases.
Continuous Distributions
Antilognormal 77 HalfNormal(A,B) 51
Bell curve 85 HyperbolicSecant(A,B) 59
Beta(A,B,C,D) 9 Inverse Gaussian 61
Bilateral exponential 63 InverseNormal(A,B) 61
Bradford(A,B,C) 15 Laplace(A,B) 63
Burr(A,B,C,D) 17 Logistic(A,B) 75
Cauchy(A,B) 19 LogLogistic 37
Chi(A,B,C) 21 LogNormal(A,B) 77
Chi-square 41 LogWeibull 49
Cobb-Douglas 77 Lorentz 19
Cosine(A,B) 23 Maxwell 21
Double-exponential 63 Negative exponential 29
DoubleGamma(A,B,C) 25 Nakagami(A,B,C) 79
DoubleWeibull(A,B,C) 27 Non-central Chi 39
Erlang 41 Normal(A,B) 85
Error function 85 Pareto(A,B) 99
Exponential(A,B) 29 Power-function 9
Extreme-value 119 Rayleigh 21
ExtremeLB(A,B,C) 35 Reciprocal(A,B) 105
Fisher-Tippett 49 Rectangular 113
Fisk(A,B,C) 37 Sech-squared 75
FoldedNormal(A,B) 39 Semicircular(A,B) 107
Frechet 119 StudentsT(A,B,C) 109
Gamma(A,B,C) 41 Triangular(A,B,C) 111
Gaussian 85 Uniform(A,B) 113
GenLogistic(A,B,C) 43 Wald 61
Gompertz 49 Weibull(A,B,C) 119
Gumbel(A,B) 49
A-5
Continuous Mixtures
Name Page
Double double-exponential 65
Expo(A,B)&Expo(A,C) 31
Expo(A,B)&Uniform(A,C) 33
HNormal(A,B)&Expo(A,C) 53
HNormal(A,B)&HNormal(A,C) 55
HNormal(A,B)&Uniform(A,C) 57
Laplace(A,B)&Laplace(C,D) 65
Laplace(A,B)&Laplace(A,C) 67
Laplace(A,B)&Uniform(C,D) 69
Laplace(A,B)&Uniform(A,C) 71
Normal(A,B)&Laplace(C,D) 87
Normal(A,B)&Laplace(A,C) 89
Normal(A,B)&Normal(C,D) 91
Normal(A,B)&Normal(A,C) 93
Normal(A,B)&Uniform(C,D) 95
Normal(A,B)&Uniform(A,C) 97
Uniform(A,B)&Uniform(C,D) 115
Uniform(A,B)&Uniform(A,C) 117
Schuhl 31
Discrete Distributions
Name Page
Binomial(A,B) 11
Furry 45
Geometric(A) 45
Logarithmic(A) 73
NegativeBinomial(A,B) 81
Pascal 81
Poisson(A) 101
Polya 81
Discrete Mixtures
Name Page
Binomial(A,C)&Binomial(B,C) 13
Geometric(A)&Geometric(B) 47
NegBinomial(A,C)&NegBinomial(B,C) 83
Poisson(A)&Poisson(B) 103
A-6
Description of Included Items
Each distribution is illustrated with at least one example. In this figure, the parameters
used are shown in parentheses, in the order listed in the header. Expressions are then given
for the PDF and CDF. Remaining subsections, as appropriate, are as follows:
Parameters
This is an interpretation of the meaning of each parameter, with the usual literature
symbol (if any) given in parentheses.
Moments, etc.
Provided that there are closed forms, the mean, variance, skewness, kurtosis, mode,
median, first quartile (Q1), and third quartile (Q3) are described along with the
quantiles for the mean (qMean) and mode (qMode). If random variates are
computable with a closed-form expression, the latter is also given.
Note that the mean and variance have their usual units while the skewness and
kurtosis are dimensionless. Furthermore, the kurtosis is referenced to that of a
standard Normal distribution (kurtosis = 3).
Notes
These include any relevant constraints, cautions, etc.
Characterizations
This list is far from exhaustive. It is intended simply to convey a few of the more
important situations in which the distribution is particularly relevant.
A-7
Legend
Shown below are definitions for some of the less common functions and abbreviations
used in this Appendix.
N
k number of combinations of N things taken k at a time
γ EulerGamma = 0.57721566...
~ (is) distributed as
A-8
Beta(A,B,C,D) A < y < B, C, D > 0
3.0
H0, 1, 6, 2L
2.5
2.0 H0, 1, 1, 2L
H0, 1, 2, 2L
PDF
1.5
1.0
0.5
0.0
0.0 0.2 0.4 0.6 0.8 1.0
Y
y–A
CDF = I , C, D
B–A
Moments, etc.
Mean = A D + B C
C+D
2
CD B–A
Variance = 2
C+D+1 C+D
2CD D–C
Skewness = 3
2
C+D
3
C+D+1 C+D+2 CD
2
C+D C+D+1
A-9
C2 D + 2 + 2 D2 + C D D – 2 C+D+1
Kurtosis = 3 –1
CD C+D+2 C+D+3
A D–1 +B C–1
Mode = , unless C = D = 1
C+D–2
Notes
1. Although C and D have no upper bound, in fact, they seldom exceed 10. If optimum
values are much greater than this, the response will often be nearly flat.
2. If both C and D are large, the distribution is roughly symmetrical and some other model
is indicated.
3. The beta distribution is often used to mimic other distributions. When suitably
transformed and normalized, a vector of random variables can almost always be modeled
as Beta.
Characterizations
A-10
Binomial(A,B) y = 0, 1, 2, …, 0 < A < 1, y, 2 ≤ B
0.25
0.15
PDF
0.10
0.05
0.00
0 2 4 6 8 10
Y
B B–y
PDF = Ay 1 – A
y
Moments, etc.
Mean = A B
Variance = A 1 – A B
Mode = int A B + 1
A-11
Notes
Characterizations
A-12
Binomial(A,C)&Binomial(B,C)
y = 0, 1, 2, …, 0 < B < A < 1, y, 3 ≤ C, 0 < p < 1
0.25
0.20
0.10
0.05
0.00
0 2 4 6 8 10
Y
C C–y C–y
PDF = p Ay 1 – A + 1 – p By 1 – B
y
Moments, etc.
Mean = C p A + 1 – p B
2
Variance = C p A 1 – A + 1 – p p C A – B + B 1 –B
A-13
Notes
Characterizations
A-14
Bradford(A,B,C) A < y < B, C > 0
3.0
H0, 1, 5L
2.0
PDF
1.0 H0, 1, 1L
0.0
0.0 0.5 1.0
Y
PDF = C
C y – A + B – A log C + 1
C y–A
log 1 +
B–A
CDF =
log C + 1
C B–A +k A C+1 –B
Mean =
Ck
2
B–A C k–2 +2k
Variance =
2 C k2
A-15
2 12 C 2 – 9 k C C + 2 + 2 k 2 C C + 3 + 3
Skewness =
C C k–2 +2k 3C k–2 +6k
C 3 k – 3 k 3 k – 16 + 24 + 12 k C 2 k – 4 k – 3 + 6 C k 2 3 k – 14 + 12 k 3
Kurtosis = 2
3C C k–2 +2k
Mode = A
Median = 1 A C + 1 – B + B – A C+1
C
Q1 = 1 A C + 1 – B + B – A Q3 = 1 A C + 1 – B + B – A
4 4 3
C+1 C+1
C C
log C
log C + 1
qMean = qMode = 0
log C + 1
RandVar = 1 A C + 1 – B + B – A C + 1
u
C
Notes
Characterizations
1. The Bradford distribution has been used to model the distribution of references among
several sources.
A-16
Burr(A,B,C,D) y > A, B > 0, 0 < C, D ≤ 100
0.80
H0, 1, 2, 1L
0.60
PDF
H0, 2, 3, 2L
0.40
0.20
0.00
0 1 2 3 4 5 6 7 8
Y
–D–1
–C–1 –C
y–A y–A
PDF = C D 1+
B B B
–D
–C
y–A
CDF = 1 +
B
2 2 1 2 1
Moments, etc. k ≡ Γ D Γ 1 – C Γ C + D – Γ 1 – C Γ C + D
2
BΓ 1– 1 Γ 1 +D
C C
Mean = A +
Γ D
2
Variance = k2 B
Γ D
A-17
Γ3 D 2 Γ3 1 – 1 Γ3 1 + D 3Γ 1– 2 Γ 1– 1 Γ 1 +D Γ 2 +D Γ 1– 3 Γ 3 +D
C C C C C C C C
Skewness = – +
k3 Γ3 D Γ2 D Γ D
Kurtosis =
Γ4 D – 3 Γ4 1 – 1 Γ4 1 + D 6 Γ 1 – 2 Γ2 1 – 1 Γ2 1 + D Γ 2 + D
C C C C C C
–3+ + –
k2 Γ D
4
Γ D
3
Γ4 D 4Γ 1– 3 Γ 1– 1 Γ 1 +D Γ 3 +D Γ 1– 4 Γ 4 +D
C C C C C C
–
k2 Γ D
2
Γ D
Mode = A + B
C C D – 1 , iff C D > 1, else Mode = A (and qMode = 0)
C+1
D – 1
C
Median = A + B 2 –1
– 1
D – 1 D 4 –1 C
C
Q1 = A + B 4 –1 Q3 = A + B
3
–D
–D
qMean = 1 + Γ C
D Γ –C
1– 1 Γ –C 1 +D qMode = 1 + C + 1
C C CD–1
– 1
RandVar = A + B D1 – 1
C
Notes
1. The Burr distribution, with D = 1, is often called the Fisk or LogLogistic distribution.
Characterizations
A-18
Cauchy(A,B) B>0
0.6
0.5
0.4
PDF
0.3
H3, 0.6L
0.2
H0, 1L
0.1
0.0
-8 -6 -4 -2 0 2 4 6 8
Y
PDF = 1
2
y–A
πB 1+
B
1 tan – 1 y – A
CDF = 1 + π
2 B
Moments, etc.
This distribution has no finite moments because the corresponding integrals do not
converge.
Median = Mode = A
Q1 = A – B Q3 = A + B
qMode = 0.5
A-19
RandVar = A + B tan π u – 1
2
Notes
1. Since there are no finite moments, the location parameter (ostensibly the mean) does not
have its usual interpretation for a symmetrical distribution.
Characterizations
A-20
Chi(A,B,C) y > A, B > 0, 0 < C ≤ 100
H0, 1, 1L
0.8
H0, 1, 2L
0.6
PDF
0.4
H0, 1, 3L
0.2
0.0
0 1 2 3 4
Y
C–1 2
y–A y–A
exp – 1
B 2 B
PDF =
22 –1 B Γ C
C
2
y–A
CDF = Γ C , 1
2 2 B
Moments, etc.
2 BΓ C+1
2
Mean = A +
Γ C
2
2 Γ2 C + 1
2 2
Variance = B C –
Γ2 C
2
A-21
2 4 Γ3 C + 1 + Γ2 C 2Γ C+3 –3CΓ C+1
2 2 2 2
Skewness = 3
2
2Γ C+1
2
2
Γ3 C C–
2 2 C
Γ
2
2 C 1 – C Γ 4 C – 24 Γ 4 C + 1 + 8 2 C – 1 Γ 2 C Γ 2 C + 1
2 2 2 2
Kurtosis = 2
CΓ 2 C – 2 Γ2 C + 1
2 2
Mode = A + B C – 1
Notes
1. In the literature, C > 0. The restrictions shown above are required for convergence when
the data are left-skewed and to ensure the existence of a Mode.
Characterizations
A-22
Cosine(A,B) A – π B ≤ y ≤ A + π B, B > 0
0.4
0.3
H0, 1L
PDF
0.2
0.1
0.0
-3 -2 -1 0 1 2 3
Y
PDF = 1 1 + cos y – A
2πB B
y–A y–A
CDF = 1 π + + sin
2π B B
Moments, etc.
Mean = Median = Mode = A
Variance = π – 2 B 2
2
Skewness = 0
– 6 π 4 – 90
Kurtosis = 2 ≈ – 0.5938
5 π2 – 6
Q1 ≈ A – 0.8317 B Q3 ≈ A + 0.8317 B
A-23
qMean = qMode = 0.5
Notes
Characterizations
A-24
DoubleGamma(A,B,C) B, C > 0
H0, 1, 3L
0.14
0.12
0.10
0.08
PDF
0.06
0.04
0.02
0.00
-10 -5 0 5 10
Y
C–1
1 y–A y–A
PDF = exp –
2BΓ C B B
1 – 1 Γ C, y – A , y≤A
2 2 B
CDF =
1 + 1 Γ C, y – A , y>A
2 2 B
Moments, etc.
Mean = Median = A
Variance = C C + 1 B 2
Skewness = 0
A-25
Q1, Q3: no simple closed form
qMean = 0.5
Notes
Characterizations
A-26
DoubleWeibull(A,B,C) B, C > 0
0.6
H0, 1, 3L
0.5
0.4
PDF
0.3
0.2
0.1
0.0
-2 -1 0 1 2
Y
C–1 C
y–A y–A
PDF = C exp –
2B B B
C
1 exp – y – A , y≤A
2 B
CDF =
C
y–A
1 – 1 exp – , y>A
2 B
Moments, etc.
Mean = Median = A
Variance = Γ C + 2 B 2
C
Skewness = 0
A-27
Q1, Q3: no simple closed form
qMean = 0.5
Notes
Characterizations
A-28
Exponential(A,B) y ≥ A, B > 0
1.0
0.8 H0, 1L
0.6
PDF
0.4
0.2 H0, 2L
0.0
0 2 4 6 8
Y
A–y
PDF = 1 exp
B B
A–y
CDF = 1 – exp
B
Moments, etc.
Mean = A + B
Variance = B 2
Skewness = 2
Kurtosis = 6
Mode = A
Median = A + B log 2
A-29
Q1 = A + B log 4 Q3 = A + B log 4
3
RandVar = A – B log u
Notes
Characterizations
1. If the future lifetime of a system at any time, t, has the same distribution for all t, then this
distribution is the Exponential distribution. This is known as the memoryless property.
A-30
Expo(A,B)&Expo(A,C) y ≥ A, B, C > 0, 0 < p < 1
1.0
0.8
0.4
0.2
0.0
0 1 2 3 4 5 6
Y
y–A y–A
CDF = p stdExponentialCDF + 1 – p stdExponentialCDF
B C
Moments, etc.
Mean = A + p B + 1 – p C
2
Variance = C 2 + 2 B B – C p – B – C p 2
Mode = A
RandVar: determined by p
A-31
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
1. This mixture, when applied to traffic analysis, is often called the Schuhl distribution.
Characterizations
A-32
Expo(A,B)&Uniform(A,C) A ≤ y < C, B > 0, 0 < p < 1
1.0
0.8
0.4
0.2
0.0
0 1 2 3 4 5 6
Y
y–A y–A
p stdExponentialCDF + 1–p , y<C
B C–A
CDF =
y–A
p stdExponentialCDF + 1–p , y≥C
B
Moments, etc.
A + C + p (A + 2 B – C)
Mean =
2
2
2
p–1 A +AC+C 2 A+C+p A+2B–C
2
Variance = p B 2 + A + B – –
3 4
Mode = A
A-33
Quantiles, etc.: no simple closed form
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-34
ExtremeLB(A,B,C) y > A, B > 0, 0 < C ≤ 100
0.9
0.8 H1, 1, 2L
0.7
0.6
0.5
PDF
0.4
0.3
H1, 2, 3L
0.2
0.1
0.0
1 2 3 4 5 6
Y
–C–1 –C
y–A y–A
PDF = C exp –
B B B
–C
y–A
CDF = exp –
B
Variance = B 2 Γ C – 2 – Γ 2 C – 1
C C
Γ C – 3 – 3 Γ C – 2 Γ C – 1 + 2 Γ3 C – 1
C C C C
Skewness = 3
Γ C – 2 – Γ2 C – 1
C C
A-35
Γ C – 4 – 4 Γ C – 3 Γ C – 1 + 3 Γ2 C – 2
C C C C
Kurtosis = – 6 + 2
Γ C – 2 – Γ2 C – 1
C C
Mode = A + B
C C
1+C
Median = A + B
C
log 2
Q1 = A + B Q3 = A + B
C
log 4 log 4
C
– 1
C
RandVar = A + B – log u
Notes
1. The name ExtremeLB does not appear in the literature. It was chosen here simply to
indicate one type of extreme-value distribution with a lower bound.
2. In the literature, C > 0. The restriction shown above is required for convergence when
the data are left-skewed.
3. Moment k exists if C > k.
Characterizations
A-36
Fisk(A,B,C) y > A, B > 0, 0 < C ≤ 100
0.7
0.6 H0, 1, 2L
0.5
0.4
PDF
0.3
0.2 H0, 2, 3L
0.1
0.0
0 1 2 3 4 5 6 7 8
Y
C–1
y–A
B
PDF = C
B C
2
y–A
1+
B
CDF = 1
–C
y–A
1+
B
C C C C
A-37
2 π 2 csc 3 π – 6 C π csc π csc 2 π + 3 C 2 csc 3 π
C C C C
Skewness = 3
2
π 2 C csc 2 π – π csc 2 π
C C
Kurtosis =
– 3 π 3 csc 4 π – 12 C 2 π csc π csc 3 π + 4 C 3 csc 4 π + 6 C π 2 csc 3 π sec π
C C C C C C
2 –3
π π csc 2 π – 2 C csc 2 π
C C
Mode = A + B
C C–1
C+1
Median = A + B
Q1 = A + CB Q3 = A + B C 3
3
qMean = 1 qMode = C – 1
π csc π
2C
–C
1+
C C
RandVar = A + B C u
1–u
Notes
Characterizations
A-38
FoldedNormal(A,B) y ≥ 0, A ≥ 0, B > 0
H1, 1L
0.5
0.4
0.3
PDF
0.2
0.1 H1, 2L
0.0
0 1 2 3 4 5
Y
2 2
PDF = 1 2 cosh A y exp – 1 y + A
B π B2 2 B2
y–A –y–A
CDF = Φ –Φ
B B
Parameters -- A (µ): Location, B (σ): Scale, both for the corresponding unfolded Normal
Moments, etc.
2
Mean = B 2 1 A –A 1–2Φ A
π exp – 2 B B
A-39
B 2 3 A2 2
4 B 2 – π exp A 2 2 A2 + B2
π exp – 2 B 2 B
Skewness = +
π Var 3
2 2
2 A erf A 6 B 2 exp – A 2 + 3 2 π A B exp – A 2 erf A + π A 2 erf 2 A –1
2B B 2B 2B 2B
π Var 3
Kurtosis = – 3 +
2
A4 + 6 A2 B2 + 3 B4 + 6 A2 2 A2 A
+ B2 B π exp – 2 B 2 + A erf 2B
–
Var 2
4
3 B 2 A2 A
Var 2 π exp – 2 B 2 + A erf 2B
–
2
4 exp – A 2 2 A2 A 2 A 2 + 2 B 2 + A A 2 + 3 B 2 exp A 2 erf A
B
B π + A exp 2 B 2 erf 2B
B π 2 B2 2B
Var 2
Notes
Characterizations
A-40
Gamma(A,B,C) y > A, B > 0, 0 < C ≤ 100
1.0
0.8
H0, 1, 1L
0.6
PDF
H0, 1, 2L
0.4
0.2
0.0
0 1 2 3 4 5 6
Y
C–1
1 y–A A–y
PDF = exp
BΓ C B B
y–A
CDF = Γ C,
B
Moments, etc.
Mean = A + B C
Variance = B 2 C
Skewness = 2
C
Kurtosis = 6
C
Mode = A + B C – 1
Notes
Characterizations
A-42
GenLogistic(A,B,C) B, C > 0
0.4
0.3
0.2
0.1
0.0
-4 -2 0 2 4 6 8
Y
A–y
exp
B
PDF = C
B C+1
A–y
1 + exp
B
CDF = 1
C
A–y
1 + exp
B
Moments, etc.
Mean = A + γ + ψ C B
Variance = π + ψ′ C B 2
2
ψ′′ C + 2 ζ 3
Skewness = 3
π 2 + ψ′ C 2
6
A-43
12 π 4 + 15 ψ′′′ C
Kurtosis = 2
5 π 2 + 6 ψ′ C
Mode = A + B log C
C
Median = A – B log 2 –1
Q1 = A – B log C
4 –1 Q3 = A – B log
C 4 –1
3
–C
C
qMean = 1 + exp – H C – 1 qMode = C
C+1
RandVar = A – B log 1 –1
C
u
Notes
Characterizations
1. The Generalized Logistic has been used in the analysis of extreme values.
A-44
Geometric(A) y = 1, 2, 3, …, 0 < A < 1
H0.45L
0.5
0.4
0.3
PDF
0.2
0.1
0.0
0 2 4 6 8 10
Y
y–1
PDF = A 1 – A
Moments, etc.
Mean = 1
A
Variance = 1 – 2A
A
Mode = 1
A-45
Notes
Characterizations
A-46
Geometric(A)&Geometric(B) y = 1, 2, 3, …, 0 < B < A < 1, 0 < p < 1
0.20
0.10
0.05
0.00
0 2 4 6 8 10 12 14 16 18 20
Y
y–1 y–1
PDF = p A 1 – A + 1–p B 1–B
Moments, etc.
p 1–p
Mean = +
A B
2
A2 1 – B + p B A – B A – 2 – p2 A – B
Variance =
A2 B2
Mode = 1
A-47
Notes
Characterizations
A-48
Gumbel(A,B) B>0
0.40
0.35
0.30 H0, 1L
0.25
PDF
0.20
0.15
0.10
H1, 2L
0.05
0.00
-4 -2 0 2 4 6 8 10
Y
A–y A–y
PDF = 1 exp exp – exp
B B B
A–y
CDF = exp – exp
B
Moments, etc.
Mean = A + γ B
Variance = 1 π B
2
6
12 6 ζ 3
Skewness = ≈ 1.1395
π3
Kurtosis = 12
5
Mode = A
A-49
Median = A – B log log 2
Notes
Characterizations
1. Extreme-value distributions are the limiting distributions, as N --> infinity, of the greatest
value among N i.i.d. variates selected from a continuous distribution. By replacing y
with –y, the smallest values may be modeled.
2. The Gumbel distribution is often used to model maxima when the random variable is
unbounded.
A-50
HalfNormal(A,B) y ≥ A, B > 0
0.8
H0, 1L
0.7
0.6
0.5
PDF
0.4
0.3
0.2 H0, 2L
0.1
0.0
0 1 2 3 4 5 6
Y
2
PDF = 1 2 1 y–A
B π exp – 2 B
y–A
CDF = 2 Φ –1
B
Moments, etc.
Mean = A + B 2
π
2
Variance = B 2 1 – π
2 4–π
Skewness = 3
≈ 0.9953
π–2
8 π–3
Kurtosis = 2
≈ 0.8692
π–2
A-51
Mode = A
Median ≈ A + 0.6745 B
Q1 ≈ A + 0.3186 B Q3 ≈ A + 1.150 B
Notes
1. The HalfNormal distribution is a special case of both the Chi and the FoldedNormal
distributions.
Characterizations
1. If X ~Normal(A, B) is folded (to the right) about its mean, A, the resulting distribution is
HalfNormal(A, B).
A-52
HNormal(A,B)&Expo(A,C) y ≥ A, B, C > 0, 0 < p < 1
0.8
0.7
0.5
PDF
0.4
0.3
0.2
0.1
0.0
0 1 2 3 4 5 6
Y
2
2 p 1 y–A 1–p A–y
PDF = π B exp – 2 + exp
B C C
y–A y–A
CDF = p stdHalfNormalCDF + 1 – p stdExponentialCDF
B C
Moments, etc.
Mean = A + p B 2
π + 1–p C
1 p π – 2 p B2 + 2 p p – 1 B C 2 π – p2 – 1 π C2
Variance = π
Mode = A
RandVar: determined by p
A-53
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. The alternate, Expo(A,B)&HNormal(A,C) distribution may be obtained by switching
identities in the parameter dialog. In this case, the parameters shown above in the
Moments section must be reversed (cf. E&E and H&H).
Characterizations
A-54
y ≥ A, B, C > 0, 0 < p < 1
HNormal(A,B)&HNormal(A,C)
0.7
0.5
0.4
PDF
0.3
0.2
0.1
0.0
0 1 2 3 4 5
Y
2 2
2 p 1 y–A 1–p y–A
PDF = π B exp – 2 + exp – 1
B C 2 C
y–A y–A
CDF = p stdHalfNormalCDF + 1 – p stdHalfNormalCDF
B C
Moments, etc.
Mean = A + 2
π pB+ 1–p C
2
2 pB+ 1–p C
Variance = p B 2 + 1 – p C 2 – π
Mode = A
RandVar: determined by p
A-55
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-56
HNormal(A,B)&Uniform(A,C) A ≤ y < C, B > 0, 0 < p < 1
0.8
0.7
0.5
PDF
0.4
0.3
0.2
0.1
0.0
0 1 2 3 4
Y
2
2 p 1 y–A 1–p y<C
PDF = π B exp – 2 +
B C–A
y–A y–A
p stdHalfNormalCDF + 1–p , y<C
B C–A
CDF =
y–A
p stdHalfNormalCDF + 1–p , y≥C
B
Moments, etc.
Mean = 1 A + C + p A – C + 2 B 2
π
2
2
12 p B 2 π – 2 p + 12 p B A – C 1 – p 2π +π A–C 1–p 1+3p
Variance =
12 π
Mode = A
A-57
Quantiles, etc.: no simple closed form
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-58
HyperbolicSecant(A,B) B>0
0.6
0.5
0.4
PDF
0.3
H3, 0.6L
0.2
H0, 1L
0.1
0.0
-8 -6 -4 -2 0 2 4 6 8
Y
y–A
sech
B
PDF =
πB
2 tan – 1 exp y – A
CDF = π
B
Moments, etc.
Mean = Median = Mode = A
Variance = 1 π B
2
4
Skewness = 0
Kurtosis = 2
Q1 = A – B log 1 + 2 Q3 = A + B log 1 + 2
Notes
Characterizations
A-60
InverseNormal(A,B) y > 0, A, B > 0
1.2
1.0 H1, 1L
0.8
PDF
0.6
0.4
0.2 H2, 3L
0.0
0 1 2 3 4 5 6
Y
2
PDF = B exp – B y – A
2 π y3 2y A
Moments, etc.
Mean = A
3
Variance = A
B
Skewness = 3 A
B
Kurtosis = 15 A
B
Mode = A 9 A2 + 4 B2 – 3 A
2B
A-61
Median, Q1, Q3, qMean, qMode: no simple closed form
Notes
1. There are several alternate forms for the PDF, some of which have more than two
parameters.
Characterizations
A-62
Laplace(A,B) B>0
1.0
0.8
0.6
H3, 0.6L
PDF
0.4
H0, 1L
0.2
0.0
-6 -4 -2 0 2 4 6
Y
PDF = 1 exp – y – A
2B B
1 exp y – A , y≤A
2 B
CDF =
A–y
1 – 1 exp , y>A
2 B
Moments, etc.
Mean = Median = Mode = A
Variance = 2 B 2
Skewness = 0
Kurtosis = 3
Q1 = A – B log 2 Q3 = A + B log 2
A-63
qMean = qMode = 0.5
Notes
Characterizations
A-64
Laplace(A,B)&Laplace(C,D) B, D > 0, 0 < p < 1
0.4
0.3
0.2
0.1
0.0
-4 -2 0 2 4 6
Y
y–A y–C
CDF = p stdLaplaceCDF + 1 – p stdLaplaceCDF
B D
Moments, etc.
Mean = p A + 1 – p C
2
Variance = p 2 B 2 – p – 1 A – C – 2 p – 1 D2
RandVar: determined by p
A-65
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-66
Laplace(A,B)&Laplace(A,C) B, C > 0, 0 < p < 1
0.5
0.4
H0, 1, 2, 0.7L
0.3
PDF
0.2
0.1
0.0
-6 -4 -2 0 2 4 6
Y
y–A y–A
CDF = p stdLaplaceCDF + 1 – p stdLaplaceCDF
B C
Moments, etc.
Mean = Median = Mode = A
Variance = 2 p B 2 + 1 – p C 2
Skewness = 0
6 p B4 + 1 – p C4
Kurtosis = 2 –3
p B2 + 1 – p C2
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-68
Laplace(A,B)&Uniform(C,D) B > 0, C < D, 0 < p < 1
0.5
0.4
0.2
0.1
0.0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Y
y–A
p stdLaplaceCDF , y≤C
B
y–A y–C
CDF = p stdLaplaceCDF + 1–p , C<y<D
B D–C
y–A
p stdLaplaceCDF + 1–p , y≥D
B
Moments, etc.
1–p C+D
Mean = p A +
2
2
1–p C+D
Variance = p A 2 + 2 B 2 – p A + + 1 1 – p C2 + C D + D2
2 3
A-69
Quantiles, etc.: no simple closed form
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-70
Laplace(A,B)&Uniform(A,C) B, C > 0, 0 < p < 1
0.6
0.5
H0, 1, 1, 0.7L
0.4
PDF
0.3
0.2
0.1
0.0
-4 -3 -2 -1 0 1 2 3 4
Y
y–A
p stdLaplaceCDF , y≤A–C
B
y–A y–A+C
CDF = p stdLaplaceCDF + 1–p , A–C<y<A+C
B 2C
y–A
p stdLaplaceCDF + 1–p , y≥A+C
B
Moments, etc.
Variance = 1 6 p B 2 + 1 – p C 2
3
A-71
Skewness = 0
9 120 p B 4 + 1 – p C 4
Kurtosis = 2 –3
2 2
5 6pB + 1–p C
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. Note that the parameters for the Uniform component are different from those in the
Uniform(A,B) distribution. Here, A is the Mean and C is the half-width.
Characterizations
A-72
Logarithmic(A) y = 1, 2, 3, …, 0 < A < 1
H0.45L
0.8
0.7
0.6
0.5
PDF
0.4
0.3
0.2
0.1
0.0
0 1 2 3 4 5 6 7 8
Y
PDF = – Ay
log 1 – A y
Moments, etc.
Mean = –A
1 – A log 1 – A
– A A + log 1 – A
Variance = 2
1 – A log 2 1 – A
Mode = 1
A-73
Notes
Characterizations
1. The Logarithmic distribution has been used to model the numbers of items of a product
purchased by a buyer in a given period of time.
A-74
Logistic(A,B) B>0
0.5
0.4
0.3
H3, 0.6L
PDF
0.2
H0, 1L
0.1
0.0
-8 -6 -4 -2 0 2 4 6 8
Y
y–A
exp
B
PDF = 1
B 2
y–A
1 + exp
B
CDF = 1
A–y
1 + exp
B
Moments, etc.
Mean = Median = Mode = A
Variance = 1 π B
2
3
Skewness = 0
Kurtosis = 6
5
A-75
Q1 = A – B log 3 Q3 = A + B log 3
RandVar = A + B log u
1–u
Notes
Characterizations
dP = α P – β P 2
dt
whence
P0 P∞
Pt =
P0 + P∞ – P0 exp – α t
where P0 is the initial and P∞ = αβ the final population size. This sigmoidal function,
P(t), reduces to exponential growth when α >> β. After appropriate normalization, it
becomes the CDF given above.
2. If lo and hi are the minima and maxima of a random sample (size = N) then,
as N --> infinity, the asymptotic distribution of the midrange = (hi – lo)/2 is ~Logistic.
A-76
LogNormal(A,B) y > 0, B > 0
1.0
0.8 H0.5, 2L
0.6
PDF
0.4
H0, 1L
0.2
0.0
0 1 2 3 4 5 6
Y
1 log y – A
PDF = exp – 1
By 2π 2 B
log y – A
CDF = Φ
B
Moments, etc.
2
Mean = exp A + B
2
Mode = exp A – B 2
A-77
Median = exp A
Notes
Characterizations
1. As the PDF suggests, the LogNormal distribution is the distribution of a random variable
which, in log space, is ~Normal.
A-78
Nakagami(A,B,C) y > A, B > 0, 0 < C ≤ 100
1.6
H0, 1, 1L
1.4
1.2
1.0
PDF
0.8
0.6 H0, 2, 3L
0.4
0.2
0.0
0 1 2 3 4
Y
2C–1 2
2 CC y–A y–A
PDF = exp – C
BΓ C B B
2
y–A
CDF = Γ C, C
B
Moments, etc.
Γ C+ 1
2
Mean = A + B
CΓ C
Γ2 C + 1
2
Variance = 1 – B2
CΓ C
2
A-79
2 Γ3 C + 1 + Γ2 C Γ C + 3 – 3 C Γ C + 1
2 2 2
Skewness = 3
2
C Γ2 C + 1
2
Γ C C 1–
3
Γ C+1
2
– 6 Γ4 C + 1 – 3 C2 Γ4 C + Γ3 C Γ C + 2 + 23 – 4 C 4 C – 1 π Γ2 2 C
2
Kurtosis = 2
Γ C + 1 – C Γ2 C
2
2
Mode = A + B 2C–1
2C
Notes
1. In the literature, C > 0. The restrictions shown above are required for convergence when
the data are left-skewed and to ensure the existence of a Mode.
Characterizations
A-80
NegativeBinomial(A,B) y = 1, 2, 3, …, 0 < A < 1, 1 ≤ B ≤ y
0.12
0.10 H0.45, 5L
0.08
PDF
0.06
0.04
0.02
0.00
5 10 15 20 25
Y
y–1 y–B
PDF = AB 1 – A
B–1
Moments, etc.
Mean = B
A
B 1–A
Variance =
A2
Mode = int A + B – 1
A
A-81
Notes
1. The NegativeBinomial is also known as the Pascal distribution. The latter name is
restricted to the case (as here) in which B is an integer.
2. It is also known as the Polya distribution.
3. If B = 1, the NegativeBinomial distribution becomes the Geometric distribution.
Characterizations
1. If Prob(success) = A, the number of Bernoulli trials required to realize the Bth success is
~NegativeBinomial(A, B).
A-82
NegBinomial(A,C)&NegBinomial(B,C)
y = 1, 2, 3, …, 0 < B < A < 1, 1 ≤ C ≤ y, 0 < p < 1
0.010
0.005
0.000
10 15 20 25 30 35 40 45 50
Y
Moments, etc.
p 1–p
Mean = C +
A B
C p B2 1 + C 1 – p – A2 1 – p B – p C – 1 – p A B B + 2 C 1 – p
Variance =
A2 B2
A-83
Notes
Characterizations
A-84
Normal(A,B) B>0
0.7
0.6
0.4
PDF
0.3 H0, 1L
0.2
0.1
0.0
-4 -2 0 2 4 6
Y
2
PDF = 1 exp – 1 y – A
B 2π 2 B
y–A
CDF = Φ
B
Moments, etc.
Mean = Median = Mode = A
Variance = B 2
Skewness = Kurtosis = 0
Q1 ≈ A – 0.6745 B Q3 ≈ A + 0.6745 B
A-85
Notes
y–A
z =
B
B = N s
N–1
where N is the sample size.
Characterizations
1. Let Z1, Z2, …, ZN be i.i.d. random variables with finite values for their mean (µ) and
variance (σ2). Then, for any real number (z),
N
Zi – µ
lim Prob
N→∞
1
N
Σ
i=1 σ ≤z = Φ z
known as the Central Limit Theorem. Loosely speaking, the sum of k random variables,
from the same distribution, tends to be ~Normal, and more so as k increases.
2. If X ~Normal(A, B), then Y = a X + b ~Normal(a A + b, a B).
3. If X ~Normal(A, B) and Y ~Normal(C, D), then S = X + Y
2 2
(i.e., the convolution of X and Y) is ~ Normal A + C, B + D .
4. Errors of real-valued observations are often ~Normal or ~Laplace.
A-86
Normal(A,B)&Laplace(C,D) B, D > 0, 0 < p < 1
0.3
0.2
PDF
0.1
0.0
-4 -2 0 2 4 6
Y
2
p y–A 1–p y–C
PDF = exp – 1 + exp –
B 2π 2 B 2D D
y–A y–C
CDF = p Φ + 1 – p stdLaplaceCDF
B D
Moments, etc.
Mean = p A + 1 – p C
2
Variance = p B 2 – p – 1 A – C – 2 p – 1 D2
RandVar: determined by p
A-87
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. The alternate, Laplace(A,B)&Normal(C,D) distribution may be obtained by switching
identities in the parameter dialog. In this case, the parameters shown above in the
Moments section must be reversed (cf. L&L and N&N).
Characterizations
A-88
Normal(A,B)&Laplace(A,C) B, C > 0, 0 < p < 1
0.4
0.2
0.1
0.0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Y
2
p y–A 1–p y–A
PDF = exp – 1 + exp –
B 2π 2 B 2C C
y–A y–A
CDF = p Φ + 1 – p stdLaplaceCDF
B C
Moments, etc.
Mean = Median = Mode = A
Variance = p B 2 + 2 1 – p C 2
Skewness = 0
3 p B 4 + 24 1 – p C 4
Kurtosis = 2
–3
p B2 + 2 1 – p C2
A-89
qMean = qMode = 0.5
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. The alternate, Laplace(A,B)&Normal(A,C) distribution may be obtained by switching
identities in the parameter dialog. In this case, the parameters shown above in the
Moments section must be reversed (cf. L&L and N&N).
Characterizations
A-90
Normal(A,B)&Normal(C,D) B, D > 0, 0 < p < 1
0.3
0.2
PDF
0.1
0.0
-4 -2 0 2 4 6
Y
2 2
p y–A 1–p y–C
PDF = exp – 1 + exp – 1
B 2π 2 B D 2π 2 D
y–A y–C
CDF = p Φ + 1–p Φ
B D
Moments, etc.
Mean = p A + 1 – p C
2
Variance = p B 2 – p – 1 A – C – p – 1 D2
RandVar: determined by p
A-91
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. Whether or not this much-studied mixture is bimodal depends partly upon parameter p.
Obviously, if p is small enough, this mixture will be unimodal regardless of the
remaining parameters. If
2 2
A – C > 8 2B D 2
2
B +D
then there will be some values of p for which this mixture is bimodal. However, if
A–C
2
< 27 B 2 D 2
4 B2 + D2
Characterizations
A-92
Normal(A,B)&Normal(A,C) B, C > 0, 0 < p < 1
0.4
0.2
0.1
0.0
-6 -4 -2 0 2 4 6
Y
2 2
p y–A 1–p y–A
PDF = exp – 1 + exp – 1
B 2π 2 B C 2π 2 C
y–A y–A
CDF = p Φ + 1–p Φ
B C
Moments, etc.
Mean = Median = Mode = A
Variance = p B 2 + 1 – p C 2
Skewness = 0
3 p 1 – p B2 – C2
Kurtosis = 2
p B2 + 1 – p C2
A-93
qMean = qMode = 0.5
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-94
Normal(A,B)&Uniform(C,D) B > 0, C < D, 0 < p < 1
0.5
0.4
0.2
0.1
0.0
-4 -3 -2 -1 0 1 2 3 4
Y
p y–A
2 1–p C<y<D
PDF = exp – 1 +
B 2π 2 B D–C
y–A
pΦ , y≤C
B
y–A y–C
CDF = pΦ + 1–p , C<y<D
B D–C
y–A
pΦ + 1–p , y≥D
B
Moments, etc.
1–p C+D
Mean = p A +
2
2
1–p C+D
Variance = p A 2 + B 2 – p A + + 1 1 – p C2 + C D + D2
2 3
A-95
Quantiles, etc.: no simple closed form
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
Characterizations
A-96
Normal(A,B)&Uniform(A,C) B, C > 0, 0 < p < 1
0.5
0.3
PDF
0.2
0.1
0.0
-3 -2 -1 0 1 2 3
Y
y–A
pΦ , y≤A–C
B
y–A y–A+C
CDF = pΦ + 1–p , A–C<y<A+C
B 2C
y–A
pΦ + 1–p , y≥A+C
B
Moments, etc.
Mean = Median = Mode = A
3 p B2 + 1 – p C2
Variance =
3
A-97
Skewness = 0
9 15 p B 4 + 1 – p C 4
Kurtosis = 2 –3
2 2
5 3pB + 1–p C
RandVar: determined by p
Notes
1. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
2. Warning! Mixtures usually have several local optima.
3. It is especially difficult to get the optimum value for parameter C in this distribution.
There is an unusual amount of ambiguity in this case. It might be best to start out with
parameter C Constant at its likely value (if known) and proceed from there.
Characterizations
A-98
Pareto(A,B) 0 < A ≤ y, B > 0
2.00
1.75 H1, 2L
1.50
1.25
PDF
1.00
0.75
0.50
H1, 1L
0.25
0.00
1 2 3 4 5
Y
B
PDF = BBA+ 1
y
B
CDF = 1 – A
y
Variance = A2 B
(B – 2) (B – 1) 2
2 B+1 B–2
Skewness =
B–3 B
6 B3 + B2 – 6 B – 2
Kurtosis =
B B 2 – 7 B + 12
A-99
Mode = A
Median = A B 2
Q1 = A
B 4 Q3 = A B 4
3
B
qMean = 1 – B – 1 qMode = 0
B
RandVar = BA
u
Notes
Characterizations
1. The Pareto distribution is often used as an income distribution. Thus, the probability that
a random income, in some defined population, exceeds a minimum, A, is ~Pareto.
A-100
Poisson(A) y = 0, 1, 2, …, A > 0
0.25
H3.5L
0.20
0.15
PDF
0.10
0.05
0.00
0 2 4 6 8 10 12
Y
exp – A A y
PDF =
y!
Moments, etc.
Mean = Variance = A
Mode = int A
A-101
Notes
Characterizations
A-102
Poisson(A)&Poisson(B) y = 0, 1, 2, …, 0 < A < B, 0 < p < 1
0.10
0.09
0.07
0.06
PDF
0.05
0.04
0.03
0.02
0.01
0.00
0 2 4 6 8 10 12 14 16 18 20
Y
p exp – A A y + 1 – p exp – B B y
PDF =
y!
Moments, etc.
Mean = p A + 1 – p B
2
Variance = p A A + 1 + 1 – p B B + 1 – p A – B + B
A-103
Notes
Characterizations
A-104
Reciprocal(A,B) 0<A≤y≤B
0.25
0.20
0.15
PDF
0.10
0.00
0 20 40 60 80 100
Y
PDF = 1
y log B
A
log A
y
CDF =
log A
B
Parameters -- A, B: Shape
A-105
2
2 12 d A – B + d 2 A 2 2 d – 9 + 2 A B d + B 2 2 d + 9
Skewness = 3
2
3d A–B A d–2 +B d+2
3 2
– 36 A – B + 36 d A – B A + B – 16 d 2 A 3 – B 3 + 3 d 3 A 2 + B 2 A + B
Kurtosis = 2 –3
3 A–B A d–2 +B d+2
Mode = A
Median = AB
4 4
Q1 = A3 4 B Q3 = 4
A B3
RandVar = A 1 – u B u
Notes
Characterizations
A-106
Semicircular(A,B) A – B ≤ y ≤ A + B, B > 0
H0, 1L
0.7
0.6
0.5
0.4
PDF
0.3
0.2
0.1
0.0
-1 0 1
Y
2
2 y–A
PDF = 1–
Bπ B
2
1 y–A
CDF = 1 + π 1–
y–A
+ asin
y–A
2 B B B
Moments, etc.
Mean = Median = Mode = A
2
Variance = B
4
Skewness = 0
Kurtosis = – 1
Q1 ≈ A – 0.4040 B Q3 ≈ A + 0.4040 B
Characterizations
A-108
StudentsT(A,B,C) B > 0, 0 < C ≤ 100
0.8
0.6
0.5
PDF
0.4
0.3 H0, 1, 2L
0.2
0.1
0.0
-5 -4 -3 -2 -1 0 1 2 3 4 5
Y
– C+1
2 2
y–A
Γ C+1
2 B
PDF = 1+
C
B πC Γ C
2
1I C , C, 1 , y–A
t≡ ≤0
2 C + t2 2 2 B
CDF =
1– 1 I C , C, 1 , t ≡ y – A > 0
2 C + t2 2 2 B
Variance = C B2
C–2
Skewness = 0
A-109
C–2 Γ C –2
2
2
Kurtosis = 3 –1
4Γ C
2
Notes
1. In the literature, C > 0. The bounds shown here are used to prevent divergence.
2. The Student’s-t distribution approaches the Normal distribution asymptotically as
C -> infinity.
3. If the optimum model has C = 100, use Normal instead.
4. Moment k exists if C > k.
Characterizations
1. The Student’s-t distribution is used to characterize small samples (typically, N < 30) from
a Normal population.
A-110
Triangular(A,B,C) A < y, C < B
0.3
0.2
PDF
H-4, 4, 2L
0.1
0.0
-4 -3 -2 -1 0 1 2 3 4
Y
2 y–A
, y<C
B–A C–A
PDF =
2 B–y
, y≥C
B–A B–C
2
y–A
, y<C
B–A C–A
CDF =
A B – C + B C – 2 y + y2
, y≥C
A–B B–C
Moments, etc. d ≡ A + B – B C + C – A B + C
2 2 2
Mean = A + B + C
3
A-111
Variance = d
18
Kurtosis = – 3
5
Mode = C
2
B+C–2A
, C≥ A+B
9 B–A C–A 2
qMean =
A2 + 5 A B – 5 B2 – 7 A C + 5 B C + C2 , C < A + B
9 A–B B–C 2
qMode = C – A
B–A
Notes
Characterizations
A-112
Uniform(A,B) A<y<B
H-1.6, -1.4L
3
PDF
H-0.5, 0.5L
H1, 3L
1
0
-2 -1 0 1 2 3
Y
PDF = 1
B–A
y–A
CDF =
B–A
Moments, etc.
Mean = Median = A + B
2
Variance = 1 B – A
2
12
Skewness = 0
Kurtosis = – 6
5
Mode = none
Q1 = 3 A + B Q3 = A + 3 B
4 4
A-113
qMean = 0.5
RandVar = A + u B – A
Notes
1. The parameters for the Uniform distribution are sometimes given, equivalently, as the
mean and half-width of the domain.
Characterizations
1. The Uniform distribution is commonly used to describe total ignorance within a bounded
interval.
2. Pseudo-random number generators typically return X ~Uniform(0, 1) as their primary
output.
A-114
Uniform(A,B)&Uniform(C,D) A < y < B or D, A < C < B, D, 0 < p < 1
0.3
0.1
0.0
-3 -2 -1 0 1 2 3
Y
where
p y–A
cdf1 = p if y > B , else and
B–A
1–p y–C
cdf2 = 0 if y < C , else 1 – p if y > D , else
D–C
Moments, etc.
p A+B + 1–p C+D
Mean =
2
2
p A2 + A B + B2 – 3 p A + B + 1 – p C + D + 1 – p C2 + C D + D2
4
Variance =
3
Notes
1. As implemented here, this distribution requires overlap between the two components.
Also, Component #1 must cover min(y) although either Component may cover max(y).
2. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
3. Warning! Mixtures usually have several local optima.
Characterizations
A-116
Uniform(A,B)&Uniform(A,C) A – C < y < A + C, B < C, 0 < p < 1
0.5
H0, 1, 2, 0.7L
0.4
0.3
PDF
0.2
0.1
0.0
-2 -1 0 1 2
Y
1–p
y–A+C , y≤A–B
2C
p 1–p
CDF = y–A+B + y–A+C , A–B<y<A+B
2B 2C
1–p
p+ y–A+C , y≥A+B
2C
Moments, etc.
Mean = Median = A
p B2 + 1 – p C2
Variance =
3
Skewness = 0
A-117
9 p B4 + 1 – p C4
Kurtosis = 2 –3
2 2
5 pB + 1–p C
Mode = none
qMean = 0.5
RandVar: determined by p
Notes
1. Note that the parameters are defined differently here than in all other Uniform
distributions discussed elsewhere in this document.
2. Binary mixtures may require hundreds of data points for adequate optimization and, even
then, often have unusually wide confidence intervals. In fact, the criterion response is
sometimes flat over a broad range, esp. with respect to parameter p.
3. Warning! Mixtures usually have several local optima.
Characterizations
A-118
Weibull(A,B,C) y > A, B, C > 0
1.0
0.8 H1, 1, 1L
0.6
H1, 2, 3L
PDF
0.4
0.2
0.0
1 2 3 4 5
Y
C–1 C
y–A y–A
PDF = C exp –
B B B
C
y–A
CDF = 1 – exp –
B
Moments, etc.
Mean = A + B Γ C + 1
C
Variance = B 2 Γ C + 2 – Γ 2 C + 1
C C
2 Γ3 C + 1 – 3 Γ C + 1 Γ C + 2 + Γ C + 3
C C C C
Skewness = 3
Γ C + 2 – Γ2 C + 1
C C
A-119
– 3 Γ4 C + 1 + 6 Γ2 C + 1 Γ C + 2 – 4 Γ C + 1 Γ C + 3 + Γ C + 4
C C C C C C
Kurtosis = 2 –3
Γ 2 C+1 –Γ C+2
C C
A, C≤1
Mode =
A+B
C C–1 , C>1
C
C
Median = A + B log 2
log 4
C C
Q1 = A + B Q3 = A + B log 4
3
RandVar = A + B C – log u
Notes
Characterizations
C
y–A
1. If X = B is ~ standard Exponential, then y ~Weibull(A, B, C). Thus, the
Weibull distribution is a generalization of the Exponential distribution.
2. Weibull(–y) is an extreme-value distribution for variates with an upper bound.
A-120