- Source: Hypoexponential distribution
In probability theory the hypoexponential distribution or the generalized Erlang distribution is a continuous distribution, that has found use in the same fields as the Erlang distribution, such as queueing theory, teletraffic engineering and more generally in stochastic processes. It is called the hypoexponetial distribution as it has a coefficient of variation less than one, compared to the hyper-exponential distribution which has coefficient of variation greater than one and the exponential distribution which has coefficient of variation of one.
Overview
The Erlang distribution is a series of k exponential distributions all with rate
λ
{\displaystyle \lambda }
. The hypoexponential is a series of k exponential distributions each with their own rate
λ
i
{\displaystyle \lambda _{i}}
, the rate of the
i
t
h
{\displaystyle i^{th}}
exponential distribution. If we have k independently distributed exponential random variables
X
i
{\displaystyle {\boldsymbol {X}}_{i}}
, then the random variable,
X
=
∑
i
=
1
k
X
i
{\displaystyle {\boldsymbol {X}}=\sum _{i=1}^{k}{\boldsymbol {X}}_{i}}
is hypoexponentially distributed. The hypoexponential has a minimum coefficient of variation of
1
/
k
{\displaystyle 1/k}
.
= Relation to the phase-type distribution
=As a result of the definition it is easier to consider this distribution as a special case of the phase-type distribution. The phase-type distribution is the time to absorption of a finite state Markov process. If we have a k+1 state process, where the first k states are transient and the state k+1 is an absorbing state, then the distribution of time from the start of the process until the absorbing state is reached is phase-type distributed. This becomes the hypoexponential if we start in the first 1 and move skip-free from state i to i+1 with rate
λ
i
{\displaystyle \lambda _{i}}
until state k transitions with rate
λ
k
{\displaystyle \lambda _{k}}
to the absorbing state k+1. This can be written in the form of a subgenerator matrix,
[
−
λ
1
λ
1
0
…
0
0
0
−
λ
2
λ
2
⋱
0
0
⋮
⋱
⋱
⋱
⋱
⋮
0
0
⋱
−
λ
k
−
2
λ
k
−
2
0
0
0
…
0
−
λ
k
−
1
λ
k
−
1
0
0
…
0
0
−
λ
k
]
.
{\displaystyle \left[{\begin{matrix}-\lambda _{1}&\lambda _{1}&0&\dots &0&0\\0&-\lambda _{2}&\lambda _{2}&\ddots &0&0\\\vdots &\ddots &\ddots &\ddots &\ddots &\vdots \\0&0&\ddots &-\lambda _{k-2}&\lambda _{k-2}&0\\0&0&\dots &0&-\lambda _{k-1}&\lambda _{k-1}\\0&0&\dots &0&0&-\lambda _{k}\end{matrix}}\right]\;.}
For simplicity denote the above matrix
Θ
≡
Θ
(
λ
1
,
…
,
λ
k
)
{\displaystyle \Theta \equiv \Theta (\lambda _{1},\dots ,\lambda _{k})}
. If the probability of starting in each of the k states is
α
=
(
1
,
0
,
…
,
0
)
{\displaystyle {\boldsymbol {\alpha }}=(1,0,\dots ,0)}
then
H
y
p
o
(
λ
1
,
…
,
λ
k
)
=
P
H
(
α
,
Θ
)
.
{\displaystyle Hypo(\lambda _{1},\dots ,\lambda _{k})=PH({\boldsymbol {\alpha }},\Theta ).}
Two parameter case
Where the distribution has two parameters (
λ
1
≠
λ
2
{\displaystyle \lambda _{1}\neq \lambda _{2}}
) the explicit forms of the probability functions and the associated statistics are:
CDF:
F
(
x
)
=
1
−
λ
2
λ
2
−
λ
1
e
−
λ
1
x
−
λ
1
λ
1
−
λ
2
e
−
λ
2
x
{\displaystyle F(x)=1-{\frac {\lambda _{2}}{\lambda _{2}-\lambda _{1}}}e^{-\lambda _{1}x}-{\frac {\lambda _{1}}{\lambda _{1}-\lambda _{2}}}e^{-\lambda _{2}x}}
PDF:
f
(
x
)
=
λ
1
λ
2
λ
1
−
λ
2
(
e
−
x
λ
2
−
e
−
x
λ
1
)
{\displaystyle f(x)={\frac {\lambda _{1}\lambda _{2}}{\lambda _{1}-\lambda _{2}}}(e^{-x\lambda _{2}}-e^{-x\lambda _{1}})}
Mean:
1
λ
1
+
1
λ
2
{\displaystyle {\frac {1}{\lambda _{1}}}+{\frac {1}{\lambda _{2}}}}
Variance:
1
λ
1
2
+
1
λ
2
2
{\displaystyle {\frac {1}{\lambda _{1}^{2}}}+{\frac {1}{\lambda _{2}^{2}}}}
Coefficient of variation:
λ
1
2
+
λ
2
2
λ
1
+
λ
2
{\displaystyle {\frac {\sqrt {\lambda _{1}^{2}+\lambda _{2}^{2}}}{\lambda _{1}+\lambda _{2}}}}
The coefficient of variation is always less than 1.
Given the sample mean (
x
¯
{\displaystyle {\bar {x}}}
) and sample coefficient of variation (
c
{\displaystyle c}
), the parameters
λ
1
{\displaystyle \lambda _{1}}
and
λ
2
{\displaystyle \lambda _{2}}
can be estimated as follows:
λ
1
=
2
x
¯
[
1
+
1
+
2
(
c
2
−
1
)
]
−
1
{\displaystyle \lambda _{1}={\frac {2}{\bar {x}}}\left[1+{\sqrt {1+2(c^{2}-1)}}\right]^{-1}}
λ
2
=
2
x
¯
[
1
−
1
+
2
(
c
2
−
1
)
]
−
1
{\displaystyle \lambda _{2}={\frac {2}{\bar {x}}}\left[1-{\sqrt {1+2(c^{2}-1)}}\right]^{-1}}
These estimators can be derived from the methods of moments by setting
1
λ
1
+
1
λ
2
=
x
¯
{\displaystyle {\frac {1}{\lambda _{1}}}+{\frac {1}{\lambda _{2}}}={\bar {x}}}
and
λ
1
2
+
λ
2
2
λ
1
+
λ
2
=
c
{\displaystyle {\frac {\sqrt {\lambda _{1}^{2}+\lambda _{2}^{2}}}{\lambda _{1}+\lambda _{2}}}=c}
.
The resulting parameters
λ
1
{\displaystyle \lambda _{1}}
and
λ
2
{\displaystyle \lambda _{2}}
are real values if
c
2
∈
[
0.5
,
1
]
{\displaystyle c^{2}\in [0.5,1]}
.
Characterization
A random variable
X
∼
H
y
p
o
(
λ
1
,
…
,
λ
k
)
{\displaystyle {\boldsymbol {X}}\sim Hypo(\lambda _{1},\dots ,\lambda _{k})}
has cumulative distribution function given by,
F
(
x
)
=
1
−
α
e
x
Θ
1
{\displaystyle F(x)=1-{\boldsymbol {\alpha }}e^{x\Theta }{\boldsymbol {1}}}
and density function,
f
(
x
)
=
−
α
e
x
Θ
Θ
1
,
{\displaystyle f(x)=-{\boldsymbol {\alpha }}e^{x\Theta }\Theta {\boldsymbol {1}}\;,}
where
1
{\displaystyle {\boldsymbol {1}}}
is a column vector of ones of the size k and
e
A
{\displaystyle e^{A}}
is the matrix exponential of A. When
λ
i
≠
λ
j
{\displaystyle \lambda _{i}\neq \lambda _{j}}
for all
i
≠
j
{\displaystyle i\neq j}
, the density function can be written as
f
(
x
)
=
∑
i
=
1
k
λ
i
e
−
x
λ
i
(
∏
j
=
1
,
j
≠
i
k
λ
j
λ
j
−
λ
i
)
=
∑
i
=
1
k
ℓ
i
(
0
)
λ
i
e
−
x
λ
i
{\displaystyle f(x)=\sum _{i=1}^{k}\lambda _{i}e^{-x\lambda _{i}}\left(\prod _{j=1,j\neq i}^{k}{\frac {\lambda _{j}}{\lambda _{j}-\lambda _{i}}}\right)=\sum _{i=1}^{k}\ell _{i}(0)\lambda _{i}e^{-x\lambda _{i}}}
where
ℓ
1
(
x
)
,
…
,
ℓ
k
(
x
)
{\displaystyle \ell _{1}(x),\dots ,\ell _{k}(x)}
are the Lagrange basis polynomials associated with the points
λ
1
,
…
,
λ
k
{\displaystyle \lambda _{1},\dots ,\lambda _{k}}
.
The distribution has Laplace transform of
L
{
f
(
x
)
}
=
−
α
(
s
I
−
Θ
)
−
1
Θ
1
{\displaystyle {\mathcal {L}}\{f(x)\}=-{\boldsymbol {\alpha }}(sI-\Theta )^{-1}\Theta {\boldsymbol {1}}}
Which can be used to find moments,
E
[
X
n
]
=
(
−
1
)
n
n
!
α
Θ
−
n
1
.
{\displaystyle E[X^{n}]=(-1)^{n}n!{\boldsymbol {\alpha }}\Theta ^{-n}{\boldsymbol {1}}\;.}
General case
In the general case
where there are
a
{\displaystyle a}
distinct sums of exponential distributions
with rates
λ
1
,
λ
2
,
⋯
,
λ
a
{\displaystyle \lambda _{1},\lambda _{2},\cdots ,\lambda _{a}}
and a number of terms in each
sum equals to
r
1
,
r
2
,
⋯
,
r
a
{\displaystyle r_{1},r_{2},\cdots ,r_{a}}
respectively. The cumulative
distribution function for
t
≥
0
{\displaystyle t\geq 0}
is given by
F
(
t
)
=
1
−
(
∏
j
=
1
a
λ
j
r
j
)
∑
k
=
1
a
∑
l
=
1
r
k
Ψ
k
,
l
(
−
λ
k
)
t
r
k
−
l
exp
(
−
λ
k
t
)
(
r
k
−
l
)
!
(
l
−
1
)
!
,
{\displaystyle F(t)=1-\left(\prod _{j=1}^{a}\lambda _{j}^{r_{j}}\right)\sum _{k=1}^{a}\sum _{l=1}^{r_{k}}{\frac {\Psi _{k,l}(-\lambda _{k})t^{r_{k}-l}\exp(-\lambda _{k}t)}{(r_{k}-l)!(l-1)!}},}
with
Ψ
k
,
l
(
x
)
=
−
∂
l
−
1
∂
x
l
−
1
(
∏
j
=
0
,
j
≠
k
a
(
λ
j
+
x
)
−
r
j
)
.
{\displaystyle \Psi _{k,l}(x)=-{\frac {\partial ^{l-1}}{\partial x^{l-1}}}\left(\prod _{j=0,j\neq k}^{a}\left(\lambda _{j}+x\right)^{-r_{j}}\right).}
with the additional convention
λ
0
=
0
,
r
0
=
1
{\displaystyle \lambda _{0}=0,r_{0}=1}
.
Uses
This distribution has been used in population genetics, cell biology, and queuing theory.
See also
Phase-type distribution
Coxian distribution
References
Further reading
M. F. Neuts. (1981) Matrix-Geometric Solutions in Stochastic Models: an Algorthmic Approach, Chapter 2: Probability Distributions of Phase Type; Dover Publications Inc.
G. Latouche, V. Ramaswami. (1999) Introduction to Matrix Analytic Methods in Stochastic Modelling, 1st edition. Chapter 2: PH Distributions; ASA SIAM,
Colm A. O'Cinneide (1999). Phase-type distribution: open problems and a few properties, Communication in Statistic - Stochastic Models, 15(4), 731–757.
L. Leemis and J. McQueston (2008). Univariate distribution relationships, The American Statistician, 62(1), 45—53.
S. Ross. (2007) Introduction to Probability Models, 9th edition, New York: Academic Press
Kata Kunci Pencarian:
- Hypoexponential distribution
- Exponential distribution
- Phase-type distribution
- List of convolutions of probability distributions
- Hyperexponential distribution
- List of statistics articles
- Multivariate normal distribution
- Elliptical distribution
- Multivariate Laplace distribution
- Wilks's lambda distribution