# A general stochastic model for bivariate episodes driven by a gamma sequence

## Abstract

We propose a new stochastic model describing the joint distribution of (X,N), where N is a counting variable while X is the sum of N independent gamma random variables. We present the main properties of this general model, which include marginal and conditional distributions, integral transforms, moments and parameter estimation. We also discuss in more detail a special case where N has a heavy tailed discrete Pareto distribution. An example from finance illustrates the modeling potential of this new mixed bivariate distribution.

## Introduction

We propose a new stochastic model describing the joint distribution of

$$(X,N)\stackrel{d}{=}\left(\sum\limits_{i=1}^{N}X_{i},N\right),$$
(1)

where N is an integer-valued random variable on $$\mathbb {N} = \{1, 2, \ldots \}$$ while the {Xi} are independent and identically distributed (IID) random variables on $$\mathbb {R}_{+}$$, independent of N. Our new model is a generalization of the BEG and the BGG models introduced in Kozubowski and Panorska (2005) and Barreto-Souza (2012), respectively, where N was geometrically distributed with the probability mass function (PMF)

$$P(N=n)=p(1-p)^{n-1},\,\,\, n\in\mathbb{N},$$
(2)

and the {Xi} were either exponential (BEG), with the probability density function (PDF)

$$f(x)=\beta e^{-\beta x}, \,\,\, x\in \mathbb{R}_{+},$$
(3)

or gamma distributed (BGG), with the PDF

$$f(x)=\frac{\beta^{\alpha} x^{\alpha-1}}{\Gamma(\alpha)}e^{-\beta x}, \,\,\, x\in \mathbb{R}_{+}.$$
(4)

Models of this kind were shown to be useful in various areas, including finance, actuarial science, hydro-climatic studies, and others (see, e.g., Arendarczyk et al. 2018; Barreto-Souza 2012; Barreto-Souza and Silva 2019; Biondi et al. 2005, 2008; Kozubowski and Panorska 2005, 2010; Kozubowski et al. 2008, 2009 and reference therein). In modeling, the vector (X,N) in (1) is often called an “episode” with magnitude X and duration N. For example, in finance we are interested in growth/decline episodes of a certain investment, which happen when the consecutive log-returns {Xi} are positive/negative. The sequence of N consecutive positive log-returns is called a“growth/decline episode of duration N and magnitude $$X = \sum \limits _{i=1}^{N}X_{i}$$”.

The majority of work connected with bivariate distributions given by (1) concerned cases where the distributions of X and N were both light-tailed, with the exception of Arendarczyk et al. (2018), where N was (light tailed) geometric and X was heavy tailed with Lomax distribution (also known as Pareto Type II), given by the survival function (SF)

$$S(x)=P(X>x)=\left(\frac{1}{1+x/\sigma}\right)^{1/\delta}, \,\,\, x\in \mathbb{R}_{+}, \; \sigma > 0, \; \delta \geq 0.$$
(5)

In contrast, in our generalization studied in this paper, both X and N in (1) are heavy tailed. We achieve this by keeping the {Xi} in (1) gamma distributed with PDF (4) as in Barreto-Souza (2012), while changing N to be a discrete version of the continuous Pareto distribution (5) with the same survival function (see, e.g., Buddana and Kozubowski 2014; Krishna and Singh Pundir 2009). The PMF of this discrete Pareto (DP) distribution can be written as

$$P(N=n) = \left(\frac{1}{1-\delta(n-1)\log(1-p)}\right)^{\frac{1}{\delta}}- \left(\frac{1}{1-\delta n \log(1-p)}\right)^{\frac{1}{\delta}},\,\,\, n\in \mathbb{N},$$
(6)

where p(0,1) is a “size” parameter and δ[0,) is a tail parameter. We note that as δ→0+, then the DP distribution (6) converges to a geometric distribution (2). Thus, while our model (1) with DP distributed N offers more flexibility compared with the BEG model of Kozubowski and Panorska (2005) or the BGG model of Barreto-Souza (2012), it turns into the latter two in the special case with δ=0 and exponential or gamma distributed {Xi} in (1).

In addition to presenting basic properties of the new model described above, we also summarize fundamental properties of the entire class of such models with gamma distributed {Xi} and arbitrary distribution of the random variable N, which drives the dependence structure of the (X,N) vector in (1). These properties include infinite divisibility, the tail behavior of X, the conditional distributions of N given X=x as well as N given X>x, and estimation. In particular, we establish the existence and uniqueness of maximum likelihood estimators of α and β within this class of models. Our general results concerning the conditional distributions have interesting connections with hidden truncation (see, e.g., Arnold ) and play important role in goodness-of-fit analysis for these models.

Our article is organized as follows. In Section 2 we present fundamental properties of general class of bivariate distributions (1), driven by an IID gamma sequence {Xi}. In Section 3, we summarize basic facts about the special case with discrete Pareto distributed N, including parameter estimation. An application from finance is presented in Section 4, which is followed by concluding remarks in Section 5. The last section contains selected proofs and technical lemmas.

## Bivariate episodes driven by an IID gamma sequence

In this section we summarize the properties which are common to all bivariate distributions describing the episodes (1) driven by a sequence {Xi} of IID gamma variables with the PDF (4), denoted by $$\mathcal {GAM}(\alpha, \beta)$$. We assume that the variable N, which describes the duration of an episode (X,N), is independent of the {Xi} and supported on the set of positive integers $$\mathbb {N} = \{1, 2, \ldots \}$$. Clearly, given $$N=n \in \mathbb {N}$$, the variable X is the sum of n IID gamma variables $$\mathcal {GAM}(\alpha, \beta)$$, so it is also gamma distributed, with parameters αn and β. Thus, the joint PDF of (X,N) in (1) is of the form

$$f_{X,N}(x,n) =\frac{\beta^{\alpha n}x^{\alpha n-1}e^{-\beta x}}{\Gamma(\alpha n)}f_{N}(n), \,\,\, x\in \mathbb{R}_{+}, n\in \mathbb{N},$$
(7)

where fN(·) is the PMF of N. This leads to the following result, whose proof is elementary.

### Proposition 1

The cumulative distribution function (CDF) and the survival functions (SF) of (X,N) in (1) are given by

$$F_{X,N}(x,n) = \mathbb{P}(X\leq x,N\leq n)=\sum\limits_{k=1}^{n}\frac{\gamma(\alpha k, \beta x)}{\Gamma(\alpha k)}f_{N}(k),\,\,\, x\in \mathbb{R}_{+}, n\in\mathbb{N},$$

and

$$\begin{array}{@{}rcl@{}} S_{X,N}(x,n) = \mathbb{P}(X> x,N> n) = \sum\limits_{k=n+1}^{\infty} \left(1 - \frac{\gamma(\alpha k, \beta x)}{\Gamma(\alpha k)}\right) f_{N}(k),\,\,\, x\in \mathbb{R}_{+}, \\n\in\mathbb{N}_{0} = \{0,1,\ldots\}, \end{array}$$

where fN(·)is the PMF of N and

$$\gamma(\alpha, x)=\int\limits_{0}^{x}t^{\alpha -1}e^{-t}dt,\,\,\, \alpha, x>0,$$

is the (lower) incomplete gamma function.

In view of the above, we conclude that the marginal distribution of X is an infinite mixture of gamma distributions, as the PDF of X takes on the form

$$f_{X}(x) = \sum\limits_{n=1}^{\infty} \frac{\beta^{\alpha n}x^{\alpha n-1}e^{-\beta x}}{\Gamma(\alpha n)}f_{N}(n), \,\,\, x\in \mathbb{R}_{+}.$$
(8)

### Remark 1

In the special case where N is geometric (2) we recover the Bivariate Gamma-Geometric (BGG) distribution of Barreto-Souza (2012). In this case, the PDF of X can be written in terms of 2-parameter Mittag-Leffler (ML) special function

$$E_{\alpha, \beta}(z) = \sum_{n=0}^{\infty}\frac{z^{n}}{\Gamma(\alpha n +\beta)}, \,\,\, z\in C, Re(\alpha)>0, Re(\beta)>0,$$
(9)

which is an entire function on the complex plain C (see, e.g., Haubold et al. 2011). Indeed, simple algebra shows that the PDF (8) takes on the form

$$f_{X}(x) = \beta p (\beta x)^{\alpha-1} e^{-\beta x} E_{\alpha,\alpha}\left((1-p)(\beta x)^{\alpha}\right), \,\,\, x\in \mathbb{R}_{+}.$$
(10)

Further, in the special case α=1, when the {Xi} in (1) are exponentially distributed, we obtain the BEG model of Kozubowski and Panorska (2005), where the term “BEG” stands for Bivariate distribution with Exponential and Geometric margins. Here, the ML special function becomes the exponential function, E1,1(z)=ez, and the marginal density of X in (10) reduces to the exponential PDF (3) with β replaced with pβ.

Let us also account for the joint Laplace transform ψ(t,s) connected with the distribution of (X,N) in (1), where

$$\psi(t,s) = \mathbb{E} \left(e^{-tX-sN}\right), \,\,\, t, s\in \mathbb{R}_{+}.$$
(11)

Standard conditioning argument shows that this function can be conveniently expressed in terms of the probability generating function (PGF) of the random variable N, $$G_{N}(s) = \mathbb {E} (s^{N}), s \in [0,1]$$.

### Proposition 2

The LT of (X,N)in (1) driven by an IID gamma $$\mathcal {GAM}(\alpha, \beta)$$ sequence {Xi} is given by

$$\psi(t,s) = G_{N}\left\{ e^{-s}\left(\frac{\beta}{\beta+t}\right)^{\alpha}\right\}, \,\,\, t, s\in \mathbb{R}_{+},$$
(12)

where GN(·) is the PGF of N.

### Remark 2

When the variable N is geometric (2) where we have

$$G_{N}(s) = \frac{ps}{1-(1-p)s}, \,\,\, s \in [0,1],$$

then (12) turns into the LT of the BGG model of Barreto-Souza (2012),

$$\psi(t,s) = \frac{p \beta^{\alpha} e^{-s}}{(\beta+t)^{\alpha} -(1-p)\beta^{\alpha} e^{-s}}, \,\,\, t, s\in \mathbb{R}_{+}.$$

### Remark 3

As can be seen from the proof of Proposition 2, if the IID variables {Xi} are non-negative but not necessarily gamma distributed, then the LT in (12) is given by $$\phantom {\dot {i}\!}\psi (t,s) = G_{N} [e^{-s} \psi _{X_{1}}(t)]$$, where $$\phantom {\dot {i}\!}\psi _{X_{1}}(\cdot)$$ is the LT of the {Xi}.

### 2.1 Moments and tail behavior

Here we consider moments and the tail behavior of the variables X and N. Clearly, if the counting variable N is light tailed, and all of its moments $$\mathbb {E}[N^{\eta }]$$ exist for η>0, then the mixed moments $$\mathbb {E}[X^{r}N^{\eta }]$$ exist as well for any r,η>0. However, this is not so if the distribution of N is heavy tailed, and its survival function behaves like a power law,

$$S_{N}(x) = \mathbb{P}(N>x) \sim c x^{-v} \,\,\, (x\rightarrow \infty),$$
(13)

where c is a positive constant and f(x)g(x)(x) means that the ratio f(x)/g(x) converges to 1 as x. Indeed, if the tail probability satisfies (13) for some v>0, then for positive η>0 the expectation $$\mathbb {E}[N^{\eta }]$$ exists only for η<v. Moreover, in this case of heavy tailed N, the variable X in (1) is heavy tailed as well, with essentially the same tail behavior as N.

### Proposition 3

If the variable N in (1) satisfies (13) with some c,v>0 and the {Xi} in (1) are IID gamma distributed $$\mathcal {GAM}(\alpha, \beta)$$, then we have

$$S_{X}(x) = \mathbb{P}(X>x) \sim c (\alpha/\beta)^{v} x^{-v} \,\,\, (x\rightarrow \infty).$$

Under the conditions of Proposition 3, the moments $$\mathbb {E}[X^{r}]$$ with r>0 are finite only for r<v. More generally, standard calculations involving either the joint PDF (7) or conditioning arguments viz. the representation (1) show that the joint moments $$\mathbb {E}[X^{r}N^{\eta }]$$ with positive r,η>0 exist only for r+η<v, and their values are provided in the following result.

### Proposition 4

Under the conditions of Proposition 3, the moments $$\mathbb {E}[X^{r}N^{\eta }]$$ with r,η>0 are finite only for r+η<v, in which case we have

$$\mathbb{E}[X^{r}N^{\eta}] = \frac{1}{\beta^{r}} \sum\limits_{n=1}^{\infty}n^{\eta} \frac{\Gamma(r + \alpha n)}{\Gamma(\alpha n)} f_{N}(n),$$
(14)

where fN(·) is the PMF of N.

If they exist, then the mean vector and the covariance matrix of (X,N) in (1) can be deduced from Proposition 4. The result below provides relevant information concerning this.

### Proposition 5

If the variable N in (1) has a finite mean $$\mathbb {E}[N]$$, then so does X, and we have $$\mathbb {E}[X] = (\alpha /\beta) \mathbb {E}[N]$$. Moreover, if the variance $$\mathbb {V} ar[N]$$ is finite, then so is the covariance matrix Σ of (X,N), and we have

$$\boldsymbol{\Sigma} = \mathbb{E}[N] \left[ \begin{array}{cc} \alpha/\beta^{2} & 0 \\ 0 & 0 \end{array} \right] + \mathbb{V} ar[N] \left[ \begin{array}{cc} \alpha^{2}/\beta^{2} & \alpha/\beta \\ \alpha/\beta & 1 \end{array} \right].$$

### Remark 4

We note that if N is geometric (2), so that $$\mathbb {E}[N]=1/p$$ and $$\mathbb {V} ar[N]=(1-p)/p^{2}$$, we recover the mean vector and the covariance matrix of the BGG model of Barreto-Souza (2012).

### Remark 5

If the variance of N is finite, then the correlation of X and N is a simple function of the parameter α and the index of dispersion of N,

$$\rho = \frac{1}{\sqrt{1+\frac{1}{\alpha}\frac{\mathbb{E}[N]}{\mathbb{V} ar[N]}}},$$

which can fall anywhere in the interval (0,). In the BGG case, this reduces to $$\rho =\sqrt {(1-p)/(1-p+p/\alpha)}$$, which can be either greater than or less than $$\sqrt {1-p}$$, as noted by Barreto-Souza (2012). The latter is the correlation in the BEG model, arising in the special case α=1 of the BGG.

### 2.2 Conditional distributions

The derivation of basic conditional distributions in this bivariate model is straightforward. Clearly, the conditional distribution of X given N=n is $$\mathcal {GAM}(\alpha n, \beta)$$, with the PDF

$$f_{X|N=n}(x) = \frac{\beta^{\alpha n}}{\Gamma(\alpha n)}x^{\alpha n - 1}e^{-\beta x}, \,\,\, x\in \mathbb{R}_{+}.$$
(15)

Similarly, the PMF of the conditional distribution of N given X=x can be obtained by dividing the joint PDF of (X,N) in (7) by the marginal PDF of X, given by (8). Interestingly, this conditional PMF is proportional to the product of two PMFs as follows:

$$f_{N|X=x}(n) = (1/c)f_{N}(n)f_{\alpha, \beta x}(n-1), \,\,\, n\in \mathbb{N},$$
(16)

where c>0 is a normalizing constant that does not involve n, fN(·) is the PMF of N, and

$$f_{\alpha, \lambda}(n) = \frac{\lambda^{\alpha(n+1)}}{\Gamma(\alpha(n+1))} \frac{\lambda^{-\alpha}}{E_{\alpha, \alpha}(\lambda^{\alpha})}, \,\,\, n\in \mathbb{N}_{0}.$$
(17)

The function Eα,α(·) in (17) is the Mittag-Leffler special function defined in (9) with α=β>0. The function in (17) is a genuine probability distribution on the set of non-negative integers $$\mathbb {N}_{0}$$ for any choice of α,λ>0, and it reduces to the classical Poisson distribution with mean λ when α=1. We shall denote the distribution with the PMF (17) by $$\mathcal {EP}(\alpha, \lambda)$$, which stands for Extended Poisson.

### Remark 6

This distribution is an example of a “power-series” type distribution (see, e.g., Noack 1950), where we used the power series of the Mittag-Leffler special function (9) with β=α and z=λα. A similar generalization of Poisson distribution, referred to as “hyper-Poisson” by the authors, was considered in Bardwell and Crow (1964), where the parameter α in (9) was taken to be α=1, leading to the PMF

$$g_{\beta, \lambda}(n) = \frac{\lambda^{n}}{\Gamma(n+\beta)} \frac{1}{E_{1, \beta}(\lambda)}, \,\,\, n\in \mathbb{N}_{0}.$$

This reduces to the Poisson PMF with parameter λ when β=1.

Let us note an important stochastic interpretation of an integer-valued random variable Y with PMF proportional to the product of two PMFs.

### Lemma 1

If we have that

$$f_{Y}(n) = (1/c) f_{N}(n) f_{Q}(n), \,\,\, n\in \mathbb{N}_{0},$$
(18)

where fY(·),fN(·), and fQ(·) are the PMFs of Y, N, and Q, respectively, then we have the representation

$$Y \stackrel{d}{=} N | N=Q,$$
(19)

where the variables N and Q are mutually independent.

In view of the above lemma, we get the following result regarding the conditional distribution of N given X=x.

### Proposition 6

Let (X,N) have a joint distribution with the PDF (7). Then the conditional distribution of N given X=x is the same as that of the random variable Y in (19), where N and Q are independent and Q−1 has extended Poisson distribution $$\mathcal {EP}(\alpha, \lambda)$$ with λ=βx, given by the PMF (17).

### Remark 7

We note that in the special case of the BGG model, where N is geometric (2), the PDF of N given X=x in (16) is proportional to [(1−p)1/αβx]αn/Γ(αn). We conclude that this conditional distribution is the same as that of 1+R, where $$R \sim \mathcal {EP}(\alpha, \lambda)$$ with λ=(1−p)1/αβx. In the special case of the BEG model, where we have α=1, the variable R is Poisson with mean λ=(1−p)βx, and N|X=x is R shifted up by 1.

The information about the conditional distribution of X given N=n provided above is useful in goodness-of-fit analysis for this bivariate model. Indeed, given data (Xi,Ni) from the model, one can isolate the pairs with a particular value of Ni=n and check the fit of the theoretical conditional distribution of X|N=n against the empirical distribution of the {Xi|Ni=n}. Unfortunately, one cannot proceed in a similar fashion using the conditional distribution of N given X=x, since the distribution of X is continuous. With this limitation in mind, we now derive the conditional distribution of N given X>x, which will be more useful in goodness-of-fit analysis.

Observe that the PMF of this conditional distribution can be written as

$$\mathbb{P}(N=n|X>x) = \frac{\mathbb{P}(N = n,X > x)}{\mathbb{P}(X >x)} = \frac{H(n)f_{N}(n)}{\mathbb{P}(X >x)},\,\,\, n\in \mathbb{N},$$
(20)

where

$$H(n) = \int_{x}^{\infty} \frac{\beta^{\alpha n}}{\Gamma(\alpha n)}t^{\alpha n - 1}e^{-\beta t}dt = \mathbb{P}(X_{1}+\cdots + X_{n}>x)$$
(21)

and the {Xi} are IID gamma $$\mathcal {GAM}(\alpha, \beta)$$ distributed random variables. This is because the integrand above is the PDF of the sum X1++Xn. Consider a renewal process {N(x), xR+} defined through the sequence {Xi}, so that

$$N(x)\geq n \,\,\, \text{if and only if} \,\,\, X_{1}+\cdots + X_{n} \leq x, \,\,\, n\in \mathbb{N}, x\in \mathbb{R}_{+}.$$

Then, the probability on the right-hand-side of (21) coincides with the probability $$\mathbb {P}(N(x)+1\leq n)$$, which in turn is the same as the H(n) in (20), with the latter taking the form

$$H(n) = \mathbb{P}(N(x) + 1 \leq n) = \frac{\Gamma(\alpha n, \beta x)}{\Gamma(\alpha n)}, \,\,\, n\in \mathbb{N}.$$
(22)

The function in the numerator on the right-hand-side in (22) is the (upper) incomplete gamma function,

$$\Gamma(\alpha, x)=\int\limits_{x}^{\infty}t^{\alpha -1}e^{-t}dt,\,\,\, \alpha, x>0.$$

Thus, the function H(n) above is a genuine CDF of a discrete random variable N(x)+1. In turn, in view of (20), the conditional PMF of N given X>x is proportional to the product of the CDF H(n) and the PMF fN(n). Distributions with such a structure have an important stochastic representation related to the hidden truncation (see, e.g., Arnold, 2009), given below.

### Lemma 2

If we have that

$$f_{Y}(n) = (1/c) f_{N}(n) F_{Q}(n), \,\,\, n\in \mathbb{N}_{0},$$

where fY(·) and fN(·) are the PMFs of Y and N, respectively, while FQ(·) is the CDF of a random variable Q, then we have the representation

$$Y \stackrel{d}{=} N | N\geq Q,$$
(23)

where the variables N and Q are mutually independent.

In view of the above lemma, we have the following result concerning the conditional distribution of N given X>x in our model.

### Proposition 7

Let (X,N) have a joint distribution with the PDF (7). Then the conditional distribution of N given X>x is the same as that of the random variable Y in (23) with N and Q independent and Q=dN(x)+1, where {N(x), xR+} is a renewal process defined through an IID gamma $$\mathcal {GAM}(\alpha,\beta)$$ sequence {Xi}.

We note that in the special case where α=1, so that the variables {Xi} in (1) are exponential with parameter β, the renewal process {N(x), xR+} is actually a Poisson process, with rate β. If in addition the variable N is geometrically distributed with the PMF (2), the distribution of (X,N) in (1) turns into the BEG model studied by Kozubowski and Panorska (2005). We have the following new result concerning the conditional distribution of N given X>x in this special case.

### Proposition 8

Let (X,N) have a joint distribution with the PDF (7) where α=1 and N is geometric, so that $$f_{N}(n) = p(1-p)^{n-1}, n\in \mathbb {N}$$. Then the conditional distribution of N given X>x is the same as that of the random variable Y=N+T, where T is Poisson with parameter (1−p)βx, independent of N.

The above results are useful in goodness-of-fit analysis when our bivariate distribution is fitted to data. Indeed, given data (Xi,Ni) from the model, one can isolate the pairs where Xi>x for a particular x>0, and check the fit of the theoretical conditional distribution of N|X>x against the empirical distribution of the {Ni} that are part of these pairs. One simple strategy is to compare certain characteristics, such as the mean, the mode, or the variance, of the theoretical and the empirical distributions across different values of x. For example, in the BEG model, by Proposition 8, the conditional expectation

$$\mathbb{E} (N|X>x) = \mathbb{E} N +\mathbb{E} T = 1/p + (1-p)\beta x$$

is a simple linear function of x, which can be plotted against x and compared with the corresponding plot of the sample mean of N|X>x for the same x. Then the goodness-of-fit of the BEG model may be judged by comparing the two plots.

We finally note that the conditional distribution of N given X>x studied above is one of the two marginal distributions in the bivariate model (X,N)|X>x. The full joint distribution of this model can be described by the joint PDF, which can be shown to be of the form

$$g(t,n) = g(t|n)f_{Y}(n), \,\,\, n\in \mathbb{N},t\in \mathbb{R}_{+},$$
(24)

where fY(·) is the PMF (20) of the distribution of N|X>x and

$$g(t|n) =\frac{\beta^{\alpha n}}{\Gamma(\alpha n)} t^{\alpha n-1}e^{-\beta t}I_{(x,\infty)}(t)/H(n), \,\,\, t\in \mathbb{R}_{+},$$
(25)

with H(n) as in (21), is the PDF of gamma distribution $$\mathcal {GAM}(\alpha n, \beta)$$ truncated below at x. The joint CDF of this distribution can be written in terms of (upper) incomplete gamma function as follows:

$$\begin{array}{*{20}l} G(y,n)&=\mathbb{P}(X\leq y, N\leq n|X>x)\\ &= \frac{ \sum_{j=1}^{n} \left\{ [\Gamma(\alpha j, \beta x) - \Gamma(\alpha j, \beta y)]/\Gamma(\alpha j)\right\} f_{N}(j)} { \sum_{j=1}^{\infty} \left\{ \Gamma(\alpha j, \beta x)/\Gamma(\alpha j)\right\} f_{N}(j)}, \,\,\, x\leq y, \, n\in \mathbb{N}. \end{array}$$

Observe that, in view of (24), while the PDF of this distribution admits a hierarchical structure in the spirit of (7), this conditional distribution generally does not have a stochastic representation analogous to (1). On the other hand, the bivariate distribution describing (X,N)|N>u, where $$u\in \mathbb {N}_{0}$$, can actually be related to (1). Indeed, it is easy to see that the joint PDF of this distribution is of the form

$$f_{X,\tilde{N}}(x,n) = \frac{\beta^{\alpha n}}{\Gamma(\alpha n)}x^{\alpha n - 1}e^{-\beta x} f_{\tilde{N}}(n), \,\,\, x\in \mathbb{R}_{+}, n\in \mathbb{N},$$
(26)

where

$$f_{\tilde{N}}(n) = \left\{ \begin{array}{ll} f_{N}(n)/\mathbb{P}(N>u) & \text{for $$n=u+1, u+2, \ldots$$}\\ 0 & \text{for $$n \leq u$$} \end{array} \right.$$
(27)

is the PMF of $$\tilde {N} \stackrel {d}{=} N | N>u$$. Thus, this conditional distribution admits the representation (1) with N replaced by $$\tilde {N}$$. For any $$x\in \mathbb {R}_{+}$$ and nu, the joint survival function of this distribution takes on the form

$$S_{X,\tilde{N}}(x,n) = \mathbb{P}(X> x,N> n|N>u) = \sum\limits_{k=n+1}^{\infty} \left(1 - \frac{\gamma(\alpha k, \beta x)}{\Gamma(\alpha k)}\right) f_{\tilde{N}}(k)$$
(28)

with $$f_{\tilde {N}}(\cdot)$$ provided in (27). We note the following stochastic representation of this distribution involving the excess random variable Nu=dNu|N>u, defined for any $$u\in \mathbb {N}_{0}$$, which, like N, is also supported on the set of positive integers $$\mathbb {N}$$.

### Proposition 9

Let (X,N) have a joint distribution with the PDF (7) and let Nu=dNu|N>u where $$u \in \mathbb {N}_{0}$$. Then we have the following stochastic representation:

$$(X,N)|N>u \stackrel{d}{=} \left(\sum_{j=1}^{u} X_{1j}, u \right) + \left(\sum_{j=1}^{N_{u}}X_{2j}, N_{u} \right),$$

where the {Xij} are IID gamma $$\mathcal {GAM}(\alpha,\beta)$$ distributed random variables, independent of Nu.

Thus, if the support of N is the set of positive integers, then the conditional distribution of (X,N)|N>u can be expressed as a convolution of two distributions with similar structure, one of which is degenerate and the other is also connected with a discrete distribution supported on $$\mathbb {N}$$. In particular, if N is geometrically distributed with the PMF (2), then due to its memoryless property, we have that Nu=dN for each $$u\in \mathbb {N}_{0}$$, so that

$$(X,N)|N>u \stackrel{d}{=} (G_{u},u) + (X,N),$$
(29)

where Gu is a gamma $$\mathcal {GAM}(\alpha u,\beta)$$ distributed random variable, independent of (X,N).

We finish this section with a brief account for two other related conditional distributions, one of (X,N)|Xx, and another of (X,N)|Nu. The following result, which is straightforward to derive, can be found in Amponsah (2017) (see equations (2.36) and (2.38) therein) in its special case of discrete Pareto random variable N.

### Proposition 10

Let $$u,n\in \mathbb {N}$$ and x>0, with nu. Then the conditional CDF of (X,N) given Nu is given by

$$\mathbb{P}(X\leq x, N\leq n |N\leq u) = \frac{\sum\limits_{j=1}^{n}\frac{\gamma(\alpha j, \beta x)}{\Gamma(\alpha j)}f_{N}(j)}{\mathbb{P}(N\leq u)}.$$
(30)

Similarly, for 0<yx and $$n\in \mathbb {N}$$, the conditional CDF of (X,N) given Xx is given by

$$\mathbb{P}(X\leq y, N\leq n |X\leq x) = \frac{\sum\limits_{j=1}^{n}\frac{\gamma(\alpha j, \beta y)}{\Gamma(\alpha j)}f_{N}(j)}{\sum\limits_{j=1}^{\infty}\frac{\gamma(\alpha j, \beta x)}{\Gamma(\alpha j)}f_{N}(j)}.$$
(31)

### 2.3 Divisibility properties

It was shown in Barreto-Souza (2012) and Kozubowski and Panorska (2005) that the BGG and BEG models are infinitely divisible (ID). Below, we show that this is a general property of this model, as long as the ID property holds for N as well as the underlying IID sequence {Xi}.

Recall that a random vector X (and its probability distribution) is said to be ID if for each $$n\in \mathbb {N}$$ the vector X can be “decomposed” into the sum

$$\mathbf{X} \stackrel{d}{=} \mathbf{X}_{1}^{(n)} + \cdots + \mathbf{X}_{n}^{(n)},$$

where the random vectors $$\{ \mathbf {X}_{i}^{(n)}\}$$ are IID. We assume that the non-negative variables {Xi} in (1) are ID, so that $$X_{1}\stackrel {d}{=}Y_{1}^{(n)} + \cdots + Y_{n}^{(n)}$$ for some IID sequence $$\{Y_{i}^{(n)}\}$$ for each $$n\in \mathbb {N}$$. In this situation, the variables $$Y_{i}^{(n)}$$ are also non-negative and their common LT is given by $$\phantom {\dot {i}\!}\psi _{n}(t) = [\psi _{X_{1}}(t)]^{1/n}$$, where $$\phantom {\dot {i}\!}\psi _{X_{1}}(\cdot)$$ is the LT of the {Xi}. We also assume that the variable N−1 is discrete ID, so that the relation $$N-1 \stackrel {d}{=}N_{1}^{(n)} + \cdots + N_{n}^{(n)}$$ holds for each $$n\in \mathbb {N}$$ with IID variables $$\{N_{i}^{(n)}\}$$ whose distribution is supported on the set $$\mathbb {N}_{0}$$ of non-negative integers. This implies that N itself is ID as well, and can be decomposed into the sum of n IID terms of the form $$1/n + N_{i}^{(n)}, i=1, \ldots n$$. Moreover, the common PGF of the $$\{N_{i}^{(n)} \}$$ is of the form Gn(s)=[GN(s)/s]1/n, where GN(·) is the PGF of N. Under this notation, we have the following result.

### Proposition 11

Let N be an integer-valued random variable supported on $$\mathbb {N}$$, given by the PGF GN(·). Further, suppose that (X,N)is defined viz. (1), where the {Xi} are nonnegative, IID, and ID random variables, given by LT $$\psi _{X_{1}}(\cdot)$$, independent of N. Suppose also that the distribution of N−1 is discrete ID. Then, the distribution of (X,N) is ID and for each $$n\in \mathbb {N}$$ we have the stochastic representation

$$(X,N) \stackrel{d}{=}\sum_{j=1}^{n}(R_{j}^{(n)},V_{j}^{(n)}),$$
(32)

where the $$\{ (R_{j}^{(n)},V_{j}^{(n)})\}$$ are IID. Moreover, we have $$V_{j}^{(n)}\stackrel {d}{=}1/n+T_{j}^{(n)}$$, where the $$\{ T_{j}^{(n)}\}$$ are IID, integer-valued random variables with the PGF Gn(s)=[GN(s)/s]1/n, and

$$R_{j}^{(n)} = \sum_{j=1}^{T_{j}^{(n)}} X_{ij} + G_{j}^{(n)},$$
(33)

where all the variables on the right-hand-side of (33) are mutually independent, the $$\{G_{j}^{(n)}\}$$ are IID non-negative variables given by the LT $$\phantom {\dot {i}\!}\psi _{n}(t) = [\psi _{X_{1}}(t)]^{1/n}$$, and the {Xij} are IID with the LT $$\phantom {\dot {i}\!}\psi _{X_{1}}(\cdot)$$.

When we restrict the above result to the case where the variables {Xi} are gamma $$\mathcal {GAM}(\alpha, \beta)$$ distributed (and thus ID), we obtain the following generalizations of the results in Barreto-Souza (2012).

### Corollary 1

Let N be an integer-valued random variable supported on $$\mathbb {N}$$, given by the PGF GN(·). Further, suppose that (X,N) is defined viz. (1), where the {Xi} are IID gamma $$\mathcal {GAM}(\alpha, \beta)$$ distributed, and independent of N. Suppose also that the distribution of N−1 is discrete ID. Then, the distribution of (X,N) is ID and for each $$n\in \mathbb {N}$$ we have the stochastic representation (32) where the $$\{ (R_{j}^{(n)},V_{j}^{(n)})\}$$ are IID. Moreover, we have $$V_{j}^{(n)}\stackrel {d}{=}1/n+T_{j}^{(n)}$$, where the $$\{ T_{j}^{(n)}\}$$ are IID, integer-valued random variables with the PGF Gn(s)=[GN(s)/s]1/n, and the $$\{R_{j}^{(n)}\}$$ are given by (33), where all the variables on the right-hand-side of (33) are mutually independent, the $$\{G_{j}^{(n)}\}$$ are IID gamma $$\mathcal {GAM}(\alpha /n, \beta)$$ distributed, and the {Xij} are IID gamma $$\mathcal {GAM}(\alpha, \beta)$$ distributed.

### Remark 8

In Kozubowski and Panorska (2005) and Barreto-Souza (2012), the variable N was geometric (2) while the {Xi} were exponentially or gamma distributed, respectively. In this case, the variables $$T_{j}^{(n)}$$ in (33) have negative binomial $$\mathcal {NB}(r,p)$$distribution with r=1/n, whose PDF and PGF are given by

$$f(n) = \binom{r+n-1}{n}p^{r}(1-p)^{n}, \,\,\, n\in \mathbb{N}_{0},$$

and

$$G(s) = \left(\frac{p}{1-(1-p)s}\right)^{r}, \,\,\, s\in [0,1],$$

respectively. We note that in the gamma case the LT of the $$\{(R_{j}^{(n)},V_{j}^{(n)})\}$$ in (32) is of the form [ψ(t,s)]1/n, where ψ(t,s) is the LT of (X,N), given by (12). More generally, the ID property of the distribution of (X,N) established above implies that [ψ(t,s)]r is a genuine LT for any $$r\in \mathbb {R}_{+}$$, and describes the marginal distribution of a bivariate Lévy motion $$\{(R(r), V(r)), \, r\in \mathbb {R}_{+}\}$$, which is a stochastic process with independent and stationary increments started at the origin, built upon the distribution of (X,N). Such processes connected with the BEG and BGG models were studied in Kozubowski et al. (2008) and Barreto-Souza (2012), respectively. Their analogs in the more general case described above can be studied along the same lines.

### 2.4 Parameter estimation

In this section we consider maximum likelihood estimation of the parameters of the model (1) where the {Xi} are IID with gamma $$\mathcal {GAM}(\alpha, \beta)$$ distribution. We shall assume that the distribution of N depends on k unknown parameters ω1,…,ωk, where $$\boldsymbol {\omega } = (\omega _{1}, \ldots, \omega _{k})\in \Omega \subset \mathbb {R}^{k}$$. Under this notation, our model has k+2 unknown parameters θi,i=1,…,k+2, where θ1=α,θ2=β, and θi=ωi−2 for i=3,…,k+2, with $$\boldsymbol {\theta } = (\theta _{1}, \ldots, \theta _{k+2})\in \Theta = \mathbb {R}_{+}^{2} \times \Omega \subset \mathbb {R}^{k+2}$$. Under this notation, the PMF of the random variable N will be denoted by fN(·;ω).

Let us start with the Fisher information matrix I(θ) corresponding to this bivariate model. Note that the logarithm of the joint PDF in (7),

$$\log f_{X,N}(x,n) = \alpha n \log \beta + (\alpha n-1)\log x - \beta x - \log \Gamma (\alpha n) + \log f_{N}(n; \boldsymbol{\omega}),$$

is the sum of two terms,

$$v_{1}(x,n; \alpha, \beta) = \alpha n \log \beta + (\alpha n-1)\log x - \beta x - \log \Gamma (\alpha n)\,\,\, \text{and} \,\,\, v_{2}(n; \boldsymbol{\omega}) = \log f_{N}(n; \boldsymbol{\omega}),$$

with each of them depending on different parameters. Consequently, if standard regularity conditions for the family {fN(·;ω), ωΩ} are satisfied and appropriate expectations exist, then the Fisher information I(θ) will have a block-diagonal structure,

$$I(\boldsymbol{\theta}) = \left[ \begin{array}{cc} \bf W & \boldsymbol{0}\\ \boldsymbol{0} & \bf J(\boldsymbol{\omega}) \end{array} \right],$$
(34)

where W=[wij] is a 2 by 2 matrix where

$$w_{ij} = - \mathbb{E} \left[\frac{\partial^{2}}{\partial \theta_{i} \partial \theta_{j}} v_{1}(X,N; \theta_{1}, \theta_{2})\right], \,\,\, i,j=1,2,$$
(35)

and J(ω) is the Fisher information matrix connected with the family {fN(·;ω), ωΩ}. Straightforward algebra shows that the entries of W are as follows:

$$w_{11} = \mathbb{E} [ N^{2}\psi'(\alpha N)], \,\,\, w_{12}=w_{21} = - \mathbb{E}[N]/\beta, \,\,\, w_{22} = \alpha \mathbb{E}[N]/\beta^{2},$$
(36)

where ψ(x)=d/dx logΓ(x) is the digamma function.

### Remark 9

In the special case of the BGG model studied in Barreto-Souza (2012) the variable N is geometric (2) so that k=1 and ω1=pΩ=(0,1). Here, J(ω) in (34) is a 1 by 1 matrix [1/(p2(1−p))] and $$\mathbb {E}[N] = 1/p$$, so that by (36) we have

$$I(\boldsymbol{\theta}) = \frac{1}{p} \left[ \begin{array}{ccc} w(\alpha,p) & \frac{-1}{\beta} & 0\\ \frac{-1}{\beta} & \frac{\alpha}{\beta^{2}} & 0\\ 0 & 0 & \frac{1}{p} \frac{1}{1-p} \end{array} \right],$$

where $$w(\alpha,p) = p^{2}\sum _{n=1}^{\infty } n^{2} (1-p)^{n-1}\psi '(\alpha n)$$ (cf., Barreto-Souza 2012, eq. (15), p. 135).

Next, we let (X1,N1),…,(Xn,Nn) be a random sample from the bivariate model with the PDF (7). By straightforward calculations, we find the log-likelihood function to be of the form

$$l(\boldsymbol{\theta}) =R(\alpha, \beta) + Q(\boldsymbol{\omega}) - \sum\limits_{i=1}^{n} \log(X_{i}),$$
(37)

where

$$R(\alpha, \beta) = n\alpha\bar{N}_{n}\log(\beta) - n\beta\bar{X}_{n} + \alpha \sum\limits_{i=1}^{n}N_{i} \log(X_{i}) - \frac{1}{n} \sum\limits_{i=1}^{n} \log\left[ \Gamma(\alpha N_{i})\right],$$
(38)
$$Q(\boldsymbol{\omega}) = \sum\limits_{i=1}^{n}\log \left[f_{N}(N_{i}; \boldsymbol{\omega}) \right],$$
(39)

and $$\bar {N}_{n}, \bar {X}_{n}$$ are the sample means of the {Ni} and {Xi}, respectively. In order to find the maximum likelihood estimator (MLE) of θ, we need to maximize the log-likelihood function (37) with respect to $$\boldsymbol {\theta } \in \Theta = \mathbb {R}_{+}^{2} \times \Omega \subset \mathbb {R}^{k+2}$$. Since θ=(α,β,ω), by the special structure of the log-likelihood function (37), we can proceed by maximizing the functions R(α,β) and Q(ω) with respect to their arguments independently of each other. Consequently, the MLE of ω is only dependent on the {Ni}, and is connected with the family {fN(·;ω), ωΩ}. Additionally, the MLEs of α and β will be of the same form regardless of the particular distribution of N, although their asymptotic properties may depend on the latter.

#### 2.4.1 The MLEs of α and β

Our focus are the MLEs of α and β. By the above discussion, to find them one needs to maximize the function R(α,β) in (38) with respect to $$(\alpha, \beta) \in \mathbb {R}_{+}^{2}$$. To do this, we first maximize the function h(β)=R(α,β) with respect to $$\beta \in \mathbb {R}_{+}$$ when α>0 is held fixed. Since h(β) is a continuous, differentiable function of β with the derivative of the form

$$h'(\beta) =\frac{n\alpha \bar{N}_{n}}{\beta}-n\bar{X}_{n},\,\,\, \beta \in \mathbb{R}_{+},$$

it is easy to see that h(β)>0 iff β<β(α) and h(β)<0 iff β>β(α), where

$$\beta(\alpha)=\frac{\alpha\bar{N}_{n}}{\bar{X}_{n}}.$$
(40)

Thus, the function h(β) has a unique maximum value on (0,), which is attained at β=β(α) given by (40). Next, we optimize the function R(α,β(α)) with respect to $$\alpha \in \mathbb {R}_{+}$$. Straightforward algebra shows that $$R(\alpha, \beta (\alpha)) = n\bar {N}_{n}g(\alpha)$$, where for $$\alpha \in \mathbb {R}_{+}$$ we have

$$g(\alpha)=\alpha \log(\alpha)+\alpha\log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right)-\alpha+\frac{\alpha}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}N_{i} \log(X_{i})- \frac{1}{n\bar{N}_{n}} \sum\limits_{i=1}^{n} \log\left[ \Gamma(\alpha N_{i})\right].$$

This is a continuous, differentiable function on (0,), with the derivative of the form

$$\begin{array}{@{}rcl@{}} g'(\alpha)=&\log(\alpha) -\frac{1}{n\bar{N}_{n}} \sum\limits_{i=1}^{n} N_{i}\psi\left[ \alpha N_{i}\right] + \log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) + \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}N_{i} \log(X_{i}), \end{array}$$
(41)

where ψ(x)=d/dx logΓ(x) is the digamma function. The following result provides key properties of the above function

### Proposition 12

Let (X1,N1),…,(Xn,Nn) be IID copies of (X,N) in (1). Then, the function (41) has the following properties:

• g(α) is continuous and monotonically decreasing in α(0,);

• $${\lim }_{\alpha \rightarrow 0^{+}}g'(\alpha)=\infty$$;

• $${\lim }_{\alpha \rightarrow \infty }g'(\alpha)= \frac {1}{n\bar {N}_{n}}\sum \limits _{i=1}^{n} N_{i} \left [\log \left (\frac {X_{i}}{N_{i}}\right) \right ]-\log \left (\frac {\bar {X}_{n}}{\bar {N}_{n}}\right)\leq 0$$;

• For $$n=1, {\lim }_{\alpha \rightarrow \infty }g'(\alpha) = 0$$;

• For $$n\geq 2, {\lim }_{\alpha \rightarrow \infty }g'(\alpha) < 0$$ almost surely.

In view of the above result, we conclude that whenever the limit in Part (iii) is negative, which occurs with probability 1 for any finite sample of size n≥2, the function g(α) in monotonically increasing on the interval $$(0,\hat {\alpha }_{n})$$ and monotonically decreasing on the interval $$(\hat {\alpha }_{n}, \infty)$$, where $$\hat {\alpha }_{n}$$ is a unique solution for α of the equation g(α)=0. This establishes the existence and uniqueness of the MLEs of α and β for this model.

### Corollary 2

Let (X1,N1),…,(Xn,Nn) be IID copies of (X,N)in (1), where n≥2. Then, the MLEs of αand β exist and are unique, where the MLE of $$\alpha, \hat {\alpha }_{n}$$, solves the equation

$$\log(\hat{\alpha}_{n})-\frac{1}{n\bar{N}_{n}} \sum\limits_{i=1}^{n} N_{i}\psi\left[\hat{\alpha}_{n} N_{i}\right] + \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}N_{i} \log(X_{i}) +\log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) = 0$$
(42)

while the MLE of β is given by $$\hat {\beta }_{n} = \hat {\alpha }_{n} \bar {N}_{n}/ \bar {X}_{n}$$.

### Remark 10

When the parameter α is known, the MLE of β always exists, is unique, and its value is provided by (40). In particular this recovers the results of the BEG model of Kozubowski and Panorska (2005), where we have α=1.

### Remark 11

Since the function in (42) is continuous and monotonic, the solution of this equation is straightforward to compute using standard numerical methods.

### Remark 12

In the special case when the sample size is n=1, the equation (42) does not admit a finite solution. However, the equation is satisfied in the limiting sense, with $$\hat {\alpha }_{n} = \infty$$. What the above analysis shows is that the function R(α,β) in (38) is maximized “at infinity”, when α and β=β(α) is given by (40). A practical interpretation of this is that the gamma variables that drive the model of (X,N) in (1) reduce to a point mass at c, where c=X1/N1, so that the distribution of (X,N) is a degenerate one, with (X,N)=(cN,N). Similar interpretation applies when n≥2 and the limit in Part (iii) in Proposition 12 is zero, in which case the solution of (42) is also $$\hat {\alpha }_{n} = \infty$$. As can be seen from the proof of this result, this can only happen if all the ratios Xi/Ni in the sample are equal to some c>0. However, this degenerate case is not of a concern in practice, as this event occurs with probability zero.

### Remark 13

Compared with the results of Barreto-Souza (2012), who considered a special case in (1) with geometrically distributed N (see eq. (14) therein), our results are stronger and more general, as they ensure the existence and uniqueness of the MLEs of α and β in any model of the form (1), going beyond geometrically distributed N.

### Remark 14

If the family of distributions {fN(·;ω), ωΩ} is such that the function Ψ=(ψ1,…,ψk+2) with ψi=−/θi logfX,N(x,n),i=1,…,k+2, satisfies standard regularity conditions of large-sample theory (see, e.g., conditions A0-A6 in Bickel and Doksum (2015), pp. 384–385) then the MLEs of α and β are consistent, asymptotically normal, and efficient (see, e.g., Theorem 6.2.2 in Bickel and Doksum (2015), p. 386). In particular, due to the special structure of the Fisher information matrix (34), we will have the following convergence in distribution

$$\sqrt{n}[(\hat{\alpha}_{n}, \hat{\beta}_{n}) - (\alpha, \beta)]\stackrel{d}{\rightarrow} \mathbf{Z} \sim N_{2}(\mathbf{0}, {\boldsymbol{W}}^{-1})\,\,\, \text{as $$n\rightarrow \infty$$},$$
(43)

where Nd(μ,Σ) denotes d-variate normal distribution with mean vector μ and covariance matrix Σ and W is the 2 by 2 matrix with entries given by (36). This asymptotic normal distribution can be used to derive (approximate) confidence intervals for the parameters.

## Bivariate episodes with discrete Pareto duration

Here, we illustrate the general results above with a special case studied in Amponsah (2017), where the random variable N in (1) has discrete Pareto (DP) distribution with parameters δ≥0 and p(0,1), given by the PMF (6) and denoted by $$\mathcal {DP}(\delta, p)$$. Virtually all the results discussed in Section 2 carry-over to this special case, where we use the PMF (6) in place of a generic PMF fN(·). First, we provide a formal definition of this particular stochastic model.

### Definition 1

A random vector (X,N) with the stochastic representation (1), where the {Xi} are IID gamma variables with the PDF (4) and N is DP with the PMF (6), independent of the {Xi}, is said to have a GMDP distribution with parameters α>0,β>0,δ≥0, and p(0,1), denoted by $$\mathcal {GMDP}(\alpha, \beta,\delta, p)$$. The term “GMDP” stands for distribution with $$\mathcal {\mathbf {G}}$$amma-$$\mathcal {\mathbf {M}}$$ixture and $$\mathcal {\mathbf {D}}$$iscrete $$\mathcal {\mathbf {P}}$$areto marginals.

### Remark 15

In the special case with δ=0, the DP distribution turns into a geometric distribution (2), and the GMDP distribution reduces to the BGG model proposed by Barreto-Souza (2012). If in addition to δ=0 we also have α=1, so that the {Xi} in (1) are IID exponential with parameter β>0, the GMDP model turns into the BEG distribution studied in Kozubowski and Panorska (2005). The GMDP model offers more flexibility, as it has an additional parameter, and can be useful for modeling heavy tailed data, as its marginal distributions of X and N are both heavy tailed for any δ>0, as will be seen below.

The joint PDF of $$(X,N)\sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$ is given explicitly by (7), with the DP PMF (6) in place of fN(·). In turn, the CDF and the SF are given by the expressions in Proposition 1. Further, by (8), we obtain the PDF of X for $$x\in \mathbb {R}_{+}$$,

\begin{aligned} f_{X}(x) = \sum\limits_{n=1}^{\infty} \frac{\beta^{n\alpha}x^{n\alpha-1}}{e^{\beta x}\Gamma(n\alpha)}\left[ \left(\frac{1}{1-\delta(n-1)\log(1-p)}\right)^{\frac{1}{\delta}}- \left(\frac{1}{1-\delta n\log(1-p)}\right)^{\frac{1}{\delta}}\right], \end{aligned}
(44)

which takes a particularly simple form in case of the BEG model (δ=0,α=1), where X is exponential with parameter βp. We also note that (for $$x\in \mathbb {R}_{+}$$) the CDF of X is of the form

\begin{aligned} F_{X}(x) &= \mathbb{P}(X\leq x)\\ & =\sum\limits_{n=1}^{\infty}\frac{\gamma(\alpha n, \beta x)}{\Gamma(\alpha n)} \left[ \left(\frac{1}{1-\delta(n-1)\log(1-p)}\right)^{\frac{1}{\delta}}- \left(\frac{1}{1-\delta n\log(1-p)}\right)^{\frac{1}{\delta}}\right]. \end{aligned}

### 3.1 Laplace transform

The LT of the GMDP model follows from general theory, discussed in Proposition 2 and the remark following it. When we take into account the particular form of the PGF connected with DP distribution (see Proposition 2.6 in Buddana and Kozubowski 2014), we obtain the following result.

### Proposition 13

The LT (11) of $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$is of the form

$$\psi(t, s) = \left(e^{-s}\psi_{X_{1}}(t)\right) - \left[1 - e^{-s}\psi_{X_{1}}(t)\right] \sum\limits_{n=1}^{\infty}\frac{\left[ e^{-s}\psi_{X_{1}}(t)\right]^{n}}{[1-\delta n\log(1-p)]^{1/\delta}}, \,\,\, t,s \in \mathbb{R}_{+},$$

where $$\psi _{X_{1}}(t)=[\beta /(\beta + t)]^{\alpha }, t \in \mathbb {R}_{+}$$, is the LT of gamma $$\mathcal {GAM}(\alpha, \beta)$$ distribution.

### 3.2 Moments and tail behavior

Recall that the SF of $$N\sim \mathcal {DP}(\delta, p)$$ is of the form (see Buddana and Kozubowski 2014, Proposition 2.4)

$$S(x) = \mathbb{P}(N > x) = \left(\frac{1}{1-\delta\log(1-p)\lfloor x\rfloor}\right)^{\frac{1}{\delta}}, \,\,\, x\geq 1,$$

where x is the floor function (the largest integer that is less than or equal to x). This shows that DP distribution is heavy tailed with tail index v=1/δ, that is the SF satisfies (13) with v=1/δ. Further the constant c in (13) is equal to c=[−δ log(1−p)]−1/δ. Thus, by Proposition 3, we have the following result.

### Proposition 14

If $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$ with δ>0 then X is heavy tailed with index 1/δ, and we have

$$S_{X}(x) = \mathbb{P}(X>x) \sim [-\delta \log (1-p)\beta/\alpha]^{-1/\delta} x^{-1/\delta} \,\,\, (x\rightarrow \infty).$$

In turn, by Proposition 4, the joint moments $$\mathbb {E}[X^{r}N^{\eta }]$$ with positive r,η>0 exist only for r+η<1/δ, and their values are provided in (14) with fN(·) replaced by the PMF (6) of DP distribution. Further, we also have an analog of Proposition 5, concerning the mean vector and the covariance matrix of $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$. To formulate the result, we need the mean and the variance of $$N\sim \mathcal {DP}(\delta, p)$$, which can be deduced from Proposition 2.7 in Buddana and Kozubowski (2014). Namely, the mean of N exists whenever δ<1, in which case we have

$$\mathbb{E}[N] = \sum_{n=0}^{\infty}\left(\frac{1}{1-\delta \log(1-p) n}\right)^{\frac{1}{\delta}} = C \zeta\left(\frac{1}{\delta}, - \frac{1}{-\delta \log(1-p)}\right),$$
(45)

where C=[−δ log(1−p)]−1/δ and

$$\zeta(s,q) = \sum_{n=0}^{\infty}\left(\frac{1}{n+q}\right)^{s}$$

is the Hurwitz-zeta special function. Similarly, the variance of N exists whenever δ<1/2, in which case we have

$$\mathbb{V} ar[N] = \sum\limits_{n=1}^{\infty}\left\lbrace (2n+1)\left[ \frac{1}{1-\delta n \log(1-p)}\right]^{\frac{1}{\delta}} \right\rbrace - \left[\sum\limits_{n=0}^{\infty} \left(\frac{1}{1-\delta n \log(1-p)}\right)^{\frac{1}{\delta}}\right]^{2}.$$
(46)

### Proposition 15

Let $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$. The mean vector μof $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$ exists only if δ<1, in which case we have $$\boldsymbol {\mu } = (\alpha /\beta, 1) \mathbb {E}[N]$$ with $$\mathbb {E}[N]$$ given by (45). Moreover, the covariance matrix Σ of (X,N) exists only if δ<1/2, in which case we have

$$\boldsymbol{\Sigma} = \mathbb{E}[N] \left[ \begin{array}{cc} \alpha/\beta^{2} & 0 \\ 0 & 0 \end{array} \right] + \mathbb{V} ar[N] \left[ \begin{array}{cc} \alpha^{2}/\beta^{2} & \alpha/\beta \\ \alpha/\beta & 1 \end{array} \right],$$

where $$\mathbb {E}[N]$$ and $$\mathbb {V} an[N]$$ are given by (45) and (46), respectively.

### 3.3 Conditional distributions

The conditional distributions of $$(X, N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$ follow general theory laid out in Section 2. First, we note that the conditional distribution of X given N=n is gamma $$\mathcal {GAM}(\alpha n, \beta)$$, with the PDF in (15). This conditional distribution is the same in any model of the form (1) driven by an IID gamma sequence {Xi}, regardless of a particular distribution of N. Second, the conditional distribution of N given X=x has the PMF given by (16) with fN(·) replaced by the DP PMF (6), the function fα,λ(·) given by (17), and the normalizing constant c is given by

{}\begin{aligned} c = \frac{(\beta x)^{-\alpha}}{E_{\alpha,\alpha}[(\beta x)^{\alpha}]} \sum_{n=1}^{\infty} \frac{(\beta x)^{\alpha n}}{\Gamma(\alpha n)} \left[ \left(\frac{1}{1-\delta(n-1)\log(1-p)}\right)^{\frac{1}{\delta}}- \left(\frac{1}{1-\delta n \log(1-p)}\right)^{\frac{1}{\delta}}\right]. \end{aligned}

Moreover, according to Proposition 6, this conditional distribution is the same as that of the random variable N|N=Q, where N and Q are independent and Q−1 has extended Poisson distribution $$\mathcal {EP}(\alpha, \lambda)$$ with λ=βx, given by the PMF (17). Further, the conditional distribution of N|X>x has the PMF of the form (20), and, in view of Proposition 7, is the same as the distribution of N|NQ with independent N and Q, where this time Q=dN(x)+1 and {N(x), xR+} is a renewal process defined through an IID gamma $$\mathcal {GAM}(\alpha,\beta)$$ sequence {Xi}. The latter distribution is one of the two marginal distributions in the bivariate model (X,N)|X>x. The full joint distribution of this conditional model has the PDF of the form (24) where fY(·) is the PMF (20) of the distribution of N|X>x and g(·|n) is the PDF of gamma distribution $$\mathcal {GAM}(\alpha n, \beta)$$ truncated below at x, with the PDF given by (25). As discussed in Section 2, the two conditional distributions, one of X|N=n and one for N|X>x, provide useful aids in goodness-of-fit analysis when the GMDP model is fitted to data.

We now turn to the conditional distribution of (X,N)|N>u, where $$u\in \mathbb {N}_{0}$$. As in the general case discussed in Section 2, this conditional distribution admits representation (1) with N replaced by $$\tilde {N}$$, where the latter has a truncated DP distribution, with the PMF given by (27). The PDF and the SF of this distribution are given by (26) and (28), respectively. This case is very special, since the corresponding excess random variable, Nu=dNu|N>u is also DP distributed (see Proposition 3.3 in Buddana and Kozubowski 2014). In view of this result, we have the following analog of Proposition 9.

### Proposition 16

If $$(X,N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$ then we have

$$(X,N)|N>u \stackrel{d}{=} (G_{u},u) + (X_{u},N_{u}),$$
(47)

where Gu and (Xu,Nu) are independent, $$G_{u} \sim \mathcal {GAM}(\alpha u,\beta)$$, and $$(X_{u},N_{u}) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p_{u})$$ with

$$p_{u} = 1- e^{\frac{\log (1-p)}{1- u \delta \log(1-p)}}.$$

### Remark 16

By the above result, if (X,N) has GMDP distribution then the conditional distribution of (X,N)|N>u is a convolution of two distributions, of which one is degenerate and the other is also a GMDP distribution, with different “size” parameter p. In the special case when δ=0 (so that N is geometrically distributed) the variable (Xu,Nu) in (47) has actually the same distribution as (X,N) itself, due to the memoryless property of the geometric distribution.

Additional two conditional distributions, (X,N)|Xx and (X,N)|Nu, are described in the next result which follows from Proposition 10.

### Proposition 17

Suppose that $$(X,N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$.

(i) Let $$u,n\in \mathbb {N}$$ and x>0, with nu. Then the conditional CDF of $$(X,N) \sim \mathcal {GMDP}(\alpha,\beta, \delta, p)$$given Nu is given by (30), where fN(·) is the DP PMF (6) and

$$\mathbb{P}(N\leq u) = 1-\left(\frac{1}{1-\delta n\log(1-p)}\right)^{\frac{1}{\delta}}.$$

(ii) Similarly, for 0<yx and $$n\in \mathbb {N}$$, the conditional CDF of (X,N)given Xx is given by (31), where again fN(·) is given by (6).

### 3.4 Infinite divisibility and stochastic representations

In this section we discuss an important stochastic representation of the GMDP distribution as well as its infinite divisibility property. Both of these are related to the fact that the DP distribution is a mixture of geometric distributions. Indeed (see, e.g., shown in Buddana and Kozubowski 2014, Proposition 3.4), the DP distribution of $$N\sim \mathcal {DP}(\delta,p)$$ admits the following hierarchical structure:

• $$Z\sim \mathcal {GAM}(1/\delta, \gamma)$$, where γ=[−δ log(1−p)]−1,

• N|Z=z has a geometric distribution (2) with parameter p=1−ez.

Since a DP distributed random variable N is an important component of the GMDP distribution, a similar hierarchical structure carries over to the latter, as shown in the result below.

### Proposition 18

The distribution of (X,N)GMDP(α,β,δ,p) admits the following hierarchical structure:

• $$Z\sim \mathcal {GAM}(1/\delta, \gamma)$$, where γ=[−δ log(1−p)]−1,

• (X,N)|Z=zGMDP(α,β,0,q) with q=1−ez.

### Remark 17

According to the above result, a GMDP distribution can be thought of as being conditionally a BGG distribution, which is a GMBD distribution with δ=0, so that the variable N in the stochastic representation is a geometric one. In other words, this N is conditionally geometric, so that its unconditional distribution is DP, as seen above, which is precisely why we recover the GMDP model. Another way to describe this is through the density functions. Namely, given a gamma $$\mathcal {GAM}(1/\delta, \gamma)$$ distributed Z in Proposition 18, with density given by

$$f_{Z}(z) = \frac{[ - \delta \log(1-p)]^{-1/\delta}}{\Gamma (1/\delta)} z^{1/\delta-1} e^{z/[\delta \log(1-p)]},\,\,\, z\in \mathbb{R}_{+},$$

the PMF of (X,N)|Z=z in Proposition 18 will be

$$f_{X,N|Z}(x,n) = \frac{\beta^{\alpha n}}{\Gamma(\alpha n)}x^{\alpha n -1}e^{-\beta x}(1-e^{-z})e^{-zn}, \,\,\, x\in \mathbb{R}_{+}, n\in \mathbb{N}.$$

Then, the distribution of (X,N)GMDP(α,β,δ,p) can be seen as the marginal distribution of (X,N) in a trivariate model (X,N,Z) given by the PDF

$$f_{X,N,Z}(x,n,z) = f_{X,N|Z}(x,n)f_{Z}(z), \,\,\, x,z\in \mathbb{R}_{+}, n\in \mathbb{N},$$

so that

$$f_{X,N}(x,n) = \int_{0}^{\infty} \frac{\beta^{\alpha n}}{\Gamma(\alpha n)}x^{\alpha n -1}e^{-\beta x}(1-e^{-z})e^{-zn}f_{Z}(z)dz, \,\,\, x\in \mathbb{R}_{+}, n\in \mathbb{N}.$$

Further properties of the GMDP model can be developed by means of this hierarchical structure coupled with standard conditioning arguments.

Another consequence of the hierarchical structure of the DP distribution is its infinite divisibility (cf., Corollary 3.5 in Buddana and Kozubowski 2014), which, in view of Corollary 1, immediately implies the same property of the GMDP model. In the following result we need the PGF of $$N\sim \mathcal {DP}(\delta,p)$$, which is known to be of the form (see Buddana and Kozubowski 2014, Proposition 2.6):

$$G_{N}(s) = s - (1-s) \sum_{n=1}^{\infty} \left(\frac{1}{1-\delta \log(1-p)n}\right)^{\frac{1}{\delta}} s^{n}, \,\,\, s\in(0,1).$$
(48)

### Corollary 3

The distribution of (X,N)GMDP(α,β,δ,p) is ID and for each $$n\in \mathbb {N}$$ we have the stochastic representation (32), where the $$\{ (R_{j}^{(n)},V_{j}^{(n)})\}$$ are IID. Moreover, we have $$V_{j}^{(n)}\stackrel {d}{=}1/n+T_{j}^{(n)}$$, where the $$\{ T_{j}^{(n)}\}$$ are IID, integer-valued random variables with the PGF Gn(s)=[GN(s)/s]1/n and GN(·) given by (48). Further, the $$\{R_{j}^{(n)}\}$$ are given by (33), where all the variables on the right-hand-side of (33) are mutually independent, the $$\{G_{j}^{(n)}\}$$ are IID gamma $$\mathcal {GAM}(\alpha /n, \beta)$$ distributed, and the {Xij} are IID gamma $$\mathcal {GAM}(\alpha, \beta)$$ distributed.

### 3.5 Parameter estimation

The discussion of estimation for the general case in Section 2 generally carries over to the special case of the GMDP distribution. Here, the variable $$N \sim \mathcal {DP}(\delta, p)$$ depends on a 2-dimensional vector-parameter ω=(δ,p)Ω, where we take the parameter space Ω to be an open set Ω=(0,)×(0,1). The PDF of N, which is given by (6), will be denoted by fN(·;ω). Clearly, our model has four unknown parameters, which can be conveniently described as a vector-parameter θ, where θ1=α,θ2=β,θ3=δ, and θ4=p. With this notation, the parameter space is an open set Θ=(0,)3×(0,1).

#### 3.5.1 Fisher information

According to general results in Section 2, if the appropriate expectations exist, the Fisher information matrix I(θ) connected with the $$\mathcal {GMDP}(\alpha, \beta, \delta, p)$$ distribution has a block-diagonal structure,

$$\bf I(\boldsymbol{\theta}) = \left[ \begin{array}{cccc} w_{11} & w_{12} & 0 & 0\\ w_{21} & w_{22} & 0 & 0\\ 0 & 0 & j_{11} & j_{12} \\ 0 & 0 & j_{21} & j_{22} \end{array} \right],$$
(49)

where the entries {wij}, generally given by (35), depend only on the parameters α and β, and the 2×2 matrix

$$\bf J(\boldsymbol{\omega}) = \left[ \begin{array}{cc} j_{11} & j_{12} \\ j_{21} & j_{22} \end{array} \right]$$
(50)

is the Fisher information matrix connected with the $$\mathcal {DP}(\delta, p)$$ model. Since in our situation the expectation of N is given by (45), according to (36), we have

$$\begin{array}{*{20}l} w_{11} & = \sum_{n=1}^{\infty} n^{2}\psi'(\alpha n) \left[ \left(\frac{1}{1-\delta(n-1)\log(1-p)}\right)^{\frac{1}{\delta}}- \left(\frac{1}{1-\delta n \log(1-p)}\right)^{\frac{1}{\delta}}\right], \\ w_{12} & = w_{21} = -\frac{1}{\beta} \sum_{n=0}^{\infty}\left(\frac{1}{1-\delta \log(1-p) n}\right)^{\frac{1}{\delta}}, \\ w_{22} & = \frac{\alpha}{\beta^{2}}\sum_{n=0}^{\infty}\left(\frac{1}{1-\delta \log(1-p) n}\right)^{\frac{1}{\delta}}, \end{array}$$

where ψ(·) is the digamma function. We note that the above quantities are well defined only when δ<1, due to the heavy tail of the DP distribution. Indeed, clearly the two series describing w12=w21 and w22 converge only if 1/δ>1, so that we must have δ<1. In addition, we have the asymptotic relation ψ(z)1/z as z (see, e.g., formula 6.4.12 in Abramowitz and Stegun 1972, p. 260) and the PMF of the DP distribution given by the expression in the square bracket in the series describing w11 behaves asymptotically as the power n−(1+1/δ) at infinity. Thus, that series also converges only if δ<1. Further, standard calculations produce the information matrix (50) of the DP model, which is given in the result below.

### Proposition 19

The Fisher information matrix (50) of the $$\mathcal {DP}(\delta, p)$$ model is well defined for any δ>0 and p(0,1), and is given by

$$\bf J(\boldsymbol{\omega}) = \mathbf{D}^{\top} \mathbf{A} \mathbf{D}$$
(51)

where

$$\mathbf{D} = \left[ \begin{array}{cc} - \frac{1}{\delta^{2}} & 0 \\ -\log(1-p) & \frac{\delta}{1-p} \end{array} \right],$$
(52)

A=[aij] is a symmetric matrix with entries of the form

$$\begin{array}{@{}rcl@{}} a_{11} & = \mathbb{E}\left\{ \frac{[g(\delta, p, N) - g(\delta, p, N-1)]^{2}}{[f_{N}(N; \delta, p)]^{2}}\right\}, \\ a_{12} = a_{21} & = \frac{1}{\delta}\mathbb{E} \left\{\frac{[g(\delta, p, N) - g(\delta, p, N-1)][v(\delta, p, N) - v(\delta, p, N-1)]}{[f_{N}(N; \delta, p)]^{2}}\right\}, \\ a_{22} & = \frac{1}{\delta^{2}}\mathbb{E} \left\{\frac{[v(\delta, p, N) - v(\delta, p, N-1)]^{2}}{[f_{N}(N; \delta, p)]^{2}}\right\}, \end{array}$$

the function fN(·;δ,p)is the PDF (6) of the $$\mathcal {DP}(\delta, p)$$ distribution, and

\begin{aligned} g(\delta, p, n) = \frac{\log[1-\delta n \log(1-p)]}{[1-\delta n \log(1-p)]^{1/\delta}}, \,\,\, v(\delta, p, n) = \frac{n}{[1-\delta n \log(1-p)]^{1+1/\delta}}, \,\,\,\\ \delta\in (0,\infty), p\in (0,1), n \in \mathbb{N}_{0}. \end{aligned}

#### 3.5.2 Maximum likelihood estimation

Let (X1,N1),…,(Xn,Nn) be a random sample from a GMDP distribution. According to general results on estimation discussed in Section 2, maximum likelihood estimation can be carried out separately for the two pairs, α,β, and δ,p. Regarding the former, the MLEs of α and β always exist and are unique, and their values are provided in Corollary 2. Moreover, while the calculation of the MLEs involves solving a one-dimensional non-linear equation, stated in (42), the relevant function is rather well behaved and finding the solution by standard numerical methods is straightforward. Further, since the family of DP distributions {fN(·;ω), ω(0,)×(0,1)} is such that the function Ψ=(ψ1,…,ψ4) with ψi=−/θi logfX,N(x,n),i=1,…,4, satisfies standard regularity conditions of large-sample theory, the MLEs of α and β are consistent, asymptotically normal, and efficient. Due to the special structure of the Fisher information matrix (49), we will have the convergence in distribution (43) where W is the 2×2 matrix with entries {wij} given above. This asymptotic normal distribution can be used to derive (approximate) confidence intervals for the two parameters.

Let us now move to estimation of parameters δ and p. According to general results on estimation discussed in Section 2, this requires optimization of the function Q(ω) in (39) with respect to ω=(δ,p)(0,)×(0,1). Since the function

$$Q(\delta, p)= \sum\limits_{i=1}^{n}\log\left[ \left(\frac{1}{1-\delta(N_{i}-1)\log(1-p)}\right)^{\frac{1}{\delta}}-\left(\frac{1}{1-\delta(N_{i})\log(1-p)}\right)^{\frac{1}{\delta}} \right]$$
(53)

is the log-likelihood (LL) function of an univariate random sample {Ni} from a DP distribution, one can use methods available for this distribution in order to get the MLEs. In particular, we can proceed by solving the likelihood equations, obtained by setting to zero the two partial derivatives of the LL function in (53). The calculation of the derivatives is straightforward; after some algebra, we obtain the following likelihood equations:

$$\begin{array}{@{}rcl@{}} &\sum\limits_{i=1}^{n}\frac{N_{i}}{h(\delta, p, N_{i})}\left\{ \left[ \frac{h(\delta, p, N_{i})}{h(\delta, p, N_{i}-1)}\right]^{\frac{1}{\delta}} - 1\right\}^{-1} = \sum\limits_{i=1}^{n}\frac{N_{i}-1}{h(\delta, p, N_{i}-1)}\left\{ 1 - \left[ \frac{h(\delta, p, N_{i}-1)}{h(\delta, p, N_{i})}\right]^{\frac{1}{\delta}} \right\}^{-1}, \end{array}$$
$$\begin{array}{@{}rcl@{}} &\sum\limits_{i=1}^{n}\log h(\delta, p, N_{i}) \left\{ \left[ \frac{h(\delta, p, N_{i})}{h(\delta, p, N_{i}-1)}\right]^{\frac{1}{\delta}} - 1\right\}^{-1} \!= \sum\limits_{i=1}^{n}\log h(\delta, p, N_{i}-1) \left\{ 1 - \left[ \frac{h(\delta, p, N_{i}-1)}{h(\delta, p, N_{i})}\right]^{\frac{1}{\delta}} \right\}^{-1}, \end{array}$$

where

$$h(\delta, p, n) = 1-\delta n \log (1-p), \,\,\, \delta \in (0,\infty), p\in (0,1), n\in \mathbb{N}_{0}.$$

While the above equations can be solved for δ and p by standard numerical methods, the actual calculations may be computationally intensive. An alternative approach for finding the MLEs in this case is the standard EM algorithm, as the DP distribution is conveniently represented as a mixture of geometric distributions. Such an algorithm has been recently developed and tested in Amponsah and Kozubowski (2021), and was found to be working quite well in this setting. We recommend that it be used for this estimation problem.

### Remark 18

A careful analysis of the case with n=1 shows that the likelihood equations above do not admit any solution within the parameter space. However, the function Q(δ,p) is maximized by $$\hat {\delta }_{n} = 0$$ and $$\hat {p}_{n}=1/\bar {N}$$. These values point towards the limiting geometric distribution of N, which arises when δ=0 in the DP model.

## An application from finance

In this section, we show an application of the GMDP model to a financial data set. We chose the daily closing price of Bahrain all share Index, quoted in Bahrain dinar. The data set was obtained from Investing.com, and consisted of the daily closing prices, from May 24, 2010, to February 20, 2019. We converted our daily data to (n=2171) daily log-returns, which are the logarithms of the ratios of the closing prices for every two consecutive days. Next, we computed the growth episodes and obtained (n=508) observations (Xi,Ni),i=1,2,…,508, where the Ni is the duration and Xi is the magnitude (total log-return over the period of the ith episode) of the ith growth episode. Duration data from this data set was previously fitted with the DP model in Amponsah and Kozubowski (2021). In this work we fit the bivariate GMDP model to the duration and magnitude of the growth episodes.

Assuming the data are IID $$\mathcal {GMDP}(\alpha,\beta, \delta, p)$$, the MLEs computed by solving likelihood equations (see Section 2.4) are $$\hat {\alpha } = 0.9420, \hat {\beta }=281.4498,$$$$\hat {\delta }=0.1153$$, and $$\hat {p}=0.5355$$. The EM algorithm of Amponsah and Kozubowski (2021) was used to compute $$\hat {\delta }$$ and $$\hat {p}$$. To assess the goodness-of-fit of our model to the growth episodes of the Bahrain all share Index, we used the MLEs described above to compute the parameters of all the marginal and conditional distributions we fit below. Considering this fact, the generally reasonable fit of the marginals and analyzed conditional models seems quite remarkable. We analyzed the fit of the following marginal and conditional distributions: (1) marginal of duration, (2) marginal of magnitude, (3) conditional distribution of X|N=n for n=1,2,3,4, and (4) conditional distribution of N|X>x for thresholds x=0.0001,0.0015, and 0.0025.

### 4.1 Goodness-of-fit analysis

We started with the fit of DP model to the marginal of duration. Table 1 contains frequency, relative frequency, and model probabilities for the values of N observed in the data, namely N=1,2,...,6, and N≥7.

Figure 1 shows a plot of the DP model probabilities (the last row of Table 1) versus the observed relative frequencies (the third row of Table 1), with a diagonal line y=x added for visualization purposes. The DP probabilities are very close to the relative frequencies, and the mean (absolute) percentage error is about 10.598%. A formal chi-square goodness-of-fit test based on the information provided in Table 1 was performed as well, resulting in the p-value of 0.27. Thus, the null hypothesis of DP Pareto distribution of N is not rejected at 5% significance level. Overall, this analysis shows a reasonable fit of the DP model to the marginal of duration.

Next, we checked if the positive daily log-returns are well fit by the gamma distribution with parameter α and β. We used graphical tools (histogram, q-q plots), and formal Kolmogorov-Smirnov (KS) test to confirm the goodness-of-fit. Figure 2 shows a histogram of positive daily log-returns overlayed with fitted gamma PDFs and a q-q plot of the model and empirical quantiles. Both plots show a good fit. We also tested the fit of the model to the positive daily log-returns. The KS statistic and its p-value are 0.0397 and 0.3905, respectively. Therefore, the null hypothesis that the positive daily log-returns follow gamma distributions with parameters α and β, is not rejected at significance level of 5%. Overall, we believe that the positive log-returns may be assumed to have come from gamma distribution with parameters α and β.

Further, we investigated the fit of the model marginal distribution of magnitude to the data. The model marginal distribution of X is a mixture of gamma distributions with scale parameter β and shape parameters $$\alpha n, n\in \mathbb {N}$$, with the mixing weights given by P(N=n), where NDP(δ,p). The plot in Fig. 3 shows a histogram of the marginal of magnitude overlaid with the gamma mixture model and the corresponding q-q plot. The KS test statistic and p-value are 0.0453 and 0.6751, respectively. Thus, the null hypothesis that the model marginal follows a mixture of gamma distributions with PDF (44) is not rejected at significance level of 5%. We believe the results show a reasonable fit of the marginal distribution.

Next, we turned to goodness-of-fit analysis of the GMDP model’s conditional distributions of X given N=k, which are gamma with parameters kα and β, to the corresponding data. Figure 4 shows graphical illustration of fit via histograms overlaid with fitted gamma PDFs and q-q plots for durations N=1,2,3, and 4. The plots show a reasonable fit of the conditional distributions of the GMDP model to the corresponding data.

In addition, we performed the Kolmogorov-Smirnov goodness-of-fit test of these conditional models to the data. The results of the KS tests of the null hypothesis that the conditional distributions of X given N=k are gamma with parameters kα and β, are given in Table 2. None of the tests rejected the null hypothesis on the significance level of 5%. We believe the results confirm a reasonable fit of the conditional distributions.

Similarly, we performed a goodness-of-fit analysis of the GMDP model’s conditional distributions of N|X>x for thresholds x=0.0001,0.0015 and 0.0025, which should have a hidden truncated discrete Pareto distributions as described in Proposition 6.

Tables 3, 4 and 5 show the frequencies, relative frequencies and estimated model probabilities of all the observed growth periods’ durations given that X>x, for the thresholds of x=0.0001, 0.0015, and 0.0025, respectively. Figure 5 shows plots of the model probabilities versus the corresponding relative frequencies reported in these tables. The estimated conditional model probabilities are very close to the relative frequencies, and the mean of the (absolute) percentage errors for the thresholds x=0.0001,0.0015, and 0.0025, are 10.5728%, 10.0313% and 10.0817%, respectively. The chi-square goodness-of-fit tests based on the information in Tables 3 - 5 resulted in the p-values of 0.45, 0.52, 0.49, respectively. We conclude that these conditional theoretical models show reasonable fit to the data.

Overall, we believe the GMDP model fits the growth periods of the Bahrain index rather well. This example provides evidence for the considerable modeling potential of the GMDP model.

## Concluding remarks

This work provides fundamental properties of a general stochastic model describing the joint distribution of (X,N), where N is a counting variable while X is the sum of N independent gamma random variables. These properties include marginal and conditional distributions, Laplace transform, moments and tail behavior, and divisibility properties. The existence and uniqueness of maximum likelihood estimators of the relevant parameters are established as well, along with their asymptotic properties. A particular example with a heavy tailed discrete Pareto distributed N illustrates general properties of this stochastic model, while a data example from finance shows the modeling potential of this new mixed bivariate distribution.

## \thelikesection Appendix

This section contains (selected) proofs and auxiliary results.

### Lemma 3

For x>0, let h(x)= log(x)−ψ(x) where ψ(x) is the digamma function, which is the derivative of logΓ(x). Then h(x) is continuous and strictly decreasing on the interval (0,), with

$${\lim}_{x\rightarrow 0^{+}}h(x)=\infty \,\,\, \text{and} \,\,\, {\lim}_{x\rightarrow \infty} h(x)=0.$$

Proof of Lemma 3. Since the trigamma function ψ(x) satisfies the inequality (see, e.g., Corollary 5.5 in Laforgia and Natalini 2013, p. 501)

$$\psi'(x) > \frac{1}{x} \,\,\, \text{for $$x>0$$},$$

we conclude that

$$h'(x) = \frac{1}{x}- \psi'(x)<0 \,\,\, \text{for $$x>0$$},$$

which proves the monotonicity of h(x) on (0,). The limits follow easily from the inequality (see, e.g., Remark 5.3 in Laforgia and Natalini 2013, p. 500)

$$\frac{1}{2x} < \log(x)-\psi(x)<\frac{1}{x}, \,\,\, x>0.$$

This concludes the proof.

Proof of Proposition 2. By standard conditioning argument, we have

$$\psi(t,s) = \mathbb{E}\left(e^{-tX-sN}\right) = \mathbb{E}\left\lbrace \mathbb{E}\left(e^{-sN} e^{-tX}|N\right) \right\rbrace = \mathbb{E}\left\lbrace e^{-sN}H(N) \right\rbrace,$$

where

$$H(n) = \mathbb{E}\left(e^{-t\sum\limits_{i=1}^{N}X_{i}} |N=n \right) =\mathbb{E}\left(e^{-t \sum\limits_{i=1}^{n}X_{i}} \right) = \prod\limits_{i=1}^{n} \mathbb{E}(e^{-tX_{i}}) = [\psi_{X_{1}}(t)]^{n}$$

and $$\psi _{X_{1}}(\cdot)$$ is the LT of the {Xi}. This leads to

$$\psi(t,s) = \mathbb{E} \left\{ \left[ e^{-s}\psi_{X_{1}}(t)\right]^{N} \right\} = G_{N}\left(e^{-s}\psi_{X_{1}}(t)\right).$$

Since for gamma distributed {Xi} we have $$\psi _{X_{1}}(t) = 1/(1+t/\beta)^{\alpha }$$, the result follows.

Proof of Proposition 3. The result follows from Theorem 3.2 in Robert and Segers (2008), as the conditions (3.10) or (3.11) of that theorem hold for our N due to the fact that N satisfies (13) while the {Xi} are light tailed gamma distributed.

Proof of Proposition 5. The result follows from straightforward application of Proposition 4.

Proof of Lemma 1. We proceed by showing that the PMF of the random variable on the right-hand-side in (19) coincides with the expression on the right-hand-side in (18). Indeed, by independence of N and Q, for any $$n\in \mathbb {N}_{0}$$, we have

$$\mathbb{P}(N=n|N=Q) = \frac{\mathbb{P}(N=n, N=Q)}{\mathbb{P}(N=Q)} = \frac{\mathbb{P}(N=n, Q =n)}{\mathbb{P}(N=Q)} = \frac{f_{N}(n)f_{Q}(n)}{\mathbb{P}(N=Q)},$$

so that the above probability must coincide with the PMF of Y in (18) with $$c=\mathbb {P}(N=Q)$$.

Proof of Proposition 8. We know, by Proposition 7, that N given X>x has the same distribution as N given NN(x)+1, where N(x) is independent of N, and, in our case, Poisson with parameter λ=βx. The PMF of this conditional distribution is of the form

\begin{aligned} h(n) = \mathbb{P}(N=n|N\geq N(x)+1) &= \frac{\mathbb{P}(N=n, N\geq N(x)+1)}{\mathbb{P}(N-1 \geq N(x))}\\& = \frac{\mathbb{P}(N=n)\mathbb{P} (N(x)\leq n-1)}{\mathbb{P}(N-1 \geq N(x))}. \end{aligned}
(54)

By the independence of N and N(x), the denominator on the right-hand-side in (54) is equal to

\begin{aligned} \mathbb{P}(N-1 \geq N(x)) &= \sum\limits_{n=1}^{\infty} \mathbb{P}(N=n, N-1 \geq N(x))\\ &= \sum\limits_{n=1}^{\infty} \mathbb{P}(N=n, n-1 \geq N(x)) = \sum\limits_{n=0}^{\infty} p(1-p)^{n} \mathbb{P}(N(x) \leq n). \end{aligned}

By inserting

$$\mathbb{P}(N(x) \leq n) = \sum_{j=0}^{n} \mathbb{P}(N(x)=j) = \sum_{j=0}^{n} \frac{(\beta x)^{j} e^{-\beta x}}{j!}$$
(55)

into the above expression and interchanging the order of the summation, followed by summing-up the resulting geometric series, we arrive at

\begin{aligned} \mathbb{P}(N-1 \geq N(x)) = \sum_{j=0}^{\infty} \frac{(\beta x)^{j} e^{-\beta x}}{j!}(1-p)^{j} = e^{(1-p)\beta x - \beta x}. \end{aligned}
(56)

We now calculate the PGF of Y=dN|NN(x)+1, which, in view of (54) and (56), can be expressed as

$$G(s) = \mathbb{E}[s^{Y}] = \sum_{n=1}^{\infty} s^{n} h(n) = \frac{s p \sum_{n=0}^{\infty} s^{n}(1-p)^{n} \mathbb{P}(N(x)\leq n)}{e^{(1-p)\beta x - \beta x}}, \,\,\, s\in [0,1].$$
(57)

By proceeding as above, and calculating the numerator in (57) by inserting there the expression (55), followed by interchanging the order of the summation and subsequently summing-up the resulting geometric series, we find the numerator in (57) to be

$$s p \sum_{n=0}^{\infty} s^{n}(1-p)^{n} \mathbb{P}(N(x)\leq n) = \frac{sp}{1-(1-p)s} e^{s(1-p)\beta x - \beta x}.$$

Thus, we obtain

$$G(s) = \frac{sp}{1-(1-p)s} e^{- (1-p)\beta x(1-s)}, \,\,\, s\in [0,1].$$

Since this is the product of the PGF of N and the PGF of T, we obtain the result.

Proof of Proposition 9. The result follows by direct computation of the LTs on each side in (29), and subsequence verification that they coincide.

Proof of Proposition 11. We proceed by showing that the joint LT of the expression on the right-hand-side in (32) coincides with the LT of (X,N), with the latter given by $$\phantom {\dot {i}\!}\psi (t,s) = G_{N} [e^{-s} \psi _{X_{1}}(t)]$$, as stated in the second remark following Proposition 2. We start with the LT of the pair $$(R_{j}^{(n)},V_{j}^{(n)})$$. By standard conditioning argument, we have

$$\mathbb{E}\left(e^{-tR_{j}^{(n)}-sV_{j}^{(n)}}\right) = \mathbb{E}\left\lbrace \mathbb{E}\left(e^{-tR_{j}^{(n)}-sV_{j}^{(n)}} | T_{j}^{(n)} \right) \right\rbrace = \mathbb{E}\left\lbrace H(T_{j}^{(n)}) \right\rbrace,$$

where

\begin{aligned} H(k) = \mathbb{E}\left(e^{-t\left(\sum\limits_{i=1}^{k}X_{ij} + G_{j}^{(n)}\right)} e^{-s(k+1/n)}\right)& = [\psi_{X_{1}}(t)]^{k+1/n}e^{-s(k+1/n)}\\ &= [e^{-s}\psi_{X_{1}}(t)]^{k} [e^{-s}\psi_{X_{1}}(t)]^{1/n}. \end{aligned}

Since the PGF of $$\{ T_{j}^{(n)}\}$$ is equal to Gn(s)=[GN(s)/s]1/n, we have

\begin{aligned} \mathbb{E} \left\lbrace H(T_{j}^{(n)}) \right\rbrace &= [e^{-s}\psi_{X_{1}}(t)]^{1/n} \mathbb{E} \left([e^{-s}\psi_{X_{1}}(t)]^{T_{j}^{(n)}}\right)\\ & = [e^{-s}\psi_{X_{1}}(t)]^{1/n} G_{n}[e^{-s}\psi_{X_{1}}(t)] = \left\{ G_{N} [e^{-s}\psi_{X_{1}}(t)]\right\}^{1/n}. \end{aligned}

Finally, since variables under the sum in (32) are IID, the LT of the sum becomes

$$\left[ \mathbb{E}\left(e^{-tR_{j}^{(n)}-sV_{j}^{(n)}}\right) \right]^{n} = \left[ \left\{ G_{N} [e^{-s}\psi_{X_{1}}(t)]\right\}^{1/n} \right]^{n} = G_{N} [e^{-s}\psi_{X_{1}}(t)],$$

as desired. This completes the proof.

Proof of Theorem 12. First, we re-write the function g(α) as follows:

{}\begin{aligned} g'(\alpha)&=\log(\alpha) -\frac{1}{n\bar{N}_{n}} \sum\limits_{i=1}^{n} N_{i}\psi\left[ \alpha N_{i}\right] + \log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) + \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}N_{i} \log(X_{i})\\ & = \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}\left[ N_{i} \left(\log(\alpha) -\psi\left[ \alpha N_{i}\right] + \log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) + \log(X_{i})\right)\right]\\ &=\frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}\left[ N_{i} \left(\log(\alpha N_{i}) -\psi\left[ \alpha N_{i}\right] + \log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) + \log\left(\frac{X_{i}}{N_{i}}\right) \right)\right]\\ &=\frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}\left[ N_{i} \left(\log(\alpha N_{i}) -\psi\left[ \alpha N_{i}\right]\right)\right] + \log\left(\frac{\bar{N}_{n}}{\bar{X}_{n}}\right) + \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n} N_{i} \left[\log\left(\frac{X_{i}}{N_{i}}\right) \right]\\ &=\frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n}\left[ N_{i} \left(\log(\alpha N_{i}) -\psi\left[ \alpha N_{i}\right]\right)\right] + \frac{1}{n\bar{N}_{n}}\sum\limits_{i=1}^{n} N_{i} \left[\log\left(\frac{X_{i}}{N_{i}}\right) \right]- \log\left(\frac{\bar{X}_{n}}{\bar{N}_{n}}\right). \end{aligned}

In view of the above and Lemma 3, we immediately obtain Parts (i) and (ii) as well as the limit in Part (iii). To see that the latter is less then or equal to zero, it is enough to use Jensen’s inequality,

$$\sum_{i=1}^{n} a_{i} \varphi(b_{i}) \leq \varphi\left(\sum_{i=1}^{n} a_{i} b_{i}\right),$$
(58)

where φ(·) is a concave function and the positive weights {ai} sum up to one. Indeed, we get the result upon setting $$\varphi (x) = \log x, a_{i}=N_{i}/(n\bar {N}_{n})$$, and bi=Xi/Ni in (58). Part (iv) is straightforward. Finally, Part (v) is obtained by noting that for a non-linear function φ(·), we have an equality in (58) if and only if b1==bn, which in our case becomes Xi/Ni=c,i=1,…,n, for some c>0. However, this event has probability zero since the {Xi/Ni} are IID and their distribution is continuous. This concludes the proof.

Proof of Proposition 19. Consider a re-parameterization of a DP model, where the new parameters are τ1=1/δ and τ2=−δ log(1−p). Then, the Fisher information matrix of the DP model in the δ,p parameterization is given by (51), where D is the Jacobian matrix of the transformation (δ,p)→(τ1,τ2), which is given by (52), and A is the Fisher information matrix of the DP model in the τ1,τ2 parameterization. Routine calculations show that the entries of the latter matrix, written in terms of δ and p, are the quantities {aij} specified in the proposition, and all the relevant expectations are finite for any δ(0,) and p(0,1).

## Availability of data and materials

The data set used in this study was obtained from Investing.com and is available from the corresponding author on reasonable request.

Declarations

## Abbreviations

BEG:

Bivariate (distribution) with exponential and geometric margins

BGG:

Bivariate gamma geometric

CDF:

Cumulative distribution function

DP:

Discrete Pareto

$$\mathcal {EP}(lpha,eta)$$ :

Extended Poisson distribution with parameters α, β

$$\mathcal {GAM}(lpha,eta)$$ :

gamma distribution with parameters α, β

GMDP:

(Distribution) with gamma-mixture and discrete Pareto margins

ID:

Infinitely divisible

IID:

Independent, identically distributed

LT:

Laplace transform

ML:

Mittag-Leffler

MLE:

maximum likelihood estimator

PDF:

probability density function

PGF:

Probability generating function

PMF:

Probability mass function

SF:

Survival function

## References

1. Abramowitz, M., Stegun, I. A.: Handbook of Mathematical Functions with Formulas, Graphs and Mathematical Tables. Dover Publications, New York (1972).

2. Arnold, B. C.: Flexible univariate and multivariate models based on hidden truncation. J. Statist. Plann. Infer. 139, 3741–3749 (2009).

3. Amponsah, C. K.: A Bivariate Gamma Mixture Discrete Pareto Distribution. Thesis, University of Nevada (2017).

4. Amponsah, C. K., Kozubowski, T. J.: A computational approach to estimation of discrete Pareto parameters, working paper (2021).

5. Arendarczyk, M., Kozubowski, T. J., Panorska, A. K.: A bivariate distribution with Lomax and geometric margins. J. Korean Statist. Soc. 47, 405–422 (2018).

6. Bardwell, G. E., Crow, E. L.: A two-parameter family of hyper-Poisson distributions. J. Amer. Statist. Assoc. 59(305), 133–141 (1964).

7. Barreto-Souza, W.: Bivariate gamma-geometric law and its induced Lévy process. J. Multivar. Anal. 109, 130–145 (2012).

8. Barreto-Souza, W., Silva, R. B.: A bivariate infinitely divisible law for modeling the magnitude and duration of monotone periods of log-returns. Statist. Neerlandica. 73, 211–233 (2019).

9. Bickel, P. J., Doksum, K. A.: Mathematical Statistics: Basic Ideas and Selected Topics, Vol I, 2nd ed. Chapman & Hall/CRC Press, Boca Raton (2015).

10. Biondi, F., Kozubowski, T. J., Panorska, A. K.: A new model for quantifying climate episodes. Intern. J. Climatol. 25, 1253–1264 (2005).

11. Biondi, F., Kozubowski, T. J., Panorska, A. K., Saito, L: A new stochastic model of episode peak and duration for eco-hydro-climatic applications. Ecol. Model. 211, 383–395 (2008).

12. Buddana, A., Kozubowski, T. J.: Discrete Pareto distribution. Econ. Qual. Control. 29, 143–156 (2014).

13. Haubold, H. J., Mathai, A. M., Saxena, R. K.: Mittag-Leffler functions and their applications. J. Appl. Math. 2011, ID 298628 (2011).

14. Kozubowski, T. J., Panorska, A. K.: A Mixed bivariate distribution with exponential and geometric marginals. J. Statist. Plann. Infer. 134, 501–520 (2005).

15. Kozubowski, T. J., Panorska, A. K., Biondi, F.: Mixed multivariate models for random sums and maxima. In: SenGupta, A. (ed.)Advances in Multivariate Statistical Methods, Vol 4, pp. 145–171. World Scientific, Singapore (2009).

16. Kozubowski, T. J., Panorska, A. K., Podgórski, K.: A bivariate Lévy process with negative binomial and gamma marginals. J. Multivar. Anal. 99, 1418–1437 (2008).

17. Kozubowski, T. J., Panorska, A. K.: A note on the joint distribution involving Poissonian sum of exponential variables. Adv. Appl. Statist. Sci. 1(2), 157–165 (2010).

18. Krishna, H., Singh Pundir, P.: Discrete Burr and discrete Pareto distributions. Statist. Meth. 6, 177–188 (2009).

19. Laforgia, A., Natalini, P.: Exponential, gamma and polygamma functions: Simple proofs of classical and new inequalities. J. Math. Anal. Appl. 407, 495–500 (2013).

20. Noack, A.: A class of random variables with discrete distributions. Ann. Math. Statist. 21(1), 127–132 (1950).

21. Robert, C. Y., Segers, J.: Tails of random sums of a heavy-tailed number of light-tailed terms. Insur. Math. Econ. 43, 85–92 (2008).

## Acknowledgements

We thank the two reviewers for their helpful and constructive comments that help improve the paper.

Not applicable.

## Author information

Authors

### Contributions

Most of the results were derived by CKA, who also did data analysis and produced the figures.

TJK and AKP contributed some results, drafted the final paper, and reviewed all the work through the final version of the manuscript. All authors read and approved the final manuscript.

### Corresponding author

Correspondence to Tomasz J. Kozubowski.

## Ethics declarations

### Competing interests

The authors declare that they have no competing interests. 