Journal of Statistical Theory and Applications

Volume 18, Issue 3, September 2019, Pages 295 - 302

Discriminating Between Exponential and Lindley Distributions

Authors
V. S. Vaidyanathan*, A Sharon Varghese
Department of Statistics, Pondicherry University, Puducherry, India
*Corresponding author. Email: vaidya.stats@gmail.com
Corresponding Author
V. S. Vaidyanathan
Received 11 October 2017, Accepted 5 June 2018, Available Online 3 September 2019.
DOI
10.2991/jsta.d.190818.006How to use a DOI?
Keywords
Hellinger distance; Lindley distribution; Meijer G-function; Probability of correct selection; Pseudo; Likelihood estimator; Ratio of maximum likelihoods
Abstract

In literature, Lindley distribution is considered as an alternate to the exponential distribution. In the present work, a methodology is developed to discriminate between exponential and Lindley distributions based on the ratio of the maximum likelihoods. Asymptotic distribution of the test statistic under the null hypothesis is derived and the minimum sample size required to discriminate between the two distributions for a user specified probability of correct selection is obtained. Numerical illustrations of the methodology are given through simulated and real life data sets.

Copyright
© 2019 The Authors. Published by Atlantis Press SARL.
Open Access
This is an open access article distributed under the CC BY-NC 4.0 license (http://creativecommons.org/licenses/by-nc/4.0/).

1. INTRODUCTION

In early days of research in reliability theory, exponential distribution is used widely to model failure time data because of its analytical simplicity. However, constant failure rate property of exponential distribution is not always suitable in most of the modern reliability engineering problems. Therefore, many life time distributions have been proposed and studied in literature as an alternative to exponential distribution to model lifetime data. Lindley [1] proposed a probability distribution to show that Bayes' distribution and fiducial distribution are not always same. Ghitany et al. [2] studied various characteristics of Lindley distribution and has shown that Lindley distribution can be used as a better alternative to exponential distribution. Cumulative distribution function (c.d.f) and probability density function (p.d.f) of Lindley distribution with parameter β, denoted by Li (β), is given by

F(x)=11+β+βx1+βeβx,x0,β>0
f(x)=β21+β(1+x)eβx,x0,β>0.

Some generalizations and modifications of Lindley distribution have been suggested by Zakerzadeh and Dolati [3], Nadarajah et al. [4], Bakouch et al. [5], Vaidyanathan and Sharon Varghese [6]. Krishna and Kumar [7], Mazucheli and Achcar [8], Al-Mutairi et al. [9] have studied applications of Lindley distribution in analyzing lifetime data.

From (2), it is clear that the p.d.f of Li(β) is a mixture of exponential (β) and Gamma (β,2) with respective weights β(1+β) and 1(1+β). Thus for larger values of β both the Lindley and exponential distributions can fit the underlying data reasonably well. However, other life time characteristics such as hazard rate function, mean time to failure, etc., are different for these distributions. From a statistical modelling perspective, it is of interest to know which distribution is more appropriate for the given data. This serves as a motivation to develop a discrimination procedure to discriminate between exponential and Lindley distributions using the sample data. In literature, many authors have considered ratio of maximum likelihoods (RML) to discriminate between various distributions. RML is used by Cox [10] and Cox [11] for discriminating between two non-nested models. Dumonceaux and Antle [12] have used RML to discriminate between log-normal and Weibull distributions. A procedure for discriminating between Weibull and log-normal distributions is developed by Kundu and Manglick [13]. Kundu et al. [14] have considered discriminating between log-normal and generalized exponential distributions. To the best of authors' knowledge, no work has been done in discriminating between exponential and Lindley distributions. Thus the aim of this paper is to develop a discrimination procedure based on RML to discriminate between exponential and Lindley distributions. Rest of the paper is organized as follows. In Section 2, discrimination procedure based on RML is developed and the probability of correct selection (PCS) of the procedure is defined. Section 3 discusses asymptotic properties of the RML under null hypotheses. In Section 4, proposed methodology is validated using simulation study. Section 5 deals with the numerical illustration of the developed procedure using two real life data sets. A procedure to determine the minimum sample size required to discriminate between the two distributions is proposed in Section 6. Conclusion of the paper is given in Section 7.

2 DISCRIMINATION PROCEDURE

In this section, likelihood ratio test statistic for testing whether the given data is from the exponential distribution (exp(α)) with p.d.f

f(x)=αeαx,x0,α>0.
or Lindley distribution (Li(β)) is derived. Suppose x1,x2,...,xn is a random sample of n observations from a population having distribution either exp(α) or Li(β). An experimenter wishes to know from which distribution the given data has arisen. For this, we consider the RML defined as
RML=Lexpα^LLiβ^,
where Lexp(α^) and LLi(β^) are likelihood functions corresponding to (3) and (2) respectively and α^ and β^ are the corresponding maximum likelihood estimators (MLEs) of α and β. Substituting respective likelihood functions in (4) and taking natural logarithm on both sides, we get logarithm of RML, denoted by T as
T=nlog(α^)α^i=1nxi2nlogβ^i=1nlog(1+xi)+β^i=1nxi+nlog(1+β^),
where α^ and β^ are given by
α^=1x¯β^=x¯1+x¯12+8x¯2x¯.

The proposed procedure is as follows: Select exponential distribution if T>0, otherwise select Lindley distribution. T>0 implies that numerator of RML is larger than the denominator. This means that the sample observations are more likely to have been generated from exponential distribution than from Lindley distribution and hence we choose exp(α) to model the given data. If T<0, then denominator of RML will be larger than the numerator, which implies Li (β) is a better fit to the data.

To get an insight into the performance of the discrimination procedure, PCS is used. PCS, like the notion of power in hypothesis testing, gives the probability of selecting the null distribution under consideration. In the current context, PCS is defined as

Pexp(Li)Lexp(α^)>(<)LLi(β^)|data is from exp(α) (Li(β)).

In terms of T, PCS can be defined as

Pexp(T>0)=PT>0|data is from exp(α).

Similarly, in the case of Lindley distribution, PCS is given as

PLi(T<0)=PT<0|data is from Li(β).

3 ASYMPTOTIC PROPERTIES OF THE RML UNDER NULL HYPOTHESES

In this section, asymptotic distribution of the logarithm of RML under null hypotheses is obtained under two different cases.

Case 1: The null hypothesis is that given data is from exp(α) and the alternative is that given data is from Li(β). The following lemma is helpful in finding the asymptotic distribution of T.

Lemma 1.

Under the assumption that data is from exp(α), as n

  1. α^α a.s.

  2. β^β a.s, where β is that value of β which maximizes Eexp(logfLi(x;β)).

  3. Let T=log(Lexp(α)LLi(β)). Then TEexp(T)n is asymptotically equivalent to (TEexp(T))n.

Proof.

  1. Using the convergence property of MLE, it is clear that under null hypothesis α^α a.s.

  2. Under the null hypothesis, it is observed that β^ is a pseudo MLE of β. Hence β^β. However β^β a.s, where β is the value of β that minimizes Kullback–Leibler information criteria. i.e., β is obtained by minimizing

    Eexplogfexp(x;α)fLi(x;β).

    It is clear that minimizing the above expectation is same as maximizing Eexp(logfLi(x;β)) and calculus approach yields β as

    β=α1+1+6α+α22.

  3. We have to show that [(TEexp(T))n][(TEexp(T))n]0 in probability in order to show that TEexp(T)n and (TEexp(T))n are asymptotically equivalent. Since α^α a.s and β^β a.s, and applying 2c4 XIII in Radhakrishna Rao [15], we get

    T=logi=1nfexp(xi;α^)i=1nfLi(xi;β^)=i=1nlogfexp(xi;α^)i=1nlogfLi(xi;β^)i=1nlogfexp(xi;α)i=1nlogfLi(xi;β)=T in probability.

    Also by applying corollary 5.7 in Bartle [16], we get

    E(T)=xlogi=1nfexp(xi;α^)i=1nfLi(xi;β^)dFexp(x)xlogi=1nfexp(xi;α)i=1nfLi(xi;β)dFexp(x)=E(T)in probability.

    Therefore, TEexp(T)n is asymptotically equivalent to (TEexp(T))n.

Theorem 1.

Under the assumption that the data is from a exp(α), T is asymptotically distributed as normal with mean Eexp(T) and variance Vexp(T).

Proof.

Using central limit theorem, it is observed that (TEexp(T))n is asymptotically distributed as normal with variance nVexpT. By part (c) of Lemma 1, it is clear that TEexp(T)n is asymptotically equivalent to (TEexp(T))n. Therefore, TEexp(T)n is also asymptotically distributed as normal with variance nVexp(T).

The expressions for asymptotic mean and variance are obtained as,

limnEexp(T)n=logα12logβeαΓ(0,α)+βα+log(1+β),
limnVexp(T)n=2βαG2,33,1α1,01,1,1eαΓ(0,α)α+eα2G2,33,0α1,10,0,0eαΓ(0,α)2+βα2α2,
where G and Γ denote Meijer G-function and incomplete gamma function respectively. Both these special functions can be evaluated using “gsl” package in R program. Using (10) and (11), asymptotic PCS can be obtained as
Pexp(T>0)=1ΦEexp(T)Vexp(T),
where Φ denotes c.d.f of the standard normal distribution.

Case 2: The null hypothesis is that given data is from Li(β) and the alternative hypothesis is that given data is from exp(α). The following lemma is needed to find the asymptotic distribution of T.

Lemma 2.

Under the assumption that data is from Li(β), as n

  1. β^β a.s.

  2. α^α a.s, where α is that value of α which maximizes ELi(logfexp(x;α)).

  3. Let T=log(Lexp(α)LLi(β)). Then TELi(T)n is asymptotically equivalent to (TELi(T))n.

Proof.

The proof is similar to that of Lemma 1 with α given by

α=β(β+1)β+2.

Theorem 2.

Under the assumption that the data is from a Li(β), T is asymptotically distributed as normal with mean ELi(T) and variance VLi(T).

Proof.

By central limit theorem, it can be shown that (TELi(T))n is asymptotically distributed as normal with variance nVLi(T). By part (c) of Lemma 2, it is clear that TELi(T)n is asymptotically equivalent to (TELi(T))n. Therefore, TELi(T)n is also asymptotically distributed as normal with variance nVLi(T).

The expressions for asymptotic mean and variance under case 2 are obtained as

limnELi(T)n=1β(β+1)G2,33,1β1,21,1,3+α(β+2)+ββlogαlogαββlog(β+1)+2(β+1)log(β)log(β+1)+eββΓ(0,β)2,
limnVexp(T)n=1(β+1)22βαβ2(β+1)G2,33,1β3,23,3,0+G2,33,1β2,12,2,0(β+2)G2,33,1β1,01,1,1+eβΓ(0,β)+2eβ(β+1)G3,44,0β1,1,10,0,0,2G2,33,1β0,10,0,2+eββΓ(0,β)2+β2+4β+2βα2β2.

Using (14) and (15), asymptotic PCS can be obtained as

PLi(T<0)=ΦELi(T)VLi(T).

4. SIMULATION STUDY

This section deals about how asymptotic results work for finite sample sizes using simulated data sets. PCS based on both simulation and asymptotic results is calculated for different choice of parameter values of the null distributions. R programming language version 3.3.2 is used to perform numerical computations. The algorithm for calculating PCS based on simulation under case 1 is given below.

  1. Generate a sample of size n from exp(α).

  2. Compute MLEs for both α and β as given in (6).

  3. Compute T using (5).

  4. Repeat the steps 1–3 for R times and count the number of positive T's.

An estimate of PCS based on simulation can be computed as

PCS^=Number of positive TsR.

In this study, R is taken to be 1000. PCS computed using simulation and asymptotic results using (12) under case 1 for different choice of parameter values namely, α=0.2,0.4,0.8,1,1.5,3,6 and different samples sizes n=20,40,60,80,100 is presented in Table 1. The entries in parenthesis represent asymptotic PCS values.

α\n

20 40 60 80 100
0.2 0.657(0.733) 0.757(0.810) 0.821(0.859) 0.881(0.893) 0.921(0.918)
0.4 0.578(0.672) 0.695(0.735) 0.731(0.780) 0.779(0.813) 0.825(0.840)
0.8 0.509(0.511) 0.550(0.516) 0.653(0.519) 0.658(0.522) 0.733(0.525)
1 0.479(0.592) 0.572(0.630) 0.587(0.657) 0.620(0.680) 0.692(0.700)
1.5 0.454(0.564) 0.506(0.590) 0.507(0.610) 0.569(0.627) 0.591(0.641)
3 0.375(0.530) 0.416(0.542) 0.464(0.552) 0.465(0.560) 0.527(0.567)
6 0.383(0.512) 0.410(0.517) 0.405(0.520) 0.460(0.524) 0.433(0.526)
Table 1

PCS based on simulation and asymptotic results under case 1.

For case 2, PCS based on simulation can be obtained using the following algorithm.

  1. Generate a sample of size n from Li(β).

  2. Compute MLEs for both α and β as given in (6).

  3. Compute T using (5).

  4. Repeat the steps 1–3 for R times and count the number of negative T's.

An estimate of PCS based on simulation can be computed as

PCS^=Number of negative TsR.

Here R is taken as 1000. PCS based on simulation and asymptotic results under case 2 are calculated for different parameters choices namely, β=0.2,0.4,0.8,1,1.5,3,6 and different samples sizes n=20,40,60,80,100 is presented in Table 2.

β\n

20 40 60 80 100
0.2 0.855(0.804) 0.898(0.886) 0.930(0.938) 0.951(0.956) 0.974(0.972)
0.4 0.820(0.740) 0.851(0.815) 0.902(0.864) 0.907(0.898) 0.935(0.922)
0.8 0.740(0.660) 0.771(0.717) 0.809(0.759) 0.840(0.793) 0.831(0.819)
1 0.719(0.630) 0.756(0.684) 0.763(0.722) 0.780(0.751) 0.797(0.776)
1.5 0.670(0.590) 0.683(0.628) 0.727(0.655) 0.732(0.678) 0.735(0.697)
3 0.656(0.540) 0.649(0.557) 0.644(0.569) 0.626(0.580) 0.622(0.589)
6 0.651(0.510) 0.632(0.520) 0.618(0.530) 0.598(0.529) 0.618(0.532)
Table 2

PCS based on simulation and asymptotic results under case 2.

From Tables 1 and 2, it is clear that PCS based on simulation and asymptotic results do not vary much even for moderate sample sizes.

5 REAL LIFE ILLUSTRATION

In this section, the proposed discrimination procedure is illustrated using two real life data sets.

Data set 1: The data set represents the failure times in days of the air conditioning system of 30 aeroplanes (Linhart and Zucchini [17]). For this data, MLEs are calculated using (6) and respective log likelihoods and T are obtained as

In Lexpα^=132.1914,In LLiβ^=127.3682,T=4.8232.

Since T<0, it is more appropriate to use Lindley distribution to model this data. To calculate estimated PCS value, 1000 bootstrap samples are generated from the data and the estimated PCS is obtained as 0.9363. This means that the chance of selecting the wrong model using the proposed procedure is only 6.37%. Next, p value for the observed T based on asymptotic results is calculated under case 1 and case 2. First, exponential distribution is considered as the null distribution. Using (10) and (11), asymptotic mean and variance of T are obtained as 4.4184 and 12.6784 respectively. Using (12), the observed p value is found to be 0.0047. For case 2, using (14) and (15), asymptotic mean and variance are obtained as 3.0483 and 62.9556 respectively. Using (16), the observed p value is found to be 0.5885. It is clear that p values in both the cases favor using Lindley distribution to model the data.

Data set 2: This data set consists of waiting time in minutes of 100 customers in a bank (Ghitany et al. [2]). For this data set, MLEs are calculated using (6) and respective log likelihoods and T are obtained as

In Lexpα^=72.52,In LLiβ^=71.6543,T=0.8745.

Since T<0, Lindley distribution can be used to model the data. To calculate PCS value, 1000 bootstrap samples are generated from the data and the estimated PCS is obtained as 0.8009. This implies that the chance of selecting the wrong model using the proposed procedure is 19.91%.

Now p value for the observed T using asymptotic results is calculated under both the cases. First, exponential distribution is considered as the null distribution. Using (10) and (11) asymptotic mean and variance of T are found to be 2.3021 and 5.8099 respectively. Using (12), the corresponding p value is found to be 0.0937. For case 2, Lindley distribution is considered as the null distribution and using (14) and (15), the values of asymptotic mean and variance of T are calculated as 1.8359 and 54.3614 respectively. Using (16), p value for the observed T is found to be 0.4481. From p values, it is clear that Lindley distribution is the better fit to the data.

6 CALCULATION OF MINIMUM SAMPLE SIZE

In Section 4, it is found that PCS based on simulation and asymptotic results do not vary much even for moderate sample sizes. Therefore, using the asymptotic results, a method to calculate the minimum sample size required to discriminate between exponential and Lindley distribution for a user specified PCS and threshold level is proposed in this section. The distance between two distributions can be measured using Kolmogorov–Smirnov distance, Hellinger distance, etc. It is intuitive that if two distributions are very close (far away), large (small) sample is required to discriminate between them. Thus, in practice, it is of interest to find the minimum sample size required to discriminate between two distributions for a user specified PCS and threshold level based on some distance measure.

The procedure for finding minimum sample size required to discriminate between the two distributions under case 1 is given as follows. Assume the threshold level to be D and a PCS of at least p. Using (12), we get

1ΦEexp(T)Vexp(T)>=pΦnEexp(T)nnVexp(T)n>=1p.

Taking Φ1 on both the sides, squaring and rearranging the terms, the minimum sample size (n) is obtained as

n=Φ1(1p)2Vexp(T)nEexp(T)n2,
where Φ1 is the quantile function of the standard normal distribution and Eexp(T)n and Vexp(T)n are obtained from (10) and (11) respectively.

In a similar manner, minimum sample size n required under case 2 is obtained as

n=Φ1(p)2VLi(T)nELi(T)n2,
where ELi(T)n and VLi(T)n correspond to (14) and (15) respectively. As a numerical illustration, minimum sample size required to discriminate between exponential and Lindley distributions under both cases is calculated for PCS p=0.75 and at different choices of parameter values and is given in Tables 3 and 4. In both the cases, Hellinger distance is used as the distance measure.

α 0.2 0.4 0.8 1 1.5 3 6
n 23 27 90 145 345 1574 9971
Hellinger distance 0.103 0.0732 0.0457 0.038 0.0262 0.0121 0.0047
Table 3

Minimum sample size under case 1.

β 0.2 0.4 0.8 1 1.5 3 6
n 18 33 81 117 252 1306 9860
Hellinger distance 0.1272 0.0965 0.0627 0.0525 0.0359 0.0159 0.0058
Table 4

Minimum sample size under case 2.

Suppose a practitioner wants to discriminate between exponential and Lindley distributions for a user specified threshold level, say D=0.04 of the Hellinger distance and PCS p=0.75. From Table 3, it can be seen that Hellinger distance is greater than 0.04 if α<1. Therefore, if the null distribution is exponential, then for tolerance level D=0.04, one needs to draw a minimum sample of size n=90 in order to meet the PCS level p=0.75. Similarly, from Table 4, it can be observed that Hellinger distance is greater than 0.04 if β1. Therefore, if the null distribution is Lindley, then for tolerance level D=0.04, the user needs to draw a minimum sample of size n=117 in order to meet the PCS level p=0.75.

7 CONCLUSION

In this paper, a discrimination procedure is developed using RML to discriminate between exponential and Lindley distributions based on a given random sample of observations. Also, the asymptotic distribution of the logarithm of the RML statistic is derived and asymptotic expression for PCS is obtained. From simulation study and real life application, it is found that even for moderate sample size the discrimination procedure works quite well. Subsequently, a procedure to calculate minimum sample size required to discriminate between the distributions for given PCS value and threshold level is developed using the asymptotic results. From numerical results, it is observed that for large parameter values, Hellinger distance between the two distributions decreases, thereby requiring more samples to discriminate between them.

CONFLICT OF INTEREST

The authors declare no conflict of interest.

ACKNOWLEDGMENTS

The authors thank the editor and anonymous referees for their useful suggestions and comments that helped in improving the article.

REFERENCES

3.H. Zakerzadeh and A. Dolati, J. Math. Ext., Vol. 3, 2009, pp. 13-25.
10.D.R. Cox, in Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability (California), Vol. 2, 1961, pp. 105-123.
17.H. Linhart and W. Zucchini, Model Selection, Wiley, New York, 1986.
Journal
Journal of Statistical Theory and Applications
Volume-Issue
18 - 3
Pages
295 - 302
Publication Date
2019/09/03
ISSN (Online)
2214-1766
ISSN (Print)
1538-7887
DOI
10.2991/jsta.d.190818.006How to use a DOI?
Copyright
© 2019 The Authors. Published by Atlantis Press SARL.
Open Access
This is an open access article distributed under the CC BY-NC 4.0 license (http://creativecommons.org/licenses/by-nc/4.0/).

Cite this article

TY  - JOUR
AU  - V. S. Vaidyanathan
AU  - A Sharon Varghese
PY  - 2019
DA  - 2019/09/03
TI  - Discriminating Between Exponential and Lindley Distributions
JO  - Journal of Statistical Theory and Applications
SP  - 295
EP  - 302
VL  - 18
IS  - 3
SN  - 2214-1766
UR  - https://doi.org/10.2991/jsta.d.190818.006
DO  - 10.2991/jsta.d.190818.006
ID  - Vaidyanathan2019
ER  -