Figures
Abstract
The estimation of the entropy of a random system or process is of interest in many scientific applications. The aim of this article is the analysis of the entropy of the famous Kumaraswamy distribution, an aspect which has not been the subject of particular attention previously as surprising as it may seem. With this in mind, six different entropy measures are considered and expressed analytically via the beta function. A numerical study is performed to discuss the behavior of these measures. Subsequently, we investigate their estimation through a semi-parametric approach combining the obtained expressions and the maximum likelihood estimation approach. Maximum likelihood estimates for the considered entropy measures are thus derived. The convergence properties of these estimates are proved through a simulated data, showing their numerical efficiency. Concrete applications to two real data sets are provided.
Citation: Al-Babtain AA, Elbatal I, Chesneau C, Elgarhy M (2021) Estimation of different types of entropies for the Kumaraswamy distribution. PLoS ONE 16(3): e0249027. https://doi.org/10.1371/journal.pone.0249027
Editor: Feng Chen, Tongii University, CHINA
Received: November 27, 2020; Accepted: March 9, 2021; Published: March 30, 2021
Copyright: © 2021 Al-Babtain et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the manuscript.
Funding: This project is supported by Researchers Supporting Project number (RSP-2020/156) King Saud University, Riyadh, Saudi Arabia. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
Information theory provides natural mathematical tools for measuring the uncertainty of random variables and the information shared by them. In this regard, entropy and mutual information are two fundamental concepts. More precisely, the probability distribution of a random variable is associated with some sort of uncertainty, and entropy is used to quantify it. The concept of entropy was formerly proposed by [1]. Since that publication, many areas of study such as statistics, neurobiology, cryptography, bioinformatics, quantum computer science and linguistics, have developed various entropy-based measures. Modern and exhaustive reviews on the ‘entropy universe’ can be found in [2–6].
In applied probability and statistics, many authors have conducted their studies for diverse and important distributions based on entropy. The essential references in this regard are briefly presented below. Reference [7] used the concept of entropy to communicate on the probability distribution of electric charge between atoms observed in a certain condition. Reference [8] derived the entropy for the Feller-Pareto family and presented the entropy ordering property for some related sample minimum and maximum. Reference [9] estimated the entropy of the Weibull distribution by considering different loss functions based on a generalized progressively hybrid censoring scheme. Reference [10] discussed the entropy for the generalized half-logistic distribution based on the type II censored samples. References [11] and [12] proposed estimates for the entropy of absolutely continuous random variables. Reference [13] presented an indirect method using a decomposition to simplify the entropy’s calculation under the progressive type II censoring. Reference [14] derived a nonparametric kernel estimator for the general Shannon entropy. Reference [15] estimated the entropy for several exponential distributions and extended the results to other circumstances. Reference [16] estimated the Shannon entropy of the Rayleigh model under doubly generalized type-II hybrid censoring, and evaluated its performance by two criteria. Reference [17] derived a nonparametric wavelet estimator for the general Shannon entropy. Reference [18] provided an exact expression for entropy information contained in both types of progressively hybrid censored data and applied it in the setting of the exponential distribution. Reference [19] investigated entropy measures for weighted and truncated weighted exponential distributions. Reference [20] presented the estimation of entropy for inverse Weibull distribution under multiple censored data. Reference [21] introduced estimation of entropy for inverse Lomax distribution under the multiple censored scheme. Reference [22] examined Bayesian and non-Bayesian methods to estimate the dynamic cumulative residual Rényi entropy for the Lomax distribution.
Surprisingly, to our knowledge, the entropy of the famous Kumaraswamy distribution has not been studied in depth. In this article, we fill this gap both probabilistically and statistically. The specificities and interests of the Kumaraswamy distribution are described below. First, it was introduced by [23], and was motivated as an alternative to the beta distribution which are (i) mathematically simpler, without special function in particular, and (ii) more suited to the modeling of various hydrological phenomena observed at low frequency (daily rainfall, daily flow of rivers, etc.). Mathematically, the probability density function (pdf) of the Kumaraswamy distribution is specified by
(1)
with f(x;a, b) = 0 otherwise, where a, b > 0. This pdf is unimodal if a, b > 1, uniantimodal if a, b < 1, increasing if a > 1, b ≤ 1, decreasing if a ≤ 1, b > 1 or constant if a = b = 1, in the same way as the beta distribution. The corresponding cumulative distribution and quantile functions are quite simple; they are defined without special function contrary to those of the beta distribution. Special cases of the Kumaraswamy distribution correspond to the distribution of minimum or maximum of uniform samples. We may refer the reader to [24] for all the known features of this distribution. Also, the kumaraswamy distribution has generated many flexible distributions with various domains and number of parameters through the generalized Kumaraswamy class elaborated by [25].
In a sense, this study complements the work of [24] by investigating the overall concept of entropy of the Kumaraswamy distribution, which has never been studied before. More precisely, we consider six well-referenced entropy measures. We derive their analytical expressions by using the well-known beta function. We compare them numerically by considering different parameter values. Then, we propose an efficient strategy based on the maximum likelihood approach to estimate these entropy measures. A simulation study is done to see how effective our strategy is. Graphical and numerical comparisons are performed. We end the study by two illustrative examples on real data sets, showing how the methodology can be applied in a concrete statistical setting.
The following sections make up the document. Section 2 presents a result on a special integral, and shows how it is related to important entropy measures of the Kumaraswamy distribution. Numerical values of these entropy measures with different values of the parameters are also given. Section 3 studies the estimation of these entropy measures. Then, using generated values from the Kumaraswamy distribution, graphical and numerical comparisons are discussed. The entropy of the random characteristics behind two real data sets is investigated. Finally, conclusions are presented in section 4.
2 Entropy of the Kumaraswamy distribution
2.1 An integral result
The following result shows that a certain integral involving the pdf of the Kumaraswamy distribution can be expressed in terms of the classical beta function. The connection between this integral and the considered entropy measures will be developed later.
Proposition 1 Let δ > 0, f(x;a, b) be specified by Eq (1) and
Then, Iδ(a, b) exists if and only if min(a, b)>max(1 − 1/δ, 0), and it is expressed as where B(u, v) denotes the classical beta function, that is
for u, v > 0.
Proof. Owing to Eq (1), we have
When x tends to 0, we have xδ(a−1)(1 − xa)δ(b−1) ∼ xδ(a−1), which is integrable in the neighborhood of 0 if and only if δ(1 − a)<1 by the Riemann integral criteria. Similarly, when x tends to 1, we have
which is integrable in the neighborhood of 1 if and only if δ(1 − b)<1 by the Riemann integral criteria. In summary, Iδ(a, b) exists if and only if δmax(1 − a, 1 − b)<1, which is equivalent to min(a, b)>1 − 1/δ. Now, under this assumption, by applying the change of variables y = xa, that is
with
, we obtain
This ends the proof of Proposition 1.
In fact, the beta function is implemented in most of the mathematical software (see the function beta of the package stat of R, the Beta function of Mathematica, etc.). Therefore, thanks to Proposition 1, the computation of Iδ(a, b) can be done quite efficiently with little effort. Also, the existing results on the beta functions allow a mathematical control of this integral. Some related results are presented below.
- Through the use of the standard Euler gamma function given as
, one can write
- Also, assuming that δ(1 − 1/a) + 1/a and δ(b − 1) + 1 are positive integers, the following formula holds:
- By virtue of the main result in [26], if δ(a − 1) ≥ a − 1 and b ≥ 1, then we have
with the best possible constants α = 0 and β = 0.08731…. Therefore, for not too large value of δ, the following numerical approximation seems acceptable:
In our study, the interest of Proposition 1 is that Iδ(a, b) is the main ingredient in the definitions of various entropy measures of the Kumaraswamy distribution, as developed in the next part.
2.2 Various entropy measures
The entropy of the Kumaraswamy distribution can be measured in different manners. The most useful entropy measures of the literature are recalled in Table 1 for a general distribution with pdf denoted by f(x;φ), φ representing a possible vector of parameters. Also, we suppose that δ > 0 and δ ≠ 1 as basic assumptions in this general case.
For the two entropy measures proposed by [27], it is supposed that is finite and well identified.
From Table 1, we see that the integral is central to determine the considered entropy measures. Now, we present the corresponding entropy measures of the Kumaraswamy distribution. Based on Proposition 1, it is supposed that a, b and δ satisfy min(a, b)>max(1 − 1/δ, 0).
Rényi entropy.
Based on Table 1, Eq (1) and Proposition 1, the Rényi entropy of the Kumaraswamy distribution can be expressed as
Havrda and Charvát entropy.
From Table 1, Eq (1) and Proposition 1, the Havrda and Charvát entropy of the Kumaraswamy distribution can be expressed as
Arimoto entropy.
Again, from Table 1, Eq (1) and Proposition 1, the Arimoto entropy of the Kumaraswamy distribution is specified by
Tsallis entropy.
Based on Table 1, Eq (1) and Proposition 1, the Tsallis entropy of the Kumaraswamy distribution can be expressed as
Awad and Alawneh 1 entropy.
From Table 1, Eq (1) and Proposition 1, the Awad and Alawneh 1 entropy of the Kumaraswamy distribution is given as
(2)
Before going further, we need to determine sup0<x<1 f(x;a, b). The following lemma provides the necessary in this regard.
Lemma 2 Let f(x;a, b) be given as Eq (1). Then, sup0<x<1 f(x;a, b) is finite if and only if a ≥ 1 and b ≥ 1 with ab ≠ 1, and in this case, we have
Therefore, f′(x*;a, b) = 0 implies that
Since f′(x;a, b)>0 for x < x* and f′(x;a, b)<0 for x > x*, x* is a maximum point for f(x;a, b). Hence,
Note that, for a = 1, with the convention 00 = 1, we have f(x*;a, b) = b(b − 1)b−1(b − 1)1−b = b and for b = 1, we have . This ends the proof of Lemma 2.
Based on Lemma 2, if a > 1 and b > 1, Eq (2) becomes
Awad and Alawneh 2 entropy.
From Table 1, Eq (1), Proposition 1 and Lemma 2, the Awad and Alawneh 2 entropy of the Kumaraswamy distribution is given as
Theoretically, it is complicated to study the behavior of these entropy measures. For this reason, a numerical study is proposed in the next section.
2.3 Numerical values
We now investigate the numerical values for the six entropy measures presented in Subsection 2.2 under the following configuration of the parameters: Configuration 1: a = 2, b ∈ Υ with Υ = {1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0, 5.5, 6.0} and δ = 0.5, Configuration 2: a = 2, b ∈ Υ and δ = 1.5, Configuration 3: a = 2, b ∈ Υ and δ = 2.5, Configuration 4: a ∈ Υ, b = 2 and δ = 0.5, Configuration 5: a ∈ Υ, b = 2 and δ = 1.5, and Configuration 6: a ∈ Υ, b = 2 and δ = 2.5. The findings of all the six entropy measures are presented for these configurations in Tables 2–7, respectively.
In view of Tables 2–7, the following comments can be formulated.
First, we recall that Tables 2–4 indicate the values of the entropy measures of the Kumaraswamy distribution for a fixed value of a and different values for b and δ. In this context,
- the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when b is increasing while the Awad and Alawneh 2 entropy is increasing when b is increasing.
- the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing.
Tables 5–7 show the values of the entropy of the Kumaraswamy distribution for a fixed value of b and different values for a and δ. In this setting,
- the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when a is increasing while the Awad and Alawneh 2 entropy is increasing when a is increasing.
- the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing.
3 Maximum likelihood estimation
The inference on the six considered entropy measures of the Kumaraswamy distribution is now investigated via the maximum likelihood technique. This technique is well-known and has proved itself in various modern studies such as those in [32–34].
3.1 Estimation of the entropy measures
The estimation of the parameters of the Kumaraswamy model through the maximum likelihood technique is well-known and the details can be found in [24]. The minimal theory is recalled below. Based on n values x1, …, xn supposed to be observed from a random variable X with the Kumaraswamy distribution with parameters a and b, the maximum likelihood estimates (MLEs) of a and b, say and
, are defined by
where ℓ(a, b) denotes the log-likelihood function specified by
These MLEs are also the solutions of the two following equations according to a and b:
That is, and
satisfy the following simple relation:
Then, the properties of these MLEs follow from the usual maximum likelihood theory. In particular, thanks to the functional invariance of the MLEs, one can deduce easily the MLEs of the entropy measures. More concretely, based on the six entropy measures described in Subsection 2.2, is the MLE of Rδ(a, b),
is the MLE of HCδ(a, b),
is the MLE of Aδ(a, b),
is the MLE of Tδ(a, b),
is the MLE of AA1δ(a, b), and
is the MLE of AA2δ(a, b).
3.2 Simulation
We now investigate the numerical behavior of the MLEs of the entropy measures via the use of simulated values. That is, we consider N = 5000 samples of values from a random variable X with the Kumaraswamy distribution of parameters a and b with different samples sizes; n = 100, 200, 300 and 1000 are considered. The following configurations on the parameters are considered: Configuration1: a = 3, b = 3 and δ ∈ Ξ with Ξ = {0.5, 1.5, 2.5}, and Configuration 2: a = 3, b = 5 and δ ∈ Ξ.
In each configuration, for each sample, the MLEs and
are determined. Then, based on the N samples of fixed size, we determine the average of the N MLEs and use it to define the entropy estimates. The corresponding mean squared error (MSE) and mean deviation (MD) defined by the following generic formulas: MSE = sum(exact value—estimate)2 / N and MD = sum abs(exact value—estimate) / N, respectively, are also calculated. These assessment criteria are often used quite effectively to make a full comparison of models. In this regard, we can refer the reader to the useful works of [35–37].
The results on the Rényi entropy under Configurations 1 and 2 are given in Tables 8 and 9, respectively, results on the Havrda and Charvat entropy under Configurations 1 and 2 are indicated in Tables 10 and 11, respectively, results on the Arimoto entropy under Configurations 1 and 2 are presented in Tables 12 and 13, respectively, results on the Tsallis entropy under Configurations 1 and 2 are given in Tables 14 and 15, respectively, results on the Awad and Alawneh 1 entropy under Configurations 1 and 2 are given in Tables 16 and 17, respectively, and results on the Awad and Alawneh 2 entropy under Configurations 1 and 2 are indicated in Tables 18 and 19.
Based on Tables 8–19, in all the situations, we see that the MLEs of the entropy measures are close to the target values and, as anticipated, the MSEs and MDs decrease and approach 0 as n increases. This proves the accuracy of the proposed estimation methods in the context of the Kumaraswamy distribution. Also, one can notice that the MSEs and MDs increase as δ increases.
For a visual approach, the behavior of the MSEs and MDs are illustrated in Figs 1–12, for the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 1 and Awad and Alawneh 2 entropy measures following the settings of Tables 8–19, respectively.
Figs 1–12 support the claims formulated about the results of Tables 8–19.
3.3 Illustrative examples
In this Section, two real life data sets are used to illustrate the proposed methodology. The considered data sets are described below.
The first data set.
The data set consists of 48 rock samples from an oil reservoir. It corresponds to twelve oil tank cores that were sampled by four cross sections. Each core was measured for permeability and each cross section has the following variables: total pore area, total pore perimeter, and shape. We analyze the perimeter of the shape by a squared variable (area). It has been analyzed by [38], among others. Explicitely, the data set is: {0.0903296, 0.2036540, 0.2043140, 0.2808870, 0.1976530, 0.3286410, 0.1486220, 0.1623940, 0.2627270, 0.1794550, 0.3266350, 0.2300810, 0.1833120, 0.1509440, 0.2000710, 0.1918020, 0.1541920, 0.4641250, 0.1170630, 0.1481410, 0.1448100, 0.1330830, 0.2760160, 0.4204770, 0.1224170, 0.2285950, 0.1138520, 0.2252140, 0.1769690, 0.2007440, 0.1670450, 0.2316230, 0.2910290, 0.3412730, 0.4387120, 0.2626510, 0.1896510, 0.1725670, 0.2400770, 0.3116460, 0.1635860, 0.1824530, 0.1641270, 0.1534810, 0.1618650, 0.2760160, 0.2538320, 0.2004470}.
The second data set.
This data set contains 20 observations of flood data. It was analyzed by [39]. The data set is listed as follows: {0.265, 0.392, 0.297, 0.3235, 0.402, 0.269, 0.315, 0.654, 0.338, 0.379, 0.418, 0.423, 0.379, 0.412, 0.416, 0.449, 0.484, 0.494, 0.613, 0.74}.
In order to check the adequateness of the Kumaraswamy distribution to these data, we apply the Kolmogorov-Smirnov test. We find p-value 0.2092 and p-value = 0.3359 for the first and second data sets, respectively. Since both satisfy p-values >0.05, the two considered data set are in adequateness with the Kumaraswamy distribution.
Now, Tables 20 and 21 present the estimations of the six entropy measures considered in Subsection 2.2, following the methodology described in Subsection 3.1, for the first and second data sets, respectively.
We can notice that, under our framework, the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 2 entropy measures are decreasing when δ is increasing while the Awad and Alawneh 1 entropy is increasing when δ is increasing.
Tour knowledge, it is the first time that the entropy of the uncertainty behind these data sets are evaluated. They can be taken into account for further statistical analysis in the future.
4 Conclusion
For the first time, this article proposed a special focus on the entropy of the Kumaraswamy distribution. Both theoretical and practical aspects were covered, though complementary works. In particular, six different entropy measures were investigated. After determining the closed-form expressions of these measures, an estimation strategy was developed to evaluate them in a practical setting. A simulation study ensured the convergence of the obtained estimates. Two real-life data sets are used to show how the related entropy can be concretely estimated. The finding of this study aims to be applied by the statistician to assess the entropy of diverse data with values on the unit interval, such as modern rate, percentage and proportion type data.
The limitation of current research remains on the classicity of the statistical framework considered. Directions for future research include the estimation of the entropy of the Kumaraswamy distribution in more sophisticated statistical schemes with physical motivations, such as the progressive type II censoring scheme, generalized progressively hybrid censoring scheme, etc., or taking into account generalized versions of the Kumaraswamy distribution, such as the one proposed by [40].
References
- 1. Shannon CE. A mathematical theory of communication. The Bell System Technical Journal. 1948, 27, 1, 379–423.
- 2. Golan A. Information and entropy econometrics-a review and synthesis. Foundations and Trends in Econometrics. 2008, 2, 1–45.
- 3.
Cover TM and Thomas JA. Elements of Information Theory. Hoboken, NJ: John Wiley & Sons, 2012.
- 4. Amigo JM, Balogh SG and Hernandez S. A brief review of generalized entropies. Entropy, 2018, 20, 813. pmid:33266537
- 5. Namdari A and Li Z. A review of entropy measures for uncertainty quantification of stochastic processes. Advances in Mechanical Engineering. 2019, 11, 6, 1–14.
- 6. Ribeiro M, Henriques T, Castro L, Souto A, Antunes L, Costa-Santos C, et al. The entropy universe. Entropy. 2021, 23, 222. pmid:33670121
- 7. Siamak N and Ehsan S. Shannon entropy as a new measure of aromaticity. Shannon aromaticity, Physical Chemistry Chemical Physics. 2010, 12, 4742–4749.
- 8. Tahmasebi S and Behboodian J. Shannon entropy for the Feller-Pareto (FP) family and order statistics of FP subfamilies. Applied Mathematical Sciences. 2010, 4, 10, 495–504.
- 9. Cho Y, Sun H and Lee K. Estimating the entropy of a Weibull distribution under generalized progressive hybrid censoring, Entropy. 2015, 17, 102–122.
- 10. Seo JI and Kang SB. Entropy estimation of generalized half-logistic distribution (GHLD) based on type-II censored samples. Entropy. 2014, 16, 1424–1433.
- 11. Ebrahimi N, Pflughoef K and Soofi E. Two measures of sample entropy. Statistics and Probability Letters. 1994, 20, 225–234.
- 12. Alizadeh Noughabi H. A new estimator of entropy and its application in testing normality. Journal of Statistical Computation and Simulation. 2010, 80, 10, 1151–1162.
- 13. Abo-Eleneen ZA. The entropy of progressively censored samples, Entropy. 2011, 13, 437–449.
- 14.
Bouzebda S and Elhattab I. Uniform-in-bandwidth consistency for kernel-type estimators of Shannon’s entropy. Electronic Journal of Statistics. 2011, 5, 440–459.
- 15. Kayal S and Kumar S. Estimation of the Shannon’s entropy of several shifted exponential populations. Statistics and Probability Letters. 2013. 83, 1127–1135.
- 16. Cho Y, Sun H and Lee K. An estimation of the entropy for a Rayleigh distribution based on doubly- generalized type-II hybrid censored samples. Entropy. 2014, 16, 3655–3669.
- 17. Chesneau C, Navarro F and Serea O. A note on the adaptive estimation of the differential entropy by wavelet methods. Commentationes Mathematicae Universitatis Carolinae. 2017, 58,1, 87–100.
- 18. Bader A. On the entropy of progressive hybrid censoring schemes. Applied Mathematics & Information Sciences. 2017, 11, 6, 1811–1814.
- 19. Basit A, Riaz A, Iqbal Z and Ahmad M. On comparison of entropy measures for weighted and truncated weighted exponential distributions. Advances and Applications in Statistics. 2017, 50, 6, 477–495.
- 20. Hassan AS and Zaky AN. Estimation of entropy for inverse Weibull distribution under multiple censored data. Journal of Taibah University for Science. 2019, V, 13, 1, 331–337.
- 21. Bantan RAR, Elgarhy M, Chesneau C and Jamal F. Estimation of entropy for inverse Lomax distribution under multiple censored data. Entropy, 2020, 22, 6, 601, 1–15. pmid:33286373
- 22. Al-Babtain AA, Hassan AS, Zaki AN, Elbatal I and Elgarhy M. Dynamic cumulative residual Rényi entropy for Lomax distribution: Bayesian and non-Bayesian methods. AIMS Mathematics, 2021, 6, 4, 3889–3914.
- 23. Kumaraswamy P. Generalized probability density-function for double-bounded random-processes. Journal of Hydrology. 1980, 46, 79–88.
- 24. Jones MC. Kumaraswamy’s distribution: A beta-type distribution with some tractability advantages, Statistical Methodology. 2009, 6, 1, 70–91.
- 25. Cordeiro GM and de Castro M. A new family of generalized distributions. Journal of Statistics Computation and Simulation. 2011, 81, 883–893.
- 26. Alzer H. Sharp inequalities for the beta function. Indagationes Mathematicae, 2001, 12, 15–21.
- 27. Awad AM and Alawneh AJ. Application of entropy to a life-time model. IMA Journal of Mathematical Control and Information. 1987, 4, 2, 143–148.
- 28.
Rényi A. On measures of entropy and information. Proc. 4th Berkeley Symposium on Mathematical Statistics and Probability. 1960, 1, 47-561.
- 29. Havrda J and Charvát F. Quantification method of classification processes, concept of structural a-entropy. Kybernetika. 1967, 3, 1, 30–35.
- 30. Arimoto S. Information-theoretical considerations on estimation problems. Information and Control. 1971, 19, 3, 181–194.
- 31. Tsallis C. Possible generalization of Boltzmann-Gibbs statistics. Journal of Statistical Physics, 1988, 52, (1-2), 479–487.
- 32.
Dong B, Ma X, Chen F and Chen S. Investigating the differences of single- and multi-vehicle accident probability using mixed logit model. Journal of Advanced Transportation. 2018, Article ID 2702360, 9 pages.
- 33. Chen F and Chen S. Injury severities of truck drivers in single- and multi-vehicle accidents on rural highway. Accident Analysis and Prevention. 2011, 43, 1677–1688. pmid:21658494
- 34. Chen F, Chen S and Ma X. Analysis of hourly crash likelihood using unbalanced panel data mixed logit model and real-time driving environmental big data. Journal of Safety Research. 2018, 65, 153–159. pmid:29776524
- 35. Huang H, Song B, Xu P, Zeng Q, Lee J and Abdel-Aty M. Macro and micro models for zonal crash prediction with application in hot zones identification. Journal of Transport Geography. 2016, 54, 248–256.
- 36. Wen H, Zhang X, Zeng Q and Sze NN. Bayesian spatial-temporal model for the main and interaction effects of roadway and weather characteristics on freeway crash incidence. Accident Analysis and Prevention. 2019, 132, 105249. pmid:31415995
- 37. Zeng Q, Wang X, Wen H and Yuan Q. An empirical investigation of the factors contributing to local-vehicle and non-local-vehicle crashes on freeway. Journal of Transportation Safety & Security. 2020, 1–15.
- 38. Cordeiro GM and Brito RDS. The beta power distribution. Brazilian Journal of Probability and Statistics. 2012, 26, 1, 88–112.
- 39. Dumonceaux RH and Antle CE. Discriminating between the log-normal and Weibull distribution. Technometrics. 1973, 15, 4, 923–926.
- 40.
Carrasco J, Ferari SLP and Cordeiro GM. A new generalized Kumaraswamy distribution. arXiv:1004.0911v1[stat.ME].