Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Estimation of different types of entropies for the Kumaraswamy distribution

  • Abdulhakim A. Al-Babtain ,

    Contributed equally to this work with: Abdulhakim A. Al-Babtain, Ibrahim Elbatal, Christophe Chesneau, Mohammed Elgarhy

    Roles Methodology

    Affiliation Department of Statistics and Operations Research, King Saud University, Riyadh, Saudi Arabia

  • Ibrahim Elbatal ,

    Contributed equally to this work with: Abdulhakim A. Al-Babtain, Ibrahim Elbatal, Christophe Chesneau, Mohammed Elgarhy

    Roles Methodology

    Affiliation Department of Mathematics and Statistics, College of Science, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia

  • Christophe Chesneau ,

    Contributed equally to this work with: Abdulhakim A. Al-Babtain, Ibrahim Elbatal, Christophe Chesneau, Mohammed Elgarhy

    Roles Methodology

    christophe.chesneau@unicaen.fr

    Affiliation Department of Mathematics, Université de Caen, LMNO, Campus II, Science 3, Caen, France

  • Mohammed Elgarhy

    Contributed equally to this work with: Abdulhakim A. Al-Babtain, Ibrahim Elbatal, Christophe Chesneau, Mohammed Elgarhy

    Roles Methodology

    Affiliation The Higher Institute of Commercial Sciences, Al mahalla Al kubra, Algarbia, Egypt

Abstract

The estimation of the entropy of a random system or process is of interest in many scientific applications. The aim of this article is the analysis of the entropy of the famous Kumaraswamy distribution, an aspect which has not been the subject of particular attention previously as surprising as it may seem. With this in mind, six different entropy measures are considered and expressed analytically via the beta function. A numerical study is performed to discuss the behavior of these measures. Subsequently, we investigate their estimation through a semi-parametric approach combining the obtained expressions and the maximum likelihood estimation approach. Maximum likelihood estimates for the considered entropy measures are thus derived. The convergence properties of these estimates are proved through a simulated data, showing their numerical efficiency. Concrete applications to two real data sets are provided.

1 Introduction

Information theory provides natural mathematical tools for measuring the uncertainty of random variables and the information shared by them. In this regard, entropy and mutual information are two fundamental concepts. More precisely, the probability distribution of a random variable is associated with some sort of uncertainty, and entropy is used to quantify it. The concept of entropy was formerly proposed by [1]. Since that publication, many areas of study such as statistics, neurobiology, cryptography, bioinformatics, quantum computer science and linguistics, have developed various entropy-based measures. Modern and exhaustive reviews on the ‘entropy universe’ can be found in [26].

In applied probability and statistics, many authors have conducted their studies for diverse and important distributions based on entropy. The essential references in this regard are briefly presented below. Reference [7] used the concept of entropy to communicate on the probability distribution of electric charge between atoms observed in a certain condition. Reference [8] derived the entropy for the Feller-Pareto family and presented the entropy ordering property for some related sample minimum and maximum. Reference [9] estimated the entropy of the Weibull distribution by considering different loss functions based on a generalized progressively hybrid censoring scheme. Reference [10] discussed the entropy for the generalized half-logistic distribution based on the type II censored samples. References [11] and [12] proposed estimates for the entropy of absolutely continuous random variables. Reference [13] presented an indirect method using a decomposition to simplify the entropy’s calculation under the progressive type II censoring. Reference [14] derived a nonparametric kernel estimator for the general Shannon entropy. Reference [15] estimated the entropy for several exponential distributions and extended the results to other circumstances. Reference [16] estimated the Shannon entropy of the Rayleigh model under doubly generalized type-II hybrid censoring, and evaluated its performance by two criteria. Reference [17] derived a nonparametric wavelet estimator for the general Shannon entropy. Reference [18] provided an exact expression for entropy information contained in both types of progressively hybrid censored data and applied it in the setting of the exponential distribution. Reference [19] investigated entropy measures for weighted and truncated weighted exponential distributions. Reference [20] presented the estimation of entropy for inverse Weibull distribution under multiple censored data. Reference [21] introduced estimation of entropy for inverse Lomax distribution under the multiple censored scheme. Reference [22] examined Bayesian and non-Bayesian methods to estimate the dynamic cumulative residual Rényi entropy for the Lomax distribution.

Surprisingly, to our knowledge, the entropy of the famous Kumaraswamy distribution has not been studied in depth. In this article, we fill this gap both probabilistically and statistically. The specificities and interests of the Kumaraswamy distribution are described below. First, it was introduced by [23], and was motivated as an alternative to the beta distribution which are (i) mathematically simpler, without special function in particular, and (ii) more suited to the modeling of various hydrological phenomena observed at low frequency (daily rainfall, daily flow of rivers, etc.). Mathematically, the probability density function (pdf) of the Kumaraswamy distribution is specified by (1) with f(x;a, b) = 0 otherwise, where a, b > 0. This pdf is unimodal if a, b > 1, uniantimodal if a, b < 1, increasing if a > 1, b ≤ 1, decreasing if a ≤ 1, b > 1 or constant if a = b = 1, in the same way as the beta distribution. The corresponding cumulative distribution and quantile functions are quite simple; they are defined without special function contrary to those of the beta distribution. Special cases of the Kumaraswamy distribution correspond to the distribution of minimum or maximum of uniform samples. We may refer the reader to [24] for all the known features of this distribution. Also, the kumaraswamy distribution has generated many flexible distributions with various domains and number of parameters through the generalized Kumaraswamy class elaborated by [25].

In a sense, this study complements the work of [24] by investigating the overall concept of entropy of the Kumaraswamy distribution, which has never been studied before. More precisely, we consider six well-referenced entropy measures. We derive their analytical expressions by using the well-known beta function. We compare them numerically by considering different parameter values. Then, we propose an efficient strategy based on the maximum likelihood approach to estimate these entropy measures. A simulation study is done to see how effective our strategy is. Graphical and numerical comparisons are performed. We end the study by two illustrative examples on real data sets, showing how the methodology can be applied in a concrete statistical setting.

The following sections make up the document. Section 2 presents a result on a special integral, and shows how it is related to important entropy measures of the Kumaraswamy distribution. Numerical values of these entropy measures with different values of the parameters are also given. Section 3 studies the estimation of these entropy measures. Then, using generated values from the Kumaraswamy distribution, graphical and numerical comparisons are discussed. The entropy of the random characteristics behind two real data sets is investigated. Finally, conclusions are presented in section 4.

2 Entropy of the Kumaraswamy distribution

2.1 An integral result

The following result shows that a certain integral involving the pdf of the Kumaraswamy distribution can be expressed in terms of the classical beta function. The connection between this integral and the considered entropy measures will be developed later.

Proposition 1 Let δ > 0, f(x;a, b) be specified by Eq (1) and

Then, Iδ(a, b) exists if and only if min(a, b)>max(1 − 1/δ, 0), and it is expressed as where B(u, v) denotes the classical beta function, that is for u, v > 0.

Proof. Owing to Eq (1), we have

When x tends to 0, we have xδ(a−1)(1 − xa)δ(b−1)xδ(a−1), which is integrable in the neighborhood of 0 if and only if δ(1 − a)<1 by the Riemann integral criteria. Similarly, when x tends to 1, we have which is integrable in the neighborhood of 1 if and only if δ(1 − b)<1 by the Riemann integral criteria. In summary, Iδ(a, b) exists if and only if δmax(1 − a, 1 − b)<1, which is equivalent to min(a, b)>1 − 1/δ. Now, under this assumption, by applying the change of variables y = xa, that is with , we obtain

This ends the proof of Proposition 1.

In fact, the beta function is implemented in most of the mathematical software (see the function beta of the package stat of R, the Beta function of Mathematica, etc.). Therefore, thanks to Proposition 1, the computation of Iδ(a, b) can be done quite efficiently with little effort. Also, the existing results on the beta functions allow a mathematical control of this integral. Some related results are presented below.

  • Through the use of the standard Euler gamma function given as , one can write
  • Also, assuming that δ(1 − 1/a) + 1/a and δ(b − 1) + 1 are positive integers, the following formula holds:
  • By virtue of the main result in [26], if δ(a − 1) ≥ a − 1 and b ≥ 1, then we have with the best possible constants α = 0 and β = 0.08731…. Therefore, for not too large value of δ, the following numerical approximation seems acceptable:

In our study, the interest of Proposition 1 is that Iδ(a, b) is the main ingredient in the definitions of various entropy measures of the Kumaraswamy distribution, as developed in the next part.

2.2 Various entropy measures

The entropy of the Kumaraswamy distribution can be measured in different manners. The most useful entropy measures of the literature are recalled in Table 1 for a general distribution with pdf denoted by f(x;φ), φ representing a possible vector of parameters. Also, we suppose that δ > 0 and δ ≠ 1 as basic assumptions in this general case.

thumbnail
Table 1. Important entropy measures of a distribution with pdf f(x;φ) at δ.

https://doi.org/10.1371/journal.pone.0249027.t001

For the two entropy measures proposed by [27], it is supposed that is finite and well identified.

From Table 1, we see that the integral is central to determine the considered entropy measures. Now, we present the corresponding entropy measures of the Kumaraswamy distribution. Based on Proposition 1, it is supposed that a, b and δ satisfy min(a, b)>max(1 − 1/δ, 0).

Rényi entropy.

Based on Table 1, Eq (1) and Proposition 1, the Rényi entropy of the Kumaraswamy distribution can be expressed as

Havrda and Charvát entropy.

From Table 1, Eq (1) and Proposition 1, the Havrda and Charvát entropy of the Kumaraswamy distribution can be expressed as

Arimoto entropy.

Again, from Table 1, Eq (1) and Proposition 1, the Arimoto entropy of the Kumaraswamy distribution is specified by

Tsallis entropy.

Based on Table 1, Eq (1) and Proposition 1, the Tsallis entropy of the Kumaraswamy distribution can be expressed as

Awad and Alawneh 1 entropy.

From Table 1, Eq (1) and Proposition 1, the Awad and Alawneh 1 entropy of the Kumaraswamy distribution is given as (2)

Before going further, we need to determine sup0<x<1 f(x;a, b). The following lemma provides the necessary in this regard.

Lemma 2 Let f(x;a, b) be given as Eq (1). Then, sup0<x<1 f(x;a, b) is finite if and only if a ≥ 1 and b ≥ 1 with ab ≠ 1, and in this case, we have

Proof. We have

Therefore, f′(x*;a, b) = 0 implies that

Since f′(x;a, b)>0 for x < x* and f′(x;a, b)<0 for x > x*, x* is a maximum point for f(x;a, b). Hence,

Note that, for a = 1, with the convention 00 = 1, we have f(x*;a, b) = b(b − 1)b−1(b − 1)1−b = b and for b = 1, we have . This ends the proof of Lemma 2.

Based on Lemma 2, if a > 1 and b > 1, Eq (2) becomes

Awad and Alawneh 2 entropy.

From Table 1, Eq (1), Proposition 1 and Lemma 2, the Awad and Alawneh 2 entropy of the Kumaraswamy distribution is given as

Theoretically, it is complicated to study the behavior of these entropy measures. For this reason, a numerical study is proposed in the next section.

2.3 Numerical values

We now investigate the numerical values for the six entropy measures presented in Subsection 2.2 under the following configuration of the parameters: Configuration 1: a = 2, bΥ with Υ = {1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0, 5.5, 6.0} and δ = 0.5, Configuration 2: a = 2, bΥ and δ = 1.5, Configuration 3: a = 2, bΥ and δ = 2.5, Configuration 4: aΥ, b = 2 and δ = 0.5, Configuration 5: aΥ, b = 2 and δ = 1.5, and Configuration 6: aΥ, b = 2 and δ = 2.5. The findings of all the six entropy measures are presented for these configurations in Tables 27, respectively.

thumbnail
Table 2. Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 0.5.

https://doi.org/10.1371/journal.pone.0249027.t002

thumbnail
Table 3. Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 1.5.

https://doi.org/10.1371/journal.pone.0249027.t003

thumbnail
Table 4. Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 2.5.

https://doi.org/10.1371/journal.pone.0249027.t004

thumbnail
Table 5. Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 0.5.

https://doi.org/10.1371/journal.pone.0249027.t005

thumbnail
Table 6. Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 1.5.

https://doi.org/10.1371/journal.pone.0249027.t006

thumbnail
Table 7. Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 2.5.

https://doi.org/10.1371/journal.pone.0249027.t007

In view of Tables 27, the following comments can be formulated.

First, we recall that Tables 24 indicate the values of the entropy measures of the Kumaraswamy distribution for a fixed value of a and different values for b and δ. In this context,

  • the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when b is increasing while the Awad and Alawneh 2 entropy is increasing when b is increasing.
  • the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing.

Tables 57 show the values of the entropy of the Kumaraswamy distribution for a fixed value of b and different values for a and δ. In this setting,

  • the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when a is increasing while the Awad and Alawneh 2 entropy is increasing when a is increasing.
  • the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing.

3 Maximum likelihood estimation

The inference on the six considered entropy measures of the Kumaraswamy distribution is now investigated via the maximum likelihood technique. This technique is well-known and has proved itself in various modern studies such as those in [3234].

3.1 Estimation of the entropy measures

The estimation of the parameters of the Kumaraswamy model through the maximum likelihood technique is well-known and the details can be found in [24]. The minimal theory is recalled below. Based on n values x1, …, xn supposed to be observed from a random variable X with the Kumaraswamy distribution with parameters a and b, the maximum likelihood estimates (MLEs) of a and b, say and , are defined by where (a, b) denotes the log-likelihood function specified by

These MLEs are also the solutions of the two following equations according to a and b:

That is, and satisfy the following simple relation:

Then, the properties of these MLEs follow from the usual maximum likelihood theory. In particular, thanks to the functional invariance of the MLEs, one can deduce easily the MLEs of the entropy measures. More concretely, based on the six entropy measures described in Subsection 2.2, is the MLE of Rδ(a, b), is the MLE of HCδ(a, b), is the MLE of Aδ(a, b), is the MLE of Tδ(a, b), is the MLE of AA1δ(a, b), and is the MLE of AA2δ(a, b).

3.2 Simulation

We now investigate the numerical behavior of the MLEs of the entropy measures via the use of simulated values. That is, we consider N = 5000 samples of values from a random variable X with the Kumaraswamy distribution of parameters a and b with different samples sizes; n = 100, 200, 300 and 1000 are considered. The following configurations on the parameters are considered: Configuration1: a = 3, b = 3 and δΞ with Ξ = {0.5, 1.5, 2.5}, and Configuration 2: a = 3, b = 5 and δΞ.

In each configuration, for each sample, the MLEs and are determined. Then, based on the N samples of fixed size, we determine the average of the N MLEs and use it to define the entropy estimates. The corresponding mean squared error (MSE) and mean deviation (MD) defined by the following generic formulas: MSE = sum(exact value—estimate)2 / N and MD = sum abs(exact value—estimate) / N, respectively, are also calculated. These assessment criteria are often used quite effectively to make a full comparison of models. In this regard, we can refer the reader to the useful works of [3537].

The results on the Rényi entropy under Configurations 1 and 2 are given in Tables 8 and 9, respectively, results on the Havrda and Charvat entropy under Configurations 1 and 2 are indicated in Tables 10 and 11, respectively, results on the Arimoto entropy under Configurations 1 and 2 are presented in Tables 12 and 13, respectively, results on the Tsallis entropy under Configurations 1 and 2 are given in Tables 14 and 15, respectively, results on the Awad and Alawneh 1 entropy under Configurations 1 and 2 are given in Tables 16 and 17, respectively, and results on the Awad and Alawneh 2 entropy under Configurations 1 and 2 are indicated in Tables 18 and 19.

thumbnail
Table 8. Numerical values of the simulation related to the Rényi entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t008

thumbnail
Table 9. Numerical values of the simulation related to the Rényi entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t009

thumbnail
Table 10. Numerical values of the simulation related to the Havrda and Charvat entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t010

thumbnail
Table 11. Numerical values of the simulation related to the Havrda and Charvat entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t011

thumbnail
Table 12. Numerical values of the simulation related to the Arimoto entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t012

thumbnail
Table 13. Numerical values of the simulation related to the Arimoto entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t013

thumbnail
Table 14. Numerical values of the simulation related to the Tsallis entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t014

thumbnail
Table 15. Numerical values of the simulation related to the Tsallis entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t015

thumbnail
Table 16. Numerical values of the simulation related to the Awad and Alawneh 1 entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t016

thumbnail
Table 17. Numerical values of the simulation related to the Awad and Alawneh 1 entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t017

thumbnail
Table 18. Numerical values of the simulation related to the Awad and Alawneh 2 entropy for Configuration 1 (a = 3, b = 3).

https://doi.org/10.1371/journal.pone.0249027.t018

thumbnail
Table 19. Numerical values of the simulation related to the Awad and Alawneh 2 entropy for Configuration 2 (a = 3, b = 5).

https://doi.org/10.1371/journal.pone.0249027.t019

Based on Tables 819, in all the situations, we see that the MLEs of the entropy measures are close to the target values and, as anticipated, the MSEs and MDs decrease and approach 0 as n increases. This proves the accuracy of the proposed estimation methods in the context of the Kumaraswamy distribution. Also, one can notice that the MSEs and MDs increase as δ increases.

For a visual approach, the behavior of the MSEs and MDs are illustrated in Figs 112, for the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 1 and Awad and Alawneh 2 entropy measures following the settings of Tables 819, respectively.

thumbnail
Fig 1. Plots of the (a) MSEs and (b) MDs for the Rényi entropy in the setting of Table 8.

https://doi.org/10.1371/journal.pone.0249027.g001

thumbnail
Fig 2. Plots of the (a) MSEs and (b) MDs for the Rényi entropy in the setting of Table 9.

https://doi.org/10.1371/journal.pone.0249027.g002

thumbnail
Fig 3. Plots of the (a) MSEs and (b) MDs for the Havrda and Chardat entropy in the setting of Table 10.

https://doi.org/10.1371/journal.pone.0249027.g003

thumbnail
Fig 4. Plots of the (a) MSEs and (b) MDs for the Havrda and Chardat entropy in the setting of Table 11.

https://doi.org/10.1371/journal.pone.0249027.g004

thumbnail
Fig 5. Plots of the (a) MSEs and (b) MDs for the Arimoto entropy in the setting of Table 12.

https://doi.org/10.1371/journal.pone.0249027.g005

thumbnail
Fig 6. Plots of the (a) MSEs and (b) MDs for the Arimoto entropy in the setting of Table 13.

https://doi.org/10.1371/journal.pone.0249027.g006

thumbnail
Fig 7. Plots of the (a) MSEs and (b) MDs for the Tsallis entropy in the setting of Table 14.

https://doi.org/10.1371/journal.pone.0249027.g007

thumbnail
Fig 8. Plots of the (a) MSEs and (b) MDs for the Tsallis entropy in the setting of Table 15.

https://doi.org/10.1371/journal.pone.0249027.g008

thumbnail
Fig 9. Plots of the (a) MSEs and (b) MDs for the Awad and Alawneh 1 entropy in the setting of Table 16.

https://doi.org/10.1371/journal.pone.0249027.g009

thumbnail
Fig 10. Plots of the (a) MSEs and (b) MDs for the Awad and Alawneh 1 entropy in the setting of Table 17.

https://doi.org/10.1371/journal.pone.0249027.g010

thumbnail
Fig 11. Plots of the (a) MSEs and (b) MDs for the Awad and Alawneh 2 entropy in the setting of Table 18.

https://doi.org/10.1371/journal.pone.0249027.g011

thumbnail
Fig 12. Plots of the (a) MSEs and (b) MDs for the Awad and Alawneh 2 entropy in the setting of Table 19.

https://doi.org/10.1371/journal.pone.0249027.g012

Figs 112 support the claims formulated about the results of Tables 819.

3.3 Illustrative examples

In this Section, two real life data sets are used to illustrate the proposed methodology. The considered data sets are described below.

The first data set.

The data set consists of 48 rock samples from an oil reservoir. It corresponds to twelve oil tank cores that were sampled by four cross sections. Each core was measured for permeability and each cross section has the following variables: total pore area, total pore perimeter, and shape. We analyze the perimeter of the shape by a squared variable (area). It has been analyzed by [38], among others. Explicitely, the data set is: {0.0903296, 0.2036540, 0.2043140, 0.2808870, 0.1976530, 0.3286410, 0.1486220, 0.1623940, 0.2627270, 0.1794550, 0.3266350, 0.2300810, 0.1833120, 0.1509440, 0.2000710, 0.1918020, 0.1541920, 0.4641250, 0.1170630, 0.1481410, 0.1448100, 0.1330830, 0.2760160, 0.4204770, 0.1224170, 0.2285950, 0.1138520, 0.2252140, 0.1769690, 0.2007440, 0.1670450, 0.2316230, 0.2910290, 0.3412730, 0.4387120, 0.2626510, 0.1896510, 0.1725670, 0.2400770, 0.3116460, 0.1635860, 0.1824530, 0.1641270, 0.1534810, 0.1618650, 0.2760160, 0.2538320, 0.2004470}.

The second data set.

This data set contains 20 observations of flood data. It was analyzed by [39]. The data set is listed as follows: {0.265, 0.392, 0.297, 0.3235, 0.402, 0.269, 0.315, 0.654, 0.338, 0.379, 0.418, 0.423, 0.379, 0.412, 0.416, 0.449, 0.484, 0.494, 0.613, 0.74}.

In order to check the adequateness of the Kumaraswamy distribution to these data, we apply the Kolmogorov-Smirnov test. We find p-value 0.2092 and p-value = 0.3359 for the first and second data sets, respectively. Since both satisfy p-values >0.05, the two considered data set are in adequateness with the Kumaraswamy distribution.

Now, Tables 20 and 21 present the estimations of the six entropy measures considered in Subsection 2.2, following the methodology described in Subsection 3.1, for the first and second data sets, respectively.

thumbnail
Table 20. Estimates of the considered entropy measures with different values of δ for the first data set.

https://doi.org/10.1371/journal.pone.0249027.t020

thumbnail
Table 21. Estimates of the considered entropy measures with different values of δ for the second data set.

https://doi.org/10.1371/journal.pone.0249027.t021

We can notice that, under our framework, the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 2 entropy measures are decreasing when δ is increasing while the Awad and Alawneh 1 entropy is increasing when δ is increasing.

Tour knowledge, it is the first time that the entropy of the uncertainty behind these data sets are evaluated. They can be taken into account for further statistical analysis in the future.

4 Conclusion

For the first time, this article proposed a special focus on the entropy of the Kumaraswamy distribution. Both theoretical and practical aspects were covered, though complementary works. In particular, six different entropy measures were investigated. After determining the closed-form expressions of these measures, an estimation strategy was developed to evaluate them in a practical setting. A simulation study ensured the convergence of the obtained estimates. Two real-life data sets are used to show how the related entropy can be concretely estimated. The finding of this study aims to be applied by the statistician to assess the entropy of diverse data with values on the unit interval, such as modern rate, percentage and proportion type data.

The limitation of current research remains on the classicity of the statistical framework considered. Directions for future research include the estimation of the entropy of the Kumaraswamy distribution in more sophisticated statistical schemes with physical motivations, such as the progressive type II censoring scheme, generalized progressively hybrid censoring scheme, etc., or taking into account generalized versions of the Kumaraswamy distribution, such as the one proposed by [40].

Acknowledgments

The authors thank the two reviewers for their detailed and constructive comments.

References

  1. 1. Shannon CE. A mathematical theory of communication. The Bell System Technical Journal. 1948, 27, 1, 379–423.
  2. 2. Golan A. Information and entropy econometrics-a review and synthesis. Foundations and Trends in Econometrics. 2008, 2, 1–45.
  3. 3. Cover TM and Thomas JA. Elements of Information Theory. Hoboken, NJ: John Wiley & Sons, 2012.
  4. 4. Amigo JM, Balogh SG and Hernandez S. A brief review of generalized entropies. Entropy, 2018, 20, 813. pmid:33266537
  5. 5. Namdari A and Li Z. A review of entropy measures for uncertainty quantification of stochastic processes. Advances in Mechanical Engineering. 2019, 11, 6, 1–14.
  6. 6. Ribeiro M, Henriques T, Castro L, Souto A, Antunes L, Costa-Santos C, et al. The entropy universe. Entropy. 2021, 23, 222. pmid:33670121
  7. 7. Siamak N and Ehsan S. Shannon entropy as a new measure of aromaticity. Shannon aromaticity, Physical Chemistry Chemical Physics. 2010, 12, 4742–4749.
  8. 8. Tahmasebi S and Behboodian J. Shannon entropy for the Feller-Pareto (FP) family and order statistics of FP subfamilies. Applied Mathematical Sciences. 2010, 4, 10, 495–504.
  9. 9. Cho Y, Sun H and Lee K. Estimating the entropy of a Weibull distribution under generalized progressive hybrid censoring, Entropy. 2015, 17, 102–122.
  10. 10. Seo JI and Kang SB. Entropy estimation of generalized half-logistic distribution (GHLD) based on type-II censored samples. Entropy. 2014, 16, 1424–1433.
  11. 11. Ebrahimi N, Pflughoef K and Soofi E. Two measures of sample entropy. Statistics and Probability Letters. 1994, 20, 225–234.
  12. 12. Alizadeh Noughabi H. A new estimator of entropy and its application in testing normality. Journal of Statistical Computation and Simulation. 2010, 80, 10, 1151–1162.
  13. 13. Abo-Eleneen ZA. The entropy of progressively censored samples, Entropy. 2011, 13, 437–449.
  14. 14. Bouzebda S and Elhattab I. Uniform-in-bandwidth consistency for kernel-type estimators of Shannon’s entropy. Electronic Journal of Statistics. 2011, 5, 440–459.
  15. 15. Kayal S and Kumar S. Estimation of the Shannon’s entropy of several shifted exponential populations. Statistics and Probability Letters. 2013. 83, 1127–1135.
  16. 16. Cho Y, Sun H and Lee K. An estimation of the entropy for a Rayleigh distribution based on doubly- generalized type-II hybrid censored samples. Entropy. 2014, 16, 3655–3669.
  17. 17. Chesneau C, Navarro F and Serea O. A note on the adaptive estimation of the differential entropy by wavelet methods. Commentationes Mathematicae Universitatis Carolinae. 2017, 58,1, 87–100.
  18. 18. Bader A. On the entropy of progressive hybrid censoring schemes. Applied Mathematics & Information Sciences. 2017, 11, 6, 1811–1814.
  19. 19. Basit A, Riaz A, Iqbal Z and Ahmad M. On comparison of entropy measures for weighted and truncated weighted exponential distributions. Advances and Applications in Statistics. 2017, 50, 6, 477–495.
  20. 20. Hassan AS and Zaky AN. Estimation of entropy for inverse Weibull distribution under multiple censored data. Journal of Taibah University for Science. 2019, V, 13, 1, 331–337.
  21. 21. Bantan RAR, Elgarhy M, Chesneau C and Jamal F. Estimation of entropy for inverse Lomax distribution under multiple censored data. Entropy, 2020, 22, 6, 601, 1–15. pmid:33286373
  22. 22. Al-Babtain AA, Hassan AS, Zaki AN, Elbatal I and Elgarhy M. Dynamic cumulative residual Rényi entropy for Lomax distribution: Bayesian and non-Bayesian methods. AIMS Mathematics, 2021, 6, 4, 3889–3914.
  23. 23. Kumaraswamy P. Generalized probability density-function for double-bounded random-processes. Journal of Hydrology. 1980, 46, 79–88.
  24. 24. Jones MC. Kumaraswamy’s distribution: A beta-type distribution with some tractability advantages, Statistical Methodology. 2009, 6, 1, 70–91.
  25. 25. Cordeiro GM and de Castro M. A new family of generalized distributions. Journal of Statistics Computation and Simulation. 2011, 81, 883–893.
  26. 26. Alzer H. Sharp inequalities for the beta function. Indagationes Mathematicae, 2001, 12, 15–21.
  27. 27. Awad AM and Alawneh AJ. Application of entropy to a life-time model. IMA Journal of Mathematical Control and Information. 1987, 4, 2, 143–148.
  28. 28. Rényi A. On measures of entropy and information. Proc. 4th Berkeley Symposium on Mathematical Statistics and Probability. 1960, 1, 47-561.
  29. 29. Havrda J and Charvát F. Quantification method of classification processes, concept of structural a-entropy. Kybernetika. 1967, 3, 1, 30–35.
  30. 30. Arimoto S. Information-theoretical considerations on estimation problems. Information and Control. 1971, 19, 3, 181–194.
  31. 31. Tsallis C. Possible generalization of Boltzmann-Gibbs statistics. Journal of Statistical Physics, 1988, 52, (1-2), 479–487.
  32. 32. Dong B, Ma X, Chen F and Chen S. Investigating the differences of single- and multi-vehicle accident probability using mixed logit model. Journal of Advanced Transportation. 2018, Article ID 2702360, 9 pages.
  33. 33. Chen F and Chen S. Injury severities of truck drivers in single- and multi-vehicle accidents on rural highway. Accident Analysis and Prevention. 2011, 43, 1677–1688. pmid:21658494
  34. 34. Chen F, Chen S and Ma X. Analysis of hourly crash likelihood using unbalanced panel data mixed logit model and real-time driving environmental big data. Journal of Safety Research. 2018, 65, 153–159. pmid:29776524
  35. 35. Huang H, Song B, Xu P, Zeng Q, Lee J and Abdel-Aty M. Macro and micro models for zonal crash prediction with application in hot zones identification. Journal of Transport Geography. 2016, 54, 248–256.
  36. 36. Wen H, Zhang X, Zeng Q and Sze NN. Bayesian spatial-temporal model for the main and interaction effects of roadway and weather characteristics on freeway crash incidence. Accident Analysis and Prevention. 2019, 132, 105249. pmid:31415995
  37. 37. Zeng Q, Wang X, Wen H and Yuan Q. An empirical investigation of the factors contributing to local-vehicle and non-local-vehicle crashes on freeway. Journal of Transportation Safety & Security. 2020, 1–15.
  38. 38. Cordeiro GM and Brito RDS. The beta power distribution. Brazilian Journal of Probability and Statistics. 2012, 26, 1, 88–112.
  39. 39. Dumonceaux RH and Antle CE. Discriminating between the log-normal and Weibull distribution. Technometrics. 1973, 15, 4, 923–926.
  40. 40. Carrasco J, Ferari SLP and Cordeiro GM. A new generalized Kumaraswamy distribution. arXiv:1004.0911v1[stat.ME].