Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
Many information measures are suggested in the literature. Among these measures are the Shannon H,(O) and the Awad A,(O) entropies. In this work we suggest a new entropy measure, B~(0), which is based on the maximum likelihood function. These three entropies were calculated from the gamma distribution and its normal approximation, the binomial and its Poisson approximation, and the Poisson and its normal approximation. The relative losses in these three entropies are used as a criterion for the appropriateness of the approximation. Copyright © 1996 Elsevier Science Ltd.
Physica A: Statistical Mechanics and its Applications, 2017
Highlights 1. Two estimation methods of Fisher Information Measure (FIM) and Shannon entropy (SE) are analysed. 2. One is based on discretizing FIM and SE formulae; the other on the kernel-based estimation of the probability density function. 3. FIM (SE) estimated by using the discrete-based approach is approximately constant with , but decreases (increases) with the bin number L. 4. FIM (SE) estimated by using the kernel-based approach is very close to the theoretic value for any .
2000
This memo,contains proofs that the Shannon entropy is the limiting case of both the Renyi entropy and the information. These results are also confirmed experimentally. We conclude with some general observations on the utility of entropy measures.
arXiv (Cornell University), 2024
In this research work, a total of 45 different estimators of the Shannon differential entropy were reviewed. The estimators were mainly based on three classes, namely: window size spacings (m), kernel density estimation (KDE) and k-nearest neighbour (kNN) estimation. A total of 16, 5 and 6 estimators were selected from each of the classes, respectively, for comparison. The performances of the 27 selected estimators, in terms of their bias values and root mean squared errors (RMSEs) as well as their asymptotic behaviours, were compared through extensive Monte Carlo simulations. The empirical comparisons were carried out at different sample sizes of n = 10, 50, and 100 and different variable dimensions of d = 1, 2, 3, and 5, for three groups of continuous distributions according to their symmetry and support. The results showed that the spacings based estimators generally performed better than the estimators from the other two classes at d = 1, but suffered from non existence at d ≥ 2. The kNN based estimators were generally inferior to the estimators from the other two classes considered but showed an advantage of existence for all d ≥ 1. Also, a new class of optimal window size was obtained and sets of estimators were recommended for different groups of distributions at different variable dimensions. Finally, the asymptotic biases, variances and distributions of the 'best estimators' were considered.
Estimating the entropy of finite strings has applications in areas such as event detection, similarity mea-surement or in the performance assessment of compression algorithms. This report compares a variety of computable information measures for finite strings that may be used in entropy estimation. These include Shannon's n-block entropy, the three variants of the Lempel-Ziv production complexity, and the lesser known T-entropy. We apply these measures to strings derived from the logistic map, for which Pesin's identity allows us to deduce corresponding Shannon entropies (Kolmogorov-Sinai entropies) without resorting to probabilistic methods.
International Statistical Review, 2020
SummaryStarting from the pioneering works of Shannon and Weiner in 1948, a plethora of works have been reported on entropy in different directions. Entropy‐related review work in the direction of statistical inference, to the best of our knowledge, has not been reported so far. Here, we have tried to collect all possible works in this direction during the last seven decades so that people interested in entropy, specially the new researchers, get benefited.
Asian Journal of Fuzzy and Applied Mathematics, 2019
In this research article, a new two parametric measure of Entropy () and its corresponding code word length () has been developed. The developed measures are the generalizations to some well known existing measures. Besides, some noiseless coding theorems for discrete noiseless channel have been developed, and the results thus obtained have been verified with the support of an numerical example. Also, at the end of this research article, a comparative study in terms of monotonic behavior among the proposed entropy (), Matahi's entropy () and Tsallis entropy () together with their respective average code word length measures have been made and graphically displayed.
International Journal of Modern Physics B, 2010
Journal of Administrative Sciences and Policy Studies, 2015
Statistical entropy is a measure of dispersion or spread of a random variable. Especially when the random variable is nominal, classical measures of dispersion like standard deviation can not be computed. In such cases, measures of variation, including entropy-based statistics;computed by using cell frequencies of a distributionmust be used. The asymptotic properties of entropy statistics have long been studied in literature. Relative entropy plays an important role in evaluating the degree of fit. In other words, relative entropy is a measure of goodness fit of an empirical distribution to a theoretical or hypothesized distribution. In this study for some frequently-used probability distributions,some relative entropy measures are derived by exploiting additivity property of Kullback-Leibler divergence and Jeffreys divergence. Their asymptotic properties under certain assumptions have been discussed. In the end, by some applications, the close relation between relative entropy statistics and other classical test statistics have been emphasized.
HAL (Le Centre pour la Communication Scientifique Directe), 2022
In this work, we propose to use varentropy, an information measure defined from a generalization of thermodynamic entropy, for the calculation of informational entropy in order to avoid negative entropy in case of continuous probability distribution.
Lecture Notes in Computer Science, 2009
It is commonly assumed that computers process information. But what is information? In a technical, important, but nevertheless rather narrow sense, Shannon's information theory gives a first answer to this question. This theory focuses on measuring the information content of a message. Essentially this measure is the reduction of the uncertainty obtained by receiving a message. The uncertainty of a situation of ignorance in turn is measured by entropy. This theory has had an immense impact on the technology of information storage, data compression, information transmission and coding and still is a very active domain of research.
1998
It is claimed that the way the cogentropy defines the probabilities of events, in a certain data set, assuming that only a possibly non-representative subset of the entire set is known, provides better estimates than using those conventionally applied in the Shannon entropy 113. In this paper, we apply both entropy measures to estimate the expected uncertainty in the data when the whole data set is known. This expected uncertainty value is nothing but the Shannon entropy when the entire set is known. The effect of the size of the subset on the approximations computed by both entropy measures is investigated. The study takes into consideration the conditional and unconditional cases.
arXiv: Other Statistics, 2019
Starting from the pioneering works of Shannon and Weiner in 1948, a plethora of works have been reported on entropy in different directions. Entropy-related review work in the direction of statistics, reliability and information science, to the best of our knowledge, has not been reported so far. Here we have tried to collect all possible works in this direction during the period 1948-2018 so that people interested in entropy, specially the new researchers, get benefited.
h i g h l i g h t s • Two estimation methods (discretization and kernel-based approach) are applied to FIM and SE. • FIM (SE) estimated by discrete approach is nearly constant with σ. • FIM (SE) estimated by discrete approach decreases (increases) with the bin number. • FIM (SE) estimated by kernel-based approach is close to the theory value for any σ. a b s t r a c t The performance of two estimators of Fisher Information Measure (FIM) and Shannon entropy (SE), one based on the discretization of the FIM and SE formulae (discrete-based approach) and the other based on the kernel-based estimation of the probability density function (pdf) (kernel-based approach) is investigated. The two approaches are employed to estimate the FIM and SE of Gaussian processes (with different values of σ and size N), whose theoretic FIM and SE depend on the standard deviation σ. The FIM (SE) estimated by using the discrete-based approach is approximately constant with σ , but decreases (increases) with the bin number L; in particular, the discrete-based approach furnishes a rather correct estimation of FIM (SE) for L ∝ σ. Furthermore, for small values of σ , the larger the size N of the series, the smaller the mean relative error; while for large values of σ , the larger the size N of the series, the larger the mean relative error. The FIM (SE) estimated by using the kernel-based approach is very close to the theoretic value for any σ , and the mean relative error decreases with the increase of the length of the series. Comparing the results obtained using the discrete-based and kernel-based approaches, the estimates of FIM and SE by using the kernel-based approach are much closer to the theoretic values for any σ and any N and have to be preferred to the discrete-based estimates.
Entropy
This study attempts to extend the prevailing definition of informational entropy, where entropy relates to the amount of reduction of uncertainty or, indirectly, to the amount of information gained through measurements of a random variable. The approach adopted herein describes informational entropy not as an absolute measure of information, but as a measure of the variation of information. This makes it possible to obtain a single value for informational entropy, instead of several values that vary with the selection of the discretizing interval, when discrete probabilities of hydrological events are estimated through relative class frequencies and discretizing intervals. Furthermore, the present work introduces confidence limits for the informational entropy function, which facilitates a comparison between the uncertainties of various hydrological processes with different scales of magnitude and different probability structures. The work addresses hydrologists and environmental engineers more than it does mathematicians and statisticians. In particular, it is intended to help solve information-related problems in hydrological monitoring design and assessment. This paper first considers the selection of probability distributions of best fit to hydrological data, using generated synthetic time series. Next, it attempts to assess hydrometric monitoring duration in a netwrok, this time using observed runoff data series. In both applications, it focuses, basically, on the theoretical background for the extended definition of informational entropy. The methodology is shown to give valid results in each case.
Entropy, 2004
Entropy has been the main tool in the analysis of the concept of information since information theory was conceived in the work of Shannon more than fifty years ago. There were some attempts to find more general measure of information, but their outcomes were more of formal, theoretical interest, and neither has provided better insight into the nature of information. The strengths of entropy seemed so obvious that no much effort has been made to find an alternative to entropy which gives different values, but which is consistent with entropy in the sense that the results obtained in information theory thus far can be reproduced with the new measure. In this article the need for such an alternative measure is demonstrated based on historical review of the problems with conceptualization of information. Then, an alternative measure is presented in the context of modified definition of information applicable outside of the conduit metaphor of Shannon's approach, and formulated without reference to uncertainty. It has several features superior to those of entropy. For instance, unlike entropy it can be easily and consistently extended to the continuous probability distributions, and unlike differential entropy this extension is always positive and invariant with respect to linear transformations of coordinates.
Journal of the Iranian Statistical Society Jirss, 2010
In this paper we propose an estimator of the entropy of a continuous random variable. The estimator is obtained by modifying the estimator proposed by . Consistency of estimator is proved, and comparisons are made with Vasicek's estimator (1976), van Es's estimator (1992), Ebrahimi et al. 's estimator (1994) and Correa's estimator (1995). The results indicate that the proposed estimator has smaller mean squared error than above estimators.
HAL (Le Centre pour la Communication Scientifique Directe), 2008
β. This paper presents an analytical investigation of this maximizable entropy for several distributions such as stretched exponential distribution, κ-exponential distribution and Cauchy distribution.
Cryptography, Information Theory, and Error‐Correction, 2021
In this note, we look at several definitions of entropy and some of their consequences in information theory. We also obtain the entropy and relative entropy for general error distribution which is used to model errors which are not normal. H(Y) = H(f) = − f (y) log f (y)dy,
Asian Research Journal of Mathematics
In this paper, a new estimator of the Shannon entropy of a random variable X having a probability density function \(\mathit{f}\)(\(\mathit{x}\)) is obtained based on window size spacings. Under the standard normal, standard exponential and uniform distributions, the estimator is shown to have relative low bias and low RMSE through extensive simulation study at sample sizes 10, 20, and 30. Based on the results, it is recommended as a good estimator of the entropy. Also, the new estimator is applied in goodness-of-fit test to normality. The statistic is affine invariant and consistent and the results show that it is a good statistic for assessing univariate normality of datasets.
Loading Preview
Sorry, preview is currently unavailable. You can download the paper by clicking the button above.