% IMPORTANT: The following is UTF-8 encoded. This means that in the presence % of non-ASCII characters, it will not work with BibTeX 0.99 or older. % Instead, you should use an up-to-date BibTeX implementation like “bibtex8” or % “biber”. @ARTICLE{Grassberger:908873, author = {Grassberger, Peter}, title = {{O}n {G}eneralized {S}chürmann {E}ntropy {E}stimators}, journal = {Entropy}, volume = {24}, number = {5}, issn = {1099-4300}, address = {Basel}, publisher = {MDPI}, reportid = {FZJ-2022-02887}, pages = {680 -}, year = {2022}, abstract = {We present a new class of estimators of Shannon entropy for severely undersampleddiscrete distributions. It is based on a generalization of an estimator proposed by T. Schürmann,which itself is a generalization of an estimator proposed by myself.For a special set of parameters,they are completely free of bias and have a finite variance, something which is widely believedto be impossible. We present also detailed numerical tests, where we compare them with otherrecent estimators and with exact results, and point out a clash with Bayesian estimators for mutualinformation.}, cin = {JSC}, ddc = {510}, cid = {I:(DE-Juel1)JSC-20090406}, pnm = {5111 - Domain-Specific Simulation $\&$ Data Life Cycle Labs (SDLs) and Research Groups (POF4-511)}, pid = {G:(DE-HGF)POF4-5111}, typ = {PUB:(DE-HGF)16}, pubmed = {35626564}, UT = {WOS:000801648200001}, doi = {10.3390/e24050680}, url = {https://juser.fz-juelich.de/record/908873}, }