Have a personal or library account? Click to login
Entropy as a measure of dependency for categorized data Cover

Entropy as a measure of dependency for categorized data

Open Access
|Dec 2018

Abstract

Data arranged in a two-way contingency table can be obtained as a result of many experiments in the life sciences. In some cases the categorized trait is in fact conditioned by an unobservable continuous variable, called liability. It may be interesting to know the relationship between the Pearson correlation coefficient of these two continuous variables and the entropy function measuring the corresponding relation for categorized data. After many simulation trials, a linear regression was estimated between the Pearson correlation coefficient and the normalized mutual information (both on a logarithmic scale). It was observed that the regression coefficients obtained do not depend either on the number of observations classified on a categorical scale or on the continuous random distribution used for the latent variable, but they are influenced by the number of columns in the contingency table. In this paper a known measure of dependency for such data, based on the entropy concept, is applied.

DOI: https://doi.org/10.2478/bile-2018-0014 | Journal eISSN: 2199-577X | Journal ISSN: 1896-3811
Language: English
Page range: 233 - 243
Published on: Dec 14, 2018
Published by: Polish Biometric Society
In partnership with: Paradigm Publishing Services
Publication frequency: 2 issues per year

© 2018 Ewa Skotarczak, Anita Dobek, Krzysztof Moliński, published by Polish Biometric Society
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 3.0 License.