Abstract
Based on the notion of mutual information between the components of a random vector, we construct, for data reduction reasons, an optimal quantization of the support of its probability measure. More precisely, we propose a simultaneous discretization of the whole set of the components of the random vector which takes into account, as much as possible, the stochastic dependence between them. Examples are presented.
Similar content being viewed by others
References
ACZÉL, J., and DARÓCZY, Z. (1975), On Measures of Information and Their Characterizations, New York: Academic Press.
ADHIKARI, B.P., and JOSHI, D.D. (1956), “Distance, Discrimination et Résumé exhaustif”, Publications de l’Institut de Statistique de l’Université de Paris, 5, 57–74.
ALI, S.M., and SILVEY, S.D. (1966), “A General Class of Coefficients of Divergence of One Distribution from Another”, Journal of the Royal Statistical Society, B28, 131–142.
BEIRLANT, J., DUDEWICZ, E.J., GYORFI, L., and VAN DER MEULEN, E.C. (1997), “Nonparametric Entropy Estimation: An Overview”, International Journal of Mathematical and Statistical Sciences, 6 (1), 17–39.
BERTSEKAS, D.P. (1999), Nonlinear Programming (2nd ed.), Belmont MA: Athena Scientific.
BOSQ, D., and LECOUTRE, J.P. (1987), Théorie de l’Estimation fonctionnelle, London: Economica.
CSISZÁ R, I. (1967), “Information-Type Measures of Difference of Probability Distributions and Indirect Observations”, Studia Scientiarum Mathematicarum Hungarica, 2, 299–318.
CSISZÁR, I. (1972), “A Class of Measures of Informativity of Observation Channels”, Periodica Mathematica Hungarica, 2(1–4), 191–213.
CSISZÁR, I. (1977), “Information Measures : A Critical Survey”, Transactions of the seventh Prague Conference on Information Theory, Statistical Decision Functions and Random Processes, Vol. A, Prague: Publishing House of the Czechoslovak Academy of Sciences, pp. 73–86.
DARBELLAY, G.A. (1999), “An Estimator of the Mutual Information Based on a Criterion for Conditional Independence”, Computational Statistics & Data Analysis, 32, 1–17.
DARBELLAY, G.A., and VAJDA, I. (1999), “Estimation of the Information by an Adaptive Partitioning of the Observation Space”, IEEE Transactions on Information Theory 45(4), 1315–1321.
ELANDT-JONHSON, R.C. (1976), “Conditional Failure Time Distributions Under Competing Risk Theory with Dependent Failure Times and Proportional Hazard Rates”, Scandinavian Actuarial Journal, 1, 37–51.
FRECHET, M. (1951), “Sur les tableaux de corrélation dont les marges sont données”, Annals of the University of Lyon, 3(14), 53–77.
GAVURIN, M.K. (1968), “On the Value of Information”, Vestuik Leningrad University Series, 4, 1963, 27–34, and Translation in Selected Translations in Mathematical Statistics and Probability, 7, 193–202.
GOËL, P.K. (1981), “Information Measures and Bayesian Hierarcichal Models”, Technical Report, #81-4-1, Department of Statistics, Purdue University, West Lafayette, IN.
HAUSSLER, D., and OPPER, M. (1997) “Mutual information, Metric Entropy and Cumulative Relative Entropy Risk”, The Annals of Statistics, 25(6), 2451–2492.
LEHMAN, E.L (1991), Theory of Point Estimation, Pacific Grove CA: Wadsworth &Brooks.
LIESE, F., and VAJDA, I. (2006), “On Divergences and Information in Statistics and Information Theory”, IEEE Transactions on Information Theory, 52(10), 4394–4412.
LIESE, F., MORALES, D., and VAJDA, I. (2006), “Asymptotically Sufficient Partitions and Quantizations”, IEEE Transactions on Information Theory, 52(12), 5599–5606.
MCELIECE, R.J. (1977), “The Theory of Information Coding”, in Encyclopedia of Mathematics and Its Applications, eds. R. Doran, M. Ismail, T.-Y. Lam, and E. Lutwak, Reading MA: Addison Wesley.
MODDEMEIJER, R. (1999), “A Statistic to Estimate the Variance of the Histogram-Based Mutual Information Estimator, Based on Dependent Pairs of Observations”, Signal Processing, 75, 51–63.
MODDEMEIJER, R. (1989), “On Estimation of Entropy and Mutual Information of Continuous Distributions”, Signal Processing, 16, 233–248.
MORALES, D., PARDO, L., and VAJDA, I. (1995), “Asymptotic Divergence of Estimates of Discrete Distributions”, Journal of Statistical Planning and Inference, 48, 347–369.
ÖSTERREICHER, F., and VAJDA, I. (2003), “A New Class of Metric Divergences on Probability Spaces and Its Applicability in Statistics”, Annals of the Institute of Statistical Mathematics, 55(3), 639–653.
PINSKER, M.S. (1964), Information and Information Stability of Random Variables and Processes, San Francisco: Holden-Day.
PÖTZLBERGER, K. (2003), “Asymptotic Quantization of Probability Distribution”, Analysis in Theory and Applications, 19(4), 355–364.
RÉNYI, A. (1959), “On Measures of Dependence”, Acta Mathematica Hungararica, 10, 441–451.
RÉNYI., A. (1961), “On Measures of Entropy and Information”, Proceedings of the Fourth Berkeley Symposium of Mathematical Statistics and Probability, (1), Berkeley: University of California Press, pp. 547–561.
SERFLING, R.J. (1980), Approximation Theorems of Mathematical Statistics, New York: Wiley.
SKLAR, A. (1959), “Fonctions de répartition à n dimensions et leurs marges”, Publications de l’Institut de Statistique de l’Université de Paris, 8, 229–231
VAJDA, I. (2002), “On Convergence of Information Contained in Quantizied Observations”, IEEE Transactions on Information Theory, 48(8), 2163–2172.
ZAKAI, J., and ZIV, M. (1973), “On Functionals Satisfying a Data-Processing Theorem”, IEEE Transactions, IT-19, 275–282.
ZOUTENDIJK, G. (1960), Methods of Feasible Directions, Amsterdam: Elsevier, and Princeton NJ: D. VanNostrand.
Author information
Authors and Affiliations
Corresponding author
Additional information
The authors are grateful to anonymous referees for a careful reading and detailed revision of the original manuscript. Their valuable comments contributed to improve the presentation of our work.
Rights and permissions
About this article
Cite this article
Colin, B., Dubeau, F., Khreibani, H. et al. Optimal Quantization of the Support of a Continuous Multivariate Distribution based on Mutual Information. J Classif 30, 453–473 (2013). https://doi.org/10.1007/s00357-013-9127-6
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00357-013-9127-6