@parismic

Normalized Entropy Aggregation for Inhomogeneous Large-Scale Data

, and . Theory and Applications of Time Series Analysis, page 19--29. Cham, Springer International Publishing, (2019)

Abstract

ItConceicão Costa, Maria da wasMacedo, Pedro already in the fifties of the last century that the relationship between information theory, statistics and maximum entropy was established, following the works of Kullback, Leibler, Lindley and Jaynes. However, the applications were restricted to very specific domains and it was not until recently that the convergence between information processing, data analysis and inference demanded the foundation of a new scientific area, commonly referred to as Info-Metrics 1, 2. As a huge amount of information and large-scale data have become available, the term ``big data'' has been used to refer to the many kinds of challenges presented in its analysis: many observations, many variables (or both), limited computational resources, different time regimes or multiple sources. In this work, we consider one particular aspect of big data analysis which is the presence of inhomogeneities, compromising the use of the classical framework in regression modelling. A new approach is proposed, based on the introduction of the concepts of info-metrics to the analysis of inhomogeneous large-scale data. The framework of information-theoretic estimation methods is presented, along with some information measures. In particular, the normalized entropy is tested in aggregation procedures and some simulation results are presented.

Description

Normalized Entropy Aggregation for Inhomogeneous Large-Scale Data | SpringerLink

Links and resources

Tags