Abstract
Intuitively, there is a relation between measures of spatial dependence and information theoretical measures of entropy. For instance, we can provide an intuition of why spatial data is special by stating that, on average, spatial data samples contain less than expected information. Similarly, spatial data, e.g., remotely sensed imagery, that is easy to compress is also likely to show significant spatial autocorrelation. Formulating our (highly specific) core concepts of spatial information theory in the widely used language of information theory opens new perspectives on their differences and similarities and also fosters cross-disciplinary collaboration, e.g., with the broader AI/ML communities. Interestingly, however, this intuitive relation is challenging to formalize and generalize, leading prior work to rely mostly on experimental results, e.g., for describing landscape patterns. In this work, we will explore the information theoretical roots of spatial autocorrelation, more specifically Moran’s I, through the lens of self-information (also known as surprisal) and provide both formal proofs and experiments.
| Original language | English |
|---|---|
| Title of host publication | 16th International Conference on Spatial Information Theory (COSIT 2024) |
| Editors | Benjamin Adams, Amy L. Griffin, Simon Scheider, Grant McKenzie |
| ISBN (Electronic) | 9783959773300 |
| DOIs | |
| Publication status | Published - Sept 2024 |
Austrian Fields of Science 2012
- 507003 Geoinformatics
- 102001 Artificial intelligence
- 102035 Data science
Keywords
- Self-Information
- Surprisal
- Information Theory
- Spatial Autocorrelation
- Moran’s I
Fingerprint
Dive into the research topics of 'Probing the Information Theoretical Roots of Spatial Dependence Measures'. Together they form a unique fingerprint.Prizes
-
Best paper award
Wang, Z. (Recipient), Janowicz, K. (Recipient), Mai, G. (Recipient) & Majic, I. (Recipient), 20 Sept 2024
Prize: Prize, award or honor