Algorithmic information theory
| Authors |
|
|---|---|
| Publication date | 2008 |
| Number of pages | 37 |
| Publisher | Amsterdam: Institute for Logic, Language and Computation |
| Organisations |
|
| Abstract |
We introduce algorithmic information theory, also known as the theory of Kolmogorov complexity. We explain the main concepts of this quantitative approach to defining `information'. We discuss the extent to which Kolmogorov's and Shannon's information theory have a common purpose, and where they are fundamentally different. We indicate how recent developments within the theory allow one to formally distinguish between `structural' (meaningful) and `random' information as measured by the Kolmogorov structure function, which leads to a mathematical formalization of Occam's razor in inductive inference. We end by discussing some of the philosophical implications of the theory.
|
| Document type | Working paper |
| Note | Information on published version available at http://www.elsevier.com/wps/find/bookdescription.cws_home/716648/description#description |
| Published at | http://arxiv.org/abs/0809.2754 |
| Downloads | |
| Permalink to this page | |