arXiv Analytics

Sign in

arXiv:cond-mat/0102181AbstractReferencesReviewsResources

Regularities Unseen, Randomness Observed: Levels of Entropy Convergence

James P. Crutchfield, David P. Feldman

Published 2001-02-09Version 1

We study how the Shannon entropy of sequences produced by an information source converges to the source's entropy rate. We synthesize several phenomenological approaches to applying information theoretic measures of randomness and memory to stochastic and deterministic processes by using successive derivatives of the Shannon entropy growth curve. This leads, in turn, to natural measures of apparent memory stored in a source and the amounts of information that must be extracted from observations of a source in order for it to be optimally predicted and for an observer to synchronize to it. One consequence of ignoring these structural properties is that the missed regularities are converted to apparent randomness. We demonstrate that this problem arises particularly for small data sets; e.g., in settings where one has access only to short measurement sequences.

Comments: 35 pages, 21 fgures, Santa Fe Institute Working Paper 01-02-012
Categories: cond-mat.stat-mech
Related articles: Most relevant | Search more
arXiv:cond-mat/0201273 (Published 2002-01-16)
Can randomness alone tune the fractal dimension?
Effect of randomness in logistic maps
arXiv:cond-mat/0401048 (Published 2004-01-06)
Does randomness in multifractals imply latent dimensions?