In What Sense is the Kolmogorov-Sinai Entropy a Measure for Chaotic Behaviour? - Bridging the Gap Between Dynamical Systems Theory and Communication Theory.
On an influential account, chaos is explained in terms of random behaviour; and random behaviour in turn is explained in terms of having positive Kolmogorov-Sinai entropy (KSE). Though intuitively plausible, the association of the KSE with random behaviour needs justification since the definition of the KSE does not make reference to any notion that is connected to randomness. I provide this justification for the case of Hamiltonian systems by proving that the KSE is equivalent to a generalized version of Shannon's communication-theoretic entropy under certain plausible assumptions. I then discuss consequences of this equivalence for randomness in chaotic dynamical systems.
Monthly Views for the past 3 years
Monthly Downloads for the past 3 years
Actions (login required)