Complexity Explorer Santa Few Institute

Explore


Shannon information

Shannon information, described by mathematician Claude Shannon, is a measure of the unpredictability of a message, given a message source, communication channel, and reciever. This unpredictability, or entropy, relates to the sum of the probabilities of each element in the message. The higher the entropy, the more information that is present.


Topics
Entropy, Information Theory
Difficulty
1
Sources
[]