Some Bits of Information Theory =============================== Entropy ------- Information is linked to probability: an unlikely event is associated with a high information content. Claude Shannon was the first to give a solid mathematical model for this intuition. For an event $A$ with probability $P(A)$ he defined **self information** .. math:: I(A) = -\log_b(P(A)) An event with probability one has zero self information. Nowing that it will rain every single day does not really provide any information. An event with very low probability is so rare that in case it does occur carries with it a lot of information. Note that an event with $P(A)=0$ is not really an event in the sense that it will never occur. In information theory we will only consider events for which $0