Lecture 7: Information and Entropy
Note
My greatest concern was what to call it. I thought of calling it ‘information,’ but the word was overly used, so I decided to call it ‘uncertainty.’ When I discussed it with John von Neumann, he had a better idea. Von Neumann told me, ‘You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, no one really knows what entropy really is, so in a debate you will always have the advantage. Claude Shannon, 1971
Warning
This lecture corresponds to Chapter 15 of the textbook.
Summary
Attention
In this lecture, we look at how one can quantify information. We all have a pretty good intuitive understanding of the amount of information available in one claim and can easily tell, among two claims, which one has the most information. For example, take the two pieces of information: “I live on Earth.” and “I live in NY State”. It is clear that the two pieces of information do not convey the same amount of new knowledge. Claude Shannon realized that the amount of information is proportional to the inverse of the probability for the claim to occur. In other words, if a less likely event takes place, you will get more information if someone tells you something about that event compared to added knowledge about the more likely on.
Formally, this leads to the definition of information (in units of bits that has a probability to take place:
We understand the need for the logarithmic function: if you are given two independent statements, knowing the two statements increases the chances by multiplying the probabilities.
This leads to the notion of average information, or Shannon Entropy:
This definition is reminiscent of Gibbs’ definition of entropy we saw in Lecture 6: Entropy. (the difference is that is no longer the Boltzmann constant).
The big leap is that information, since it carries entropy, can be considered as a physical quantity (Rolf Landauer). After all, this is not surprising since, in thermodynamic, we defined entropy as a measure of a number of microstates a system can be in to realize a macrostate. This uncertainty (that is: lack of knowledge) is certainly related to information!
Interestingly, this allows us to resolve the issue with Maxwell’s demon related to the irreversibility of the Joule expansion we saw in the previous lecture. The demon must lose information and thus increase entropy during the sorting of the gas molecules!
Finally, in this lecture, we look into the issues of data compression and discussed a couple of examples of application of Bayes theorem for conditional probabilities:
Where we defined , as the independent probabilities of and as the probability of given is true. Likewise, is the probability of given is true.
Learning Material
Copy of Slides
The slides for Lecture 7 are available in pdf format here: pdf
Screencast
Test your knowledge
Consider the following two statements. (a) Students who graduate with a bachelor in physics do so by passing IQM and (b) Students who graduate with a bachelor in applied physics do so by passing IQM. Statement (a) occurs with probability and statement (b) occurs with a probability . What are is Shannon information of each statement, in bits (we use basis and suppose )?
and bits. There is more information in statement b.
and bits. There is more information in statement b.
and bits. There is more information in statement a.
and bits. There is more information in statement b.
Mrs. Bonnie T. has three kittens. Two of them are male. What is the probability that the third one is a female?
75%
50%
37.5%
25%
Mrs. Bonnie T. has three kittens. The two tallest ones are male. What is the probability that the third one is a female?
75%
50%
37.5%
25%
The less information you know about a system, the largest its entropy.
True
False
It depends
Hint
Find the answer keys on this page: Answers to selected test your knowledge questions. Don’t cheat! Try solving the problems on your own first!
Homework Assignment
Solve the following problems from the textbook: