Information Theory: A Tutorial Intorudction, Chs 1-2 by James V. Stone

December 21, 2022

1. Useful information 

    • (maximize) informative redundancy 

    • (minimize) non-informative redundancy 


2. Noise
 

3. Minimizing "effort" required for transmitting each portion of an informative signal: selection in biology/language/culture 
  1. i) articulatory? how much effort? (e.g., Spanish vs. English vowels) 

    i) cognitive effort? 

    Iii) motor plan for producing similar sounds, 

    v) sounds including some noise 


4. Bits and binary digits 

m outcomes = 2^n binary choices 

n binary choices = log2(m outcomes) 

n bits = log2(m outcomes) 

 

 

5. Random variable: e.g., X(head) = 1, X(tall) = 0 / a kind of function  

6. Rather than an actual variable ? 

7. Schema for message transmission: a message is an ordered sequence of symbols corresponding to outcome values of random variables 



8. Surprisal and entropy 

9. Entropy 

10. Entropy of a system

11. Some assumptions
  • Independent and Identically Distributed (a.k.a. i.i.d.)
  • Stationary: doesn't change through time
  • Ergodic: system can be represented by a reasonably large number of measurements 
12. Uniform Information Density Hypothesis: if language is optimized for information transmission, it should tend toward equal surprisal for symbols per unit time 

Discussion: The actual entropy of written English letters lower than that of English estimated from the individual prob of letters from a big sample of text? Cuz that is not independent! More skewed because of the context! 

Categories: Book Summary, Information Theory

Original post: https://cheonkamjeong.blogspot.com/2022/12/book-summary-information-theory.html