…information is received in inverse proportion to its predictability.

Mike Ford, This American Life #258: Leaving the Fold. (Keep that one around the next time you’re underestimated.) (via heather-rivers)

I was flicking through Cryptography: An Introduction by Nigel Smart this morning to keep my crypto-sec neurons from rusting over and this quote rings true with information theory. Information is entropy: you learn information when something you couldn’t predict happens.

Submitted by marek on Tue, 02/10/2009 - 12:53