NATS 1700 6.0 COMPUTERS,  INFORMATION  AND  SOCIETY

Lecture 6: Information : Shannon's Definition

| Previous | Next | Syllabus | Selected References | Home |

| 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 | 13 | 14 | 15 | 16 | 17 | 18 | 19 | 20 | 21 | 22 |

 

Introduction

  • For a nice introduction to the concept of information, read  Read !  Information Theory and Practice: Once Upon a Time Information Didn't Even Exist, a review of two recent books on the subject, on Salon 21st. Other classical texts are briefly reviewed by Laurence Heglar in  Read !  Information Theory.
  • A rather unique project, How Much Information?, has been undertaken by a group of "faculty and students at the School of Information Management and Systems at the University of California at Berkeley....This study is an attempt to measure how much information is produced in the world each year...The world's total yearly production of print, film, optical, and magnetic content would require roughly 1.5 billion gigabytes of storage. This is the equivalent of 250 megabytes per person for each man, woman, and child on earth." Luckily, much of this 'information' has probably little 'meaning' for most people.
  • A rather famous introduction to communication is provided by Colin Cherry in On Human Communication (3rd edition. 1978 MIT Press).
  • For a good summary of Shannon's ideas read  Read !  Brief Excerpts from Warren Weaver’s Introduction to Claude Shannon’s The Mathematical Theory of Communication.
  • It is always wise to approach anything with some measure of humour. Rich Tennant, "the father of the computer cartoon," can certainly help us in this respect. Look for is work at The 5th Wave.

 
Topics

  • The word information is used everywhere nowadays. 'Information technology' (and its acronym 'IT'), 'the information age,' 'the information revolution,' and other variations on the theme are waved about like flags. We also know that information and computers somehow belong together. It is therefore necessary to ask ourselves if the concept behind this word is indeed still the familiar one. Is information just another term for facts, experience, knowledge?
  • Although the two concepts are by no means identical, we will assume as a given that information and communication are related. For example, in The Evolution of Communication (The MIT Press, 1997, p. 6), Marc Hauser writes: "The concepts of information and signal form integral components of most definitions of communication...Thus information is a feature of an interaction between sender and receiver. Signals carry certain kinds of information content, which can be manipulated by the sender and differentially acted upon by the receiver.." Statements such as this seem pretty clear, but they are qualitative, and may not be suitable in situations where it is necessary to quantify information, signals, etc. This is obviously the case when we deal with computers and other devices that process information. Various attempts to sharpen the concept of information were thus made, mostly in the first part of the 20th century.
  • For instance, in 1917 and later in 1928, Nyquist found the theoretical limits on how much information can be transmitted with a telegraph, and the frequency band (or bandwidth) needed to transmit a given amount of information. In this case, Nyquist defined information in terms of number of dots and dashes a telegraph system can exchange in a given time. In 1928, Hartley defined the information content of a message as a function (the logarithm) of the number of messages which might have occurred. For example, if we compare two situations, one in which six messages are possible, and another in which fifty-two messages are possible (think of a dice, and of a deck of cards, respectively), Hartley would say that a message about which face of the dice came up would carry less information than a message about which card came up on top of the deck of cards.
  • In 1948, Claude E Shannon, and American engineer, took a fresh look at information from the point of view of communication. For Shannon, not unlike Hartley, information is that which reduces a receiver's uncertainty. In a given situation, any message can be thought of as belonging to a set of possible messages. For example, if I ask a weather forecaster how's the sky over a certain region, the possible answers I could receive would be something like 'clear,' 'partly cloudy,' 'overcast.' Before I get the answer, I am uncertain as to which of the three possibilities is actually true. The forecaster's answer reduces my uncertainty. The larger the set of possible messages, the greater is the reduction of uncertainty. If the set contains only one possibility, then I already know the answer, I am certain about it, and Shannon would say that no information is conveyed to me by the actual answer. This last observation is important, because in our daily lives we may not be prepared to accept it. A message may not contain any new factual information, but the expression of my interlocutor's face or the tone of his voice do carry information that is meaningful to me. We must distinguish therefore between the vernacular sense of information, and the technical one used in telecommunications. As it always happens in science, to quantity the description of a phenomenon makes that description usable in a precise, predictable, controllable way. But this happens at the expenses of most of their familiar denotations and connotations. That's what we mean when we complain that science is drier, poorer than life. To make concepts unambiguous, so that they can be logically manipulated, we have no choice but to discard the ambiguities that language allows.

    Claude Shannon

    Claude Shannon

  • For Shannon therefore information is a form of processed data about things, events, people, which is meaningful to the receiver of such information because it reduces the receiver's uncertainty. Here is another way to express Shannon's definition: "The more uncertainty there is about the contents of a message that is about to be received, the more information the message contains. He was not talking about meaning. Instead, he was talking about symbols in which the meaning is encoded. For example, if an AP story is coming over the wire in Morse code, the first letter contains more information than the following letters because it could be any one of 26. If the first letter is a 'q,' the second letter contains little information because most of the time 'q' is followed by 'u.' In other words, we already know what it will be."
  • Shannon formalized this definition in a formula which is now so famous that it deserves at least one appearance in these lectures:  H = log2n  . H is the quantity of information carried by one of n possible (and equally probable, for simplicity) messages; log2n is the logarithm in base 2 of n. For example, log28 = 3, log232 = 5, etc. Notice in particular that when n = 1, then H = 0. In words, when a given message is the only possible one, then no information is conveyed by such message. Finally, notice that if there are only two possible messages, say 'yes' and 'no,' then the amount of information of each message will be  H = log22 = 1. In this case we say that the message carries 1 bit of information. The word "bit' is a contraction of the expression 'binary digit.'
  • Shannon was able to prove that his formula  H = log2n  represents the maximum amount of information that a message can carry. Noise and other factors will usually reduce the actual figure.
  • Let's now return to communication. What happens when two people communicate? They do exchanges messages, of course. But there is more. For example, they may be speaking to each other on a subway car during rush hour, and the surrounding buzz may make communications difficult. Or lightning may introduce spurious signals in the telegraph wires. Or the cat may have spilled some milk on your newspaper smudging some of the words. In fact noise is always present, and in a fundamental sense it can never be completely eliminated. Shannon proceeded therefore to formalize the concept of noise, but I will spare you the technicalities.
  • Here, in conclusion, is the model of communication that Shannon adopted.

Shannon's Model

Questions and Exercises
  • Apply Shannon's model of communication to a person reading the newspaper.
  • What is the difference between 'information' and 'meaning'?
  • If 'meaning' is excluded from the technical definition of information, and if information so defined is what computers process, what limitations will computers have?

 


Picture Credit: University of Michigan
Last Modification Date: 07 July 2008