25.08.2015 Views

In the Beginning was Information

6KezkB

6KezkB

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

as expressing <strong>the</strong> highest possible freedom from interference. <strong>In</strong> this waya very exciting basic phenomenon of physiology could be understood bymeans of <strong>the</strong> new concepts of information <strong>the</strong>ory.”It should now be clear that Shannon’s information <strong>the</strong>ory is very importantfor evaluating transmission processes of messages. But, as far as <strong>the</strong> messageitself is concerned, it can only say something about its statisticalproperties, and nothing about <strong>the</strong> essential nature of information. This isits real weakness as well as its inherent propensity for leading to misunderstandings.The German cyberneticist Bernhard Hassenstein rightlycriticises it in <strong>the</strong> following words: “It would have been better to devisean artificial term, ra<strong>the</strong>r than taking a common word and giving it a completelynew meaning.” If we restrict Shannon’s information to one of <strong>the</strong>five aspects of information, <strong>the</strong>n we do obtain a scientifically sound solution[G5]. But without <strong>the</strong> extension to <strong>the</strong> o<strong>the</strong>r four levels of information,we are stuck with <strong>the</strong> properties of a transmission channel. No science,apart from communication technology, should limit itself to just <strong>the</strong>statistical level of information.Natural languages may be analysed and compared statistically by meansof Shannon’s <strong>the</strong>ory, as we will now proceed to do.A1.4 Statistical Analysis of LanguageIt is possible to calculate certain quantitative characteristics of languagesby means of Shannon’s information <strong>the</strong>ory. One example of such a propertyis <strong>the</strong> average information content of a letter, a syllable, or a word. <strong>In</strong>equation (9) this numerical value is denoted by H, <strong>the</strong> entropy.1. Letters: If, for <strong>the</strong> sake of simplicity, we assume that all 26 letters plus<strong>the</strong> space between words occur with <strong>the</strong> same frequency, <strong>the</strong>n we haveH 0 = lb 27 = log 27/log 2 = 4.755 bits/letter. (11)But it is known that <strong>the</strong> frequency of occurrence of <strong>the</strong> different letters ischaracteristic of <strong>the</strong> language we are investigating [B2 p 4]. The probabilityp i of occurrence of single letters and <strong>the</strong> space are given for Englishand German in Table 1, as well as <strong>the</strong> average information content per letter,H. On applying equation (9) to <strong>the</strong> various letter frequencies p i in German,<strong>the</strong> average information content (= entropy) of a symbol is given by19630H 1 = ∑ p i x lb(1/p i ) = 4.112 95 bits/letter. (12)i=1

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!