25.08.2015 Views

In the Beginning was Information

6KezkB

6KezkB

SHOW MORE
SHOW LESS
  • No tags were found...

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

A1.2 Ma<strong>the</strong>matical Description of Statistical<strong>In</strong>formationA1.2.1 The Bit: Statistical Unit of <strong>In</strong>formationOne of <strong>the</strong> chief concerns in science and technology is to express resultsas far as possible in a numerical form or in a formula. Quantitative measuresplay an important part in <strong>the</strong>se endeavours. They comprise twoparts: <strong>the</strong> relevant number or magnitude, and <strong>the</strong> unit of measure. The latteris a pre-determined unit of comparison (e. g. metre, second, Watt)which can be used to express o<strong>the</strong>r similarly measurable quantities.The bit (abbreviated from binary digit) is <strong>the</strong> unit for measuring informationcontent. The number of bits is <strong>the</strong> same as <strong>the</strong> number of binary symbols.<strong>In</strong> data processing systems information is represented and processedin <strong>the</strong> form of electrical, optical or mechanical signals. For this purpose itis technically extremely advantageous, and <strong>the</strong>refore customary, toemploy only two defined (binary) states and signals. Binary states have<strong>the</strong> property that only one of <strong>the</strong> two binary symbols can be involved at acertain moment. One state is designated as binary one (1), and <strong>the</strong> o<strong>the</strong>r asbinary nought (0). It is also possible to have different pairs of binary symbolslike 0 and L, YES and NO, TRUE and FALSE, and 12 V and 2 V. <strong>In</strong>computer technology a bit also refers to <strong>the</strong> binary position in a machineword. The bit is also <strong>the</strong> smallest unit of information that can be representedin a digital computer.When text is entered in a computer, it is transformedinto a pre-determined binary code and also stored in this form.One letter usually requires 8 binary storage positions, known as a byte.The information content (= storage requirement) of a text is <strong>the</strong>ndescribed in terms of <strong>the</strong> number of bits required. Different pieces of textare thus accorded <strong>the</strong> same information content, regardless of sense andmeaning. The number of bits only measures <strong>the</strong> statistical quantity of <strong>the</strong>information, with no regard to meaningfulness.Two computer examples will now illustrate <strong>the</strong> advantages (e. g. to helpdetermine <strong>the</strong> amount of storage space) and <strong>the</strong> disadvantages (e. g. ignoring<strong>the</strong> semantic aspects) of Shannon’s definition of information:Example 1: Storage of biological information: The human DNA molecule(body cell) is about two metres long when fully stretched and it containsapproximately 6 x 10 9 nucleotides (<strong>the</strong> chemical letters: adenin,cytosin, guanin, and thymin). How much statistical information is thisaccording to Shannon’s defintion? The N = 4 chemical letters, A, C, G,and T occur nearly equally frequently; <strong>the</strong>ir mean information content isH = lb 4 = (log 4)/(log 2) = 2 bits. The entire DNA thus has an informationcontent of I tot = 6 x 10 9 nucleotides x 2 bits/nucleotide = 1.2 x 10 10 bits178

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!