Anything that stores inbformation is a channel. Books, CDs etc. As such DNA
seems to fit into this category quite well. A channel can output
information, but this does not make it a source of information. Easy mistake
to make.
> 2. How much can you compress a DNA sequence
> Random mutations will either *increase* or *decrease* the length
> of the compressed string. (On balance, there will be an increase.)
Compression is a useful measure of the informatin content, unfortunately the
generic compression algoirthms do not give the optimum compression. The true
compression can only be obtained by using a good model of the source and
this is obviously not done in generic compression programs.
> 3. DNA as source
> Random mutations will *increase* the information content of the source.
This is just the wrong way of looking at it. My corrections to Glenn's use
of information theory on this topic should show that. As nobody tried to
correct my maths then I assume that Glenn now sees that his use of the
simplistic formula was incorrect?
>
> In a straightforward (and oversimplistic) biology example, suppose we
> have two parents (AA) and (aA). How much information is there in knowing
> the DNA of their offspring?
>
> Well, it could be (AA) or (aA). Probabilities are .5, and .5,
> and the information resolved in learning the actual sequence is just
> 1 bit.
>
> Now suppose there is a small probability that (a->b). Now we have
> (AA), (aA), and (bA). The information is slightly higher now, due to
> the fact that the logarithm is sublinear. The information resolved
> in observing *any of the three* is identical.
This is true if you are transmitting this data, it does not follow that this
is the true information of the source.
To find the source information you need the BEST model possible, the generic
equation for zero memory systems is almost never the best (or even close).
The better method of modeling this is to take the random mutation of (a->b)
as noise on a channel going from parent to offspring. In fact any randomness
of a source is modeled as a source and a noisey channel.
I want to repeat here that sources are NOT random. They are deliberate
meaningful information. This seems to be a major misunderstanding here.
Information theory does not ascribe meaning to a source, but it definitely
ASSUMES that it is meaningful. The purpose of information theory is
therefore to transmit the MEANING of the data in as efficient a manner as
possible.
Of course random noise has no meaning and so the most efficient manner of
transmitting it is just not to....
>
> One of the reasons it is oversimplistic, of course, is that we are usually
> concerned with populations, and having a new (bA) in the population
> can make it much more interesting.
>
> DNA-as-channel is interesting if organisms care explicitly about the
> sequence of their DNA. Since we've only known what it is for a few
> decades, it is usually taken for granted that this is biologically
> unrealistic. That is, organisms care more about their offsprings'
> well-being, and not at all about their DNA. So the 'meaningfulness'
> that guides to a particular model tends to make DNA-as-channel less
> apropriate. (Certainly for somatic protein-manufacture it is
> inappropriate.)
>
ANY information storage system can be modeled as a channel.
> > _____________________________________________________________
> > Do you aggree or disagree with the following statments?
> >
> > "Information theory is pure nonsense! Noise is usually modelled
> as a random
> > source and a random source contains the most information since
> all symbols
> > are equiprobable. Thus the most informative information source
> is noise."
> > _____________________________________________________________
>
> Looks like a trick question. The source conveys lots of information,
> but its not very "informative," since that's something information theory
> lays no claim to be about.
>
Not a trick question. Why don't you have a go at answering it? This question
is used in exams as a general understanding question.
the question is: Does noise have more information than any other signal?
Glenn (and others) previously stated that it had maximal information. That
means that they agree with the above statement (The most informative source
is noise).
Do you want to the Lecturers response to this question? (Dr Roberto Togneri,
http://www.ee.uwa.edu.au/staff/togneri.r.html/)
--------------------------------------------
Brad Jones
3rd Year BE(IT)
Electrical & Electronic Engineering
University of Western Australia