I agree with
@Warden_of_the_Storm that this is incorrect. (And yes, what you quoted Dawkins as saying is wrong.)
just to be clear, it was his quote- also writing "The machine code of the genes is uncannily computer-like. Apart from differences in jargon, the pages of a molecular biology journal might be interchanged with those of a computer engineering journal."
DNA is not pure digital information. For example, protein-coding genes have a digital signal that tells the cell's machinery to stop, ending the protein. So far, completely digital.
yes, just like memory buffers in computer code, that information is digital.
Sometimes, however, the machinery makes a mistake
absolutely, the machinery, or medium that carries the digital information, is distinct from the information itself, and of course can corrupt the digital information
just as an old mechanical hard drive can wear out and begin to make mistakes
That's a pretty small error, however.
I do take your point, it would be more precise for Dawkins to say that the information
stored in DNA is purely digital-
but I believe that was his intent.
A bigger problem is with your statement is that DNA requires 'precise sequences' to function. In reality, a single protein (i.e. a specific sequence of amino acids) can be coded for in an immense number of ways, thanks to the redundancy of the genetic code, and the number of similar proteins that perform exactly the same function with a somewhat different string of amino acids is much larger still. And that's sticking with what is essentially the same protein. Now add in all of the unrelated proteins that could perform the same function with completely different amino acid sequences and you'll have a combinatorial explosion of possible sequences doing the same thing.
What you're describing also conforms to Dawkins observation that 'The machine code of the genes is uncannily computer-like'.
The reason there is redundancy, is because there are only 4 nucleotides to describe 22 amino acids. This is done by grouping the nucleotides into groups of three called codons giving you 4x4x4 combinations.
Computer characters were encoded with a similar system, called ASCII, using 8 bits (I think 16 now) to describe each character. Also Unicode to give a much larger set of symbols
And yes, redundancy is possible since there are more combinations than things to describe
en.wikipedia.org
Unicode has a certain amount of duplication of
characters. These are pairs of single Unicode code points that are
canonically equivalent.
I take your point that on one hand, this would seem to allow more room for error, though in the larger context this is negligible (redundancy in Unicode doesn't really help a chimp to write war and peace accidentally)
But it does presents a very prominent problem regarding the addition of new volumes of genetic information as we talked about earlier in this thread.
Because yes, if you merely mutate a nucleotide, a point mutation, you have some chance that the codon will still read as it did.
But if you ADD a new nucleotide, aka an insertion mutation, this now throws off the frame reference for each proceeding codon,
which is why such insertion mutations are called
nonsense mutations. They essentially scramble everything that follows. It's actually one of they key hurdles in macroevolution v microevolution-
adding something new rather than merely degrading what you have
Here's your real problem: you have a fundamental misunderstanding of what evolution does. Why would you calculate the odds of a particular sequence occurring by chance? Evolution isn't aiming at a particular sequence; it's not aiming for anything. What you have to do is calculate the probability that any possible mutation in a species (given the wide range of genetic variation already present) will produce any beneficial effect for the organism.
Well no, that's a fundamental and very common misunderstanding about evolution, which we covered earlier in the thread, don't want to make this post too long, but in short:
There are many mutations that are entirely probable through random replication errors, whereby functionality of a protein is
destroyed, providing a benefit. And this is generally what we see in micro-evolution.
e.g. A bear losing it's ability to produce pigment in it's fur, may certainly provide a benefit, in a niche environment at least.
But you see the problem; you cannot 'macro-evolve' a bacteria to a human being by merely breaking things in the bacteria..