Vance suggested I start a thread on information. What you will find are four posts (all lengthy I'm afraid) covering what I know of information theory (not much) and genetic information, followed by a critique of the Carl Weiland article "Variation, Information and the Created Kind" to which Liberty Wing referred us. If you are reading this before I have completed all four posts, please refrain from posting a reply until all four introductory posts are up. Then feel free to critique.
Part One
The latest attempt of YECists to refute evolution is the concept of "no new information" illustrated recently by the brief appearance of Liberty Wing on Vance's "Why I post" thread. Vance has suggested that it is time we had a thread on genetic information.
Typically, creationists never define "information" or provide any means of measuring it. Yet they claim that information decreases but never increases. Clearly the claim cannot be backed up without a measuring tool.
Scientists, on the other hand, have defined information and measured it. There are two recognized theories of information, and they each treat information in a slightly different (though compatible) ways.
One is the Kolmogorov-Chaitin (K-C) theory. It was developed from their work on computer systems. The problem K-C faced was that of a computer receiving information from another computer. How does the receiving computer know it has received the whole message and not just a part of it?
This breaks down into two questions. The first is: What is the total amount of information in the message?" Since computer messages are transmitted in binary code, there is no problem with defining the unit of information. It is the "bit". Each bit is either a 1 or a 0. The total amount of information in a message is the total number of bits. And as you know, every time you create a file, your computer counts how many bits there are in the message.
The tougher question is How does the receiving computer know it is getting the right bits? No point having the right number of bits if the bits are wrong. If one is receiving 1s when one ought to be receiving 0s.
On the other hand, if each bit needs to be described separately, the information about what information is in the message is as long as the message itself. The description of the content of the message needs to be compressed as far as possible. In the K-C system, the amount of information in a message is generally measured by how much computer code it takes to describe the content.
Repetitious information is easily described in a single line. If you have a string of 20 0s it can be described as
"0 x 20"
If you double the number of 0s, the string can be described as
"0 x 40"
Both descriptions take the same amount of code to describe. It is in this sense that duplication can be said to add "no new information".
Repetitious patterns can alse be described with few lines of code. If you have a string that consists of ten repetitions of 110 it can be described as
1. "1 x 2"
2. "0 x 1"
3. Repeat 1 & 2 10 times
Again, duplicating this pattern merely involves changing the actual number of repetitions. It takes no more code to describe 100 repetitions than it takes to describe 10.
One thing creationists seem not to notice: while it is trivially true that if you measure information by how much code it takes to describe it, duplications add no new information, the reverse is also true. If you decrease the number of repetitions, it still takes the same amount of code to describe the reduced number of bits as to describe the larger number. So if you reduce the string from 20 to 10 repetitions---there has been no loss of information either!
Finally, as K-C noted, repetition is compressible. It takes no more code to describe 500 repetitions of a bit or a repeated pattern than to describe 5. The reverse of this is that random strings are not as easily compressible. If a string reads something like:
1001110101100101
a line of code must be created for each switch from 1 to 0.
This string would need 11 lines of code as compared to the one or three lines for the duplicated strings above. By this measure, a random string contains more information than a string of duplications or duplicated patterns. Randomness=MORE information.
Now, let's look at the Shannon theory. Shannon's theory was developed out of his work for Bell Labs studying ways of overcoming "noise" in a system of information transfer.
The problem with transmitting a message is that the system gets between the sender and the receiver and adds "noise" to the message. Sometimes this is quite obvious to the people at either end of transmission as static. But even a very quiet system still has a discernable level of noise. Noise damages or destroys or blocks or garbles the message so that what is received is not precisely what was sent.
In Shannon theory "randomness" describes the level of information-destroying noise in the system. So in the Shannon system Randomness=LESS information.
Still a third way to measure information (though I hesitate to call it a scientific measure) is by what Dembski calls its "specified complexity". For example, take the following letters of the Enlgish alphabet: A, C, E, N, O. In addition to this alphabetical arrangement, once can create 119 other orders if repeats of the same letter are not allowed. (5^5 if repeats are allowed.) But most of these are meaningless strings of letters. Only an arrangement such as CANOE or OCEAN is specified as having a meaning, as conveying useful information.
Note that both K-C and Shannon handle this question of specificity differently. In K-C the question of meaning does not arise. ACENO, NOECA, OCEAN, CANOE all contain the same amount of information. In Shannon theory, if the original message is OCEAN, then all other arrangements, including CANOE, are equally random and represent loss of information.
Part One
The latest attempt of YECists to refute evolution is the concept of "no new information" illustrated recently by the brief appearance of Liberty Wing on Vance's "Why I post" thread. Vance has suggested that it is time we had a thread on genetic information.
Typically, creationists never define "information" or provide any means of measuring it. Yet they claim that information decreases but never increases. Clearly the claim cannot be backed up without a measuring tool.
Scientists, on the other hand, have defined information and measured it. There are two recognized theories of information, and they each treat information in a slightly different (though compatible) ways.
One is the Kolmogorov-Chaitin (K-C) theory. It was developed from their work on computer systems. The problem K-C faced was that of a computer receiving information from another computer. How does the receiving computer know it has received the whole message and not just a part of it?
This breaks down into two questions. The first is: What is the total amount of information in the message?" Since computer messages are transmitted in binary code, there is no problem with defining the unit of information. It is the "bit". Each bit is either a 1 or a 0. The total amount of information in a message is the total number of bits. And as you know, every time you create a file, your computer counts how many bits there are in the message.
The tougher question is How does the receiving computer know it is getting the right bits? No point having the right number of bits if the bits are wrong. If one is receiving 1s when one ought to be receiving 0s.
On the other hand, if each bit needs to be described separately, the information about what information is in the message is as long as the message itself. The description of the content of the message needs to be compressed as far as possible. In the K-C system, the amount of information in a message is generally measured by how much computer code it takes to describe the content.
Repetitious information is easily described in a single line. If you have a string of 20 0s it can be described as
"0 x 20"
If you double the number of 0s, the string can be described as
"0 x 40"
Both descriptions take the same amount of code to describe. It is in this sense that duplication can be said to add "no new information".
Repetitious patterns can alse be described with few lines of code. If you have a string that consists of ten repetitions of 110 it can be described as
1. "1 x 2"
2. "0 x 1"
3. Repeat 1 & 2 10 times
Again, duplicating this pattern merely involves changing the actual number of repetitions. It takes no more code to describe 100 repetitions than it takes to describe 10.
One thing creationists seem not to notice: while it is trivially true that if you measure information by how much code it takes to describe it, duplications add no new information, the reverse is also true. If you decrease the number of repetitions, it still takes the same amount of code to describe the reduced number of bits as to describe the larger number. So if you reduce the string from 20 to 10 repetitions---there has been no loss of information either!
Finally, as K-C noted, repetition is compressible. It takes no more code to describe 500 repetitions of a bit or a repeated pattern than to describe 5. The reverse of this is that random strings are not as easily compressible. If a string reads something like:
1001110101100101
a line of code must be created for each switch from 1 to 0.
This string would need 11 lines of code as compared to the one or three lines for the duplicated strings above. By this measure, a random string contains more information than a string of duplications or duplicated patterns. Randomness=MORE information.
Now, let's look at the Shannon theory. Shannon's theory was developed out of his work for Bell Labs studying ways of overcoming "noise" in a system of information transfer.
The problem with transmitting a message is that the system gets between the sender and the receiver and adds "noise" to the message. Sometimes this is quite obvious to the people at either end of transmission as static. But even a very quiet system still has a discernable level of noise. Noise damages or destroys or blocks or garbles the message so that what is received is not precisely what was sent.
In Shannon theory "randomness" describes the level of information-destroying noise in the system. So in the Shannon system Randomness=LESS information.
Still a third way to measure information (though I hesitate to call it a scientific measure) is by what Dembski calls its "specified complexity". For example, take the following letters of the Enlgish alphabet: A, C, E, N, O. In addition to this alphabetical arrangement, once can create 119 other orders if repeats of the same letter are not allowed. (5^5 if repeats are allowed.) But most of these are meaningless strings of letters. Only an arrangement such as CANOE or OCEAN is specified as having a meaning, as conveying useful information.
Note that both K-C and Shannon handle this question of specificity differently. In K-C the question of meaning does not arise. ACENO, NOECA, OCEAN, CANOE all contain the same amount of information. In Shannon theory, if the original message is OCEAN, then all other arrangements, including CANOE, are equally random and represent loss of information.