And if you'll notice, Alate was discussing the situation you have presented (where we have all the data) as opposed to the explanation I was giving to BJ where we did not.
So? You're attempting to dodge, but I'm not gonna let you. Alate_One asked which of two strings had more information content. This is well after you had dug-in and declared that you're sticking with Tom Schneider's definition of "information" which only applies to the
transmission of data, not the source data itself. And you answered A_O, (correctly, according to standard usage of the word) stating that the information content is dependent on the compressibility of the source data. Then you denied that you did this very thing. Now you're trying to dodge it.
I answered the question correctly and completely in synch with what I've been saying all along.
Nope, you're equivocating, using whichever definition lets you squirm out of admitting your previous mistakes.
You've just decided that I cannot talk about the compressibility while still insisting that the i-word is only to be understood as reduction in uncertainty at the receiver.
Not at all, I will let you talk about the
compressibility of data, or its
entropy, but I will remind you that you have already stated that the definition of
information that you yourself are sticking to cannot be applied to the source data.
Here's what you can do. You can decode the genome of an organism. But in that process we are the receiver and must account for noise. We can decode the genome from another organism of the same kind. Again we are the receiver and must account for noise.
It's not necessary to account for noise, when the accuracy of our ability to read it makes any tiny errors negligible.
Once we have the two genomes written down we can assume they are accurate.
Good - you are agreeing with what I just said, about any potential errors being negligible.
After we have assumed they are accurate we can compare them to see which contains more information.
We can't compare the
information in the two sets of data, because according to your previous declaration, that word is off-limits for the discussion of sets of data.
Defining the i-word as reduction of uncertainty at receiver does not prohibit me from comparing two sets of data.
You can certainly compare two sets of data, measuring the compressibility or the entropy, but your own definition prohibits you from comparing the
information in two sets of data. You've said this yourself, and you keep switching when it suits you.
I can when we know that the entropy is equal to the reduction of uncertainty at receiver. Just because I define the product as the result of multiplication, doesn't mean I can't find a product by using addition.
That first sentence makes no sense, and the second doesn't help shed any light. Care to restate that?
But I cannot do this in the real world because in the real world we must always account for noise.
We've already discussed the effect of noise. Since we can read the data as carefully as we want, we can arbitrarily reduce the noise so that it's below any relevant threshold. The noise can be neglected when we're talking about sets of data like text on a page or the sequence of DNA.