**… proved without reference to infinity and the empty string.**

Some readers have objected to my simple proof that computable transformation of a binary string can result in an infinite increase of algorithmic specified complexity (ASC). Here I give a less-simple proof that there is no upper bound on the difference in ASC of and To put it more correctly, I show that the difference can be any positive real number.

**Updated 12/8/2019:** The assumptions of my theorem were unnecessarily restrictive. I have relaxed the assumptions, without changing the proof.

**Definitions.** Let the set of all binary strings. For all in the *algorithmic specified complexity* of is defined

where is an element of and is a distribution of probability over Here is the algorithmic complexity of conditioned on

**Theorem.** Let be a constant function on Let and be elements of for which holds. Also let be a positive real number. There exists a distribution of probability over such that

[**12/8/2019:** To generalize, replace the first two sentences of the theorem with the following: Let be an element of and let and be elements of such that Let be a computable function on with ]

**Proof.** Noting that by hypothesis,

where and In the case of set and solve for in terms of :

Observe that is precisely equal to if and only if and is strictly less than when Thus distributes probability mass

over the nonempty set excluding and In the case of set and similarly solve for in terms of *QED.*

Your random equations are meaningless. You might as well calculate how many angels fit on a pin head. It’s true that your ID opponents talk about this nonsense called “specified complexity” and “information” (thank you Shannon for screwing up so many generations by mislabeling ‘data’ as ‘information’), but this doesn’t absolve you from clarifying and VALIDATING your assumptions before launching into nonsense “proofs”.

They aren’t meaningless. Dr. English is establishing the AIT conservation of randomness proof doesn’t apply to ASC, which is correct.

You replied too soon. Read again (edited).

Could you expand on this, Eric?

What do you mean by “AIT conservation of randomness proof ” (text reference is fine as an answer).

Why don’t you think the proof is addressing ASC?

Tom: Suppose we limit the domain script-X to finite bit strings. Is there any reason in that case to say c has to be an element of that domain?

Here is my reason: I agree with you and many of the posters that we need to be able to provide a biological interpretation of ASC to understand whether it has any scientific relevance for NS mechanism of evolution.

But as I said in my post in your previous thread, I think that interpretation has to based on populations of genomes and how the relationship to the niche matters. For example, the bit string x could be the concatenation of actual genome bit strings and the Context C somehow captures the niche and so would have nothing to do with interpretation of domain of x. (Of course for infinite populations, C in the domain still applies by coincidence).

(Advertisement) A post by me will appear today at Panda’s Thumb expressing my skepticism that the ASC measure can be used to place any limits on adaptation or on the increase of fitness by natural selection. Basically: why is description by a simple program connected to fitness? It’s not, as far as I can see,

I look forward to the post. I assume you want replies at PT only to avoid having to deal with two different places to answer questions and concerns. I hope Eric will post there.

By the way, my quick and dirty answer to your question about fitness and ASC would be this:

if some bit string C=context somehow captures the niche of an organism

and if the ASC of a bit string representation of one genome is greater than that of a second bit-stringed genome for that context,

then one could argue that the first genome is fitter than the second for the niche represented by that context.

Or at least so it seems to a newbie like me.

So fitness is the “meaning” of the context. I use scare quotes around ‘meaning’ because I agree with Tom that Eric’s use of the term meaning to deone a difference in KC appears to have nothing to do with usual understanding of meaning, eg in natural language or model theory.

Sorry about the misunderstanding here. A string is a finite sequence, by definition. That is is the set of all finite sequences of 0s and 1s. There is a definition of K(x) for some infinite sequences x of 0s and 1s, namely, the computable numbers. But I have no idea of how to handle an infinite sequence c of 0s and 1s in K(x|c). I cannot recall seeing anything but a (finite) binary string c. But my not seeing anything doesn’t rule out the possibility of an infinite sequence.

Sorry, I’m not following. Let me take a shot in the dark. Whatever the environment is, C has to be a binary encoding of it.

I’m mainly concerned with how your OP for this thread relates to Eric’s claims in section 4.2. However, I understand from your long reply in the other thread that you are planning a new OP that will address that in more detail.

Right, I mentioned that in my long post on this in the other thread.

But I should start by explaining what I am trying to do. Based on on that goal, you can then decide whether it is worth your time to reply to this post.

Mainly as an intellectual exercise, I am trying to to see if there is a way to capture evolution by the mechanism of NS using one of these three ASC formalisms and results.

1. Levin’s result of conservation of mutual information. That’s the one Eric used in his earlier OPs here. To try to use that formalism to model NS, I believe you have to interpret the two bit strings x and y and the function f(x) that cannot increase the mutual information of x and y.

2. The hypothesis testing results in section 3 of Eric’s paper (which I think are a generalization of Milosavljevic’s work, although I need to study them both further to confirm that). For that case, we have the probability distributions p, q and the context C. I need to do more work to understand if and how C relates to the two

distributions.

3. Some version of Eric’s claims about conservation of ASC. Here we would need to interpret the x, f(x), p(x), and the context C.

I believe the math of the first two is not in dispute. This is not true of the third, but one could try to interpret the formalism regardless.

As was pointed out in the thread responding to Eric’s earlier attempt, and as Joe re-iterates in his PT post, to capture NS using the formalism of any of the above, one must be able to model a population of genomes and how NS uses fitness in a niche when operating on that population.

If it is possible to capture NS in one of the formalism, then I think that would be a response to Joe’s post in PT. I am not saying that the resulting model would be of any practical use to biologists. I am only saying that it would show it is possible to provide a biological interpretation of some approaches to ASC which captures NS and its role in population genetics.

I have updated the opening post.The place to look first is the opening of Section 4 of the article by Nemati and Holloway. Eric says that there’s something wrong with my argument in Evo-Info 4 that data processing can result in an infinite increase in ASC, i.e., that the difference of and can be infinite. He has generalized the argument slightly, replacing my constant function where denotes the empty (length-zero) string, with the constant function where can be any binary string. (That’s why, in my original statement of the theorem in the opening post, I stipulated a constant function I thought that it was best to go with a theorem that was clearly related to the opening of Section 4. Today I realized, while considering how to reply to your comment, that I should have started with the more general theorem, and then stated the less general theorem, with a constant function, as a corollary.)

The most important thing to understand is that Eric acknowledged, in his comment above, that the theorem is correct, but is now engaging in rhetorical shenanigans to conceal the fact that it refutes a claim that he made hither and yon in cyberspace. You asked him just the right question, in your response above. The reason that you did not get an answer is that it would have exposed what Eric wants desperately to hide. And the reason that Eric is desperate is that he has proclaimed, hither and yon, his great expertise in ID and algorithmic information theory. I do not relish the thought that I am threatening his self-image. And I tell you, in all sincerity, that it’s distressing to see that as Eric fights to preserve his self-image and his public image, his ethical core is evaporating. But I am not going to let up, inasmuch as the Discovery Institute (foolishly) has cast Eric as a public intellectual, and many more people than Eric himself are hurt by his crankery.

You can see, in his (yters’s) comments on Joe Felsenstein’s post at

The Panda’s Thumb, including the one quoted below, that Eric has begun constructing a false narrative in which it was not he, but instead I, who introduced the notion that ASC is conserved in the sense that algorithmic mutual information is conserved — what he has started calling “AIT non growth”:Some points in response:

1. Eric has silently introduced a distinction of “AIT non growth” and the “non growth” that he establishes in Section 4.2. He has not made this distinction before, having previously said that the “law of information conservation (nongrowth),” from algorithmic information theory (AIT), held for ASC as for algorithmic mutual information. You can see at the end of the (familiar) annotated screenshot, below, that he believed that he had refuted, with his new “non growth” result, my refutation of what he now calls “AIT non growth.”

2. In the screenshot, below, it is obvious that Eric defines fASC very differently from ASC. How great is the difference? In the example given at the beginning of Section 4, for which ASC is infinite, fASC is a negative number. Eric then derives an upper bound on fASC, and decorates the mathematical result with rhetoric regarding ASC. Nowhere does he prove any relation of fASC and ASC.

3. In short, Eric does not want to admit to error. So he turns things around, and falsely attributes to me the claim that I have refuted his upper bound on fASC — a result that he egregiously mischaracterizes as “non growth of ASC.”

4. My non-conservation theorems devastate the notion, articulated in the book by Marks, Dembski, and Ewert (

Introduction to Evolutionary Informatics), that ASC is a measure of meaningful information. I otherwise would not have addressed non-conservation of ASC in part 4 of my response to the book.5. As far as I can see, my non-conservation theorems say nothing about the utility of ASC as a statistic. As I indicated in Evo-Info 4, a statistical interpretation of ASC makes sense to me. Three weeks after my post appeared, Bio-Complexity published an article in which George Montañez adopted a perspective similar to mine. In fact, George later added to his article citations of an important hypothesis-testing precedent, from 24 years ago, that I had discovered, and had discussed in my post.

6. Section 2 of the article by Nemati and Holloway is an essay on why Eric feels deeply that ASC is a measure of meaningful information. It is probably based on his Ph.D. dissertation. Section 3 treats ASC not as a measure of meaningful information, but instead as a statistic. It is probably based on Nemati’s master’s thesis. Section 4 of the article, which is clearly due to Eric, and is framed as a response to the first of my non-conservation results, treats ASC in pure abstraction. However, non-conservation of ASC is irrelevant to Section 3, including the “conservation of expected ASC” result. Also, the upper bound for

fASC, not ASC, that Eric derives in Section 4.2 is irrelevant to Section 3.7. In short, my non-conservation results, including the one in the OP, say “something important and negative about ASC” as a measure of meaningful information, not ASC as a statistic. They say nothing at all about Section 4.2, because that is where Eric derives a result for something that is very different from ASC, fASC.

Not to discourage further study, but Section 3 addresses

expectedASC, and Milosavljevic addresses algorithmic mutual information, not its expectation.The “conservation of ASC” is just rhetoric that Eric has attached to a result for fASC, which is very different from ASC. No one has ever applied fASC. It serves no purpose but to provide Eric with a result that he can mischaracterize as “conservation of ASC.”

Why would you do anything at all with

fASC? The “conservation of ASC,” which is nothing but an upper bound onfASC, is an almost-trivial consequence of the definition offASC. Eric obviously built the upper bound into the definition offASC, just so that he would have an upper bound that he could mischaracterize as “conservation of ASC.” Please do not give his crankery any credence.As for your first option:

A gazillion people have answered the ID protestation “But where does the information come from!?” with “The environment!” There is conceivably some sense in relating changes in the frequency distribution of a

populationover fitnesses with changes in the algorithmic mutual information of a (binary encoded)populationand its (binary encoded) environment. However, I want to say that, although natural selection is a (the only known) mechanism of adaptive evolution, I don’t believe that it’s correct to equate the fitness of a genotype with the adaptedness of the corresponding phenotype (or distribution of phenotypes, given that the genotype typically does not fully determine the phenotype) to the environment.If you model the environment explicitly, and your model lacks something that looks like feedback from the environment to the population, then I’d say that your model is categorically wrong. Perhaps

Joe Felsensteinwill disagree with me here. But, notwithstanding my inexpertise on the general topic, I’m fairly confident on this particular point. So if you see the mutual information of the population and environment increasing, the increase is attributable to interaction of the population with its environment. In terms used by Dembski and Marks, the evolutionary process poses queries with offspring, and the environment informs the process with its responses to the queries. However, if you take that literally, then you are reifying the model. (Note that Dembski and Marks go off the rails by asserting that evolution is a search for a prespecified target, and saying that the environment would not have the information to guide the search to the prespecified target unless there were an informed choice of environment. This, loosely speaking, is conservation of active information.)I’m pretty sure that I’ve seen a mathematician attempt to relate adaptation to increase in mutual information. He’s a very bright guy, but I cannot recall his name at the moment. I’ll leave it to you to do some googling.

BruceS,One other thing. You have to understand that Eric believes that his ideas come from Jesus. You think that I’m being sarcastic? Here’s the first sentence of the acknowledgments section of his master’s thesis:

He evidently feels no compunction about pronouncing on matters without reviewing the relevant literature. For instance, you’ll get gobs of hits if you use Google Scholar to search for papers on meaningful information or semantic information — even with restriction of search terms. For instance, I get 94 results in a search for the precise phrase “measure of semantic information”. The engineers in the ID crowd, including Eric, are ignoring a large volume of prior work. (Ironically, the few scientists in the ID movement engage mostly in critical reviews of the literature.)

Have a look at the Google Scholar results for “mutual information” evolution adaptation environment. Read some papers. Please do not make Eric’s “inspired” speculations your starting point.