Eric Holloway has littered cyberspace with claims, arrived at by faulty reasoning, that the “laws of information conservation (nongrowth)” in data processing hold for algorithmic specified complexity as for algorithmic mutual information. It is essential to understand that there are infinitely many measures of algorithmic specified complexity. Nemati and Holloway would have us believe that each of them is a quantification of the meaningful information in binary strings, i.e., finite sequences of 0s and 1s. If Holloway’s claims were correct, then there would be a limit on the increase in algorithmic specified complexity resulting from execution of a computer program (itself a string). Whichever one of the measures were applied, the difference in algorithmic specified complexity of the string output by the process and the string input to the process would be at most the program length plus a constant. It would follow, more generally, that an approximate upper bound on the difference in algorithmic specified complexity of strings  and
 and  is the length of the shortest program that outputs
 is the length of the shortest program that outputs  on input of
 on input of  Of course, the measure must be the same for both strings. Otherwise it would be absurd to speak of (non-)conservation of a quantity of algorithmic specified complexity.
 Of course, the measure must be the same for both strings. Otherwise it would be absurd to speak of (non-)conservation of a quantity of algorithmic specified complexity.
For conventional computers, the shortest of programs immediately halts the computational process. The output is the empty string, no matter what string is input. However, the length of the program is not an approximate upper bound on the difference in algorithmic specified complexity of empty and nonempty strings. In “Evo-Info 4: Non-Conservation of Algorithmic Specified Complexity,” I showed that the empty string is, by some measures, infinitely greater in algorithmic specified complexity than all nonempty strings. My result not only refutes Holloway’s claims, but also destroys the prospect of establishing an alternative “conservation law” that holds for each and every measure of algorithmic specified complexity (ASC). Yet Nemati and Holloway frame Section 4 of their article as a response to my argument, saying that it “does not disprove conservation of ASC.” They neither state the result that I derived, nor venture even to cite the post containing it. Instead, they derive a new expression of “conservation of complexity,” obfuscating an application of different measures of algorithmic specified complexity to two strings. If they had made a direct comparison of their result to mine, then they necessarily would have exposed their change from one ASC measure to another. In a remarkable touch, at the end of the passage on display below, the authors attribute error to me in not pulling the switcheroo that they do.
The gist of the passage above is that the algorithmic specified complexity of a string “properly” depends on how the string is expressed. In my argument, the ASC measure is restricted in such a way that the algorithmic specified complexity of a string is infinite only when the string is empty. If we refer to the empty string as  (the lowercase Greek letter lambda), then Nemati and Holloway agree that the empty string is infinitely greater in algorithmic specified complexity than all nonempty strings. But if we instead refer to the empty string as
 (the lowercase Greek letter lambda), then Nemati and Holloway agree that the empty string is infinitely greater in algorithmic specified complexity than all nonempty strings. But if we instead refer to the empty string as  where
 where  is a function with
 is a function with  for all strings
 for all strings  then the authors change the ASC measure in such a way that the algorithmic specified complexity of the empty string is negative.  Their “conservation of complexity” inequality (43), in the passage above, is tacitly an upper bound on the difference in the ASC of
 then the authors change the ASC measure in such a way that the algorithmic specified complexity of the empty string is negative.  Their “conservation of complexity” inequality (43), in the passage above, is tacitly an upper bound on the difference in the ASC of  according to the modified measure, and the ASC of
 according to the modified measure, and the ASC of  according to the original measure. If you find it hard to believe that the authors would believe that the measure should depend on how a string is expressed, then you should read “Stark Incompetence,” the first part of my review of the article.
 according to the original measure. If you find it hard to believe that the authors would believe that the measure should depend on how a string is expressed, then you should read “Stark Incompetence,” the first part of my review of the article.
The next section of this post gives a simple explanation of algorithmic specified complexity, and then describes how the authors use the expression of a string to modify the ASC measure. The following two sections step through my annotations of the passage above, and delve into the misconception that the measure of algorithmic specified complexity should depend on the way in which a string is expressed. (There is some redundancy, owing to the fact that the sections originated in separate attempts at writing this post.) The final section, “Supplement for the Mathematically Adept,” not only provides details undergirding my remarks in the earlier sections, but also
- tightens the “conservation of complexity” bound (43),
- improves upon “the paper’s main result” (16), and
- corrects some errors in Section 4.2.
The authors should cite this post, and note that it improves upon some of their results, in the statement of errata that they submit to Bio-Complexity. They have no excuse, when publishing in a scholarly journal, for omitting citations of posts of mine that they draw upon.
A provisional response
Was mich nicht umbringt, macht mich stärker.
— Friedrich Nietzsche, Maxims and Arrows, 8 [translate]
The significance of provisional is that I am suppressing some details. In later sections, you will get graduated doses of the straight dope, and hopefully will acquire greater and greater tolerance of it. However, there is one technical point that I cannot bring myself to suppress. The computer I refer to is “ordinary” in the sense that the shortest of all programs halts with the empty string as output, no matter what string is input. It is possible to contrive a computer for which the shortest program that merely halts is much longer than some other programs that halt. But what matters here is that there do exist computers for which the “do nothing but to halt” program is no longer than any “do something and halt” program.
As I have emphasized, algorithmic specified complexity is an infinite class of measures. The particular measure is determined by parameters referred to as the chance hypothesis and the context. The chance hypothesis corresponds to one scheme for recoding of data (strings of bits), and the context corresponds to another. It might be that the recoding of data  in the scheme associated with the chance hypothesis is
 in the scheme associated with the chance hypothesis is  and that the recoding in the scheme associated with the context is
  and that the recoding in the scheme associated with the context is  The algorithmic specified complexity of the data is the difference in the numbers of bits in the two recodings of the data. In this case, the difference is
 The algorithmic specified complexity of the data is the difference in the numbers of bits in the two recodings of the data. In this case, the difference is  bits. When Nemati and Holloway switch from one measure of algorithmic specified complexity to another, it is actually just the chance hypothesis that they change. So we will be concerned primarily with the chance hypothesis.
 bits. When Nemati and Holloway switch from one measure of algorithmic specified complexity to another, it is actually just the chance hypothesis that they change. So we will be concerned primarily with the chance hypothesis.
The chance hypothesis, so called, is an assignment  of probabilities to data that might be encountered. That is,
 of probabilities to data that might be encountered. That is,  is the assumed probability of encountering data
 is the assumed probability of encountering data  In the recoding scheme associated with the chance hypothesis, the number of bits in the recoded data is ideally the assumed improbability of the data,
 In the recoding scheme associated with the chance hypothesis, the number of bits in the recoded data is ideally the assumed improbability of the data,  expressed on a logarithmic scale (base 2). In the example above, the 6-bit recoding
 expressed on a logarithmic scale (base 2). In the example above, the 6-bit recoding  of the data
 of the data  corresponds to improbability
 corresponds to improbability  the reciprocal of the assumed probability
 the reciprocal of the assumed probability  of encountering the data. The recoding scheme is ideal in the sense that the average length of the recoded data is minimized if the assumed probabilities are correct. Given that only the length of the recoded data enters into the calculation of algorithmic specified complexity, the recoding is not always performed. It suffices to use the ideal length, i.e., the log improbability of the data, even though it sometimes indicates a fractional number of bits. However, in an example to come, the data will be a digital image, and the recoding scheme associated with the chance hypothesis will be conversion of the image to PNG format.
 of encountering the data. The recoding scheme is ideal in the sense that the average length of the recoded data is minimized if the assumed probabilities are correct. Given that only the length of the recoded data enters into the calculation of algorithmic specified complexity, the recoding is not always performed. It suffices to use the ideal length, i.e., the log improbability of the data, even though it sometimes indicates a fractional number of bits. However, in an example to come, the data will be a digital image, and the recoding scheme associated with the chance hypothesis will be conversion of the image to PNG format.
The context parameter of the measure of algorithmic specified complexity is a string of bits. In the recoding scheme associated with the context, the context is used in description of the data. To gain some insight into the scheme, consider following scenario: Alice and Bob possess identical copies of the context, and use the same kind of computer. In order to communicate data  Alice transmits to Bob a program that outputs
 Alice transmits to Bob a program that outputs  on input of the context. On receipt of the program, Bob runs it with the context as input, and obtains the data
 on input of the context. On receipt of the program, Bob runs it with the context as input, and obtains the data  as the output. Ideally, Alice transmits the shortest of all programs that output
 as the output. Ideally, Alice transmits the shortest of all programs that output  on input of the context. In the recoding scheme associated with the context, the recoded data is the shortest program that outputs the data on input of the context. Returning to the example above, the data
 on input of the context. In the recoding scheme associated with the context, the recoded data is the shortest program that outputs the data on input of the context. Returning to the example above, the data  is  recoded as the 3-bit program
 is  recoded as the 3-bit program  which is the shortest of all programs that output
 which is the shortest of all programs that output  on input of the context.
 on input of the context.
Suppose that the 7-bit string  is the shortest program that outputs the data on input of the empty (zero-bit) string. Then there is a clear operational sense in which the context provides 4 bits of information about the data: description of the data with a program requires 4 fewer bits when the context is available to the program than when it is unavailable. The reduction in shortest-program length when the program has the context, rather than the empty string, as input is a measure of the algorithmic information that the context provides about the data. Writing
 is the shortest program that outputs the data on input of the empty (zero-bit) string. Then there is a clear operational sense in which the context provides 4 bits of information about the data: description of the data with a program requires 4 fewer bits when the context is available to the program than when it is unavailable. The reduction in shortest-program length when the program has the context, rather than the empty string, as input is a measure of the algorithmic information that the context provides about the data. Writing  for the context,
 for the context,  for the empty string, and
 for the empty string, and  for the length of the shortest program that outputs string
 for the length of the shortest program that outputs string  on input of string
 on input of string  the quantity of algorithmic information that the context  provides about the data
 the quantity of algorithmic information that the context  provides about the data  is
 is
 (A)    ![Rendered by QuickLaTeX.com \[I(x; C) = K(x|\lambda) - K(x|C) \text{ bits}. \]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-e2312c9c40b18cd27a9e8e96562555d1_l3.png)
Returning to Alice and Bob,  is the minimum number of bits in the program that Alice transmits to Bob as a description of how to produce the data, given that Bob executes the received program with the context as input. If there were no shared context, and Bob executed the received program with the empty string as input, then Alice would have to transmit at least
 is the minimum number of bits in the program that Alice transmits to Bob as a description of how to produce the data, given that Bob executes the received program with the context as input. If there were no shared context, and Bob executed the received program with the empty string as input, then Alice would have to transmit at least  program bits. Alice may convey the data to Bob with
 program bits. Alice may convey the data to Bob with  fewer program bits when Bob inputs the context, rather than the empty string, to the received program. Thus the context effectively conveys
 fewer program bits when Bob inputs the context, rather than the empty string, to the received program. Thus the context effectively conveys  bits of information about the data.
 bits of information about the data.
Now we replace  in Equation (A) with
 in Equation (A) with  the log improbability of
 the log improbability of  under the chance hypothesis
 under the chance hypothesis  The algorithmic specified complexity of the data
 The algorithmic specified complexity of the data  is
 is
 (41*)    ![Rendered by QuickLaTeX.com \[ASC(x, C, p) = I_p(x) - K(x|C) \text{ bits} \]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-573bf964115cc3f35352cd1c57691bc7_l3.png)
when the context is  and the chance hypothesis is
 and the chance hypothesis is  The greater the assumed improbability of the data, and the lesser the length of the shortest program that outputs the data on input of the context, the greater is the algorithmic specified complexity of the data. You might guess that for all contexts and chance hypotheses, algorithmic specified complexity is an alternative measure of the quantity of information that the context provides about the data. But you would be wrong. No matter how little algorithmic information the context provides about the data, the algorithmic specified complexity of the data can be made arbitrarily large by assuming that the data is highly improbable. There is a lower limit on
 The greater the assumed improbability of the data, and the lesser the length of the shortest program that outputs the data on input of the context, the greater is the algorithmic specified complexity of the data. You might guess that for all contexts and chance hypotheses, algorithmic specified complexity is an alternative measure of the quantity of information that the context provides about the data. But you would be wrong. No matter how little algorithmic information the context provides about the data, the algorithmic specified complexity of the data can be made arbitrarily large by assuming that the data is highly improbable. There is a lower limit on  because the shortest program that outputs the data on input of the context is not much  greater in length than the data itself. But there is no upper limit on
 because the shortest program that outputs the data on input of the context is not much  greater in length than the data itself. But there is no upper limit on  because the assumed improbability of the data is not constrained in any way by the data. The algorithmic specified  complexity of the data tends to infinity as the assumed probability of the data approaches zero, no matter how uninformative the context.
 because the assumed improbability of the data is not constrained in any way by the data. The algorithmic specified  complexity of the data tends to infinity as the assumed probability of the data approaches zero, no matter how uninformative the context.
It is easy to see that the algorithmic specified complexity of the empty string depends only on the chance hypothesis, given that the shortest program that halts with the empty string as output does so for all inputs. That is, the length  of the shortest program that outputs the empty string on input of the context does not depend on the context. It follows that the algorithmic specified complexity of the empty string does not depend on the context. That is, the algorithmic specified complexity of the empty string is the logarithm of its assumed improbability, minus a constant. The notion that this quantity is the amount of meaningful information in the empty string is ludicrous.
 of the shortest program that outputs the empty string on input of the context does not depend on the context. It follows that the algorithmic specified complexity of the empty string does not depend on the context. That is, the algorithmic specified complexity of the empty string is the logarithm of its assumed improbability, minus a constant. The notion that this quantity is the amount of meaningful information in the empty string is ludicrous.
In the argument of mine that Nemati and Holloway address, the assumed probability  of encountering data
 of encountering data  is zero if and only if
 is zero if and only if  is the empty string. As mentioned above, there is no upper limit on algorithmic specified complexity as the assumed probability of the data approaches zero. So I write
 is the empty string. As mentioned above, there is no upper limit on algorithmic specified complexity as the assumed probability of the data approaches zero. So I write  when
 when  irrespective of the context
 irrespective of the context  It follows that the difference in algorithmic specified complexity of the empty string and any nonempty string is infinite. (For readers who are bothered by my references to infinity, I note that the difference can be made arbitrarily large for an arbitrarily large number of nonempty strings by assuming that the empty string has very low probability instead of zero probability.) In my argument, I also define a function
 It follows that the difference in algorithmic specified complexity of the empty string and any nonempty string is infinite. (For readers who are bothered by my references to infinity, I note that the difference can be made arbitrarily large for an arbitrarily large number of nonempty strings by assuming that the empty string has very low probability instead of zero probability.) In my argument, I also define a function  with
 with  for all strings
 for all strings 
It ought to go without saying that everything I have just said about the empty string holds also for  inasmuch as
 inasmuch as  is, for all strings
 is, for all strings  identical to the empty string. However, Nemati and Holloway seem to mistake the expression
 identical to the empty string. However, Nemati and Holloway seem to mistake the expression  as indicating that something is done to the string
 as indicating that something is done to the string  in order to transform it into the empty string. I will have much more to say about this in a later section. Here I address the way in which the authors modify the chance hypothesis when the empty string is referred to as
 in order to transform it into the empty string. I will have much more to say about this in a later section. Here I address the way in which the authors modify the chance hypothesis when the empty string is referred to as  instead of
 instead of  The modified chance hypothesis indicates that the encountered data is the result of a deterministic transformation
 The modified chance hypothesis indicates that the encountered data is the result of a deterministic transformation  of a random, unobserved string. The probability that string
 of a random, unobserved string. The probability that string  enters into the process is
 enters into the process is  where
 where  is the original chance hypothesis. The probability of encountering data
 is the original chance hypothesis. The probability of encountering data  is the probability that some string
 is the probability that some string  for which
 for which  enters into the transformation process. That is, the probability
 enters into the transformation process. That is, the probability  assigned to data
 assigned to data  by the modified chance hypothesis
 by the modified chance hypothesis  is the sum of probabilities
 is the sum of probabilities  for all strings
 for all strings  such that
 such that  is the outcome of the transformation.
 is the outcome of the transformation.
Given that  for all strings
 for all strings  in my definition of function
 in my definition of function  the only data ever encountered, according to the modified chance hypothesis, is the empty string. That is, when the empty string is referred to as
 the only data ever encountered, according to the modified chance hypothesis, is the empty string. That is, when the empty string is referred to as  instead of
 instead of  Nemati and Holloway change its assumed probability from
 Nemati and Holloway change its assumed probability from  to
 to  in the measure of algorithmic specified complexity. Then the log improbability
 in the measure of algorithmic specified complexity. Then the log improbability  of the empty string is zero, and the algorithmic specified complexity
 of the empty string is zero, and the algorithmic specified complexity  of the empty string is just
 of the empty string is just  Where I write
 Where I write
      ![Rendered by QuickLaTeX.com \[ASC(f(x), C, p) - ASC(x, C, p) = \infty\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-dd4a4ecc27d3fe3ae3b77b09974584c5_l3.png)
for all nonempty strings  Nemati and Holloway write
 Nemati and Holloway write
 (43)    ![Rendered by QuickLaTeX.com \[f\!ASC(x, C, p, f) < I(x) \]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-be1f18af47fa35f31adefc3e8141fcad_l3.png)
for all strings  Inequality (43) is what the authors refer to as “conservation of complexity.” When deobfuscated, in the next section of this post, the inequality turns out to be  precisely equivalent to
 Inequality (43) is what the authors refer to as “conservation of complexity.” When deobfuscated, in the next section of this post, the inequality turns out to be  precisely equivalent to
      ![Rendered by QuickLaTeX.com \[ASC(f(x), C, q) - ASC(x, C, p) < K(x|C).\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-5fea80a028a19d776dab3bfb61acaf98_l3.png)
What the authors have managed to bound is the difference in the ASC of  according to a measure with chance hypothesis
 according to a measure with chance hypothesis  and the ASC of
 and the ASC of  according to a measure with chance hypothesis
 according to a measure with chance hypothesis 
Deobfuscating “conservation of complexity”
The first principle is that you must not fool yourself—and you are the easiest person to fool. So you have to be very careful about that. After you’ve not fooled yourself, it’s easy not to fool other scientists. You just have to be honest in a conventional way after that.
— Richard P. Feynman, “Cargo Cult Science“
In “Stark Incompetence,” I provided general readers with a detailed parse of the mathematics in a passage like the one on display above. It is impossible to do the same here. Instead, I will break the passage into parts, and provide math-lite commentary. In broad overview, the authors have hidden, with egregious abuse of notation, a switch from one chance hypothesis to another. Most of my annotation of the passage is substitution of appropriate expressions for misleading ones.
Section 4.2 begins with a remark that what comes at the end of the preceding subsection, addressed below, “gives us the basis to see why [infinite algorithmic specified complexity of  for all strings
 for all strings  ] does not disprove conservation of ASC.” However, my proof of non-conservation identifies cases in which the ASC of the
] does not disprove conservation of ASC.” However, my proof of non-conservation identifies cases in which the ASC of the  is finite and the ASC of
 is finite and the ASC of  measured the same way as for
 measured the same way as for  is infinite. The authors omit the crucial point that there is no upper bound on the difference in algorithmic specified complexity of
 is infinite. The authors omit the crucial point that there is no upper bound on the difference in algorithmic specified complexity of  and
 and  that holds for all measures. They perhaps do this because their own expression of “conservation of ASC”  does not make the difference explicit. We will find that there are hidden within the expression different measures of algorithmic specified complexity for the input data and the output data
 that holds for all measures. They perhaps do this because their own expression of “conservation of ASC”  does not make the difference explicit. We will find that there are hidden within the expression different measures of algorithmic specified complexity for the input data and the output data  and
 and 
To reiterate an important point, algorithmic specified complexity is an infinite class of measures. The particular measure is determined by the settings of parameters referred to as the context and the chance hypothesis. The algorithmic specified complexity of data  when the context is
 when the context is  and the chance hypothesis is
 and the chance hypothesis is  denoted
 denoted  is defined as the difference of two quantities, the first of which depends on the chance hypothesis, and the second of which depends on the context. It is easy to see, in the passage above, that the symbol
 is defined as the difference of two quantities, the first of which depends on the chance hypothesis, and the second of which depends on the context. It is easy to see, in the passage above, that the symbol  does not appear in the ostensible definition of ASC. The authors should have written
 does not appear in the ostensible definition of ASC. The authors should have written  instead of
 instead of  making the dependence on the chance hypothesis explicit as they do elsewhere in the article. This is not nitpicking. In mathematical analysis, one should begin by making all details explicit, and later suppress details when there is no danger of confusion. Such discipline is a major aspect of how one avoids fooling the easiest person to fool.
 making the dependence on the chance hypothesis explicit as they do elsewhere in the article. This is not nitpicking. In mathematical analysis, one should begin by making all details explicit, and later suppress details when there is no danger of confusion. Such discipline is a major aspect of how one avoids fooling the easiest person to fool.
As explained in the preceding section of this post, and as shown in the box above,  is the log improbability of the data
 is the log improbability of the data  Although other investigators of algorithmic specified complexity refer to this term as the complexity of
 Although other investigators of algorithmic specified complexity refer to this term as the complexity of  Nemati and Holloway refer to it as the surprisal. As explained previously,
 Nemati and Holloway refer to it as the surprisal. As explained previously,  is the length of the shortest program that outputs the data on input of the context. Other investigators of algorithmic specified complexity have referred to
 is the length of the shortest program that outputs the data on input of the context. Other investigators of algorithmic specified complexity have referred to  as a measure of the specification of the data.
 as a measure of the specification of the data.
Now we examine the end of Section 4.1, which is referred to in the opening of Section 4.2. Recall that we corrected the definition of algorithmic specified complexity by making explicit the dependence of  on the chance hypothesis
 on the chance hypothesis  Nemati and Holloway themselves use this notation elsewhere in the article, so there is no excuse for them not to have used it here. What they do instead is to engage in the most egregious abuse of notation that I ever have seen in a published paper. The authors make up a special rule for evaluating the expression
 Nemati and Holloway themselves use this notation elsewhere in the article, so there is no excuse for them not to have used it here. What they do instead is to engage in the most egregious abuse of notation that I ever have seen in a published paper. The authors make up a special rule for evaluating the expression  the effect of which is to transform the chance hypothesis with the function
 the effect of which is to transform the chance hypothesis with the function  We write the frightening expression
 We write the frightening expression  for the transformed chance hypothesis, and replace instances of the inexplicit
 for the transformed chance hypothesis, and replace instances of the inexplicit  with the explicit
 with the explicit 
Now we pick up where we left off in Section 4.2. After repeating the definition of ASC, the authors define something new, called fASC. The very fact that they do this should send up a red flag. Why introduce fASC when addressing conservation of ASC?  All readers should take a look at the annotation above. The authors have no good reason to repeat a definition that they gave earlier in the paper. What they do, by juxtaposing definitions in which the chance hypotheses are inexplicit, is to foster the false impression that the latter is obtained merely by replacing instances of  in the former with the expression
 in the former with the expression 
Switching from inexplicit to explicit notation in Equation (42), we discover that fASC serves only to conceal a change in chance hypothesis. That is, fASC is just an obfuscated expression of the algorithmic specified complexity of  according to the ASC measure with context
 according to the ASC measure with context  and chance hypothesis
 and chance hypothesis 
Eliminating fASC from what Nemati and Holloway call “conservation of complexity for ASC,” the result is a statement that the algorithmic specified complexity of  according to a measure with chance hypothesis
 according to a measure with chance hypothesis  is less than the log improbability
 is less than the log improbability  of
 of  under the chance hypothesis
 under the chance hypothesis  This is very strange, because we naturally expect to see a comparison of two quantities of ASC in an expression of “conservation.”
 This is very strange, because we naturally expect to see a comparison of two quantities of ASC in an expression of “conservation.”
As shown at the top of the box, it takes just a bit of algebra to derive an expression of  in terms of algorithmic specified complexity. At the bottom of the box,
 in terms of algorithmic specified complexity. At the bottom of the box,  is rewritten accordingly. With just a bit more algebra, we find that (43) is tacitly an upper bound on a difference of the ASC of
 is rewritten accordingly. With just a bit more algebra, we find that (43) is tacitly an upper bound on a difference of the ASC of  according to a measure with
 according to a measure with  as the chance hypothesis, and the ASC of
 as the chance hypothesis, and the ASC of  according to a measure with
 according to a measure with  as the chance hypothesis:
 as the chance hypothesis:
      ![Rendered by QuickLaTeX.com \[ASC(f(x), \, C, \, p \circ f^\leftarrow) - ASC(x, \, C,\,  p) < K(x|C).\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-cf31b2dd8c066d7804432ac4444e8ef7_l3.png)
(I derive a tighter bound in the last section of this post.) As I said in the introduction, it would be absurd to speak to speak of (non-)conservation of a quantity of algorithmic specified complexity if different measures were applied to two strings. What Nemati and Holloway have derived is a relation of two different ASC measures. Their characterization of (43) as “conservation of complexity” is indeed absurd.
Wherefore art thou f(Montague)?
What’s in a name? that which we call a rose
By any other name would smell as sweet…—William Shakespeare, Romeo and Juliet, 2.2
It is important to understand that a mathematical function  defined on the set of all strings is not a computer program. The expression
 defined on the set of all strings is not a computer program. The expression  indicates a particular relation of two strings. It does not indicate that something is done to one string in order to produce another. If we were to execute a program for calculation of
 indicates a particular relation of two strings. It does not indicate that something is done to one string in order to produce another. If we were to execute a program for calculation of  with
 with  as the input, then the output
 as the input, then the output  would be equal to
 would be equal to  That would be “data processing.” But it should be clear that
 That would be “data processing.” But it should be clear that  and
 and  are two ways of denoting a particular string, and that the properties of the string do not depend on how we express it. Most importantly, the algorithmic specified complexity of the string does not depend on whether we refer to it as
 are two ways of denoting a particular string, and that the properties of the string do not depend on how we express it. Most importantly, the algorithmic specified complexity of the string does not depend on whether we refer to it as  or as
 or as  Furthermore, there are infinitely many computable functions
 Furthermore, there are infinitely many computable functions  for which
 for which  is the string
 is the string  and in all cases the difference in ASC of
 and in all cases the difference in ASC of  and
 and  is the difference in ASC of
 is the difference in ASC of  and
 and 
Now, if you get what I am saying, then you probably are wondering why I would bother to say it. The reason is that Nemati and Holloway do not get it. In the passage above, the authors characterize the expression  in my proof of non-conservation of ASC as “applying a function” to the string
 in my proof of non-conservation of ASC as “applying a function” to the string  What they have done is to introduce an alternative expression of algorithmic specified complexity, fASC, that treats the function
 What they have done is to introduce an alternative expression of algorithmic specified complexity, fASC, that treats the function  as something that is done to
 as something that is done to  Simply put, they make the measure of algorithmic specified complexity of a string depend on how the string is expressed. In my proof, the string
 Simply put, they make the measure of algorithmic specified complexity of a string depend on how the string is expressed. In my proof, the string  is infinite in algorithmic specified complexity. According to the authors, the string has infinite ASC when expressed as
 is infinite in algorithmic specified complexity. According to the authors, the string has infinite ASC when expressed as  and “properly” has negative ASC, by a different measure, when expressed as
 and “properly” has negative ASC, by a different measure, when expressed as 
The authors make the strange claim that my “error lies in not properly transforming the random variable  with the function.” You need not know what a random variable is to understand my response. Two months before the authors submitted their paper for review, I told Holloway that there is no random variable in the definition of algorithmic specified complexity. As a matter of fact, the authors define ASC in Equation (10), and first write the symbol
 with the function.” You need not know what a random variable is to understand my response. Two months before the authors submitted their paper for review, I told Holloway that there is no random variable in the definition of algorithmic specified complexity. As a matter of fact, the authors define ASC in Equation (10), and first write the symbol  in Equation (16) (15), without defining it. More precisely, what they define is
 in Equation (16) (15), without defining it. More precisely, what they define is  the algorithmic specified complexity of string
 the algorithmic specified complexity of string  when the context is
 when the context is  and the chance hypothesis is
 and the chance hypothesis is  As explained above,
 As explained above,  is the assumed probability of encountering string
 is the assumed probability of encountering string  Far into the article, the authors claim that the probability assigned to string
 Far into the article, the authors claim that the probability assigned to string  by the chance hypothesis has been, all along, the probability that a random variable
 by the chance hypothesis has been, all along, the probability that a random variable  has
 has  as its value:
 as its value:
So when we write
what we mean is
(34)
Excuse me for laying a technical term on you, but people in my field refer to this as making shit up. You may think of  as a randomly arising string, and of the right-hand side of the equation as the probability that the string turns out to be
 as a randomly arising string, and of the right-hand side of the equation as the probability that the string turns out to be  It follows from Equation (34) that the probability assigned to the string
 It follows from Equation (34) that the probability assigned to the string  by the chance hypothesis is the probability that the randomly arising string turns out to be
 by the chance hypothesis is the probability that the randomly arising string turns out to be  :
:
      ![Rendered by QuickLaTeX.com \[p(\underbrace{f(x)}_{=y}) = p(y) = \Pr[\mathcal{X} = y].\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-ef63a9d4e2403b05858d65e0b0757778_l3.png)
The authors, who are on the losing end of a struggle with the rudiments of probability theory (I refer you again to “Stark Incompetence”), fail to see this. Evidently mistaking  as an expression of doing something to the string
 as an expression of doing something to the string  they insist on doing the same thing to the randomly arising string
 they insist on doing the same thing to the randomly arising string  That is, they believe that the probability assigned to
 That is, they believe that the probability assigned to  by the chance hypothesis ought to be the probability that
 by the chance hypothesis ought to be the probability that  turns out to be
 turns out to be 
      ![Rendered by QuickLaTeX.com \[\Pr[f(\mathcal{X}) = f(x)] \neq p(f(x)).\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-2f2c41912478aeb113c0ae78a7d2d73a_l3.png)
What the authors refer to as “properly transforming the random variable  with the function” is replacing
 with the function” is replacing  with
 with  when the string is referred to as
 when the string is referred to as  instead of
 instead of 
When the chance hypothesis is defined in terms of a random variable, a transformation of the random variable is necessarily a transformation of the chance hypothesis. Fortunately, and amusingly, with simplification of what ensues in the article, the random variable vanishes. The transformation of the chance hypothesis can be expressed, somewhat frighteningly, as  It is much easier to explain the physical significance of the transformed chance hypothesis than to parse the expression. In essence, an unobserved process deterministically transforms a randomly arising string into an observed string. The transformation is modeled by the function
 It is much easier to explain the physical significance of the transformed chance hypothesis than to parse the expression. In essence, an unobserved process deterministically transforms a randomly arising string into an observed string. The transformation is modeled by the function  String
 String  enters unobserved into the process with probability
 enters unobserved into the process with probability  String
 String  is observed with probability
 is observed with probability  which is equal to the sum of the probabilities
 which is equal to the sum of the probabilities  of all strings
 of all strings  that the process transforms into
 that the process transforms into 
By way of explaining the context parameter of the measure  and also of driving home the point that
 and also of driving home the point that  merely expresses a relation of strings, I turn to a pictorial example. The context
 merely expresses a relation of strings, I turn to a pictorial example. The context  can be any string, and here we take it to be an image-processing program that contains the Wikimedia Commons. Also, we adopt the chance hypothesis
 can be any string, and here we take it to be an image-processing program that contains the Wikimedia Commons. Also, we adopt the chance hypothesis  given by Ewert, Dembski, and Marks in “Measuring Meaningful Information in Images: Algorithmic Specified Complexity” (cited by Nemati and Holloway). That is, the log improbability of encountering an image is the number of bits in the image when it is converted to PNG format (and stripped of the header).
 given by Ewert, Dembski, and Marks in “Measuring Meaningful Information in Images: Algorithmic Specified Complexity” (cited by Nemati and Holloway). That is, the log improbability of encountering an image is the number of bits in the image when it is converted to PNG format (and stripped of the header).
 The Apotheosis of Meaning (at reduced resolution). The full-resolution image (3456 × 2592 pixels) contains 206 megabits of “meaningful information.”
The Apotheosis of Meaning (at reduced resolution). The full-resolution image (3456 × 2592 pixels) contains 206 megabits of “meaningful information.”The displayed image, which corresponds to a string  “looks random.” But there actually is nothing random about it. In particular, I did not sample the space of all images according to the chance hypothesis, i.e., with the probability
 “looks random.” But there actually is nothing random about it. In particular, I did not sample the space of all images according to the chance hypothesis, i.e., with the probability  of drawing image
 of drawing image  and then transform the image into
 and then transform the image into  I wrote a short program that, on input of the context, outputs the image
 I wrote a short program that, on input of the context, outputs the image  (The program does this virtually: I have not constructed the context literally.) Referring to the function implemented by the program as
 (The program does this virtually: I have not constructed the context literally.) Referring to the function implemented by the program as  we can write
 we can write 
As explained above, the algorithmic specified complexity of data  is the difference in length of two recodings of the data. Here the data is an image, and the recoding associated with the chance hypothesis is the image in PNG format. The recoding associated with the context is the shortest program that outputs the image on input of the image-processing program containing the Wikimedia Commons. The length of the program that I used to transform the context into image
 is the difference in length of two recodings of the data. Here the data is an image, and the recoding associated with the chance hypothesis is the image in PNG format. The recoding associated with the context is the shortest program that outputs the image on input of the image-processing program containing the Wikimedia Commons. The length of the program that I used to transform the context into image  is an upper bound on the length of the shortest program that does the trick. I do not know the precise length of my program, but I do know that it is less than 0.5 megabits. Thus, in my calculation that the algorithmic specified complexity
 is an upper bound on the length of the shortest program that does the trick. I do not know the precise length of my program, but I do know that it is less than 0.5 megabits. Thus, in my calculation that the algorithmic specified complexity  of image
 of image  is approximately 206 megabits, I set
 is approximately 206 megabits, I set  to 0.5 megabits. The string
 to 0.5 megabits. The string  representing the image titled The Apotheosis of Meaning, is an irreversible corruption of string
 representing the image titled The Apotheosis of Meaning, is an irreversible corruption of string  representing the following image from the Wikimedia Commons.
 representing the following image from the Wikimedia Commons.

E.T. Views Rushmore (at reduced resolution). The full-resolution image (3456 × 2592 pixels) contains 102 megabits of “meaningful information.” Photo credit: Volkan Yuksel, via Wikimedia Commons (CC BY-SA 3.0).
The selection of an image of Mount Rushmore contained by the context  is neither random nor arbitrary. My example is designed to embarrass Robert Marks (see why). The algorithmic specified complexity of
 is neither random nor arbitrary. My example is designed to embarrass Robert Marks (see why). The algorithmic specified complexity of  is approximately 102 megabits, and the difference in ASC of
 is approximately 102 megabits, and the difference in ASC of  and
 and  is approximately 104 megabits. Thus Marks, as well as his former students Nemati and Holloway, would have us believe that The Apotheosis of Meaning is twice as meaningful as E.T. Views Rushmore.
 is approximately 104 megabits. Thus Marks, as well as his former students Nemati and Holloway, would have us believe that The Apotheosis of Meaning is twice as meaningful as E.T. Views Rushmore.
Recall that  denotes the function computed by my program that transforms the context into the image
 denotes the function computed by my program that transforms the context into the image  The Apotheosis of Meaning. That is,
 The Apotheosis of Meaning. That is,  is equal to
 is equal to  . Nemati and Holloway presumably would not insist on changing the chance hypothesis to
. Nemati and Holloway presumably would not insist on changing the chance hypothesis to  when
 when  is referred to as
 is referred to as  However, my program is easily modified to output
 However, my program is easily modified to output  on input of E.T. Views Rushmore, i.e., to compute a function
 on input of E.T. Views Rushmore, i.e., to compute a function  for which
 for which  is equal to
 is equal to  and
 and  (The modified program cumulatively sums the RGB values in the input to produce the output, as explained in Evo-Info 4.) Now Nemati and Holloway object to my observation that the difference in algorithmic specified complexity of
 (The modified program cumulatively sums the RGB values in the input to produce the output, as explained in Evo-Info 4.) Now Nemati and Holloway object to my observation that the difference in algorithmic specified complexity of  and
 and 
      
is vastly greater than the length of a program for performing the “data processing”  They say that my “error lies in not properly transforming the random variable
 They say that my “error lies in not properly transforming the random variable  with the function.” In other (and better) words, I should have changed the chance hypothesis to
 with the function.” In other (and better) words, I should have changed the chance hypothesis to  in the measure of the algorithmic specified complexity of
 in the measure of the algorithmic specified complexity of  But you know that I did not obtain
 But you know that I did not obtain  by “applying a function”
 by “applying a function”  to a randomly selected string
 to a randomly selected string  as the transformed chance hypothesis would indicate. I have told you that I obtained
 as the transformed chance hypothesis would indicate. I have told you that I obtained  by selecting a program, and running it with the context as input. You might say that I tacitly selected a function
 by selecting a program, and running it with the context as input. You might say that I tacitly selected a function  and applied it to the context. You know that the expression
 and applied it to the context. You know that the expression  reflects how The Apotheosis of Meaning, represented by string
 reflects how The Apotheosis of Meaning, represented by string  came to appear on this page because I have told you so, not because the expression itself tells you so.  Similarly, the expression
 came to appear on this page because I have told you so, not because the expression itself tells you so.  Similarly, the expression  indicates only a mathematical relation of
 indicates only a mathematical relation of  to
 to  It does not say that
 It does not say that  is the outcome of a process transforming
 is the outcome of a process transforming  into
 into  let alone that
 let alone that  is a random occurrence.
 is a random occurrence.
Supplement for the mathematically adept
Nemati and Holloway use the symbol  to denote both the set of all binary strings and a random variable taking binary strings as values. In the following, I write
 to denote both the set of all binary strings and a random variable taking binary strings as values. In the following, I write  for the set
 for the set  and
 and  for the random variable. I refer to elements of
 for the random variable. I refer to elements of  as strings. The function
 as strings. The function  is defined on the set of all strings, i.e.,
 is defined on the set of all strings, i.e., 
 is the algorithmic specified complexity of string
 is the algorithmic specified complexity of string  when the context is
 when the context is  and the chance hypothesis
 and the chance hypothesis  The context is a string, and
 The context is a string, and  is the algorithmic complexity of
 is the algorithmic complexity of  conditioned on
 conditioned on  (loosely, the length of the shortest computer program, itself a string, that outputs
 (loosely, the length of the shortest computer program, itself a string, that outputs  on input of
 on input of  ). There is actually another parameter here, the universal prefix machine in terms of which conditional algorithmic complexity is defined. But leaving it inexplicit is unproblematic. The chance hypothesis is a distribution of probability over
). There is actually another parameter here, the universal prefix machine in terms of which conditional algorithmic complexity is defined. But leaving it inexplicit is unproblematic. The chance hypothesis is a distribution of probability over  The surprisal
 The surprisal  is the base-2 logarithm of
 is the base-2 logarithm of 
In the “definition” of fASC (42), the authors do not indicate that the probability distribution of the random variable  in Equation (39) is the parameter
 in Equation (39) is the parameter  But their claims are correct only if we assume as much. In (39), the chance hypothesis, left inexplicit in the notation
 But their claims are correct only if we assume as much. In (39), the chance hypothesis, left inexplicit in the notation  is the distribution of the random variable
 is the distribution of the random variable  In my expression
 In my expression  of the distribution of
 of the distribution of  I abuse notation slightly by writing
 I abuse notation slightly by writing  in place of the corresponding probability measure
 in place of the corresponding probability measure  on the space
 on the space  with
 with  
  for all
 for all  in
 in  I should explain also that
 I should explain also that  is a function from
 is a function from  to
 to  with
 with
      ![Rendered by QuickLaTeX.com \[f^\leftarrow(y) = \{x \in \mathcal{X} \mid f(x) = y\},\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-5cc6e84a41315627408b00412a84349b_l3.png)
the set of all preimages of  under the function
 under the function  So for all
 So for all  in
 in 
      ![Rendered by QuickLaTeX.com \begin{align*} p \circ f^\leftarrow(y) & = \mu_p(f^\leftarrow(y)) \\ & = \sum_{x \in f^\leftarrow(y)} \mu_p(\{x\}) \\ & = \sum_{x \in f^\leftarrow(y)} p(x) \\ & = \sum_{x : f(x) = y} p(x) \\ & = \sum_{x : f(x) = y} \Pr[X = y] \\ & = \Pr[f(X) = y]. \end{align*}](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-0c1299a83aaf5a2e34104a3f5fdac20e_l3.png)
I emphasize that there is no random variable in the definition of ASC. What the authors refer to as “properly transforming the random variable  with the function”  is actually replacing the chance hypothesis with another. The new chance hypothesis essentially says that observed data result from unobserved processing
 with the function”  is actually replacing the chance hypothesis with another. The new chance hypothesis essentially says that observed data result from unobserved processing  of randomly occurring data
 of randomly occurring data  That is, the
 That is, the  –valued random variable
–valued random variable  modeling the unobserved input to
 modeling the unobserved input to  follows distribution
 follows distribution  and
 and  modeling the observed output, follows distribution
 modeling the observed output, follows distribution  It is important to understand that if I select
 It is important to understand that if I select  and
 and  in order to make
 in order to make  large, and provide you with just the binary string
 large, and provide you with just the binary string  then your observation of large
 then your observation of large  gives you reason to reject
 gives you reason to reject  as a model of how
 as a model of how  arose. If I tell you that I generated
 arose. If I tell you that I generated  by applying function
 by applying function  to
 to  you do not have reason to replace the chance hypothesis with
 you do not have reason to replace the chance hypothesis with  inasmuch as I did not obtain
 inasmuch as I did not obtain  by sampling the distribution
 by sampling the distribution 
Tightening the “conservation of ASC” bound
It is easy, and furthermore useful, to derive a “conservation of ASC” bound tighter than the one that Nemati and Holloway provide in (43):
      ![Rendered by QuickLaTeX.com \begin{align*} & ASC(f(x), C, p \circ f^\leftarrow) - ASC(x, C, p) \\ & \quad = I_{p \circ f^\leftarrow}(f(x)) - K(f(x)|C) - [ I_{p}(x) - K(x|C)] \\ & \quad = I_{p \circ f^\leftarrow}(f(x)) - I_p(x) + K(x|C) - K(f(x)|C) \\ & \quad \leq K(x|C) - K(f(x)|C). \end{align*}](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-019d94a875e32d9a98b14cbaa9ac63eb_l3.png)
The inequality in the last step follows from (40). If  is the unique preimage of
 is the unique preimage of  under
 under  i.e.,
 i.e.,  then (40) holds with equality, and the last line in the preceding derivation holds with equality. [Added equation to justify:]
 then (40) holds with equality, and the last line in the preceding derivation holds with equality. [Added equation to justify:]
      
If, more restrictively,  is the unique preimage of
 is the unique preimage of  under
 under  then
 then
      
Thus the difference of  and
 and  can be made large by selecting a string
 can be made large by selecting a string  for which
 for which  is large, along with a function
 is large, along with a function  under which
 under which  is the unique preimage of the empty string.
 is the unique preimage of the empty string.
Improving upon the main result of the article
The following box contains what Nemati and Holloway describe as “the paper’s main result.” As indicated above, the symbol  is used to denote both a set and a random variable. Furthermore,
 is used to denote both a set and a random variable. Furthermore,  is not defined as a set or a random variable, either one. Again, I write
 is not defined as a set or a random variable, either one. Again, I write  for the random variable, and
 for the random variable, and  for the set. In (15), the set
 for the set. In (15), the set  must be a subset of
 must be a subset of  given the definition of
 given the definition of  In fuller context,
 In fuller context,  is necessarily equal to
 is necessarily equal to 
In the expression ![Rendered by QuickLaTeX.com E_\mathcal{X}[ASC(\mathcal{X}, C, p)],](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-d0d0b77ced482b611cb2eaff018b4509_l3.png) 
   must be a random variable. The only way to make the first equality in derivation (16) coherent is to assume that
 must be a random variable. The only way to make the first equality in derivation (16) coherent is to assume that  is a distribution of probability over
 is a distribution of probability over  and that random variable
 and that random variable  follows distribution
 follows distribution  The second line of the derivation would be the negative Kullback-Leibler divergence of
 The second line of the derivation would be the negative Kullback-Leibler divergence of  from
 from  if
 if  were a proper distribution. To improve upon the result, turning the inequality into an equality, we need only normalize (regularize) the semimeasure
 were a proper distribution. To improve upon the result, turning the inequality into an equality, we need only normalize (regularize) the semimeasure  to produce a probability distribution. Define probability distribution
 to produce a probability distribution. Define probability distribution
      ![Rendered by QuickLaTeX.com \[\mathbf{m}_{/C}^*(x) := \frac{ \mathbf{m}_{/C}(x) } { \mathbf{m}_{/C}(\mathcal{X}) }\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-ab71cb6e7ae7734dd79d89c8e3fba8f1_l3.png)
for all  in
 in  where
 where
      ![Rendered by QuickLaTeX.com \[\mathbf{m}_{/C}(\mathcal{X}) := \sum_{x \in \mathcal{X}} \mathbf{m}_{/C}(x).\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-f274c63823d76b6a10265605f0f1c772_l3.png)
Writing  for the random variable, and continuing from the second line of the derivation,
 for the random variable, and continuing from the second line of the derivation,
      ![Rendered by QuickLaTeX.com \begin{align*} & E_X[ASC(X, C, p)] \\ & \quad = \sum_{x \in \mathcal{X}} p(x) \log_2 \left( \frac{\mathbf{m}_{/C}(x)}{p(x)} \right) \\ & \quad = \sum_{x \in \mathcal{X}} p(x) \log_2 \left( \frac{\mathbf{m}^*_{/C}(x) \, \mathbf{m}_{/C}(\mathcal{X})}{p(x)} \right) \\ & \quad = \sum_{x \in \mathcal{X}} p(x) \log_2 \left( \frac{\mathbf{m}^*_{/C}(x)}{p(x)} \right) + \sum_{x \in \mathcal{X}} p(x) \log_2 \mathbf{m}_{/C}(\mathcal{X})\\ & \quad = \sum_{x \in \mathcal{X}} p(x) \log_2 \mathbf{m}_{/C}(\mathcal{X}) - \sum_{x \in \mathcal{X}} p(x) \log_2 \left( \frac{p(x)}{\mathbf{m}^*_{/C}(x)} \right)  \\ & \quad = \log_2 \mathbf{m}_{/C}(\mathcal{X}) \sum_{x \in \mathcal{X}} p(x) - D(p||\mathbf{m}^*_{/C})   \\ & \quad = \underbrace{\log_2 \mathbf{m}_{/C}(\mathcal{X})}_{< 0} - \underbrace{D(p||\mathbf{m}^*_{/C})}_{\geq 0}. \end{align*}](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-bdc059fb815494da7c7702b7f6d2a286_l3.png)
The Kullback-Leibler divergence of  from
 from  
  is zero if and only if the two probability distributions  are identical.
 is zero if and only if the two probability distributions  are identical.
Correcting some errors in Section 4.2
On display below is the passage that immediately follows the one on display above.
Again, we write  for the random variable that Nemati and Holloway denote
 for the random variable that Nemati and Holloway denote  Recall that the distribution of
 Recall that the distribution of  is
 is  When the inequality in (45) is revised to read
 When the inequality in (45) is revised to read
      ![Rendered by QuickLaTeX.com \[E_L[ASC(L, \, C, \, p \circ f^\leftarrow)] < 0,\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-e446ec66026bbe090cbed09ce89325a0_l3.png)
it does follow from inequality (16). That is, as explained above, (16) is coherent only if the random variable is distributed according to the chance hypothesis. Furthermore, the equality in (45) holds because, as explained above,
      
Similarly, when (46) is revised to read
      ![Rendered by QuickLaTeX.com \[\Pr[ASC(L, \,C, \,p \circ f^\leftarrow) \geq \alpha] \leq 2^{-\alpha},\]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-42ee345617007bd661918c41bcbef4f1_l3.png)
it follows from an emendation of Theorem 1 of Ewert, Marks, and Dembski, “On the Improbability of Algorithmic Specified Complexity”:
Theorem 1. The probability of obtaining an object exhibiting
bits of ASC is less then [sic] or equal to
(2)
Interestingly, Ewert et al. also neglect to identify  as a string-valued random variable following probability distribution
 as a string-valued random variable following probability distribution  In the immediately preceding paragraphs of their paper, they identify
 In the immediately preceding paragraphs of their paper, they identify  as “the object or event or under consideration,” and also write
 as “the object or event or under consideration,” and also write  which indicates that the “object or event” actually must be an element of
 which indicates that the “object or event” actually must be an element of  Thus it would seem that junk mathematics is being handed down from one Marks advisee to the next.
  Thus it would seem that junk mathematics is being handed down from one Marks advisee to the next.
The most important of my annotations in the box above is the note that the equality in (45) is a bald assertion. If the authors had attempted a derivation, then they probably would have failed, and perhaps would have discovered how to correct their errors. I cannot imagine how they would have derived the equality without laying bare the fact that they modify the chance hypothesis. Note that with only  written for the chance hypotheses in (45) and (46), there is nary a clue that that chance hypothesis is modified.
 written for the chance hypotheses in (45) and (46), there is nary a clue that that chance hypothesis is modified.
There are some similar errors in Sections 4.3 and 4.4. I will not spell them out. The authors should be able to find them, knowing now what they got wrong in Section 4.2, and address them in the errata that they submit to Bio-Complexity. I have spotted plenty of other errors. But life is short.
 
			








![Rendered by QuickLaTeX.com \begin{align*} p(x) & = \Pr[\mathcal{X} = x]  \\ & \phantom{::} \vdots \\ \end{align*}](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-643e7d6744fa06f0c604a94e580d990a_l3.png)




![Rendered by QuickLaTeX.com \[\Pr[ASC(X, C, P) \geq \alpha] \leq 2^{-\alpha} \]](http://theskepticalzone.com/wp/wp-content/ql-cache/quicklatex.com-8a896bf9f2b3f84b1f4b271f7374ab96_l3.png)
I have searched the OP carefully and found nowhere near any indications this applies to biology…
As I recollect correctly, EricHM has stated on more than one occasion that he doesn’t know whether his paper applies to real life… (biology/genetics).
If this is the case, why drag on this nonsense, if none of you is even willing to investigate the real life application?
Eric has a family of 5 to feed… Why would he even consider responding to this OP?
That’s pretty much my reaction to all of the ID arguments about information.
Don’t blame Tom, as he is just responding to those arguments.
Well, that’s just because of your ignorance…
Read my, and Eric’s, comments again!
Do you need links???
I agree with J-Mac, very long, unclear article. Tom, can you provide a tl;dr that points out what my error is so I can respond to that?
Or better yet, submit a manuscript to Bio-C. Then it’ll get filtered for coherence. Once it’s passed that gate, then it’ll be clear to me it is worth my time to analyze and respond.
There’s a concise response in the last section, “Supplement for the Mathematically Adept.” Admittedly, I have argued that you are not a member of the target audience. Are you conceding the point?
Thanks for taking the time to write this up, Tom. Nice to have the concept explained again.
From a comment of mine on “Questions for Eric Holloway”:
Eric Holloway has an ethical obligation to submit errata to Bio-Complexity when he is aware of errors in his article. Trying to shift the burden of submission to me is itself unethical.
You will find here a case in which some ID creationists, including the editor-in-chief of Bio-Complexity, attached an entire page of errata to a paper of theirs, in response to errors that I identified on my personal blog. They replaced five plots in the paper. (In fact, I diagnosed precisely the error they had made in processing the data that they plotted.) Interestingly, Eric Holloway recently referred to this paper. But rather than link to the freely available version containing the statement of errata, at the website of the Evolutionary Informatics Lab, he linked to a pay-to-view version without the errata. But surely that was merely an oversight on his part, for Eric is an honorable man.
I will point out, also, that Eric Holloway has benefitted from research that I have published on this blog. It was I, not any ID creationist, who discovered work by A. Milosavljecvic from 25 years ago that is quite similar to some published applications of algorithmic specified complexity. My discussion of Milosavljecvic comes toward the end of a long post, Evo-Info 4, that contains more math than does “The Old Switcheroo.” Eric Holloway happily benefits from that long post. But now, when a post identifies errors in an article of his, he claims that he does not have time to read it.
Ever suggest that to gpuccio?