(Preamble: I apologize in advance for cluttering TSZ with these three posts. There are very few people on either side of the debate that actually care about the details of this “conservation of information” stuff, but these posts make good on some claims I made at UD.)

For the past three years Dembski has been promoting his Law of Conservation of Information (LCI), most recently here. The paper he most often promotes is this one, which begins as follows:

Laws of nature are universal in scope, hold with unfailing regularity, and receive support from a wide array of facts and observations. The Law of Conservation of Information (LCI) is such a law.

Dembski hasn’t proven that the LCI is universal, and in fact he claims that it can’t be proven, but he also claims that to date it has always been confirmed. He doesn’t say whether he as actually tried to find counterexamples, but the reality is that they are trivial to come up with. This post demonstrates one very simple counterexample.

## Definitions

First we need to clarify Dembski’s terminology. In his LCI math, a *search* is described by a probability distribution over a sample space Ω. In other words, a search is nothing more than an Ω-valued random variable. Execution of the search consists of a single *query*, which is simply a realization of the random variable. The search is deemed successful if the realized outcome resides in target T ⊆ Ω. (We must be careful to not read teleology into the terms *search*, *query*, and *target*, despite the terms’ connotations. Obviously, Dembski’s framework must not presuppose teleology if it is to be used to detect design.)

If a search’s parameters depend on the outcome of a preceding search, then the preceding search is a *search for a search*. It’s this hierarchy of two searches that is the subject of the LCI, which we can state as follows.

Given a search S, we define:

*q*as the probability of S succeeding*p*as the probability that S would succeed_{2}*if it were a uniform distribution**p*as the probability that a uniformly distributed search-for-a-search would yield a search at least as good as S_{1}

The LCI says that **p _{1} ≤ p_{2}/q**.

## Counterexample

In thinking of a counterexample to the LCI, we should remember that this two-level search hierarchy is nothing more than a chain of two random variables. (Dembski’s search hierarchy is like a Markov chain, except that each transition is from one state space to another, rather than within the same state space.) One of the simplest examples of a chain of random variables is a *one-dimensional random walk*. Think of a system that periodically changes state, with each state transition represented by a shift to the left or to the right on an state diagram. If we know at a certain point in time that it is in one of, say, three states, namely n-1 or n or n+1, then after the next transition it will be in n-2, n-1, n, n+1, or n+2, as in the following diagram:

Assume that the system is always equally likely to shift left as to shift right, and let the “target” be defined as the center node n. If the state at time t is, say, n-1, then the probability of success q is 1/2. Of the three original states, two (namely n-1 and n+1) yield this probability of success, so p_{1} is 2/3. Finally, p_{2} is 1/5 since the target consists of only one of the final five states. The LCI says that p_{1} ≤ p_{2}/q. Plugging in our numbers for this example, we get 2/3 ≤ (1/5)/(1/2), which is clearly false.

Of course, the LCI *does* hold under certain conditions. To show that the LCI to biological evolution, Dembski needs to show that his mathematical model of evolution meets those conditions. This model would necessarily include the higher-level search that gave rise to the evolutionary process. As will be shown in the next post, the good news for Dembski is that *any* process can be modeled such that it obeys the LCI. The bad news is that any process can also be modeled such that it violates the LCI.

DiEB claims that Dembski made a rookie mistake in arithmetic, and Joe has challenged that claim.

I’m not a very sophisticated programmer, but it looked like something I cold swing at. So heres my set of six virtual machines loaded with Ys and Ns to what I think are Dembski’s specifications. The results of a few tens of thousands of searches are included.

Option Explicit

Option Base 1

Private m(6, 6, 10) As String

Private Sub LoadIt()

Dim i As Integer, j As Integer, k As Integer

‘ load 5 bad machines

For i = 1 To 5

For j = 1 To 6

For k = 1 To 10

Select Case k

Case 1 To 9

m(i, j, k) = “n”

Case Else

m(i, j, k) = “y”

End Select

Next k

Next j

Next i

‘ load 1 good machine

For j = 1 To 6

For k = 1 To 10

Select Case k

Case 1 To 5

m(6, j, k) = “n”

Case Else

m(6, j, k) = “y”

End Select

Next k

Next j

End Sub

Private Sub searchIt()

Dim f As Integer, i As Integer, j As Integer, k As Integer, result As String

Dim Ys As Integer, Ns As Integer, outp As Integer

LoadIt

Ys = 0

Ns = 0

For f = 1 To 10000

Randomize

i = Int(6 * Rnd + 1)

j = Int(6 * Rnd + 1)

k = Int(10 * Rnd + 1)

result = m(i, j, k)

Select Case result

Case “n”

Ns = Ns + 1

Case “y”

Ys = Ys + 1

End Select

Next f

outp = 0

outp = (Ys / 10000) * 100

Debug.Print “Ys = ” & Ys & ” (” & outp & “%) — Ns = ” & Ns

End Sub

Results:

Ys = 1720 (17%) — Ns = 8280

Ys = 1876 (19%) — Ns = 8124

Ys = 1563 (16%) — Ns = 8437

Ys = 1685 (17%) — Ns = 8315

Ys = 1800 (18%) — Ns = 8200

Ys = 1691 (17%) — Ns = 8309

Ys = 1742 (17%) — Ns = 8258

Ys = 1758 (18%) — Ns = 8242

Ys = 1405 (14%) — Ns = 8595

I seem to have stumbled on the expected answer, but I defer to others whether this is Dembski’s array of six machines.

The BASIC, it burns!

I see that R0b has posted a simpler and more accurate model. I could change my arrays to match his, but the results would be the same. None of this is relevant to evolution because the question of probabilities is empirical.

Here’s a version that conforms to R0b’s post. I think it is exactly what Dembski described.

Option Explicit

Option Base 1

Private m(6, 10) As Integer

Private Sub LoadIt()

Dim j As Integer, k As Integer

‘ load 6 machines

For j = 1 To 6

For k = 1 To 10

Select Case k

Case 1 To 5

m(j, k) = j

Case Else

m(j, k) = IIf(k – 5 >= j, k – 4, k – 5)

End Select

Debug.Print m(j, k);

Next k

Debug.Print “”

Next j

End Sub

Private Sub searchIt()

Dim f As Integer, i As Integer, j As Integer, Result() As Integer

Dim k As Integer, OutP As Integer

LoadIt

For i = 1 To 10

ReDim Result(6) As Integer

For f = 1 To 10000

Randomize

j = Int(6 * Rnd + 1)

k = Int(10 * Rnd + 1)

Result(m(j, k)) = Result(m(j, k)) + 1

Next f

OutP = 0

OutP = (Result(6) / 10000) * 100

Debug.Print “Sixes = ” & Result(6) & ” (” & OutP & “%)”

Next i

End Sub

Results:

1 1 1 1 1 2 3 4 5 6

2 2 2 2 2 1 3 4 5 6

3 3 3 3 3 1 2 4 5 6

4 4 4 4 4 1 2 3 5 6

5 5 5 5 5 1 2 3 4 6

6 6 6 6 6 1 2 3 4 5

Sixes = 1459 (15%)

Sixes = 1827 (18%)

Sixes = 1685 (17%)

Sixes = 1678 (17%)

Sixes = 1719 (17%)

Sixes = 1562 (16%)

Sixes = 1753 (18%)

Sixes = 1485 (15%)

Sixes = 1530 (15%)

Sixes = 1686 (17%)

1 1 1 1 1 2 3 4 5 6

2 2 2 2 2 1 3 4 5 6

3 3 3 3 3 1 2 4 5 6

4 4 4 4 4 1 2 3 5 6

5 5 5 5 5 1 2 3 4 6

6 6 6 6 6 1 2 3 4 5

Sixes = 1654 (17%)

Sixes = 1679 (17%)

Sixes = 1540 (15%)

Sixes = 1822 (18%)

Sixes = 1592 (16%)

Sixes = 1827 (18%)

Sixes = 2029 (20%)

Sixes = 1952 (20%)

Sixes = 1768 (18%)

Sixes = 1574 (16%)

1 1 1 1 1 2 3 4 5 6

2 2 2 2 2 1 3 4 5 6

3 3 3 3 3 1 2 4 5 6

4 4 4 4 4 1 2 3 5 6

5 5 5 5 5 1 2 3 4 6

6 6 6 6 6 1 2 3 4 5

Sixes = 1842 (18%)

Sixes = 1407 (14%)

Sixes = 1736 (17%)

Sixes = 1510 (15%)

Sixes = 1725 (17%)

Sixes = 1845 (18%)

Sixes = 1526 (15%)

Sixes = 1982 (20%)

Sixes = 1723 (17%)

Sixes = 1680 (17%)

Dieb is right (it hardly needs a computer programme to prove it). But I don’t think it is big problem for Dembski. I think his point would be that having got a six via the machine which increases your chances of getting a six you might think you had improved your overall odds. In fact the chances of getting six

via that routeare lower than by simple throwing the dice.Actually this is just another way of looking at R0b’s balls (if you will pardon the expression) otherwise known as Betrand’s Box which I think is a brilliant way of thinking clearly aboout this.

Suppose there was a large number of six different colours of balls all in the same proportion. You want a gold one. You have two options. You can just choose a ball at random from the master bag. Or you can choose a box which has 5 of one colour and one of each other colour – but you don’t know which colour – and then choose a ball. It is rather obvious that your chances of getting a gold ball remain the same.

That is essentially the NFL theory. The LCI adds a corollary that the probability of selecting a box which increases your chances of getting a gold ball is always so low that your overall chances of getting a gold ball by that route are not increased. It is a slightly different statement and to my mind not very interesting except outside the bizarre interpretation he has put on it.

The point being that the LCI does not apply at all to the evolution of life or indeed most of the real world. In evolution of life a box already exists which greatly increases the chances of gold balls (it is called the environment it works by eliminating the other colours). The proportion of gold balls before entering this box is absolutely unrelated to the probability of this box existing.

I just thought of an even simper way of expressing this using Dembski’s own example. His example assumes that we know we want a six in advance. The way evolution works we don’t know what number we want. There is a machine out there (and in his model one such machine must exist) which makes one number more likely. We discover which number it is because that’s the one the machine creates. The machine (the environment) defines the target (whatever is needed to survive) – not the other way round.

The error itself isn’t devastating, but it punctures the aura of infallibility that Dembski has in the ID community, not the least, because it slipped through ID’s peer review. It might mean very little in absolute terms, but it exposes the inability of the ID community to think clearly about probability.

Dembski has devoted quite a bit of time and effort to crying “squirrel”! Look over there! It’s a shell game where the object is obfuscation and misdirection.

Perhaps the most revealing thing about this episode is not what it says about Dembski but what it says about his followers.