ev

Recent discussions of genetic algorithms here and Dave Thomas’ evisceration of Winston Ewert’s review of several genetic algorithms at The Panda’s Thumb prompted me to dust off my notes and ev implementation.

Introduction

In the spring of 1984, Thomas Schneider submitted his Ph.D thesis demonstrating that the information content of DNA binding sites closely approximates the information required to identify the sites in the genome. In the week between submitting his thesis and defending it, he wrote a software simulation to confirm that the behavior he observed in biological organisms could arise from a subset of known evolutionary mechanisms. Specifically, starting from a completely random population, he used only point mutations and simple fitness-based selection to create the next generation.

The function of ev is to explain and model an observation about natural systems.
— Thomas D. Schneider

Even with this grossly simplified version of evolution, Schneider’s simulator, tersely named ev, demonstrated that the information content of a DNA binding site, R_{sequence}, consistently and relatively quickly evolves to be approximately equal to the information required to identify the binding sites in a given genome, R_{frequency}, just as is seen in the biological systems that were the subject of his thesis.

Schneider didn’t publish the details of ev until 2000, in response to creationist claims that evolution is incapable of generating information.

Core Concepts

Before discussing the implementation, it’s important to understand exactly what is being simulated. Dr. Schneider’s thesis is quite readable. The core concepts of ev are binding sites, R_{sequence}, and R_{frequency}.

Binding Sites

A binding site is a location on a strand of DNA or RNA where a protein can attach (bind). Binding sites consist of a sequence of nucleotides that together provide the necessary chemical bonds to hold the protein.

A good example of binding sites in action is the synthesis of messenger RNA (mRNA) by RNA polymerase (RNAP). RNAP binds to a set of a few tens of base pairs on a DNA strand which triggers a series of chemical reactions that result in mRNA. This mRNA is then picked up by a ribosome (which also attaches to a binding site) that transcribes a protein from it.

The bases that make up a binding site are best described by a probability distribution, they are not a fixed set requiring an exact match.

R_{frequency}

R_{frequency} is the simplest of the two information measures in ev. Basically, it is the number of bits required to find one binding site out of set of binding sites in a genome of a certain length. For a genome of length G with \gamma binding sites, this is -log_2(\gamma / G)

For example, consider a genome of 1000 base pairs containing 5 binding sites. The average distance between binding sites is 200 bases, so the information needed to find them is -log_{2}200 which is approximately 7.64 bits.

R_{sequence}

R_{sequence} is the amount of information in the binding site itself. There are two problems with computing R_{sequence}. The first is the definition of “information.” Schneider uses Shannon information, a clearly defined, well-respected metric with demonstrated utility in the study of biological systems.

The second problem is that binding sites for the same protein don’t consist of exactly the same sequence of bases. Aligned sequences are frequently used to identify the bases that are most common at each location in the binding site, but they don’t tell the whole story. An aligned sequence that shows an A in the first position may reflect a set of actual sites of which 70% have A in the first position, 25% C, and 5% G. R_{sequence} must take into account this distribution.

The Shannon uncertainty of a base in a binding site is:

(1)   \begin{equation*} H_g = \sum_{b}^{A,C,G,T}(p(b)log_{2}p(b)) \end{equation*}

where p(b) is the probability of a base b at that location in the genome. This will be approximately 0.25, equal probability for all bases, for the initial, randomly generated genome. The initial uncertainty at a binding site is therefore:

(2)   \begin{equation*} H_{before} = H_{g}L = 4(0.25)(log_{2}(0.25)L = -2L \end{equation*}

where L is the width of the site.

R_{sequence}, the increase in information, is then H_{after} - H_{before}, where:

(3)   \begin{equation*} H_{after} = \sum_{l = 1}^{L}(H_{g}(l)) \end{equation*}

computed from the observed probabilities.

There is one additional complexity with these formulas. Because of the small sample size, an adjustment must be computed for H_g:

(4)   \begin{equation*} H_g = \sum_{l = 1}^{L}(E(H_{nb}) - H_{g}(L)) \end{equation*}

or

(5)   \begin{equation*} H_{after} = \sum_{l = 1}^{L}\bigg((e(n(l)) - \sum_{b}^{A,C,G,T}f(b,l) log_{2}f(b,l)\bigg) \end{equation*}

measured in bits per site.

The math behind the small sample adjustment is explained in Appendix 1 of Schneider’s thesis. Approximate values for E(H_{nb}) for binding site widths from 1 to 50 are available pre-computed by a program available on Schneider’s site:

For a random sequence, R_{sequence} will be near 0. This will evolve to R_{frequency} over an ev run.

Schneider’s ev Implementation

Schneider’s implementation is a fairly standard genetic algorithm, with an interesting fitness function. The virtual genomes contain, by default, 256 potential binding sites. The genomes are composed of characters from an alphabet of four letters (A, C, G, and T). The default number of optimal binding sites, \gamma, is 16. The locations of these sites are randomly generated at the beginning of each run and remain the same throughout the run. Given this configuration, R_{frequency}, the amount of information required to identify one of these sites in a genome of length G = 256 is -log_2(\gamma / G) which equals 4. Per Schneider’s Ph.D thesis, R_{sequence}, the information in the binding site itself, should evolve to and remain at approximately this value during a run.

To determine the number of binding sites actually present, a portion of the genome codes for a recognizer as well as being part of the set of potential binding sites. This recognizer, which is subject to the same mutation and selection as the rest of the genome, is applied at each base to determine if that base is the start of a binding site. If a base is not correctly identified as the start of a binding site, the fitness of the genome is decreased by one. If a base is incorrectly identified as the start of a binding site, the fitness of the genome is also decreased by one. Schneider notes that changing this weighting may affect the rate at which R_{sequence} converges to R_{frequency} but not the final result.

After all genomes are evaluated, the half with the lowest fitness are eliminated and the remaining are duplicated with mutation. Schneider uses a relatively small population size of 64.

The recognizer is encoded as a weight matrix of 4xL two’s-complement integers, where L is the length of a binding site (6 by default). Schneider notes that:

At first it may seem that this is insufficient to simulate the complex processes of transcription, translation, protein folding and DNA sequence recognition found in cells. However the success of the simulation, as shown below, demonstrates that the form of the genetic apparatus does not affect the computed information measures. For information theorists and physicists this emergent mesoscopic property will come as no surprise because information theory is extremely general and does not depend on the physical mechanism. It applies equally well to telephone conversations, telegraph signals, music and molecular biology.

Since ev genomes consist only of A, C, G, and T, these need to be translated to integers for the weight matrix. Schneider uses the straightforward mapping of (A, C, G, T) to (00, 01, 10, 11). The default number of bases for each integer is Bp = 5. Given these settings, AAAAA encodes the value 0, AAAAC encodes 1, and TTTTT encodes -1 (by two’s-complement rules).

When evaluating a genome, the first 4 x L x Bp bases are read into the 4 x L weight matrix. The next Bp bases represent a threshold value that is used to determine whether or not the value returned by the recognizer is a binding site match. This is also a two’s-complement integer with the same mapping. The recognizer is then applied from the first base in the genome to the last that could possibly be the start of a binding site (given the binding site length).

It’s worth re-emphasizing that the recognizer and the threshold are part of the genome containing the binding sites. The length of the full genome is therefore G + L - 1 bases, with only the first G bases being potential binding sites.

The recognizer calculates a total value for the potential site starting at a given point by summing the values of the matching bases in the weight matrix. The weight matrix contains a value for each base at each position in the site, so for a binding site length of 7 and a potential binding site consisting of the bases GATTACA, the total value is:

w[0]['G'] + w[1]['A'] + w[2]['T'] + w[3]['T'] + w[4]['A'] + w[5]['C'] + w[6]['A']

If this value exceeds the threshold, the recognizer identifies the bases as a binding site.

This implementation of the recognizer is an interesting way of encapsulating the biological reality that binding sites don’t always consist of exactly the same sequence of bases. Schneider notes, though, that “the exact form of the recognition mechanism is immaterial because of the generality of information theory.”

Schneider’s Results

Using his default settings of:

  • Genome length: G = 256
  • Number of binding sites: \gamma = 16
  • Binding site length: L = 6
  • Bases per integer: Bp = 5

Schneider found that:

When the program starts, the genomes all contain random sequence, and the information content of the binding sites, R_{sequence}, is close to zero. Remarkably, the cyclic mutation and selection process leads to an organism that makes no mistakes in only 704 generations (Fig. 2a). Although the sites can contain a maximum of 2L = 12 bits, the information content of the binding sites rises during this time until it oscillates around the predicted information content, R_{frequency} = 4 bits . . . .

The Creationist Response

30 years after the original implantation and 16 years after it was published, Intelligent Design Creationists (IDCists) are still attempting to refute ev and are still getting it wrong.

Dembski In 2001

In 2001, William Dembski claimed that ev does not demonstrate an information increase and further claimed that Schneider “smuggled in” information via his rule for handling ties in fitness. Schneider reviewed and rebutted the first claim and tested Dembski’s second claim, conclusively demonstrating it to be false.

Schneider wryly addresses this in the ev FAQ:

Does the Special Rule smuggle information into the ev program?

This claim, by William Dembski, is answered in the on-line paper Effect of Ties on the evolution of Information by the ev program. Basically, changing the rule still gives an information gain, so Dembski’s prediction was wrong.

Has Dembski ever acknowledged this error?

Not to my knowledge.

Don’t scientists admit their errors?

Generally, yes, by publishing a retraction explaining what happened.

Montanez, Ewert, Dembski, and Marks In 2010

Montanez, Ewert, Dembski, and Marks published A Vivisection of the ev Computer Organism: Identifying Sources of Active Information in the IDCist’s pseudo-science journal BIO-Complexity in 2010. Despite its title, the paper doesn’t demonstrate any understanding of the ev algorithm or what it demonstrates:

  • The authors spend a significant portion of the paper discussing the efficiency of the ev algorithm. This is a red herring — nature is profligate and no biologist, including Schneider, claims that evolutionary mechanisms are the most efficient way of achieving the results observed.
  • Related to the efficiency misdirection, the authors suggest alternative algorithms that have no biological relevance instead of addressing the actual algorithm used by ev.
  • The authors do not use Shannon information, instead substituting their idiosyncratic “active information”, including dependencies on Dembski’s concept of “Conservation of Information” which has been debunked by Wesley Elsberry and Jeffrey Shallit in Information Theory, Evolutionary Computation, and Dembski’s “Complex Specified Information”, among others.
  • The authors note that “A common source of active information is a software oracle”. By recognizing that an oracle enables evolutionary mechanisms to work in software, they are admitting that those same mechanisms can explain what we observe in biological systems because the real world environment is just such an oracle. The environment provides information about what works and what doesn’t by ensuring that organisms less suited to it will tend to leave fewer descendants.
  • The authors repeatedly claim that the “perceptron” used as a recognizer makes the ev algorithm more efficient. Their attempted explanation of why this is the case completely ignores the actual implementation of ev. They seem so caught up in Schneider’s description of the recognizer as a perceptron that they miss the fact that it’s nothing more than a weight matrix that models the biological reality that a binding site is not a fixed set of bases.

Basically the paper is a rehash of concepts the authors have discussed in previous papers combined with the hope that some of it will be applicable to ev. None of it is.

Schneider soundly refuted the paper in Dissection of “A Vivisection of the ev Computer Organism: Identifying Sources of Active Information”. He succinctly summarized the utter failure of the authors to address the most important feature of ev:

They do not compute the information in the binding sites. So they didn’t evaluate the relevant information (R_{sequence}) at all.

In a response to that refutation, Marks concedes that “Regardless, while we may have different preferred techniques for measuring information, we agree that the ev genome does in fact gain information.”

After that damning admission, Marks still doubles down on his spurious claim that the “Hamming oracle” makes ev more efficient:

Schneider addresses the hamming oracle issue by assuming that nature provides a correct fitness function (a hamming function) that allows for positive selection in the direction of our target. He argues that this fitness is based on a

biologically sensible criteria: having functional DNA binding sites and not having extra ones.

But this describes a target; this is the desired goal of the simulation. The fitness function actually being used is a distance to this target. This distance makes efficient information extraction possible.

That’s not a target. It provides no details about what a solution would look like or how to reduce the distance measured, it simply indicates how far away a genome is from being a solution. In fact, it does less than that because it doesn’t provide any information about the difference between an existing recognizer and an ideal recognizer. It also says nothing at all about the relationship between R_{frequency} and R_{sequence}.

Even as he tries to salvage the tatters of his paper, Marks makes another concession:

Reaching that point requires a particular shape to the fitness landscape to guide evolution.

He admits again that evolution does work in certain environments. The real world is one of those.

Nothing in Marks’ response changes the accuracy of Schneider’s summary in his refutation:

Aside from their propensity to veer away from the actual biological situation, the main flaw in this paper is the apparent misunderstanding of what ev is doing, namely what information is being measured and that there are two different measures. The authors only worked with R_{frequency} and ignored R_{sequence}. They apparently didn’t compute information from the sequences. But it is the increase of R_{sequence} that is of primary importance to understand. Thanks to Chris Adami, we clearly understand that information gained in genomes reflects ‘information’ in the environment. I put environmental ‘information’ in quotes because it is not clear that information is meaningful when entirely outside the context of a living organism. An organism interprets its surroundings and that ‘information’ guides the evolution of its genome.

Ewert in 2014

Winston Ewert published Digital Irreducible Complexity: A Survey of Irreducible Complexity in Computer Simulations in 2014, again in the IDCists’ house journal BIO-Complexity. This paper constituted 25% of the output of that august publication in that year. Ewert reviewed Avida, ev, Dave Thomas’ Steiner Trees, a geometric algorithm by Suzanne Sadedin, and Adrian Thompson’s Digital Ears, attempting to demonstrate that none of them generated irreducible complexity.

Michael Behe defined “irreducible complexity” in his 1996 book Darwin’s Black Box:

By irreducibly complex I mean a single system composed of several well-matched, interacting parts that contribute to the basic function, wherein the removal of any one of the parts causes the system to effectively cease functioning. An irreducibly complex system cannot be produced directly (that is, by continuously improving the initial function, which continues to work by the same mechanism) by slight, successive modifications of a precursor system, because any precursor to an irreducibly complex system that is missing a part is by definition nonfunctional.

Dave Thomas has shredded Ewert’s discussion of Steiner Trees, demonstrating that Ewert addressed a much simpler problem (spanning trees) instead.

Richard B. Hoppe has similarly destroyed Ewert’s claims about Avida.

Schneider does explicitly claim that ev demonstrates the evolution of irreducible complexity:

The ev model can also be used to succinctly address two other creationist arguments. First, the recognizer gene and its binding sites co-evolve, so they become dependent on each other and destructive mutations in either immediately lead to elimination of the organism. This situation fits Behe’s definition of ‘irreducible complexity’ exactly . . . .

Ewert quotes this claim and attempts to refute it with:

It appears that Schneider has misunderstood the definition of irreducible complexity. Elimination of the organism would appear to refer to being killed by the model’s analogue to natural selection. Given destructive mutations, an organism will perform less well than its competitors and “die.” However, this is not what irreducible complexity is referring to by “effectively ceasing to function.” It is true that in death, an organism certainly ceases to function. However, Behe’s requirement is that:

If one removes a part of a clearly defined, irreducibly complex system, the system itself immediately and necessarily ceases to function.

The system must cease to function purely by virtue of the missing part, not by virtue of selection.

It appears that Ewert is the one with the misunderstanding here. If there is a destructive mutation in the genes that code for the recognizer, none of the binding sites will be recognized and, in the biological systems that ev models, the protein will not bind and the resulting capability will not be provided. It will “immediately and necessarily” cease to function. This makes the system irreducibly complex by Behe’s definition.

Binding sites are somewhat less brittle, simply because there are more of them. However, if there is a destructive mutation in one or more of the binding sites, the organism with that mutation will be less fit than others in the same population. In a real biological system, the function provided by the protein binding will be degraded at best and eliminated at worst. The organism will have effectively ceased to function. It is this lack of function that results in the genome being removed from the gene pool in the next generation.

Given that the recognizer and binding sites form a set of “well-matched, interacting parts that contribute to the basic function” and that “the removal of any one of the parts causes the system to effectively cease functioning”, ev meets Behe’s definition of an irreducibly complex system.

The IDCist Discovery Institute touted Ewert’s paper as evidence of “an Excellent Decade for Intelligent Design” in the ten years following the Dover trial. That case, of course, is the one that showed conclusively that ID is simply another variant of creationism and a transparent attempt to violate the separation of church and state in the United States. If Ewert’s paper is among the top achievements of the IDC movement in the past ten years then it’s clear that reality-based observers still have no reason to take any IDCist pretensions to scientific credibility seriously. The political threat posed by intelligent design and other variants of creationism is, unfortunately, still a significant problem.

An Alternative ev Implementation

I have implemented a variant of Dr. Schneider’s ev in order to reproduce his results and explore the impact of some alternative approaches. My version of ev uses the GA Engine I wrote to solve Dave Thomas’ Steiner Network design challenge. This engine operates on bit strings rather than the characters used by Dr. Schneider’s implementation.

As described in the GA engine documentation, applying the GA engine to a problem consists of following a few simple steps:

  1. Create a class to represent the characteristics of the problem

    The problem class ev-problem contains the parameters for genome length (G), number of binding sites (\gamma), binding site length (L), and bases per integer (Bp).

  2. Implement a method to create instances of the problem class

    The make-ev-problem function creates an instance of ev-problem.

  3. Implement the required generic functions for the problem:
    • genome-length

    The genome length is (G + L - 1) * 2, using two bits to encode each base pair.

    • fitness

    The fitness of a genome is the number of mistakes made by the recognizer, the total of missed and spurious binding sites.

    • fitness-comparator

    This problem uses the greater-comparator provided by the GA engine.

  4. Implement a terminator function

    This problem uses the generation-terminator provided by the GA engine, stopping after a specified number of generations.

  5. Run solve

Initial Results

In my implementation, Schneider’s default settings and selection mechanism are configured like this:

(defparameter *default-ev-problem*
  (make-ev-problem 256 16 6 5))

(let* ((problem *default-ev-problem*)
       (gene-pool (solve problem 64 (generation-terminator 3000)
                         :selection-method :truncation-selection
                         :mutation-count 1
                         :mutate-parents t
                         :interim-result-writer #'ev-interim-result-writer))
       (best-genome (most-fit-genome gene-pool (fitness-comparator problem))))
  (format t "~%Best = ~F~%Average = ~F~%~%"
          (fitness problem best-genome)
          (average-fitness problem gene-pool)))

This creates an instance of the ev-problem with 256 potential binding sites, 16 actual binding sites, a binding site width of 6 bases, and 5 bases per integer. It then evolves this population for 3000 generations using truncation selection (taking the top 50% of each gene pool to seed the next generation), changing one base in each genome, including the parent genomes, per generation.

The results are identical to those reported by Schneider. Over ten runs, each with a different random seed, the population evolves to have at least one member with no mistakes within 533 to 2324 generations (the mean was 1243.6 with a standard deviation of 570.91). R_{sequence} approaches R_{frequency} throughout this time. As maximally fit genomes take over the population, R_{sequence} oscillates around R_{frequency}.

While my implementation lacks the GUI provided by Schneider’s Java version, the R_{sequence} values output by ev-interim-result-writer show a similar distribution.

Variations

There are many configuration dimensions that can be explored with ev. I tested a few, including the effect of population size, selection method, mutation rate, and some problem-specific parameters.

Population Size

Increasing the population size from 64 to 256 but leaving the rest of the settings the same reduces the number of generations required to produce a maximally fit genome to between 251 and 2255 with a mean of 962.23 and a standard deviation of 792.11. A population size of 1000 results in a range of 293 to 2247 generations with a lower mean (779.4) and standard deviation (689.68), at a higher computation cost.

Selection Method

Schneider’s ev implementation uses truncation selection, using the top 50% of a population to seed the next generation. Using tournament selection with a population of 250, a tournament size of 50, and a mutation rate of 0.5% results in a maximally fit individual arising within 311 to 4561 generations (with a mean of 2529.9 and standard deviation of 1509.01). Increasing the population size to 500 gives a range of 262 to 4143 with a mean of 1441.2 and standard deviation of 1091.95.

Adding crossover to tournament selection with the other parameters remaining the same does not seem to significantly change the convergence rate.

Changing the tournament selection to mutate the parents as well as the children of the next generation does, however, have a significant impact. Using the same population size of 500 and mutation rate of 0.5% but mutating the parents results in a maximally fit individual within 114 to 1455 generations, with a mean of 534.1 and a standard deviation of 412.01.

Roulette wheel selection took much longer to converge, probably due to the fact that a large percentage of random genomes have identical fitness because no binding sites, real or spurious, are matched. This makes the areas of the wheel nearly equal for all genomes in a population.

Mutation Rate

In the original ev, exactly one base of 261 in each genome is modified per generation. This explores the fitness space immediately adjacent to the genome and is essentially simple hill climbing. This mutation rate is approximately 0.2% when applied to a string of bases represented by two bits each.

Changing the mutation count to a mutation rate of 1% results in ev taking an order of magnitude more generations to produce a maximally fit individual. Rates of 0.5% and 0.2% result in convergence periods closer to those seen with a single mutation per genome, both with truncation and tournament selection, particularly with larger population sizes. As Schneider notes, this is only about ten times the mutation rate of HIV-1 reverse transcriptase.

Binding Site Overlap

By default, binding sites are selected to be separated by at least the binding site width. When this restriction is removed, surprisingly the number of generations required to produce the first maximally fit genome ranges does not change significantly from the non-overlapping case.

Alternative Implementation Results

Population size seems to have the largest impact on the number of generations required to reach equilibrium. Mutation rate has a smaller effect, but can prevent convergence when set too high. Tournament selection takes a bit longer to converge than truncation selection, but the two are within the same order of magnitude. Roulette selection does not work well for this problem.

Unlike the Steiner network and some other problems, crossover doesn’t increase the convergence rate. Mutating the parent genomes before adding them to the next generation’s gene pool does have a measurable impact.

Regardless of selection method, mutation rate, or other parameters, R_{sequence} always evolves to and then oscillates around R_{frequency}.

Source Code

The code is available on GitHub. The required files are:

  • ga-package.lisp
  • ga.lisp
  • examples/ga-ev-package.lisp
  • examples/ga-ev.lisp
  • examples/load-ev.lisp

To run from the command line, make the examples directory your working directory and then call

ccl64 - -load load-ev.lisp`

if you’re using Clozure CL or

sbcl - -load load-ev.lisp`

if you’re using Steel Bank Common Lisp.

If you need a refresher on Common Lisp programming, Peter Seibel’s Practical Common Lisp is an excellent book.

Summary

In addition to being a good test case for evolutionary algorithms, ev is interesting because of its biological relevance. As Schneider points out in his Results section:

Although the sites can contain a maximum of 2L = 12 bits, the information content of the binding sites rises during this time until it oscillates around the predicted information content R_{frequency} = 4 bits, with R_{sequence} = 3.983 \pm 0.399 bits during the 1000 to 2000 generation interval.

With this, ev sticks a spoke in the tires of creationists who complain that GAs like Richard Dawkins’ weasel program only demonstrate “directed evolution”. There is nothing in the ev implementation that requires that R_{sequence} evolve to R_{frequency}, yet it does every time.

It’s well worth running the Java version of ev to see the recognizer, threshold, and binding sites all co-evolving. This is a clear answer to creationist objections about how “irreducibly complex” parts could evolve.

The common creationist argument from incredulity based on the complexity of cell biochemistry is also touched on by ev. Even with a genome where the recognizer and binding sites all overlap indiscriminately, a maximally fit recognizer evolves in a tiny population within a small number of generations.

Despite numerous attempts, Intelligent Design Creationists haven’t been able to refute any of Dr. Schneider’s claims or the evidence provided by ev. His history of responses to creationists is both amusing and devastating to his critics.

Unlike his IDCist critics, Schneider uses a clear, unambiguous, measurable definition of information and demonstrates that even the simplest known evolutionary mechanisms can increase it. Shannon information is produced randomly in the context of the environment but is preserved non-randomly by selection. Differential reproductive success does, therefore, generate information. As Schneider succinctly puts it:

Replication, mutation and selection are necessary and sufficient for information gain to occur.
This process is called evolution.
— Thomas D. Schneider

Please contact me by email if you have any questions, comments, or suggestions.

360 thoughts on “ev

  1. keiths: Alan:

    My point was that Shannon information does not refer to information content.

    But it does. Messages contain Shannon information, which is why they reduce the Shannon uncertainty of the receiver.

    Is it not clear from my comments that by content I meant meaning? Or are you telling me that Shannon information can distinguish between a meaningful message and a randomly generated one?

    ETA Shannon “information” should be “uncertainty” as per Dr Schneider.

  2. Regarding your last two comments:

    They’re just mistakes, Alan. Why not acknowledge them and move on?

  3. Alan Fox: Which people use CSI? How do they use CSI? And what is CSI, by the way?

    Everyone who can read and write uses CSI, Alan. All scientific papers are CSI, Alan. Computer programs are CSI, Alan. Newspapers are full of CSI as are encyclopedias.

    CSI is information in the normal use of the word- ie it is information with meaning and/ or function. Just as you have been told many, many times,

    With respect to biology CSI is as Crick says but it has to be a long enough sequence to be considered complex.

  4. Patrick: I know!I know!CSI is Dembski’s bogus metric described in his 2005 paper that no one has ever been able to calculate, in part because of the P(T|H) term.Some intelligent design creationists try to claim otherwise, but that’s the actual definition.

    LoL! @ Patrick! No, Patrick, that papers does not even deal with CSI. Search the paper for CSI or complex specified information. You won’t find it.

    The term P(T|H) is a conditional probability based on a materialistic hypothesis.

    This is the problem with ID critics- they are totally unaware of what they are trying to criticize.

  5. keiths,

    But it does. Messages contain Shannon information, which is why they reduce the Shannon uncertainty of the receiver

    Shannon never cared for meaning. True a message has meaning but Shannon didn’t care about that:

    The word information in this theory is used in a special mathematical sense that must not be confused with its ordinary usage. In particular, information must not be confused with meaning.– Warren Weaver, one of Shannon’s collaborators

    That is because the machines sending and receiving do not care about meaning.

  6. Crick defined information with respect to biology. Shannon showed us how to measure information in any given context. Kolmogorov provided the background on complexity. CSI wrt biology is just complex Crick information (Crick information is already specified as it is part of a function). And it exists regardless of how it arose.

    The point of CSI and ID is that no one has ever observed stochastic processes producing it, in any context. Yet we have experience and direct observations of intentional agencies producing it. Every time we have observed CSI and knew the cause it was via an intentional agency. It is a 100% observation. We wouldn’t even know how to test the claim that stochastic processes can produce it. So we can safely infer that CSI is an indication of intentional design. And yes, if someone steps up and demonstrates that stochastic processes can indeed produce CSI then the mere existence of CSI cannot be used to infer ID.

  7. Neil Rickert: Link needed (to this definition).

    Seriously? It is found in ID books- Dembski and Meyer. It has been discussed in this thread yesterday:

    Information means here the precise determination of sequence, either of bases in the nucleic acid or on amino acid residues in the protein. Sir Francis Crick 1958 Central Dogma

  8. Frankie: Seriously? It is found in ID books- Dembski and Meyer. It has been discussed in this thread yesterday:

    Well sure. Since it is obviously well known, I googled for “crick information”.

    It gave me page of information about Crick. It even offered to search for “cricket information”. But it did not give a link to the definition of “Crick information”.

    Since you think this well known, I’m assuming that you can provide the link that google didn’t have.

  9. Neil Rickert: Well sure.Since it is obviously well known, I googled for “crick information”.

    It gave me page of information about Crick.It even offered to search for “cricket information”.But it did not give a link to the definition of “Crick information”.

    Since you think this well known, I’m assuming that you can provide the link that google didn’t have.

    Wow. Talk about humping a strawman. I never said it was well known and I quoted Crick and gave the reference.

    That said it is well known to those who have actually read what Dembski and Meyer have said about CSI. I would think that such a vocal opponent would actually understand the concepts he is opposing. Silly me.

  10. Information means here the precise determination of sequence, either of bases in the nucleic acid or on amino acid residues in the protein. Sir Francis Crick 1958 “Central Dogma”

  11. keiths:
    Regarding your last two comments:

    They’re just mistakes, Alan. Why not acknowledge them and move on

    I see you’ve ignored what I’ve written and I therefore see no point in wasting further keystrokes on you.

  12. Frankie: Everyone who can read and write uses CSI, Alan.

    Presumably without realising it. Though as you haven’t made any attempt to explain what CSI is yet, I’ll wait for your definition to see how this could be so.

    All scientific papers are CSI, Alan. Computer programs are CSI, Alan. Newspapers are full of CSI as are encyclopedias.

    Didn’t have to wait long, did I? So biological systems are in a category with computer programs, newspapers and encyclopedias? Ooh, and scientific papers!

    CSI is information in the normal use of the word- ie it is information with meaning and/ or function.

    So meaningful information and not just nucleotide counts or Shannon information/uncertainty, eh?

    Just as you have been told many, many times,

    Repetition, like peppering abuse into comments, is not a convincing argument.

    With respect to biology CSI is as Crick says but it has to be a long enough sequence to be considered complex

    Now, now! We both know Crick has never used the phrase “complex specified information”. And if you are now claiming the functional content (expressed as a useful RNA or protein molecule) is dependent only on the length of the DNA sequence as per “500 bit threshold” or some such nonsense, then that is plainly daft. Is the sequence that codes for oxytocin complex or not?

  13. Frankie: The point of CSI and ID is that no one has ever observed stochastic processes producing it, in any context.

    I have an example of a stochastic process producing what I believe to be valid CSI.

    Are you Frankie, able to test that claim of mine? How?

    If you can describe, with a worked example, how you would test if CSI was indeed being produced, I’ll be happy to demonstrate stochastic processes producing CSI,

  14. Frankie: That said it is well known to those who have actually read what Dembski and Meyer have said about CSI.

    Do you have a page reference?

  15. Alan Fox,

    LoL! Crick didn’t have to use CSI, Alan. What the heck are you getting at? And no, I am not making the claim you say.

    Look, Alan, yours cannot account for living organisms. It can’t even account for ATP synthase. It has nothing but equivocations.

    As for convincing arguments, you don’t have any of those.

  16. Frankie: It can’t even account for ATP synthase

    For you “it was designed” is apparently a sufficient explanation, but not for everyone. Some people want just a little more then something that can equally be applied to everything we see.
    The moon? It was designed.
    ATP? It was designed.
    and so on.

    The worth of an answer is directly proportional to the number of questions it can answer. It was designed can answer any question of origin for anything, hence it has 0 utility or use.

  17. keiths:

    Regarding your last two comments:

    They’re just mistakes, Alan. Why not acknowledge them and move on?

    Alan:

    I see you’ve ignored what I’ve written and I therefore see no point in wasting further keystrokes on you.

    I haven’t ignored what you’ve written — I’m rolling my eyes at it. Hence my question:

    They’re just mistakes, Alan. Why not acknowledge them and move on?

  18. Alan,

    Since you insist on dragging this out, let me restate the obvious.

    1. You claimed, in response to Frankie, that Shannon is a red herring because he wasn’t concerned with biology:

    Kolmorogov was concerned with computing And Shannon worked on communication channels. Neither Kolmogorov nor Shannon were concerned with biology. Red herrings.

    That’s incorrect, and the fact that you were responding to Frankie doesn’t somehow magically make it correct.

    2. You confused channel capacity with Shannon information:

    Shannon information refers only to the carrying capacity of some information channel.

    Channel capacity is the rate at which Shannon information can be carried by the channel. It’s a distinct concept.

    3. This statement is also incorrect:

    My point was that Shannon information does not refer to information content.

    Shannon information does refer to information content, and it is the information content of messages that allows them to reduce the Shannon uncertainty of the receiver.

    4. You ask:

    Is it not clear from my comments that by [information] content I meant meaning?

    If you mean ‘meaning’ and not ‘information content’, then why not write ‘meaning’ instead of ‘information content’?

  19. Patrick,

    At the beginning of each run the binding site indexes are randomly generated. These can either be far enough apart to prevent overlap of binding sites or not. I tested both and got the same results, even with heavy overlap.

    Could you elaborate on that? How does one site gain fitness without penalizing an overlapping one?

    ETA: Is it that the recognizer evolves to favor sequences with the right sort of symmetry to “satisfy” both sites at once, in the case of overlap?

  20. OMagain: I have an example of a stochastic process producing what I believe to be valid CSI. Are you Frankie, able to test that claim of mine? How?

    We should have a contest.

    A test for CSI vs. a test for the power of cumulative selection. I would say the test for CSI at least has a head start, as there is a mathematical definition of CSI.

  21. Mung: We should have a contest.

    A test for CSI vs. a test for the power of cumulative selection. I would say the test for CSI at least has a head start, as there is a mathematical definition of CSI.

    That would be the CSI definition no IDiot has ever produced in 10+ years of asking.

  22. keiths:
    Patrick,

    Could you elaborate on that? How does one site gain fitness without penalizing an overlapping one?

    ETA: Is it that the recognizer evolves to favor sequences with the right sort of symmetry to “satisfy” both sites at once, in the case of overlap?

    That seems to be what’s happening. I was surprised that allowing overlap didn’t significantly affect the number of generations before a maximally fit genome appeared nor did it change the fact that Rsequence approaches then oscillates around Rfrequency.

    The weight matrix that comprises the bulk of the recognizer probably does penalize overlap in some sense, but because it models the biological reality that binding sites don’t require exact matches, the weights evolve to handle that case.

  23. Adapa: That would be the CSI definition no IDiot has ever produced in 10+ years of asking.

    To be fair, Dembski has provided a definition of CSI:

        \[\chi = -log_2[10^{120} \cdot \phi_S(T) \cdot P(T|H)]\]

    (See Lizzie’s post Belling the Cat from 2013.)

    The problem, of course, is that P(T|H) cannot be calculated. This causes intelligent design creationists to attempt to obfuscate by claiming that CSI really means something else. What that something else is, none specify in sufficient detail to actually test. They seem to consider that lack of testability a feature rather than a bug.

  24. Patrick: To be fair, Dembski has provided a definition of CSI:

    (See Lizzie’s post Belling the Cat from 2013.)

    The problem, of course, is that cannot be calculated.This causes intelligent design creationists to attempt to obfuscate by claiming that CSI really means something else.What that something else is, none specify in sufficient detail to actually test.They seem to consider that lack of testability a feature rather than a bug.

    No Patrick, that has nothing to do with CSI. P(T|H) is a conditional probability for any given materialistic hypothesis. And the reason it cannot be calculated in most cases is because of the failure of providing a testable materialistic hypothesis. IOW Patrick, no one knows how to test the claims of your position.

  25. keiths,

    Shannon information does refer to information content, and it is the information content of messages that allows them to reduce the Shannon uncertainty of the receiver.

    Except Warren Weaver, one of Shannon’s collaborators, says otherwise. Shannon did not care about meaning.

    The word information in this theory is used in a special mathematical sense that must not be confused with its ordinary usage. In particular, information must not be confused with meaning.– Warren Weaver, one of Shannon’s collaborators

  26. Patrick: The problem, of course, is that P(T|H) cannot be calculated.

    So when Dawkins calculated the chance of finding the weasel phrase he wasn’t calculating P(T|H)?

  27. Mung: So when Dawkins calculated the chance of finding the weasel phrase he wasn’t calculating P(T|H)?

    Interesting question. In the context of his cumulative selection demonstration, it is possible to calculate P(T|H) since all mechanisms are known and the problem is extremely simple.

    In the context of biology on Earth, the context of Dembski’s paper where he defines CSI, it is not possible to calculate P(T|H). That’s been beaten to death here in the threads I linked to earlier. CSI is demonstrably a bogus metric when applied to biological artifacts.

  28. Patrick: Interesting question.In the context of his cumulative selection demonstration, it is possible to calculate P(T|H) since all mechanisms are known and the problem is extremely simple.

    In the context of biology on Earth, the context of Dembski’s paper where he defines CSI, it is not possible to calculate P(T|H).That’s been beaten to death here in the threads I linked to earlier.CSI is demonstrably a bogus metric when applied to biological artifacts.

    LoL! The paper Patrick refers to never discusses CSI. It never discusses complex specified information. So it never defines CSI.

    And when it isn’t possible to calculate P(T|H) it is because your position cannot provide any testable hypotheses.

  29. Frankie: And when it isn’t possible to calculate P(T|H) it is because your position cannot provide any testable hypotheses.

    You spend an awful lot of time defeating something that sounds like it’s never existed in the first place.

  30. keiths:
    Frankie,

    You’re making the same mistake as Alan.‘Information content’ and ‘meaning’ aren’t synonymous.

    I am just stating the case as Alan is stumbling over the words.

    Also “content” is synonymous with “meaning”:

    http://www.merriam-webster.com/dictionary/content – scroll down to 4 content:

    2 a : substance, gist
    b : meaning, significance

    So I don’t understand your argument.

  31. Shannon measures information carrying capacity- as in how many bits of information content are available to carry meaning or perform a function.

  32. Frankie,

    A message has “information content” if it contains information. Shannon information is information. Therefore Alan’s statement is incorrect:

    My point was that Shannon information does not refer to information content.

  33. keiths,

    A message has “information content” if it contains information.

    It isn’t a message without information.

    Shannon information is information.

    Shannon just devised a way to measure information. His was a special mathematical sense. He never cared about meaning.

    So if the word “content” = “meaning”, then Alan is correct.

  34. Jesus, Frankie. This is not that difficult.

    Something has “water content” if it contains water. Something has “information content” if it contains information. Shannon information is information. Shannon information is not synonymous with meaning, as Weaver points out in the quote that you yourself provided.

    Therefore Alan’s statement is incorrect:

    My point was that Shannon information does not refer to information content.

    Messages that contain Shannon information do have information content.

  35. CSI has been defined. P(T|H) can be calculated. ev is not a biological artifact.

    This could be the best year ever at TSZ. 🙂

  36. Mung:
    CSI has been defined. P(T|H) can be calculated. ev is not a biological artifact.

    This could be the best year ever at TSZ. :)

    CSI has been defined by Dembski. The only people who don’t agree with that are the intelligent design creationists who, despite their dull intellects, have learned after repeated demonstrations that making their claims testable never ends well for them.

  37. Mung: A test for CSI vs. a test for the power of cumulative selection.

    Well, WEASEL exists already. Your go!

  38. Frankie: Except Warren Weaver, one of Shannon’s collaborators, says otherwise.

    This is meaningless nonsense.

    I know it is meaningless nonsense, because my computer received it as Shannon information. And, according to Frankie, Shannon information doesn’t have meaning.

  39. Well, WEASEL exists already. Your go!

    So is Weasel a better example of CSI or cumulative selection? Why?

  40. Neil Rickert: This is meaningless nonsense.

    I know it is meaningless nonsense, because my computer received it as Shannon information.And, according to Frankie, Shannon information doesn’t have meaning.

    Wrong, again. I said that Shannon didn’t care about information. I never said anything about Shannon information. I don’t even know what that is.

    Shannon just devised a metric to measure information

  41. Patrick: CSI has been defined by Dembski.The only people who don’t agree with that are the intelligent design creationists who, despite their dull intellects, have learned after repeated demonstrations that making their claims testable never ends well for them.

    Yes CSI has been defined by Dembski and misrepresented by the likes of you and your ilk.

  42. Frankie,

    It depends on how Alan was using the word “content”.

    No, it depends on whether a message containing Shannon information can be said to have “information content”. The answer is obviously “yes”, since Shannon information is information.

    Thus Alan’s statement is false:

    My point was that Shannon information does not refer to information content.

  43. colewd: So is Weasel a better example of CSI or cumulative selection? Why?

    WEASEL was created as an example of cumulative selection. If you like I can paste in the words of the author attesting to this.

  44. Frankie: Yes CSI has been defined by Dembski and misrepresented by the likes of you and your ilk.

    Tell us what the proper, correct usage of CSI is then. What useful work can be done with it?

  45. keiths:
    Frankie,

    No, it depends on whether a message containing Shannon information can be said to have “information content”. The answer is obviously “yes”, since Shannon information is information.

    Thus Alan’s statement is false:

    If “content” = “meaning” then Alan’s statement is correct. And there isn’t any such thing as “Shannon information”. He just devised a way to measure it. And again in order to be a message it has to have meaning.

Leave a Reply