I was recently sent a link to yet another of Dembski's wretched writings about specified complexity, titled Specification: The Pattern The Signifies Intelligence.
While reading this, I came across a statement that actually changes my opinion of Dembski. Before reading this, I thought that Dembski was just a liar. I thought that he was a reasonably competent mathematician who was willing to misuse his knowledge in order to prop up his religious beliefs with pseudo-intellectual rigor. I no longer think that. I've now become convinced that he's just an idiot who's able to throw around mathematical jargon without understanding it.
In this paper, as usual, he's spending rather a lot of time avoiding defining specification. Purportedly, he's doing a survey of the mathematical techniques that can be used to define specification. Of course, while rambling on and on, he manages to never actually say just what the hell specification is - just goes on and on with various discussions of what it could be.
Most of which are wrong.
"But wait", I can hear objectors saying. "It's his theory! How can his own definitions of his own theory be wrong? Sure, his theory can be wrong, but how can his own definition of his theory be wrong?" Allow me to head off that objection before I continue.
Demsbki's theory of specicfied complexity as a discriminator for identifying intelligent design relies on the idea that there are two distinct quantifiable properties: specification, and complexity. He argues that if you can find systems that posess sufficient quantities of both specification and complexity, that those systems cannot have arisen except by intelligent intervention.
But what if Demsbki defines specification and complexity as the same thing? Then his definitions are wrong: because he requires them to be distinct concepts, but he defines them as being the same thing.
Throughout this paper, he pretty ignores the complexity to focus on specification. He's pretty careful never to say "specification is this", but rather "specification can be this". If you actually read what he does say about specification, and you go back and compare it to some of his other writings about complexity, you'll find a positively amazing resemblance.
But onwards. Here's the part that really blew my mind.
One of the methods that he purports to use to discuss specification is based on Kolmogorov-Chaitin algorithmic information theory. And in his explanation, he demonstrates a profound lack of comprehension of anything about KC theory.
First - he purports to discuss K-C within the framework of probability theory. K-C theory has nothing to do with probability theory. K-C theory is about the meaning of quantifying information; the central question of K-C theory is: How much information is in a given string? It defines the answer to that question in terms of computation and the size of programs that can generate that string.
Now, the quotes that blew my mind:
Consider a concrete case. If we flip a fair coin and note the occurrences of heads and tails in
order, denoting heads by 1 and tails by 0, then a sequence of 100 coin flips looks as follows:(R) 11000011010110001101111111010001100011011001110111 00011001000010111101110110011111010010100101011110.
This is in fact a sequence I obtained by flipping a coin 100 times. The problem algorithmic
information theory seeks to resolve is this: Given probability theory and its usual way of
calculating probabilities for coin tosses, how is it possible to distinguish these sequences in terms
of their degree of randomness? Probability theory alone is not enough. For instance, instead of
flipping (R) I might just as well have flipped the following sequence:(N) 11111111111111111111111111111111111111111111111111 11111111111111111111111111111111111111111111111111.
Sequences (R) and (N) have been labeled suggestively, R for "random," N for "nonrandom."
Chaitin, Kolmogorov, and Solomonoff wanted to say that (R) was "more random" than (N). But
given the usual way of computing probabilities, all one could say was that each of these
sequences had the same small probability of occurring, namely, 1 in 2100, or approximately 1 in
1030. Indeed, every sequence of 100 coin tosses has exactly this same small probability of
To get around this difficulty Chaitin, Kolmogorov, and Solomonoff supplemented conventional
probability theory with some ideas from recursion theory, a subfield of mathematical logic that
provides the theoretical underpinnings for computer science and generally is considered quite far
removed from probability theory.
It would be difficult to find a more misrepresentative description of K-C theory than this. This has nothing to do with the original motivation of K-C theory; it has nothing to do with the practice of K-C theory; and it has pretty much nothing to do with the actual value of K-C theory. This is, to put it mildly, a pile of nonsense spewed from the keyboard of an idiot who thinks that he knows something that he doesn't.
But it gets worse.
Since one can always describe a sequence in terms of itself, (R) has the descriptioncopy '11000011010110001101111111010001100011011001110111 00011001000010111101110110011111010010100101011110'.
Because (R) was constructed by flipping a coin, it is very likely that this is the shortest
description of (R). It is a combinatorial fact that the vast majority of sequences of 0s and 1s have
as their shortest description just the sequence itself. In other words, most sequences are random
in the sense of being algorithmically incompressible. It follows that the collection of nonrandom
sequences has small probability among the totality of sequences so that observing a nonrandom
sequence is reason to look for explanations other than chance.
This is so very wrong that it demonstrates a total lack of comprehension of what K-C theory is about, how it measures information, or what it says about anything. No one who actually understands K-C theory would ever make a statement like Dembski's quote above. No one.
But to make matters worse - this statement explicitly invalidates the entire concept of specified complexity. What this statement means - what it explicitly says if you understand the math - is that specification is the opposite of complexity. Anything which posesses the property of specification by definition does not posess the property of complexity.
In information-theory terms, complexity is non-compressibility. But according to Dembski, in IT terms, specification is compressibility. Something that possesses "specified complexity" is therefore something which is simultaneously compressible and non-compressible.
The only thing that saves Dembski is that he hedges everything that he says. He's not saying that this is what specification means. He's saying that this could be what specification means. But he also offers a half-dozen other alternative definitions - with similar problems. Anytime you point out what's wrong with any of them, he can always say "No, that's not specification. It's one of the others." Even if you go through the whole list of possible definitions, and show why every single one is no good - he can still say "But I didn't say any of those were the definition".
But the fact that he would even say this - that he would present this as even a possibility for the definition of specification - shows that Dembski quite simply does not get it. He believes that he gets it - he believes that he gets it well enough to use it in his arguments. But there is absolutely no way that he understands it. He is an ignorant jackass pretending to know things so that he can trick people into accepting his religious beliefs.