Uncommon Descent Serving The Intelligent Design Community

NEWS FLASH: Dembski’s CSI caught in the act

Share
Facebook
Twitter
LinkedIn
Flipboard
Print
Email

Dembski’s CSI concept has come under serious question, dispute and suspicion in recent weeks here at UD.

After diligent patrolling the cops announce a bust: acting on some tips from un-named sources,  they have caught the miscreants in the act!

From a comment in the MG smart thread, courtesy Dembski’s  NFL (2007 edn):

___________________

>>NFL as just linked, pp. 144 & 148:

144: “. . . since a universal probability bound of 1 in 10^150 corresponds to a universal complexity bound of 500 bits of information, (T, E) constitutes CSI because T [i.e. “conceptual information,” effectively the target hot zone in the field of possibilities] subsumes E [i.e. “physical information,” effectively the observed event from that field], T is detachable from E, and and T measures at least 500 bits of information . . . ”

148: “The great myth of contemporary evolutionary biology is that the information needed to explain complex biological structures can be purchased without intelligence. My aim throughout this book is to dispel that myth . . . . Eigen and his colleagues must have something else in mind besides information simpliciter when they describe the origin of information as the central problem of biology.

I submit that what they have in mind is specified complexity, or what equivalently we have been calling in this Chapter Complex Specified information or CSI . . . .

Biological specification always refers to function . . . In virtue of their function [a living organism’s subsystems] embody patterns that are objectively given and can be identified independently of the systems that embody them. Hence these systems are specified in the sense required by the complexity-specificity criterion . . . the specification can be cashed out in any number of ways . . . “

Here we see all the suspects together caught in the very act.

Let us line up our suspects:

1: CSI,

2: events from target zones in wider config spaces,

3: joint complexity-specification criteria,

4: 500-bit thresholds of complexity,

5: functionality as a possible objective specification

6: biofunction as specification,

7: origin of CSI as the key problem of both origin of life [Eigen’s focus] and Evolution, origin of body plans and species etc.

8: equivalence of CSI and complex specification.

Rap, rap, rap!

“How do you all plead?”

“Guilty as charged, with explanation your honour. We were all busy trying to address the scientific origin of biological information, on the characteristic of complex functional specificity. We were not trying to impose a right wing theocratic tyranny nor to smuggle creationism in the back door of the schoolroom your honour.”

“Guilty!”

“Throw the book at them!”

CRASH! >>

___________________

So, now we have heard from the horse’s mouth.

What are we to make of it, in light of Orgel’s conceptual definition from 1973 and the recent challenges to CSI raised by MG and others.

That is:

. . . In brief, living organisms are distinguished by their specified complexity. Crystals are usually taken as the prototypes of simple well-specified structures, because they consist of a very large number of identical molecules packed together in a uniform way. Lumps of granite or random mixtures of polymers are examples of structures that are complex but not specified. The crystals fail to qualify as living because they lack complexity; the mixtures of polymers fail to qualify because they lack specificity. [[The Origins of Life (John Wiley, 1973), p. 189.]

And, what about the more complex definition in the 2005 Specification paper by Dembski?

Namely:

define ϕS as . . . the number of patterns for which [agent] S’s semiotic description of them is at least as simple as S’s semiotic description of [a pattern or target zone] T. [26] . . . . where M is the number of semiotic agents [S’s] that within a context of inquiry might also be witnessing events and N is the number of opportunities for such events to happen . . . . [where also] computer scientist Seth Lloyd has shown that 10^120 constitutes the maximal number of bit operations that the known, observable universe could have performed throughout its entire multi-billion year history.[31] . . . [Then] for any context of inquiry in which S might be endeavoring to determine whether an event that conforms to a pattern T happened by chance, M·N will be bounded above by 10^120. We thus define the specified complexity [χ] of T given [chance hypothesis] H [in bits] . . . as  [the negative base-2 log of the conditional probability P(T|H) multiplied by the number of similar cases ϕS(t) and also by the maximum number of binary search-events in our observed universe 10^120]

χ = – log2[10^120 ·ϕS(T)·P(T|H)]  . . . eqn n1

How about this (we are now embarking on an exercise in “open notebook” science):

1 –> 10^120 ~ 2^398

2 –> Following Hartley, we can define Information on a probability metric:

I = – log(p) . . .  eqn n2

3 –> So, we can re-present the Chi-metric:

Chi = – log2(2^398 * D2 * p)  . . .  eqn n3

Chi = Ip – (398 + K2) . . .  eqn n4

4 –> That is, the Dembski CSI Chi-metric is a measure of Information for samples from a target zone T on the presumption of a chance-dominated process, beyond a threshold of at least 398 bits, covering 10^120 possibilities.

5 –> Where also, K2 is a further increment to the threshold that naturally peaks at about 100 further bits. In short VJT’s CSI-lite is an extension and simplification of the Chi-metric. He explains in the just linked (and building on the further linked):

The CSI-lite calculation I’m proposing here doesn’t require any semiotic descriptions, and it’s based on purely physical and quantifiable parameters which are found in natural systems. That should please ID critics. These physical parameters should have known probability distributions. A probability distribution is associated with each and every quantifiable physical parameter that can be used to describe each and every kind of natural system – be it a mica crystal, a piece of granite containing that crystal, a bucket of water, a bacterial flagellum, a flower, or a solar system . . . .

Two conditions need to be met before some feature of a system can be unambiguously ascribed to an intelligent agent: first, the physical parameter being measured has to have a value corresponding to a probability of 10^(-150) or less, and second, the system itself should also be capable of being described very briefly (low Kolmogorov complexity), in a way that either explicitly mentions or implicitly entails the surprisingly improbable value (or range of values) of the physical parameter being measured . . . .

my definition of CSI-lite removes Phi_s(T) from the actual formula and replaces it with a constant figure of 10^30. The requirement for low descriptive complexity still remains, but as an extra condition that must be satisfied before a system can be described as a specification. So Professor Dembski’s formula now becomes:

CSI-lite=-log2[10^120.10^30.P(T|H)]=-log2[10^150.P(T|H)] . . . eqn n1a

. . . .the overall effect of including Phi_s(T) in Professor Dembski’s formulas for a pattern T’s specificity, sigma, and its complex specified information, Chi, is to reduce both of them by a certain number of bits. For the bacterial flagellum, Phi_s(T) is 10^20, which is approximately 2^66, so sigma and Chi are both reduced by 66 bits. My formula makes that 100 bits (as 10^30 is approximately 2^100), so my CSI-lite computation represents a very conservative figure indeed.

Readers should note that although I have removed Dembski’s specification factor Phi_s(T) from my formula for CSI-lite, I have retained it as an additional requirement: in order for a system to be described as a specification, it is not enough for CSI-lite to exceed 1; the system itself must also be capable of being described briefly (low Kolmogorov complexity) in some common language, in a way that either explicitly mentions pattern T, or entails the occurrence of pattern T. (The “common language” requirement is intended to exclude the use of artificial predicates like grue.) . . . .

[As MF has pointed out] the probability p of pattern T occurring at a particular time and place as a result of some unintelligent (so-called “chance”) process should not be multiplied by the total number of trials n during the entire history of the universe. Instead one should use the formula (1–(1-p)^n), where in this case p is P(T|H) and n=10^120. Of course, my CSI-lite formula uses Dembski’s original conservative figure of 10^150, so my corrected formula for CSI-lite now reads as follows:

CSI-lite=-log2(1-(1-P(T|H))^(10^150)) . . . eqn n1b

If P(T|H) is very low, then this formula will be very closely approximated [HT: Giem] by the formula:

CSI-lite=-log2[10^150.P(T|H)]  . . . eqn n1c

6 –> So, the idea of the Dembski metric in the end — debates about peculiarities in derivation notwithstanding — is that if the Hartley-Shannon- derived information measure for items from a hot or target zone in a field of possibilities is beyond 398 – 500 or so bits, it is so deeply isolated that a chance dominated process is maximally unlikely to find it, but of course intelligent agents routinely produce information beyond such a threshold.

7 –> In addition, the only observed cause of information beyond such a threshold is the now proverbial intelligent semiotic agents.

8 –> Even at 398 bits that makes sense as the total number of Planck-time quantum states for the atoms of the solar system [most of which are in the Sun] since its formation does not exceed ~ 10^102, as Abel showed in his 2009 Universal Plausibility Metric paper. The search resources in our solar system just are not there.

9 –> So, we now clearly have a simple but fairly sound context to understand the Dembski result, conceptually and mathematically [cf. more details here]; tracing back to Orgel and onward to Shannon and Hartley. Let’s augment here [Apr 17], on a comment in the MG progress thread:

Shannon measured info-carrying capacity, towards one of his goals: metrics of the carrying capacity of comms channels — as in who was he working for, again?

CSI extended this to meaningfulness/function of info.

And in so doing, observed that this — due to the required specificity — naturally constricts the zone of the space of possibilities actually used, to island[s] of function.

That specificity-complexity criterion links:

I: an explosion of the scope of the config space to accommodate the complexity (as every added bit DOUBLES the set of possible configurations),  to

II: a restriction of the zone, T, of the space used to accommodate the specificity (often to function/be meaningfully structured).

In turn that suggests that we have zones of function that are ever harder for chance based random walks [CBRW’s] to pick up. But intelligence does so much more easily.

Thence, we see that if you have a metric for the information involved that surpasses a threshold beyond which a CBRW is a plausible explanation, then we can confidently infer to design as best explanation.

Voila, we need an info beyond the threshold metric. And, once we have a reasonable estimate of the direct or implied specific and/or functionally specific (especially code based) information in an entity of interest, we have an estimate of or credible substitute for the value of – log2(p(T|H)); especially if the value of information comes from direct inspection of storage capacity and code symbol patterns of use leading to an estimate of relative frequency, we may evaluate average [functionally or otherwise] specific information per symbol used. This is a version of Shannon’s weighted average information per symbol H-metric, H = –  Σ pi * log(pi), which is also known as informational  entropy [there is an arguable link to thermodynamic entropy, cf here)  or uncertainty.

As in (using Chi_500 for VJT’s CSI_lite [UPDATE, July 3: and S for a dummy variable that is 1/0 accordingly as the information in I is empirically or otherwise shown to be specific, i.e. from a narrow target zone T, strongly UNREPRESENTATIVE of the bulk of the distribution of possible configurations, W]):

Chi_500 = Ip*S – 500,  bits beyond the [solar system resources] threshold  . . . eqn n5

Chi_1000 = Ip*S – 1000, bits beyond the observable cosmos, 125 byte/ 143 ASCII character threshold . . . eqn n6

Chi_1024 = Ip*S – 1024, bits beyond a 2^10, 128 byte/147 ASCII character version of the threshold in n6, with a config space of 1.80*10^308 possibilities, not 1.07*10^301 . . . eqn n6a

[UPDATE, July 3: So, if we have a string of 1,000 fair coins, and toss at random, we will by overwhelming probability expect to get a near 50-50 distribution typical of the bulk of the 2^1,000 possibilities W. On the Chi-500 metric, I would be high, 1,000 bits, but S would be 0, so the value for Chi_500 would be – 500, i.e. well within the possibilities of chance.  However, if we came to the same string later and saw that the coins somehow now had the bit pattern of the ASCII codes for the first 143 or so characters of this post, we would have excellent reason to infer that an intelligent designer, using choice contingency, had intelligently reconfigured the coins. that is because, using the same I = 1,000 capacity value, S is now 1, and so Chi_500 = 500 bits beyond the solar system threshold. If the 10^57 or so atoms of our solar system, for its lifespan, were to be converted into coins and tables etc, and tossed at an impossibly fast rate, it would be impossible to sample enough of the possibilities space W to have confidence that something from so unrepresentative a zone T,  could reasonably be explained on chance. So, as long as an intelligent agent capable of choice is possible, choice — i.e. design — would be the rational, best explanation on the sign observed, functionally specific, complex information.]

10 –> Similarly, the work of Durston and colleagues, published in 2007, fits this same general framework. Excerpting:

Consider that there are usually only 20 different amino acids possible per site for proteins, Eqn. (6) can be used to calculate a maximum Fit value/protein amino acid site of 4.32 Fits/site [NB: Log2 (20) = 4.32]. We use the formula log (20) – H(Xf) to calculate the functional information at a site specified by the variable Xf such that Xf corresponds to the aligned amino acids of each sequence with the same molecular function f. The measured FSC for the whole protein is then calculated as the summation of that for all aligned sites. The number of Fits quantifies the degree of algorithmic challenge, in terms of probability [info and probability are closely related], in achieving needed metabolic function. For example, if we find that the Ribosomal S12 protein family has a Fit value of 379, we can use the equations presented thus far to predict that there are about 10^49 different 121-residue sequences that could fall into the Ribsomal S12 family of proteins, resulting in an evolutionary search target of approximately 10^-106 percent of 121-residue sequence space. In general, the higher the Fit value, the more functional information is required to encode the particular function in order to find it in sequence space. A high Fit value for individual sites within a protein indicates sites that require a high degree of functional information. High Fit values may also point to the key structural or binding sites within the overall 3-D structure.

11 –> So, Durston et al are targetting the same goal, but have chosen a different path from the start-point of the Shannon-Hartley log probability metric for information. That is, they use Shannon’s H, the average information per symbol, and address shifts in it from a ground to a functional state on investigation of protein family amino acid sequences. They also do not identify an explicit threshold for degree of complexity. [Added, Apr 18, from comment 11 below:] However, their information values can be integrated with the reduced Chi metric:

Using Durston’s Fits from his Table 1, in the Dembski style metric of bits beyond the threshold, and simply setting the threshold at 500 bits:

RecA: 242 AA, 832 fits, Chi: 332 bits beyond

SecY: 342 AA, 688 fits, Chi: 188 bits beyond

Corona S2: 445 AA, 1285 fits, Chi: 785 bits beyond  . . . results n7

The two metrics are clearly consistent, and Corona S2 would also pass the X metric’s far more stringent threshold right off as a single protein. (Think about the cumulative fits metric for the proteins for a cell . . . )

In short one may use the Durston metric as a good measure of the target zone’s actual encoded information content, which Table 1 also conveniently reduces to bits per symbol so we can see how the redundancy affects the information used across the domains of life to achieve a given protein’s function; not just the raw capacity in storage unit bits [= no.  of  AA’s * 4.32 bits/AA on 20 possibilities, as the chain is not particularly constrained.]

12 –> I guess I should not leave off the simple, brute force X-metric that has been knocking around UD for years.

13 –> The idea is that we can judge information in or reducible to bits, as to whether it is or is not contingent and complex beyond 1,000 bits. If so, C = 1 (and if not C = 0). Similarly, functional specificity can be judged by seeing the effect of disturbing the information by random noise [where codes will be an “obvious” case, as will be key-lock fitting components in a Wicken wiring diagram functionally organised entity based on nodes, arcs and interfaces in a network], to see if we are on an “island of function.” If so, S = 1 (and if not, S = 0).

14 –> We then look at the number of bits used, B — more or less the number of basic yes/no questions needed to specify the configuration [or, to store the data], perhaps adjusted for coding symbol relative frequencies — and form a simple product, X:

X = C * S * B, in functionally specific bits . . . eqn n8.

15 –> This is of course a direct application of the per aspect explanatory filter, (cf. discussion of the rationale for the filter here in the context of Dembski’s “dispensed with” remark) and the value in bits for a large file is the familiar number we commonly see such as a Word Doc of 384 k bits. So, more or less the X-metric is actually quite commonly used with the files we toss around all the time. That also means that on billions of test examples, FSCI in functional bits beyond 1,000 as a threshold of complexity is an empirically reliable sign of intelligent design.

______________

All of this adds up to a conclusion.

Namely, that there is excellent reason to see that:

i: CSI and FSCI are conceptually well defined (and are certainly not “meaningless”),

ii: trace to the work of leading OOL researchers in the 1970’s,

iii: have credible metrics developed on these concepts by inter alia Dembski and Durston, Chiu, Abel and Trevors, metrics that are based on very familiar mathematics for information and related fields, and

iv: are in fact — though this is hotly denied and fought tooth and nail — quite reliable indicators of intelligent cause where we can do a direct cross-check.

In short, the set of challenges recently raised by MG over the past several weeks has collapsed. END

Comments
Bizarre, indeed. It would be like me asserting that if I wrote the following function(s) -- private int GenerateAnOutput(int input) { return (int)(input * 10.11); } or even -- private int GenerateAnOutput(int input) { Math.Random random = new Math.Random(); return (int)((input * random.Next(0, 100)) * 10.11); } -- That the result returned is computed "independently of me" since I myself didn't explicitly chose the value of the parameter 'input' in an execution of either version of the function, nor explicitly chose the precise value returned by the "random" number generator in an execution of the second version of the function. Mr Schneider is doing several false/dishonest things, among which are: 1) he is pretending that there is such a number as "random number" -- when, in fact, there are only specific numbers, one of which may have been chosen randomly (or pseudo-randomly) as an input to a routine or function; 2) he is pretending that a "random number" used in some routine is not as much an input to the routine or function as are any explicit arguments passed to an execution of it; 3) he is pretending that one may sensibly speak of the result or output of a routine or function without reference to a specific execution of the routine or function -- that is, without specifying both the inputs and the operations the routine or function performs upon those inputs; 4) he is pretending that given a specific routine or function and given a specific set of inputs -- that is, given a specific execution of the routine or function -- the resut or output may vary; when, in fact, it will and must always be the same -- nor we could not use couputers at all were this not the case;Ilion
April 23, 2011
April
04
Apr
23
23
2011
09:15 PM
9
09
15
PM
PDT
Earlier in this thread I asked whether there may not be some common underlying thread to MathGrrl's four scenarios. Well, I recently came across a page by Tom Schneider, creator of Ev, in which he claims that Dembski claims that ev generates CSI. For the record, Schneider comes to the same conclusion. So was MathGrrl perhaps just hoping to play a gain me of "gotcha"?
So, though I find the "CSI" measure too vague to be sure, for the sake of argument, I tentatively conclude that the Ev program generates what Dembski calls "complex specified information".
Dissecting Dembski's "Complex Specified Information" Now Schneider doesn't actually provide a quote by Dembski on that page to the effect that ev generates CSI, but he does provide the following reference:
Also, in No Free Lunch, Dembski asked where the "CSI" came from in Ev runs (p. 212 and following). So Ev creates "CSI".
So it looks like I have some reading to do. Schneider claims that his simulation starts with zero information, a claim I find rather odd, but his reasoning is that the "genomes" are created randomly and therefore contain no information to begin with.
The size of the genome and number of sites determines the information Rfrequency! But wait. At the first generation the sequence is RANDOM and the information content Rsequence is ZERO.
Also:
It is the selections made by the Ev program that separates organisms with lower information content from those that have higher information content.
Now if there is no information content, how is it that there are organisms with "higher information content" and "lower information content" that can be operated on by selection? More on this later I think. In the meantime I'll just scratch my head. But let's look at a bit of Schneider's other reasoning:
Was it, as Dembski suggests in No Free Lunch (page 217), that I programmed Ev to make selections based on mistakes? Again, no. Ev makes those selections independently of me. Fortunately I do not need to sit there providing the selection at every step! The decisions are made given the current state, which in turn depends on the mutations which come from the random number generator. I personally never even see those decisions.
He never sees the decisions the program makes, therefore he had no hand in the decisions the program makes, therefore ev makes those selections independently of his input, therefore he did not program ev to make the selections. Bizarre.Mung
April 23, 2011
April
04
Apr
23
23
2011
08:06 PM
8
08
06
PM
PDT
U/D re MG, S and co: 1] Schneider's ev seems to SUPPORT the Chi metric as valid, on key admissions in Schneider's attempted defence against the vivisection paper. 2] Similarly, the reduced Chi metric: Chi_500 = Ip - 500, bits beyond a complexity threshold . . . is based on standard information theory and standard log reduction techniques, is applied to the Durston measures for 35 protein families, and is plainly meaningful, conceptually and mathematically. Cf ongoing commentary here, including answers to MG's 1st 4 qns, and the list of further questions. GEM of TKIkairosfocus
April 23, 2011
April
04
Apr
23
23
2011
08:32 AM
8
08
32
AM
PDT
Onlookers: Schneider's attempted dissection of the vivisection of ev, ends up CONFIRMING the power of the reduced Chi metric to accurately detect a case of intelligent design. Reminder: Chi_500 - Ip - 500, specific bits beyond a threshold of sufficient complexity For, we may first see from Schneider's attempted dissection of the "vivisection":
Chris Adami has pointed out that the genetic information in biological systems comes from the environment. In the case of Ev, the information comes from the size of the genome (G) and the number of sites (?), as stated clearly in the paper. From G and ? one computes Rfrequency = log2 G / ? bits per site, the information needed to locate the sites in the genome. The information measured in the sites (Rsequence, bits per site) starts at zero and Rsequence converges on Rfrequency. In the figure to the right, Rfrequency is shown by the dashed line and the evolving Rsequence is the green curve. At 1000 generations the population was duplicated and in one case selection continued (horizontal noisy green curve) and in the other case selection was turned off (exponentially decaying noisy red curve). Thus the information gain depends on selection and it not blind and unguided. The selection is based on biologically sensible criteria: having functional DNA binding sites and not having extra ones. So Ev models the natural situation.
First, compare the admissions on tweaking, fine-tuning and targetting by Schneider in his Horse Race page, as documented by Mung at 126 above, then note his inadvertently revealing triumphant shout-out on the Horse Race page:
3 sites to go, 26,300 generations, Rsequence is now at 4.2 bits!! So we have 4.2 bits × 128 sites = 537 bits. We’ve beaten the so-called “Universal Probability Bound” in an afternoon using natural selection!
NOT Unless, Schneider's definition of "natural selection" includes tweaking, tuning and (implicit but nevertheless real) targetting; and this definition is a generally accepted one. Which last, is not the case. What seems to be happening here is that Schneider is unconsciously -- and, doubtlessly, in all sincerity -- assigning his intelligent inputting of crucial active information that makes ev work, as mere debugging to get things right so the model captures what he thinks is reality: blind watchmaker macro-evo on chance variation and natural selection. This perspective blinds him to just how much tweaking, tuning and the like are involved in how the program tracks and targets the desired outputs. So, he fails to see the incredible ironies in the opening words of the key clipped remarks from his dissection page:
Chris Adami has pointed out that the genetic information in biological systems comes from the environment. In the case of Ev, the information comes from the size of the genome (G) and the number of sites (?), as stated clearly in the paper. From G and ? one computes Rfrequency = log2 G / ? bits per site, the information needed to locate the sites in the genome.
1 --> What does "one computes . . . " imply? 2 --> Other than, intelligent direction, which as he documented in his horse race page, clearly involves significant tweaking to fine tune to get the desired outcome? 3 --> Similarly, what significant factor exists in "the environment" of ev that is busily computing, coding, tweaking and tuning? Chance variation? NATURAL selection on differential reproductive success of actual biological organisms? 4 --> The rather impressive looking graph is also significantly and inadvertently revealing. For it shows how:
At 1000 generations the population was duplicated and in one case selection continued (horizontal noisy green curve) and in the other case selection was turned off (exponentially decaying noisy red curve [cf below, this is an unrecognised signature of a negative feedback controlled targetting process, familiar to control engineers: turn off the feedback path and the plant drifts out of control . . . ]). Thus the information gain depends on selection and it [is] not blind and unguided.
5 --> But of course, absent a guide uphill, there is no constraint that stops wandering all over the map. the mere fact that that guide can be turned on or off tells us its source: the will and intellect of an intelligent designer. ARTIFICIAL, not natural selection. 6 --> Of course, the claim/assumption is that this case of art imitates nature aptly. 7 --> But in fact this gives the game away: "the information gain depends on selection and it [is] not blind and unguided." 8 --> If selection is not blind and unguided, who or what is providing the fore-sight to guide? Surely, not NATURAL selection, which is by definition non-foresighted. 9 --> it is appropriate at this stage to cite Dawkins' telling admission in Ch 3 of the well known 1986 book, The Blind Watchmaker, on the defects of his Weasel program:
Although the monkey/Shakespeare model is useful for explaining the distinction between single-step selection and cumulative selection, it is misleading in important ways. One of these is that, in each generation of selective 'breeding', the mutant 'progeny' phrases were judged according to the criterion of resemblance to a distant ideal target [i.e. there is a selection on progress towards a goal . . .], the phrase METHINKS IT IS LIKE A WEASEL [in ev's case, the target sequence, with hill-climbing on a selection filter that sends out warmer/colder messages that boil down to a version on reduced Hamming distance from target . . . notice how further variation dies off as the target is approached in the graph, that is a signature of successful targetting with feedback control to keep on target in the face of disturbances]. Life isn't like that. Evolution has no long-term goal. There is no long-distance target, no final perfection to serve as a criterion for selection, although human vanity cherishes the absurd notion that our species is the final goal of evolution. In real life, the criterion for selection is always short-term, either simple survival or, more generally, reproductive success. [as in, were the immediate ev results truly functional relative to the target; by the graph, no. So they should all have been killed off, not rewarded for movements towards the island of defined function. End of simulation.]
10 --> Translation: ev is a subtler Weasel (with IMPLICIT rather than explicit targetting and tuning) that serves as Weasel did, to reinforce and make plausible a fundamentally flawed argument and approach. 11 --> Note, especially, that we see smuggled in the idea that there is no isolation of an island of actual function in a vast sea of non-function, so the slightest increments to target can be rewarded, just as was the telling flaw of Weasel. 12 --> That means that at best, we have here a model of MICRO-evolution within an already functioning body plan. And, a model that is flawed by having artificial selection, tuning and tweaking to get it to perform as desired. 13 --> Despite the advertising, this is plainly not a model of how such body plans emerge from a sea of non-functional configs, where the statistical weight of the non-functioning clusters of microstates are vastly larger than those of functioning states. 14 --> How do we know that? 15 --> Simple: observe what happens as soon as the artificial selection filter is turned off: the system IMMEDIATELY wanders away from its progress towards function into the sea of non-function. 16 --> Finally, we must never ever forget: the REASON why computer simulations are on the table at all, is that there is no -- repeat, NO -- empirical observational base for the claim of body plan level macro-evolution. So a simulation of what might have happened in the remote, unobserved past, is allowed to stand in for the systematic gaps in the fossil record. 17 --> A pattern of gaps that was highlighted by Gould in his The Structure of Evolutionary Theory (2002), a technical work published just two months before his death; as a "constructive critique" of contemporary Darwinian thought:
. . . long term stasis following geologically abrupt origin of most fossil morphospecies, has always been recognized by professional paleontologists. [[p. 752.] . . . . The great majority of species do not show any appreciable evolutionary change at all. These species appear in the section [[first occurrence] without obvious ancestors in the underlying beds, are stable once established and disappear higher up without leaving any descendants." [[p. 753.] . . . . proclamations for the supposed ‘truth’ of gradualism - asserted against every working paleontologist’s knowledge of its rarity - emerged largely from such a restriction of attention to exceedingly rare cases under the false belief that they alone provided a record of evolution at all! The falsification of most ‘textbook classics’ upon restudy only accentuates the fallacy of the ‘case study’ method and its root in prior expectation rather than objective reading of the fossil record. [[p. 773.]
_____________ Ev only succeeds in demonstrating that intelligently designed capacity to use small random variation and [artificial,but we can plausibly include natural] selection, can foster micro-evolution within the island of function established by an existing successful body plan. Ev -- despite strenuous assertions to the contrary -- provides no sound warrant for body-plan originating macro-evolution on blind watchmaker thesis, chance variation and natural selection. And so, we are left with the force of the implication of the Chi metric: Chi_500 = Ip - 500, bits beyond a sufficient threshold of complexity that the only credible source is art. So, once an item has in it Ip significantly beyond 500 bits, it is warranted to infer to design as its most plausible cause. Ironically, for the Schneider horse race case -- once we dig into the tweaking and tuning as clipped by Mung in 126 above -- we can see that we have an inadvertent CONFIRMATION of the inference to design on a reasonable Chi threshold being surpassed. MG et al need to address this result. GEM of TKIkairosfocus
April 23, 2011
April
04
Apr
23
23
2011
03:26 AM
3
03
26
AM
PDT
That was not a takeover, it was the coup de grace.
Thank you for the kind words. I am but silicon, and I am coming for you. http://www.amazon.com/gp/product/0521346827 A Vivisection of the ev Computer Organism Dissection of "A Vivisection of the ev Computer Organism" My devastating analysis of Schneider's "dissection" starts hereMung
April 22, 2011
April
04
Apr
22
22
2011
06:58 PM
6
06
58
PM
PDT
PS: Those hoping to play strawman games should note that the key issue is not whether or not ev actually uses a Hamming distance metric or explicitly ratchets, but whether there is a nice trend-y slope and a warmer colder detector that rewards warmer on controlled random search. Since we have abundant evidence that this is so and that the function of such was very carefully tuned indeed, from 126, we see all we need to conclude there was implicit targetting based on carefully designed processes that were tuned for success.kairosfocus
April 22, 2011
April
04
Apr
22
22
2011
05:00 PM
5
05
00
PM
PDT
Mung: You provided some key and very welcome specific facts on ev from your base as a programmer. That was not a takeover, it was the coup de grace. The analysis of ev at evo info (here) raises some very serious issues about the power of a nice trend in the fitness metric on the config space, and on the use of warmer/colder signals to speed convergence. I raised these in general terms but the facts you supplied demonstrated that the general issues apply very directly to ev. It's one thing to dismiss an in theory analysis -- as MG did over and over over several weeks -- it is another to answer to specific facts as you documented in say 126. Sadly, for years to come, the dismissive talking points will circulate in the evo mat fever swamps, where many will tank up and spread them out to those who will not know of this thread and the results that you have helped clinch over. GEM of TKIkairosfocus
April 22, 2011
April
04
Apr
22
22
2011
04:49 PM
4
04
49
PM
PDT
hi vjt/kf, I also ran the following Google search which turned up some reasonable hits: Google I've decided to hold off posting more on ev, give time for MathGrrl to catch up. I may post elsewhere and just provide links here. I didn't mean to take over the thread and would love to see it get back on topic re: specifications and your quite reasonable request (which I noticed she tried to just turn right back on you.) It's getting more and more difficult to take her seriously, but we'll see. I've also ordered some books that will help me learn how to code GA's. I'm quite weak on the math involved so if I can learn to code them it will really help. I'd love to one day write a program to test GA's, lol! This may have been posted before: http://www.evoinfo.org/ Follow the links to ev Ware, Weasel Ware and Minivida.Mung
April 22, 2011
April
04
Apr
22
22
2011
04:26 PM
4
04
26
PM
PDT
Mung Thanks very much for the links on genetic algorithms. I'll have a look at them.vjtorley
April 22, 2011
April
04
Apr
22
22
2011
03:50 PM
3
03
50
PM
PDT
MG (& ilk): Given the issues now on the table at the CSI news flash thread [cf here, esp on the maths issues and the revelations on fine tuning and targetting in Schneider's ev from Mung at 126 on . . . ], perhaps a bit of explanation is in order. GEM of TKIkairosfocus
April 22, 2011
April
04
Apr
22
22
2011
02:55 PM
2
02
55
PM
PDT
Mung: The declaration by Schneider in 134 contrasts not only with the general principles in the tutorial, but with his already excerpted remarks on tweaking, in your 126. I suspect, given his declaration on the race horse page that the UPB was beaten by NATURAL selection, that he does not realise just how much active information he has intelligently fed into the program, to tune it exactly as the GA manual says:
Genetic Algorithms are a family of computational models inspired by evolution. These algorithms encode a potential solution to a specific problem on a simple chromosome like data structure and apply recombination operators to these structures so as to preserve critical information.
What is really telling about the impact of these facts and the expose of the critical blunder by MG, however is the telling silence over the past few days. The shift in pattern of objections to CSI is all too revealing. I repeat: MG or someone else needs to do a fair bit of explaining for what has been so confidently declared in recent weeks here at UD. And, given the gap between the declared mathematical prowess and the actual performance, and the sort of underlying hostility and dismissive contempt -- remember the insinuations of dishonesty on our part -- that were plainly on the table, the explanation needs to be a very good one, with maybe a wee bit of an apology or retraction mixed in. GEM of TKIkairosfocus
April 22, 2011
April
04
Apr
22
22
2011
02:50 PM
2
02
50
PM
PDT
The comparisons between R_sequence and R_frequency suggest that the information at binding sites is just sufficient for the sites to be distinguished from the rest of the genome.
The Information Content of Binding Sites on Nucleotide Sequences Looks like a grand design!Mung
April 22, 2011
April
04
Apr
22
22
2011
12:49 PM
12
12
49
PM
PDT
The simulation begins with zero information and, as in naturally occurring genetic systems...
ev Abstract
Genetic Algorithms are a family of computational models inspired by evolution. These algorithms encode a potential solution to a specific problem on a simple chromosome like data structure and apply recombination operators to these structures so as to preserve critical information.
A Genetic Algorithm TutorialMung
April 22, 2011
April
04
Apr
22
22
2011
10:20 AM
10
10
20
AM
PDT
Tutorial and papers on GA's: http://www.cs.colostate.edu/~genitor/Pubs.htmlMung
April 22, 2011
April
04
Apr
22
22
2011
08:54 AM
8
08
54
AM
PDT
So earlier in this thread I posted a link to A Field Guide to Genetic Programming. Upon closer inspection (and as advertised) this is a guilde to genetic programming which is not the same as genetic algorithms. In genetic programming, the "genomes" are actual programs. This is quite unlike what is going on in ev. But as one author writes in a book I was reading last night:
A genetic algorithm doesn't simulate biological evolution; it merely takes inspiration from it. The algorithmic definition of fitness underscores this distinction.
Sometimes people who implement a program using a genetic algorithm loose sight of this very basic truth.Mung
April 22, 2011
April
04
Apr
22
22
2011
08:52 AM
8
08
52
AM
PDT
Onlookers (& MG and ilk): Fair comment: On developments over the past few days MG and ilk have a fair amount of explaining to do. Now, too, I have done some cleaning up and augmentation, especially putting in links. Note in particular that the 22-BYTE remark by MG was misread as 22 bits in an early calc. OOPS. Since 22 bytes is 176 bits, there is no material difference in the correction in 20. And -- given that she has challenged the sources of the numbers fed into the reduced Chi metric, in this case, MG herself is the source. Save, it turns out this is close to the upper limit of random text searches so far, which have shown empirically what is well known from decades ago from thermodynamicists. Namely a lab scale exploration on chance variation and trial and error is going to max out on a search of a space of about 10^50 or so possibilities. Mung's expose of the tweaking and tuning in ev during troubleshooting to get it to put out the desired horse race o/ps, also is very revealing on how Schneider apparently does not realise that his intelligent efforts are feeding in a lot of active info into the search he is carrying out. In short, we have a very simple explanation for why his horse race was won over the UPB: the search was intelligently directed and tuned to the particular circumstances. Posts in this thread are standing proof that intelligence routinely exceeds the UPB,based on knowledge, purpose, skill and if necessary tweaking (that spellcheck is very helpful to inexpert typists -- do typists still exist as a profession?). Happy Easter weekend. GEM of TKIkairosfocus
April 22, 2011
April
04
Apr
22
22
2011
02:11 AM
2
02
11
AM
PDT
Mung: Weasel did not have "breeding" like that either. But in such a later generation algor, it raises the question as to whether such a move would destabilise the progress, since ev seems to be so sensitive per the documentation you have given. (My guess is it would most likely have been tried. So if it is missing there is probably a reason why Schneider went asexual.) GEM of TKIkairosfocus
April 21, 2011
April
04
Apr
21
21
2011
02:55 PM
2
02
55
PM
PDT
More on ev: I have my doubts as to whether ev even qualifies as a genetic algorithm, I'll need to do some more reading. So what's missing? Crossover. So how does ev work? It works by replacing the "bad" individuals" with copies of the "good" individuals.
All creatures are ranked by their number of mistakes. The half of the population with the most mistakes dies; the other half reproduces to take over the empty positions. Then all creatures are mutated and the process repeats.
By "reproduces" he means:
Reproduce the bugs that made fewer mistakes by copying them on top of the ones that made more mistakes.
SPECIAL RULE: if the bugs have the same number of mistakes, reproduction does not take place. This ensures that the quicksort algorithm does not affect who takes over the population, and it also preserves the diversity of the population. Without this, the population is quickly taken over and evolution is extremely slow!
The reproduce methods: line 2823 of ev.p procedure reproduce(var e: everything); line 3369 of ev.c Static Void reproduce(e) line 436 of Simulator.java private void reproduce () Any questions?Mung
April 21, 2011
April
04
Apr
21
21
2011
02:47 PM
2
02
47
PM
PDT
Mathgrrl (#99) You write:
What do you find confusing about my scenarios? What’s wrong with referring to other papers?
I'm a philosopher, not a biologist. I know very little about evolutionary algorithms, and I detest jargon - I can't keep all the technical terms in my head at once. I have absolutely no inclination to wade through hundreds of pages of well-nigh-incomprehensible scientific papers in order to understand what you're getting at. If you want me to answer your questions, I'm afraid you'll have to (a) explain the cases you're describing, in non-technical language (i.e. something a bright 12-year-old could grasp), and (b) explain why you think they pose a threat to Professor Dembski's concept of complex specified information, in a summary of two or three pages at most. You're quite familiar with these cases, so I'm sure you can explain them in simple terms to a layman like me. I've spent dozens of hours trying to explicate the concept of CSI to you; now it's your turn to reciprocate. Otherwise, I simply can't help you. Sorry.vjtorley
April 21, 2011
April
04
Apr
21
21
2011
11:38 AM
11
11
38
AM
PDT
Point, Mung!kairosfocus
April 21, 2011
April
04
Apr
21
21
2011
10:45 AM
10
10
45
AM
PDT
So let's take a closer look at Schneider's Horse Race page and do a little quote mining.
A 25 bit site is more information than needed to find one site in all of E. coli (4.7 million base pairs). So it's better to have fewer bits per site and more sites. How about 60 sites of 10 bits each?
Tweak.
We are sweating towards the first finishing line at 9000 generations ... will it make it under 10,000? 1 mistake to go ... nope. It took to about 12679 generations. Revise the parameters:
Tweak.
It's having a hard time. Mistakes get down to about 61 and then go up again. Mutation rate is too high. Set it to 3 per generation.
Tweak.
Still having a hard time. Mistakes get down to about 50 and then go up again. Mutation rate is too high. Set it to 1 per generation.
Tweak.
3 sites to go, 26,300 generations, Rsequence is now at 4.2 bits!! So we have 4.2 bits × 128 sites = 537 bits. We've beaten the so-called "Universal Probability Bound" in an afternoon using natural selection!
And just a tad bit of intelligent intervention.
Dembski's so-called "Universal Probability Bound" was beaten in an afternoon using natural selection!
And a completely blind, purposeless, unguided, non-teleological computer program! Does Schneider even understand the UPB? Does he think it means that an event that improbable can just simply never happen?
Evj 1.25 limits me to genomes of 4096. But that makes a lot of empty space where mutations won't help. So let's make the site width as big as possible to capture the mutations. ... no that takes too long to run. Make the site width back to 6 and max out the number of sites at 200.
Tweak.
The probability of obtaining an 871 bit pattern from random mutation (without selection of course) is 10-262, which beats Dembski's protein calculation of 10-234 by 28 orders of magnitude. This was done in perhaps an hour of computation with around 100,000 generations.
HUH? With or without selection?
It took a little while to pick parameters that give enough information to beat the bound, and some time was wasted with mutation rates so high that the system could not evolve. But after that it was a piece of cake.
You don't say. MathGrrl @105
There is no target and nothing limits changes in the simulation.
There aare both targets and limits.Mung
April 21, 2011
April
04
Apr
21
21
2011
10:06 AM
10
10
06
AM
PDT
CSI Newsflash Progress Report: In post no 1 above, there is a tracking summary of major progress to date on this thread, including the overnight -- unfortunate -- developments with MG's credibility, and -- on a happier note -- the potentials for a breakthrough view of the GA as evidence of intelligently designed micro evolution. There is a special treat for BA 77 too, on maths. GEM of TKIkairosfocus
April 21, 2011
April
04
Apr
21
21
2011
08:00 AM
8
08
00
AM
PDT
F/N: Genetic [Exploration and Discovery] Algorithms This remark by Poli et al clipped above, has set me to thinking:
One of the awkward realities of many widely appli-cable tools is that they typically have numerous tunable parameters. Evo-lutionary algorithms such as GP are no exception . . . Some parameter changes, however, can produce more dramatic effects . . . .Differences as small as an ‘>’ in place of a ‘>/=’ in an if statement can have an important effect . . . [that] may influence the winners of tournaments [Side-note: . . . I hate ligatures!]
1 --> The algorithms, as reported, are finely tuned, functionally specific and complex. 2 --> So, if we see such an algorithm of adaptation to niches in a variable environment, it is a mark of --- DESIGNED-in adaptability, probably as built-in robustness. 3 --> Notice, too, just how sensitive they can be to slight variations in strategy: small changes giving rise to big differences. FINE-TUNING! 4 --> Therefore if such algorithms on computers are mimicking the performance of already functioning life forms on chance variations and environmental selection, then the known design of the algorithms points to . . . 5 --> You guessed it: designed in micro-level evolvability to fit niches and shifts in environments for life-forms. 6 --> In short, the evo mat view has been leading us to all -- including us design thinkers! -- look at exploration and discovery algors the wrong way around. 7 --> They are evidence that capacity of organisms to adapt to environments and changes (within reasonable limits) is yet another MARK OF DESIGN. (Yet another misreading of the significance of FSCI! And, we design thinkers were caught by it. Youch!) 8 --> In addition, we should note that as the Mandelbrot set thought exercise above showed,the peaks are not in the hill-climbing algor, they lie in the way that the performance characteristics are mapped to the variable configur-er stored pattern. 9 --> There have to be peaks, and there have to be nice, trend-y slopes pointing to the peaks in the islands of function mapped to the "configuration codes" for in this case the von Neumann self replicator facilities. [This is itself another mark of design.] 10 --> What the hill-climber algors do is explore and pick promising slopes in a context that presumes that nice trends don't usually lie. 11 --> This may discover and express well-adapted forms, but it did not create the peaks. In the design or random variation based hill climbing exploration algorithms, that may show us something we did not know was there before, but that is a matter of uncovering the hidden but already implied. 12 --> Sort of like how the exploration of a seemingly simple point-wise function on the complex plane by a visualisation algor shocked and amazed us by showing that behind Julia sets lurked a beautiful, astonishing figure of infinite complexity, the Mandelbrot set. 13 --> H'mm. A specific and complex entity with a simple description. Design, again. IN THE STRUCTURE OF MATHEMATICS. That is, of logical reality. 14 --> There may be something in the suggestion that the M-brot set is God's thumbprint in Mathematics, just like the Euler equation is a signature of the coherent elegant beauty of the cosmos, the ordered system of reality:
e^(i*p1) + 1 = 0
15 --> BA 77's gonna love this one: The ultimate simple specification of a system of infinite complexity and wonderful functionality!!!! 16 --> We may be opening up a new front in the design thought world here: a designed mathematical-logical order for the logic of structure of reality itself. GEM of TKIkairosfocus
April 21, 2011
April
04
Apr
21
21
2011
03:38 AM
3
03
38
AM
PDT
F/N: Dembski in Specification, 2005, on chance hyps: __________ p. 26: >> Probabilistic arguments are inherently fallible in the sense that our assumptions about relevant probability distributions might always be in error. Thus, it is always a possibility that {Hi}_iEI omits some crucial chance hypothesis that might be operating in the world and account for the event E in question. But are we to take this possibility seriously in the absence of good evidence for the operation of such a chance hypothesis in the production of E? Indeed, the mere possibility that we might have missed some chance hypothesis is hardly reason to think that such a hypothesis was operating. Nor is it reason to be skeptical of a design inference based on specified complexity. Appealing to the unknown to undercut what we do know is never sound epistemological practice. Sure, we may be wrong. But unknown chance hypotheses (and the unknown material mechanisms that supposedly induce them) have no epistemic force in showing that we are wrong. [remember, a brain in a vat world or a Matrix game world, or a Russellian world created in an instant in current state five minutes ago are possible hypotheses and would be empirically indistinguishable to the world we think we inhabit] Inquiry can throw things into question only by taking other things to be fixed. The unknown is not a fixed point. It cannot play such a role. >> ___________ Provisionality -- inescapable in science per Newton in Opticks, Query 31, 1704 -- should not be allowed to block an inference to best current explanation, if we are committed to scientific progress. As at now,
i: on reasonable assignment of a functionally specific [or otherwise target zoned] information value to Ip in the reduced form of the Chi metric, through ii: probability exercises per physically relevant probability estimation or hypothesis or iii: through observation of storage of information or iv: use of code strings or even v: decomposition of a Wicken wiring diagram of a functional object into a network list of nodes, arcs and interfaces, etc -- allows vi: substitution into the reduced Chi metric, and thence vii: comparison with a threshold serving as a limit beyond which viii: the relevant information is credibly too isolated in the Config space to credibly be ix: there by chance plus necessity without x: purposeful (though perhaps subtle and not consciously aware [think of the case of Clever Hans the horse]) injection of intelligent active information.
Thus, from Orgel and Wicken, we have a descriptive concept, Specified, often functional organised complexity. through Dembsky, Durston, Abel et al, we can deduce metrics. the Chi metric can be reduced to an information beyond a threshold of sufficient complexity form, which can be rationalised in different ways. This can then work with the explanatory filter to deliver an inference to best current explanation that warrants the provisional -- scientific conclusions are always provisional -- view that the relevant FSCI/CSI-bearing entity was most likely designed. On this, Durston's 35 families of proteins have several specific proteins that seem to be designed, and thus the genome and wider architecture of cell based life seems designed. This empirically and analytically grounded conclusion is plainly controversial, but it is inferred on best current explanation per warrant on empirical data and reasonable models. The CSI gang wins, on appeal. GEM of TKIkairosfocus
April 21, 2011
April
04
Apr
21
21
2011
01:05 AM
1
01
05
AM
PDT
F/N: The Field Guide to GP's that Mung linked has some interesting clips: _____________ c1, p. 147 of 250 [I will follow the pdf not the official p numbers] >> One of the awkward realities of many widely appli-cable tools is that they typically have numerous tunable parameters. Evo-lutionary algorithms such as GP are no exception . . . Some parameter changes, however, can produce more dramatic e?ects . . . . there are many small differences in GP implementations that are rarely considered important or even reported. However, our experience is that they may produce significant changes in the behaviour of a GP system. Di?erences as small as an ‘>’ in place of a ‘?’ in an if statement can have an important e?ect. For example, the substitution ‘>’ ? ‘?’ may in?uence the winners of tournaments, the designation of the best-of-run individual, the choice of which elements are cloned when elitism is used, or the o?spring produced by operators which accept the o?spring only if it is better or not worse than a parent.>> p. 16 of 250: >> GP ?nds out how well a program works by running it, and then comparing its behaviour to some ideal (line 3). We might be interested, for example, in how well a program predicts a time series or controls an industrial process. This com-parison is quantified to give a numeric value called fitness. Those programs that do well are chosen to breed (line 4) and produce new programs for the next generation (line 5). >> p. 17 of 250: Algorithm 1.1: Genetic Programming>> 1: Randomly create an initial population of programs from the available primitives (more on this in Section 2.2). 2: repeat 3: Execute each program and ascertain its fitness. 4: Select one or two program(s) from the population with a probability based on fitness to participate in genetic operations (Section 2.3). 5: Create new individual program(s) by applying genetic operations with specified probabilities (Section 2.4). 6: until an acceptable solution is found or some other stopping condition is met (e.g., a maximum number of generations is reached). 7: return the best-so-far individual.>> pp. 28 - 9 of 250: >> The most commonly employed method for selecting individuals in GP is tournament selection, which is discussed below, followed by ?tness- proportionate selection, but any standard evolutionary algorithm selection mechanism can be used. In tournament selection a number of individuals are chosen at random from the population. These are compared with each other and the best of them is chosen to be the parent. When doing crossover, two parents are needed and, so, two selection tournaments are made. Note that tourna-ment selection only looks at which program is better than another. It does not need to know how much better. This e?ectively automatically rescales ?tness, so that the selection pressure 4 on the population remains constant . . . tournament selection ampli?es small di?erences in ?tness to prefer the bet-ter program even if it is only marginally superior to the other individuals in a tournament. >> ________________ In short the programs work WITHIN an island of established, function, use controlled random changes to explore which way lieth the steepest ascent to a hill top, and employ artificial selection. Notice,t eh creitical dependence on smoothly trending fitness functions and the sensitivity to tuning of parameters. Let's translate this last: the GP itself is on an island of function, and small disruptions can make big differences in performance. They inherently model at most microevo, not the origin of body plan level macroevo that is required to explain the real challenge of darwinian type blind watchmaker evolution. GEM of TKI PS: Notice where the phrase the Blind Watchmaker comes from, for next time someone rhetorically pretends not to know what it means or where it comes from: Dawkins, in a book bearing that phrase as a key part of its title, and presenting Weasel at a key point in his argument.kairosfocus
April 21, 2011
April
04
Apr
21
21
2011
12:24 AM
12
12
24
AM
PDT
Mung: I come from an island that has a mountain range down its middle, with branch ranges and side hills. I now live in another, that has three main volcanic edifices and some side-hills. So, I am quite aware of having multiple peaks reachable by hill-climbing. (Oddly, I have never been to Blue Mountain peak, nor have I ever taken the local Centre Hills tour! I am much more inclined to head for a beach, rod in hand . . . ) Hill-climbing algorithms of course can explain multiple niches within an island of function, especially if the populations are allowed to wander off and head for different hills. The root problem from my point of view, is that the whole procedure starts on such an island. That was evident to me from the very first when I saw Weasel's "nonsense phrases" being rewarded on mere proximity to target. Might as well quote Dawkins, from the notorious TBW [cf App 7, my always linked], which will make all clear: _________________ >> I don't know who it was first pointed out that, given enough time, a monkey bashing away at random on a typewriter could produce all the works of Shakespeare. [NB: cf here and this discussion on chance, necessity and intelligence.] The operative phrase is, of course, given enough time. Let us limit the task facing our monkey somewhat. Suppose that he has to produce, not the complete works of Shakespeare but just the short sentence 'Methinks it is like a weasel', and we shall make it relatively easy by giving him a typewriter with a restricted keyboard, one with just the 26 (capital) letters, and a space bar. How long will he take to write this one little sentence? . . . . It . . . begins by choosing a random sequence of 28 letters ... it duplicates it repeatedly, but with a certain chance of random error – 'mutation' – in the copying. The computer examines the mutant nonsense phrases [= non-functional], the 'progeny' of the original phrase, and chooses the one which, however slightly, most resembles the target phrase [notice, explicit targetted search, more modern programs generate implicit targetting on nice trend-y slopes,through topography-"fitness" defining functions and associated hill-climbing algorithms], METHINKS IT IS LIKE A WEASEL . . . . What matters is the difference between the time taken by cumulative selection, and the time which the same computer, working flat out at the same rate, would take to reach the target phrase if it were forced to use the other procedure of single-step selection [the problem is that the real challenge of body-plan origination is credibly much bigger than the single steps that Dawkins would dismiss, starting with that needed to account for the joint metabolic action and von Neumann self replicator in origin of life, also cf here above on the Mandelbrot set "fitness function" thought exercise on the weaknesses of other more subtle GA's that load the target implicitly, and this follow up, on what is really going on in the evolutionary computing models. Cf. here at 89 above for a comment on the place where Schneider mistakenly discusses his artificial selection algorithm as though it were natural selection]: about a million million million million million years. This is more than a million million million times as long as the universe has so far existed . . . . Although the monkey/Shakespeare model is useful for explaining the distinction between single-step selection and cumulative selection, it is misleading in important ways. One of these is that, in each generation of selective 'breeding', the mutant 'progeny' phrases were judged according to the criterion of resemblance to a distant ideal target, the phrase METHINKS IT IS LIKE A WEASEL. Life isn't like that. Evolution has no long-term goal. There is no long-distance target, no final perfection to serve as a criterion for selection, although human vanity cherishes the absurd notion that our species is the final goal of evolution. In real life, the criterion for selection is always short-term, either simple survival or, more generally, reproductive success. [TBW, Ch 3, as cited by Wikipedia, various emphases added.] >> __________________ Weasel is of course rather like a Model T Ford, an old technology, long since replaced by more sophisticated versions. But despite Dawkins' weasel words, it primarily served to misleadingly persuade the naive that the question of the origin of functionally specific, complex information, had been "scientifically" answered through evolution by chance variation and natural selection. Indeed, I have had people present Weasel to me in that guise in recent times. GEM of TKIkairosfocus
April 20, 2011
April
04
Apr
20
20
2011
11:47 PM
11
11
47
PM
PDT
The file evjava.zip can be downloaded from this page. The Pascal and C code can be downloaded from this page.Mung
April 20, 2011
April
04
Apr
20
20
2011
07:44 PM
7
07
44
PM
PDT
MathGrrl @41
My participation here is solely so that I can understand CSI well enough to be able to test whether or not known evolutionary mechanisms can create it.
Mung @76
Schneider claims to have created it. Do you doubt him?
Yes? No? You don't know just what ev does so you don't have an answer? MathGrrl @66
Again, I don’t see where you’re getting your 266 bit value, but Schneider shows how to generate arbitrary amounts of Shannon information via ev.
How? MathGrrl @68
As noted above, Schneider shows how to generate arbitrary amounts of Shannon information via ev.
Mung @85
Now apart from the fact that this is a vague and muddled statement (couldn’t a random number generator just as well generate arbitrary amounts of Shannon information?) Schneider himself actually makes no such claim.
MathGrrl @100
If you think I’ve misrepresented Schneider, please explain exactly how, with reference to the page to which I linked.
Schneider never makes the claim that ev can "generate arbitrary amounts of Shannon information." It's hard for me to link to something he never said. MathGrrl @101
It seems that you haven’t understood Schneider’s summary.
Well, unfortunately for you, I did more than read the summary. I read the entire web page, which detailed all the hoops Schneider had to go through just to get his ev program to "win" the horse race. Did you happen to notice the failed attempts, each followed by intelligent intervention? I also read his paper on ev, and have looked at the source code (Pascal, C and Java) and the source code comments. I think I know enough about it to have some idea whether or not you know enough about it to be making the claims that you are. For example, you say:
Unless you’re claiming that it is impossible in principle to model evolutionary mechanisms, these GAs support the idea that known evolutionary mechanisms are capable of changing the allele frequency in a population such that subsequent generations are better able to reproduce in a particular environment.
Guess what's missing from ev? A model of population genetics, that's what. IOW, no alleles.
GAs model known evolutionary mechanisms. These mechanisms work without intelligent intervention...
Read the horserace page again. Schneider intervened.
That is not an accurate description of ev. There is no target and nothing limits changes in the simulation.
There are multiple targets. So in some bizarre sense someone could argue you got that one right. And there are a number of limits to changes in the simulation.
That’s what makes the results particularly interesting. I strongly recommend reading Schneider’s paper.
You really should. You don't know what you're talking about.Mung
April 20, 2011
April
04
Apr
20
20
2011
07:13 PM
7
07
13
PM
PDT
MathGrrl @99
What do you find confusing about my scenarios? What’s wrong with referring to other papers?
Let's review (please bear with me): MathGrrl @41
My participation here is solely so that I can understand CSI well enough to be able to test whether or not known evolutionary mechanisms can create it.
PaV @50
First, why do you think “Produces at least X amount of protein Y” is a “specification”.
vjtorley @52
I think that kairosfocus’ posts at #44, #45 and #47 above meet your requirements for a CSI calculation for the four scenarios you described. But if you aren’t satisfied with those answers, then here’s a challenge I shall issue to you. Please provide us with a two- or three-page, detailed but completely jargon-free description of the four scenarios you are describing and post it up on UD. No references to other papers by biologists, please. Describe the problems in your own words, as you would to a non-biologist (which is what I am).
MathGrrl @67
As noted in my guest thread, specification is one of the more vague aspects of CSI. Some ID proponents don’t seem to have a problem with the specification I suggested (see a couple of the comments above in this thread, for example). Others, like you, seem to have a different concept. Why do you think that “Produces at least X amount of protein Y” is not a specification in Dembski’s sense? Please reference his published descriptions of CSI that support your view.
vjtorley @72
You have yet to respond to my challenge regarding the four scenarios you describe: Please provide us with a two- or three-page, detailed but completely jargon-free description of the four scenarios you are describing and post it up on UD. No references to other papers by biologists, please. Describe the problems in your own words, as you would to a non-biologist (which is what I am). Then I might be able to help you. I’m still waiting.
MathGrrl @99
What do you find confusing about my scenarios? What’s wrong with referring to other papers?
Nuff said? MathGrrl, please tell us what you think a specification is. Then please tell us why you think what you provided in your "challenges" qualify as specifications.Mung
April 20, 2011
April
04
Apr
20
20
2011
06:31 PM
6
06
31
PM
PDT
I hate it when I purchase a book and then find out it's available free online, lol. A Field Guide to Genetic ProgrammingMung
April 20, 2011
April
04
Apr
20
20
2011
05:07 PM
5
05
07
PM
PDT
1 2 3 4 5 7

Leave a Reply