On Mon, 18 Aug 2003, Axel Harvey wrote:
> I presume you have seen the bet in Gerard 't Hooft's web pages. It
> seems to be aimed squarely at the GCP. He even links your site.
>
> http://www.phys.uu.nl/~thooft/parabet.html
> http://www.phys.uu.nl/~thooft/paraplue.html
>
> (The link to the GCP site is reached from the latter URL. Go to the
> bottom of that page and click on "scientifically detected".)
>
> Maybe you have mentioned this already in your website or emails -
> sorry if I've missed it. What do you think?
I appreciate Gerard 't Hooft's interest and his willingness
to take the time to think about the GCP database and results.
It is a large corpus and thus it is perhaps understandable that
he is not (apparently) aware of some aspects of the database
that are relevant, and has incorrect impressions about the
types of events we assess.
I had not seen his web page yet, though I have been in contact
with 't Hooft once directly and other times indirectly by way
of my colleague Dick Bierman. Thus I am aware of his basic idea.
It is not a bad proposal in some respects, but it is ungainly and
demonstrably unnecessary for a substantial proportion of the GCP
database, even if one believes there are grounds for his concern.
In addition, because he is looking for 5 sigma, it would, according
to our experience (and statistical power estimates), require on
the order of 100 events (about 2.5 years at our average rate) to
have even .5 beta of reaching such a level. His plan also requires
that everybody be blind to the data and that "An experiment must
be analyzed in one single procedure, immediately after disclosure
of its results." Together, if I understand him correctly, these
conditions would require that we run the experiment blind for
two or three years.
I have suggested to 't Hooft an alternative procedure that is
also immune to the problem he is concerned about, namely:
"In this particular [GCP] case, the point is that whatever
'emotionally distressing events' are, is defined while the
experiment takes place."
That statement or description is not true for a substantial proportion
of the GCP events that constitute the experiment. I suggest
that for cases that are specified a priori, that is, before
the event takes place, none of the concerns or preferred
"explanations" in 't Hooft's description are valid. Thus, I
propose that the results for the a priori cases be examined
and compared with those that are necessarily specified after
the fact. This is a simple and instructive test of the
viability of 't Hooft's criticism.
About 45% of the formal predictions can be and are established
before the event. For example, we have each year made predictions
about the New Year transition period. Obviously, these can be
made ahead of time. Similarly, for web- or media-organized global
peace marches, or major religious gatherings or meditation events,
we know the publicized,
planned time and can make an a priori prediction. There actually
are a considerable number of potentially interesting, often
unique events for which such a priori specification is possible:
The moment of the Dow Jones Industrial Average first breaking
the 10,000 barrier; the major solar eclipse over Europe and the Middle
East in August 1999;
openings of the Olympic games; the Kumbh Mela in India; Earth
Day celebrations; and so on. We have as of August 2003 a total
of 147 formal predictions and tests of the general hypothesis that there
will be departures from expected random behavior in the GCP data that
are correlated with major global events. Of the 147 events or cases,
67 are defined by a priori predictions, and 80
are post facto because
they address accidents, explosions, disasters, events of war,
etc., which cannot be subject to a priori specification.
So, 't Hooft's description, above, is not and cannot be
relevant to a substantial number of cases in the GCP database.
The question is whether these cases show less evidence for
the correlations we see overall in the database. The answer
is no -- indeed they present a somewhat more persuasive bottom line.
The a priori cases have a smaller mean probability
(0.354 vs 0.410 for the post facto subset), and despite the
smaller database size, they actually have a more extreme composite
p-value (1x10e-6 vs 9x10e-4 for the post facto subset).
Going beyond this simple demonstration, Dick Bierman and I
are discussing and preparing to test a procedure that meets,
even for the post hoc cases, the standards 't Hooft proposes,
while also allowing the project to proceed with little
interference. Although I don't intend to engage in the
Randi-style "bet" that 't Hooft uses to express his
depth of conviction, I will keep him informed.
|