Re: Mersenne: Re: Factoring benefit/cost ratio

2001-12-05 Thread Steve Harris

Richard,

Your first interpretation of "verified" residues is correct, they are
retested until two residues match. Any time a double-check reports in a
residue which is different from the first LL test, the exponent is returned
to the database to be tested again. This means that at least one of the
residues is incorrect, and happens (relatively) often, I believe about two
percent of the time.  However, as has been pointed out before, the odds of
two LL tests on different machines returning the _same_  incorrect residues
are astronomical (although, of course, still non-zero).

Steve

-Original Message-
From: [EMAIL PROTECTED] <[EMAIL PROTECTED]>
To: [EMAIL PROTECTED] <[EMAIL PROTECTED]>
Cc: [EMAIL PROTECTED] <[EMAIL PROTECTED]>
Date: Wednesday, December 05, 2001 8:34 PM
Subject: Re: Mersenne: Re: Factoring benefit/cost ratio


>Brian,
>
>I'm wondering whether we may be misunderstanding each other's
>contentions here.  I thought you object to at least some of what I
>claimed, but now it seems that you're presenting arguments and
>evidence that support what I'm claiming.
>
>Since my previous postings may have had careless wording or otherwise
>obscured my intentions, and I did not earlier realize the importance
>of certain details to the discussion, let me restate what I've meant
>to claim:
>
>1. It is more valuable to know a specific factor of a Mnumber than to
>know that that Mnumber is composite without knowing any specific
>factor.
>
>(There's little dispute about #1.)
>
>2. Claim #1 is true not only from the viewpoint of mathematics in
>general, but also from the narrower viewpoint of the GIMPS search for
>Mersenne primes.
>
>3. One (but not the only) justification for claim #2 is that, _in
>current practice_, a composite status derived by GIMPS from finding a
>specific factor is (slightly) more reliable than a composite status
>derived by GIMPS from matching nonzero residues from Lucas-Lehmer
>tests.
>
>That is, although in theory, or ideally, those two methods of
>determining compositeness are equally reliable, there currently exists
>a slight difference in reliability, in favor of the factor, from a
>practical standpoint.
>
>4. Our experience ("the record"), as documented in the Mersenne
>mailing list or GIMPS history, supports claim #3.
>
>- - - - -
>
>Brian Beesley wrote:
>>> > AFAIK our record does _not_ show any such thing.
>>>
>>> Oh?  It doesn't?
>>
>> There is no evidence of any verified residuals being incorrect.
>
>Wait a second -- just yesterday you wrote that you had "triple-checked
>thousands of small exponents" (which means they had already been
>double-checked) and that "A very few (think fingers of one hand)
>instances of incorrectly matched residuals have come to light -
>completing the double-check in these cases proved that one of the
>recorded residuals was correct".
>
>So it seems that the meaning you're assigning to "verified" is
>something like "retested and retested until two residuals match".
>Is that a correct interpretation?  If not, what is?
>
>My claim #3 means that in practice, factors require fewer verification
>runs to produce matching results than do L-L residues, on average.
>Do you disagree with that?  If not, then don't we agree about claim
>#3?
>
>Furthermore, my claim #4 means that the demonstration that factors
>require fewer verification runs to produce matching results than do
>L-L residues, on average, rests on the observed history _including the
>paragraph you wrote from which I just quoted above!_  Do you disagree?
>
>Also, in that same paragraph you wrote, "... - some of the ones where
>the accepted residual was recorded to only 16 bits or less, which
>makes the chance of an undetected error _much_ greater (though still
>quite small) ..."  Am I correct in interpreting this to mean that you
>think that using 64-bit residuals is more reliable than using 16-bit
>residuals?  If so, then surely you'll grant that 256-bit residuals
>would be even more reliable yet, meaning that there's still room for
>error in our practice of using 64-bit residuals.  But a specific
>factor
>is a _complete value_, not some truncation, and so its reliability is
>not damaged by the incompleteness which you admit keeps the L-L
>residues from being totally reliable - right?
>
>Then you wrote "so far no substantive errors in the database have come
>to light", but seemingly contradicted that in the very next sentence,
>"A very few (think fingers of one hand) instances of incorrectly
>matched residuals have come to light - completing the double-check in
>these cases proved that one of the recorded residuals was correct."
>... And thus _the other_ recorded residual was _incorrect_.
>
>> Neither is there any evidence that any verified factors are
>incorrect.
>
>Depends on the meaning of "verified", of course.  :-)
>
>Will Edgington (I think) has reported finding errors in his factor
>data base ... even though he verifies factors before adding them.
>Mistakes happen.  But I think the

Re: Mersenne: New exponents

2001-12-05 Thread George Woltman

At 07:15 PM 12/5/2001 -0800, Mary Conner wrote:
> > No.  The server never contacts the client.  That's too much of a security
> > risk in my book.
>
>That isn't exactly what I meant.  Given that his exponent has been expired
>and assigned to me, if he then checks in later to report further progress
>on the exponent, will the server tell his client that the exponent has
>been expired and assigned to someone else, or will it tell me the next
>time I check in that he is still working on it?

If he checks in his result, primenet will return error 11 - but your 
computation
will continue.

>Or do both of our clients continue happily chugging away on it?

I think prime95 removes it from worktodo.ini only if you have not
started the LL test.

Obviously there is some room for improvement here.  The current scheme
works OK for first time checks since yours would then become a valid
double-check.

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: New exponents

2001-12-05 Thread Mary Conner



On Wed, 5 Dec 2001, George Woltman wrote:
> >On a purely technical note,  In the event that the other person does
> >eventually check back in, is there a mechanism in place to either tell his
> >machine or mine that it should abandon the exponent
> 
> No.  The server never contacts the client.  That's too much of a security
> risk in my book.

That isn't exactly what I meant.  Given that his exponent has been expired
and assigned to me, if he then checks in later to report further progress
on the exponent, will the server tell his client that the exponent has
been expired and assigned to someone else, or will it tell me the next
time I check in that he is still working on it?  Or do both of our clients
continue happily chugging away on it?


_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: New exponents

2001-12-05 Thread Nathan Russell

At 08:40 PM 12/5/2001 -0500, George Woltman wrote in reply to Mary Conner:
>>On a purely technical note,  In the event that the other person does
>>eventually check back in, is there a mechanism in place to either tell his
>>machine or mine that it should abandon the exponent
>
>No.  The server never contacts the client.  That's too much of a security
>risk in my book.

However, when the client does contact the server (every 28 days by default, 
IIRC), will it not get an "this assignment does not belong to us"?

I know that I had that happen while I had QA and primenet work queued on 
the same machine, and in fact it happened often enough to be rather annoying.

Nathan

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: Re: Factoring benefit/cost ratio

2001-12-05 Thread ribwoods

Brian,

I'm wondering whether we may be misunderstanding each other's
contentions here.  I thought you object to at least some of what I
claimed, but now it seems that you're presenting arguments and
evidence that support what I'm claiming.

Since my previous postings may have had careless wording or otherwise
obscured my intentions, and I did not earlier realize the importance
of certain details to the discussion, let me restate what I've meant
to claim:

1. It is more valuable to know a specific factor of a Mnumber than to
know that that Mnumber is composite without knowing any specific
factor.

(There's little dispute about #1.)

2. Claim #1 is true not only from the viewpoint of mathematics in
general, but also from the narrower viewpoint of the GIMPS search for
Mersenne primes.

3. One (but not the only) justification for claim #2 is that, _in
current practice_, a composite status derived by GIMPS from finding a
specific factor is (slightly) more reliable than a composite status
derived by GIMPS from matching nonzero residues from Lucas-Lehmer
tests.

That is, although in theory, or ideally, those two methods of
determining compositeness are equally reliable, there currently exists
a slight difference in reliability, in favor of the factor, from a
practical standpoint.

4. Our experience ("the record"), as documented in the Mersenne
mailing list or GIMPS history, supports claim #3.

- - - - -

Brian Beesley wrote:
>> > AFAIK our record does _not_ show any such thing.
>>
>> Oh?  It doesn't?
>
> There is no evidence of any verified residuals being incorrect.

Wait a second -- just yesterday you wrote that you had "triple-checked
thousands of small exponents" (which means they had already been
double-checked) and that "A very few (think fingers of one hand)
instances of incorrectly matched residuals have come to light -
completing the double-check in these cases proved that one of the
recorded residuals was correct".

So it seems that the meaning you're assigning to "verified" is
something like "retested and retested until two residuals match".
Is that a correct interpretation?  If not, what is?

My claim #3 means that in practice, factors require fewer verification
runs to produce matching results than do L-L residues, on average.
Do you disagree with that?  If not, then don't we agree about claim
#3?

Furthermore, my claim #4 means that the demonstration that factors
require fewer verification runs to produce matching results than do
L-L residues, on average, rests on the observed history _including the
paragraph you wrote from which I just quoted above!_  Do you disagree?

Also, in that same paragraph you wrote, "... - some of the ones where
the accepted residual was recorded to only 16 bits or less, which
makes the chance of an undetected error _much_ greater (though still
quite small) ..."  Am I correct in interpreting this to mean that you
think that using 64-bit residuals is more reliable than using 16-bit
residuals?  If so, then surely you'll grant that 256-bit residuals
would be even more reliable yet, meaning that there's still room for
error in our practice of using 64-bit residuals.  But a specific
factor
is a _complete value_, not some truncation, and so its reliability is
not damaged by the incompleteness which you admit keeps the L-L
residues from being totally reliable - right?

Then you wrote "so far no substantive errors in the database have come
to light", but seemingly contradicted that in the very next sentence,
"A very few (think fingers of one hand) instances of incorrectly
matched residuals have come to light - completing the double-check in
these cases proved that one of the recorded residuals was correct."
... And thus _the other_ recorded residual was _incorrect_.

> Neither is there any evidence that any verified factors are
incorrect.

Depends on the meaning of "verified", of course.  :-)

Will Edgington (I think) has reported finding errors in his factor
data base ... even though he verifies factors before adding them.
Mistakes happen.  But I think the error rate for factors has been
significantly lower than for L-L residuals.

> Whatever theory states, the experimental evidence is that verified
> factors are no more (or less) reliable than verified LL tests.

Then why don't we triple-check factors as often as we triple-check L-L
results?  Oh, wait ... depends on the meaning of "verified", again.

> Suppose a taxi firm runs 10 Fords and 10 Hondas for a year.
[ snip ]

Let's have an example in which the number and nature of the units is
closer to the gigabytes of data items we're slinging around.

> Besides which, checking a factor takes a few microseconds, whereas
> checking a LL test likely takes a few hundred hours.

... which tends to support my claim #3.

> If anything goes wrong during a factoring run, we would be far more
> likely to miss a factor which we should have found rather than vice
> versa.

... which is in agreement with claim #3.

>> How does that compare t

Re: Mersenne: New exponents

2001-12-05 Thread George Woltman

At 02:40 PM 12/5/2001 -0800, Mary Conner wrote:
> > David Slowinski discovered that M1257787 was prime
> > - when George's own computer was only a few days from finishing that
> > very exponent!

One other "what could have been" note.

I owned two computers at the time.The P-90 was testing 1257xxx and
the PPro-200 was testing 1258xxx.  If only I'd assigned the ranges the
other way around :)


_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: New exponents

2001-12-05 Thread George Woltman

At 02:40 PM 12/5/2001 -0800, Mary Conner wrote:
> > To make matters worse, Slowinski delayed the announcement of the prime
> > as he "was out of town" for a while - which turned out to be the better
> > part of half a year. I think I would have flipped. Living for half a
> > year with such a freak incident on your mind and not even being able to
> > tell anyone.
> > OTOH, it's an impressive example of how to keep a secret.

Keeping the secret was easy.  I had promised David.

>Ooof, so if Slowinski had gone out of town without contacting George, or
>had contacted someone else for independant verification, George's computer
>would have found the prime, and could have announced the discovery before
>Slowinski returned, and boy, wouldn't that have been a huge snarl over who
>got credit.

Probably not.  One of the reasons we tell past Mersenne discoverers
immediately is to "stake our claim".  I know David contacted Richard
Crandall immediately and he probably contacted others.

If had not contacted me directly, when my computer found M#34 and
I then relayed the exciting news to Dr. Crandall, he would have given me
the bad news.

Slightly off topic, someone at the time (a non-GIMPSer) said no way I just
missed out on M#34.  The odds must be one in a million.  So I calculated that
 a) given there is a Mersenne prime in the 1.2 millions, and
 b) given that a Cray found it, and
 c) given that there were about 40 GIMPS Pentiums working in
 the 1.2 million area at the time, and
 d) given the time it takes the average Pentium to do an LL test (I've
 since forgotten that value), then
What is the chance that the two competing groups would both find the
Mersenne prime within a week of each other?

The answer turned out to be 3%.  Not likely, but a far cry from 1 in a million.

>On a purely technical note,  In the event that the other person does
>eventually check back in, is there a mechanism in place to either tell his
>machine or mine that it should abandon the exponent

No.  The server never contacts the client.  That's too much of a security
risk in my book.

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: Re: Factoring benefit/cost ratio

2001-12-05 Thread bjb
On 5 Dec 2001, at 6:09, [EMAIL PROTECTED] wrote:

> Brian Beesley wrote:
> > On 3 Dec 2001, at 20:38, [EMAIL PROTECTED] wrote:
> [... snip ...]
> > > I think our record shows that a verified factor is still
> > > slightly (by a minute but nonzero margin) more reliable an
> > > indicator of compositeness than two matching nonzero LL
> > > residues.
> >
> > AFAIK our record does _not_ show any such thing.
> 
> Oh?  It doesn't?

There is no evidence of any verified residuals being incorrect.  Neither is there any evidence that any verified factors are incorrect.
Whatever theory states, the experimental evidence is that verified  factors are no more (or less) reliable than verified LL tests.

Suppose a taxi firm runs 10 Fords and 10 Hondas for a year. None  of them break down. On that basis alone, there is no evidence  whatsoever that one make is more reliable than the other.  Naturally, other companies' experimental evidence may vary.
>
> [ big snip ]
> There is a small chance that we may accept an incorrect factor even
> after double-checking it, but that chance is even smaller than the
> small chance that we may accept an incorrect double-checked L-L
> residual.

I doubt very much that we would accept an incorrect factor. The  double-checking is done with completely different code. Besides  which, checking a factor takes a few microseconds, whereas  checking a LL test likely takes a few hundred hours.

If anything goes wrong during a factoring run, we would be far more  likely to miss a factor which we should have found rather than vice  versa. This is relatively unimportant from the point of view of finding  Mersenne primes; the effect is a small loss of efficiency.

> How does that compare to the observed rate of incorrect factors
> discovered after triple-checking _them_?

AFAIK no-one bothers to triple-check factors. Nowadays factors  are verified on the server at the time they're reported. I'm not privy  to the server logs, so I simply don't know how many get rejected  (except for the server _database_ problem reported recently - not  related to the actual factor checking code - which causes problems  with genuine factors > 32 digits in length). However, I can think of  at least one way of pushing up the rejection rate.
> 
> How many of those problems caused errors during L-L verifications,
> and how many caused errors during factor verifications?
> 
All during LL first tests, or QA runs (which are LL & DC done in  parallel with intermediate crosscheck points).

> However, you may not have spent anywhere near as much time doing
> factor verifications as you have doing L-L verifications, so it may
> not be valid to draw any conclusion about comparative error rates on
> your system.

I've spent no time at all verifying factors - it would take less than a  minute to verify everything in the factors database. The total  factoring effort I've put in (ignoring ECM & P-1 on small exponents)  is only about 3% of my total contribution, so I would expect not to  have had any factoring errors. Besides which, trial factoring  _should_ have a lower error rate than LL testing, due to the lower  load on the FPU (which is usually the CPU element most sensite  to excess heat) and the smaller memory footprint (less chance of  data getting clobbered by rogue software or random bit-flips).


Regards
Brian Beesley
_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers


Re: Mersenne: New exponents

2001-12-05 Thread Mary Conner



On Wed, 5 Dec 2001, Alexander Kruppa wrote:
> David Slowinski contacted George, asking him wether Prime95 could test
> numbers >1 million bits. He had just discovered that M1257787 was prime
> - when George's own computer was only a few days from finishing that
> very exponent! David also asked for an independet verification of the
> prime, so suddenly the LL run on George's computer that had almost
> discovered GIMPS' first prime was no more than a double check for
> David's success.
> 
> To make matters worse, Slowinski delayed the announcement of the prime
> as he "was out of town" for a while - which turned out to be the better
> part of half a year. I think I would have flipped. Living for half a
> year with such a freak incident on your mind and not even being able to
> tell anyone.
> OTOH, it's an impressive example of how to keep a secret.

Ooof, so if Slowinski had gone out of town without contacting George, or
had contacted someone else for independant verification, George's computer
would have found the prime, and could have announced the discovery before
Slowinski returned, and boy, wouldn't that have been a huge snarl over who
got credit.  Not to mention that someone other than George or Slowinski
could have found it too.

On a purely technical note, I've just been assigned an exponent that
expired from someone else.  I looked at the active exponent report from
just before it was expired, and it looks like the exponent was about one
third done before it expired.  In the event that the other person does
eventually check back in, is there a mechanism in place to either tell his
machine or mine that it should abandon the exponent (it's a double check
LL assignment, so two tests don't need to be run), or would we just end up
both continuing on the same exponent, ending with it triple checked?  
Given how long this person had the exponent compared to the progress on
it, it is likely that I would finish long before him anyway.



_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: New exponents

2001-12-05 Thread Alexander Kruppa

[EMAIL PROTECTED] wrote:
> 
> On 4 Dec 2001, at 17:59, George Woltman wrote:
> 
> > >Case 1: I finish first, find a prime and announce my discovery. I did
> > >the work but the exponent is assigned to you! Who gets the
> > >credit???
> >
> > You, get the credit.  User b will be mighty disheartened.  I know first hand.
> > Slowinski's Cray beat my own Pentium-90 by just a few days in the discovery
> > of M#34.
> 
> Ooof. I didn't know about that.

I read about this on the mailing list archives. The info was scattered
over several posts, hopefully I recall everything correctly, and that
all the story bits belong to the same Mersenne prime! I hope George will
correct me if I'm wrong.

David Slowinski contacted George, asking him wether Prime95 could test
numbers >1 million bits. He had just discovered that M1257787 was prime
- when George's own computer was only a few days from finishing that
very exponent! David also asked for an independet verification of the
prime, so suddenly the LL run on George's computer that had almost
discovered GIMPS' first prime was no more than a double check for
David's success.

To make matters worse, Slowinski delayed the announcement of the prime
as he "was out of town" for a while - which turned out to be the better
part of half a year. I think I would have flipped. Living for half a
year with such a freak incident on your mind and not even being able to
tell anyone.
OTOH, it's an impressive example of how to keep a secret.

Alex
_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



RE: Mersenne: HD crash

2001-12-05 Thread Aaron Blosser

Best idea is to look at your account status page and find the exponents
for that machine on there.

I made the mistake of rebuilding my laptop the other day and while I had
backed up everything else, I forgot to backup the directory with
ntprime. Argh... fortunately it wasn't too far along on the current
exponent.

Just rebuilt the worktodo.ini file with the appropriate test=,xx
lines (optionally add the ,1 on the end if it had already completed the
p-1 factoring) and let 'er rip.

At least on my machines at home I try to be better about at least making
weekly backups of the prime directories on there.  For my machines at
work, they don't suffer the same "wipe and rebuild" fate as many of my
home test machines, so I don't bother backing them up much, which of
course is why I lost the stuff on my work laptop. :)

Aaron

> -Original Message-
> From: [EMAIL PROTECTED]
[mailto:mersenne-invalid-
> [EMAIL PROTECTED]] On Behalf Of Jud McCranie
> Sent: Wednesday, December 05, 2001 12:35 PM
> To: [EMAIL PROTECTED]
> Subject: Mersenne: HD crash
> 
> My hard drive crashed, and I have almost certainly lost all of the
GIMPS
> data for the exponent I was working on and 4 more I had in the queue.
The
> initial trial factorization had been done on all of them and the first
one
> was just about 4 days from completion.  What should I do about these
lost
> exponents? (I don't know which ones they were)

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Mersenne: HD crash

2001-12-05 Thread Jud McCranie

My hard drive crashed, and I have almost certainly lost all of the GIMPS 
data for the exponent I was working on and 4 more I had in the queue.  The 
initial trial factorization had been done on all of them and the first one 
was just about 4 days from completion.  What should I do about these lost 
exponents? (I don't know which ones they were)

+-+
| Jud McCranie|
| |
|"Thought I saw angels, but I could have been wrong." Ian Anderson|
+-+


_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: New exponents

2001-12-05 Thread bjb

On 4 Dec 2001, at 17:59, George Woltman wrote:

> >Case 1: I finish first, find a prime and announce my discovery. I did
> >the work but the exponent is assigned to you! Who gets the
> >credit???
> 
> You, get the credit.  User b will be mighty disheartened.  I know first hand.
> Slowinski's Cray beat my own Pentium-90 by just a few days in the discovery
> of M#34.

Ooof. I didn't know about that.
> 
> As to legal issues, the disclaimer section of the download page states:
> 
> We are not responsible for lost prize money, fame, credit, etc. should 
> someone accidentally or maliciously test the number you are working on and 
> find it to be prime.

The case I describe might not fall under the legal definition of 
"accidental" or "malicious". It would be possible for one or other of 
the parties to argue that they have been assigned the work by 
PrimeNet and that PrimeNet was therefore liable for lost prize 
money. On the "belt and braces" principle I think that the wording 
should be reinforced.

As to people working independently - I don't see how you can cover 
that one, since the independent party will not be covered by the 
disclaimer if they never downloaded any variant of Prime95. In that 
case it _is_ accidental that PrimeNet should assign work which is 
being replicated elsewhere without its knowledge.

Nowadays it's at least a reasonable assumption that people with 
interests in this field _would_ obtain assignments through 
PrimeNet. In the early days it would be much more likely that 
people were replicating each other's work without knowing it. 
Presumably such a situation is what led to your disappointment.


Regards
Brian Beesley
_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: Re: Mersenne Digest V1 #913

2001-12-05 Thread bjb

On 4 Dec 2001, at 20:36, Gordon Spence wrote:

> >I've triple-checked thousands of small exponents - some of the
> >ones where the accepted residual was recorded to only 16 bits or
> >less, which makes the chance of an undetected error _much_
> >greater (though still quite small) - so far no substantive errors in the
> >database have come to light. A very few (think fingers of one hand)
> >instances of incorrectly matched residuals have come to light -
> >completing the double-check in these cases proved that one of the
> >recorded residuals was correct.
> 
> Currently my team report cleared list shows 338 double checks and 12 double 
> checked factored including this monster

I'm not talking about missed factors. The database shows that all 
the small exponents ( < 1 million) have been factored at least a bit 
or two deeper than the "calculated optimum", so I haven't even 
been trying. I've found quite a few factors of these small exponents 
by running P-1, that's a different story.

My point here is that if we have database entries with at most one 
64-bit residual (so that the matching residuals depend on only the 
bottom 16 bits), so far when I've run a triple-check LL test the 
bottom 16 bits have always matched; indeed when there is already 
one 64-bit residual to compare with, my triple-check has always 
matched that.

The work I'm doing in this area is a marginally useful way of using 
systems which are rather too limited to do other work.
> 
> 6630223 87 DF 195139088771490335223859559 07-Apr-01 07:58 trilog
> 
> (In fact when it was checked in PrimeNet initially rejected it because it 
> was longer than this sort of check was supposed to find! Has anyone found a 
> factor bigger than 87 bits using Prime95?)

Interesting, though George's answer explains this. I wrote the factor 
validation code running on the server; I specifically wrote it to be 
able to handle factors of any size (subject to system memory at 
approx 8 bytes per decimal digit of the factor), with a rider that run 
times for extremely large factors (>> 50 digits) might be 
problematic, given that the server cannot afford to spend very long 
running each validation.

My record factor found using Prime95 is

[Fri Sep 07 21:33:06 2001]
ECM found a factor in curve #199, stage #2
Sigma=6849154397631118, B1=300, B2=3.
UID: beejaybee/Simon2, P1136 has a factor: 
9168689293924594269435012699390053650369

I've actually found a couple of larger factors using P-1, but these 
don't count as the factors found were composite. This can happen 
because P-1 depends on finding the GCD of a calculated value and 
the number being factored. (So does ECM.) If you're (un?)lucky you 
can find two factors at once, in which case the result of the GCD is 
their product. This is what ECM uses the lowm & lowp files for - 
ECM is often used to try to find more factors of a number with 
some factors already known; when you get a GCD > 1 you have to 
divide out any previously known factors to find if you've discovered a 
new one.
> 
> Of course some of these may be because the original check went to a lower 
> bit depth than the version of Prime95 that I used.

Naturally.

> I know from doing "deep 
> factoring" in the 60m range that one more bit of factoring can find a "lot" 
> of extra factors...

Over ranges of reasonable size, the number of factors you find 
between 2kp+1 and 2Kp+1 should be independent of p, i.e. the 
expected distribution of smallest factors is logarithmic. For factors 
of a particular absolute size, larger exponents make finding factors 
easier. The effort involved is (ignoring complications resulting from 
computer word length, which are certainly not insignificant!) 
dependent on the range of k, not the range of 2kp+1.

> So if we say that as a ballpark figure half of these are 
> due to an increase in factoring depth, then the error rate from this 
> admittedly small sample is 1.78% or in other words of the current 137,924 
> exponents less than 20m with only a single LL test we can expect to find 
> just under 2500 exponents with an incorrect result.

This is an interesting finding - roughly in line with other estimates of 
raw error rates - but I'm not sure I entirely understand the logic. I 
simply don't see how "missed factor" errors during trial factoring 
are related to incorrect residuals resulting from LL testing - except 
that the LL test wouldn't have been run if the factor wasn't missed.


Regards
Brian Beesley
_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Mersenne: Correction regarding 'p-1 records' thread

2001-12-05 Thread Nathan Russell

Several list members have been kind enough to point out to me that 2^n is 
the smallest n+1 bit number - not the smallest n bit number - in the saeme 
way that 10^1 is the smallest 2-digit number.

Nathan

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: Re: Factoring benefit/cost ratio

2001-12-05 Thread ribwoods

Brian Beesley wrote:
> On 3 Dec 2001, at 20:38, [EMAIL PROTECTED] wrote:
[... snip ...]
> > I think our record shows that a verified factor is still
> > slightly (by a minute but nonzero margin) more reliable an
> > indicator of compositeness than two matching nonzero LL
> > residues.
>
> AFAIK our record does _not_ show any such thing.

Oh?  It doesn't?

If our record does _not_ show that a verified factor is more reliable
an indicator than DC LLs, then what does it show as the (higher,
according to this hypothesis) error rate for verified factors?

Is the observed error rate in reported factors as great as the ~1-1.5%
error rate in reported LL results?  How often does double-checking a
reported factor find that the reported factor was in error?

> In _theory_ there is a very small chance that we _may_ have accepted
> an incorrect residual even after double-checking. There is a small
> chance in such a case that the residual should have been zero, i.e.
> we missed a Mersenne prime.

... and my contention is that those small chances, for L-L results,
exceed the corresponding probabilities in the case of a reported
factor.

There is a small chance that we may accept an incorrect factor even
after double-checking it, but that chance is even smaller than the
small chance that we may accept an incorrect double-checked L-L
residual.

> I've triple-checked thousands of small exponents - some of the
> ones where the accepted residual was recorded to only 16 bits or
> less, which makes the chance of an undetected error _much_
> greater (though still quite small) - so far no substantive errors in
> the database have come to light. A very few (think fingers of one
> hand) instances of incorrectly matched residuals have come to light
-

How does that compare to the observed rate of incorrect factors
discovered after triple-checking _them_?

> > Some error sources would be more likely to affect the longer LL
> > test than the shorter factor verification.)
>
> Indeed - if errors occur at random intervals, results should get
> less reliable as runs take progressively longer.

... and L-L verification runs take a lot longer than factor
verification runs, so we can reasonably expect that this class of
error will affect L-L verification runs (and thus lower such runs'
reliability) a lot more than they do factor verification runs, all
else being equal (such as use of the same hardware systems).

> However there does seem to be a wide variation in the reliability
> of systems. Some seem to have a lot of problems, some very few
> (if any).

So, while a factor verification run on a system that has a lot of
problems may be less reliable than an L-L verification run on a
system that has very few or no problems, if we were to compare
L-L verification and factor verification runs on the same system or
on systems of equal problem probability, we would expect to find that
errors whose rates were proportional to time would affect the L-L
verification runs more than the factor verification runs because of
the significant difference in average run times between the two
categories, making the L-L verification runs less reliable than
the factor verification runs.  Agreed?

> I've detected four problems on my own systems so far; two of these
> were caused by a failed cooling fan on a P100 system (unlike
> current systems, the cooling was _almost_ sufficient even without
> forced ventilation); one was caused by a configuration error -
> running Glucas on an Alpha system I inadvertently allowed "fast"
> arithmetic and turned off error checking, which was a really stupid
> thing to do on a system with a 53-bit mantissa FPU - as this was a
> QA run on a 33 million range exponent, I was lucky to lose only 3
> months work. The last one I never got to the bottom of, but I
> strongly suspect that Prime95's workspace memory was clobbered
> during a software installation on a Windows 98 system.

How many of those problems caused errors during L-L verifications,
and how many caused errors during factor verifications?

However, you may not have spent anywhere near as much time doing
factor verifications as you have doing L-L verifications, so it may
not be valid to draw any conclusion about comparative error rates on
your system.

Richard B. Woods


_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



RE: Mersenne: p-1 records

2001-12-05 Thread Paul Leyland


> Is this a bug in the reporting software?  I don't have the 
> tools to work it out exactly, but a 103-bit number should be slightly
larger 
> than 2^103, or

Nope.  A 103-bit number N should lie in the range 2^102 <= N < 2^103.

> Something really odd is going on.


Perhaps this small example will make things clearer:

N(dec)  N(bin)  bits

 1   1 1
 2  10 2
 3  11 2
 4 100 3
 5 101 3
 6 110 3
 7 111 3
 81000 4
 91001 4
101010 4
111011 4
121100 4
131101 4
141110 4
15 4
16   1 5
...



Paul
_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers



Re: Mersenne: p-1 records

2001-12-05 Thread Henk Stokhorst

Nathan Russell wrote:

>> 12348829 103   F  9722991869324431663702571958950  22-Feb-01 07:48
>> SCUM   C7375CE26
>
>
> Is this a bug in the reporting software?  I don't have the tools to 
> work it out exactly, but a 103-bit number should be slightly larger 
> than 2^103, or
> 10141204801825835211973625643008.
> the factor
> 9722991869324431663702571958950
> is one digit too short, and thus appears to be truncated. 

2^103 is the upper limit I believe, not the lower limit.

YotN,

Henk

_
Unsubscribe & list info -- http://www.ndatech.com/mersenne/signup.htm
Mersenne Prime FAQ  -- http://www.tasam.com/~lrwiman/FAQ-mers