mersenneforum.org

mersenneforum.org (https://www.mersenneforum.org/index.php)
-   Software (https://www.mersenneforum.org/forumdisplay.php?f=10)
-   -   Does the LL test:s factorization save or waste CPU time? (https://www.mersenneforum.org/showthread.php?t=75)

svempasnake 2002-09-14 15:19

[quote="toferc"]You need to add 1, not 0.5.[/quote]Dang! :surprised:ops: Youre right. Thougt I tripplechecked it on a bunch of numbers, but I must have been blind. With 1 it works. Thanks for seeing it!

[quote="toferc"]The reason your code is working with huge numbers is that they are implicitly cast to a floating-point representation before the log operation. [/quote]Ah! I tested now. Yup, it fails after 308 decimal digits, so this makes sence.

binarydigits 2002-09-15 03:41

[quote="svempasnake"][quote="binarydigits"]The factor was 40498974660725214481 which is more than 9% higher than 2^66.[/quote]
Not quiet :shock: : 2^66 is 7.4e19. The factor mentioned here is 4.0e19 or 65 bits.[/quote]

Oops, I calculated it correctly but worded my message wrong. The factor is 9% higher than 2^65, which still makes it a 66-bit factor. Sorry for the confusion.

svempasnake 2002-09-16 21:32

Re: Does the LL test:s factorization save or waste CPU time?
 
[quote="dswanson"]...For this data set, George's formula looks pretty darn good. :D But there are some trends in the results that make me suspicious of generalizing the conclusion. In particular, the percentage of numbers factored within a range decreases as the M-number increases. But George's formula predicts that this percentage should be a constant. I suspect that if I had a similar data set in the vicinity of say 30-33M that George's formula would look much worse, at least in the 2^52 to 2^57 range.

If I were really ambitious I'm sure I could piece together a much more thorough analysis by combining the data in the factors and nofactors files. Are there any enterprising souls out there who'd like to take on this challenge...?[/quote]

OK, this is a part of what I can get out of it:
------------------------------------------------------
Probability that TF up to n bits will find a factor, under assumption TF already is done up to n-1 bits. Based on historical data from factors & nofactors - Sept.08-2002[code:1]
Fact bits: Exponent range
Predicted 00M-79M 10M-15M 15M-20M 30M-35M
n 1/n: ( =ALL)
50 2,00% 2,05% 1,99% 1,98% 2,03%
51 1,96% 2,01% 1,97% 1,95% 1,93%
52 1,92% 1,97% 1,90% 1,96% 1,96%
53 1,89% 1,89% 1,89% 1,90% 1,88%
54 1,85% 1,85% 1,80% 1,81% 1,86%
55 1,82% 1,80% 1,81% 1,78% 1,84%
56 1,79% 1,79% 1,77% 1,73% 1,76%
[/code:1]
When looking here, Georges 1/n formula seem to work very good. I guess the standard deviations are normal for the actual sample sizes. For reference, there are around 2800 factors behind the percentages at the 5M samples, and around 44000 per figure in the "ALL" collumn.

I choosed to show up to 56 bits only, just because I then feel pretty safe that we look at "apples vs apples" rather than anything uncomparable. Nofactors states that every exponent up to 79M is trial factored to at least 56 bits.

An assumtion I made about the factors table, is that when a factor is shown, it is supposed to be the smallest existing one. Anyone knowing if that's right?

Prime95 2002-09-16 22:10

Re: Does the LL test:s factorization save or waste CPU time?
 
[quote="svempasnake"]An assumtion I made about the factors table, is that when a factor is shown, it is supposed to be the smallest existing one. Anyone knowing if that's right?[/quote]

There may be isolated cases where that is not the case, but rare enough to not affect your excellent analysis.

dswanson 2002-09-18 04:07

Hey Svempasnake, great work! :D Looks like George is fully exonerated. The trend I saw must have a statistical fluke due to the limited size of my dataset.

I'm not convinced that it was necessary to exclude all numbers above 56 bits. As long as at each level you work only with numbers that have either had a factor found at that level or that have been trial factored through that level without finding a factor, the results for that level should be valid. The only problem is that the sample size gets smaller and smaller as you reach levels that have been less than fully trial-factored over the entire 0-79M range, making the results less statistically significant at each subsequent level.

On second thought, P-1 factoring is also going to corrupt the results at higher levels. If a number has been trial factored to n bits but subsequent P-1 factoring finds a factor at n+3 bits, this new factor will be placed in the factors database with no indication that the range from n+1 to n+3 hasn't been trial factored.

So I guess I've just talked myself back around to your assumption that the number in the factors table is indeed the smallest factor. And as more and more factors are found via P-1 factoring, the validity of this method of data-mining the factors database will become worse and worse.

George, is a record kept anywhere of which factors were found using trial factoring versus using P-1?

Prime95 2002-09-18 14:35

[quote="dswanson"]is a record kept anywhere of which factors were found using trial factoring versus using P-1?[/quote]

Nope

svempasnake 2002-09-21 10:11

Are only TF:s counted in nofactors?
 
[quote="Prime95"][quote="dswanson"]is a record kept anywhere of which factors were found using trial factoring versus using P-1?[/quote]

Nope[/quote]On the other hand, I believe we will know this anyway, most of the time! Even better, if nofactors is updated the way I hope and belive, the cases where we can't see if a factor was found with TF or P-1 should be very rare.

Let's have a look at nofactors, which can help us inderectly:[code:1]
All exponents 8.25M - 13.38M in nofactors (TF goes up to 64 bits)

Factored to bits / Number of exponents
59 13
60 15
61 7
62 9
63 6472
64 113487

All exponents 13.38M - 17.85M in nofactors (TF goes up to 65 bits)
Factored to bits / Number of exponents
57 21
58 59
59 149
60 56
61 30
62 6
64 49
65 102436
[/code:1]It looks to me as nofactors purely consist of TF:s, and P-1:s aren't updated here. Does anyone know if thats how it works :question:
If that's the case, we can almost allways tell whether a factor was found by TF or P-1. Especially in exponent ranges where TF is as 99%-complete as above:

Take binarydigits 4.0e19 factor as an example. First we count its bits carefully ;) and see it's a 66 bit factor. Since we know that TF ends at 65 bits for 16 M exponents, this factor must have been found trough P-1 and not trough TF.
Another example: A 17.0 M exponent has a 65 bit factor. Assuming TF went up to 65 bits, this factor was found trough TF and not trough P-1.
Third example: a 12 M exponent has a factor of 64 bits. Assuming TF went to 64 bits, this factor was also found trough TF.
A table like the one above should help us knowing how good or bad such assumtions are.
[quote="dswanson"]I'm not convinced that it was necessary to exclude all numbers above 56 bits.[/quote]Yeah, I now agree my 56 bit limitation was too conservative. I will make another post so you can see some more results. I will probably begin a new topic in the math group in that case. But first it would be nice to know for sure whether P-1 doesn't or does update nofactors.

svempasnake 2002-09-21 17:19

[quote="toferc"]Even more interesting is a factor currently under discussion in the Team Prime Rib Perpetual Thread (part 3) in the forums at ars Technica:

The factor 48091790614964383575080990595271931709835968599049593 of M16486579 was recently found. It was reported as:
[code:1]
16486579 F 102 4809179061496438357508099059527 09-Sep-2002 18:38 horus2_02 horus2[/code:1]
which as you can see has been rather truncated. [/quote]Sorry but I think I have to add some confusion here: I was curious to see wheter the original or the truncated one gets into factors.zip. To my surprise neither of these are here. Instead a smaller factor, 13329478389458153107343, is there! The only good explanation I can think of is that this smaller factor was found later. Would be fun if someone could check if this smaller factor actually divides M16486579.

akruppa 2002-09-21 17:48

[quote="svempasnake"][quote="toferc"]Even more interesting is a factor currently under discussion in the Team Prime Rib Perpetual Thread (part 3) in the forums at ars Technica:

The factor 48091790614964383575080990595271931709835968599049593 of M16486579 was recently found. It was reported as: [...]
[/quote]Sorry but I think I have to add some confusion here: I was curious to see wheter the original or the truncated one gets into factors.zip. To my surprise neither of these are here. Instead a smaller factor, 13329478389458153107343, is there! The only good explanation I can think of is that this smaller factor was found later. Would be fun if someone could check if this smaller factor actually divides M16486579.[/quote]

The large factor is composite.

48091790614964383575080990595271931709835968599049593 =
13329478389458153107343 * 3607927422951418869297495045751

I, too, checked factors.zip to see if the composite factor made it there, but the smaller of the two prime factors is correcty listed.

A p53 by P-1 would just have been too nice..

Alex

jocelynl 2002-09-21 17:54

[quote="svempasnake"][quote="toferc"]Even more interesting is a factor currently under discussion in the Team Prime Rib Perpetual Thread (part 3) in the forums at ars Technica:

The factor 48091790614964383575080990595271931709835968599049593 of M16486579 was recently found. It was reported as:
[code:1]
16486579 F 102 4809179061496438357508099059527 09-Sep-2002 18:38 horus2_02 horus2[/code:1]
which as you can see has been rather truncated. [/quote]Sorry but I think I have to add some confusion here: I was curious to see wheter the original or the truncated one gets into factors.zip. To my surprise neither of these are here. Instead a smaller factor, 13329478389458153107343, is there! The only good explanation I can think of is that this smaller factor was found later. Would be fun if someone could check if this smaller factor actually divides M16486579.[/quote]

in some cases p-1 factoring will find factors that are not prime. So my guess is that this one was factored again and that the smaller factor was found.

garo 2002-09-21 19:59

Re: Are only TF:s counted in nofactors?
 
[quote="svempasnake"]
Let's have a look at nofactors, which can help us inderectly:[code:1]
All exponents 8.25M - 13.38M in nofactors (TF goes up to 64 bits)

Factored to bits / Number of exponents
59 13
60 15
61 7
62 9
63 6472
64 113487

All exponents 13.38M - 17.85M in nofactors (TF goes up to 65 bits)
Factored to bits / Number of exponents
57 21
58 59
59 149
60 56
61 30
62 6
64 49
65 102436
[/code:1]

Ha! I might actually be worth it to do the factoring for atleast the exponents which have been factored to 57-63 bits only. The smaller bit factoring will go through pretty fast I imagine. I looked in the status.txt file and most of these exponents have been tested once and are not yet in range to be doublechecked. I wonder if the prime95 client reported incorrectly or if the folks who did the first test turned off trial factoring. Maybe George can tell us:
[code:1]
9995243,60
9996509,60
9996673,60
9996703,60
9996859,60
12026057,60
12195593,60
12376337,60
12376363,60
12722923,60
12750373,60
13019759,60

12718319
12771757
12808949
12821153
12837259
12862439
12865469
12867941
12872239
12876709
12885793
12904979
13182061
[/code:1]
I am wondering if it's worth getting a box to dedicate itself to factoring these exponents?? In the 13.38-17.85M range several of the exponents are still in the process of being factored so touching them at this time is not prudent but I am definitely going to try and factor the exponents listed above if anyone cannot give me a reason not to.
Garo


All times are UTC. The time now is 16:47.

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.