![]() |
|
|
#23 | |
|
"Viliam FurÃk"
Jul 2018
Martin, Slovakia
54 Posts |
Quote:
|
|
|
|
|
|
|
#24 | |
|
Jul 2003
wear a mask
167610 Posts |
Quote:
|
|
|
|
|
|
|
#25 | |
|
"David Kirkby"
Jan 2021
Althorne, Essex, UK
6608 Posts |
Quote:
Code:
Daytime P-1/P+1/ECM stage 2 memory in GB (0.300000): 380 Please enter a value between 0.000000 and 338.915039. 338 Nighttime P-1/P+1/ECM stage 2 memory in GB (0.300000): 380 Please enter a value between 0.000000 and 338.915039. 338 Yes, simply edit local.txt and restart mprime. The restriction is there to prevent users from creating a RAM thrashing situation. I'm sure George would not have told me how to set a limit higher than 338.915039 GB if that would have misconfigured the software. |
|
|
|
|
|
|
#26 | ||
|
Jun 2003
13DE16 Posts |
Quote:
Quote:
Code:
[Jul 20 13:27:30] Stage 2 init complete. 12157 transforms. Time: 98.042 sec. ... [Jul 20 13:52:27] M105207229 stage 2 is 9.08% complete. Time: 98.176 sec. Then: Code:
[Jul 20 13:53:16] M105207229 stage 2 is 9.09% complete. [Jul 20 13:59:47] M105207229 stage 2 is 11.39% complete. Time: 98.387 sec. But let's say your numbers are accurate. You say that the other worker which did whole stage 2 with reduced memory would finish in 16600. Did it by any chance use a smaller B2? If not, then you should consider the increased projected time for the first one as a penalty for restarting with reduced memory -- a penalty which you wouldn't have paid had you run the whole test at reduced memory. Also a quick sanity check: 50% @ 16300 rate and 50% @ 17100 rate is very roughly 16700 rate. So not sure how you say that you will gain time by allowing it to run at full memory for half the time. Anyway, I stand by my original point. If you're going to allocate all the memory to a single worker, you should ensure that no other worker enters stage 2. Otherwise, limit the per-worker memory. That way, you don't have to worry about all the penalties and stuff. |
||
|
|
|
|
|
#27 | |
|
"David Kirkby"
Jan 2021
Althorne, Essex, UK
24·33 Posts |
Quote:
Code:
[Worker #1 Jul 21 15:09] Optimal P-1 factoring of M105213601 using up to 376832MB of memory. [Worker #1 Jul 21 15:09] Assuming no factors below 2^76 and 1 primality test saved if a factor is found. [Worker #1 Jul 21 15:09] Optimal bounds are B1=434000, B2=21338000 [Worker #1 Jul 21 15:09] Chance of finding a factor is an estimated 3.6% [Worker #2 Jul 21 15:09] Optimal P-1 factoring of M105212323 using up to 376832MB of memory. [Worker #2 Jul 21 15:09] Assuming no factors below 2^76 and 1 primality test saved if a factor is found. [Worker #2 Jul 21 15:09] Optimal bounds are B1=434000, B2=21338000 [Worker #2 Jul 21 15:09] Chance of finding a factor is an estimated 3.6% [Worker #3 Jul 21 15:09] Optimal P-1 factoring of M105212539 using up to 376832MB of memory. [Worker #3 Jul 21 15:09] Assuming no factors below 2^76 and 1 primality test saved if a factor is found. [Worker #3 Jul 21 15:09] Optimal bounds are B1=434000, B2=21338000 [Worker #3 Jul 21 15:09] Chance of finding a factor is an estimated 3.6% [Worker #4 Jul 21 15:09] Optimal P-1 factoring of M105212563 using up to 376832MB of memory. [Worker #4 Jul 21 15:09] Assuming no factors below 2^76 and 1 primality test saved if a factor is found. [Worker #4 Jul 21 15:09] Optimal bounds are B1=434000, B2=21338000 [Worker #4 Jul 21 15:09] Chance of finding a factor is an estimated 3.6% Any thoughts about what will happen when they enter stage 2? I will let you know later! Last fiddled with by drkirkby on 2021-07-21 at 14:40 |
|
|
|
|
|
|
#28 | |
|
Jun 2003
508610 Posts |
Quote:
Where's the popcorn smiley
|
|
|
|
|
|
|
#29 | ||
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,437 Posts |
Quote:
Re George's response on how to set >90% of ram as available, IIRC it's improper to quote a PM without consent, and SOP to indicate consent to quote in the same post if it was given. George gives the user the benefit of the doubt as to knowing what they are doing, or learning, or supporting experimentation. He also supports, and codes for, many usages that are not optimal. That is I believe a project-level optimization. Happier users are more likely to remain engaged and apply more cycles of more hardware to the overall project progress. Quote:
I just conducted the following experiment in v30.6b4 prime95. Remove all per-worker memory caps from local.txt. Restart prime95 with only its global memory cap 110GB. Only one of 4 workers is running P-1. Code:
[Jul 21 11:04] Waiting 15 seconds to stagger worker starts. [Jul 21 11:04] Worker starting [Jul 21 11:04] Setting affinity to run worker on CPU core #13 [Jul 21 11:04] Optimal P-1 factoring of M105198883 using up to 112640MB of memory. [Jul 21 11:04] Assuming no factors below 2^76 and 2 primality tests saved if a factor is found. [Jul 21 11:04] Optimal bounds are B1=882000, B2=50252000 [Jul 21 11:04] Chance of finding a factor is an estimated 4.61% [Jul 21 11:04] [Jul 21 11:04] Setting affinity to run helper thread 1 on CPU core #14 [Jul 21 11:04] Setting affinity to run helper thread 2 on CPU core #15 [Jul 21 11:04] Using AVX FFT length 5600K, Pass1=896, Pass2=6400, clm=1, 4 threads [Jul 21 11:04] Setting affinity to run helper thread 3 on CPU core #16 [Jul 21 11:04] Ignoring suggested B1 value, using B1=857000 from the save file [Jul 21 11:04] Ignoring suggested B2 value, using B2=44190000 from the save file [Jul 21 11:04] Available memory is 111776MB. [Jul 21 11:04] Using 32763MB of memory. CUDAPm1 and GpuOwl also resist bounds changes midstream. Getting initially or later, exactly optimal bounds is nice but not very important. It's important to remember that a) the differing bounds for considerably different memory allocations are not that far apart on a log scale, b) the slopes (partial derivative of run time vs B1 or B2, probability of factor found, probable net time savings) are small near a given optimal for given parameters c) it only affects stage 2 of P-1, which is ~1/60 of a P-1/PRP combination. So the memory change effect is ~ (small delta)3. I have a heterogenous 100M / 100Mdigit combination queued up for test. Last fiddled with by kriesel on 2021-07-21 at 16:50 |
||
|
|
|
|
|
#30 |
|
"David Kirkby"
Jan 2021
Althorne, Essex, UK
24×33 Posts |
Okay, this is what happened ...Worker 2 got completed first, and grabbed 207828 MB (203 GB) RAM
Code:
[Worker #2 Jul 21 15:42] Stage 1 GCD complete. Time: 46.141 sec. [Worker #2 Jul 21 15:42] Available memory is 376632MB. [Worker #2 Jul 21 15:42] D: 2310, relative primes: 4743, stage 2 primes: 1313481, pair%=94.34 [Worker #2 Jul 21 15:42] Using 207828MB of memory. Next worker 1 grabbed 168867 MB (165 GB) of RAM, which is 99.98% of the reaming available memory. Code:
[Worker #1 Jul 21 15:42] Stage 1 GCD complete. Time: 45.777 sec. [Worker #1 Jul 21 15:42] Available memory is 168900MB. [Worker #1 Jul 21 15:42] D: 1470, relative primes: 3853, stage 2 primes: 1313481, pair%=95.43 [Worker #1 Jul 21 15:42] Using 168867MB of memory. Code:
[Worker #2 Jul 21 15:43] Restarting worker with new memory settings. [Worker #1 Jul 21 15:43] Restarting worker with new memory settings. Code:
[Worker #2 Jul 21 15:43] Using 82273MB of memory. [Worker #3 Jul 21 15:43] Using 125569MB of memory. [Worker #1 Jul 21 15:43] Using 70802MB of memory. [Worker #4 Jul 21 15:43] Using 98163MB of memory. Code:
[Worker #2 Jul 21 16:43] M105212323 stage 2 is 81.24% complete. Time: 576.928 sec. [Worker #1 Jul 21 16:47] M105213601 stage 2 is 80.33% complete. Time: 613.766 sec. [Worker #3 Jul 21 16:49] M105212539 stage 2 is 96.66% complete. Time: 534.885 sec. [Worker #4 Jul 21 16:49] M105212563 stage 2 is 95.41% complete. Time: 539.106 sec Note after the RAM fights, worker 3 ended up with the most RAM and worker 1 the least. Just as they are near the end of completing stage 2, one can see the percentage completed is highest for the worker with the most RAM, and lowest for the worker with the least RAM. Code:
[Worker #3 Jul 21 16:52] Stage 2 GCD complete. Time: 47.464 sec. [Worker #4 Jul 21 16:53] Stage 2 GCD complete. Time: 46.851 sec. [Worker #2 Jul 21 16:57] Stage 2 GCD complete. Time: 48.068 sec. [Worker #1 Jul 21 17:02] Stage 2 GCD complete. Time: 47.487 sec. All workers had the same number of cores (13). Last fiddled with by drkirkby on 2021-07-21 at 16:37 |
|
|
|
|
|
#31 | |
|
"David Kirkby"
Jan 2021
Althorne, Essex, UK
6608 Posts |
Quote:
Looking at the data from https://www.mersenne.ca/exponent/105212563 I would intuitively think it is better to spend 9.3765 GHz days to get a 5.3973% chance of finding a factor, rather than 9.0809 GHz days for the 3.6% chance reported by mprime. However, mprime has changed the P-1 factoring a lot in the latest version, so I don't know if it is valid to compare with the data at https://www.mersenne.ca/ any more. |
|
|
|
|
|
|
#32 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,437 Posts |
|
|
|
|
|
|
#33 | |
|
If I May
"Chris Halsall"
Sep 2002
Barbados
263616 Posts |
Quote:
I've had my CPU Colab workers clearing out candidates which didn't have a P-1 done before the FTC. A few recent examples are 104510041, 104509267 and 104180299 Interestingly, these were all "cleared" by way of C-PRP. But, at the same time, the column showing "Available: P-1" for 104M hasn't changed for days. So, empirically, those candidates I'm doing P-1 on are not the ones included in this column. Any insight the Primenet "gods" might be able to provide would be appreciated. My OCD doesn't like having values in that column before the Cat 0 wavefront... |
|
|
|
|
![]() |
| Thread Tools | |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| P-1 on small exponents | markr | PrimeNet | 18 | 2009-08-23 17:23 |
| Large small factor | Zeta-Flux | Factoring | 96 | 2007-05-14 16:59 |
| Problems with Large FFT but not Small FFT's? | RichTJ99 | Hardware | 2 | 2006-02-08 23:38 |
| Small range with high density of factors | hbock | Lone Mersenne Hunters | 1 | 2004-03-07 19:51 |
| Small win32 program, range of time to do a TF | dsouza123 | Programming | 1 | 2003-10-09 16:04 |