![]() |
|
|
#12 | |
|
Sep 2010
Scandinavia
3·5·41 Posts |
Quote:
In short; In order to take advantage of the superior factoring capabilities of a GPU, I would have to invest of CPU cycles anyway. Still more efficient given that I want to factor the number at hand. But it's certainly a different deal. |
|
|
|
|
|
|
#13 | |||||
|
Sep 2010
Scandinavia
3×5×41 Posts |
Quote:
Quote:
Also, not that that makes a big difference either, when memory is sufficient; Suyama's extension can be used. Quote:
Are you saying that the time a pass takes is roughly proportional to the amount of relative primes processed during it? Quote:
And it doesn't make sense to spend extra cycles and time only to not allocate available memory. But I get your point, more generally. I have 8GB RAM. Let's say I insist on Intel, the price difference between dual and quad would correspond to ~1.5GB of RAM... Speaking of hardware; do you know which parameters affect P-1 and/or ECM performance most? Like timings etc.. (Any other comments on the hardware portion of my original post?) Quote:
Fix B1, and memory use will not increase strictly with B2... I have even less understanding of ECM than of P-1. |
|||||
|
|
|
|
|
#14 | |||||||||
|
Jun 2003
7·167 Posts |
Quote:
Quote:
Quote:
Quote:
Quote:
Quote:
Quote:
On my machine I usually set the core that is doing stage 2 to a high priority, well above standard, so as to shift all other processes onto the other core. Still I find I accumulate stage 1 work over time, and occasionally switch the other core to some other task. Quote:
Quote:
|
|||||||||
|
|
|
|
|
#15 | ||||
|
Sep 2010
Scandinavia
26716 Posts |
Quote:
Someone said it well... along the lines of; (relative primes) mod (primes per pass) should be either 0 or ~<(primes per pass) I'm a little confused about blocksizes... IIRC the run I finished this night used 432relative primes, and E=12. That took ~5GB btw. Quote:
Quote:
I'd like to get input from someone with hardware expertise. Right now I can read at 10.4GB/s and write at 7.8GB/s. Quote:
Any way, in ECM; the memory use as a function of B2 is a discontinuous function. |
||||
|
|
|
|
|
#16 | |||
|
"Richard B. Woods"
Aug 2002
Wisconsin USA
22·3·641 Posts |
Quote:
Quote:
Quote:
Mr. P-1's "Yes, but it's not smooth. Increasing ..." response was about "per pass overhead", not total time per pass. Your "roughly proportional to the amount of relative primes processed during it" applies to total time per pass. |
|||
|
|
|
|
|
#17 | |||
|
"Richard B. Woods"
Aug 2002
Wisconsin USA
22·3·641 Posts |
Quote:
Quote:
Quote:
1) Implicitly, as part of a "Test=" LL assignment for an exponent for which P-1 has not yet been performed, 2) Explicitly with "Pfactor=" (this is the worktodo keyword for a P-1 assignment from PrimeNet), 3) Explicitly with "Pminus1=" In case 3), the Pminus1= line explicitly specifies B1 and B2. Prime95 does no bounds-choosing, but uses what's specified. This is not the case you're talking about here. In cases 1) and 2), prime95 does choose the optimal B1 and B2 bounds. One of the inputs to the bounds-choosing algorithm is the user-specified Available Memory. The algorithm will calculate how many workareas can be accommodated in the Available Memory, and whether or not the Suyama extension will yield a worthwhile improvement. Then, it tries varying values of B1 and B2 to find the optimal ratio of (probability of finding a factor at that B1/B2)* to (estimated elapsed time for stage 1 plus stage 2 at that B1/B2). In general, increasing the amount of Available Memory allows the algorithm to make its estimates with increasing numbers of workareas and, if high enough, Suyama extension. The latter numbers, in turn, affect both the probability of finding a factor and the estimated elapsed run time for stages 1 and 2. So, in the bounds-choosing algorithm, it's not that increasing B2 => using more memory. It's that allowing more memory => being able to use more workareas and Suyama extension, which become more effective at raising the probability-to-time ratio for higher B2 values. - - - * - Actually, the ratio is (estimated time saved by not running LL(s) if a factor is found) to (estimated elapsed time for stage 1 plus stage 2 at that B1/B2). However, (estimated time saved by not running LL(s) if a factor is found) = (probability of finding a factor at that B1/B2) * (estimated time for remaining LL(s) if no factor is found). Since the algorithm is always concerned with only one exponent at a time, the (estimated time for remaining LL(s) if no factor is found) is constant for all choices of B1/B2, and can be omitted from a description comparing one set of B1/B2 to another set of B1/B2 for optimal P-1 on that exponent. Last fiddled with by cheesehead on 2011-01-08 at 09:20 |
|||
|
|
|
|
|
#18 |
|
Sep 2010
Scandinavia
26716 Posts |
A bigger area changes the optimal bounds. That makes sense.
|
|
|
|
|
|
#19 | |
|
Oct 2007
Manchester, UK
53·11 Posts |
Quote:
Additionally, if you disable all of the EIST (speedstep) and the various sleep states so that the CPU can't downclock itself when there's minimal activity, but you leave Turbo mode enabled, then on some motherboards you effectively gain access to permanent Turbo mode, where it can't downclock the CPU once it's on. If this applies to you then you could get to 3,840 MHz without changing the BCLK at all. However, you would likely want to alter the cooling first. Unless you're relying on Turbo for your overclock though, when you get to higher frequencies (greater than 3.5GHz say) I recommend disabling it. Before changing any more settings, I would recommend running the Intel Burn Test. It generally finds faults quicker than the Prime95 torture test: http://downloads.guru3d.com/IntelBur...load-2047.html Run at least 5 rounds with as much memory as you have available, and close all other programs to put as much stress on the CPU as possible. Intel say that the safe upper limit for the i3 VCore is 1.4 V, but a lot of people stay below 1.35 or 1.3 V for 24/7 use on air cooling. You should be able to get a lot of mileage out of your CPU before 1.3 V though, providing your cooling can keep up. |
|
|
|
|
|
|
#20 |
|
Sep 2010
Scandinavia
26716 Posts |
Thank you very much for input on the hardware!
A little update; Vcore ranges 1.128~1.144. P=50.21W most of the time. (very stable) I don't think any automatic downclocking is enabled. Except that overheat thing, the name of which I can't remember. Doesn't seem to ever have kicked in though. I think 58°C was a one time thing. It didn't go there again for 30+hrs of almost completely uninterrupted full load. So I reset the log; it reached 56 three hours ago. Besides that it has been below 56°C for a little over 3hrs now. Guessing it would reach 57 with a little more disk and GPU activity. 160*20 has been thoroughly tested. I think by now I can say that about the 8-9-9-23 timings too. (Been running several P-1 sessions allocating 4~6GB for many hours) The system failed after 1~2days at 160*21, but I'm almost positive that was caused by too aggressive RAM tweaking. All cooling is stock. But I put in two additional fans. And set high RPM on all. One moving air out (2000~2100RPM, 120mm), below the PSU. The other pulling air in (3000~3100, 90mm) through a "tube" directed at the CPU(CPU fan mostly at 2000~2100RPM, even though it drops when idle, which I don't like). I also removed an older disk that was producing heat. GPU doesn't go much further than 40°C the way I use it. My only internal disk barely reaches body temp. And the motherboard is almost indistinguishable from RT since the extra fans. Everything I've described so far seems stable. Moving on; The program you linked says XP/Vista. I dl'd it, should it be ok on 64bit Win7? BIOS allows BCLK to be set to 80-500. CPU mult. is 9-23. RAM is 10, 8 or lower.. Is BCLK somehow sensitive in itself? Or does it just set the pace for CPU, RAM, DMI etc.? I haven't been keeping up with hardware for a few CPU generations... Should maxing out the CPU be considered to be a problem with 1, 2 or 3 variables? Multiplier, Voltage, and BCLK. In other words; is it enough to maximize BCLK*multiplier? Or...? You get the point. Should I leave CPU voltage on [Auto] while setting mult. to 21? Any comments/tips regarding RAM? Thank you very much! |
|
|
|
|
|
#21 | |||
|
Jun 2003
7·167 Posts |
Quote:
Quote:
Quote:
The quantity being optimised is the difference, not the ratio, between the cost of the computation and the expected benefit. The bounds are optimal when this quantity is minimised, which happens when the partial derivatives of this quantity with respect to B1 and B2 are both zero. If then and so So there is an optimal ratio, but it is between the delta values. or equivalently between the partial derivatives (with respect to both B1 and B2) of cost and benefit, not between cost and benefit per se Last fiddled with by Mr. P-1 on 2011-01-08 at 14:23 |
|||
|
|
|
|
|
#22 | ||
|
Jun 2003
7·167 Posts |
Correct.
Quote:
Quote:
Last fiddled with by Mr. P-1 on 2011-01-08 at 14:19 |
||
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Large FFT tweaking | Zerowalker | Information & Answers | 8 | 2013-04-19 15:01 |
| Tweaking polynomial search for C197 | fivemack | Msieve | 38 | 2011-07-08 08:12 |
| Tweaking and compiling the Kleinjung siever | Batalov | Factoring | 57 | 2010-11-30 18:03 |