![]() |
|
|
#529 |
|
Random Account
Aug 2009
195610 Posts |
|
|
|
|
|
|
#530 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,419 Posts |
No. I have two, on the same machine. They came with different default clocks, 701 and 725. The 725 I downclock to 702. The 701 has been reliable; the 725/702 has repeatable memory errors in the middle of the address range, that at one time were reduced by downclocking but no longer are. So I use it only for trial factoring, which occupies memory not affected by the errors. I've become an advocate of testing as much gpu memory as possible, from what I've learned on that second GTX480.
|
|
|
|
|
|
#531 | |
|
Random Account
Aug 2009
22·3·163 Posts |
Quote:
|
|
|
|
|
|
|
#532 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,419 Posts |
One of the two goes AWOL at varying intervals. I found that to get reliable p-1 or LL tests, it required making the 702/memory error one device zero. If it was device one, and p-1 or LL were set to run on device zero, when the one goes AWOL, the bad-memory one drops to device zero and causes problems with a p-1 or LL run. When I say AWOL, it's physically there, but GPU-Z only finds the one device, and a running GPU-Z already set to track device one ceases displaying its sensor readings, Windows event log shows a driver restart, and restarted cudapm1 and cudalucas don't find a device one. Clearing that up requires a shutdown/restart, in command line, shutdown -r.
|
|
|
|
|
|
#533 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,419 Posts |
Hi,
Has anyone experimented with running more than one instance of CUDAPm1 on a single GPU? Reason I ask is I'm used to seeing 100% GPU load in GPU-Z, with a single instance of CUDALucas or CUDAPm1 per GPU, but on a GTX1070 it varies 99-100%. Also I have found gains in running multiple Mfaktc instances, raising the GPU load from 98 to 100%, on a GTX480. In sharing a single GTX480 GPU between simultaneous single instances of CUDALucas and CUDAPm1, in a quick test, I'm calculating more combined throughput than either running alone, by several percent. Since I'm running numerous GPUs, if that holds up, it's the equivalent of adding another GPU. Any light you can shed on effects of multiple instances, such as confirming results, or negative results, on various GPU models, would be appreciated. |
|
|
|
|
|
#534 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
152B16 Posts |
I have never seen a 450 clock rate on either of my GTX480's, or a lower clock after driver restart or program reset. I have seen them drop from 70x to 405, and then some seconds later down to 50.6, when there's little or no GPU processing load, and go back up with load.
|
|
|
|
|
|
#535 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
5,419 Posts |
Quote:
|
|
|
|
|
|
|
#536 |
|
Random Account
Aug 2009
22·3·163 Posts |
Code:
Transforms: 2024 M83496293, 0x4b29a8a5677c72db, n = 4608K, CUDAPm1 v0.20 err = 0.17578 (0:40 real, 19.5816 ms/tran, ETA 0:04) Last fiddled with by storm5510 on 2017-12-05 at 01:20 |
|
|
|
|
|
#537 |
|
"/X\(‘-‘)/X\"
Jan 2013
1011011100102 Posts |
Use smaller fonts?
|
|
|
|
|
|
#538 |
|
Random Account
Aug 2009
36448 Posts |
|
|
|
|
|
|
#539 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
10101001010112 Posts |
Quote:
CUDA reports 830M of 1024M GPU memory free. Index 64 Using threads: norm1 256, mult 256, norm2 32. Using up to 720M GPU memory. Selected B1=690000, B2=12420000, 3.05% chance of finding a factor Starting stage 1 P-1, M84861479, B1 = 690000, B2 = 12420000, fft length = 4608K Doing 995519 iterations Iteration 5000 M84861479, 0x85dcbca418bb3656, n = 4608K, CUDAPm1 v0.20 err = 0.27344 (3:03 real, 36.6115 ms/iter, ETA 10:04:24) ... Iteration 995000 M84861479, 0xb98ed42b48260d4a, n = 4608K, CUDAPm1 v0.20 err = 0.25000 (3:02 real, 36.5191 ms/iter, ETA 0:18) M84861479, 0x4a2093b79c7bf108, n = 4608K, CUDAPm1 v0.20 Stage 1 complete, estimated total time = 10:06:31 Starting stage 1 gcd. M84861479 Stage 1 found no factor (P-1, B1=690000, B2=12420000, e=0, n=4608K CUDAPm1 v0.20) Starting stage 2. Using b1 = 690000, b2 = 12420000, d = 2310, e = 2, nrp = 10 Zeros: 554802, Ones: 637038, Pairs: 121194 Processing 1 - 10 of 480 relative primes. ... Stage 2 complete, 1766191 transforms, estimated total time = 9:31:47 Starting stage 2 gcd. M84861479 Stage 2 found no factor (P-1, B1=690000, B2=12420000, e=2, n=4608K CUDAPm1 v0.20) Weird, but I'll take it. A couple other things I had thought of to try were matching OS and system ram on another box & GPU and retrying there. System that ran the problem 84.2m exponents to completion had a newer Windows OS and twice the system ram. ... |
|
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| mfaktc: a CUDA program for Mersenne prefactoring | TheJudger | GPU Computing | 3497 | 2021-06-05 12:27 |
| World's second-dumbest CUDA program | fivemack | Programming | 112 | 2015-02-12 22:51 |
| World's dumbest CUDA program? | xilman | Programming | 1 | 2009-11-16 10:26 |
| Factoring program need help | Citrix | Lone Mersenne Hunters | 8 | 2005-09-16 02:31 |
| Factoring program | ET_ | Programming | 3 | 2003-11-25 02:57 |