![]() |
|
|
#1497 | |
|
Random Account
Aug 2009
13·151 Posts |
Quote:
I have only used it for P-1 tests. I just have to make sure the "F" in "PFactor" is a capital. I think PrimeNet issues these in lower case. It took me quite a while to figure out how to customize the bounds. Once done, no problems...
|
|
|
|
|
|
|
#1498 | |
|
P90 years forever!
Aug 2002
Yeehaw, FL
19×397 Posts |
Quote:
I chose 1000 as Mr. Gerbicz original threads used that value calculating a 0.2% overhead. A block size of 400 has a 0.5% overhead. I understand frequent errors make a smaller block size desirable. Prime95 automatically reduces the block size when an error occurs. I'm not suggesting this feature -- a bit of overkill. The P-1 error I was getting: GPU->host read failed (check 61e4 vs 3f07) |
|
|
|
|
|
|
#1499 |
|
"Robert Gerbicz"
Oct 2005
Hungary
2·743 Posts |
|
|
|
|
|
|
#1500 | |
|
"Mihai Preda"
Apr 2015
25338 Posts |
Quote:
The P-1 error -- strange, I don't understand why you were getting it, it seems that the memory transfer (reading from the GPU) or the synchronization around it (i.e. waiting for it to finish) was failing. |
|
|
|
|
|
|
#1501 |
|
P90 years forever!
Aug 2002
Yeehaw, FL
19·397 Posts |
Once you pass a Gerbicz error check (or fail and rollback to a save file that passed a check) you are essentially in a virgin state where you can select any block size you want going forward.
|
|
|
|
|
|
#1502 | |
|
"Robert Gerbicz"
Oct 2005
Hungary
2×743 Posts |
Quote:
it is trivial that f(s+t)=f(t)^(2^s), so you can start a new blocklength=new L at an error checked residue at iteration=t using a new "base"=f(t). (why? because you are trusted, that at iteration=t with high probability you have a good residue). The only difference with this is that at error check you need to multiple with f(t) and not with the smallish a=3. So in the leading wavefront exponents you'd need ~100-250 more mulmods (almost nothing in computation time). |
|
|
|
|
|
|
#1503 | |
|
∂2ω=0
Sep 2002
República de California
19×613 Posts |
Quote:
|
|
|
|
|
|
|
#1504 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
10101001111012 Posts |
Quote:
https://www.mersenneforum.org/showpo...83&postcount=7 and https://www.mersenneforum.org/showpo...3&postcount=15 are about gpuowl versions and residue types. Last fiddled with by kriesel on 2019-12-03 at 01:00 |
|
|
|
|
|
|
#1505 |
|
"Mr. Meeseeks"
Jan 2012
California, USA
87816 Posts |
I've been having a weird issue with gpuowl, I have a system(RX570) which I run headless most of the time... When an assignment finishes, gpuowl will write the result then do nothing... until I login with RDP, then gpuowl immediately starts the next assignment. Tried running mfakto... zero issues.
Code:
2019-12-02 02:25:51 core 92912081 P2 2880/2880: setup 1128 ms; 5931 us/prime, 9223 primes
2019-12-02 02:25:51 core waiting for background GCDs..
2019-12-02 02:25:51 core 92912087 FFT 5120K: Width 256x4, Height 64x4, Middle 10; 17.72 bits/word
2019-12-02 02:25:51 core OpenCL args "-DEXP=92912087u -DWIDTH=1024u -DSMALL_HEIGHT=256u -DMIDDLE=10u -DWEIGHT_STEP=0x9.b3f5913600238p-3 -DIWEIGHT_STEP=0xd.311c9cb7274a8p-4 -DWEIGHT_BIGSTEP=0x9.837f0518db8a8p-3 -DIWEIGHT_BIGSTEP=0xd.744fccad69d68p-4 -DORIG_X2=1 -I. -cl-fast-relaxed-math -cl-std=CL2.0"
2019-12-02 02:25:53 core OpenCL compilation in 2060 ms
2019-12-02 02:26:39 core 92912081 P2 GCD: no factor
2019-12-02 02:26:39 core {"exponent":"92912081", "worktype":"PM1", "status":"NF", "program":{"name":"gpuowl", "version":"v6.11-11-gfaaa2f2"}, "timestamp":"2019-12-02 10:26:39 UTC", "user":"kracker", "computer":"core", "aid":"----", "fft-length":5242880, "B1":720000, "B2":13680000}
2019-12-02 06:07:51 core 92912087 P1 B1=720000, B2=13680000; 1038539 bits; starting at 0
2019-12-02 06:08:38 core 92912087 P1 10000 0.96%; 4698 us/sq; ETA 0d 01:21; b195a86475b0f7e5
|
|
|
|
|
|
#1506 |
|
"Robert Gerbicz"
Oct 2005
Hungary
101110011102 Posts |
|
|
|
|
|
|
#1507 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
153D16 Posts |
It's been reported that on Radeon VII, running two instances improves total throughput, and throughput per watt-hour.
I found a case where very different instances result in ~95% of single instance throughput. This case combines very different gpuowl versions, computation (LL vs. PRP3), exponent and so fft length. Windows 10, Lenovo Thinkstation D30, XFX Radeon VII; stock settings. gpuowl v0.6 alone 1.005ms/iter (50330737 LL DC, 4M fft) = 995. iter/sec alone v6.11 alone 1.193 ms/iter (89260099 PRP, 5M fft) = 838 iter/sec alone Two disparate instances run together: gpuowl v0.6 2.161 ms/iter; 463 iter/sec; throughput 463/995 = 0.4651 of solo; v6.11 2.458 ms/iter; 407 iter/sec; throughput 407/838 = 0.4855 of solo; combined, 0.4651 + 0.4855 = 0.9506 < 1. Slower running together, noticeably. Last fiddled with by kriesel on 2019-12-05 at 22:59 |
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| mfakto: an OpenCL program for Mersenne prefactoring | Bdot | GPU Computing | 1676 | 2021-06-30 21:23 |
| GPUOWL AMD Windows OpenCL issues | xx005fs | GpuOwl | 0 | 2019-07-26 21:37 |
| Testing an expression for primality | 1260 | Software | 17 | 2015-08-28 01:35 |
| Testing Mersenne cofactors for primality? | CRGreathouse | Computer Science & Computational Number Theory | 18 | 2013-06-08 19:12 |
| Primality-testing program with multiple types of moduli (PFGW-related) | Unregistered | Information & Answers | 4 | 2006-10-04 22:38 |