![]() |
![]() |
#2773 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
2×29×127 Posts |
![]() Quote:
Also, investigate why your system is having errors that make restart from last checkpoint an issue. Test the cpu/system side memory as well as the gpu side, etc. Does prime95/mprime work reliably on it? SaveAllCheckpoints=1 reportiterations=100000 checkpointiterations=500000 checkpoints at 500000 iteration intervals as intended (up to 6pm 4/12/19 in my test folder) reduced checkpointiterations to 200000 in ini file and restarted; increased reportiterations to 500000 via Y keyboard input twice; checkpoints then occurred at 1M intervals (which is when screen and checkpoint nominal intervals coincide; least common multiple) (up to 1am 4/13/19 in my test) reduced reportiterations to 200000 via y keyboard input at 1am checkpoint occurred at 400K after the last one at 102am. (the software correctly did a checkpoint at the next n*200k point after the change) after that checkpoints and screen reports are at 200000 intervals as intended (UP TO 7:16AM) reduced reportiterations to 100000 via y keyboard input at 716am, gave checkpoints at 200k, screen reports at 100k as intended Last fiddled with by kriesel on 2019-04-13 at 14:20 |
|
![]() |
![]() |
![]() |
#2774 |
Einyen
Dec 2003
Denmark
343410 Posts |
![]()
Anyone know what causes this error which stops the CUDALucas run?
Code:
CUDALucas.cu(1989) : cudaSafeCall() Runtime API error 6: the launch timed out and was terminated. Resetting device and restarting from last checkpoint. CUDALucas.cu(1115) : cudaSafeCall() Runtime API error 46: all CUDA-capable devices are busy or unavailable. |
![]() |
![]() |
![]() |
#2775 | |||
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
11100110001102 Posts |
![]() Quote:
Edit: better search turned up https://www.mersenneforum.org/showpo...postcount=2207 and owftheevil responds in https://www.mersenneforum.org/showpo...postcount=2208 Quote:
Quote:
Last fiddled with by kriesel on 2019-04-13 at 17:32 |
|||
![]() |
![]() |
![]() |
#2776 | |
"Ghetto_Child"
Jul 2014
Montreal, QC, Canada
41 Posts |
![]()
@kriesel:
I have not tested 2.06 yet, I get 1-5 errors per day on a single instance of 2.05.1 when I run more than one instance simultaneously; (currently running a 4M FFT and 14M FFT exponents each instance). Error are more likely to trigger when I run the Chrome browser while the two instances are processing. A single instance can go many days without errors unless ambient temps get too high. Errors are not difficult to manage or avoid, they frequently go away when the instance retests the iteration in question. It's just the size of reprocessing or how far back it restarts that can be a huge waste. I tested again the difference between using only ini file checkpoint and screen output intervals vs specifying switches/flags in command-line. Results are the same, it uses the screen output value instead of checkpoint value no matter which one is smaller; and no matter if the values are specified only in the ini file or in both ini file and command-line. Again this is on version 2.05.1 as I have not used 2.06 yet but I will try the newer one later tonight. Quote:
Last fiddled with by GhettoChild on 2019-04-14 at 02:51 |
|
![]() |
![]() |
![]() |
#2777 |
"Ghetto_Child"
Jul 2014
Montreal, QC, Canada
4110 Posts |
![]()
I get identical results in v2.05.1 and v2.06beta. The screen output iteration value overrides the checkpoint value no matter if specified in the .ini file or in the command-line or in both simultaneously. If I set screen output to 500,000 and checkpoint to 100,000, the checkpoint will only update every 500,000.
I decided to use the checkpoint files from 2.05.1 to continue processing in 2.06beta. I didn't want to toss weeks of processing without confirmation they are incompatible. Right now I have CUDAPm1 v0.20 CUDA5.5 running a 14432K FFT and CUDALucas v2.06beta CUDA8.0. I can't run CUDA10 because it makes a crt dll error prompt. Similar with CUDAPm1 v.22. |
![]() |
![]() |
![]() |
#2778 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
2·29·127 Posts |
![]() Quote:
Continuing with CUDALucas v2.06 from a V2.05.1 start or even slightly earlier is no problem. (Mfaktc version or class count change will restart from the beginning of a bit level.) I recommend NOT running large fft lengths until you get your rig and config more stable at modest fft lengths. (Note, the self tests only go up to 8M fft length.) Running CUDAPm1 on a 4GB gpu along with another CUDA GIMPS program is probably asking for trouble, especially in a host-system-RAM-starved system. CUDAPm1 is described as ALPHA software (less mature than beta, which is less mature than a release). P-1 stage 2 is the most memory-hungry computation type there is in GIMPS and will try to use most or all the gpu ram. CUDAPm1 checks available gpu memory early in a run, and apparently assumes an equal amount will be available hours or days later. It decides how many primes between B1 and B2 to work on simultaneously, at the beginning of stage 2, based partly on available memory. If another program is comiing and going and varying requirements, and therefore varying the amount of memory available, that could be a problem. It sounds like the issues you're having are more than negating any plausible slight gain from running multiple applications. Multiple instances on one gpu seems to work with mfaktc and provide some aggregate throughput increase. Mfaktc does not need much gpu memory or system memory. What vintage is your computer? Three GB sounds like too little ram to me, and consistent with perhaps a Core2 Duo system. Running prime95 primality tests on old systems can be very expensive per test (~$12/85M primality test without any restarts), much more than cloud computing costs, even if only considering the electricity. I've bought used system boxes (workstation grade, solidly built, well documented and labeled, toolfree maintenance, dual-Xeon, Windows OS, etc) with 12GB to 128GB for US$200-800 and some included an older gpu or two in that; capable of doing the same test for a lot less $. I don't think I have any hardware with less host system ram than the ram contained in an individual gpu installed in it. Asking the OS and driver to reliably handle gpu to host transfers spilling directly to the paging file seems risky to me. |
|
![]() |
![]() |
![]() |
#2779 | |
"Marv"
May 2009
near the Tannhäuser Gate
22×3×67 Posts |
![]() Quote:
What video board and what exponent are you using? Does this happen soon after starting? Windows/linux? If windoze, could a virus check or update be going on? I have seen this when too much is going on in the system. |
|
![]() |
![]() |
![]() |
#2780 |
Einyen
Dec 2003
Denmark
2×17×101 Posts |
![]()
It is an old Titan Black and it mostly happened when I used CUDA 5.0 version of CUDALucas, so I stopped using that version. But I'm about to scrap this card now, it keeps giving bad results randomly even though when I test it with GPUmemtest and CUDALucas extended self tests and mfaktc extended self tests there are no errors.
|
![]() |
![]() |
![]() |
#2781 |
"Ghetto_Child"
Jul 2014
Montreal, QC, Canada
518 Posts |
![]()
Core 2 Quad Q6600. It was a test install that became my main PC; when my install medium died during install and my main PC (same generation) had a main drive die. My PC is only pulling 375Watts according to the UPS. 4yrs ago I was running a GTX 295 the same way and it was pulling 350Watts. Electricity price in my home is arbitrary but I don't have air-conditioning so the power consumption is no worse that those with A/C. I turn off all this high drain number processing during heat waves.
As for why I want really large/slow screen outputs and much smaller/faster checkpoint updates. I run these CUDA programs for days and weeks at a time. I observe performance, errors and ETA. I don't want to scroll through pages and pages of screens full of output lines per instance/program to observe its performance trend. 100,000/hr for 3-5days and more of output. That's a waste. |
![]() |
![]() |
![]() |
#2782 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
736610 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#2783 |
"Jerry"
Nov 2011
Vancouver, WA
1,123 Posts |
![]()
Hello all!
I compiled a new windows CUDALucas 2.06 for all CUDA 4.0 - 10.1. Support for win32 ended at CUDA 6.5. No big changes, removed nmvl requirement and made small updates for new CUDA versions. 2.06 is here Lib files are here, if you need them I was able to do basic testing, but please test thoroughly before using for production. Everyone, let me know what you find that needs to be fixed and what you would like changed. ~Cheers ![]() |
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Don't DC/LL them with CudaLucas | LaurV | Data | 131 | 2017-05-02 18:41 |
CUDALucas / cuFFT Performance on CUDA 7 / 7.5 / 8 | Brain | GPU Computing | 13 | 2016-02-19 15:53 |
CUDALucas: which binary to use? | Karl M Johnson | GPU Computing | 15 | 2015-10-13 04:44 |
settings for cudaLucas | fairsky | GPU Computing | 11 | 2013-11-03 02:08 |
Trying to run CUDALucas on Windows 8 CP | Rodrigo | GPU Computing | 12 | 2012-03-07 23:20 |