![]() |
|
|
#2663 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
31·173 Posts |
Quote:
Look at NVIDIA developer support or download sites for nvml.h, and the linux equivalent of the nvml.dll required on Windows? Note from flashjh at https://sourceforge.net/p/cudalucas/code/102/ says: Updates for GPU UUID from NVML. nvml.dll is required and included with the .7z file. Have not compiled or tested linux yet. Also included code to exclude 1.x GPUs for CUDA >7.0 and future code to exclude 2.x when CUDA does not support. https://download.mersenne.ca/ appears not to have any linux CUDALucas executables. https://sourceforge.net/projects/cudalucas/files/ has a tgz file of precompiled V2.05.1 for linux, which is from before the added code requiring nvml. Or see https://stackoverflow.com/questions/...r-file-missing from some years back, about the effects of a missing nmvl.h. Good luck! Last fiddled with by kriesel on 2018-04-04 at 18:46 |
|
|
|
|
|
|
#2664 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
14F316 Posts |
Quote:
It appears to me that different apps handle the jumpy versus averaging ETA differently. Prime95 appears to take both forks; worker window ETA fluctuates with current time per iteration and user behavior, while Status appears to be the long term averaging over the history of the exponent run approach. In my experience this often results in the actual progress beating the Status prediction to completion by days or sometimes weeks. I find the more responsive indication more useful. Others may not. As far as requesting too much, I prefer to think of the lists I've made as representing abundant opportunity. Any would be code author or maintainer can choose how to spend his time, what his personal priorities are, and what to tackle. Any progress that's made and shared benefits all the users of a software title and the project. |
|
|
|
|
|
|
#2665 |
|
Basketry That Evening!
"Bunslow the Bold"
Jun 2011
40<A<43 -89<O<-88
722110 Posts |
Right, the subsequent posts about "generic he" have been moved to a more appropriate location.
|
|
|
|
|
|
#2666 | |
|
27248 Posts |
Quote:
With mprime I have noted that even on a dedicated system, with no user input, the ETA fluctuates with system load, obviously because of the system daemons that periodically occupy the CPU. Theoretically with maximum priority set for mprime, still the ETA should be fluctuating. |
|
|
|
|
#2667 | ||
|
Mar 2018
Shenzhen, China
2·32 Posts |
Quote:
I downloaded file CUDALucas.cu of version 2.05.1 (r94) and the source finally compiled. I already confirmed the executable works well. As for the precompiled executables, they just won't run on my machine: Quote:
|
||
|
|
|
|
|
#2668 | |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
31·173 Posts |
Quote:
Reviewing my offline notes, I see there was a stretch of posts 2642-2648 about the same error message you encountered. Note that success on one fft length, or one run, while a good sign, does not ensure success on another. Running the available residue checks on the fft lengths you're likely to use is recommended. (See the -r option.) Running a separate gpu memory test program, or thorough coverage with CUDALucas's -memtest option, is also a useful precaution. There are certain known-bad residues that sometimes appear, indicating a problem. Continuing an exponent run after any such appear is a waste of processing time. Restarting from a saved interim file from before such appear is recommended. For CUDALucas, these known problem residues are: 0x0000000000000000, 0x0000000000000002, 0xfffffffffffffffd 0x0 is ok for a final residue if it does not occur before then; appearing only as the final residue is the sign of a rare prime result. Appearing earlier is a sign of something gone wrong, perhaps a host to gpu or gpu to host failed memory copy. Residues that repeat from one iteration to the next would also be a symptom of a problem. As I recall, the April 18 2017 2.06beta did not include the nvml requirement, but did include at least some checks for some of those known-bad residues. I suggest you try compiling that version for more likely valid results. Please consider making your compiled executable(s) available to James Heinrich for posting on his software mirror. http://download.mersenne.ca/CUDALucas/ For performance tuning, if you haven't yet, you may want to run -fftbench and -threadbench for the range of fft lengths you're likely to use. I would use at least 3 passes on a fast card like yours, and expand the range outward to powers of two including the intended usage. For example, if intending to use 4608k-7168k, benchmark fft and threads over 4096k-8192k. You may find skimming through the bug and wish list useful; last posted http://www.mersenneforum.org/showpos...postcount=2658 Please note any other bugs you might find, or features that would be useful, and I'll add them to the list. Also a run-time scaling guide is available at post 2659. Last fiddled with by kriesel on 2018-04-07 at 18:32 |
|
|
|
|
|
|
#2669 | ||
|
Sep 2009
1000000111102 Posts |
Quote:
Quote:
Code:
ls -l CUDALucas-2.05.1-CUDA6.5-linux-x86_64 file CUDALucas-2.05.1-CUDA6.5-linux-x86_64 ldd CUDALucas-2.05.1-CUDA6.5-linux-x86_64 uname -a Chris |
||
|
|
|
|
|
#2670 |
|
Apr 2018
1 Posts |
Just a tiny FYI on what was needed to build the latest CUDALucas from source on RHEL7.4 with CUDA 9.1.
I had to add "#include <nvml.h>" to CUDALucas.cu and also it needed to be linked with /usr/local/cuda/targets/x86_64-linux/lib/stubs/libnvidia-ml.so to the final build command was: gcc CUDALucas.o parse.o -O1 -Wall -fPIC -L/usr/local/cuda/lib64 -L/usr/local/cuda/targets/x86_64-linux/lib/stubs/ -lcufft -lcudart -lnvidia-ml -lm -o CUDALucas and for ppc64le it was approximately the same command: gcc CUDALucas.o parse.o -O1 -Wall -fPIC -L/usr/local/cuda/lib64 -L/usr/local/cuda/targets/x86_64-linux/lib/stubs/ -lcufft -lcudart -lnvidia-ml -lm -o CUDALucas oh.. and in the Makefile I also specified "--generate-code arch=compute_50,code=sm_50" which is the capability of my 940MX, and compute_60,cm_60 for the Tesla P100. On the ppc64le I am seeing some strange stuff (crash on -threadbench) or if I don't specify "-threads". But it seems to be working fine for LL. |
|
|
|
|
|
#2671 | ||||
|
Mar 2018
Shenzhen, China
100102 Posts |
Quote:
Quote:
Quote:
Quote:
|
||||
|
|
|
|
|
#2672 | ||
|
Mar 2018
Shenzhen, China
2×32 Posts |
Quote:
Since the NVIDIA Management Library (nvidia-ml, nvml) was in different location on my machine, my build commands were: Quote:
Last fiddled with by Lexicographer on 2018-04-09 at 04:55 |
||
|
|
|
|
|
#2673 |
|
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
536310 Posts |
A recovery method tested the last few days on CUDAPm1 and GTX480 may also apply here. See the detailed writeup at http://www.mersenneforum.org/showpos...8&postcount=37
|
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Don't DC/LL them with CudaLucas | LaurV | Data | 131 | 2017-05-02 18:41 |
| CUDALucas / cuFFT Performance on CUDA 7 / 7.5 / 8 | Brain | GPU Computing | 13 | 2016-02-19 15:53 |
| CUDALucas: which binary to use? | Karl M Johnson | GPU Computing | 15 | 2015-10-13 04:44 |
| settings for cudaLucas | fairsky | GPU Computing | 11 | 2013-11-03 02:08 |
| Trying to run CUDALucas on Windows 8 CP | Rodrigo | GPU Computing | 12 | 2012-03-07 23:20 |