![]() |
|
|
#1915 |
|
Jun 2012
17 Posts |
Hi Owftheevil,Manpowre
My recomended mod to the threads in invocation of the kernel does not give the right results. It for example says M110503 is composite instead of prime. So i would suggest leaving that at 128 until someone who understands the cudalucas 2.05Alpha code comes along and joins in the frey. However my mod for Fermat numbers can be done in version 2.03 and 2.05 alpha if n is calculated in code as 2^2^n+1 instead of 2^p-1. Does any one know which lines of code that n=2^p-1 is calculated or is it inherent in the cufft calls? Thank you! PS For example 2.03 version runs M61787581 in ETA 60 hrs. will check for a few days if that is prime? Also to be added to parse.c is trial division by small primes of Mp. This can be done simply in parse.c by implementing it in GMP. Also is there any GPU code that does a Pepins' test on Fermat numbers? |
|
|
|
|
|
#1916 |
|
"Carl Darby"
Oct 2012
Spring Mountains, Nevada
32·5·7 Posts |
Well, I certainly understand R.D. Silverman better now.
|
|
|
|
|
|
#1917 | |
|
Jul 2003
So Cal
22·32·59 Posts |
Quote:
|
|
|
|
|
|
|
#1918 | |
|
Romulan Interpreter
Jun 2011
Thailand
100101101111112 Posts |
Quote:
James' site is a perfect starting point. You can sort it by other columns too, by clicking on the column head (:P) [edit: for TF, as I did not see the first time that you asked for TF, the link is here. Read about the last two columns, very good comparison criteria!] Last fiddled with by LaurV on 2013-05-29 at 06:45 |
|
|
|
|
|
|
#1919 | |
|
"Svein Johansen"
May 2013
Norway
3·67 Posts |
Quote:
The only way I can get to the iterations as low as less than 4ms each, is to take memory clock back up to stock, but that doesnt make the Titan stable as I understand the memory on the back side of the card heats up too much. |
|
|
|
|
|
|
#1920 | |
|
"Mr. Meeseeks"
Jan 2012
California, USA
41708 Posts |
Quote:
Also, on James's site I saw that the 7970 GHz beats the 580... is that true?
|
|
|
|
|
|
|
#1921 |
|
"James Heinrich"
May 2004
ex-Northern Ontario
D6316 Posts |
|
|
|
|
|
|
#1922 |
|
Romulan Interpreter
Jun 2011
Thailand
3·3,221 Posts |
Indeed. If you look for the results on the bitcoin forums, I always wondered why Radeons did not beat Nvidias at TF, from the very beginning...
They are better at integer math, I mean if you don;t ask them to do much DP calculus, but as soon as someone port the FFT package to OpenCL - there is already one done by Apple, I posted a link somewhere - they may surprise us with LL test's speed too... It hurts me a bit to say that, I am pure Nvidia/cuda guy...
|
|
|
|
|
|
#1923 | |
|
"James Heinrich"
May 2004
ex-Northern Ontario
23·149 Posts |
Quote:
The FFT benchmark is useful for deciding which FFT size to use for a given exponent, but it's not very useful for helping to predict performance. |
|
|
|
|
|
|
#1924 | |
|
"Carl Darby"
Oct 2012
Spring Mountains, Nevada
32·5·7 Posts |
Quote:
This would also be useful for CUDALucas itself. There could be cases where shorter ffts take longer, but yield better iteration times. |
|
|
|
|
|
|
#1925 | |
|
"James Heinrich"
May 2004
ex-Northern Ontario
65438 Posts |
Quote:
![]() If you have any suggestions for methodology for determining suitable exponents then I don't mind preparing that list for you. My method would be guess-and-check: try different exponents and see what FFT size is selected, populate my chart, and then try and fill in the gaps by guess-and-checking at what exponent would best fill the in-between FFT sizes. Would that work? |
|
|
|
|
![]() |
| Thread Tools | |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Don't DC/LL them with CudaLucas | LaurV | Data | 131 | 2017-05-02 18:41 |
| CUDALucas / cuFFT Performance on CUDA 7 / 7.5 / 8 | Brain | GPU Computing | 13 | 2016-02-19 15:53 |
| CUDALucas: which binary to use? | Karl M Johnson | GPU Computing | 15 | 2015-10-13 04:44 |
| settings for cudaLucas | fairsky | GPU Computing | 11 | 2013-11-03 02:08 |
| Trying to run CUDALucas on Windows 8 CP | Rodrigo | GPU Computing | 12 | 2012-03-07 23:20 |