![]() |
|
|
#2509 |
|
Mar 2010
19B16 Posts |
It would seem that despite the severely limited fp64 performance of high-end Pascal GPUs like 1080 and Titan X, they're a match for the Kepler Titans after all.
I suspect it has something to do with higher memory clocks & bandwidth, two generations gap and lower fab ![]() Of course, the GP100 PCI-E Tesla GPU, which is scheduled to be released Q4 2016, should be even better at dp fp calculus, given its fp64 performance equal to 1/2 of fp32, but it will be pricey as usual. |
|
|
|
|
|
#2510 | |
|
Aug 2010
Republic of Belarus
2×89 Posts |
Quote:
|
|
|
|
|
|
|
#2511 |
|
Mar 2010
3·137 Posts |
Hiya Lorenzo!
Here's the benchmark data: Code:
Using threads: square 256, splice 256. Starting M332220523 fft length = 19208K | Date Time | Test Num Iter Residue | FFT Error ms/It Time | ETA Done | | Aug 27 13:02:36 | M332220523 3000 0x3821367649db99f0 | 19208K 0.11328 18.3241 0:19 | 72:16:28:31 0.00% | | Aug 27 13:02:54 | M332220523 4000 0xd257f742cbc095ef | 19208K 0.11328 18.3239 0:18 | 72:03:05:48 0.00% | | Aug 27 13:03:13 | M332220523 5000 0xb6a474948e0d0da4 | 19208K 0.12109 18.3275 0:18 | 71:19:08:00 0.00% | | Aug 27 13:03:31 | M332220523 6000 0x553f2f945aac8b5f | 19208K 0.12109 18.3260 0:19 | 71:13:48:01 0.00% | | Aug 27 13:03:49 | M332220523 7000 0x6f41d82484cef29c | 19208K 0.12109 18.3255 0:18 | 71:09:58:59 0.00% | | Aug 27 13:04:08 | M332220523 8000 0x1fa05e9cbb4a8709 | 19208K 0.12109 18.3265 0:18 | 71:07:07:49 0.00% | | Aug 27 13:04:26 | M332220523 9000 0x53b63e5e6e665bf6 | 19208K 0.12109 18.3228 0:19 | 71:04:52:21 0.00% | Guess non-Tesla Pascals can't shine everywhere at once, but it's too early to say without a different compute capability benchmark, namely Titan X (P). |
|
|
|
|
|
#2512 | |
|
Sep 2010
So Cal
2·52 Posts |
Quote:
|
|
|
|
|
|
|
#2513 |
|
Mar 2010
6338 Posts |
2.516 ms/iter for M76 makes Titan X faster than Titan Black
![]() And since the residues match, well, congrats, money well spent! If only it had 1/2 fp64 performance of fp32, instead of 1/32. |
|
|
|
|
|
#2514 | |
|
Aug 2010
Republic of Belarus
2×89 Posts |
Quote:
|
|
|
|
|
|
|
#2515 | |
|
Sep 2010
So Cal
5010 Posts |
Quote:
|
|
|
|
|
|
|
#2517 | |
|
"David"
Jul 2015
Ohio
10058 Posts |
Quote:
|
|
|
|
|
|
|
#2518 |
|
Banned
"Luigi"
Aug 2002
Team Italia
32·5·107 Posts |
Anybody tested mfaktc for the 980? smile:
|
|
|
|
|
|
#2519 | |
|
"Kieren"
Jul 2011
In My Own Galaxy!
2·3·1,693 Posts |
(tangential to the current discussion)
I recently tried again to find settings that would work on my MSI 580. I got good -memtest results, but it never finished -r 1. I went as low as 775 MHz core (Factory OC 833), 1500 MHz RAM, with substantially increased Vcore. What I don't know is if the string of error 30 results which ended these attempts were error-driven, or if this was just another timeout situation. This card is still great with mfaktc, and can run at 900 MHz or higher doing that. Meanwhile, the faithful Gigabyte 460 is OC to 833 MHz core, RAM clocked down to 1700. It has churned out a 40.9M DC every 3.5 days for weeks, with perhaps one mismatch, which I haven't checked back on. It never does the timeout thing. Is that only 580s? I just ran -r 1 again, with the settings above. It lasted some minutes, but ended this way: Quote:
I wish I could get this card to do CuLu, but previous experience has not been good, with proven bad results on some runs. Last fiddled with by kladner on 2016-08-31 at 19:45 |
|
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Don't DC/LL them with CudaLucas | LaurV | Data | 131 | 2017-05-02 18:41 |
| CUDALucas / cuFFT Performance on CUDA 7 / 7.5 / 8 | Brain | GPU Computing | 13 | 2016-02-19 15:53 |
| CUDALucas: which binary to use? | Karl M Johnson | GPU Computing | 15 | 2015-10-13 04:44 |
| settings for cudaLucas | fairsky | GPU Computing | 11 | 2013-11-03 02:08 |
| Trying to run CUDALucas on Windows 8 CP | Rodrigo | GPU Computing | 12 | 2012-03-07 23:20 |