![]() |
|
|
#474 | |
|
(loop (#_fork))
Feb 2006
Cambridge, England
2·7·461 Posts |
Quote:
|
|
|
|
|
|
|
#475 |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
22×13×107 Posts |
It's quite possible it was 11e6, now that you mention it. I think at the time I was checking different values, 11e3, 11e4, 11e5, etc. I might have gotten mixed up, but I was rather certain that I had made note that the 11e7 stage 1 on the GPU Colab instance took longer at 40 minutes than the ecmpi runs locally at 35 minutes. Hmm, maybe I stopped it at 40 minutes, because it took longer. My memory is really great, but for too short a spell. Always a jab to remind me to keep better notes. . .
|
|
|
|
|
|
#476 |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
22·13·107 Posts |
In playing with my Colab instance of GMP-ECM with GPU enabled (K80 GPU), I have made some observations:
The K80 shows 832 cores. If I run threads up to half of the cores, vs. over half the cores, the time for the run just about doubles. It is pretty close to constant within those halves. For example, running stage 1 for 5+3,1185L, it takes just about 22 seconds to complete <417 curves at 11e4. >416 curves takes about 41 seconds. Of note, I am using multiples of 32 cores. Running a single curve without the GPU takes about .4 second. Obviously, running more curves on a single CPU core would sequentially add time - 416 curves by CPU would take about 166 seconds. This appears to indicate that while the GPU sounds impressive with all its cores, a quad core CPU would keep up with the K80 GPU in stage 1 runs. Am I missing something here? |
|
|
|
|
|
#477 | |
|
"Bob Silverman"
Nov 2003
North of Boston
5·17·89 Posts |
Quote:
|
|
|
|
|
|
|
#478 | |
|
Just call me Henry
"David"
Sep 2007
Liverpool (GMT/BST)
17FD16 Posts |
Quote:
I seem to recall that it was possible to compile versions of gpu-ecm that ran faster with a limit of ~256 or ~512 bits. I am not sure whether ~768 worked. I think it needed to be a power of 2. |
|
|
|
|
|
|
#479 | |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
22·13·107 Posts |
Quote:
|
|
|
|
|
|
|
#480 |
|
Just call me Henry
"David"
Sep 2007
Liverpool (GMT/BST)
137758 Posts |
Be careful if you go down the route of trying gpu-ecm with different limits. I think there was some concern about whether it was appearing to work but wasn't finding all expected factors.
|
|
|
|
|
|
#481 | |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
22×13×107 Posts |
Quote:
Thanks! |
|
|
|
|
|
|
#482 |
|
"Curtis"
Feb 2005
Riverside, CA
2·2,927 Posts |
Henry is right- GPU-ECM is the same speed with any input within its bounds (1019 bits or smaller). So, it's relatively a waste of time for small numbers, but quite nice for 900+ bit inputs. When I had a working CUDA setup, I used it for lots of 900-1000 bit inputs. Even in your test case, it doubles the production of a quad-core by doing stage1 on GPU and stage2 on CPU (with a couple cores left over, since stage 2 runs faster than stage 1).
|
|
|
|
|
|
#483 | |
|
(loop (#_fork))
Feb 2006
Cambridge, England
645410 Posts |
Quote:
(though my 1080Ti is specified as 3584 cores and runs 1792 curves at a time, so there's a small constant factor there in any case) |
|
|
|
|
|
|
#484 | |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
22×13×107 Posts |
Thanks for all the info, everyone.
Quote:
I think one of the times I got a P100, it showed 4000+ with my ECM run. Since I only get two Xeon cores to back up all the GPU stage 1s, unless I pull the residues off, it's kind of less than ideal to run much in a GPU-ECM Colab session. |
|
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Running CUDA on non-Nvidia GPUs | Rodrigo | GPU Computing | 3 | 2016-05-17 05:43 |
| Error in GMP-ECM 6.4.3 and latest svn | ATH | GMP-ECM | 10 | 2012-07-29 17:15 |
| latest SVN 1677 | ATH | GMP-ECM | 7 | 2012-01-07 18:34 |
| Has anyone seen my latest treatise? | davieddy | Lounge | 0 | 2011-01-21 19:29 |
| Latest version? | [CZ]Pegas | Software | 3 | 2002-08-23 17:05 |