![]() |
![]() |
#1 |
"Ed Hall"
Dec 2009
Adirondack Mtns
3·1,229 Posts |
![]()
Back in the Pentium4 days I was able to get multiple machines to run msieve LA (-nc2) and save time.
Recently I revisited this with a couple i7s to see if there could be a gain of anything. The answer that we all probably already knew, is not with Gigabit. With a relations set that took ~10 hours on one machine, it showed ~12 hours to ~22 hours (after settling) for the two machines, depending on various thread/grid combinations. The one area that might be of use is if I can increase my memory capability by using two machines. This was something brought up by VBCurtis a while back. I haven't explored that to any extent, but if I need more than the 16G of one machine, perhaps I can use two and only lose some time (although possibly as much as 20%). |
![]() |
![]() |
![]() |
#2 |
"Curtis"
Feb 2005
Riverside, CA
10010100000012 Posts |
![]()
1. Older-generation infiniband cards are not very expensive; if one were serious about tackling jobs too large for one's best machine, a pair of IB cards with a single cable should net good msieve-MPI results.
2. I think I have openMPI installed and configured on my Z620, but msieve-MPI (self-compiled) does not function. Might you email me your msieve-MPI linux binary? |
![]() |
![]() |
![]() |
#3 | |
"Ed Hall"
Dec 2009
Adirondack Mtns
1110011001112 Posts |
![]() Quote:
2. As to openmpi, if you're running Ubuntu 18.04, the repository openmpi is broken. It will not work if you try to use more than the localhost, which makes it rather useless. I've tried installing the latest version from the source site, but never got it to run, either. This is actually keeping me from upgrading all my 16.04 machines. I will try to search out the binaries and send them your way. I know there are three main files to install - openmpi-common, openmpi-bin and libopenmpi-dev. Give me a day or so. Locate and whereis didn't turn up anything. |
|
![]() |
![]() |
![]() |
#4 |
"Curtis"
Feb 2005
Riverside, CA
3·1,579 Posts |
![]()
I believe I am running 18.04 on the Z620; though for my use case localhost-only does help, as others have found that using MPI for each socket is much faster than using a single pool of threads. I'm running -t 20 right now on the dual-10-core, where I'd like to be running MPI 2x10-threads.
Good to know it's likely openMPI that I should try to address, rather than my build of msieve. |
![]() |
![]() |
![]() |
#5 | |
"Ed Hall"
Dec 2009
Adirondack Mtns
3·1,229 Posts |
![]() Quote:
openmpi-bin, openmpi-common and libopenmpi-dev installed, and then compile msieve, with MPI=1, it should work as long as you don't include a hostfile. |
|
![]() |
![]() |
![]() |
#6 |
Tribal Bullet
Oct 2004
24×13×17 Posts |
![]()
If you are working with MPI over multiple machines, can you see if the postprocessing runs to completion using the latest Msieve svn? Greg has reported that it does not (the square root doesn't like the data it's given). It doesn't have to be for a large problem, just a C100 on a 1x2 or 2x2 grid would be enough.
Last fiddled with by jasonp on 2019-11-04 at 15:22 |
![]() |
![]() |
![]() |
#7 | |
"Ed Hall"
Dec 2009
Adirondack Mtns
3×1,229 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#8 | |
"Ed Hall"
Dec 2009
Adirondack Mtns
3·1,229 Posts |
![]() Quote:
Code:
. . . Mon Nov 4 20:30:25 2019 reading relations for dependency 63 Mon Nov 4 20:30:25 2019 read 0 cycles Mon Nov 4 20:30:25 2019 reading relations for dependency 64 Mon Nov 4 20:30:25 2019 read 0 cycles Mon Nov 4 20:30:25 2019 sqrtTime: 119 Mon Nov 4 20:30:25 2019 elapsed time 00:02:00 |
|
![]() |
![]() |
![]() |
#9 |
"Ed Hall"
Dec 2009
Adirondack Mtns
3·1,229 Posts |
![]() A run without mpi worked fine:
Code:
. . . Mon Nov 4 20:49:50 2019 initial square root is modulo 4203473 Mon Nov 4 20:49:56 2019 sqrtTime: 28 Mon Nov 4 20:49:56 2019 p50 factor: 26221114229909593079493944061795669970670518225931 Mon Nov 4 20:49:56 2019 p51 factor: 116701385250110252294900265085979409741229672958141 Mon Nov 4 20:49:56 2019 elapsed time 00:00:28 To note, I ran -nc1 once and saved all the files to another directory. Then I copied them into the working directory before each -nc2 run. |
![]() |
![]() |
![]() |
#10 | ||
"Ed Hall"
Dec 2009
Adirondack Mtns
3·1,229 Posts |
![]() Quote:
Quote:
![]() 2. I was confused as to which binary(ies) you were interested in. Rather troubling, since you were specific! I'm assuming you're not interested right now, since the mpi msieve seems to be having some trouble, per previous posts and I believe your issue is really the 18.04 openmpi issue? |
||
![]() |
![]() |
![]() |
#11 |
"Curtis"
Feb 2005
Riverside, CA
3·1,579 Posts |
![]()
Agree to both on #2; the 18.04 openMPI is likely the problem, and the msieve bug means I shouldn't be trying it on the C207 matrix. My original intent was for you to send an msieve binary, because it didn't occur to me that MPI might be broken.
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
ecmpi with openmpi on Ubuntu? | EdH | GMP-ECM | 2 | 2020-10-05 16:25 |
Is Anyone Here Successfully Using OpenMPI with Ubuntu 18.04? | EdH | Linux | 0 | 2019-09-13 02:03 |
More openmpi questions... | EdH | Msieve | 5 | 2017-01-16 17:22 |
Running msieve LA with openmpi - do all machines need to be same/similar | EdH | Msieve | 32 | 2013-11-08 17:57 |
Current status | fivemack | NFSNET Discussion | 90 | 2006-11-13 13:37 |