![]() |
|
|
#12 |
|
"William"
May 2003
New Haven
2·7·132 Posts |
|
|
|
|
|
|
#13 |
|
Jul 2003
So Cal
2×34×13 Posts |
Here are the results of measurements I did a short while ago. This was done on an 8-CPU 32-core Opteron system using shared memory. Although the scaling for CADO BW was slightly better than msieve, note that the BW algorithm is much less efficient than BL, so the overall run would take nearly twice as long at each point as msieve's BL.
|
|
|
|
|
|
#14 | |
|
Bamboozled!
"πΊππ·π·π"
May 2003
Down not across
2A0116 Posts |
Quote:
My experience with a MPI-implementation of BL running on a cluster with up to 32 processes was that performance scaled roughly as the square root of the number of processes. The machine architecture was different and maybe that's relevant. The MSR cluster had sixteen dual-cpu machines; each cpu had only one core and no hyperthreading. Paul |
|
|
|
|
|
|
#15 |
|
Noodles
"Mr. Tuch"
Dec 2007
Chennai, India
23518 Posts |
If the msieve Linear Algebra is working efficiently by using many computers in parallel over a closely coupled network connection, then the immediate next target should be M1061
![]() I would be glad to contribute some sieving for this number itself... I have a pending exercise in CUDA that I have to submit next week for which the next week is the deadline... I am learning about CUDA programming right now actually... but that people are not teaching anything at all... we have to learn everything on our own... NVIDIA CUDA... not much references available over the web online at all... This exercise is actually for effective parallel matrix multiplication... msieve with GPU (Graphical Processing Unit?) in clearly parallel execution MPI libraries are now being released... Good luck! Last fiddled with by Raman on 2010-03-15 at 11:22 |
|
|
|
|
|
#16 | |
|
Tribal Bullet
Oct 2004
3,541 Posts |
Quote:
Paul, I suspect that when the computations are all on one machine then the slowdown with many threads reflects caches getting saturated. The caches on your cluster were probably much larger in the aggregate. Also, it was never faster for msieve to find more than 64 dependencies, whereas IIRC the CWI parallel BL code liked 128 dependencies better than 64. |
|
|
|
|
|
|
#17 |
|
Jul 2003
So Cal
210610 Posts |
Paul, that's exactly what happens. Beyond 8 threads CADO BW gets no faster, and msieve actually slows.
I've also informally tried CADO BW on the beowulf cluster consisting of quad-core Core 2's connected by GigE and I saw similar behavior. But I seem to recall in that case speed got significantly worse once I moved from 4 threads to 5 because of communication speeds (4 threads uses shared memory on one node, 5 threads requires communication over GigE), and it never recovered. That cluster is due a memory upgrade soon, so I'll try it again with a bigger matrix when that happens. |
|
|
|
|
|
#18 | |
|
Noodles
"Mr. Tuch"
Dec 2007
Chennai, India
23518 Posts |
Quote:
It is hundred pages, cannot be read out entirely for a small example project. It is quite often rigorous, with mostly containing unnecessary or unwanted text, rather than demonstrations with examples. There are plenty of languages, could I learn them all? I could only do so that are very popular, and then are being more frequently used up. I am quite good at programming within C, C++, Java, Visual Basic. But can I say "good"? No. In the sense that I can't write programs as efficient as msieve, GGNFS, GMP-ECM, YAFU, etc. Of course, for those things understanding about algorithm is almost as important, with all those optimizations till now, as the algorithm gets much better everyday, right? Then, I cannot understand even the rigorous code of such projects. Where to start up with, actually? There is no limit upon learning around with the things in this world at all. I am quite pretty sure that even the top project developers do not know about everything within this world at all, actually! |
|
|
|
|
|
|
#19 | |
|
Bamboozled!
"πΊππ·π·π"
May 2003
Down not across
10,753 Posts |
Quote:
No-one has been able to understand everything of importance for several hundred years. All you can do is concentrate on the things you think are both important and interesting. Paul |
|
|
|
|
|
|
#20 | ||
|
Jun 2005
lehigh.edu
210 Posts |
Quote:
Quote:
and I'm just finishing up a 3-day 7.15M^2 matrix on our fast xeon SMP, two dual quadcores (SunFire x2270; "2x 2.93GHz Intel Xeon X5570 (Nehalem)"; "Connected using 1000Mbps Intel PRO/1000"). I stick to 4 threads (-t 4), and top shows that the job has accumulated Code:
2568m 2.5g 1184 R 401.0 5.2 12862:10 msieveH More precisely, "top -H" shows threads, and Code:
2568m 2.5g 1184 R 101.7 5.2 2892:31 msieveH 2568m 2.5g 1184 R 99.8 5.2 4012:01 msieveH 2568m 2.5g 1184 R 99.8 5.2 2975:16 msieveH 2568m 2.5g 1184 R 99.8 5.2 2976:37 msieveH three days, so far; with three other processes (17567, 17568, 17570) having contributed somewhat more than two cpu days (at 1440 cpu minutes/day). Actually, this is a fairly stunning read, with all four threads at c. 100%. I've gotten 3 + 3*2 = 9 cpudays, from 3 days on the four cores. (Ah, starting at 12:20 noon, read at 9:35; that much less than 3 days). As the other processors are sieving, even perfect scaling to 8 threads, one main, seven side processes wouldn't be worthwhile (I'm not in any particular hurry); but my recolllection was that the side processes also tail off, and the total cputime goes up. Four are worthwhile, eight not-so-much. Just a liberal arts point-of-view, as a math person. -bd |
||
|
|
|
|
|
#21 | |
|
Oct 2004
Austria
2·17·73 Posts |
Quote:
Off-topic question: Is there any windows equivalent to the linux command line "wc -l *.out" for counting relations in a bunch of relation files? |
|
|
|
|
|
|
#22 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
72×131 Posts |
You might want to look at the thread around
http://www.mersenneforum.org/showpos...2&postcount=14 and at http://www.mersenneforum.org/showthread.php?t=12861 |
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| msieve on KNL | frmky | Msieve | 3 | 2016-11-06 11:45 |
| Msieve on a Mac (Help) | pxp | Msieve | 1 | 2013-02-28 14:56 |
| Using msieve with c | burrobert | Msieve | 9 | 2012-10-26 22:46 |
| msieve help | em99010pepe | Msieve | 23 | 2009-09-27 16:13 |
| fun with msieve | masser | Sierpinski/Riesel Base 5 | 83 | 2007-11-17 19:39 |