![]() |
![]() |
#232 |
"Curtis"
Feb 2005
Riverside, CA
3·29·53 Posts |
![]()
As Q rises, our sec/rel performance falls. By the time we get to Q=200M, performance will be half of what it was at Q=20M.
That doesn't fully explain the performance difference, though, since we're only at 40M now (we started at 8M). It's possible that your current instance is sharing the machine with something that uses hyperthreads more; "CPU time" that CADO records doesn't to my knowledge know anything about HT, so running 8 HT threads on a 4-core will "look" 75% slower, but actually be 25% faster (not verified percentages). Finally, you may be on an older architecture. Compare the performance of the machine "9800" (a 9800 processor, current generation) to "Supercomputer" (haswell i7-5820k), "TheMachine" (ivy bridge xeon, running only a few more threads than cores), or "z600" (sandy bridge xeon, running ~20 threads on 12 cores). My guess is your speed difference is 20% from larger Q values, 80% from different architecture / sharing the machine. Workunits doubled in size around Q=20M. That's why all the "unlucky" badges are from people who began running before we changed WU size. |
![]() |
![]() |
![]() |
#233 |
Sep 2008
Kansas
23×34×5 Posts |
![]() |
![]() |
![]() |
![]() |
#234 |
"Luke Richards"
Jan 2018
Birmingham, UK
25·32 Posts |
![]() |
![]() |
![]() |
![]() |
#235 |
"Curtis"
Feb 2005
Riverside, CA
461110 Posts |
![]() |
![]() |
![]() |
![]() |
#236 |
"Luke Richards"
Jan 2018
Birmingham, UK
28810 Posts |
![]()
Have tweaked some settings and using a clientid of lukerichards-pre1-test1 for a day or two. Will probably check in on Sunday to see how that has fared.
|
![]() |
![]() |
![]() |
#237 |
"Carlos Pinho"
Oct 2011
Milton Keynes, UK
10010111011112 Posts |
![]()
Anyone knows if it is quicker to run individual instances rather than multi threads ( call 8x with t 1 instead of t 8). TIA
|
![]() |
![]() |
![]() |
#238 |
"Curtis"
Feb 2005
Riverside, CA
10010000000112 Posts |
![]()
The siever is designed to be multi-threaded, and the documentation explains that CADO defaults to -t 2 because there is no penalty.
I think there is a small penalty for running, say, 20 threads rather than 2 or 4, and I think that penalty is related to the size of the factor base/size of the job. I wouldn't run 4-threaded sievers for a C110, and my guess is that on this C207 multiple 4- or 8-threaded instances would be a bit faster than a single 16+ threaded instance. Fivemack did some testing on a C193 in the CADO thread, but I didn't see clear evidence of perfect scaling to megathreads; then again, he's smarter than I am, so perhaps he's convinced and this job is fine up to 16 or 20 threads per job. It may be that they're the only ones making full use of HT and are also running older architecture, but I note that Vebis and birch4 are the slowest relations-per-cpuday on the stats, and both are running 16+ threads per instance. |
![]() |
![]() |
![]() |
#239 |
"Carlos Pinho"
Oct 2011
Milton Keynes, UK
113578 Posts |
![]()
Another question: how to run the client without having to have the terminal window open. This is very important
![]() |
![]() |
![]() |
![]() |
#240 |
"Carlos Pinho"
Oct 2011
Milton Keynes, UK
10010111011112 Posts |
![]()
From stats I’m also one of the slowest rels/sec running 8 threads per instance. Maybe I’ll try 6 or even stay with 4 since I suppose I’m reaching bandwidth limit. BTW, ETA will soon drop, stay tuned.
Edit: Not sure what happened, double posted. Upper one can be deleted. Last fiddled with by pinhodecarlos on 2019-06-16 at 09:28 |
![]() |
![]() |
![]() |
#241 |
"Curtis"
Feb 2005
Riverside, CA
3×29×53 Posts |
![]()
I use "screen" before the cado invocation, and then ctrl-a, ctrl-d to detach it from the terminal window.
8 threads on a HT 4-core machine is likely faster than 6 or 4; rel/CPU-sec appears slower, but rel/wall-clock-sec is faster. Seth detailed this early on- that using HT and 8 threads means about a 20-25% reduction in wall clock time (while CPU-time appears to rise by 75% because the CADO timer just counts threads * wall clock, I think) |
![]() |
![]() |
![]() |
#242 | |
"Seth"
Apr 2019
5×47 Posts |
![]() Quote:
Also 15%! |
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Coordination thread for redoing P-1 factoring | ixfd64 | Lone Mersenne Hunters | 80 | 2021-01-19 17:17 |
big job planning | henryzz | Cunningham Tables | 16 | 2010-08-07 05:08 |
Sieving reservations and coordination | gd_barnes | No Prime Left Behind | 2 | 2008-02-16 03:28 |
Sieved files/sieving coordination | gd_barnes | Conjectures 'R Us | 32 | 2008-01-22 03:09 |
Special Project Planning | wblipp | ElevenSmooth | 2 | 2004-02-19 05:25 |