![]() |
|
|
#34 |
|
Mar 2008
5·11 Posts |
No, I thought the working directory needed to previously exist, so it was there, but empty. This then showed up when running bwc.pl with :complete.
Thanks for the information. As I get further along I may have more errors to ask about. |
|
|
|
|
|
#35 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
72·131 Posts |
OK: I'm using the command line
Code:
% /home/nfsslave2/cado/cado-nfs-20090528-r2167/build/cow/linalg/bwc/u128_bench snfs.small -impl bucket u128 T0 snfs.small: 5582216 rows 5582056 cols 259518490 coeffs 22 iterations in 101s, 4.61/1, 17.77 ns/coeff u64 T0 snfs.small: 5582216 rows 5582056 cols 259518490 coeffs 38 iterations in 102s, 2.67/1, 10.29 ns/coeff u64k says 'T0 : Check failed Aborted' u64n also says this. I assume u128 would want to do exactly half as many iterations as u64, so would be quicker in total; should I be getting a 'k' or 'n' parameter to u64k or u64n in some way? If u128 does 5582216/128 iterations, the total runtime would be ~200k seconds, which seems pretty good since msieve lanczos took 108242 wall-time seconds with four threads - but I'm not sure whether there's not another factor two hiding somewhere in the block Wiedemann algorithm. So, time to try threading. Code:
/home/nfsslave2/cado/cado-nfs-20090528-r2167/build/cow/linalg/balance --in snfs.small --out cabbage --nslices 2x2 --ramlimit 8G Code:
taskset 0f /home/nfsslave2/cado/cado-nfs-20090528-r2167/build/cow/linalg/bwc/u128_bench -impl bucket -nthreads 4 -- cabbage.h0.v0 cabbage.h1.v0 cabbage.h0.v1 cabbage.h1.v1 19 iterations in 102s, 5.35/1, 20.62 ns/coeff Does this mean that threads are treading on one another's toes and four threads are slower than one, or that each thread has done 19 iterations in 102 seconds for a total speed of effectively 5.16 ns/coeff ? |
|
|
|
|
|
#36 |
|
Mar 2008
5×11 Posts |
Ok. A little farther.
Now I have had the following: Computing trsp(x)*M^100 ..........Warning: Doing many iterations with bad code Warning: Doing many iterations with bad code Warning: Doing many iterations with bad code Warning: Doing many iterations with bad code Then a little later... Failed check at iteration 100 /cado-nfs/linalg/bwc/u64_krylov: exited with status 1 Tried with a different seed, and it failed at iteration 1900. I know I had trouble with the msieve version of block lanczos if the matrix was too sparse, I believe it was. Is there a similar condition here which could cause it to fail? Last fiddled with by joral on 2009-06-02 at 23:47 |
|
|
|
|
|
#37 | ||||||
|
May 2009
2×11 Posts |
Quote:
Quote:
Quote:
Quote:
Quote:
Quote:
E. |
||||||
|
|
|
|
|
#38 | |||
|
May 2009
2×11 Posts |
Quote:
Quote:
Care for sharing your matrix ? Quote:
E. |
|||
|
|
|
|
|
#39 |
|
May 2009
2×11 Posts |
|
|
|
|
|
|
#40 | |
|
May 2009
2×11 Posts |
Quote:
E. |
|
|
|
|
|
|
#41 | |
|
Mar 2008
5·11 Posts |
Quote:
If I leave the seed parameter unchanged, it always fails at the same iteration. It's about a 280 Mb matrix file ungzipped, so I'll see what it compresses to and where I can put it. |
|
|
|
|
|
|
#42 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
72·131 Posts |
The machine I'm doing the benchmarks on is a single-socket 2.66GHz Core i7 (256k 10-cycle L2 cache per core + 8192k 19-cycle L3 cache per four cores + 12G DDR3/8500); I am a little surprised that I don't have to give a load of cache parameters to bench, if it's running one thread blocking for the 256k cache rather than the 8192k one then I could understand it being a bit slow.
Will try more sensible benchmarks (correct transpose parameters, trying 1x4 2x2 4x1 decompositions on four cores and 1x8 2x4 4x2 8x1 decompositions on eight-threads-on-four-cores) with new tarball tonight; I've left a make-matrix-from-relations job running today on a set of relations from a very large SNFS job, and will mention if that falls over in interesting ways. It's using an awful lot of memory (17G vsize, 10G rsize), but I have an awful lot of memory and a fast swap disc. Last fiddled with by fivemack on 2009-06-03 at 11:15 |
|
|
|
|
|
#43 |
|
Tribal Bullet
Oct 2004
3,541 Posts |
In case it becomes an issue: the latest GGNFS lattice sievers do not print all the factors of relations; they skip multiplicity beyond 1 and skip printing factors smaller than 1000, so that both of these have to be rediscovered by any relation-reading code.
|
|
|
|
|
|
#44 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
72×131 Posts |
I issue the command
Code:
nfsslave2@cow:/scratch/fib1039/with-cado$ /home/nfsslave2/cado/cado-nfs-20090603-r2189/build/cow/linalg/bwc/u128_bench -t --impl bucket snfs.small Code:
Lsl 56 cols 3634827..3699734 w=778884, avg dj=7.2, max dj=34365, bucket hit=1/1834.7-> too sparse Switching to huge slices. Lsl 56 to be redone Flushing 56 large slices Hsl 0 cols 3634827..5582056 (30*64908) .............................. w=16383453, avg dj=0.3, max dj=29376, bucket block hit=1/10.2 u128_bench: /home/nfsslave2/cado/cado-nfs-20090603-r2189/linalg/bwc/matmul-bucket.cpp:610: void split_huge_slice_in_vblocks(builder*, huge_slice_t*, huge_slice_raw_t*, unsigned int): Assertion `(n+np)*2 == (size_t) (spc - sp0)' failed. Aborted |
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| CADO-NFS on windows | jux | CADO-NFS | 25 | 2021-07-13 23:53 |
| CADO help | henryzz | CADO-NFS | 4 | 2017-11-20 15:14 |
| CADO and WinBlows | akruppa | Programming | 22 | 2015-12-31 08:37 |
| CADO-NFS | skan | Information & Answers | 1 | 2013-10-22 07:00 |
| CADO | R.D. Silverman | Factoring | 4 | 2008-11-06 12:35 |