![]() |
|
|
#12 |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
73518 Posts |
I swapped over to a newer commit (Aug 5) and remembered why I wasn't using it - It won't communicate properly with clients:
Code:
ERROR:root:Invalid workunit file: Error: key STDOUT not recognized |
|
|
|
|
|
#13 |
|
Aug 2006
Monza, Italy
73 Posts |
It seems the good guys at INRIA are already looking into my report. They don't seem to require more information for now.
|
|
|
|
|
|
#14 |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
11×347 Posts |
I'll read the posts when I get my digest version. For now, I'm going to run my September commit and see what shows up later. I'll check the latest git again later on and see if the client communication issue has disappeared.
|
|
|
|
|
|
#15 |
|
Aug 2020
3×5×19 Posts |
Unfortunately, I ran into that error on a C153 which ran over the weekend:
Code:
Warning:Command: Process with PID 849626 finished with return code -6 Error:Filtering - Duplicate Removal, removal pass: Program run on server failed
with exit code -6 Error:Filtering - Duplicate Removal, removal pass: Command line was: /home/flori
an/Math/cado-nfs/build/florian-Precision-3640-Tower/filter/dup2 -poly ./workdir/AL30081984/1971-C153/c155.poly -nrels 62519376 -renumber ./workdir/AL30081984/19
71-C153/c155.renumber.gz ./workdir/AL30081984/1971-C153/c155.dup1//0/dup1.0.0000.gz ./workdir/AL30081984/1971-C153/c155.dup1//0/dup1.0.0001.gz > ./workdir/AL300
81984/1971-C153/c155.dup2.slice0.stdout.4 2> ./workdir/AL30081984/1971-C153/c155.dup2.slice0.stderr.4 Error:Filtering - Duplicate Removal, removal pass: Stderr output (last 10 lines
only) follow (stored in file ./workdir/AL30081984/1971-C153/c155.dup2.slice0.std
err.4):
Error:Filtering - Duplicate Removal, removal pass: antebuffer set to /home/
florian/Math/cado-nfs/build/florian-Precision-3640-Tower/utils/antebuffer
Error:Filtering - Duplicate Removal, removal pass: [checking true duplicate
s on sample of 750234 cells]
Error:Filtering - Duplicate Removal, removal pass: Allocated hash table of
75023359 entries (286MiB)
Error:Filtering - Duplicate Removal, removal pass: Constructing the two fil
elists...
Error:Filtering - Duplicate Removal, removal pass: 2 files (2 new and 0 alr
eady renumbered)
Error:Filtering - Duplicate Removal, removal pass: Reading files already re
numbered:
Error:Filtering - Duplicate Removal, removal pass: Reading new files (using
3 auxiliary threads for roots mod p):
Error:Filtering - Duplicate Removal, removal pass: terminate called after t
hrowing an instance of 'renumber_t::corrupted_table'
Error:Filtering - Duplicate Removal, removal pass: what(): Renumber tabl
e is corrupt: cannot find p=0x4a2bfa9, r=0xd70340 on side 1; note: vp=0x4a2bfb6,
vr=0xd70340
Error:Filtering - Duplicate Removal, removal pass:
Traceback (most recent call last):
File "./cado-nfs.py", line 122, in <module>
factors = factorjob.run()
File "./scripts/cadofactor/cadotask.py", line 6131, in run
last_status = task.run()
File "./scripts/cadofactor/cadotask.py", line 3845, in run
raise Exception("Program failed")
Exception: Program failed
It seems I can still use the relations by having msieve continue the work? How would I do that? According to https://www.mersenneforum.org/showth...48&page=21#227 it seems I can cat the gz files and have msieve process them. But if one of the files is apparently corrupted, how do I find out which one? They all have a size between 3 and 7 MB. I did a zcat | grep and the missing 4a2bfa9 prime is present in some relation, but does that help? Please don't tell me all is lost... Last fiddled with by bur on 2021-05-03 at 11:07 |
|
|
|
|
|
#16 |
|
Aug 2020
28510 Posts |
So I just ignored the cado error message and used the relations with msieve. In case someone has the same problem in the future:
All required files are in workdir/cxxx.upload. First combine all gz compressed relations into one rels.dat: Code:
zcat *.gz > rels.dat Code:
convert_poly -if cado -of msieve < c155.poly > c155.fb Code:
../msieve/msieve -i c155.n -s rels.dat -l c155msieve.log -nf c155.fb -t 10 -nc1 ../msieve/msieve -i c155.n -s rels.dat -l c155msieve.log -nf c155.fb -t 10 -nc2 ../msieve/msieve -i c155.n -s rels.dat -l c155msieve.log -nf c155.fb -t 10 -nc3 For sake of completeness, if not enough relation are found, see https://www.mersenneforum.org/showth...48&page=21#230 for how to make cado-nfs do more sieving. After that it should be possible use msieve as explained above. |
|
|
|
|
|
#17 | |
|
Apr 2020
11×31 Posts |
Quote:
|
|
|
|
|
|
|
#18 |
|
"Ed Hall"
Dec 2009
Adirondack Mtns
11·347 Posts |
Good post!
I thought I had posted a "How I ..." on using CADO-NFS for poly/sieving and Msieve for LA, but apparently I've been slacking. This is how I run all my larger jobs. I had originally written my own conversion (for the .fb), before I learned of the provided one. For some of my scripts, I do a check for *.cyc after the -nc1 step. The scripts use the existence of that file to tell whether filtering succeeded or not. Then the scripts can either call -nc2 or call for more sieving. Not sure if you know this (you probably do), but if -nc2 is interrupted, use -ncr to continue. If you use -nc2 again, it will start LA from scratch. |
|
|
|
|
|
#19 | ||
|
Aug 2020
3·5·19 Posts |
Thanks, it's basically your linked post with the small addition of how to convert poly to fb. I'm glad this error can easily be worked out, otherwise I'd be quite nervous on longer jobs.
Not sure why cado-nfs chokes on othe rels while msieve has no problem with them. Quote:
Quote:
|
||
|
|
|
|
|
#20 | ||
|
Apr 2020
1010101012 Posts |
Quote:
Quote:
|
||
|
|
|
|
|
#21 |
|
Aug 2020
1000111012 Posts |
Ah, that's good to know!
Maybe a stupid question, but since msieve is open source why is the implementation of cado-nfs linear algebra not just taken from msieve? |
|
|
|
|
|
#22 |
|
"Curtis"
Feb 2005
Riverside, CA
4,861 Posts |
CADO's algorithm features less interprocess communication during the (longest) first stage of matrix solving than msieve, which allows jobs to be split among machines fruitfully. This allows larger jobs to be run on regular hardware.
An ideal solution would be to have an -msieve flag in CADO which runs the matrix using msieve within the cado-nfs.py wrapper. |
|
|
|
![]() |
| Thread Tools | |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| CADO-NFS error (exit code -9) | RedGolpe | CADO-NFS | 6 | 2020-09-01 12:29 |
| Is there an error code listing for msieve? | EdH | Msieve | 2 | 2019-11-14 22:58 |
| CADO-NFS Square Root Error | Ferrier | CADO-NFS | 3 | 2019-11-01 23:51 |
| Error Code 40 | storm5510 | Software | 19 | 2016-11-14 15:59 |
| HRF3.TXT now has computer-id and error code | GP2 | Data | 2 | 2003-10-09 06:46 |