![]() |
|
|
#1 | |
|
Nov 2003
22×5×373 Posts |
Quote:
Now let's (collectively perhaps?) tackle the next first hole in the table. Let's push to finish base 2 to 1000 bits. Perhaps the IGG will stop doing easy factorizations and help out. Or perhaps not...... |
|
|
|
|
|
|
#2 | |
|
Aug 2005
Seattle, WA
2·877 Posts |
Quote:
And perhaps the barrier to performing such factorizations is not the lack of willing helpers, but rather the need for coordination. Perhaps you can volunteer in that regard. Or perhaps not.... |
|
|
|
|
|
|
#3 | |
|
Nov 2003
22·5·373 Posts |
Quote:
Actually it has been stated repeatedly by members of the IGG (within this forum!) that they do not want to work on difficult factorizations because it is "boring". One sieves for a long period and nothing happens. They want more immediate results; hence "instant gratification". Or maybe they just have ADD? We have coordination: NFS@Home. Last fiddled with by R.D. Silverman on 2014-05-05 at 18:54 Reason: pagination |
|
|
|
|
|
|
#4 |
|
Aug 2005
Seattle, WA
110110110102 Posts |
Well if you're satisfied with the factorizations that NFS@Home is doing, then what's the problem? And if you're not (because you'd rather work on different numbers), then I guess we don't really have coordination, do we?
Again I ask whether you'd care to volunteer your services toward coordinating the factorization of a number you care about? If not, then I scarcely see how you have cause to complain about others' indolence. |
|
|
|
|
|
#5 | |
|
"Mike"
Aug 2002
25×257 Posts |
Quote:
We would be willing to allocate one (Linux) computer to your effort, if you decide to coordinate something. We bet others would as well. |
|
|
|
|
|
|
#6 |
|
I moo ablest echo power!
May 2013
6E916 Posts |
I've got 10 to 18 (effective) cores to add with some guidance.
|
|
|
|
|
|
#7 | |
|
Nov 2003
1D2416 Posts |
Quote:
I can coordinate, but I can not send data nor receive data owing to a very strict firewall and corporate policy. I can not give out assignments except at a level of course granularity. And while I have CPU resources internally at work they are strictly devoted to projects. The only machine on which I could place data that I received from an outside source is my home PC. It is inadequate for large post processing. "The problem" is that NFS@Home has limited resources. It is not doing large/first hole Cunningham numbers since it is focused at the moment on a C221 via GNFS. More resources would allow work on the former. Actually, I do not understand the criteria by which they have been selecting their numbers over the past 1-2 years. At one time they were doing wanted Cunningham numbers. I lack the resources to perform leading edge factorizations. I do numbers that are within reach of my resources; currently these are Homogeneous Cunningham first holes. Last fiddled with by R.D. Silverman on 2014-05-05 at 21:09 Reason: pagination |
|
|
|
|
|
|
#8 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
191316 Posts |
The Cunningham tables appear to be looking after themselves, thanks mostly to Ryan Propper who clearly has access to some very substantial resources.
On the whole I would rather run large jobs than complain that other people aren't using their superior resources to run the particular jobs that I'm interested in; convincing people to do what you want is a diplomatic rather than a mathematical problem. NFS@home's criterion seems fairly simple: on the large-jobs side they do GNFS at the edge of the practical, on the small-jobs side they do whatever people express interest in: I've a few times sent them a decent polynomial for a C17x and then done the linear algebra myself. Cunningham first-holes are bigger than small jobs, so don't get done. I'm also unsure, in this world where computers do clock themselves down and save significant power when idle, that using computers other than the personally-owned ones of explicit volunteers is necessarily the right way to go. Last fiddled with by fivemack on 2014-05-05 at 22:26 |
|
|
|
|
|
#9 | |
|
Aug 2005
Seattle, WA
2·877 Posts |
Quote:
1) Assignments are given out (actually, just taken) at a level of coarse granularity. 2) The sieving work need not be done at all by the coordinator (though Tom did in fact do quite a bit of the sieving). That should address your concerns, at least up to the point of sieving. I suspect you can find willing volunteers to store the data and do the post-processing, though I concede I might be wrong about that. It would probably be worth checking with Tom to get some idea of how much iron he thinks is actually required for post-processing job on e.g. 2,983+. The real value that Tom added was: 1) Coming up with a good polynomial (in the case of GNFS jobs). 2) Doing a bunch of test-sieving in order to come up with good parameters. 3) Maintaining the thread which contains reservations. 4) Doing periodic relation maintenance (e.g. duplicate elimination, post-processing test runs). 5) Maintaining a server to act as a relation dump. 6) Doing the post-processing. 7) Maybe some other stuff I don't know about, but he would most likely share. Do any of those, other than 6, seem too onerous for you? Again, I would be surprised if you didn't find someone willing to take on 6. I think you'll find plenty of people here will jump at the chance to join in on a factorization if someone (e.g. you) can perform these steps. What do you say? |
|
|
|
|
|
|
#10 |
|
(loop (#_fork))
Feb 2006
Cambridge, England
72×131 Posts |
The linear algebra for a kilobit SNFS job takes somewhere between a month and a season on an i7-4930K or similar (six cores, four memory controllers, you probably do want 32G of memory); the relations take low-three-digit gigabytes of disc space, maybe a week of download over slow ADSL.
RDS is entirely aware of what resources are required - he's coordinated some enormous projects - but is not in a position to buy them. His perfectly reasonable point is that individuals are unwilling to commit to six-month runs, but he considers this a problem to be ranted about rather than a political truth to be worked around. nfs@home works around it by using spare cycles from people who couldn't care less about factorisation, but there's only space for one nfs@home, someone else has done the tedious work to set it up, and the reward for that tedious work is that he gets to set the priorities. I'm happy to run linear algebra for as many 1000ish-bit SNFS jobs as nfs@home is prepared to run, but Greg has perfectly reasonably chosen to run limit-of-the-possible GNFS jobs instead. Last fiddled with by fivemack on 2014-05-05 at 23:05 |
|
|
|
|
|
#11 | ||||||||
|
Nov 2003
164448 Posts |
Quote:
Or I could post them here. Quote:
format. Quote:
Quote:
However: Who would create the input files? I do not have Msieve/GGNFS. (and am not allowed to install them) Quote:
Quote:
I am behind a very strict firewall/restrictions at work. I can't accept data from an external source. Quote:
Quote:
|
||||||||
|
|
|
![]() |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| 1000 TFLOPs | ramgeis | PrimeNet | 2 | 2014-04-08 10:27 |
| Base-6 speed for prime testing vs. base-2 | jasong | Conjectures 'R Us | 36 | 2010-08-03 06:25 |
| 1000 < k < 2000 | Flatlander | Riesel Prime Data Collecting (k*2^n-1) | 17 | 2009-01-28 16:57 |
| Base 2 to 768 bits | R.D. Silverman | NFSNET Discussion | 14 | 2007-03-10 12:47 |
| The 1000-day club | GP2 | Data | 15 | 2004-06-11 01:57 |