![]() |
![]() |
#1 |
Nov 2002
Anchorage, AK
3·7·17 Posts |
![]()
As a favor to this community I've been archiving and hosting the hourly Primenet reports (summary.txt, status.txt. cleared.txt) as well as mirroring the software ftp directory here: http://www.mersenneforum.org/primenet
As of today this archive has grown to about 30GB. I've been burning DVD's regularly and actually sending a backup set over to Xyzzy for "safe keeping", but I'm wondering if it'd be a good idea to let another entity crawl the site and keep the older files in their archivess. Naturally I'm talking about www.archive.org. I've restricted all search engines from crawling the site (with the use robots.txt) to minimize traffic and also not to bog down those indexing/archiving servers with a massive file load. Basically restricting these sites was done as a favor to them so they don't all of a sudden end up with these 30GB archived Primenet files (archived files are compressed with bzip2). So what I'd like to hear from the community is whether it would be helpful to keep these files around for the long term? Are these files worth saving for the future, or are they only helpful for the short term? I will still keep this server up as long as I have the resources available, but just want to hear what people think about long term archiving. |
![]() |
![]() |
![]() |
#2 |
Aug 2002
Termonfeckin, IE
24×173 Posts |
![]()
I think it is useful to keep files for the long term though perhaps not 24x3 files a day. It would be much more practical and easier for you to say keep one or two files a day. Or even maybe come up with a way to keep diffs of the hourly files instead of entire files. That is if space is an issue.
I usually find such files useful to refer back to for at least one year past their date. |
![]() |
![]() |
![]() |
#3 |
Nov 2002
Anchorage, AK
16516 Posts |
![]()
It's no problem for me to archive these files and keep collecting them on the hourly basis since I have the resources available to me right now (I have it on it's own dedicated server actually). But should I open up the robots.txt to archive.org's crawler and have it try to archive what we already have? This is so there would be a long term site having the older files stored away in case something does happen where we can't put up or host the huge archive of older files. As an example, last year's archive (2004) from June-December was 18G (started archiving again in June).
|
![]() |
![]() |
![]() |
#4 |
"Richard B. Woods"
Aug 2002
Wisconsin USA
22×3×641 Posts |
![]()
I'm in favor of keeping them all, somewhere. You never know when they'll be very valuable for some reason.
... though probably just to some grad student doing a thesis on the history of distributed computing. Last fiddled with by cheesehead on 2005-08-25 at 00:34 |
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
How Big Can an SNFS Constant Term Be? | wblipp | Factoring | 14 | 2015-03-31 23:05 |
Using long long's in Mingw with 32-bit Windows XP | grandpascorpion | Programming | 7 | 2009-10-04 12:13 |
I think it's gonna be a long, long time | panic | Hardware | 9 | 2009-09-11 05:11 |
long-standing PrimeNet issues and how to fix them | ixfd64 | PrimeNet | 16 | 2008-11-17 07:53 |
Short term goal | em99010pepe | No Prime Left Behind | 94 | 2008-03-24 21:02 |