mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware

Closed Thread
 
Thread Tools
Old 2008-09-11, 09:19   #78
xilman
Bamboozled!
 
xilman's Avatar
 
"π’‰Ίπ’ŒŒπ’‡·π’†·π’€­"
May 2003
Down not across

22×5×72×11 Posts
Default

Quote:
Originally Posted by jasong View Post
As to the hardware crash, one of the problems with RAID arrays is that everything tends to be backed up in the same location. So things like ceiling leaks and fallen over shelves have the potential to destroy enough hard drives in a RAID array to make them unrecoverable.
With respect, anyone running a serious operation knows that putting data on a RAID array is not a good backup strategy.

Data loss can occur not only through damage to hardware which, as you describe, can take out an entire array. Data can also be corrupted by finger-trouble, by errant programs, by broken hardware elsewhere in the system and by enemy action. RAID will not protect your data from any of those.

I know of what I speak. 18 months ago we were unlucky enough to have one disk in an array fail hard and the buggy RAID controller then corrupted one of the remaining set. We were out of action for a couple of days while Dell ran diags and shipped us a new disk and RAID controller. Restoring the data from tapes took one of the two days. The backup tapes were brought in from an off-site store. This precaution guards against events that take out our entire facility. Replacing hardware is cheap and easy. Restoring data accumulated over 15 years of operation is hard and expensive in the absence of adequate backups.

Sadly, all too many people don't think about instigating a proper backup mechanism until they've learned the hard way.


Paul
xilman is offline  
Old 2008-09-11, 18:02   #79
Jeff Gilchrist
 
Jeff Gilchrist's Avatar
 
Jun 2003
Ottawa, Canada

49516 Posts
Default

Quote:
Originally Posted by xilman View Post
Sadly, all too many people don't think about instigating a proper backup mechanism until they've learned the hard way.
And now there are even inexpensive ways to get off-site backups, especially for home users via services such as MozyBackup and many others that give you unlimited data, encryption, and even file versioning. So there really aren't any excuses left (except lazyiness I guess).
Jeff Gilchrist is offline  
Old 2008-09-16, 19:15   #80
jrk
 
jrk's Avatar
 
May 2008

3×5×73 Posts
Default

Quote:
Originally Posted by jasonp View Post
Almost all special purpose hardware does not implement double precision, and so there is no possibility to achieve the aforementioned enormous performance.
I really hate to ask this now, because I can already guess the answer, but I read this on the tech specs page about the NVidia 8800 family of GPUs:
Quote:
Originally Posted by http://www.nvidia.com/page/8800_tech_specs.html
Full 128-bit floating point precision through the entire rendering pipeline
I don't suppose this actually means you could do double precision float computations, it is more likely a clever marketing lie, right?

Thanks.
jrk is offline  
Old 2008-09-16, 21:38   #81
jasonp
Tribal Bullet
 
jasonp's Avatar
 
Oct 2004

3×1,181 Posts
Default

Quote:
Originally Posted by jrk View Post
I don't suppose this actually means you could do double precision float computations, it is more likely a clever marketing lie, right?
I looked around the page, and without reading actual technical documents on that card the most you can conclude is that the datapath to the execution units is 128 bits wide. Most likely this means 4 x 32-bit single-precision floating point, with a possibility of much slower 2 x 64-bit double precision.

When we start seeing graphics applications that demand 64-bit precision, then we'll start seeing cards that trumpet this as a feature. I can't imagine an app that needs that and is purely graphical and has mass appeal.
jasonp is offline  
Old 2008-09-16, 21:50   #82
retina
Undefined
 
retina's Avatar
 
"The unspeakable one"
Jun 2006
My evil lair

24·389 Posts
Default

If we take the quote "Full 128-bit floating point precision" at face value we would have to assume that the precision is the mantissa. Then when we add the exponent and sign bit, that would suggest ~ 18 byte data format. It doesn't seem very likely.
retina is online now  
Old 2008-09-16, 22:15   #83
fivemack
(loop (#_fork))
 
fivemack's Avatar
 
Feb 2006
Cambridge, England

3×2,141 Posts
Default

Quote:
Originally Posted by jrk View Post
I really hate to ask this now, because I can already guess the answer, but I read this on the tech specs page about the NVidia 8800 family of GPUs:

I don't suppose this actually means you could do double precision float computations, it is more likely a clever marketing lie, right?
You can do double-precision FP with CUDA on the GT200 family; see http://www.mathematik.tu-dortmund.de...dprecision.pdf and http://www.mathematik.tu-dortmund.de...ION08-long.pdf

They're working on partial differential equations, and a year or so ago published a paper where they did some number of iterations in single-precision on the GPU then copied across to the CPU and 'polished' the result in double-precision. The current paper does the polishing on the GPU too.

http://www.theinquirer.net/gb/inquir...dra-2009-sizes has benchmark results from Sandra 2009, which includes basic Mandelbrot in SP and DP; I suspect that the 'float performance' and 'double performance' axes are the wrong way round in the graphs presented.

http://forums.nvidia.com/index.php?s...483&hl=fourier indicates that 32-bit integer performance is already quite interesting ... they're done some large searches for sets of bases with large smallest pseudoprime using a GPU implementation of a^b%p on 64-bit p.

Last fiddled with by fivemack on 2008-09-16 at 22:20
fivemack is offline  
Old 2008-09-17, 06:46   #84
Tumo
 
Jan 2005

19 Posts
Question Using GPUs for other things

Hi,
First, thank you all for the great info... it's more than I can swallow. Perhaps some day.

I do have a question....

GPU's which use 16 bit pipes, say my nvidia 6600, do a great job in spite of its age. I have been told it's because of the depth and because of the number of parallel 16 bit internal registers. I know the 8800's and (I hear) soon 9xxx cards are going to be WILD! I do hope to upgrade and tag along.

Reading the '128 bit vs 2 x 64 vs 4 x 32' discussion portion made me wonder. I heard one of the cards has 9 x 16 internal (registers or paths). If so, can that be of any help...

Would someone please explain? I'm wondering because we all talk about the primality tests, but what about the sieve/factoring portion? Can that be put in the card as well and would it help ?

With thanks and thanks in advance,
Ty
Tumo is offline  
Old 2008-09-17, 06:59   #85
jrk
 
jrk's Avatar
 
May 2008

109510 Posts
Default

Thank you for the insights, jasonp and retina & fivemack.

Last fiddled with by jrk on 2008-09-17 at 06:59
jrk is offline  
Old 2008-11-26, 04:34   #86
RMAC9.5
 
RMAC9.5's Avatar
 
Jun 2003

15310 Posts
Default

Ernst,
Here is an article which may interest you http://it.tmcnet.com/news/2008/11/18/3796116.htm.
I got it from a posting on aceshardware.freeforums.org.
RMAC9.5 is offline  
Old 2008-12-01, 03:58   #87
jasong
 
jasong's Avatar
 
"Jason Goatcher"
Mar 2005

DB316 Posts
Default

Quote:
must use fast Fourier transforms (FFTs) implemented in double precision (DP) floating point.
This is wrong. I don't say this because I'm an expert, I'm not. I say this because the person programming the GPU prime-finding algorithm told me this isn't true. He is programming the GPU to work with integers

Now, of course, you'll ask,"Jason, if you don't understand the math, then how the hell do you know he isn't BSing you?" Simple, he's done major things in the prime-finding community in the past, and I'm simply going by his reputation.

Last fiddled with by jasong on 2008-12-01 at 04:02
jasong is offline  
Old 2008-12-01, 07:39   #88
cheesehead
 
cheesehead's Avatar
 
"Richard B. Woods"
Aug 2002
Wisconsin USA

22·3·641 Posts
Default

Quote:
Originally Posted by jasonp View Post
The problem is that in order to have a chance of achieving that incredible performance, code must not use any feature of the special purpose processor that is not implemented in dedicated hardware. There are projects that can abide by this restriction (distributed.net, folding@home), but any project that works by multiplying large numbers (LLR, GIMPS) must use fast Fourier transforms (FFTs) implemented in double precision (DP) floating point.
Quote:
Originally Posted by jasong View Post
This is wrong. I don't say this because I'm an expert, I'm not. I say this because the person programming the GPU prime-finding algorithm told me this isn't true. He is programming the GPU to work with integers
As you correctly point out, the claim in post #1 that "... any project that works by multiplying large numbers (LLR, GIMPS) must use fast Fourier transforms (FFTs) implemented in double precision (DP) floating point" needs to be amended. It is, of course, quite possible to use integers entirely in the Fourier transforms. The statement as it stands is especially misleading in a FAQ about computations performed by hardware other than general-purpose CPUs.

I suggest a more-correct replacement:

"Projects that involve multiplying large numbers (LLR, GIMPS) take into account that the fastest methods for doing that use a type of mathematical operation known as a Fast Fourier Transform (FFT).

On recent Intel systems, or any others with a similar floating-point-vs-integer arithmetic performance ratio, the fastest way of implementing FFTs uses double-precision (DP) floating point.

It is quite possible to use single-precision floating point, or to use all-integer arithmetic, to implement FFTs, but on the most common current CPUs those methods will be significantly slower than using DP floating-point. Furthermore, single-precision floating-point becomes impractical to use at all for FFTs on numbers above a certain range, and that range is routinely exceeded by current LLR and GIMPS work.

However, the superiority of DP over single-precision, or of floating-point over integer arithmetic, for Intel-like CPUs may not be so true, or at least may not be a limiting consideration, for special-purpose dedicated hardware such as graphics cards."

I welcome anyone's correction or improvement of that suggestion.
cheesehead is offline  
Closed Thread



Similar Threads
Thread Thread Starter Forum Replies Last Post
New PC dedicated to Mersenne Prime Search Taiy Hardware 12 2018-01-02 15:54
The prime-crunching on dedicated hardware FAQ (II) jasonp Hardware 46 2016-07-18 16:41
How would you design a CPU/GPU for prime number crunching? emily Hardware 4 2012-02-20 18:46
DSP hardware for number crunching? ixfd64 Hardware 15 2011-08-09 01:11
Optimal Hardware for Dedicated Crunching Computer Angular Hardware 5 2004-01-16 12:37

All times are UTC. The time now is 21:19.


Sun Aug 1 21:19:40 UTC 2021 up 9 days, 15:48, 0 users, load averages: 1.81, 1.68, 1.60

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.