mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware > GPU Computing

Reply
 
Thread Tools
Old 2016-06-10, 22:03   #12
Madpoo
Serpentine Vermin Jar
 
Madpoo's Avatar
 
Jul 2014

2×13×131 Posts
Default

Quote:
Originally Posted by airsquirrels View Post
Ok I got the 1080 Founders Edition running.

It is still failing 43/107 self tests using 8.0 and compute 6.1, so it seems the mfaktc issue is still present. I skipped the self test to get performance numbers and I'm seeing 1071 GhzDay/Day up to 75 bits, which is 23% faster than my Fury X results.

Also compiled CudaLucas. M49 (4M FFT) runs at 3.4ms/iter, 70 hour ETA
70 hours to do M74207281 isn't bad. It sounds like despite the skepticism because it's 1/32 FP64, it's still not a bad LL testing bit of silicon. I mean, it has a lot of FP cores so 1/32 of a lot is still a good amount... isn't that how it works?

Sure, there are faster things out there, and in the end it may be more useful to have something like that doing TF work (more overall throughput), but I say use it for whatever you have more fun with.
Madpoo is offline   Reply With Quote
Old 2016-06-10, 23:17   #13
airsquirrels
 
airsquirrels's Avatar
 
"David"
Jul 2015
Ohio

11·47 Posts
Default

70 hours at a mere 180 watts. That's the most impressive part.
airsquirrels is offline   Reply With Quote
Old 2016-06-11, 11:36   #14
henryzz
Just call me Henry
 
henryzz's Avatar
 
"David"
Sep 2007
Liverpool (GMT/BST)

614110 Posts
Default

The switch from 28nm to 16nm will have helped a lot.
Hopefully they will keep up to Intel in future rather than slip so far behind.
henryzz is offline   Reply With Quote
Old 2016-06-11, 11:56   #15
jasong
 
jasong's Avatar
 
"Jason Goatcher"
Mar 2005

3×7×167 Posts
Default

Quote:
Originally Posted by henryzz View Post
The switch from 28nm to 16nm will have helped a lot.
Hopefully they will keep up to Intel in future rather than slip so far behind.
Is Intel legitimately ahead, or did they cheat in some way that I don't know about.

I'm specifically thinking of them bribing companies into delaying or not allowing Athlon-based computers to be sold.

Athlon is a brand of cpu that came out while Intel was building P4s. AMD did significantly worse than it could have because Intel used it's clout to prevent companies from selling the Athlons.

It's kind of like the Windows tax inflating the Windows numbers even while people immediately install Linux on the computers they bring home.
jasong is offline   Reply With Quote
Old 2016-06-11, 16:46   #16
airsquirrels
 
airsquirrels's Avatar
 
"David"
Jul 2015
Ohio

10058 Posts
Default

Interestingly enough, performance at the 2048K FFT is actually even more impressive:

Code:
FFTSize: 2048K Exponent: 38410247 (0.10%) Error: 0.18750 ms: 1.7378 eta: 18:27:14
Card 1 (Graphics Device - 75.00C, 100% Load [1833/607]@145.70W/180.00W, M38410247 using 2048K) GhzDay: 68.23
That's not far off from my Titans, and significantly better than my Fury Xs.
airsquirrels is offline   Reply With Quote
Old 2016-06-11, 17:14   #17
Madpoo
Serpentine Vermin Jar
 
Madpoo's Avatar
 
Jul 2014

1101010011102 Posts
Default

Quote:
Originally Posted by jasong View Post
Is Intel legitimately ahead, or did they cheat in some way that I don't know about.

I'm specifically thinking of them bribing companies into delaying or not allowing Athlon-based computers to be sold.

Athlon is a brand of cpu that came out while Intel was building P4s. AMD did significantly worse than it could have because Intel used it's clout to prevent companies from selling the Athlons.
....
As far as I know, Athlons were always horrible at FP, nothing new about that. Their other stats were impressive and they made some decent advances that pushed both Intel and AMD ahead by competing. But I don't think they were ever competitive in the precision computing arena, were they?
Madpoo is offline   Reply With Quote
Old 2016-06-11, 21:43   #18
ATH
Einyen
 
ATH's Avatar
 
Dec 2003
Denmark

345210 Posts
Default

That is very impressive performance considering FP64= 1/32 * FP32.

But it still seems like a waste to use this for LL when it is the fastest mainstream graphic card for trial factoring.

Is there actually anything faster than this except the new Tesla P100 for trial factoring?

Last fiddled with by ATH on 2016-06-11 at 21:45
ATH is offline   Reply With Quote
Old 2016-06-11, 22:42   #19
mackerel
 
mackerel's Avatar
 
Feb 2016
UK

26×7 Posts
Default

Quote:
Originally Posted by Madpoo View Post
As far as I know, Athlons were always horrible at FP, nothing new about that. Their other stats were impressive and they made some decent advances that pushed both Intel and AMD ahead by competing. But I don't think they were ever competitive in the precision computing arena, were they?
http://s5.postimg.org/jg4d7uciv/llr_perf.png

That's a performance comparison I did in 2005. I never gave architecture much thought at the time and just wanted something fast. I guess if you were to look up the actual clock and estimate the values from the image (it is unlikely I will ever find the original sheet if I even still have it), you could estimate an IPC for the processors for prime finding application the time. The software I used at the time was LLRnet, so presumably that was a network wrapper for LLR. I was running what was the rieselsieve project at the time. Without in depth analysis, it does look like this might be an area where Intel's higher clock actually gave some advantage. Even HT looks like it helped throughput some although I didn't use that for LLR.

For context, Intel was still in full Netburst swing. Multi-core hadn't come along yet. AMD had come out with 64 bit extensions although no one used it at the time. I was likely running either Windows 2000 or XP then depending on the age of the system.

Fast forward, I think where Intel really jumped ahead in CPU FP was with Sandy Bridge, with another big jump from Haswell. I got a feeling that the post ATI merger AMD wanted FP to be done by the GPU allowing them to cut back on spending CPU resource there.

Last fiddled with by mackerel on 2016-06-11 at 22:43
mackerel is offline   Reply With Quote
Old 2016-06-11, 23:33   #20
Mark Rose
 
Mark Rose's Avatar
 
"/X\(‘-‘)/X\"
Jan 2013
https://pedan.tech/

24·199 Posts
Default

Yeah, once the Athlon came out, it was king until the Core arrived in 2006. That's when Intel dropped Netburst and went back and extended the PIII design.
Mark Rose is offline   Reply With Quote
Old 2016-06-12, 01:14   #21
jasong
 
jasong's Avatar
 
"Jason Goatcher"
Mar 2005

3×7×167 Posts
Default

Quote:
Originally Posted by Madpoo View Post
As far as I know, Athlons were always horrible at FP, nothing new about that. Their other stats were impressive and they made some decent advances that pushed both Intel and AMD ahead by competing. But I don't think they were ever competitive in the precision computing arena, were they?
Good point, I temporarily forgot the forum I was in when I asked that question.
jasong is offline   Reply With Quote
Old 2016-06-12, 13:30   #22
henryzz
Just call me Henry
 
henryzz's Avatar
 
"David"
Sep 2007
Liverpool (GMT/BST)

3×23×89 Posts
Default

Quote:
Originally Posted by mackerel View Post
http://s5.postimg.org/jg4d7uciv/llr_perf.png

That's a performance comparison I did in 2005. I never gave architecture much thought at the time and just wanted something fast. I guess if you were to look up the actual clock and estimate the values from the image (it is unlikely I will ever find the original sheet if I even still have it), you could estimate an IPC for the processors for prime finding application the time. The software I used at the time was LLRnet, so presumably that was a network wrapper for LLR. I was running what was the rieselsieve project at the time. Without in depth analysis, it does look like this might be an area where Intel's higher clock actually gave some advantage. Even HT looks like it helped throughput some although I didn't use that for LLR.

For context, Intel was still in full Netburst swing. Multi-core hadn't come along yet. AMD had come out with 64 bit extensions although no one used it at the time. I was likely running either Windows 2000 or XP then depending on the age of the system.

Fast forward, I think where Intel really jumped ahead in CPU FP was with Sandy Bridge, with another big jump from Haswell. I got a feeling that the post ATI merger AMD wanted FP to be done by the GPU allowing them to cut back on spending CPU resource there.
AMD really lost out LLR/Prime95 performance until their SSE2 implementation caught up in speed with Phenom II. I remember moving from a 2.4 Ghz Athlon 64 to 2.4 Ghz Q6600 and getting an 8x speed improvement.
henryzz is offline   Reply With Quote
Reply



All times are UTC. The time now is 15:02.


Fri Jul 7 15:02:30 UTC 2023 up 323 days, 12:31, 0 users, load averages: 1.59, 1.29, 1.18

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2023, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.

≠ ± ∓ ÷ × · − √ ‰ ⊗ ⊕ ⊖ ⊘ ⊙ ≤ ≥ ≦ ≧ ≨ ≩ ≺ ≻ ≼ ≽ ⊏ ⊐ ⊑ ⊒ ² ³ °
∠ ∟ ° ≅ ~ ‖ ⟂ ⫛
≡ ≜ ≈ ∝ ∞ ≪ ≫ ⌊⌋ ⌈⌉ ∘ ∏ ∐ ∑ ∧ ∨ ∩ ∪ ⨀ ⊕ ⊗ 𝖕 𝖖 𝖗 ⊲ ⊳
∅ ∖ ∁ ↦ ↣ ∩ ∪ ⊆ ⊂ ⊄ ⊊ ⊇ ⊃ ⊅ ⊋ ⊖ ∈ ∉ ∋ ∌ ℕ ℤ ℚ ℝ ℂ ℵ ℶ ℷ ℸ 𝓟
¬ ∨ ∧ ⊕ → ← ⇒ ⇐ ⇔ ∀ ∃ ∄ ∴ ∵ ⊤ ⊥ ⊢ ⊨ ⫤ ⊣ … ⋯ ⋮ ⋰ ⋱
∫ ∬ ∭ ∮ ∯ ∰ ∇ ∆ δ ∂ ℱ ℒ ℓ
𝛢𝛼 𝛣𝛽 𝛤𝛾 𝛥𝛿 𝛦𝜀𝜖 𝛧𝜁 𝛨𝜂 𝛩𝜃𝜗 𝛪𝜄 𝛫𝜅 𝛬𝜆 𝛭𝜇 𝛮𝜈 𝛯𝜉 𝛰𝜊 𝛱𝜋 𝛲𝜌 𝛴𝜎𝜍 𝛵𝜏 𝛶𝜐 𝛷𝜙𝜑 𝛸𝜒 𝛹𝜓 𝛺𝜔