mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware > GPU Computing

Reply
 
Thread Tools
Old 2019-06-04, 09:18   #1
M344587487
 
M344587487's Avatar
 
"Composite as Heck"
Oct 2017

10010111012 Posts
Default Radeon Pro Vega II Duo (look at this monster)

https://www.amd.com/en/graphics/work...on-pro-vega-ii


https://old.reddit.com/r/Amd/comment...card_infinity/


https://old.reddit.com/r/Amd/comment...rt_of_new_mac/


  • Two MI60 chips on the same board with an 84GB/s infinity fabric link connecting them
  • 128 CU's
  • "Up to 64GB HBM2" meaning 2x16GB or 2x32GB



Each chip should have slightly more throughput than a Radeon VII when doing disjoint tasks due to the extra CU's. There is an outside chance that a single test could span both chips (not efficient, maybe not too detrimental to throughput in theory not that anyone's going to take the time coding something so niche). There's definitely not going to be a "gamer" version of this card so the price is going to be hideous.
M344587487 is offline   Reply With Quote
Old 2019-06-17, 08:04   #2
SELROC
 

2×1,217 Posts
Default The new Radeon RX 5700 XT

Quote:
Originally Posted by M344587487 View Post
https://www.amd.com/en/graphics/work...on-pro-vega-ii


https://old.reddit.com/r/Amd/comment...card_infinity/


https://old.reddit.com/r/Amd/comment...rt_of_new_mac/


  • Two MI60 chips on the same board with an 84GB/s infinity fabric link connecting them
  • 128 CU's
  • "Up to 64GB HBM2" meaning 2x16GB or 2x32GB



Each chip should have slightly more throughput than a Radeon VII when doing disjoint tasks due to the extra CU's. There is an outside chance that a single test could span both chips (not efficient, maybe not too detrimental to throughput in theory not that anyone's going to take the time coding something so niche). There's definitely not going to be a "gamer" version of this card so the price is going to be hideous.







https://www.amd.com/en/products/grap...eon-rx-5700-xt
  Reply With Quote
Old 2019-06-17, 09:38   #3
M344587487
 
M344587487's Avatar
 
"Composite as Heck"
Oct 2017

5·112 Posts
Default

Navi like the RX 5700 above is on the RDNA architecture, something somewhat based on GCN but tailored more to graphics than compute. I expect it to be much more similar to current nvidia cards than AMD cards in that the single precision will make it performant for TF but low DP performance would make it a poor choice for PRP/P-1. I've heard mixed things about Navi and my expectations are accordingly low to mid. RDNA is going to be used in Samsung phones so the arch is scalable, if that translates to good power scaling in desktop parts it could be surprisingly effective for our niche.
M344587487 is offline   Reply With Quote
Old 2019-06-18, 05:54   #4
SELROC
 

44010 Posts
Default

Quote:
Originally Posted by M344587487 View Post
Navi like the RX 5700 above is on the RDNA architecture, something somewhat based on GCN but tailored more to graphics than compute. I expect it to be much more similar to current nvidia cards than AMD cards in that the single precision will make it performant for TF but low DP performance would make it a poor choice for PRP/P-1. I've heard mixed things about Navi and my expectations are accordingly low to mid. RDNA is going to be used in Samsung phones so the arch is scalable, if that translates to good power scaling in desktop parts it could be surprisingly effective for our niche.

Anyway the new Navi GPUs require PCI 4.0, which means they need new mainboards.
  Reply With Quote
Old 2019-06-18, 12:25   #5
henryzz
Just call me Henry
 
henryzz's Avatar
 
"David"
Sep 2007
Cambridge (GMT)

26×89 Posts
Default

Quote:
Originally Posted by SELROC View Post
Anyway the new Navi GPUs require PCI 4.0, which means they need new mainboards.
Require or support? I have run a pcie 3 card in a 1.1 socket. It would surprise me if it won't run on current motherboards.
henryzz is offline   Reply With Quote
Old 2019-06-18, 12:35   #6
SELROC
 

22·32·167 Posts
Default

Quote:
Originally Posted by henryzz View Post
Require or support? I have run a pcie 3 card in a 1.1 socket. It would surprise me if it won't run on current motherboards.

Yes, probably, but with lower performance.
  Reply With Quote
Old 2019-06-18, 12:50   #7
M344587487
 
M344587487's Avatar
 
"Composite as Heck"
Oct 2017

5×112 Posts
Default

They don't require PCIe 4.0 motherboards, it's a backwards compatible standard. A PCIe 4.0 compatible motherboard is required to use PCIe 4.0 but current PCIe 3.0 compatible motherboards will take these fine and run at PCIe 3.0 speeds. In fact for power efficiency we are probably better off using older motherboards or 3.0 speeds anyway, PCIe 4.0 and the signal boosting it requires bumps the chipset power requirements to the point of requiring a fan under load. There's only one, maybe two passively cooled x570 motherboards out there indicating that the chipset power consumption is probably 5-10 watts under load.
M344587487 is offline   Reply With Quote
Old 2019-06-18, 12:57   #8
kriesel
 
kriesel's Avatar
 
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest

23·32·59 Posts
Default

Quote:
Originally Posted by SELROC View Post
Yes, probably, but with lower performance.
Could you post some numbers for linux and your gpu, of bus load?
On Windows 7 pro x64, my GTX1080, running CUDALucas, 125M LL, on a PCIe 2.0 connection, GPU-Z is showing only 1% bus interface load while the GPU load is 100% and the gpu memory controller load is 76%. Running gpuowl v6.5 on 332M PRP on the same setup, it's showing rapidly varying 4-50% bus interface load, steady GPU load 100%, GPU memory controller load steady 52%. Gpuowl V6.5 PRP, on same 125M exponent, gpu load 100%, gpu memory controller load 60%, bus interface load varying from 8-52% (22% average).

Last fiddled with by kriesel on 2019-06-18 at 13:13
kriesel is offline   Reply With Quote
Old 2019-06-18, 13:05   #9
SELROC
 

19·47 Posts
Default

Quote:
Originally Posted by kriesel View Post
Could you post some numbers for linux and your gpu, of bus load?
On my GTX1080, running CUDALucas, 125M LL, on a PCIe 2.0 connection, GPU-Z is showing only 1% bus interface load while the GPU load is 100% and the gpu memory controller load is 76%. Running gpuowl v6.5 on 332M PRP on the same setup, it's showing rapidly varying 4-50% bus interface load, steady GPU load 100%, GPU memory controller load steady 52%.

Sorry I have omitted important information: in PRP case the bus load could only affect GEC data transfers I think.


I don't know GPU-Z, on linux there are various gpu diagnostic programs, with varying levels of confidence.
  Reply With Quote
Old 2019-06-18, 13:29   #10
kriesel
 
kriesel's Avatar
 
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest

23×32×59 Posts
Default

Quote:
Originally Posted by SELROC View Post
Sorry I have omitted important information: in PRP case the bus load could only affect GEC data transfers I think.

I don't know GPU-Z, on linux there are various gpu diagnostic programs, with varying levels of confidence.
What I observe while running gpuowl 6.5 in GPU-Z is a rapid and significant bus load variation on the gpu, from one 2.5-second sensor sampling period to the next, faster than the ~9.5 second PRP block size (1000 iterations) period and far faster than the usual (~3.2 minutes) 20,000 iteration screen output period or the ~2.6 hour (1,000,000 iterations) GEC period of the 125M PRP. Don't know why. Does not occur in CUDALucas, and the gpu in question is not connected to a display. GPU-Z info at https://www.techpowerup.com/gpuz/

Last fiddled with by kriesel on 2019-06-18 at 13:35
kriesel is offline   Reply With Quote
Old 2019-06-18, 14:00   #11
SELROC
 

7,687 Posts
Default

Quote:
Originally Posted by kriesel View Post
What I observe while running gpuowl 6.5 in GPU-Z is a rapid and significant bus load variation on the gpu, from one 2.5-second sensor sampling period to the next, faster than the ~9.5 second PRP block size (1000 iterations) period and far faster than the usual (~3.2 minutes) 20,000 iteration screen output period or the ~2.6 hour (1,000,000 iterations) GEC period of the 125M PRP. Don't know why. Does not occur in CUDALucas, and the gpu in question is not connected to a display. GPU-Z info at https://www.techpowerup.com/gpuz/

GPU-Z shows all information in one window, if you have a graphics interface on linux there is https://github.com/Eliovp/amdmemorytweak for AMD GPUs, also it can be used from the command line but I have not tested yet.
  Reply With Quote
Reply

Thread Tools


Similar Threads
Thread Thread Starter Forum Replies Last Post
Radeon VII (2nd gen consumer Vega GPU) M344587487 GPU Computing 276 2020-08-02 16:52
Vega 20 announced with 7.64 TFlops of FP64 M344587487 GPU Computing 4 2018-11-08 16:56
RX Vega performance xx005fs GPU Computing 5 2018-01-17 00:22
Monster cruncher with 96GB RAM and 20 HDs Jeff Gilchrist Hardware 9 2010-08-09 00:23
Monsters and Monster farms Unregistered Data 6 2004-08-12 00:28

All times are UTC. The time now is 05:02.

Sat Aug 15 05:02:50 UTC 2020 up 2 days, 1:38, 0 users, load averages: 2.45, 2.66, 2.72

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2020, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.