mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware > GPU Computing

Reply
 
Thread Tools
Old 2018-06-28, 10:02   #12
mackerel
 
mackerel's Avatar
 
Feb 2016
UK

26×7 Posts
Default

The use case for these is that most cards will be mounted off board, connected using PCIe riser cables.

There are some driver/OS limits on number of GPUs supported so that would be the next problem.
mackerel is offline   Reply With Quote
Old 2018-06-28, 10:20   #13
SELROC
 

2×4,583 Posts
Default

Quote:
Originally Posted by mackerel View Post
The use case for these is that most cards will be mounted off board, connected using PCIe riser cables.

There are some driver/OS limits on number of GPUs supported so that would be the next problem.



On Debian with amdgpu-pro there are no limits, but you would have to install also the nvidia-driver to mix GPUs on the same main board.
  Reply With Quote
Old 2018-10-02, 09:02   #14
SELROC
 

22×7×269 Posts
Default

Quote:
Originally Posted by SELROC View Post
On Debian with amdgpu-pro there are no limits, but you would have to install also the nvidia-driver to mix GPUs on the same main board.

I have to retract on this one. Mixing drivers will certainly cause more problems. I have not attempted that myself yet as I don't own a CUDA GPU.


However, the amdgpu driver is still having problems when I want to use more that 4 GPUs together. It shows a persisting error in this case.


I have been able to avoid the error by using only 4 GPUs at the same time.


Is this a driver limit? I don't know.
  Reply With Quote
Old 2018-10-08, 13:04   #15
niconiconi
 
niconiconi's Avatar
 
Oct 2018
Beijing, China

3 Posts
Default

Quote:
Originally Posted by SELROC View Post
I have to retract on this one. Mixing drivers will certainly cause more problems.
The problem of mixing drivers can be overcame by using virtualization technology.

IOMMU, also known as VT-d on Intel processor, is a hardware mechanism to map and isolate PCI-e in its own address space. This can improve security by stopping malicious hardware from access the main memory directly, but it is mainly used for accessing hardware directly from a virtual machine.

The setup is simple - you enable IOMMU on Linux host, and run two Linux KVM guests with PCI passthrough, on separate virtual machines, one for Nvidia, another one for AMD. Then the drivers and GPUs can work independently without interference. The overhead is negligible, because most work is done on GPUs, even computation on CPU is nearly as-fast-as native, since virtualization itself is accelerated by VT-x.

Here is a guide, https://bluehatrecord.wordpress.com/2015/12/05/performing-iommu-based-gpu-pci-passthrough-with-fedora-22/

The only problem is that, VT-d requires CPU support. Most Core i5 and Core i7 CPU does support it, but this feature is removed on CPUs ending with "K". Probably it is an intentional plan by Intel, to prevent cost-effective enthusiasts CPUs to be used in production. But no problem if you use Xeon.
niconiconi is offline   Reply With Quote
Old 2018-10-08, 13:40   #16
SELROC
 

215458 Posts
Default

Quote:
Originally Posted by mackerel View Post
The use case for these is that most cards will be mounted off board, connected using PCIe riser cables.

There are some driver/OS limits on number of GPUs supported so that would be the next problem.



And to mount 20 GPUs you need 3 power supplies on the mainboard. Problem is switching PSUs are not made to run in parallel...


we tried and got the beginnings of a fire outbreak.
  Reply With Quote
Old 2018-10-09, 00:29   #17
niconiconi
 
niconiconi's Avatar
 
Oct 2018
Beijing, China

3 Posts
Default

Quote:
Originally Posted by SELROC View Post
Problem is switching PSUs are not made to run in parallel...
Connecting +12V/Vcc together is surely problematic for most power supplies in general, but it shouldn't be a problem if you power each row of GPUs independently (don't interconnect +12V in between different supplies, e.g. each row of six GPUs to be powered by +12V from a single 1200W power supply). For those PCI-e riser card extender with Molex power connectors and USB 3.0 cables to carry PCI-e signal, the power is effectively isolated, and only the GND on PCI-e signal circuitry is interconnected. See https://nerdralph.blogspot.com/2017/...ning-rigs.html All don't forget to bridge a few control wires (e.g. Power-Good) from three power supplies together with two dual PSU adapters, to ensure they starts properly at the same time.

Last fiddled with by niconiconi on 2018-10-09 at 00:50
niconiconi is offline   Reply With Quote
Old 2018-10-09, 05:06   #18
SELROC
 

33×283 Posts
Default

Quote:
Originally Posted by niconiconi View Post
Connecting +12V/Vcc together is surely problematic for most power supplies in general, but it shouldn't be a problem if you power each row of GPUs independently (don't interconnect +12V in between different supplies, e.g. each row of six GPUs to be powered by +12V from a single 1200W power supply). For those PCI-e riser card extender with Molex power connectors and USB 3.0 cables to carry PCI-e signal, the power is effectively isolated, and only the GND on PCI-e signal circuitry is interconnected. See https://nerdralph.blogspot.com/2017/...ning-rigs.html All don't forget to bridge a few control wires (e.g. Power-Good) from three power supplies together with two dual PSU adapters, to ensure they starts properly at the same time.



The mainboard has 3 PSU connectors and it is supposed to power 18 GPU with 3 1200W PSU. The PSU are not IBM Dell or HP but absolutely CoolerMaster with 94% efficiency.


A total of 3600W will flow in the mainboard.


Personal advice: to do cryptomining you only run 1 program which exploits all GPUs, but with Mersenne you need to run 1 program for each GPU. It is difficult to manage and you need to develop scripts to start/stop the programs.
  Reply With Quote
Old 2018-10-09, 09:17   #19
niconiconi
 
niconiconi's Avatar
 
Oct 2018
Beijing, China

3 Posts
Default

Quote:
Originally Posted by SELROC View Post
The mainboard has 3 PSU connectors.
I didn't realize the mainboard has all the PSU connectors. Instead of being powered externally, all the power flows into the board?! Now I'm suspicious about the stability of this setup, 3300 W is even beyond the rating of a standard 230 VAC outlet/connector. I think it's better to rather stay around six cards, which is a time-proven setup used by many in for GPU computing...

Quote:
Originally Posted by SELROC View Post
Personal advice: to do cryptomining you only run 1 program which exploits all GPUs, but with Mersenne you need to run 1 program for each GPU. It is difficult to manage and you need to develop scripts to start/stop the programs.
Good point. It shouldn't be too difficult though. If anyone wants to mix AMD and Nvidia GPU, use KVM virtual machines w/ PCI passthough, and control them from host via virtual network interface with a script should work.

Nevertheless, just my two cents... I've only played with some BOINC projects, weak SHA bruteforcing analysis, GPU video encoding, and only became interested in factoring some numbers recently - all of these are just nearly fire-and-forgot tasks without needs for scheduling. I have absolutely no authority on GIMPS, nor large GPU clusters.

Last fiddled with by niconiconi on 2018-10-09 at 09:19
niconiconi is offline   Reply With Quote
Old 2018-10-09, 09:36   #20
SELROC
 

31×179 Posts
Default

Quote:
Originally Posted by niconiconi View Post
I didn't realize the mainboard has all the PSU connectors. Instead of being powered externally, all the power flows into the board?! Now I'm suspicious about the stability of this setup, 3300 W is even beyond the rating of a standard 230 VAC outlet/connector. I think it's better to rather stay around six cards, which is a time-proven setup used by many in for GPU computing...



Good point. It shouldn't be too difficult though. If anyone wants to mix AMD and Nvidia GPU, use KVM virtual machines w/ PCI passthough, and control them from host via virtual network interface with a script should work.

Nevertheless, just my two cents... I've only played with some BOINC projects, weak SHA bruteforcing analysis, GPU video encoding, and only became interested in factoring some numbers recently - all of these are just nearly fire-and-forgot tasks without needs for scheduling. I have absolutely no authority on GIMPS, nor large GPU clusters.



You are right to be suspicious, that system *has* stability problems.


Cryptomining is done with a proprietary system, I think they use a proprietary driver (not AMD) and that driver supports 18 gpus. The AMD driver shows errors with more than 4 gus.


After all, that makes sense, gaming mainboards come with maximum 4 Gen3 slots.
  Reply With Quote
Old 2018-10-09, 09:45   #21
SELROC
 

7·839 Posts
Default

News from AMD: https://www.phoronix.com/scan.php?pa...m-Polaris-Soon
  Reply With Quote
Old 2018-10-09, 23:57   #22
mackerel
 
mackerel's Avatar
 
Feb 2016
UK

44810 Posts
Default

Quote:
Originally Posted by SELROC View Post
The mainboard has 3 PSU connectors and it is supposed to power 18 GPU with 3 1200W PSU. The PSU are not IBM Dell or HP but absolutely CoolerMaster with 94% efficiency.

A total of 3600W will flow in the mainboard.
It doesn't work that way. The bulk of the power will be provided directly from the PSU to the GPUs via PCIe 6 or 8 pin power connectors. A small amount of power may be provided via the PCIe slot, but that looks like it can be taken care of either by the 4 pin molex connectors on the mobo, or risers boards often have their own separate power input.

The ATX 24-pin PSU connectors, likely one is the main one that actually powers the rest of the mobo, and the other two are provided to send a switch signal to the other two PSUs. CPU is powered by the EPS connector. In modern systems not a lot of power goes through the ATX power connectors.
mackerel is offline   Reply With Quote
Reply



Similar Threads
Thread Thread Starter Forum Replies Last Post
Prime95 and graphics cards keithschmidt Information & Answers 45 2016-09-10 10:08
New Linux rootkit leverages graphics cards for stealth. swl551 Lounge 0 2015-05-08 14:06
Nvidia's next-generation graphics cards ixfd64 GPU Computing 22 2014-11-15 04:25
how do graphics cards work so fast? ixfd64 Hardware 1 2004-06-02 03:01
Chance to use modern Graphics Cards as.. Marco Hardware 28 2003-11-02 23:21

All times are UTC. The time now is 03:46.


Fri Jul 7 03:46:27 UTC 2023 up 323 days, 1:15, 0 users, load averages: 0.79, 0.86, 1.10

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2023, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.

≠ ± ∓ ÷ × · − √ ‰ ⊗ ⊕ ⊖ ⊘ ⊙ ≤ ≥ ≦ ≧ ≨ ≩ ≺ ≻ ≼ ≽ ⊏ ⊐ ⊑ ⊒ ² ³ °
∠ ∟ ° ≅ ~ ‖ ⟂ ⫛
≡ ≜ ≈ ∝ ∞ ≪ ≫ ⌊⌋ ⌈⌉ ∘ ∏ ∐ ∑ ∧ ∨ ∩ ∪ ⨀ ⊕ ⊗ 𝖕 𝖖 𝖗 ⊲ ⊳
∅ ∖ ∁ ↦ ↣ ∩ ∪ ⊆ ⊂ ⊄ ⊊ ⊇ ⊃ ⊅ ⊋ ⊖ ∈ ∉ ∋ ∌ ℕ ℤ ℚ ℝ ℂ ℵ ℶ ℷ ℸ 𝓟
¬ ∨ ∧ ⊕ → ← ⇒ ⇐ ⇔ ∀ ∃ ∄ ∴ ∵ ⊤ ⊥ ⊢ ⊨ ⫤ ⊣ … ⋯ ⋮ ⋰ ⋱
∫ ∬ ∭ ∮ ∯ ∰ ∇ ∆ δ ∂ ℱ ℒ ℓ
𝛢𝛼 𝛣𝛽 𝛤𝛾 𝛥𝛿 𝛦𝜀𝜖 𝛧𝜁 𝛨𝜂 𝛩𝜃𝜗 𝛪𝜄 𝛫𝜅 𝛬𝜆 𝛭𝜇 𝛮𝜈 𝛯𝜉 𝛰𝜊 𝛱𝜋 𝛲𝜌 𝛴𝜎𝜍 𝛵𝜏 𝛶𝜐 𝛷𝜙𝜑 𝛸𝜒 𝛹𝜓 𝛺𝜔