mersenneforum.org

mersenneforum.org (https://www.mersenneforum.org/index.php)
-   GPU Computing (https://www.mersenneforum.org/forumdisplay.php?f=92)
-   -   mfaktc: a CUDA program for Mersenne prefactoring (https://www.mersenneforum.org/showthread.php?t=12827)

kriesel 2020-02-07 15:57

RTX2080 Super mfaktc tune, likes 2047Mib GPUSieveSize or maybe more
 
It would apparently benefit from a further increase in max GPUSieveSize.[CODE]RTX2080 Super mfaktc tuning; 2047-enabled CUDA 10 version
M441000023 from 2^80 to 2^81

Starting from
GPUSieveSize=2047, GPUSieveProcessSize=32, GPUSievePrimes=100000 2936.74

GPUSieveProcessSize=32 GPUSieveSize 2040 2936.37
GPUSieveProcessSize=24 GPUSieveSize 2040 2925.78
GPUSieveProcessSize=16 GPUSieveSize 2040 3021.20 *
GPUSieveProcessSize=8 GPUSieveSize 2040 2996.12

GPUSieveProcessSize=16, GPUSievePrimes=100000 vary GPUSieveSize
GPUSieveSize=2047 2936.67 *
GPUSieveSize=1536 2914.86
GPUSieveSize=1024 2897.07
GPUSieveSize=512 2810.90
GPUSieveSize=256 2648.74

GPUSieveProcessSize=16, GPUSieveSize=2047
GPUSievePrimes=80000 2931.64
GPUSievePrimes=90000 2936.92
GPUSievePrimes=100000 2936.43
GPUSievePrimes=94000 2930.264
GPUSievePrimes=92000 2937.64 *

2 instances
1 1506.48 tuned
2 1475.14 tuned
total 2981.62
2981.62 / 2937.64 = 1.015 ratio 2-instance/1-instance
A third instance would probably help a little.

nvidia-smi indicates 99% gpu load not 100%, indicating there's still some untapped capacity with tune plus two-instance operation, 98% with one.

recheck later, unchanged tune
1 1497.12
2 1486.10
total 2983.22, still 99%; 97-102%TDP, up to 1875Mhz
[/CODE]

storm5510 2020-04-14 16:09

[QUOTE=MrRepunit;529390]Hi all,
just want to announce here in this thread that I finished the generalized repunits version of mfaktc, the long wished for generalization of my base 10 repunits mfaktc variant.
It supports only positive bases (b>=2).

I did not want to hijack this thread, so I created my own:
[URL]https://www.mersenneforum.org/showthread.php?t=24901[/URL]

Feel free to test it.[/QUOTE]

I have not been here is quite some time. Has there been any updates to [I]gr-mfaktc[/I]? I seem to recall a throughput issue when compared to the CUDA10 version of [I]mfaktc[/I]. It runs around above 1000 GHz-days/day. [I]gr-mfaktc[/I] ran in the low 700's. This is on my hardware; GTX 1080 with Win 10 v1903.

kriesel 2020-04-15 09:52

I've been running into gpu stoppages lately on my little miner rig. It's not clear what's cause and what's effect, but I'm seeing the mfaktc instances stop working, mmff too, Windows TDR events, and the known-bad-factor occur. Sometimes the failures are when switching gpu-z from tab to tab.

[CODE]Date Time | class Pct | time ETA | GHz-d/day Sieve Wait
Apr 14 22:38 | 3273 70.8% | [B]43.799[/B] 3h24m | 757.24 92725 n.a.%
M332233123 has a factor: [B]38814612911305349835664385407[/B]
ERROR: cudaGetLastError() returned 6: the launch timed out and was terminated
batch wrapper reports mfaktc exited at Tue 04/14/2020 22:38:56.98
[/CODE]

storm5510 2020-04-25 23:17

[QUOTE=kriesel;542746]I've been running into gpu stoppages lately on my little miner rig. It's not clear what's cause and what's effect, but I'm seeing the mfaktc instances stop working, mmff too, Windows TDR events, and the known-bad-factor occur. Sometimes the failures are when switching gpu-z from tab to tab.

[CODE]Date Time | class Pct | time ETA | GHz-d/day Sieve Wait
Apr 14 22:38 | 3273 70.8% | [B]43.799[/B] 3h24m | 757.24 92725 n.a.%
M332233123 has a factor: [B]38814612911305349835664385407[/B]
ERROR: cudaGetLastError() returned 6: the launch timed out and was terminated
batch wrapper reports mfaktc exited at Tue 04/14/2020 22:38:56.98
[/CODE][/QUOTE]

Are you over-clocking? What I see in the first paragraph would tend to indicate you are. If so, ease off the throttle a little.

kriesel 2020-04-25 23:37

[QUOTE=storm5510;543823]Are you over-clocking? What I see in the first paragraph would tend to indicate you are. If so, ease off the throttle a little.[/QUOTE]Nope; I never overclock. I'll underclock on occasion to get reliability. The results you responded about were stock clock throughout.

James Heinrich 2020-04-25 23:50

[QUOTE=kriesel;543824]Nope; I never overclock. I'll underclock on occasion to get reliability. The results you responded about were stock clock throughout.[/QUOTE]Still, the point may be valid, I find that hardware (both GPUs and CPUs) tend to "wear out" and tolerate lower clocks/heat as the years go by, irrelevant of cleanliness & cooling. It would [COLOR="Magenta"]not[/COLOR] hurt to try underclocking the questionable GPU and see if the problem goes away.

kriesel 2020-04-26 07:54

[QUOTE=James Heinrich;543826]Still, the point may be valid, I find that hardware (both GPUs and CPUs) tend to "wear out" and tolerate lower clocks/heat as the years go by, irrelevant of cleanliness & cooling. It would hurt to try underclocking the questionable GPU and see if the problem goes away.[/QUOTE]Thanks. It was a new gpu on a new 1200W PSU. The wrong factor has not occurred in the past 10 days. No clock change. That gpu is still on the system but various extenders have been shuffled around. It might have been a failing PCIe extender pad.

storm5510 2020-04-26 15:35

[QUOTE=kriesel;543824]Nope; I never overclock. I'll underclock on occasion to get reliability. The results you responded about were stock clock throughout.[/QUOTE]

I never did either. I did not want to risk damaging something.

[QUOTE=James Heinrich]Still, the point may be valid, I find that hardware (both GPUs and CPUs) tend to "wear out" and tolerate lower clocks/heat as the years go by, irrelevant of cleanliness & cooling. It would hurt to try underclocking the questionable GPU and see if the problem goes away. [/QUOTE]

I've had my GTX 1080 for two years now. It does [U]not[/U] perform the way it used to. What was once 1200+ GHz-days/day using some programs is now down to 950 to 1050. I believe James left out a word in what he wrote above. "It would [U]not[/U] hurt to try..." I do not run mine at 100% capacity now and haven't for quite some time. I use [I]MSI Afterburner [/I]to reduce it to 80% of capacity. Doing so has just a slight affect on throughput and reduces the operating temperature by 12°C, on average.

lalera 2020-05-02 21:25

hi,
i have a gtx 580 that is about 8 years old
running nearly the same speed as new
but at higher temperatures
sometimes i had problems with drivers
i use this card sometimes for trial factoring with mfaktc

ZFR 2020-05-06 23:11

Sorry to derail the thread a bit, but quick question: does changing CheckpointDelay to a lower value like 5 or 10 add a lot of overhead? Will it affect performance?

Thanks.

kladner 2020-05-09 02:46

[QUOTE=lalera;544476]hi,
i have a gtx 580 that is about 8 years old
running nearly the same speed as new
but at higher temperatures
sometimes i had problems with drivers
i use this card sometimes for trial factoring with mfaktc[/QUOTE]
Thermal compound is probably shot. If you went after that you could also deep clean the cooler.


All times are UTC. The time now is 22:42.

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.