mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > PrimeNet > GPU to 72

Reply
 
Thread Tools
Old 2022-05-30, 21:38   #34
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

2·33·197 Posts
Default

Quote:
Originally Posted by TheJudger View Post
Hi, do you have CUDA toolkit 11.x installed on that system? If so just grab the code, perhaps adjust some pathes in the makefile and run make. That is the prefered way. P.S. this isn't specific to A100 cards, every Ampere card needs a CUDA 11 compiled binary of mfaktc.
Thanks greatly, Oliver.

I have no idea what is currently "under" current Colab instances. I'll climb into one tomorrow by way of reverse SSH, and take a look.

There's nothing quite like talking directly to the developer to get high-bandwidth guidance!
chalsall is offline   Reply With Quote
Old 2022-05-31, 22:40   #35
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

2·33·197 Posts
Default

Quote:
Originally Posted by chalsall View Post
While I'm writing though... For the last couple of weeks, I have been getting nothing *BUT* T4s! For ~ 3.5 hours per day.
I found this article interesting...

Correlation does not mean causality... 9-)
chalsall is offline   Reply With Quote
Old 2022-06-01, 19:45   #36
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

2·33·197 Posts
Default

Quote:
Originally Posted by chalsall View Post
I have no idea what is currently "under" current Colab instances. I'll climb into one tomorrow by way of reverse SSH, and take a look.
Hey Oliver et al... So, to answer your question, it looks like the current (at least "Free Tier") Colab instances are installed with cuda-10.0. It /might/ be possible to replace this with a different version during the ephemeral run...

root@Colab_CVH:/# lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 18.04.5 LTS
Release: 18.04
Codename: bionic

root@Colab_CVH:~# lspci
00:00.0 Host bridge: Intel Corporation 440FX - 82441FX PMC [Natoma] (rev 02)
00:01.0 ISA bridge: Intel Corporation 82371AB/EB/MB PIIX4 ISA (rev 03)
00:01.3 Bridge: Intel Corporation 82371AB/EB/MB PIIX4 ACPI (rev 03)
00:03.0 Non-VGA unclassified device: Red Hat, Inc. Virtio SCSI
00:04.0 3D controller: NVIDIA Corporation Device 1eb8 (rev a1)
00:05.0 Ethernet controller: Red Hat, Inc. Virtio network device
00:06.0 Unclassified device [00ff]: Red Hat, Inc. Virtio RNG

root@Colab_CVH:/usr/local/cuda-10.0/bin# ./nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2018 NVIDIA Corporation
Built on Sat_Aug_25_21:08:01_CDT_2018
Cuda compilation tools, release 10.0, V10.0.130

root@Colab_CVH:~# nvidia-smi
Failed to initialize NVML: Driver/library version mismatch


Does anyone have any suggestions?
chalsall is offline   Reply With Quote
Old 2022-06-02, 11:37   #37
LaurV
Romulan Interpreter
 
LaurV's Avatar
 
"name field"
Jun 2011
Thailand

10,093 Posts
Default

Yarrr...

Isn't that a waste?...

BTW, my smi shows cuda 11 available ?? (are you in the wrong folder? hihi)
(you are talking foreign languages anyhow, so sorry if I am out of line)


Click image for larger version

Name:	nvsmi.JPG
Views:	47
Size:	73.9 KB
ID:	26954 Click image for larger version

Name:	nvsmi2.JPG
Views:	36
Size:	63.8 KB
ID:	26955 Click image for larger version

Name:	nvsmi3.JPG
Views:	37
Size:	82.4 KB
ID:	26956

Last fiddled with by LaurV on 2022-06-02 at 11:37
LaurV is online now   Reply With Quote
Old 2022-08-13, 14:18   #38
DrobinsonPE
 
Aug 2020

2238 Posts
Default

This happened this morning with 20 minutes of repeating selftests before I caught the issue. I changed the work type in GPU72 from "LMH breadth first" back to "GPU72 decides", re-launched colab and got an assignment.

My other colab has not been getting any GPU work for the last few days so I am going to change it from "LMH normal" back to "GPU72 decides" to see if it will start getting a GPU again.

Code:
20220813_134952 ( 0:19): no factor for M332698771 from 2^76 to 2^77 [mfaktc 0.21 barrett87_mul32_gs]
20220813_134952 ( 0:19): tf(): time spent since restart:   16m  4.768s
20220813_134952 ( 0:19):       estimated total time spent: 44m  6.220s
20220813_134952 ( 0:19): Fetching initial work...
20220813_134956 ( 0:19): Running GPU type Tesla T4

20220813_134956 ( 0:19): running a simple selftest...
20220813_135003 ( 0:19): Selftest statistics
20220813_135003 ( 0:19):   number of tests           107
20220813_135003 ( 0:19):   successfull tests         107
20220813_135003 ( 0:19): selftest PASSED!
20220813_135003 ( 0:19): Fetching initial work...
20220813_135004 ( 0:19): Running GPU type Tesla T4

20220813_135004 ( 0:19): running a simple selftest...
20220813_135010 ( 0:19): Selftest statistics
20220813_135010 ( 0:19):   number of tests           107
20220813_135010 ( 0:19):   successfull tests         107
20220813_135010 ( 0:19): selftest PASSED!
20220813_135010 ( 0:19): Fetching initial work...
20220813_135011 ( 0:19): Running GPU type Tesla T4
DrobinsonPE is offline   Reply With Quote
Old 2022-08-13, 19:40   #39
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

2×33×197 Posts
Default

Quote:
Originally Posted by DrobinsonPE View Post
This happened this morning with 20 minutes of repeating selftests before I caught the issue.
Sorry... SPE on my part...

The 332M range had a stupid SELECT statement. When the imported candidates at 76 bits were exhausted, the query failed.

I have had GPU72 reserve from Primenet an additional 1,956 candidates in the 332.7M range. And, also fixed the query such that if this happens again in the future, an assignment will still be given.

BTW... For anyone following along... The TF'ing is now ***well*** ahead of the FTC wavefront. The P-1 workers are similarly comfortable.

I would ask for some guidance from those who understand the maths far better than I do...

Should we just keep taking ranges up to 77 bits? Or perhaps invest some resources in the 332M range? Or the Cleanup sub-sub-sub project?

It is all great fun. I enjoy being a very small part of it.
chalsall is offline   Reply With Quote
Reply

Thread Tools


Similar Threads
Thread Thread Starter Forum Replies Last Post
Colab question David703 GPU to 72 302 2022-07-01 03:41
Revisiting msieve GPU with Colab EdH Msieve 4 2020-08-24 02:22
GPU72 + Colab multiple assignments windowfrog GPU to 72 2 2020-07-07 19:47
Alternatives to Google Colab kriesel Cloud Computing 11 2020-01-14 18:45
AVX2 Troubles with Colab Instance EdH YAFU 5 2019-11-12 01:49

All times are UTC. The time now is 12:08.


Sun Sep 25 12:08:51 UTC 2022 up 38 days, 9:37, 0 users, load averages: 0.88, 1.14, 1.11

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2022, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.

≠ ± ∓ ÷ × · − √ ‰ ⊗ ⊕ ⊖ ⊘ ⊙ ≤ ≥ ≦ ≧ ≨ ≩ ≺ ≻ ≼ ≽ ⊏ ⊐ ⊑ ⊒ ² ³ °
∠ ∟ ° ≅ ~ ‖ ⟂ ⫛
≡ ≜ ≈ ∝ ∞ ≪ ≫ ⌊⌋ ⌈⌉ ∘ ∏ ∐ ∑ ∧ ∨ ∩ ∪ ⨀ ⊕ ⊗ 𝖕 𝖖 𝖗 ⊲ ⊳
∅ ∖ ∁ ↦ ↣ ∩ ∪ ⊆ ⊂ ⊄ ⊊ ⊇ ⊃ ⊅ ⊋ ⊖ ∈ ∉ ∋ ∌ ℕ ℤ ℚ ℝ ℂ ℵ ℶ ℷ ℸ 𝓟
¬ ∨ ∧ ⊕ → ← ⇒ ⇐ ⇔ ∀ ∃ ∄ ∴ ∵ ⊤ ⊥ ⊢ ⊨ ⫤ ⊣ … ⋯ ⋮ ⋰ ⋱
∫ ∬ ∭ ∮ ∯ ∰ ∇ ∆ δ ∂ ℱ ℒ ℓ
𝛢𝛼 𝛣𝛽 𝛤𝛾 𝛥𝛿 𝛦𝜀𝜖 𝛧𝜁 𝛨𝜂 𝛩𝜃𝜗 𝛪𝜄 𝛫𝜅 𝛬𝜆 𝛭𝜇 𝛮𝜈 𝛯𝜉 𝛰𝜊 𝛱𝜋 𝛲𝜌 𝛴𝜎𝜍 𝛵𝜏 𝛶𝜐 𝛷𝜙𝜑 𝛸𝜒 𝛹𝜓 𝛺𝜔