mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > PrimeNet > GPU to 72

Reply
Thread Tools
Old 2019-11-05, 17:08   #4379
James Heinrich
 
James Heinrich's Avatar
 
"James Heinrich"
May 2004
ex-Northern Ontario

11×311 Posts
Default

Quote:
Originally Posted by storm5510 View Post
I managed to muddle my way though it. It's running.
...I've tread into an area I have no experience with.
I likewise. Perhaps there is another introductory thread that I missed where this is all explained?

Some FAQ:
  • do I need to leave the browser window open, or does it run in the background?
  • it seemed to fetch 3 assignments, does it quit after than and I need to restart it, or does it keep fetching more work?
  • I see manual results on gpu72.com that need to be manually submitted (at least until Chris gets around to automating it)
  • can I run more than one instance (per Google account)?
James Heinrich is offline   Reply With Quote
Old 2019-11-05, 17:34   #4380
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

100110001001112 Posts
Default

Quote:
Originally Posted by James Heinrich View Post
I likewise. Perhaps there is another introductory thread that I missed where this is all explained?
Thanks for the "ping" guys. This is a "hoot", but things have been moving so quickly that there isn't yet a FAQ. Thank you for your below:

Quote:
Originally Posted by James Heinrich View Post
Some FAQ:
  • do I need to leave the browser window open, or does it run in the background?
  • it seemed to fetch 3 assignments, does it quit after than and I need to restart it, or does it keep fetching more work?
  • I see manual results on gpu72.com that need to be manually submitted (at least until Chris gets around to automating it)
  • can I run more than one instance (per Google account)?
1. Any "interactive" Notebook sessions are shut down shortly after the browser is closed.

1.1. Kaggle lets you "commit" a Notebook, where-in every Section runs, in order, until the last executable exists.

1.2. TL;DR: Leave your browser open if possible.

2. The GPU72_TF Notebook fetches three (3#) TF assignments initially and then gets to work.

2.1. Assignments are first "reissued" from previous Notebook runs which have been "killed" (RIP), and then new assignments as specified by the AKey's work preference.

2.2. Once an assignment is completed is reported back to GPU72, and another assignment is fetch.

3. Yeah... Sorry. I subscribe strongly to "Never send a human to do a machine's job". But often achieving that ideal involves a human. In this case, it involves my time...

3.1. I have mapped in my head a solution space for this (read: automatically submitting back to Primenet the Instance(s)' results), but things have been a little hectic in the last few weeks.

3.1.1. Still on one of my whiteboards, as well as in my pen-and-paper workbook.

4. Nominally ill-advised. Although there could be some workflows where this would make sense (constrained human resources, for example).

4.1. Empirical experimentation suggests that each Colab Account gets ~12 to 16 hours of GPU compute per day.

4.2. Kaggle is contrained to ~30 hours of P100 GPU per week per account. If you're creative, you can actually get ~38.99 hours...

4.3. Interestingly, different Google Accounts seem to be thusly individually temporally constrained. Even when running within the same browser context (and thus OS fingerprint, IP address, and even MAC address).

Last fiddled with by chalsall on 2019-11-05 at 17:44 Reason: Second 4.2 -> 4.3.
chalsall is offline   Reply With Quote
Old 2019-11-05, 18:00   #4381
James Heinrich
 
James Heinrich's Avatar
 
"James Heinrich"
May 2004
ex-Northern Ontario

11·311 Posts
Default

Thanks, that helps.
I also discovered I don't need to copy-paste code per Uncwilly's post, I just need to click the magic Colaboratory link on gpu72.com after creating a NAK and copy-paste in the Access Key.
I always have two browsers open with my home and work Google accounts signed in, so I fired up a second instance on my other browser and it seems to run fine (except my first attempt got me a "Tesla P100-PCIE-16GB" (1140 GHd/d) and the second a notably slower "Tesla K80" (390 GHd/d), luck of the draw I guess).

Last fiddled with by James Heinrich on 2019-11-05 at 18:02
James Heinrich is offline   Reply With Quote
Old 2019-11-05, 18:09   #4382
petrw1
1976 Toyota Corona years forever!
 
petrw1's Avatar
 
"Wayne"
Nov 2006
Saskatchewan, Canada

22×3×17×23 Posts
Default

Looking at your charts in www.mersenne.ca for GPU-TF vs. GPU-LL performance it seems these Tesla 100 and K80 are relatively much better at LL than TF. I assume LL includes P-1.

For this reason, I would prefer to use these GPUs (especially the K80) for P-1 rather than TF.

Have people had much luck running CUDA-P1 in CoLab or Kaggle?
petrw1 is offline   Reply With Quote
Old 2019-11-05, 18:32   #4383
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

976710 Posts
Default

Quote:
Originally Posted by petrw1 View Post
For this reason, I would prefer to use these GPUs (especially the K80) for P-1 rather than TF.
As the "owner" of the resources, you're free to do whatever you want with them.

Please know, though, that Primenet is not currently lacking in either LL'ing nor P-1'ing resources.

Quote:
Originally Posted by petrw1 View Post
Have people had much luck running CUDA-P1 in CoLab or Kaggle?
My understanding is that both CUDA P-1 and LL code have been successfully built and run on both Colab and Kaggle.

I also (possibly correctly; possibly not) understand that the OpenCL LL code implementation is actually more efficient than the native CUDA one. Outside of my experience space to understand why.

To say again what I've said before... The GPU72_TF experiment was a "proof-of-concept". Just seeing if what we thought might be possible actually was.

Once that knowledge was established, other things can then be done...
chalsall is offline   Reply With Quote
Old 2019-11-05, 21:46   #4384
petrw1
1976 Toyota Corona years forever!
 
petrw1's Avatar
 
"Wayne"
Nov 2006
Saskatchewan, Canada

22·3·17·23 Posts
Default

Quote:
Originally Posted by chalsall View Post
Please know, though, that Primenet is not currently lacking in either LL'ing nor P-1'ing resources.
I'd have to agree.

Thx
petrw1 is offline   Reply With Quote
Old 2019-11-06, 16:20   #4385
storm5510
Random Account
 
storm5510's Avatar
 
Aug 2009

22·3·163 Posts
Default

I got up early this morning and found my colaboratory instance had stopped. Looking at the details, I saw "spider" so I figured someone had been working on it during the wee hours of the morning. The spider appeared to be functioning properly the last time I checked.

I am still running 274 locally. It is getting close to 98-million. I am wondering what happens when the 99's are complete. I changed the "High" value in the GPU72config file to 110,000,000. However, I do not know if the allocation from PrimeNet goes that far.

If the allocation does not go that far, then I imagine there will be a wrap-around back to smaller exponents running to 275. That will be fine. At 276, I will stop because my colab instance can run those quite a bit faster than my 1080.

In the interim, something else may come down the road.
storm5510 is offline   Reply With Quote
Old 2019-11-06, 17:56   #4386
James Heinrich
 
James Heinrich's Avatar
 
"James Heinrich"
May 2004
ex-Northern Ontario

11×311 Posts
Default

Quote:
Originally Posted by James Heinrich View Post
my first attempt got me a "Tesla P100-PCIE-16GB" (1140 GHd/d) and the second a notably slower "Tesla K80" (390 GHd/d), luck of the draw I guess.
I lost my connection to the P100 and I've got a K80 on both accounts now.
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my mfaktc table, which makes sense.
James Heinrich is offline   Reply With Quote
Old 2019-11-06, 18:39   #4387
kracker
 
kracker's Avatar
 
"Mr. Meeseeks"
Jan 2012
California, USA

23·271 Posts
Default

Quote:
Originally Posted by James Heinrich View Post
I lost my connection to the P100 and I've got a K80 on both accounts now.
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my mfaktc table, which makes sense.
Might be just coincidence, but I tried resetting and restarting the runtimes and after a try or two usually get a P100...
kracker is offline   Reply With Quote
Old 2019-11-06, 18:51   #4388
James Heinrich
 
James Heinrich's Avatar
 
"James Heinrich"
May 2004
ex-Northern Ontario

11×311 Posts
Default

Quote:
Originally Posted by kracker View Post
Might be just coincidence, but I tried resetting and restarting the runtimes and after a try or two usually get a P100...
Lucky you. I tried restarting 5 times each and get nothing but K80.
James Heinrich is offline   Reply With Quote
Old 2019-11-07, 00:47   #4389
storm5510
Random Account
 
storm5510's Avatar
 
Aug 2009

111101001002 Posts
Default

Quote:
Originally Posted by James Heinrich View Post
I lost my connection to the P100 and I've got a K80 on both accounts now.
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my mfaktc table, which makes sense.
Perhaps what's going on there is that you have two accounts running on a single Public IP address. Colab probably sees this as a double-instance. Therefore, K80.

I switched browsers on my HP earlier today so both my desktops would be using Firefox. They keep each other synced. I also got a K80 on the HP. I ended up deleting my instance on Colab and recreated it with the same code. P100 the first try. I only run one computer with it. So, you would probably have to drop one as well to get a P100 again.
storm5510 is offline   Reply With Quote
Reply



Similar Threads
Thread Thread Starter Forum Replies Last Post
Status Primeinator Operation Billion Digits 5 2011-12-06 02:35
62 bit status 1997rj7 Lone Mersenne Hunters 27 2008-09-29 13:52
OBD Status Uncwilly Operation Billion Digits 22 2005-10-25 14:05
1-2M LLR status paulunderwood 3*2^n-1 Search 2 2005-03-13 17:03
Status of 26.0M - 26.5M 1997rj7 Lone Mersenne Hunters 25 2004-06-18 16:46

All times are UTC. The time now is 09:34.


Mon Aug 2 09:34:37 UTC 2021 up 10 days, 4:03, 0 users, load averages: 1.01, 1.18, 1.29

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.