![]() |
[QUOTE=storm5510;529725]I managed to muddle my way though it. It's running.
...I've tread into an area I have no experience with. :smile:[/QUOTE]I likewise. Perhaps there is another introductory thread that I missed where this is all explained? Some FAQ:[LIST][*]do I need to leave the browser window open, or does it run in the background?[*]it seemed to fetch 3 assignments, does it quit after than and I need to restart it, or does it keep fetching more work?[*]I see manual results on gpu72.com that need to be manually submitted (at least until Chris gets around to automating it)[*]can I run more than one instance (per Google account)?[/LIST] |
[QUOTE=James Heinrich;529731]I likewise. Perhaps there is another introductory thread that I missed where this is all explained?[/QUOTE]
Thanks for the "ping" guys. This is a "hoot", but things have been moving so quickly that there isn't yet a FAQ. Thank you for your below: [QUOTE=James Heinrich;529731]Some FAQ:[LIST][*]do I need to leave the browser window open, or does it run in the background?[*]it seemed to fetch 3 assignments, does it quit after than and I need to restart it, or does it keep fetching more work?[*]I see manual results on gpu72.com that need to be manually submitted (at least until Chris gets around to automating it)[*]can I run more than one instance (per Google account)?[/LIST][/QUOTE] 1. Any "interactive" Notebook sessions are shut down shortly after the browser is closed. 1.1. Kaggle lets you "commit" a Notebook, where-in every Section runs, in order, until the last executable exists. 1.2. TL;DR: Leave your browser open if possible. 2. The GPU72_TF Notebook fetches three (3#) TF assignments initially and then gets to work. 2.1. Assignments are first "reissued" from previous Notebook runs which have been "killed" (RIP), and then new assignments as specified by the AKey's work preference. 2.2. Once an assignment is completed is reported back to GPU72, and another assignment is fetch. 3. Yeah... Sorry. I subscribe strongly to "Never send a human to do a machine's job". But often achieving that ideal involves a human. In this case, it involves my time... :wink: 3.1. I have mapped in my head a solution space for this (read: automatically submitting back to Primenet the Instance(s)' results), but things have been a little hectic in the last few weeks. 3.1.1. Still on one of my whiteboards, as well as in my pen-and-paper workbook. 4. Nominally ill-advised. Although there could be some workflows where this would make sense (constrained human resources, for example). 4.1. Empirical experimentation suggests that each Colab Account gets ~12 to 16 hours of GPU compute per day. 4.2. Kaggle is contrained to ~30 hours of P100 GPU per week per account. If you're creative, you can actually get ~38.99 hours... 4.3. Interestingly, different Google Accounts seem to be thusly individually temporally constrained. Even when running within the same browser context (and thus OS fingerprint, IP address, and even MAC address). |
Thanks, that helps.
I also discovered I don't need to copy-paste code per [i]Uncwilly[/i]'s post, I just need to click the magic Colaboratory link on gpu72.com after creating a NAK and copy-paste in the Access Key. I always have two browsers open with my home and work Google accounts signed in, so I fired up a second instance on my other browser and it seems to run fine (except my first attempt got me a "Tesla P100-PCIE-16GB" (1140 GHd/d) and the second a notably slower "Tesla K80" (390 GHd/d), luck of the draw I guess). |
Looking at your charts in [url]www.mersenne.ca[/url] for GPU-TF vs. GPU-LL performance it seems these Tesla 100 and K80 are relatively much better at LL than TF. I assume LL includes P-1.
For this reason, I would prefer to use these GPUs (especially the K80) for P-1 rather than TF. Have people had much luck running CUDA-P1 in CoLab or Kaggle? |
[QUOTE=petrw1;529740]For this reason, I would prefer to use these GPUs (especially the K80) for P-1 rather than TF.[/QUOTE]
As the "owner" of the resources, you're free to do whatever you want with them. Please know, though, that Primenet is not currently lacking in either LL'ing nor P-1'ing resources. [QUOTE=petrw1;529740]Have people had much luck running CUDA-P1 in CoLab or Kaggle?[/QUOTE] My understanding is that both CUDA P-1 and LL code have been successfully built and run on both Colab and Kaggle. I also (possibly correctly; possibly not) understand that the OpenCL LL code implementation is actually more efficient than the native CUDA one. Outside of my experience space to understand why. To say again what I've said before... The GPU72_TF experiment was a "proof-of-concept". Just seeing if what we thought might be possible actually was. Once that knowledge was established, other things can then be done... |
[QUOTE=chalsall;529742]Please know, though, that Primenet is not currently lacking in either LL'ing nor P-1'ing resources.
[/QUOTE] I'd have to agree. Thx |
I got up early this morning and found my colaboratory instance had stopped. Looking at the details, I saw "spider" so I figured someone had been working on it during the wee hours of the morning. The spider appeared to be functioning properly the last time I checked.
I am still running 2[SUP]74[/SUP] locally. It is getting close to 98-million. I am wondering what happens when the 99's are complete. I changed the "High" value in the [I]GPU72config[/I] file to 110,000,000. However, I do not know if the allocation from [I]PrimeNet[/I] goes that far. If the allocation does not go that far, then I imagine there will be a wrap-around back to smaller exponents running to 2[SUP]75[/SUP]. That will be fine. At 2[SUP]76[/SUP], I will stop because my colab instance can run those quite a bit faster than my 1080. In the interim, something else may come down the road. :popcorn: |
[QUOTE=James Heinrich;529738]my first attempt got me a "Tesla P100-PCIE-16GB" (1140 GHd/d) and the second a notably slower "Tesla K80" (390 GHd/d), luck of the draw I guess.[/QUOTE]I lost my connection to the P100 and I've got a K80 on both accounts now. :sad:
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my [url=https://www.mersenne.ca/mfaktc.php?filter=K80]mfaktc table[/url], which makes sense. |
[QUOTE=James Heinrich;529816]I lost my connection to the P100 and I've got a K80 on both accounts now. :sad:
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my [url=https://www.mersenne.ca/mfaktc.php?filter=K80]mfaktc table[/url], which makes sense.[/QUOTE] Might be just coincidence, but I tried resetting and restarting the runtimes and after a try or two usually get a P100... |
[QUOTE=kracker;529823]Might be just coincidence, but I tried resetting and restarting the runtimes and after a try or two usually get a P100...[/QUOTE]Lucky you. I tried restarting 5 times each and get nothing but K80. :sad:
|
[QUOTE=James Heinrich;529816]I lost my connection to the P100 and I've got a K80 on both accounts now. :sad:
What I noticed is that the K80 is a dual-GPU model and mfaktc is of course using only one GPU, so the throughput is half what is shown on my [URL="https://www.mersenne.ca/mfaktc.php?filter=K80"]mfaktc table[/URL], which makes sense.[/QUOTE] Perhaps what's going on there is that you have two accounts running on a single Public IP address. [I]Colab[/I] probably sees this as a double-instance. Therefore, K80. I switched browsers on my HP earlier today so both my desktops would be using Firefox. They keep each other synced. I also got a K80 on the HP. I ended up deleting my instance on [I]Colab[/I] and recreated it with the same code. P100 the first try. [U] I only run one computer with it[/U]. So, you would probably have to drop one as well to get a P100 again. |
| All times are UTC. The time now is 23:10. |
Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.