mersenneforum.org

mersenneforum.org (https://www.mersenneforum.org/index.php)
-   Cloud Computing (https://www.mersenneforum.org/forumdisplay.php?f=134)
-   -   Google Diet Colab Notebook (https://www.mersenneforum.org/showthread.php?t=24646)

Corbeau 2019-10-17 19:34

@kriesel Can I run your script while running the GPU72 trial factoring script on Colab? If so, do I need to manually add work in the worktodo file, or will it get assignments from the GPU72-reserved LL assignments?

kriesel 2019-10-17 19:44

Google drive etc mounting locally
 
One of the barriers, or rather a slowdown, I've run into initially in using Colab is that, handy as drag and drop of a file from a local host that I own to Google drive and vice versa is, compared to other methods, it still constitutes some overhead, and the file duplication seems likely to cause errors. What I'd really like is:
[LIST=1][*]Ability to mount the relevant Google drives on a local system here[*]Easy setup for that[*]Multiple OS support for the host system OS[*]Low host system overhead[*]Multiple cloud storage type support, to unify usage of other cloud options such as Box, OneDrive, etc[*]Adequate security that it does not compromise the local host system[*]Acceptable convenience and performance[*]Low cost (free is ideal)[*]No requirement to encrypt the cloud storage, which could interfere with the Colab or other cloud application being able to usefully process the data and return useful results.[*](What did I miss?)[/LIST]I think CloudMounter qualifies, after a quick initial skim. There are alternatives.
[URL]https://www.nextofwindows.com/how-to-mount-dropbox-onedrive-or-google-drive-as-local-disk-on-windows[/URL]
Windows [URL]https://cloudmounter.net/mount-cloud-drive-win.html[/URL]
Mac [URL]https://cloudmounter.net/[/URL]
linux [URL]https://cloudmounter.net/mount-cloud-drive-linux.html[/URL]

Some of the following may qualify, while some are quite a stretch, or are useful utilities for other purposes alongside mounting a cloud drive. [URL]https://www.topbestalternatives.com/cloudmounter/[/URL]

Favorite solutions? Thoughts? Experience?

(Please avoid OS-bashing. Different people use different OSes for different reasons, that make sense to them.)

PhilF 2019-10-17 19:45

[QUOTE=kriesel;528241]And the rapid progress of this collaborative effort has been something to see.[/QUOTE]

That's probably what Google and Kaggle are thinking too...

kriesel 2019-10-17 19:52

[QUOTE=Corbeau;528244]@kriesel Can I run your script while running the GPU72 trial factoring script on Colab? If so, do I need to manually add work in the worktodo file, or will it get assignments from the GPU72-reserved LL assignments?[/QUOTE]I don't use GPU72, so have no idea about compatibility with it. I'm also unsure which script you're referring to; the mprime resume one at [URL]https://www.mersenneforum.org/showpost.php?p=527910&postcount=3?[/URL] If so, it uses primenet, as set up by a variation of Dylan14's initial setup script, gets work, reports work, on its own, given enough 12-hour sessions to finish an assignment. My first Colab 87M primality test is weeks away from completion at twice-daily restarts; it takes persistence. My understanding is if using cpu and gpu application script sections in Colab in the same work section on the Colab web page, one of them needs to be run as a subprocess, then the other run as the script process. I'm about to try that as ath has described.
If you're using chalsall's reverse tunneling, you could probably instead use that to run mprime.
It appeals to me to have one Colab script to relaunch, that runs both the subprocess and the other to occupy the VM cpu and the VM gpu for the duration.
For gpu applications, the same or similar client management applies in the Colab environment as when running on our own gpus. A single pass through a Python script for result reporting and ensuring adequate work seems like a natural fit to me. There are other possibilities.
I regard the Colab thread scripts I've posted on my blog as a collaborative effort, not "mine", and try to give credit there for the originator.

kriesel 2019-10-17 19:54

[QUOTE=PhilF;528249]That's probably what Google and Kaggle are thinking too...[/QUOTE]Hence the disappearance of T4 availability? ("Give them only the older K80s, let's see what they come up with.")

LaurV 2019-10-18 04:29

Well, we should try to "come up with" running cudaLucas on it.:razz:
K80 is a [SIZE=6][COLOR=Red][B]waste[/B][/COLOR][/SIZE] if used for TF. This card is flying like a rocket at LL.

chalsall 2019-10-18 04:33

[QUOTE=kriesel;528241]And the rapid progress of this collaborative effort has been something to see.[/QUOTE]

The family's asleep. Back At Console for a little bit...

A book that I seriously internalized as a "youngin" was [URL="https://en.wikipedia.org/wiki/The_Mythical_Man-Month"]The Mythical Man-Month[/URL] by Brooks.

Empirically, I have found that small ad-hock teams can often do better work faster than larger groups with hierarchical communication channels can.

YMMV.

chalsall 2019-10-18 04:35

[QUOTE=LaurV;528271]Well, we should try to "come up with" running cudaLucas on it.:razz:[/QUOTE]

Don't disagree.

Why don't you code that up? :chalsall:

axn 2019-10-18 04:46

[QUOTE=chalsall;528273]Why don't you code that up? :chalsall:[/QUOTE]

BTDT. Got 3 DCs out of it. Major pain; Colab gets a conniption if you run it for long, and then you don't get GPU instance for a while.
Pretty fast, though. I estimated that if you run it full time, you could get about 60 GhzDay/day, which is pretty much in line with [url]https://www.mersenne.ca/cudalucas.php[/url]

EDIT:- Even more impressive is Kaggle and its P100. That is not a 2x GPU like K80, and so in theory you should get the full 160 GhzDay/day. But for that, you need a file hosting location to download the appropriate CUDALucas executables and libraries (cudart and cufft) -- google drive won't work.

bayanne 2019-10-18 08:46

[QUOTE=chalsall;528273]Don't disagree.

Why don't you code that up? :chalsall:[/QUOTE]

Chapeau from me if you can :)

LaurV 2019-10-18 10:00

Well, I am digging, but Linux was never my strong point, and here the pain is to store and retrieve the checkpoint files. But I am learning...


All times are UTC. The time now is 22:51.

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.