mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware > Cloud Computing

Reply
Thread Tools
Old 2019-12-31, 15:04   #749
kriesel
 
kriesel's Avatar
 
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest

5,437 Posts
Default

Quote:
Originally Posted by PhilF View Post
If that was in writing I would. However, their silence about it is deafening.
It is if you read between the lines. https://research.google.com/colaboratory/faq.html
Quote:
Why are hardware resources such as T4 GPUs not available to me?
The best available hardware is prioritized for users who use Colaboratory interactively rather than for long-running computations. Users who use Colaboratory for long-running computations may be temporarily restricted in the type of hardware made available to them, and/or the duration that the hardware can be used for. We encourage users with high computational needs to use Colaboratory’s UI with a local runtime.
Please note that using Colaboratory for cryptocurrency mining is disallowed entirely, and may result in being banned from using Colab altogether.
kriesel is offline   Reply With Quote
Old 2019-12-31, 15:12   #750
PhilF
 
PhilF's Avatar
 
Feb 2005
Colorado

2·7·47 Posts
Default

Quote:
Originally Posted by kriesel View Post
It is if you read between the lines.
I guess since we are such a fringe case we should be used to having to read between the lines when it comes to what is acceptable behavior and what isn't.

But Colab, unlike Kaggle, has never answered any of Chris' repeated emails about whether or not they wish we would just go away.
PhilF is offline   Reply With Quote
Old 2019-12-31, 16:12   #751
axn
 
axn's Avatar
 
Jun 2003

32×5×113 Posts
Default

Quote:
Originally Posted by kriesel View Post
It is if you read between the lines. https://research.google.com/colaboratory/faq.html
It deosn't even require much reading between the lines, if you ask me. That's damn near explicit.
axn is offline   Reply With Quote
Old 2019-12-31, 16:44   #752
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

2×67×73 Posts
Default

Quote:
Originally Posted by PhilF View Post
But Colab, unlike Kaggle, has never answered any of Chris' repeated emails about whether or not they wish we would just go away.
Correct. I've tried reaching out to them at least four different times, with not a single response -- not even a "Your email has been received". But then Google is (in)famous for being unreachable unless you're spending money with them (AdSense, GCE, etc).

Also, it appears there is definitely some kind of a "per account" (NOT per user) quota'ing going on. Yesterday I was able to get four instances for 10 hours each -- three appearing to be here in Barbados, and one (RPi) tunneled into the States. All P100s.

This morning none of the accounts are getting instances, but a Virtual Machine based browser (CentOS) tunneling into another IP in the States just got a back-end. I hadn't used this for the last two days.

Also, I see twelve (12) different people have successfully had access to a GPU backend over the last fourteen (14) hours; six (6) people are running right now.

Oh well... While the apparent randomness of the access can be a bit annoying, the good news is we haven't been banned. And, again, we can hardly complain about free compute.
chalsall is offline   Reply With Quote
Old 2019-12-31, 19:29   #753
kriesel
 
kriesel's Avatar
 
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest

124758 Posts
Default

That which is not prohibited is allowed.
kriesel is offline   Reply With Quote
Old 2020-01-01, 01:33   #754
Chuck
 
Chuck's Avatar
 
May 2011
Orange Park, FL

3×5×59 Posts
Default

For a couple of days I was getting bumped off of colab after 20 minutes. I normally alternate between a pair of signons which run for ten hours each. I waited a day, and now it looks like things are back to "normal". I just had one session run for ten hours and I have successfully started the other one.
Chuck is offline   Reply With Quote
Old 2020-01-01, 02:05   #755
PhilF
 
PhilF's Avatar
 
Feb 2005
Colorado

2·7·47 Posts
Default

This saga kind of reminds me of Whack-a-Mole.
PhilF is offline   Reply With Quote
Old 2020-01-01, 16:57   #756
Fan Ming
 
Oct 2019

5F16 Posts
Default

The *** buffer overflow detected *** problem of CUDALucas on colab using P100 is solved now.
When I first encountered this "buffer overflow detected" error on colab, I made a joke to my roommate that it must because the name of P100 is too long - Tesla P100-PCIE-16GB.
Today when I tried to compile CUDALucas on colab I noticed that the compiler warnings that some sprintf may cause buffer overflow, and I found that the size of many char arrays is 32, which cannot accommodate some information that contains device name if the name of device is too long...
I enlarged the size of some array, and yes, it worked on colab P100 without "buffer overflow" error now... The compiler just told me everything, alas...
Attached file contains the modified source file CUDALucas.cu, the binary CUDALucas.exe I compiled and the Makefile I used.
In detail, I changed sizes of all char arrays which may cause overflow by sprintf to 371 or 742. Please notice me if I left some unchanged or I made some bad changes.
I'm using the compiled binary to do a double check. About 13 hours for ~50M exponent. Seems a little slower than excepted to be.
Attached Files
File Type: zip CUDALucas-colab.zip (206.0 KB, 70 views)

Last fiddled with by Fan Ming on 2020-01-01 at 17:06
Fan Ming is offline   Reply With Quote
Old 2020-01-01, 19:37   #757
kriesel
 
kriesel's Avatar
 
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest

124758 Posts
Default

Quote:
Originally Posted by Fan Ming View Post
The *** buffer overflow detected *** problem of CUDALucas on colab using P100 is solved now.
When I first encountered this "buffer overflow detected" error on colab, I made a joke to my roommate that it must because the name of P100 is too long - Tesla P100-PCIE-16GB.
Today when I tried to compile CUDALucas on colab I noticed that the compiler warnings that some sprintf may cause buffer overflow, and I found that the size of many char arrays is 32, which cannot accommodate some information that contains device name if the name of device is too long
Interesting.
I count "Tesla P100-PCIE-16GB" as 20,
and "GeForce GTX 1080 Ti" as 19.
Code:
CUDALucas v2.06beta 64-bit build, compiled May  5 2017 @ 13:02:54

binary compiled for CUDA  8.0
CUDA runtime version      8.0
CUDA driver version       8.0

------- DEVICE 0 -------
name                GeForce GTX 1080 Ti
But, "GeForce GTX 1060 3GB" as 20 also, and no problem.
Code:
CUDALucas v2.06beta 64-bit build, compiled May  5 2017 @ 13:00:15

binary compiled for CUDA  6.50
CUDA runtime version      6.50
CUDA driver version       8.0

------- DEVICE 0 -------
name                GeForce GTX 1060 3GB
Those are from flashjh's compiles for Windows.

Last fiddled with by kriesel on 2020-01-01 at 19:39
kriesel is offline   Reply With Quote
Old 2020-01-02, 00:19   #758
ATH
Einyen
 
ATH's Avatar
 
Dec 2003
Denmark

2·1,579 Posts
Default

Quote:
Originally Posted by Fan Ming View Post
Attached file contains the modified source file CUDALucas.cu, the binary CUDALucas.exe I compiled and the Makefile I used.
In detail, I changed sizes of all char arrays which may cause overflow by sprintf to 371 or 742. Please notice me if I left some unchanged or I made some bad changes.
I'm using the compiled binary to do a double check. About 13 hours for ~50M exponent. Seems a little slower than excepted to be.
Thanks, I used your CUDALucas.cu and compiled it as well and it works.

But you should ONLY use CUDALucas on the P100 if you want to do double check LL tests, because gpuowl is faster for first time tests:
I'm getting 1.17ms/iteration on 5120K FFT on gpuowl and 1.44 ms/iteration on 5184K FFT on CUDALucas (and even more on 5120K).
ATH is offline   Reply With Quote
Old 2020-01-02, 08:00   #759
Fan Ming
 
Oct 2019

5×19 Posts
Default

Quote:
Originally Posted by ATH View Post
Thanks, I used your CUDALucas.cu and compiled it as well and it works.

But you should ONLY use CUDALucas on the P100 if you want to do double check LL tests, because gpuowl is faster for first time tests:
I'm getting 1.17ms/iteration on 5120K FFT on gpuowl and 1.44 ms/iteration on 5184K FFT on CUDALucas (and even more on 5120K).
Yes, gpuowl is significantly faster.
Fan Ming is offline   Reply With Quote
Reply

Thread Tools


Similar Threads
Thread Thread Starter Forum Replies Last Post
Alternatives to Google Colab kriesel Cloud Computing 11 2020-01-14 18:45
Notebook enzocreti enzocreti 0 2019-02-15 08:20
Computer Diet causes Machine Check Exception -- need heuristics help Christenson Hardware 32 2011-12-25 08:17
Computer diet - Need help garo Hardware 41 2011-10-06 04:06
Workunit diet ? dsouza123 NFSNET Discussion 5 2004-02-27 00:42

All times are UTC. The time now is 05:45.


Fri Aug 6 05:45:47 UTC 2021 up 14 days, 14 mins, 1 user, load averages: 3.20, 3.02, 2.90

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.