![]() |
|
|
#67 | ||||
|
"Teal Dulcet"
Jun 2018
2×17 Posts |
Quote:
Quote:
Quote:
Quote:
Thanks for the feedback! No problem, happy to clear up any confusion. |
||||
|
|
|
|
|
#68 | |
|
Romulan Interpreter
Jun 2011
Thailand
7·1,373 Posts |
Quote:
Already reported different types of work completed (including a GPU LLDC done in the 60M). The things are not so bad as I describe them, but if I paint them as minor things, you will never care. Now, if I paint them in black, I will make you angry, and you will try to prove me wrong... ![]() But we like the toys, otherwise we would just ignore them and not use them. We also learned a couple of things or two from them. So, |
|
|
|
|
|
|
#69 | |
|
If I May
"Chris Halsall"
Sep 2002
Barbados
100110000000112 Posts |
Quote:
I, like most, am very busy. But I have to read hundreds of pages of language (some human, some deterministic) every single day. One /possible/ motivation of you promoting your Free Extension which many of us have argued is again the "spirit" of the Colab Terms of Service is it /might/ assist in getting your Notebook to find the next MP by someone who is both using your Notebook and your extension. I could, of course, be entirely incorrect in that assessment. I'm simply posting based on my own position, and what I observe. Personally, I tend to error on the side of caution in situtations like this. Last fiddled with by chalsall on 2021-03-22 at 21:12 Reason: s/MB/MP/; # Typing too quickly... |
|
|
|
|
|
|
#70 |
|
Romulan Interpreter
Jun 2011
Thailand
258B16 Posts |
Hey Teal, Dan,
How do I pass a key from my keyboard to your colab cudaLucas copy? Beside of taking the pliers, pull the key out of the keyboard and throw it hard around the globe to reach google hq's. ![]() Why do I bother? Well... Here attached there is a digest of the FFT sizes, with times per iteration, for all the five cards that colab offers. The cards are so different, and the optimum FFT sizes for them are different too. If you start a LL test with some card, but later you got offered another card, you may lose up to 50% for the speed, because the FFT chosen by the first card is not the optimum value for the second one, and there is no (easy) way to change it. For example (see excel file inside the zip), your K80 just finished a test, and starts the next one, which by any chance is an exponent in, say, 112M, the K80 will start doing this with a FFT=6144, as that is the best choice for a K80 for this exponent size, with about 7.25 ms/iter (line 61 in the excel file). Then your time expires, and next time you are extremely lucky to get a P100, the P100 will continue the test with FFT=6144, which is a terrible unlucky choice of a size for it, getting about 2.1 ms/iter, when a larger FFT could be used: FFT=6272 with 1.7ms/iter. If you continue the test with the P100, then you got a huge penalty. This happens the other way around too. If you start a 65M test with P100, it will chose the size 3584, but after few minutes you are out, and next time you get a K80, you will continue with this size, at about 4.2 ms/iter, when a smaller FFT could be used for this card, for only 3.8 ms/iter. Another example, say you pay your money to Gugu, and get only good cards, and you decide to do a current 100M-digits assignment. Then you get a P100 which will chose the FFT=19683 (line 111 in the table), the smallest and fastest it can use for a 332M exponent, for which it spends about 5.8 ms/iter. Next time you get a V100, which will continue testing at this size, getting about 4.6 ms/iter for the next 20 days (line 286 in the "Threads" table in the excel file, second sheet), when you could use a larger FFT=20736 at 3.73 ms/iter and finish your job in 16 days instead of 20. On the reverse case, I can find a much worse example, but you got the idea. Now, cudaLucas is very clever, when it runs locally with the "-k" command line switch, we can use the keyboard to increase/decrease FFT size (and other parameters, like how often the screen output, how often to save checkpoints, etc), and we always can chose the best FFT on the fly! by pressing few keys (uppercase F, lowercase f, ). In fact, in the past, before gpuOwl era, I was using it intensively like that, always trying to push the FFT as low as possible, to get the fastest times, and back off when the rounding got in the dangerous area. Most of the tests can be run with lower/faster FFT, if you know what you are doing, the limits are "for safety", and to cover strange cases, but in real life, strange cases are few. So. Can you implement a similar feature, for example, I can write some text file directly in the drive's folders from which cudaLucas will read (as it can't read my keyboard) periodically, and adjust its parameters? Or offer a way to pass the text I type to it (yes, I can click in the window and type some commands in the square box that appears, but I 'ave no idea where those commands go, if that's actually possible, please enlighten me/us). Last fiddled with by LaurV on 2021-04-05 at 07:19 |
|
|
|
|
|
#71 |
|
P90 years forever!
Aug 2002
Yeehaw, FL
2·53·71 Posts |
Warning: Server will soon refuse to give first time LL tests. I haven't thought through all the details, most likely a double-check will be assigned.
The server will still accept first-time LL results. |
|
|
|
|
|
#72 | ||||
|
"Teal Dulcet"
Jun 2018
2×17 Posts |
Quote:
Quote:
Your examples provide another good reason to switch to GpuOwl. We did not initially notice this issue with CUDALucas, since when doing wavefront first time primality tests on Colab Pro, both the P100 and V100 GPUs happen to be optimal at the 6272K FFT length. Quote:
Quote:
I am assuming you are referring to already assigned first time LL tests or do you mean our PrimeNet script could rewrite new first time PRP assignments into LL tests and the server would still accept the results? We completely understand that this is not what you want users to do, but as I explained, unfortunately many Colab Pro users and people doing 100 million digit tests do not have much other choice. Our only other option would be to allow users to set the proof power as you suggested. However, that would obviously be very unfair to whoever has to do the proof certifications since these users would need to use proof powers of 5 or 6, which is why our notebooks currently do not support it. |
||||
|
|
|
|
|
#73 | |
|
P90 years forever!
Aug 2002
Yeehaw, FL
2×53×71 Posts |
Quote:
I'd prefer you do double-checks instead -- first time LL requests will get turned into LL double-check assignments. Proof power 5 or 6 is still an excellent option for the disk-constrained. A certification at 1/32 or 1/64th the cost of a first time test is still a huge savings. |
|
|
|
|
|
|
#74 | |||||
|
Romulan Interpreter
Jun 2011
Thailand
100101100010112 Posts |
Quote:
![]() Quote:
![]() Quote:
Quote:
![]() Quote:
And for PRP-CF CERTs? PRP-CF-DC CERTs? Last fiddled with by LaurV on 2021-04-06 at 04:04 |
|||||
|
|
|
|
|
#75 |
|
P90 years forever!
Aug 2002
Yeehaw, FL
2·53·71 Posts |
|
|
|
|
|
|
#76 | |||
|
"Teal Dulcet"
Jun 2018
2×17 Posts |
Quote:
Quote:
Quote:
No problem. I updated our GPU notebook with your requested change. As suggested by @Prime95, I also added an option to both notebooks so users can select the PRP proof power. Feedback is welcome. |
|||
|
|
|
|
|
#77 |
|
Romulan Interpreter
Jun 2011
Thailand
7·1,373 Posts |
Wow! it works!
You (two) are my heroes for this weekend! Albeit a little bit too complicate, first it didn't work, as I had the "CPU and GPU" output (sure! I want to see what BOTH of them are doing!), then I looked in the code and seen that you use the "-k" switch only when the output is "GPU Only", so, ok, stop the test, switch to "GPU Only" mode, restart the test, press the "f/F/t/T/etc" until "OCD satisfied", then let it run for 20 minutes to see that the output and speed is indeed what I want, stop, switch back to "CPU and GPU" output, restart the test. It works a marvell, as Dave would say! Now the tests will be in average ~10% to ~15% faster if I am clever enough to tune the FFT every time the GPU changes. I didn't want to modify the code, as I don't understood the implications, it may be an omission on your side, or you may have a very good reason why the "-k" is active only for the "GPU Only" output, but I didn't have the time (and skill) to look deeper into it. It works. Full stop. Thanks. Last fiddled with by LaurV on 2021-04-11 at 17:09 Reason: link |
|
|
|
![]() |
| Thread Tools | |
Similar Threads
|
||||
| Thread | Thread Starter | Forum | Replies | Last Post |
| Google Diet Colab Notebook | Corbeau | Cloud Computing | 1157 | 2021-05-20 19:47 |
| Primality testing of numbers k*b^n+c | Viliam Furik | Math | 3 | 2020-08-18 01:51 |
| Alternatives to Google Colab | kriesel | Cloud Computing | 11 | 2020-01-14 18:45 |
| Google Notebooks -- Free GPUs!!! -- Deployment discussions... | chalsall | Cloud Computing | 3 | 2019-10-13 20:03 |
| a new primality testing method | jasong | Math | 1 | 2007-11-06 21:46 |