mersenneforum.org  

Go Back   mersenneforum.org > Great Internet Mersenne Prime Search > Hardware > Cloud Computing

Reply
 
Thread Tools
Old 2019-09-08, 17:14   #34
xx005fs
 
"Eric"
Jan 2018
USA

21210 Posts
Default Tesla T4 or K80 for PRP

It turns out the K80 is significantly faster than the T4 (yesterday didn't manage to get an actual T4 instance to benchmark the speed difference). Generally I can get around 4.5ms/it with the K80 while the T4 is about 6ms/it after it starts to throttle down. Here are the results

K80:
Code:
2019-09-08 16:32:02 90396263 OK 27360000 30.27%; 4559 us/sq; ETA 3d 07:49; 3ec593b85e44fb66 (check 1.12s)
2019-09-08 16:35:06 90396263 OK 27400000 30.31%; 4561 us/sq; ETA 3d 07:49; ddc1e8d47986dac7 (check 1.11s)
2019-09-08 16:38:09 90396263 OK 27440000 30.36%; 4561 us/sq; ETA 3d 07:45; d7ad382a0b7037d3 (check 1.11s)
2019-09-08 16:41:13 90396263 OK 27480000 30.40%; 4560 us/sq; ETA 3d 07:42; 30103cf3945858fc (check 1.10s)
2019-09-08 16:44:17 90396263 OK 27520000 30.44%; 4561 us/sq; ETA 3d 07:39; 496945fc83650272 (check 1.11s)
T4:
Code:
2019-09-08 17:04:58 90396473 OK 27625600 30.56%; 5578 us/sq; ETA 4d 01:15; eb92599fdef067db (check 1.29s)
2019-09-08 17:06:24 90396473 OK 27640000 30.58%; 5884 us/sq; ETA 4d 06:35; ec8ff248229a7a0d (check 1.41s)
2019-09-08 17:10:27 90396473 OK 27680000 30.62%; 6038 us/sq; ETA 4d 09:11; 62cd89f65e138ebc (check 1.38s)
2019-09-08 17:14:30 90396473 OK 27720000 30.66%; 6038 us/sq; ETA 4d 09:08; f9c1164a0935332f (check 1.38s)
xx005fs is offline   Reply With Quote
Old 2019-09-08, 17:33   #35
hansl
 
hansl's Avatar
 
Apr 2019

5·41 Posts
Default

Quote:
Originally Posted by xx005fs View Post
Similarly, I am using... Colab to run gpuowl
Just curious why gpuowl and not CUDALucas? I haven't tried either yet so I hardly know anything about them, but I guess I would assume that CUDA would generally beat OpenCL on nvidia GPUs?
hansl is offline   Reply With Quote
Old 2019-09-08, 18:47   #36
xx005fs
 
"Eric"
Jan 2018
USA

22×53 Posts
Default

Quote:
Originally Posted by hansl View Post
Just curious why gpuowl and not CUDALucas? I haven't tried either yet so I hardly know anything about them, but I guess I would assume that CUDA would generally beat OpenCL on nvidia GPUs?
I personally think the fact that PRP has a reliable error check algorithm makes it superior to LL. Secondly, gpuowl seems to run faster on Nvidia cards that are bandwidth starved (for example my Titan V is significantly faster on gpuowl than CUDALucas, down from 1.12ms/it to 0.83ms/it with the switch. The Tesla K80 is no exception since it has a 1:3 FP32:FP64 rate but a relatively low memory bandwidth, the same should apply.), and obviously I would want maximum throughput. I would try CUDALucas sometimes if it's faster than gpuowl on the K80.
xx005fs is offline   Reply With Quote
Old 2019-09-08, 22:49   #37
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

224048 Posts
Default

Quote:
Originally Posted by Corbeau View Post
Be careful using more than one Google account. Apparently people on the LCZero project were banned from using CoLab because they did that.
I would really like to get a handle on what Google considers fair-use for this kind of thing. I'll reach out to them to try to get some clarity.

But... Doing some research on this I came across another Google offering: Kaggle. I registered (using my Google account), and after authenticating by way of a SMS message I was given a very similar Notebook environment.

Pasted in my beta Notebook and clicked run, and it happily gave me a Tesla P100-PCIE-16GB producing ~1,200 GHzD/D of TF'ing -- no changes to my code. Clicked on "Commit and Run" and I was allowed to launch two more batch runs, each with another P100!

The interactive session is limited to nine (9) hours, while the batch runs are limited to six (6). They also provide 5 GB of persistent storage per Notebook.

Like, wow!

Last fiddled with by chalsall on 2019-09-08 at 22:50 Reason: Smelling mistake.
chalsall is offline   Reply With Quote
Old 2019-09-08, 23:00   #38
xx005fs
 
"Eric"
Jan 2018
USA

22×53 Posts
Default

Quote:
Originally Posted by chalsall View Post
I would really like to get a handle on what Google considers fair-use for this kind of thing. I'll reach out to them to try to get some clarity.

But... Doing some research on this I came across another Google offering: Kaggle. I registered (using my Google account), and after authenticating by way of a SMS message I was given a very similar Notebook environment.

Pasted in my beta Notebook and clicked run, and it happily gave me a Tesla P100-PCIE-16GB producing ~1,200 GHzD/D of TF'ing -- no changes to my code. Clicked on "Commit and Run" and I was allowed to launch two more batch runs, each with another P100!

The interactive session is limited to nine (9) hours, while the batch runs are limited to six (6). They also provide 5 GB of persistent storage per Notebook.

Like, wow!
That's amazing, the P100 is an amazing GPU for PRP/LL. I would rather run those workloads instead of TF since that should be left to the Turing GPUs. Any tips on how to set it up since I can't seem to run gpuowl from uploading files.

Last fiddled with by xx005fs on 2019-09-08 at 23:23
xx005fs is offline   Reply With Quote
Old 2019-09-08, 23:03   #39
Dylan14
 
Dylan14's Avatar
 
"Dylan"
Mar 2017

2×281 Posts
Default

Quote:
Originally Posted by Dylan14 View Post
A true acid test would be to get BOINC set up on the notebook. It would prove handy for quite a lot of projects (thinking NFS@Home with the Xeon, or the PPS sieve on the GPU for PrimeGrid).

Well, butter my biscuit, I have created code for this. This took a bit of tinkering, but here it is:


Code:
import os.path
#Use apt-get to get boinc
!apt-get install boinc boinc-client
#cp boinc, boinccmd to working directory
!cp /usr/bin/boinc /content
!cp /usr/bin/boinccmd /content
#create a slots directory if it doesn't exist(otherwise boinc doesn't work)
if not os.path.exists('/content/slots'):
  !mkdir slots
#launch the client
#attach to projects as desired (here I used NFS@home)
if not os.path.exists('/content/slots/0'):
  !boinc --attach_project https://escatter11.fullerton.edu/nfs/ (your account key here)
else:
  !boinc
replace (your account key here) with your actual account key. Note you'll have to register for the project first. Also, the processor will default to the blank location: make sure you have the preferences set correctly!
Dylan14 is offline   Reply With Quote
Old 2019-09-09, 21:03   #40
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

22·23·103 Posts
Default

Quote:
Originally Posted by Dylan14 View Post
Well, butter my biscuit, I have created code for this. This took a bit of tinkering, but here it is:
Absolutely mind-blowingly awesome!!!

I have reached out to both Colaboratory and Kaggle, pointing to this thread and asking them if what we're doing here is considered OK by them, and if they have any comments or feedback.

I consider this to be much like running compute on an employer's or client's kit -- best to get explicit permission.
chalsall is offline   Reply With Quote
Old 2019-09-10, 06:01   #41
LaurV
Romulan Interpreter
 
LaurV's Avatar
 
Jun 2011
Thailand

927210 Posts
Default

grrr... now because of you, they will nerf it...
LaurV is offline   Reply With Quote
Old 2019-09-10, 06:53   #42
GP2
 
GP2's Avatar
 
Sep 2003

1010000101112 Posts
Default

FAQ says:

Quote:
What is Colaboratory?
Colaboratory is a research tool for machine learning education and research.

Why are hardware resources such as T4 GPUs not available to me?
The best available hardware is prioritized for users who use Colaboratory interactively rather than for long-running computations. Users who use Colaboratory for long-running computations may be temporarily restricted in the type of hardware made available to them, and/or the duration that the hardware can be used for. We encourage users with high computational needs to use Colaboratory’s UI with a local runtime.
Please note that using Colaboratory for cryptocurrency mining is disallowed entirely, and may result in being banned from using Colab altogether.
So it sounds like they somewhat tolerate compute-intensive usage even though the purpose of the thing is machine learning.
GP2 is offline   Reply With Quote
Old 2019-09-11, 21:42   #43
chalsall
If I May
 
chalsall's Avatar
 
"Chris Halsall"
Sep 2002
Barbados

100101000001002 Posts
Default

Quote:
Originally Posted by GP2 View Post
So it sounds like they somewhat tolerate compute-intensive usage even though the purpose of the thing is machine learning.
I haven't heard back from either of them, and I continue to be allowed to compute, so...

I could really use a couple of more beta testers. If you have a GPU72 account and a Google account, and are willing to help out, please PM me. It's pretty simple.

As always, you get all the credit for the work done on your behalf.
chalsall is offline   Reply With Quote
Old 2019-09-12, 21:13   #44
De Wandelaar
 
De Wandelaar's Avatar
 
"Yves"
Jul 2017
Belgium

3·17 Posts
Default

For the Kaggle user's, see https://www.kaggle.com/general/108481.
From now on, GPU usage is limited to 30 hours per week.
De Wandelaar is offline   Reply With Quote
Reply

Thread Tools


Similar Threads
Thread Thread Starter Forum Replies Last Post
Alternatives to Google Colab kriesel Cloud Computing 11 2020-01-14 18:45
Notebook enzocreti enzocreti 0 2019-02-15 08:20
Computer Diet causes Machine Check Exception -- need heuristics help Christenson Hardware 32 2011-12-25 08:17
Computer diet - Need help garo Hardware 41 2011-10-06 04:06
Workunit diet ? dsouza123 NFSNET Discussion 5 2004-02-27 00:42

All times are UTC. The time now is 07:45.

Sun Feb 28 07:45:15 UTC 2021 up 87 days, 3:56, 0 users, load averages: 1.50, 1.29, 1.31

Powered by vBulletin® Version 3.8.11
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.