Register FAQ Search Today's Posts Mark Forums Read

 2019-09-08, 17:14 #34 xx005fs   "Eric" Jan 2018 USA 22×53 Posts Tesla T4 or K80 for PRP It turns out the K80 is significantly faster than the T4 (yesterday didn't manage to get an actual T4 instance to benchmark the speed difference). Generally I can get around 4.5ms/it with the K80 while the T4 is about 6ms/it after it starts to throttle down. Here are the results K80: Code: 2019-09-08 16:32:02 90396263 OK 27360000 30.27%; 4559 us/sq; ETA 3d 07:49; 3ec593b85e44fb66 (check 1.12s) 2019-09-08 16:35:06 90396263 OK 27400000 30.31%; 4561 us/sq; ETA 3d 07:49; ddc1e8d47986dac7 (check 1.11s) 2019-09-08 16:38:09 90396263 OK 27440000 30.36%; 4561 us/sq; ETA 3d 07:45; d7ad382a0b7037d3 (check 1.11s) 2019-09-08 16:41:13 90396263 OK 27480000 30.40%; 4560 us/sq; ETA 3d 07:42; 30103cf3945858fc (check 1.10s) 2019-09-08 16:44:17 90396263 OK 27520000 30.44%; 4561 us/sq; ETA 3d 07:39; 496945fc83650272 (check 1.11s) T4: Code: 2019-09-08 17:04:58 90396473 OK 27625600 30.56%; 5578 us/sq; ETA 4d 01:15; eb92599fdef067db (check 1.29s) 2019-09-08 17:06:24 90396473 OK 27640000 30.58%; 5884 us/sq; ETA 4d 06:35; ec8ff248229a7a0d (check 1.41s) 2019-09-08 17:10:27 90396473 OK 27680000 30.62%; 6038 us/sq; ETA 4d 09:11; 62cd89f65e138ebc (check 1.38s) 2019-09-08 17:14:30 90396473 OK 27720000 30.66%; 6038 us/sq; ETA 4d 09:08; f9c1164a0935332f (check 1.38s)
2019-09-08, 17:33   #35
hansl

Apr 2019

5·41 Posts

Quote:
 Originally Posted by xx005fs Similarly, I am using... Colab to run gpuowl
Just curious why gpuowl and not CUDALucas? I haven't tried either yet so I hardly know anything about them, but I guess I would assume that CUDA would generally beat OpenCL on nvidia GPUs?

2019-09-08, 18:47   #36
xx005fs

"Eric"
Jan 2018
USA

22·53 Posts

Quote:
 Originally Posted by hansl Just curious why gpuowl and not CUDALucas? I haven't tried either yet so I hardly know anything about them, but I guess I would assume that CUDA would generally beat OpenCL on nvidia GPUs?
I personally think the fact that PRP has a reliable error check algorithm makes it superior to LL. Secondly, gpuowl seems to run faster on Nvidia cards that are bandwidth starved (for example my Titan V is significantly faster on gpuowl than CUDALucas, down from 1.12ms/it to 0.83ms/it with the switch. The Tesla K80 is no exception since it has a 1:3 FP32:FP64 rate but a relatively low memory bandwidth, the same should apply.), and obviously I would want maximum throughput. I would try CUDALucas sometimes if it's faster than gpuowl on the K80.

2019-09-08, 22:49   #37
chalsall
If I May

"Chris Halsall"
Sep 2002

22×32×5×53 Posts

Quote:
 Originally Posted by Corbeau Be careful using more than one Google account. Apparently people on the LCZero project were banned from using CoLab because they did that.
I would really like to get a handle on what Google considers fair-use for this kind of thing. I'll reach out to them to try to get some clarity.

But... Doing some research on this I came across another Google offering: Kaggle. I registered (using my Google account), and after authenticating by way of a SMS message I was given a very similar Notebook environment.

Pasted in my beta Notebook and clicked run, and it happily gave me a Tesla P100-PCIE-16GB producing ~1,200 GHzD/D of TF'ing -- no changes to my code. Clicked on "Commit and Run" and I was allowed to launch two more batch runs, each with another P100!

The interactive session is limited to nine (9) hours, while the batch runs are limited to six (6). They also provide 5 GB of persistent storage per Notebook.

Like, wow!

Last fiddled with by chalsall on 2019-09-08 at 22:50 Reason: Smelling mistake.

2019-09-08, 23:00   #38
xx005fs

"Eric"
Jan 2018
USA

22×53 Posts

Quote:
 Originally Posted by chalsall I would really like to get a handle on what Google considers fair-use for this kind of thing. I'll reach out to them to try to get some clarity. But... Doing some research on this I came across another Google offering: Kaggle. I registered (using my Google account), and after authenticating by way of a SMS message I was given a very similar Notebook environment. Pasted in my beta Notebook and clicked run, and it happily gave me a Tesla P100-PCIE-16GB producing ~1,200 GHzD/D of TF'ing -- no changes to my code. Clicked on "Commit and Run" and I was allowed to launch two more batch runs, each with another P100! The interactive session is limited to nine (9) hours, while the batch runs are limited to six (6). They also provide 5 GB of persistent storage per Notebook. Like, wow!
That's amazing, the P100 is an amazing GPU for PRP/LL. I would rather run those workloads instead of TF since that should be left to the Turing GPUs. Any tips on how to set it up since I can't seem to run gpuowl from uploading files.

Last fiddled with by xx005fs on 2019-09-08 at 23:23

2019-09-08, 23:03   #39
Dylan14

"Dylan"
Mar 2017

2·7·41 Posts

Quote:
 Originally Posted by Dylan14 A true acid test would be to get BOINC set up on the notebook. It would prove handy for quite a lot of projects (thinking NFS@Home with the Xeon, or the PPS sieve on the GPU for PrimeGrid).

Well, butter my biscuit, I have created code for this. This took a bit of tinkering, but here it is:

Code:
import os.path
#Use apt-get to get boinc
!apt-get install boinc boinc-client
#cp boinc, boinccmd to working directory
!cp /usr/bin/boinc /content
!cp /usr/bin/boinccmd /content
#create a slots directory if it doesn't exist(otherwise boinc doesn't work)
if not os.path.exists('/content/slots'):
!mkdir slots
#launch the client
#attach to projects as desired (here I used NFS@home)
if not os.path.exists('/content/slots/0'):
!boinc --attach_project https://escatter11.fullerton.edu/nfs/ (your account key here)
else:
!boinc
replace (your account key here) with your actual account key. Note you'll have to register for the project first. Also, the processor will default to the blank location: make sure you have the preferences set correctly!

2019-09-09, 21:03   #40
chalsall
If I May

"Chris Halsall"
Sep 2002

22×32×5×53 Posts

Quote:
 Originally Posted by Dylan14 Well, butter my biscuit, I have created code for this. This took a bit of tinkering, but here it is:
Absolutely mind-blowingly awesome!!!

I have reached out to both Colaboratory and Kaggle, pointing to this thread and asking them if what we're doing here is considered OK by them, and if they have any comments or feedback.

I consider this to be much like running compute on an employer's or client's kit -- best to get explicit permission.

 2019-09-10, 06:01 #41 LaurV Romulan Interpreter     Jun 2011 Thailand 100100101010012 Posts grrr... now because of you, they will nerf it...
2019-09-10, 06:53   #42
GP2

Sep 2003

A1C16 Posts

FAQ says:

Quote:
 What is Colaboratory? Colaboratory is a research tool for machine learning education and research. Why are hardware resources such as T4 GPUs not available to me? The best available hardware is prioritized for users who use Colaboratory interactively rather than for long-running computations. Users who use Colaboratory for long-running computations may be temporarily restricted in the type of hardware made available to them, and/or the duration that the hardware can be used for. We encourage users with high computational needs to use Colaboratory’s UI with a local runtime. Please note that using Colaboratory for cryptocurrency mining is disallowed entirely, and may result in being banned from using Colab altogether.
So it sounds like they somewhat tolerate compute-intensive usage even though the purpose of the thing is machine learning.

2019-09-11, 21:42   #43
chalsall
If I May

"Chris Halsall"
Sep 2002

22×32×5×53 Posts

Quote:
 Originally Posted by GP2 So it sounds like they somewhat tolerate compute-intensive usage even though the purpose of the thing is machine learning.
I haven't heard back from either of them, and I continue to be allowed to compute, so...

I could really use a couple of more beta testers. If you have a GPU72 account and a Google account, and are willing to help out, please PM me. It's pretty simple.

As always, you get all the credit for the work done on your behalf.

 2019-09-12, 21:13 #44 De Wandelaar     "Yves" Jul 2017 Belgium 3·17 Posts For the Kaggle user's, see https://www.kaggle.com/general/108481. From now on, GPU usage is limited to 30 hours per week.

 Similar Threads Thread Thread Starter Forum Replies Last Post kriesel Cloud Computing 11 2020-01-14 18:45 enzocreti enzocreti 0 2019-02-15 08:20 Christenson Hardware 32 2011-12-25 08:17 garo Hardware 41 2011-10-06 04:06 dsouza123 NFSNET Discussion 5 2004-02-27 00:42

All times are UTC. The time now is 20:31.

Wed Apr 21 20:31:39 UTC 2021 up 13 days, 15:12, 0 users, load averages: 2.12, 1.77, 1.80