![]() |
![]() |
#2861 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
24·461 Posts |
![]() Quote:
Or buy a used rack style server with a row of blowers included. And either way, deal with limited driver availability. (Few Linux distros, no Windows.) |
|
![]() |
![]() |
![]() |
#2862 | ||
Jun 2003
2·2,719 Posts |
![]() Quote:
Quote:
I am not sure how much cheaper these will be in used market, but yes, if you can get some of these cheap enough, it _might_ be worth it. "Might", because, by the time they become cheap enough, the latest gen consumer card might still offer better performance/$ |
||
![]() |
![]() |
![]() |
#2863 |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
1CD016 Posts |
![]()
No purchase cost, and no utility cost, but higher end user time cost, is a tradeoff that may be very palatable for some with limited funds or cooling capacity. Plus it offers a try-before-you-buy experience. I'm routinely running 7 Colab free tabs, getting usually a 3.5 hour GPU & CPU session & multiple CPU-only sessions per day on each tab, for a combined throughput of ~ >1. T4 24/7. It's not for everyone.
|
![]() |
![]() |
![]() |
#2864 |
Jul 2003
So Cal
22·3·7·31 Posts |
![]()
Likely it will. While the RX 7900 XTX is overall a much faster card (a MI60 is based on the same architecture as the Radeon VII) the 1/16 FP64/FP32 ratio on the RX 7900 XTX slows it down for gpuOwl.
|
![]() |
![]() |
![]() |
#2865 |
Mar 2022
Earth
7316 Posts |
![]() |
![]() |
![]() |
![]() |
#2866 |
"Yuki@karoushi"
Feb 2020
Japan, Chiba pref
2·3·5 Posts |
![]()
https://www.techpowerup.com/gpu-spec...7900-xtx.c3941
3.848 TFLOPS (1:16) i think PRP much faster than my RTX4090 |
![]() |
![]() |
![]() |
#2867 | |
Mar 2022
Earth
11510 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#2868 |
Sep 2017
USA
2×5×19 Posts |
![]()
Hi, is it currently possible to run P-1 with B1 only parallel with PRP on gpuOwl, and then copy the P-1 save file over to mprime/prime95 for the new and improved B2 w/ lots of RAM? Thank you.
|
![]() |
![]() |
![]() |
#2869 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
24×461 Posts |
![]() Quote:
There's a reference info post for that: Save file (in)compatibility https://www.mersenneforum.org/showpo...8&postcount=26 I looked at implementing CUDAPm1 -> MNEF ->prime95 some time ago, and found they used not merely different formats, but different data structures, such that what prime95 expected in a resume was not present from CUDAPm1. IIRC Mihai has contemplated implementing split P-1 (perhaps a future v8) Gpuowl S1 -> mprime/prime95 S2, but I've not seen any announcement of the capability yet. And see also the exponent limits reference info post. |
|
![]() |
![]() |
![]() |
#2870 | |
"Mihai Preda"
Apr 2015
5·172 Posts |
![]() Quote:
All this is still work-in-progress to a large degree. It's not very documented, and it's not polished, etc. But the basics are there. I do use it myself this way: - I run P-1 first-stage on the GPU (on Radeon VII), with B1=2M at the wavefront, - I continue with second-stage on the CPU with mprime, with either 256GB or 128GB of RAM, with a B2 between 300M and 1000M. The second stage takes around 1h - 1.5h. I plan to post more about how I run it. Last fiddled with by preda on 2022-12-09 at 18:21 |
|
![]() |
![]() |
![]() |
#2871 | |
"Mihai Preda"
Apr 2015
5×172 Posts |
![]()
So we need to feed mprime with tasks originating for gpuowl.
Gpuowl runs a P-1 first-stage on an exponent, using an assignment that looks like: Quote:
Once gpuowl is done with P-1 first-stage, it needs to forward the same assignment line to mprime. It does this by appending the assignment to a worktodo.add file in the folder that is set with the "-mprimeDir <dir>" argument to gpuowl. (Note: normally the arguments to gpuowl are not passed on the command line, but are instead read from the config.txt file that is found in the folder passed with -dir <folder> to gpuowl) Now there's a problem with frequent writes to mprime's worktodo.add while mprime is doing P-1 second-stage: mprime, realizing there's a worktodo.add, will interrupt the P-1 second-stage, integrate the worktodo.add, and continue the second-stage, and this interruption is rather expensive (on the order of 5-10 minutes). This problem will likely be fixed in a future release of mprime (by not interrupting the second-stage in the middle just for the sake of an worktodo.add). But in the meantime, the workaround that I use is this: 1. I accumulate the assignments in a separate worktodo.add (that is in a different folder, not in mprime's folder). Thus mprime is not interrupted. 2. Using a cron job, every 24h I move the big worktodo.add into mprime's folder. This will cause an interruption, but only once in 24h which is ok. (to be continued) Last fiddled with by preda on 2022-12-09 at 18:08 |
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
mfakto: an OpenCL program for Mersenne prefactoring | Bdot | GPU Computing | 1719 | 2023-01-16 15:51 |
GPUOWL AMD Windows OpenCL issues | xx005fs | GpuOwl | 0 | 2019-07-26 21:37 |
Testing an expression for primality | 1260 | Software | 17 | 2015-08-28 01:35 |
Testing Mersenne cofactors for primality? | CRGreathouse | Computer Science & Computational Number Theory | 18 | 2013-06-08 19:12 |
Primality-testing program with multiple types of moduli (PFGW-related) | Unregistered | Information & Answers | 4 | 2006-10-04 22:38 |