20221120, 16:08  #2850 
Mar 2022
Earth
5·23 Posts 
That Radeon Instinct M250XAldebaran XT is stupid fast compared to anything else on the market!
I hope the 7900 XTX offers performance that is at least somewhat close to it on GPUOWL. Last fiddled with by Magellan3s on 20221120 at 16:09 
20221120, 16:45  #2851  
Jun 2003
2·2,719 Posts 
Quote:
Assuming a 2.7x (SWAG) improvement over 6950, it would be about 193 microseconds, which would be significantly faster than A100. 

20221120, 18:18  #2852 
Jul 2003
So Cal
101000101010_{2} Posts 
Also the MI250X (and MI250) are really two GPUs packaged together, like the nVidia K80. AMD hides this a bit in their marketing. So expect only half of that performance in a single gpuOwl run. And AMD is a bit behind nVidia in getting data from GPU memory to the cores, so the actual performance is closer to the A100 than the raw FLOPS numbers would suggest for most workloads.

20221120, 19:10  #2853  
Jul 2009
Germany
11·61 Posts 
Quote:
And then there's the famous Bottleneck probably too for MI50 and MI60. Last fiddled with by moebius on 20221120 at 19:13 

20221120, 20:05  #2854 
Jul 2009
Germany
11×61 Posts 
This one is rather close to the Instinct MI210. I've entered my predictions for the 7900 graphics cards into the list too.
Last fiddled with by moebius on 20221120 at 20:16 
20221120, 20:15  #2855 
Jul 2003
So Cal
5052_{8} Posts 
A MI250 is really two MI210's in the same package. Since gpuOwl doesn't support multiple GPUs, a single run on a MI250 will run on one of them and give the same times as a MI210. You can just run two simultaneously, one on each GPU. Same for the MI250X.

20221120, 20:54  #2856  
Jul 2009
Germany
11·61 Posts 
Quote:
I conclude that mfakto will scale just as well as on a MI100, but gpuowl much better. But I'm only human, I can also be wrong. 

20221120, 21:17  #2857 
Jul 2003
So Cal
2×1,301 Posts 
But to the user they appear as two distinct GPUs with separate memory spaces. Data transfer between the two over Infinity Fabric is much slower than HBM2, and applications must be coded to support multiple GPUs to use both.
https://chipsandcheese.com/2022/09/1...architecture/ (5th paragraph) https://twitter.com/projectphysx/sta...623746?lang=en 
20221120, 21:54  #2858 
Jul 2009
Germany
11·61 Posts 
In any case it would be interesting to know if these two discrete GPU's are as fast in wavefront prptest as on the card with only one discrete GPU, then they would be at least twice as good for the project.

20221121, 01:39  #2859 
Jun 2003
2×2,719 Posts 
These beasts (Instincts, Teslas, Quadros, etc.) are only of theoretical interest to the project. They are f***ing expensive! You're better off building one or more mulitGPU PCs with that kind of money.

20221121, 02:20  #2860 
Jul 2003
So Cal
A2A_{16} Posts 
32GB MI60's are available for under $1000. But they are passively cooled so you'd have to deal with rigging a cooler for it.

Thread Tools  
Similar Threads  
Thread  Thread Starter  Forum  Replies  Last Post 
mfakto: an OpenCL program for Mersenne prefactoring  Bdot  GPU Computing  1719  20230116 15:51 
GPUOWL AMD Windows OpenCL issues  xx005fs  GpuOwl  0  20190726 21:37 
Testing an expression for primality  1260  Software  17  20150828 01:35 
Testing Mersenne cofactors for primality?  CRGreathouse  Computer Science & Computational Number Theory  18  20130608 19:12 
Primalitytesting program with multiple types of moduli (PFGWrelated)  Unregistered  Information & Answers  4  20061004 22:38 