![]() |
![]() |
#2146 |
Sep 2002
Database er0rr
23×449 Posts |
![]()
I have been running two instances at 5632K. Now I have a new set of assignments at 6144K. Running two different FFT sizes has two effects.: The smaller runs faster and the larger runs much slower. It is more efficient to run one instance. Will this imbalance be redressed when I have equal 6144K assignments?
|
![]() |
![]() |
![]() |
#2147 | |
P90 years forever!
Aug 2002
Yeehaw, FL
23·919 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#2148 | |
Sep 2002
Database er0rr
E0816 Posts |
![]() Quote:
Code:
./gpuowl ./gpuowl: /usr/lib/x86_64-linux-gnu/libstdc++.so.6: version `GLIBCXX_3.4.26' not found (required by ./gpuowl) |
|
![]() |
![]() |
![]() |
#2149 |
Sep 2002
Database er0rr
23×449 Posts |
![]()
I have two compilers. I think I installed gcc 9 manually for the source and 8 is native. Anyway for my purposes I hard wired g++-8 into the make file and all is hunky dory now.
Last fiddled with by paulunderwood on 2020-05-05 at 13:57 |
![]() |
![]() |
![]() |
#2150 |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
7·701 Posts |
![]()
For what it's worth, this matching LL DC with shift 0 was performed in gpuowl v6.11-264 on the RX480 in the same system as was previously having frequent GEC errors on rx550s.
https://www.mersenne.org/report_expo...5487503&full=1 |
![]() |
![]() |
![]() |
#2151 | |
∂2ω=0
Sep 2002
República de California
3×53×73 Posts |
![]() Quote:
1. Both @5632K; 2. One each @5632K and @6144K; 3. Both @6144K. Presumably you already know the timing for the first 2 ... you could temporarily move one of your queued-up 6144K assignments to top of the worktodo file for the current 5632K run to get both @6144K. If the slowdown for 2 compared to 1 and 3 really is as bad as you describe, I wonder if it's something to do with context-switching on the GPU between tasks that have different memory mappings: 2 jobs at same FFT length have different run data and e.g. DWT weights but have the same memory profile and GPU resources usage. Edit: I tried the above three 2-jobs scenarios on my own Radeon7, using expos ~107M to trigger the 6M FFT length. Here are the per-iteration timings: 1. Both @5632K: 1470 us/iter for each, total throughput 1360 iter/sec; 2. One each @5632K,@6144K: 1530,1546 us/iter resp., total throughput 1300 iter/sec; 3. Both @6144K: 1615 us/iter for each, total throughput 1238 iter/sec. So no anomalous slowdowns for me at any of these combos, and the per-iteration timings hew very closely to what one would expect based on an n*log(n) per-autosquaring scaling. Last fiddled with by ewmayer on 2020-05-05 at 21:33 |
|
![]() |
![]() |
![]() |
#2152 | |
Sep 2002
Database er0rr
23·449 Posts |
![]() Quote:
2. One each @5632K and @6144K ----> the latter was ~2300us/it (very slow); the former 1125us/it At the moment (with latest commit) it is running ~1200us/it (103.9M) and ~1800us/it (104.9M). They were running at the average earlier until I restarted them. It is my last 103.9M exponent. Last fiddled with by paulunderwood on 2020-05-05 at 22:13 |
|
![]() |
![]() |
![]() |
#2153 |
Sep 2002
Database er0rr
23·449 Posts |
![]()
Was ~1200us/it (103.9M) and ~1800us/it (104.9M).
Now 1440us/it each -- both at 104.9M. |
![]() |
![]() |
![]() |
#2154 |
"Eric"
Jan 2018
USA
D416 Posts |
![]()
It seems that for factored PM1 results out of GPUOWL, primenet won't be able to understand it.
Code:
{"status":"F", "exponent":"98141611", "worktype":"PM1", "B1":"750000", "B2":"15000000", "fft-length":"5767168", "factors":"["****"]", "program":{"name":"gpuowl", "version":"v6.11-258-gb92cdfd"}, "computer":"TITAN V-0", "aid":"******", "timestamp":"2020-05-06 07:29:29 UTC"} |
![]() |
![]() |
![]() |
#2155 |
Sep 2006
Brussels, Belgium
13·127 Posts |
![]() |
![]() |
![]() |
![]() |
#2156 | |
Sep 2002
Database er0rr
23×449 Posts |
![]() Quote:
nvm: I found this on how to do it Gnome without having to remember and use unicodes. Thanks for prompting me! Last fiddled with by paulunderwood on 2020-05-07 at 18:47 |
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
mfakto: an OpenCL program for Mersenne prefactoring | Bdot | GPU Computing | 1668 | 2020-12-22 15:38 |
GPUOWL AMD Windows OpenCL issues | xx005fs | GpuOwl | 0 | 2019-07-26 21:37 |
Testing an expression for primality | 1260 | Software | 17 | 2015-08-28 01:35 |
Testing Mersenne cofactors for primality? | CRGreathouse | Computer Science & Computational Number Theory | 18 | 2013-06-08 19:12 |
Primality-testing program with multiple types of moduli (PFGW-related) | Unregistered | Information & Answers | 4 | 2006-10-04 22:38 |