![]() |
![]() |
#188 |
"Mihai Preda"
Apr 2015
17·79 Posts |
![]()
I'm personally not using 6.x myself, thus I don't have a lot of motivation to improve it. From my POV, 7.x is now better in a couple of ways than 6.x, and I prefer to focus my (limited) resources on 7.x. OTOH in the open spirit I'm not against others having different oppinions, and using different versions in different ways etc.
|
![]() |
![]() |
![]() |
#189 |
Jul 2003
So Cal
2,069 Posts |
![]()
P-1 factors seem to always be reported as found by stage 1. This leads to confusing reports like https://www.mersenne.ca/exponent/104592713. Can this be fixed easily or is this behavior intentional?
|
![]() |
![]() |
![]() |
#191 | |
Random Account
"Norman D. Powell"
Aug 2009
Indiana, USA.
23·5·47 Posts |
![]() Quote:
![]() |
|
![]() |
![]() |
![]() |
#192 |
"Oliver"
Sep 2017
Porta Westfalica, DE
23·5·11 Posts |
![]()
Yes, that would exclude everybody who is doing P-1 for exponent with known residues.
Last fiddled with by kruoli on 2020-12-19 at 21:15 Reason: Conjunctive. |
![]() |
![]() |
![]() |
#193 |
Random Account
"Norman D. Powell"
Aug 2009
Indiana, USA.
188010 Posts |
![]()
I run first-time P-1's from Primenet by manual reservation. The only thing done to them prior is TF to a specific bit level, Currently, it is 2^76. I can also get these from GPUto72's website. A P-1 with a "known residue" would indicate to me that is has a factor, meaning it is composite and requires no further effort.
|
![]() |
![]() |
![]() |
#194 |
May 2020
2510 Posts |
![]()
Does anyone have any benchmarks running gpuowl on the new gtx 30 series or the new AMD 6800x series gpus?
|
![]() |
![]() |
![]() |
#195 |
1976 Toyota Corona years forever!
"Wayne"
Nov 2006
Saskatchewan, Canada
22·17·67 Posts |
![]() |
![]() |
![]() |
![]() |
#196 |
"Viliam FurÃk"
Jul 2018
Martin, Slovakia
19016 Posts |
![]() |
![]() |
![]() |
![]() |
#197 |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
22·32·137 Posts |
![]()
Among the v7.2 builds I've made and run on Win 10, -53 is fastest. But for straight PRP, v6.11-364 or -380 are faster still. There are some caveats to that result summary; tested on one system & one OS, with multiple fast gpus supported by a slow cpu. But some cases were substantial enough to more than erase the ostensible advantage of doing PRP & P-1 stage 1 simultaneously with many of the same squarings, compared to running separate V6.11 P-1 to normal bounds followed by V6.11 PRP. That's comparing time per iteration after initialization time, bringing gpu up to temperature for stable ~steady state iteration times, etc. Additional initialization in V7.2 puts it at an additional disadvantage (trig table setup for example).
In all V7.2 tests mentioned here, PRP iterations were comparable to V6.11 iterations; PRP was continuing in iterations after P-1 had already completed, or P-1 was suppressed by B1=0,B2=0; in the worktodo line or tests-saved=0 in the worktodo line for V7.2-x timing tests. Part one PRP iteration speed regression for gpuowl-win v7.2-x on Radeon VII, Win 10 Pro x64 Variation ~3-7% 623M 7312/7087 = 1.032 927M 12263/11458 = 1.070 Code:
gpuowl us/it us/it version 623M 927M V7.2-21 7119 11960 V7.2-69 7312 12257 V7.2-63 7311 12263 V7.2-53 x x system restart after these attempts v7.2-53 7088 11459 V7.2-39 fatal cl compile errors, so no timings possible V7.2-21 7106 11955 * good repeatability after restart and other versions v7.2-13 7141 12365 v7.2-53 7087 11458 * fastest and very good repeatability of timing, saves ~8 days on 927M fft size 36M 52M split 4k:9:512 4K:13:512 gpu # 2 3 This multigpu slow-cpu system sometimes exhibits loss of communication with a gpu, which can bog down other gpus also while system interrupts occupy 1 of the 2 cpu cores for ~1 hour or until a restart clears the condition. Note, gpus on this system are configured to run at reduced electrical power. Part two 100M, 300M, 900M, v6.11 & 7.2-53 PRP iteration speed regression for gpuowl-win v6.11-x on Radeon VII, Win 10 Pro x64 Variation 1.2 - 5.2% 100M 869/826 = 1.052, v6.11-380 fastest, > ~1.025 cost of usual-bounds (P-1+PRP) / PRP 300M 3360/3310 = 1.015, v6.11-364 fastest, > ~1.01 cost of usual-bounds (P-1 stage 1 + PRP) / PRP 900M 10895/10763 = 1.012, V6.11-364 -380 tie for fastest, ~1.01 cost of usual-bounds (P-1 stage 1 + PRP) / PRP Differences between v6.11-364 and v6.11-380 were +/- 1 count, not significant. Code:
gpuowl us/it us/it us/it version 300M 103M 900M 6.11-380 3311 826 10763 6.11-364 3310 827 10763 7.2-53 3360 869 10895 6.11-318 3327 836 10815 fft size 16M 5.5M 52M split 1k:8:1k 1K:11:256 4k:13:512 gpu # 4 4 4 Note, gpus on this system are configured to run at reduced electrical power. Last fiddled with by kriesel on 2021-03-07 at 20:10 |
![]() |
![]() |
![]() |
#198 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
22·32·137 Posts |
![]() Quote:
Code:
-iters <N> : run next PRP test for <N> iterations and exit. Multiple of 10000. It may be faster though to stick with v6.11-380 for P-1, if the system is reliable enough and the exponent P-1 quick enough not to need the better error detection in P-1 of V7.2. Last fiddled with by kriesel on 2021-03-07 at 20:32 |
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
GpuOwl PRP-Proof changes | preda | GpuOwl | 20 | 2020-10-17 06:51 |
gpuowl: runtime error | SELROC | GpuOwl | 59 | 2020-10-02 03:56 |
gpuOWL for Wagstaff | GP2 | GpuOwl | 22 | 2020-06-13 16:57 |
gpuowl tuning | M344587487 | GpuOwl | 14 | 2018-12-29 08:11 |
How to interface gpuOwl with PrimeNet | preda | PrimeNet | 2 | 2017-10-07 21:32 |