![]() |
![]() |
#1 |
Oct 2007
Manchester, UK
17×79 Posts |
![]()
Does anyone know of some utility that can handle P-1 on these monsters?
v29.8 of prime95 says it only accepts exponents up to 595,800,000, which I assume corresponds to its maximum FFT size. A few years ago LaurV made a post about potentially implementing P-1 in CUDA which sounds encouraging, but I don't know if he or anyone else got much further. https://www.mersenneforum.org/showpo...3&postcount=11 At least Prime95 can give optimal bounds for P-1. If I put in a candidate TF'd to 86 bits, it recommends B1=B2=44,680,000, no stage 2 due to RAM limitations I believe. This doesn't sound completely unreasonable, and offers a 3.53% chance of a factor, this is slightly higher than the chance of a factor from continuing TF up to 89 bits (3/89 ~ 3.37%). If I say the candidate has been TF'd to 91 bits instead (which seems to be vaguely where GPU TFing should probably stop), then Prime95 offers the bounds B1=B2=30,920,000 with a 2.07% chance of a factor. Seems a bit odd to me that the bounds are LOWER when TF has progressed more, but alright. |
![]() |
![]() |
![]() |
#2 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
2·11·229 Posts |
![]()
On an FMA3-capable system prime95 should be capable of going to 920M (50M fft since V29.2). I'm running 701M now on 29.7b1 x64.
https://www.mersenneforum.org/showpo...&postcount=218 CUDAPm1 has been around for years but doesn't reach that high, due to various issues, although it nominally supports sufficiently large fft lengths, on gpus with sufficient ram. https://www.mersenneforum.org/showthread.php?t=23389 Quote:
Last fiddled with by kriesel on 2019-06-04 at 06:26 |
|
![]() |
![]() |
![]() |
#3 |
Oct 2007
Manchester, UK
101001111112 Posts |
![]() |
![]() |
![]() |
![]() |
#4 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
2×11×229 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#5 |
Oct 2007
Manchester, UK
17·79 Posts |
![]()
Perhaps in time the limits will be raised such that stage 1 for these numbers will be possible. Though I completely understand why enabling such functionality is not exactly top priority.
For the memory usage, is there any possibility that it could be lowered if the second stage was broken down into multiple chunks, similar to stage 2 of ECM? |
![]() |
![]() |
![]() |
#6 | |
"TF79LL86GIMPS96gpu17"
Mar 2017
US midwest
2×11×229 Posts |
![]() Quote:
Of course, they are all going to be impacted by the roughly p2.2 run time scaling also. The 701M P-1 run on my i7-8750H (all cores one worker) took 32.3 days at NRP~25, while a recent 430M P-1 on a 3GB GTX1060 took ~5. days at NRP=5 (for both stages, no factor found). Those would scale to ~989. days and ~449. days respectively, per P-1 on a gigadigit candidate. Note that run time also lengthens when NRP goes toward 1 due to memory size limitations. (CUDAPm1 reference info https://www.mersenneforum.org/showthread.php?t=23389 re prime95 see https://www.mersenneforum.org/showthread.php?t=23900) Last fiddled with by kriesel on 2019-06-11 at 19:26 |
|
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
A couple of 15e candidates | fivemack | NFS@Home | 1 | 2014-11-30 07:52 |
How to calculate FFT lengths of candidates | pepi37 | Riesel Prime Search | 8 | 2014-04-17 20:51 |
No available candidates on server | japelprime | Prime Sierpinski Project | 2 | 2011-12-28 07:38 |
Adding New Candidates | wblipp | Operation Billion Digits | 6 | 2011-04-10 17:45 |
new candidates for M...46 and M48 | cochet | Miscellaneous Math | 4 | 2008-10-24 14:33 |