mersenneforum.org Certificate : For all indents and purposes
 Register FAQ Search Today's Posts Mark Forums Read

 2017-08-15, 17:34 #232 pakaran     Aug 2002 3·83 Posts Incidentally, I'm somewhat annoyed with the quantity of these numbers that have showed up, especially since it looks like someone's just submitting 306 dd PRPs in sequence, which are almost at the point of being less trouble to generate than they are to download.
2017-08-15, 20:01   #233
EdH

"Ed Hall"
Dec 2009

1101011110012 Posts

Quote:
 Originally Posted by pakaran Incidentally, I'm somewhat annoyed with the quantity of these numbers that have showed up, especially since it looks like someone's just submitting 306 dd PRPs in sequence, which are almost at the point of being less trouble to generate than they are to download.
I would wonder if the certificate needs the original index to be accepted by the db, though. I am tied up ATM, but am looking for a way to help without duplicating your effort. If I can find a way to d/l only a set well above the lower bounds where you are working, I will add some help. Unfortunately, the db's limit of 10k doesn't help in this. By the time I would grab 10k and try to work only the upper half, you would be finished with the lower 4096 and downloading the same ones I'm working. Incidentally, I'm just becoming able to assign several machines back to the certificate task once again.

OTOH, since I can have a more timely direct control over my machines, would you rather I work the 306 dds and you take the others?

For now, I'll search for a way to d/l well above the lower bounds and see what I come up with.

Anyone else, feel free to add to this...

2017-08-15, 20:09   #234
pakaran

Aug 2002

3·83 Posts

Quote:
 Originally Posted by EdH I would wonder if the certificate needs the original index to be accepted by the db, though. I am tied up ATM, but am looking for a way to help without duplicating your effort. If I can find a way to d/l only a set well above the lower bounds where you are working, I will add some help. Unfortunately, the db's limit of 10k doesn't help in this. By the time I would grab 10k and try to work only the upper half, you would be finished with the lower 4096 and downloading the same ones I'm working. Incidentally, I'm just becoming able to assign several machines back to the certificate task once again. OTOH, since I can have a more timely direct control over my machines, would you rather I work the 306 dds and you take the others? For now, I'll search for a way to d/l well above the lower bounds and see what I come up with. Anyone else, feel free to add to this...
If you would like to take the 306's, and I take everything over (through however high I go before more appear), that's fine. I suppose you could use FactorDB's multiple directories option to get (e.g.) 3k numbers to distribute to each of 3 machines, if they were of comparable speed.

As such, I'm uploading all the work I completed in the 306 dd numbers, and will start working upwards from 307.

2017-08-16, 01:24   #235
EdH

"Ed Hall"
Dec 2009

65718 Posts

Quote:
 Originally Posted by pakaran If you would like to take the 306's, and I take everything over (through however high I go before more appear), that's fine. I suppose you could use FactorDB's multiple directories option to get (e.g.) 3k numbers to distribute to each of 3 machines, if they were of comparable speed. As such, I'm uploading all the work I completed in the 306 dd numbers, and will start working upwards from 307.
Sounds like a workable idea. I have started ten machines, for now, with 1000 each. They aren't identical, but until bedtime I'll keep up with them. I do need to modify a couple scripts to make the assignments and collections easier. Then again, maybe I'll just collect a little less often than my other setup. I suppose a bright side might be how much of an increase in proven primes there will be from this. But, unfortunately, I see no real reason someone did this other than it being a "fun" experiment.

 2017-08-16, 18:42 #236 pakaran     Aug 2002 111110012 Posts Alright, I just uploaded my work to date. I went by increasing input file size (a decent proxy for number size, and easy to sort by in Primo), and got as far as several 2k digit numbers. I'm now re-downloading and will restart from 307, picking up a several dozen new numbers that have appeared. Should get back into the area where I was pushing hard earlier (2200-2500) in a day or less, eyeballing the list. I'm leaving for you the few numbers of 300-305 dd inclusive, currently six of them. I assume it'll be easy enough to do your next sweep in a way that catches them, but I'm happy to do them if you like.
2017-08-17, 03:53   #237
EdH

"Ed Hall"
Dec 2009

344910 Posts

Quote:
 Originally Posted by pakaran ... I'm leaving for you the few numbers of 300-305 dd inclusive, currently six of them. I assume it'll be easy enough to do your next sweep in a way that catches them, but I'm happy to do them if you like.
I have been limiting my runs to just 306, so you can grab anything else around them.

Several of my fastest machines have gone on strike for some reason. I just got one of the better ones back on line, but have three others that are misbehaving. Sticking with ten for now, I have at least swapped a much faster machine for my slowest of the ten. It's still going to take some time to burn through all these 306 entries. This is where the 10000 limit is too restrictive.

 2017-08-17, 16:37 #238 pakaran     Aug 2002 3×83 Posts I know what you mean, and because Primo is GUI, you can't just set up a master-worker architecture where one machine does all the fetching (and maybe stores already-assigned numbers in a hash table or something, to discard future duplicates). :(
 2017-08-17, 18:54 #239 Stargate38     "Daniel Jackson" May 2011 14285714285714285714 613 Posts Maybe the 10k limit should be raised to 100k. EdH would then be able to download all the 306-digit entries at once, and you (pakaran) would be able to run a higher set of 100k certs (maybe 307 digits) without having to worry about work collisions.
 2017-08-17, 19:26 #240 schickel     "Frank <^>" Dec 2004 CDP Janesville 83216 Posts One other possible workload split would be to do a full 10k download, than one person pull the "bottom" x% and the other the "top" x% based on filename (which is based on factordb id #); I used to do that when there wasn't much below the 3k dd limit: I would pull 1000 primes and do them in batches. that was I felt I was processing the primes that had been waiting the longest.
2017-08-17, 22:14   #241
EdH

"Ed Hall"
Dec 2009

3,449 Posts

Quote:
 Originally Posted by pakaran I know what you mean, and because Primo is GUI, you can't just set up a master-worker architecture where one machine does all the fetching (and maybe stores already-assigned numbers in a hash table or something, to discard future duplicates). :(
Actually, I have automated the primo runs using xdotool in a script on each machine. The sticking point is that I can't figure out the db call to d/l 10k, 306dd candidates with a 12 way split.

Quote:
 Originally Posted by Stargate38 Maybe the 10k limit should be raised to 100k. EdH would then be able to download all the 306-digit entries at once, and you (pakaran) would be able to run a higher set of 100k certs (maybe 307 digits) without having to worry about work collisions.
I think if you looked a little closer you would find about 470k, 306dd entries and no 307s. With me running 306 and Pakaran running elsewhere, the only collisions currently occurring are probably from the random pulls.

Quote:
 Originally Posted by schickel One other possible workload split would be to do a full 10k download, than one person pull the "bottom" x% and the other the "top" x% based on filename (which is based on factordb id #); I used to do that when there wasn't much below the 3k dd limit: I would pull 1000 primes and do them in batches. that was I felt I was processing the primes that had been waiting the longest.
The splitting feature of the db would be my choice for an easier split and that's what I use to spread the candidates between my current machines.

My machines are running 10k in less than an hour. I'm able to interface more frequently than Pakaran, so I can get more runs per day. But, I'm not available all the time so there are longer periods, including not being able to link with my machines while I was out today due to laptop application failure.

2017-08-17, 22:40   #242
EdH

"Ed Hall"
Dec 2009

3,449 Posts

Quote:
 Originally Posted by EdH ... The sticking point is that I can't figure out the db call to d/l 10k, 306dd candidates with a 12 way split...
I think I have figured this one out. I might be able to totally automate this...
Code:
http://www.factordb.com/primobatch.php?digits=306&files=10000&parts=12&start=Generate%20zip

 Similar Threads Thread Thread Starter Forum Replies Last Post wblipp FactorDB 1 2012-05-28 03:16 IvanP FactorDB 3 2012-05-11 12:17 jasong Science & Technology 10 2007-01-19 19:04 Unregistered Information & Answers 13 2004-04-28 06:24

All times are UTC. The time now is 20:39.

Fri Nov 27 20:39:11 UTC 2020 up 78 days, 17:50, 3 users, load averages: 2.09, 1.62, 1.62

This forum has received and complied with 0 (zero) government requests for information.

Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later version published by the Free Software Foundation.
A copy of the license is included in the FAQ.