![]() |
![]() |
#2542 |
"Kieren"
Jul 2011
In My Own Galaxy!
1014810 Posts |
![]()
THe 460 is CC 2.1. I considered the possibility that, but perhaps did not look closely enough. I just now did some searching for specific floating point capability, but despite lengthy charts in various places, I did not get the differences sorted out.
|
![]() |
![]() |
![]() |
#2543 | |
Just call me Henry
"David"
Sep 2007
Cambridge (GMT/BST)
16A216 Posts |
![]() Quote:
https://en.wikipedia.org/wiki/Kepler_(microarchitecture) states that consumer Kepler cards have 1/24 double precision speed. https://en.wikipedia.org/wiki/Maxwel...roarchitecture) states that consumer Maxwell cards have 1/32 double precision speed. https://en.wikipedia.org/wiki/Pascal_(microarchitecture) basically states that consumer Pascal cards have 1/32 double precision speed. The 460 is Fermi and 1/8 and the 750Ti is Maxwell and 1/32. There are variations with some of the Titan cards and server cards which allow things like 1/2 or 1/3 depending on the generation. The TDP of the 460 is also around 2.5x that of the 750Ti. The 750Ti is only a 60 watt card which is low for a gpu. It is almost as efficient as most of the 900 series gpus(Maxwell gen 2) |
|
![]() |
![]() |
![]() |
#2544 |
"Kieren"
Jul 2011
In My Own Galaxy!
22×43×59 Posts |
![]()
Thanks! I looked at related sources, but obviously not carefully enough. Actually, I think I read the general Wiki on CUDA.
Last fiddled with by kladner on 2016-11-06 at 20:30 |
![]() |
![]() |
![]() |
#2545 | ||
"Kieren"
Jul 2011
In My Own Galaxy!
22×43×59 Posts |
![]() Quote:
Quote:
|
||
![]() |
![]() |
![]() |
#2546 |
Just call me Henry
"David"
Sep 2007
Cambridge (GMT/BST)
10110101000102 Posts |
![]() |
![]() |
![]() |
![]() |
#2547 |
"David"
Jul 2015
Ohio
20516 Posts |
![]()
I see this pulled out over and over again about GPUs. "More suited to TF work"
I fail to see how this it backed up by much in terms of actual productivity (exponents cleared - not GhzDays, TF GhzDays are a joke) for many modern cards. I believe the math is "For exponents currently factored < [card specific crossover point], it is better to use your card to TF those exponents. For the giant piles of exponents needing double or first time checks already factored beyond those cross-over points, the card is just as productive doing LL work." For most modern cards we are already at or nearly at those cross-over points with lots of room to spare, meaning it is equally productive for a person with such a GPU to choose LL from the already factored pile or TF work from the far future list of exponents needing more factoring. It is true that GPUs are far better than CPUs at TF, however we only need enough current generation GPUs doing TF to keep ahead of the CPU/other GPU LL demand pool. What is especially interesting on the AMD side, and somewhat on the NVIDIA side, is that the older generations of cards did 200-300 GhzDay/Day of TF, but are close to the same as modern cards at LL work due to 1/2 or 1/3 DP units (40-50GhzDay/day). These cards are better suited to DC or LL work than TF compared to a modern card, as in they have a lower cross over point, however even modern cards have crossover points that we often have large reserves of exponents that are factored to that level. My counter-argument to continuing to factor far ahead with current or aging GPUs is that it isn't particularly power efficient. By the time we actually need those exponents factored we will likely be 1-2 generations of GPUs newer, or current high-end GPUs will be more accessible and migrated down to be widely in use. To henryzz's point, the direction technology is going does seem to be widening the gap between TF and LL and thus raising the cross-over point. That means newer GPUs will likely be just slightly better at LL but perhaps 2x as good at TF. Better to use the GPUs we have now balanced between filling the reserves and checking the exponents and save far-future TF for the next generation of cards that will be even more efficient at it. |
![]() |
![]() |
![]() |
#2548 | |
If I May
"Chris Halsall"
Sep 2002
Barbados
222728 Posts |
![]() Quote:
For a while (read: several years) the GPU TF'ing effort was playing catch-up. Oliver and Bdot (and NVidia and AMD) completely changed the game with their programs and the respective hardware. Heck, we didn't even initially know how deep the GPU TF'ing should go until James stepped in with his analysis of the TF'ing vs. LL'ing and DC'ing cross-over points. This might sound strange coming from the GPU72 guy, but just looking at the Primenet Exponent Status Distribution Map it is clear that the TF'ing is *well* ahead of the LL'ing, DC'ing and even the P-1'ing. GPU TF'ing will always be needed, but this is a resource management and optimization problem. As GIMPS' stated goal is to find Mersenne Primes (not factors), perhaps it is time for more GPUs be directed to LL'ing or DC'ing (or Carl's P-1 GPU program). But, as always, this is a volunteer effort. Everyone is encouraged to do whatever kind of work rocks their boat. At the end of the day all the work will be needed and useful. |
|
![]() |
![]() |
![]() |
#2549 |
"Kieren"
Jul 2011
In My Own Galaxy!
22·43·59 Posts |
![]()
Thanks to everyone for the discussion and explanations. The GTX 460 may not be very efficient, but it has proven to be rock steady at DC work. I am more leery of CuLu on the 580 because of the odd glitch which affects CC 2.0 cards. I know it's not supposed to spoil the result, but I had more than one proven bad mismatch on that card. That probably means that I did not go far enough in "detuning" it for stability.
![]() |
![]() |
![]() |
![]() |
#2550 | |
"Jerry"
Nov 2011
Vancouver, WA
1,123 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#2551 |
"Serge"
Mar 2008
Phi(4,2^7658614+1)/2
242916 Posts |
![]()
I could use one...
|
![]() |
![]() |
![]() |
#2552 |
"Jerry"
Nov 2011
Vancouver, WA
1,123 Posts |
![]() |
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Don't DC/LL them with CudaLucas | LaurV | Data | 131 | 2017-05-02 18:41 |
CUDALucas / cuFFT Performance on CUDA 7 / 7.5 / 8 | Brain | GPU Computing | 13 | 2016-02-19 15:53 |
CUDALucas: which binary to use? | Karl M Johnson | GPU Computing | 15 | 2015-10-13 04:44 |
settings for cudaLucas | fairsky | GPU Computing | 11 | 2013-11-03 02:08 |
Trying to run CUDALucas on Windows 8 CP | Rodrigo | GPU Computing | 12 | 2012-03-07 23:20 |