![]() |
![]() |
#1 |
Apr 2010
Over the rainbow
11×229 Posts |
![]()
http://www.techtimes.com/articles/13...sing-power.htm
Based on a number of slides from an independent researcher, the Nvidia Pascal GPU100 features Stacked DRAM (1 TB/s) giving it as much as 12 TFLOPs of Single-Precision (FP32) compute performance. The flagship GPU is purportedly able to provide four TFLOPs of Double-Precision (FP64) compute performance as well. Last fiddled with by wblipp on 2016-02-21 at 23:40 |
![]() |
![]() |
![]() |
#2 |
Romulan Interpreter
Jun 2011
Thailand
2×19×241 Posts |
![]()
Yarrrr !!!
![]() |
![]() |
![]() |
![]() |
#3 | |
Oct 2015
26610 Posts |
![]() Quote:
|
|
![]() |
![]() |
![]() |
#4 | |
Einyen
Dec 2003
Denmark
57078 Posts |
![]()
12 TFLOPS FP32 and 4 TFLOPS FP64 *drool* It will be hard to decide whether to run factoring or LL on it.
Quote:
|
|
![]() |
![]() |
![]() |
#5 | |
Romulan Interpreter
Jun 2011
Thailand
100011110001102 Posts |
![]() Quote:
Trial Factoring... (maybe.... you can use 8 of those to keep the precision of 80 to 88 bits, depending of content, but you will need about 72 multiplications to multiply those 8x8 "digits", with some karatsuba-like stuff (edit: can it be done in 72 multiplications?? don't forget that you multiply 11 bits times 11 bits and get 11 bits result, not 22 bits), so you will only get a third of a teraflop, like a gtx 560 or so. OTOH, I assume they will make a kill at integer arithmetic too, so FP16 will not be the best choice for TF either ...) Last fiddled with by LaurV on 2016-02-22 at 08:40 |
|
![]() |
![]() |
![]() |
#6 |
Jun 2003
37·131 Posts |
![]()
I am gonna go out on a limb and predict that there will be no consumer/prosumer version that offers 4 TFLOPS DP (i.e. not even a 1000$ Titan variety will offer 4 TFLOPS)
|
![]() |
![]() |
![]() |
#7 |
Feb 2016
UK
3×7×19 Posts |
![]()
Looking at previous high rate DP cards and their release dates:
R9 280X, ~1 TFLOP, Oct. 2013 Titan ~1.5 TFLOP, Feb. 2013 Titan Black 1.7 TFLOP, Feb. 2014 Could they manage to get it up to 4 in 2 years? I think there's more than a possibility they can, if they want to, in a higher end card. Especially now they're finally moving onto smaller manufacturing process again. |
![]() |
![]() |
![]() |
#8 |
Einyen
Dec 2003
Denmark
BC716 Posts |
![]() |
![]() |
![]() |
![]() |
#9 |
Jun 2003
113578 Posts |
![]() |
![]() |
![]() |
![]() |
#10 | |
"Marv"
May 2009
near the Tannhäuser Gate
2×293 Posts |
![]() Quote:
http://www.mersenneforum.org/showpos...&postcount=604 Do you want 3 Tflops FP64 from an Nvidia board? It's been available for a year on their K80 Tesla! The catch is it costs 5,000 dollars and requires intense, cooling found in servers located in frigid computer rooms. You can bet your boots Pascal chips with gobs of FP64 are destined for Teslas and not for the great unwashed masses ( us ). Even Nvidia's expected April announcement will probably be a "tease" in that regard. BTW, FP16 is there for Deep Learning Neural Nets, which is the hottest thing in AI right now. Researchers have done some truly amazing things with these such as driving cars and beating a GO master. Nvidia has very nice libraries for these. They require zillions of small FP values for all the weights used during training. They can tolerate loss of precision FP16 provides and are willing to trade that for having twice as many in memory as FP32 values. |
|
![]() |
![]() |
![]() |
#11 |
Feb 2016
UK
3×7×19 Posts |
![]()
Go on then, I'll take the optimistic route that they will put this in a consumer device, perhaps a future Titan something.
The fastest single-chip compute device they do is the K40, which appears to use the same chip as the Titan Black. They can still differentiate between the products in other ways. It's not like they're going to stand still in compute either, and they can't risk AMD not crippling their offering and looking bad. |
![]() |
![]() |
![]() |
Thread Tools | |
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
Pascal's OPN roadblock files | ThomRuley | Factoring | 898 | 2021-01-25 01:21 |
Passive Pascal | Xyzzy | GPU Computing | 1 | 2017-05-17 20:22 |
Tesla P100 — 5.4 DP TeraFLOPS — Pascal | Mark Rose | GPU Computing | 52 | 2016-07-02 12:11 |
Calculating perfect numbers in Pascal | Elhueno | Homework Help | 5 | 2008-06-12 16:37 |
No Notice- Binomial Coefficients, Pascal's triangle | Vijay | Miscellaneous Math | 5 | 2005-04-09 20:36 |