Pages:
Author

Topic: GK104: nVidia's Kepler to be the First Mining Card? (Read 8346 times)

Vbs
hero member
Activity: 504
Merit: 500
Newegg listed "Shader clock: 2012MHz".

Interesting... 1536 ALUs at 2012MHz would give it more processing power than a HD 6990. Either the GTX 680 is going to be the fastest GPU for mining, or it is a blatant mistake from Newegg or the microarchitecture has undisclosed limitations that would prevent exploiting all this apparent power.

Per the NDA lifted today, this was a blatant error. The ALUs will run at 1006-1058 MHz, which should allow this card to mine at an upper bound of 450-470 Mh/s (80-85% the speed of a HD 7970.) This is assuming of course thar Nvidia added a BFI_INT-like instruction to the architecture, which is not certain. If not, performance would be much lower...

WOW. This would be heaven if true.

Can you say 14 GPU Nvidia mining rig with a relatively cheap motherboard ?

The only thing that may kill it is the crap MH/$ value but prices will drop anyway ...

http://www.techpowerup.com/162935/ZOTAC-Working-On-GeForce-GTX-680-with-2-GHz-Core-Clock-Speed.html  Grin
hero member
Activity: 518
Merit: 500
Newegg listed "Shader clock: 2012MHz".

Interesting... 1536 ALUs at 2012MHz would give it more processing power than a HD 6990. Either the GTX 680 is going to be the fastest GPU for mining, or it is a blatant mistake from Newegg or the microarchitecture has undisclosed limitations that would prevent exploiting all this apparent power.

Per the NDA lifted today, this was a blatant error. The ALUs will run at 1006-1058 MHz, which should allow this card to mine at an upper bound of 450-470 Mh/s (80-85% the speed of a HD 7970.) This is assuming of course thar Nvidia added a BFI_INT-like instruction to the architecture, which is not certain. If not, performance would be much lower...

WOW. This would be heaven if true.

Can you say 14 GPU Nvidia mining rig with a relatively cheap motherboard ?

The only thing that may kill it is the crap MH/$ value but prices will drop anyway ...
mrb
legendary
Activity: 1512
Merit: 1028
Newegg listed "Shader clock: 2012MHz".

Interesting... 1536 ALUs at 2012MHz would give it more processing power than a HD 6990. Either the GTX 680 is going to be the fastest GPU for mining, or it is a blatant mistake from Newegg or the microarchitecture has undisclosed limitations that would prevent exploiting all this apparent power.

Per the NDA lifted today, this was a blatant error. The ALUs will run at 1006-1058 MHz, which should allow this card to mine at an upper bound of 450-470 Mh/s (80-85% the speed of a HD 7970.) This is assuming of course thar Nvidia added a BFI_INT-like instruction to the architecture, which is not certain. If not, performance would be much lower...

Edit: for more accurate perf estimations see https://bitcointalksearch.org/topic/gtx-680-perf-estimations-73627
Vbs
hero member
Activity: 504
Merit: 500
This is gonna be fun, since nVidia has got huge margins on this GPU. This was supposed to be the "performance/$" oriented card, like the GTX 560 Ti and GTX 570 were. I'm expecting heavy price drops on this one as soon as supply stabilizes.
donator
Activity: 1218
Merit: 1079
Gerald Davis
549 Euros for a card w/ retail price of $500?  Ouch.  Guess some people had to have it.

I hope AMD counters with a drop in 7970 prices.  If they do, I just might go with some 7970s instead of waiting for the 7990.

Agreed.  Lets go pricing war! Lets go!
hero member
Activity: 675
Merit: 514
549 Euros for a card w/ retail price of $500?  Ouch.  Guess some people had to have it.
499 Euro in Germany, including tax:
http://www.alternate.de/html/product/MSI/N680GTX-PM2D2GD5-OC/995698/?
Still not cheap
hero member
Activity: 896
Merit: 1000
Seal Cub Clubbing Club
549 Euros for a card w/ retail price of $500?  Ouch.  Guess some people had to have it.

I hope AMD counters with a drop in 7970 prices.  If they do, I just might go with some 7970s instead of waiting for the 7990.
donator
Activity: 1218
Merit: 1079
Gerald Davis
549 Euros for a card w/ retail price of $500?  Ouch.  Guess some people had to have it.
hero member
Activity: 840
Merit: 529
You can already buy them in Portugal: http://pcdiga.com/pcdiga/Produto.asp?Familia=&Artigo=9711 .
For anyone wondering the cards sold out in like half an hour!
donator
Activity: 1218
Merit: 1079
Gerald Davis
For my watercooled servers I am buying 1 more 5970 (and maybe buy a hot spare) and then I likely will only consider new 7970s or 7990s going forward.  The bad news is for watercooling only dual GPU make any sense so I will need to wait.

My pump and radiator has enough capacity for 2 more rigs of 5970s but honestly their age is starting to concern me.  Now if the price drops to sub $300 that concern will go out the window. Smiley
legendary
Activity: 2044
Merit: 1000
I don't understand the people buying brand new 7970's thinking the resell value will be better, it's going to lose a few hundred $ in value the first few months, while a 5970 still maintains the majority if not all of it's value if you bought one when the 7970 came out.  Anyway more for me I guess.

Very simplistic view.  Most 5970's are 2 years old, and many have been overclocked the entire time.  It is a gamble paying $400 for a card that could go bad at any moment, thus taking the re-sale value to nearly zero. 

Many people would rather take a garaunteed 150-200 dollar hit on a new 7970 that will very rarely crap out (and has a warranty even if it does), over a very sketchy 5970 that could cost you 100% of your investment. 

I have about a 50/50 split in my farm on 5970/7970.....so I have thought this through extensively. 
donator
Activity: 1218
Merit: 1079
Gerald Davis
Newegg listing is obviously a misquote.

http://www.tomshardware.com/reviews/geforce-gtx-680-review-benchmark,3161-4.html

Quote
First, we launch a single run of the Central Park level at 1920x1080 in DirectX 11 mode, without anti-aliasing. We get a 72.3 FPS result, and we observe GPU Boost pushing the GeForce GTX 680 between 1071 and 1124 MHz during the run (up from the 1006 MHz base).

The top chart shows that we’re bouncing around the upper end of GK104’s power ceiling. So, we increase the target board power by 15%. The result is a small jump to 74.2 FPS, along with clocks that vacillate between 1145 and 1197 MHz.

Figuring the power target boost likely freed up some thermal headroom, we then increase the offset by 100 MHz, which enables even better performance—76.1 FPS. This time, however, we get a constant 1215 MHz. Nvidia says this is basically as fast as the card will go given our workload and the power limit.

So why not up the target power again? At 130% (basically, the interface’s 225 W specification), performance actually drops to 75.6 FPS, and the graph over time shows a constant 1202 MHz. We expected more performance, not less. What gives? This is where folks are going to find a problem with GPU Boost. Because outcome is dependent on factors continually being monitored, performance does change over time. As a GPU heats up, current leakage increases. And as that happens, variables like frequency and voltage are brought down to counter a vicious cycle.

The effect is similar to heat soak in an engine. If you’re on a dynamometer doing back to back pulls, you expect to see a drop in horsepower if you don’t wait long enough between runs. Similarly, it’s easy to get consistently-high numbers after a few minute-long benchmarks. But if you’re gaming for hours, GPU Boost cannot be as effective.

Our attempt to push a 200 MHz offset demonstrates that, even though this technology tries to keep you at the highest frequency under a given power ceiling, increasing both limits still makes it easy to exceed the board’s potential and seize up.

Sliding back a bit to a 150 MHz offset gives us stability, but performance isn’t any better than the 100 MHz setting. No doubt, it’ll take more tinkering to find the right overclock with GPU Boost in the mix and always on.
hero member
Activity: 546
Merit: 500
Tom's hardware has its review posted.

http://www.tomshardware.com/reviews/geforce-gtx-680-review-benchmark,3161-2.html

Quote
Kepler’s shaders run at the processor’s frequency (1:1)
Vbs
hero member
Activity: 504
Merit: 500
Newegg listed "Shader clock: 2012MHz".

Interesting... 1536 ALUs at 2012MHz would give it more processing power than a HD 6990. Either the GTX 680 is going to be the fastest GPU for mining, or it is a blatant mistake from Newegg or the microarchitecture has undisclosed limitations that would prevent exploiting all this apparent power.

Shaders are default clocked at 1411MHz on the highest profile, although there seems to be a big overclock headroom. Need some serious reviews to come up! Tongue
mrb
legendary
Activity: 1512
Merit: 1028
Newegg listed "Shader clock: 2012MHz".

Interesting... 1536 ALUs at 2012MHz would give it more processing power than a HD 6990. Either the GTX 680 is going to be the fastest GPU for mining, or it is a blatant mistake from Newegg or the microarchitecture has undisclosed limitations that would prevent exploiting all this apparent power.
legendary
Activity: 1274
Merit: 1004
Looks like the NDA lifts tomorrow morning. Not sure which Canadian time zone.
http://www.hardwarecanucks.com/forum/video-cards/52615-12-more-hours.html
legendary
Activity: 1274
Merit: 1004
Well, after the Tom's leak it looks like Newegg had their cards listed early. It's been pulled now, but it was saved.



We'll know more once the NDA lifts (tonight?), but if gaming performance really is 10%+ better than Tahiti with lower power draw, I can't see how AMD could continue to sell many 7970s above $500.
legendary
Activity: 1484
Merit: 1005
So according to that article, it is horrible at FP ops, and the 7970 beats the pants off of it in that benchmark. Wonder if they compensated by giving better integer ops? One can only hope....

Not surprising, they nerfed the FPUs in GK104 to about 1/8 of those of the GTX570, with the intent of making it a fast gaming card without much in the way of FP compute.  GK100 will be the one with very high FP/DPFP compute ability.
sr. member
Activity: 256
Merit: 250
SHA1 is based on a Merkle-Damgard construction. It doesn't matter much whether you hash a lot of data once or a small amount of data multiple times. In both cases you are calculating multiple times the same compression function. Of course, there are some differences like hashing small amount of data can be more cache-friendly (with GPUs that could mean less __global reads). There are quite a lot of optimizations that can be done when input is fixed and small enough like what's the case with bitcoin. Anyway I believe that the ratios would be more or less the same provided that the graphs are correct.

However, this does not take into consideration whether the code is optimized better for a specific platform, the quality of the drivers and the OpenCL stack (which does not perform as well as CUDA on NVidia).
Pages:
Jump to: