Power, Temperature, & Noise

As always, last but not least is our look at power, temperature, and noise. Next to price and performance of course, these are some of the most important aspects of a GPU, due in large part to the impact of noise. All things considered, a loud card is undesirable unless there’s a sufficiently good reason – or sufficiently good performance – to ignore the noise.

GeForce GTX 780 Series Voltages
GTX 780 Ti Boost Voltage GTX 780 Boost Voltage GTX 780 Ti Base Voltage
1.187v 1.1625v 1.012v

Taking a quick look at voltages, we find that our GTX 780 Ti operates at a slightly higher voltage at its maximum boost bin than the original GTX 780 did. The difference is minor, but the additional voltage may be necessary to hit the slightly higher clockspeeds GTX 780 Ti operates at relative to GTX Titan and GTX 780.

GeForce GTX 780 Ti Average Clockspeeds
Max Boost Clock 1020MHz
Metro: LL
1000MHz
CoH2
997MHz
Bioshock
954MHz
Battlefield 3
980MHz
Crysis 3
980MHz
Crysis: Warhead
1000MHz
TW: Rome 2
950MHz
Hitman
993MHz
GRID 2
967MHz
Furmark
823MHz

Moving on to clockspeeds, we find that the GTX 780 Ti does very well when it comes to boosting. With a maximum boost clock of 1020MHz, we have 2 benchmarks averaging 1000MHz, and another 4 averaging 980MHz or better.

With all of our GK110 cards sharing a common design, at idle there’s very little to differentiate them. Other than GTX Titan’s extra 3GB of VRAM, we’re essentially looking at identical cards when idling.

Moving on to load power, we can see the power/heat ramifications of the slight clockspeed increase coupled with the activation of the 15th SMX. Even with the further optimizations NVIDIA has put into the new revision of GK110, power consumption has gone up in accordance with the higher performance of the card, just as we’d expect. Since NVIDIA doesn’t notably alter their power efficiency here, that increased performance has to come at the cost of increased power consumption. Though in this benchmark it’s worth pointing out that we’re measuring from the wall and that GTX 780 Ti outperforms GTX Titan by 8%, so some of that 29W power difference will come from the higher CPU load caused by the increased framerates.

As for the GTX 780 Ti SLI, here we see power level off at 556W, 20W more than the GTX 780 SLI. Some (if not most) of that is going to be explained by the increased CPU power consumption from the GTX 780 Ti SLI’s higher framerates. Coupled with that is the fact that in SLI setups these cards get hotter, and hence have to downclock a bit more to maintain equilibrium, which helps to offset the increased power requirements of GTX 780 Ti and keep the SLI results so close to the GTX 780 SLI results.

Switching over to FurMark, we find that power consumption is also up, but only slightly. With GPU Boost 2.0 clamping down on power consumption all of our GK110 cards should be clamped at 250W here, and with a difference between GTX 780 and GTX 780 Ti of under 10W, that’s exactly what appears to be happening here.

On a side note, it’s interesting to note here that under FurMark we’re seeing the GTX 780 Ti draw more power than the Radeon R9 290X. Despite the fact that the 290X has a higher rated TDP, in the card’s default quiet mode the card can’t actually dissipate as much heat (and thereby consume as much power) as the GTX 780 Ti can.

For idle temperatures we’re once again looking at cards that are for all intents and purposes identical. At 30C the GTX 780 Ti easily stays nice and cool.

As we mentioned in our look at the GTX 780 Ti hardware, NVIDIA has increased their default temperature throttle point from 80C on the GTX Titan/780 to 83C on the GTX 780 Ti. The end result is that in all of our temperature limited tests the GTX 780 Ti will peak at 83C-84C, whereas the older GK110 cards will peak at 80C-81C.

FurMark reiterates what we saw with Crysis 3. The temps are up a bit across the board, while the GK110 cards are holding near their throttle points. The SLI setups meanwhile approach the upper-80s at 88C, reflecting the fact that even with blowers, there’s some impact on neighboring cards in high load situations.

Our last idle scenario, we once again see all of our GK110 cards performing similarly, with idle noise levels in the 38dB-39dB range.

Moving on to our gaming load noise results, we can see the full repercussions of the GTX 780 Ti’s higher average power consumption coupled with the card’s higher temperature throttle point. Moving the throttle point along the same curve has the end result of moving higher the equilibrium point and thereby the card’s operating noise levels. As the fastest single-GPU card on this card, the GTX 780 Ti is still doing very well for itself and for a blower based design at 51.7dB, though at 1.5dB louder than GTX Titan and 4.2dB louder than GTX 780 the noise tradeoff for the card’s higher performance is very clear. Meanwhile the fact that it’s tied with the GTX 780 SLI comes with its own bit of irony.

Speaking of the GTX 780 SLI, we can see the noise impact of SLI configurations too. The GTX 780 Ti SLI levels out at 53.7dB, 2dB louder than our single-card configuration and 2dB louder than the GTX 780 SLI. At this point it’s just a bit louder than the 290X and quieter than a number of other 290 series setups.

Finally with load noise levels under FurMark we can see where our various cards will peak at for noise levels. The GTX 780 Ti creeps up to 52.3dB, essentially tying with the GTX 780 and GTX Titan. Otherwise it comes in just behind the 290X, and the start of the pack for our multi-GPU setups.

As for the GTX 780 Ti SLI, like our single-card comparison points its up slightly as compared to the GTX 780 SLI.

Overall, our look at power, temperatures, and noise has been a rather straightforward validation of our earlier suspicions. GTX 780 Ti’s higher performance leads to higher power consumption, and will all other factors being held equal – including the cooler – power, temps, and noise levels all rise a bit as compared to GTX Titan and GTX 780. There’s no such thing as a free lunch here, and while GPU Boost 2.0 will keep the maximum levels suitably in check, on average GTX 780 Ti is going to be a bit worse than the other GK110 cards due to those factors. Though even with the increased noise levels in particular, GTX 780 Ti is still able to outperform 290X on noise while also delivering better gaming performance, which makes this another tidy victory for NVIDIA.

Compute Overclocking
Comments Locked

302 Comments

View All Comments

  • fewafwwaefwa - Thursday, November 7, 2013 - link

    sterven.
  • looncraz - Thursday, November 7, 2013 - link

    When game producers author the games they will do it with a mind towards Mantle and exploiting the AMD GPU characteristics exposed by Mantle on PCs for their console games.

    When creating portable software you create as thin of an abstraction layer as possible, that layer will now be much closer to the metal with unoptimized DirectX alternatives being manually added. That could very well mean that AMD hardware will have a noticeable advantage on PCs and game producers will only need to do a little extra work to become compatible with other DX-10/11 compatible video cards on Windows/Linux - so nVidia will become something of a "don't forget about me!" rather than "let's build to a generic platform and pull in the nVidia GPU extensions..."
  • Basstrip - Friday, November 8, 2013 - link

    I think they've ALWAYS programmed directly to the core. I think it's safe to assume that the processes translate fairly well and that although they might not be the same, they are similar.

    It just seems so economic to streamline the whole process. Less of a headache than to constantly try optimize things for multiple platforms.

    AMD chips on consoles may not be able to support mantle on the hardware side but programming for consoles and for pc will definitely NOT be 2 completely different things.
  • elajt_1 - Friday, November 8, 2013 - link

    Something I read on Extremetech: Feedback we’ve gotten from other sources continues to suggest that Microsoft’s low-level API for the Xbox One is extremely similar to Mantle, and the difference between the two is basically semantic. This doesn’t square very well with Microsoft’s own statements; we’ll continue to investigate.
    http://www.extremetech.com/gaming/168671-xbox-one-...
  • klmccaughey - Monday, November 11, 2013 - link

    The difference is a couple of header files. Izzy Wizzy! And you have your API calling code in Xbox transferable to a PC, the header files compile the API's to Mantle API - but both API's are essentailly the same. It couldn't be easier.
  • polaco - Friday, November 8, 2013 - link

    The point of mantle I think is to provide an easy way to port from PC to console or Console to PC. So giving the possibility to allow an easier cross compilation.
  • L33T BEANS - Friday, November 8, 2013 - link

    Basing someones intelligence on a single statement is unwise.
  • Totally - Sunday, November 10, 2013 - link

    Reading these comments makes me wonder, if these people slinging mantle around like a buzzword actually know what it does, because going by the comments alone trying to pitting it against g-sync they clearly don't. Mantle is as relevant to gamers as Cuda is. Yes it does have a direct impact but the benefits aren't for the end user.
  • klmccaughey - Monday, November 11, 2013 - link

    You do not understand. The API on the consoles is basically "Mantle". Mantle copies verbatim the API calls for the consoles. They just call it the API on the console. Port the code across, change a few headers, and you have your Mantle calls ;)
  • MonkeyM - Sunday, November 10, 2013 - link

    They will sell DIY kits, you don't need a new monitor, as per the press conference.

Log in

Don't have an account? Sign up now