Meet The GeForce GTX 660

For virtual launches it’s often difficult for us to acquire reference clocked cards since NVIDIA doesn’t directly sample the press with reference cards, and today’s launch of the GeForce GTX 660 launch is one of those times. The problem stems from the fact that NVIDIA’s partners are hesitant to offer reference clocked cards to the press since they don’t want to lose to factory overclocked cards in benchmarks, which is an odd (but reasonable) concern.

For today’s launch we were able to get a reference clocked card, but in order to do so we had to agree not to show the card or name the partner who supplied the card. As it turns out this isn’t a big deal since the card we received is for all practical purposes identical to NVIDIA’s reference GTX 660, which NVIDIA has supplied pictures of. So let’s take a look at the “reference” GTX 660.

The reference GTX 660 is in many ways identical to the GTX 670, which comes as no great surprise given the similar size of their PCBs, which in turn allows NVIDIA to reuse the same cooler with little modification. Like the GTX 670, the reference GTX 660 is 9.5” long, with the PCB itself composing just 6.75” of that length while the blower and its housing composes the rest. The size of retail cards will vary between these two lengths as partners like EVGA will be implementing their own blowers similar to NVIDIA’s, while other partners like Zotac will be using open air coolers not much larger than the reference PCB itself.

Breaking open one of our factory overclocked GTX 660 (specifically, our EVGA 660 SC using the NV reference PCB), we can see that while the GTX 670 and GTX 660 are superficially similar on the outside, the PCB itself is quite different. The biggest change here is that while the 670 PCB made the unusual move of putting the VRM circuitry towards the front of the card, the GTX 660 PCB once more puts it on the far side. With the GTX 670 this was a design choice to get the GTX 670 PCB down to 6.75”, whereas with the GTX 660 it requires so little VRM circuitry in the first place that it’s no longer necessary to put that circuitry at the front of the card to find the necessary space.

Looking at the GK106 GPU itself, we can see that not only is the GPU smaller than GK104, but the entire GPU package itself has been reduced in size. Meanwhile, not that it has any functional difference, but GK106 is a bit more rectangular than GK104.

Moving on to the GTX 660’s RAM, we find something quite interesting. Up until now NVIDIA and their partners have regularly used Hynix 6GHz GDDR5 memory modules, with that specific RAM showing up on every GTX 680, GTX 670, and GTX 660 Ti we’ve tested. The GTX 660 meanwhile is the very first card we’ve seen that’s equipped with Samsung’s 6GHz GDDR5 memory modules, marking the first time we’ve seen non-Hynix memory on a GeForce GTX 600 card. Truth be told, though it has no technical implications we’ve seen so many Hynix equipped cards from both AMD and NVIDIA that it’s refreshing to see that there is in fact more than one GDDR5 supplier in the marketplace.

For the 2GB GTX 660, NVIDIA has outfit the card with 8 2Gb memory modules, 4 on the front and 4 on the rear. Oddly enough there aren’t any vacant RAM pads on the 2GB reference PCB, so it’s not entirely clear what partners are doing for their 3GB cards; presumably there’s a second reference PCB specifically built to house the 12 memory modules needed for 3GB cards.

Elsewhere we can find the GTX 660’s sole PCIe power socket on the rear of the card, responsible for supplying the other 75W the card needs. As for the front of the card, here we can find the card’s one SLI connector, which like previous generation mainstream video cards supports up to 2-way SLI.

Finally, looking at display connectivity we once more see the return of NVIDIA’s standard GTX 600 series display configuration. The reference GTX 660 is equipped with 1 DL-DVI-D port, 1 DL-DVI-I port, 1 full size HDMI 1.4 port, and 1 full size DisplayPort 1.2. Like GK104 and GK107, GK106 can drive up to 4 displays, meaning all 4 ports can be put into use simultaneously.

The GeForce GTX 660 Review Just What Is NVIDIA’s Competition & The Test
Comments Locked

147 Comments

View All Comments

  • chizow - Sunday, September 16, 2012 - link

    But SB had no impact on 980X pricing, Intel is very deliberate in their pricing and EOL schedules so these parts do not lose much value before they gracefully go EOL. Otherwise, 980X still offered benefit over SB with 6 cores, something that was not replaced until SB-E over 1 year later. Even then, there was plenty of indication before Intel launched their SB-E platform to mitigate any sense of buyer's remorse.

    As for the German site you're critical of, you need to read German to be able to understand numbers and bar graphs? Not to mention Computerbase is internationally acclaimed as one of the best resources for PC related topics. I linked their sites because they were one of the first to use such easy to read performance summaries and even break them down by resolution and settings.

    If you prefer since you listed it, TechPowerUp has similar listings, they only copied the performance summaries of course after ground-breaking sites like Computerbase were using them for some years.

    http://www.techpowerup.com/reviews/NVIDIA/GeForce_...

    As you can see, even at your lower 1680x1050 resolution, the GTX 280 still handily outclasses the 4870 512MB by ~21%, so I guess there goes your theory? I've seen more recent benchmarks in games like Skyrim or any title with 4xAA or high-res textures where the gap widens as the 4870 chokes on the size of the requisite framebuffer.

    As for your own situation and you're wife's graphics card, TigerDirect has the same GTX 670 GC card I bought for $350 after rebate. Not as good as the deal I got but again, there are certainly new ones out there to be had for cheap. I'm personally going to wait another week or two to see if the GTX 660 price and its impact on AMD prices (another round of cuts expected next week, LOL) forces Nvidia to drop their prices as I also need to buy another GPU for the gf.
  • Galidou - Sunday, September 16, 2012 - link

    And then once again you're off the subject, you send me relative performance of the gtx 280 to the 4870 2 years after their launch...... We were speaking of the rebate issued in 2008 for the performance it had back then. All the links I sent you were from 4 years ago and there's a reason to it and they're showing the 280 on average 10% difference in performance, and sometimes loosing big time.

    Sure the 1gb ram of the gtx 280 in the long run paid off. But we're speaking of 2008 situation that forced them to issue rebates..... which is 1 month after it's launch there's a part that performs ''similarly'' that costs less than HALF of it's price. computerbase is a good website, not the first time I see it, but it's the only one(and I never use only one website to base on the REAL average) that shows 20% difference in performance for a part that did still cost 115% more than the radeon 4870... 115 FREAKING % within one month!! nothing else to say.

    From a % point of view the 7970 and gtx 680 was a REALLY different fight.... and it was 3 month separating them which is something we commonly see in video card industry. while 115% more pricey parts that performs let's say 15% for your pleasure average from all websites than another part.....

    ''980X still offered benefit over SB with 6 cores''

    Never said it didn't, that's why I precised: ''A gamer buying an i7 980x 1 week prior to the sandy bridge launch''. For the gamer there was NO benefit at ALL, it even lost to the sandy bridge in games.

    But that's true, the 980x was out for a while unlike the gtx 280 who didn't have almost any time to keep it's amazing lead from last gen parts. The reason why they HAD to issue rebates, and the reason why I switched back to ATI from my 7800gt that died.

    The gtx 660 ti is a fine card, I'm just worried about the ROPs for future proofing, she'll keep the card a VERY long time so I regret not buying the 670 in the first place.
  • chizow - Sunday, September 16, 2012 - link

    No I'm not off the subject, you're obviously basing your performance differences based on a specific low-resolution setting that was important to you at launch while I'm showing performance numbers of all resolutions that have only increased over time. The TPU link I provided was from a later review because as I already stated, Computerbase was one of the first sites to use these aggregate performance numbers, only later did other sites like TPU follow suit. The GTX 280 was always the best choice for enthusiasts running higher resolutions and more demanding AA and those differences only increased over time, just as I stated.

    Nvidia didn't feel they needed to drop the price any more than the initial cut on the 280 because after the cut they had a "similar" part to compete with the 4870 with their own GTX 260. Once again, AMD charged too little for their effort, but that has no bearing on the fact that the 280's launch price was *JUSTIFIED* based on relative performance to last-gen parts, unlike the situation with AMD's 28nm launch prices.

    As for the 680 and 7970? It just started driving home the fact the 7970 was grossly overpriced, as it offered 10-15% *more* performance than the 7970 at *10%* less price which began the tumble on AMD prices we see today. I've also been critical of the GTX 680 though, as it only offers ~35-40% increase in performance over GTX 580 at 100% of the price, which is still the worst increase in the last 10 years for Nvidia, but still obviously better than the joke AMD launched with Tahiti. 115% performance for 110% of the price compared to last-gen after 18 months is an absolute debacle.

    As for the 980X and SB, again the whole tangent is irrelevant. What would make it applicable would be if AMD launched a bulldozer variant that offered 90% of 980X performance at $400 price point and forced Intel to drop prices and issue rebates, but that obviously didn't happen. You're comparing factors that Intel has complete control over where in the case of the GTX 280, Nvidia obviously had no control over what AMD decided to do with the 4870.
  • chizow - Sunday, September 16, 2012 - link

    There were numerous other important resolutions that took advantage of the 280's larger frame buffer, 1600x1200, 1920x1080 and 1920x1200. While they were obviously not as prevalent as they are now, they were certainly not uncommon for anyone shopping for a $300+ or $500 video card.

    As for the 7970 asking price, are you kidding? I had 10x as many AMD fanboys saying the 7970 price was justified at launch (not just Rarson), and where do you get 70% more perf? Its 50% being generous.

    So you got 150% performance for 150% of last-gen AMD price compared to 6970, how is that a good deal? Or similarly, you got 120% more performance for 110% the price compared to GTX 580, both last-gen parts.

    What you *SHOULD* expect is 150-200% performance for 100% of last-gen price, which is what the GTX 280 offered relative to 8800GTX, which is why I stated its pricing was justified.

    We've already covered the RV770, AMD could've easily priced it higher, even matched the GTX 260 price at $400 and still won, but they admittedly chose to go after market/mindshare instead after being beaten so badly by Nvidia since R600. Ever since then, they have clearly admitted their pricing mistake and have done everything in their power to slowly creep those prices upwards, culminating in the HUGE price increase we saw with Tahiti (see 150% price increase from 6970).
  • Galidou - Sunday, September 16, 2012 - link

    They went for price related to the size of the Die. The radeon 4870 was more than half the size of the gtx 280 thus costing less than half to produce then justifying the value of the chip by the size and not the perdformance.

    We all know why this is happening now, AMD was battling to get back for competition against the top because they left this idea by building smaller die with the HD 3xxx, leaving the higher end to double chip boards, end parts below 400$ prices. So I knew it had to happen one day or another.

    So if you're really into making a wikipedia about pricing scheme for video cards and developping about it, go on. But in my opinion, as long as they do not sell us something worse than last gen for a higher price, I'll leave it for people to discern what they need. With all the competition in the market, it's hard to settle for anything that's a real winner, it's mostly based on personnal usage and money someone is willing to spend.

    If someone want to upgrade his video card for xxx$, only thing he have to do is look at the benchmark for the game(s) he plays. Not looking at the price of the last generation of video cards to see if the price is relevant to the price he pays now. Usually a good 30 minutes looking at 3-4 different web sites looking at the graphs reading a little will give you good indication without sending you in the dust by speaking and arguing about last gen stuff compared to what's out now...

    You speak like you're trying to justify what people should buy now because of how things were priced in the past..... Not working. You take your X bucks check out benchies, go out there and buy the card you want end of the freaking line. No video card is interesting you now, wait until something does. Stop living in the past and get to another chapter ffs...
  • chizow - Sunday, September 16, 2012 - link

    Heh Wikipedia page? Obviously its necessary to set the record straight as revisionists like yourself are only going to emphasize the lowlights rather than the highlights.

    What you don't seem to understand is that transactions in a free market are not conducted in a vacuum, so the purchases of others do directly impact you if you are in the same market for these goods.

    Its important for reviewers to emphasize such important factors like historical prices and changes in performance, otherwise it reinforces and encourages poor pricing practices like we've seen from AMD. It just sets a bad precedent.

    Obviously the market has reacted by rejecting AMD's pricing scheme, and as a result we see the huge price drops we've seen over the last few months on their 28nm parts. All that's left is all the ill will from the AMD early adopters. You think all those people who got burned are OK with all the price drops, and that AMD won't have to deal with those repercussions later?

    You want to get dismissive and condescending, if all it took was a good 30 minutes looking at 3-4 different websites to get it, why haven't people like you and rarson gotten it yet?
  • rarson - Tuesday, September 18, 2012 - link

    Yeah, it was rhetorical. It was also pointless and off-topic.

    "And even after the launch of 28nm, they still held their prices because there was no incentive or need to drop in price based on relative price and performance?"

    Your problem is that you just don't pay any goddamn attention. You have the attention span of a fruit fly. Let me refresh your memory. Kepler "launched" way back in March. All throughout April, the approximate availability of Kepler was zero. AMD didn't drop prices immediately because Kepler only existed in a few thousand parts. They dropped prices sometime around early May, when Kepler finally started appearing in decent quantities, because by then, the cards had already been on the market for FOUR MONTHS. See, even simple math escapes you.

    "You have no idea what you're talking about, stop typing."

    You're projecting and need to take your own advice.
  • CeriseCogburn - Thursday, November 29, 2012 - link

    AMD's 7970/7950 series supply finally became "available" on average a few days before Kepler launched.
    LOL
    You said something about amnesia ?
    You rarson, are a sad joke.
  • CeriseCogburn - Thursday, November 29, 2012 - link

    I love it when the penny pinching amd fanboy whose been whining about 5 bucks in amd "big win!" pricing loses their mind and their cool and starts yapping about new technology is expensive, achieving the highest amd price apologist marks one could hope for.
    LOL
    It's awesome seeing amd fanboys with zero cred and zero morality.
    The GTX570 made the 7850 and 7870 the morons choice from the very first date of release.
    You cannot expect the truth from the amd fans. It never happens. If there's any exception to that hard and fast rule, it's a mistake, soon to be corrected, with a vengeance, as the brainwashing and emotional baggage is all powerful.
  • Galidou - Monday, September 17, 2012 - link

    While speaking about all that, pricing of the 4870 and 7970 do you really know everything around that, because it seems not when you are arguing, you just seem to put everything on the shoulder of a company not knowing any of the background.

    Do you know the price of the 4870 was already decided and it was in correlation with Nvidia's 9000 series performance. That the 4870 was supposed to compete against 400$ cards and not win and the 4850 supposed to compete against 300$ series card and not win. You heard right, the 9k series, not the GTX 2xx.

    The results even just before the coming out of the cards were already ''known''. The real things were quite different with the final product and last drivers enhancements. The performance of the card was actually a surprise, AMD never thought it was supposed to compete against the gtx 280, because they already knew the performance of the latter and that it was ''unnaittanable'' considering the size of the thing. Life is full of surprise you know.

    Do you know that after that, Nvidia sued AMD/ATI for price fixing asking for more communications between launch and less ''surprises''. Yes, they SUED them because they had a nice surprise... AMD couldn't play with prices too much because they were already published by the media and it was not supposed to compete against gtx2xx series. They had hoped that at 300$ it would ''compete'' against the gtx260 and not win against i thus justifying the price of the things at launch. And here you are saying it's a mistake launching insults at me, telling me I have a low intelligence and showing you're a know it all....

    Do you know that this price fixing obligation is the result of the pricing of the 7970, I bet AMD would of loved to price the latter at 400$ and could do it but it would of resulted in another war and more suing from Nvidia that wanted to price it's gtx 680 500$ 3 month after so to not break their consumers joy, they communicate A LOT more than before so everyone is happy, except now it hurts AMD because you compare to last gen and it makes things seems less of a deal. But with things back to normal we will be able to compare last gen after the refreshed radeon 7xxx parts and new gen after that.

    Nvidia the ''giant'' suing companies on the limit of ''extinction'', nice image indeed. Imagine the rich bankers starting to sue people in the streets, and they are the one you defend so vigorously. If they are that rich, do you rightly think the gtx 280 was well priced even considering it was double the last generation..

    It just means one thing, they could sell their card for less money but instead they sue the other company to take more money from our pockets, nice image.... very nice..... But that doesn't mean I won't buy an Nvidia card, I just won't defend them as vigorously as you do.... For every Goliath, we need a David, and I prefer David over Goliath.... even if I admire the strenght of the latter....

Log in

Don't have an account? Sign up now