AMD Polaris architecture to succeed Graphics Core Next

With as small as the 14nm chips are, would it not be possible to have 6 dies stack of HBM or even 8?

That would bring the bus width to 6144 or 8192 in case of 6 and 8 stacks and increase the available bandwidth by 1.5 or 2 times. This is completely impractical for a GPU which is unlikely to be much faster than Fiji. Considering that the P10 die is likely to be small it may also be a problem to fit all the connections for such a wide bus on it.

I think that Polaris 10 will do just fine with GDDR5X and Maxwell like memory efficiency improvements. HBM is a more expensive tech which is better to be used for the "big die" Vega 10 which will actually require the bandwidth it provides.
 
That seems VERY unlikely.

Nvidia first Pascal is the 1080 one, that will be incredibly expensive. It's not very plausible that AMD will release nothing to fight in that slot.

Nor it's very plausible that the big market (970/290) will get absolutely nothing until 2017. Come on.

Nvidia also released the 950/960 AFTER the 980/970. It makes NO SENSE that they are releasing once again low end graphic cards.

What I'm saying is that Polaris 10 and 11 are lower power and their price will reflect that (we hope). Doesn't mean one of these GPUs won't match a 290. The high end AMD cards that will blow that out the water are coming 2017, according to the roadmap shown. Only that is packing HBM2. It will likely be a similar story for Nvidia.
 
That seems VERY unlikely.

Nvidia first Pascal is the 1080 one, that will be incredibly expensive. It's not very plausible that AMD will release nothing to fight in that slot.
Nor it's very plausible that the big market (970/290) will get absolutely nothing until 2017. Come on.

Seems on point to me. The 1080 rumour you're referring to relates to the GP104 part, which is almost certainly going to power the x80 and x70 products, follow ups to the 980 and 970 that debuted late summer 2014. ~18 months to 2 years gap.
I dispute "incredibly expensive". That product is probably in the $500 area and the x70 in the $350 area.

In September 2014 when the 980 and 970 released, AMD did not release anything new in that price bracket. AMD's response at the time was a price cut to their 200 line. 300 series and Fury did not arrive until Summer 2015. It makes sense that AMD's mid-high end GPUs may not be replaced until 18 months-2 years from the 300 release.

Nvidia also released the 950/960 AFTER the 980/970. It makes NO SENSE that they are releasing once again low end graphic cards.

The first Maxwell GPU released was the 750 based on GM107. There is precedent for new designs being debuted with lower performing parts. Not to mention that due to how semi-conductor fabrication goes, the parts with fewest transistors are likely to have the best yields and therefore are likely to be ready for sale sooner.
 
Seems on point to me. The 1080 rumour you're referring to relates to the GP104 part, which is almost certainly going to power the x80 and x70 products, follow ups to the 980 and 970 that debuted late summer 2014. ~18 months to 2 years gap.
I dispute "incredibly expensive". That product is probably in the $500 area and the x70 in the $350 area.

And so you say these won't be ready until 2017? That's unlikely.

We're talking merely of 16nm here. That's the jump. And whether or not there will be mainline cards (970,290 class) 16nm released this year.

The post I replied said that for this whole year only low-end cards will be released on 16nm. And I said I consider this very unlikely, especially with DX12 looming and Nvidia taking a beating on all the recent big releases.
 
As much as I would love to support AMD my next card will be another Nvidia due to the fact that I just purchased a GSync monitor, and I don't change monitors very often.
 
Raja reiterated the company’s plan to bring both GPUs to market availability by mid 2016. One is a high-end “enthusiast” version, while the other is a frugal power sipping console-class gaming GPU meant to deliver that level of performance to thin and light notebook form factors for the very first time.

So, that's it.
 
So high-end cards (500$+) and a mobile GPU for 2016? That's it?

My guess is that the new cards will be priced slightly higher than current ones, due to new tech.

So, taking Nvidia as reference, we'll see a 970 class going for at least $400-450 and the 1080 for at least $650.

If AMD releases just one model to compete in that slot, then it will be around $500.

What's uncertain is when or whether or not Nvidia releases the 970-like in the summer or more closer to the end of the year. Which is the ugly case, but possible one.
 
And so you say these won't be ready until 2017? That's unlikely.

We're talking merely of 16nm here. That's the jump. And whether or not there will be mainline cards (970,290 class) 16nm released this year.

The post I replied said that for this whole year only low-end cards will be released on 16nm. And I said I consider this very unlikely, especially with DX12 looming and Nvidia taking a beating on all the recent big releases.

Here's the post you replied to:

Round up for those confused:

Polaris (formerly Greenland) 10 and 11 GPUs are releasing this year. They are lower power GPUs.Polaris 11 is aimed for laptops. Polaris 10 is more powerful but a lower-end desktop part with some very impressive perf/watt.

Vega GPUs are coming 2017. These are high-end. Expect nvidia to roll out lower power Pascal GPUs first as well.

Radeon Pro Duo will have to satisfy AMD fans looking for new enthusiast products this year.

Nowhere does it say that there will only be low end 16nm products in 2017 from both vendors. From AMD? Yes, they have told you as much with their roadmap and their 390/Fury products released less than 1 year ago.

Nvidia is looking like they will release x80 and x70 products based on GP104 in 2016. This makes sense given the age of GM204 and the rumours we've seen recently. Additionally, the very first Pascal GPU released may be on the low end.
 
Nowhere does it say that there will only be low end 16nm products in 2017 from both vendors. From AMD? Yes, they have told you as much with their roadmap and their 390/Fury products released less than 1 year ago.

Nvidia is looking like they will release x80 and x70 products based on GP104 in 2016. This makes sense given the age of GM204 and the rumours we've seen recently. Additionally, the very first Pascal GPU released may be on the low end.

So not only you say Nvidia will release both 1070 and 1080 this year, along with a third low-end model. But that AMD instead just have the low-end one?

So AMD will surrender the whole market to Nvidia for another full year?

Come on.
 
So high-end cards (500$+) and a mobile GPU for 2016? That's it?
Couple of high end cards (Fiji+25%) on Polaris 10 and a couple of lower end cards (probably around R9 380X territory) on P11 seems right. But all of this is just guesswork for now.
 
Couple of high end cards (Fiji+25%) on Polaris 10 and a couple of lower end cards (probably around R9 380X territory) on P11 seems right. But all of this is just guesswork for now.

They could do a bigger die cut on P10 to have it give two tiers (4 cards) and then P11 can have two tiers (4 lower end cards)

Time will tell though

Edit:

My guess is 490x/490 and 480x/480 will be P10 and 470x/470 and 460x460 will be P11. New Fury will come early next year.
 
Actually if you look at what we have we also have a likely scenario:

we know Nvidia will release this summer the high end 1080 model.

we know that AMD has for this year a high end and low end model (the previously quoted article).

Result: no mid-range/mainstream model announced.

It means neither Nvidia nor AMD have announced something that will go in the 300-450 price range sweet spot. It doesn't mean that a model won't be out this year, as they are likely waiting to see each other moves, to determine what to do in the mainstream slice.

AMD will wait Nvidia to release the 1080, so they know how to market their own high end compared to that one.
 
So not only you say Nvidia will release both 1070 and 1080 this year, along with a third low-end model. But that AMD instead just have the low-end one?

So AMD will surrender the whole market to Nvidia for another full year?

Come on.

Having done more reading, I think there will be something fast in the ~$300-450 bracket from AMD this year. Well, I hope so. I want progress.
We know that Vega is coming in 2017 and at least one product will sport HBM2. Vega 10 is probably an enthusiast GPU a la GP100 to fill the $600+ bracket.
Of course, AMD are not giving specifics yet, so there's we're all forced to speculate. My guess is:

2016
Polaris 11 - Lower power GPU for laptops and portables, low end discrete.
Polaris 10 - Mid to high tier GPU. May use HBM1 at the higher end, GDDR5 at the lower.

2017
Vega 10 - Enthusiast GPU using HBM2.
 
Having done more reading, I think there will be something fast in the ~$300-450 bracket from AMD this year.

It's not likely because the only thing Nvidia has hinted is the 1080, that surely won't be in the 300-450 range, as it will be higher (though one wonders what happens to the 980 Ti).

If then AMD was to release a midrange model, then it would be them owning the market. It's more likely that AMD model will be trying to match the 1080.
 
I'm not an expert but I think that it would be a mistake not to have a solid 300-500 card ready to go for the VR graphics card buying bonanza this summer.
 
I don't care much for HBM1/2 and large VRAM pools [and even power efficiency, up to a point], since I plan on sticking to 1080p resolution for considerable time.

As long as this summer I can get my hands on ~300-350€ card that is well above gtx970 range, I will be very happy. This purchase will ensure I can enjoy games and VR for years to come. I'm sure that upcoming 16nm GPUs will serve me well. For now I aim to get a Radeon, but we will see...
 
Round up for those confused:

Polaris (formerly Greenland) 10 and 11 GPUs are releasing this year. They are lower power GPUs.Polaris 11 is aimed for laptops. Polaris 10 is more powerful but a lower-end desktop part with some very impressive perf/watt.

Vega GPUs are coming 2017. These are high-end. Expect nvidia to roll out lower power Pascal GPUs first as well.

Radeon Pro Duo will have to satisfy AMD fans looking for new enthusiast products this year.

Greenland was originally supposed to be the flagship GPU in the Arctic Islands family. This will take on Pascal GP100 that also uses HBM2.

I believe Vega with HBM2 *is* Greenland. Polaris 11 being the laptop GPU and Polaris 10 being the mid range desktop part.

Navi with Nextgen memory is a new next gen GPU architecture,it's scaleable.and therefore a new family. This is what will fight Nvidia's Volta.
 
It's not likely because the only thing Nvidia has hinted is the 1080, that surely won't be in the 300-450 range, as it will be higher (though one wonders what happens to the 980 Ti).

If then AMD was to release a midrange model, then it would be them owning the market. It's more likely that AMD model will be trying to match the 1080.

Of course the AMD model will be trying to match the x80 in performance, that's what I mean by high end. GTX 980, R9 390X. High end.

The prices I wrote in that post are speculative AMD prices. Nvidia prices are going to be higher. As I said earlier in the thread, Nvidia x80 usually comes in around $500-$550 and x70 around $350.

Nothing will happen to the 980 Ti pricing. Nvidia just stop making it and prices stay about the same from then onwards. You can pay over £500 for a new 780 Ti today if you like.
 
As much as I would love to support AMD my next card will be another Nvidia due to the fact that I just purchased a GSync monitor, and I don't change monitors very often.

This is one of the various reasons I'm moving to AMD. Nvidia is moving towards a walled garden approach with many of their actions including gameworks and Gsync. A big part of my decision was their refusal to support adaptive sync as an alternative when there's essentially nothing stopping them from doing so... aside from an anti consumer, and anti competitve stance that "locks" people into buying Nvidia GPU's for the life of their monitors. Like you for instance. I would have no issue If they offered both and then showed why G-sync is better as an upgraded feature. But trying to restrict a feature that is becoming a standard and is easy to implement as its a base feature of Display Port now just strikes me as wrong. (we may even see adaptive sync on TV's at some poin) Nvidia is forcing me to buy into a brand with this and I'm not ok with that. I want to be able to decide what GPU fits my needs best for price and performance not which one may or may not "work " with my monitor.
 
Actually if you look at what we have we also have a likely scenario:

we know Nvidia will release this summer the high end 1080 model.

we know that AMD has for this year a high end and low end model (the previously quoted article).

Result: no mid-range/mainstream model announced.

It means neither Nvidia nor AMD have announced something that will go in the 300-450 price range sweet spot. It doesn't mean that a model won't be out this year, as they are likely waiting to see each other moves, to determine what to do in the mainstream slice.

AMD will wait Nvidia to release the 1080, so they know how to market their own high end compared to that one.

I don't think this will be the case for NV.
GP106 will come alongside or not long after GP104.
And for AMD this may not be the case either depending on how good P11 will end up.

I also think that you're overestimating the "high end" offering of the first wave of 14/16nm GPUs. A cut down GP104 part is quite likely to fill the $450 position at least. Maxwell->Pascal transition won't be much different from any previous NV's node switch, with the whole lineup but the lowest end being upgraded to Pascal by the end of 2016 most likely. And both vendors are likely to introduce dual chip cards akin to 7990 and 690 for the top end $600+ range to fill this range until GP100 and Vega 10 will be ready for production.
 
AMD Polaris 10 to feature 2304 Stream Processors and 8GB GDDR5 memory

These look rather low. I wonder if Sisoft Ranker isn't able to properly detect the card yet.

Also: AMD confirms Polaris 10 is Ellesmere and Polaris 11 is Baffin
Vega 10 is likely to be what was Greenland previously.

Isn't Polaris 10 expected to be a mid-range desktop/high-end laptop chip? That would correlate with what we're seeing there. Given AMD's typical CU organisations, I'd imagine that it's a binned 40CU die, which should compete with or beat the 390X in its full format (given both increased clocks from the smaller process and architectural improvements). Sounds about right for a die intended for the 480 and 480X.

The clock speed reported sounds a bit odd, but could be explained either on the basis that it's very early silicon, or a laptop variant of the chip. The bandwidth's a little low, too, but that could simply be that they're using GDDR5 on the prototypes as they don't have GDDR5X chips to test with yet (as they use the same interface and GDDR5X has only just started sampling). Of course, we've seen AMD reduce bandwidth in like-for-like die successors before (Tonga vs Tahiti, for example), so I could definitely see them running the chip on much less bandwidth than Hawaii had.
 
Isn't Polaris 10 expected to be a mid-range desktop/high-end laptop chip? That would correlate with what we're seeing there. Given AMD's typical CU organisations, I'd imagine that it's a binned 40CU die, which should compete with or beat the 390X in its full format (given both increased clocks from the smaller process and architectural improvements). Sounds about right for a die intended for the 480 and 480X.
If P10 will be just a Hawaii replacement then this will be a disappointment for me as I expect GP104 to be faster than GM200 and with AMD's plans to launch only P10 and P11 this year this means that both GP104 and GP100 won't have any competition on the market until Vega 10 next year. Which means only bad things for us.

The clock speed reported sounds a bit odd, but could be explained either on the basis that it's very early silicon, or a laptop variant of the chip. The bandwidth's a little low, too, but that could simply be that they're using GDDR5 on the prototypes as they don't have GDDR5X chips to test with yet (as they use the same interface and GDDR5X has only just started sampling). Of course, we've seen AMD reduce bandwidth in like-for-like die successors before (Tonga vs Tahiti, for example), so I could definitely see them running the chip on much less bandwidth than Hawaii had.
Clock speeds are especially odd but this may be an engineering sample. I'm expecting Polaris cards to run on higher clocks than that of 28nm GCN meaning that they should be able to hit ~1.2GHz in reference.

As for bandwidth this may be both a result of Sisoft tool not detecting GDDR5X properly and/or Polaris memory system using extensive b/w optimizations akin to those used in Maxwell by NV.
 
If P10 will be just a Hawaii replacement then this will be a disappointment for me as I expect GP104 to be faster than GM200 and with AMD's plans to launch only P10 and P11 this year this means that both GP104 and GP100 won't have any competition on the market until Vega 10 next year. Which means only bad things for us.

I'll admit I haven't been following the rumours lately, but purely based on the progress of the 14nm/16nm processes, I don't see 980Ti/Fury X beating cards this year. You'd be looking at 300mm²+ die sizes, and I don't see them being affordable by the end of the year.

Clock speeds are especially odd but this may be an engineering sample. I'm expecting Polaris cards to run on higher clocks than that of 28nm GCN meaning that they should be able to hit ~1.2GHz in reference.

As for bandwidth this may be both a result of Sisoft tool not detecting GDDR5X properly and/or Polaris memory system using extensive b/w optimizations akin to those used in Maxwell by NV.

Focussing on bandwidth optimisations in Polaris/Vega would make sense for them, as it would further benefit their APUs, which are heavily bandwidth bottlenecked on DDR3/4 (the higher-end APUs next year may see HBM being used to solve this problem, but for more affordable chips they'll still need to work within the limits of DDR).
 
I'll admit I haven't been following the rumours lately, but purely based on the progress of the 14nm/16nm processes, I don't see 980Ti/Fury X beating cards this year. You'd be looking at 300mm²+ die sizes, and I don't see them being affordable by the end of the year.

GDC P10 demo of Hitman was running at Fury X performance basically. So I expect it to be at least on Fury's level from this. 300-400mm^2 dies should be ok this year if they'll cost $500+ on the market which is what I'm expecting. Then again a 400mm^2 14nm die should be able to fit more SPs than that of Fiji so seeing it perform just on Fiji's level or even worse - be a Hawaii replacement - would be a disappointment. It would also be a disappointment as I firmly believe that GP104 will beat GM200 which in itself is faster than Fiji.
 
If P10 will be just a Hawaii replacement then this will be a disappointment for me as I expect GP104 to be faster than GM200 and with AMD's plans to launch only P10 and P11 this year this means that both GP104 and GP100 won't have any competition on the market until Vega 10 next year. Which means only bad things for us.

I highly doubt GP100 is releasing this year.

Also, I think this data is most likely on the binned like not the full on.

So if these were 490x/490 I think the 2304SP's is for the 490.
 
I just want AMD to release something with HDMI 2.0 so I can actually upgrade my 7870... Being so strongly attached to team red, I feel dirty seriously contemplating a 970 as a stop-gap... but even that is giving me pause since I keep hearing "1070/80 just around the corner!"
 
GDC P10 demo of Hitman was running at Fury X performance basically. So I expect it to be at least on Fury's level from this. 300-400mm^2 dies should be ok this year if they'll cost $500+ on the market which is what I'm expecting. Then again a 400mm^2 14nm die should be able to fit more SPs than that of Fiji so seeing it perform just on Fiji's level or even worse - be a Hawaii replacement - would be a disappointment. It would also be a disappointment as I firmly believe that GP104 will beat GM200 which in itself is faster than Fiji.

It ran at 60fps on a 1440p screen using DX12. According to computerbase.de, the 390X averages just over 59fps at 1440p using DX12. Slightly outperforming the 390X would make sense for a 480X card.

I just want AMD to release something with HDMI 2.0 so I can actually upgrade my 7870... Being so strongly attached to team red, I feel dirty seriously contemplating a 970 as a stop-gap... but even that is giving me pause since I keep hearing "1070/80 just around the corner!"

I'd say all their new Polaris/Vega cards should support HDMI 2.0. Then again, the Fury line not supporting it was an odd move, given how much focus they put on 4K performance.
 
Polaris 10 has the same ram bandwidth as the PS4 & 2x the Stream Processors & clocked at the same clock speed.

14nm is 1/2 of 28nm do the math

None of this is likely to be correct.

Including the rather strange description of what 14nm FinFET GloFo is compared to TSMC's 28HP High-K.

It ran at 60fps on a 1440p screen using DX12. According to computerbase.de, the 390X averages just over 59fps at 1440p using DX12. Slightly outperforming the 390X would make sense for a 480X card.

Well, just slightly outperforming not the most performant card of this generation which cost $400 would definitely be a disappointing start of 14nm for AMD from my point of view.
 
Well, just slightly outperforming not the most performant card of this generation which cost $400 would definitely be a disappointing start of 14nm for AMD from my point of view.

Depending on architectural improvements in Polaris and the clocking abilities of Samsung/GF's 14nm process they may hit somewhere between Hawaii and Fiji in final performance, but I'd err on the side of caution on both counts.

The issue is that it's not really up to AMD or Nvidia as to when they can release high-end 14nm/16nm cards, it's up to Samsung, Global Foundries and TSMC. Yields need to be good enough to make large dies viable, and until they are AMD and Nvidia wouldn't be able to release their 1080 or 490 cards however much they wanted to.

Lets have a look at current and confirmed future chips planned for Samsung/GF's 14LPP process (which is what AMD will be using for Polaris):

Snapdragon and Exynos 8890: Just shipping now, being used in high-end smartphones (Galaxy S7 starts at $670 off-contract). No confirmation yet, but both likely to be in and around 100mm². Likely to be clock-speed binned, but I don't believe either will be binned with disabled blocks.

AMD Zen: Up to eight-core CPUs (w/o IGP) reported to launch in October. Die sizes likely to top out somewhere between 150mm² to 200mm² (judging by Intel's 14nm core sizes, no integrated graphics, and likely smaller cores than Intel due to reported 32 core APUs on 14nm). Likely to be binned both with disabled cores and various clock speeds, which will increase yields. Full eight-core chip could be pretty expensive. If they're genuinely competing with Intel on IPC, then the comparable Intel chip (5960X) costs about $1000.

IBM POWER9: Very large die (POWER8 is 650mm² at 22nm) with a high core count and big cache, due sometime 2017. Likely to both bin with disabled cores and multiple clock options. IBM doesn't sell POWER CPUs directly, but they'd be several thousand dollars a piece if they did.

AMD Polaris 10 & 11: Judging by AMD's presentation, probably shipping in the second half of the year. Each will be binned both with disabled CUs (for non-X versions) and on clock speeds (laptop versions). The report you linked to suggests a 40CU part for the larger die (Polaris 10), which would put the die size in and around 200mm². Desktop Hawaii+ performance would probably see a $300-$350 or so price, but it's likely that they're largely targeting laptops, where the same chip would go for perhaps twice as much, and the TDP-reduction benefits of 14nm would be worth the manufacturing cost increase over 28nm.

AMD Vega 10 & 11: As far as I'm aware these aren't officially confirmed to be on 14nm, so they could actually be TSMC 16nm chips. We know they're arriving in 2017, and given their position above Polaris, they're likely to push above 200mm², probably 300-400mm² if they want to comfortably outperform Fiji.

And the confirmed chips on TSMC's 16nm:

Apple A9: Also produced on Samsung's older 14LPE process, the TSMC version is 104.5mm² and was first introduced last September in the iPhone 6s (starting at $649 off-contract). One or both of these are also used in the cheaper iPhone SE (starting at $399), although other than the SoC and fingerprint scanner the phone is almost identical to the 3 and a half year old iPhone 5. Unlikely to be binned with disabled blocks, there may be some clock speed binning between the 6S and SE variants of the chip.

Apple A9X: First released in November last year in the $799+ 13" iPad Pro, this is a 147mm² chip, it's the largest chip yet mass produced at 16nm or 14nm by anyone other than Intel. Also used in the soon to be released 10" iPad Pro, which is $100 more expensive than 10" iPads are typically introduced at. They're adding pencil support to justify the Pro moniker, but the price is likely to be almost entirely down to the exceptionally high cost of the SoC. Also unlikely to be binned with disabled cores, but there might be some clock speed binning between the larger and smaller iPad Pro.

Nvidia Pascal: As far as I'm aware there's been no official word on availability other than that some Pascal chips will become available in 2016 (and I don't believe they even confirmed that much in recent presentations such as the one at CES 2016). The rumours are kind of all over the place (and often come across more as hopeful speculation than insider reports).

AMD Vega 10 & 11: As mentioned above, it's possible these will be on TSMC's 16nm process, although we don't know much about them other than that they're coming in 2017.

Then there's Intel's troubles with their 14nm FinFET node. Despite their enormous R&D budget, they still haven't managed to put out a chip larger than 168mm² on 14nm (a process which was initially intended to be replaced in just a few months time). Broadwell was significantly delayed, we don't have any Iris Pro Skylake chips yet, and there's still no sign of 14nm Broadwell-E Xeons (which usually wouldn't have an issue with relatively poor yields, due to ample binning and high margins). And we've just had confirmation from them that the slow progress on 14nm and 10nm has caused them to abandon their two-year node cycle and replace it with a three-year cycle from one node to the next.

GPUs are in an unusual transition between the long-lived 28nm node and the slowly improving yields of the 14nm/16nm node. AMD and Nvidia have squeezed about as much as possible out of the 28nm node, with both top-end GPU dies being around 600mm², which puts a high hurdle up for the new chips to jump over. The new nodes, though, are combining very slow yield growth with likely a decent jump in wafer costs due to the more advanced FinFET process. This means a much longer wait before it's economical to build high-end GPUs on the new nodes. Back in early 2012, even with reportedly low yields on the 28nm process, Tahiti was the first major product that came to market on the process, and within a few months both AMD and Nvidia had brought pretty much their entire line-ups onto the node. The reason was that, even with poorer yields than 40nm, and higher wafer costs, the increased density meant that quite quickly it actually became cheaper to get a given performance on 28nm than 40nm.

That's not the case this time around. While for TSMC's previous node jumps, it was almost always GPUs that were the first major chips to market, this time it's all high-end phone and tablet SoCs (and the same for Samsung/GF's 14nm). The reason for this is that these new nodes aren't cheaper per unit performance than 28nm, and they're unlikely to be for some time. High-end phones are heavily driven by performance per watt, both in terms of offering performance in a tight thermal environment, but also to reduce power consumption to increase battery life and allow for thinner devices. Hence, they're willing to pay a premium over 28nm and 20nm in order to compete on those terms.

This is why I believe AMD and Nvidia will initially introduce Polaris and Pascal chips which primarily target laptops. High-end desktop chips are likely to be infeasible for some time, and even for mid-range cards, they're unlikely to be offer an improvement in performance per dollar for AMD or Nvidia. In a laptop form-factor, though, the TDP-reduction benefits would become worth it. Laptop GPUs sell for significantly higher prices than their desktop equivalents, and benefit highly from being able to squeeze more performance out of the constraints of operating within a laptop.
 
The issue is that it's not really up to AMD or Nvidia as to when they can release high-end 14nm/16nm cards, it's up to Samsung, Global Foundries and TSMC. Yields need to be good enough to make large dies viable, and until they are AMD and Nvidia wouldn't be able to release their 1080 or 490 cards however much they wanted to.
A GPU of the same size as GK104 or Tahiti made on 14/16nm should be able to beat both GM200 and Fiji from a pure theoretical point of view. Yields has nothing to do with this as a) we don't know them and b) it's a question of retail pricing which decides if yields are fine or not and we don't know that either.

Lets have a look at current and confirmed future chips planned for Samsung/GF's 14LPP process (which is what AMD will be using for Polaris):

Snapdragon and Exynos 8890: Just shipping now, being used in high-end smartphones (Galaxy S7 starts at $670 off-contract). No confirmation yet, but both likely to be in and around 100mm². Likely to be clock-speed binned, but I don't believe either will be binned with disabled blocks.
Mobile chips are using a different version of 14nm process.

AMD Zen: Up to eight-core CPUs (w/o IGP) reported to launch in October. Die sizes likely to top out somewhere between 150mm² to 200mm² (judging by Intel's 14nm core sizes, no integrated graphics, and likely smaller cores than Intel due to reported 32 core APUs on 14nm). Likely to be binned both with disabled cores and various clock speeds, which will increase yields. Full eight-core chip could be pretty expensive. If they're genuinely competing with Intel on IPC, then the comparable Intel chip (5960X) costs about $1000.
We don't know much about Zen yet to make any kind of prognosis here. But if AMD is planning to sell them for $1000 they're delusional.

IBM POWER9: Very large die (POWER8 is 650mm² at 22nm) with a high core count and big cache, due sometime 2017. Likely to both bin with disabled cores and multiple clock options. IBM doesn't sell POWER CPUs directly, but they'd be several thousand dollars a piece if they did.
Power CPUs are for supercomputers only these days and their costs there are obviously of no issue for the most part. I.e. a bad reference point for a consumer level GPU chips. You can use POWER9 chips as a reference for a possible GP100 introduction to the same HPC market this year however.

AMD Polaris 10 & 11: Judging by AMD's presentation, probably shipping in the second half of the year. Each will be binned both with disabled CUs (for non-X versions) and on clock speeds (laptop versions). The report you linked to suggests a 40CU part for the larger die (Polaris 10), which would put the die size in and around 200mm². Desktop Hawaii+ performance would probably see a $300-$350 or so price, but it's likely that they're largely targeting laptops, where the same chip would go for perhaps twice as much, and the TDP-reduction benefits of 14nm would be worth the manufacturing cost increase over 28nm.
You're assuming that Polaris will be just a shrink of GCN3. Which it won't be, not from architectural point of view and not from the production point of view either as GloFo's 14LPE is not a shrink process to TSMC's 28HP. There are just too many variables to guess on right now so this is mostly pointless.

All I can be sure of is that porting Hawaii to 14LPE should give AMD a die of ~200mm^2 - which seems a bit on the low side from my point of view as all AMD GPU dies on a new production process for the last several years have been bigger than that. They usually start the new process with a die of approximately 300mm^2 in size which would be analogous to a chip of ~640mm^2 size on 28HP process - which is pretty close to how large Fiji is.

Hence my possible disappointment with P10 if it'll end up being a Hawaii "port" and will fall somewhere between Hawaii and Fiji in performance. I pretty much expect P10 to be a "port" of Fiji with higher clocks and optimized architecture driving it above Fiji in performance, not below it.

AMD Vega 10 & 11: As far as I'm aware these aren't officially confirmed to be on 14nm, so they could actually be TSMC 16nm chips. We know they're arriving in 2017, and given their position above Polaris, they're likely to push above 200mm², probably 300-400mm² if they want to comfortably outperform Fiji.
It would be a disaster if AMD will wait till 2017 to introduce something faster than Fury X. In any case, I think that we shouldn't read too much into a marketing slide measuring some unknown perf/watt.

GPUs are in an unusual transition between the long-lived 28nm node and the slowly improving yields of the 14nm/16nm node. AMD and Nvidia have squeezed about as much as possible out of the 28nm node, with both top-end GPU dies being around 600mm², which puts a high hurdle up for the new chips to jump over. The new nodes, though, are combining very slow yield growth with likely a decent jump in wafer costs due to the more advanced FinFET process. This means a much longer wait before it's economical to build high-end GPUs on the new nodes. Back in early 2012, even with reportedly low yields on the 28nm process, Tahiti was the first major product that came to market on the process, and within a few months both AMD and Nvidia had brought pretty much their entire line-ups onto the node. The reason was that, even with poorer yields than 40nm, and higher wafer costs, the increased density meant that quite quickly it actually became cheaper to get a given performance on 28nm than 40nm.

That's not the case this time around. While for TSMC's previous node jumps, it was almost always GPUs that were the first major chips to market, this time it's all high-end phone and tablet SoCs (and the same for Samsung/GF's 14nm). The reason for this is that these new nodes aren't cheaper per unit performance than 28nm, and they're unlikely to be for some time. High-end phones are heavily driven by performance per watt, both in terms of offering performance in a tight thermal environment, but also to reduce power consumption to increase battery life and allow for thinner devices. Hence, they're willing to pay a premium over 28nm and 20nm in order to compete on those terms.

This is why I believe AMD and Nvidia will initially introduce Polaris and Pascal chips which primarily target laptops. High-end desktop chips are likely to be infeasible for some time, and even for mid-range cards, they're unlikely to be offer an improvement in performance per dollar for AMD or Nvidia. In a laptop form-factor, though, the TDP-reduction benefits would become worth it. Laptop GPUs sell for significantly higher prices than their desktop equivalents, and benefit highly from being able to squeeze more performance out of the constraints of operating within a laptop.

Well, we can't really talk about yields because we don't know anything on this. Yields are highly product related anyway and thus it's pointless to talk about general "process yields" as there is no such thing.

While you are definitely correct on the prices of wafers for newer processes rising all the time and essentially pushing bigger chips into the future where they will be feasible to produce what you're missing is that 200mm^2 isn't a typical starting point on a new process. It may be that both AMD and NV decided to use the first generation of 14/16 chips for mobile solutions only but I personally find this rather unlikely as their key market is dGPUs. Thus the wait for the ability to produce a GPU which will be better that the top dog of the previous gen has already been done most likely - and it's probably the reason both of them have passed on 20nm process for GPUs.

This means that we should expect chips which will be bigger than 200mm^2 from the start as only such chips will be able to beat Fiji/GM200 - which is crucial for them as this allows to price the products on these chips accordingly. When you talk about high end and mid range you seem to be thinking solely in terms of chip sizes while these things are not even related - the market segment of a product is decided upon from a chip's performance, not its size.

Thus I'd argue that building a 300mm^2 chip and using it for a card which will retail for $500-700 while performing better than modern $500-700 solutions is actually a better way to handle the economies of the new production process compared to a chip of ~200mm^2 which won't beat the current high end card and because of that will have to be sold in the $300-500 bracket.

It's also rather important for either IHV not to undershoot in a segment when producing a new GPU on 14/16nm because that would mean that the competitor has gotten himself a year or so of complete market domination in this segment. And I'm pretty sure that NV will try to beat 980Ti with GP104 as this is the only way I see for them to handle the process node transition without loosing a lot of margins.

So yeah, I understand what you're saying but I expect a better future. P10 being slower than Fiji will be very disappointing. Same for GP104 being slower than GM200.
 
Vega 10 or similar, with all that HBM2, would probably be a ps5 sort of thing. I can't see sony doing any major memory shifts, save for gddr5 to gddr5+ or gddr5x.
 
The Vega 10 GPU (H1 2017) pretty much confirmed to be 'Greenland' with 4096 Stream Processors, strong FP64 compute performance and upto 32 GB HBM2.

-FirePro cards for HPC market.
-HPC APUs
-Radeon cards
-probably future Radeon Pro Duo cards


http://videocardz.com/58665/amds-greenlandvega-10-gpu-to-feature-4096-stream-processors
http://www.3dcenter.org/news/amds-greenland-chip-mit-4096-shader-einheiten-bestaetigt
http://wccftech.com/amd-vega-10-4096-stream-processors/
This is definitely disappointing. 4096 is Fiji's number and this basically mean that P10 will likely be a Hawaii "port". Well, I'll still hope for faster h/w for now since nothing is officially confirmed yet.

Is the Vega 10 remotely possible in a console?
Technically sure why not. A GPU of Fiji's complexity on 14nm process would fit into ~300mm^2 meaning that there's a lot of room left for CPUs to make APU from it.
Economically and from power consumption point of view however this is somewhat unlikely - in 2018 or 2019 maybe but not this year and not 2017 I'd say.
 
I will, as always be going with the card the offers the best bang for buck that meets meets my needs. Currently I'm holding off until VR settles in a bit and we get a good feel of what game requirements will be.
 
This is definitely disappointing. 4096 is Fiji's number and this basically mean that P10 will likely be a Hawaii "port". Well, I'll still hope for faster h/w for now since nothing is officially confirmed yet.

I don't know why you keep saying it's disappointing simply based on the number of shaders. Maxwell GPUs had less CUDA Cores than their Kepler equivalents but in practice were much faster, due to higher clocks and shader efficiency gains. It's not out of the question to believe the same will probably happen with Polaris.
 
A GPU of the same size as GK104 or Tahiti made on 14/16nm should be able to beat both GM200 and Fiji from a pure theoretical point of view. Yields has nothing to do with this as a) we don't know them and b) it's a question of retail pricing which decides if yields are fine or not and we don't know that either.

You're correct in that we don't know yields, but we can infer a lot from what we do know, which is the timescale of confirmed product launches on 14nm/16nm, and likely pricing information. The point of my post is that all the publicly confirmed information we have points to a very slow yield growth and a process which will still be expensive even for relatively small dies for the rest of the year.

Mobile chips are using a different version of 14nm process.

No, the A9 and Samsung's Exynos 7420 (used in the S6) are made on the older 14LPE process. The Exynos 8890 and the Snapdragon 820 are made on the newer 14LPP process, and are being used in currently shipping phones like the S7. The Snapdragon 820 in particular is the first 14LPP chip mass-produced by Global Foundries, as far as I'm aware.

We don't know much about Zen yet to make any kind of prognosis here. But if AMD is planning to sell them for $1000 they're delusional.

Of course I'd expect AMD will undercut Intel on price with Zen. My point, though, was that they could undercut Intel even if the cost of an 8-core die pushed the price all the way up to $800 or so. The cost of a quad-core chip will have to be under $300 to compete with consumer-level i7's, but the quad-core die will be truly tiny (i.e. likely sub 100mm²).

Power CPUs are for supercomputers only these days and their costs there are obviously of no issue for the most part. I.e. a bad reference point for a consumer level GPU chips. You can use POWER9 chips as a reference for a possible GP100 introduction to the same HPC market this year however.

The bolded is exactly my point. IBM can afford to push out large-die POWER chips pretty much straight out of the gate on new processes, because they're heavily binned and command large margins. POWER8 was the first chip on their/GF's 22nm node, POWER7+ was the first on their 32nm node, POWER7 was the first on their 45nm node, etc., etc.

Yet here we are with the 14nm process they'll be using for POWER9 (Global Foundries' 14nm) starting to push out chips, and IBM's new chip is a year or more away. This isn't a design delay, either, as the three year wait between POWER8 and POWER9 will actually be the longest ever between generations of IBM's server chips.

So that's a data point in the expected yields of the 14nm node: IBM don't expect a heavily binned large 14nm die (potentially around 600mm²) to be profitable at several thousand dollars a piece until 2017. This may not seem like it's relevant to smaller 300-400mm² dies, but it again paints a picture of a very slowly maturing node.

You're assuming that Polaris will be just a shrink of GCN3. Which it won't be, not from architectural point of view and not from the production point of view either as GloFo's 14LPE is not a shrink process to TSMC's 28HP. There are just too many variables to guess on right now so this is mostly pointless.

You're right in that it's not a simple shrink, and different kinds of chips will scale differently. We do, however, have a couple of cases where AMD and Nvidia have either produced a direct die shrink of a GPU, or have produced an almost identical die on a smaller process, which we can use as a baseline for our expectations.

The first is Nvidia's G92, which was first produced on a 65nm process, and later shrunk to a 55nm process. The die was 324mm² on 65nm and 260mm² on 55nm, so that's a 19.8% reduction in size over a single node jump (28nm to 14nm is of course a two node jump). The second is AMD's RV790, which was a 282mm² die on 55nm. It was replaced by Juniper, which wasn't a direct die shrink, but shared the same architecture (Terascale 1) and ALU configuration (although it did have a narrower 128-bit GDDR5 interface, which would have saved some die space). Juniper was 170mm² on a 40nm process, so that was a 39.7% shrink (probably closer to 35% accounting for the smaller memory interface).

So, for a single-node jump, assuming no architectural changes, we have a die shrink range from about 20% to about 35%. For a two-node jump, then, our expectation should be in the range from about 36% to about 58%, or let's say 47% (+-11).

Still assuming no architectural changes for the moment, this would place a straight shrink of Hawaii at 232mm² (+-48).

Once you account for architectural improvements, though, the die size is more likely to go up than down. Attempting to estimate those changes would be pure speculation, so I'll leave it at this:

Based on the evidence available to us, a 44CU Polaris die would likely exceed 200mm², perhaps by a large margin. It would certainly outperform Hawaii, although to what extent is impossible to predict.

All I can be sure of is that porting Hawaii to 14LPE should give AMD a die of ~200mm^2 - which seems a bit on the low side from my point of view as all AMD GPU dies on a new production process for the last several years have been bigger than that. They usually start the new process with a die of approximately 300mm^2 in size which would be analogous to a chip of ~640mm^2 size on 28HP process - which is pretty close to how large Fiji is.

Hence my possible disappointment with P10 if it'll end up being a Hawaii "port" and will fall somewhere between Hawaii and Fiji in performance. I pretty much expect P10 to be a "port" of Fiji with higher clocks and optimized architecture driving it above Fiji in performance, not below it.

What AMD usually start on a new process with is irrelevant, as this isn't 28nm or 40nm or any older process. If it was, we'd already have a full range of Polaris/Vega and Pascal GPUs on store shelves.

It would be a disaster if AMD will wait till 2017 to introduce something faster than Fury X. In any case, I think that we shouldn't read too much into a marketing slide measuring some unknown perf/watt.

That "marketing slide" is pretty much the only piece of confirmed information we have (i.e. not rumour) on the timescale for the rollout of AMD's new GPUs. And that slide appears to show Polaris releasing in late 2016 and Vega in early/mid 2017.

(Regarding the perf/W, I would imagine the higher figure for Vega is largely due to its use of HBM2 compared to more power-hungry GDDR5(X) for Polaris).

Well, we can't really talk about yields because we don't know anything on this. Yields are highly product related anyway and thus it's pointless to talk about general "process yields" as there is no such thing.

There very much is such a thing as process yields. Different chips will have slightly different yields depending on their design (for example 1mm² of SRAM will be less likely to develop a fault than 1mm² of custom ALU logic), but general improvements in the manufacturing process over the life of a node help all dies in fairly equal measure.

The main difference in yield is binning strategies, which are employed on almost all but the smallest of dies. Speed binning will increase yields by a relatively small amount (it's more generally used to squeeze higher profits out of the best dies), whereas binning with disabled functional blocks (i.e. CPU cores or GPU ALUs) will increase yields by a larger amount. Block level binning will of course be used on Polaris, Vega and Pascal, but it has no greater benefit than it did on their 28nm products, or on AMD's Zen processors or APUs, or on Intel's forthcoming 14nm Xeons, or on IBM's POWER9. If slow yield growth is preventing those kinds of chips from releasing in a timely fashion, then we can expect the same slow yield growth to affect AMD and Nvidia's plans.

While you are definitely correct on the prices of wafers for newer processes rising all the time and essentially pushing bigger chips into the future where they will be feasible to produce what you're missing is that 200mm^2 isn't a typical starting point on a new process. It may be that both AMD and NV decided to use the first generation of 14/16 chips for mobile solutions only but I personally find this rather unlikely as their key market is dGPUs. Thus the wait for the ability to produce a GPU which will be better that the top dog of the previous gen has already been done most likely - and it's probably the reason both of them have passed on 20nm process for GPUs.

Their key market is desktop GPUs, but laptop GPUs command significantly higher margins, and are highly dependent on perf/W. If AMD's claims of 2.5x improvement in power efficiency over GCN1.2 are true, then a mobile Polaris 11 chip that matched their current Tonga-based M395X could be squeezed into a 50W envelope, not much more than the M370X used in the current MacBook Pro. And of course they could put out Polaris 11 based mobile GPUs which substantially outperform any current laptop chips at around the 100W envelope.

Now let's say Nvidia waits until a 980Ti-beater is feasible before releasing any Pascal chips. Why would AMD wait as well, if they could release a 400M line which could gain them significant market-share in a profitable market before Nvidia's FinFET GPUs arrive? Ditto for Nvidia, although I suppose they have less incentive to move early as they already have a larger share of the laptop market.

This means that we should expect chips which will be bigger than 200mm^2 from the start as only such chips will be able to beat Fiji/GM200 - which is crucial for them as this allows to price the products on these chips accordingly. When you talk about high end and mid range you seem to be thinking solely in terms of chip sizes while these things are not even related - the market segment of a product is decided upon from a chip's performance, not its size.

Thus I'd argue that building a 300mm^2 chip and using it for a card which will retail for $500-700 while performing better than modern $500-700 solutions is actually a better way to handle the economies of the new production process compared to a chip of ~200mm^2 which won't beat the current high end card and because of that will have to be sold in the $300-500 bracket.

It's important to note that die costs aren't linear wrt size, in fact on an immature node they're close to exponential. Not only is in entirely possible for a 200mm² die to be financially viable for a $350 product well before a 300mm² die is viable for a $500+ product, but if those 200mm² dies will be sold as both desktop and mobile GPUs, the average selling price per chip will increase substantially (i.e. even if Polaris 10 isn't actually viable for a $350 desktop card, it may be worth entering mass production on it if a large proportion of them actually sell for twice as much as laptop GPUs).

It's also rather important for either IHV not to undershoot in a segment when producing a new GPU on 14/16nm because that would mean that the competitor has gotten himself a year or so of complete market domination in this segment. And I'm pretty sure that NV will try to beat 980Ti with GP104 as this is the only way I see for them to handle the process node transition without loosing a lot of margins.

So yeah, I understand what you're saying but I expect a better future. P10 being slower than Fiji will be very disappointing. Same for GP104 being slower than GM200.

The only way I could see Nvidia entering the market with high-end cards substantially sooner than AMD do is if:

(a) TSMC's 16nm process yields make large dies affordable far before they are on GF's 14nm process

and

(b) AMD are dropping TSMC altogether irrespective of their 16nm yields.

The first isn't completely impossible (see A9X being fabbed by TSMC rather than Samsung, although as discussed previously it's still an expensive chip). The second, though, I would find very strange. AMD have been a long-time customer of TSMC, and would have kept themselves very well informed on the expected yield growth on their new process. If it was going to offer substantially better yields than 14nm then AMD would be using it, particularly for their larger dies (hence my speculation that Vega may be on TSMC's 16nm). Even if 14LPP offers better performance/Watt for their CPUs, APUs and small/mobile GPUs, AMD have shown they're willing to shop around between fabs, and if their high-end desktop GPUs were able to get better yields on TSMC, that's where they'd go.
 
Ugh. I really hope one of the two release a faster-than-980ti card this summer for < $700. Lots of people have bought one of the two VR solutions, and 970 is the minimum requirement (not even the recommended), SLI/XF is not recommended, so I hope one of them releases one that is quite a bit better than 980ti for around 980 ti's launch price. If either company does, they get a buy from me.
 
Ugh. I really hope one of the two release a faster-than-980ti card this summer for < $700. Lots of people have bought one of the two VR solutions, and 970 is the minimum requirement (not even the recommended), SLI/XF is not recommended, so I hope one of them releases one that is quite a bit better than 980ti for around 980 ti's launch price. If either company does, they get a buy from me.
Sli/cf is not recommended but both gpu vendors are pushing specialized SFR solutions through their respective VR APIs. Given the resource requirements of VR I'd expect more conscientious developers to use it.

Of course, that remains to be seen.
 
Top Bottom