--> http://www.nvidia.com/object/gpu_tec...onference.html
watch the opening keynote with Jen-Hsun Huang. He says it runs on Fermi. A lot better quality than youtube too.
And it's really interesting.
Printable View
--> http://www.nvidia.com/object/gpu_tec...onference.html
watch the opening keynote with Jen-Hsun Huang. He says it runs on Fermi. A lot better quality than youtube too.
And it's really interesting.
http://www.xbitlabs.com/news/video/d...703090833.html
Original tape out in april, then A0...10 weeks later they respin to A1 which is what they have now... ;)Quote:
The first working sample of a chip carries A0 revision, while companies usually launch A2 revisions commercially. It usually takes several – up to 10 – weeks to build a new chip revision, which means that it is unlikely that the G80 would be production-ready by September.
Also, I'm positive the oem NV15 was definitely revision A0.
explain to me how 512 shaders is not over double 240 shaders. the bandwidth increased by 50% too. the theoretical numbers are not that impressive but you completely missed a lot of factors and posted wrong information. nvidia also said 1.5ghz is a conservative estimate for clockspeed.
Are we talking about metal or silicon respins? Dunno why nV has only one letter and one number in the spin code - while ATi lists both silicon and metal spins. Anyways, if ATi's A0 was first revision then R600 would have gotten unrealistic total of 4 four respins, as early samples were A11 (1st rev silicon, 1st rev metal) when retail chips were A13. More likely, ATi's first rev is A11 meaning R600 had two metal respins (A11 -> A12 -> A13).
And, I haven't seen any nV, nor ATi, chips marked A0...
512 shaders is over double 240 (x2.13 to be exact). But 48 ROPs is not over double 32 (x1.5 to be exact). And 230 MB/s is not over double 141 (x1.63 to be exact). So overall, it's not over double the specs of the previous one. I don't think it's so hard to get what I've said there, and I don't get where I've said anything about CPs not being double (I think I have mentioned +113%). I would also like to know what are all those lot of factors that I've missed and what wrong information I've posted, based on what we know at the moment.
And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.
People were also talking about G200 being similar to G92 in clocks and look what happened at the first gen...
Fact is, when I hear Nvidia's own engineers claim that the design is delayed because it's incredibly hard, i'm not holding my breath on getting incredible clocks especially since they've had their own struggles going to 40nm
games are bound by shaders in the majority of cases. you can see that clearly in the 5870. they are running games at ridiculously high resolutions on a single card and still its bandwidth that really bottlenecks pixel fillrates. the factors you missed were new memory hierarchy, better scheduling logic, predication, and instruction set improvements.
i would trust nvidia more than i trust you for the clockspeed.
Here ya go guys...
http://www.hardocp.com/news/2009/10/...es_eyefinity63
I still don't care about multi-monitor for gaming until they make multi panel monitors into one frame, but good for those who do care.
I didn't miss that factors. They simply don't take any part in anything that I've said. And when it take it, I have mentioned them and considered them. Take the "trouble" of reading my posts and trying to understand them before quoting me, please, to not put things in my mouth.
And I don't know how to use HD5870 to know how games are shader bottlenecked since the proportion in which they have improved shader processing power it's the same than texture processing power, rasterizing operations processing power, and so.
There are more things involved in the 3D rendering process apart from shaders and memory bandwidth. ;)
Yeah, no doubt. But I think you have misunderstood them when you have the idea that they are talking about a clock of 1500MHz for the GPU core.:yepp:Quote:
i would trust nvidia more than i trust you for the clockspeed.
Um, you don't have to double EVERYTHING to get doubled performance. More than anything this depends on the particular application you're running, and where the bottlenecks lie within it.
If you look at a past example where performance WAS doubled, like the 8800GTX, let's compare that to the previous gen flagship, the 7900GTX. The 8800GTX had almost exactly twice the GFLOPs of the 7900GTX, even taking into account the nearly useless MUL op. The 8800GTX had 69% more memory bandwidth, and get this, only 33% more pixel fillrate, and 18% more bilinear texture fillrate.
The GF100 is more of an improvement in raw specs over the GTX 285 than the 8800GTX was over the 7900GTX. So doubling performance is more than possible.
Of course you haven't, A0 is usually in-house only. Only chip I can think of that released as an A0 from NVidia is the NV15. Usually it takes a few revisions before they can release.
Also, the R600 DID take several respins before it could release if you remember. You're talking about a card that was 6 months+ late. ;)
Now again, all that could have changed since then, but I've never heard or read anything to tell me that. I'd ask the reps, but that's likely information they aren't willing to let out.
Where in the post you're quoting I say that you have to double everything to double performance? I'm aswering a specific question.
And you can't compare G80 with previous generation, as it's a completely different architecture. Starting by the unified shader processors (instead of units that only could calculate vertex or pixel shaders), with a completely different architecture, and the same for TMUs and ROPs.
Again, I've never said that doubling is not possible (why is everybody putting that words in my mouth? It's at least the 3rd person who says that, and I'm starting to be tired or repeating it). You can read it yourself in my post quoted by Chumbucket843 (that I should add it's taken from a conversation including more posts before and after).
I have only said that there is not a single evidence which grants that the GT300 is going to be more than twice the performance of GT200.
But oh, well. If all of you are getting hurt by hearing it, I'll correct myself and let's finish with this: "GT300 is going to be obligatory at least 2x the performance of GTX285, and probably more". ¿Happy there?
EDIT: I have edited the former paragraphs to give a much more accurate response.
Of course it was a completely different architecture, as is the GF100. It may not be to the same extent as the G80 enjoyed, but it makes up for that by increasing raw specs more than the G80 did.
It really doesn't bother me when people say the GF100 won't perform as well as such-and-such or whatever, because no one knows, and everyone's entitled to their opinion. The main thing that bothers me is how much importance you place in ROPs, TMUs and bandwidth, when those are insignificant factors in games that are GPU-limited. Granted, there aren't that many of those anymore, thanks to consoles and the perceived threat of piracy.
my reference to the 5870 was to show that rops are where they should be. too much and youre just wasting die space. they are running games at 7680x3200. the rop's were added to help texture filtering quality which wont double the performance in either gpu. if you dont believe me look at the ratio of shaders to rops over the past 5 years.
this is the statement i was referring to:
Quote:
Consider that HD5870 is exactly double the HD4890 (+100% everything at the same clocks) except bandwidth (aprox. +30%) and it's far from double the real world performance (that's one of the most recent proves that doubling everything doesn't mean doubling real world performance), and NVIDIA is not even doubling processing units.
i responded to this part of your statement about shader clocks and you somehow got the idea i was talking about core?
Quote:
And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.
That's exactly what I was trying to say.
I have never said "GF100 won't perform as well as papapa".
Exactly that's what I'm talking about.
Somebody said "GF100 is going to perform at least twice as well as" and I asked him "Why? What's the reason why do you think that? What info which we have now lead you to take that for granted?".
And then some of you started quoting me puting words in my mouth.
But oh, you know what? The guilt is all mine:
I should have started to reply "I know it. I didn't say otherwise. Read it again" since the very beginning.
Processing units. CUDA cores are processing units. Texture units are processing units. Raster Operation Processors are processing units. So no, they are not "doubling processing units".
And regarding the clocks, obviously. If you understand it like shaders clock I don't know how it's an argument to say that they have doubled processing units power.
EDIT: And from my part, discussion about what I've said or left to say is over. My (at the present time favourable, and I think not unrealistic) opinions about GT300 are pretty clear at posts in previous pages (some of them quoted on this one), even when some people is absolutely determined to misunderstand them.