MMM
Results 1 to 25 of 1035

Thread: The official GT300/Fermi Thread

Hybrid View

  1. #1
    Xtreme Cruncher
    Join Date
    May 2009
    Location
    Bloomfield
    Posts
    1,968
    Quote Originally Posted by Farinorco View Post
    Yeah but... why do you expect it to have 2x the performance (I'm suppose you're talking about real world performance) if it's going to have +113% CPs more but only +50% ROP more, +60% mem bandwidth more...

    Consider that HD5870 is exactly double the HD4890 (+100% everything at the same clocks) except bandwidth (aprox. +30%) and it's far from double the real world performance (that's one of the most recent proves that doubling everything doesn't mean doubling real world performance), and NVIDIA is not even doubling processing units.

    Can they improve the performance per unit and per clock? Sure. Maybe. But how much and why, I think is way soon with the info we have to say it's going to be 2x real world performance of a GTX285. I even would say I hugely doubt it, given that they are more focused in get the new (future?) HPC market before Intel has their Larrabee working (if it happens to be on this century).
    explain to me how 512 shaders is not over double 240 shaders. the bandwidth increased by 50% too. the theoretical numbers are not that impressive but you completely missed a lot of factors and posted wrong information. nvidia also said 1.5ghz is a conservative estimate for clockspeed.

  2. #2
    Xtreme Member
    Join Date
    Jul 2009
    Location
    Madrid (Spain)
    Posts
    352
    Quote Originally Posted by Chumbucket843 View Post
    explain to me how 512 shaders is not over double 240 shaders. the bandwidth increased by 50% too. the theoretical numbers are not that impressive but you completely missed a lot of factors and posted wrong information. nvidia also said 1.5ghz is a conservative estimate for clockspeed.
    512 shaders is over double 240 (x2.13 to be exact). But 48 ROPs is not over double 32 (x1.5 to be exact). And 230 MB/s is not over double 141 (x1.63 to be exact). So overall, it's not over double the specs of the previous one. I don't think it's so hard to get what I've said there, and I don't get where I've said anything about CPs not being double (I think I have mentioned +113%). I would also like to know what are all those lot of factors that I've missed and what wrong information I've posted, based on what we know at the moment.

    And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.

  3. #3
    Xtreme Cruncher
    Join Date
    May 2009
    Location
    Bloomfield
    Posts
    1,968
    Quote Originally Posted by Farinorco View Post
    512 shaders is over double 240 (x2.13 to be exact). But 48 ROPs is not over double 32 (x1.5 to be exact). And 230 MB/s is not over double 141 (x1.63 to be exact). So overall, it's not over double the specs of the previous one. I don't think it's so hard to get what I've said there, and I don't get where I've said anything about CPs not being double (I think I have mentioned +113%). I would also like to know what are all those lot of factors that I've missed and what wrong information I've posted, based on what we know at the moment.

    And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.
    games are bound by shaders in the majority of cases. you can see that clearly in the 5870. they are running games at ridiculously high resolutions on a single card and still its bandwidth that really bottlenecks pixel fillrates. the factors you missed were new memory hierarchy, better scheduling logic, predication, and instruction set improvements.

    i would trust nvidia more than i trust you for the clockspeed.

  4. #4
    Xtreme Member
    Join Date
    Jul 2009
    Location
    Madrid (Spain)
    Posts
    352
    Quote Originally Posted by Chumbucket843 View Post
    games are bound by shaders in the majority of cases. you can see that clearly in the 5870. they are running games at ridiculously high resolutions on a single card and still its bandwidth that really bottlenecks pixel fillrates. the factors you missed were new memory hierarchy, better scheduling logic, predication, and instruction set improvements.
    I didn't miss that factors. They simply don't take any part in anything that I've said. And when it take it, I have mentioned them and considered them. Take the "trouble" of reading my posts and trying to understand them before quoting me, please, to not put things in my mouth.

    And I don't know how to use HD5870 to know how games are shader bottlenecked since the proportion in which they have improved shader processing power it's the same than texture processing power, rasterizing operations processing power, and so.

    There are more things involved in the 3D rendering process apart from shaders and memory bandwidth.

    i would trust nvidia more than i trust you for the clockspeed.
    Yeah, no doubt. But I think you have misunderstood them when you have the idea that they are talking about a clock of 1500MHz for the GPU core.

  5. #5
    Xtreme Cruncher
    Join Date
    May 2009
    Location
    Bloomfield
    Posts
    1,968
    Quote Originally Posted by Farinorco View Post
    I didn't miss that factors. They simply don't take any part in anything that I've said. And when it take it, I have mentioned them and considered them. Take the "trouble" of reading my posts and trying to understand them before quoting me, please, to not put things in my mouth.

    And I don't know how to use HD5870 to know how games are shader bottlenecked since the proportion in which they have improved shader processing power it's the same than texture processing power, rasterizing operations processing power, and so.

    There are more things involved in the 3D rendering process apart from shaders and memory bandwidth.



    Yeah, no doubt. But I think you have misunderstood them when you have the idea that they are talking about a clock of 1500MHz for the GPU core.
    my reference to the 5870 was to show that rops are where they should be. too much and youre just wasting die space. they are running games at 7680x3200. the rop's were added to help texture filtering quality which wont double the performance in either gpu. if you dont believe me look at the ratio of shaders to rops over the past 5 years.

    this is the statement i was referring to:
    Consider that HD5870 is exactly double the HD4890 (+100% everything at the same clocks) except bandwidth (aprox. +30%) and it's far from double the real world performance (that's one of the most recent proves that doubling everything doesn't mean doubling real world performance), and NVIDIA is not even doubling processing units.

    i responded to this part of your statement about shader clocks and you somehow got the idea i was talking about core?
    And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.

  6. #6
    Xtreme Member
    Join Date
    Jul 2009
    Location
    Madrid (Spain)
    Posts
    352
    Quote Originally Posted by Chumbucket843 View Post
    my reference to the 5870 was to show that rops are where they should be. too much and youre just wasting die space. they are running games at 7680x3200. the rop's were added to help texture filtering quality which wont double the performance in either gpu. if you dont believe me look at the ratio of shaders to rops over the past 5 years.

    this is the statement i was referring to:



    i responded to this part of your statement about shader clocks and you somehow got the idea i was talking about core?
    Processing units. CUDA cores are processing units. Texture units are processing units. Raster Operation Processors are processing units. So no, they are not "doubling processing units".

    And regarding the clocks, obviously. If you understand it like shaders clock I don't know how it's an argument to say that they have doubled processing units power.

    EDIT: And from my part, discussion about what I've said or left to say is over. My (at the present time favourable, and I think not unrealistic) opinions about GT300 are pretty clear at posts in previous pages (some of them quoted on this one), even when some people is absolutely determined to misunderstand them.
    Last edited by Farinorco; 10-02-2009 at 02:26 PM.

  7. #7
    Xtreme Addict
    Join Date
    Apr 2004
    Posts
    1,640
    Quote Originally Posted by Farinorco View Post
    512 shaders is over double 240 (x2.13 to be exact). But 48 ROPs is not over double 32 (x1.5 to be exact). And 230 MB/s is not over double 141 (x1.63 to be exact). So overall, it's not over double the specs of the previous one. I don't think it's so hard to get what I've said there, and I don't get where I've said anything about CPs not being double (I think I have mentioned +113%). I would also like to know what are all those lot of factors that I've missed and what wrong information I've posted, based on what we know at the moment.

    And regarding clock speed, I would take it like talking about the shaders clock. I wouldn't expect much higher clocks than GTX285, if at all.
    Um, you don't have to double EVERYTHING to get doubled performance. More than anything this depends on the particular application you're running, and where the bottlenecks lie within it.

    If you look at a past example where performance WAS doubled, like the 8800GTX, let's compare that to the previous gen flagship, the 7900GTX. The 8800GTX had almost exactly twice the GFLOPs of the 7900GTX, even taking into account the nearly useless MUL op. The 8800GTX had 69% more memory bandwidth, and get this, only 33% more pixel fillrate, and 18% more bilinear texture fillrate.

    The GF100 is more of an improvement in raw specs over the GTX 285 than the 8800GTX was over the 7900GTX. So doubling performance is more than possible.
    DFI LANParty DK 790FX-B
    Phenom II X4 955 BE (1003GPMW) @ 3.8GHz (19x200) w/1.36v
    -cooling: Scythe Mugen 2 + AC MX-2
    XFX ATI Radeon HD 5870 1024MB
    8GB PC2-6400 G.Skill @ 800MHz (1:2) 5-5-5-15 w/1.8v
    Seagate 1TB 7200.11 Barracuda
    Corsair HX620W


    Support PC gaming. Don't pirate games.

  8. #8
    Xtreme Mentor
    Join Date
    Oct 2005
    Posts
    2,788
    Quote Originally Posted by Cybercat View Post
    Um, you don't have to double EVERYTHING to get doubled performance. More than anything this depends on the particular application you're running, and where the bottlenecks lie within it.
    Not always, for example the shaders are more efficient being MIMD/FMA.

    Also you have to keep in mind, while the ROPs and TMUs were doubled on RV870, ask yourself, doubled to what? 32/80 respectively. GT200 already has 32/80.
    Asus Rampage II Gene | Core i7 920 | 6*2GB Mushkin 998729 | BFG GTX280 OCX | Auzentech X-Fi Forte | Corsair VX550
    —Life is too short to be bound by the moral, ethical and legal constraints imposed on us by modern day society.

  9. #9
    Xtreme Member
    Join Date
    Jul 2009
    Location
    Madrid (Spain)
    Posts
    352
    Quote Originally Posted by Cybercat View Post
    Um, you don't have to double EVERYTHING to get doubled performance. More than anything this depends on the particular application you're running, and where the bottlenecks lie within it.

    If you look at a past example where performance WAS doubled, like the 8800GTX, let's compare that to the previous gen flagship, the 7900GTX. The 8800GTX had almost exactly twice the GFLOPs of the 7900GTX, even taking into account the nearly useless MUL op. The 8800GTX had 69% more memory bandwidth, and get this, only 33% more pixel fillrate, and 18% more bilinear texture fillrate.

    The GF100 is more of an improvement in raw specs over the GTX 285 than the 8800GTX was over the 7900GTX. So doubling performance is more than possible.
    Where in the post you're quoting I say that you have to double everything to double performance? I'm aswering a specific question.

    And you can't compare G80 with previous generation, as it's a completely different architecture. Starting by the unified shader processors (instead of units that only could calculate vertex or pixel shaders), with a completely different architecture, and the same for TMUs and ROPs.

    Again, I've never said that doubling is not possible (why is everybody putting that words in my mouth? It's at least the 3rd person who says that, and I'm starting to be tired or repeating it). You can read it yourself in my post quoted by Chumbucket843 (that I should add it's taken from a conversation including more posts before and after).

    I have only said that there is not a single evidence which grants that the GT300 is going to be more than twice the performance of GT200.

    But oh, well. If all of you are getting hurt by hearing it, I'll correct myself and let's finish with this: "GT300 is going to be obligatory at least 2x the performance of GTX285, and probably more". ¿Happy there?

    EDIT: I have edited the former paragraphs to give a much more accurate response.
    Last edited by Farinorco; 10-02-2009 at 02:02 PM.

  10. #10
    Xtreme Addict
    Join Date
    Apr 2004
    Posts
    1,640
    Quote Originally Posted by Farinorco View Post
    And you can't compare G80 with previous generation, as it's a completely different architecture. Starting by the unified shader processors (instead of units that only could calculate vertex or pixel shaders), with a completely different architecture, and the same for TMUs and ROPs.
    Of course it was a completely different architecture, as is the GF100. It may not be to the same extent as the G80 enjoyed, but it makes up for that by increasing raw specs more than the G80 did.

    It really doesn't bother me when people say the GF100 won't perform as well as such-and-such or whatever, because no one knows, and everyone's entitled to their opinion. The main thing that bothers me is how much importance you place in ROPs, TMUs and bandwidth, when those are insignificant factors in games that are GPU-limited. Granted, there aren't that many of those anymore, thanks to consoles and the perceived threat of piracy.
    DFI LANParty DK 790FX-B
    Phenom II X4 955 BE (1003GPMW) @ 3.8GHz (19x200) w/1.36v
    -cooling: Scythe Mugen 2 + AC MX-2
    XFX ATI Radeon HD 5870 1024MB
    8GB PC2-6400 G.Skill @ 800MHz (1:2) 5-5-5-15 w/1.8v
    Seagate 1TB 7200.11 Barracuda
    Corsair HX620W


    Support PC gaming. Don't pirate games.

  11. #11
    Xtreme Member
    Join Date
    Jul 2009
    Location
    Madrid (Spain)
    Posts
    352
    Quote Originally Posted by Cybercat View Post
    It really doesn't bother me when people say the GF100 won't perform as well as such-and-such or whatever, because no one knows, and everyone's entitled to their opinion. The main thing that bothers me is how much importance you place in ROPs, TMUs and bandwidth, when those are insignificant factors in games that are GPU-limited. Granted, there aren't that many of those anymore, thanks to consoles and the perceived threat of piracy.
    That's exactly what I was trying to say.

    I have never said "GF100 won't perform as well as papapa".

    Exactly that's what I'm talking about.

    Somebody said "GF100 is going to perform at least twice as well as" and I asked him "Why? What's the reason why do you think that? What info which we have now lead you to take that for granted?".

    And then some of you started quoting me puting words in my mouth.

    But oh, you know what? The guilt is all mine:

    I should have started to reply "I know it. I didn't say otherwise. Read it again" since the very beginning.

  12. #12
    I am Xtreme
    Join Date
    Dec 2008
    Location
    France
    Posts
    9,060
    Quote Originally Posted by Chumbucket843 View Post
    nvidia also said 1.5ghz is a conservative estimate for clockspeed.
    Do you honestly believe in 1.5Ghz GT300 on stock settings?
    Donate to XS forums
    Quote Originally Posted by jayhall0315 View Post
    If you are really extreme, you never let informed facts or the scientific method hold you back from your journey to the wrong answer.

  13. #13
    Xtreme Mentor
    Join Date
    Oct 2005
    Posts
    2,788
    Quote Originally Posted by zalbard View Post
    Do you honestly believe in 1.5Ghz GT300 on stock settings?
    For the shaders, not the core. Obviously the core won't be 1.5GHz.
    Asus Rampage II Gene | Core i7 920 | 6*2GB Mushkin 998729 | BFG GTX280 OCX | Auzentech X-Fi Forte | Corsair VX550
    —Life is too short to be bound by the moral, ethical and legal constraints imposed on us by modern day society.

Bookmarks

Bookmarks

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •