your right that consoles have similar situations, and with the ideas expressed in this thread it sounds like people are scared the PCs will follow the same path.
with consoles we do see games that work for both, and games that are exclusive, and the funny part is, the kind of uprising we see when one company who use to be exclusive, starts to design multi platform games (SquareEnix comes to mind). for games that work on both, differences are usually negligible, but we do have cases where there is a difference (does anyone know of any titles that are on PS3/360 where one was massively favored over the other? i dont keep up with consoles enough)
i still believe people are too scared a company will let themselves suffer profit loses to favor a company, and if they mess up big enough, reviews will show it, and sales will be hurt big time. look what happened on here over ingame AA, when you still have a work around by enabling it in CCC. (what kind of perf loss was noticed between the two methods?)
On multi platform console games the most recent game that specifically favored one platform was Batman: AA(ironic, isn't it) which lets you play as the Joker for specific challenges on the PS3 as DLC which won't be released on the x360.
UE3 engine performs really bad with AA (at least with nvidia, dunno about ati) so that's the reason I think AA was not added as ingame option for starters. 200 series handles it a lot better than G92 tho, UT3 got even somewhat unplayable with AA4x from being more than perfectly playable with AA off with a 8800GT, there's a lot less performance hit with a GTX 260 but it's still huge compared to other games.
Last edited by RPGWiZaRD; 09-29-2009 at 02:45 PM.
Intel? Core i5-4670K @ 4.3 GHz | ASRock Extreme6 Z87 | G.Skill Sniper 2x8GB @ DDR4-1866 CL9 | Gigabyte GTX 970 OC Windforce 3x | Super Flower Titanium 1000W | ViewSonic VX2268wm 120Hz LCD | Phanteks PH-TC14PE | Logitech MX-518 | Win 7 x64 Professional | Samsung 850 EVO & 840 Pro SSDs
If all people would share opinions in an objective manner, the world would be a friendlier place
yah i've seen the tv commercial for that "play as the joker exclusively only on PS3"
Wouldn't it be neat if you could play as the joker exclusively on a gtx260?
[SIGPIC][/SIGPIC]Bring... bring the amber lamps.
I agree with Nvidia if ATI want it why don't they spend the money on developing teams?
another way nvidia screwin ati buyers
the good thing is no need for crappy chipsets no more
i buy only ati now
i dont give a crap for physx either
nice game rocks on ati too
maximus IV extreme gtx580
gigabut p67-ud7
p67 sabertooth
2500k+2600k
antec 1200watt
EVGA classified 760
920 Batch# 3849B018 4.985ghz@1.52v gtx285 ftw sli
OCZ3RPR1866LV6GK hypers
dfi ut p35 rampage extreme
gigabut p35c-ds3r bios suks
gigabut x38-d6q dead thank god
ballistix 8500 1240mhz@2.02v
Joker playable only on nvidia hardware !!! : 'why only nvidia ? cause ati didn't support us when we created joker scenario'.... I really expect such things from now on
5930k, R5E, samsung 8GBx4 d-die, vega 56, wd gold 8TB, wd 4TB red, 2TB raid1 wd blue 5400
samsung 840 evo 500GB, HP EX 1TB NVME , CM690II, swiftech h220, corsair 750hxi
ATI does spend money on developing teams, as NVIDIA. Not in every developing team, of course (as NVIDIA). The difference is that until the moment (that we now) neither company had used their support to "lock" the features that they are supporting to their own hw.
For example, let me insist. ATI (sorry, AMD) is spending money on Codemasters to implement DX11 features (as the Batman AA is a DX9/10 feature). I wouldn't like that they include a check to not allow NVIDIA cards to run those effects (when NVIDIA cards have DX11 support, of course).
AMD is spending money on Havok to develope their OpenCL GPGPU acceleration. I wouldn't like that they include a check to not allow NVIDIA cards to use OpenCL GPGPU acceleration with Havok.
AMD is spending money on Bullet Physics to develope their OpenCL GPGPU acceleration. See above.
ATI (AMD) is spending money on those things, and that doesn't mean they are going to prevent NVIDIA users to enable those effects. Supporting developers to ensure their games are run bug free or well optimized (at least for their own hw), or to make easier for them to include features that you are interested for whatever reason they have, is all good. That's what "supporting developers" is meant to be. Not the case discussed here.
I don't know that I can say for sure that nVidia did anything wrong without having been at their discussions, but there is someone to blame: the game developer. Sure ATI could have sent in someone to work with them, but if a game developer can't test a feature like AA on ATI hardware, they shouldn't be making games in the first place. ATi should be able to come in to help *optimize* it, but in terms of ensuring that features as basic as AA work.. that should be something they can do.
Dual CCIE (Route\Switch and Security) at your disposal. Have a Cisco-related or other network question? My PM box is always open.
Xtreme Network:
- Cisco 3560X-24P PoE Switch
- Cisco ASA 5505 Firewall
- Cisco 4402 Wireless LAN Controller
- Cisco 3502i Access Point
I think that Nvidia should keep to their own, and simply get the hell out of the software delvelopment - PERIOD! Because this is the first and crucial step against open standards, which could end up in a scenario where one must have a cirtain GFX to be able to enjoy a cirtain game.
If you ask me, it should be illegal to work closely so together as Eidos and Nvidia have done in this case, without explicitly "warning" buyer's of both game and GFX, that this partnership has been going on behind the curtans. If not, then it's only a matter of time till the before mentioned scenario with x gfx to play x game becomes a reality.
So yeah, give Nvidia the finger, don't buy the game or any of the numerus Batman-theme'd or bundled Nvidia cards out there, and move on.
Last edited by Mads321; 09-29-2009 at 03:19 PM.
Yup.
Nvidia does a lot of nasty crap, but this doesn't sound like a case of it.
Although the way they're trying to play the high and mighty card is kind of pathetic, but whatever, that's PR for you.
Developers shouldn't need gpu manufacturer's to hold their 's. Are the days of devs taking any form of responsibility for their crappy games dead already?
how can u say that this whole game dosnt reek of the way its ment to be played, there is no AA for non NV, then in the demo physX worked on the low lvl with no paper on the cpu but when its retail then physX has an ass load of paper eveyware making the cpu not work. u say that both of those were in the devs best interest to do
5930k, R5E, samsung 8GBx4 d-die, vega 56, wd gold 8TB, wd 4TB red, 2TB raid1 wd blue 5400
samsung 840 evo 500GB, HP EX 1TB NVME , CM690II, swiftech h220, corsair 750hxi
Well PhysX is just pure unadulterated crap imo. I'm just past the point of caring about it, Nvidia knows it's a scam, we all know it's a scam. Just that there's nothing that can be done about it.
I just don't see Nvidia telling the devs, "btw you'd better not enable AA for ATI or we'll stop giving you checks."
That would cause such a *censor*storm, and for what? So they could have "exclusive" rights to AA in Batman? All my friends who play PC games daily don't even understand what AA is, they're not going to care either way franky. And I suspect that's pretty representative of the general population.
Seems like the simplest explanation is Nvidia made sure AA worked on their hardware so the devs enabled it. ATI didn't (to no fault of their own, they shouldn't have to babysit the developers), so they devs decided to be lazy buggers and disable it by default.
Could be wrong, but I just don't see what Nvidia could gain by this.
Last edited by Sly Fox; 09-29-2009 at 04:05 PM.
This is what happens when you use deferred rendering. Same thing with STALKER, AA is dodgy.
Well, that's half the truth. What happens if you could cover you'r potential losses (let's face it, ppl with ATI GFX will still buy this game), with a great deal on getting ur game bundled with numerous GFX
Do you honestly think the batman game is in there curtesy of the develpor, or might Nvidia have bought a load of games from Eidos, in favor for something else?
I don't know what you make of that, but the only thing i see is:
Last edited by Mads321; 09-29-2009 at 07:34 PM.
ha ha i got the game code free,today i installed the game played for 1hour and it thuoght jeez you can have it,then i uninstall game nobodys missing anything.
todays games are so god awful its enuogh to just get a console and dont buy into this crap,and for me to say that is very bad hating consoles.
nobody wants a game that runs 150fps+ plays well+ great online play but just looks "good" so they got to sell it to you good.
_________________
No game is awful enough to push me to consoles when console games are the same but 2x the price.
Is there a filter for posts from people who don't know jack about game development and therefore shouldn't comment on these topics? This thread would be about 6 posts long.
Agreed
The game is ... I don't see why you guys are kicking up such a fuss about it. The devs went around it the wrong way, nVidia saw an opportunity and made the most of what they could (which is what any company would do) and ATi ended up getting the short end of the stick. But tbh, this game is such a god awful load of crap, I don't see why nVidia would even bother with it.
What I'd like to see is ATi stamped all over Modern Warfare 2 though
Unreal 3 can support AA natively if implemented. (I've seen it many times, Last example: Mirror's Edge)
Bookmarks