One thing is clear in all those reviews: AA scaling is so much better in the 4850 compared to RV670. The perfomance hit is on par with NV cards, which is nice.
One thing is clear in all those reviews: AA scaling is so much better in the 4850 compared to RV670. The perfomance hit is on par with NV cards, which is nice.
Friends shouldn't let friends use Windows 7 until Microsoft fixes Windows Explorer (link)
![]()
nah they wernt oced, at least to there test system specs:
but they used win vista for all testsTestsystem und Konfiguration
CPU: Intel Core 2 Duo E8500 @ 3.600 MHz (400x9)
Board: Asus P5N-D (Nforce-750i-SLI-Chipsatz)
RAM: 4x 1.024 MiByte DDR2-800 (5-5-5-15)
OS: Windows Vista 64 Bit samt SP1
Driver:
• Forceware 177.34 (HQ)
• Catalyst 8.5 respektive 8.6 bei HD 4850 (AI def.)
VGA:
• Geforce GTX 280, 1.024 MiB GDDR3, 602/1.296/1.107 MHz
• Geforce 9800 GX2, 2x 512 MiB GDDR3, 600/1.512/1.000 MHz
• Geforce 8800 Ultra, 768 MiB GDDR3, 612/1.512/1.080 MHz
• Geforce 9800 GTX, 512 MiB GDDR3, 675/1.674/1.100 MHz
• HD 4850, 512 MiB GDDR3, 625/993 MHz (inklusive Crossfire)
• HD 3870 X2, 2x 512 MiB GDDR3, 825/901 MHz
• HD 3870, 512 MiB GDDR4, 776/1.125 MHz
http://www.pcgameshardware.de/aid,64...=648091&page=1
I said the PCPER article, where they used the BFG 8800GT OCX and BFG 9800GTX and didnt state those two cards were both OC'd.
The PCGH article didn't use OC'd cards. But i'm still confused as to how the 4850 straight up beat the GTX280 while using the Qarls Texture Pack at 1920 x 1200. I thought cards with < 700MB choke with that pack, but the 4850 is eating a 1 GB card at 8xAA no less.
You are completely right...however I have been watching GDDR5 since quimonda first released specs(in such a way that I know that up pic there is from a quimonda document)...I fully understand it's possible(I was saying yes months ago, I had a big part in the shared framebuffer rumour, when ATI guys were saying no), but I question only whether the pcb design will affect overall cost as much as it should...given the intended prices for such cards. I see the same cost as making 512-bit membus pcb...which gives me a $575 retail price for 4870x2's...I am hoping for less cost..$499...
Again, I'm the ATI fanboi, but i like to keep it real. If they have pulled it off as I hinted like months ago, they've got nVidia under thier boot. If they don't have shared framebuffer, i'll still buy if i get more than 512mb of framebuffer for each gpu.
The PCGH review is the only inconsistent review out there. Yes, they could be the only review with the correct figures, but im gonna take the most likely possibility, there was a problem in their test setup. When you get the 4850 barely outperforming the hd 3870 and even falling behin in the higher res cod4 benchmark, it should tell you something. Am i a fanboy? I dont really care but I think the 9800gtx+ will do better than the hd4850 as the reviews show us, at normal resolutions of 1280*1024 and 1680*1050, the 9800gtx at the moment is pretty much neck and neck.
Yes there custom time demo might be different, but why should the relative positions be so significantly different? It would only affect absolute figures. Look at the pcgh graphs, the 4850 is absolutely awful in some of the games.
The HD3870x2 also does abnormally well as well. Maybe somethign to do with ATI drivers being better optimised in that particular scenario.The PCGH article didn't use OC'd cards. But i'm still confused as to how the 4850 straight up beat the GTX280 while using the Qarls Texture Pack at 1920 x 1200. I thought cards with < 700MB choke with that pack, but the 4850 is eating a 1 GB card at 8xAA no less.
Last edited by ghost101; 06-19-2008 at 10:15 AM.
Q9300 l 4GB DDR2 l HD 4850 l GA-X38-DQ6 l 2.5TB HD l VX550 l Dell S2409W l Vista X64
.
TJ07 | Corsair HX1000W | Gigabyte EX58 Extreme | i7 930 @ 4ghz | Ek Supreme | Thermochill PA 120.3 | Laing DDC 12v w/ mod plexi top | 3x2gb Corsair 1600mhz | GTX 680 | Raid 0 300gb Velociraptor x 2 | Razer Lachesis & Lycosa | Win7 HP x64 | fluffy dice.
I don't consider the competition. I have rigs from both makers, but really could care less about what nvidia offers in this respect. Two different companies with different goals leads to very different products, and different apps will be better on either system.
but playing the ATI fanboi, I won't talk nV stuff often.
August seems fair...if 4870 is really dealyed until july though...4807x2 may take longer as board partners are responsible for pcb design, no? they've not had the gpu's for that long yet...
Last edited by cadaveca; 06-19-2008 at 10:22 AM.
No, $499,-
For instance:
http://www.tgdaily.com/content/view/37093/135/When it will become available the 4870 X2 will hit the market for $499.
System Specs: * i7 2700K @ 4.8 Ghz * Zalman CPNS9900-A LED * Asus Maximus IV Extreme -Z * 16 GB Corsair Dominator GT CMT16GX3M4X2133C9 * Sapphire HD7970 crossfire * Creative X-Fi Titanium Fatality Pro [PCI-E] * Corsair AX 1200W * WDC WD1002FAEX + WDC WD1002FAEX * Optiarc AD 5240S * Dell U3010 @ 2560 x 1600 [DVI-D] * Steelseries 7G * Logitech G9 * Steelseries SX * Coolermaster Stacker STC T01 * Logitech Z-5500 * Sennheiser HD598 * Windows 7 Ultimate x64 SP1*
Wow, guys! These cards are really something! I love the peak power consumption. No need to upgrade my PSU!
For my part I know nothing with any certainty, but the sight of the stars makes me dream.
.
.
well,when there will be any 4870 for sale ?
Well, GDDR5 can use different trace lengths to memory chips, which I'm sure you know. If the shared frame buffer rumor is true I'm guessing that the trace lengths between a memory chip and the two RV770s will need to be equal, but even then board design should probably be less complex and have less layers if the memory placement is good.
STOP THE PRESS
TAKE A LOOK AT THIS
http://www.tweaktown.com/articles/14...nce/index.html
so a lot of review are pure sh_t -.-
there's a bug here, I suppose
Last edited by ORBR; 06-19-2008 at 10:52 AM.
Well, that certainly helps explain things!
System Specs: * i7 2700K @ 4.8 Ghz * Zalman CPNS9900-A LED * Asus Maximus IV Extreme -Z * 16 GB Corsair Dominator GT CMT16GX3M4X2133C9 * Sapphire HD7970 crossfire * Creative X-Fi Titanium Fatality Pro [PCI-E] * Corsair AX 1200W * WDC WD1002FAEX + WDC WD1002FAEX * Optiarc AD 5240S * Dell U3010 @ 2560 x 1600 [DVI-D] * Steelseries 7G * Logitech G9 * Steelseries SX * Coolermaster Stacker STC T01 * Logitech Z-5500 * Sennheiser HD598 * Windows 7 Ultimate x64 SP1*
Yep, you are think as I am...trace length from one gpu to mem to next gpu might have to stay the same, or one gpu will always get "priority". I do forsee a way around this, but it's not good for cooling, imho, so who knows. It's a techinical challenge that should be shouted from the mountaintops if they pull it off...as they should have patents for such that would prevent nVidia from ever pulling off the same thing...making ATI king of dual gpus.
But since R300 ATI gpu's have been capable of multi-rendering(renderbeast), so they have far more experience in this field, with products in the marketplace far longer than NV has...so I will put nothing past them.
But let me say this much...R600 was reviewed as having UVD...even some reviews posted results of UVD(which were completely faked)...but no card actually was capable. Today this review is still posted in this fashion...which makes me wonder about AMD's rumour control...everything out in the public domain right now is just rumour, IMHO, and ones I helped start aren't gonna get much attention from me.
No, I was just poniting out that thier info need to be taken with a grain of salt, as does all other info.
I mean really, perfect example is GTX260/280 release prices...they were supposed to be what? And are actually how much more? It's just speculation in that article, as where the specs they listed, the prices, and everything else. That article is far too old to have any real info other than stuff purposely leaked to find holes in NDAs.
They even hype split clock domains...
"Our sources"....in other words, the info was not from ATI/AMD.
Last edited by cadaveca; 06-19-2008 at 10:55 AM.
System Specs: * i7 2700K @ 4.8 Ghz * Zalman CPNS9900-A LED * Asus Maximus IV Extreme -Z * 16 GB Corsair Dominator GT CMT16GX3M4X2133C9 * Sapphire HD7970 crossfire * Creative X-Fi Titanium Fatality Pro [PCI-E] * Corsair AX 1200W * WDC WD1002FAEX + WDC WD1002FAEX * Optiarc AD 5240S * Dell U3010 @ 2560 x 1600 [DVI-D] * Steelseries 7G * Logitech G9 * Steelseries SX * Coolermaster Stacker STC T01 * Logitech Z-5500 * Sennheiser HD598 * Windows 7 Ultimate x64 SP1*
WTF?! Why word it like that? How about something like: "The crossfire mode on the P45 seems to have some kind of problem". Oh wait. Is it because it would have negative implications for AMD and that might not be great for his shares? If he wanted to hide a crossfire problem with the P35/P45 chipset he should have just left that part out.An intensive game like Crysis sees the cards get it at all resolutions, and quite significantly at that. The X48 is really able to stand out when compared to the P45.
Hmm P45 is PCIe v2.0 2x 8 which is equivalent in bandwidth to 2x 16 PCIe v1.1. Its interesting if these cards actually require the additional bandwidth of full PCIe v2.0 16x. If this is truly the case and not merely a driver deficiency, I'm glad I didn't wait for a P45 and stuck with x48. 4870/X2s in crossfire will really hate P45 if this is the case.
@ghost
A single 4850 will stuff run full speed in 1 slot so it won't be bottlenecked.
Feedanator 7.0
CASE:R5|PSU:850G2|CPU:i7 6850K|MB:x99 Ultra|RAM:8x4 2666|GPU:980TI|SSD:BPX256/Evo500|SOUND:2i4/HS8
LCD:XB271HU|OS:Win10|INPUT:G900/K70 |HS/F:H115i
@cadaveca: I was just in the shower when I had a "Eureka!" moment.
Wait for it....
Wait for it....
Who said that both rv770s need to write to the framebuffer?![]()
Intel i7 920 C0 @ 3.67GHz
ASUS 6T Deluxe
Powercolor 7970 @ 1050/1475
12GB GSkill Ripjaws
Antec 850W TruePower Quattro
50" Full HD PDP
Red Cosmos 1000
Bookmarks