What AGESA does the 9905 ASUS BIOS use?
Printable View
What AGESA does the 9905 ASUS BIOS use?
Don't believe in miracles like 50-100% boost by only flashing new BIOS.
Testing games at such low resolution is complete fail and I'm not listening to the BS that it puts the emphasis on the cpu. Benchmarks are supposed to be indicative of real life performance and nobody uses fast graphics cards and cpu's at low resolution. Furthermore, by not benching at enthusiast-level resolutions you run the risk of failing to fully stress the cpu's in that role and will give a false impression as to how much better or worse a certain cpu is at gaming. There's no excuse for it.
so you think they way to test CPU performance in games is by limiting the overall performance by the GPU? like AMD did on their slides (HD6870 1080p)?
I think there is value in both methods, no one is hiding anything, clearly the FX is more than good enough to play these games, as a Phenom II X4 is or i3 2100 are,
but you need to consider that they probably had limited time for testing, and that maybe a dual gpu configuration can have the same performance difference as this on a higher res, or a new GPU next year and so on...
ideally we could have more test situations varying the resolution, and perhaps more well known CPU intensive games, but anyway, I think by now everybody knows that gaming is mostly limited by GPU if you have a good enough CPU,
but still is a valid test, showing what the CPUs can achieve under these conditions,
With todays high end video cards people are more often cpu limited than they think.
People say it's supposed to be a cpu benchmark but there are plenty of those in reviews. It's supposed to be a cpu benchmark in *gaming*. That means you don't negate the graphics card.
It is quite likely that SB will score 50% higher fps than Bulldozer in a game like Dirt 3 at low res, but lose by 5-10% at 1080p. What is the better gaming cpu?
By your logic you would test Crysis at 1080p , and conclude that Phenom x4 970 is only marginally slower then I7 sandy , when in reality in other game where CPU is important Phenom is significantly slower.
Thats why you bench at lower res to eliminate GPU dependency and measure CPU potential.
Attachment 121006
Looks like the retail pricing turns out to be the best avg benchmark for cpu performance.
Would have been nice to see some 2nd pass x264 results.
I was going to build a BD rig but I'll probably wait and see what PD brings to the table at this point.
Both Intel and AMD are going to need to deliver something more compelling, I'm just not feeling the excitement to upgrade.
Finally legit preview of bulldozer! Can't say the same for windows 7.
:P
Can't wait, 3 more days to see more!
:lsfight:
http://lab501.ro/wp-content/uploads/...0/spi_5066.jpg
If AMD are calling people tomorrow they are probably trying to ensure this low-res gaming skullduggery isn't happening, which imo is their right as they are simply ensuring that the gaming benchmarks are representative of the cpu's while gaming. Also, if BD is strong in minimum framerates then AMD would want that message to be getting across, which is also well within their rights.
Any site using low res gaming benchmarks has an axe to grind with AMD, or an intel shill. Why do you think the unnamed blogger benchmarked Unigine at low res to "prove" his point? I've never seen anyone else do that, like ever. But even at that BD was massively better in minimum framerates.
http://3.bp.blogspot.com/-pyUWXUhTaB...ld_unigine.jpg
You can bench at HI res but you have to find cpu dependent point.It is not easy as you think.
compare with that
http://img718.imageshack.us/img718/2159/36073117.png
@down
maximum settings from game ultra/high GTX580
http://img717.imageshack.us/img717/8346/farg.png
Your pictures show the use of built-in benchmark that is GPU test.
have you even read any recent reviews?
Most bigger review sites state in there test methods why they use lower resolutions for games and often also provide benches in full resolution and also mention that this is not real life scenario..
And here you have a perfect example why testing with realworld resolutions doesn't tells you anything about the cpu at all... you only test the gpu:
Attachment 121007Attachment 121008Attachment 121009
Granted this test is older (march 2010 from when the X6 was released and tested with a HD5870) you would think if you look at "real world resulsts" all cpus are pretty much even, but what happens when you plug in a dual 580 instead a 5870... you suddenly run into the gpu limitation again and your "even" cpu gets smacked hard.
Both test have validity, its just that recently people negate low res test for whatever reason...
Judging by Combined average intel system will have higher minimum on average ,without fraps FPS history graph . those minimum and maximum mean nothing ,all it takes is one split second dib and benchmarking program will register it as minimum , rest of the benchmark sandy will have higher fps then buldozer.
What exactly is the point in benchmarking at such low resolutions? RE2 and HAWX are 50% faster on SB than BD in these tests. Is that representative of the other results? Is that representative of a gaming benchmark?
Or is it simply representative of a useless benchmark of a game at low resolution while giving no indication of how the chips perform during actual gaming?
Do you think that might be because nobody games at 800x600 on enthusiast level hardware?Quote:
Both test have validity, its just that recently people negate low res test for whatever reason...
It's not to directly assess real gaming performance as it is to showcase and isolate the cpu's ability to pump fps rather than showing how well the video card performs.
Turning up the resolutions shifts the bottleneck to the gpu, if you want to bench the cpu you need to isolate it from gpu limitations.
Most review sites do a range of resolutions and settings anymore and when NDA is over you won't have anything to complain about.
As we are here, CB has a nice articel to this..
http://www.computerbase.de/artikel/g...frameverlaeufe
If you jack up resolution and AA/AF its pretty much irrelevant what cpu you use. It all depends on the GPU... on a HD6850 even a PII X2 produces nearly the same framerates as a 2600K@4,5ghz...
So why not benchmark it to see if that's true or not?
If SB is 50% faster than BD then why is it not 50% faster in anything else (except superpi :p)? Why only very low resolution gaming? It is obvious to me that low res gaming is not indicative of actual cpu performance. It's an abberation.
Again, Anandtech's cpu benches are :banana::banana::banana::banana:ty.
I HAVE NEVER PLAYED CRYSIS on MEDIUM SETTINGS like they use in that :banana::banana::banana::banana:ty review. THEY ARE ALSO USING A NOW ANCIENT GTX280 in that review. To top it all off they also don't even mention minimum framerates.
ANANDTECH'S CPU BENCHMARKS ARE USELESS.
They might as well just post a 3d mark bench.
It is indicative of the cpu's ability to generate game & frame data for the gpu. If you use high or low resolution the cpu still generates the same game and frame data but at high resolution the cpu is limited by the gpu's speed for rendering the frames which means all you see is more of a gpu limitation than a cpu limitation at higher resolutions.
In the end the cpu that can pump the highest fps period, raw cpu speed unhindered, has more processing power but that doesn't mean it translates into raw performance for other workloads.
Here is what I saw in actual gameplay with the settings that I actually use not some :banana::banana::banana::banana:ty canned bench like what Hornet just posted like thats supposed mean something to me.
This is what I would expect from a "professional review" I should have left out the maximum framerate though.
This is an i5 750 at 4ghz and an i5 2500k at 4.7ghz.
http://i308.photobucket.com/albums/k...0vsi52500k.png
Testing prefetching on the cpu.
Most games have one heavy render thread, only BF3 and some other are using multithreaded rendering (a dx11 feature). Most games are using half or less than half of total CPU power. For most games data is rather static, maps etc are data loaded from disk and not modified during gameplay.
The main work will be calculating dots sent to the gpu. Long trains of data is calculated and sent to the gpu in chunks with commands informing the gpu how to work with those dots.
The amount of work the gpu needs to do for each frame is much less on low resolutions. It could be faster than the cpu and then it starts to wait for the cpu to process all dots.
Prefetching will increase performance a lot working with long trains of data. The cpus guess where the next data will be and fetches that while current data is being processed.
i5/i7 prefetches data to L1 and L2 cache, phenom is only using the L1 with some small prefetching functionality.
This is the reason why i5/i7 can produce much more fps on low resolutions. It's about prefetching and a fast cache.
If the game are using more threads and maps are dynamic and complex. Data isn't at all that predictable and prefething isn't as important any more.
Yes cause the custom benchmarks of CB (that even have fram graphs) are so canned... :rolleyes: