maybe something that could calculate the amount of milliseconds (?) between each frame or something?
maybe something that could calculate the amount of milliseconds (?) between each frame or something?
LEO!!!! amd phenom II x6 1100T | gigabyte 990fxa-ud3 . . 2x2gb g.skill 2133c8 | 128gb g.skill falcon ssd sapphire ati 5850 | x-fi xtrememusic. . . samsung f4 2tb | samsung dvdrw . . corsair tx850w | windows 7 64-bit. ddc3.25 xspc restop | ek ltx | mc-tdx | BIP . . lycosa-g9-z2300 | 26" 1920x1200 lcd .
go into fraps and tell it to log frame times. this is what you get.
AMD 1090T@4.0ghz
Enzotech sapphire/Mo-Ra extreme rad
Asus Crosshair IV Formula
ht 2400mhz / nb 2400mhz
12gb Gskill 1300mhz
HIS HD5970
Enermax Evo Galaxy 1250
case: XCLIO A380PLUS-BK
4.61ghz water
4.5ghz superpi 1M 15.585
http://i271.photobucket.com/albums/j...rpi4500mhz.jpg
25,396 06 Phenom 965@ 4.4ghz HIS 5970@960/1260
21,893 Vantage, Phenom 965 4.2ghz HIS 5970 @960/1260
Phenom 2 125w 965 test results
http://futuremark.yougamers.com/foru...d.php?t=117414
Phenom 2 140w 965 test results
http://futuremark.yougamers.com/foru...d.php?t=109214
Phenom 2 AM2+ 940 cold air results
http://futuremark.yougamers.com/foru...ad.php?t=97430
If I dont get every single drop out of my cpu I feel like someone is stealing from me
LEO!!!! amd phenom II x6 1100T | gigabyte 990fxa-ud3 . . 2x2gb g.skill 2133c8 | 128gb g.skill falcon ssd sapphire ati 5850 | x-fi xtrememusic. . . samsung f4 2tb | samsung dvdrw . . corsair tx850w | windows 7 64-bit. ddc3.25 xspc restop | ek ltx | mc-tdx | BIP . . lycosa-g9-z2300 | 26" 1920x1200 lcd .
I'm not a pro with excel but I have enough know-how to make it do calculations and some charts and stuff, if someone wanted to run these tests with fraps and send me the data from fraps I'd be glad to do some data analysis of it. Maybe fraps data from Chew's testing?
The Cardboard Master Crunch with us, the XS WCG team
Intel Core i7 2600k @ 4.5GHz, 16GB DDR3-1600, Radeon 7950 @ 1000/1250, Win 10 Pro x64
If you knew anything about benchmarking then you would know lower resolution tests use the CPU more and the higher resolutions shift to the GPU.
Many CPU reviews use lower resolutions benchmarks to show this.
Meaning when you want to show CPU gains you use lower resolution benches.
CPU: Intel i5-3570K @ 4.2ghz (1.064V)
GPU: SLI ASUS GTX 660 Ti DCII 2GB @ 1215/7012
LCD: BenQ XL2420TE (144Hz)
Mobo: ASRock Z77 Extreme6
Sound: SoundBlaster ZXR + Yamaha RX-V863 (LPCM) + Polk Audio Monitor Series II Speakers
RAM: G.SKILL Sniper Series DDR3 2133 4x4GB
Storage: Samsung 840 Pro 128GB + 1TB Seagate Barracuda +1TB WD Black Caviar
PSU: Corsair HX 750W 80+ Silver (62A)
UPS: Cyberpower CP1200AVR (720W)
OS: Microsoft Windows 7 Ultimate x64
Cooling: Corsair 650D + TT Water2.0 Pro + 2x Silverstone FM121
i agree, the guy you're quoting has nothing to do with this forum, this is not about who gets higher frame rates at what resolution, it's about:
Hey, this CPU gets 90fps average in this game, but goes all the way down to 30fps every now and then
While this CPU gets 75fps average, but only goes down to 55fps every now and then
someone wanted the definition of smoothness, there's your smoothness, the difference between lows and highs on many occasions, i mean really, who cares if you get 200fps if every 10 seconds it goes down to 1fps, i wouldn't want to play like that
http://forums.overclockers.co.uk/sho...php?t=17896754
Maybe someone can ask the Author for that program as it will work out the avg deviation for you from fraps.
The DL links are dead.
New link http://www.mediafire.com/download.php?2moyugibhez
there where already many threads about that topic which came to similar ideas. When in fact the only thing that influences the "smoothness" factor is the difference between max/min fps, then there is a easy fix for the intel system -> limit max. fps or enable vsync.
If that hypothesis would turn out to be true, a intel rig would be sometimes even more smoother cause of the often higher min fps rate then a amd rig or the other way riudn if the amd rig has a higher min fps and the max fps is limited.
Imho its the "smoothness" debate is the same thing as the "they eye cant see more then 30fps" debate. It is based to much on the individual to make a general statment, just like people say they can see the difference between 60fps and 100fps. Maybe some individual can but the majority can't, its the same with audiophiles, play joeavarage a 192kbit/s encoded mp3 and a flac file on a highend headphones, they won't notice the difference, hell maybe they would even say the mp3 sounded better.
well just like said above, there are often cases where the min fps is higher on the intel rig -> enable vsync, limit max. fps and you have the same smoothness as the amd rig, but your artificially limiting your gfx card of showing what power it has.
But with same/lower min fps for AMD systems this doesn't clarify anything. It's more like this:
System A has avg fps of 120 but has min fps at 40 which happens every five seconds.
System B has avg fps 100, has min fps of 35 but this happens every 40 seconds.
System B would probably be a more fluid experience than A. The main issue is, as maany has already stated; fps stability. If a system fluctuates between 50 and 150 fps dropping to 35 sometimes, averaging 120, it will probably be considered more stuttering than a system fluctuating between 50 and 70 fps, dropping occasionally to 35, avg fps at 55. Microstuttering should be a similar thing.
afaik JumpingJack already did show framrate over time with a QX9650 and a P1 9850 on quadfire.
edit:
just look at that for example:
Company of Heros
Racedriver Grind (done on a 8800 GTX)
also this smashes the theory that systems behave differently form each other, if a system gets a drop every 30secs the other system also will get a drop every 30sces, when the same situation is played/displayed.
And if you knew anything about benchmarking you would know that using results from a different setup with different settings at a different resolution is totally useless here.
The tests AT did wasnt benchmarking they were actually playing the games using real world resolutions.
You cant say a game felt smoother by watching a benchmark
Last edited by Titan7171; 02-04-2009 at 05:35 AM.
AMD 1090T@4.0ghz
Enzotech sapphire/Mo-Ra extreme rad
Asus Crosshair IV Formula
ht 2400mhz / nb 2400mhz
12gb Gskill 1300mhz
HIS HD5970
Enermax Evo Galaxy 1250
case: XCLIO A380PLUS-BK
4.61ghz water
4.5ghz superpi 1M 15.585
http://i271.photobucket.com/albums/j...rpi4500mhz.jpg
25,396 06 Phenom 965@ 4.4ghz HIS 5970@960/1260
21,893 Vantage, Phenom 965 4.2ghz HIS 5970 @960/1260
Phenom 2 125w 965 test results
http://futuremark.yougamers.com/foru...d.php?t=117414
Phenom 2 140w 965 test results
http://futuremark.yougamers.com/foru...d.php?t=109214
Phenom 2 AM2+ 940 cold air results
http://futuremark.yougamers.com/foru...ad.php?t=97430
If I dont get every single drop out of my cpu I feel like someone is stealing from me
The Cardboard Master Crunch with us, the XS WCG team
Intel Core i7 2600k @ 4.5GHz, 16GB DDR3-1600, Radeon 7950 @ 1000/1250, Win 10 Pro x64
TJ07 | Corsair HX1000W | Gigabyte EX58 Extreme | i7 930 @ 4ghz | Ek Supreme | Thermochill PA 120.3 | Laing DDC 12v w/ mod plexi top | 3x2gb Corsair 1600mhz | GTX 680 | Raid 0 300gb Velociraptor x 2 | Razer Lachesis & Lycosa | Win7 HP x64 | fluffy dice.
Very interesting graphs, and as a general statement the AMD systems do have less framerate variance from a glance.
I wouldn't necessarily say those graphs "smash" anything though; we are only seeing 1/40th of the available data. Each connected data point is avg fps per second, so any hiccups would be included in that average. we would need to see a non-smoothed raw data graph at least, but a histogram of frame render delay would be absolutely ideal.
Asus G73- i7-740QM, Mobility 5870, 6Gb DDR3-1333, OCZ Vertex II 90Gb
You need to calculate on more detailed level to understand how smooth the game is. The graphic card and CPU works asyncronosly. If you calculate based on when picture is drawn from the GPU this isn't the same as when the CPU calculates points for the picture. The GPU is always behind the CPU. The GPU will also smear time spans between when the CPU calculate the picture.
To get a better value you need to time the EndScene API, when they are called if we are talking about Direct3D.
How the game feels depends on when the picture is calculated, not when it is drawn on the screen. Also the mouse is important, reading mouse data needs to be very exact.
Last edited by gosh; 02-04-2009 at 08:21 AM.
You know Guys Graph & Charts do not prove everything![]()
SuperMicro X8SAX
Xeon 5620
12GB - Crucial ECC DDR3 1333
Intel 520 180GB Cherryville
Areca 1231ML ~ 2~ 250GB Seagate ES.2 ~ Raid 0 ~ 4~ Hitachi 5K3000 2TB ~ Raid 6 ~
the COH chart from hornet actually 'proves' this thread's point.
I think the word instead of smooth might also be "Stable framerates" as the drops and speedups in fps for the Q9550 are more noticeable while playing and thus not so smooth.
obiously vert. sync will help but then we're doing apples vs oranges in this thread.
Current:
24/7 Torrent & Internet Box: 0.975V -0.800V. Brisbane @2Ghz w. CrystalnQ // Biostar TA690G // 2x 1Gb Geil 800@4-4-4 // IDE DVDRW // 250Gb WD S16 Sata // Coolermaster iGreen 500W & 2x 120mm Antec case fans.
Main PC: 3Ghz BE 5000+ @1.45Vcore 273x11 // Asus M32R2-MVP // eVGA 8800GTS SC 320mb // 4x 1GB DDR1066 OCZ D9GKX // several HDDs & optical // Soprano DX & Zalman cnps9500nt // FSP FX700-GLN // Acer AL2216Wbd 22"
My 6600GT AGP Ghetto PrescottCooler Mod
old Main rig:
Desktop Barton 2600+ @202x11 with Lapped CM Jet7 - GA-7NNXP - 2x512MB 2.5-3-3-11 Value Kingstons - PNY 6600GT AGP @ 605/1135 - 3x120GB & 1x200GB HDDs (WDC/Maxtor/Seagate) - 2xopticals burners (DVD R/RW 8-4-32 +/- Liteons) - iBest "crap" 600Watt Dual fan - Logitech MX Cordless Duo - Logitec Quickcam Orbit
Very old Server:
Intel P4 willamette 1.4Ghz - 2x128MB & 2x64MB PC800 RDRAM - Intel D850GB Mobo - 10GB & 80GB seagate HDD's - jetway 9600SE - Win2003 Server
I think that the chart you displayed with the frames and the time would be much more useful than the "max/min" idea or any graph of FPS over time.
Using the data you displayed you could make a graph of the time DELTA (change in time) between frames. (So it would show the time spent between each frame.)
Optimally this graph should be as flat as possible. Although realistically it won't be due to the operating system, rendering different things in the game, communication between CPU cores in a multi-threaded games, cache contention in the CPU, and a zillion other factors. (i.e., some factors related to the CPU, some are not. A good tester would attempt to eliminate most factors except for the CPU.)
BUT if you look at a graph of the delta vs the frame generally the system with the flatter graph would probably be the "smoother" system. Or if one system had a major dip in this graph on occasion and the other system did not. Of course this would have to be something that can be reproduced; a good tester would throw out weird anomalies that only happen one time and not use them.
NOTE: I know this image I am showing below is NOT REALLY RELATED. I am throwing it here into this post as an example of something similar to what we might expect to see by creating the type of graph I propose. POP QUIZ: If this WAS a graph of the time delta versus frame... which system would you rather have? (And amusingly the system with the higher min and max in this non-related example is also not the "better" system.)
![]()
Last edited by keithlm; 02-04-2009 at 10:00 AM.
FX-8350, Powercolor ATI R9 290X LCS, OCZ Vertex 4, Crosshair V Forumula-Z, AMD Radeon DDR3-2133 2x8Gb, Corsair HX1000W, Thermaltake Xaser VI, Xonar D2X, Water Cooling 140.3
<sarcasm mode ON>
We're all just shocked that you, of all people, find this to be a non-issue.
<sarcasm mode OFF>
This is most especially true since we almost have a constructive conversation going on. That can NOT be allowed. You'll need to run to the other forum subsection and gather some of your cohorts.
Actually I'm sure I speak for many people when I say your comment is worthy of a very large yawn.
And changing the vsync won't actually change anything. It might hide the issue from any method that could possibly be used to determine the difference. (But then hiding the issue is probably acceptable to some people.)
Last edited by keithlm; 02-04-2009 at 10:07 AM.
FX-8350, Powercolor ATI R9 290X LCS, OCZ Vertex 4, Crosshair V Forumula-Z, AMD Radeon DDR3-2133 2x8Gb, Corsair HX1000W, Thermaltake Xaser VI, Xonar D2X, Water Cooling 140.3
Bookmarks