Comments on benchmarking, and a request (typically longwinded post from Robo)

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
Lately, I have been doing a lot of thinking (no, your car isn't burning oil, smarta$$!)

The way I see it, there are 3 main problems with the benchmark-centric graphics card reviews of today:

1) There is usually no way to measure "minimum" framerate

2) There isn't a standardized way to measure "average framerate under heavy stress"

3) Super-high framerates in non-detailed areas are given undue weight in a benchmark "fps average" score, when they don't really add anything to your gaming experience

As far as framerate is concerned, when does framerate matter while gaming? (gaming, not benchmarking)

It matters when it gets LOW. When we are maintaining a steady 100, 80 or maybe even 60 fps, we usually won't notice what our framerate is (although 60 is pushing it, IMHO). However, when we drop to 30 or 20 (blech!!!), our framerate becomes VERY VERY noticeable, kinda like a 3-year old throwing a tantrum. It gets our attention in a NEGATIVE way. Hitting 150 fps in a low detail scene does nothing for my gaming experience, but maintaining 45-50 fps in the heaviest of situations certainly does.

Are today's benchmarks really telling us how our systems will perform under heavy stress? I say "no" for the above 3 mentioned reasons.

re: # 1) - Not too many games today have a "minimum" framerate measure. Evolva and UT come to mind presently, but, I don't know of any OGL games *at all* that do. We haven't, in the past, had ANY that gave us an idea of how our systems performed when we needed them to, so we "resorted" to #2...

re: # 2) "Killer timedemos" - crusher.dm2 and massive1.dm2 used to be "the standard" for benchmarking. Why? Because if your graphics card could perform well under those stresses, then it could handle ANYTHING any OGL game (and most games, in general) would throw at you.

No one used q2demo1 as a benchmark, because it was "too weak" and didn't really represent the true strength of the card. Yet today, we are using "q2demo1-type" benchmarks for means of comparison. q3's demo001 and MDK2's default test really don't tell us a whole lot, do they? Having an average of 90 fps is nice, but that doesn't tell us that it includes those few seconds you jumped to 200+ fps in a room with no enemies (which didn't help you), it doesn't tell us about the time you dropped to 20 after gibbing that guy at point blank range (and then got fragged yourself cuz your system slowed to a crawl), nor does it tell you how your framerate will suffer in a close RA3 match in the castle map with a boatload of h0m0s trying to rocketjump off your head.

re: # 3) - When we do our benchmarking, today's "big 3" (5500, Radeon and GTS) are fast enough to break 100 fps quite easily in the vast majority of games at most resolutions and 125 fps in many situations. Now, I will argue tooth and nail against someone who says there's no difference between 60 fps and 30fps. They're grossly incorrect. I can tell quite easily the difference between 60 fps and 100 fps even. But past 100 fps, I start to lose track. Unless I have a framerate counter, I can't tell if I'm getting 150 fps or 125 fps or even 180 fps.

In Q3 demo001, for example, do you break 200 fps in the little corridor just before you gib that guy with the railgun? (/seta r_BraggingBitch 1) I do (/seta r_BraggingBitch 0). Does breaking 200fps help the smoothness of the game? Nope. Does it have a heavy weight in your final benchmarking score? Sure as hell does.

So what? Well, it goes without stating that maintaining framerate is one of the most important considerations when building your gaming rig. In-game, I want no hiccups or stutters. I want perfectly smooth gameplay.

Unfortunately, the way benchmarks are run today, it is no longer about "measuring smooth gaming", it's about "statistics" (of the "lies, damn lies, and..." fame)

(shaddup, I'm getting to the point!)

So what can we do about the 3 "issues" I've raised?

1) Not much we can do. Either the developers stick in the "minimum/average/maximum" framerate calculations, or they don't. If the game has it, then let's use it and report it, as that gives a MUCH better picture of performance than just a simple average framerate, IMHO.

2) Someone, preferably a large website host (HINT HINT HINT!!!) should develop one (or more) &quot;killer demos&quot; to be used for benchmarking, much like Brett Jacobs did with massive1 and crusher. If it's good, it'll become a standard. Just don't give yourself too much credit, or other websites won't want to use it <g>

When I was comparing graphics cards, I made my own demo. Q3tourney4, 10 bots, me on god mode (cuz I sUx0rZ), fraglimit of 50. It really told me QUITE a bit about how the cards performed when I needed them to. The level detail was high, the action was intense, and there was NO time when my framerate jumped real high, so I didn't get a 'biased' report of my framerate, I got a good representation of how my system performed in intense situations when I needed it to perform well for me.

The difference between an average of 80 fps and 90 fps on demo001 is minimal. The difference between averaging 30 fps and 40 fps on my homemade 'q3crusher' demo is another frag added to your total vs. respawning with the gauntlet and running frantically trying to find a &quot;real&quot; weapon. The killer demo tells you how your system will perform in worst-case scenario, when you really need it to come through for you.

3) In lieu of #1 and #2, let's do something &quot;drastic&quot;. Let's benchmark with vsync ENABLED. yes, it goes against EVERYTHING we've ever thought, but if you have a monitor that can do 125 Hz, setting vsync means you are effectively capping your frames @ 125 fps. Does anyone REALLY NEED MORE???

Well, why bother capping your framerate @ 125?

Simple. A &quot;lower low&quot; can be made up for by a &quot;higher high&quot;. The &quot;lower low&quot; has a negative effect on your gaming, but the &quot;higher high&quot; doesn't help out at all.

Simplified hypothetical example:

Card 1 - high = 145 fps
low = 25 fps
ave = 85 fps

Card 2 - high = 120
low = 50
ave = 85

Ideally, a benchmark will report that card 1 dropped down twice as low as card 2. Usually, all we'll see is the average score. The average might lead one to believe that both cards performed about equal, but that obviously isn't the case.

Now apply vsync.
Card 1 - high = 125
low = 25
ave = 75

Card 2 - high = 120
low = 50
ave = 85

By appying vsync, we give a more accurate report of the true noticeable performance of the card. Obviously the #'s are hypothetical and assumed, but you get the message (I hope)

Is this is good as being able to report a &quot;high/low/average&quot;? Probably not. Is this as good as having a &quot;killer demo&quot;? Probably not. But it's as close as it gets anymore.


This primarily applies to the &quot;middle of the road&quot; resolutions, like 1024x768 (depending upon the game), where you can have massively high spikes in low detail situations, but your framerate can drop rock-bottom in intense situations. The spikes don't affect your gaming at all, but the drops sure hurt it!

Another way to look at it is to measure how fast a car is. Hey, my little Dodge Neon can break 100 MPH while going down a nice hill no problem! Let's not give that too much weight when asking &quot;can the car move, or what?&quot;

Anyway, if you agree, then say so. If you disagree, then piss off. <G>

Now then, a request:

If you have one of the &quot;big 3&quot; cards (or hell, any card) and you have a monitor that can manage 125 Hz at 1024 and 1280, could you run some benchmarks with vsync enabled and report your scores? I'm wondering if this will reveal something interesting.

Q3, Evolva, MDK2, and UT all have downloadable demos with benchmarks, so even if you don't have the game, you can run the benchmark. Obviously, CPU and RAM type and speeds, mobo, OS, etc. are important side information.

Feedback appreciated, and discussion welcome, as always.


 

pen^2

Banned
Apr 1, 2000
2,845
0
0
wow, first time (i guess) in complete agreement with ever single points you made! yeah your post is kinda longwinded as always ;) but at the same time it didnt have any unnecessary whining/ranting, eloquent and concise i must say :)

just to beef up the arguments:

1) another thing that comes into play is consistency; say, on one card the fps might stay high for most part but dip low at times due to texture thrashing or whatever. on another card, unlike the former, the fps stay generally lower the entire time, yet the lowest being the same. from a gamers viewpoint, occasional hiccups could be tolerable/less noticable compared to overall slower/jerkier visuals. again, nothing you could reflect this on the fps marks, maybe the games should come with a log generator that records fps for every fraction of seconds or so... obvioulsy this shouldnt be as easy as it sounds i am afraid...

2) 10fps differntial at lower fps not only affects your frag counts, but at the same time it could greatly influence 'fun factor' of the game as the whole..... personally, gettin sub 30fps in deathmatch games distracts me from immersing into the gameplay, disgusting me and turning me off at the same time (bleh)

3) i am all for enabling the vsync even in the benchmarks..... it sickens me to see misguided people leaving vsync off in casual gaming for the sake of higher fps :Q
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0


<< wow, first time (i guess) in complete agreement with ever single points you made! >>

well, it's about time you started to be sensible!!! :D

what the truly ULTIMATE way would be is for a game developer, say JC for example were to program into his benchmarking code for a game, say Doom3, a way to graph the rendering speed of each frame.

for example, say the demo runs for 100 frames. By measuring the speed that each frame is rendered, you can get a frame-by-frame rendering speed measurement. Graph this, and you get a PERFECT way to truly measure rendering speed throughout the demo. You can see not just the &quot;low&quot;, but you can set a &quot;lower limit&quot;, for example, 60 fps, and then check to see just how long the card is rendering slower than your lower limit.

You can also superimpose the graph of one card over that of another, and get some very precise quantitative data.

Now if we can only get someone to do this. HA!!!!!!

/me wakes up from dream

 

BFG10K

Lifer
Aug 14, 2000
22,709
3,007
126
I've allways wanted two seperate sets of benchmarks: one set of standard ones (to show the full muscle of the system) and another set of benchmarks with vsync on and reporting the minimum framerates, all the while running a crusher style benchmark.

One of the reasons for advocating the highest possible average FPS is because we don't have have the second method available to test the worst-case scenario. Therefore the higher the average FPS the less chance the minimum framerates will drop too low.

That's why I get so annoyed at those &quot;experts&quot; who make up some magic FPS number and claim anybody who gets more than it is wasting their time because &quot;your eyes can't see a difference&quot;. They just don't have a clue about fluctuating gameplay and the desire to achieve constant smooth gaming by using vsync to cap maximum framerates and using a high average framerate to keep the minimum FPS up.
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
soooo.....does anyone have a Radeon, GTS or GTS-Ultra or DDr they can do some benchmarking with vsync enabled @ at least 100 Hz?
 

BFG10K

Lifer
Aug 14, 2000
22,709
3,007
126
ever get that MX working properly?

Sigh... no I returned it a few days ago. It just didn't like my monitor or something. Anyway it's back to the good ol' V3 for me (for the meantime at least).
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
how much did you pay for the MX?

You see that the 5500 can be picked up for <$145 these days, right?

:)

 

Weyoun

Senior member
Aug 7, 2000
700
0
0
that would be, indeed, the end of the world robo :D the year 2000 may indeed be a year of change, but the words voodoo5 and bfg together, just dont work :) j/k

that sounds like a good solution, although your cpu seems to be a little mediocre for a card of that calibre. hell im running a 466 and considered buying a v5 :) if 3dfx slashed the prices on their v4s, they'd be absolute killers in the sales department, im sure BFG'd pick one up, i'd be able to get my friend to get one too, they're just too expensive...

BTW BFG, where to you order your parts from?
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
heh...yeah, no kidding

however, after finally using an nvidia card, he's back with 3dfx.

:)

he's not the first to do that, either

 

Weyoun

Senior member
Aug 7, 2000
700
0
0
I'm seriously considering getting a V5 5500 for my brother's new comp (t-bird, ddr when/if it ever comes out here) basically because of these new rumoured drivers, the performance increase Dave has brought to light (thanx dude :) ) FSAA and image quality. He plays a helluva lot of counterstrike, racing games, some quake3 and RPGs, i honestly thought that'd be the best choice for him :) Even though he said he wants a geforce for the w@nkery value, im sure a V5 would be sufficient for him :)

Also, being the first rig and AMD system i've built, even though i manage to keep our current machine in working, tweaked order, I didn't want to think about any compatibility issues so many nVidia AMD users are encountering.

It really is a pitty that 3dfx can't pick up too many OEM sales, I think they truly deserve *some* reward for their effort, especially after what they've just been through....
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
Weyoun, I agree with you

one of 2 things are going to happen with Rampage

1) The single Rampage chip will be SOOOOOO awesome, it'll garner a ton of OEM deals.

or

2) The single Rampage chip will be good, but they'll NEED the SLI Rampage + SAGE to really compete

If # 1 happens, 3dfx will be &quot;back on top&quot;, because if the single chip kicks ass, the SLI will frickin' own.

But if #2 happpens, even if the SLI Rampage kicks ass, 3dfx is gonna be REALLY hurting

plain and smiple - OEM's dont' like SLI. They like a simple card that has tons of checkbox features. The 5500 didn't provide that, and the 4500 didn't, and was/is slow.

 

BFG10K

Lifer
Aug 14, 2000
22,709
3,007
126
RoboTECH:

how much did you pay for the MX?

$359 NZD. I got a full refund though.

You see that the 5500 can be picked up for <$145 these days, right?

Is there something you're trying to tell me? ;)

however, after finally using an nvidia card, he's back with 3dfx.

If you told me that a month ago I wouldn't have believed you.

Weyoun:

but the words voodoo5 and bfg together, just dont work j/k

I still don't like the multi-CPU implementation of the V5 5500 and the associated problems with it (heat, size, power requirements etc). That part hasn't changed.

although your cpu seems to be a little mediocre for a card of that calibre.

Indeed. That's why I mentioned in another thread that my next upgrade is going to be a Coppermine Pentium 3 (133 MHz FSB). It was inevitable anyway given my motherboard fully supports 133 MHz FSB and my RAM is PC 133.

if 3dfx slashed the prices on their v4s, they'd be absolute killers in the sales department, im sure BFG'd pick one up,

Yep, I sure would.

BTW BFG, where to you order your parts from?

Anything I can I order from the U.S (PC Warehouse, Cyberian Outpost, etc). If it's too expensive (or risky) to import it I go to my local computer store which built my computer.
 

BenSkywalker

Diamond Member
Oct 9, 1999
9,140
67
91
&quot;If you told me that a month ago I wouldn't have believed you.&quot;

3dfx's drivers hosed your system, I thought that you knew that that would be the case beforehand? It did seem that they(drivers) did quite a nice job on your system, but it isn't anything that isn't commonplace when trying to get rid of a 3dfx board. They should include a utility to remove the drivers with virus scans(no, I'm not kidding).

Format and get rid of the 3dfx crap on your HD and you would have had no problems(though I know that wasn't an option for you, it is the only thing I know to do outside of what I suggested). If you try to move to a Radeon, you may well run into the exact same problems(unless they overwrite all the junk installed by the 3dfx offerings which is possible, they both have enormous driver files).
 

BFG10K

Lifer
Aug 14, 2000
22,709
3,007
126
3dfx's drivers hosed your system, I thought that you knew that that would be the case beforehand?

I did realise that and I did clear my HD and registry from all traces of them before I installed the GF2 MX. Maybe I did leave something behind but I doubt it. I'm not blaming 3dfx or nVidia for the problem. The point is the video card didn't work for whatever reason and I returned it.
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
Ben, you speak as though this is &quot;unique&quot; to 3dfx?

same stuff happens when you try to move from an nvidia card

oh wait, you haven't done that in about 4 years, eh? <g>

 

BenSkywalker

Diamond Member
Oct 9, 1999
9,140
67
91
&quot;same stuff happens when you try to move from an nvidia card&quot;

Bounce back and forth between ATi and nVidia all the time, never have any problems at all. Move over to Matrox and back, no issues. Throw 3dfx into the mix and everything goes south.
 

RoboTECH

Platinum Member
Jun 16, 2000
2,034
0
0
interesting. I've had no problems going from ATi to 3dfx and back (Rage Pro cheapo). Had a hellacious time going from 3dfx to nvidia, and from nvidia to 3dfx.

even checked the GeForce FAQ to figure out how to get all of nvidia's crap out of the registry and I still ended up reformatting. Couldn't get rid of these goofy lines on the screen with the 5500 initially. Checked dejanews and found that there were several posts (even found a website review where this happened, can't remember off the top of my head tho) that said that is common when moving from nvidia to 3dfx.

go figure. Kinda like Intel's bridge chip not working with non-Intel chipset based mobo's, heh.....

interestingly tho, when moving from the 5500 to the GTS, I had zero problems (with 5.32) until I installed 6.18. I thought that perhaps there was still some driver junk leftover from the 5500, so I reformatted, Still no go.

You do know you have to uninstall 3dfx tools and do a registry check for glide and 3dfx stuff, right?