kb3edk
Senior member
LOL I didn't even look at the dates on the earlier posts! Not that it matters, I'm not the one who revived the thread 😛Originally posted by: Insidious
OK, who's the dork that revied this thread 😛
LOL I didn't even look at the dates on the earlier posts! Not that it matters, I'm not the one who revived the thread 😛Originally posted by: Insidious
OK, who's the dork that revied this thread 😛
Originally posted by: Insidious
OK, who's the dork that revied this thread 😛
All joking aside, time for some live and let live. If anyone doesn't want to run it they aren't. And If anyone wants to run it, they are.
Big Deal
-Sid
Originally posted by: Wolfsraider
Ok, let me ask a question here.
If I had 2 exactly the same computers running rosetta@home,
opteron 165 in each one.
identical in every way.
one is running the standard client.
the other is running the sse2 client.
the one using the optimized client supports my hardware and actually helps it score where it should because it supports my processors instruction bit. The other actually is a blanket for all cpu's and has no code to help it.
Here is my question:
Is this any different from ram tweaking, or different processors or crunchers or operating systems, doing better in the result field, and why?
I am not trying to start anything here so please refrain from evaluating others posts. I simply want a better idea about what I may not see.
Originally posted by: Rattledagger
Originally posted by: Wolfsraider
...
Anyway, one resent change to BOINC added after Rosetta@home's wish is so a project can set max claimed credit, any result with higher claimed credit will automatically be flagged as invalid and therefore get no credit...[
I didn't know that. Sounds like a good 'first step' in a good direction to me! :thumbsup:
Any progress on the flop counting methods that were getting talked about so much for a while?
-Sid
Originally posted by: Insidious
Any progress on the flop counting methods that were getting talked about so much for a while?