• We’re currently investigating an issue related to the forum theme and styling that is impacting page layout and visual formatting. The problem has been identified, and we are actively working on a resolution. There is no impact to user data or functionality, this is strictly a front-end display issue. We’ll post an update once the fix has been deployed. Thanks for your patience while we get this sorted.

[www.youtube.com/paulshardware] Maxwell 750 ti video with benchmarks.

VirtualLarry

No Lifer
http://www.youtube.com/watch?v=vpVEA6zBrcc&list=UUvWWf-LYjaujE50iYai8WgQ

Paul (of NeweggTV fame) shows off some 750 ti cards, including benchmarks.

Among the benchmarks is the R7 260X. The 750 ti is around equal, wins some, loses some (especially at high res - the R7 260X has more memory bandwidth).

Edit: My bad, I was thinking of the AT article benchmarks on Maxwell. In Pauls video, the 750 ti is ahead of the R7 260X in nearly all of the benchmarks. The plain 750 is slower in pretty much every benchmark.

That was the basis of my hypothesis in the other thread, which was basically, trying to decide between these two cards, in the price bracket.

1GB R7 260X $130
http://www.newegg.com/Product/Product.aspx?Item=N82E16814161454
1GB 750 OC $125
http://www.newegg.com/Product/Product.aspx?Item=N82E16814127782

Based on Pauls benchmarks, it seems like, if I have a 6-pin, I would be better off with the R7 260X.
 
Last edited:
The 750 is way to expensive, needs to drop $30. The 260x is cheaper and faster/on par and the 265 is the same price with 20-30% higher performance.
 
Factory overclocked 750 Ti vs stock 260x.

That model of the 750ti (EVGA 750 Ti FTW) in the review is 179$ on newegg and sold out.
 
Is rather a efficiency and feature matter than a gaming value matter. Viewing by this approach, the 750/750Ti looks not a bad value.
 
...and the ultimate low-profile / HTPC card!

I agree. For that purpose, it looks to have no equal.

I would love to get a 750 / 750ti low-profile card for my G630 Pentium Sandy Bridge slimline Gateway PC. It currently has a GT 430 (I think 128-bit DDR3).
 
Btw, I believe that this is the beginning of a new era in GPU technology. Where GPUs are no longer rated by their performance, or price / performance, but by performance / watt. NV is taking a page out of Intel's playbook here. Imagine future GPU generations, that are no more powerful than today's GPUs (partially because they will hit a CPU bottleneck wall in a few generations, because the absolute performance of the CPUs that drive them have also stagnated), but instead, take 30-50% less power.
 
Agreed, and it's also helped by the ability to use multiple of these cards in the same system to increase performance accordingly. The drivers keep getting better for multi-GPU.
 
What's the point in ever upgrading if absolute performance doesn't increase? I'm not going to pay $110 for a card every 3 years to save me $20 a year in energy costs. I save much more by changing my lights or fridge.
 
I think $130+ in the right gpu direction is fine for a gaming gpu but none of those games were play able so not sure what the hype is about ,but no doubt under clockers will blow their load at a gaming card that can't play games , but will reach a all time low watts record for a non gaming card in game benches , but gaming on a smart phone still beats watts per fps it so whats the deal...imo
 
Last edited:
Btw, I believe that this is the beginning of a new era in GPU technology. Where GPUs are no longer rated by their performance, or price / performance, but by performance / watt. NV is taking a page out of Intel's playbook here. Imagine future GPU generations, that are no more powerful than today's GPUs (partially because they will hit a CPU bottleneck wall in a few generations, because the absolute performance of the CPUs that drive them have also stagnated), but instead, take 30-50% less power.

That's depends on the segment.

The true gaming crow and top dog performance crow don't care about perf/watt ratio.
 
That's depends on the segment.

The true gaming crow and top dog performance crow don't care about perf/watt ratio.

But what if GPU companies themselves start to only cater to the performance / watt crowd? Remember, "mobile is the future, desktop is dead".
 
Back
Top