What a great early birthday present!the [HD 4000] driver will improve performance up to 10% while lowering power consumption at the same time.
What a great early birthday present!the [HD 4000] driver will improve performance up to 10% while lowering power consumption at the same time.
The only difference between those 2 cards is DDR3 vs GDDR5. And its 1800Mhz DDR3 vs 4Ghz GDDR5. Yet board TDP is 36% higher under maximum load and 10% higher under idle. Assuming the DDR3 draws 4W and the GDDR5 draws 20W under peak load.
GDDR is simply a power hog. Speed over power consumption.
Haswell has OpenCL 1.2 support. Ivy Bridge does not.
It just doesn't make financial sense to use GDDR5 as system RAM. Look in the memory and storage forum, people are freaking out about standard DDR3 memory prices rising over their historic lows. People aren't going to pay several times current prices to put memory in a budget system.
Do you remember when the P4 required RDRAM? That didn't work out so well for Intel. If Intel can't push expensive memory, AMD certainly can't.
And how is AMD going to sell their customers? "Hey Dell, HP, Lenovo, build systems using our low price CPU. Ignore the fact that the price difference plus more is eaten up by the cost of memory. Just pass it on to your customers, they will be fine with it because it has faster graphics than Intel".
If AMD tries to go down this route they will just hasten their decline.
OpenCL gets translated on the fly with the driver to the hardware. Just like Java and .Net. Or DirectX, DirectCompute and so on. OpenCL is also on top of CUDA on nVidia as well.
Its certainly not closer to bare metal.
![]()
Hmm, didn't know that by a brief look at the code. Still, no VM, and I'm pretty sure you need to know the GPU arch to get the best performance; whereas in Java and .net you need to understand the JVM to get the best performance.
I'm writing CUDA code right now, it's much closer to the metal, there is a CUDA runtime and the CUDA compiler (nvcc.exe) does produce an IL code (PTX) that runs on that runtime/driver. Still, it's allot closer to the hardware than something like Java being 'compiled' by a JIT and then run on the JVM.
If if don't code for the arch I'm running on, or build the code to handle multiple uArchs, it stands a good chance of running sub-optimally on a different GPU than I'm using (Kepler 3.0). (**would love a Titan** - maybe if there will be some price cuts after Thanksgiving, and hopefully Malta rocks with the new ~June release of Catalyst).
Hmm, didn't know that by a brief look at the code. Still, no VM, and I'm pretty sure you need to know the GPU arch to get the best performance; whereas in Java and .net you need to understand the JVM to get the best performance.
I'm writing CUDA code right now, it's much closer to the metal, there is a CUDA runtime and the CUDA compiler (nvcc.exe) does produce an IL code (PTX) that runs on that runtime/driver. Still, it's allot closer to the hardware than something like Java being 'compiled' by a JIT and then run on the JVM.
If if don't code for the arch I'm running on, or build the code to handle multiple uArchs, it stands a good chance of running sub-optimally on a different GPU than I'm using (Kepler 3.0). (**would love a Titan** - maybe if there will be some price cuts after Thanksgiving, and hopefully Malta rocks with the new ~June release of Catalyst).
JIT code is JIT code. About the only difference I'd see is that the VM on a graphics card would compile for the specific hardware (ie, how much memory is available), whereas the Java VM has to be manually tuned by the user.
It can be tuned by the user, but never really is, it is tuned by the developer(I wil instruct the JVM to run my program with these flags ....).
Unless you are running a cluster of weblogics or whatever you, as an user, will never deal with "tuning the jvm".
As far as I know, there is no such thing as JIT code, yet on interpretation it could be something I never quite understood, as in "why not"? When the JVM compiles(or JIT's) some code, why not write that binary to disc for next time? Why do the JIT compilation on the same code over and over again each time the app is executed .. It is not like we all dont see the lag !
Most likely most systems will use DDR3. But perhaps AMD will try to score some "premium" gaming systems with high end APUs with mandatory GDDR5. (high end being >$700 - $1000, a range in which intel igps are still quite common)
That APU system will cost more and have lower performance than using a discreet GPU.
1) That lag isn't important (I care, and you care, but Oracle doesn't), and 2) it's spending way more time executing it, and 3) in the case of standard Java, much of the code stays interpreted, rather than being compiled.It can be tuned by the user, but never really is, it is tuned by the developer(I wil instruct the JVM to run my program with these flags ....).
Unless you are running a cluster of weblogics or whatever you, as an user, will never deal with "tuning the jvm".
As far as I know, there is no such thing as JIT code, yet on interpretation it could be something I never quite understood, as in "why not"? When the JVM compiles(or JIT's) some code, why not write that binary to disc for next time? Why do the JIT compilation on the same code over and over again each time the app is executed .. It is not like we all dont see the lag !
That's just wrong. I'm using OpenCL 1.2 on my HD4000 since several months. Every 15.31 driver out there support this.
It just doesn't make financial sense to use GDDR5 as system RAM. Look in the memory and storage forum, people are freaking out about standard DDR3 memory prices rising over their historic lows. People aren't going to pay several times current prices to put memory in a budget system.
Do you remember when the P4 required RDRAM? That didn't work out so well for Intel. If Intel can't push expensive memory, AMD certainly can't.
And how is AMD going to sell their customers? "Hey Dell, HP, Lenovo, build systems using our low price CPU. Ignore the fact that the price difference plus more is eaten up by the cost of memory. Just pass it on to your customers, they will be fine with it because it has faster graphics than Intel".
If AMD tries to go down this route they will just hasten their decline.
That APU system will cost more and have lower performance than using a discreet GPU.
There may be volume behind this to drive down prices, or the repurposing of existing engineering designs to reduce design costs.
In that case isn't is strange that Sony and MS didn't go for a discrete GPU with limited amount of GDDR5 RAM in the PS4 and XBOX720?
Instead they choose 8 GB GDDR5 RAM and an APU. Just like what a possible Kaveri based system that we are now discussing could look like.
Yes, if both AMD, Sony and Microsoft think it's a good idea to go with this solution to use a large amount of GDDR5 RAM together with an APU, then I guess they must have though of something.
Just looking at current price of high end GDDR5 RAM, it does seem expensive. But likely they are primarily focusing on using lower speed GDDR5 RAM which is cheaper. And maybe they are seeing something looking ahead, like GDDR5 prices falling. Just look at what has happened to the DDR3 prices. Not that long ago it was quite expensive, but now it's insanely cheap. Today some people even get 16-64 GB RAM systems "just because they can"...
Also, if they will be able to use 8 GB GDDR5 RAM for the PS4 and XBOX720, it seems like they must be able to get it at reasonable prices. After all, their budget must also include an APU, HD, Blu-ray drive, motherboard, chassis, controller, etc. And all of it should perhaps be priced at around $400. In addition, what's been said is that they don't intend to take any initial cost penalty this time, selling the units with a loss like Sony did with the PS3.
So far the Xbox720 is rumoured to go with 8GB DDR3.
It's also rumoured to have some form of on-package EDRAM to help with bandwidth issues... which ties in quite nicely to the original topic!
So far the Xbox720 is rumoured to go with 8GB DDR3.
In that case isn't is strange that Sony and MS didn't go for a discrete GPU with limited amount of GDDR5 RAM in the PS4 and XBOX720?
Instead they choose 8 GB GDDR5 RAM and an APU. Just like what a possible Kaveri based system that we are now discussing could look like.