PCGH: So, I can just take my CUDA program, port the host-based part of code to support the ARM-architecture and then it'll run on a smartphone for example.
Bill Dally: That's the goal, yes. We want CUDA to run across the entire product line. CUDA everywhere. [Except everyone who owns an ATI graphics card of course]
PCGH: So it is more important to you to support CUDA everywhere, than to make a decision between the host architecture, between ARM and x86 for example?
Bill Dally: I don't think the instruction set of the CPU matters a whole lot. In running the CUDA program, which is running on the GPU, even on the GPU part we have an abstract with PTX which will then translate to support multiple generations of Streaming Multiprocessors that have slightly different native assembly languages. So it doesn't matter what the native assembly language of the host is.