*update*
well my machine spent most of yesterday and part of today testing 5 simultaneous GPU tasks in multiple configurations. i started w/
5 GPU tasks +
4 CPU tasks, and took note that my GPU was seeing 95% utilization, and GPU task run times were ~201 minutes. i then tried
5 GPU tasks +
3 CPU tasks, and took note that my GPU was now seeing 96% utilization, and GPU task run times were now only ~200 minutes. naturally i tried
5 GPU tasks +
2 CPU tasks next, and took note that my GPU was seeing 97% utilization, and GPU task run times were now down to ~198 minutes.finally, i tried the "
5 GPU tasks +
1 CPU task" &
"5 GPU tasks +
0 CPU tasks" configurations , but my GPU utilization never went beyond 97%, and my GPU task run times never dipped below 198 minutes.
at this point, i took all my data and plugged it into the following excel spreadsheet:
i then highlighted the most GPU-compute-efficient configurations and began to contemplate them. these included 3, 4, and 5 simultaneous GPU task configurations, the improvements in compute efficiency of which all centered around the 37.5% mark, give or take no more than 0.6%. knowing that the ever-so-slight differences in compute efficiency and PPD between some of these configurations would be negligible in the short run, and would hardly amount to anything significant in the long run, i decided to focus on which configuration would give me the best combination of CPU and GPU compute efficiency. the most CPU tasks running simultaneously in any of these configurations is 4, and the least 2. if you note the CPU task run time column, you'll note that CPU task run times remained the same across the testing of multiple configurations. given that CPU task run times are the same whether i run 2 simultaneously or 4 simultaneously, i obviously want to choose one of the configurations that involves running 4 simultaneous CPU tasks. that eliminated quite a few configurations, and left only two - the "
4 GPU tasks +
4 CPU tasks" configuration and the "
5 GPU tasks +
4 CPU tasks" configuration. if the same amount of CPU work gets done in the same amount of time in either of these configurations, then it comes back down to the GPU tasks as the deciding factor. if running 5 simultaneous GPU tasks doesn't do any more work in any less time than running 4 simultaneous GPU tasks, then it would seem to me that the "
4 GPU tasks +
4 CPU tasks" configuration is the logical choice. even though GPU utilization is the same in both scenarios, the "
5 GPU tasks +
4 CPU tasks" configuration uses more VRAM, and theoretically should cause the GPU to draw more power.
at any rate, i know that everyone's hardware/software setup is a little different, and i don't expect anyone to be able to duplicate these exact results...of course this info was just meant for reference anyways. hopefully it'll give interested folks an idea of what to expect w/ their ATI/AMD GPUs...and if not for the Einstein@Home project itself, i hope folks can at least use this model as a template for testing DC platforms w/ both CPU and GPU clients.
Eric