No one is denying the 7970 isn't a good deal at $200 (unless it's been thrashed by mining), so you're basically arguing with yourself.
You also seem to be another person hung up on numbers. AMD/ATI cards have had the edge over Nvidia boards in double precision (and single for that matter) on some time, yet it still isn't reflected in the wider community. Why? Because AMD's architectures are totally reliant upon OpenCL for the most part, and OpenCL support is spotty at best.
Say what you will about CUDA, but the software ecosystem is in place and it works.
Blender from their own
FAQ:
AFAIK, OpenCL (working) support isn't overly prevalent. Even Lux, which is touted as a poster child for OpenCL has ongoing issues, and generally where both CUDA and OpenCL are supported, it is the former that is generally more mature. The Blender sentiment isn't a lone voice (
pdf)
So, basically, hardware performs as well as the coding allows. AMD is tied to OpenCL, and OpenCL is tied to third parties for its advancement, which it is very much a case of YMMV. What clouds the issue further is that mainstream (gaming) sites use OpenCL apps only to compare AMD and Nvidia cards which distorts the overall picture, since using the CUDA path for Nvidia hardware invariably means a better result.
Note this render test using Premiere where the GTX 670 (using CUDA) and the R9 290X (using OpenCL) are basically equal in time to render. On raw numbers the 290X should have it all over the GTX 670, after all the AMD card has
5.8 TFlops of processing power to the 670's
2.46 TFlops - well over double!
So the 7970 might represent great value with the right application, and it certainly is affordable at $200. On the other side of the ledger, the GTX 580 -also because of its decreasing price (and its ability to run CUDA apps), makes a compelling buy for people who want to use CUDA coded render/CAD apps. Is the Titan the be-all-and-end-all ? Of course not, and I don't see anyone saying it is. What I see is people comparing two current top tier GPUs because....well, because they are the flavour of the week.
Numbers on the page don't always translate that well in real life scenarios. Harping back to point regarding double precision. It's use is governed by the same coding environment. Can't say I've seen many FP64-only benchmarks outside of HPC, most consumer apps tend to involve both single and double precision calculation rather than FP64 solely, and those that do find their way into benchmarks, are again devoid of using the CUDA path. HPC is the environment for widespread use of double precision, and the
ratio of Nvidia to AMD GPUs there is a pretty telling story.
Now, you realise that these two statements you made are contradictory:
The majority of compute application Nvidia cards use
ARE CUDA code. These
machines, this
machine, this
machine, these
machines are all geared for content creation. You say its a waste of money, these people beg to differ. Different requirements equals different usage patterns.