An interesting tidbit of info just surfaced on the web through Bright Side of News, regarding NVIDIA's upcoming GT300 chip (high-end graphics card part. I'll do the blunt copy/paste here:
nVidia GT300 chip is a computational beast like you have never seen before. In fact, we would go as far out and state that this is as closest as GPU can be to a CPU in the whole history of graphics technology. Now, time will tell whatever GT300 was the much needed revolution.
Beside the regular NV70 and GT300 codenames [codename for the GPU], nVidia's insiders called the GPU architecture - Fermi. Enrico Fermi was an Italian physicist who is credited with the invention of nuclear reactor. That brings us to one of codenames we heard for one of the GT300 board itself - "reactor".
When it comes to boards themselves, you can expect to see configurations with 1.5, 3.0 GB and 6GB of GDDR5 memory, but more on that a little bit later.
This is the meat part you always want to read fist. So, here it how it goes:
- 3.0 billion transistors
- 40nm TSMC
- 384-bit memory interface
- 512 shader cores [renamed into CUDA Cores]
- 32 CUDA cores per Shader Cluster
- 1MB L1 cache memory [divided into 16KB Cache - Shared Memory]
- 768KB L2 unified cache memory
- Up to 6GB GDDR5 memory
- Half Speed IEEE 754 Double Precision
As you can read for yourself, the GT300 packs three billion transistors of silicon real estate, packing 16 Streaming Multiprocessor [new name for former Shader Cluster] in a single chip. Each of these sixteen multiprocessors packs 32 cores and this part is very important - we already disclosed future plans in terms to this cluster in terms of future applications. What makes a single unit important is the fact that it can execute an integer or a floating point instruction per clock per thread.
TSMC was in charge of manufacturing the three billion transistor mammoth, but it didn't stop there. Just like the G80 chip, nVidia GT300 packs six 64-bit memory controllers for a grand total of 384-bit, bringing back the odd memory capacity numbers. The memory controller is a GDDR5 native controller, which means it can take advantage of built-in ECC features inside the GDDR5 SDRAM memory and more importantly, GT300 can drive GDDR5 memory in the same manner as AMD can with its really good Radeon HD 5800 series. The additional two memory interfaces will have to wait until 28nm or 22nm full node shrinks, if we get to them with an essentially unchanged architecture. You can expect that the lower-end variants of GT300 architecture will pack less dense memory controller for more cost efficiency, especially on the memory side.
Ferni architecture natively supports C [CUDA], C++, DirectCompute, DirectX 11, Fortran, OpenCL, OpenGL 3.1 and OpenGL 3.2. Now, you've read that correctly - Ferni comes with a support for native execution of C++. For the first time in history, a GPU can run C++ code with no major issues or performance penalties and when you add Fortran or C to that, it is easy to see that GPGPU-wise, nVidia did a huge job [via bsn].