Nvidia released the PCI-Express version of the P100 in the form of the Tesla P100. The hpc-card was revealed back in April already however was based on a Mezzanine-connector.
Nvidia is to release two versions of the PCI-Express models with 16GB HBM2 memory, and one with 12 GB HBMs (one stack disabled). The cards can communicate over the NVLINk interconnect with each other. The cards will get a slightly lower boost frequency compared to the Mezzanine NVLINK version, the TDP is lower as well at 250 Watt.
Nvidia Tesla | ||||||
---|---|---|---|---|---|---|
Tesla P100 (Mezzanine) |
Tesla P100 (16GB) |
Tesla P100 (12GB) |
Tesla M40 | |||
Gpu | GP100 (610mm2) |
GP100 (610mm2) |
GP100 (610mm2) |
GM200 | ||
Architecture | Pascal | Pascal | Pascal | Maxwell 2 | ||
Core | 1328MHz | - | - | 948MHz | ||
Streamp processors | 3584 | 3584 | 3584 | 3072 | ||
Boost | 1480MHz | 1300MHz | 1300MHz | 1114MHz | ||
Mem. | 1,4Gbit/s HBM2 | 1,4Gbit/s HBM2 | 1,4Gbit/s HBM2 | 6Gbit/s gddr5 | ||
Mem bus | 4096-bit | 4096-bit | 3072-bit | 384-bit | ||
Mem bandwidth | 720GB/sec | 720GB/sec | 540GB/sec | 288GB/sec | ||
Mem MGB | 16GB | 16GB | 12GB | 12GB | ||
Half Precision | 21,2 tflops | 18,7 tflops | 18,7 tflops | 6,8 tflops | ||
Single Precision | 10,6 tflops | 9,3 tflops | 9,3 tflops | 6,8 tflops | ||
Double Precision | 5,3 tflops (1/2 rate) |
4,7 tflops (1/2 rate) |
4,7 tflops (1/2 rate) |
213 gflops (1/32 rate) |
||
Transistors | 15,3 Billion | 15,3 Billion | 15,3 Billion | 8 Billion | ||
Tdp | 300W | 250W | 250W | 250W | ||
Formfactor | Mezzanine | pci-e | pci-e | pci-e | ||
Cooler | N/A | passive | passive | passive | ||
Fab | tsmc 16nm finfet | tsmc 16nm finfet | tsmc 16nm finfet | tsmc 28nm |
The 12GB version thus will have a 3072 bits wide memory bus, the 16GB version has the full 4096 bit wide memory bus. These cards are intended for high performance computing of course.