Nvidia Announces PCIe version Tesla P100

Published by

Click here to post a comment for Nvidia Announces PCIe version Tesla P100 on our message forum
https://forums.guru3d.com/data/avatars/m/204/204118.jpg
Check spelling^
https://forums.guru3d.com/data/avatars/m/80/80129.jpg
I wish they had a vendor list of companies that are buying this, for my own curiosity. I know Google has their own Tensor Processing Unit which they claim is the best in the industry for deep learning. I wonder how these stack up.
https://forums.guru3d.com/data/avatars/m/226/226864.jpg
I assume we could see something similar as the PCIe versions for the Geforce Ti and Titan Pascal variants, respectively.
data/avatar/default/avatar27.webp
I assume we could see something similar for the Geforce Ti and Titan Pascal variants, respectively.
Exactly what I hope for and can't wait to.
https://forums.guru3d.com/data/avatars/m/246/246171.jpg
I wish they had a vendor list of companies that are buying this, for my own curiosity. I know Google has their own Tensor Processing Unit which they claim is the best in the industry for deep learning. I wonder how these stack up.
I'm not sure I understand. Teslas are as similar to TPUs as they are to CPUs. Teslas are meant for high-precision highly-parallel number crunching. The TPUs, to my knowledge, are meant for rapid approximations. In the server world, you buy what does your workload fastest, and that's why architectures like PPC, SPARC, and AMD's Bulldozer are still relevant. Intel is really the only company that has any interest in general-purpose servers.
https://forums.guru3d.com/data/avatars/m/80/80129.jpg
I'm not sure I understand. Teslas are as similar to TPUs as they are to CPUs. Teslas are meant for high-precision highly-parallel number crunching. The TPUs, to my knowledge, are meant for rapid approximations. In the server world, you buy what does your workload fastest, and that's why architectures like PPC, SPARC, and AMD's Bulldozer are still relevant. Intel is really the only company that has any interest in general-purpose servers.
Tesla's were meant for high precision workload crunching, but with Pascal they added mixed precision (FP16) high performance crunching too. That's why they give an FP16 (Half Precision) number of 21Tflops. Nvidia specifically added that functionality for deep learning. Now obviously, I'd imagine that an entire chip dedicated to deep learning would perform better than one that is good across a wide variety of stuff. But regardless, I'd like to see how the Nvidia chips would stack up against Google's custom built hardware.
https://forums.guru3d.com/data/avatars/m/256/256969.jpg
I might be flamed but... Can you play games with this card? I know about all the deep-learning focus, but "what if" ?
https://forums.guru3d.com/data/avatars/m/246/246171.jpg
Tesla's were meant for high precision workload crunching, but with Pascal they added mixed precision (FP16) high performance crunching too. That's why they give an FP16 (Half Precision) number of 21Tflops.
Ah I wasn't aware of FP16 or Nvidia targeting deep learning. But yeah, I can't imagine that performing as well as a chip dedicated toward it. On the other hand, Google's TPUs are ASICs. They were meant for Google's own specific purposes. What Nvidia is doing is allowing for a general-purpose deep-learning design. So, you sacrifice a little bit of performance but the hardware will still work for other tasks, while also probably being a little more lenient about development. I suppose it's a lot like comparing games developed for consoles vs PC. The console games ditch the graphics APIs, which give them better performance on weaker hardware. Meanwhile, PCs sacrifice performance for freedom of choice.
I might be flamed but... Can you play games with this card? I know about all the deep-learning focus, but "what if" ?
Probably, but it doesn't have any display connectors.
https://forums.guru3d.com/data/avatars/m/256/256969.jpg
Probably, but it doesn't have any display connectors.
lol, ok, i guess not xD Most expensive PhysX card on the market lol
https://forums.guru3d.com/data/avatars/m/80/80129.jpg
It's also passively cooled by the server fans. So unless you have some wind tunnel case cooling going on, you're going to have a bad time. Then there is also driver support and stuff. Oh and it probably costs over $10,000+ per card. You probably can but it would require some dedication.
https://forums.guru3d.com/data/avatars/m/216/216490.jpg
It's also passively cooled by the server fans. So unless you have some wind tunnel case cooling going on, you're going to have a bad time. Then there is also driver support and stuff. Oh and it probably costs over + per card. You probably can but it would require some dedication.
Or simply wait for the gaming Ti variant of the card for 1/10 of the cost. 😀
https://forums.guru3d.com/data/avatars/m/258/258664.jpg
So my question is, and I know I'm hitting a nerve here probably with asking, but are the Titans and 1080TIs as we expect them going to be based on that GP100 chip, or are we getting some kind of cut down GP102 chip, missing the double precision part and nothing else, as in full single precision and memory bandwidth and everything?
https://forums.guru3d.com/data/avatars/m/80/80129.jpg
So my question is, and I know I'm hitting a nerve here probably with asking, but are the Titans and 1080TIs as we expect them going to be based on that GP100 chip, or are we getting some kind of cut down GP102 chip, missing the double precision part and nothing else, as in full single precision and memory bandwidth and everything?
I expect it to be different. There is very little point of doing HBM2 on a gaming card. It's added cost and it yields very little in return, the extra bandwidth is unnecessary, the power savings is minimal compared to GDDR5x on 16nmFF (10-15w at most). The biggest reason is the die size (memory controller on HBM2 is much smaller on die) but I'm not even sure that is worth the cost of manufacturing it over GDD5x. Further, GP100 has some under the hood architectural changes that didn't get carried over to GP104. In GP100 the SM is split in half and nearly every part of the SM got doubled, registers and all. GP104 isn't split -- they did change the number of SM's per GPC but that's about it. I expect cut down GP100's used in the Tesla P100 will end up in some other Tesla variant (they are already kind of doing it here with the HBM). I'm not really sure what form the Titan will take. I expect the Ti to have a 384 bit GDDR5x bus and 3840 cuda cores (or less if it's cut down) and probably also run at a similar clockspeed as the 1080 does.
https://forums.guru3d.com/data/avatars/m/216/216490.jpg
I expect it to be different. There is very little point of doing HBM2 on a gaming card. It's added cost and it yields very little in return, the extra bandwidth is unnecessary, the power savings is minimal compared to GDDR5x on 16nmFF (10-15w at most). The biggest reason is the die size (memory controller on HBM2 is much smaller on die) but I'm not even sure that is worth the cost of manufacturing it over GDD5x. Further, GP100 has some under the hood architectural changes that didn't get carried over to GP104. In GP100 the SM is split in half and nearly every part of the SM got doubled, registers and all. GP104 isn't split -- they did change the number of SM's per GPC but that's about it. I expect cut down GP100's used in the Tesla P100 will end up in some other Tesla variant (they are already kind of doing it here with the HBM). I'm not really sure what form the Titan will take. I expect the Ti to have a 384 bit GDDR5x bus and 3840 cuda cores (or less if it's cut down) and probably also run at a similar clockspeed as the 1080 does.
If history is to be trusted(it almost always is) something tells me we are gonna see the same/similar scenario as 970SLI vs 980Ti performance wise. Which means GTX 1070 will offer equal or a bit more performance over GTX 1080Ti. Which is almost 980Ti SLI performance in a single card! That's completely bonkers! 😀
https://forums.guru3d.com/data/avatars/m/251/251862.jpg
I wish they had a vendor list of companies that are buying this, for my own curiosity.
I don't think they will give you a list, but they claim they are selling a bunch to the U.S. Department of Energy. In 6 months, you can probably look at the top500 and see if anyone is using them. http://www.top500.org/ https://www.youtube.com/watch?v=J6oE5Knk8K8
https://forums.guru3d.com/data/avatars/m/175/175902.jpg
I might be flamed but... Can you play games with this card? I know about all the deep-learning focus, but "what if" ?
yes you can as PhysX card 🙂 (tested myself 🙂)
I wish they had a vendor list of companies that are buying this, for my own curiosity.
yes of course 🙂
https://forums.guru3d.com/data/avatars/m/175/175902.jpg
I'd like to see how the Nvidia chips would stack up against Google's custom built hardware.
Probably less, it is not too much made to compete with this (despite it can : as seen in some custom gigantic build). but don't forget: if it's Google, it is evil :bat:
https://forums.guru3d.com/data/avatars/m/235/235224.jpg
So we might see a Titan or GeForce version this year? 😀
https://forums.guru3d.com/data/avatars/m/90/90667.jpg
I think there is a chance that nvidia will counter Vega with a cut down titan version, if they can't produce enough units for the full version. Or they release it in Feb-March.
https://forums.guru3d.com/data/avatars/m/267/267531.jpg
Gp100 isn't coming to GeForce guys, they're not going to waste all that die area on fp64, not to mention that the size of SMs are cut in half, so register file size is doubled across the chip. HBM won't be worth the cost. There's probably going to be another gpu called gp102 for GeForce and quadro variants with 12/24 gb g5x