Nvidia might be moving to Multi-Chip-Module GPU design
With Moore's law becoming more difficult each year technology is bound to change. At one point it will be impossible to shrink transistors even further, hence companies like Nvidia already are thinking about new methodologies and technologies to adapt to that. Meet the Multi-Chip-Module GPU design.
Nvidia published a paper that shows how they can connect multiple parts (GPU modules) with an interconnect. According to the research, this will allow for bigger GPUs with more processing power. Not only will is help tackling the common problems, it would also be cheaper to achieve as fabbing four dies that you connect is cheaper to do than to make one huge monolithic design.
Thinking about it, AMD is doing exactly this with Threadripper and EPYC processors where they basically connect two to four Summit Ridge (ZEN) dies with that wide PCIe lane link (they use 64 PCie lanes per link with 128 available), Infinity Fabric.
According to the researchers, as an example a GPU with four GPU modules they recommend three architecture optimizations that will allow for minimal loss off data-communication in-between the different modules. According to the paper the loss in performance compared to a monolithic single die chip would be merely 10%
Of course when you think about it, in essence SLI is already a similar methodology (not technology), however as you guys know it can be rather inefficient and challenging in scaling and compatibility. The paper states this MCM design would be performing 26.8% better compared to any multi-GPU solution. If and when Nvidia is going to fab MCM multi GPU module based chips is not known, for now this is just a paper on the topic. The fact that they publish it indicates it is bound to happen at one point in time though.
Sorry, I could not resist ... ;)
Rumor: Nvidia Mobile Pascal-GPUs during Computex 2016 - not Desktop - 02/26/2016 01:31 PM
It's been topic for discussion for a while now. Personally I think we'll see some soft of announcement in April during the GTC, and later on a broad announcement in the Computex timeframe. Likely, a...
Nvidia might release GTX 980MX and 970MX for laptops - 01/19/2016 10:26 AM
Nvidia is likely planning the successors to the 970M and 980M in the 2nd half of 2016. The successors will be the 980MX and 970MX and are based on GM204....
Nvidia might be working on their own VR headset - 06/07/2015 01:03 PM
Nvidia might release their own VR headset, this is now speculated as information surfaced showing that Nvidia holds a patent for a headset with six camera and two displays, each for one eye....
Microsoft Confirms DirectX 12 MIX AMD and Nvidia Multi-GPUs - 03/13/2015 08:38 AM
It's not exactly new news, but Microsoft actually confirms it at this stage. Microsoft technical support states that DirectX 12 will support “multi-GPU configurations between Nvidia and AM...
Nvidia Maxwell GM200 Flagship GPU caught on photo - 01/16/2015 04:10 PM
NVIDIA’s flagship Maxwell GPU (to be released) GM200 GPU core has been spotted and somebody took a photo. The product will end up in Quadro and hopefully GeForce graphics cards. The card tha...
Senior Member
Posts: 6125
Joined: 2003-07-26
both companies need to mask the amount of gpus from the OS driver level so the system only sees 1 and the onboard bios of the gpu decides out how the gpu dishes out the utilization otherwise we will be stuck waiting and hoping the developers figure it out
Same goes for CPU i really want to find the documents on this it was discussed way back in mid 2000's how its possible but no one wants to do it..
and from what I can find it has been done back in the earlier days aka voodoo and someother company forget which one where os and drivers only seen it as 1
Senior Member
Posts: 686
Joined: 2010-03-27
Hello Voodoo 5 ?
Senior Member
Posts: 1781
Joined: 2003-10-27
Good stuff, no more masses of video cards crammed into cases overheating all over the place.
This could be soo good, SLI in the one card done properly, but I guess it hasn't happend due to the current tech limitations.
Senior Member
Posts: 1309
Joined: 2003-09-14
yay for a Voodoo 5

This is understandable, especially what with the Fab Processes being late, and farther and farther apart.
If you can't rely on reducing the size of the chips/transistors, the you have to go MCM, or Multi-chip-on-board a la Voodoo, for the High end cards.
even if its only for the datacenter GPU cards at first (they can pay for the R&D with the prices the cards sell for) and then the consumer will get a trickle down effect later on.
I suspect it hasn't been done lately, because they didn't plan to, officially.
It also means dedicating on-die space to something that won't be used in Single chip boards, which would be wasted.
I could see a medium sized GPU with a 128bit GDDR/X bus, or a Single HBM Stack (either 512bit or 1024bit wide) and adding up to four of them together on a single card.
The HBCC might be a good fit for this, since there would be one per chip, and one of them could become 'master' to the others, and give them orders.
NVLink is probably set up for this as well.
Senior Member
Posts: 3501
Joined: 2011-05-10
Thanks for the link, it's a great read. Looking forward to seeing what they achieve with their first commercial Multi-Chip-Module GPU.