Guru3D.com
  • HOME
  • NEWS
    • Channels
    • Archive
  • DOWNLOADS
    • New Downloads
    • Categories
    • Archive
  • GAME REVIEWS
  • ARTICLES
    • Rig of the Month
    • Join ROTM
    • PC Buyers Guide
    • Guru3D VGA Charts
    • Editorials
    • Dated content
  • HARDWARE REVIEWS
    • Videocards
    • Processors
    • Audio
    • Motherboards
    • Memory and Flash
    • SSD Storage
    • Chassis
    • Media Players
    • Power Supply
    • Laptop and Mobile
    • Smartphone
    • Networking
    • Keyboard Mouse
    • Cooling
    • Search articles
    • Knowledgebase
    • More Categories
  • FORUMS
  • NEWSLETTER
  • CONTACT

New Reviews
ASUS ROG Radeon RX 6750 XT STRIX review
AMD FidelityFX Super Resolution 2.0 - preview
Sapphire Radeon RX 6650 XT Nitro+ review
Sapphire Radeon RX 6950 XT Sapphire Nitro+ Pure review
Sapphire Radeon RX 6750 XT Nitro+ review
MSI Radeon RX 6950 XT Gaming X TRIO review
MSI Radeon RX 6750 XT Gaming X TRIO review
MSI Radeon RX 6650 XT Gaming X review
Deepcool AS500 PLUS CPU Cooler Review
Kioxia Exceria Pro 2 TB M.2 NVMe SSD Review

New Downloads
HWiNFO Download v7.24
GeForce 512.77 WHQL driver download
Intel HD graphics Driver Download Version: 30.0.101.1960
AMD Radeon Software Adrenalin 22.5.1 WHQL driver download
3DMark Download v2.22.7359 + Time Spy
Prime95 download version 30.8 build 15
AIDA64 Download Version 6.70
PCMark 10 Download v2.1.2556
GPU-Z Download v2.46.0
Display Driver Uninstaller Download version 18.0.5.0


New Forum Topics
AMD Zen 4 CPU with 5.2 GHz Boost and RDNA 2 iGPU surfaces Info Zone - gEngines, Ray Tracing, DLSS, DLAA, TSR, FSR, XeSS, DLDSR etc. NVIDIA GeForce 512.77 WHQL driver download & Discussion AMD Software Preview Driver May 2022 driver download and discussion Samsung Electronics Showcases New Era of Micro LED Technology AMD FidelityFX Super Resolution 2.0 - Deathloop preview The AMD Ryzen All In One Thread /Overclocking/Memory Speeds & Timings/Tweaking/Cooling Part 2 AMD Radeon Software Adrenalin 22.5.1 WHQL driver download and discussion [3rd-Party Driver] Amernime Zone Radeon Insight 22.4.1 WHQL Driver Pack (P/V/N 22.5.1 ...) Free to grab: Prey at Epic Games Store




Guru3D.com » News » NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads

NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads

by Hilbert Hagedoorn on: 04/13/2021 06:32 AM | source: | 16 comment(s)
NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads

NVIDIA announced its first data center CPU, an Arm-based processor that incorporates state-of-the-art graphics and memory, such as the LPDDR5X.

Nvidia is developing CPUs. With "Grace" aka Grace Hopper, based on ARM, the manufacturer wants to take on the competition from 2023 in the HPC segment. Grace will rely on next-gen ARM cores and should primarily solve bottlenecks in the connection to the GPU and memory. That should make Nvidia even stronger in high-performance computing. According to a first building blobk, Grace could have around 80 cores. 84 are visible or countable, eight of them are not clear as to what they are, 76 is also conceivable. However, the manufacturer does not provide any technical details for the time being. To what extent there will be different versions based on Grace, nothing is known about that either.

With this memory configuration, NVIDIA indicates that it is multiplied x2 the bandwidth compared to DDR4 memory, at the same time that it offers 10 times more energy efficiency. Also present will be the 4th Generation of NVIDIA NVLink interconnection technology, which will unite Grace processors with graphics at 900 GB / s, which means increasing the bandwidth of current servers by 30 times.

- NVIDIA - 

The result of more than 10,000 engineering years of work, the NVIDIA Grace CPU is designed to address the computing requirements for the world's most advanced applications—including natural language processing, recommender systems and AI supercomputing—that analyze enormous datasets requiring both ultra-fast compute performance and massive memory. It combines energy-efficient Arm CPU cores with an innovative low-power memory subsystem to deliver high performance with great efficiency.

"Leading-edge AI and data science are pushing today's computer architecture beyond its limits - processing unthinkable amounts of data," said Jensen Huang, founder and CEO of NVIDIA. "Using licensed Arm IP, NVIDIA has designed Grace as a CPU specifically for giant-scale AI and HPC. Coupled with the GPU and DPU, Grace gives us the third foundational technology for computing, and the ability to re-architect the data center to advance AI. NVIDIA is now a three-chip company."

Grace is a highly specialized processor targeting workloads such as training next-generation NLP models that have more than 1 trillion parameters. When tightly coupled with NVIDIA GPUs, a Grace CPU-based system will deliver 10x faster performance than today's state-of-the-art NVIDIA DGX -based systems, which run on x86 CPUs.

While the vast majority of data centers are expected to be served by existing CPUs, Grace—named for Grace Hopper, the U.S. computer-programming pioneer—will serve a niche segment of computing.

The Swiss National Supercomputing Centre (CSCS) and the U.S. Department of Energy's Los Alamos National Laboratory are the first to announce plans to build Grace-powered supercomputers in support of national scientific research efforts.

NVIDIA is introducing Grace as the volume of data and size of AI models are growing exponentially. Today's largest AI models include billions of parameters and are doubling every two-and-a-half months. Training them requires a new CPU that can be tightly coupled with a GPU to eliminate system bottlenecks.

NVIDIA built Grace by leveraging the incredible flexibility of Arm's data center architecture. By introducing a new server-class CPU, NVIDIA is advancing the goal of technology diversity in AI and HPC communities, where choice is key to delivering the innovation needed to solve the world's most pressing problems.

"As the world's most widely licensed processor architecture, Arm drives innovation in incredible new ways every day," said Arm CEO Simon Segars. "NVIDIA's introduction of the Grace data center CPU illustrates clearly how Arm's licensing model enables an important invention, one that will further support the incredible work of AI researchers and scientists everywhere."

Grace's First Adopters Push Limits of Science and AI
CSCS and Los Alamos National Laboratory both plan to bring Grace-powered supercomputers, built by Hewlett Packard Enterprise, online in 2023.

"NVIDIA's novel Grace CPU allows us to converge AI technologies and classic supercomputing for solving some of the hardest problems in computational science," said CSCS Director Prof. Thomas Schulthess. "We are excited to make the new NVIDIA CPU available for our users in Switzerland and globally for processing and analyzing massive and complex scientific datasets."

"With an innovative balance of memory bandwidth and capacity, this next-generation system will shape our institution's computing strategy," said Thom Mason, director of the Los Alamos National Laboratory. "Thanks to NVIDIA's new Grace CPU, we'll be able to deliver advanced scientific research using high-fidelity 3D simulations and analytics with datasets that are larger than previously possible."

Delivering Breakthrough Performance
Underlying Grace's performance is fourth-generation NVIDIA NVLink interconnect technology, which provides a record 900 GB/s connection between Grace and NVIDIA GPUs to enable 30x higher aggregate bandwidth compared to today's leading servers.

Grace will also utilize an innovative LPDDR5x memory subsystem that will deliver twice the bandwidth and 10x better energy efficiency compared with DDR4 memory. In addition, the new architecture provides unified cache coherence with a single memory address space, combining system and HBM GPU memory to simplify programmability.

Grace will be supported by the NVIDIA HPC software development kit and the full suite of CUDA and CUDA-X libraries, which accelerate more than 2,000 GPU applications, speeding discoveries for scientists and researchers working on the world's most important challenges.

Availability is expected in the beginning of 2023.



NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads




« Download: MSI Afterburner 4.6.4 Beta 2 · NVIDIA Announces ARM based Grace CPU for Giant AI and High Performance Computing Workloads · Spire Launches Value Eagleforce 80 Plus Certified Power Supplies »

Related Stories

Nintendo Switch Pro Based on 5nm NVIDIA Ada Lovelace SoC - 03/25/2021 02:28 PM
Nintendo has plans to launch their Switch Pro at the end of this year with support for DLSS technology, a well-known leaker indicated that the company's new console would make use of an NVIDIA chip w...

NVIDIA Announces Financial Results for Fourth Quarter and Fiscal 2021 - 02/25/2021 09:40 AM
NVIDIA reported a record revenue for the fourth quarter ended January 31, 2021, of $5.00 billion, up 61 percent from $3.11 billion a year earlier, and up 6 percent from $4.73 billion in the previous q...

NVIDIA Announces GeForce RTX 3060 Graphics cards (updated) - 02/15/2021 01:11 PM
And that would be the non-Ti model. The Ampere card series will be based on 12GB, much like the leaks have indicated. So that also means a 6Gb version is inbound at a certain point in time. ...

NVIDIA and AMD Graphics card shortages to last thoughout H1 2021 - 01/13/2021 06:19 PM
It has been a pain for many; if you can even get a new graphics card, they are too expensive due to low availability. Both AMD and NVIDIA have extensive shortages for numerous reasons. Both companies ...

Recap: NVIDIA at CES 2021 RTX Mobile, RTX 3060, ReSize BAR and new RTX Games - 01/13/2021 08:03 AM
NVIDIA held its conference yesterday at a CES 2021 event, there were no surprises. The corporation announced its RTX 30 Mobile series, a 12GB GeForce RTX 3060, and gave more details on the arrival of ...


4 pages 1 2 3 4


Lowice
Member



Posts: 57
Joined: 2014-01-24

#5904345 Posted on: 04/13/2021 09:23 AM
Add it to the new Nvidia Shield TV and I'll buy it :)

DeskStar
Senior Member



Posts: 1307
Joined: 2011-01-11

#5904378 Posted on: 04/13/2021 11:45 AM
Wholly WOWO!!!

This is truly something to behold!!

10,000 years of engineering!!! Amazing how we can use compute power of today in order to attain information of tomorrow!!

Love to see which CPU Nvidia is using to only get 200GB/s of throughput?!?

Ummmmmmmm..... Pretty sure my 3900X is capable of handling information in the realm of thousands of gigabytes per second?!?!

Maybe I'm seeing something differently?!?

Maybe?

theoneofgod
Senior Member



Posts: 4652
Joined: 2014-01-17

#5904390 Posted on: 04/13/2021 12:58 PM
Wholly WOWO!!!

This is truly something to behold!!

10,000 years of engineering!!! Amazing how we can use compute power of today in order to attain information of tomorrow!!

Love to see which CPU Nvidia is using to only get 200GB/s of throughput?!?

Ummmmmmmm..... Pretty sure my 3900X is capable of handling information in the realm of thousands of gigabytes per second?!?!

Maybe I'm seeing something differently?!?

Maybe?

900 GB/s connection between Grace and NVIDIA GPUs to enable 30x higher aggregate bandwidth compared to today's leading servers

Denial
Senior Member



Posts: 13711
Joined: 2004-05-16

#5904392 Posted on: 04/13/2021 01:06 PM
Wholly WOWO!!!

This is truly something to behold!!

10,000 years of engineering!!! Amazing how we can use compute power of today in order to attain information of tomorrow!!

Love to see which CPU Nvidia is using to only get 200GB/s of throughput?!?

Ummmmmmmm..... Pretty sure my 3900X is capable of handling information in the realm of thousands of gigabytes per second?!?!
Maybe I'm seeing something differently?!?

Maybe?

They are talking about the bandwidth between interfaces, not cache bandwidth. Also why is your picture so massive.

Noisiv
Senior Member



Posts: 8175
Joined: 2010-11-16

#5904401 Posted on: 04/13/2021 01:49 PM

Ummmmmmmm..... Pretty sure my 3900X is capable of handling information in the realm of thousands of gigabytes per second?!?!



I am not.

The poor thing needs 4,000 pixels of height for a single Aida64 screenie.

4 pages 1 2 3 4


Post New Comment
Click here to post a comment for this news story on the message forum.


Guru3D.com © 2022