Guru3D.com
  • HOME
  • NEWS
    • Channels
    • Archive
  • DOWNLOADS
    • New Downloads
    • Categories
    • Archive
  • GAME REVIEWS
  • ARTICLES
    • Rig of the Month
    • Join ROTM
    • PC Buyers Guide
    • Guru3D VGA Charts
    • Editorials
    • Dated content
  • HARDWARE REVIEWS
    • Videocards
    • Processors
    • Audio
    • Motherboards
    • Memory and Flash
    • SSD Storage
    • Chassis
    • Media Players
    • Power Supply
    • Laptop and Mobile
    • Smartphone
    • Networking
    • Keyboard Mouse
    • Cooling
    • Search articles
    • Knowledgebase
    • More Categories
  • FORUMS
  • NEWSLETTER
  • CONTACT

New Reviews
Far Cry New Dawn PC graphics performance benchmark review
WD Black SN750 NVME SSD (1TB) Review
Battlefield V: DLSS PC Performance Update
Metro Exodus: PC graphics performance benchmarks
Team Group Delta S TUF RGB SSD Review
T-Force Delta TUF Gaming RGB Memory Review
AMD Radeon VII 16 GB review
DeepCool Captain 240 PRO review
Guru3D Rig of the Month - January 2019
Combo deal: Office 2016 Pro and W10 for $34

New Downloads
Display Driver Uninstaller Download version 18.0.0.9
AMD Radeon Adrenalin Edition 19.2.2 driver Download
GeForce 418.91 WHQL driver download
3DMark Download v2.8.6446 + Port Royale
Prime95 download version 29.5 build 10
Corsair Utility Engine Download (iCUE) Download v3.12.118
Cinebench R15 Extreme Edition Download
AMD Radeon Adrenalin Edition 19.2.1 driver Download
GeForce 418.81 WHQL driver download
ATTO Disk Benchmark download v4.00.0f2


New Forum Topics
How to find out if a program locked up? RX Vega Owners Thread, Tests, Mods, BIOS & Tweaks ! Intel Flagship Core i9-9990XE has only 14 activated cores, not 18 Monitor LG 29UM68-P (G-SYNC stutter) Low GPU usage , low fps in some games How to boot old drive in new system Review: Far Cry New Dawn PC graphics benchmark analysis Download: Guru3D RTSS Rivatuner Statistics Server 7.2.0 Unreal Engine adds support for raytracing MSI Afterburner causing mouse input lag?




Guru3D.com » News » NVIDIA Announces Tesla T4 Tensor Core GPU

NVIDIA Announces Tesla T4 Tensor Core GPU

by Hilbert Hagedoorn on: 09/13/2018 08:00 AM | source: | 12 comment(s)
NVIDIA Announces Tesla T4 Tensor Core GPU

Fueling the growth of AI services worldwide, NVIDIA today launched an AI data center platform that delivers the industry’s most advanced inference acceleration for voice, video, image and recommendation services.

The NVIDIA TensorRT Hyperscale Inference Platform features NVIDIA Tesla T4 GPUs based on the company’s breakthrough NVIDIA Turing™ architecture and a comprehensive set of new inference software.

Delivering the fastest performance with lower latency for end-to-end applications, the platform enables hyperscale data centers to offer new services, such as enhanced natural language interactions and direct answers to search queries rather than a list of possible results.

“Our customers are racing toward a future where every product and service will be touched and improved by AI,” said Ian Buck, vice president and general manager of Accelerated Business at NVIDIA. “The NVIDIA TensorRT Hyperscale Platform has been built to bring this to reality — faster and more efficiently than had been previously thought possible.”

Every day, massive data centers process billions of voice queries, translations, images, videos, recommendations and social media interactions. Each of these applications requires a different type of neural network residing on the server where the processing takes place.

To optimize the data center for maximum throughput and server utilization, the NVIDIA TensorRT Hyperscale Platform includes both real-time inference software and Tesla T4 GPUs, which process queries up to 40x faster than CPUs alone.

NVIDIA estimates that the AI inference industry is poised to grow in the next five years into a $20 billion market.

Industry’s Most Advanced AI Inference Platform
The NVIDIA TensorRT Hyperscale Platform includes a comprehensive set of hardware and software offerings optimized for powerful, highly efficient inference. Key elements include:

  • NVIDIA Tesla T4 GPU – Featuring 320 Turing Tensor Cores and 2,560 CUDA® cores, this new GPU provides breakthrough performance with flexible, multi-precision capabilities, from FP32 to FP16 to INT8, as well as INT4. Packaged in an energy-efficient, 75-watt, small PCIe form factor that easily fits into most servers, it offers 65 teraflops of peak performance for FP16, 130 teraflops for INT8 and 260 teraflops for INT4.
  • NVIDIA TensorRT 5 – An inference optimizer and runtime engine, NVIDIA TensorRT 5 supports Turing Tensor Cores and expands the set of neural network optimizations for multi-precision workloads.
  • NVIDIA TensorRT inference server – This containerized microservice software enables applications to use AI models in data center production. Freely available from the NVIDIA GPU Cloud container registry, it maximizes data center throughput and GPU utilization, supports all popular AI models and frameworks, and integrates with Kubernetes and Docker.

Supported by Technology Leaders Worldwide
Support for NVIDIA’s new inference platform comes from leading consumer and business technology companies around the world.

“We are working hard at Microsoft to deliver the most innovative AI-powered services to our customers,” said Jordi Ribas, corporate vice president for Bing and AI Products at Microsoft. “Using NVIDIA GPUs in real-time inference workloads has improved Bing’s advanced search offerings, enabling us to reduce object detection latency for images. We look forward to working with NVIDIA’s next-generation inference hardware and software to expand the way people benefit from AI products and services.”

Chris Kleban, product manager at Google Cloud, said: “AI is becoming increasingly pervasive, and inference is a critical capability customers need to successfully deploy their AI models, so we’re excited to support NVIDIA’s Turing Tesla T4 GPUs on Google Cloud Platform soon.”

More information, including details on how to request early access to T4 GPUs on Google Cloud Platform, is available here.



NVIDIA Announces Tesla T4 Tensor Core GPU NVIDIA Announces Tesla T4 Tensor Core GPU




Rate this story
Rating:

« Fractal Design Define R6 Line Gets More Diverse · NVIDIA Announces Tesla T4 Tensor Core GPU · Forza Horizon 4 Demo Available »

Related Stories

NVIDIA Announces GeForce RTX 2070, 2080 and 2080 Ti - 08/20/2018 06:25 PM
It has been a long time coming, moments ago NVIDIA has announced the GeForce RTX 2080 and 2080 Ti. During its Gamescom 2018 event the CEO of Nvidia took the stage and announced the two Turing pr...

NVIDIA Announces Financial Results for Second Quarter Fiscal 2019 - 08/17/2018 10:07 AM
NVIDIA today reported revenue for the second quarter ended July 29, 2018, of $3.12 billion, up 40 percent from $2.23 billion a year earlier, and down 3 percent from $3.21 billion in the previous quart...

Nvidia announces Turing architecture for gpu's Quadro RTX8000, 6000, 5000 - 08/14/2018 07:17 AM
Nvidia has made a number of announcements over at SIGGRAPH, one of them. Turing architecture for GPUs. In addition, the company has announced a number of graphics cards, not intended for the consumer...

NVIDIA AIB Manli registers GA104-400 - Ampere? And Lists GeForce GTX 2070 and 2080 - 08/02/2018 04:03 PM
Holy moley, we're in for some gossip and chatter. Manli Technology Group, yes from the graphics cards, has a EEC certification upcoming products. And if you read closely you'll notice the mention ...

NVIDIA adds far more visual option to not install GeForce Experience - 07/16/2018 06:13 PM
Do you remember how Geforce Experience first was a feature, then an optional install and then pretty much became a mandatory installation toed to cloud logins like Facebook etc? Well, there always has...


Fox2232
Senior Member



Posts: 8375
Joined: 2012-07-20

#5584047 Posted on: 09/13/2018 08:17 AM
Marketing:
"Tesla T4 GPUs, which process queries up to 40x faster than CPUs alone."

Did they compare it to iP3@800MHz? Random C2Q? Or 28C/56T chilled core chip?

But I hope everyone sees how much they fit in... Someone should start getting info about sizes of each unit. I bet quite few people here would rather have GPU with 1/2 of SPs RTX2080Ti has and double tensor/RT cores. (That's if they can be decoupled.)

Valken
Senior Member



Posts: 1432
Joined: 2011-01-05

#5584058 Posted on: 09/13/2018 09:17 AM
They are probably talking about parallel or OpenCL type operations and not x86 code. In those cases, GPUs are vastly superior for highly paralleled compute code.

Fox2232
Senior Member



Posts: 8375
Joined: 2012-07-20

#5584059 Posted on: 09/13/2018 09:22 AM
They are probably talking about parallel or OpenCL type operations and not x86 code. In those cases, GPUs are vastly superior for highly paralleled compute code.

Once more, read slowly...
What are they comparing that GPU to? Till we know that, their entire statement is irrelevant.

Valken
Senior Member



Posts: 1432
Joined: 2011-01-05

#5584066 Posted on: 09/13/2018 09:45 AM
I don't want to argue with you but its right there in their statement. We can guess what data centers use so many processors, CPU or GPU streams for: AI data processing...

"AI data center platform that delivers the industry’s most advanced inference acceleration for voice, video, image and recommendation services."

It says on a cloud with queries, lets say AI derived web searchs or photo retouching or map generation or scientific calculations, those are highly paralleled tasks and as per my statement, GPUs in general, regardless of any vendors are usually better than x86 CISC CPUs, multicore, multithread or not. Hell, better than RISC CPUs as well due to pure scaling of stream "processors".

Google is listed as a test case for trial runs... No need to be hostile about a discussion.

Fox2232
Senior Member



Posts: 8375
Joined: 2012-07-20

#5584067 Posted on: 09/13/2018 09:54 AM
I don't want to argue with you but its right there in their statement. We can guess what data centers use so many processors, CPU or GPU streams for: data processing...

It says on a cloud with queries, lets say search engine lookups, those are highly paralleled tasks and as per my statement, GPUs in general, regardless of any vendors are usually better than x86 CISC CPUs, multicore, multithread or not. Hell, better than RISC CPUs as well due to pure scaling of stream "processors".

Google is listed as a test case for trial runs... No need to be hostile about a discussion.
CPU in my cellphone is up to 200times faster than PC CPU. It is true, but moment you get what CPU I compared it to...
yes, this statement I made is baseless because it makes comparison to "smoke" same way nVidia did.

If I was to guess what is relevant comparison in case no contestants in comparison are shown, it would be 1CU in GPU vs 1CU in CPU. But I am sure that they did compare quite a few SMs (CUs) to something like intel's 4C/8T low power server chip.
In best case scenario (as close to relevant as possible) they took some 75W server CPU and compared it to their 75W GPU in some very specific workload where that GPU excels and CPUs generally suck.

asder
Junior Member



Posts: 17
Joined: 2018-04-30

#5584069 Posted on: 09/13/2018 09:58 AM
It's at the bottom of the slide Xeon Gold 6140

NaturalViolence
Member



Posts: 66
Joined: 2009-10-01

#5584115 Posted on: 09/13/2018 01:02 PM
Oh thank christ somebody here actually reads the source material before commenting.

tunejunky
Senior Member



Posts: 798
Joined: 2017-08-18

#5584150 Posted on: 09/13/2018 03:21 PM
i'm laughing so hard right now. of course Google is the target...and any other big data/AI company.
they can fill an entire production run just from companies in a 40km radius from where i live.

this is the money market, they will receive higher priority than any AIB gaming manufacturer by a factor of at least two in shipments. cost is literally no object (from a gamer perspective) to these companies as they're paid by the data they crunch, mo' data mo' money.

i would even go so far as to say the gaming market is tertiary to Nvidia at this point, as they are making more money from automotive than AIB's.

Stefem
Member



Posts: 31
Joined: 2016-07-19

#5584167 Posted on: 09/13/2018 04:17 PM
It's at the bottom of the slide Xeon Gold 6140
Actually reading before commenting seems too much for many nowadays

Stefem
Member



Posts: 31
Joined: 2016-07-19

#5584172 Posted on: 09/13/2018 04:28 PM
i'm laughing so hard right now. of course Google is the target...and any other big data/AI company.
they can fill an entire production run just from companies in a 40km radius from where i live.

this is the money market, they will receive higher priority than any AIB gaming manufacturer by a factor of at least two in shipments. cost is literally no object (from a gamer perspective) to these companies as they're paid by the data they crunch, mo' data mo' money.

i would even go so far as to say the gaming market is tertiary to Nvidia at this point, as they are making more money from automotive than AIB's.
It does not, more than half of the revenue of NVIDIA came from gaming alone

Caesar
Senior Member



Posts: 382
Joined: 2015-06-18

#5584188 Posted on: 09/13/2018 05:16 PM
.............the first customer (one of them) is GOOGLE....

Check this: early access program
-----------------------------------------------------------------------------------------------
“AI is becoming increasingly pervasive, and inference is a critical capability customers need to successfully deploy their AI models,” said Chris Kleban, product manager at Google Cloud, “so we’re excited to support
NVIDIA’s Turing Tesla T4 GPUs
on Google Cloud Platform soon.”

Source:
https://nvidianews.nvidia.com/news/new-nvidia-data-center-inference-platform-to-fuel-next-wave-of-ai-powered-services


tunejunky
Senior Member



Posts: 798
Joined: 2017-08-18

#5584532 Posted on: 09/14/2018 03:58 PM
It does not, more than half of the revenue of NVIDIA came from gaming alone


if you analyze the trends and the marketplace, that gaming revenue is sliding more rapidly than the growth of the two segments i mentioned. furthermore, the profitability is far greater in every other segment. in 2019 the graph chart will have flipped.

AIB sales are down and they've been trending down for years as simpler games (i.e. android/ iOS) dominate the gaming market by revenue. concurrently, Big Data discovered Pascal gpu's (by the thousands) as did A.I. and the auto industries.

as an investor invested in Nvidia (with Intel, TSMC, Qualcomm and AMD) i keep abreast of these things.
go read some quarterly reports and listen in on earnings calls first.

Post New Comment
Click here to post a comment for this news story on the message forum.


Guru3D.com © 2019