The NVIDIA GTC 2018 Keynote Live Blog
by Ryan Smith & Nate Oh on March 27, 2018 11:50 AM EST12:02PM EDT - I'm here at the San Jose Convention Center for NVIDIA's 2018 GPU Technology Conference
12:03PM EDT - Kicking off this first full day of events is of course NVIDIA CEO Jen-Hsun Huang's keynote address, which always moves fast and is packed with major NVIDIA announcements
12:04PM EDT - The show is set to start at 9am, but it typically runs 10 minutes or so late
12:05PM EDT - Jen-Hsun has taken away the tables for the press this year, so we're all doing this on lap-tops, literally
12:05PM EDT - And here we go
12:07PM EDT - It's a long song, it seems
12:07PM EDT - “This is the greatest show”
12:08PM EDT - Joining me remotely today is AnandTech GPU Editor Nate Oh, who is covering photos
12:08PM EDT - WiFi connectivity at GTC is always spotty. But this year is proving worse than most
12:10PM EDT - Okay, and now the song is over and a video is rolling
12:10PM EDT - I think this may be an updated version of last year's video?
12:12PM EDT - "I am AI"
12:12PM EDT - And Jen-Hsun has now taken the stage
12:14PM EDT - Today's topics: amazing graphics, amazing science, amazing AI, and amazing robots
12:14PM EDT - First topic: graphics
12:14PM EDT - Looks like we're starting with photorealism and ray tracing
12:15PM EDT - (Captain Phasma may not have done anything of importance in the movies, but she sure shows up a lot in ray tracing talks thanks to that reflective chrome suit)
12:16PM EDT - Jen-Hsun is noting the amount of compute time required to render a single image for a film, never mind all of the frames
12:17PM EDT - NVIDIA wants to further close the gap between rasterization and ray tracing
12:18PM EDT - Jen-Hsun is going over all the tricks currently used to fake lighting in games/rasterization. Pre-baked lighting, AO shaders, etc
12:19PM EDT - As opposed to real reflections, real transparencies, real global illumination, etc
12:25PM EDT - Sorry for the interruption. WiFi is trying its best to die
12:26PM EDT - Jen-Hsun and his tech are going over everything involved in the Star Wars ray tracing tech demo from last week
12:27PM EDT - NVIDIA and ILM do this in real time on a single DGX Station. So 4 GPUs
12:27PM EDT - Confirming that this is in real time, moving the camera around to prove it
12:28PM EDT - Announcing NVIDIA RTX technology
12:28PM EDT - Perhaps we'll get a more hardware-focused look at last week's RTX announcement?
12:29PM EDT - Announciung Quadro GV100
12:30PM EDT - API support: DirectX Raytracing, NVIDIA OptiX, and Vulkan
12:30PM EDT - Quadro GV100 is the "world's largest GPU"
12:31PM EDT - (It looks like the same size as always)
12:31PM EDT - Confirming NVLink 2 support. 2 connectors on the board
12:31PM EDT - Discussing how two GV100 cards can be teamed together, including shared memory pools
12:32PM EDT - 5120 CUDA cores. 640 tensor cores
12:32PM EDT - 32GB of HBM2 memory
12:33PM EDT - Now back to rendering technology
12:37PM EDT - “The more GPUs to buy, the more money you save”
12:38PM EDT - Discussing the cost benefits of GPU rendering versus a large CPU farm
12:38PM EDT - Now talking about the rise of GPU computing
12:39PM EDT - Discusisng how GPU's have become much more flexible over the years
12:40PM EDT - "We're at the tipping point": of GPU computing
12:41PM EDT - The SJC Convention Center is a packed house. 8500 registrations
12:42PM EDT - "The world needs larger computers"
12:43PM EDT - Recapping various supercomputers and what projects they've worked on
12:44PM EDT - Fermi vs. Volta
12:45PM EDT - This is supposed to be a 5 year comparison, but I'd note that the Kepler-based Tesla K20 launched more than 5 years ago
12:46PM EDT - Moral of the story? GPU perf has thus far grown faster than Moore's Law (in most case)
12:50PM EDT - "The more you buy, the more you save" once again. This time reflecting the pace of the Tesla line and Volta V100 servers
12:50PM EDT - (WiFi gets overloaded, dies, dumps the problem machines, and comes back)
12:50PM EDT - Now talking about medical applications of GPU computing via medical imaging
12:52PM EDT - NVIDIA Project Clara. A "medical imaging supercomputer"
12:54PM EDT - Demo time
12:54PM EDT - EKG of a heart
12:55PM EDT - Showing how NVIDIA is taking the scan data off of a 15 year old machine and enhancing the resulting imagery
12:55PM EDT - Using algorithms to produce a volumetric representation
12:56PM EDT - Inferring where the left ventrical is in 3D basedon the 2D data
12:58PM EDT - Medical instruments have a long life cycle. NVIDIA wants to increase their usefulness by post-processing the data on GPUs to infer more data from the scans
12:59PM EDT - Shifting gears to AI
01:00PM EDT - "Deep learning has revolutionized modern AI"
01:01PM EDT - NVIDIA is doing everything they can to expand the platform (and capture the market)
01:03PM EDT - Volta V100 is getting a memory bump to 32GB of HBM2 per card
01:03PM EDT - Available now
01:03PM EDT - DGXs are getting the upgraded cards as well
01:03PM EDT - This isn't a new SKU to sit beside the 16GB SKUs. Tesla V100 is being replaced with 32GB cards in the market wholesale
01:05PM EDT - "The world wants a gigantic GPU"
01:05PM EDT - "The world's largest GPU"
01:05PM EDT - This is definitely a misnomer
01:05PM EDT - 16 Tesla V100s in a cluster together
01:06PM EDT - All connected together via a new item called the NVSwitch
01:06PM EDT - Unified memory space
01:06PM EDT - So this is all 16 cards in one system
01:07PM EDT - Aggregate HBM2 capacity of 512GB
01:07PM EDT - 2 PFLOPs
01:07PM EDT - That's 2 PFLOPs of tensor core throughput
01:09PM EDT - A new NVIDIA DGX server. "The world's largest GPU"
01:09PM EDT - NVSwitch has wB transistors, made on TSMC 12nm FFN
01:10PM EDT - 18 NVLInk ports per switch
01:10PM EDT - 900GB of full duplex bandwidth on a switch
01:10PM EDT - 12 switches on the new server
01:10PM EDT - Every GPU can communicate at 300GB/sec. 20x the bandwidth of PCIe
01:11PM EDT - "It's not a network. It's a switch" a fabric, to be precise
01:11PM EDT - So all the NVLink benefits of shared memory, coherency, etc are retained
01:11PM EDT - DGX-2
01:11PM EDT - 8x EDR IB/100 GigE
01:11PM EDT - 10 Kilowatt power consumption!
01:12PM EDT - 350 lbs
01:12PM EDT - 16 V100s backed by a pair of Xeon Platinum CPUs
01:12PM EDT - "No human can lift it"
01:14PM EDT - Uses a new NVLink plane card to connect two trays of Tesla V100s
01:16PM EDT - “This is what an engineer finds beautiful”
01:16PM EDT - (NV apparently filmed the blow-apart animation using stop-motion rather than CGI. Why? Because they could)
01:16PM EDT - 10x faster than DGX-1 in 6 months
01:17PM EDT - How much should we charge?
01:18PM EDT - Answer: $1,500,000
01:18PM EDT - Available in Q3
01:18PM EDT - Oh, no, $1.5M was a joke
01:18PM EDT - $399K is the real price
01:22PM EDT - "We are all in on deep learning"
01:22PM EDT - Comparing AlexNet training time on the GTX 580 versus the DGX-2
01:22PM EDT - Days versus minutes
01:23PM EDT - Now talking about NVIDIA's software container service, the NVIDIA GPU Cloud
01:24PM EDT - 30 different containers
01:24PM EDT - Google, Alibaba, AWS, and more support the containers
01:25PM EDT - "Plaster"
01:26PM EDT - "We need a computing model that dramatically reduces the energy requirements"
01:29PM EDT - “Inference is complicated”
01:29PM EDT - (This is an understatement, to ssy the least)
01:30PM EDT - Announcing a new batch of tools and libraries
01:30PM EDT - TensorRT 4.0
01:30PM EDT - Deep TensorFlow integration
01:31PM EDT - A fully trained network can be run right on NV's architecture in an optimized fashion
01:31PM EDT - TensorRT also adds support for the Onyx format and WinML
01:33PM EDT - Now discussing the various acceleration factors versus CPU-only execution
01:33PM EDT - (They're all big integers)
01:34PM EDT - "We're going to save a lot of money"
01:34PM EDT - Kubernetes on NVIDIA GPUs
01:34PM EDT - IMost people here don't know what Kubernetes is)
01:35PM EDT - An orchestration layer to allocate work over a datacenter
01:36PM EDT - Kubernetes is now GPU-aware, allowing large tasks to be allocated over multiple servers and supported at multiple cloud services
01:38PM EDT - Demoing TensorRT on one V100
01:38PM EDT - Now on 8 servers via Kubernetes in real time
01:39PM EDT - Not an 8x speedup, but still a lot faster
01:40PM EDT - Saturn V is up to 660 DGX-1 nodes
01:41PM EDT - Saturn V is up to 660 DGX-1 nodes
01:42PM EDT - Demoing resiliency
01:43PM EDT - And that's a wrap on AI inference
01:44PM EDT - Titan V is still officailly out of stock
01:44PM EDT - NV is working on it
01:45PM EDT - Now heaping praise on the NVIDIA Research group
01:46PM EDT - JHH's favorite project? Noise-to-noise denoising of ray tracing
01:48PM EDT - Another project: using GANs to synthesize realistic environments
01:50PM EDT - And now time to talk about autonomy and NV's self-driving projects
01:50PM EDT - "It's hard, it's really, really hard"
01:50PM EDT - But "this work is vitally important" as well
01:52PM EDT - Sensor fusion can overcome the limitations of individual sensor types
01:52PM EDT - But it's a hard problem. One that NV is working on
01:53PM EDT - Recapping the DRIVE platform and the end-to-end training process
01:54PM EDT - Recapping the DRIVE platform and the end-to-end training process
01:55PM EDT - “Simulation is going to be the foundation of success”
01:55PM EDT - The NVIDIA Perception Infrastructure
01:55PM EDT - Labeling 1mil images per month
01:56PM EDT - The end product of the system is networks
01:58PM EDT - (Ultimately this seems to be NV's response to last week's self-driving Uber incident)
01:58PM EDT - (Which, I hear from a source, was using NVIDIA DRIVE hardware; no idea whose algorithms though)
02:00PM EDT - All cars are mapping cars
02:01PM EDT - "We're going to work on this for another 2 or 3 years before we ship in volume cars"
02:01PM EDT - Now rolling a video
02:07PM EDT - Monitoring not only the car but the passengers as well
02:08PM EDT - Recapping the DRIVE hardware roadmap
02:09PM EDT - Both Xavier and Pegasus in production by the end of the year
02:09PM EDT - ASIL-D functional safety
02:10PM EDT - "We're not stopping here"
02:10PM EDT - Next generation DRIVE: Orin
02:11PM EDT - Get Pegasus down to a couple of small Orins
02:13PM EDT - Jen-hsun is now making the point that it would take many thousands of test cars to get to a fraction of the real world miles driven by humans
02:13PM EDT - Which makes it hard to do enough testing
02:13PM EDT - The solution? Simulation
02:14PM EDT - This looks to be a continuation of NV's simulation-driven training program revealed at GTC 2017
02:15PM EDT - The simulation needs to offer enough fidelity to properly tax the sensors
02:16PM EDT - Now showing the DRIVE system view
02:18PM EDT - Can use the artificial nature of the sim to ensure weird, atypical situations are much more frequent than in real life
02:20PM EDT - NVIDIA DRIVE Sim and Constellation, an AV validation system
02:20PM EDT - 10K Constallations drive 3B miles per year
02:22PM EDT - 370 partners on DRIVE
02:22PM EDT - Cars are the first, but robots are coming
02:23PM EDT - (Jen-Hsun is running long)
02:23PM EDT - NVIDIA will be relasing their Isaac robotics platofrm, first introduced at GTC 2017
02:24PM EDT - Rolling video
02:28PM EDT - One last thing. NV is demoing using VR to take control of an autonomous car
02:28PM EDT - This is a live demo
02:29PM EDT - The car is out back. The driver is on the stage
02:31PM EDT - This has been a sci-fi staple for years. But it still looks cool
02:33PM EDT - VR + Autonomous machines = teleportation
02:33PM EDT - Now recapping the keynote
02:36PM EDT - And that's a warp. Off to Q&A. Thanks for joining us.
27 Comments
View All Comments
lane42 - Tuesday, March 27, 2018 - link
What a way to run a company.............Ya sell a product that you don't have. No wonder the stock tanked today.
mode_13h - Wednesday, March 28, 2018 - link
Stock was hit because of self-driving-car backlash.mode_13h - Wednesday, March 28, 2018 - link
And they're primarily a chip company. In semiconductors, most of the sales cycle happens before you have working silicon.mode_13h - Wednesday, March 28, 2018 - link
So, what about NVDLA, or was that not announced in the keynote?I'm really interested in their strategy with that. All this talk about it being open source... are they planning to make money by selling TensorRT licenses for it, or something? Trained models? What's their angle?
Yojimbo - Wednesday, March 28, 2018 - link
Huang talked more about NVDLA at the Investor's Day. He said something along the lines of it was easy for them to make and they gave it away because they didn't see it as beneficial for them to try to build an ecosystem around it. He said it's available to be used for free, and the plan is that it will help others create IOT devices, which will create data that will need to be processed in data centers on NVIDIA's GPUs. He said ARM has picked it up for use in their AI chips (that has been sparsely reported on since the announcement by ARM a few days ago). So they aren't planning on directly monetizing the NVDLA in any way (except when it's sold in their own chips, I guess, such as Xavier).I doubt the plan is to charge anything for TensorRT. TensorRT is part of what one gets when one buys NVIDIA hardware. TensorRT allows that hardware to be used more effectively for inference much like CuDNN allows it to be used for training neural networks and graphics drivers allow it to be used for playing games. NVIDIA supports the NVDLA in TensorRT and I'd assume they will continue to do so, as it sounds like the NVDLA might be comparatively useless without it, or some equivalent to it.
mode_13h - Friday, March 30, 2018 - link
Thanks!mariajohn - Tuesday, December 18, 2018 - link
Sometime NVDIA NOT Support Rendering Software i'm designer i work daily basic NVidia should have to upgrade to improve rendering You can check here https://3danimationrendering.com/ my previews projects it's to heavy can you suggest me what can i do improve my speed?