Top Banner
Pedro Mario Cruz e Silva Solutions Architect Manager, Latin América | Global Energy Team "ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS"
86

ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

Jul 31, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

Pedro Mario Cruz e SilvaSolutions Architect Manager, Latin América | Global Energy Team

"ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS"

Page 2: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

2

200B CORE HOURS OF LOST SCIENCEData Center Throughput is the Most Important Thing for HPC

Source: NSF XSEDE Data: https://portal.xsede.org/#/galleryNU = Normalized Computing Units are used to compare compute resources across supercomputers and are based on the result of the High Performance LINPACK benchmark run on each system

0

50

100

150

200

250

300

350

400

2009 2010 2011 2012 2013 2014 2015

Computing Resources Requested

Computing Resources Available

Norm

alized U

nit

(Billions)

National Science Foundation (NSF XSEDE) Supercomputing Resources

Page 3: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

3

1980 1990 2000 2010 2020

GPU-Computing perf

1.5X per year

1000X

by

2025

RISE OF GPU COMPUTING

Original data up to the year 2010 collected and plotted by M. Horowitz, F. Labonte, O. Shacham, K.

Olukotun, L. Hammond, and C. Batten New plot and data collected for 2010-2015 by K. Rupp

102

103

104

105

106

107

Single-threaded perf

1.5X per year

1.1X per year

APPLICATIONS

SYSTEMS

ALGORITHMS

CUDA

ARCHITECTURE

Page 4: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

4

1

10

100

1000

Mar-12 Mar-13 Mar-14 Mar-15 Mar-16 Mar-17 Mar-18

Re

lati

ve

Pe

rfo

rm

an

ce

Mar-19

2013

BEYOND MOORE’S LAW

Base OS: CentOS 6.2

Resource Mgr: r304

CUDA: 5.0

Thrust: 1.5.3

2019

Accelerated Server

With FermiAccelerated Server

with Volta

NPP: 5.0

cuSPARSE: 5.0

cuRAND: 5.0

cuFFT: 5.0

cuBLAS: 5.0

Base OS: Ubuntu 16.04

Resource Mgr: r384

CUDA: 10.0

NPP: 10.0

cuSPARSE: 10.0

cuSOLVER: 10.0

cuRAND: 10.0

cuFFT: 10.0

cuBLAS: 10.0

Thrust: 1.9.0

Progress Of Stack In 6 Years

GPU-Accelerated Computing

CPU

Moore’s Law

2013 2014 2015 2016 2017 2018 2019March

Rela

tive P

erf

orm

ance

Page 5: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

5

APPS &FRAMEWORKS

CUDA-XNVIDIA SDK & LIBRARIES)

NVIDIA DATA CENTER PLATFORMSingle Platform Drives Utilization and Productivity

VIRTUAL GPU

CUDA & CORE LIBRARIES - cuBLAS | NCCL

DEEP LEARNING

cuDNN

HPC

cuFFTOpenACC

+550 Applications

Amber

NAMD

CUSTOMER USE CASES

VIRTUAL GRAPHICS

Speech Translate Recommender

SCIENTIFIC APPLICATIONS

Molecular Simulations

WeatherForecasting

SeismicMapping

CONSUMER INTERNET & INDUSTRY APPLICATIONS

ManufacturingHealthcare Finance

MACHINE LEARNING

cuMLcuDF cuGRAPH cuDNN CUTLASS TensorRTvDWS vPC

Creative & Technical

Knowledge Workers

vAPPS

+600 Applications

TESLA GPUs & SYSTEMS

SYSTEM OEM CLOUDTESLA GPU NVIDIA HGXNVIDIA DGX FAMILY

Page 6: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

6

DEEP LEARNING

Page 7: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

7

LEARNING FROM DATAAND SOME BUZZ WORDS

ARTIFICALINTELLIGENCE

MACHINELEARNING DEEP

LEARNING

Knowledge & Reason

Learning

Planning

Communicating

Perceiving

Learning from data

Expert systems

Handcrafted features

Learning from data

Neural networks

Computer learned features

Page 8: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

8

A NEW COMPUTING MODEL

“Label”

Input

Training Data

Output

Trained NeuralNetwork

Trained NeuralNetwork

“Label”

OutputInput

TRAINING

INFERENCE

Page 9: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

9

A NEW COMPUTING MODELOutperform experts, facts, rules with software that writes software

Deep Learning Object DetectionDNN + Data + GPU

Traditional Computer VisionExperts + Time

Deep Learning Achieves “Superhuman” Results

0%

10%

20%

30%

40%

50%

60%

70%

80%

90%

100%

2009 2010 2011 2012 2013 2014 2015 2016

Traditional CV

Deep Learning

ImageNet

Page 10: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

10

TESLA REVOLUTIONIZES DEEP LEARNING

GOOGLE BRAIN APPLICATION

BEFORE TESLA AFTER TESLA

Cost $5,000K $200K

Servers 1,000 Servers 16 Tesla Servers

Energy 600 KW 4 KW

Performance 1x 6x

Page 11: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

11

Page 12: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

12

NEW AI DRIVING

Training on DGX-1

Driving with DriveWorks

KALDI

LOCALIZATION

MAPPING

DRIVENET

DAVENET

NVIDIA DGX-1 NVIDIA DRIVE PX

WATCH VIDEO

Page 13: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

13

NVIDIA DRIVE PEGASUSFirst AI Computer to Make Robotaxis a Reality

WATCH VIDEO

Page 14: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

14

Case Study: BERT(Bidirectional Encoder Representations from Transformer)

Page 15: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

15

Deep Learning Models Increasing in Complexity

Image Recognition

NLP

NLP – Generative Tasks

ChatbotsE-mail auto-completionDocument Summarization

Autonomous VehiclesSocial TaggingVisual Search

Q&ASentimentTranslation

1.5Bn

26M340M

Next-Level Use-Cases Require Gigantic Models

https://github.com/NVIDIA/Megatron-LM

Project Megatron

8.3B parameters

8-way Model Parallel

64-way Data Parallel

24x larger than BERT

Speech Recognition

Translation

Object Detection

Page 16: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

16

NVIDIA BREAKS RECORDS IN AI PERFORMANCEAt Scale and Per Accelerator

Record Type Benchmark Record

Max Scale

(Minutes To

Train)

Object Detection (Heavy Weight) Mask R-CNN 18.47 Mins

Translation (Recurrent) GNMT 1.8 Mins

Reinforcement Learning (MiniGo) 13.57 Mins

Per Accelerator

(Hours To Train)

Object Detection (Heavy Weight) Mask R-CNN 25.39 Hrs

Object Detection (Light Weight) SSD 3.04 Hrs

Translation (Recurrent) GNMT 2.63 Hrs

Translation (Non-recurrent)Transformer 2.61 Hrs

Reinforcement Learning (MiniGo) 3.65 Hrs

Per Accelerator comparison using reported performance for MLPerf 0.6 NVIDIA DGX-2H (16 V100s) compared to other submissions at same scale except for MiniGo where NVIDIA DGX-1 (8 V100s) submission was

used| MLPerf ID Max Scale: Mask R-CNN: 0.6-23, GNMT: 0.6-26, MiniGo: 0.6-11 | MLPerf ID Per Accelerator: Mask R-CNN, SSD, GNMT, Transformer: all use 0.6-20, MiniGo: 0.6-10

Page 17: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

Enterprise NLP Trend

Unstructured content represents as much as 80%of enterprise information resources.

A recent Gartner Research Circle survey on data and analytics trends shows that organizations are actively developing text analytics as part of their data and analytics strategies.

80% of survey respondents either have text analytics in use or plan to use it within the next two years.

Page 18: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

“BERT is a method of pre-training language representations, meaning that we train a general-purpose "language

understanding" model on a large text corpus (like Wikipedia), and then use that model for downstream NLP tasks that we

care about (like question answering).

BERT outperforms previous methods because it is the first unsupervised, deeply bidirectional system for pre-training NLP.”

18

BERT: Flexibility + Accuracy for NLP Tasks

Super Human Question & Answering

9th October, Google submitted GLUE benchmark

● Sentence Pair Classification: MNLI, QQP, QNLI, STS-B, MRPC, RTE, SWAG

● Single Sentence Classification: SST-2, CoLA

● Question Answering: SQuAD

● Single Sentence Tagging: CoNLL - 2003 NER

Page 19: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

19

DGX REFERENCE ARCHITECURES

Page 20: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

TESLA V100 TENSOR CORE GPUWorld’s Most Powerful Data Center GPU

5,120 CUDA cores

640 NEW Tensor cores

7.8 FP64 TFLOPS | 15.7 FP32 TFLOPS

| 125 Tensor TFLOPS

20MB SM RF | 16MB Cache

32 GB HBM2 @ 900GB/s |

300GB/s NVLink

Page 21: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

21

TENSOR CORE4x4x4 matrix multiply and accumulate

Page 22: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

22

Page 23: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

23

NVSWITCHWorld’s Highest Bandwidth On-node Switch

7.2 Terabits/sec or 900 GB/sec

18 NVLINK ports | 50GB/s per

port bi-directional

Fully-connected crossbar

2 billion transistors |

47.5mm x 47.5mm package

Page 24: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

24

NVIDIA DGX-2THE LARGEST GPU EVER CREATED

2 PFLOPS | 512GB HBM2 | 10 kW | 350 lbs

Page 25: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

25

40 PetaFLOPS Peak FP64 Performance | 660 PetaFLOPS DL FP16 Performance | 660 NVIDIA DGX-1 Server Nodes

ANNOUNCING NVIDIA SATURNV WITH VOLTA

DGX-PODNVIDIA SATURNV WITH VOLTA

Page 26: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

26

NVIDIA DGX SUPERPOD

Mellanox EDR 100G InfiniBand Network

Mellanox Smart Director Switches

In-Network Computing Acceleration Engines

Fast and Efficient Storage Access with RDMA

Up to 130Tb/s Switching Capacity per Switch

Ultra-Low Latency of 300ns

Integrated Network Manager

Terabit-Speed InfiniBand Networking per Node

Rack 1 Rack 16

ComputeBackplane

Switch

Storage Backplane

Switch

64 DGX-2

GPFS

200 Gb/s per node

800 Gb/s per node

White paper: https://nvidia.highspot.com/items/5d073ad681171721086b2788

Page 27: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

27

PLATFORM FOR AI INFERENCE

Page 28: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

28

320 Turing Tensor Cores

2,560 CUDA Cores

65 FP16 TFLOPS | 130 INT8 TOPS | 260 INT4 TOPS

16GB | 320GB/s

70 W

TESLA T4WORLD’S MOST EFFICIENT GPU FOR MAINSTREAM SERVERS

Page 29: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

29

NEW TURING TENSOR CORE

MULTI-PRECISION FOR AI TRAINING AND INFERENCE

65 TFLOPS FP16 | 130 TeraOPS INT8 | 260 TeraOPS INT4

Page 30: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

30

Kernel

Auto-TuningOptimal kernels selected

by activation precision

Layer &

Tensor Fusion

Dynamic Tensor

MemoryEfficient usage by GPU

Precision

Selection FP32, FP16, INT32

CalibrationINT8

NVIDIA TensorRT 5 INFERENCE PLATFORM

Accelerates Throughput On Leading Industry Platforms

Embedded

Automotive

Data center

Jetson

Drive

Tesla

TESLA V100

DRIVE PX 2

TESLA P4

JETSON TX2

NVIDIA DLA

Optimizer Runtime

TensorRT

FRAMEWORKS GPU PLATFORMS

Page 31: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

31

NVIDIA GPUs IN THE CLOUDAVAILABLE ON-DEMAND FROM THE TOP CLOUD SERVICE PROVIDERS

• Immediate access to NVIDIA GPU

infrastructure for data science in the

cloud

• Wide variety of deployment and

management options using

containers, Kubernetes, Kubeflow,

support for cloud native services, and

more

Page 32: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

32

NGCNVIDIA GPU CLOUD

OPTIMIZED CONTAINERS

Page 33: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

33

NGC: GPU-OPTIMIZED SOFTWARE HUBReady-to-run GPU Optimized Software, Anywhere

50+ ContainersDL, ML, HPC

60 Pre-trained ModelsNLP, Image Classification, Object Detection & more

Industry WorkflowsMedical Imaging, Intelligent Video

Analytics

15+ Model Training ScriptsNLP, Image Classification, Object Detection &

more

NGC

CloudOn-prem Hybrid Cloud Multi-cloud

Page 34: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

34

NVIDIA GPU CLOUDDeep Learning Containers

docker pull nvcr.io/nvidia/pytorch:19.09-py3nvidia-docker run nvcr.io/nvidia/pytorch:19.09-py3

Page 35: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

35

NVIDIA GPU CLOUDMachine Learning Containers

docker pull nvcr.io/nvidia/rapidsai/rapidsai:0.9-cuda10.0-runtime-ubuntu18.04nvidia-docker run nvcr.io/nvidia/rapidsai/rapidsai:0.9-cuda10.0-runtime-ubuntu18.04

Page 36: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

36

EDGE COMPUTING

Page 37: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

37

JETSON SUCCESS STORIES

Industrial Aerospace/Defense Healthcare Construction Agriculture Smart City

Inspection ServiceRetail Logistics Inventory Mgmt Delivery

Page 38: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

38

ANNOUNCING: JETSON NANOSmall, low-power AI Computer

128 CUDA Cores | 4 Core CPU

4 GB Memory

472 GFLOPs

70x45mm

5W | 10W

$129

Page 39: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

39

JETSON TX1 → JETSON TX2 4 GB7 - 15W

1 – 1.3 TFLOPS (FP16)50mm x 87mm

$299

JETSON TX2 8GB | Industrial7 – 15W

1.3 TFLOPS (FP16)50mm x 87mm

$399 - $749

JETSON AGX XAVIER10 – 30W

10 TFLOPS (FP16) | 32 TOPS (INT8)100mm x 87mm

$1099

JETSON NANO5 - 10W

0.5 TFLOPS (FP16)45mm x 70mm

$129

THE JETSON FAMILYFrom AI at the Edge to Autonomous Machines

Multiple devices - Same software

AI at the edge Fully autonomous machines

Listed prices are for 1000u+ | Full specs at developer.nvidia.com/jetson

Page 40: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

40

JETSON NANO DEVELOPER KITAI Computer

128 CUDA Cores | 4 Core CPU

472 GFLOPs

5W | 10W

Available from nvidia.com and

distributors worldwide

Page 41: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

41

DIGITAL SCIENCEHPC + AI + DATA

Page 42: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

42

Mixed-Precision Computing

TENSOR CORES FOR SCIENCE

7.815.7

125

0

20

40

60

80

100

120

140

V100 TFLOPS

FP64+ MULTI-PRECISIONPLASMA FUSION APPLICATION

FP16 Solver

3.5x faster

EARTHQUAKE SIMULATION

FP16-FP21-FP32-FP64

25x faster

MIXED PRECISION WEATHER PREDICTION

FP16/FP32/FP64

4x faster

Page 43: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

43

NVIDIA POWERS WORLD'S FASTEST SUPERCOMPUTER

27,648Volta Tensor Core GPUs

Summit Becomes First System To Scale The 100 Petaflops Milestone

122 PF 3 EFHPC AI

Page 44: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

44

NVIDIA POWERS TODAY’S FASTEST SUPERCOMPUTERS

22 of Top 25 Greenest

Piz DaintEurope’s Fastest

5,704 GPUs| 21 PF

ORNL SummitWorld’s Fastest

27,648 GPUs| 149 PF

Total Pangea 3Fastest Industrial

3,348 GPUs| 18 PF

ABCIJapan’s Fastest

4,352 GPUs| 20 PF

LLNL SierraWorld’s 2nd Fastest

17,280 GPUs| 95 PF

Page 45: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

45

NVIDIA POWER GORDON BELL WINNERS & 5 OF 6 FINALISTS

GPU Acceleration Critical To HPC At Scale Today

Material Science300X HigherPerformance

Genomics 2.36 ExaFLOPS

Seismic1st Soil & Structure

Simulation

Quantum Chromodynamics

<1% of Uncertainty Margin

Weather1.13 ExaFLOPS

Winner Winner

Page 46: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

46

Petrobras’s supercomputer Fênix is among

the world’s 500 biggest computers and ranks

first in Latin America. The list was compiled

by Top500.org, based on the machines’

performance in data processing, and

features Fênix at the 142nd position

worldwide.

1º LATIN AMERICA SUPER-COMPUTER:

PETROBRAS FENIX

Page 47: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

47

Page 48: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

48

Page 49: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

49

SEISMIC INTERPRETATIONNew Deep Learning approaches - Salt

Features (Seismic) Labels

Page 50: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

50

SEISMIC INTERPRETATION

Detection Prob. Map

New Deep Learning approaches - Salt

Page 51: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

51

Page 52: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

52

Page 53: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

53

2-D ELASTIC WAVE PROPAGATIONModel Properties Vp, Vs, Rho

• Jupyter• Matplotlib• OpenACC• GPU

Page 54: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

54

2-D ELASTIC WAVE PROPAGATIONRecorded Wavefield

• Jupyter• Matplotlib• OpenACC• GPU

Page 55: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

55

NVIDIA POWERS ACCELERATED DATA

SCIENCE

Page 56: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

56

THE BIG PROBLEM IN DATA SCIENCE

All

DataETL

Manage Data

Structured

Data Store

Data Preparation

Training

Model Training

Visualization

Evaluate

Inference

Deploy

Slow Training Times for Data Scientists

Page 57: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

57

ACCELERATING MACHINE LEARNINGThe RAPIDS Ecosystem

Open Source Community

Enterprise Data Science Platforms

StartupsDeep Learning

Integration

GPU Servers Storage Partners

Page 58: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

58

RAPIDS — OPEN GPU DATA SCIENCESoftware Stack

Data Preparation VisualizationModel Training

CUDA

PYTHON

APACHE ARROW

DASK

DEEP LEARNING

FRAMEWORKS

CUDNN

RAPIDS

CUMLCUDF CUGRAPH

Page 59: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

59

PILLARS OF DATA SCIENCE PERFORMANCE

CUDA Architecture NVLink/NVSwitch CUDA-X AI

Massively Parallel Processing High Speed Connecting between GPUs for Distributed Algorithms

NVIDIA GPU Acceleration Libraries for Data Science and AI

NVSwitch

6x NVLink

CUDA-X

PYTHON

APACHE ARROW on GPU Memory

DASK

cuDNN

RAPIDS

cuGraphcuMLcuDF

DL

FRAMEWORKS

Page 60: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

60

GPU-ACCELERATED DATA SCIENCE WORKFLOW WITH RAPIDS

Built on CUDA-X AI

DATA

DATA PREPARATION

GPUs accelerated compute for in-memory data preparation

Simplified implementation using familiar data science tools

Python drop-in pandas replacement built on CUDA C++.

GPU-accelerated Spark (in development)

PREDICTIONS

Page 61: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

61

GPU-ACCELERATED DATA SCIENCE WORKFLOW WITH RAPIDS

Built on CUDA-X AI

MODEL TRAINING

GPU-acceleration of today’s most popular ML algorithms such as XGBoost

Also available are PCA, K-means, k-NN, DBScan, tSVD, and many more

Easy-to-adopt, scikit-learn like interface

DATA

PREDICTIONS

Page 62: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

62

GPU-ACCELERATED DATA SCIENCE WORKFLOW WITH RAPIDS

Built on CUDA-X AI

VISUALIZATION

Effortless exploration of datasets, billions of records in milliseconds

Dynamic interaction with data = faster ML model development

Data visualization ecosystem (Graphistry & OmniSci), integrated with RAPIDS

DATA

PREDICTIONS

Page 63: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

63

XGBOOST: THE WORLD’S MOST POPULAR MACHINE LEARNING ALGORITHM

The leading algorithm for tabular data

Outperforms most ML algorithms on regression, classification and ranking

Winner of many data science Kaggle competitions

InfoWorld Technology of the Year Award, 2019

Well known in data science community and widely used for forecasting, fraud detection, recommender engines, and much more

Versatile and High Performance

Page 64: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

64

HOW CAN XGBOOST BE IMPROVED?

CPU processing is slow, creating issues for large data sets or when timeliness is crucial (e.g. intraday requirements for financial services)

Hyperparameter search is very slow, making search not feasible

Prediction speed limits the depth and number of trees in time sensitive applications

XGBoost Performance is Constrained by CPU Limitations

Page 65: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

65

GPU-ACCELERATED XGBOOST

Faster Time To InsightXGBoost training on GPUs is significantly faster than CPUs, completely transforming the timescales of machine learning workflows.

Better Predictions, SoonerWork with larger datasets and perform more model iterations without spending valuable time waiting.

Lower CostsReduce infrastructure investment and save money with improved business forecasting.

Easy to UseWorks seamlessly with the RAPIDS open source data processing and machine learning libraries and ecosystem for end-to-end GPU-accelerated workflows.

Unleashing the Power of NVIDIA GPUs for Users of XGBoost

Page 66: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

66

LOADING DATA INTO A GPU DATAFRAME

Create an empty DataFrame, and add a column

USE WITH MINIMAL CODE CHANGESGPU-Acceleration with the same XGBoost Usage

BEFORE AFTER

import xgboost as xgb

params = {'max_depth': 3,

'learning_rate': 0.1}

dtrain = xgb.DMatrix(X, y)

bst = xgb.train(params, dtrain)

import xgboost as xgb

params = {‘tree_method’: ‘gpu_hist’,

'max_depth': 3,

'learning_rate': 0.1}

dtrain = xgb.DMatrix(X, y)

bst = xgb.train(params, dtrain)

Page 67: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

67

XGBOOST: GPU VS. CPU

Take advantage of parallel processing with multiple GPUs

Scale to multiple nodes

GPU implementation is more memory efficient (half of CPU)

Improved accuracy by allowing time for more iterations, ability to leverage hyperparameter search, and reduced scale out needs

Tremendous Performance Improvements and Better Accuracy

A single DGX-2 with GPU-accelerated XGBoost is 10x Faster than 100 CPU nodes

Page 68: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

68

TRADITIONAL DATA SCIENCE CLUSTER

Workload Profile:

Fannie Mae Mortgage Data:

• 192GB data set

• 16 years, 68 quarters

• 34.7 Million single family mortgage loans

• 1.85 Billion performance records

• XGBoost training set: 50 features

300 Servers | $3M | 180 kW

Page 69: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

69

GPU-ACCELERATEDDATA SCIENCE CLUSTER

1 DGX-2 | 10 kW

1/8 the Cost | 1/15 the Space

1/18 the Power

GPU-accelerated XGBoost with DGX-2

0 2,000 4,000 6,000 8,000 10,000

20 CPU Nodes

30 CPU Nodes

50 CPU Nodes

100 CPU…

DGX-2

5x DGX-1

End-to-End

Page 70: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

70

NVIDIA GPUS ARE PROVENFASTER FOR DATA SCIENCE

2,290

1,956

1,999

1,948

169

157

0 500 1,000 1,500 2,000 2,500

20 CPU Nodes

30 CPU Nodes

50 CPU Nodes

100 CPU Nodes

DGX-2

5x DGX-1

0 2,000 4,000 6,000 8,000 10,000

20 CPU Nodes

30 CPU Nodes

50 CPU Nodes

100 CPUNodes

DGX-2

5x DGX-1

cuML — XGBoost

2,741

1,675

715

379

42

19

0 1,000 2,000 3,000

20 CPU Nodes

30 CPU Nodes

50 CPU Nodes

100 CPU Nodes

DGX-2

5x DGX-1

End-to-EndcuIO/cuDF —Load and Data Preparation

Benchmark

200GB CSV dataset; Data preparation includes joins, variable transformations.

CPU Cluster Configuration

CPU nodes (61 GiB of memory, 8 vCPUs, 64-bit platform), Apache Spark

DGX Cluster Configuration

5x DGX-1 on InfiniBand network

Time in seconds — Shorter is better

cuIO / cuDF (Load and Data Preparation) Data Conversion XGBoost

8,762

6,148

3,925

3,221

322

213

Page 71: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

71

DISTRIBUTED XGBOOST

GPU-acceleration for XGBoost with Apache Spark and Dask

Multiple nodes and multiple GPUs per node

Explore and prototype models on a PC, workstation, server, or cloud instance and scale to two or more nodes for production training

An ideal solution for GPU-accelerated clusters and enterprise scale workloads

Try out Dask support immediately using Google Cloud Dataproc

Download for on-prem and cloud deployments

GPU-Accelerated XGBoost for Large Scale Workloads

Page 72: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

72

GPU-ACCELERATED DATA SCIENCEA Solution for Every User and Every Organization

PRODUCTION DATA CENTER

NVIDIA-Powered

Data Science

Workstations

T4

Enterprise

ServersGeForce TITAN RTX

ML EXPERIMENTATION

Cloud

DGX Station,

DGX-1 / HGX-1DGX-2 / HGX-2

Page 73: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

73

GPU-ACCELERATEDDATA SCIENCE PLATFORMS

Unparalleled Performance and Productivity

Benefit

Ease of getting

started, low/no

barrier to entry,

elasticity of

resources

Enthusiast PC

solution, easy to

acquire, low cost,

great performance

The ultimate PC

GPU for data

scientists. Easy to

acquire, deploy

and get started

experimenting.

Enterprise

workstation for

experienced data

scientists

Standard GPU-

accelerated data

center

infrastructures

with the world’s

leading servers

Enterprise server,

proven 4 or 8-way

configuration,

modular approach for

scale-up, fastest

multi-GPU & multi-

node training

Largest compute

and memory

capacity in a

single node,

fastest training

solution

Typical GPU Memory

(system dependent)

varies depending on offering

22GB 48GB 96GB 64 GB(4 x 16 GB)

128GB-256GB 512GB

GPU Fabricvaries

depending on offering

2-way NVLink

2-way NVLink

2-way NVLink

PCIe 3.04- and 8-way

NVLink16-way

NVSwitch

Enterprise Data Center

NVIDIA-Powered

Data Science

Workstations

Enterprise Desktop

Max Performance

T4 Enterprise

Servers

Max Flexibility

ML Enthusiast

GeForce

High-end PCs

TITAN RTX

ML in the Cloud

NVIDIA GPUs

in the Cloud

All the top CSPs

DGX Station,

DGX-1 / HGX-1DGX-2 / HGX-2

Individual Workstations Shared Infrastructure for Data Science Teams

Page 74: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

74

NVIDIA ACCELERATED DATA SCIENCEKEY USE CASES

RECOMMENDER SYSTEMSFRAUD DETECTIONFORECASTING

Courtesy Pixar

Data science significantly increases the efficiency and

accuracy of forecasting, directly contributing to bottom line

growth.

Data science dramatically improves fraud detection,

saving money for organizations and customers alike.

Data science has revolutionized the connected experience,

delighting users with personalized experiences.

Page 75: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

75

With >100,000 different products in its 4,700 U.S. stores,

the Walmart Labs data science team predicts demand for

500 million item-by-store combinations every week.

By performing forecasting with the open-source RAPIDS

data processing and machine learning libraries built

on CUDA-X AI on NVIDIA GPUs, Walmart speeds

up feature engineering 100x and trains machine

learning algorithms 20x faster, resulting in faster

delivery of products, real-time reaction to

shopper trends, and inventory cost

savings at scale.

IMPROVINGDEMAND FORECASTS

Page 76: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

76

AI & DATA SCIENCE FOR NETWORK OPERATIONSA wireless network operator had access to terabytes of data

daily but no efficient way to gain insights from it. That is

until a deep learning solution powered by NVIDIA DGX POD,

RAPIDS, and software from Datalogue and OmniSci,

changed the way they collect, process, visualize

and understand data.

Data prep improved from 8 days to 4 minutes and

the company’s new AI models predict high-surge

Wi-Fi usage and detect anomalies

with 99% accuracy.

Page 77: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

77

SUPERCHARGING GENOMIC ANALYTICSChina’s healthcare industry is turning to AI to address the

needs of its elderly population. Genetics giant BGI—which

has over 1PB of data—is classifying targetable peptides

for personalized immunotherapy for cancer patients.

By running the open source RAPIDS data processing and

machine learning libraries built on CUDA-X AI on an

NVIDIA DGX-1 AI supercomputer, BGI sped up

analysis 18x using cuDF, and 10x using XGBoost.

The company is now expanding analysis

to millions of peptide

candidates.

Page 78: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

78

FOR MORE INFORMATION

nvidia.com/datascience rapids.ai nvidia.com/en-us/technologies/cuda-x/

Page 79: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

79

LEARN & SHARE MORE

Page 80: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

80

CONNECT

Connect with hundreds of experts from top industry, academic, startup, and government organizations

LEARN

Gain insight and valuable hands-on training through over 500+ sessions

DISCOVER

See how GPU technology is creating breakthroughs in deep learning, cybersecurity, data science, healthcare and more

INNOVATE

Explore disruptive innovations that can transform your work

JOIN US AT GTC 2020 | USE VIP CODE XXXXX FOR 25% OFF

March 22—26, 2020 | Silicon Valley

Don’t miss the premier AI conference.

www.nvidia.com/gtc

Page 81: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

81

March 22 | Full-Day Workshops

March 23 - 26 | Conference & Training

Get the hands-on experience you need to transform the

future of AI, high-performance computing and more with

NVIDIA’s Deep Learning Institute (DLI).

Register for GTC 2020 to earn certification in full-day

workshops, join instructor-led sessions, and start self-

paced training.

www.nvidia.com/en-us/gtc/sessions/training/

THE LATEST DEEP LEARNING

DEVELOPER TOOLS

Page 82: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

82

JOINT MACHINE LEARNING WORKSHOPSBGf & SEG, 12-13 May 2020, Rio de Janeiro, Brazil

Page 83: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

83

developer.nvidia.com

Page 84: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

84

Deep Learning Fundamentals

Game Development & Digital Content

Finance

NVIDIA DEEP LEARNING INSTITUTE

Hands-on self-paced and instructor-led training in deep learning and accelerated computing for developers

Request onsite instructor-led workshops at your organization: www.nvidia.com/requestdli

Take self-paced labs online: www.nvidia.com/dlilabs

Download the course catalog, view upcoming workshops, and learn about the University Ambassador Program: www.nvidia.com/dli

Intelligent Video Analytics

Medical Image Analysis

Autonomous Vehicles

Accelerated Computing Fundamentals

More industry-specific training coming soon…

Genomics

Page 85: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

85

NVIDIA HW GRANT PROGRAM

Titan V Volta

• Robotics

• Autonomous Machines

Jetson TX2(Dev Kit)

• Scientific Visualization

• Virtual Reality

Quadro P6000

• Scientific Computing

• HPC

• Deep Learning

https://developer.nvidia.com/academic_gpu_seeding

Page 86: ACCELERATING PYTHON DATA SCIENCE WITH NVIDIA RAPIDS...3 1980 1990 2000 2010 2020 GPU-Computing perf 1.5X per year 1000X by 2025 RISE OF GPU COMPUTING Original data up to the year 2010

ObrigadoGracias

Thank you

[email protected]