Top Banner
A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work with Alexandru Costan, Gabriel Antoniu, Luc Bougé
19

A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Jan 11, 2016

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

A Performance Evaluation of Azure and Nimbus Clouds forScientific Applications

Radu TudoranKerData TeamInria RennesENS Cachan 10 April 2012

Joint work with Alexandru Costan, Gabriel Antoniu, Luc Bougé

Page 2: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Outline

• Context and motivation• 2 cloud environments: Azure and Nimbus• Metrics• Evaluation and discussions• Conclusion

10 April 2012Performance Evaluation of Azure and Nimbus - 2

Page 3: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Scientific Context for Clouds

• Up to recent time, scientists mainly relied on grids

and clusters for their experiments• Clouds emerged as an alternative to these due to:

- Elasticity

- Easier management

- Customizable environment

- Experiments’ repeatability

- Larger scales

10 April 2012Performance Evaluation of Azure and Nimbus - 3

Page 4: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Requirements for Science Applications

• Performance: throughput, computation power, etc.• Control• Cost• Stability• Large storage capacity• Reliability• Data access and throughput• Security• Intra-machine communication

10 April 2012Performance Evaluation of Azure and Nimbus - 4

Page 5: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Stages of a Scientific Application

10 April 2012Performance Evaluation of Azure and Nimbus - 5

1

2

3

4

5

6

Computation Nodes

Cloud Storage

Local Host

Page 6: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Public Clouds: Azure

• On demand – pay as you go• Computation (Web/Worker Role) separated from storage (Azure

BLOBs)• HTTP for storage access • BLOB are structured in containers• Multitenancy model

10 April 2012Performance Evaluation of Azure and Nimbus - 6

VM Type CPU Cores Memory Disk

Small 1 1.75GB 225GB

Medium 2 3.5GB 490GB

Large 4 7GB 1000GB

ExtraLarge 8 14GB 2040GB

VM Characteristics

Page 7: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Private Nimbus-powered Clouds

• Deployed in a controlled infrastructure• Allows to lease a set of virtualized computation resources and

to customize the environment• 2 Phases deployment:

- The cloud environment

- The computation VMs• Cumulus API

- allows multiple storage

- quota based storage

- VM image repository

10 April 2012Performance Evaluation of Azure and Nimbus - 7

Page 8: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Focus of Our Evaluation

• Initial phase- Deploying the environment (hypervisors , application, etc.)

- Data staging

- Metric: Total time

• Applications’ Performance and Variability

- Computation - Real application ABrain - Metric: Makespan

- Data transfer - Synthetic benchmarks

- Metric: Throughput• Cost

- Pay-as you go

- Infrastructure & Maintenance

10 April 2012Performance Evaluation of Azure and Nimbus - 8

Page 9: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

10 April 2012Performance Evaluation of Azure and Nimbus - 9

• Deployment time - Azure – 10 – 20 minutes

- Nimbus - Phase 1 – 15 minutes

- Phase 2 – 10 minutes (on Grid5000)

Pre-processing

0.1 1 10 1001

10

100

1000

10000

Local->AzureBlobs AzureBlobs->Local Local->Cumulus Cumulus->Local

Size GB

Tim

e (s

econ

ds)

Data moved from Local to Cloud Storage

Page 10: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

10 April 2012Performance Evaluation of Azure and Nimbus - 10

Pre-processing (2)

0.1 1 10 1001

10

100

1000

10000

14161

ExtraLarge->AzureBlobs AzureBlobs->ExtraLargeVM->Cumulus Cumulus->VM

Size GB

Tim

e (s

econ

ds)

Data moved from Cloud Storage to Computing nodes

Page 11: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

ABrain

10 April 2012Performance Evaluation of Azure and Nimbus - 12

p( ),

Genetic dataBrain image

Y

q~105-6

N~2000

Xp~106

– Anatomical MRI– Functional MRI– Diffusion MRI

– DNA array (SNP/CNV)– gene expression data– others...

finding associations:

Page 12: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Application Performance – Computation Time

10 April 2012Performance Evaluation of Azure and Nimbus - 13

Repeated a run of ABrain 1440 times on each machine(Operations on large matrices)

Evaluate only the local resources (CPU, Memory, Local storage)

Page 13: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

10 April 2012Performance Evaluation of Azure and Nimbus - 14

Computation Variability vs. Fairness

Multitenancy model

Variability of a VM sharing a node with others with respect to an isolated one

Page 14: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Data Transfer Throughput

• TCP - default and most commonly used

- reliable transfer mechanism of data between VM instances

• RPC - based on HTTP

- used by many applications as communication paradigm between

their distributed entities

10 April 2012Performance Evaluation of Azure and Nimbus - 15

Page 15: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Application Performance – Data Transfers

10 April 2012Performance Evaluation of Azure and Nimbus - 16

Delivered TCP throughput at application level

Page 16: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Application Performance – Data Transfers

10 April 2012Performance Evaluation of Azure and Nimbus - 17

𝑐𝑣=𝑠𝑡𝑑

𝑚𝑒𝑎𝑛%

RPC (HTTP) – delivered throughput at application level

HTTP traffic control on nodes

Page 17: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Cost Analysis

• Although scientists usually don’t pay for the private

infrastructures that they access – these are not free• Hard to compute for private infrastructures • Direct for public clouds

• euros/hour (=0.0899)

• euros/hour (=0.0086) (p=0.11 ; )

• Cost private = 0.0985 Azure: 13.5%

• Cost public = 0.0852 CHEAPER

10 April 2012Performance Evaluation of Azure and Nimbus - 18

Page 18: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

Conclusions

• Compared Nimbus and Azure clouds from the point of view of

scientific applications stages• Azure

- lower cost

- good TCP throughput and stability

- faster transfer from storage to VM• Nimbus

- additional control

- good RPC (HTTP) throughput

- in general better data staging • An analysis of how multitenancy model affects the fairness in

public clouds

10 April 2012Performance Evaluation of Azure and Nimbus - 19

Page 19: A Performance Evaluation of Azure and Nimbus Clouds for Scientific Applications Radu Tudoran KerData Team Inria Rennes ENS Cachan 10 April 2012 Joint work.

thank you!

Alexandru Costan

Gabriel Antoniu

Radu Tudoran

Luc Bougé

A Performance Evaluation of Azure and Nimbus

Clouds forScientific Applications