Top Banner
REFERENCE ARCHITECTURES WHAT’S NEW Brent Compton Director Storage Solution Architectures Red Hat Red Hat Storage Day NYC Oct 2016
39

Red Hat Storage Day New York - New Reference Architectures

Jan 11, 2017

Download

Technology

Red_Hat_Storage
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Red Hat Storage Day New York - New Reference Architectures

REFERENCE ARCHITECTURES WHAT’S NEW

Brent Compton Director Storage Solution Architectures Red Hat

Red Hat Storage Day NYC Oct 2016

Page 2: Red Hat Storage Day New York - New Reference Architectures

Reference Architecture Work

2

MYSQL & HADOOP SOFTWARE-DEFINED NAS DIGITAL MEDIA REPOS

Page 3: Red Hat Storage Day New York - New Reference Architectures

Reference Architecture Work

3

MYSQL & HADOOP SOFTWARE-DEFINED NAS DIGITAL MEDIA REPOS

Page 4: Red Hat Storage Day New York - New Reference Architectures

(link)

Page 5: Red Hat Storage Day New York - New Reference Architectures

Appetite for Storage-Centric Cloud Services

EC2 EBS S3 KVM RBD RGW (Nova) (Cinder) (Swift)

RDS EMR MySQL Hadoop

Page 6: Red Hat Storage Day New York - New Reference Architectures

Cloud Storage Features Useful for MySQL

Page 7: Red Hat Storage Day New York - New Reference Architectures

HEAD-TO-HEAD LAB TEST ENVIRONMENTS

•  EC2 r3.2xlarge and m4.4xlarge

•  EBS Provisioned IOPS and GPSSD

•  Percona Server

•  Supermicro servers

•  Red Hat Ceph Storage RBD

•  Percona Server

Page 8: Red Hat Storage Day New York - New Reference Architectures

30.0 29.8

3.6

25.6 25.7

4.1

0.0

5.0

10.0

15.0

20.0

25.0

30.0

35.0

P-IOPS m4.4xl

P-IOPS r3.2xl

GP-SSD r3.2xl

100% Read

100% Write

AWS IOPS/GB BASELINE: ~ AS ADVERTISED

Page 9: Red Hat Storage Day New York - New Reference Architectures

IOPS/GB PER MYSQL INSTANCE

30

252

150

26

78

19

0

50

100

150

200

250

300

P-IOPS m4.4xl

Ceph cluster 1x "m4.4xl"

(14% capacity)

Ceph cluster 6x "m4.4xl"

(87% capacity)

MySQL IOPS/GB Reads MySQL IOPS/GB Writes

Page 10: Red Hat Storage Day New York - New Reference Architectures

FOCUSING ON WRITE IOPS/GB AWS IO THROTTLING LEVEL FOR DETERMINISTIC PERFORMANCE

26

78

19

0 10 20 30 40 50 60 70 80 90

P-IOPS m4.4xl

Ceph cluster 1x "m4.4xl"

(14% capacity)

Ceph cluster 6x "m4.4xl"

(87% capacity)

Page 11: Red Hat Storage Day New York - New Reference Architectures

EFFECT OF CEPH CLUSTER LOADING ON IOPS/GB

78

37

25 19

0

10

20

30

40

50

60

70

80

90

Ceph cluster (14%

capacity)

Ceph cluster (36%

capacity)

Ceph cluster (72%

capacity)

Ceph cluster (87%

capacity)

IOPS

/GB

100% Write

Page 12: Red Hat Storage Day New York - New Reference Architectures

$/STORAGE-IOP* FOR COMPRABLE CONFIGS

$2.40

$0.80 $0.78 $1.06

$-

$1.00

$2.00

$3.00

Stor

age

$/IO

P (S

ysbe

nch

Writ

e)

AWS EBS Provisioned-IOPS

Ceph on Supermicro FatTwin 72% Capacity

Ceph on Supermicro MicroCloud 87% Capacity

Ceph on Supermicro MicroCloud 14% Capacity * Ceph configs do not include power, cooling, or admin costs

Page 13: Red Hat Storage Day New York - New Reference Architectures

18 18 19

6

0

5

10

15

20

25

Ceph cluster 80 cores 8 NVMe

(87% capacity)

Ceph cluster 40 cores 4 NVMe

(87% capacity)

Ceph cluster 80 cores 4 NVMe

(87% capacity)

Ceph cluster 80 cores 12 NVMe

(84% capacity)

IOPS

/GB

100% Write

CONSIDERING CORE-TO-FLASH RATIO

Page 14: Red Hat Storage Day New York - New Reference Architectures

8x Nodes in 3U chassis

Model: SYS-5038MR-OSDXXXP

Per Node Configuration: CPU: Single Intel Xeon E5-2630 v4 Memory: 32GB NVMe Storage: Single 800GB Intel P3700 Networking: 1x dual-port 10G SFP+

+ +

1x CPU + 1x NVMe + 1x SFP

SUPERMICRO MICRO CLOUD CEPH MYSQL PERFORMANCE SKU

Page 15: Red Hat Storage Day New York - New Reference Architectures

Enhancing On-premise MySQL Scalability

Page 16: Red Hat Storage Day New York - New Reference Architectures

Appetite for Storage-Centric Cloud Services

EC2 EBS S3 KVM RBD RGW (Nova) (Cinder) (Swift)

RDS EMR MySQL Hadoop

Page 17: Red Hat Storage Day New York - New Reference Architectures

Trend: Disaggregating Hadoop Compute and Storage

§  Hadoop retained data growing at faster pace than hadoop compute needs.

§  Don’t want to waste money on un-needed compute in more hadoop data nodes.

§  Driving trend to disaggregate storage from traditional hadoop nodes. (eBay blog on tiering – here)

§  Multiple disaggregation architecture options

Page 18: Red Hat Storage Day New York - New Reference Architectures

Data Flow Options (Traditional, Partial Disaggregation)

ingres copy

?

?

HDFS

HDFS

aging

retrieval Ceph

MapReduce/Pig, Spark, Hbase/Hive

ingres copy S3A

MapReduce, Cold Data

S3A MapReduce/Pig, Spark, Hbase/Hive

Page 19: Red Hat Storage Day New York - New Reference Architectures

Data Flow Options (Full Disaggregation)

Ceph

MapReduce/Pig, Hive/HBase MapReduce, Spark Hot Data

Non-Hadoop Tools

ingres

S3

S3A

S3

RBD

HDFS over RBD

volumes

Page 20: Red Hat Storage Day New York - New Reference Architectures

Reference Architecture Work

20

MYSQL & HADOOP SOFTWARE-DEFINED NAS DIGITAL MEDIA REPOS

•  Digital object repo •  Digital file repo (WIP)

Page 21: Red Hat Storage Day New York - New Reference Architectures

(link)

Page 22: Red Hat Storage Day New York - New Reference Architectures

- 20,00040,00060,00080,000100,000

UHDmovie

Blue-raymovie

HDmovie

DVDmovie

AudioCD

MP3song

e-book

100,000

25,000

12,000

3,000

750

4

1

MB

Different Servers Yield 10x Ceph Performance

1 DVD movie/sec with 3-node* cluster A

1 Blue-ray movie/sec with 3-node* cluster B

*Both A & B cluster nodes are 2U servers

Page 23: Red Hat Storage Day New York - New Reference Architectures

Ceph Nodes Saturating 80GbE Pipes

1MBSeq.Read=28.5GB/sec 1MBSeq.Write=6.2GB/sec

Page 24: Red Hat Storage Day New York - New Reference Architectures

… and note optimal CPU/SSD ratio for IOPS

4KBRandomRead=693KIOPS 4KBRandomWrite=87.8KIOPS

Page 25: Red Hat Storage Day New York - New Reference Architectures

Sample High Throughput Config for Ceph

• 2x Intel E5-269xv3 (up to 145W per CPU) • 4x-24x 2.5” hot swap Samsung NVMe SSDs • 16x DDR4 2133MHz L/RDIMM & up to 1024GB • 2x 16-lane PCIe Gen3 slots • 2x dual 40 GbE NICs (with 100 GbE option) • EIA 310-D 19” in 2U

Page 26: Red Hat Storage Day New York - New Reference Architectures

Scaling Ceph Object Storage (RGW) to NIC Saturation

Page 27: Red Hat Storage Day New York - New Reference Architectures

Reference Architecture Work

27

MYSQL & HADOOP SOFTWARE-DEFINED NAS DIGITAL MEDIA REPOS

Page 28: Red Hat Storage Day New York - New Reference Architectures

(link)

Page 29: Red Hat Storage Day New York - New Reference Architectures

Server Category Nomenclature ContemporaryStorage

ServerChassisCategoriesFlashBlades FlashArray Standard Dense Ultra-dense

StoragemediaSSD

(NVMehi-perf)SSD

(NVMemid-perf)HDD+SSD HDD+SSD HDD

Mediadrives/node 1-4 4-24 12-16 24-46 60-90

Servernodesperchassis 2-8 1 1 1-2 1

CPUsizing*** 10coresperSSD 4coresperSSD 1coreper2HDD 1coreper2HDD 1coreper2HDD

Servernetworking 10GbE 40GbE+ 10GbE10GbE(archive)40GbE(active)

10GbE(archive)40GbE(active)

TargetIOpattern smallrandomIO largesequentialIO mixed mixed largesequentialIO

VendorExamples

SupermicroMicrocloudSamsungSierra2U/4*

SanDiskInfiniFlash**SamsungSierra2U/24

Supermicro2U/12Quanta/QCT1U/12DellR730XDCiscoC240MLenovoX3650

Supermicro4U/36Quanta/QCT4U/35x2DellDSS7000/2CiscoC3260

Supermicro4U/72Quanta/QCT4U/76DellDSS7000/1CiscoC3160

*smallerflasharray **JBOFwithservers ***characterizedforCeph

Page 30: Red Hat Storage Day New York - New Reference Architectures

- 20,00040,00060,00080,000100,000

UHDmovie

Blue-raymovie

HDmovie

DVDmovie

AudioCD

MP3song

e-book

100,000

25,000

12,000

3,000

750

4

1

MB

Size of Common Things

Page 31: Red Hat Storage Day New York - New Reference Architectures

Jumbo File Performance Comparison (4GB files with 4MB sequential IO - think DVD video size)

0 50 100 150 200 250

Manufacturer spec (HDD)

Standard, baseline (1xRAID6 vol, no Gluster)

Dense, baseline (3xRAID6 vol, no Gluster)

Standard (EC4:2), JBOD Bricks

Dense, (2xRep), RAID6 Bricks

Standard, (2xRep) RAID6 Bricks

Dense (EC4:2), RAID6 bricks

Standard (EC4:2), RAID6 Bricks

MB/sec per Drive (HDD)

Read

Write

Page 32: Red Hat Storage Day New York - New Reference Architectures

Jumbo File Price-Performance Comparison (4GB files with 4MB sequential IO - think DVD video size)

0 50 100 150 200 250 300 350

Standard (EC4:2), JBOD Bricks

Dense, (2xRep), RAID6 Bricks

Standard, (2xRep) RAID6 Bricks

Dense (EC4:2), RAID6 bricks

Standard (EC4:2), RAID6 Bricks

MB/sec per $

Read

Write

Page 33: Red Hat Storage Day New York - New Reference Architectures

3 Year TCO (incl. support)

TCO COMPARISON

For 1PB Usable Capacity Throughput-optimized Solutions Configuration Highlights:

•  HDD-only media

•  2x replication with RHGS

•  8:3 Erasure Coding with

EMC Isilon

•  Higher CPU-to-media ratio

than archive-optimized

X-210 12LFF 12LFF (standard)

X-410 36LFF 36LFF (dense)

Pricing Sources: EMC Isilon: Gartner Competitive Profiles, as of 2/16/16) & Supermicro: Thinkmate, as of 1/13/16)

Page 34: Red Hat Storage Day New York - New Reference Architectures

Comparing Throughput and Costs at Scale

STORAGE PERFORMANCE SCALABILITY

STORAGE COSTS SCALABILITY

Number of Storage Nodes Number of Storage Nodes

Tota

l Sto

rage

Cos

ts ($

)

Rea

ds/W

rite

s Th

roug

hput

(mB

ps)

Software Defined Scale out Storage

Traditional Enterprise NAS

Storage

Traditional Enterprise NAS

Storage

Software Defined Scale out Storage

Page 35: Red Hat Storage Day New York - New Reference Architectures

Small File Performance Comparison (50KB files - think small jpeg image size)

0 100 200 300 400 500 600

Dense (EC4:2), Tiered (2xSSD/svr)

Dense, no tiering

Dense, Tiered (2xSSD/svr)

Standard, no tiering

Standard, Tiered (4xSSD/svr)

Standard, Tiered (2xNVMe/svr), 70%

Standard, Tiered (2xNVMe/svr)

Standard, Tiered (1xNVMe/svr)

File Operations/Second per drive (HDD), 50KB Files

Read

Create

Page 36: Red Hat Storage Day New York - New Reference Architectures

Small File Performance Comparison (50KB files - think small jpeg image size)

0 100 200 300 400 500 600

Dense (EC4:2), Tiered (2xSSD/svr)

Dense, no tiering

Dense, Tiered (2xSSD/svr)

Standard, no tiering

Standard, Tiered (4xSSD/svr)

Standard, Tiered (2xNVMe/svr), 70% full

Standard, Tiered (2xNVMe/svr)

Standard, Tiered (1xNVMe/svr)

File Operations/Second per drive (HDD), 50KB Files

Read

Create

Page 37: Red Hat Storage Day New York - New Reference Architectures

Small File Performance Comparison (50KB files - think small jpeg image size)

0 100 200 300 400 500 600

Dense (EC4:2), Tiered (2xSSD/svr)

Dense, no tiering

Dense, Tiered (2xSSD/svr)

Standard, no tiering

Standard, Tiered (4xSSD/svr)

Standard, Tiered (2xNVMe/svr), 70%

Standard, Tiered (2xNVMe/svr)

Standard, Tiered (1xNVMe/svr)

File Operations/Second per drive (HDD), 50KB Files

Read

Create

Page 38: Red Hat Storage Day New York - New Reference Architectures

Test drive: bit.ly/cephtestdrive bit.ly/glustertestdrive

Test drive:

Try it Building your first software-defined storage cluster

Page 39: Red Hat Storage Day New York - New Reference Architectures

THANK YOU