Top Banner
© 2011 IBM Corporation IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague A practical Introduction to Disk Storage System Performance Gero Schmidt, ATS System Storage IBM European Storage Competence Center
94

GSchmidt - IntroDiskPerformance_V1.63

Oct 02, 2014

Download

Documents

Pavel Alexei
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation

IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague

A practical Introduction toDisk Storage System Performance

Gero Schmidt, ATS System StorageIBM European Storage Competence Center

Page 2: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation2 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Disclaimers

IBM has not formally reviewed this document. While effort has been made to verify the information, this document may contain errors. IBM makes no warranties or representations with respect to the content hereof and specifically disclaim any implied warranties of merchantability or fitness for any particular purpose. IBM assumes no responsibility for any errors that may appear in this document. The information contained in this document is subject to change without any notice. IBM reserves the right to make any such changes without obligation to notify any person of such revision or changes. IBM makes no commitment to keep the information contained herein up to date.

Note: This presentation is intended for IBMers and IBM BPs only. As IBM has not formally reviewed this document it may be presented but should not be handed out to clients (especially not as ppt version). A pdf version of the presentation without slides 33, 52 to 54 and 79 may be suitable as hand-out for clients at one's own responsibility using one's own best judgement.

If you have any suggestions or corrections please send comments to: [email protected]

Page 3: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation3 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs• Application I/O & Workload Characteristics• Hard Disk Drive (HDD) Basics – It’s all mechanic• HDD Performance & Capacity Aspects (SATA vs FC/SAS)• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)• New Trends & Directions: 2.5" & Solid State Drive (SSD)• Basic Principles for Planning Logical Configurations• Performance Data Collection and Analysis

Page 4: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation4 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: 2.5" & Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 5: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation5 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

IBM System Storage Disk Subsystems – Making a Choice

Selecting a storage subsystem:entry-level, midrange or enterprise classsupport for host systems and interfacesoverall capacity & growth considerationsoverall box performanceadvanced features and copy servicesprice, costs / TCO, footprint, etc.

needs to meet client & application requirements

Subsystem performance:overall I/O processing capabilityoverall bandwidth

choosing the rightnumber and type of

Disk Drives

EnterpriseDS3000 DS5000 Storwize V7000 DS8000

MidrangeEntry-levelXIV

Page 6: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation6 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Storage Subsystem Specs – Data Rate (MBps)

max. throughput may be achieved with a relatively low no. of disk drivessubsystem architecture: frontend / backend bandwidth capabilities are keySATA may be considered for applications requiring throughput

Note: Results as of 6-26-2006. Source of information from Engenio and not confirmed by IBM. Performance results achieved under ideal circumstances in a benchmark test environment. Actual customer results will vary based on configuration and infrastructure components.

The number of drives used for MB/s performance does not reflect an optimized test config. The number of drives required could be lower/higher.

Page 7: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation7 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Storage Subsystem Specs – I/O Rate (IOps)

Note: Results as of 6-26-2006. Source of information from Engenio and not confirmed by IBM. Performance results achieved under ideal circumstances in a benchmark test environment. Actual customer results will vary based on configuration and infrastructure components.

Drives were “short-stroked” to optimize for IOPs performance. Real-life may take more drives to achieve the numbers listed..

max. IOps performance requires a high no. of fast FC/SAS disk drivessubsystem architecture: I/O processing capability >> disk drives IOps capability SATA is not a good fit for enterprise class applications requiring transaction performance

Page 8: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation8 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Storage Performance Council (SPC) - Benchmarks

The Storage Performance Council (SPC) is a vendor-neutral standards body focused on the storage industry. It has created the first industry-standard performance benchmark targeted at the needs and concerns of the storage industry. From component-level evaluation to the measurement of complete distributed storage systems, SPC benchmarks will provide a rigorous, audited and reliable measure of performance.

http://www.storageperformance.org

Page 9: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation9 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: 2.5" & Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 10: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation10 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Avg. Access Time for an I/O operation:CPU cycle < 0.000001msMEMORY < 0.001msDISK (HDD) < 10ms

Application I/O – An Overview

HardwareHardwareSetup

Software

System Software

ApplicationSoftware

↔ ↔

Application

File Systems

SANFC

iSCSIIB

SASSATASCSI

SERVER

MEMORy

STORAGE

CACHE

Device DriversVolume Manager

Storage subsystemCache hit: < 1 msPhysical HDD: ~ 5...15 ms

Storage I/O performance:Proper data placement is key!

Application I/O performance: Efficient memory usage is key!Access to memory is >10000 times faster than disk access!

Disk access is ‚SLOW‘compared to

CPU and MEMORY

Page 11: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation12 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Application I/O – On a typical System Time Scale

CPU 1ns (1GHz) = 0.000000001s

MEMORY 100ns = 0.000000100s

DISK 10ms = 0.010000000s

Page 12: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation13 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Application I/O – On a human Time Scale

‚SLOW‘

MEMORY 1:40 minutes

DISK 116 days

CPU 1 cycle := 1 second

Page 13: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation14 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Application I/O – Where does it come from?

Transaction ProcessingA single end-user is capable of initiating only a moderate number of transactions with a limited amount of data changes per minute Thousands of end-users can already initiate thousands of transactions and generate high I/O rates with only low data ratesEnd-users are directly affectedby the application response time People‘s work time is expensiveExcellent overall response time of the application is business criticaland requires low I/O response times at high I/O rates

Batch JobsA single batch job can alreadygenerate a considerable amount of disk I/O operations in terms ofI/O rate and data rateMultiple batchjobs can createa huge amount of disk activityBatch jobs should not interact with end-user transactionsand are typically run outsideend-user business hoursTime frames for batch jobseven during nights / weekendsare limitedOverall job runtime is critical andmostly dependent on achievedoverall data rate

I/O I/O

Page 14: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation15 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Throughput dependent workloads

Application I/O – Workload Characteristics

I/O rate in IO/s (IOps) Data rate in MB/s (MBps)

! Time to data is critical

! Dependent on number and type of disk drives

! Data transfer rate enables performance

! Dependent on internal controller bandwidth

typical for transaction processing workloads with random, small-block I/O requests, e.g. OLTP – online transaction processing,databases, mail servers – the majority of enterprise applicationsavg. I/O response time is most important here (RT < 10ms is a good initial choice)number and speed of of disk drives is essential(e.g. 73GB15k FC drives as best choice) SATA disk drives not generally recommended, high speed FC/SAS/SCSI disk drives preferredbalanced system configuration and volume layout is key to utilize all disk spindles

typical for throughput dependent workloads with sequential, large-block I/O requests, e.g. HPC, seismic processing, data mining, streaming video applications, large file access, backup/restore, batch jobsavg. I/O response time is less important(high overall throughput required)bandwidth requirements (no. of adapters and host ports, link speed) must be metnot necessarily a high number of disk drives requiredSATA disk drives may be a suitable choicebalanced system configuration and volume layout is important to utilize full system bandwidth

Transaction processing workloads

Page 15: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation16 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Application I/O – Workload Performance Characteristics

Basic workload performance characteristics:I/O rate [IOps] (transactions) or data rate [MBps] (throughput)

Random access or sequential access workload pattern

Read:write ratio (percentage of read:write I/O requests, e.g. 70:30)

average I/O request size (average I/O transfer size or block size, e.g. 8kB for Oracle DB, 64kB or larger for streaming applications, 256kB for TSM)

Additional workload performance characteristics / objectives:

Read cache hit ratio (percentage of read cache hits)

average response time (RT) requirements (e.g. RT < 10ms)

Page 16: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation17 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: 2.5" & Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 17: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation18 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Hard Disk Drive (HDD) Basics – It’s all mechanic...

Read / write cache hits are in the range < 1ms

Physical disk I/O operations are in the range of > 5ms because mechanicalcomponents such as head movements and spinning disks are involved

Each hard disk drive (HDD) can only process a limited no. of random I/O operations per second, mainly determined by :– Average Seek Time [ms] (head movement to required track)– Rotational Latency [ms] (disk platter spinning until the first sector

addressed passes under the r/w heads; avg. time = half a rotation)– Transfer Time [ms] (read/write data sectors, 1 sector = 512 Byte)

Start SeekTime

RotationalLatency

TransferTime

Page 18: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation19 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Simple IOps Calculation per Hard Disk Drive (HDD)

Avg. Seek Time = see manufacturer specs (typical: 4-10ms)

Rotational Latency = ½ × (60000/RPM) [ms] (typical: 2-4ms)

Transfer Time = 1000 × sectors × sector size / avg. Transfer Rate [ms](typically << 1ms for small I/O request sizes ≤ 32kB)

Page 19: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation20 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Manufacturer Specs for Hard Disk Drives

Source: www.seagate.com (2008)This is just an example for getting a view on typical disk drive characteristics. The chosen disk types above do not necessarily represent the characteristics of the disk drive modules used in IBM System Storage systems.

Page 20: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation21 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

A single disk driveis only capable of processing a

limited number of I/O operations per second!

Example Random IOps Calculation per Hard Disk Drive

769 ms4.2 ms7200 rpmSATA2

500GB7.2k

1255 ms3 ms10000 rpmFC

146GB10k

1674 ms2 ms15000 rpmFC

146GB15k

IOpsAvg. Seek Time

Rotational LatencySpeedDisk Drive

Rules of Thumb - Random IOps/HDD (conservative estimate to start with):

• FC 15k DDM : ~160 IOps• FC 10k DDM : ~120 IOps• SATA2 7.2k DDM: ~75 IOps

Page 21: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation22 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Efforts to improve HDD Performance

Efforts to reduce HDD access times (mechanical delays)Disk Drive: Introduce Command Queuing and Re-Ordering of I/Os

SATA: NCQ (Native Command Queuing)SCSI: TCQ (Tagged Command Queuing)

Disk Drive Usage: 'Short Stroking' of HDDsDisk Subsystem: Subsystem Cache

Intelligent Cache Page Replacement & Prefetching AlgorithmsStandard: LRU (least recently used) / LFU (least frequently used)IBM System Storage DS8000 - Advanced Caching Algorithms

2004 – ARC (Adaptive Replacement Cache) 2007 – AMP (Adaptive Multi-stream Prefetching)2009 – IWC (Intelligent Write Caching)IBM Almaden Research Center - Storage Systems Caching Technologieshttp://www.almaden.ibm.com/storagesystems/projects/arc/technologies/

Seek latency optimization

Caching / Cache Hits

Page 22: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation23 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Increase HDD Performance - Command Queuing

Tagged Command Queuing (TCQ, SCSI-2) & Native Command Queuing (NCQ, SATA2)further improves disk drive random access performance

by re-ordering the I/O commands so that workloads can experience seek times which are considerably less than the nominal seek times

Queue Depth: SATA2 (NCQ): 32 in-flight commands, SCSI-3 (TCQ): 2^64 in-flight commands

Page 23: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation24 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Increase HDD Performance - Short Stroking

Short Stroking:

Approach to achieve maximum possible performance from an HDD bylimiting the overall head movement and thus minimizing the average seek time.

Implementation:

- Use only a small portion of overall capacity

- Use tracks on outer edge with higher data density

Disadvantage:

- Typically large number of HDDs involved

- Only small portion of storage capacity usedTypical usage:

Applications with high access densities (IOps/GB) that require high random I/O rates at low response times but with only a comparatively small amount of data.

Page 24: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation25 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Short stroking – DS8800 Single Rank Performance

100%100% 25%25%50%50%100% random read100% random read

DS8800 R6.0DS8800 R6.0

Page 25: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation26 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Increase HDD Performance - Subsystem Cache

Disk Subsystem Cache–Read Cache Hits–Write Cache Hits / Write behind–Sequential Prefetch Algorithms

Intelligent Cache Page Replacement & Prefetch Algorithms–What data should be stored in cache based upon the

recent access and frequency needs of the hosts (LRU/LFU)? –Determine what data in cache can be removed to

accommodate newer data.–Predictive algorithms to anticipate data prior to a host request and

loading it into cache.

1:5001:5001:10001:1000

Page 26: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation27 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Sample Random IOps Calculation with reduced ⅓ Seek Times

9 ms

5 ms

4 ms

Avg. Seek Time

9/3 ms

5/3 ms

4/3 ms

⅓ Reduced Seek Time

1384.2 ms7200 rpmSATA2

500GB7.2k

2143 ms10000 rpmFC

146GB10k

3002 ms15000 rpmFC

146GB15k

IOps

Red. SeekRotational LatencySpeedDisk Drive

Even with reduced average seek times you cannot expect more thana few hundred random I/O operations per second from a single HDD.

So a single HDD can only process a limited number of random IOps withaverage access times in the typical range of 5...15ms due to the mechanical delays associated with spinning disks (HDDs).

Page 27: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation28 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Storage Disk Subsystem – Typical I/O Rate & Response Time Relation

Response Time versus I/O Rate

0

5

10

15

20

25

30

0

1000

2000

3000

4000

5000

6000

7000

8000

9000

1000

0

1100

0

Total I/O [IO/s]

Res

pons

e Ti

me

[ms]

+/- 10% change in I/O rate

Page 28: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation29 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: 2.5" & Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 29: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation30 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Subsystem Sizing – Meeting Performance and Capacity Requirements

Capacity:– number of disk drives to meet capacity requirements– only low no. of large capacity disks required to meet capacity needs

Performance:– number and speed of disk drives (spindles) to meet IOps requirements– high no. of fast, low capacity drives required to meet performance needs

Cost:

COST

CapacityPerformance

higher lower 146GB15k drives arean excellent trade offbetween performanceand capacity needs

IOpsIOps

GBGB

no. of drives drive capacity

Page 30: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation31 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

20050.7

Average Access Density over recent Years

Access Density is a measure of I/O throughput per unit of usable storage capacity (backstore). The primary use of access density is to identify a range on a response time curves to give the typical response time expected by the average customer, based on the amount of total usable storage in their environment. The average industry value for access density in the year 2005 is thought to be approximately 0.7 I/Os per second per GB. Year-to-year industry data is incomplete, but the value has been decreasing as companies acquire usable storage faster than they access it.

Source: IBM data, other consultants

IOpsIOps

GBGB

= Access Density

[IOps/GB]

cold

hot

data

Page 31: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation32 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

1x 1TB 7.2k SATA(75 IOps/HDD; 9.8W)

75 IOps1000 GB

9.8 W

7x 146GB15k FC(160 IOps/HDD; 15W)

1120 IOps1022 GB105 W

Subsystem Sizing – Meeting Performance and Capacity Requirements

Application: Capacity 1000GB; Performance 1000 IOps (1.0 IOps/GB)

FC

SATA

14x 1TB 7.2k SATA(75 IOps/HDD; 9.8W)

1050 IOps14000 GB (!)

137.2 W

SATA SATAAccess Density:

1.1 IOps/GBAccess Density:0.075 IOps/GB

Page 32: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation33 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS4300: SATA vs. FC - Read IOps (RAID5)

2 Drawer FC Disks2 Drawer FC Disks

1 Drawer FC Disks1 Drawer FC Disks

SATA-1

IOps performance increases with no. of disks

Page 33: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation34 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

SATA vs FC - HDD Performance Positioning

up to 80%

around 45% of FC 15kSATA 7.2k

Fibre Channel (FC) disk drives / Serial Attached SCSI (SAS)Offer highest enterprise-class performance, reliability, and availability for business-critical applications requiring high I/O transaction performance (high access densities)

Serial Advanced Technology Attachment (SATA) disk drivesPrice-attractive alternative to the enterprise class FC drives for near-line applications with lower production costs, larger capacities but also lower specifications (e.g. rotational speeds, data rates, seek times)

SATA vs. FC Drive Positioning & Considerations

sequential workloads: SATA drives perform quite well with only about 20% reduction inthroughput compared to FC drives.

random workloads: SATA drive transaction performance is considerably below FCdrives and their use in environments with critical online transactionworkloads and lowest response times is not generally recommended!

SATA drives typically are very well suited for various fixed content, data archival, reference data, and near-line applications that require large amounts of data at low cost, e.g. bandwidth / streaming applications, audio/video streaming, surveillance data, seismic data, medical imaging or secondary storage. They also can be a reasonable choice for business critical applications in selected environments with less critical IOps performance requirements (e.g. low access densities).

Page 34: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation35 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)• New Trends & Directions: Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 35: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation36 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

RAID Level Comparison - RAID5 vs RAID10

RAID5• cost-effective with regard to performance and usable capacity (87.5% usable capacity for 7+P)• provides fault tolerance for one disk drive failure• data is striped across all drives in the array with the parity being distributed across all the drives• A single random small block write operation typically causes a “RAID5 write penalty”, initiating

four I/O operations to the disk back-end by reading the old data and the old parity block before finally writing the new data and the new parity block (this is kind of a worst-case scenario – it may take less operations when writing partial or even full stripes dependent on the I/Os in cache).

• On modern disk systems write operations are generally cached by the storage subsystem and thus handled asynchronously so that RAID5 write penalties are generally shielded from the users in terms of disk response time. However, with steady and heavy random write workloads, the cache destages to the back-end may still become a limiting factor so that either more disks or a RAID10 configuration might be required to provide sufficient disk back-end write performance.

RAID10• best choice for fault-tolerant, write-sensitive environments at the cost of 50% usable capacity• can tolerate at least one, and in most cases even multiple disk failures. • data is striped across several disks and the first set of disk drives is mirrored to an identical set.• each write operation initiates two write operations at the disk back-end

Page 36: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation37 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

RAID5 – Writing a single data block

ParityRAID5 (7+P) ARRAY

RAID5 - Read-Modify-Write: RAID5 Write PenaltyWorst case scenario with one write operation requiring four disk operations to array

(1) read old data (2) read old parity [ MODIFY ] (3) write new data (4) write new parity

= data being read from disk = data being written to disk

(1) (2)(3) (4)Cache

performing XOR calculation

Page 37: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation38 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Cache

RAID5 – Writing a full stripe

ParityRAID5 (7+P) ARRAY

RAID5 - Full Stripe WriteEspecially with large I/O transfer sizes or sequential workoads full stripe writescan be accomplished with RAID5 where the parity can be calculated on the fly

without the need to read any old data from the array prior to the write operation

= data being read from disk = data being written to disk

Page 38: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation39 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

RAID5 vs RAID10 – Backend I/O rate calculation example

Sustained front-end I/O rate: 1000 IOps (70:30:50)Example for a typical 70:30:50 random, small-block application workload(Read:write ratio = 70:30; Read cache hit ratio = 50%)

Sustained back-end I/O rate: 1550 IOps RAID5 vs 950 IOps RAID10

RAID5 : 1000 logical random IOps– 700 reads × 50% cache hits = 350 reads– 300 writes × 4 (write penalty: read old data/parity, write new data/parity)

= 1200 reads & writes– a total of 1550 physical IOps on the disks at the physical backend

RAID10 : 1000 logical random IOps– 700 Reads × 50% Cache Hits = 350 Reads– 300 Writes × 2 (two mirrored writes) = 600 Writes– a total of 950 physical IOps on the disks at the physical backend

RAID10 already outperforms RAID5 in a typical 70-30-50 workload.!!! Consider using RAID10 if random write percentage is higher than 35% !!!

Page 39: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation40 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

RAID5 vs RAID10 – Performance summary

RAID5 vs RAID10 - PerformanceRAID5 and RAID10 basically deliver a comparable performance for read operations.RAID5 tends to perform better than RAID10 for large block sequential writes.RAID10 always performs better than RAID5 for small block random writes.

RAID5 vs RAID10 - SelectionRAID5 is a good choice for most environments requiring high availability and fewer writes than reads (e.g. multi-user environments with transaction database applications and a high read activity).RAID10 should be considered for fault-tolerant and performance-critical, write-sensitive transaction processing environments with a high random write percentage above 35%.

50.0%o+++RAID10

87.5%++o+RAID5

Capacity 8-DDMs

Sequential Write

Sequential Read

Random Write

Random ReadRAID level

Page 40: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation41 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

RAID6 - Overview

675%2RAID-6, 6+P+Q

250%at least 1RAID-10, 4+4

487.5%1RAID-5, 7+P

Write penalty (Disk ops)Space efficiencyReliability

(#Erasures)RAID Level

RAID6: Dual parity RAID– DS8000: 5+P+Q+S or 6+P+Q arrays (using modified EVENODD code)– Survives 2 “erasures”

• 2 drive failures• 1 drive failure plus a medium error, such as during rebuild

(especially with large capacity drives)– Like RAID5, parity is distributed in stripes, with the parity blocks in a different place in

each stripe– RAID6 does have a higher performance penalty on write operations than RAID5 due to

the additional parity calculations.

RAID Level Comparison:

Page 41: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation42 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS8000 - Single Rank RAID Performance (1/2)

DS8000 R4.0(no IWC)

full stroke

DS8000 R4.0(no IWC)

full stroke

Page 42: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation43 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS8000 - Single Rank RAID Performance (2/2)

RAID6RAID6 RAID10RAID10RAID5RAID5

DS8000 R4.0(no IWC)

full stroke

DS8000 R4.0(no IWC)

full stroke

increasing write penalty

Page 43: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation44 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: 2.5" & Solid State Drive (SSD)• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 44: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation45 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

New Trends – Small-form-factor (2.5") SAS disk drivesHigh-density small-form-factor (2.5") SAS HDDs are replacing LFF (3.5") FC/SAS HDDs providing a high level of performance and energy efficiency on a smaller footprint:allowing higher packing densities with more disks in the same footprint consuming considerably less power than 3.5" 15k drivesincreasing system-level performance up to 115 percent over same speed drives offering both high transactional performance and low power consumption thus improving IOPS/W ratio considerably over comparable 3.5" 15k drives

Example: DS8800 (2.5" 146GB15k SAS HDDs) vs. DS8700 (3.5" 146GB15k HDDs)

Fully configured with a base frame and two expansion frames, the new DS8800 can reduce floor space requirements by 40% and energy requirements by over 35%, all while supporting more drives than a five-frame DS8700 model (e.g. 1056x 2.5" disks in 3 frames in DS8800 vs. 1024x 3.5" DDMs in 5 frames DS8700). The small-form-factor drives offer better performance at the same rotational speeds, as well as better energy usage per drive and at a lower cost per gigabyte than the large-form-factor enterprise Fibre Channel drives available on most high-end systems today.

10.2 W19.4 WPower per Disk

245 W310 WPower per Enclosure

24x 2.5" HDDs

DS8800

16x 3.5" HDDs

DS8700Estimated Storage Enclosure powerTable takes into account controller card power, power efficiencies, power for cooling, and power for disks.

Page 45: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation46 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

New Trends – Small-form-factor (2.5") SAS Disk Drive Specs

Source: www.seagate.com (2010)This is just an example for getting a view on typical disk drive characteristics. The chosen disk types above do not necessarily represent the characteristics of the disk drive modules used in IBM System Storage systems.

2.5" 146GB15k 3.5" 146GB15k

=6.95W

Page 46: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation47 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS8800 - Single Rank RAID5 performance: 2.5" vs. 3.5"

RAID6RAID6 RAID10RAID10RAID5RAID5

DS8700 R5.0 / DS8800 R6.0 full stroke

DS8700 R5.0 / DS8800 R6.0 full stroke

3.5" HDDs3.5" HDDs 2.5" HDDs2.5" HDDs

Page 47: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation48 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Processing Capabilities and Disk Performance over 50 years

1956 IBM RAMAC (1st disk drive) 5 MB storage, 1200 RPMdata transfer rate 8800 characters per second

2010 Enterprise FC Hard Disk Drive (HDD)600GB storage capacity, 15000 RPMdata transfer rate 122 to 204 MB/s

Last 50 years of HDD technology:HDD RPM: 12.5 xHDD Capacity 120 000 x

Ope

ratio

ns p

er s

econ

d

Time

Disk Performance

Processor C

apabilities

Performance Gap

New: SSD drives (STEC-inc):

0.1 MHz

4 GHz

Page 48: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation49 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

What are solid-state drives?

• Semiconductor (NAND flash, non-volatile)

• No mechanical read/write interface, no rotating parts:i.e. no seek time or rotational delays

• Electronically erasable medium

• Random access storage

• Capable of driving tens of thousands of IOpswith response times less than 1ms

• Absence of mechanical moving parts makes SSDssignificantly more reliable than HDDs

• Wear issues are overcome through over-provisioningand intelligent controller algorithms (Wear-Levelling)

Application benefits

Increased performance for transactional applications with high random IO rates (IOps):Online Banking / ATM / Currency Trading, Point-of-Sale Transactions / Processing, Real-time data mining

Solid state disks in DS8000 offer a new higher performance option for enterprise applications.

Best suited for cache-unfriendly data with high access densities (IOps/GB) requiring low response times

Additional benefit of lower energy consumption, cooling and space requirements (data center footprint)

New Trends & Directions - Solid State Drives (SSD)

Page 49: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation50 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid State Drives (SSD) - DS8300 R4.2 Single Rank PerformanceSingle RAID5 Rank - Random Read Single RAID5 Rank - Random I/O

Single RAID5 Rank - Random Read

Single RAID5 Rank - Sequential I/O

Random I/O: SSDs >> HDDsRandom I/O: SSDs >> HDDs

Sequential I/O: SSDs ~ HDDsSequential I/O: SSDs ~ HDDs

RAID5 Write-Penalty(1:4 Backend Ops)

RAID5 Write-Penalty(1:4 Backend Ops)

SSDs show exceptionally low response timesSSDs show exceptionally low response times

Source: IBM Whitepaper, IBM System Storage DS8000 with SSDs - An In-Depth Look at SSD Performance in the DS8000, http://www.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP101466

Page 50: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation51 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Note:A single rank is only serviced by a single Device Adapter (DA) of a DA pair and managed by either CEC#0 (rank group 0) or CEC#1 (reank group 1) after assignment to an extent pool.

Two or more ranks are required to be able to utilize the full I/O bandwidth of a DA pair by assigning half of the ranks of each DA pair to even extent pools (P0, P2, P4, ..., managed by CEC#0 / rank group 0) and half of the ranks to odd extent pools (P1, P3, P5, ..., managed by CEC#1 / rank group 1).

Solid State Drives (SSD) - DS8700 R5.1 Single Rank Performance

Source: IBM Whitepaper, IBM® System Storage™ DS8700™ Performance Whitepaper, ftp://public.dhe.ibm.com/common/ssi/sa/wh/n/tsw03053usen/TSW03053USEN.PDF

64kB Sequential IO

4kB Random IO

New 750GX Device Adapter

Page 51: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation52 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid State Drives (SSD) - DS8700 Single Rank Performance

IOps (12x)

DS8700 R5.1full stroke

DS8700 R5.1full stroke

RT

Page 52: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation53 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid State Drives (SSD) - DS8700 Single Rank Performance

IOps (>8x)

DS8700 R5.1full stroke

DS8700 R5.1full stroke

RT

Page 53: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation54 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid State Drive (SSD) – Latency and IOps vs I/O block size

Adding writes and/or increasing transfer size reduces SSDthroughput and increases latency substantially

Source: Session sDS10, Storage Performance Made Easy with Easy Tier and SSDs, IBM, IBM STG Technical Conference, Lyon, 2010

Page 54: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation56 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

SSD 11½ days

Application I/O – On a human Time Scale (SSDs vs. HDDs)

>10x fasterthan HDD

MEMORY 1:40 minutes

CPU 1 cycle := 1 second

(<1ms)>100xmore IOps than HDD

Page 55: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation57 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid State Drive (SSD) - Tiered Storage Concepts

Solid State Drive technology remains more expensive than traditional spinning disks, so the two technologies will coexist in hybrid configurations for several years.

Tiered storage is an approach of utilizing different types of storage throughout the storage infrastructure. Using the right mix of tier 0, 1, and 2 drives will provide optimal performance at the minimum cost, power, cooling and space usage.

Data Placement is key! To maximize the benefit of SSDs it is important to analyze application workloads and only place data which requires high access densities (IOps/GB) and low response times on them.

IBM System Storage DS8000 with SSDs - An In-Depth Look at SSD Performance in the DS8000http://www.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP101466Driving Business Value on Power Systems with Solid State Drivesftp://ftp.software.ibm.com/common/ssi/sa/wh/n/pow03025usen/POW03025USEN.PDF

7200 RPM HDDs (SATA):Lowest performance and cost/GBTier 2

15k RPM HDDs (FC/SAS):High performance lower cost/GBTier 1

Solid State Drives (SSD): Highest performance and cost/GBTier 0

SSDwhitepapers

cold

hot

data

Page 56: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation59 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Solid-state drives (SSDs) offer significantly improved performance compared to mechanical disk drives... but it takes more than just supporting SSDs in a disk subsystem for clients to achieve the full benefit:

Task: Optimizing data placement across tiers of drives with different price and performance attributes can help clients operate at peak price/performance. Implementing this type of optimization is a three-step process:

(1) Data performance information must be collected.(2) Information must be analyzed to determine optimal data placement.(3) Data must be relocated to the optimal tier.

Solution: With DS8700 R5.1 IBM introduced IBM System Storage Easy Tierwhich automates data placement throughout the DS8700 disk pool (including multiple drive tiers) to intelligently align the system with current workload requirements. This includes the ability for the system to automatically and nondisruptively relocate sub-volume data (at the extent level) across drive tiers, and the ability to manually relocate full volumes or merge extent pools. Easy Tier enables smart data placement and optimizes SSD deployments with minimal costs. The additional Storage Tier Advisor Tool provides guidance for SSD capacity planning based on existing client workloads on the DS8700.

IBM System Storage DS8700 R5.1 Announcement Letter (Easy Tier)http://www.ibm.com/common/ssi/rep_ca/5/877/ENUSZG10-0125/ENUSZG10-0125.PDF

IBM Redpaper: IBM System Storage DS8700 Easy Tierhttp://www.redbooks.ibm.com/abstracts/redp4667.html?Open

IBM DS8700 R5.1 Solid-State Storage Optimization with Easy Tier

Page 57: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation60 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Easy Tier optimizes SSD deployments by balancing performance AND cost requirements

Easy Tier delivers the full promise of SSD performance while balancing the costs associated with over provisioning this expensive resource

“Just Right”“Slower, inexpensive” “Fast, expensive”

IBM Easy Tier

LUN Heatmap

Page 58: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation61 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Smart data placement with Easy Tier: SPC-1 (SATA/SDD)

First ever Storage Performance Council (SPC-1) benchmark submission with SATA and SSD technology

Source:Storage Performance Council, April 2010: http://www.storageperformance.org/results/benchmark_results_spc1#a00092IBM Whitepaper, May 2010: IBM® System Storage™ DS8700™ Performance with Easy Tier®, http://www.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/WP101675

Easy Tier

0:00 2:00 4:00 6:00 8:00 10:00 12:00 14:00 16:00 18:00

Time

Thro

ughp

ut (I

O/s

)Over 3x IOPSImprovement

Increase ofover 3X!

System configuration: 16x SSD + 96x 1TB SATA

Page 59: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation62 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Smart data placement with Easy Tier: SPC-1 (SATA/SDD)

0.00

5.00

10.00

15.00

0 10000 20000 30000 40000 50000 60000

Throughput (IO/s)

Res

pons

e Ti

me

(ms)

192 FC HDD 96 SATA + 16 SSD

Improves RT in rangeof ordinary use

Dual frames Single Frame

Improves RT in range of ordinary use

Improves RT in range of ordinary use

96 SATA HDD RAID10 (no SSD)

192 FC HDD RAID5

96 SATA HDD RAID10 plus 16 SSD RAID5

SSD + SATA + Easy Tier Config vs. FC 15K HDDs Config (same capacity)

DS8700 R5.1 DS8700 R5.1

Page 60: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation63 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Smart data placement with Easy Tier: SPC-1 Backend I/O Migration

0

1

2

3

4

5

6

0 10 20 30 40 50 60 70 80 90

% Backend IO migrated

% C

apac

ity m

igra

ted

With approximately 4.9% of the SPC-1 data being migrated to SSDs about 76% of all backend IOs were moved to SSDs!

With approximately 4.9% of the SPC-1 data being migrated to SSDs about 76% of all backend IOs were moved to SSDs!

Page 61: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation64 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Sizing for Easy Tier – Skew Level

1.0

3.5

2.0

7.0

Skew Level

Disk Magic provides

3 predefined skew levels:

heavily skewed medium skewed lightly skewed

to predict the amount of the I/Oworkload that can be servicedby Solid State Drives (SSDs).

2.0

3.5

7.0

Skew Value

37%20%Light

55%20%Medium

80%20%Heavy

IOps on SSD

Capacity on SSD

Skew

20% capacityon SSD

55% IOps on SSD

37% IOps on SSD

80% IOps on SSD

no skew

(homogen

eous

acc

ess d

ensit

y)

Page 62: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation65 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Disk Magic model with Easy Tier

Page 63: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation66 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations• Performance Data Collection and Analysis

Page 64: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation67 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Logical Configuration - Basic PrinciplesThree major principles for the logical configuration to optimize storage subsystem performance:

Workload isolation (e.g. on extent pool and array level)– dedicate a subset of hardware resources to a high priority workload in order to reduce impacts of

less important workloads (protect the loved ones) and meet given service level agreements (SLAs)– limit low priority workloads which tend to fully utilize given resources to only a subset of hardware

resources in order to avoid impacting other more important workloads (isolate the badly behaving ones)

– provides guaranteed availability of the dedicated hardware resources but also limits the isolated workload to only a subset of the total subsystem resources and overall subsystem performance

Workload resource sharing– multiple workloads share a common set of subsystem hardware resources, such as arrays,

adapters, ports– single workloads now can utilize more subsystem resources and experience a higher performance

than with only a smaller subset of dedicated resources if the workloads do not show contention with each other

– good approach when workload information is not available, with workloads that do not try to consume all the hardware resources available, or with workloads that show workload peaks at different times

Workload spreading– most important principle of performance optimization, applies to both isolated workloads and

resource-sharing workloads– simply means using all available resources of the storage subsystem in a balanced manner by

spreading the workload evenly across all available resources that are dedicated to that workload, e.g. arrays, controllers, disk adapters, host adapters, host ports

– host-level striping and multi-pathing software may further help to spread workloads evenly

(1) Workload isolation (2) Workload resource-sharing (3) Workload spreading(1) Workload isolation (2) Workload resource-sharing (3) Workload spreading

Page 65: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation68 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Logical Configuration – DS8100/DS8300/DS8700 Examples

using Rotate Volumes

or Rotate Extents

Extent Allocation Method (eam)

Page 66: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation69 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Logical Configuration – DS8100/DS8300/DS8700 Examples

using Rotate Extents

Extent Allocation Method (eam)

(Storage Pool Striping)

Page 67: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation70 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Logical Configuration - DS8800 new high-density enclosures

3.5” (LFF) Fibre Channel, 2Gbps FCSupports 16 disks per enclosure

2.5” (SFF) SAS, 6Gbps SAS to disksSupports 24 disks per enclosure

DS8700 Megapack DS8800 Gigapack

Page 68: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation71 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Agenda

• Disk Storage System Selection & Specs

• Application I/O & Workload Characteristics

• Hard Disk Drive (HDD) Basics – It’s all mechanic

• HDD Performance & Capacity Aspects (SATA vs FC/SAS)

• RAID Level Considerations (RAID-5 / RAID-6 / RAID-10)

• New Trends & Directions: Solid State Drive (SSD)

• Basic Principles for Planning Logical Configurations

• Performance Data Collection and Analysis

Page 69: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation72 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Analyzing Disk Subsystem I/O PerformanceQuestions to ask when performance problems occur:

What exactly is considered to perform poorly? Which application, server, volumes?Is there a detailed description of the performance problem and environment available?What is the actual business impact of the performance problem?What was the first occurrance of the problem and were there any changes in the environment?When does the problem typically occur, e.g. during daily business hours or nightly batch runs?What facts indicate that the performance problem is related to the storage subsystem?What would be the criteria for the problem to be considered as solved? Any expectations?

Data to collect and analyze:description & config of the architecture (application – server – SAN – storage)application characteristics, logical and physical volume layout (usage, mapping server/storage)I/O performance data collection during problem occurrance on server and storage subsystem:(a) Server Performance Data Collection:

AIX # iostat [–sT|-sTD] [interval] [no. of intervals]# filemon –o fmon.log –O lv,pv; sleep 60; trcstop

Linux # iostat –x [interval] [no. of intervals]

Windows # perfmon→ GUI, then select Physical Disk Counters

(b) Storage Subsystem Performance Data Collection:DS3k/DS4k/DS5k (SMcli), XIV (XCLI), DS6k/DS8k and other (TPC for Disk)

Page 70: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation73 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 Performance Monitor

• only counters for quantity of processed I/Os up to current point in time• no counters for quality of processed I/Os as, for example, I/O service times• additional host system performance statistics required for I/O response times

Page 71: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation74 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 Performance Metrics

DS3000/DS4000/DS5000 Performance Metrics:

Total IOs (total number of processed IOs since start of data collection) Read Percentage (read percentage of all processed IOs since start of data collection) Cache Hit Percentage (read cache hit percentage of all processed read IOs) Current kB/second (average data rate in binary kB/s for current measurement interval)Maximum kB/second (maximum data rate in binary kB/s since start) Current IO/second (average I/O rate for the current measurement interval) Maximum IO/second (maximum I/O rate since start)

Please note: The Read Percentage and the (Read) Cache Hit Percentage provided by these native DS3000/DS4000/DS5000 performance statistics refer to the total number of I/Os (Total IOs) which have been processed during the the whole measurement so far (i.e. from the start of the performance data collection up to the current measurement interval). They do not solely refer to the current measurement interval. Read percentage and read cache hit percentage for the I/O rate of the current measurement interval can be derived from these values. However, due to the limited decimals for these percentages the calculation will lack accuracy with a growing number of Total IOs. If the change of Total IOsduring a measurement interval becomes less than 0.1% it is impossible to correctly calculate the read and read cache hit percentage for this interval anymore.

Page 72: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation75 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 Performance Data Collection

SMcli script for continuous performance data collection over given time frame:

on error stop; set performanceMonitor interval=60 iterations=1440; upload storageSubsystem file="c:\perf01.txt" content=performanceStats;

>smcli [IP-Addr. Ctr.A] [IP-Addr. Ctr.B] -f perfmon.scrPerforming syntax check...Syntax check complete.Executing script...Script execution complete.SMcli completed successfully.

Always collect the

Performance Statistics

together with latest

Subsystem Profile

to document the actualsubsystem configuration

used during data collection

perfmon.scr

Page 73: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation76 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 Performance Data Collection Example

Performance Monitor Statistics for Storage Subsystem: 174290U-13F1217-AS_DC1Date/Time: 4/13/05 7:20:12 AMPolling interval in seconds: 60

Devices,Total,Read,Cache Hit,Current,Maximum,Current,Maximum,IOs,Percentage,Percentage,KB/second,KB/second,IO/second,IO/secondCapture Iteration: 1Date/Time: 4/13/05 7:20:13 AMCONTROLLER IN SLOT A,593368.0,15.4,20.2,1516.6,1516.6,164.8,164.8,Logical Drive AIX01_09,38.0,28.9,54.5,0.1,0.1,0.0,0.0,Logical Drive AIX01_15,119.0,61.3,75.3,2.2,2.2,0.0,0.0,Logical Drive AIX02_08,59.0,27.1,37.5,0.1,0.1,0.0,0.0,[...]CONTROLLER IN SLOT B,2347017.0,59.4,34.5,16469.9,16469.9,651.8,651.8,Logical Drive AIX01_08,107.0,63.6,80.9,2.1,2.1,0.0,0.0,Logical Drive AIX01_10,112.0,67.0,73.3,2.2,2.2,0.0,0.0,Logical Drive AIX01_14,109.0,73.4,75.0,2.2,2.2,0.0,0.0,[...]STORAGE SUBSYSTEM TOTALS,2940385.0,50.5,33.6,17986.5,17986.5,816.5,816.5,[...]

For more information about how to collect and process these DS4000 performance statistics please see:How to collect performance statistics on IBM DS3000 and DS4000 subsystems (on IBM Techdocs)IBMers http://w3.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/TD103963IBM BPs http://partners.boulder.ibm.com/src/atsmastr.nsf/WebIndex/TD103963

Example of performance statistics file collected on DS4000 with firmware < v7.xx

Page 74: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation77 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 Performance Data Collection Example

"Performance Monitor Statistics for Storage Subsystem: DS4700_PFE1 -Date/Time: 12.02.08 10:29:13 - Polling interval in seconds: 20"

"Storage Subsystems ","Total IOs ","Read Percentage ","Cache Hit Percentage ","Current KB/second ","Maximum KB/second ","Current IO/second ","Maximum IO/second"

"Capture Iteration: 1","","","","","","","""Date/Time: 12.02.08 10:29:14","","","","","","","""CONTROLLER IN SLOT A","0.0","0.0","0.0","0.0","0.0","0.0","0.0""Logical Drive Data_1","0.0","0.0","0.0","0.0","0.0","0.0","0.0""Logical Drive Data_3","0.0","0.0","0.0","0.0","0.0","0.0","0.0"[...]"CONTROLLER IN SLOT B","0.0","0.0","0.0","0.0","0.0","0.0","0.0""Logical Drive Data_2","0.0","0.0","0.0","0.0","0.0","0.0","0.0""Logical Drive Data_4","0.0","0.0","0.0","0.0","0.0","0.0","0.0"[...]"STORAGE SUBSYSTEM TOTALS","0.0","0.0","0.0","0.0","0.0","0.0","0.0"[...]

Example of performance statistics file collected on DS4000 with v7.xx firmware

For more information about how to collect and process these DS4000 performance statistics please see:How to collect performance statistics on IBM DS3000 and DS4000 subsystems (on IBM Techdocs)IBMers http://w3.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/TD103963IBM BPs http://partners.boulder.ibm.com/src/atsmastr.nsf/WebIndex/TD103963

(same format as DS3000/DS5000 performance statistics)

Page 75: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation78 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 – Performance Data Analysis

Subsystem total IOps / MBps (average / peak)Controller A and B total IOps / MBpsIdentify busiest volumesIdentify busiest arrays

Verify if- Array/volume configuration - RAID level- Disk typeis appropriate for the workload

Verify if workload distributionis balanced across all arraysand both controllers

Evaluate response times with appropriate Disk Magic models

Page 76: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation79 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3/4/5000 – Performance Analyzer Tool (IBM internal only)

Excel based DS4000 Performance Analyze Tool http://w3-03.ibm.com/support/techdocs/atsmastr.nsf/WebIndex/PRS3088

MS Excel spreadsheet for aquick import and analysisof DS4000 performance

statistic outputs and DS4000 profile with export feature for

generating a html report

MS Excel spreadsheet for aquick import and analysisof DS4000 performance

statistic outputs and DS4000 profile with export feature for

generating a html report

Page 77: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation80 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS3000/4000/5000 – Performance Data CollectionStorage Explorer Lite (LSI):

Free tool from LSI to collect subsystem performance statistics from a selection of DS3000/4000/5000 subsystems providing historical performance information:

IOps, MBps, Read Cache Hit, Read:Write Ratio, Average I/O Size

Runs on Windows 32-bit platforms only, including Windows XP and Windows Vista. Installation utility installs the Microsoft .NET 3.5 Framework and Microsoft SQL Server Express 2008 (requires SP2 for Windows XP) from active internet connection.

Requires registration at: http://www.lsi.com/DistributionSystem/User/Login.aspx

Page 78: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation81 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

XIV – XIVGUI Performance Data Collection

Page 79: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation82 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

XIV – XCLI Performance Data Collection

XCLI (one command line):>xcli -m IPADDR -u USER -p PASSWD -s -y

statistics_get start=2009-10-07.11:00 count=300interval=1 resolution_unit=minute > C:\xiv_20091007.csv

Page 80: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation83 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DS6000/DS8000 – DSCLI Performance Metrics Examplesdscli> showfbvol -metrics 2000Date/Time: 24. April 2007 14:32:15 CEST IBM DSCLI Version: 5.2.2.224 DS: IBM.2107-7503461ID 2000Date 04/24/2007 14:30:25 CESTnormrdrqts 17normrdhits 5normwritereq 121050normwritehits 121050seqreadreqs 0seqreadhits 0seqwritereq 151127seqwritehits 151127cachfwrreqs 0cachfwrhits 0cachefwreqs 0cachfwhits 0inbcachload 0bypasscach 0DASDtrans 29seqDASDtrans 0cachetrans 33315NVSspadel 0normwriteops 0seqwriteops 0reccachemis 2qwriteprots 0CKDirtrkac 0CKDirtrkhits 0cachspdelay 0timelowifact 0phread 25phwrite 33420phbyteread 5phbytewrite 2082recmoreads 2sfiletrkreads 0contamwrts 0PPRCtrks 0NVSspallo 272177timephread 28timephwrite 40138byteread 0bytewrit 8508timeread 4timewrite 4061

dscli> showrank -metrics r2Date/Time: 24. April 2007 14:37:43 CEST IBM DSCLI Version: 5.2.2.224 DS: IBM.2107-7503461ID R2Date 04/24/2007 14:35:53 CESTbyteread 587183bytewrit 287002Reads 1176760Writes 315629timeread 2509716timewrite 392892

dscli> showioport -metrics I001Date/Time: 24. April 2007 14:41:47 CEST IBM DSCLI Version: 5.2.2.224 DS: IBM.2107-7503461ID I0001Date 04/24/2007 14:39:56 CESTbyteread (FICON/ESCON) 0bytewrit (FICON/ESCON) 0Reads (FICON/ESCON) 0Writes (FICON/ESCON) 0timeread (FICON/ESCON) 0timewrite (FICON/ESCON) 0bytewrit (PPRC) 0byteread (PPRC) 0Writes (PPRC) 0Reads (PPRC) 0timewrite (PPRC) 0timeread (PPRC) 0byteread (SCSI) 56586bytewrit (SCSI) 454426Reads (SCSI) 414404Writes (SCSI) 4906333timeread (SCSI) 2849timewrite (SCSI) 111272

Page 81: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation84 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Subsystem Performance Monitoring

The IBM Tivoli Storage Productivity Center (TPC) is a suite of storage infrastructure management tools for storage environments by centralizing, simplifying and automating storage tasks associated with storage systems, Storage Area Networks (SAN), replication services and capacity management. IBM Tivoli Storage Productivity Center for Disk (TPC for Disk) is an optional component of TPC, that is designed to manage multiple SAN storage devices and to monitor the performance of SMI-S compliant storage subsystems from a single user interface.IBM Tivoli Storage Productivity Center Standard Edition includes three components of the TPC suite as one bundle at a single price: TPC for Data, Fabric and Disk.New customers with IBM System Storage Productivity Center (SSPC) which includes the pre-installed (but separately purchased) IBM Tivoli Storage Productivity Center Basic Edition only need to purchase the additional „TPC for Disk“ component to be able to collect performance statistics from their supported IBM storage subsystems.TPC for Disk is the official IBM product for clients requiring performance monitoring of their IBM storage subsystems (e.g. DS4k, DS5k, DS6k, DS8k, SVC, ESS, 3584 Tape, ...)TPC V4.x introduced Tivoli Common Reporting (TCR) for creating customized reports from TPC database with BIRT (Business Intelligence Reporting Tools) & IBM Cognos 8 Business Intelligence (Version 8.4)

Page 82: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation85 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Subsystem Performance Reports

Select to initiate the report creation

3

1

2

4

Page 83: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation86 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Subsystem Performance Reports

Select for creating a chart

Page 84: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation87 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Subsystem Performance Reports

Page 85: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation88 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Export Subsystem Performance Reports

Select to export performance data as CSV output fileusing ‚File > Export Data‘ dialog

Page 86: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation89 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Analyzing Reports in a Spreadsheet

Page 87: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation90 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Reports of Interest by Subsystem

ESS, DS6000 and DS8000: By Storage SubsystemBy ControllerBy ArrayBy VolumeBy Port

SAN Volume Controller:By Storage SubsystemBy IO GroupBy NodeBy Managed Disk GroupBy VolumeBy Managed DiskBy Port

DS4000 and other supported SMI-S compliant storage subsystems:By Storage SubsystemBy VolumeBy Port

Some reports may give more or less data, depending on

the exact level of SMI-S compliance by the vendor

supplied CIM agents.

Don’t forget to export a complete set of reports for the subsystem of interest, e.g. for a DS8000:

20080131-75APNK1-subsystem.csv,

20080131-75APNK1-controller.csv,

20080131-75APNK1-ports.csv,

20080131-75APNK1-arrays.csv,

20080131-75APNK1-volumes.csv

Limit the reports to a representative time frame as the amount of data especially for the volume report can be

extremly large!

Page 88: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation91 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – How to start with Performance MonitoringSimply start monitoring and thus understanding the current workload patterns (workload rangeand workload profile) developing over the day/week/month for normal operation conditions where no end-user complaints are present. Develop an understanding of the expected behaviour. I/O rates and response times may vary considerably from hour to hour or day to day simply due to various application loads, business times and changes in the workload profile. You may even experience times with high I/O rates and extremly low response times (e.g. high cache hit ratios) as well as times with only moderate I/O rates but higher response times (e.g. lower cache hit ratios) still not being of any concern. Appropriate thresholds for I/O rates and response times can be derived from these statistics based on particular application and business requirements.Regularly collect selected data sets for historical reference and do projections of workload trends. Evaluate trends in I/O rate and response time and plan for growth accordingly. Typically response times increase with increasing I/O rates. Historical performance data is the best source for performance and capacity planning.Watch for any imbalance of the overall workload distribution across the subsystem resources. Avoid single resources from becoming overloaded (hot spots). Redistribute workload if needed.When end-user performance complaints arise simply compare current and historical data and look for appropriate changes in the workload that may lead to performance impacts.Additional performance metrics may help to better understand the workload profile behind the changes in I/O rates and response times:

Read:Write ratioRead Cache Hit Percentage [%]avg. Read/Write/Overall Transfer Size [kB] per I/O operation

Required for appropriateDisk Magic models and

performance evaluations

Page 89: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation92 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TPC for Disk – Basic Performance MetricsThere are lots of performance metrics available. Which ones are best to start with?Most important metrics for a storage subsystem are:

I/O Rate: number of I/O operations per second [IOps or IO/s]Response Time (RT): average service time per I/O operation in milliseconds [ms]

These metrics are typically available for read operations, write operations and the total number of processed I/O operationson subsystem, controller, port, array, volume, I/O group, node, mdisk & mdisk group level

Basic performance statistics to look at for storage subsystems are in principle: front-end I/O statistics on subsystem level for overview of system overall workloadfront-end I/O statistics on volume level for selected critical applications / host systemsbackend I/O statistics on array level (i.e. on the physical disk level / spindles)

General thresholds for front-end statistics are difficult to provide, becauseI/O rate thresholds depend on workload profile and subsystem capabilitiesRT thresholds depend on application, customer requirements, business hours

Additional metric is Data Rate: throughput in megabytes per second [MBps]on subsystem level for overview of overall throughputon port level together with Port RT for overview of port and I/O adapter utilization

Page 90: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation93 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

In general, there do not exist typical values or fixed thresholds for all performance metrics as they typically strongly depend on the nature of the workload:

Online Transaction Processing (OLTP) workloads (e.g. database)- small transfer sizes (4kB...16kB) with high I/O rates- low front-end response times around 5ms commonly expected

Backup, batch or sequential-like workloads- large transfer sizes (32kB...256kB) with low I/O rates but high data rates- high front-end response times even up to 30ms still can be acceptable

Subsystem level front-end metrics (subsystem total average):- Overall Response Time < 10ms

Array level back-end metrics (physical disk access):- Back-end Read Response Time < 25ms- Disk Utilization Percentage << 80%- I/O rate: depends on RAID level, workload profile, number and speed of DDMs

considered very busy with I/O rates near or above 1000 I/Os (DS8000/DS6000)Volume level front-end metrics (I/O performance as experienced by the host systems):

- Overall Response Time < 15ms (depends on application requirements and workload)- Write-cache Delay Percentage < 3% (typically should be 0%)

These values are just some suggestions as rules of thumb

to start with. In general appropriate thresholds

need to be based on the client‘s particular environment

and application requirements.

TPC for Disk – Basic Guidelines for DS8000

Page 91: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation94 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

Page 92: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation95 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

ESCC – Storage is our profession!

Client trainingworkshops,

seminars

Channel / skillenablement,certification

Lab validation,proof of concept

Showcases,remote demo,new products

Client strategyworkshops

Custom software& solutions

End-to-endclient support

& services

Storage technical

assistanceUsergroups,

Client councils

http://escc.mainz.de.ibm.com

Page 93: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation96 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

DisclaimerCopyright © 2011 by International Business Machines Corporation.

No part of this document may be reproduced or transmitted in any form without written permission from IBM Corporation. Product data has been

reviewed for accuracy as of the date of initial publication. Product data is subject to change without notice. This information could include technical

inaccuracies or typographical errors. IBM may make improvements and/or changes in the product(s) and/or programs(s) at any time without notice.

Any statements regarding IBM's future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only.

References in this document to IBM products, programs, or services does not imply that IBM intends to make such products, programs or services

available in all countries in which IBM operates or does business. Any reference to an IBM Program Product in this document is not intended to state or

imply that only that program product may be used. Any functionally equivalent program, that does not infringe IBM's intellectually property rights, may

be used instead. It is the user's responsibility to evaluate and verify the operation of any on-IBM product, program or service.

The performance information contained in this document was derived under specific operating and environmental conditions. The results obtained by

any party implementing the products and/or services described in this document will depend on a number of factors specific to such party’s operating

environment and may vary significantly. IBM makes no representation that these results can be expected in any implementation of such products

and/or services. Accordingly, IBM does not provide any representations, assurances, guarantees, or warranties regarding performance.

THE INFORMATION PROVIDED IN THIS DOCUMENT IS DISTRIBUTED "AS IS" WITHOUT ANY WARRANTY, EITHER EXPRESS OR IMPLIED. IBM

EXPRESSLY DISCLAIMS ANY WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR NONINFRINGEMENT.

IBM shall have no responsibility to update this information. IBM products are warranted according to the terms and conditions of the agreements (e.g.,

IBM Customer Agreement, Statement of Limited Warranty, International Program License Agreement, etc.) under which they are provided. IBM is not

responsible for the performance or interoperability of any non-IBM products discussed herein.

The provision of the information contained herein is not intended to, and does not, grant any right or license under any IBM patents or

copyrights. Inquiries regarding patent or copyright licenses should be made, in writing, to:

IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY, 10504-1785, U.S.A.

Page 94: GSchmidt - IntroDiskPerformance_V1.63

© 2011 IBM Corporation97 IBM Systems Technical University & STG Technical Enablement Conference, April 2011, Prague2011-04-04

A practical Introduction to Disk Storage System Performance

TrademarksThe following are trademarks of the International Business Machines Corporation in the United States, other countries, or both.

* All other products may be trademarks or registered trademarks of their respective companies.

Notes:

Performance is in Internal Throughput Rate (ITR) ratio based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput that any user will experience will vary depending upon considerations such as the amount of multiprogramming in the user's job stream, the I/O configuration, the storage configuration, and the workload processed. Therefore, no assurance can be given that an individual user will achieve throughput improvements equivalent to the performance ratios stated here. IBM hardware products are manufactured from new parts, or new and serviceable used parts. Regardless, our warranty terms apply.All customer examples cited or described in this presentation are presented as illustrations of the manner in which some customers have used IBM products and the results they may have achieved. Actual environmental costs and performance characteristics will vary depending on individual customer configurations and conditions.This publication was produced in the United States. IBM may not offer the products, services or features discussed in this document in other countries, and the information may be subject to change without notice. Consult your local IBM business contact for information on the product or services available in your area.All statements regarding IBM's future direction and intent are subject to change or withdrawal without notice, and represent goals and objectives only.Information about non-IBM products is obtained from the manufacturers of those products or their published announcements. IBM has not tested those products and cannot confirm the performance, compatibility, or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products.Prices subject to change without notice. Contact your IBM representative or Business Partner for the most current pricing in your geography.

The following are trademarks or registered trademarks of other companies.Adobe, the Adobe logo, PostScript, and the PostScript logo are either registered trademarks or trademarks of Adobe Systems Incorporated in the United States, and/or other countries.Cell Broadband Engine is a trademark of Sony Computer Entertainment, Inc. in the United States, other countries, or both and is used under license therefrom. Java and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other countries, or both. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both.Intel, Intel logo, Intel Inside, Intel Inside logo, Intel Centrino, Intel Centrino logo, Celeron, Intel Xeon, Intel SpeedStep, Itanium, and Pentium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries.UNIX is a registered trademark of The Open Group in the United States and other countries. Linux is a registered trademark of Linus Torvalds in the United States, other countries, or both. ITIL is a registered trademark, and a registered community trademark of the Office of Government Commerce, and is registered in the U.S. Patent and Trademark Office.IT Infrastructure Library is a registered trademark of the Central Computer and Telecommunications Agency, which is now part of the Office of Government Commerce. LSI is a trademark or registered trademark of LSI Corporation.

AS/400®, e business(logo)®, eServer, FICON, IBM®, IBM (logo)®, iSeries®, OS/390®, pSeries®, RS/6000®, S/30, VM/ESA®, VSE/ESA, WebSphere®, xSeries®, z/OS®, zSeries®, z/VM®, System i, System i5, System p, System p5, System x, System z, System z9®, BladeCenter®, System Storage, System Storage DS®, TotalStorage®

For a complete list of IBM Trademarks, see www.ibm.com/legal/copytrade.shtml

Not all common law marks used by IBM are listed on this page. Failure of a mark to appear does not mean that IBM does not use the mark nor does it mean that the product is not actively marketed or is not significant within its relevant market.

Those trademarks followed by ® are registered trademarks of IBM in the United States; all others are trademarks or common law marks of IBM in the United States.