7/26/2019 Exchange_best_practice.pdf
1/41
Engineering White Paper
EMC CLARiiON Storage Solutions
Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines
Abstract
This white paper presents the latest storage configuration guidelines and best practices for Microsoft Exchangeon CLARiiON storage systems. It is focused on Exchange 2003, but most material also applies to Exchange
2000.
Published 8/8/2005
7/26/2019 Exchange_best_practice.pdf
2/41
8/8/2005
Copyright 2005 EMC Corporation. All rights reserved.
EMC believes the information in this publication is accurate as of its publication date. The information is
subject to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED AS IS. EMC CORPORATION
MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THEINFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED
WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicablesoftware license.
Part Number H1363
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 2
7/26/2019 Exchange_best_practice.pdf
3/41
8/8/2005
Table of Contents
Executive Summary............................................................................................5
Intended Audience..............................................................................................5Introduction.........................................................................................................5
Environmental Parameters for Storage Design ........................................................................... 5
User Community Information.................................................................................................... 5Backup/Recovery Requirements.............................................................................................. 6Other Organizational Requirements or Constraints ................................................................. 7
Planning Storage for the Exchange Production Data......................................7
Exchange Storage Groups........................................................................................................... 7
How Many ESGs per Server? .................................................................................................. 7How Many Databases per ESG? ............................................................................................. 8How Many LUNs per ESG?...................................................................................................... 8
Calculating the Base I/O per User Requirement.......................................................................... 9
Calculating the IOPS Requirement for an Exchange Environment ........................................... 10
RAID Types and the Read/Write Ratio ...................................................................................... 10
Other Factors That May Impact I/O ........................................................................................... 11
Calculating the Capacity Requirement for Database LUNs....................................................... 12
Choosing a RAID and Disk Type ............................................................................................... 13
Comparing RAID 1/0 to RAID 5.............................................................................................. 14Comparing 10K rpm to 15K rpm............................................................................................. 16Comparing 73 GB, 146 GB, and 300 GB ............................................................................... 16Capacity Check ...................................................................................................................... 16Summary ................................................................................................................................ 17
MetaLUNs .................................................................................................................................. 17
Building Blocks ....................................................................................................................... 18Log LUN Configuration .............................................................................................................. 18
Additional Storage Considerations for the Exchange Production Data ..................................... 19
Public Folders......................................................................................................................... 19SMTP Queue.......................................................................................................................... 20Keeping EDB Files and STM Files Together ......................................................................... 20Smaller Exchange Environments........................................................................................... 20
Planning Storage for Local Recovery .............................................................21
SnapView for Disk-Based Replication ....................................................................................... 21
Clone-Based Replication........................................................................................................ 21Snapshot-Based Replication.................................................................................................. 22Comparison of Local Replication Options .............................................................................. 23
Online Backup to Disk................................................................................................................ 23Recovery Storage Groups ......................................................................................................... 24
Planning Storage for Local Message Archiving.............................................24
Storage-System Considerations......................................................................25
iSCSI Guidelines........................................................................................................................ 26
CLARiiON Storage Systems Comparison ................................................................................. 27
Putting It All Together ......................................................................................28
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 3
7/26/2019 Exchange_best_practice.pdf
4/41
8/8/2005
Consider Site-Specific Constraints......................................................................................... 28Configure the Cleanest Looking Layout Diagram .................................................................. 28Plan Throughout for Operational Resiliency .......................................................................... 28Validate the Design ................................................................................................................ 29
Additional Recommendations for Optimal Performance ..............................29
Storage-System Tuning ............................................................................................................. 29
Exchange Server and Windows Environment ........................................................................... 30Windows File-System Alignment ............................................................................................... 30
Conclusion ........................................................................................................31
Appendix A: Storage Design Examples..........................................................32
Example 1 .................................................................................................................................. 32
RAID-Adjusted Back-end Disk IOPS Calculation................................................................... 32Capacity Calculation with 200 MB Mailboxes (73 GB Drives) ............................................... 32Capacity Calculation with 400 MB Mailboxes (73 GB Drives) ............................................... 32Matching Up the IOPS and Capacity Requirements.............................................................. 33
Example 2 .................................................................................................................................. 33
Capacity Calculation with 200 MB Mailboxes ........................................................................ 34
Matching Up the IOPS and Capacity Requirements.............................................................. 34Example 3 .................................................................................................................................. 34
Calculations............................................................................................................................ 35Appendix B: Quantifying Exchange User Activity .........................................39
Determining the Peak Activity Period ........................................................................................ 39
Measuring IOPS per User.......................................................................................................... 40
Read/Write Ratio........................................................................................................................ 40
Performance Counter Guidelines............................................................................................... 40
Appendix C: Additional Resources .................................................................41
EMC White Papers .................................................................................................................... 41
Microsoft White Papers.............................................................................................................. 41
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 4
7/26/2019 Exchange_best_practice.pdf
5/41
8/8/2005
Executive SummaryThis white paper proceeds through the step-by-step process of planning the storage layout for Exchangedata on an EMCCLARiiONstorage system. It offers considerations and the latest best practice
recommendations along the way. The approach taken is to design a layout that meets the following goals:
Optimal performance During peak periods, user response times are still acceptable and there is nobuildup of mail queues.
Efficient backup and rapid recovery Backups complete within the allotted window, with anacceptable impact on the production environment. Local recovery meets the Service Level Agreement
(SLA) requirement.
Simplicity of design The resulting configuration is straightforward to implement and easy tomanage.
In addition to recommendations for production data storage layout, the paper includes considerations for
configuring CLARiiON storage for Exchange backup, local replication, and archiving.
Intended Audience
The intended audience for this white paper is system engineers who have customers interested inimplementing Microsoft Exchange using EMC CLARiiONFibre Channel storage.
The reader should have a general knowledge of Microsoft Exchange and Windows technology, as well as
an understanding of basic CLARiiON features and terminology.
IntroductionIts said that if you ask 10 consultants to architect an Exchange storage design for an organization, you willget back 10 (or more) different design proposals. CLARiiON storage systems offer a great deal of
flexibility with various combinations of RAID and disk types. Many Exchange storage configurations will
work well on CLARiiON systems, but some may cause unforeseen bottlenecks. As new versions of
Exchange are released and as CLARiiON array technology advances, best practice information for
Exchange storage design is constantly evolving.
Environmental Parameters for Storage DesignSeveral factors figure into the storage design for an Exchange environment. The more you know about an
organizations use of the existing messaging system and the better defined the requirements are for the new
implementation, the closer you should be able to come to constructing a useful design.
This section describes data items to gather. Whenever possible, it is always valuable to have this data
supported by empirical measurements (concurrent users, IOPS, read/write ratio, log files/day, etc.) from the
current environment.
User Community Information
The information gathered in this category should lead to a good estimate of the I/O profile for a set of usersover time. It should also lead to determining the users storage requirements. Appropriate tools and
counters for quantifying the following information are included in Appendix B:
How many total users?
Today
Anticipated growth over the next few years
How many concurrent users during the peak period?
How many mailboxes not associated with an individual user (such as a central help desk mailbox)?
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 5
7/26/2019 Exchange_best_practice.pdf
6/41
8/8/2005
What mail client is used?
Outlook (2003 cached, or other)
Outlook Web Access
Mobile devices (Blackberry)
When are the peak activity periods?
What is the typical working day? Is there geographic dispersal of the users across time zones?
What is the Exchange activity level of the users?
Categorization of the user types leads to estimated base IOPS demand (see Table 1 on page 9)
Measured I/O in the existing environment gives the best starting point
Are there special category userswith different security, performance, or backup/recoveryrequirements?
What are the mailbox size limits?
Is there anything else pertinent that helps to describe the user profile for this organization?
Heavy use of personal folders?
Do users often send large documents?
Integrated use of voice mail?
Considerable use of Outlook 2003 shared folders?
What are the characteristics of public folder usage?
Size of the public store
Replication activity among public stores
Backup/Recovery Requirements
The choice of backup and recovery method will play an important part in the resulting storage design.
Once again, measurement of the existing environment will provide the best starting point for the new
design.
What is the deleted item retention period?This is the time period that Exchange maintains an item after a user has deleted ittypically 10 to 30
days.
What is the chosen backup method?
Backup to disk using standard Exchange online backup
Backup to tape using standard Exchange online backup
Clone-based replication (with archival backup to disk or tape)
Uses an EMC application such as Replication Manager SE (RMSE) or Replication Manager (RM)to create physical copies of the Exchange data on CLARiiON clones (BCVs). Archival backup to
disk or tape performs an offline copy of the Exchange data files (databases and logs) from the
BCV.
Snapshot-based replication (with archival backup to disk or tape)
Uses an EMC application such as RMSE or RM to create copies of the Exchange data onCLARiiON snaps. Archival backup to disk or tape performs an offline copy of the Exchange data
files (databases and logs) from the snapshot.
What is the timing of the backup activity?
What are the requirements (service-level agreement) for recovery?
Is a distance replication or disaster recovery solution planned?
DR site distance
Network connection
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 6
7/26/2019 Exchange_best_practice.pdf
7/41
8/8/2005
Other Organizational Requirements or Constraints
This category covers any additional pertinent factors that have already been decided, or have been added as
a requirement.
What are the type, number, and location of Exchange servers?
Are the Exchange servers clustered?
What is the planned Exchange front-end/back-end server layout?
What is the SAN/network structure?
Is there an existing CLARiiON storage system?
What other software will be operating in the Exchange environment?
Antivirus
E-mail archiving solution
Applications integrated with Exchange (e.g., for workflow, etc.)
Exchange-integrated third-party tools (e.g., for mailbox recovery, enhanced indexing, etc.)
Planning Storage for the Exchange Production DataThis section provides guidelines for configuring storage to handle the production Exchange data. Whendesigning a storage configuration for Exchange, the first disk measurement to consider must be the I/O
operations per second (IOPS) that the Exchange environment requires. Once you have calculated thenumber of drives necessary to meet the I/O demand, then determine the capacity requirement and adjust the
drive count upward if necessary.
Exchange Storage GroupsThe Exchange storage group (ESG) is the fundamental unit for layout planning. When backing up
Exchange, the elements of an ESG should be treated together.
How Many ESGs per Server?
In the past, Microsoft has recommended that to make efficient use of CPU and memory, the administratormust fill up an ESG with users before creating additional Exchange storage groups. However,
improvements in Exchange (starting in Exchange 2000 SP2) allow the use of Exchanges maximum of fourproduction storage groups without running out of system resources. Because there is a single set of log
files for all databases within one storage group, there are advantages to configuring more ESGs on a server
even when you dont have to. With Exchange 2003, in most cases it will be best to use all four storage
groups.
Following are some considerations for the various ESG configuration options.
Using All Four ESGs
Works well with most servers deployed todayat least dual processor and 1 GB memory.
Offers the best granularity for performance. Using multiple Exchange storage groups results in morelog operations in parallel. Since database performance depends on log file performance, increasing the
number of logs can increase overall performance.
Offers the most granular management. Using four storage groups allows maintenance and recoveryoperations that are fully compartmentalized and affect the least amount of users possible.
Allows for full separate treatment of a set of users with different performance, security, orbackup/recovery requirements.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 7
7/26/2019 Exchange_best_practice.pdf
8/41
8/8/2005
Using the Fewest ESGs Possible
Makes the most efficient use of server resources and thus is better for underpowered servers. This hasbecome a less-significant (potentially negligible) factor with newer servers and Exchange 2003.
Fewer ESGs may be easier to manage.
This comes at the expense of lost flexibility and added exposure. Downtime or data loss caused by theloss of a particular database or storage group affects more users.
Using Two or Three ESGs
If a server does not have the resources to handle four ESGs, spreading users across an extra storagegroup or two can be a compromise.
Appropriate for a smaller number of mailboxes on the server (~1500 or less).
Some organizations may prefer to reserve the fourth storage group for growth, low-risk testing, orsome recovery scenarios.
In some cases, the optimal disk layout may align better with two or three ESGs for performance orcapacity reasons.
How Many Databases per ESG?
There can be up to five databases in each Exchange 2000 or 2003 storage group. Following are someconsiderations for database configuration.
Using Five Databases
A particular database can become logically corrupt. With five databases, logical problems on onedatabase will affect the data of the fewest number of people.
For some backup methods, a single database can be restored without affecting the users in otherdatabases in the ESG. Thus database recovery affects the fewest number of people.
Makes for databases of the smallest possible size. If you are allowing space for offlinedefragmentation, you can keep all databases on the same LUN, but need to allow defrag space for thesize of just one database (the largest).
Using Less Than Five Databases
Administration may be a little easier when operating manually.
This is one way of reserving for growth.
Single-instance store is on a per-database basis. Spreading users across fewer databases may providestorage and performance benefits (e.g., when sending a large document to a large distribution list) from
single-instance storage.
In some cases, the optimal disk layout may align better with three or four ESGs for performance orcapacity reasons.
How Many LUNs per ESG?
In most cases, two LUNs should be allocated for each ESGone for the transaction logs, and one for the
databases (edband stmfiles).
The reasons for placing all ESG databases together, and some considerations for variations, are described inthe next sections.
Maintaining All Databases on the Same LUN
Less complexity for backup/recovery and disk configuration.
Offers the optimal capacity utilization for offline database defragmentation.
Allows growth space to be shared for any of the databases.
Best performance for volume shadow copies. Fewer LUNs provide additional safety margin forcompleting VSS split operations within the allotted 10-second time window.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 8
7/26/2019 Exchange_best_practice.pdf
9/41
8/8/2005
Distributing Databases on Multiple LUNs
Allows for more granular restore from clones (but during recovery, all databases in the ESG must bedown anyway because of the resetting of the checkpoint file).
In the case of very large mailboxes, LUN sizes can approach 1 TB or more. Splitting the ESGdatabases across two or three LUNs can improve clone synchronization times and the performance of
incremental SAN Copy from a clone.
Calculating the Base I/O per User RequirementThe best way to provide enough I/O to your application, especially in a large Exchange environment, is toknow your users usage profile. Sizing of the storage infrastructure should be based on a careful analysis of
the number of current and anticipated users, and their messaging habits and patterns. The fundamental
calculation concerns I/Os per second (IOPS) per user.
Table 1 describes four Exchange user categories and provides an estimate of the IOPS demand per user for
each1.
Table 1. Exchange User Profiles
Typical UserProfile
Description Mailbox Size Expected I/Os per Second
Light POP3 Hosted Internet mail < 25 MB 0.08
Light MAPI Infrequent e-mail access
Small mailboxes
2000 users)
Large mailboxes (>200 MB)Start with a rough estimate that for each doubling of the mailbox size over 100 MB you increase the
IOPS per user by about one third.
Regularly sending very large documents (>5 MB)Integrated voice mail is roughly equivalent to large documents.
Blackberry client usersCount each Blackberry user as the equivalent of two to three typical users.
JournalingThis adds significant overhead. Start with an estimate double the I/O.
1IOPS numbers referenced from the MicrosoftExchange Server 2003 Performance and Scalability Guide
white paper.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 9
7/26/2019 Exchange_best_practice.pdf
10/41
8/8/2005
A typical read/write ratio in Exchange is from 2:1 to 3:1. A ratio lower than this (higher percentage of
writes) will also increase the IOPS/user on RAID storage. This is discussed inRAID Types and the
Read/Write Ratioon page 10.
Organizations typically use many mailboxes in Exchange that are not associated with an individual user.
These mailboxes often serve as the central contact point for a group, for conference rooms, or mailboxesused by integrated applications. Depending on the number of these unassociated mailboxes (it can be
significant10 percent or more) and their activity level (which can vary significantly), it may beappropriate to factor these into the IOPS calculation. By default, treat these mailboxes as equivalent to thetypical user mailboxes, and always include them in the capacity calculation.
Calculating the IOPS Requirement for an Exchange EnvironmentThis is a key step in configuring CLARiiON storage for good Exchange performance. It is a best practice
to configure dedicated disk drives for the Exchange databases. Calculate the periods of highest I/O demandduring the day by looking at the anticipated cumulative effect of user activity, system activity (virus
checkers), and background activity (local or remote replication). Balance the I/O where possible with
scheduling (backup during off-peak hours) and even distribution of users across ESGs. Then, plan a designthat will handle the resulting peak I/O load.
Start with the measurement or estimate of the I/O profile of the Exchange users in the organization. Plan
for the peak user load timetypically mid-morning on Monday.
RAID Types and the Read/Write RatioDepending on the particular organizational requirements, there are two possible RAID type options that can
be appropriate for production Exchange database LUNs:
RAID 1/0 This offers the best performance with high protection, but only 50 percent of the RAIDgroup capacity is usable. It is frequently recommended because it provides sufficient space across thenumber of spindles required for handling peak I/O load with todays larger disk drives. On a RAID
1/0 LUN, there are two physical I/O operations for each write requested (a write to each mirrored
disk), described as a write penalty of two.
RAID 5 This configuration offers a higher usable capacity per RAID group than RAID 1/0. It canbe effective for environments with very large mailboxes and/or lower IOPS requirements. However, in
a RAID 5 group there are four physical I/O operations for each write requested (two reads to calculateparity, one write for data, and one write for parity).
Regardless of the RAID type chosen, it is important to configure enough drives to handle the I/O demand.
Table 2. Write Penalty by RAID Type
RAID Type Write Penalty
RAID 1/0 (Striping + Mirroring) 2
RAID 5 (Striping + Parity) 4
An example with 3,000 Exchange users, separated evenly into four storage groups, runs through the setof calculations in the next few sections. The example is highlighted in a series of text boxes.
1000 heavy users at a peak of 1 IOPS each [1000 IOPS]
2000 typical users at a peak of .5 IOPS each [1000 IOPS]
A general read/write ratio of 2:1
Maximum concurrency of active users at 90%
This results in a requirement of [(1000 + 1000) x .9] = 1,800 host-based IOPS for the 3,000 users
during their peak activity period.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 10
7/26/2019 Exchange_best_practice.pdf
11/41
8/8/2005
Use this formula to adjust the base user I/O requirements in the ESG by figuring in the write penalty for
each RAID group:
(Base IOPS x Read %) + (Base IOPS x Write % x Write Penalty)
= RAID-Adjusted Back-End IOPS
Carrying on the example above, with 1800 total base IOPS and a 2:1 read/write ratio:For RAID 1/0:
(1800 x 2/3) + (1800 x 1/3 x 2) = 1200 + 1200 = 2400 IOPS
For RAID 5:
(1800 x 2/3) + (1800 x 1/3 x 4) = 1200 + 2400 = 3600 IOPS
Other Factors That May Impact I/OThere are other administrative operations that may impact I/O. Several of these should be scheduled totake place only during off-peak times (see the following examples). This additional I/O activity must be
accounted for by increasing the IOPS capacity by some amount over the RAID-adjusted IOPS requirement.To estimate the amount of I/O overhead these additional activities will cost, it is best to perform tests in anenvironment that match the target production environment as closely as possible. The less sure you are of
this overhead, the more capacity you should assign to ensure good performance.
Examples of Background I/O Activity That Cannot Be Scheduled to Off-Peak Times
High load on the server The more active mailboxes a server is managing and the less memory it has,the less likely any particular users mailbox will be cached. If this has not been figured into the per-user IOPS requirement, it should be figured in here. For example, going from 2,500 to 4,000 users on
a system can increase the IOPS per user by about 10 percent.
Server-based antivirus protection Besides the extra reads, antivirus software can add 20 percent ormore to the CPU utilization of the Exchange server.
Integrated features and applications The impact here depends on the number and type of any
integrated features (such as content indexing) or applications (such as workflow), and the amount oftheir use.
Synchronous or asynchronous mirroring If a mirroring solution is used for distance replication ordisaster recovery, it should be factored in very carefully.
The total I/O demand during peak user load can be calculated by adding the cumulative overhead of the
background activity to the RAID-adjusted IOPS requirement. This overhead is often calculated simply as
an added percentage, but some activities (such as virus checkers) involve primarily reads, in which case its
more accurate to ignore the write penalty in the calculation.
RAID-Adjusted User IOPS + I/O Overhead = IOPS Requirement at Peak User Activity
For example, if operating on an active 3000 user Exchange server running a frequently used
workflow application, it would be reasonable to start with an estimated overhead percentage of 20
percent. Continuing the example to calculate the IOPS requirement during peak user activity:
RAID 1/0:
2400 + 20% = 2880 IOPS
RAID 5:
3600 + 20% = 4320 IOPS
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 11
7/26/2019 Exchange_best_practice.pdf
12/41
8/8/2005
There are other schedulable activities that can significantly increase I/O demand on the Exchange LUNs.
You should understand their impact. It is important to factor these activities into the peak I/O requirements
if they are scheduled to take place during a period of high user activity.
Examples of Schedulable Activities
Online backup to disk or tape
This places heavy read activity on the production LUNs. There is added overhead if the Exchange
server is used to manage the backup.
Local clone-based replication
Clone-based replication using an application such as RMSE involves synchronizing all source
Exchange LUNs in the ESG to their clones. During the incremental synchronization part of the
process, there is heavy back-end read activity against the production LUNS.
Once the copy is complete on the clones, Eseutil performs an integrity check on each page of the
database replicas. The validated copy may also then be archived or copied (via SAN Copy) to aremote site. These checking and archiving functions cause high read activity, but it is to the clones
rather than the production LUNs. As long as best practice is followed and the clone LUNs are on
separate drives from production LUNs, this heavy read activity comes into calculations only in termsof backup scheduling and use of overall array resources. A single Eseutil process can use up to 30
percent of the CPU of one SP.
Local snap-based replication
Snaps will have a more significant impact on production Exchange LUNs than clone-based replication.
If snaps are chosen as a backup method, it is particularly important to factor in the associated impact.
This is discussed further in the section Snapshot-Based Replicationon page 22.
Exchange online maintenance
By default, Exchange schedules online database maintenance for a 4-hour period nightly to performfunctions that include clearing out deleted mailboxes and deleted items that have gone past their
retention period, plus online defragmentation. This timing and duration can be adjusted.
Because of the heavy I/O that online maintenance adds, schedule it to take place during the period of
lightest activity. At the beginning of online maintenance, Exchange performs an Active Directorylookup for each user in the database. Slightly offsetting the online maintenance start times of the
databases will reduce the impact of these searches on the Active Directory.
You cannot perform a backup on a database at the same time it is undergoing online maintenance.Maintenance will pause until the backup job for that database completes, but it will not extend
operation past its allotted time window. Take this into consideration to ensure that online maintenancegets enough time each day.
In summary, take additional I/O activity into consideration when calculating the anticipated demand. Then,
design to accommodate the peak I/O load with that overhead factored in.
Calculating the Capacity Requirement for Database LUNsUsually the I/O requirements will determine the number of drives required, but it is also necessary to know
the space requirement for the databases in the storage group. For certain environments (such as large
mailboxes or low IOPS per user) the storage capacity requirements may call for more disk space thanperformance needs dictate. Comparatively, this is an easy calculation.
For each category of users in the ESG, multiply the maximum allowed mailbox size by the number of users
in that category. If the number of users is planned to grow, factor that in here. Allow an additionalpercentage of the database size for deleted item retention (~ 10 percent for a typical 30-day retention
period). Because its important not to run out of space on the LUN, allow an additional buffer of 10 to 20
percent of the sum of the databases on the LUN.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 12
7/26/2019 Exchange_best_practice.pdf
13/41
8/8/2005
Offline DefragmentationWith Exchange 2003, in most cases it is no longer necessary to perform offline defragmentation of the
databases. Normal online maintenance will defragment the database, but it will not compact the size of the
file. The only way to actually shrink the database size is to perform offline defragmentation, where the
database is dismounted and Eseutil is used to rebuild a new copy. To have the shortest time offline, thisrebuild is performed on the same LUN as the source. There must be free space equaling at least 110 percent
of the size of the database for the rebuild. If multiple databases are stored on the same LUN, enough space
must be allowed to handle the rebuild of the database with the largest size.
Summarizing the CalculationSpace Required for the ESG Database LUNs =
Maximum Mailbox Size * Number of mailboxes
+ Extra space for the deleted item retention
+ Public Folder space (if part of the ESG)
+ 10% to 20% free space for growth protection
+ Space for offline defragmentation if required
When four or five databases are planned for each storage group, deleted item retention is typical, and spaceis required for offline defragmentation; you can perform a quick capacity calculation by simply allocating
free space equal to 50 percent of the mailbox total (62.5 x 1.5 = 93.75).
For example, with 750 users in one ESG, distributed evenly across five databases but all on the sameLUN (no offline defrag space included):
250 heavy users @ 100 MB Mailbox 25 GB
500 typical users @ 75 MB Mailbox 37.5 GB
Sum requirement for mailboxes 62.5 GB
Add 10% for deleted item retention 62.5 x 1.1 = 68.8 GB
Size of each database 68.8 / 5 = 13.8 GB
Add 15% for extra free space 68.8 x 1.15 = 79.1 GB
To include space for offline defragmentation, add space equivalent to the size of the largest database.
With space for offline defragmentation 79.1 + 13.8 = 92.9 GB(extra free space is already available)
Choosing a RAID and Disk TypeRegardless of RAID type, a physical disk can handle a certain number of Exchange-style IOPS. The IOPS
capacity of disks continues to improve with new disk models, but the performance improvement has not
kept pace with the increase in storage capacity. Consequently, most Exchange disk configurations today
are determined by I/O requirements rather than capacity.There is not consistent agreement on the IOPS capability of a disk drive. Although some sequential I/O
tests have indicated that a CLARiiON 10K rpm drive can perform at a speed greater than 300 IOPS, a more
practical value to use with the Exchange 4 KB random I/Os on CLARiiON has been determined to be in thevicinity of 130 IOPS. Similarly, a practical value to use for 15K rpm drives is 180 IOPS.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 13
7/26/2019 Exchange_best_practice.pdf
14/41
8/8/2005
Table 3. IOPS per Spindle
Disk rpm CLARiiON Disk I/O Capacity
with Exchange Databases
10K 130 IOPS
15K 180 IOPS
Using these values, you can divide into the RAID-adjusted IOPS requirement to construct a table indicating
the number of drives needed to handle the I/O demand of the ESG. Although in the end you may be
required to round up the number of drives to meet RAID 1/0 or RAID 5 requirements, the following tableleaves the number as calculated (rounded up to the next whole drive).
RAID-Adjusted IOPS / IOPS per Disk = Drive Count for the Exchange Database LUNs
Continuing the example, calculating the drive requirement for the RAID 1/0 IOPS total (2880) and RAID 5
IOPS total (4320):
Table 4. Disk-Drive Requirements by RAID Type and Disk Speed All Users
RAID 1/0 RAID 5
10K rpm 23 (2880/130) 34 (4320/130)
15K rpm 16 (2880/180) 24 (4320/180)
Since its advisable to lay out Exchange databases on dedicated drives, the most straightforward design is
to dedicate one or more RAID groups for each ESG.
If the 3000 users are spread across four storage groups, applying the example to some dedicated RAID
configurations:
Table 5. Disk-Drive Requirements by RAID Type and Disk Speed Per ESG
RAID 1/0 RAID 5
10K rpm 6 (3+3) 10 (4+1)s
9 (4+1) and (3+1)
15K rpm 4 (2+2) 6 (5+1)
This assumes that the I/O requirements for each storage group are the same. In practice, some fine tuning
is often required because of varying user counts and activity level.
Comparing RAID 1/0 to RAID 5
It is an accepted notion that RAID 1/0 is a better choice for random-write environments like Exchange. Theeffect is somewhat subtle; since all writes hit the write cache, RAID 1/0 and RAID 5 RAID groups perform
equally well until the storage system is sufficiently busy and the write cache becomes saturated. The
advantage of using RAID 1/0 rather than RAID 5 with Exchange is that RAID 1/0 groups can flush the
cache of Exchanges random write load about 15 percent to 30 percent faster than RAID 5 groups. Thisequates to cache-speed performance at a higher random-write load. Additionally, rebuild times and rebuild
impact is reduced with RAID 1/0 in the event of disk failures (see Table 6.,reproduced from the CLARiiON
Best Practices for Fibre Channel Storagewhite paper).
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 14
7/26/2019 Exchange_best_practice.pdf
15/41
8/8/2005
Table 6. RAID Types and Relative Performance in Failure Scenarios
RAID Type Rebuild IOPS Loss Rebuild Time Impact of Second Failure
during Rebuild
RAID 5 50% 15% to 50% slower than RAID 1/0 Loss of data
RAID 1/0 20% - 25% 15% to 50% faster than RAID 5 Loss of data 14% of time in aneight-disk group (1/[n-1])
RAID 1* 20% - 25% 15% to 50% faster than RAID 5 Loss of data
* RAID 1 is not a recommended RAID type for Exchange database LUNs, but it may be appropriate for some smaller
log LUN configurations.
The downside may be cost. RAID 1/0 requires more drives for a given capacity, but that extra capacity may
not be valuable when the spindle count is determined by I/O throughput. Consider the following
calculation to determine the number of users that one physical drive can handle:
User IOPS per Drive = IOPS per Drive x Host-Based IOPS / Back-End IOPS
The ratio of host-based IOPS to back-end IOPS depends on the read/write ratio and the RAID type. Theeasiest way to get this figure is to add the left and right sides of the read/write ratio and divide by the (read
ratio + write ratio x write penalty).
RAID 1/0, User IOPS per Drive with a 3:1 read/write ratio on a 10K drive:
130 x (3 + 1) / (3 + 1 x 2) = 130 x 4 / 5 = 104 IOPS
RAID 5, User IOPS per Drive with a 3:1 read/write ratio on a 10K drive:
130 x (3 + 1) / (3 + 1 x 4) = 130 x 4 / 7 = 74 IOPS
Users per Drive = User IOPS per Drive / IOPS per User
RAID 1/0, .8 IOPS per user:
104 / .8 = 130 users per drive
RAID 5, .8 IOPS per user:
74 / .8 = 92 users per drive
Table 7.7 uses these calculations for six drives, configured as RAID 1/0 and RAID 5, with a few different
mailbox sizes, allowing 50 percent free space on the LUN for deleted item retention, growth, and offlinedefragmentation.
Table 7. Capacity Comparison for Six 73 GB Drives by RAID Type and Mailbox Size
Required Capacity
Usable
Capacity
Users @ .8
IOPS75 MB
Mailboxes
150 MB
Mailboxes
250 MB
Mailboxes
RAID 1/0 (3+3)10K
198 GB 780 (130x6) 88 GB 176 GB 293 GB
RAID 5 (5+1)10K 330 GB 552 (92x6) 62 GB 124 GB 207 GB
Note that once the mailbox size gets much over 150 MB, there is not enough space on the RAID 1/0 groupto accommodate the number of users whose I/O it can handle. This does not take into consideration the
fact that the increased mailbox size causes the IOPS per user to increase.
RAID 5 becomes more appropriate when the capacity requirements are high, relative to the I/O demand,
such as when the I/O demand is low (< .4 IOPS) or the mailbox limit is high (>250 MB).
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 15
7/26/2019 Exchange_best_practice.pdf
16/41
8/8/2005
Comparing 10K rpm to 15K rpm
Another choice for buying performance is the 15K rpm drive. The 15K rpm drive offers up to 30 percent
better performance over 10K rpm drives in the kind of random-access case that Exchange presents. The
increased speed assures the write cache avoids saturation and keeps writes going at cache speeds. This doesnot apply to the sequential access log devices, where the two RAID types perform about the same.
Comparing 73 GB, 146 GB, and 300 GB
Smaller drives offer more performance per gigabyte. However, the performance/gigabyte curve drops aslarger drives are deployed. Base the disk-size decision for production Exchange drives on:
The I/O capacity of the drive (number of users it will support, when averaged across the RAID group).
The average maximum mailbox size for those users.
Design simplicity and flexibility.
For example, if a RAID 1/0 3+3 group will support the I/O demand of 600 Exchange users with an averagemailbox maximum of 100 MB, 73 GB drives provide more than enough storage capacity. In some cases,
where the predominant disk on the storage system is a larger size or faster speed, it may make sense to
standardize on that type to simplify LUN layout. Refer to Appendix A for some comparative examples.
Capacity Check
The amount of usable space in a RAID group can be calculated with the following formula2:
RAID 1/0:
Usable space of a RAID 1/0 Group = Usable Drive Space x (# Drives / 2)
RAID 5:
Usable space of a RAID 5 Group = Usable Drive Space x (# Drives 1)
Table 8 compares the usable capacity of 10 drives configured as RAID 1/0 and as RAID 5.
Table 8. Usable Capacity of 10 Drives by Disk Size and RAID Type
Raw Capacity Usable Capacity
per Drive
10 Spindles
Usable Capacity
5+5 R1/0
10 Spindles
Usable Capacity
Two 4+1 R5
36 GB * 33 165 (33x10/2) 264 (2x33x[5-1])
73 GB 66 330 528
146 GB 134 670 1072
300 GB 268 1340 2144
* No longer sold
Typically, a disk layout that meets the I/O requirements for an ESG will contain more than enough capacityto meet the ESG storage requirements. However, if the mailboxes are particularly large, the I/O demand per
user is very low, or if a large amount of space is required for offline defragmentation, the storage
requirements may affect the design.
2For exact figures, especially with the use of vault drives, refer to the CLARiiON Capacity Calculator athttp://clariipub.corp.emc.com/cse/capacitycalc/capacitycalc.htm
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 16
http://clariipub.corp.emc.com/cse/capacitycalc/capacitycalc.htmhttp://clariipub.corp.emc.com/cse/capacitycalc/capacitycalc.htm7/26/2019 Exchange_best_practice.pdf
17/41
8/8/2005
Referring to Table 8.8, if the required storage for the ESG went beyond 330 GB, the RAID 1/0 spindle
count of 10 for 36 GB and 73 GB drives would not be adequate. Options would be to increase the drive
count, move up to 146 GB drives, or switch to RAID 5.
Using the calculation of 92.9 GB per ESG from the example, disk space requirements would not be a
factor except for a RAID 1/0 2+2 configuration of 36 GB drives.
Appendix A includes some additional examples of I/O and capacity calculations.
Summary
Because storage capacity of disk drives has outpaced their increase in I/O throughput, IOPS capacity is the
standard to use today when determining the required number of drives. Since a RAID 1/0 group canperform better than that of a RAID 5 group under certain I/O loads with the same number of spindles, and
since the space disadvantage of RAID 1/0 has become less significant, we recommend RAID 1/0 as the
default choice for Exchange database volumes. RAID 5 may be appropriate for some customers depending
on I/O load and cost considerations.
The decision to choose between 10K or 15K drives will likely come down to cost. Since IOPS are the
determining factor in how many disks you need, the number of 15K drives required will usually be lessthan the 10K drive requirement. Balance the additional cost per drive against savings on additional drives,
DAEs, and possibly cabinets.
MetaLUNsCLARiiON storage systems can combine multiple LUNs into a larger metaLUNthat spans multiple RAIDgroups. MetaLUNs offer two primary advantages: I/O load balancing and expandability. Usually you will
configure an ESG for its maximum anticipated size at the start, but it is possible to use a metaLUN to
handle gradual growth.
The main advantage is the ability to distribute I/O over many spindles without resorting to host striping.
Striped volumes are particularly advantageous with workloads such as Exchange that are random andbursty, and metaLUNs make the use of striped volumes simple. Suppose that you have planned two ESGs
to reside on their own 3+3 RAID 1/0 groups, for a total of 12 spindles. An alternative design would be tocreate a metaLUN for each of the ESGsboth spanning the two 3+3 groups. The same 12 spindles wouldstill be handling the I/O of the two ESGs, but in this case, if the I/O demand of one ESG is higher than the
other, the combined load will be balanced across all of the drives. This can help to avoid an I/O bottleneck
for a particular ESG. The two storage groups would also share the cost of a disk rebuild. By spanning two
RAID groups, they double the risk of being affected by a rebuild but a rebuild will only affect half the disks
of the metaLUN.
Figure 1. MetaLUNs Sharing Two RAID Groups
Another choice would be a single 6+6 RAID group. It would yield the same performance, but with less
growth potential. To accommodate growth, additional free space on the existing RAID group set can be
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 17
7/26/2019 Exchange_best_practice.pdf
18/41
8/8/2005
used by concatenating the metaLUN. RAID group expansion also provides room for metaLUN growth (by
concatenating the metaLUN), along with added performance.
For added load balancing across a RAID group set, you can interleave the data of multiple storage groups
on the RAID set by creating metaLUNs for each ESG in the following order (illustrated in Figure 2):
Stripe the first component of ESG1
Stripe the first component of ESG2 Concatenate the next component of ESG1
Concatenate the next component of ESG2
Figure 2. Interleaving MetaLUNs
Building Blocks
Configuring metaLUNs can add a level of complexity to an Exchange design. In practice, the metaLUNconfiguration performs well and is easier to manage if you choose from a limited number of proven RAID
group types and sizes. These groups serve as flexible building blocks that will be easy to multiply out in
the full Exchange storage design.
Following are recommended building block elements that are small enough to be flexible, allow for growth,
and have been proven to perform well:
RAID 10: 3+3; 4+4; 5+5
RAID 5: 4+1
Its possible to overuse metaLUNs. While offering the benefits of flexibility and load balancing,metaLUNs residing on the same RAID group all share the risk of a physical disk failure in that group. It isworth considering the consequences of a RAID group failure when you plan your metaLUN layout.
Typically, two or three ESG metaLUNs sharing a RAID group set become a practical limit.
Log LUN ConfigurationWhen configuring disks for Exchange, most attention is paid to the database LUNs because they typically
represent the highest risk of performance bottleneck. But the database performance depends on the log
response time. Database transactions are gated by the completion of the associated log write.
When choosing a RAID type for log file LUNs, I/O performance and data protection are the overriding
factors rather than capacity. RAID 1/0 is the best RAID type to use for log LUNs. It provides better
response time than RAID 5 in degraded situations. In the case of a disk failure, RAID 1/0 rebuildscomplete faster than RAID 5. The longer the rebuild period, the more vulnerability there is to data loss.
Data loss always occurs if a second drive is lost during rebuild of a RAID 5 or RAID 1 group (see Table 6on page 15).
Although writes to the log LUN are sequential, performance tests have shown that you can take best
advantage of a set of drives by sharing a set of log LUNs on them. A rough rule of thumb to use for log
drives is to allocate one-eighth to one-tenth the number of spindles you have allocated for the databases,rounding up for RAID 1/0. For example, if you have calculated the need for 36 drives to handle the
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 18
7/26/2019 Exchange_best_practice.pdf
19/41
8/8/2005
databases of four Exchange storage groups on a server, you could allocate four drives in a RAID 1/0 2+2
group to handle the four transaction log LUNs for that server.
To avoid added recovery complications in the unlikely case of the physical loss of a RAID group, the log
files for multiple servers should be stored on separate drives.
There are some other important considerations in the design of the disk layout for the Exchange transactionlogs:
There is one set of transaction logs for each ESG (i.e., the logs for all databases in the ESG arecombined).
For manageability and flexibility, each set of log files should reside on its own LUN. This is an actualrequirement for VSS-based backups.
The transaction log files for an ESG should always be on a separate LUN from their associateddatabases. The log LUN should never share the same spindles as the database LUN for the same
ESGeven on small systems. This is for protection rather than performance. If something should
happen to the database, the log files are essential to recover transactions since the last backup. If those
log files reside on the same physical disk and that disk is damaged, this option is lost.
Log I/Os are 100 percent writes. They are most frequently 512-byte writes, but can be up to 64 KB orlarger.
Host I/O to the log LUN equals approximately 10 to 15 percent of the host I/O to the database LUNs.
The size of each Exchange log file is 5 MB. Most online Exchange backup processes will delete logfiles whose transactions have been committed to the database. It is important to confirm that
committed log files are being deleted to avoid running out of space on the log LUN.
Circular logging is an Exchange feature that causes log files to be deleted after their transactions havebeen committed to the database. Only a handful of the logs are maintained at any time to save space.
However, this sacrifices the ability to recover a database up to the minute. It is off by default andshould never be turned on.
Calculating Log LUN Storage Capacity RequirementsYou can calculate storage capacity requirements for a log LUN by multiplying 5 MB by the maximum
number of log files maintained before being pruned.
If you dont know the maximum number of log files generated in a day, you can use a rough rule of thumbof one log file per user per day. (Microsoft uses an estimate of two logs per day for configuration; EMCs
actual log file rate is considerably less than one per day.)
Unless you specify otherwise, typical Exchange online backups will prune the log files on each run. For
example, if a backup is run nightly, the log LUN for a 1,000-user ESG would minimally need space for
1,000 log files, or 5 GB. Be sure to allow extra capacity to ensure that the log file LUN never runs out of
space.
Additional Storage Considerations for the Exchange ProductionDataThe previous configuration guidelines have covered design recommendations for database and log LUNs
for medium-to-large Exchange implementations. This section describes some additional considerations.
Public Folders
Its difficult to provide general guidelines for configuring public folder storage because their usage varies
so widely. Some organizations practically ignore the existence of Exchanges public folder capability,while others use them extensively for shared document repositories, discussions groups, shared calendars,
and several other purposes. By default, the public store is contained within a dedicated database in the first
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 19
7/26/2019 Exchange_best_practice.pdf
20/41
8/8/2005
storage group; but when used actively, it is often configured on its own Exchange server with at least one
replicated copy to another Exchange server.
The best starting point for planning public folder storage for a newly migrated Exchange 2003 environment
is to examine the current I/O, storage usage, and growth rate for public folders in the current environment.
Adjust these measurements as necessary based on any planned changes to the public folder use policy (suchas adding a new integrated application that makes significant use of public folders, or switching shared
documents to file shares). Then, deploy using the principles described in this paper as with any otherExchange storage group.
SMTP Queue
The SMTP message queue should be placed on CLARiiON storage. It is not necessary to create replicas of
the SMTP queue. However, the queue can be placed on one of the database LUNs for the server as long asthe additional I/O and capacity requirements are factored in.
Keeping EDB Files and STM Files Together
An Exchange message store (database) consists of an EDB filecontaining all content generated by MAPI
clients, indexed message properties, and moreand a streaming file containing content generated by
Internet clients. Since the EDB file and STM file compose a complete message store, it is advisable tokeep them together on the same LUN.
Smaller Exchange Environments
If a relatively small (fewer than 20 database drives or 1,000 users) Exchange system is implemented on a
CLARiiON system, the same I/O requirements apply for the database LUNs, and thus it remains important
to allocate enough drives. It still may make sense to use RAID 1/0 for the database drives.
Database and log files should still be kept on separate RAID groups. For these smaller configurations, itmay be appropriate to use a RAID 1 pair for log files. If the number of drives is tight, the extra capacity on
the log drives could be used for some other light purpose.
If there are only two RAID groups available for Exchange data, the users can be split across two ESGs and
the logs placed with the databases for the alternate ESG. However, you must then factor in the log I/Orequirement when calculating the spindle count.
Figure 3. Sharing a Database LUN with the Logs from an Alternate Storage Group
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 20
7/26/2019 Exchange_best_practice.pdf
21/41
8/8/2005
Planning Storage for Local RecoveryThis section provides guidelines for configuring storage to handle local replicas and disk-based backups ofExchange production data.
SnapView for Disk-Based ReplicationSnapViewis an optional software package for the EMC CLARiiON storage system3. Using SnapView,
users can create a point-in-time viewor multiple viewsof a LUN, which can subsequently be made
accessible to another server, or simply held as a point-in-time copy for possible restoration. For instance, a
system administrator can make the SnapView replica accessible to a backup server so that the productionserver can continue processing without the downtime traditionally associated with backup processes. In the
event of a data corruption on the source LUN, SnapView replicas can be used to restore the contents of a
corrupted LUN to the point-in-time creation of the replica. SnapView can create replicas using either
clones (BCVs) or snapshots.
There are currently two products sold by EMC that facilitate the management and integration with
Exchange 2003 and the Windows Volume Shadow Copy Service (VSS) to create verified SnapView
replicas of an ESG, ready to be restored immediately:
Replication Manager/SE (RMSE)
Replication Manager (RM)
Clone-Based Replication
The backup option that provides the most rapid recovery of an Exchange database today is clone-basedreplication. SnapView clones provide users the ability to create fully populated copies of LUNs within a
single array. Once synchronized, clones can be fractured from their source, and then presented to a
secondary server for read and write access. Following the initial synchronization, clones can beincrementally resynchronized, where only the data that has changed on the source since the clone was
fractured will be copied to the clone. In the event of a data corruption on the source LUN, clones can be
used to restore the source LUN via a reverse synchronizationoperation. This returns the source LUN to the
point-in-time view of the source as it was when the clone was fractured. In the unlikely event of a
hardware error on the LUN (for instance, a multiple-drive failure), the clone can be repurposed as theproduction LUN. Thus, clones provide protection against both software errors and hardware errors.
RM and RMSE allow you to configure up to eight clones for each of the Exchange production LUNs.
Consider the following when clone-based replication is chosen as the backup method:
Know the characteristics of the clone backups youll be using.
Number of clones for each production LUN Each extra clone maintains a validated backup copyfor a different point in time.
Timing and frequency of the backup operation Spread out backups of the storage groups andtime them for low-usage periods to take best advantage of array resources.
Fibre drives are recommended for clones. Their performance is well-suited to cloneresynchronizations. ATA drives are not recommended for clone LUNs in an active Exchange
environment because the clone resynchronization operation to an ATA drive is considerably (up toseveral times) slower. The likelihood of affecting the production environment during this time
increases. Additionally, with an IOPS rate of less than half that of fibre drives, the backup window
will also be extended on ATA drives by an Eseutil check that takes longer to complete.
Clone LUNs can be bound on drives that differ in size, speed, RAID geometry, or even drive type. Forinstance, some users may elect to use RAID 1/0 for their production LUNs, and RAID 5 for their
3SnapView is supported on all CLARiiON models, CX300 and higher. The only CLARiiON platform on
which SnapView is not supported is the CX200.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 21
7/26/2019 Exchange_best_practice.pdf
22/41
8/8/2005
clones. RAID 5 is recommended for clones because they dont have the same IOPS requirements, and
provide greater capacity. Additionally, some users may elect to put production data on 15K rpm
drives, and use 10K rpm drives for their clones.
Eseutil will be the most I/O-intensive activity on the database clones. Use the Eseutil throughputrequirements when determining the spindle count here.
When using RAID 5 with clones, take modified RAID 3 (MR3) support into consideration. RAID 5
4+1 sets offer a good balance of rebuild time with MR3 support4.
146 GB and 300 GB drives may also be more appropriate with clones. The combination of RAID 5and larger drives allows you to configure sufficient extra capacity to store multiple clones on the sameRAID group. Since clone synch operations are scheduled, the I/O capacity for a set of drives can be
shared to handle multiple backups occurring at different times.
Do not place clone LUNs in the same RAID group that contains their source LUN.
Plan clone layouts to avoid backups occurring simultaneously on different LUNs configured on thesame RAID group. LUN resynchronizations and Eseutil integrity checks both involve a heavy amount
of I/O. Trying to pair two or more of these activities at the same time will slow each other down and
possibly affect user response times.
Consider using metaLUNs for clones to provide more spindles to improve Eseutil performance.
Keep in mind the building blocks discussion on page 18. For simplicity, it is advisable to try to selectyour RAID group from the recommended choices, and expand using this as a base.
To add some extra protection, if you are configuring multiple clones for an Exchange LUN, its best toalternate the clones on separate spindle sets.
Snapshot-Based Replication
RM and RMSE can create a VSS shadow copy via a SnapView snapshot. As with clones, snapshots
provide users a readable and writeable LUN replica. Snapshots, however, are not fully populated copies.
They use a pointer-and-copy-based design, where pointers map to data regions on the source LUN untilthey are changed, at which point the original data is copied to a reserved area and the pointers are
redirected accordingly. In this way, users have only to allocate sufficient disk space to accommodate the
changes to the source LUN.
The process of allocating the pointers according to a particular point in time is referred to asstartinga
SnapViewsession. To see the contents of a particular session, a user can activate a snapshotto the session.
The reserved LUNis the private LUN that contains the original source data, and the process of copying thatdata is referred to as copy on first write(since it must only occur on the initial change to the source LUN).
As with clones, SnapView session data can be used to restore a corrupted source LUN. Much like the
reverse synchronizationthat clones offer, SnapView sessions can be rolled backto the point-in-time view
of when the session was started.
Snap-based replication is not an ideal backup method to use with Exchange for the following reasons:
SnapView copy on first write (COFW) must be completed before allowing a change to productiondata. This causes additional overhead on the production LUNs, especially for the period shortly afterthe snap session is started, and can noticeably impact performance. Even if the replication is
performed during off hours, the snap session remains active during the day to hold the backup copy. In
that case, the highest COFW activity takes place when users become active the following morning. The Eseutil integrity check on the snap results in very heavy read activity to the production database
LUN, which can cause elevated disk response times.
The replica taken is not a completely separate physical copy. An unlikely physical loss of theproduction LUN results in a loss of the backup as well.
4For information on MR3 writes, refer to the CLARiiON RAID 5 Optimizationssection of theEMCCLARiiON Fibre Channel Storage Fundamentals white paper.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 22
7/26/2019 Exchange_best_practice.pdf
23/41
8/8/2005
The space required for the reserved LUN pool (snap cache) is larger than typicalpossibly equivalentto the cumulative size of all files on the source LUNbecause of the random nature of Exchange I/O.
Comparison of Local Replication Options
When comparing the benefits of clone-based to snapshot-based replication of Exchange data, the clone
option has several advantages. It is the recommended solution for Exchange local replication. The
following tables provide a brief comparison of clone/snapshot capabilities and impact, and of the currentlyreleased EMC local replication solutions.
Table 9. Comparison of SnapView Clones and Snapshots for Exchange Local Replication
Feature Clones Snapshots
Accessible to secondary server Yes Yes
Recovery in event of software error Yes Yes
Maximum replicas 8 8
Can be used with MirrorView
No Yes
Recovery in event of hardware error Yes No
Disk space, as a percentage of thesource LUN
100% per clone Varies widely. With a daily snapshotfor local Exchange replication, the
snapshot may approach 100%.
Performance impact on source Only during resync and impactis typically less than COFWactivity
Yes, for the duration of session. Anyactivity on the snap (such as Eseutil or
backup to tape) directly impacts theproduction LUN.
Table 10. Comparison of VSS-Integrated EMC Local Replication Products
Product Supported
Exchange
Versions
Supported OS Supported Array Replicas
RMSE a 5.5
2000
2003
Windows 2000
Windows 2003
CLARiiON Up to eight clones orsnapshots
Replication
Managerb
5.5
2000
2003
Windows 2000
Windows 2003
CLARiiON
Symmetrix
Some third-party arrays(seeEMC Support Matrix)
Up to eight clones or
snapshots
aRMSE also supports replication of SQL and Windows file systems.
bReplication Manager also supports additional operating systems and replication of additional applications.
Online Backup to DiskDisk-based Exchange online backup has become a competitive alternative to tape backup, offering faster
performance and higher reliability. Consider the following when online backup to disk is chosen as theprimary backup method.
Capacity requirements are determined by the size and frequency of the backups and the number ofcopies maintained before archiving to tape.
Online backups perform sequential writes. A configuration that optimizes this type of I/O will performbest.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 23
7/26/2019 Exchange_best_practice.pdf
24/41
8/8/2005
CLARiiON ATA disk drives are effective for backup to disk. They perform well with sequential I/Ooperations especially with FLARERelease 13 or higher and a RAID 3 configuration. They are alsomost competitive with the cost of tape backup.
If keeping more than one backup of an ESG on disk, you can alternate copies across two differentRAID groups for added protection.
Most testing has been done on RAID 5 4+1 (RAID 3 4+1 for ATAs). It has been determined to be the
sweet spot for performance with backup to disk.
Multiple streams within the same LUN will often result in better overall throughput, although the per-stream performance will understandably decline. If each stream is sent to a separate LUN on the
RAID group, the data in a particular stream will be more contiguous on the disk and thus restoresomewhat faster (by 5 MB/s to 10 MB/s).
For the most rapid recovery and the least user impact, individual Exchange databases should be assmall as possible with the smallest possible number of users. If the backup method will be online
backup to disk, you have added incentive to use the maximum number of storage groups and databases
within those ESGs.
If online backup is performed during a period of active production, be sure to take this additionalactivity into account when calculating I/O requirements, also paying attention to overall array
limitations.
For more detail on this topic, refer to the white paperEMC CLARiiON Backup Storage Solutions CXSeries: Backup-to-Disk Performance Guide.
Recovery Storage GroupsWith the addition of the recovery storage group (RSG) feature in Exchange Server 2003, a separate server
for mailbox recovery is no longer required. Within an RSG, an administrator can mount a second copy of amailbox database and use it to recover any data it contains.
When allocating storage for RSGs, plan a single LUN to handle the size of the largest regular Exchange
storage group, plus existing logs for that ESG. Log files can reside in the same LUN as the databases since
users cannot log on to these and mail cannot be delivered to them.
With local replication to clones, its convenient to mount a snapshot of the clones for use with a recovery
storage group. Of course this is useful only if the data you need to recover is recent enough to exist on theclone replicas. You cannot mount the clone directly for use with an RSG because mounting the Exchange
database would make it unusable for the standard VSS recovery. When planning to mount this replica, be
sure to allocate some snap cache for this purpose.
For more detail on this topic, refer to the Microsoft white paper Using Exchange Server 2003 RecoveryStorage Groups.
Planning Storage for Local Message ArchivingMessage archiving on CLARiiON storage has been a growing component of new messaging system
designs. The archiving implementation generally involves adding one or more servers to the environment
to manage the movement of the content of messages past a set age out of the standard Exchange database
structure. The archiving servers also manage the near-line retrieval of these messages when a user calls for
one.
EMC LegatoEmailXtender software provides this archiving capability. The EmailXtender manual
provides a formula to estimate the amount of data to be archived.
Container File Storage (disk space for archived messages) =
Number of users
x Messages per user per day
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 24
7/26/2019 Exchange_best_practice.pdf
25/41
8/8/2005
x Days per work week
x Number of weeks mail is retained
x Average message size (KB)
x .000001 (to convert the result to GB)
For example:3200 x 20 x 5 x 250 x 50 x .000001 = 4000 GB
EmailXtender also estimates an approximately 20 percent overhead for the installation and Message Center
drive, plus .5 GB for queuing. The space required on the installation LUN in this case would be:
4000 x .2 + .5 = 800.5 GB
Because there is usually a very large quantity of data archived, and the access to this data is less time-
critical, this is a very appropriate application for CLARiiON ATA drives.
Storage-System ConsiderationsOnce you have determined the I/O requirements of the new messaging system and settled upon an
appropriate number and type of disk drive, the next step is to consider the throughput and features of thestorage system.
The best throughput and data protection from the storage system results from a design that aims for
balanced use of the array resources, considering both the layout of the data and the timing of schedulableactivities. Consider the following when planning your disk layout:
Avoid configuring Exchange database LUNs on the CLARiiON persistent storage manager (PSM)drives (drives 0-2). However, these drives may be suitable for the lower I/O requirements of a
properly configured set of log LUNs.
Its not required to have the log LUN and database LUN for an ESG managed by opposite CLARiiONstorage processors (SPs). It matters more that the overall I/O demand is balanced across the two SPs.
There is a small performance advantage (three to four percent) to be gained by binding the RAID 1/0primaries and secondaries on different back-end buses. The main advantage to this approach is that the
administrator does not have to be cognizant of which LUN is on which back-end bus. The back-end isbalanced by virtue of the RAID group layout.
Dont forget to include hot spares in the configuration. Typically, you should add one hot spare forevery 30 drives.
Clone LUNs must be assigned to the same SP as their source. (It is allowed to configure clones on theopposite SP to its source, but the clone would be trespassed for synchronizations in this case.) Ensure
that both the current SP owner and preferred owner are the same for both the source and target LUNs.
MirrorView (synchronous or asynchronous) cannot be used with any LUN that is part of a clone group(source or target).
Some activities within the Exchange environment place heavy use of the CLARiiON SP. Theseinclude performing an Eseutil check against a database, and running CLARiiON layered applications
particularly when acting upon several LUNs at once. When configuring an array for Exchange and
scheduling Exchange administrative operations, its important to consider the limits of the SP CPUresources.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 25
7/26/2019 Exchange_best_practice.pdf
26/41
8/8/2005
iSCSI GuidelinesThe CLARiiON CX500i and CX300i models provide direct support for iSCSI connections. Note the
following considerations before choosing an iSCSI model CLARiiON storage system for Exchange:
Compared to Fibre Channel, iSCSI offers a lower cost hurdle for customers moving from direct-attached or internal server-based storage to SAN. Connectivity components are less expensive and IP
network expertise is more common.
There is inherently more processing required with the iSCSI protocol than Fibre Channel and thedelivery of iSCSI packets is less regular. This introduces extra latency into the I/O stream. Exchange
activity can generate high disk I/O, and it has a low tolerance for slow disk response.
The iSCSI models support the ability to boot directly from the storage system, but only with an iSCSIHBA (TCP/IP Offload Engine, or TOE) installed in the server.
The iSCSI models support only up to a two-node Exchange cluster.
The CLARiiON Disk Library is fibre-based and requires the network infrastructure of a fibre SAN.
Local replicationincluding VSS-supported shadow copies for Exchange 2003 via ReplicationManageris supported. These storage system models are internally identical to their equivalent Fibre
Channel models; therefore, internal performance such as clone synchronizations should be the same.
Remote replication using CLARiiON layered applications (SAN Copy or MirrorView) is not supportedto or from the iSCSI models.
Front-end ports on iSCSI models (1 Gb versus 2 Gb for fibre models) have the potential to be abottleneck during high I/O activity. Eseutil, for example, can process an Exchange database at the rate
of 10 GB per minute. All CX models have Fibre Channel back ends.
When configuring an iSCSI model CLARiiON storage system for Exchange, note the following
recommendations:
When working with the Microsoft iSCSI Initiator Service, use the Initiator Control Panel to configurethe LUNs on the CLARiiON storage system as persistent targets. This is necessary to automaticallyreestablish a connection to the storage system upon a restart.
Use Gigabit Ethernet and separate network traffic from storage traffic. Dedicated Gigabit Ethernetoffers the best throughput to the CLARiiON iSCSI models.
Configure redundant controllers and switches for high availability.
While it is common to use a network interface card (NIC) in servers for iSCSI connections to savecost, consider replacing the NIC with an iSCSI HBA. A TOE handles the overhead added by the
TCP/IP processing, which can be particularly important on an active Exchange server.
If you use a NIC, install software in this order:
1. Microsoft iSCSI initiator
2. Navisphere Agent
3. PowerPath
During the installation of Navisphere Agent, make sure to answer yeswhen asked if the system uses
the Microsoft iSCSI Initiator.
If you use a TOE, the recommended order is:
1. HBA software (such as QLogic SANsurfer)
2. Navisphere Agent
3. PowerPath
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 26
7/26/2019 Exchange_best_practice.pdf
27/41
8/8/2005
During the installation of Navisphere Agent, make sure to answer nowhen asked if the system uses the
iSCSI initiator.
The bottom line for performance is to minimize traffic latency in the iSCSI connections betweenservers and storage system, and to test Exchange performance in the environment to ensure that
response time is as required.
For further detail on Microsoft iSCSI guidelines, refer to Knowledge Base article 839686, Support for
iSCSI Technology Components in Exchange Server.For information on Microsoft iSCSI cluster supportrequirements refer to the FAQ at:
http://www.microsoft.com/WindowsServer2003/technologies/storage/iscsi/i
scsicluster.mspx
CLARiiON Storage Systems ComparisonTable 11 lists useful specifications for the five current CLARiiON CX models.
Table 11. CLARiiON CX Series Storage Systems Feature Summary
Feature CX700 CX500 CX300 CX500i CX300iMaximum Disks 240 120 60 120 60
Storage Processors (SP) 2 2 2 2 2
CPUs/SP2x3GHz
2x1.6GHz
1x800MHz
2x1.6GHz
1x800MHz
Front-End Ports/SP
4 @ 2Gb
(fibre)
2 @ 2Gb
(fibre)
2 @ 2Gb
(fibre)
2 @ 1Gb
(iSCSI)
2 @ 1Gb
(iSCSI)
Back-End Fibre Channel Ports/SP4 @ 2
Gb
2 @ 2
Gb
1 @ 2
Gb
2 @ 2
Gb
1 @ 2
Gb
I/O Buses 4 2 1 2 1
Array Cache 8 GB 4 GB 2 GB 4 GB 2 GB
Highly Available Hosts 256 128 64 128 c 64 c
Maximum LUNs 2048 1024 512 1024 512
MirrorView Images (Total Primary + Secondary) a 100 50 n/a n/a n/a
Snapshot LUNs 300 150 100 150 100
Clone Groups 50 25 25 25 25
Clone Objects * 100 50 50 50 50
SAN Copy Concurrent Sessions (and MaxDestinations)
16(100)
8 (50) 4 (50) n/a n/a
Max Incremental SAN Copy Source LUNs 100 50 25 n/a n/a
Exchange Users with no replicationb
20,000 10,000 5,000 6,000 3,000
Exchange Users with local replication (RMSE) 10,000 5,000 2,500 TBD TBDa
Clone objects and MV images share the max.bAs described in this paper, the number of Exchange users a storage system can handle will vary greatly.
The number in this table is provided primarily as a starting point and for comparison between the models. It
refers to a storage system dedicated to Exchange users (.4 IOPS/user). While it is very unusual for a
CLARiiON storage processor to fail, it is advisable to plan for and test an SP failure scenario to understandhow the environment would run in degraded mode, and then to size accordingly.
cFor iSCSI, each NIC/TOE connected to the storage system counts as a connection.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 27
http://www.microsoft.com/WindowsServer2003/technologies/storage/iscsi/iscsicluster.mspxhttp://www.microsoft.com/WindowsServer2003/technologies/storage/iscsi/iscsicluster.mspxhttp://www.microsoft.com/WindowsServer2003/technologies/storage/iscsi/iscsicluster.mspxhttp://www.microsoft.com/WindowsServer2003/technologies/storage/iscsi/iscsicluster.mspx7/26/2019 Exchange_best_practice.pdf
28/41
8/8/2005
Be aware of these specifications when determining the appropriate storage system(s) for the new Exchange
storage design. For example, you can create up to 50 clone objects on a CLARiiON CX500. Each LUN in
a clone groupincluding the sourcecounts as one clone object. If the local replication plan calls for two
clone copies of each Exchange production LUN, the CX500 can support up to 16 Exchange LUNs (each
production LUN and its two clones represent three clone objects, for a total of 48 clone objects). If theExchange design follows the standard recommendation of two LUNs per storage group, the CX500 can
handle eight ESGs. The number of servers does not matter here. The CX500 could support four Exchange
servers with two ESGs each, two servers with four ESGs, or any other combination totaling eight ESGs. Ifonly one clone is used per production LUN, the CX500 could support up to 12 ESGs, assuming that no
other resource limits are reached.
Putting It All TogetherThis section presents some final suggestions for designing a successful Exchange storage design.
Consider Site-Specific ConstraintsThe resulting storage design must obviously take into account any requirements or restrictions in theenvironment where the messaging system will be implemented.
For example, the organization may have an existing CX500 available that they want to use before any
additional array is purchased. They may not be able to dedicate an entire array to Exchange use, or mayhave a certain type of drive already in place for the Exchange data. There are likely to be many other
decisions made already that will affect the storage design, such as number and location of Exchange
servers, number of ESGs per server, network capacity, etc.
Regardless of the constraints, the core requirement of providing enough drives to meet peak I/O demand
remains, as does the strong recommendation to keep log and database LUNs for the same ESG on separate
spindles.
Configure the Cleanest Looking Layout Diagram
Using a building block style, draw up a clean-looking storage layout diagram. This will ease understanding
of the design, help identify possible weaknesses, and aid in the storage administration of the
implementation.
Plan Throughout for Operational Resiliency
Increasing availability through the elimination of single failure points is easier in SAN environments.
Consider all possible points of failure and distribute users from the same functional area to minimize theimpact of planned or unplanned downtime.
Separate ESGs width first (more ESGs), depth second (databases per ESG)
Separate Exchange servers typically 4,000 or fewer users per mailbox server
Separate storage systems spread out users (see following example)
Geographically dispersed locations for larger organizations
Upper management and other priority users and key group mailboxes may have stricter SLAs than the rest
of the organization. These users should also be distributed, in addition to the extra performance or HAconfigurations they are provided.
ExampleFor an organization of 7,000 typical Exchange 2003 users, a single CX500 could handle their basic mailbox
requirements. A CX700 could handle their mailbox requirements, including local replication to clones. Asan alternative to the single CX700, consider designing the configuration with two CX500s handling 3,500
users each. Distributing these 3,500 users over all four ESGs places fewer than 900 users per Exchange
storage group and under 200 per database, minimizing user impact and maintaining conservative databasesizes.
EMC CLARiiON Storage Solutions Microsoft Exchange 2003 Best PracticesStorage Configuration Guidelines 28
7/26/2019 Exchange_best_practice.pdf
29/41
8/8/2005
Additionally, each of the CX500s could be configured to handle disaster recovery for the peer storage
system in the unlikely event one storage system experiences a significant problem. By providing additional
storage space and temporarily postponing