Top Banner
RHIC Computing Facility Michael Ernst Michael Ernst DOE/Nuclear Physics Review of RHIC DOE/Nuclear Physics Review of RHIC Science and Technology Science and Technology 18 18 - - 20 July 2007 20 July 2007
50

Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

Aug 01, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

RHIC Computing Facility

Michael ErnstMichael Ernst

DOE/Nuclear Physics Review of RHIC DOE/Nuclear Physics Review of RHIC Science and TechnologyScience and Technology

18 18 -- 20 July 200720 July 2007

Page 2: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 2

RHIC Computing Facility (RCF)

Organizationally established in 1997Organizationally established in 1997

Staffed as a Group in Physics DepartmentStaffed as a Group in Physics Department

Equipment physically located at Brookhaven Computing Facility (BEquipment physically located at Brookhaven Computing Facility (BCF)CF)BCF operated by ITD

Currently coCurrently co--located and colocated and co--operated with the ATLAS Computing operated with the ATLAS Computing Facility (ACF), the U.S. ATLAS TierFacility (ACF), the U.S. ATLAS Tier--1 Regional Center 1 Regional Center

ACF ramping up quickly, currentlyo ACF capacities are ~ 65% for processing, 121% for disk capacity

o ACF staff level ~ 75% of RCF

Page 3: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 3

RCF Mission and Scale

MissionMissionOnline Recording of Raw Data

Production reconstruction of Raw Data

Primary Facility for Data Selection and Analysis

Long time Archiving and Serving of all Data

ScaleScaleAuthorized staff of 20 FTE’s

Historically ~$2M/year equipment replacement funding (25% annualreplacement) – Last year limited to $1.3M, current year limited to $1.7M

o Addressing obsolescence

Growth beyond originally planned scale will require an increase in the funding

Page 4: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 4

Experiment / RCF Interaction

Weekly Liaison MeetingWeekly Liaison MeetingAddressing operations issuesReview recent performance and problemsPlan for scheduled interventions

Experiments / RCF Annual Series of Meetings to develop Capital SExperiments / RCF Annual Series of Meetings to develop Capital Spending Planpending PlanEstimate scale of need for current/coming runDetails of distribution of equipment to be procuredMost recent in early Spring for FY-07 funds

Periodic Topical Meetings, examplesPeriodic Topical Meetings, examples~Annual Linux Farm OS upgrade planningReplacement of Central Disk Storage

Other User InteractionsOther User InteractionsWeb siteTicket System (Request Tracker (RT – Open Source)

o Fully replaced in-house developed Trouble Ticket System (CTS) o ~3000 Tickets for RHIC & ATLAS Services (last 12 months)

Page 5: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 5

Computing Requirements EstimateA Comprehensive Long Range Estimate done by PHENIX, RCF and A Comprehensive Long Range Estimate done by PHENIX, RCF and STAR in Fall / Winter 2005STAR in Fall / Winter 2005

Conclusions published as part of “Mid-Term Strategic Plan: 2006-2011 For the Relativistic Heavy Ion Collider”, February 14, 2006Needs to be revisited/updated

o Lack of disk space has an obvious impact on analysis performance

Input is Raw Data Volume for Each Species & Experiment by YearInput is Raw Data Volume for Each Species & Experiment by Year

Model for Requirements ProjectionModel for Requirements ProjectionAssume Facility resource needs scale with Raw Data volumeWith adjustable parameters reflecting expected relative …

o Richness of data set (density of interesting events)o Maturity of processing softwareo Number of reconstruction passes

… for each experiment, species, and year

Page 6: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 6

Computing Cost Estimate

Requirements Model Output used as input to Cost EstimateRequirements Model Output used as input to Cost Estimate

Costing Model is based onCosting Model is based onRecent procurements

Historic Trends (Moore’s Law and similar technology based trends)

Use of inexpensive disk for bulk of storageo Linux processor farm distributed disko Raid 6/ZFS based Storage Farms

Assume use of obvious technology evolution (multi-core processors), etc.)

For running scenarios considered, capacity growth associated with replacement of obsolete equipment meets increased capacity requirements in 2007 but increase of equipment funding is required in 2008 and beyond

o Required capacities by year and a funding profile allowing them to be achieved are shown on following slide

Page 7: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 7

Page 8: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 8

Principal RCF Services

General Collaboration and User SupportGeneral Collaboration and User Support

Processing Services (Linux Farm)Processing Services (Linux Farm)Programmatic Production processingIndividual and Group Analysis

Online Storage (Disk)Online Storage (Disk)Data storage for work area (Read / Write)Data serving for Analysis (> 90% Read)

Mass Storage (Robotic Tape System)Mass Storage (Robotic Tape System)Raw Data recording and archivingDerived Data Archiving

Grid & Network ServicesGrid & Network Services

Page 9: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 9

RCF StaffCurrent authorized staff level: 20 Current authorized staff level: 20

FTEFTE’’s s

Excellent synergy in the context Excellent synergy in the context of a coof a co--located ATLAS Tierlocated ATLAS Tier--1 1 Center in terms of operationsCenter in terms of operations

Very high level of commonalityA dramatic divergence in technical directions could change this, but this seems very unlikely

It does not allow for aggressive It does not allow for aggressive involvement in new technologiesinvolvement in new technologies

Effort spent primarily on Integration and Operation

Current Current FTEFTE’’ss

Target Target FTEFTE’’ss

Linux FarmsLinux Farms 3.53.5 3.53.5

Mass StorageMass Storage 4.24.2 4.24.2

DiskDisk 2.62.6 2.62.6

User SupportUser Support 2.92.9 2.92.9

Fabric InfrastructureFabric Infrastructure 2.12.1 2.62.6

Wide Area ServicesWide Area Services 1.81.8 1.81.8

AdminAdmin 2.52.5 2.52.5

TotalTotal 19.519.5 20.020.0

Page 10: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 10

Compute ServersThree Generations of Linux CPU rack mount systemsThree Generations of Linux CPU rack mount systems

Dual CPU (single core) systems (2600 SI2k per box, bought in 2002)

Dual CPU (dual core) systems (4600 SI2k – 10,000 SI2k per box) x 8

Dual CPU (quad core) systems (20,000 SI2k per box)

Currently 1,400 compute servers with 2,800 CPU’s (4200 cores)

Lack of funding does not allow a timely “refresh” of equipmentRequires more space, power and cooling than anticipated

~100 additional Dual CPU / Quad Core machines (8 cores / box) with 2 MSI2k – Delivery expected by end August

o Multi-core CPU technology also addresses power/cooling barrier by finessing non- linearity of power consumption with clock speed

Expect to address future requirements by continuing to follow Moore’s Law price/performance in commodity market (multi-core, 64 bit advances)

Page 11: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 11

Expected Computing Capacity Evolution

0

10000

20000

30000

40000

50000

60000

70000

FY

' 06

FY

' 07

FY

' 08

FY

' 09

FY

' 10

FY

' 11

FY

' 12

KSI

2K RHIC

USATLAS

Page 12: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 12

Resource Utilization Issues in 2006

Page 13: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 13

Resource Sharing among Experiments

Goal was to make idle cycles available in processor farms Goal was to make idle cycles available in processor farms to other user communities without impact to to other user communities without impact to ““ownerowner””

Mechanism is to evict Mechanism is to evict ““guestguest”” jobs when jobs when ““ownerowner”” needs needs cyclescycles

Consider extended rights for guestso Allow guest job to complete by grace period (implemented but currently

not used)

Page 14: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 14

Resource Utilization in 2007

Average load of 77% for the past 12 months.Average load of 77% for the past 12 months.

Average load of ~62% for 07/05 to 07/06 (2006 review).Average load of ~62% for 07/05 to 07/06 (2006 review).

Excluding interactive nodes, maximum possible load is ~94%.Excluding interactive nodes, maximum possible load is ~94%.

Page 15: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 15

Condor Occupancy (PHENIX)

LeftLeft--hand plot is for late Junehand plot is for late June’’06 to 12/31/06.06 to 12/31/06.

RightRight--hand plot is for 01/01/07 to 07/03/07.hand plot is for 01/01/07 to 07/03/07.

Occupancy rose from 81% to 89% between the two periods.Occupancy rose from 81% to 89% between the two periods.

Upper Limit

Page 16: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 16

Condor/LSF Occupancy (STAR)

LeftLeft--hand plot is for late Junehand plot is for late June’’06 to 12/31/06.06 to 12/31/06.

RightRight--hand plot is for 01/01/07 to 07/03/07.hand plot is for 01/01/07 to 07/03/07.

Occupancy rose from 105% to 126% between the two periods.Occupancy rose from 105% to 126% between the two periods.

Page 17: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 17

Condor Occupancy (RACF)

LeftLeft--hand plot is for late Junehand plot is for late June’’06 to 12/31/06.06 to 12/31/06.

RightRight--hand plot is for 01/01/07 to 07/03/07.hand plot is for 01/01/07 to 07/03/07.

Occupancy rose from 83% to 94% between the two periods.Occupancy rose from 83% to 94% between the two periods.

Created general queue in 2006 to increase occupancy.Created general queue in 2006 to increase occupancy.

4,200 Job Slots

Page 18: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 18

Online (Disk) Storage

Historic RCF model was Sun / NFS served RAID 5 SAN Historic RCF model was Sun / NFS served RAID 5 SAN connected central disk for all storage areasconnected central disk for all storage areas

Growth in demand drove disk costs to match and exceed Growth in demand drove disk costs to match and exceed CPU costsCPU costs

Current strategy: Differentiate disk technology by functionCurrent strategy: Differentiate disk technology by functionCentral Disk

o Limited amount of NFS “full function” (fully Posix compliant) disk for Read/Write

o Working on a backup solution (selective)

“Read only” Disko Majority on less expensive distributed disk (on Farm nodes) and

integrated in storage farms for “mostly Read” of data on secure medium (tape)

Page 19: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 19

“Full Function” Disk Service

Read/Write (Read/Write (PosixPosix compliant), reliable, high performance and high compliant), reliable, high performance and high availability availability –– NFS served RAID systemsNFS served RAID systems

Historicallyo ~150 TB of Sun served RAID 5 disko ~70 TB of Panasas (appliance) served RAID 5 disk

Acquisition in 2006o ~100 TB of Nexsan & Aberdeen Linux served RAID 5/6 disk

Movement to lower Tier of RAID disk vendors last year o Product from expensive vendor failed to fulfill expectationso Inexpensive RAID systems unable to sustain the load

Too many concurrent processes

Very bad situation in early 2007o Many service disruptions due to old and unreliable equipmento Services distributed on too many different products

Negative impact on user efficiency (losing jobs, eventually losing data) Two FTE’s constantly occupied to keep the service operational

Page 20: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 20

Central Disk Failures over Time

GCE Failures 1/04 - 12/07

0

5

10

15

20

25

30

Failures 4 6 2 8 7 8 8 8 3 15 7 18 12 11 7 8 3 11 7 5 5 9 13 5 14 14 24 9 16 14 12 11 10 8 5 18 13 14 10 9 9 7 1 0 0 0 0 0

Jan-04

Feb-04

Mar-04

Apr-04

May-04

Jun-04

Jul-04

Aug-04

Sep-04

Oct-

04

Nov-04

Dec-04

Jan-05

Feb-05

Mar-05

Apr-05

May-05

Jun-05

Jul-05

Aug-05

Sep-05

Oct-

05

Nov-05

Dec-05

Jan-06

Feb-06

Mar-06

Apr-06

May-06

Jun-06

Jul-06

Aug-06

Sep-06

Oct-

06

Nov-06

Dec-06

Jan-07

Feb-07

Mar-07

Apr-07

May-07

Jun-07

Jul-07

Aug-07

Sep-07

Oct-

07

Nov-07

Dec-07

Page 21: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 21

Consolidation of Central DiskHave initiated a Storage Evaluation ProjectHave initiated a Storage Evaluation Project

Multiple vendors of appliances and disk storage backends were asked to bring in equipment for evaluationRan relevant low level and (physics) application driven tests

o Massive concurrency (100’s of clients)o Read/Write performance oriented testso Resiliency and fail-over tests

Tests took longer than expectedo Though recommended by the vendor of NAS Head the backend storage

performance of SATA based disk backends was poorUnable to fix, despite massive amount of effort spent by vendorsFC disk based backend the only solution satisfying RACF’s performance and resilience requirements

Purchasing a 200 TB system from BlueArc / Hitachio To replace equipment older than 3 years (Panasas, MTI, Zzyzsx)o Requires ~50% of FY’07 funds (not much left for processing)o High performance, high-availability storage at very competitive Cost (~$3.5/GB)

Page 22: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 22

“Mostly Read” Disk

Disk deployed on Linux processor farm nodesDisk deployed on Linux processor farm nodes~3 x less expensive than full function diskNo RAID, JBOD (Just a Bunch Of Disks)

Requires additional storage management softwareRequires additional storage management software

Two such storage management systems currently in use at Two such storage management systems currently in use at RCFRCF

dCache – DESY/Fermilab developed Grid-aware S/W packageo Scalable, robust, unified pool of independent storage components with

integral Mass Storage backend, posix-like data access, ROOT supporto ATLAS is major BNL user with 850 TB => 1,500 TB by end July 2007

Xrootd – SLAC, CERN, BNL + other community developerso STAR is major BNL user with ~300 TB managed capacity

Heavily used for more than 2 years

Page 23: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 23

dCache Usage by PHENIXUsage is dominated by data transfer on LANUsage is dominated by data transfer on LAN

Aggregate Throughput up to 1.5 GB/s

Repository and Archiving mechanism for data production streamRepository and Archiving mechanism for data production stream

Integrated into Integrated into ““Analysis TrainAnalysis Train””Aggregates user analysis jobs to run efficiently on common data subsets Access restricted by policy to train “operators”Increasing WAN transfer (to IN2P3)

Page 24: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 24

PHENIX Transfer Statistics

Page 25: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 25

More Information on Storage ManagementSTAR and STAR and XrootdXrootd

Xrootd / Scalla rationales areo Hope for better data access providing improved performanceo Growth of dataset size + budget constraints leading to difficult situation

Use data compression (STAR tried, implemented)Use even more inexpensive hardwareAccess the data in smarter ways

o STAR has the largest Xrootd deployment to date (still growing)

Issueso Xrootd and dCache are still in R&D

High backend MSS stability is required to utilize highly dynamic disk population modelOptimization non-trivial – STAR spent a fair amount of time to study data retrieval strategies assisted this year by RCF team

o To make it work for RHIC effort is required from RHIC projectSTAR allocates out-sourced FTE to work on Scalla

– Not a long-term solution– Dedicated effort would be much more efficient

Analysis relies on leading edge development and integration Stability questionableModel in RHIC II era questionable w/o (more) integration effort now

Page 26: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 26

Example of Dataset Usage and Access Pattern in STAR

Rich variety of Physics Data to be concurrently analyzed leading to “threshing” of disk inventory

Page 27: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 27

Expected Disk Storage Capacity Evolution

0

5000

10000

15000

20000

25000

30000

35000

40000F

Y' 0

6

FY

' 07

FY

' 08

FY

' 09

FY

' 10

FY

' 11

FY

' 12

TB

of d

isk

stor

age

RHIC

USATLAS

Disk capacity projection for RHIC as described in mid-term plan has foreseen far less space thanATLAS (despite the fact that U.S. ATLAS plans to keep all reconstructed events on Disk)

Page 28: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 28

Mass Storage SystemHPSS (High Performance Storage System) Hierarchical HPSS (High Performance Storage System) Hierarchical Storage manager from IBMStorage manager from IBM

Moving to version 6.2 in August

Sun/Sun/StorageTekStorageTek Robotic Tape LibrariesRobotic Tape LibrariesFour PowderHorn SilosOne SL8500 linear library (+1 SL8500 for ATLAS)7 PB total capacity

7 PB5.5 PB

Page 29: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 29

A lot more Raw Data this Year …

Page 30: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 30

Latency (all Experiments)

Average Wait Time (minutes)

0

20

40

60

80

100

120

7/2/2006 9/2/2006 11/2/2006 1/2/2007 3/2/2007 5/2/2007

Week

Wai

t Tim

e (M

in)

Avg Wait (Min)

… an important parameter for planning purposes (number of Tape Drives in Robotic Library)

Page 31: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 31

Tape Handling Performance

Tape Mounts Per Month

020000400006000080000

100000120000140000160000180000200000

7/2006 8/2006 9/2006 10/2006 11/2006 12/2006 1/2007 2/2007 3/2007 4/2007 5/2007 6/007

Month

Tape

Mou

nts

M ounts

[Read & Write, includes ATLAS]

Page 32: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 32

Grid and Network Services

Computing models of RHIC Experiments predate the GridComputing models of RHIC Experiments predate the GridUnlike ATLAS, they were not originally based on Grid Technology

Desire to utilize substantial distributed resources is driving evolution towards Grid Computing

o Started with simulation, moving towards analysisLBNL, Prague (working with ITD and ESnet on link), etc. for STARRiken, Vanderbilt, IN2P3, etc. for PHENIX

o Same staff engaged in U.S. ATLAS Grid effort also supports RHIC wide area distributed computing with

Support for Grid tools and services as well as network expertise– GridFTP, SRM, …– High volume network transfer optimization

Support for involvement (of STAR) in Open Science Grid (OSG)– OSG software deployment and integration of resources into OSG– OSG administration

Page 33: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 33

Physical InfrastructureMajor physical infrastructure improvements were made over the coMajor physical infrastructure improvements were made over the course urse of the past 12 monthsof the past 12 months

1.25 MW of local UPS / PDU systems added to support new procurements New chilled water feedLocal rack top cooling for new procurementsCovered by GPP funds

Have reached limit of available floor spaceHave reached limit of available floor spaceWithout additional space RCF will not be able to accommodate the next robot (due in early spring 2008)Reallocation of space to RCF/ACF allows 2007/8 expansion

o Additional power & cooling is needed each yearNeed expansion of space in 2009 and beyond

o Working with ITD, BNL Plant Engineering and BNL Management on a planVery tight scheduleProgress is not as good as we had hoped for

– Technical and organizational problemsThis is our top concern at the moment

Page 34: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 34

Severe Cooling Problems since April ‘07

A/C units for BlueGenewere turned on

A lot of sediment was stirred up due to increased cooling flow, system never recovered so far

Page 35: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 35

Infrastructure Planning at RACF

Currently available space filledCurrently available space filled

Soon running out of PowerSoon running out of Power

Now

2009

0

1500

3000

4500

6000

2007 2008 2009 2010 2011 2012

Fiscal Year

Spac

e (s

q. ft

.)

space neededspace available

0

1000

2000

3000

2007 2008 2009 2010 2011 2012

Fiscal Year

Pow

er (K

W)

power neededpower available

+ 4000 sqft.

(planned)

Page 36: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 36

Cyber Security

Facility is a Firewall protected enclave within the BNL firewallFacility is a Firewall protected enclave within the BNL firewall protected protected sitesite

Most Services provided by the Facility have a single signMost Services provided by the Facility have a single sign--on Kerberos on Kerberos based authentication infrastructurebased authentication infrastructure

Major effortsMajor effortsContributing to BNL Cyber Security Program Plan

Deploying – Facility-wide – Ordo (BNL developed host based configuration tracking/auditing tool for Unix-like system)

Concern remains of conflicts between User (Grid) requirements, Concern remains of conflicts between User (Grid) requirements, regulatory requirements, and a cyber security policy/architecturregulatory requirements, and a cyber security policy/architecture which e which does not disrupt effective facility usedoes not disrupt effective facility use

Page 37: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 37

ConclusionsPlans to evolve and expand facility services to meet expected nePlans to evolve and expand facility services to meet expected needseds

Are based on successful adjustments of technical directionsRequires agreed and planned for increases in 2008 and beyond

Continued Slippage of Funding for Infrastructure / Facility creaContinued Slippage of Funding for Infrastructure / Facility creating difficult ting difficult situation for the Experiments and the RHIC Computing Facilitysituation for the Experiments and the RHIC Computing Facility

1/3 replacement per year impossibleStretching equipment lifetime with bulk replacement potentially disruptiveCore Infrastructure (HPSS, High-end Disk) improvements need to be delayedFurther increasing burden on staff

Physical infrastructure expansions and improvements are the top Physical infrastructure expansions and improvements are the top concernconcernFacility needs new space with appropriate characteristics and services for 2008 and beyond

Grid technology is likely to change future RHIC computingGrid technology is likely to change future RHIC computingAre building on ATLAS experience

Cyber Security is a major concern Cyber Security is a major concern Security versus Usability

Page 38: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 38

Backup Slides

Page 39: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 39

Computing Resource Utilization BRAHMS & PHOBOS

Page 40: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 40

Resource Utilization PHENIX & STAR

Page 41: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 41

Condor Usage (2006)

Creation of general queue allows opportunistic usage of idle CPUCreation of general queue allows opportunistic usage of idle CPU’’s by user jobs s by user jobs not normally affiliated with CPU ownershipnot normally affiliated with CPU ownership

General queue became default queue in late 2006. Users can overrGeneral queue became default queue in late 2006. Users can override by ide by specifying other queuesspecifying other queues

General queue jobs were only 1.4% of all Condor jobs during thisGeneral queue jobs were only 1.4% of all Condor jobs during this periodperiod

Page 42: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 42

Condor Usage (2007)

Condor usage grew by a factor of 3 (in terms of number of jobs) Condor usage grew by a factor of 3 (in terms of number of jobs) and by a factor of and by a factor of 4 (in terms of CPU time) over the past year. 4 (in terms of CPU time) over the past year.

PHENIX executed over 40% of their jobs in the general queue.PHENIX executed over 40% of their jobs in the general queue.

General queue efficiency is ~ 87% (i.e., only 13% ineffective usGeneral queue efficiency is ~ 87% (i.e., only 13% ineffective use).e).

General queue jobs amounted to 21% of all Condor jobs during thiGeneral queue jobs amounted to 21% of all Condor jobs during this period.s period.

Page 43: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 43

PHENIX dCache Deployment (v1.7)

415 Read/Write Pools (shared), 36 external Write Pools or 415 Read/Write Pools (shared), 36 external Write Pools or dedicated hostsdedicated hosts

212 TB Storage, >750k files on disk212 TB Storage, >750k files on diskAdding 140 TB (usable) by end of July

3 3 GridFTPGridFTP/SRM + 1 /SRM + 1 dCapdCap door nodes, 1 admin nodedoor nodes, 1 admin node

SL3+EXT3 on Read Pools, SL3/XFS+SL4/EXT3 on SL3+EXT3 on Read Pools, SL3/XFS+SL4/EXT3 on external Write Poolsexternal Write Pools

HPSS backend interface via HIS/Carousel/PFTPHPSS backend interface via HIS/Carousel/PFTP

Page 44: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 44

STAR Mass Storage System File retrieval Performance

HPSS backend (DataCarousel) performance monitored versus file size in Xrootd / Scalla context

Page 45: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 45

Mass Storage System – High Performance Storage System (HPSS)

Tape DrivesTape Drives37 StorageTek 9940B (30 MB/s)30 LTO Gen3 (80 MB/s)

30 TB of HPSS Disk Cache30 TB of HPSS Disk Cache

InIn--house developed tape access optimization softwarehouse developed tape access optimization softwareIncreases access efficiency by sorting requests according to data placement

Page 46: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 46

HPSS Read Performance (GB/day)

Average GB Read Per Day

0

1000

2000

3000

4000

5000

6000

7000

8000

7/2/2006 9/2/2006 11/2/2006 1/2/2007 3/2/2007 5/2/2007

Week

Avg

. GB

Rea

d/D

ay

Avg GB/Day

Page 47: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 47

HPSS Read performance (# of Files)

Average Files Read Per Day

0

2000

4000

6000

8000

10000

12000

14000

16000

2006/08/06 2006/09/10 2006/12/02 2007/02/24 2007/05/20

Week

Avg

File

s R

ead/

Day Phenix 9940

Phenix LTOStar 9940Star LTOPhenix 9940Star 9940Phobos 9940Brahms 9940Archive 9840

Page 48: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 48

HPSS File Retrieval Latency (per Experiment)

Average File Retrieval Latency

0

2000

4000

6000

8000

10000

12000

14000

16000

2006/08/06 2006/09/17 2006/12/16 2007/03/18

Week

Sec

onds

Phenix 9940Phenix LTOStar 9940Star LTOPhenix 9940Star 9940Phobos 9940Brahms 9940Archive 9840

Page 49: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 49

Wide Area Network

Jan Jan ’’06 WAN last upgrade on BNL connectivity to 20 06 WAN last upgrade on BNL connectivity to 20 GbpsGbps

Funded in equal part by Funded in equal part by ESnetESnet, DOE NP, DOE HEP and BNL, DOE NP, DOE HEP and BNL

Connection still lacks desired redundancy and diversityConnection still lacks desired redundancy and diversityWill require significant additional funding not yet identified

Page 50: Michael Ernst DOE/Nuclear Physics Review of RHIC Science and … · 2007-07-16 · DOE/Nuclear Physics Review of RHIC Science and Technology 18 - ... Assume Facility resource needs

1818--20 July 200720 July 2007I M. Ernst DOE/Nuclear Physics Review of RHIC S&TI M. Ernst DOE/Nuclear Physics Review of RHIC S&T 50

Wide Area Network Architecture