RIT Colloqui um (May 23, 2007) Paul Avery 1 Paul Avery University of Florida [email protected]Physics Colloquium RIT (Rochester, NY) May 23, 2007 Open Science Grid Linking Universities and Laboratories In National Cyberinfrastructure www.opensciencegrid.org
49
Embed
RIT Colloquium (May 23, 2007)Paul Avery 1 Paul Avery University of Florida [email protected] Physics Colloquium RIT (Rochester, NY) May 23, 2007 Open.
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Powerful cyberinfrastructure neededComputation Massive, distributed CPUData storage & access Large-scale, distributed storageData movement International optical networksData sharing Global collaborations (100s – 1000s)Software Managing all of the above
How to collect, manage,access and interpret thisquantity of data?
RIT Colloquium (May 23, 2007)
Paul Avery 4
Open Science Grid: July 20, 2005Consortium of many organizations (multiple
Comp. scientists, IT specialists, physicists, biologists, etc.
WhatShared computing and storage resourcesHigh-speed production and research networksMeeting place for research groups, software experts, IT
providers
VisionMaintain and operate a premier distributed computing facilityProvide education and training opportunities in its useExpand reach & capacity to meet needs of stakeholdersDynamically integrate new resources and applications
Members and partnersMembers: HPC facilities, campus, laboratory & regional
gridsPartners: Interoperation with TeraGrid, EGEE, NorduGrid,
etc.
RIT Colloquium (May 23, 2007)
Paul Avery 7
Crucial Ingredients in Building OSG
Science “Push”: ATLAS, CMS, LIGO, SDSS1999: Foresaw overwhelming need for distributed
SDSS Digital astronomy Sloan Digital Sky Survey (Astro)
STAR Nuclear physics Nuclear physics experiment at Brookhaven
UFGrid Campus grid Campus grid at U of Florida
RIT Colloquium (May 23, 2007)
Paul Avery 12
Partners: Federating with OSG Campus and regional
Grid Laboratory of Wisconsin (GLOW) Grid Operations Center at Indiana University (GOC) Grid Research and Education Group at Iowa (GROW) Northwest Indiana Computational Grid (NWICG) New York State Grid (NYSGrid) (in progress) Texas Internet Grid for Research and Education (TIGRE) nanoHUB (Purdue) LONI (Louisiana)
National Data Intensive Science University Network (DISUN) TeraGrid
International Worldwide LHC Computing Grid Collaboration (WLCG) Enabling Grids for E-SciencE (EGEE) TWGrid (from Academica Sinica Grid Computing) Nordic Data Grid Facility (NorduGrid) Australian Partnerships for Advanced Computing (APAC)
RIT Colloquium (May 23, 2007)
Paul Avery 13
Search for Origin of Mass New fundamental forces Supersymmetry Other new particles 2007 – ?
TOTEM
LHCb
ALICE
27 km Tunnel in Switzerland & France
CMS
ATLAS
Defining the Scale of OSG:Experiments at Large Hadron Collider
LHC @ CERN
RIT Colloquium (May 23, 2007)
Paul Avery 14
CMS: “Compact” Muon Solenoid
Inconsequential humans
RIT Colloquium (May 23, 2007)
Paul Avery 15
All charged tracks with pt > 2 GeV
Reconstructed tracks with pt > 25 GeV
(+30 minimum bias events)
109 collisions/sec, selectivity: 1 in 1013
Collision Complexity: CPU + Storage
RIT Colloquium (May 23, 2007)
Paul Avery 16
CMSATLAS
LHCb
Storage Raw recording rate 0.2 – 1.5 GB/s Large Monte Carlo data samples 100 PB by ~2013 1000 PB later in decade?
Processing PetaOps (> 300,000 3 GHz PCs)
Users 100s of institutes 1000s of researchers
LHC Data and CPU Requirements
RIT Colloquium (May 23, 2007)
Paul Avery 17
CMS Experiment
OSG and LHC Global Grid
Online System
CERN Computer Center
FermiLabKorea RussiaUK
Maryland
200 - 1500 MB/s
>10 Gb/s
10-40 Gb/s
2.5-10 Gb/s
Tier 0
Tier 1
Tier 3
Tier 2
Physics caches
PCs
Iowa
UCSDCaltechU Florida
5000 physicists, 60 countries
10s of Petabytes/yr by 2009 CERN / Outside = 10-20%
FIU
Tier 4
OSG
RIT Colloquium (May 23, 2007)
Paul Avery 18
ATLAS CMS
LHC Global Collaborations
2000 – 3000 physicists per experiment USA is 20–31% of total
Integrated Database Includes: Parsed Sequence Data and
Annotation Data from Public web sources.
Results of different tools used for Analysis: Blast, Blocks, TMHMM, …
GADU using GridApplications executed on Grid as
workflows and results are stored in integrated Database.
GADU Performs:Acquisition: to acquire Genome
Data from a variety of publicly available databases and store temporarily on the file system.
Analysis: to run different publicly available tools and in-house tools on the Grid using Acquired data & data from Integrated database.
Storage: Store the parsed data acquired from public databases and parsed results of the tools and workflows used during analysis.
Bidirectional Data Flow
Public DatabasesGenomic databases available on the web.Eg: NCBI, PIR, KEGG, EMP, InterPro, etc.
Applications (Web Interfaces) Based on the Integrated Database
PUMA2Evolutionary Analysis of
Metabolism
ChiselProtein Function Analysis
Tool.
TARGETTargets for Structural analysis of proteins.
PATHOSPathogenic DB for
Bio-defense research
PhyloblocksEvolutionary analysis of
protein families
TeraGrid OSG DOE SG
GNARE – Genome Analysis Research Environment
Services to Other Groups
•SEED(Data Acquisition)
•Shewanella Consortium
(Genome Analysis)Others..
Bioinformatics: GADU / GNARE
RIT Colloquium (May 23, 2007)
Paul Avery 22
Bioinformatics (cont)
Shewanella oneidensisgenome
RIT Colloquium (May 23, 2007)
Paul Avery 23
Nanoscience Simulations
collaboration
nanoHUB.org
courses, tutorialsonline simulation
seminars
learning modules
Real users and real usage >10,100 users
1881 sim. users>53,000 simulations
RIT Colloquium (May 23, 2007)
Paul Avery 24
OSG Engagement Effort Purpose: Bring non-physics applications to OSG
Led by RENCI (UNC + NC State + Duke)
Specific targeted opportunitiesDevelop relationshipDirect assistance with technical details of connecting to
OSG
Feedback and new requirements for OSG infrastructure
(To facilitate inclusion of new communities)More & better documentationMore automation
RIT Colloquium (May 23, 2007)
Paul Avery 25
OSG and the Virtual Data Toolkit VDT: a collection of software
Grid software (Condor, Globus, VOMS, dCache, GUMS, Gratia, …)
Virtual Data SystemUtilities
VDT: the basis for the OSG software stackGoal is easy installation with automatic configurationNow widely used in other projectsHas a growing support infrastructure
RIT Colloquium (May 23, 2007)
Paul Avery 26
Why Have the VDT? Everyone could download the software from the
providers But the VDT:
Figures out dependencies between software Works with providers for bug fixesAutomatic configures & packages softwareTests everything on 15 platforms (and growing)
Debian 3.1 Fedora Core 3 Fedora Core 4 (x86, x86-64) Fedora Core 4 (x86-64) RedHat Enterprise Linux 3 AS (x86, x86-64, ia64) RedHat Enterprise Linux 4 AS (x64, x86-64) ROCKS Linux 3.3 Scientific Linux Fermi 3 Scientific Linux Fermi 4 (x86, x86-64, ia64) SUSE Linux 9 (IA-64)
WeekSGTW iSGTWFrom April 2005Diverse audience>1000
subscribers
www.isgtw.org
RIT Colloquium (May 23, 2007)
Paul Avery 39
OSG News: Monthly Newsletter
18 issues by Apr. 2007
www.opensciencegrid.org/osgnews
RIT Colloquium (May 23, 2007)
Paul Avery 40
Grid Summer Schools
Summer 2004, 2005, 20061 week @ South Padre Island, TexasLectures plus hands-on exercises to ~40 studentsStudents of differing backgrounds (physics + CS), minorities
Reaching a wider audienceLectures, exercises, video, on webMore tutorials, 3-4/yearStudents, postdocs, scientistsAgency specific tutorials
RIT Colloquium (May 23, 2007)
Paul Avery 41
Project Challenges Technical constraints
Commercial tools fall far short, require (too much) invention
Integration of advanced CI, e.g. networks
Financial constraints (slide)Fragmented & short term funding injections (recent
$30M/5 years)Fragmentation of individual efforts
Distributed coordination and managementTighter organization within member projects compared to
OSGCoordination of schedules & milestonesMany phone/video meetings, travelKnowledge dispersed, few people have broad overview
RIT Colloquium (May 23, 2007)
Paul Avery 42
Funding & Milestones: 1999 – 2007
2000 2001 2003 2004 2005 2006 20072002
GriPhyN, $12M
PPDG, $9.5M
UltraLight, $2M
CHEPREO, $4M
Grid Communications
Grid Summer Schools 2004,
2005, 2006
Grid3 start OSG
start
VDT 1.0
First US-LHCGrid
Testbeds
Digital Divide Workshops04, 05, 06
LIGO Grid
LHC startiVDGL,
$14M
DISUN, $10M
OSG, $30M NSF, DOE
VDT 1.3
Grid, networking projects Large experiments Education, outreach, training
RIT Colloquium (May 23, 2007)
Paul Avery 43
Challenges from Diversity and Growth
Management of an increasingly diverse enterpriseSci/Eng projects, organizations, disciplines as distinct
culturesAccommodating new member communities (expectations?)
Interoperation with other gridsTeraGrid International partners (EGEE, NorduGrid, etc.)Multiple campus and regional grids
Education, outreach and trainingTraining for researchers, students… but also project PIs, program officers
Operating a rapidly growing cyberinfrastructure25K 100K CPUs, 4 10 PB diskManagement of and access to rapidly increasing data stores
(slide)Monitoring, accounting, achieving high utilizationScalability of support model (slide)
RIT Colloquium (May 23, 2007)
Paul Avery 44
Rapid Cyberinfrastructure Growth: LHC
0
50
100
150
200
250
300
350
2007 2008 2009 2010Year
MS
I200
0
LHCb-Tier-2
CMS-Tier-2
ATLAS-Tier-2
ALICE-Tier-2
LHCb-Tier-1
CMS-Tier-1
ATLAS-Tier-1
ALICE-Tier-1
LHCb-CERN
CMS-CERN
ATLAS-CERN
ALICE-CERN
CERN
Tier-1
Tier-2
2008: ~140,000PCs
Meeting LHC service challenges & milestonesParticipating in worldwide simulation productions
RIT Colloquium (May 23, 2007)
Paul Avery 45
OSG Operations
Distributed modelScalability!VOs, sites, providersRigorous problem