ParallelDatorCentrum - Enabling Science PDC May 27, 2002 Lennart Johnsson Grids Next Generation Infrastructure for Research, Education and Commerce Lennart Johnsson Director, ParallelDatorCentrum Professor, Numerical Analysis and Computer Science KTH
90
Embed
What Dreams May Come as envisioned at Bell Labs 75 …johnsson/Talks/Carl_lindberg_052702_RS.pdfParallelDatorCentrum - Enabling Science PDC May 27, 2002 Lennart Johnsson The Communication
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Grids Next Generation Infrastructure for
Research, Education and Commerce
Lennart JohnssonDirector, ParallelDatorCentrum
Professor, Numerical Analysis and Computer ScienceKTH
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The Communication SkinWhat Dreams May Come as envisioned at Bell Labs 75 year
anniversary
People will have instant access to millions of movies on demand at any time of the day or night. There will be interactive TV programs and 3-D networked, interactive video games, as well as giant, flat-screen, color videophones.And that's just the beginning. We are already building the first layer of a mega-network that will cover the entire planet like a skin.
Wireless base stations will become smaller and cheaper. You'll be able to send and receive large files through the air instantly, using a variety of information appliances. With unlimited bandwidth, it will be more direct and more convenient to get electronic newspapers, magazines, books, CDs, movies, and much more. And they'll cost a lot less, because the network will eliminate the middleman, like the video rental store. Software-driven intelligent networks and wireless technology will reach people wherever they are and let them choose if a message will be an e-mail, voice mail or video clip.
The telephone will be replaced by "metaphones" the size of jewelry and controlled by voice commands. The small metaphone on your lapel will even be able to read web sites and e-mail to you. Wearable videophones are another very real possibility.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The Communication Skin (cont’d)Major advances in videoconferencing and high-speed networking will lead to a new Age of Virtuality that will transform the way people live and conduct business -- with
virtual enterprises, virtual travel, virtual business conferences, virtual offices, virtual universities, and a host of other virtual experiences. Distance will become irrelevant -- and
place will no longer be limiting.
Videoconferencing will evolve into more realistic, more "immersive," "virtual-conferencing". Thousands of 360-degree cameras and stereo microphones placed around sporting events, music concerts and business meetings will give web participants full control of what they are seeing, hearing and experiencing.
Today's Internet will transform from an avalanche of data into a smarter "High-IQ Net" with natural interfaces, active Web sites, and software agents to extract desired information via text, voice, images and video. You'll be able to embed your personal preferences -- creating your own Web persona. Personal "cyberclones" will constantly anticipate your information wants and needs, letting you know, for example, when 10 inches of fresh powder falls at your favorite ski resort. This HiQNet, which will be as immediate as dial tone is today, will be so integral to our lives it will become practically invisible. People will use anything from a TV to a wireless lapel phone for access. Expect great things!
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Cost of Computing
In 2010, the compute power of today’s top-of-the-line PC can be found in $1 consumer electronics
Today’s most powerful computers (the power of 10,000 PCs at a cost of $100M) will cost a few hundred thousand dollars
0
10000
20000
30000
40000
50000
60000
70000
80000
90000
100000
1999 180 nm
2000 2001 130 nm
2002 .
2003 .
2004 90 nm
2005 .
2008 60 nm
2011 40 nm
2014 30 nm
Functions per chip, Mtransistors
DRAM, at Production
DRAM, at Introduction
MPU, Cost-Performance, atProduction MPU, Cost-Performance, atIntroduction MPU, High-Performance, atProduction ASIC, at Production
05
1015202530354045
$/M
tran
sist
or
1999 180nm
2001 130nm
2003 .
2005 100nm
2011 40 nm
Cost/Mtransistor
DRAM, cost x 100, at Introduction,$/MtransistorsDRAM, cost x 100, at Production,$/MtransistorsMPU, Cost-Performance, atIntroduction, $/MtransistorsMPU, Cost-Performance, atProduction, $/MtransistorsMPU, High Performance, atProduction, $/Mtransistors
SIA Roadmap
SIA Roadmap
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
In 2010, $1 will buy enough disk space to store
10,000 Books 35 hrs of CDQuality audio
2 min of DVDQuality Video
IBM 9.1GB Ultra 2XP
1980 1985 1990 1995 2000 2005 2010
Year
0.001
0.01
0.1
1
10
100
1000
Pric
e/M
Byt
e, D
olla
rs
HDD DRAM Flash Paper/FilmAverage Price of Storage
IBM 18.2GB Ultrastar
IBM Deskstar 37GB
Toshiba 6.4GB
IBM Deskstar4
IBM Deskstar3
IBM 16.8GB Deskstar
IBM 8.1GB Travelstar
Seagate 8.6GB
Quant 4.5GB
64MB
IBM 9.1GB Ultrastar
96 MB Flash Camera Mem.
64MB Flash
4MB Flash
16MB Flash1MB Flash
512KB Flash256KB Flash
128KB Flash
8KB
32KB 64KB
128KB
512KB 1MB 2MB
4MB
IBM6150
Wren II Seagate ST125
Maxt170IBM0615
IBM0663
Seagate B'cuda4
Seagate ST500
oem
prc2
000a
a.pr
z
128MB Flash
64MB
Ed Grochowski at Almaden
128MB Flash
IBM 25GB Travelstar
IBM 340 MB Microdrive
IBM Deskstar 25GB
IBM Deskstar 75GXP
IBM 1 GB Microdrive
1" HDD ProjectionDataQuest 2000
Flash ProjectionDataQuest 2000
Range of Paper/Film
3.5 " HDD 2.5 " HDD
1 " HDD
Flash
DRAM
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Fiber Optic Communication
In 2010. . .
A million books can be sent across the Atlantic for 1$ in 8 seconds
All books in the American Research Libraries can be sent across the Atlantic in about 1 hr for $500
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Fiberoptic Communication Milestones
First Laser 1960
First room temperature laser, ~1970
Continuous mode commercial lasers, ~1980
Tunable lasers, ~1990
Commercial fiberoptic WANs, 1985
10 Tbps/strand demonstrated in 2000 (10% of fiber peak capacity). (10 Tbps is enough bandwidth to transmit a million high-definition resolution movies simultaneously, or over 100 million phone calls).
WAN fiberoptic cables often have 384 strands of fiber and would have a capacity of 2 Pbps. Several such cables are typically deployed in the same conduit/right-of-way
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Optical Communication costs
Larry Roberts, Caspian Networks
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Internet bandwidth Growth
Larry Roberts, Caspian Networks
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
GlobeSystem Internet - We Connect the World !TeleGlobe
Over 400,000 route miles• Next generation technology• US$5B investment over 4
years
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The Battle of the Atlantic
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The Battle of the Pacific
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Access Technologies
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Access Devices
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The World Wide Computer
“When the Network is as fast as the computer’s internal links, the machine disintegrates across the Net into a set of special purpose appliances”
E-Science: Data Gathering, Analysis, Simulation, and Collaboration
LHC
CMS
Simulated Higgs Decay
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Wireless Sensors Allow Your Body to Become an Internet Data Source
Next Step—Putting You On-Line!– Wireless Internet Transmission– Key Metabolic and Physical Variables– Model -- Dozens of 25 Processors and 60
Sensors / Actuators Inside of our CarsPost-Genomic Individualized Medicine– Combine
• Genetic Code • Body Data Flow
– Use Powerful AI Data Mining Techniques
www.bodymedia.com
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
As Our Bodies Move On-LineNew Sensors—Israeli Video Pill– Battery, Light, & Video Camera– Images Stored on Hip Mounted
Device
www.givenimaging.com
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Neptune Undersea Grid
Air Quality Measurementand Control
Surface dataRadar dataBallon dataSatellite data
Real-time data
NCAR
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Neptune - Fiberoptic Telescope to Inner Space
NEPTUNE’s 3000 kilometers of fiber-optic cable will provide power and communications to scientific instruments. The system will provide real-time flow of data and imagery to shore-based Internet sites. It will permit interactive control over robotic vehicles on site and will provide power to the instruments and vehicles. NEPTUNE may also serve as a unique testbedfor sensor and robotic systems designed to explore other oceans in the solar system.The NEPTUNE network is expected to be operational by 2006 and will cost approximately $250 million to develop, install, and operate through the first five years.The NEPTUNE system must have the following characteristics to meet the scientific requirements:
– Plate scale (covering the full Juan de Fuca tectonic plate) – Power (order of tens of kW) – Bandwidth (order of many Gbits/sec) – Real-time data return and robotic control capability – Robust design for high reliability – Precision timing at all instruments – Available for nominal 20-30 years.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
March 28, 2000 Fort Worth Tornado
Courtesy Kelvin Droegemeier
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
In 1988 … NEXRAD Was Becoming a Reality
s fn
Courtesy Kelvin Droegemeier
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Wireless Sensors Will Allow Instrumentation of Critical Civil Infrastructure
New Bay Bridge Tower with Lateral Shear Links
Cal-(IT)2 Will Develop and Install Wireless Sensor ArraysLinked to Crisis Management
Control Rooms
Source: UCSD Structural Engineering Dept.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Next Step-Allen Telescope Array
Larry SmarrComputer Science
& Engineering
SETI@Home
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Casino-21: Large Scale Monte Carlo Climate Simulations
• Ensemble Computing Varying Model Parameters• Evaluate Model Against Current Climate• Home in on Most Realistic Models by Natural Selection• Then Model 21st Century Climate Evolution• One Climate Model per PC• Currently About 20,000 PCs
Rutherford Appleton Laboratory
www.climate-dynamics.rl.ac.uk/index.html
Larry SmarrComputer Science
& Engineering
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Evolving Drugs to Match Evolving AIDS Virus ByVarying Both Target Molecules and Target Proteins
Running AutoDock Application Software on 2100 PCs
In SilicoDrug Design
Art Olson, TSRI
Larry SmarrComputer Science
& Engineering
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Protein Folding with Distributed Computing
www.stanford.edu/group/pandegroup/
10,000 PCs70 CPU-Years
as of 10/13/00
Larry SmarrComputer Science
& Engineering
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
National Digital Mammography ArchiveAbout 40 million mammograms/yr (USA) (estimates 32 – 48 million)About 250,000 new breast cancer cases detected each yearOver 10,000 units (analogue)Image size: about 48 MbytesImages per patient: 4Data set size per patient: about 200 MbytesData set per year: about 10 PbytesData set per unit, if digital: 1 Tbytes/yr, on average
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Digital MammographyAbout 40 million mammograms/yr (USA) (estimates 32 – 48 million)About 250,000 new breast cancer cases detected each yearOver 10,000 units (analogue)Resolution: up to about 25 microns/pixelImage size: up to about 4k x 6k (example: 4096 x 5624)Dynamic range: 12-bitsImage size: about 48 MbytesImages per patient: 4Data set size per patient: about 200 MbytesData set per year: about 10 PbytesData set per unit, if digital: 1 Tbytes/yr, on averageData rates/unit: 4 Gbytes/operating day, or 0.5 Gbytes/hr, or 1 MbpsComputation: 100 ops/pixel = 10 Mflops/unit, 100 Gflops total; 1000 ops/pixel = 1 Tflops total
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Grid Application Projects
ODIN
PAMELA
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
BIOMEDICAL INFORMATICS RESEARCH NETWORK (BIRN)
The BIRN is an NCRR initiative aimed at creating a testbed to address biomedical researchers' need to access and analyze data at a variety of levels of aggregation located at diverse sites throughout the country. The BIRN testbed will bring together hardware and develop software necessary for a scalable network of databases and computational resources. Issues of user authentication, data integrity, security, and data ownership will also be addressed. $20M.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
500 Å
JEOL3000-FEGLiquid He stageNSF support
Biological Imaging
No. of Particles Needed for 3-D Reconstruction
B = 100 Å2
8.5 Å 4.5 Å6,000 5,000,000
Resolution
B = 50 Å2 3,000 150,0008.5 Å Structure
of the HSV-1 Capsid
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
10 geographically distributed observatories nationwide to serve as national research platforms for integrated, cutting-edge research in field biology To enable scientists to conduct experiments on ecological systems at all levels of biological organization – from molecular genetics to whole ecosystems, and across scales – from seconds to geological time, and from microns to regions and continents. Observatories will have scalable computation capabilities and will be networked via satellite and landlines – to each other and to specialized facilities, such as supercomputer centers. By creating one virtual installation via a cutting-edge computational network, all members of the field biology research community will be able to access NEON remotely.
National Ecological Observatory Network
$ 100M FY01-06 NSF support requested
10 observatories nationwide; sponsored by NSF, Archbold Biological Station and SDSC
http://www.sdsc.edu/NEON/
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Hemodynamics
Blood flow and stint research
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
NSFnetvBNS
Internet2 AbileneTeraGrid
0
2000
4000
6000
8000
10000
1986 1996 1997 1999 2001
Year
Mbi
t/s
OC-3
OC-12
OC-48
OC-192Doubling every year
Vancouver
SeattlePortland
San Francisco
Los Angeles
San Diego
(SDSC)
NCSA
Chicago NYC
SURFnetCA*net4
AMPATH
PSC
Atlanta
IU
U Wisconsin
DTF 40Gb
NTON
NTON
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
DoE EnergySciences Network
NASA Research and
Education Network (NREN)
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
I-WIREUIC
ANL
NCSA/UIUC
UC
NU / Starlight
Star Tap
IIT
Charlie Catlett Argonne National Laboratory
• State Funded Infrastructure to support Networking and Applications Research
– $6.5M Total Funding• $4M FY00-01 (in hand)• $2.5M FY02 (approved 1-
• Texas GigaPoP established• Regional High-Speed network established
(Rice, Texas Medical Center, UH; Texas A&M)
• Co-recipient (as part of the Globus team) of the Global Information Infrastructure Next Generation Award for Advanced Computing Infrastructure
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Computing Platforms
2001 ⇒ 2030Personal Computers O[$1000]– 109 Flops/sec in 2001 ⇒ 1015 – 1017 Flops/sec by 2030
Supercomputers O[$100,000,000]– 1013 Flops/sec in 2001 ⇒ 1018 – 1020 Flops/sec by 2030
Number of Computers [global population ~1010]– SCs ⇒ 10-8 –10-6 per person ⇒ 102 – 104 systems– PCs ⇒ .1x – 10x per person ⇒ 109 – 1011 systems– Embedded ⇒ 10x – 105x per person ⇒ 1011 – 1015 systems– Nanocomputers ⇒ 0x – 1010 per person ⇒ 0 – 1020 systems
Available Flops Planetwide by 2030– 1024 – 1030 Flops/sec [assuming classical models of computation]
Courtesy Rick Stevens
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Visualization
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Petter Bjørstad, ParallabLennart Johnsson, PDC
Jari Järvinen, CSC
CSCHelsinki
ParallabBergen
PDCStockholm
Nordic Grid Consortium
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
ObjectivesProvide
Best-of-practice environment and resources for data and compute intensive research and educationTraining and user support across a broad spectrum of applications, computing platforms, data resources, and visualization systemsConvenient, transparent, secure access for eligible users
Conduct research and development in software systems for enhanced user services and efficient use of Grid resourcesBe a resourceful partner in European and intercontinental research, development and deployment of integrated large-scale data, computing and visualization resources serving academia and its partners
Nordic Grid Consortium
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Founding PartnersCSC, the Finnish Ministry of Education’s National Center
for High-Performance Computing and Networks. http://www.csc.fi
Parallab, University of Bergen’s Computational Science and High-Performance Computing Laboratory funded in part by the Norwegian Research Council. http://www.parallab.uib.no
PDC (ParallalDatorCentrum), Royal Institute of Technology (KTH), Stockholm, the lead national center serving Swedish academic research and higher education funded in part by the Swedish Research Council. http://www.pdc.kth.se
Nordic Grid Consortium
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
CAVERN
Grid Project Experiencies
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Automated Massive DNA Sequencing
PDC in collaboration with Center for Structural Biochemistry, KI1992
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Remote files system: SSF Large Scale
Genotyping Laboratory
UU/SSF National Large-Scale Genotyping Laboratory, since 1998– Data repository and long
term storage at PDC
betula.genpat.uu.se
Genotyping Machine
HSM
Archive
toad.pdc.kth.se
http://genpat.pdc.kth.se:8001
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
68,365,451 milljon hits during the race Peak simultaneous connections: ~37 thousand Peak rates: 29.2 Mbps out, 7.3 Mbps in Peak IP addresses: 165,915 Peak one day hits: 32,155,188Peak 5 minutes: 446,695 hits (89,339 hits/min)Data transferred: 160,632 GByteshttp://www.vasaloppet.se
PlansWork with sponsoring agencies to facilitate resource sharingSet up a NGC portalFirst edition shared software environment by end of April (Globus)“Friendly” users in summer of 2002Account and user access managmentLimited service in accordance with sponsor policies by the end of 2002
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The main goal of the DataGridinitiative is to develop and test the technological infrastructure that will enable the implementa-tion of scientific “collaboratories”where researchers and scientists will perform their activities regardless of geographical loca-tion. The project will devise and develop scalable software solu-tions and testbeds in order to handle many PetaBytes of distri-buted data, tens of thousand of computing resources (processors, disks, etc.), and thousands of simultaneous users from multiple research institutions. €9.3M
EU-DataGrid
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
EU-DataTag
The fundamental objective of the DataTAG project is to create a large-scale inter-continental Grid testbed in-volving the DataGRID pro-ject, several national pro-jects in Europe, and relatedGrid projects in the USA. This will allow to explore advanced networking tech-nologies and interoperability issues between different Grid domains
2002-02-16 CERN Update & DataTAG Project Olivier H. Martin (9)
NLNLSURFnet
CERN
UKUKSuperJANET4
AbileneAbilene
ESNETESNET
MRENMREN
ITITGARR-B
GEANT
NewYork
STAR-TAP
STAR-LIGHT
DataTAG project
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
GridLab (EU)
The GridLab project will de-velop a easy-to-use, flexible, generic and modular Grid Application Toolkit (GAT),enabling todays applicationsto make innovative use of global computing resources. The project is grounded by two principles, (i) the co-development of infrastruc-ture with real applications and user communities, lead-ing to working scenarios, and (ii) dynamic use of grids, with self-aware simulations adapting to their changing environment. €5M
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
E-Science (UK)
Biotechnology and Biological Sciences Research Council, £8MCentral Laboratory of the Research Councils, £5ME-Science Core Programme,
£15M from the Office of Science and Technology, £20M from the Department of Trade and Industry£15M industry match
Economic and Social Research Council, £3M Engineering and Physical Sciences Research Council, £17MMedical Research Council, £8MNatural Environment Research Council, £7M Particle Physics and Astronomy Research Council, £26MTeraFlop platform, £9M
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
The Grid for UK Particle Physics
GridPP will deliver the Grid software (middleware) and hardware infrastructure to enable testing of a prototype of the Grid for the Large Hadron Collider(LHC) project. The project will disseminate GridPPdeliverables in the multi-disciplinary e-Science environment and will seek to build collaborations with emerging Grid activities both nationally and internationally. £21M
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
AstroGrid (UK)
AstroGrid is a £5M project aimed at building a data-grid for UK astronomy, which will form the UK contribution to a global Virtual Observatory. The goals of the project are :
- A working datagrid for key UK databases - High throughput datamining facilities for interrogating those
databases - A uniform archive query and data-mining software interface - The ability to browse simultaneously multiple datasets - A set of tools for integrated on-line analysis of extracted data - A set of tools for on-line database analysis and exploration - A facility for users to upload code to run their own algorithms on
the datamining machines - An exploration of techniques for open-ended resource discovery
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
INFN-Grid
http://www.ge.infn.it/com4/grid/grid.pdf
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
DutchGridThe "Dutch Grid Infrastructure" is the sum of all the individual test beds of the contributing partners: each or-ganization is a fully autonomous admi-nistrative entity. The authentication service, the Dutch-Grid Certification Authority, is shared by (nearly) all part-ners; this CA is also an integral part of the authentication mesh of the EU DataGrid and related projects.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
SimDBNPACI Alpha_ProjectUniversity of Houston
Simulation Data Base
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
EarthScope: USArray and SAFOD
USArray: a dense array of high-capability seismometers to be deployed throughout the US to improve our resolution of the subsurface structure.San Andreas Fault Observatory at Depth (SAFOD): a 4km-deep hole into the San Andreas fault zone close to the hypocenter of the 1966 M~6 Parkfield earthquake, to access a major active fault at depthTo provide input to NSF’s Network for Earthquake Engineering Simulation (NEES) project that studies the response of the built environment to earthquakes.
$ 74.81M FY01-04 NSF support requestedNSF, USGS, NASA Consortium
EarthScope will bring real-time Earth Science data to our desktops, to provide unprecedented opportunities to unravel the structure, evolution, and dynamics of the North American continent, and to better understand earthquakes and fault systems, volcanoes and magmaticprocesses, and links between tectonics and surficalprocesses.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Network for Earthquake Engineering Simulation (NEES)
$ 81.8M FY01-04 NSF support requestedScoping study managed by NCSA; sponsored by NSF
NEES will provide a networked, national resource of geographically-distributed, shared-use, next-generation, experimental research equipment installations, with tele-observation and tele-operation capabilities. NEES will shift the emphasis of earthquake engineering research from current reliance on physical testing to integrated experimentation, computation, theory, databases, and model-based simulation using input data from EarthScope and other sources. NEES will be a collaboratory – an integrated experimental, computational, communications, and curated repository system, developed to support collaboration in earthquake engineering research and education.
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
10 geographically distributed observatories nationwide to serve as national research platforms for integrated, cutting-edge research in field biology To enable scientists to conduct experiments on ecological systems at all levels of biological organization – from molecular genetics to whole ecosystems, and across scales – from seconds to geological time, and from microns to regions and continents. Observatories will have scalable computation capabilities and will be networked via satellite and landlines – to each other and to specialized facilities, such as supercomputer centers. By creating one virtual installation via a cutting-edge computational network, all members of the field biology research community will be able to access NEON remotely.
National EcologicalObservatory Network
$ 100M FY01-06 NSF support requested
10 observatories nationwide; sponsored by NSF, Archbold Biological Station and SDSC
http://www.sdsc.edu/NEON/
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Atacama Large Millimeter Array
Prior to developing ALMA, the US conceived the MMA as an aperture-synthesis radio telescope operating in the wavelength range from 3 to 0.4 mm. ALMA will be the world’s most sensitive, highest resolution, millimeter-wavelength telescope. It will combine an angular resolution comparable to that of the Hubble Space Telescope with the sensitivity of a single antenna nearly 100 meters in diameter.ALMA will consist of no less than 64 12-meter antennas located at an elevation of 16,400 feet in Llano de Chajnantor, Chile
Atacama Large Millimeter Array (ALMA), an expanded Millimeter Array (MMA)$ 32M FY98-01 NSF support requested for Design and DevelopmentUS: National Radio Astronomy Observatory and Associated Universities, Inc. with NSF funding
Europe: European Southern Observatory, Centre National de la Recherche Scientifique, Max-Planck-Gesellschaft, Netherlands Foundation for Research in Astronomyand Nederlandse Onderzoekschool Voor Astronomie, and the UKParticle Physics and Astronomy Research Council
http://www.alma.nrao.edu/
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Construction of two detectors of the LHC: ATLAS (A Toroidal Large Angle Spectrometer) and CMS (Compact Muon Solenoid)The research, design, and prototyping of Petascale Virtual Data Grids, which will support the LHC as well as the SDSS (Sloan Digital Sky Survey) and LIGO (Laser Interferometer Gravitational-wave Observatory), is being carried out by GriPhyN, a multi- institutional team that received the largest NSF ITR grant in FY00.
Large Hadron ColliderLarge Hadron Collider (LHC)$ 80.9M FY99-04 NSF support requested
CERN (Switzerland) and international Consortium
http://lhc.web.cern.ch/lhc/
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Grid Physics Network
Team of physicists and computer scientists who plan to implementPetascale Virtual Data Grids (PVDGs) computational environments for data-intensive scienceFour physics experiments – CMS, ATLAS, LIGO, SDSS –share common challenges: massive datasets, large-scale computational resources and diverse communities of thousands of scientists spread across the globe
– The LHC CMS and ATLAS experiments will search for the origins of mass and probe matter at the smallest length scales
– LIGO will detect the gravitational waves of pulsars, supernovae and in-spiraling binary stars
– SDSS will carry out an automated sky survey enabling systematic studies of stars, galaxies, nebula, and large-scale structure
GriPhyN estimates 20 Tier 2 sites (6 CMS, 6 ATLAS, 5 LIGO and 2 SDSS), with a projected five-year cost of ~$85M-90M, half of which is for hardware
Grid Physics Network (GriPhyN)$ 11.9M FY00 NSF for R&D Development (largest ITR award)
Led by Univ Florida and Univ Chicago; includes US institutions
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
International Virtual Data Grid LaboratoryThe iVDGL will provide a global computing resource for several leading international experiments in physics and astronomy, including the Laser Interferometer Gravitational-wave Observatory (LIGO), the ATLAS and CMS experiments at CERN, the Sloan Digital Sky Survey (SDSS), and the proposed National Virtual Observatory (NVO). $13.5M
www.ivdgl.org
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Particle Physics Data Grid
PPDG will develop, acquire and deliver vitally needed Grid-enabled tools for data-intensive requirements of particle and nuclear physics. Novel mechanisms and policies will be vertically integrated with Grid middleware and experiment-specific applications and computing resources to form effective end-to-end capabilities. $11.5M
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
National Data ObservatoryThe National Virtual Observatory (NVO) will unite astronomical databases of many earthbound and orbital observatories, taking advantage of the latest computer technology and data storage and analysis techniques. The goal is to maximize the potential for newscientific insights from the data by making them available in an accessible, seamlessly unified form to professional researchers, amateur astronomers and students. $10 M.
http://us-vo.org
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Globus - Condor
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
National Middle-ware Initiative
Grids Research Integration Deployment and Support Center
NMINational Middleware Initiative
$12M
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Distributed Terascale FacilityThe DTF will perform 13.6-trillion calculations per second and store more than 450-trillion bytes of data, with a comprehensive infrastructure called the “TeraGrid” to link computers, visualization systems and data at four sites through a 40-billion bits-per-second optical network. $53M.
26
24
8
4 HPSS
5
HPSS
HPSS UniTree
External Networks
External NetworksExternal
Networks
External Networks
Site Resources Site Resources
Site Resources
Site Resources NCSA/PACI8 TF240 TB
SDSC4.1 TF225 TB
Caltech Argonne
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
NASA’s Information Power Grid
HQARC
JPL
GRC GSFC
MSFC
NGIXChicago
NREN WANTestbed
SDSC
CMU
NextGeneration
Internet
NCSA
LaRC
300 node Condor pool
MCAT/SRB
1024 CPU, O3000
DMF MDSCA
Globus and other Grid Services
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
Combining Grid and Web Services
Grid
Pro
toco
ls a
nd G
rid S
ecur
ity In
fras
truc
ture
Clie
nts
ApplicationPortals
WebServices
Grid Services:Collective and Resource Access Resources
Compute(many)
Storage(many)
Communi-cation
Instruments(various)
GRAM
GridFTPData Replica and Metadata Catalog
Grid Monitoring Architecture
GridInformation
Service
Web
Bro
wse
r
Grid Web ServiceDescription (WSDL)& Discovery (UDDI)
Grid X.509Certification
Authority
SRB/MetadataCatalogue
Condor-G
CORBA
MPI
Secure, Reliable
Group Comm.
Discipline /Application
SpecificPortals
(e.g. SDSCTeleScience)
ProblemSolving
Environments(AVS, SciRun,
Cactus)
EnvironmentManagement(LaunchPad,
HotPage)
Job Submission /Control
File Transfer
Data Management
CredentialManagement
MonitoringEvents
WorkflowManagement
other services:•visualization•interface builders•collaboration tools•numerical grid
generators, etc.
Apache Tomcat&WebSphere&Cold Fusion=JVM + servlet
instantiation + routing
CoG Kits implementingWeb Services in
servelets, servers, etc.
Python, Java, etc.,JSPs
compositionframeworks(e.g. XCAT) XM
L / S
OA
P ov
er G
rid S
ecur
ity In
fras
truc
ture
Grid
Pro
toco
ls a
nd
Grid
Sec
urity
Infr
astr
uctu
re
Apache SOAP,.NET, etc.
……
http
, http
s. e
tc.
X W
indo
ws
PDA
Grid ssh
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart Johnsson
DoE Science Grid
Initial Science Grid Configuration
NERSCSupercomputing
& Large-Scale Storage
PNNL
LBNL
ANL
ESnet
Europe
DOEScience Grid
ORNL
ESNetMDSCA
Grid Managed Resources
User InterfacesApplication Frameworks
Applications
Grid Services: Uniform access to distributed resources
Grid
In
form
atio
n
Serv
ice
Uni
form
Res
ourc
eA
cces
sB
roke
ring
Glo
bal
Que
uing
Glo
bal E
vent
Serv
ices
Co-
Sche
dulin
g
Dat
a C
atal
ogui
ngU
nifo
rm D
ata
Acc
ess
Col
labo
ratio
n an
d R
emot
e In
stru
men
t Se
rvic
esN
etw
ork
Cac
heC
omm
unic
atio
n Se
rvic
esA
uthe
ntic
atio
n A
utho
rizat
ion
Secu
rity
Serv
ices
Aud
iting
Faul
t M
anag
emen
t
Mon
itorin
g
Asia-Pacific
ParallelDatorCentrum - Enabling Science
PDC May 27, 2002 Lennart JohnssonCourtesy Larry Smarr