Top Banner
1 Lectures on Grid Computing Tuğba Taşkaya-Temizel Prof. K. Ahmad January 2005
54
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Lecture slides / notes.

1

Lectures on Grid Computing

Tuğba Taşkaya-TemizelProf. K. Ahmad

January 2005

Page 2: Lecture slides / notes.

2

Grid ComputingEverywhere

Business: Sectors like financial services, industrial manufacturing, energy…

Humanitarian works

Research : Health, Aerospace, Astronomy, Finance…

Government

Page 3: Lecture slides / notes.

3

Grid Computing

The internet took 20 years to be taken seriously by business. By comparison the grid is happening far more rapidly. Tom Hawk, IBM Insight Research says the worldwide market for grid technology and services is doubling every year and will reach $5 billion by 2008. Grid computing is just one of the technologies the UK government says, in its latest report, should receive more support and funding. (December 17,2003)

Page 4: Lecture slides / notes.

4

Grid Computing

"We really do believe that grid computing is real," CEO of Hewlett-Packard Carly Fiorina said. "It is driving the R&D in our industry. For the first time our energy is focused on something else than building a killer app or a hot box. We are more focused on making system that combines the best of IT and business. Imagine what is possible." (September 11, 2003)"The Grid will be the major new direction for IT," said Geoff Brown, technical director for ATS Core Technologies at Oracle. (October 28, 2002)

Page 5: Lecture slides / notes.

5

DEFINITIONS: Grid?

ELECTRICITY GRID:• A network of high-voltage transmission lines and connections that supply electricity from a number of generating stations to various distribution centres in a country or a region, so that no consumer is dependent on a single station.

UTILITY GRID: • (Term) used of any network that serves a similar purpose for other services.

www.oed.com

Page 6: Lecture slides / notes.

6

DEFINITIONS: Grid?

GRID: The Grid is envisaged to be

‘the computing and data management infrastructure that will provide the electronic underpinning for a global society in business, government, science and entertainment’

Berman, Fox and Hey (2003:9)

Page 7: Lecture slides / notes.

7

DEFINITIONS: Grid?

GRID: A virtual information

processing environment where the user has the ‘illusion’ of a seamless single-source computing power which is actually distributed.

Page 8: Lecture slides / notes.

8

Why should you care?

Ian Foster explains why we should care Grids in three points:

Vision

Reality

Future

Page 9: Lecture slides / notes.

9

Why should you care?

Grid is a disruptive technology [Vision] It ushers in a virtualized, collaborative,

distributed world.

Two interrelated opportunities1) Enhance economy, flexibility, access by

virtualizing computing resources2) Deliver entirely new capabilities by

integrating distributed resources

Vision

Reality

Future

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 10: Lecture slides / notes.

10

Why should you care? Virtualization

Vision

Reality

Future

Vision

Reality

Future

Application Virtualization

• Automatically connect applications to services• Dynamic & intelligent provisioning

Infrastructure Virtualization

• Dynamic & intelligent provisioning• Automatic failover

Source: The Grid: Blueprint for a New Computing Infrastructure (2nd Edition), 2004

Page 11: Lecture slides / notes.

11

Why should you care? Distributed System Integration

Vision

Reality

Future

Vision

Reality

Future

UK e-Science Centres

Source: http://www.nesc.ac.uk/centres/

Page 12: Lecture slides / notes.

12

Why should you care?Vision

Reality

Future

Vision

Reality

Future

Source: “The Anatomy of the Grid”, Foster, Kesselman, Tuecke, 2001

The real and specific problem that underlies the Grid concept is coordinated resource sharing and problem solving in dynamic, multi-institutional virtual organizations.

Page 13: Lecture slides / notes.

13

Why should you care?Terminology

Grid has strong links with “Utility Computing”, “Autonomic Computing” and “Service Oriented Architecture”.

Vision

Reality

Future

Vision

Reality

Future

Page 14: Lecture slides / notes.

14

Why should you care?

Grid addresses pain points now [Reality]Grids are built not bought, but are delivering

real benefits in commercial settings Low utilization of enterprise resources High cost of provisioning for peak demand Inadequate resources prevent use of

advanced applications Lack of information integration

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 15: Lecture slides / notes.

15

Why should you care?Early Commercial Applications

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

“Gridified” Infrastructure

FinancialServices

DerivativesAnalysis

Statistical Analysis

Portfolio Risk

Analysis

DerivativesAnalysis

Statistical Analysis

Portfolio Risk

Analysis

Manufacturing

Mechanical/ Electronic

Design

Process Simulation

FiniteElement Analysis

Failure Analysis

Mechanical/ Electronic

Design

Process Simulation

FiniteElement Analysis

Failure Analysis

LS / Bioinformatics

Cancer Research

Drug Discovery

Protein Folding

Protein Sequencing

Cancer Research

Drug Discovery

Protein Folding

Protein Sequencing

Other

Web Applications

Weather Analysis

Code Breaking/

Simulation

Academic

Web Applications

Weather Analysis

Code Breaking/

Simulation

Academic

Sources: IDC, 2000 and Bear Stearns- Internet 3.0 - 5/01 Analysis by SAI

Gri

d S

erv

ice

s M

ark

et

Op

po

rtu

nit

y 2

00

5

Energy

Seismic Analysis

Reservoir Analysis

Seismic Analysis

Reservoir Analysis

Entertainment

Digital Rendering

Digital Rendering

Massive Multi-Player

Games

Massive Multi-Player

Games

Streaming Media

Streaming Media

Leading adopters (Oct 2003) *• Financial services: 31%• Life sciences: 26%• Manufacturing: 18%

*Grids 2004: From Rocket Science To Business Service, The 451 Group

Page 16: Lecture slides / notes.

16

Why should you care?Grid Deployment Strategies

A range of excellent commercial & open source products for resource federation Federate enterprise computing resources Federate enterprise information resources Globus Toolkit®: inter-enterprise sharing

But, “Grids are built, not bought” Integration with other enterprise systems is

needed to deliver complete solution

Start small & with well-defined ROI case Grow based on experience

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 17: Lecture slides / notes.

17Image courtesy Christian Richters: Source:Wired News

Data Grids for High Energy Physics

Fastest particle accelarator: Large Hadron ColliderWhen completed in 2005, CERN's Large Hadron Collider will send protons and ions from hydrogen nuclei rushing through a 17-mile circular tunnel at speeds of up to 52,200,000 miles per hour.

Page 18: Lecture slides / notes.

18Image courtesy Harvey Newman, Caltech

Tier2 Centre ~1 TIPS

Online System

Offline Processor Farm

~20 TIPS

CERN Computer Centre

FermiLab ~4 TIPSFrance Regional Centre

Italy Regional Centre

Germany Regional Centre

InstituteInstituteInstituteInstitute ~0.25TIPS

Physicist workstations

~100 MBytes/sec

~100 MBytes/sec

~622 Mbits/sec

~1 MBytes/sec

There is a “bunch crossing” every 25 nsecs.

There are 100 “triggers” per second

Each triggered event is ~1 MByte in size

Physicists work on analysis “channels”.

Each institute will have ~10 physicists working on one or more channels; data for these channels should be cached by the institute server

Physics data cache

~PBytes/sec

~622 Mbits/sec or Air Freight (deprecated)

Tier2 Centre ~1 TIPS

Tier2 Centre ~1 TIPS

Tier2 Centre ~1 TIPS

Caltech ~1 TIPS

~622 Mbits/sec

Tier 0Tier 0

Tier 1Tier 1

Tier 2Tier 2

Tier 4Tier 4

1 TIPS is approximately 25,000

SpecInt95 equivalents

Data Grids for High Energy Physics

Page 19: Lecture slides / notes.

19

Mathematicians Solve NUG30Quadratic Assignment Problem

MetaNEOS: Argonne, Iowa, Northwestern, WisconsinSource:Shawn McKee The Grid:The Future of High Energy Physics Computing? January 7,2002

Location 1

Location 2

Location 3

Location 4

The distances are:•d(1,2) = 22, •d(1,3) = 53, •d(2,3) = 40, •d(3,4) = 55.

The required flows between facilities are:

•f(2,4) = 1, •f(1,4) = 2, •f(1,2) = 3, •f(3,4) = 4.

The permutation p corresponding to this graphical solution is ( 2, 1, 4, 3 ).

Page 20: Lecture slides / notes.

20

Mathematicians Solve NUG30Looking for the solution to the NUG30 quadratic assignment problem An informal collaboration of mathematicians and computer scientistsCondor-G delivered 3.46E8 CPU seconds in 7 days (peak 1009 processors) in U.S. and Italy (8 sites)

NUG30 Solution: 14,5,28,24,1,3,16,15,10,9,21,2,4,29,25,22,13,26,17,30,6,20,19,8,18,7,27,12,11,23

MetaNEOS: Argonne, Iowa, Northwestern, WisconsinSource:Shawn McKee The Grid:The Future of High Energy Physics Computing? January 7,2002

Page 21: Lecture slides / notes.

21

Network for Earthquake Engineering Simulation NEESgrid: national infrastructure to couple earthquake engineers with experimental facilities, databases, computers, & each otherOn-demand access to experiments, data streams, computing, archives, collaboration

NEESgrid: Argonne, Michigan, NCSA, UIUC, USC

Page 22: Lecture slides / notes.

22

The 13.6 TF TeraGrid: Computing at 40 Gb/s

26

24

8

4 HPSS

5

HPSS

HPSS UniTree

External Networks

External Networks

External Networks

External Networks

Site Resources Site Resources

Site ResourcesSite ResourcesNCSA/PACI8 TF240 TB

SDSC4.1 TF225 TB

Caltech Argonne

TeraGrid/DTF: NCSA, SDSC, Caltech, Argonne www.teragrid.org

Page 23: Lecture slides / notes.

23U.S. PIs: Avery, Foster, Gardner, Newman, Szalay www.ivdgl.orgImage courtesy of http://www.sdss.org/news/releases/20050111.yardstick.html

iVDGL:International Virtual Data Grid Laboratory

Sloan Digital Sky Survey is the most ambitious astronomical survey project ever undertaken. The survey will map in detail one-quarter of the entire sky, determining the positions and absolute brightnesses of more than 100 million celestial objects. It will also measure the distances to more than a million galaxies and quasars

Page 24: Lecture slides / notes.

24U.S. PIs: Avery, Foster, Gardner, Newman, Szalay www.ivdgl.org

iVDGL:International Virtual Data Grid Laboratory

Tier0/1 facility

Tier2 facility

10 Gbps link

2.5 Gbps link

622 Mbps link

Other link

Tier3 facility

Page 25: Lecture slides / notes.

25

Why should you care?

An open Grid is to your advantage [Future] Standards are being defined now that will determine the future of this technology

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 26: Lecture slides / notes.

26

Grid Vision, Marketing, and Reality

Vision Computing & data resources can be shared

like content on the Wb

Marketing Have we got a [Data, compute, knowledge,

information, desktop, PC, enterprise, cluster, …] Grid for you!

Reality Commercial products mostly noninteroperable Open source tools offer de facto standards,

but are also far from a complete solution

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 27: Lecture slides / notes.

27

Standards Matter!Open, standard protocols Enable interoperability Avoid product/vendor lock-in Enable innovation/competition on end points Enable ubiquity

In Grid space, must address how we Describe, discover, & access resources Monitor, manage, & coordinate, resources Account & charge for resources

For many different types of resource

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 28: Lecture slides / notes.

28

Open Grid Services Architecture

Define a service-oriented architecture … the key to effective virtualization

… that addresses vital “Grid” requirements AKA utility, on-demand, system management,

collaborative computing in particular, distributed service management

… building on Web services standards extending those standards where needed

“The Physiology of the Grid: An Open Grid Services Architecture for Distributed Systems Integration”, Foster, Kesselman, Nick, Tuecke, 2002

Vision

Reality

Future

Page 29: Lecture slides / notes.

29

A family of six Web services specifications A design pattern to

specify how to use Web services to access “stateful” components

Message-based publish-subscribe to Web services

Latest Step Forward:WS-Resource Framework

Gro

ups

References

Noti

fica

tion

Faults

Properties

Lifetim

e

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 30: Lecture slides / notes.

30

WS-Resource Framework Completes Grid-WS Convergence

Grid

Web

The definition of WSRF means that Grid and Web communities can move forward on a common base

WSRF

Started far apart in apps & tech

OGSI

GT2

GT1

HTTPWSDL,

WS-*

WSDL 2,

WSDM

Have beenconverging

Vision

Reality

Future

Source: Ian Foster’ s presentation on “The Grid” , COMDEX 2003, Las Vegas, Nevada USA, November 18, 2003

Page 31: Lecture slides / notes.

31

The Evolution of the GRID1980’s Parallel computing clusters -

improved performance from tightly coupled clusters and data sharing

1990’s Grid 1: Extend the advances in parallel computing to geographically distributed systems

2000 Grid II: Grid is a platform for integrating loosely coupled applications: some components running in parallel and some for linking disparate resources largely developed in the serial-von-Neumann paradigm - storage, visualisation, a-d/d-a converters and sensors

Page 32: Lecture slides / notes.

32

The Evolution of the GRID

Currently there are (clusters) of very powerful computing/ communications systems (i) Systems for acquiring digital data and processing data (Amazon.com or Oracle clusters)

(ii) Systems for analysing and visualising information (CERN’s large hadron collider, Protein Synthesis systems)

(iii) Systems for imaging, analysis and visualisation for distributed data (weather prediction, satellite based military civilian systems)

(iv) Systems that can link Sensors and predict on real-time information (military systems, video surveillance)

Page 33: Lecture slides / notes.

33

The Evolution of the GRIDDevelopments in networking technologies, operating systems, clustered data bases, application services and device technologies have enabled developers to build systems with literally distributed millions of nodes for providing: • Web-based services personal commercial transactions

• Content delivery networks that can cache web-pages seamlessly

• Wireless networks have spawned ad-hoc distributed systems that when linked to wide-area networks lead to a complex distributed system.

Problems of efficiency, reliability, accessibility and security are not addressed in ‘global’ terms.

Page 34: Lecture slides / notes.

34

The Evolution of the GRID

* Sputnik

1960 1970 1975 1980 1985 1990 1995 2000

* ARPANET

* Email* Ethernet

* TCP/IP* IETF

* Internet Era * WWW Era

* Mosaic

* XML

* PC Clusters* Crays * MPPs

* Mainframes

* HTML

* W3C

* P2P

* Grids

* XEROX PARC wormCO

MP

UTIN

GC

om

mu

nic

ati

on

* Web Services

* Minicomputers

* PCs

* WS Clusters

* PDAs* Workstations

* HTC

Source: www.gridbus.org

Page 35: Lecture slides / notes.

35

2100

2100 2100 2100 2100

2100 2100 2100 2100

Personal Device SMPs or SuperComputers

LocalCluster

GlobalGrid

PERFORMANCE

+

Q

o

S

•Individual•Group•Department•Campus•State•National•Globe•Inter Planet•Universe

Administrative Barriers

EnterpriseCluster/Grid

The Evolution of the GRID

Source: www.gridbus.org

Page 36: Lecture slides / notes.

36

The Evolution of the GRID

Grid is being developed not only to make distributed resources available to end-user not also to co-ordinate such usage for sharing and aggregation of resources.

Page 37: Lecture slides / notes.

37

The Evolution of the GRID

Moore’s law improvements in computing produce highly functional end-systemsThe internet and burgeoning wired and wireless provide wide-spread connectivityChanging modes of working and problem solving emphasise teamwork, computationNetwork growth produce dramatic changes in topology and geography

Page 38: Lecture slides / notes.

38

The Evolution of the GRID

The first generation involved proprietary solutions for sharing high-performance computing resourcesThe second generation introduced middleware to cope with scale and heterogeneityThe third generation introduced a service-oriented approach leading to commercial projects in addition to the scientific projects now collectively known as e-Science

Page 39: Lecture slides / notes.

39

The Evolution of the GRIDThe first generation

FAFNER, I-WAY

The second generation Technologies: Globus, Legion Distributed object systems (Jini and RMI, The

common component architecture form) Grid resource brokers and schedulers Grid portals Integrated systems Peer-to-Peer computing

The third generation Service-oriented architecture (web services, OGSA,

Agents) Information aspects: relation with the World Wide

Web Live information systems

Page 40: Lecture slides / notes.

40

The Evolution of the GRIDGlobus Toolkit® History

0

5000

10000

15000

20000

25000

30000

1997 1998 1999 2000 2001 2002

Do

wn

loa

ds

pe

r M

on

th f

rom

ftp

.glo

bu

s.o

rg

DARPA, NSF, and DOE begin funding Grid work

NASA beginsfunding Grid work,DOE adds support

The Grid: Blueprint for a New Computing

Infrastructure published

GT 1.0.0Released

Early ApplicationSuccesses Reported

NSF & European CommissionInitiate Many New Grid Projects

Anatomy of the GridPaper Released Significant

CommercialInterest inGrids

Physiology of the GridPaper Released

GT 2.0Released

Does not include downloads from:NMI, UK eScience, EU Datagrid,IBM, Platform, etc.

Source: Ian Foster’ s presentation on “The First 50 Years” , British Computer Society, Lovelace Medal Award Presentation, May, 2003

Page 41: Lecture slides / notes.

41

Building blocks of the Grid

NetworksComputational ‘nodes’ on the GridPulling it all togetherCommon infrastructure: standards

Page 42: Lecture slides / notes.

42

GRID: Key Issues

Resources Discovery, Allocation, Scheduling

Availability

Access, Security, Networks

Efficiency Economy, Management Administration.

Hardware Computers, Services, Networks

Application

Development, Testing

Page 43: Lecture slides / notes.

43

GRID: Key Issues Sharing

A biochemist will be able to exploit 10,000 computers to screen 100,000 compounds in an hour1,000 physicists worldwide will be able to pool resources for petop analyses of petabytes of data

A multidisciplinary analysis in aerospace couples code and data in geographically distributed organisations may be possibleCivil engineers colloborate to design, execute, and analyse shake table experiments

Climate scientists will be able to visualise, annotate, and analyse terabyte simulation datasets

Page 44: Lecture slides / notes.

44DOE X-ray grand challenge: ANL, USC/ISI, NIST, U.Chicago

tomographic reconstruction

real-timecollection

wide-areadissemination

desktop & VR clients with shared controls

Advanced Photon Source

GRID: Key Issues Sharing Online Access to Scientific Instruments

archival storage

Page 45: Lecture slides / notes.

45

MORE DEFINTIONS

ResourceNetwork protocolNetwork enabled serviceApplication Programming Interface(API)Software Development Kit (SDK)Syntax

Page 46: Lecture slides / notes.

46

MORE DEFINTIONS: Resource

An entity that is to be shared E.g., computers, storage, data, software

Does not have to be physical entity E.g., Condor pool, distributed file system,…

Defined in terms of interfaces, not devices E.g. scheduler such as LSF and PBS define a

compute resource Open/close/read/write define access to a

distributed file system, e.g NFS, AFS, DFS

Page 47: Lecture slides / notes.

47

MORE DEFINTIONS: Network protocol

A formal description of message formats and a set of rules for message exchange Rules may define sequence of message

exchanges Protocol may define state-change in endpoint,

e.g. file system state change

Good protocols designed to do one thing Protocols can be layered

Examples of protocols IP, TCP, TLS( was SSL), HTTP, Kerberos

Page 48: Lecture slides / notes.

48

MORE DEFINTIONS: Network enabled services

Implementation of a protocol that defines a set of capabilities Protocol defines interaction with

service All services require protocols Not all protocols are used to provide

services (e.g. IP, TLS)

Examples: FTP and Web servers

Page 49: Lecture slides / notes.

49

MORE DEFINTIONS: Application Programming Interface

(API)

A specification for a set of routines to facilitate application developmentSpec often language specific (or IDL) Routine name, number, order and type of

arguments; mapping to language constructs Behaviour or function of routine

Examples GSS API(security), MPI (message passing)

Page 50: Lecture slides / notes.

50

MORE DEFINTIONS: Software Development Kit (SDK)

A particular instantiation of APISDK consists of libraries and tools Provides implementation of API

specification

Can have multiple SDKs for an APIExamples of SDKs MPICH, Motif Widgets

Page 51: Lecture slides / notes.

51

MORE DEFINTIONS: Syntax

Rules for encoding information, e.g. XML, Condor ClassAds, Globus RSL

Distinct from protocols One syntax may be used by many

protocols

Syntaxes may be layered E.g., Condor ClassAds -> XML->ASCII

Page 52: Lecture slides / notes.

52

References

Berman F., Fox G., Hey T. (2003) Grid Computing: Making the Global Infrastructure a Reality, Chichester, John Willey & Sons Inc. http://www.computing.surrey.ac.uk/courses/csm23/list.html

Page 53: Lecture slides / notes.

53

CSM23 Assessment and WeightingComponents of Assessment

Method(s) Percentage weighting

Annotated bibliography Students are required to write a 200 word summary of each of 5 key research papers

10%

Oral Examination 20%

Laboratory Exercise Students are required to implement small-scale laboratory homework during the semester.

20%

Project Students are expected to implement a Grid project ad write IEEE formatted report about their projects. In addition, the students are asked to give a presentation.

Implementation:20%IEEE Report:20%Presentation:10%

Page 54: Lecture slides / notes.

54

CSM23 TimetableDate Topic Lecturer

17/01/2005 Overview and Motivation Mrs.Tugba Taskaya-Temizel

24/01/2005 Grid Architecture, Technologies and Resource Allocation

Mrs.Tugba Taskaya-Temizel

31/01/2005 Peer to Peer Computing Dr Nick Antonopoulos

7/02/2005 Parallel Computing

Security

Dr.Roger M A Peel, Dr.James Heather

14/02/2005 Parallel Computing

Security

Dr.Roger M A Peel, Dr.James Heather

21/02/2005 Grid Applications Mrs.Tugba Taskaya-Temizel

28/02,7/03, 14/03, 21/03

Seminars Mrs.Tugba Taskaya-Temizel