Top Banner
Integrating Web 2.0 Technologies with Scientific Simulation Codes for Real-Time Collaboration Gabrielle Allen (LSU) , Frank Loeffler (LSU), Thomas Radke (AEI), Erik Schnetter (LSU), Edward Seidel (NSF/LSU) E Cluster Computing, New Orleans, September 2009
26
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Cluster Computing Web2 Sept2009

Integrating Web 2.0 Technologies with Scientific Simulation Codes

for Real-Time Collaboration

Gabrielle Allen (LSU), Frank Loeffler (LSU),

Thomas Radke (AEI), Erik Schnetter (LSU),

Edward Seidel (NSF/LSU)

IEEE Cluster Computing, New Orleans, September 2009

Page 2: Cluster Computing Web2 Sept2009

Gravitational Wave Physics

Observations

ModelsAnalysis & Insight

Petascale problems: Full 3D general relativistic models of binary systems, supernova, gamma-ray bursts

Page 3: Cluster Computing Web2 Sept2009

3

Understanding Gravity

1600 1970 1995 2000 2009Galileo Smarr LSU PRAC

Log(Data)

Data and Collaboration increasing

Page 4: Cluster Computing Web2 Sept2009

Cactus Framework• Component-based HPC framework:

– Freely available, modular, portable, manageable environment for collaboratively developing parallel, multi-dimensional simulation codes

• Enabling applications: – Numerical Relativity/Astrophysics, CFD, Coastal,

Reservoir Engineering, Quantum Gravity, …

• Finite difference, AMR, FE/FV, multipatch, …

• Cutting edge CS:– Grid computing, petascale,

accelerators, steering, remote viz, application driver

• Active user & developer communities:– 12 year pedigree, led from LSU. Over

$10M support : NSF, EU, DOD, DOE, NASA, Microsoft, MPG, LSU, NCSA.

Page 5: Cluster Computing Web2 Sept2009

5

Cactus Structure

Core “Flesh”

Plug-In “Thorns”(components)

driver

input/output

interpolation

SOR solver

coordinates

boundary conditions

black holes

equations of state

remote steering

wave evolvers

multigrid

parameters

grid variables

error handling

scheduling

extensible APIs

make system

ANSI CFortran/C/C++

Your Physics !!

ComputationalTools !!

Page 6: Cluster Computing Web2 Sept2009

Cactus Application Environment

Adaptive mesh refinement, parallel I/O, interaction, …

Flesh: APIs, information, orchestration

Domain specific shared infrastructure

Individual research groups

Page 7: Cluster Computing Web2 Sept2009

7

Typical Black Hole Simulations

At LSU …• 300 Cactus thorns• 10,000 potential parameters• 20 different supercomputers• 100-2000 cores• Days/weeks to run (checkpoint/restart)• GBs to TBs of data (HDF5, ASCII, jpeg)

Page 8: Cluster Computing Web2 Sept2009

8

Collaborative Technologies

• Technologies to share simulation-related information developed in our group from the early 1990s– Essential to support the scientific research

• Review historical evolution of these technologies

• Show how Web 2.0 provides new tools to enable old scenarios

Page 9: Cluster Computing Web2 Sept2009

9

Web-based Mail Lists

• Mosaic web browser (1993, NCSA)– Seidel’s group at NCSA worry about content– http://archive.ncsa.illinois.edu/Cyberia/NumRel/GravWaves.html (1995)

• Collaborative Cork Board (CoCoBoard) (Mid 90’s)– Researchers have web-based “project pages”– Could attach images!! (usually 1-D plots of results)– Used till late 90’s

• Currently– Project based private wikis: parameter/output files, figures– Organize material for weekly project conference calls– Cons: network to access/edit wiki, editing slow

Page 10: Cluster Computing Web2 Sept2009

10

CoCoBoard

Page 11: Cluster Computing Web2 Sept2009

11

Simulation Web Interfaces

• Thorn “httpd”– First collaborative tool fundamentally

integrated into Cactus– Werner Benger (1999), visiting NCSA from

Germany (7 hr time difference and email)– Used socket library developed for remote viz

(John Shalf & TIKSL project)

• Thorn “HTTPD” in standard toolkit (2000)– Simulation status, variables, timing, viewport,

output files, parameter steering, etc– Thorns can include their own web content

Page 12: Cluster Computing Web2 Sept2009

12

Issues• Authorization to web pages

(username/password in parameter file is insecure and awkward, newer version uses https and can also use X.509)

• Browsers can display images in certain formats, a Visualization thorn uses gnuplot to include e.g. performance with time, physical parameters

• Problem deploying on compute nodes where web server cannot be directly accessed (port forwarding, filewalls)

• How to find and track the simulations, publicize existence to a collaboration?

Page 13: Cluster Computing Web2 Sept2009

13

Cactus HTTPD Simulation Page

Page 14: Cluster Computing Web2 Sept2009

14

Cactus HTTPD Viewport

Page 15: Cluster Computing Web2 Sept2009

15

Simulation Reports and Email• Readable report automatically

generated for each simulation (computation and physics)– Prototyped 2001 but not used (?)

• How to collect reports in one place?• Mail Thorn (sendmail)

– Email reliable and fault tolerant (spool)– Supercomputers do not allow mail to be

sent from compute nodes.

Page 16: Cluster Computing Web2 Sept2009

16

GridLab Visualization Service

Brygg Ullmer (2004)

Page 17: Cluster Computing Web2 Sept2009

17

Announcing and Grid Portals• Collaborations need

reliable, live information about long running simulations.– NSF Astrophysics Simulation

Collaboratory (ASC), 1999– Grid Portal provided centralized,

collaborative interface to submit, monitor and archive simulations

– Java, JSP, Javascript with back-end data base, contributed to GridSphere design (GridLab)

– JavaCOG to submit jobs and basic monitoring. ASC Portal (2002)

Page 18: Cluster Computing Web2 Sept2009

18

Announcing Simulation Info• Publish (application provided)

simulation information• Thorn Announce, in prototype

Cactus Worm scenario (2001)– Message from Flesh/Thorn info– Transport: XML-RPC to remote

socket (portal)

• Issues– Job IDs– Security, mapping users – Cumbersome user set parameters

(portal location, visibility of job, notification needs)

Announcing to ASC Portal (2002)

Page 19: Cluster Computing Web2 Sept2009

19

Notification

• Portal notification service

• Portal users configure at portal, simulations configure in parameter file

• Email, SMS, Instant Message– Initial experiments

generated large telecom bills!

Cool and useful, but lots of work (FTE) to develop and modify portal service, difficult to configure.

Page 20: Cluster Computing Web2 Sept2009

20

Web 2.0 Technologies

• Use for collaborative, simulation-level messaging and information archiving– Reliable, persistent, well-documented,

user-configurable, cheap, well supported, good APIs

Page 21: Cluster Computing Web2 Sept2009

21

Twitter

• March 2006• Real-time short messaging system.

Users send and receive each others updates (tweets). Wide range of devices and rudimentary social networking.

• Receivers can filter messages they see and specify how they receive them

• Twitter API (e.g. post a new Twitter message from a user)

• Free

Page 22: Cluster Computing Web2 Sept2009

22

Thorn Twitter

• Uses libcurl• Cactus parameters

for twitter username/password

• Twitter API: statuses/update

• At LSU “numrel” group account

• Messages when simulation starts and at different stages

Page 23: Cluster Computing Web2 Sept2009

23

Flickr• 2004, image hosting website for digital photographs

(and now videos). Bought by Yahoo (2005).• Professional account ($25/yr) for unlimited use• Web service API for uploading and manipulating

images– Group images into Sets and Collections– Tags, title, description, metadata from EXIF headers

• Social networking: users can comment on images, flag them, order by popularity, etc. Public/Private/Friends/Family. Blogs.

• RSS field allows quick previewing.

Page 24: Cluster Computing Web2 Sept2009

24

Thorn Flickr

• Send images from running simulation

• Uses: flickcurl, libcurl, libxml2, openssl

• Authentication more complex (api key, shared secret)

• Thorn uploads images that are generated by Cactus (and known to I/O layer), e.g. IoJpeg

• Each simulation given its own Flickr set

Page 25: Cluster Computing Web2 Sept2009

25

Future Work

• Extend capabilities, production testing• Common authentication mechanism• Social networking model (individual/shared

accounts)• Development of common tags, more metadata etc• Storing videos (Flickr, YouTube, Vimeo)

– Advantage for scientists presenting

• Lots of other possibilities: DropBox to publish files across a collaboration, WordPress for simulation reports/blogs, FaceBook to replace grid portals and aggregate services, Cloud computing APIs for “grid” scenarios, …

Page 26: Cluster Computing Web2 Sept2009

26

Conclusions• Started as a fun project (undergrad)• Web 2.0 provide reliable delivery, storage,

access, and flexible collaborative features• Can use Web 2.0 to easily prototype new

interactive and collaborative scenarios (have really missed this)– Small groups and individuals can do this too!!

• Target standard of ease-of-use for cyberinfrastructure development

• For real use need unified authentication, clear policies on data, site versions