Top Banner
Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy
36

Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Mar 26, 2015

Download

Documents

Amber Payne
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Zeuthen, 23.04.2013

Computing Seminar

Volker Gülzow, Kars Ohrenberg

Networking@desy

Page 2: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Network Topology Hamburg

Page 3: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Topology

Page 4: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Zeuthen inbound

Page 5: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Zeuthen outbound

Page 6: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.
Page 7: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite7

Bandwidth Evolution @ DFN

> DFN is upgrading the optical platform of the X-WiN■Contract awarded to ECI Telecom (http://www.ecitele.com)

■Migration work is currently underway

> High Bandwidth Capabilities■88 wave length per fiber

■Up to 100 Gbps per wave length■ thus 8.8 Tbps per fiber!

■1 Tbps Switching Fabric (aggregation of 10 Gbps lines on single 100 Gbps line)

Page 8: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite8

Growing WiN Capacities

Page 9: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite9

Bandwidth Evolution @ DFN

> Significant cheaper components for 1 Gbps and 10 Gbps components -> reduced cost for VPN connections, new DFN pricing

> New DFN conditions starting 1.7.2013■DESYs contract of 2 x 2 GBits will go to 2 x 5 Gbps without additional costs

■New cost model for Point-to-point VPNs

■1) Initial installation payment■ 10 GBps ~ 11.400 €, 40 GBps ~ 38.000 €, 100 GBps ~ 94.000 €

■2) Annual fee now depends on the distance■ Hamburg <> Berlin at ~ 20% of the current costs (for 10 Gbps)

■ Hamburg <> Berlin at ~ 80% of the current costs (for 40 Gbps)

■ Hamburg <> Karlsruhe at ~ 45% of the current costs (for 10 Gbps)

■ Hamburg <> Karlsruhe at ~ 150% of the current costs (for 40 Gbps)

Page 10: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite10

Page 11: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Geant3 topology

11

Page 12: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite12

Page 13: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Networking for LHC

13

Page 14: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite14

LHC Computing Infrastructure

> WLCG in brief:■1 Tier-0, 11 Tier-1s, ~ 140 Tier-2s, O(300) Tier-3s worldwide

Page 15: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite15

The LHC Optical Private Network

> The LHCOPN (from http://lhcopn.web.cern.ch)

■The LHCOPN is the private IP network that connects the Tier0 and the Tier1 sites of the LCG.

■The LHCOPN consists of any T0-T1 or T1-T1 link which is dedicated to the transport of WLCG traffic and whose utilization is restricted to the Tier0 and the Tier1s.

■Any other T0-T1 or T1-T1 link not dedicated to WLCG traffic may be part of the LHCOPN, assuming the exception is communicated to and agreed by the LHCOPN community

> Very closed and restricted access policy

> No Gateways

Page 16: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite16

LHCOPN Network Map

Page 17: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Data transfers

> Global transfer rates are always significant (12-15 Gb/s) – permanent on-going workloads

> CERN export rates driven (mostly) by LHC data export

> By Ian Bird, CRRB,4/13

April 16, 2013 [email protected] 17

CERN Tier 1s

Global transfers

Page 18: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Resource usage: Tier 0/1

April 16, 2013 [email protected] 18

By Ian Bird

Page 19: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Resource use vs pledge

CCRC F2F 10/01/2008 19

CERN

Tier 1s

Page 20: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Resource vs pledges: Tier 2

April 16, 2013 [email protected] 20

By Ian Bird

Page 21: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite

Connectivity (100 Gb/s)

April 16, 2013 [email protected] 21

Latency measured;No problems anticipated

By Ian Bird

Page 22: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite22

Computing Models Evolution

> The original MONARC model was strictly hierarchical

> Changes introduced gradually since 2010

> Main evolutions:■Meshed data flows: Any site can use any other

site as source of data

■Dynamic data caching: Analysis sites pull datasets from other sites „on demand“, including from Tier-2s in other regions

■Remote data access

> Variations by experiment

> LHCOPN only connects T0 and T1

Page 23: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite23

LHC Open Network Environment

> With the successful operation of the LHC accelerator and the start of the data analysis, there has come a re-evaluation of the computing and data models of the experiments

> The goal of LHCONE (LHC Open Network Environment) is to ensure better access to the most important datasets by the worldwide HEP community

> Traffic patterns have altered to the extent that substantial data transfers between major sites are regularly being observed on the General Purpose Networks (GPN)

> The main principle is to separate the LHC traffic from the GPN traffic, thus avoiding degraded performance

> The objective of LHCONE is to provide entry points into a network that is private to the LHC T1/2/3 sites.

> LHCONE is not intended to replace LHCOPN but rather to complement it

Page 24: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite24

LHCONE Achitecture

Page 25: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite25

LHCONE VRF Map (from Bill Johnston, ESNet)

Page 26: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite26

LHCONE: A global Infrastructure

Page 27: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite27

LHCONE Activities

> With the above in mind, LHCONE has defined the following activities:

> VRF-based multipoint service: a “quick-fix” to provide multipoint LHCONE connectivity, with logical separation from R&E GPN

> Layer 2 multipath: evaluate use of emerging standards such as TRILL (IETF) or Shortest Path Bridging (SPB, IEEE 802.1aq) in WAN environments

> Openflow: There was wide agreement that SDN is the most probable candidate technology for LHCONE in the long-term (but needs more investigations)

> Point-to-point dynamic circuits pilots

> Diagnostic Infrastructure: each site to have the ability to perform E2E performance tests with all other LHCONE sites

Page 28: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite28

Software-Defined Networking (SDN)

> Is a form of network virtualization in which the control plane is separated from the data plane and implemented in a software application

> This architecture allows network administrators to have programmable central control of network traffic without requiring physical access to the network's hardware devices

> SDN requires some method for the control plane to communicate with the data plane. One such mechanism is OpenFlow which is a standard interface for controlling computer networking switches

Page 29: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite29

LHCONE VRF

> Implementation of multiple logical router instances inside a physical device (virtualized Layer 3)

> Logical control plane separation between multiple clients

> VRF in LHCONE: regional networks implement VRF domains to logically separate LHCONE from other flows

> BGP peerings used inter-domain and to end-sites

Page 30: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite30

Multipath in LHCONE

> Multipath problem:■How to use the many (transatlantic) paths at Layer 2 among the many partners,

e.g. USLHCNet, GEANT, SURFnet, NORDUnet, ...

> Layer 3 (VRF) can use some BGP techniques■MED, AS padding, local preference, restricted announcements

■works in a reasonably small configuration, not clear it will scale up to O(100) end-sites

> Some approaches to Layer 2 mulitpath:■IETF: TRILL (TRansparent Interconnect of Lots of Links)

■IEEE: 802.1aq (Shortest Path Bridging)

> None of these L2 protocols is designed for WAN!■R&E needed

Page 31: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite31

LHCONE Routing Policies

> Only the networks which are announced to LHCONE are allowed to reach the LHCONE

■Networks announced by DESY:

■ 131.169.98.0/24, 131.169.160.0/21, 131.169.191.0/24 (Tier-2 Hamburg)

■ 141.34.192.0/21, 141.34.200.0/24 (Tier-2 Zeuthen)

■ 141.34.224.0/22, 141.34.228.0/24, 141.34.229.0/24, 141.34.230.0/24 (NAF Hamburg)

■ 141.34.216.0/23, 141.34.218.0/24, 141.34.219.0/24, 141.34.220.0/24 (NAF Zeuthen)

■e.g. networks announced by CERN:

■ 128.142.0.0/16 but not 137.138.0.0

> Only these networks will be reachable via the LHCONE

> Other traffic uses the public, general purpose networks

> Asymmetric routing should be avoided as this will cause problems for traffic passing (public) firewalls

Page 32: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite32

LHCONE - the current status

> Currently ~100 network prefixes■German sites currently participating in LHCONE

■ DESY, KIT, GSI, RWTH Aachen, Uni Wuppertal

■Europe■ CERN, SARA , GRIF (LAL + LPNHE), INFN, FZU, PIC, ...

■US:■ AGLT2 (MSU + UM), MWT2 (UC), BNL, ...

■Canada■ TRIUMF, Toronto, ...

■Asia■ ASGC, ICEPP, ...

> Detailed monitoring via perfSONAR

Page 33: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite33

LHCONE Monitoring

Page 34: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite34

R&D Network Trends

> Increased multiplicity of 10Gbps links in the major R&E networks: GEANT, Internet2, ESnet, various NREN, ...

> 100Gbps Backbones in place and transition now underway

■GEANT, DFN, ...

■CERN - Budapest 2 X 100G for LHC Remote Tier- 0 Center

> OpenFlow (Software-defined switching and routing) taken up by much of the network industry and R&E networks

Page 35: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite35

WAN + LHCONE Infrastructure at DESY

Page 36: Zeuthen, 23.04.2013 Computing Seminar Volker Gülzow, Kars Ohrenberg Networking@desy.

Gülzow/Ohrenberg, Zeuthen | 23.4.2013 | Seite36

Summary

> The LHC computing and data models continue to evolve towards more dynamic, less structured, on-demand data movement thus requiring different network structures

> LHCOPN and LHCONE may merge in the future

> With the evolution of the new optical platforms bandwidth will get more affordable