Regional and Campus GENI• Many projects with different needshttp://groups.geni.net/geni/wiki/SpiralTwoand /GEC8DemoSummary• All projects eventually enable experiments
between researchers and infrastructure spanning multiple locations (usually campuses)
• Most separate control, data, instrumentation planes
• Notable exceptions like Million Node GENI (may not be exceptions in the long run)
http://groups.geni.net/geni/wiki/MillionNodeGENI
• Network protocol experimentation makes layer 2 connections important for many projects
• GENI campus connections in Spiral2 mostly 1Gbps, some 10Gbps, but will grow quickly.
• About 75% of Quilt members already support GENI projects. Role of regionals expanding (e.g. GENI racks)
Nationwide Meso-scale PrototypeCurrent plans for locations & equipment
WiMAX
ShadowNet
Salt Lake CityKansas City
Washington, DCAtlanta
StanfordUCLAUC BoulderWisconsinRutgersNYU PolytechUMassColumbia
OpenFlowBackbonesSeattleSalt Lake CitySunnyvaleDenverNew York CityHoustonChicagoLos AngelesAtlanta
OpenFlowStanford
U WashingtonWisconsin U
Indiana URutgers
PrincetonClemson
Georgia Tech
Arista 7124S Switch
Toroki LightSwitch 4810
HP ProCurve 5400 SwitchJuniper MX240 Ethernet
Services Router NEC IP8800 Ethernet SwitchNEC WiMAX Base Station
Meso-Scale in the backbones • Internet2 and NLR installing 5 OpenFlow switches in
each backbone (HP ProCurve and possibly a second vendor)
• NLR 3-node backbone live for GEC8 demo
• ProtoGENI installing additional nodes (2 or more) in Internet2 and integrating prototype OpenFlow
• GENI layer 2 data planes starting to appear across backbones, regionals and campuses
– ION, FrameNet and custom network engineering "stitching" VLANs
– ORCA, IGENI demonstrating dynamic VLANs at GECs
– NLR and I2 10Gpbs layer2 interconnect at Atlanta
• ShadowNet: installing 3 Juniper M7i routers for measurements in I2 PoPs this year
• Find updates through GENI wiki Spiral Two page (http://groups.geni.net/geni/wiki/SpiralTwo)
Internet2
MX-960 conns
NLR
dwgs courtesy I2, NLR
MesoScale in the Regionals• Plain old IP
• Tunneling above layer 2 (GRE, OpenVPN)
• Layer 2
– Single VLAN (801.q tagged)
– VLAN translation (Supported in ION, not Sherpa/FrameNet)
– Layer 2 tunneling (MPLS, QinQ) (DRAGON, OpenFlow)
– Direct fiber to NLR or I2 backbone (ION, FrameNet)
– Access via FrameNet/ION procedures (duration, bw limits,first-come first-served)
• More info (how-tos, track connections, references)
http://groups.geni.net/geni/wiki/ConnectivityHome
site BBN snippet
MesoScale in the Regionals (cont)
• International
– GpENI
– iGENI (Starlight)
– K-GENI (Korea/Indiana only)
– ORBIT (Australia only)
– PrimoGENI (Brazil)
– Others brewing
• Mobile/Wireless (over L2 or IP at regional)
– WiMAx
– OpenFlow WiFi (Openroads)
– Offloading/Application Migration (e.g. Android phones) University of Wisconsin
– Million Node GENI dwgs courtesy iGENI
DRAGON
dwgs courtesy iGENI
•MAX, GpENI, TIED projects (may be others)•Dynamic resource allocation via GMPLS-enabled control plane•MAX supports sliced edge compute resources (myPLC) and network virtualization with a GENI Aggregate Manager•Layer 2 Ethernet switches with dedicated 1 GbE and 10GbE connections•Layer 3 IP (management, control and general connectivity)• Programmable network hardware — NetFPGA card installed in rack mounted PC o acts as OpenFlow switch, programmable router, packet generator, etc. o offers advanced measurement capabilities
ProtoGENI and ORCA
• ProtoGENI backbone (5 locations in I2)
– Current nodes on 10 GE dedicated connections to Infinera GENI wave
– Moving to 1 GE ION links between nodes in August (stay tuned)
– Campuses access via IP, ION or dedicated layer2 (through regional)
• ORCA clearinghouse uses NLR backbone in Spiral 2
– Campuses connect via FrameNet, iGENI, and regional optical networks
– Dynamic VLAN allocation
– ORCA broker plans to eventually support many mixed connection types
dwgs courtesy iGENI
Clemson OpenFlow Example
Multiple layer 2 & 3 connectionsVLAN 3711 connects BBN and Clemson via NLR, SoX and NoX.