Date post: | 19-Dec-2015 |
Category: |
Documents |
View: | 213 times |
Download: | 0 times |
1
ESnet Network Requirements
ASCAC Networking Sub-committee MeetingApril 13, 2007
Eli Dart
ESnet Engineering Group Lawrence Berkeley National Laboratory
2
Overview
Requirements are primary drivers for ESnet – science focused
Sources of Requirements
Office of Science (SC) Program Managers
Direct gathering through interaction with science users of the networkExample case studies (updated 2005/2006)
Magnetic Fusion Large Hadron Collider (LHC) Climate Modeling Spallation Neutron Source
Observation of the network
Other requirements
Requirements aggregation
Convergence on a complete set of network requirements
3
Requirements from SC Program Managers
• SC Program Offices have determined that ESnet future priorities must address the requirements for:
– Large Hadron Collider (LHC), CERN
– Relativistic Heavy Ion Collider (RHIC), BNL, US
– Large-scale fusion (ITER), France
– High-speed connectivity to Asia-Pacific• Climate and Fusion
– Other priorities and guidance from SC will come from upcoming per-Program Office requirements workshops, beginning this summer
• Modern science infrastructure is too large to be housed at any one institution
– Structure of DOE science assumes the existence of a robust, high-bandwidth, feature-rich network fabric that interconnects scientists, instruments and facilities such that collaboration may flourish
4
Direct Gathering Through Interaction with Stakeholders
• SC selected a representative set of applications for the 2002 Workshop
• Case studies were created for each application at the Workshop in order to consistently characterize the requirements
• The requirements collected from the case studies form the foundation for the current ESnet4 architecture
– Bandwidth, Connectivity Scope / Footprint, Services
– We do not ask that our users become network experts in order to communicate their requirements to us
– We ask what tools the researchers need to conduct their science, synthesize the necessary networking capabilities, and pass that back to our constituents for evaluation
• Per-Program Office workshops continue this process
– Workshops established as a result of ESnet baseline Lehman Review
– Workshop survey process extended to ESnet sites via Site Coordinators
• ESnet has a much larger user base (~50k to 100k users) than a typical supercomputer center (~3k users) and so has a more diffuse relationship with individual users
– Requirements gathering focused on key Principal Investigators, Program Managers, Scientists, etc, rather than a broad survey of every computer user within DOE
– Laboratory CIOs and their designates also play a key role in requirements input
5
Case Studies For Requirements
• Advanced Scientific Computing Research (ASCR)
– NERSC
– NLCF
• Basic Energy Sciences
– Advanced Light Source• Macromolecular Crystallography
– Chemistry/Combustion
– Spallation Neutron Source
• Biological and Environmental
– Bioinformatics/Genomics
– Climate Science
• Fusion Energy Sciences
– Magnetic Fusion Energy/ITER
• High Energy Physics
– LHC
• Nuclear Physics
– RHIC
• There is a high level of correlation between network requirements for large and small scale science – the only difference is bandwidth
– Meeting the requirements of the large-scale stakeholders will cover the smaller ones, provided the required services set is the same
6
Case Studies Requirements Gathering
• For all the science cases the following were identified by examining the science environment
– Instruments and facilities• Location and use of facilities, instruments, computational
resources, etc.• Data movement and storage requirements
– Process of science• Collaborations• Network services requirements• Noteworthy patterns of use (e.g. duty cycle of instruments)
– Near-term needs (now to 12 months)
– 5 year needs (relatively concrete)
– 5-10 year needs (more uncertainty)
7
Example Case Study Summary Matrix: Fusion
Feature
Science Instruments and
Facilities Process of Science
Anticipated Requirements
TimeFrame Network
Network Services and Middleware
Near-term
Each experiment only gets a few days per year - high productivity is critical
Experiment episodes (“shots”) generate 2-3 Gbytes every 20 minutes, which has to be delivered to the remote analysis sites in two minutes in order to analyze before next shot
Highly collaborative experiment and analysis environment
Real-time data access and analysis for experiment steering (the more that you can analyze between shots the more effective you can make the next shot)
Shared visualization capabilities
PKI certificate authorities that enable strong authentication of the community members and the use of Grid security tools and services.
Directory services that can be used to provide the naming root and high-level (community-wide) indexing of shared, persistent data that transforms into community information and knowledge
Efficient means to sift through large data repositories to extract meaningful information from unstructured data.
5 years 10 Gbytes generated by experiment every 20 minutes (time between shots) to be delivered in two minutes
Gbyte subsets of much larger simulation datasets to be delivered in two minutes for comparison with experiment
Simulation data scattered across United States
Transparent security Global directory and naming
services needed to anchor all of the distributed metadata
Support for “smooth” collaboration in a high-stress environment
Real-time data analysis for experiment steering combined with simulation interaction = big productivity increase
Real-time visualization and interaction among collaborators across United States
Integrated simulation of the several distinct regions of the reactor will produce a much more realistic model of the fusion process
Network bandwidth and data analysis computing capacity guarantees (quality of service) for inter-shot data analysis
Gbits/sec for 20 seconds out of 20 minutes, guaranteed
5 to 10 remote sites involved for data analysis and visualization
Parallel network I/O between simulations, data archives, experiments, and visualization
High quality, 7x24 PKI identity authentication infrastructure
End-to-end quality of service and quality of service management
Secure/authenticated transport to ease access through firewalls
Reliable data transfer Transient and transparent data replication for
real-time reliability Support for human collaboration tools
5+ years Simulations generate 100s of Tbytes ITER – Tbyte per shot, PB per year
Real-time remote operation of the experiment
Comprehensive integrated simulation
Quality of service for network latency and reliability, and for co-scheduling computing resources
Management functions for network quality of service that provides the request and access mechanisms for the experiment run time, periodic traffic noted above.
• Considers instrument and facility requirements, the process of science drivers and resulting network requirements cross cut with timelines
8
Requirements from Instruments and Facilities
• This is the ‘hardware infrastructure’ of DOE science – types of requirements can be summarized as follows– Bandwidth: Quantity of data produced, requirements for timely movement– Connectivity: Geographic reach – location of instruments, facilities, and users
plus network infrastructure involved (e.g. ESnet, Internet2, GEANT) – Services: Guaranteed bandwidth, traffic isolation, etc.; IP multicast
• Data rates and volumes from facilities and instruments – bandwidth, connectivity, services– Large supercomputer centers (NERSC, NLCF)– Large-scale science instruments (e.g. LHC, RHIC)– Other computational and data resources (clusters, data archives, etc.)
• Some instruments have special characteristics that must be addressed (e.g. Fusion) – bandwidth, services
• Next generation of experiments and facilities, and upgrades to existing facilities – bandwidth, connectivity, services– Addition of facilities increases bandwidth requirements– Existing facilities generate more data as they are upgraded– Reach of collaboration expands over time– New capabilities require advanced services
9
Requirements from Examining the Process of Science (1)
• The geographic extent and size of the user base of scientific collaboration is continuously expanding– DOE US and international collaborators rely on ESnet to reach
DOE facilities– DOE Scientists rely on ESnet to reach non-DOE facilities
nationally and internationally (e.g. LHC, ITER)– In the general case, the structure of modern scientific
collaboration assumes the existence of a robust, high-performance network infrastructure interconnecting collaborators with each other and with the instruments and facilities they use
– Therefore, close collaboration with other networks is essential for end-to-end service deployment, diagnostic transparency, etc.
• Robustness and stability (network reliability) are critical– Large-scale investment in science facilities and experiments
makes network failure unacceptable when the experiments depend on the network
– Dependence on the network is the general case
10
Requirements from Examining the Process of Science (2)
• Science requires several advanced network services for different purposes
– Predictable latency, quality of service guarantees• Remote real-time instrument control• Computational steering• Interactive visualization
– Bandwidth guarantees and traffic isolation• Large data transfers (potentially using TCP-unfriendly protocols)• Network support for deadline scheduling of data transfers
• Science requires other services as well – for example
– Federated Trust / Grid PKI for collaboration and middleware• Grid Authentication credentials for DOE science (researchers, users,
scientists, etc.)• Federation of international Grid PKIs
– Collaborations services such as audio and video conferencing
11
Science Network Requirements Aggregation Summary
Science Drivers
Science Areas / Facilities
End2End Reliability
Connectivity 2006 End2End
Band width
2010 End2End
Band width
Traffic Characteristics
Network Services
Advanced Light Source
- • DOE sites
• US Universities
• Industry
1 TB/day
300 Mbps
5 TB/day
1.5 Gbps
• Bulk data
• Remote control
• Guaranteed bandwidth
• PKI / Grid
Bioinformatics - • DOE sites
• US Universities
625 Mbps
12.5 Gbps in
two years
250 Gbps • Bulk data
• Remote control
• Point-to-multipoint
• Guaranteed bandwidth
• High-speed multicast
Chemistry / Combustion
- • DOE sites
• US Universities
• Industry
- 10s of Gigabits
per second
• Bulk data • Guaranteed bandwidth
• PKI / Grid
Climate Science
- • DOE sites
• US Universities
• International
- 5 PB per year
5 Gbps
• Bulk data
• Remote control
• Guaranteed bandwidth
• PKI / Grid
High Energy Physics (LHC)
99.95+%
(Less than 4 hrs/year)
• US Tier1 (DOE)
• US Tier2 (Universities)
• International (Europe, Canada)
10 Gbps 60 to 80 Gbps
(30-40 Gbps per US Tier1)
• Bulk data
• Remote control
• Guaranteed bandwidth
• Traffic isolation
• PKI / Grid
12
Science Network Requirements Aggregation Summary
Science Drivers
Science Areas / Facilities
End2End Reliability
Connectivity 2006 End2End
Band width
2010 End2End
Band width
Traffic Characteristics
Network Services
Magnetic Fusion Energy
99.999%
(Impossible without full
redundancy)
• DOE sites
• US Universities
• Industry
200+ Mbps
1 Gbps • Bulk data
• Remote control
• Guaranteed bandwidth
• Guaranteed QoS
• Deadline scheduling
NERSC - • DOE sites
• US Universities
• Industry
• International
10 Gbps 20 to 40 Gbps
• Bulk data
• Remote control
• Guaranteed bandwidth
• Guaranteed QoS
• Deadline Scheduling
• PKI / Grid
NLCF - • DOE sites
• US Universities
• Industry
• International
Backbone Band width parity
Backbone band width parity
• Bulk data
Nuclear Physics (RHIC)
- • DOE sites
• US Universities
• International
12 Gbps 70 Gbps • Bulk data • Guaranteed bandwidth
• PKI / Grid
Spallation Neutron Source
High
(24x7 operation)
• DOE sites 640 Mbps 2 Gbps • Bulk data
13
Example Case Studies
• By way of example, four of the cases are discussed hereMagnetic fusion Large Hadron ColliderClimate ModelingSpallation Neutron Source
• Categorization of case study information: quantitative vs. qualitative– Quantitative requirements from instruments, facilities, etc.
• Bandwidth requirements
• Storage requirements
• Computational facilities
• Other ‘hardware infrastructure’
– Qualitative requirements from the science process• Bandwidth and service guarantees
• Usage patterns
15
Magnetic Fusion Requirements – Instruments and Facilities
• Three large experimental facilities in US (General Atomics, MIT, Princeton Plasma Physics Laboratory)
– 3 GB data set per pulse today, 10+ GB per pulse in 5 years
– 1 pulse every 20 minutes, 25-35 pulses per day
– Guaranteed bandwidth requirement: 200+ Mbps today, ~1 Gbps in 5 years (driven by science process)
• Computationally intensive theory/simulation component
– Simulation runs at supercomputer centers, post-simulation analysis at ~20 other sites
– Large data sets (1 TB+ in 3-5 years)
– 10’s of TB of data in distributed archives
• ITER
– Located in France
– Groundbreaking soon, production operations in 2015
– 1 TB of data per pulse, 1 pulse per hour
– Petabytes of simulation data per year
16
Magnetic Fusion Requirements – Process of Science (1)
• Experiments today
– Interaction between large groups of local and remote users and the instrument during experiments – highly collaborative
– Data from current pulse is analyzed to provide input parameters for next pulse
– Requires guaranteed network and computational throughput on short time scales
• Data transfer in 2 minutes
• Computational analysis in ~7 minutes
• Science analysis in ~10 minutes
• Experimental pulses are 20 minutes apart
• ~1 minute of slack – this amounts to 99.999% uptime requirement
– Network reliability is critical, since each experiment gets only a few days of instrument time per year
17
Magnetic Fusion Requirements – Process of Science (2)
• Simulation
– Large, geographically dispersed data sets, more so in the future
– New long-term initiative (Fusion Simulation Project, FSP) – integrated simulation suite
– FSP will increase the computational requirements significantly in the future, resulting in increased bandwidth needs between fusion users and the SC supercomputer centers
• Both experiments and simulations rely on middleware that uses ESnet’s federated trust services to support authentication
• ITER
– Scale will increase substantially
– Close collaboration with the Europeans is essential for DOE science
18
Magnetic Fusion – Network Requirements
• Experiments– Guaranteed bandwidth requirement: 200+ Mbps today, ~1 Gbps in 5
years (driven by science process)
– Reliability (99.999% uptime)
– Deadline scheduling
– Service guarantees for remote steering and visualization
• Simulation
– Bulk data movement (310 Mbps end2end to move 1 TB in 8 hours)
• Federated Trust / Grid PKI for authentication
• ITER– Large guaranteed bandwidth requirement (pulsed operation and
science process as today, much larger data sets)
– Large bulk data movement for simulation data (Petabytes per year)
20
LHC Requirements – Instruments and Facilities
• Large Hadron Collider at CERN– Networking requirements of two experiments have been characterized – CMS
and Atlas– Petabytes of data per year to be distributed
• LHC networking and data volume requirements are unique to date– First in a series of DOE science projects with requirements of unprecedented
scale– Driving ESnet’s near-term bandwidth and architecture requirements– These requirements are shared by other very-large-scale projects that are
coming on line soon (e.g. ITER)
• Tiered data distribution model– Tier0 center at CERN processes raw data into event data– Tier1 centers receive event data from CERN
• FNAL is CMS Tier1 center for US• BNL is Atlas Tier1 center for US• CERN to US Tier1 data rates: 10 Gbps by 2007, 30-40 Gbps by 2010/11
– Tier2 and Tier3 sites receive data from Tier1 centers• Tier2 and Tier3 sites are end user analysis facilities• Analysis results are sent back to Tier1 and Tier0 centers• Tier2 and Tier3 sites are largely universities in US and Europe
21
LHCNet Security Requirements
• Security for the LHC Tier0-Tier1 network is being defined by CERN in the context of the LHC Network Operations forum
• Security to be achieved by filtering packets at CERN and the Tier1 sites to enforce routing policy (only approved hosts may send traffic)
• In providing circuits for LHC, providers must make sure that these policies cannot be circumvented
22
LHC Requirements – Process of Science
• Strictly tiered data distribution model is only part of the picture– Some Tier2 scientists will require data not available from their local Tier1 center– This will generate additional traffic outside the strict tiered data distribution tree– CMS Tier2 sites will fetch data from all Tier1 centers in the general case
• CMS traffic patterns will depend on data locality, which is currently unclear
• Network reliability is critical for the LHC– Data rates are so large that buffering capacity is limited– If an outage is more than a few hours in duration, the analysis could fall
permanently behind• Analysis capability is already maximized – little extra headroom
• CMS/Atlas require DOE federated trust for credentials and federation with LCG
• Service guarantees will play a key role– Traffic isolation for unfriendly data transport protocols– Bandwidth guarantees for deadline scheduling
• Several unknowns will require ESnet to be nimble and flexible– Tier1 to Tier1,Tier2 to Tier1, and Tier2 to Tier0 data rates could add significant
additional requirements for international bandwidth– Bandwidth will need to be added once requirements are clarified– Drives architectural requirements for scalability, modularity
23
LHC Ongoing Requirements Gathering Process
• ESnet has been an active participant in the LHC network planning and operation
– Been an active participant in the LHC network operations working group since its creation
– Jointly organized the US CMS Tier2 networking requirements workshop with Internet2
– Participated in the US Atlas Tier2 networking requirements workshop
– Participated in all 5 US Tier3 networking workshops
24
LHC Requirements Identified To Date• 10 Gbps “light paths” from FNAL and BNL to CERN– CERN / USLHCnet will provide10 Gbps circuits to Starlight, to 32 AoA, NYC
(MAN LAN), and between Starlight and NYC– 10 Gbps each in near term, additional lambdas over time (3-4 lambdas each by
2010)
• BNL must communicate with TRIUMF in Vancouver – This is an example of Tier1 to Tier1 traffic – 1 Gbps in near term– Circuit is currently being built
• Additional bandwidth requirements between US Tier1s and European Tier2s– To be served by USLHCnet circuit between New York and Amsterdam
• Reliability– 99.95%+ uptime (small number of hours per year)– Secondary backup paths – SDN for the US and possibly GLIF (Global Lambda
Integrated Facility) for transatlantic links– Tertiary backup paths – virtual circuits through ESnet, Internet2, and GEANT
production networks
• Tier2 site connectivity– Characteristics TBD, and is the focus of the Tier2 workshops– At least 1 Gbps required (this is already known to be a significant underestimate
for large US Tier2 sites)– Many large Tier2 sites require direct connections to the Tier1 sites – this drives
bandwidth and Virtual Circuit deployment (e.g. UCSD)• Ability to add bandwidth as additional requirements are clarified
25
Identified US Tier2 Sites
• Atlas (BNL Clients)
– Boston University
– Harvard University
– Indiana University Bloomington
– Langston University
– University of Chicago
– University of New Mexico Alb.
– University of Oklahoma Norman
– University of Texas at Arlington
• Calibration site
– University of Michigan
• CMS (FNAL Clients)
– Caltech
– MIT
– Purdue University
– University of California San Diego
– University of Florida at Gainesville
– University of Nebraska at Lincoln
– University of Wisconsin at Madison
26
LHC Tier 0, 1, and 2 Connectivity Requirements Summary
Denver
Su
nn
yv
ale
LA
KC
Dallas
Albuq.
CE
RN
-1G
ÉA
NT
-1G
ÉA
NT
-2C
ER
N-2
Tier 1 Centers
ESnet IP core hubs
ESnet SDN/NLR hubs
Cross connects with Internet2
CE
RN
-3
Internet2/GigaPoP nodes
USLHC nodes
ESnetSDN
Internet2 / Gigapop Footprint
Seattle
FNAL (CMS T1)
BNL (Atlas T1)
New York
Wash DC
Jacksonville
Boise
San DiegoAtlanta
Vancouver
Toronto
Tier 2 Sites
Chicago
ESnetIP Core
TRIUMF (Atlas T1, Canada)
CANARIE
GÉANT
USLHCNet
Virtual Circuits
• Direct connectivity T0-T1-T2
• USLHCNet to ESnet to Internet2
• Backup connectivity
• SDN, GLIF, VCs
27
LHC ATLAS Bandwidth Matrix as of April 2007
Site A Site Z ESnet A ESnet Z A-Z 2007 Bandwidth
A-Z 2010 Bandwidth
CERN BNL AofA (NYC) BNL 10Gbps 20-40Gbps
BNL U. of Michigan (Calibration)
BNL (LIMAN) Starlight (CHIMAN)
3Gbps 10Gbps
BNL Boston University
BNL (LIMAN)
Internet2 / NLR Peerings
3Gbps
(Northeastern Tier2 Center)
10Gbps
(Northeastern Tier2 Center)
BNL Harvard University
BNL Indiana U. at Bloomington
BNL (LIMAN)
Internet2 / NLR Peerings
3Gbps
(Midwestern Tier2 Center)
10Gbps
(Midwestern Tier2 Center)BNL U. of Chicago
BNL Langston University
BNL (LIMAN) Internet2 / NLR Peerings
3Gbps
(Southwestern Tier2 Center)
10Gbps
(Southwestern Tier2 Center)
BNL U. Oklahoma Norman
BNL U. of Texas Arlington
BNL Tier3 Aggregate BNL (LIMAN) Internet2 / NLR Peerings
5Gbps 20Gbps
BNL TRIUMF (Canadian ATLAS Tier1)
BNL (LIMAN) Seattle 1Gbps 5Gbps
28
LHC CMS Bandwidth Matrix as of April 2007
Site A Site Z ESnet A ESnet Z A-Z 2007 Bandwidth
A-Z 2010 Bandwidth
CERN FNAL Starlight (CHIMAN)
FNAL (CHIMAN)
10Gbps 20-40Gbps
FNAL U. of Michigan (Calibration)
FNAL (CHIMAN)
Starlight (CHIMAN)
3Gbps 10Gbps
FNAL Caltech FNAL (CHIMAN)
Starlight (CHIMAN)
3Gbps 10Gbps
FNAL MIT FNAL (CHIMAN)
AofA (NYC)/ Boston
3Gbps 10Gbps
FNAL Purdue University FNAL (CHIMAN)
Starlight (CHIMAN)
3Gbps 10Gbps
FNAL U. of California at San Diego
FNAL (CHIMAN)
San Diego 3Gbps 10Gbps
FNAL U. of Florida at Gainesville
FNAL (CHIMAN)
SOX 3Gbps 10Gbps
FNAL U. of Nebraska at Lincoln
FNAL (CHIMAN)
Starlight (CHIMAN)
3Gbps 10Gbps
FNAL U. of Wisconsin at Madison
FNAL (CHIMAN)
Starlight (CHIMAN)
3Gbps 10Gbps
FNAL Tier3 Aggregate FNAL (CHIMAN)
Internet2 / NLR Peerings
5Gbps 20Gbps
29
Estimated Aggregate Link Loadings, 2007-08
Denver
Seattle
Su
nn
yv
ale
LA
San Diego
Chicago
Jacksonville
KC
El Paso
Albuq.Tulsa
Clev.
Boise
Wash DC
Salt Lake City
Portland
BatonRougeHouston
Pitts.
NYC
Boston
Philadelphia
Indianapolis
Atlanta
Nashville
Existing site supplied circuits
ESnet IP core (1)ESnet Science Data Network coreESnet SDN core, NLR linksLab supplied linkLHC related linkMAN linkInternational IP Connections
Raleigh
OC48
(1)(1(3))
Layer 1 optical nodes at eventual ESnet Points of Presence
ESnet IP switch only hubs
ESnet IP switch/router hubs
ESnet SDN switch hubs
Layer 1 optical nodes not currently in ESnet plans
Lab site
13
12.5
8.5
9
13
2.5 Committed bandwidth, Gb/s
6
9
6
2.52.5
2.5
unlabeled links are 10 Gb/s
30
Layer 1 optical nodes at eventual ESnet Points of Presence
ESnet IP switch only hubs
ESnet IP switch/router hubs
ESnet SDN switch hubs
Layer 1 optical nodes not currently in ESnet plans
Lab site
ESnet IP coreESnet Science Data Network coreESnet SDN core, NLR links (existing)Lab supplied linkLHC related linkMAN linkInternational IP Connections
ESnet4 2007-8 Estimated Bandwidth Commitments
Denver
Seattle
Su
nn
yv
ale
LA
San Diego
Chicago
Raleigh
Jacksonville
KC
El Paso
Albuq.Tulsa
Clev.
Boise
Wash DC
Salt Lake City
Portland
BatonRougeHouston
Pitts.
NYC
Boston
Philadelphia
Indianapolis
Atlanta
Nashville
All circuits are 10Gb/s.
MAX
West Chicago MAN Long Island MAN
Newport News - Elite
San FranciscoBay Area MAN
LBNL
SLAC
JGI
LLNL
SNLL
NERSC
JLab
ELITE
ODU
MATP
Wash., DC
OC48(1(3))
(7)
(17)
(19) (20)
(22)(23)
(29)
(28)
(8)
(16)
(32)
(2)
(4)
(5)
(6)
(9)
(11)
(13) (25)
(26)
(10)
(12)
(3)
(21)
(27)
(14)
(24)
(15)
(0)
(1)
(30)
FNAL
600 W. Chicago
Starlight
ANL
USLHCNet
CERN
10
29(total)
2.5 Committed bandwidth, Gb/s
BNL
32 AoA, NYC
USLHCNet
CERN
10
13
5
unlabeled links are 10 Gb/s
31
Estimated Aggregate Link Loadings, 2010-11
Denver
Seattle
Su
nn
yv
ale
LA
San Diego
Chicago
Raleigh
Jacksonville
KC
El Paso
Albuq.Tulsa
Clev.
Boise
Wash. DC
Salt Lake City
Portland
BatonRougeHouston
Pitts. NYC
Boston
Philadelphia
Indianapolis
(>1 )
Atlanta
Nashville
Layer 1 optical nodes at eventual ESnet Points of Presence
ESnet IP switch only hubs
ESnet IP switch/router hubs
ESnet SDN switch hubs
Layer 1 optical nodes not currently in ESnet plans
Lab site
OC48
ESnet IP core (1)ESnet Science Data Network coreESnet SDN core, NLR links (existing)Lab supplied linkLHC related linkMAN linkInternational IP Connections
50
40
40
40
40 40
4
50
5
50
30
4040
50
50
50
50
5040(16)
30
30
5045
30
15
20
2020 5
20
5
5
20
10
2.5 Committed bandwidth, Gb/s link capacity, Gb/s40
unlabeled links are 10 Gb/slabeled links are in Gb/s
32
ESnet4 2010-11 Estimated Bandwidth Commitments
Denver
Seattle
Su
nn
yv
ale
LA
San Diego
Chicago
Raleigh
Jacksonville
KC
El Paso
Albuq.Tulsa
Clev.
Boise
Wash. DC
Salt Lake City
Portland
BatonRougeHouston
Pitts. NYC
Boston
Philadelphia
Indianapolis
(>1 )
Atlanta
Nashville
Layer 1 optical nodes at eventual ESnet Points of Presence
ESnet IP switch only hubs
ESnet IP switch/router hubs
ESnet SDN switch hubs
Layer 1 optical nodes not currently in ESnet plans
Lab site
OC48
(0)
(1)
ESnet IP core (1)ESnet Science Data Network coreESnet SDN core, NLR links (existing)Lab supplied linkLHC related linkMAN linkInternational IP ConnectionsInternet2 circuit number(20)
5
4
4
4
4 4
4
5
5
5
3
44
5
5
5
5
54
(7)
(17)
(19) (20)
(22)(23)
(29)
(28)
(8)
(16)
(32)
(2)
(4)
(5)
(6)
(9)
(11)
(13) (25)
(26)
(10)
(12)
(27)
(14)
(24)
(15)
(30)3
3(3)
(21)
2520
25
15
10
2020 5
10
5
5
80
FNAL
600 W. Chicago
Starlight
ANL
USLHCNet
CERN
40
808080100
BNL
32 AoA, NYC
USLHCNet
CERN
65
40
unlabeled links are 10 Gb/s
2.5 Committed bandwidth, Gb/s
34
Climate Modeling Requirements – Instruments and Facilities
• Climate Science is a large consumer of supercomputer time
• Data produced in direct proportion to CPU allocation– As supercomputers increase in capability and models become
more advanced, model resolution improves
– As model resolution improves, data sets increase in size
– CPU allocation may increase due to increased interest from policymakers
– Significant data set growth is likely in the next 5 years, with corresponding increase in network bandwidth requirement for data movement (current data volume is ~200TB, 1.5PB/year expected rate by 2010)
• Primary data repositories co-located with compute resources– Secondary analysis is often geographically distant from data
repositories, requiring data movement
35
Climate Modeling Requirements – Process of Science
• Climate models are run many times– Analysis improved model analysis is typical cycle
– Repeated runs of models are required to generate sufficient data for analysis and model improvement
• Current analysis is done by transferring model output data sets to scientist’s home institution for local study– Recent trend is to make data from many models widely available
– Less efficient use of network bandwidth, but huge scientific win• PCMDI (Program for Climate Model Diagnosis and Intercomparison)
generated 200 papers in a year
• Wide sharing of data expected to continue
• PCMDI paradigm of wide sharing from central locations will require significant bandwidth and excellent connectivity at those locations
– If trend of sharing data continues, more data repositories will be opened, requiring more bandwidth resources
36
Climate Modeling Requirements
• Data movement
– Large data sets must be moved to remote analysis resources
– Central repositories collect and distribute large data volumes
• Hundreds of Terabytes today
• Petabytes by 2010
• Analysis cycle
– Steady growth in network usage as models improve
• Increased use of supercomputer resources
– As computational systems increase in capability, data set sizes increase
– Increased demand from policymakers may result in increased data production
38
SNS Requirements – Instruments and Facilities
• SNS is latest instrument for Neutron Science
– Most intense pulsed neutron beams available for research
– Wide applicability to materials science, medicine, etc
– Users from DOE, Industry, Academia
– In process of coming into full production (full-power Accelerator Readiness Review imminent as of April 2007)
• SNS detectors produce 160GB/day of data in production
– Operation schedule results in about 50TB/year
– Network requirements are 640Mbps peak
– This will increase to 10Gbps peak within 5 years
• Neutron science data repository is being considered
39
SNS Requirements – Process of Science
• Productivity is critical
– Scientists are expected to get just a few days per year of instrument time
• Drives requirement for reliability
– Real-time analysis used to tune experiment in progress
• Linkage with remote computational resources
• 2Gbps network load for real-time remote visualization
• Most analysis of instrument data is expected to be done using remote computational resources
– Data movement is necessary
– Workflow management software (possibly based on Grid tools) will be necessary
• There is interest from the SNS community in ESnet’s Federated Trust services for Grid applications
40
SNS Requirements
• Bandwidth
– 2Gbps today
– 10Gbps in 5 years
• Reliability
– Instrument time is a scarce resource
– Real-time instrument interaction
• Data movement
– Workflow tools
– Potential neutron science data repository
• Federated Trust
– User management
– Workflow tools
41
Aggregation of Requirements from All Case Studies
• Analysis of diverse programs and facilities yields dramatic convergence on a well-defined set of requirements– Reliability
• Fusion – 1 minute of slack during an experiment (99.999%)• LHC – Small number of hours (99.95+%)• SNS – limited instrument time makes outages unacceptable• Drives requirement for redundancy, both in site connectivity and within ESnet
– Connectivity• Geographic reach equivalent to that of scientific collaboration• Multiple peerings to add reliability and bandwidth to interdomain connectivity• Critical both within the US and internationally
– Bandwidth• 10 Gbps site to site connectivity today• 100 Gbps backbone by 2010• Multiple 10 Gbps R&E peerings• Ability to easily deploy additional 10 Gbps lambdas and peerings• Per-lambda bandwidth of 40 Gbps or 100 Gbps should be available by 2010
– Bandwidth and service guarantees• All R&E networks must interoperate as one seamless fabric to enable end2end
service deployment• Flexible rate bandwidth guarantees
– Collaboration support (federated trust, PKI, AV conferencing, etc.)
42
Additional Bandwidth Requirements Matrix – April 2007
Site A Site Z ESnet A ESnet Z A-Z 2007 Bandwidth
A-Z 2010 Bandwidth
ANL (ALCF)
ORNL ANL (CHIMAN) ORNL (Atlanta and Chicago)
10Gbps (2008) 20Gbps
ANL (ALCF)
NERSC ANL (CHIMAN) NERSC (BAMAN) 10Gbps (2008) 20Gbps
BNL (RHIC)
CC-J, RIKEN, Japan
BNL (LIMAN) NYC (MANLAN) 1Gbps 3Gbps
• Argonne Leadership Computing Facility requirement is for large-scale distributed filesystem linking ANL, NERSC and ORNL supercomputer centers
• BNL to RIKEN traffic is a subset of total RHIC requirements, and is subject to revision as the impact of RHIC detector upgrades becomes clearer
43
Requirements Derivation from Network Observation
• ESnet observes several aspects of network traffic on an ongoing basis
– Load
• Network traffic load continues to grow exponentially
– Flow endpoints
• Network flow analysis shows a clear trend toward the dominance of large-scale science traffic and wide collaboration
– Traffic patterns
• Traffic pattern analysis indicates a trend toward circuit-like behaviors in science flows
44
Ter
abyt
es /
mo
nth
ESnet Monthly Accepted Traffic, January, 2000 – June, 2006•ESnet is currently transporting more than1 petabyte (1000 terabytes) per month•More than 50% of the traffic is now generated by the top 100 sites — large-scale science dominates all ESnet traffic
top 100 sites to siteworkflows
Network Observation – Bandwidth
45
R2 = 0.9898
0.0
0.1
1.0
10.0
100.0
1000.0
10000.0
Jan
, 9
0
Jan
, 9
1
Jan
, 9
2
Jan
, 9
3
Jan
, 9
4
Jan
, 9
5
Jan
, 9
6
Jan
, 9
7
Jan
, 9
8
Jan
, 9
9
Jan
, 0
0
Jan
, 0
1
Jan
, 0
2
Jan
, 0
3
Jan
, 0
4
Jan
, 0
5
Jan
, 0
6
ESnet Traffic has Increased by10X Every 47 Months, on Average, Since 1990
Ter
abyt
es /
mo
nth
Log Plot of ESnet Monthly Accepted Traffic, January, 1990 – June, 2006
Oct., 19931 TBy/mo.
Aug., 1990100 MBy/mo.
Jul., 199810 TBy/mo.
38 months
57 months
40 months
Nov., 2001100 TBy/mo.
Apr., 20061 PBy/mo.
53 months
46
Requirements from Network Utilization Observation
• In 4 years, we can expect a 10x increase in traffic over current levels without the addition of production LHC traffic
– Nominal average load on busiest backbone links is greater than 1 Gbps today
– In 4 years that figure will be over 10 Gbps if current trends continue
• Measurements of this kind are science-agnostic
– It doesn’t matter who the users are, the traffic load is increasing exponentially
• Bandwidth trends drive requirement for a new network architecture
– New ESnet4 architecture designed with these drivers in mind
47
Requirements from Traffic Flow Observations
• Most ESnet science traffic has a source or sink outside of ESnet– Drives requirement for high-bandwidth peering– Reliability and bandwidth requirements demand that peering be
redundant– Multiple 10 Gbps peerings today, must be able to add more flexibly and
cost-effectively
• Bandwidth and service guarantees must traverse R&E peerings– “Seamless fabric”– Collaboration with other R&E networks on a common framework is
critical
• Large-scale science is becoming the dominant user of the network– Satisfying the demands of large-scale science traffic into the future will
require a purpose-built, scalable architecture– Traffic patterns are different than commodity Internet
• Since large-scale science will be the dominant user going forward, the network should be architected to serve large-scale science
48
Aggregation of Requirements from Network Observation
• Traffic load continues to increase exponentially– 15-year trend indicates an increase of 10x in next 4 years– This means backbone traffic load will exceed 10 Gbps within 4
years requiring increased backbone bandwidth– Need new architecture – ESnet4
• Large science flows typically cross network administrative boundaries, and are beginning to dominate– Requirements such as bandwidth capacity, reliability, etc. apply
to peerings as well as ESnet itself– Large-scale science is becoming the dominant network user
49
Other Networking Requirements
• Production ISP Service for Lab Operations
– Captured in workshops, and in discussions with SLCCC (Lab CIOs)
– Drivers are an enhanced set of standard business networking requirements
– Traditional ISP service, plus enhancements (e.g. multicast)
– Reliable, cost-effective networking for business, technical, and research operations
• Collaboration tools for DOE science community
– Audio conferencing
– Video conferencing
50
Required Network Services Suite for DOE Science
• We have collected requirements from diverse science programs, program offices, and network analysis – the following summarizes the requirements:– Reliability
• 99.95% to 99.999% reliability• Redundancy is the only way to meet the reliability requirements
– Redundancy within ESnet– Redundant peerings– Redundant site connections where needed
– Connectivity• Geographic reach equivalent to that of scientific collaboration• Multiple peerings to add reliability and bandwidth to interdomain connectivity• Critical both within the US and internationally
– Bandwidth• 10 Gbps site to site connectivity today• 100 Gbps backbone by 2010• Multiple 10+ Gbps R&E peerings• Ability to easily deploy additional lambdas and peerings
– Service guarantees• All R&E networks must interoperate as one seamless fabric to enable end2end service deployment• Guaranteed bandwidth, traffic isolation, quality of service• Flexible rate bandwidth guarantees
– Collaboration support• Federated trust, PKI (Grid, middleware)• Audio and Video conferencing
– Production ISP service