U.S. Department of Energy’s Office of Science

30
U.S. epartment of Energy’s Office f Science Mary Anne Scott Program Manager Advanced Scientific Computing Research Mathematical, Information, and Computational Research Division [email protected] Connecting Scientists Connecting Scientists to collaborators, to facilities, to datastores, to collaborators, to facilities, to datastores, to to computational resources, to…. computational resources, to…. Energy Sciences Network (ESnet) Energy Sciences Network (ESnet) November 3, 2004 November 3, 2004 BERAC Meeting BERAC Meeting

description

U.S. Department of Energy’s Office of Science. Connecting Scientists …to collaborators, to facilities, to datastores, to computational resources, to…. Energy Sciences Network (ESnet) November 3, 2004 BERAC Meeting. Mary Anne Scott Program Manager Advanced Scientific Computing Research - PowerPoint PPT Presentation

Transcript of U.S. Department of Energy’s Office of Science

Page 1: U.S. Department of Energy’s Office of Science

U.S. Department of Energy’s

Office of Science

Mary Anne ScottProgram ManagerAdvanced Scientific Computing ResearchMathematical, Information, and Computational Research [email protected]

Connecting ScientistsConnecting Scientists……to collaborators, to facilities, to datastores, to to collaborators, to facilities, to datastores, to

computational resources, to….computational resources, to….

Energy Sciences Network (ESnet)Energy Sciences Network (ESnet)

November 3, 2004November 3, 2004BERAC MeetingBERAC Meeting

Page 2: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Science Drivers for Networks

Modern, large-scale science is dependent on networks Data sharing Collaboration Distributed data processing Distributed simulation Data management

Page 3: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Science Mission Critical Infrastructure

ESnet is a visible and critical piece of DOE science infrastructure If ESnet fails, tens of thousands of DOE and

University users know it within minutes if not seconds

Requires high reliability and high operational security in the network operations

Page 4: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

What is ESnet?

Mission:• Provide, interoperable, effective and reliable communications infrastructure

and leading-edge network services that support missions of the Department of Energy, especially the Office of Science

Vision:• Provide seamless and ubiquitous access, via shared collaborative information

and computational environments, to the facilities, data, and colleagues needed to accomplish their goals.

Role: • A component of the Office of Science infrastructure critical to the success of

its research programs (funded through ASCR/MICS and managed and operated by ESnet staff at LBNL).

Essentially all of the national data traffic supporting US science is Essentially all of the national data traffic supporting US science is carried by two networks—ESnet and Internet-2/Abilene (which plays carried by two networks—ESnet and Internet-2/Abilene (which plays a similar role for the university community)a similar role for the university community)

Page 5: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

What is ESnet’s user base?

Between 10,000 and 100,000 researchers in the US (guesstimate) Mainly Office of Science programs—ASCR, BER, BES, FES, HEP,

NP Also traffic for NNSA and others All the US national labs Hundreds of universities Hundreds of foreign institutions

Characteristics of the user base Many casual users Users from science disciplines that span SC interests Users concerned with any data intensive and computationally

intensive tasks Collaborators distributed geographically, small to large groups

Page 6: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

What Does ESnet Provide?

ESnet builds a comprehensive IP network infrastructure Telecommunications services ranging from T1 (1 Mb/s) to OC192 SONET (10

Gb/s) and uses these to connect core routers and sites to form the ESnet IP network

Peering access to commercial networks

High bandwidth access to DOE’s primary science collaborators: Research and Education institutions in the US, Europe, Asia Pacific, and elsewhere

Comprehensive user support, including “owning” all trouble tickets involving ESnet users (including problems at the far end of an ESnet connection) until they are resolved – 24x7 coverage

Grid middleware and collaboration services supporting collaborative science trust, persistence, and science oriented policy

Page 7: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Why is ESnet important?

Enables thousands of DOE, university and industry scientists and collaborators worldwide to make effective use of unique DOE research facilities and computing resources independent of time and geographic location Direct connections to all major DOE sites Access to the global Internet (managing 150,000 routes at 10 commercial

peering points) User demand has grown by a factor of more than 10,000 since its inception

in the mid 1990’s—a 100 percent increase every year since 1990 Capabilities not available through commercial networks

Architected to move huge amounts of data between a small number of sites High bandwidth peering to provide access to US, European, Asia-Pacific, and

other research and education networks.

Objective: Support scientific researchSupport scientific research by providing seamless

and ubiquitous access to the facilities, data, and colleagues

Page 8: U.S. Department of Energy’s Office of Science

8

TWC

JGISNLL

LBNL

SLAC

YUCCA MT

BECHTEL

PNNLLIGO

INEEL

LANL

SNLAAlliedSignal

PANTEX

ARM

KCP

NOAA

OSTIORAU

SRS

ORNLJLAB

PPPL

ANL-DCINEEL-DCORAU-DC

LLNL/LANL-DC

MIT

ANL

BNL

FNALAMES

4xLAB-DCNERSC

NR

EL

ALBHUB

LLNL

GA DOE-ALB

SDSC

Japan

GTN&NNSA

International (high speed)OC192 (10G/s optical)OC48 (2.5 Gb/s optical)Gigabit Ethernet (1 Gb/s)OC12 ATM (622 Mb/s)OC12 OC3 (155 Mb/s)T3 (45 Mb/s)T1-T3T1 (1 Mb/s)

Office Of Science Sponsored (22)NNSA Sponsored (12)Joint Sponsored (3)

Other Sponsored (NSF LIGO, NOAA)Laboratory Sponsored (6)

QWESTATM

42 end user sites

ESnet IP

GEANT - Germany - France - Italy - UK - etc Sinet (Japan)Japan – Russia(BINP)

CA*net4CERNMRENNetherlandsRussiaStarTapTaiwan (ASCC)

CA*net4KDDI (Japan)FranceSwitzerlandTaiwan (TANet2)

AustraliaCA*net4Taiwan (TANet2)Singaren

ESnet core: Optical Ring and Hubs

ELP HUB

SNV HUB CHI HUB

NYC HUB

ATL HUB

DC HUB

peering points

MAE-ES

tarligh

tChi NAP

Fix-W

PAIX-W

MAE-W

NY-NAP

PAIX-E

Euqinix

PNW

G

SEA HUB

ESnet Connects DOE Facilities and Collaborators

hubs SNV HUB

Abi

lene A

bile

ne

Abilene

Ab

ilene

Page 9: U.S. Department of Energy’s Office of Science

9

STARLI

GH

T

MAE-E

NY-NAP

PAIX-E

GA

LB

NL

ESnet’s Peering InfrastructureConnects the DOE Community With its Collaborators

ESnet Peering (connections to other networks)

NYC HUBS

SEA HUB

Japan

SNV HUB

MAE-W

FIX

-W

PAIX-W 26 PEERS

CA*net4CERNMRENNetherlandsRussiaStarTapTaiwan (ASCC)

Abilene +7 Universities

22 PEERS

MAX GPOP

GEANT - Germany - France - Italy - UK - etc SInet (Japan)KEKJapan – Russia (BINP)

AustraliaCA*net4Taiwan

(TANet2)Singaren

20 PEERS3 PEERS

LANL

TECHnet

2 PEERS

39 PEERS

CENICSDSC

PNW-GPOP

CalREN2 CHI NAP

Distributed 6TAP19 Peers

2 PEERS

KDDI (Japan)France

EQX-ASH

1 PEER

1 PEER

5 PEERS

ESnet provides access to all of the Internet by managing the full complement of Global Internet routes (about 150,000) at 10 general/commercial peering points + high-speed peerings w/ Abilene and the international R&E networks. This is a lot of work, and is very visible, but provides full access for DOE.

ATL HUB

University

International

Commercial

Abilene

EQX-SJ

Abilene

6 PEERS

Abilene

Page 10: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

How is ESnet Managed?

A community endeavor Strategic guidance from the OSC programs

Energy Science Network Steering Committee (ESSC) BER representatives are T.P. Straatsma, PNNL; Raymond McCord, ORNL

Network operation is a shared activity with the community

ESnet Site Coordinators Committee

Ensures the right operational “sociology” for success

Complex and specialized – both in the network engineering and the network management – in order to provide its services to the laboratories in an integrated support environment

Extremely reliable in several dimensions

Taken together these points make ESnet a unique facility supporting DOE science that is quite different from a commercial ISP or University network

Page 11: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

What’s a possible future?

VISION – A seamless, high-performance network infrastructure in which

science applications and advanced facilities are "n-way" interconnected to terascale computing, petascale storage, and high-end visualization capabilities.

This advanced network facilitates collaborations among researchers and interactions between researchers and experimental and computational resources,

Science, especially large-scale science, moves to a new regime that eliminates isolation, discourages redundancy, and promotes rapid scientific progress through the interplay of theory, simulation, and experiment.

Page 12: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Network and Middleware Needs of DOE Science

Focused on science requirements driving advanced network infrastructure Middleware research Network research Network governance model

Requirements for DOE science developed for a representative cross-section of the OSC scientific community Climate simulation Spallation Neutron Source facility Macromolecular Crystallography High Energy Physics experiments Magnetic Fusion Energy Sciences Chemical Sciences Bioinformatics

Page 13: U.S. Department of Energy’s Office of Science

13

Evolving Quantitative Science Requirements for Networks

Science Areas Today End2End Throughput

5 years End2End Throughput

5-10 Years End2End Throughput

Remarks

High Energy Physics

0.5 Gb/s 100 Gb/s 1000 Gb/s high bulk throughput

Climate (Data & Computation)

0.5 Gb/s 160-200 Gb/s N x 1000 Gb/s high bulk throughput

SNS NanoScience Not yet started 1 Gb/s 1000 Gb/s + QoS for control channel

remote control and time critical throughput

Fusion Energy 0.066 Gb/s(500 MB/s burst)

0.198 Gb/s(500MB/20 sec. burst)

N x 1000 Gb/s time critical throughput

Astrophysics 0.013 Gb/s(1 TBy/week)

N*N multicast 1000 Gb/s computational steering and collaborations

Genomics Data & Computation

0.091 Gb/s(1 TBy/day)

100s of users 1000 Gb/s + QoS for control channel

high throughput and steering

Page 14: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

New Strategic Directions to Address Needs of DOE Science

Focused on what is needed to achieve the science driven network requirements of the previous workshop

THE #1 DRIVER for continuing advancements in networking and middleware – Petabyte scale experimental and simulation data systems will be increasing to exabyte scale data systems.

Bioinformatics, Climate, LHC, etc.

Computational systems that process or produce data continue to advance with Moore’s Law

…..

Organized by the ESSC Workshop Chair, Roy Whitney, JLAB Workshop Editors, Roy Whitney, JLAB; Larry Price,

ANL

Page 15: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Observed Drivers for ESnet Evolution

Are we seeing the predictions of two years ago come true?

Yes!

Page 16: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

…what now???

VISION - A scalable, secure, integrated network environmentnetwork environment for ultra-scale distributed science is being developed to make it possible to combine resources and expertise to address complex questions that no single institution could manage alone.

Network StrategyProduction network

Base TCP/IP services; +99.9% reliable

High-impact network Increments of 10 Gbps; switched lambdas (other solutions); 99% reliable

Research network Interfaces with production, high-impact and other research networks; start

electronic and advance towards optical switching; very flexible

Revisit governance model SC-wide coordination Advisory Committee involvement

Page 17: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Evolution of ESnet

Upgrading ESnet to accommodate the anticipated increase from the current 100%/yr traffic growth to 300%/yr over the next 5-10 years is priority number 7 out of 20 in DOE’s “Facilities for the Future of Science – A Twenty Year Outlook”

Based on the requirements, ESnet must address

I. Capable, scalable, and reliable production IP networking University and international collaborator connectivity Scalable, reliable, and high bandwidth site connectivity

II. Network support of high-impact science provisioned circuits with guaranteed quality of service

(e.g. dedicated bandwidth)

III. Evolution to optical switched networks Partnership with UltraScienceNet Close collaboration with the network R&D community

IV. Science Services to support Grids, collaboratories, etc

Page 18: U.S. Department of Energy’s Office of Science

18

New ESnet Architecture Needed to Accommodate OSC

• The essential DOE Office of Science requirements cannot be met with the current, telecom provided, hub and spoke architecture of ESnet

• The core ring has good capacity and resiliency against single point failures, but the point-to-point tail circuits are neither reliable nor scalable to the required bandwidth

ESnetCore

New York (AOA)

Chicago (CHI)

Sunnyvale (SNV)

Atlanta (ATL)

Washington, DC (DC)

El Paso (ELP)

DOE sites

Page 19: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

A New ESnet Architecture

Goals Fully redundant connectivity for every site High-speed access for every site (at least 20 Gb/s)

Three part strategy Metropolitan Area Network (MAN) rings provide dual site

connectivity and much higher site-to-core bandwidth A Science Data Network core for

multiply connected MAN rings for protection against hub failure expanded capacity for science data a platform for provisioned, guaranteed bandwidth circuits alternate path for production IP traffic carrier circuit and fiber access neutral hubs

A High-reliability IP core (e.g. the current ESnet core)

Page 20: U.S. Department of Energy’s Office of Science

20

Site gateway routersite equip. Site gateway router

Qwest hub

ESnet IP core

ANLFNAL

ESnet MAN ArchitectureCERN (DOE funded link)

monitor

ESnet management and

monitoring

ESnet managedλ / circuit servicestunneled through the IP backbone

monitor

site equip.

ESnet production IP service

ESnet managedλ / circuit services

T320StarLight

International peerings

Site LAN Site LAN

ESnet SDN core

T320

Page 21: U.S. Department of Energy’s Office of Science

GEANT (Europe)

Asia-Pacific

ESnetIP Core

New York(AOA)

Chicago (CHI)

Sunnyvale(SNV)

Washington, DC (DC)

El Paso (ELP)

DOE/OSC Labs

New hubs

Existing hubs

ESnetScience Data

Network(2nd Core)

A New ESnet Architecture:Science Data Network + IP Core

Possible new hubs

Atlanta (ATL)

MetropolitanAreaRings

CERN

Page 22: U.S. Department of Energy’s Office of Science

22

DENDEN

ELPELP

ALBALBATLATL

MANs

High-speed cross connects with Internet2/Abilene

Major DOE Office of Science Sites

Goals for ESnet – FY07

Japan

CERNEurope

SDGSDG

AsiaPacSEASEA

SDN – ESnet hubsQwest – ESnet hubs

SNVSNV

Europe

10Gb/s30Bg/s

40Gb/s

Japan

CHICHI

High-impact science coreLab suppliedMajor international

2.5 Gbs10 Gbs

Future phases

Production IP ESnet core

DCDC

Japan

NYCNYC

Page 23: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Conclusions

ESnet is an infrastructure that is critical to DOE’s science mission

Focused on the Office of Science Labs, but serves many other parts of DOE

ESnet is working hard to meet the current and future networking need of DOE mission science in several ways: Evolving a new high speed, high reliability, leveraged

architecture Championing several new initiatives which will keep ESnet’s

contributions relevant to the needs of our community Grid middleware services for large numbers of users are hard –

but they can be provided if careful attention is paid to scaling

Page 24: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Where do you come in?

Early identification of requirements Evolving programs New facilities

Interaction with BER representatives on ESSC

Participation in management activities

Page 25: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Additional Information

Page 26: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Planning Workshops

High Performance Network Planning Workshop, August 2002http://www.doecollaboratory.org/meetings/hpnpw

DOE Workshop on Ultra High-Speed Transport Protocols and Network Provisioning for Large-Scale Science Applications, April 2003

http://www.csm.ornl.gov/ghpn/wk2003 Science Case for Large Scale Simulation, June 2003

http://www.pnl.gov/scales/ DOE Science Networking Roadmap Meeting, June 2003

http://www.es.net/hypertext/welcome/pr/Roadmap/index.html Workshop on the Road Map for the Revitalization of High End Computing, June 2003

http://www.cra.org/Activities/workshops/nitrd

http://www.sc.doe.gov/ascr/20040510_hecrtf.pdf (public report)

ASCR Strategic Planning Workshop, July 2003http://www.fp-mcs.anl.gov/ascr-july03spw

Planning Workshops-Office of Science Data-Management Strategy, March & May 2004 http://www-conf.slac.stanford.edu/dmw2004 (report coming soon)

Page 27: U.S. Department of Energy’s Office of Science

27

ESnet is Architected and Engineered to Move a Lot of Data

Annual growth in the past five years has increased from 1.7x annually to just over 2.0x annually.

TB

ytes

/M

onth

ESnet is currently transporting about 250 terabytes/mo.(250,000,000 Mbytes/mo.)

ESnet Monthly Accepted Traffic

Page 28: U.S. Department of Energy’s Office of Science

28

Traffic coming into ESnet = GreenTraffic leaving ESnet = BlueTraffic between sites% = of total ingress or egress traffic

Note that more that 90% of the ESnet traffic is OSC traffic

ESnet Appropriate Use Policy (AUP)

All ESnet traffic must originate and/or terminate on an ESnet an site (no transit traffic is allowed)

Who Generates Traffic, and Where Does it Go?ESnet Inter-Sector Traffic Summary,

Jan 2003 / Feb 2004 (1.7X overall traffic increase, 1.9X OSC increase) (the international traffic is increasing due to BABAR at SLAC and the LHC tier 1 centers

at FNAL and BNL)

Commercial

R&E (mostlyuniversities)

International

21/14%

17/10%

9/26%

14/12%

10/13%

4/6%

ESnet

~25/18%

Peering Points

DOE collaborator traffic, inc.data

72/68%

53/49%

DOE is a net supplier of data because DOE facilities are used by universities and commercial entities, as well as by DOE researchers

DOE sites

Page 29: U.S. Department of Energy’s Office of Science

Office of Science

U.S. Department of Energy

Fermila

b (US)

CERN

SLAC (US)

IN2P3 (F

R)

1 T

erab

yte/

day

SLAC (US)

INFN P

adva (I

T)

Fermila

b (US)

U. C

hicago (U

S)

CEBAF (US)

IN2P3 (F

R)

INFN P

adva (I

T) S

LAC (US)

U. Toro

nto (CA)

Ferm

ilab (U

S)

Helmholtz

-Karls

ruhe (D

E) S

LAC (US)

DOE Lab D

OE Lab

DOE Lab D

OE Lab

SLAC (US)

JANET (U

K)

Fermila

b (US)

JANET (U

K)

Argonne (U

S) Leve

l3 (US)

Argonne

SURFnet (

NL)

IN2P3 (F

R) S

LAC (US)

Fermila

b (US)

INFN P

adva (I

T)

A small number of science users

account for a significant

fraction of all ESnet traffic

Since BaBar data analysis started, the top 20 ESnet flows have consistently accounted for > 50% of ESnet’s monthly total traffic (~130 of 250 TBy/mo)

ESnet Top 20 Data Flows, 24 hr. avg., 2004-04-20ESnet is Engineered to Move a Lot of Data

Page 30: U.S. Department of Energy’s Office of Science

FNAL (US)

IN2P3 (F

R)

2.2 Tera

bytes

SLAC (U

S) IN

FN Pad

ua (I

T)

5.9

Terab

ytes

U. Toro

nto (CA)

Ferm

ilab (U

S)

0.9 Tera

bytes

SLAC (US)

Helm

holtz-K

arlsru

he (DE)

0.9 Tera

bytes

SLAC (U

S) IN

2P3

(FR)

5.3

Terab

ytes

CERN F

NAL (US)

1.3 Tera

bytes

FNAL (US)

U. N

ijmegen (N

L)

1.0 Tera

bytes

FNAL (US)

Helm

holtz-K

arlsru

he (DE)

0.6 Tera

bytes

FNAL (US)

SDSC (U

S)

0.6 Tera

bytes

U. Wisc

. (US)

FNAL (U

S)

0.6 Tera

bytes

The traffic is not transient: Daily and weekly averages are about the same.SLAC is a prototype for what will happen when Climate, Fusion, SNS, Astrophysics, etc., start to ramp up the next generation science

ESnet Top 10 Data Flows, 1 week avg., 2004-07-01