SlideShare una empresa de Scribd logo
1 de 65
Descargar para leer sin conexión
National Science Foundation Award #1826994
Jason Zurawski
zurawski@es.net
ESnet / Lawrence Berkeley National Laboratory
http://bit.ly/2vx2HiQ
Maximizing Performance and Network
Utility with a Science DMZ
Globusworld 2019
Chicago, IL
May 2nd, 2019
https://epoc.global
Outline
• Motivation
• TCP Basics
• DMZ Design Patterns
• Integration w/ Science
• Conclusions
2 – ESnet Science Engagement (engage@es.net) - 5/3/19
The Science DMZ in 1 Slide
Consists of four key components, all required:
• “Friction free” network path
• Highly capable network devices (wire-speed, deep queues)
• Virtual circuit connectivity option
• Security policy and enforcement specific to science workflows
• Located at or near site perimeter if possible
• Dedicated, high-performance Data Transfer Nodes (DTNs)
• Hardware, operating system, libraries all optimized for transfer
• Includes optimized data transfer tools such as Globus and GridFTP
• Performance measurement/test node
• perfSONAR
• Once it’s up, users often need training – Engagement is key
Details at http://fasterdata.es.net/science-dmz/
3 – ESnet Science Engagement (engage@es.net) - 5/3/19
© 2013 Wikipedia
© 2013 Globus
Michael Sinatra (ESnet)
The Science DMZ has two main purposes:
1.A security architecture which allows for better segmentation of risks, and
more granular application of controls to those segmented risks. In the case
of the Science DMZ, the goal is to limit the segmented risk profile to that of
a dedicated data transfer host.
2.An attempt to streamline tuning and troubleshooting by removing
degrees-of-freedom (from both a reliability and performance perspective)
from the active data transfer path.
4 – ESnet Science Engagement (engage@es.net) - 5/3/19
The Science DMZ in 1 Picture
5 – ESnet Science Engagement (engage@es.net) - 5/3/19
Outline
• Motivation
• TCP Basics
• DMZ Design Patterns
• Integration w/ Science
• Conclusions
6 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science DMZ Background
• The data mobility performance requirements for data intensive science are beyond
what can typically be achieved using traditional methods
• Default host configurations (TCP, filesystems, NICs)
• Converged network architectures designed for commodity traffic
• Conventional security tools and policies
• Legacy data transfer tools (e.g. SCP)
• Wait-for-trouble-ticket operational models for network performance
7 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science DMZ Background
• The Science DMZ model describes a performance-based approach
• Dedicated infrastructure for wide-area data transfer
• Well-configured data transfer hosts with modern tools
• Capable network devices
• High-performance data path which does not traverse commodity LAN
• Proactive operational models that enable performance
• Well-deployed test and measurement tools (perfSONAR)
• Periodic testing to locate issues instead of waiting for users to complain
• Security posture well-matched to high-performance science applications
8 – ESnet Science Engagement (engage@es.net) - 5/3/19
TCP – Ubiquitous and Fragile
• Networks provide connectivity between hosts – how do hosts
see the network?
• From an application’s perspective, the interface to “the other end” is a socket
• Communication is between applications – mostly over TCP
• TCP – the fragile workhorse
• TCP is (for very good reasons) timid – packet loss is interpreted as congestion
• Packet loss in conjunction with latency is a performance killer
• Like it or not, TCP is used for the vast majority of data transfer applications (more
than 95% of ESnet traffic is TCP)
9 – ESnet Science Engagement (engage@es.net) - 5/3/19
Packet/Data Loss?!
• “Wait a minute! I thought TCP was a reliable protocol? What do you mean ‘packet loss’,
where is the data going!?”
• The data isn’t loss forever, its dropped somewhere on the path.
• Usually by a device without enough buffer space to accept it, or by someone who thinks the data is
corrupted and they won’t send it
• Once its dropped, we have a way of knowing its been dropped.
• TCP is reliable, each end is keeping track of what was sent, and what was received.
• If something goes missing, its resent. Resending is what takes the time, and causes the slowdown.
• TCP is able to reliably and transparently recover from packet loss by retransmitting any/all lost
packets
• This is how it provides a reliable data transfer services to the applications
• The reliability mechanisms dramatically reduce performance when they are exercised
• We want to eliminate the causes of packet loss – so that we don’t need to test out the (slow)
way that TCP can recover.
• What is the impact of that recovery?
10 – ESnet Science Engagement (engage@es.net) - 5/3/19
A small amount of packet loss makes a huge
difference in TCP performance
Metro Area
Local
(LAN)
Regional
Continental
International
Measured (TCP Reno) Measured (HTCP) Theoretical (TCP Reno) Measured (no loss)
With loss, high performance
beyond metro distances is
essentially impossible
11 – ESnet Science Engagement (engage@es.net) - 5/3/19
How Do We Accommodate TCP?
• High-performance wide area TCP flows must get loss-
free service
• Sufficient bandwidth to avoid congestion
• Deep enough buffers in routers and switches to handle
bursts
• Especially true for long-distance flows due to packet behavior
• No, this isn’t buffer bloat
12 – ESnet Science Engagement (engage@es.net) - 5/3/19
• Equally important – the infrastructure must be verifiable so that clean service can be provided
– Stuff breaks
• Hardware, software, optics, bugs, …
• How do we deal with it in a production environment?
– Must be able to prove a network device or path is functioning correctly
• Regular active tests should be run - perfSONAR
– Small footprint is a huge win
• Fewer the number of devices = easier to locate the source of packet loss
© 2013 icanhascheezburger.com
Outline
• Motivation
• TCP Basics
• DMZ Design Patterns
• Integration w/ Science
• Conclusions
13 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science DMZ Takes Many Forms
• There are a lot of ways to combine these things – it all depends on
what you need to do
• Small installation for a project or two
• Facility inside a larger institution
• Institutional capability serving multiple departments/divisions
• Science capability that consumes a majority of the infrastructure
• Some of these are straightforward, others are less obvious
• Key point of concentration: eliminate sources of packet loss / packet
friction
14 – ESnet Science Engagement (engage@es.net) - 5/3/19
Ad Hoc DTN Deployment
15 – ESnet Science Engagement (engage@es.net) - 5/3/19
10GE10G
Site Border
Router
WAN
Building or Wiring
Closet Switch/Router
Perimeter Firewall
Site / Campus
LAN
High performance
Data Transfer Node
with high-speed storage
Global security policy
mixes rules for science
and business traffic
DTN traffic subject to firewall
limitations
perfSONAR
Test and measurement
not aligned with data
resource placement
DTN traffic subject to limitations of
general-purpose networking
equipment/config
Note: Site border
router and perimeter
firewall are often the
same device
Conflicting requirements
result in performance
compromises
Legacy Method: Ad Hoc DTN Deployment
• This is often what gets tried first
• Data transfer node deployed where the owner has space
• This is often the easiest thing to do at the time
• Straightforward to turn on, hard to achieve performance
16 – ESnet Science Engagement (engage@es.net) - 5/3/19
• If lucky, perfSONAR is at the border
– This is a good start
– Need a second one next to the DTN
• Entire LAN path has to be sized for data flows
• Entire LAN path is part of any troubleshooting exercise
• This usually fails to provide the necessary performance.
A better approach: simple Science DMZ
10GE
10GE
10GE
10GE
10G
Border Router
WAN
Science DMZ
Switch/Router
Enterprise Border
Router/Firewall
Site / Campus
LAN
High performance
Data Transfer Node
with high-speed storage
Per-service
security policy
control points
Clean,
High-bandwidth
WAN path
Site / Campus
access to Science
DMZ resources
perfSONAR
perfSONAR
perfSONAR
17 – ESnet Science Engagement (engage@es.net) - 5/3/19
Small-scale Science DMZ Deployment
• Add-on to existing network infrastructure
• All that is required is a port on the border router
• Small footprint, pre-production commitment
• Easy to experiment with components and technologies
• DTN prototyping
• perfSONAR testing
• Limited scope makes security policy exceptions easy
• Only allow traffic from partners
• Add-on to production infrastructure – lower risk
18 – ESnet Science Engagement (engage@es.net) - 5/3/19
Prototype Science DMZ Data Path
10GE
10GE
10GE
10GE
10G
Border Router
WAN
Science DMZ
Switch/Router
Enterprise Border
Router/Firewall
Site / Campus
LAN
High performance
Data Transfer Node
with high-speed storage
Per-service
security policy
control points
Clean,
High-bandwidth
WAN path
Site / Campus
access to Science
DMZ resources
perfSONAR
perfSONAR
High Latency WAN Path
Low Latency LAN Path
19 – ESnet Science Engagement (engage@es.net) - 5/3/19
Support For Multiple Projects
• Science DMZ architecture allows multiple projects to put DTNs in place
• Modular architecture
• Centralized location for data servers
• This may or may not work well depending on institutional politics
• Issues such as physical security can make this a non-starter
• On the other hand, some shops already have service models in place
• On balance, this can provide a cost savings – it depends
• Central support for data servers vs. carrying data flows
• How far do the data flows have to go?
20 – ESnet Science Engagement (engage@es.net) - 5/3/19
Multiple Projects
10GE
10GE
10GE
10G
Border Router
WAN
Science DMZ
Switch/Router
Enterprise Border
Router/Firewall
Site / Campus
LAN
Project A DTN
Per-project
security policy
control points
Clean,
High-bandwidth
WAN path
Site / Campus
access to Science
DMZ resources
perfSONAR
perfSONAR
Project B DTN
Project C DTN
21 – ESnet Science Engagement (engage@es.net) - 5/3/19
Supercomputer Center Deployment
• High-performance networking is assumed in this environment
• Data flows between systems, between systems and storage, wide area, etc.
• Global filesystem often ties resources together
• Portions of this may not run over Ethernet (e.g. IB)
• Implications for Data Transfer Nodes
• “Science DMZ” may not look like a discrete entity here
• By the time you get through interconnecting all the resources, you end up with most of the
network in the Science DMZ
• This is as it should be – the point is appropriate deployment of tools, configuration, policy
control, etc.
• Office networks can look like an afterthought, but they aren’t
• Deployed with appropriate security controls
• Office infrastructure need not be sized for science traffic
22 – ESnet Science Engagement (engage@es.net) - 5/3/19
Supercomputer Center Data Path
Virtual
Circuit
Routed
Border Router
WAN
Core
Switch/Router
Firewall
Offices
perfSONAR
perfSONAR
perfSONAR
Supercomputer
Parallel Filesystem
Front end
switch
Data Transfer
Nodes
Front end
switch
High Latency WAN Path
Low Latency LAN Path
High Latency VC Path
23 – ESnet Science Engagement (engage@es.net) - 5/3/19
Distributed Science DMZ
• Fiber-rich environment enables distributed Science DMZ
• No need to accommodate all equipment in one location
• Allows the deployment of institutional science service
• WAN services arrive at the site in the normal way
• Dark fiber distributes connectivity to Science DMZ services throughout the site
• Departments with their own networking groups can manage their own local Science DMZ
infrastructure
• Facilities or buildings can be served without building up the business network to support
those flows
• Security is potentially more complex
• Remote infrastructure must be monitored
• Solutions depend on relationships with security groups
24 – ESnet Science Engagement (engage@es.net) - 5/3/19
Distributed Science DMZ – Dark Fiber
Dark
Fiber
Dark
Fiber
10GE
Dark
Fiber
10GE
10GE
10G
Border Router
WAN
Science DMZ
Switch/Router
Enterprise Border
Router/Firewall
Site / Campus
LAN
Per-project
security policy
control points
Clean,
High-bandwidth
WAN path
Site / Campus
access to Science
DMZ resources
perfSONAR
perfSONAR
Project A DTN
(remote)
Project B DTN
(remote)
Project C DTN
(remote)
25 – ESnet Science Engagement (engage@es.net) - 5/3/19
Common Threads• Two common threads exist in all these examples
• Accommodation of TCP
• Wide area portion of data transfers traverses purpose-built path
• High performance devices that don’t drop packets
• Ability to test and verify
• When problems arise (and they always will), they can be solved if the infrastructure is
built correctly
• Small device count makes it easier to find issues
• Multiple test and measurement hosts provide multiple views of the data path
• perfSONAR nodes at the site and in the WAN
• perfSONAR nodes at the remote site
26 – ESnet Science Engagement (engage@es.net) - 5/3/19
Equipment – Routers and Switches
• Requirements for Science DMZ gear are different
• No need to go for the kitchen sink list of services
• A Science DMZ box only needs to do a few things, but do them well
• Support for the latest LAN integration magic with your Windows Active Directory
environment is probably not super-important
• A clean architecture is important
• How fast can a single flow go?
• Are there any components that go slower than interface wire speed?
• There is a temptation to go cheap
• Hey, it only needs to do a few things, right?
• You typically don’t get what you don’t pay for
• (You sometimes don’t get what you pay for either)
27 – ESnet Science Engagement (engage@es.net) - 5/3/19
Common Circumstance:
Multiple Ingress Data Flows, Common Egress
Background
traffic or
competing bursts
DTN traffic with
wire-speed
bursts
10GE
10GE
10GE
Hosts will typically send packets at the speed of their
interface (1G, 10G, etc.)
• Instantaneous rate, not average rate
• If TCP has window available and data to send, host
sends until there is either no data or no window
Hosts moving big data (e.g. DTNs) can send large bursts
of back-to-back packets
• This is true even if the average rate as measured
over seconds is slower (e.g. 4Gbps)
• On microsecond time scales, there is often
congestion
• Router or switch must queue packets or drop them
28 – ESnet Science Engagement (engage@es.net) - 5/3/19
Rant Ahead
29 – ESnet Science Engagement (engage@es.net) - 5/3/19
N.B. You are entering into rant territory on the matter of switch buffering. If you are going to take
away anything from the next section:
1. Under buffered network devices are the single greatest threat to data intensive use of the
network. You can make hosts, operating systems, and application choices perform better
for ‘free’, it will cost $$$ to fix a crappy switch or router
2. You will be steered toward non-optimal choices when you talk with the vendor community
because they don’t understand simple math (but by the end of this, you will).
3. A 1U/2U data center/racklan network device should never be in the path of your data
intensive network use case.
4. Non-passive (e.g. stateful) security devices are the same for buffering, and are actually
worse due to the processing overhead.
5. Anytime you jump around the OSI stack – add friction (e.g. routing when you don’t need to,
application layer inspection, etc.)
All About That Buffer (No Cut Through)
30 – ESnet Science Engagement (engage@es.net) - 5/3/19
All About That Buffer (No Cut Through)
31 – ESnet Science Engagement (engage@es.net) - 5/3/19
• Data arrives
from multiple
sources
• Buffers have a finite amount of memory
• Some have this per interface
• Others may have access to a shared
memory region with other interfaces
• The processing engine will:
• Extract each packet/frame from the
queues
• Pull off header information to see where
the destination should be
• Move the packet/frame to the correct
output queue
• Additional delay is possible as the
queues physically write the packet
to the transport medium (e.g.
optical interface, copper interface)
All About That Buffer (No Cut Through)
32 – ESnet Science Engagement (engage@es.net) - 5/3/19
• The Bandwidth Delay Product
• The amount of “in flight” data for a TCP connection (BDP = bandwidth * round
trip time)
• Example: 10Gb/s cross country, ~100ms
• 10,000,000,000 b/s * .1 s = 1,000,000,000 bits
• 1,000,000,000 / 8 = 125,000,000 bytes
• 125,000,000 bytes / (1024*1024) ~ 125MB
• Ignore the math aspect: its making sure there is memory to catch and send
packets
• As the speed increases, there are more packets.
• If there is not memory, we drop them, and that makes TCP sad.
All About That Buffer (No Cut Through)
33 – ESnet Science Engagement (engage@es.net) - 5/3/19
• Buffering isn’t as important on the LAN (this is why you are normally
pressured to buy ‘cut through’ devices)
• Change the math to make the Latency 1ms = 1.25MB
• ‘Cut through’ and low latency switches are designed for the data center, and
can handle typical data center loads that don’t require buffering (e.g. same to
same speeds, destinations within the broadcast domain)
• Buffering *MATTERS* for WAN Transfers
• Placing something with inadequate buffering in the path reduces the buffer
for the entire path. E.g. if you have an expectation of 10Gbps over 100ms –
don’t place a 12MB buffer anywhere in there – your reality is now ~10x less
than it was before (e.g. 10Gbps @ 10ms, or 1Gbps @ 100ms)
• Ignore the math aspect, its really just about making sure there is memory to catch
packets. As the speed increases, there are more packets. If there is not memory,
we drop them, and that makes TCP sad.
• Memory on hosts, and network gear
All About That Buffer (No Cut Through)
[ ID] Interval Transfer Bandwidth Retr Cwnd
[ 14] 0.00-1.00 sec 524 KBytes 4.29 Mbits/sec 0 157 KBytes
[ 14] 1.00-2.00 sec 3.31 MBytes 27.8 Mbits/sec 0 979 KBytes
[ 14] 2.00-3.00 sec 17.7 MBytes 148 Mbits/sec 0 5.36 MBytes
[ 14] 3.00-4.00 sec 18.8 MBytes 157 Mbits/sec 214 1.77 MBytes
[ 14] 4.00-5.00 sec 11.2 MBytes 94.4 Mbits/sec 0 1.88 MBytes
[ 14] 5.00-6.00 sec 10.0 MBytes 83.9 Mbits/sec 0 2.39 MBytes
[ 14] 6.00-7.00 sec 16.2 MBytes 136 Mbits/sec 0 3.63 MBytes
[ 14] 7.00-8.00 sec 23.8 MBytes 199 Mbits/sec 0 5.50 MBytes
[ 14] 8.00-9.00 sec 38.8 MBytes 325 Mbits/sec 0 8.23 MBytes
[ 14] 9.00-10.00 sec 57.5 MBytes 482 Mbits/sec 0 11.8 MBytes
[ 14] 10.00-11.00 sec 81.2 MBytes 682 Mbits/sec 0 16.2 MBytes
[ 14] 11.00-12.00 sec 50.0 MBytes 419 Mbits/sec 35 3.93 MBytes
[ 14] 12.00-13.00 sec 15.0 MBytes 126 Mbits/sec 0 2.20 MBytes
[ 14] 13.00-14.00 sec 11.2 MBytes 94.4 Mbits/sec 0 2.53 MBytes
[ 14] 14.00-15.00 sec 13.8 MBytes 115 Mbits/sec 1 1.50 MBytes
[ 14] 15.00-16.00 sec 6.25 MBytes 52.4 Mbits/sec 5 813 KBytes
[ 14] 16.00-17.00 sec 5.00 MBytes 41.9 Mbits/sec 0 909 KBytes
[ 14] 17.00-18.00 sec 5.00 MBytes 41.9 Mbits/sec 0 1.37 MBytes
[ 14] 18.00-19.00 sec 10.0 MBytes 83.9 Mbits/sec 0 2.43 MBytes
[ 14] 19.00-20.00 sec 17.5 MBytes 147 Mbits/sec 0 4.22 MBytes
34 – ESnet Science Engagement (engage@es.net) - 5/3/19
• What does this “look” like to a data transfer? Consider the test of iperf below
– See TCP ‘ramp up’ and slowly increase the window
– When something in the path has no more space for packets – a drop occurs. TCP will eventually react to the lost packet,
and ‘back off’
– In the example, this first occurs when we reach a buffer of around 6-8MB. Then after backoff the window is halved a
couple of times
– This happens again later – at a slightly higher buffer limit. This could be because there was cross traffic the first time, etc.
Decoding Specifications
• “The buffering behaviors of the switches and their
operating system, such as behavior under memory stress,
are typically proprietary information and not well
documented” http://www.measurementlab.net/blog/traffic-microbursts-and-their-effect-on-internet-measurement/
• “Even if you know how much packet buffer is in the
switch, assumptions on how it is deployed that are not
backed up by testing can lead to unhappiness. What we
like to say is that is the job of the network engineers to
move bottlenecks around.”
• Jim Warner
• http://people.ucsc.edu/~warner/buffer.html
35 – ESnet Science Engagement (engage@es.net) - 5/3/19
Decoding Specifications
• So lets say the spec sheet says this:
• What does ‘module’ mean?
• Typically this means the amount of memory for the entire switch
(if it’s a single unit) or a blade (if the chassis supports more than
one.
• BUT … this memory can be allocated in a number of different
ways:
• Shared between all ports
• Dedicated (smaller) amounts per-port
• Shared between ASICS, which control a bank of ports
36 – ESnet Science Engagement (engage@es.net) - 5/3/19
Decoding Specifications
•Consider this architecture
• 48 Ports
• 12 ASICS
• 4 Ports per ASIC
• 72MB total
• 6MB per ASIC
• If all ports are in use – expect that each
port has access to 1.5MB. If only one is
in use, it can use 6MB
• Additional memory is often available
in a ‘burst buffer’ in the fabric
37 – ESnet Science Engagement (engage@es.net) - 5/3/19
ASIC = application-specific integrated circuit, think 'small routing engine'
Decoding Specifications
• Recall:
https://www.switch.ch/network/tools/tcp_throughput/
• What does 6MB get you?
• 1Gbps @ <= 48ms (e.g. ½ needed for coast-to-coast)
• 10Gbps @ <= 4.8ms (e.g. metro area)
• What does 1.5MB get you?
• 1Gbps @ <= 12ms (e.g. regional area)
• 10Gbps @ <= 1.2ms (e.g. data center [or more accurately, rack or
row)]
• In either case – remember this assumes you are the only thing using that memory
… congestion is a more likely reality
38 – ESnet Science Engagement (engage@es.net) - 5/3/19
Outline
• Motivation
• TCP Basics
• DMZ Design Patterns
• Integration w/ Science
• Conclusions
39 – ESnet Science Engagement (engage@es.net) - 5/3/19
NCAR RDA Data Portal
• Let’s say I have a nice compute allocation at NERSC – climate science
• Let’s say I need some data from NCAR for my project
• https://rda.ucar.edu/
• Data sets (there are many more, but these are two):
• https://rda.ucar.edu/datasets/ds199.1/ (1.5TB)
• https://rda.ucar.edu/datasets/ds313.0/ (430GB)
• Download to NERSC (could also do ALCF or NCSA or OLCF)
NCAR RDA Performance to DOE HPC Facilities
13.9 Gbps 16.6 Gbps 11.9 Gbps
DTN
nersc#dtn
NERSC
DTN
olcf#dtn_atlas
OLCF
DTN
alcf#dtn_mira
ALCF
DTN
NCAR RDA
rda#datashare
• 1.5TB data set
• 1121 files
41 – ESnet Science Engagement (engage@es.net) - 5/3/19
Next Steps – Building On The Science DMZ
• Enhanced cyberinfrastructure substrate now exists
• Wide area networks (ESnet, GEANT, Internet2, Regionals)
• Science DMZs connected to those networks
• DTNs in the Science DMZs
• What does the scientist see?
• Scientist sees a science application
• Data transfer
• Data portal
• Data analysis
• Science applications are the user interface to networks and DMZs
• Large-scale data-intensive science requires that we build larger structures
on top of those components
42 – ESnet Science Engagement (engage@es.net) - 5/3/19
HPC Facilities: The Petascale DTN Project
• Built on top of the Science DMZ model
• Effort to improve data transfer performance between the DOE ASCR HPC
facilities at ANL, LBNL, and ORNL, and also NCSA.
• Multiple current and future science projects need to transfer data between HPC facilities
• Performance goal is 15 gigabits per second (equivalent to 1PB/week)
• Realize performance goal for routine Globus transfers without special tuning
• Reference data set is 4.4TB of cosmology simulation data
43 – ESnet Science Engagement (engage@es.net) - 5/3/19
DTN Cluster Performance – HPC Facilities
(2016)
6.8 Gbps
7.6 Gbps
6.9 Gbps
13.3 Gbps
6.0 Gbps
6.7 Gbps
11.1 Gbps
10.5 Gbps
7.3 Gbps
10.0 Gbps
13.4 Gbps
8.2 Gbps
DTN
DTN
DTN
DTN
Data set: L380
Files: 19260
Directories: 211
Other files: 0
Total bytes: 4442781786482 (4.4T bytes)
Smallest file: 0 bytes (0 bytes)
Largest file: 11313896248 bytes (11G bytes)
Size distribution:
1 - 10 bytes: 7 files
10 - 100 bytes: 1 files
100 - 1K bytes: 59 files
1K - 10K bytes: 3170 files
10K - 100K bytes: 1560 files
100K - 1M bytes: 2817 files
1M - 10M bytes: 3901 files
10M - 100M bytes: 3800 files
100M - 1G bytes: 2295 files
1G - 10G bytes: 1647 files
10G - 100G bytes: 3 files
Petascale DTN Project
March 2016
L380 Data Set
NERSC DTN cluster
Globus endpoint: nersc#dtn
ALCF DTN cluster
Globus endpoint: alcf#dtn_mira
OLCF DTN cluster
Globus endpoint: olcf#dtn_atlas
NCSA DTN cluster
Globus endpoint: ncsa#BlueWaters
44 – ESnet Science Engagement (engage@es.net) - 5/3/19
DTN Cluster Performance – HPC
Facilities (2017)
21.2/22.6/24.5
Gbps
23.1/33.7/39.7
Gbps
26.7/34.7/39.9
Gbps
33.2/43.4/50.3
Gbps
35.9/39.0/40.7
Gbps
29.9/33.1/35.5
Gbps
34.6/47.5/56.8
Gbps
44.1/46.8/48.4
Gbps
41.0/42.2/43.9
Gbps
33.0/35.0/37.8
Gbps
43.0/50.0/56.3
Gbps
55.4/56.7/57.4
Gbps
DTN
DTN
DTN
DTN
NERSC DTN cluster
Globus endpoint: nersc#dtn
Filesystem: /project
Data set: L380
Files: 19260
Directories: 211
Other files: 0
Total bytes: 4442781786482 (4.4T bytes)
Smallest file: 0 bytes (0 bytes)
Largest file: 11313896248 bytes (11G bytes)
Size distribution:
1 - 10 bytes: 7 files
10 - 100 bytes: 1 files
100 - 1K bytes: 59 files
1K - 10K bytes: 3170 files
10K - 100K bytes: 1560 files
100K - 1M bytes: 2817 files
1M - 10M bytes: 3901 files
10M - 100M bytes: 3800 files
100M - 1G bytes: 2295 files
1G - 10G bytes: 1647 files
10G - 100G bytes: 3 files
Petascale DTN Project
November 2017
L380 Data Set
Gigabits per second
(min/avg/max), three
transfers
ALCF DTN cluster
Globus endpoint: alcf#dtn_mira
Filesystem: /projects
OLCF DTN cluster
Globus endpoint: olcf#dtn_atlas
Filesystem: atlas2
NCSA DTN cluster
Globus endpoint: ncsa#BlueWaters
Filesystem: /scratch
45 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science Data Portals
• Large repositories of scientific data
• Climate data
• Sky surveys (astronomy, cosmology)
• Many others
• Data search, browsing, access
• Many scientific data portals were designed 15+ years ago
• Single-web-server design
• Data browse/search, data access, user awareness all in a single
system
• All the data goes through the portal server
• In many cases by design
• E.g. embargo before publication (enforce access control)
46 – ESnet Science Engagement (engage@es.net) - 5/3/19
Legacy Portal Design
10GE
Border Router
WAN
Firewall
Enterprise
perfSONAR
perfSONAR
Filesystem
(data store)
10GE
Portal
Server
Browsing path
Query path
Data path
Portal server applications:
· web server
· search
· database
· authentication
· data service
• Very difficult to improve performance
without architectural change
– Software components all tangled
together
– Difficult to put the whole portal in a
Science DMZ because of security
– Even if you could put it in a DMZ, many
components aren’t scalable
• What does architectural change mean?
47 – ESnet Science Engagement (engage@es.net) - 5/3/19
Architectural Examination of Data Portals
• Common data portal functions (most portals have these)
• Search/query/discovery
• Data download method for data access
• GUI for browsing by humans
• API for machine access – ideally incorporates search/query + download
• Performance pain is primarily in the data handling piece
• Rapid increase in data scale eclipsed legacy software stack capabilities
• Portal servers often stuck in enterprise network
• Can we “disassemble” the portal and put the pieces back together better?
• Use Science DMZ as a platform for the data piece
• Avoid placing complex software in the Science DMZ
48 – ESnet Science Engagement (engage@es.net) - 5/3/19
Legacy Portal Design
10GE
Border Router
WAN
Firewall
Enterprise
perfSONAR
perfSONAR
Filesystem
(data store)
10GE
Portal
Server
Browsing path
Query path
Data path
Portal server applications:
· web server
· search
· database
· authentication
· data service
49 – ESnet Science Engagement (engage@es.net) - 5/3/19
Next-Generation Portal Leverages Science
DMZ
10GE10GE
10GE
10GE
Border Router
WAN
Science DMZ
Switch/Router
Firewall
Enterprise
perfSONAR
perfSONAR
10GE
10GE
10GE
10GE
DTN
DTN
API DTNs
(data access governed
by portal)
DTN
DTN
perfSONAR
Filesystem
(data store)
10GE
Portal
Server
Browsing path
Query path
Portal server applications:
· web server
· search
· database
· authentication
Data Path
Data Transfer Path
Portal Query/Browse Path
50 – ESnet Science Engagement (engage@es.net) - 5/3/19
Put The Data On Dedicated Infrastructure
• We have separated the data handling from the portal logic
• Portal is still its normal self, but enhanced
• Portal GUI, database, search, etc. all function as they did before
• Query returns pointers to data objects in the Science DMZ
• Portal is now freed from ties to the data servers (run it on Amazon if you want!)
• Data handling is separate, and scalable
• High-performance DTNs in the Science DMZ
• Scale as much as you need to without modifying the portal software
• Outsource data handling to computing centers
• Computing centers are set up for large-scale data
• Let them handle the large-scale data, and let the portal do the orchestration of data
placement
51 – ESnet Science Engagement (engage@es.net) - 5/3/19
JGI Data Portal
52 – ESnet Science Engagement (engage@es.net) - 5/3/19
53 – ESnet Science Engagement (engage@es.net) - 5/3/19
Reasons To Scale Data Portals
• Some reasons are obvious
• Increase in size of data objects (MB à GB à 100s of GB)
• Number of data objects (many thousands per data set)
• Other reasons are paradigm shifts
• Modern data analysis on HPC can use a *lot* of data
• Today’s HPC facilities are far more capable than in the past
• Retrofit / rebuilt data portals and data repositories
• Significant wins from increased data analysis
54 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science at Scale: Genomics
55 – ESnet Science Engagement (engage@es.net) - 5/3/19
Science at Scale: Climate
56 – ESnet Science Engagement (engage@es.net) - 5/3/19
They Can Use All The Data
• Groups like these need large data sets
• Much of the data in their field is behind legacy portals
• Significant human effort to retrieve what they need
• Legacy systems perform poorly, especially at scale
• Legacy data portals are a product of their time
• We now live in the future from the perspective of those designs
• Current systems far exceed the capabilities available 15 years ago
• From the perspective of today’s systems, legacy portals are products of a bygone
past
• It is now perfectly reasonable for a scientist to want all the data
• Machine learning + HPC
• But this only works if the scientists can get to the data at scale
57 – ESnet Science Engagement (engage@es.net) - 5/3/19
IT/Engineering (not the scientists) Have To Do This
• The scientific community cannot do this for themselves
• Individual researchers do not control the resources
• Computing centers
• Data repositories
• Science networks
• Our community owns these – we have to do the work
• Integration, performance engineering, interoperability
• Science Engagement to teach scientists how to use the
better platforms
• This is the path forward
¯_(ツ)_/¯
58 – ESnet Science Engagement (engage@es.net) - 5/3/19
Networks Cannot Do This Alone
• We need a whole-community effort
• Networks
• HPC facilities
• Data repositories / Data portals
• Experimental facilities
• Science collaborations
• Science programs
• Networks can help, and must be part of the conversation
• Heavy lifting is now at the network edge, in collaboration with the
network core
• Need to get the architecture right – we know how to do this
• Engagement is critical
• We have to teach people to use the new thing
• They aren’t going to magically find out about new services
59 – ESnet Science Engagement (engage@es.net) - 5/3/19
Long-Term Vision
ESnet
(Big Science facilities,
DOE labs)
Internet2 + Regionals
(US Universities and
affiliated institutions)
International Networks
(Universities and labs in Europe,
Asia, Americas, Australia, etc.)
High-performance feature-rich
science network ecosystem
Commercial Clouds
(Amazon, Google,
Microsoft, etc.)
Agency Networks
(NASA, NOAA, etc.)
Campus HPC
+
Data
60 – ESnet Science Engagement (engage@es.net) - 5/3/19
It’s All A Bunch Of Science DMZs
High-performance feature-rich
science network ecosystem
DTN
DTN
DTN
DTN
DMZDMZ
DMZDMZ
DTN
DATA
DTN
DMZDMZ
DTN
DTN
DMZDMZ
DATA
DTN DTN
DMZDMZ
Parallel
Filesystem
DTN DTN
DTN
DTNDMZDMZ
DATA
DTN
DTN
DMZDMZ
Experiment
Data Archive
DTN
DTN
DTN
DTN
DTN DMZDMZ
DTN DTN
DMZDMZ
DATA
61 – ESnet Science Engagement (engage@es.net) - 5/3/19
It’s All A Bunch Of Science DMZs
High-performance feature-rich
science network ecosystem
DTN
DTN
DTN
DTN
DMZDMZ
DMZDMZ
DTN
DATA
DTN
DMZDMZ
DTN
DTN
DMZDMZ
DATA
DTN DTN
DMZDMZ
Parallel
Filesystem
DTN DTN
DTN
DTNDMZDMZ
DATA
DTN
DTN
DMZDMZ
Experiment
Data Archive
DTN
DTN
DTN
DTN
DTN DMZDMZ
DTN DTN
DMZDMZ
DATA
HPC Facilities
Single Lab
Experiments
Data
Portal
LHC
Experiments
University
Computing
62 – ESnet Science Engagement (engage@es.net) - 5/3/19
Outline
• Motivation
• TCP Basics
• DMZ Design Patterns
• Integration w/ Science
• Conclusions
63 – ESnet Science Engagement (engage@es.net) - 5/3/19
ESnet’s Science DMZ – Enabling Big Data Science
The Science DMZ design pattern supports secure, high-performance data movement between and among
facilities, labs, universities, data portals, and clouds
DevelopmentDeploymentPlatform
10GE10GE
10GE
10GE
Border Router
WAN
Science DMZ
Switch/Router
Firewall
Enterprise
perfSONAR
perfSONAR
10GE
10GE
10GE
10GE
DTN
DTN
API DTNs
(data access governed
by portal)
DTN
DTN
perfSONAR
Filesystem
(data store)
10GE
Portal
Server
Browsing path
Query path
Portal server applications:
· web server
· search
· database
· authentication
Data Path
Data Transfer Path
Portal Query/Browse Path
Long Term Vision
Development Support
§ Science DMZ supports development,
prototyping, and incremental
deployment of advanced network and
data services
§ Petascale DTN
Project
§ Ongoing production
operation across
science complex
2017-2018
§ Modern Research
Data Portal design
pattern
§ Pacific Research
Platform
§ Science DMZs deployed
at DOE labs and facilities
§ NSF CC-NIE and
successor programs fund
over 100 Science DMZs
at universities
2011-20162005-2010
§ Requirements analysis of DOE Office of
Science facilities and projects from network
and data perspective
§ Performance engineering, troubleshooting,
experience
§ Creation of Science DMZ design pattern
Science Networks:
Campus, Regional,
National, International
Science DMZ
Science DMZ
Science DMZ Science DMZ
DTN
DTN
DTN
DTN
DTN
DTN
DTN
DTN
Data Platform
Data Platform
§ Science Engagement model for science
support
§ OIN workshop series trains CI engineers
throughout the community
21.2/22.6/24.5
Gbps
23.1/33.7/39.7
Gbps
26.7/34.7/39.9
Gbps
33.2/43.4/50.3
Gbps
35.9/39.0/40.7
Gbps
29.9/33.1/35.5
Gbps
34.6/47.5/56.8
Gbps
44.1/46.8/48.4
Gbps
41.0/42.2/43.9
Gbps
33.0/35.0/37.8
Gbps
43.0/50.0/56.3
Gbps
55.4/56.7/57.4
Gbps
DTN
DTN
DTN
DTN
NERSC DTN cluster
Globus endpoint: nersc#dtn
Filesystem: /project
Data set: L380
Files: 19260
Directories: 211
Other files: 0
Total bytes: 4442781786482 (4.4T bytes)
Smallest file: 0 bytes (0 bytes)
Largest file: 11313896248 bytes (11G bytes)
Size distribution:
1 - 10 bytes: 7 files
10 - 100 bytes: 1 files
100 - 1K bytes: 59 files
1K - 10K bytes: 3170 files
10K - 100K bytes: 1560 files
100K - 1M bytes: 2817 files
1M - 10M bytes: 3901 files
10M - 100M bytes: 3800 files
100M - 1G bytes: 2295 files
1G - 10G bytes: 1647 files
10G - 100G bytes: 3 files
Petascale DTN Project
November 2017
L380 Data Set
Gigabits per second
(min/avg/max), three
transfers
ALCF DTN cluster
Globus endpoint: alcf#dtn_mira
Filesystem: /projects
OLCF DTN cluster
Globus endpoint: olcf#dtn_atlas
Filesystem: atlas2
NCSA DTN cluster
Globus endpoint: ncsa#BlueWaters
Filesystem: /scratch
10GE
10GE
10GE
10GE
10G
Border Router
WAN
Science DMZ
Switch/Router
Enterprise Border
Router/Firewall
Site / Campus
LAN
High performance
Data Transfer Node
with high-speed storage
Per-service
security policy
control points
Clean,
High-bandwidth
WAN path
Site / Campus
access to Science
DMZ resources
perfSONAR
perfSONAR
High Latency WAN Path
Low Latency LAN Path
perfSONAR
Production Data Movement
§ Science DMZs support large scale data
movement and data placement across
the science complex (DOE and
collaborators)
§ Performant security
Scalable Data Platform
§ Science DMZs provide
foundation for data-driven
science globally
National Science Foundation Award #1826994
Jason Zurawski
zurawski@es.net
ESnet / Lawrence Berkeley National Laboratory
http://bit.ly/2vx2HiQ
Maximizing Performance and Network
Utility with a Science DMZ
Globusworld 2019
Chicago, IL
May 2nd, 2019
https://epoc.global

Más contenido relacionado

La actualidad más candente

Enterprise Networks for Connected Buildings
Enterprise Networks for Connected BuildingsEnterprise Networks for Connected Buildings
Enterprise Networks for Connected BuildingsPanduit
 
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...ITU
 
Edge device multi-unicasting for video streaming
Edge device multi-unicasting for video streamingEdge device multi-unicasting for video streaming
Edge device multi-unicasting for video streamingTal Lavian Ph.D.
 
( Network topology) )final
( Network topology) )final( Network topology) )final
( Network topology) )finalJai Kumar
 
Peerto Peer Networks
Peerto Peer NetworksPeerto Peer Networks
Peerto Peer Networkssanjoysanyal
 
Datacom module 3: Data Communications Circuits, Arrangements, and Networks
Datacom module 3:  Data Communications Circuits, Arrangements, and NetworksDatacom module 3:  Data Communications Circuits, Arrangements, and Networks
Datacom module 3: Data Communications Circuits, Arrangements, and NetworksJeffrey Des Binwag
 
Lect 3-topologies
Lect 3-topologiesLect 3-topologies
Lect 3-topologiessomsanwal
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task ComputingEric Van Hensbergen
 
Het nets up close - Webinar with ThinkSmallCell and IXIA
Het nets up close - Webinar with ThinkSmallCell and IXIAHet nets up close - Webinar with ThinkSmallCell and IXIA
Het nets up close - Webinar with ThinkSmallCell and IXIADavid Chambers
 
Network function virtualization
Network function virtualizationNetwork function virtualization
Network function virtualizationSatish Chavan
 
WIRELESS INTERNET BY SAIKIRAN PANJALA
WIRELESS INTERNET BY SAIKIRAN PANJALAWIRELESS INTERNET BY SAIKIRAN PANJALA
WIRELESS INTERNET BY SAIKIRAN PANJALASaikiran Panjala
 
Non-Fluff Software Defined Networking, Network Function Virtualization and IoT
Non-Fluff Software Defined Networking, Network Function Virtualization and IoTNon-Fluff Software Defined Networking, Network Function Virtualization and IoT
Non-Fluff Software Defined Networking, Network Function Virtualization and IoTMark Ryan Castellani
 

La actualidad más candente (20)

Enterprise Networks for Connected Buildings
Enterprise Networks for Connected BuildingsEnterprise Networks for Connected Buildings
Enterprise Networks for Connected Buildings
 
Mini Project- Implementation & Evaluation of Wireless LANs
Mini Project- Implementation & Evaluation of Wireless LANsMini Project- Implementation & Evaluation of Wireless LANs
Mini Project- Implementation & Evaluation of Wireless LANs
 
Computer network
Computer networkComputer network
Computer network
 
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...
The Fraunhofer FOKUS 5G Playground Prototyping 5G-Ready Infrastructures and A...
 
Edge device multi-unicasting for video streaming
Edge device multi-unicasting for video streamingEdge device multi-unicasting for video streaming
Edge device multi-unicasting for video streaming
 
( Network topology) )final
( Network topology) )final( Network topology) )final
( Network topology) )final
 
Networking
NetworkingNetworking
Networking
 
Final_Report
Final_ReportFinal_Report
Final_Report
 
Bit torrent a revolution in p2p
Bit torrent a revolution in p2pBit torrent a revolution in p2p
Bit torrent a revolution in p2p
 
NETWORKING
NETWORKINGNETWORKING
NETWORKING
 
Peerto Peer Networks
Peerto Peer NetworksPeerto Peer Networks
Peerto Peer Networks
 
Networking(CPM & PERT)
Networking(CPM & PERT)Networking(CPM & PERT)
Networking(CPM & PERT)
 
Datacom module 3: Data Communications Circuits, Arrangements, and Networks
Datacom module 3:  Data Communications Circuits, Arrangements, and NetworksDatacom module 3:  Data Communications Circuits, Arrangements, and Networks
Datacom module 3: Data Communications Circuits, Arrangements, and Networks
 
Lect 3-topologies
Lect 3-topologiesLect 3-topologies
Lect 3-topologies
 
Presentation7
Presentation7Presentation7
Presentation7
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task Computing
 
Het nets up close - Webinar with ThinkSmallCell and IXIA
Het nets up close - Webinar with ThinkSmallCell and IXIAHet nets up close - Webinar with ThinkSmallCell and IXIA
Het nets up close - Webinar with ThinkSmallCell and IXIA
 
Network function virtualization
Network function virtualizationNetwork function virtualization
Network function virtualization
 
WIRELESS INTERNET BY SAIKIRAN PANJALA
WIRELESS INTERNET BY SAIKIRAN PANJALAWIRELESS INTERNET BY SAIKIRAN PANJALA
WIRELESS INTERNET BY SAIKIRAN PANJALA
 
Non-Fluff Software Defined Networking, Network Function Virtualization and IoT
Non-Fluff Software Defined Networking, Network Function Virtualization and IoTNon-Fluff Software Defined Networking, Network Function Virtualization and IoT
Non-Fluff Software Defined Networking, Network Function Virtualization and IoT
 

Similar a Tutorial: Maximizing Performance and Network Utility with a Science DMZ

The Science DMZ
The Science DMZThe Science DMZ
The Science DMZJisc
 
Future services on Janet
Future services on JanetFuture services on Janet
Future services on JanetJisc
 
Intternetworking With TCP/IP
Intternetworking With TCP/IPIntternetworking With TCP/IP
Intternetworking With TCP/IPBIT DURG
 
Network Engineering for High Speed Data Sharing
Network Engineering for High Speed Data SharingNetwork Engineering for High Speed Data Sharing
Network Engineering for High Speed Data SharingGlobus
 
Science DMZ
Science DMZScience DMZ
Science DMZJisc
 
Common Design Elements for Data Movement Eli Dart
Common Design Elements for Data Movement Eli DartCommon Design Elements for Data Movement Eli Dart
Common Design Elements for Data Movement Eli DartEd Dodds
 
Data Mobility Exhibition
Data Mobility ExhibitionData Mobility Exhibition
Data Mobility ExhibitionGlobus
 
Datacom module 2: Data Communication Architecture, Protocols, and Standards
Datacom module 2:  Data Communication Architecture, Protocols, and StandardsDatacom module 2:  Data Communication Architecture, Protocols, and Standards
Datacom module 2: Data Communication Architecture, Protocols, and StandardsJeffrey Des Binwag
 
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...SURFnet
 
Introduction to data communications
Introduction to data communicationsIntroduction to data communications
Introduction to data communicationsArunaSekar1
 
EC8551 COMMUNICATION NETWORKS
EC8551 COMMUNICATION NETWORKSEC8551 COMMUNICATION NETWORKS
EC8551 COMMUNICATION NETWORKSGOWTHAMMS6
 
Science DMZ security
Science DMZ securityScience DMZ security
Science DMZ securityJisc
 
Big Data Analytics and Advanced Computer Networking Scenarios
Big Data Analytics and Advanced Computer Networking ScenariosBig Data Analytics and Advanced Computer Networking Scenarios
Big Data Analytics and Advanced Computer Networking ScenariosStenio Fernandes
 
Science DMZ as a Service: Creating Science Super- Facilities with GENI
Science DMZ as a Service: Creating Science Super- Facilities with GENIScience DMZ as a Service: Creating Science Super- Facilities with GENI
Science DMZ as a Service: Creating Science Super- Facilities with GENIUS-Ignite
 
network basics
network basicsnetwork basics
network basicsAvin Ash
 
Active Nets Technology Transfer through High-Performance Network Devices
Active Nets Technology Transfer through High-Performance Network DevicesActive Nets Technology Transfer through High-Performance Network Devices
Active Nets Technology Transfer through High-Performance Network DevicesTal Lavian Ph.D.
 

Similar a Tutorial: Maximizing Performance and Network Utility with a Science DMZ (20)

The Science DMZ
The Science DMZThe Science DMZ
The Science DMZ
 
Future services on Janet
Future services on JanetFuture services on Janet
Future services on Janet
 
Intternetworking With TCP/IP
Intternetworking With TCP/IPIntternetworking With TCP/IP
Intternetworking With TCP/IP
 
Network Engineering for High Speed Data Sharing
Network Engineering for High Speed Data SharingNetwork Engineering for High Speed Data Sharing
Network Engineering for High Speed Data Sharing
 
Science DMZ
Science DMZScience DMZ
Science DMZ
 
Networking lec1 4
Networking lec1 4Networking lec1 4
Networking lec1 4
 
Common Design Elements for Data Movement Eli Dart
Common Design Elements for Data Movement Eli DartCommon Design Elements for Data Movement Eli Dart
Common Design Elements for Data Movement Eli Dart
 
Data Mobility Exhibition
Data Mobility ExhibitionData Mobility Exhibition
Data Mobility Exhibition
 
Networking Fundamentals.ppt
Networking Fundamentals.pptNetworking Fundamentals.ppt
Networking Fundamentals.ppt
 
Sky x technology
Sky x technologySky x technology
Sky x technology
 
Datacom module 2: Data Communication Architecture, Protocols, and Standards
Datacom module 2:  Data Communication Architecture, Protocols, and StandardsDatacom module 2:  Data Communication Architecture, Protocols, and Standards
Datacom module 2: Data Communication Architecture, Protocols, and Standards
 
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...
Research data zone: veilige en geoptimaliseerde netwerkomgeving voor onderzoe...
 
Introduction to data communications
Introduction to data communicationsIntroduction to data communications
Introduction to data communications
 
Basic networking
Basic networkingBasic networking
Basic networking
 
EC8551 COMMUNICATION NETWORKS
EC8551 COMMUNICATION NETWORKSEC8551 COMMUNICATION NETWORKS
EC8551 COMMUNICATION NETWORKS
 
Science DMZ security
Science DMZ securityScience DMZ security
Science DMZ security
 
Big Data Analytics and Advanced Computer Networking Scenarios
Big Data Analytics and Advanced Computer Networking ScenariosBig Data Analytics and Advanced Computer Networking Scenarios
Big Data Analytics and Advanced Computer Networking Scenarios
 
Science DMZ as a Service: Creating Science Super- Facilities with GENI
Science DMZ as a Service: Creating Science Super- Facilities with GENIScience DMZ as a Service: Creating Science Super- Facilities with GENI
Science DMZ as a Service: Creating Science Super- Facilities with GENI
 
network basics
network basicsnetwork basics
network basics
 
Active Nets Technology Transfer through High-Performance Network Devices
Active Nets Technology Transfer through High-Performance Network DevicesActive Nets Technology Transfer through High-Performance Network Devices
Active Nets Technology Transfer through High-Performance Network Devices
 

Más de Globus

Advanced Globus System Administration Topics
Advanced Globus System Administration TopicsAdvanced Globus System Administration Topics
Advanced Globus System Administration TopicsGlobus
 
Instrument Data Automation: The Life of a Flow
Instrument Data Automation: The Life of a FlowInstrument Data Automation: The Life of a Flow
Instrument Data Automation: The Life of a FlowGlobus
 
Building Research Applications with Globus PaaS
Building Research Applications with Globus PaaSBuilding Research Applications with Globus PaaS
Building Research Applications with Globus PaaSGlobus
 
Reliable, Remote Computation at All Scales
Reliable, Remote Computation at All ScalesReliable, Remote Computation at All Scales
Reliable, Remote Computation at All ScalesGlobus
 
Best Practices for Data Sharing Using Globus
Best Practices for Data Sharing Using GlobusBest Practices for Data Sharing Using Globus
Best Practices for Data Sharing Using GlobusGlobus
 
An Introduction to Globus for Researchers
An Introduction to Globus for ResearchersAn Introduction to Globus for Researchers
An Introduction to Globus for ResearchersGlobus
 
Introduction to Research Automation with Globus
Introduction to Research Automation with GlobusIntroduction to Research Automation with Globus
Introduction to Research Automation with GlobusGlobus
 
Globus for System Administrators
Globus for System AdministratorsGlobus for System Administrators
Globus for System AdministratorsGlobus
 
Introduction to Globus for System Administrators
Introduction to Globus for System AdministratorsIntroduction to Globus for System Administrators
Introduction to Globus for System AdministratorsGlobus
 
Introduction to Data Transfer and Sharing for Researchers
Introduction to Data Transfer and Sharing for ResearchersIntroduction to Data Transfer and Sharing for Researchers
Introduction to Data Transfer and Sharing for ResearchersGlobus
 
Introduction to the Globus Platform for Developers
Introduction to the Globus Platform for DevelopersIntroduction to the Globus Platform for Developers
Introduction to the Globus Platform for DevelopersGlobus
 
Introduction to the Command Line Interface (CLI)
Introduction to the Command Line Interface (CLI)Introduction to the Command Line Interface (CLI)
Introduction to the Command Line Interface (CLI)Globus
 
Automating Research Data with Globus Flows and Compute
Automating Research Data with Globus Flows and ComputeAutomating Research Data with Globus Flows and Compute
Automating Research Data with Globus Flows and ComputeGlobus
 
Automating Research Data Flows and Introduction to the Globus Platform
Automating Research Data Flows and Introduction to the Globus PlatformAutomating Research Data Flows and Introduction to the Globus Platform
Automating Research Data Flows and Introduction to the Globus PlatformGlobus
 
Advanced Globus System Administration
Advanced Globus System AdministrationAdvanced Globus System Administration
Advanced Globus System AdministrationGlobus
 
Introduction to Globus for System Administrators
Introduction to Globus for System AdministratorsIntroduction to Globus for System Administrators
Introduction to Globus for System AdministratorsGlobus
 
Introduction to Globus for New Users
Introduction to Globus for New UsersIntroduction to Globus for New Users
Introduction to Globus for New UsersGlobus
 
Working with Globus Platform Services and Portals
Working with Globus Platform Services and PortalsWorking with Globus Platform Services and Portals
Working with Globus Platform Services and PortalsGlobus
 
Globus Automation
Globus AutomationGlobus Automation
Globus AutomationGlobus
 
Advanced Globus System Administration
Advanced Globus System AdministrationAdvanced Globus System Administration
Advanced Globus System AdministrationGlobus
 

Más de Globus (20)

Advanced Globus System Administration Topics
Advanced Globus System Administration TopicsAdvanced Globus System Administration Topics
Advanced Globus System Administration Topics
 
Instrument Data Automation: The Life of a Flow
Instrument Data Automation: The Life of a FlowInstrument Data Automation: The Life of a Flow
Instrument Data Automation: The Life of a Flow
 
Building Research Applications with Globus PaaS
Building Research Applications with Globus PaaSBuilding Research Applications with Globus PaaS
Building Research Applications with Globus PaaS
 
Reliable, Remote Computation at All Scales
Reliable, Remote Computation at All ScalesReliable, Remote Computation at All Scales
Reliable, Remote Computation at All Scales
 
Best Practices for Data Sharing Using Globus
Best Practices for Data Sharing Using GlobusBest Practices for Data Sharing Using Globus
Best Practices for Data Sharing Using Globus
 
An Introduction to Globus for Researchers
An Introduction to Globus for ResearchersAn Introduction to Globus for Researchers
An Introduction to Globus for Researchers
 
Introduction to Research Automation with Globus
Introduction to Research Automation with GlobusIntroduction to Research Automation with Globus
Introduction to Research Automation with Globus
 
Globus for System Administrators
Globus for System AdministratorsGlobus for System Administrators
Globus for System Administrators
 
Introduction to Globus for System Administrators
Introduction to Globus for System AdministratorsIntroduction to Globus for System Administrators
Introduction to Globus for System Administrators
 
Introduction to Data Transfer and Sharing for Researchers
Introduction to Data Transfer and Sharing for ResearchersIntroduction to Data Transfer and Sharing for Researchers
Introduction to Data Transfer and Sharing for Researchers
 
Introduction to the Globus Platform for Developers
Introduction to the Globus Platform for DevelopersIntroduction to the Globus Platform for Developers
Introduction to the Globus Platform for Developers
 
Introduction to the Command Line Interface (CLI)
Introduction to the Command Line Interface (CLI)Introduction to the Command Line Interface (CLI)
Introduction to the Command Line Interface (CLI)
 
Automating Research Data with Globus Flows and Compute
Automating Research Data with Globus Flows and ComputeAutomating Research Data with Globus Flows and Compute
Automating Research Data with Globus Flows and Compute
 
Automating Research Data Flows and Introduction to the Globus Platform
Automating Research Data Flows and Introduction to the Globus PlatformAutomating Research Data Flows and Introduction to the Globus Platform
Automating Research Data Flows and Introduction to the Globus Platform
 
Advanced Globus System Administration
Advanced Globus System AdministrationAdvanced Globus System Administration
Advanced Globus System Administration
 
Introduction to Globus for System Administrators
Introduction to Globus for System AdministratorsIntroduction to Globus for System Administrators
Introduction to Globus for System Administrators
 
Introduction to Globus for New Users
Introduction to Globus for New UsersIntroduction to Globus for New Users
Introduction to Globus for New Users
 
Working with Globus Platform Services and Portals
Working with Globus Platform Services and PortalsWorking with Globus Platform Services and Portals
Working with Globus Platform Services and Portals
 
Globus Automation
Globus AutomationGlobus Automation
Globus Automation
 
Advanced Globus System Administration
Advanced Globus System AdministrationAdvanced Globus System Administration
Advanced Globus System Administration
 

Último

The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxBkGupta21
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
Sample pptx for embedding into website for demo
Sample pptx for embedding into website for demoSample pptx for embedding into website for demo
Sample pptx for embedding into website for demoHarshalMandlekar2
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESmohitsingh558521
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersRaghuram Pandurangan
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...AliaaTarek5
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate Agents
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate AgentsRyan Mahoney - Will Artificial Intelligence Replace Real Estate Agents
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate AgentsRyan Mahoney
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 

Último (20)

The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptx
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
Sample pptx for embedding into website for demo
Sample pptx for embedding into website for demoSample pptx for embedding into website for demo
Sample pptx for embedding into website for demo
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Generative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information DevelopersGenerative AI for Technical Writer or Information Developers
Generative AI for Technical Writer or Information Developers
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate Agents
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate AgentsRyan Mahoney - Will Artificial Intelligence Replace Real Estate Agents
Ryan Mahoney - Will Artificial Intelligence Replace Real Estate Agents
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 

Tutorial: Maximizing Performance and Network Utility with a Science DMZ

  • 1. National Science Foundation Award #1826994 Jason Zurawski zurawski@es.net ESnet / Lawrence Berkeley National Laboratory http://bit.ly/2vx2HiQ Maximizing Performance and Network Utility with a Science DMZ Globusworld 2019 Chicago, IL May 2nd, 2019 https://epoc.global
  • 2. Outline • Motivation • TCP Basics • DMZ Design Patterns • Integration w/ Science • Conclusions 2 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 3. The Science DMZ in 1 Slide Consists of four key components, all required: • “Friction free” network path • Highly capable network devices (wire-speed, deep queues) • Virtual circuit connectivity option • Security policy and enforcement specific to science workflows • Located at or near site perimeter if possible • Dedicated, high-performance Data Transfer Nodes (DTNs) • Hardware, operating system, libraries all optimized for transfer • Includes optimized data transfer tools such as Globus and GridFTP • Performance measurement/test node • perfSONAR • Once it’s up, users often need training – Engagement is key Details at http://fasterdata.es.net/science-dmz/ 3 – ESnet Science Engagement (engage@es.net) - 5/3/19 © 2013 Wikipedia © 2013 Globus
  • 4. Michael Sinatra (ESnet) The Science DMZ has two main purposes: 1.A security architecture which allows for better segmentation of risks, and more granular application of controls to those segmented risks. In the case of the Science DMZ, the goal is to limit the segmented risk profile to that of a dedicated data transfer host. 2.An attempt to streamline tuning and troubleshooting by removing degrees-of-freedom (from both a reliability and performance perspective) from the active data transfer path. 4 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 5. The Science DMZ in 1 Picture 5 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 6. Outline • Motivation • TCP Basics • DMZ Design Patterns • Integration w/ Science • Conclusions 6 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 7. Science DMZ Background • The data mobility performance requirements for data intensive science are beyond what can typically be achieved using traditional methods • Default host configurations (TCP, filesystems, NICs) • Converged network architectures designed for commodity traffic • Conventional security tools and policies • Legacy data transfer tools (e.g. SCP) • Wait-for-trouble-ticket operational models for network performance 7 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 8. Science DMZ Background • The Science DMZ model describes a performance-based approach • Dedicated infrastructure for wide-area data transfer • Well-configured data transfer hosts with modern tools • Capable network devices • High-performance data path which does not traverse commodity LAN • Proactive operational models that enable performance • Well-deployed test and measurement tools (perfSONAR) • Periodic testing to locate issues instead of waiting for users to complain • Security posture well-matched to high-performance science applications 8 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 9. TCP – Ubiquitous and Fragile • Networks provide connectivity between hosts – how do hosts see the network? • From an application’s perspective, the interface to “the other end” is a socket • Communication is between applications – mostly over TCP • TCP – the fragile workhorse • TCP is (for very good reasons) timid – packet loss is interpreted as congestion • Packet loss in conjunction with latency is a performance killer • Like it or not, TCP is used for the vast majority of data transfer applications (more than 95% of ESnet traffic is TCP) 9 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 10. Packet/Data Loss?! • “Wait a minute! I thought TCP was a reliable protocol? What do you mean ‘packet loss’, where is the data going!?” • The data isn’t loss forever, its dropped somewhere on the path. • Usually by a device without enough buffer space to accept it, or by someone who thinks the data is corrupted and they won’t send it • Once its dropped, we have a way of knowing its been dropped. • TCP is reliable, each end is keeping track of what was sent, and what was received. • If something goes missing, its resent. Resending is what takes the time, and causes the slowdown. • TCP is able to reliably and transparently recover from packet loss by retransmitting any/all lost packets • This is how it provides a reliable data transfer services to the applications • The reliability mechanisms dramatically reduce performance when they are exercised • We want to eliminate the causes of packet loss – so that we don’t need to test out the (slow) way that TCP can recover. • What is the impact of that recovery? 10 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 11. A small amount of packet loss makes a huge difference in TCP performance Metro Area Local (LAN) Regional Continental International Measured (TCP Reno) Measured (HTCP) Theoretical (TCP Reno) Measured (no loss) With loss, high performance beyond metro distances is essentially impossible 11 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 12. How Do We Accommodate TCP? • High-performance wide area TCP flows must get loss- free service • Sufficient bandwidth to avoid congestion • Deep enough buffers in routers and switches to handle bursts • Especially true for long-distance flows due to packet behavior • No, this isn’t buffer bloat 12 – ESnet Science Engagement (engage@es.net) - 5/3/19 • Equally important – the infrastructure must be verifiable so that clean service can be provided – Stuff breaks • Hardware, software, optics, bugs, … • How do we deal with it in a production environment? – Must be able to prove a network device or path is functioning correctly • Regular active tests should be run - perfSONAR – Small footprint is a huge win • Fewer the number of devices = easier to locate the source of packet loss © 2013 icanhascheezburger.com
  • 13. Outline • Motivation • TCP Basics • DMZ Design Patterns • Integration w/ Science • Conclusions 13 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 14. Science DMZ Takes Many Forms • There are a lot of ways to combine these things – it all depends on what you need to do • Small installation for a project or two • Facility inside a larger institution • Institutional capability serving multiple departments/divisions • Science capability that consumes a majority of the infrastructure • Some of these are straightforward, others are less obvious • Key point of concentration: eliminate sources of packet loss / packet friction 14 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 15. Ad Hoc DTN Deployment 15 – ESnet Science Engagement (engage@es.net) - 5/3/19 10GE10G Site Border Router WAN Building or Wiring Closet Switch/Router Perimeter Firewall Site / Campus LAN High performance Data Transfer Node with high-speed storage Global security policy mixes rules for science and business traffic DTN traffic subject to firewall limitations perfSONAR Test and measurement not aligned with data resource placement DTN traffic subject to limitations of general-purpose networking equipment/config Note: Site border router and perimeter firewall are often the same device Conflicting requirements result in performance compromises
  • 16. Legacy Method: Ad Hoc DTN Deployment • This is often what gets tried first • Data transfer node deployed where the owner has space • This is often the easiest thing to do at the time • Straightforward to turn on, hard to achieve performance 16 – ESnet Science Engagement (engage@es.net) - 5/3/19 • If lucky, perfSONAR is at the border – This is a good start – Need a second one next to the DTN • Entire LAN path has to be sized for data flows • Entire LAN path is part of any troubleshooting exercise • This usually fails to provide the necessary performance.
  • 17. A better approach: simple Science DMZ 10GE 10GE 10GE 10GE 10G Border Router WAN Science DMZ Switch/Router Enterprise Border Router/Firewall Site / Campus LAN High performance Data Transfer Node with high-speed storage Per-service security policy control points Clean, High-bandwidth WAN path Site / Campus access to Science DMZ resources perfSONAR perfSONAR perfSONAR 17 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 18. Small-scale Science DMZ Deployment • Add-on to existing network infrastructure • All that is required is a port on the border router • Small footprint, pre-production commitment • Easy to experiment with components and technologies • DTN prototyping • perfSONAR testing • Limited scope makes security policy exceptions easy • Only allow traffic from partners • Add-on to production infrastructure – lower risk 18 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 19. Prototype Science DMZ Data Path 10GE 10GE 10GE 10GE 10G Border Router WAN Science DMZ Switch/Router Enterprise Border Router/Firewall Site / Campus LAN High performance Data Transfer Node with high-speed storage Per-service security policy control points Clean, High-bandwidth WAN path Site / Campus access to Science DMZ resources perfSONAR perfSONAR High Latency WAN Path Low Latency LAN Path 19 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 20. Support For Multiple Projects • Science DMZ architecture allows multiple projects to put DTNs in place • Modular architecture • Centralized location for data servers • This may or may not work well depending on institutional politics • Issues such as physical security can make this a non-starter • On the other hand, some shops already have service models in place • On balance, this can provide a cost savings – it depends • Central support for data servers vs. carrying data flows • How far do the data flows have to go? 20 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 21. Multiple Projects 10GE 10GE 10GE 10G Border Router WAN Science DMZ Switch/Router Enterprise Border Router/Firewall Site / Campus LAN Project A DTN Per-project security policy control points Clean, High-bandwidth WAN path Site / Campus access to Science DMZ resources perfSONAR perfSONAR Project B DTN Project C DTN 21 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 22. Supercomputer Center Deployment • High-performance networking is assumed in this environment • Data flows between systems, between systems and storage, wide area, etc. • Global filesystem often ties resources together • Portions of this may not run over Ethernet (e.g. IB) • Implications for Data Transfer Nodes • “Science DMZ” may not look like a discrete entity here • By the time you get through interconnecting all the resources, you end up with most of the network in the Science DMZ • This is as it should be – the point is appropriate deployment of tools, configuration, policy control, etc. • Office networks can look like an afterthought, but they aren’t • Deployed with appropriate security controls • Office infrastructure need not be sized for science traffic 22 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 23. Supercomputer Center Data Path Virtual Circuit Routed Border Router WAN Core Switch/Router Firewall Offices perfSONAR perfSONAR perfSONAR Supercomputer Parallel Filesystem Front end switch Data Transfer Nodes Front end switch High Latency WAN Path Low Latency LAN Path High Latency VC Path 23 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 24. Distributed Science DMZ • Fiber-rich environment enables distributed Science DMZ • No need to accommodate all equipment in one location • Allows the deployment of institutional science service • WAN services arrive at the site in the normal way • Dark fiber distributes connectivity to Science DMZ services throughout the site • Departments with their own networking groups can manage their own local Science DMZ infrastructure • Facilities or buildings can be served without building up the business network to support those flows • Security is potentially more complex • Remote infrastructure must be monitored • Solutions depend on relationships with security groups 24 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 25. Distributed Science DMZ – Dark Fiber Dark Fiber Dark Fiber 10GE Dark Fiber 10GE 10GE 10G Border Router WAN Science DMZ Switch/Router Enterprise Border Router/Firewall Site / Campus LAN Per-project security policy control points Clean, High-bandwidth WAN path Site / Campus access to Science DMZ resources perfSONAR perfSONAR Project A DTN (remote) Project B DTN (remote) Project C DTN (remote) 25 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 26. Common Threads• Two common threads exist in all these examples • Accommodation of TCP • Wide area portion of data transfers traverses purpose-built path • High performance devices that don’t drop packets • Ability to test and verify • When problems arise (and they always will), they can be solved if the infrastructure is built correctly • Small device count makes it easier to find issues • Multiple test and measurement hosts provide multiple views of the data path • perfSONAR nodes at the site and in the WAN • perfSONAR nodes at the remote site 26 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 27. Equipment – Routers and Switches • Requirements for Science DMZ gear are different • No need to go for the kitchen sink list of services • A Science DMZ box only needs to do a few things, but do them well • Support for the latest LAN integration magic with your Windows Active Directory environment is probably not super-important • A clean architecture is important • How fast can a single flow go? • Are there any components that go slower than interface wire speed? • There is a temptation to go cheap • Hey, it only needs to do a few things, right? • You typically don’t get what you don’t pay for • (You sometimes don’t get what you pay for either) 27 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 28. Common Circumstance: Multiple Ingress Data Flows, Common Egress Background traffic or competing bursts DTN traffic with wire-speed bursts 10GE 10GE 10GE Hosts will typically send packets at the speed of their interface (1G, 10G, etc.) • Instantaneous rate, not average rate • If TCP has window available and data to send, host sends until there is either no data or no window Hosts moving big data (e.g. DTNs) can send large bursts of back-to-back packets • This is true even if the average rate as measured over seconds is slower (e.g. 4Gbps) • On microsecond time scales, there is often congestion • Router or switch must queue packets or drop them 28 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 29. Rant Ahead 29 – ESnet Science Engagement (engage@es.net) - 5/3/19 N.B. You are entering into rant territory on the matter of switch buffering. If you are going to take away anything from the next section: 1. Under buffered network devices are the single greatest threat to data intensive use of the network. You can make hosts, operating systems, and application choices perform better for ‘free’, it will cost $$$ to fix a crappy switch or router 2. You will be steered toward non-optimal choices when you talk with the vendor community because they don’t understand simple math (but by the end of this, you will). 3. A 1U/2U data center/racklan network device should never be in the path of your data intensive network use case. 4. Non-passive (e.g. stateful) security devices are the same for buffering, and are actually worse due to the processing overhead. 5. Anytime you jump around the OSI stack – add friction (e.g. routing when you don’t need to, application layer inspection, etc.)
  • 30. All About That Buffer (No Cut Through) 30 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 31. All About That Buffer (No Cut Through) 31 – ESnet Science Engagement (engage@es.net) - 5/3/19 • Data arrives from multiple sources • Buffers have a finite amount of memory • Some have this per interface • Others may have access to a shared memory region with other interfaces • The processing engine will: • Extract each packet/frame from the queues • Pull off header information to see where the destination should be • Move the packet/frame to the correct output queue • Additional delay is possible as the queues physically write the packet to the transport medium (e.g. optical interface, copper interface)
  • 32. All About That Buffer (No Cut Through) 32 – ESnet Science Engagement (engage@es.net) - 5/3/19 • The Bandwidth Delay Product • The amount of “in flight” data for a TCP connection (BDP = bandwidth * round trip time) • Example: 10Gb/s cross country, ~100ms • 10,000,000,000 b/s * .1 s = 1,000,000,000 bits • 1,000,000,000 / 8 = 125,000,000 bytes • 125,000,000 bytes / (1024*1024) ~ 125MB • Ignore the math aspect: its making sure there is memory to catch and send packets • As the speed increases, there are more packets. • If there is not memory, we drop them, and that makes TCP sad.
  • 33. All About That Buffer (No Cut Through) 33 – ESnet Science Engagement (engage@es.net) - 5/3/19 • Buffering isn’t as important on the LAN (this is why you are normally pressured to buy ‘cut through’ devices) • Change the math to make the Latency 1ms = 1.25MB • ‘Cut through’ and low latency switches are designed for the data center, and can handle typical data center loads that don’t require buffering (e.g. same to same speeds, destinations within the broadcast domain) • Buffering *MATTERS* for WAN Transfers • Placing something with inadequate buffering in the path reduces the buffer for the entire path. E.g. if you have an expectation of 10Gbps over 100ms – don’t place a 12MB buffer anywhere in there – your reality is now ~10x less than it was before (e.g. 10Gbps @ 10ms, or 1Gbps @ 100ms) • Ignore the math aspect, its really just about making sure there is memory to catch packets. As the speed increases, there are more packets. If there is not memory, we drop them, and that makes TCP sad. • Memory on hosts, and network gear
  • 34. All About That Buffer (No Cut Through) [ ID] Interval Transfer Bandwidth Retr Cwnd [ 14] 0.00-1.00 sec 524 KBytes 4.29 Mbits/sec 0 157 KBytes [ 14] 1.00-2.00 sec 3.31 MBytes 27.8 Mbits/sec 0 979 KBytes [ 14] 2.00-3.00 sec 17.7 MBytes 148 Mbits/sec 0 5.36 MBytes [ 14] 3.00-4.00 sec 18.8 MBytes 157 Mbits/sec 214 1.77 MBytes [ 14] 4.00-5.00 sec 11.2 MBytes 94.4 Mbits/sec 0 1.88 MBytes [ 14] 5.00-6.00 sec 10.0 MBytes 83.9 Mbits/sec 0 2.39 MBytes [ 14] 6.00-7.00 sec 16.2 MBytes 136 Mbits/sec 0 3.63 MBytes [ 14] 7.00-8.00 sec 23.8 MBytes 199 Mbits/sec 0 5.50 MBytes [ 14] 8.00-9.00 sec 38.8 MBytes 325 Mbits/sec 0 8.23 MBytes [ 14] 9.00-10.00 sec 57.5 MBytes 482 Mbits/sec 0 11.8 MBytes [ 14] 10.00-11.00 sec 81.2 MBytes 682 Mbits/sec 0 16.2 MBytes [ 14] 11.00-12.00 sec 50.0 MBytes 419 Mbits/sec 35 3.93 MBytes [ 14] 12.00-13.00 sec 15.0 MBytes 126 Mbits/sec 0 2.20 MBytes [ 14] 13.00-14.00 sec 11.2 MBytes 94.4 Mbits/sec 0 2.53 MBytes [ 14] 14.00-15.00 sec 13.8 MBytes 115 Mbits/sec 1 1.50 MBytes [ 14] 15.00-16.00 sec 6.25 MBytes 52.4 Mbits/sec 5 813 KBytes [ 14] 16.00-17.00 sec 5.00 MBytes 41.9 Mbits/sec 0 909 KBytes [ 14] 17.00-18.00 sec 5.00 MBytes 41.9 Mbits/sec 0 1.37 MBytes [ 14] 18.00-19.00 sec 10.0 MBytes 83.9 Mbits/sec 0 2.43 MBytes [ 14] 19.00-20.00 sec 17.5 MBytes 147 Mbits/sec 0 4.22 MBytes 34 – ESnet Science Engagement (engage@es.net) - 5/3/19 • What does this “look” like to a data transfer? Consider the test of iperf below – See TCP ‘ramp up’ and slowly increase the window – When something in the path has no more space for packets – a drop occurs. TCP will eventually react to the lost packet, and ‘back off’ – In the example, this first occurs when we reach a buffer of around 6-8MB. Then after backoff the window is halved a couple of times – This happens again later – at a slightly higher buffer limit. This could be because there was cross traffic the first time, etc.
  • 35. Decoding Specifications • “The buffering behaviors of the switches and their operating system, such as behavior under memory stress, are typically proprietary information and not well documented” http://www.measurementlab.net/blog/traffic-microbursts-and-their-effect-on-internet-measurement/ • “Even if you know how much packet buffer is in the switch, assumptions on how it is deployed that are not backed up by testing can lead to unhappiness. What we like to say is that is the job of the network engineers to move bottlenecks around.” • Jim Warner • http://people.ucsc.edu/~warner/buffer.html 35 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 36. Decoding Specifications • So lets say the spec sheet says this: • What does ‘module’ mean? • Typically this means the amount of memory for the entire switch (if it’s a single unit) or a blade (if the chassis supports more than one. • BUT … this memory can be allocated in a number of different ways: • Shared between all ports • Dedicated (smaller) amounts per-port • Shared between ASICS, which control a bank of ports 36 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 37. Decoding Specifications •Consider this architecture • 48 Ports • 12 ASICS • 4 Ports per ASIC • 72MB total • 6MB per ASIC • If all ports are in use – expect that each port has access to 1.5MB. If only one is in use, it can use 6MB • Additional memory is often available in a ‘burst buffer’ in the fabric 37 – ESnet Science Engagement (engage@es.net) - 5/3/19 ASIC = application-specific integrated circuit, think 'small routing engine'
  • 38. Decoding Specifications • Recall: https://www.switch.ch/network/tools/tcp_throughput/ • What does 6MB get you? • 1Gbps @ <= 48ms (e.g. ½ needed for coast-to-coast) • 10Gbps @ <= 4.8ms (e.g. metro area) • What does 1.5MB get you? • 1Gbps @ <= 12ms (e.g. regional area) • 10Gbps @ <= 1.2ms (e.g. data center [or more accurately, rack or row)] • In either case – remember this assumes you are the only thing using that memory … congestion is a more likely reality 38 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 39. Outline • Motivation • TCP Basics • DMZ Design Patterns • Integration w/ Science • Conclusions 39 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 40. NCAR RDA Data Portal • Let’s say I have a nice compute allocation at NERSC – climate science • Let’s say I need some data from NCAR for my project • https://rda.ucar.edu/ • Data sets (there are many more, but these are two): • https://rda.ucar.edu/datasets/ds199.1/ (1.5TB) • https://rda.ucar.edu/datasets/ds313.0/ (430GB) • Download to NERSC (could also do ALCF or NCSA or OLCF)
  • 41. NCAR RDA Performance to DOE HPC Facilities 13.9 Gbps 16.6 Gbps 11.9 Gbps DTN nersc#dtn NERSC DTN olcf#dtn_atlas OLCF DTN alcf#dtn_mira ALCF DTN NCAR RDA rda#datashare • 1.5TB data set • 1121 files 41 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 42. Next Steps – Building On The Science DMZ • Enhanced cyberinfrastructure substrate now exists • Wide area networks (ESnet, GEANT, Internet2, Regionals) • Science DMZs connected to those networks • DTNs in the Science DMZs • What does the scientist see? • Scientist sees a science application • Data transfer • Data portal • Data analysis • Science applications are the user interface to networks and DMZs • Large-scale data-intensive science requires that we build larger structures on top of those components 42 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 43. HPC Facilities: The Petascale DTN Project • Built on top of the Science DMZ model • Effort to improve data transfer performance between the DOE ASCR HPC facilities at ANL, LBNL, and ORNL, and also NCSA. • Multiple current and future science projects need to transfer data between HPC facilities • Performance goal is 15 gigabits per second (equivalent to 1PB/week) • Realize performance goal for routine Globus transfers without special tuning • Reference data set is 4.4TB of cosmology simulation data 43 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 44. DTN Cluster Performance – HPC Facilities (2016) 6.8 Gbps 7.6 Gbps 6.9 Gbps 13.3 Gbps 6.0 Gbps 6.7 Gbps 11.1 Gbps 10.5 Gbps 7.3 Gbps 10.0 Gbps 13.4 Gbps 8.2 Gbps DTN DTN DTN DTN Data set: L380 Files: 19260 Directories: 211 Other files: 0 Total bytes: 4442781786482 (4.4T bytes) Smallest file: 0 bytes (0 bytes) Largest file: 11313896248 bytes (11G bytes) Size distribution: 1 - 10 bytes: 7 files 10 - 100 bytes: 1 files 100 - 1K bytes: 59 files 1K - 10K bytes: 3170 files 10K - 100K bytes: 1560 files 100K - 1M bytes: 2817 files 1M - 10M bytes: 3901 files 10M - 100M bytes: 3800 files 100M - 1G bytes: 2295 files 1G - 10G bytes: 1647 files 10G - 100G bytes: 3 files Petascale DTN Project March 2016 L380 Data Set NERSC DTN cluster Globus endpoint: nersc#dtn ALCF DTN cluster Globus endpoint: alcf#dtn_mira OLCF DTN cluster Globus endpoint: olcf#dtn_atlas NCSA DTN cluster Globus endpoint: ncsa#BlueWaters 44 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 45. DTN Cluster Performance – HPC Facilities (2017) 21.2/22.6/24.5 Gbps 23.1/33.7/39.7 Gbps 26.7/34.7/39.9 Gbps 33.2/43.4/50.3 Gbps 35.9/39.0/40.7 Gbps 29.9/33.1/35.5 Gbps 34.6/47.5/56.8 Gbps 44.1/46.8/48.4 Gbps 41.0/42.2/43.9 Gbps 33.0/35.0/37.8 Gbps 43.0/50.0/56.3 Gbps 55.4/56.7/57.4 Gbps DTN DTN DTN DTN NERSC DTN cluster Globus endpoint: nersc#dtn Filesystem: /project Data set: L380 Files: 19260 Directories: 211 Other files: 0 Total bytes: 4442781786482 (4.4T bytes) Smallest file: 0 bytes (0 bytes) Largest file: 11313896248 bytes (11G bytes) Size distribution: 1 - 10 bytes: 7 files 10 - 100 bytes: 1 files 100 - 1K bytes: 59 files 1K - 10K bytes: 3170 files 10K - 100K bytes: 1560 files 100K - 1M bytes: 2817 files 1M - 10M bytes: 3901 files 10M - 100M bytes: 3800 files 100M - 1G bytes: 2295 files 1G - 10G bytes: 1647 files 10G - 100G bytes: 3 files Petascale DTN Project November 2017 L380 Data Set Gigabits per second (min/avg/max), three transfers ALCF DTN cluster Globus endpoint: alcf#dtn_mira Filesystem: /projects OLCF DTN cluster Globus endpoint: olcf#dtn_atlas Filesystem: atlas2 NCSA DTN cluster Globus endpoint: ncsa#BlueWaters Filesystem: /scratch 45 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 46. Science Data Portals • Large repositories of scientific data • Climate data • Sky surveys (astronomy, cosmology) • Many others • Data search, browsing, access • Many scientific data portals were designed 15+ years ago • Single-web-server design • Data browse/search, data access, user awareness all in a single system • All the data goes through the portal server • In many cases by design • E.g. embargo before publication (enforce access control) 46 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 47. Legacy Portal Design 10GE Border Router WAN Firewall Enterprise perfSONAR perfSONAR Filesystem (data store) 10GE Portal Server Browsing path Query path Data path Portal server applications: · web server · search · database · authentication · data service • Very difficult to improve performance without architectural change – Software components all tangled together – Difficult to put the whole portal in a Science DMZ because of security – Even if you could put it in a DMZ, many components aren’t scalable • What does architectural change mean? 47 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 48. Architectural Examination of Data Portals • Common data portal functions (most portals have these) • Search/query/discovery • Data download method for data access • GUI for browsing by humans • API for machine access – ideally incorporates search/query + download • Performance pain is primarily in the data handling piece • Rapid increase in data scale eclipsed legacy software stack capabilities • Portal servers often stuck in enterprise network • Can we “disassemble” the portal and put the pieces back together better? • Use Science DMZ as a platform for the data piece • Avoid placing complex software in the Science DMZ 48 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 49. Legacy Portal Design 10GE Border Router WAN Firewall Enterprise perfSONAR perfSONAR Filesystem (data store) 10GE Portal Server Browsing path Query path Data path Portal server applications: · web server · search · database · authentication · data service 49 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 50. Next-Generation Portal Leverages Science DMZ 10GE10GE 10GE 10GE Border Router WAN Science DMZ Switch/Router Firewall Enterprise perfSONAR perfSONAR 10GE 10GE 10GE 10GE DTN DTN API DTNs (data access governed by portal) DTN DTN perfSONAR Filesystem (data store) 10GE Portal Server Browsing path Query path Portal server applications: · web server · search · database · authentication Data Path Data Transfer Path Portal Query/Browse Path 50 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 51. Put The Data On Dedicated Infrastructure • We have separated the data handling from the portal logic • Portal is still its normal self, but enhanced • Portal GUI, database, search, etc. all function as they did before • Query returns pointers to data objects in the Science DMZ • Portal is now freed from ties to the data servers (run it on Amazon if you want!) • Data handling is separate, and scalable • High-performance DTNs in the Science DMZ • Scale as much as you need to without modifying the portal software • Outsource data handling to computing centers • Computing centers are set up for large-scale data • Let them handle the large-scale data, and let the portal do the orchestration of data placement 51 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 52. JGI Data Portal 52 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 53. 53 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 54. Reasons To Scale Data Portals • Some reasons are obvious • Increase in size of data objects (MB à GB à 100s of GB) • Number of data objects (many thousands per data set) • Other reasons are paradigm shifts • Modern data analysis on HPC can use a *lot* of data • Today’s HPC facilities are far more capable than in the past • Retrofit / rebuilt data portals and data repositories • Significant wins from increased data analysis 54 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 55. Science at Scale: Genomics 55 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 56. Science at Scale: Climate 56 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 57. They Can Use All The Data • Groups like these need large data sets • Much of the data in their field is behind legacy portals • Significant human effort to retrieve what they need • Legacy systems perform poorly, especially at scale • Legacy data portals are a product of their time • We now live in the future from the perspective of those designs • Current systems far exceed the capabilities available 15 years ago • From the perspective of today’s systems, legacy portals are products of a bygone past • It is now perfectly reasonable for a scientist to want all the data • Machine learning + HPC • But this only works if the scientists can get to the data at scale 57 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 58. IT/Engineering (not the scientists) Have To Do This • The scientific community cannot do this for themselves • Individual researchers do not control the resources • Computing centers • Data repositories • Science networks • Our community owns these – we have to do the work • Integration, performance engineering, interoperability • Science Engagement to teach scientists how to use the better platforms • This is the path forward ¯_(ツ)_/¯ 58 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 59. Networks Cannot Do This Alone • We need a whole-community effort • Networks • HPC facilities • Data repositories / Data portals • Experimental facilities • Science collaborations • Science programs • Networks can help, and must be part of the conversation • Heavy lifting is now at the network edge, in collaboration with the network core • Need to get the architecture right – we know how to do this • Engagement is critical • We have to teach people to use the new thing • They aren’t going to magically find out about new services 59 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 60. Long-Term Vision ESnet (Big Science facilities, DOE labs) Internet2 + Regionals (US Universities and affiliated institutions) International Networks (Universities and labs in Europe, Asia, Americas, Australia, etc.) High-performance feature-rich science network ecosystem Commercial Clouds (Amazon, Google, Microsoft, etc.) Agency Networks (NASA, NOAA, etc.) Campus HPC + Data 60 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 61. It’s All A Bunch Of Science DMZs High-performance feature-rich science network ecosystem DTN DTN DTN DTN DMZDMZ DMZDMZ DTN DATA DTN DMZDMZ DTN DTN DMZDMZ DATA DTN DTN DMZDMZ Parallel Filesystem DTN DTN DTN DTNDMZDMZ DATA DTN DTN DMZDMZ Experiment Data Archive DTN DTN DTN DTN DTN DMZDMZ DTN DTN DMZDMZ DATA 61 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 62. It’s All A Bunch Of Science DMZs High-performance feature-rich science network ecosystem DTN DTN DTN DTN DMZDMZ DMZDMZ DTN DATA DTN DMZDMZ DTN DTN DMZDMZ DATA DTN DTN DMZDMZ Parallel Filesystem DTN DTN DTN DTNDMZDMZ DATA DTN DTN DMZDMZ Experiment Data Archive DTN DTN DTN DTN DTN DMZDMZ DTN DTN DMZDMZ DATA HPC Facilities Single Lab Experiments Data Portal LHC Experiments University Computing 62 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 63. Outline • Motivation • TCP Basics • DMZ Design Patterns • Integration w/ Science • Conclusions 63 – ESnet Science Engagement (engage@es.net) - 5/3/19
  • 64. ESnet’s Science DMZ – Enabling Big Data Science The Science DMZ design pattern supports secure, high-performance data movement between and among facilities, labs, universities, data portals, and clouds DevelopmentDeploymentPlatform 10GE10GE 10GE 10GE Border Router WAN Science DMZ Switch/Router Firewall Enterprise perfSONAR perfSONAR 10GE 10GE 10GE 10GE DTN DTN API DTNs (data access governed by portal) DTN DTN perfSONAR Filesystem (data store) 10GE Portal Server Browsing path Query path Portal server applications: · web server · search · database · authentication Data Path Data Transfer Path Portal Query/Browse Path Long Term Vision Development Support § Science DMZ supports development, prototyping, and incremental deployment of advanced network and data services § Petascale DTN Project § Ongoing production operation across science complex 2017-2018 § Modern Research Data Portal design pattern § Pacific Research Platform § Science DMZs deployed at DOE labs and facilities § NSF CC-NIE and successor programs fund over 100 Science DMZs at universities 2011-20162005-2010 § Requirements analysis of DOE Office of Science facilities and projects from network and data perspective § Performance engineering, troubleshooting, experience § Creation of Science DMZ design pattern Science Networks: Campus, Regional, National, International Science DMZ Science DMZ Science DMZ Science DMZ DTN DTN DTN DTN DTN DTN DTN DTN Data Platform Data Platform § Science Engagement model for science support § OIN workshop series trains CI engineers throughout the community 21.2/22.6/24.5 Gbps 23.1/33.7/39.7 Gbps 26.7/34.7/39.9 Gbps 33.2/43.4/50.3 Gbps 35.9/39.0/40.7 Gbps 29.9/33.1/35.5 Gbps 34.6/47.5/56.8 Gbps 44.1/46.8/48.4 Gbps 41.0/42.2/43.9 Gbps 33.0/35.0/37.8 Gbps 43.0/50.0/56.3 Gbps 55.4/56.7/57.4 Gbps DTN DTN DTN DTN NERSC DTN cluster Globus endpoint: nersc#dtn Filesystem: /project Data set: L380 Files: 19260 Directories: 211 Other files: 0 Total bytes: 4442781786482 (4.4T bytes) Smallest file: 0 bytes (0 bytes) Largest file: 11313896248 bytes (11G bytes) Size distribution: 1 - 10 bytes: 7 files 10 - 100 bytes: 1 files 100 - 1K bytes: 59 files 1K - 10K bytes: 3170 files 10K - 100K bytes: 1560 files 100K - 1M bytes: 2817 files 1M - 10M bytes: 3901 files 10M - 100M bytes: 3800 files 100M - 1G bytes: 2295 files 1G - 10G bytes: 1647 files 10G - 100G bytes: 3 files Petascale DTN Project November 2017 L380 Data Set Gigabits per second (min/avg/max), three transfers ALCF DTN cluster Globus endpoint: alcf#dtn_mira Filesystem: /projects OLCF DTN cluster Globus endpoint: olcf#dtn_atlas Filesystem: atlas2 NCSA DTN cluster Globus endpoint: ncsa#BlueWaters Filesystem: /scratch 10GE 10GE 10GE 10GE 10G Border Router WAN Science DMZ Switch/Router Enterprise Border Router/Firewall Site / Campus LAN High performance Data Transfer Node with high-speed storage Per-service security policy control points Clean, High-bandwidth WAN path Site / Campus access to Science DMZ resources perfSONAR perfSONAR High Latency WAN Path Low Latency LAN Path perfSONAR Production Data Movement § Science DMZs support large scale data movement and data placement across the science complex (DOE and collaborators) § Performant security Scalable Data Platform § Science DMZs provide foundation for data-driven science globally
  • 65. National Science Foundation Award #1826994 Jason Zurawski zurawski@es.net ESnet / Lawrence Berkeley National Laboratory http://bit.ly/2vx2HiQ Maximizing Performance and Network Utility with a Science DMZ Globusworld 2019 Chicago, IL May 2nd, 2019 https://epoc.global