SlideShare una empresa de Scribd logo
1 de 33
Descargar para leer sin conexión
Ingress Scheduling in OvS-DPDK
Billy O’Mahony – Intel
Jan Scheurich – Ericsson
November 16-17, 2017 | San Jose, CA
Introduction
u Use cases for traffic prioritization in NFV
u State of the art in OvS-DPDK datapath
u Rx queue prioritization in DPDK datapath
u Traffic classification and queue selection on NIC
u Next steps
Compute Node
OvS
br-intbr-ctrl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
Compute Node
OvS
Scenario: NFVI on Converged Data Center
VIM control plane sharing physical network with tenant data
ToR A ToR B
dpdk0 dpdk1
br-intbr-ctl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
VIM = Virtual Infrastructure Manager
For example OpenStack component: Nova,
Neutron services and their local agents
Compute Node
OvS
br-intbr-ctrl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
Compute Node
OvS
Use Case 1: In-band OvS Control Plane
ToR A ToR B
dpdk0 dpdk1
br-intbr-ctl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
LACP bond supervision
LACP = Link Aggregation Control Protocol
Here: in-band heart-beat between OVS and each ToR
Compute Node
OvS
br-intbr-ctrl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
Compute Node
OvS
Use Case 1: In-band OvS Control Plane
ToR A ToR B
dpdk0 dpdk1
br-intbr-ctl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
BFD tunnel monitoring BFD packets are
sent inside tunnel
BFD = Bidirectional Forwarding Detection
Here: Heart-beat between OVS instances
connected through tunnel mesh
Compute Node
OvS
br-intbr-ctrl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
Compute Node
OvS
Use Case 2: VIM Control Plane
ToR A ToR B
dpdk0 dpdk1
br-intbr-ctl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsNova,
Neutron, …
Host
networking
bond
br-prv
tag tag
VIM Control
Plane
VIM Control Plane
In OpenStack: RabbitMQ, REST APIs calls, …
Compute Node
OvS
br-intbr-ctrl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
Compute Node
OvS
Use Case 2: VIM Control Plane
ToR A ToR B
dpdk0 dpdk1
br-intbr-ctl
Tenant
VM
Tenant
VM
vhostuser
Local Agents
Local Agents
Local AgentsVIM
Components
Host
networking
bond
br-prv
tag tag
OvS control plane:
OpenFlow and OVSDB
OpenFlow and OVSDB are special cases of
VIM control plane
OvS
Status Quo in OvS DPDK Datapath
NIC
PMD
1
PMD
2
ovs-vswitchd
thread
Tenant VM
BFD
LACP
RSS
HW
Scheduler
VIM
Components
Host
networking
br-ctl
OvS
Scenario: PMD Overload
NIC
PMD
1
PMD
2
ovs-vswitchd
thread
Tenant VM
br-ctl
BFD
LACP
RSS
PMD cycles
exhausted through
tenant data
Rx queues full. Data
and control packets
dropped.
VIM
Components
Host
networking
OvS
Scenario: Egress Link Overload
NIC
PMD
1
PMD
2
ovs-vswitchd
thread
Tenant VM
BFD
LACP
RSS
VIM
Components
Host
networking
br-ctl
PMD Tx queues
full. Tenant data
being dropped.
Separate Tx queue for
ovs-vswitchd. HW
scheduler can provide
fair share of link band-
width
Egress link band-
width exhausted
by tenant data
Offered	load	on	physical	
port	[Kpps] 2000 2200 2400 2600 2800 3200 3600 4000
Offered	load	on	phy	port	
[Gbit/s] 1.54 1.69 1.84 2.00 2.15 2.46 2.76 3.07
PMD	overload	factor	[%] 0% 9% 18% 27% 45% 64% 82%
PMD	utilization	[%] 99.95% 99.99% 100% 100% 100% 100% 100% 100%
Phy	port	rx	drop	[%] 0% 0% 8% 15% 21% 31% 39% 45%
ping	-f	average	RTT	[ms] 0.45 0.50 3.02 3.03 3.15 3.10 3.69 3.95
ping	-f	packet	drop	[%] 0 0 10% 16% 21% 37% 45% 49%
BFD	flappings	[1/min] 1.85 3.75 5.66 5.71
Num	flaps 0 0 17 17 43 20
OpenFlow	connection	
timeouts	in	OVS 0 0 0 0 0 0 3 3
Connection	closed	by	
peer	(ODL) 20 15
Connection	reset	by	peer	
(ODL) 2 0
37
5.03
PMD	polling	physical	port	overloaded	with	64B	packets
u Packet drops in the Rx queue of
physical port equally affect tenant
data, BFD and OVS control plane
packets
u “ping –f“ to br-ctl interface to
quantify control plane impact
u Ping packet drop in line with overall
packet drop
u RTT jumps from 50 us to 3 ms
u BFD flapping occurs already at
moderate overload
u The rate increases with overload
u Above 45% packet drop the
OpenFlow control channel breaks
due to missed Echo Replies
Measurements
Impact of PMD Overload
source: Ericsson
CPU: Dual socket, Xeon CPU E5-2697 v3 @2.60GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x 10Gbit/s;
OvS: version 2.6, 1 PMD, 1 phy port, 1 vhostuser port; VM: TRex DPDK traffic source/sink
u Egress link overload does not affect
the control plane
u Outgoing packets are forwarded by
the ovs-vswitchd thread, which has
its dedicated TX queue in the
Fortville NIC
u The NIC schedules packets from each
of the TX queues in some fair
manner, so that the ovs-vswitchd
queue gets sufficient bandwidth on
the link
u Incoming packets are not affected as
neither link nor PMDs are overloaded
u No BFD flapping
Measurements
Impact of Egress Link Overload
Offered	load	from	VM	[Kpps] 800 900 1000 1200 1600
Offered	load	from	VM	[Gbit/s] 9.80 11.03 12.26 14.71 19.61
Transmitted	load	on	phy	port	[Gbit/s] 9.81 9.90 9.88 9.88 9.88
Link	overload 0% 11% 24% 49% 99%
PMD	utilization	[%] 41.45% 46.30% 50.20% 56.36% 69.89%
ping	-f	average	RTT	[ms] 0.109 0.205 0.206 0.210 0.204
ping	-f	packet	drop	[%] 0% 0% 0% 0% 0%
BFD	flappings	[1/min]
Num	flaps 0 0 0 0 0
OpenFlow	connection	timeouts	in	OVS 0 0 0 0 0
Connection	closed	by	peer	(ODL)
Connection	reset	by	peer	(ODL)
10G	Link	from	OvS	overloaded	with	outgoing	traffic	from	VM
(1500	byte	packets)
source: Ericsson
CPU: Dual socket, Xeon CPU E5-2697 v3 @2.60GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x 10Gbit/s;
OvS: version 2.6, 1 PMD, 1 phy port, 1 vhostuser port; VM: TRex DPDK traffic source/sink
Use Case 3: QoS for Tenant Data
All tenant data traffic is equal!?
Well, some packets are more equal than others!
u Virtual Network Functions send/receive a large variety of network traffic
u Top prio: Critical internal control plane (e.g. cluster membership)
u …
u Min prio: Bulk user plane
u VNFs need prioritization for their critical traffic in the NFVI
u How to orchestrate and implement the necessary QoS end-to-end?
u Will need additional priority levels and packet marking (e.g. IP Diffserv)
Desired Ingress Prioritization on
Physical Ports
u Priority 1: In-band control plane
u Untagged LACP packets
u BFD packets inside tunnel based on IP DSCP of outer IP header
u Priority 2: VIM control plane
u Certain prioritized VLAN tags
u Priority 3+: Prioritized tenant data
u E.g. based on IP DSCP of outer IP header
u Base Priority
u Non-prioritized traffic spread through RSS over multiple Rx queues
PMD
“Schedulers arrange and/or rearrange packets for output.”
-- http://www.tldp.org/HOWTO/html_single/Traffic-Control-HOWTO/#e-scheduling
Ingress Scheduling
RX Queue
à TX Queue
à
?
ovs-vswitchd
BFD
LACP
Priority packet – e.g
control plane.
Ingress Scheduling - Implementation
RX Queue x2
à
TX Queue
à
?
DPDK rte_flowAPI
installs rxq
assignment filters
on supported
vendors NICs
ovs-vswitchd
BFD
LACP
PMD empties
priority queue
before reading non-
priority queue
PMD
Ingress Scheduling - Implementation
1. Move packet prioritization decision to the NIC
2. Place prioritized packets on separate RX Queue
3. Read preferentially from “priority” RxQ. Keep it simple:
u Read from priority queue until it’s empty
u Service other queues
u Repeat
Ingress Scheduling – Latency effect
~99.9 % of packets
already have have a
latency <20us
There are x10 to
x50 less packets in
any given latency
bucket – good.
But worst case
latency does not
improve.
CPU: Dual socket, Xeon(R) CPU E5-2695 v3 @2.30GHz 14 core no-HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x
10Gbit/s;
OvS: version 2.7.90, 1 PMD, 2 phy port, Hardware trafficsource/sink
Source: Intel
Ingress Scheduling – Overload protection
RX Queue x2
à
TX Queue
à
L
?
ovs-vswitchd
BFD
LACP
PMD
1
dpdk1
Ingress Scheduling – Traffic Protection
u Overload PMD through 64 byte DPDK traffic on dpdk0
à 100% PMD load in pmd-stats-show
à 25% rx packet drop on dpdk0
u Add iperf3 UDP traffic (256 bytes) in parallel over dpdk1
u Measurement result:
SUT Server
OvS
VM
dpdk
testpmd
vhostuser
iperf3
udp
server
PMD
br-prv
ToR
dpdk0dpdk1
TGen Server
BM
dpdk
pktgen
iperf3
udp
client
dpdk0
Low priorityHigh/Low
priority
source: Ericsson
Condition 1
dpdk1
low priority
Condition 2
dpdk1
high priority
iperf3 UDP
throughput
not
measured
1 Gbit/s
460 Kpps 1)
iperf3 UDP
packet loss 28% 0%
1) iperf3 throughput limited by UDP/IP stack on client side
CPU: Dual socket, Xeon CPU E5-2680 v4 @2.40GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache
NIC: Intel Fortville X710, 4 x 10Gbit/s; OvS: version 2.6, 1 PMD, all ports and VM on NUMA node 0
u $ ovs-vsctl set Interface phy1
ingress_sched:
eth_type=0x8809
Ingress Scheduling – Configuration
Field as per ovs-fields(7)
and ofctl add-flow. Not
all netdevs/NICs will
support all combinations.
Single prioritization
condition.
u $ ovs-vsctl set Interface phy1
ingress_sched:
vlan_tci=0x1123/0x1fff,ip,ip_dscp=0x5
Ingress Scheduling – Configuration (future)
Several different
prioritization
conditions. a AND b.
u $ ovs-vsctl set Interface phy1
ingress_sched:
filter=vlan_tci=0x1123/0x1fff
filter=ip,ip_dscp=0x5
Ingress Scheduling – Configuration (future)
Several different
prioritization
conditions. a OR b.
u $ ovs-vsctl set Interface phy1
ingress_sched:
prio=1,
filter,vlan_tci=0x1123/0x1fff,
filter,eth_type=0x8809,
prio=2,
filter,ip,ip_dscp=0x5,
Ingress Scheduling – Configuration (future)
Traffic Priority Levels
Support several levels of
prioritization: High and Low
but also a Critical level for
instance.
u $ ovs-vsctl set Interface phy1
ingress_sched:
prio=2,
filter,ip,ip_dscp=0x5,
prio=1,
filter=vlan_tci=0x1123/0x1fff,
filter,eth_type=0x8809
Ingress Scheduling – Configuration (future)
Filter Priority:
Filter groups are
applied in the order in
which they appear on
the configuration line.
u ovsdb-schema
<table name="Interface"…
<column name="ingress_sched" key="err">
If the specified ingress scheduling could not be
applied, Open vSwitch sets this column to an error
description in human readable form. Otherwise, Open
vSwitch clears this column.
Ingress Scheduling –Error reporting
u $ ovs-vsctl set Interface phy1
options:n_rxq=4
u $ ovs-vsctl set Interface phy1
ingress_sched:
prio=2,
filter,ip,ip_dscp=0x5,
prio=1,
filter=vlan_tci=0x1123/0x1fff,
filter,eth_type=0x8809
Ingress Scheduling – RxQ’s & RSS
RSS queues
Additional Priority
Queues
Ingress Scheduling – Next Steps
PMD PMD PMD PMD
u Avoid poor rxq -> pmd assignment
Ingress Scheduling – Next Steps
u Use rte_flow API for offload
u Extend to several priorities
u Priorities of overlapping filters
u Multiple traffic priorities
u Working with RFC ‘Flow Offload’ feature…
u …
u Prioritization to the Guest…
Summary
u OvS-DPDK in NFVI context needs ingress scheduling to
protect priority traffic against PMD overload
u SW priority queue handling in the PMD loop is effective
u Could be upstreamed first, priority configurable per port
u Off-loading classification and queue selection to NIC
through rte_flow API allows generic solution
u Interaction with RFC Flow Classification Offload
u Work in progress
u Lots left to figure out
u We are open for suggestions/collaboration
Intel Notices & Disclaimers
Intel technologies’ features and benefits depend onsystem configurationand may require enabledhardware, software or service activation.
Performance varies dependingon system configuration. Check with your system manufacturer or retailer orlearn more at intel.com.
No computer system can be absolutely secure.
Tests document performance of components on a particular test, in specific systems. Differences in hardware, software,or configurationwill affect
actual performance. Consult othersources of informationto evaluate performance as you consider your purchase. For more complete information
about performance andbenchmark results, visit http://www.intel.com/benchmarks .
Software and workloads used in performance tests may have been optimizedfor performance onlyon Intel microprocessors. Performance tests, such
as SYSmark and MobileMark, are measuredusing specific computersystems, components, software,operations and functions. Anychange to any of
those factors may cause the results to vary. You should consult otherinformation andperformance tests to assist you in fully evaluating your
contemplated purchases, including the performance of that product when combinedwith other products. For more completeinformation visit
http://www.intel.com/benchmarks .
Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to
Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not
guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel.
Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not
specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference
Guides for more information regarding the specific instruction sets covered by this notice.
Cost reduction scenarios described are intendedas examples of how a given Intel-based product, in the specified circumstances and configurations,
may affect future costs and provide cost savings. Circumstances will vary. Intel does not guarantee any costs or cost reduction.
Intel does not control or audit third-partybenchmark data or the web sites referencedin this document. Youshouldvisit the referencedweb site and
confirm whether referenced data are accurate.
© 2017 Intel Corporation.
Intel, the Intel logo, and Intel Xeonare trademarks of Intel Corporationin the U.S. and/or other countries.
*Other names and brands may be claimedas property of others.
Thank You!
Questions?
References
1. [ovs-dev] [RFC PATCH 0/3] prioritizing latency sensitive traffic
u [ovs-dev,RFC,1/3] netdev: Add set_ingress_sched to netdev api
u [ovs-dev,RFC,2/3] netdev-dpdk: Apply ingress_sched config to dpdk phy ports
u [ovs-dev,RFC,3/3] dpif-netdev: Add rxq prioritization

Más contenido relacionado

La actualidad más candente

2015 FOSDEM - OVS Stateful Services
2015 FOSDEM - OVS Stateful Services2015 FOSDEM - OVS Stateful Services
2015 FOSDEM - OVS Stateful ServicesThomas Graf
 
DPDK Support for New HW Offloads
DPDK Support for New HW OffloadsDPDK Support for New HW Offloads
DPDK Support for New HW OffloadsNetronome
 
DPDK Summit 2015 - HP - Al Sanders
DPDK Summit 2015 - HP - Al SandersDPDK Summit 2015 - HP - Al Sanders
DPDK Summit 2015 - HP - Al SandersJim St. Leger
 
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIO
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIOLF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIO
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIOLF_OpenvSwitch
 
TC Flower Offload
TC Flower OffloadTC Flower Offload
TC Flower OffloadNetronome
 
Open vSwitch - Stateful Connection Tracking & Stateful NAT
Open vSwitch - Stateful Connection Tracking & Stateful NATOpen vSwitch - Stateful Connection Tracking & Stateful NAT
Open vSwitch - Stateful Connection Tracking & Stateful NATThomas Graf
 
Virtualized network with openvswitch
Virtualized network with openvswitchVirtualized network with openvswitch
Virtualized network with openvswitchSim Janghoon
 
The n00bs guide to ovs dpdk
The n00bs guide to ovs dpdkThe n00bs guide to ovs dpdk
The n00bs guide to ovs dpdkmarkdgray
 
Quality of Service Ingress Rate Limiting and OVS Hardware Offloads
Quality of Service Ingress Rate Limiting and OVS Hardware OffloadsQuality of Service Ingress Rate Limiting and OVS Hardware Offloads
Quality of Service Ingress Rate Limiting and OVS Hardware OffloadsNetronome
 
The Next Generation Firewall for Red Hat Enterprise Linux 7 RC
The Next Generation Firewall for Red Hat Enterprise Linux 7 RCThe Next Generation Firewall for Red Hat Enterprise Linux 7 RC
The Next Generation Firewall for Red Hat Enterprise Linux 7 RCThomas Graf
 
debugging openstack neutron /w openvswitch
debugging openstack neutron /w openvswitchdebugging openstack neutron /w openvswitch
debugging openstack neutron /w openvswitch어형 이
 
BPF & Cilium - Turning Linux into a Microservices-aware Operating System
BPF  & Cilium - Turning Linux into a Microservices-aware Operating SystemBPF  & Cilium - Turning Linux into a Microservices-aware Operating System
BPF & Cilium - Turning Linux into a Microservices-aware Operating SystemThomas Graf
 
Accelerating Neutron with Intel DPDK
Accelerating Neutron with Intel DPDKAccelerating Neutron with Intel DPDK
Accelerating Neutron with Intel DPDKAlexander Shalimov
 
OpenStack networking juno l3 h-a, dvr
OpenStack networking   juno l3 h-a, dvrOpenStack networking   juno l3 h-a, dvr
OpenStack networking juno l3 h-a, dvrSim Janghoon
 
LF_OVS_17_State of the OVN
LF_OVS_17_State of the OVNLF_OVS_17_State of the OVN
LF_OVS_17_State of the OVNLF_OpenvSwitch
 
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)Thomas Graf
 
Linux Networking Explained
Linux Networking ExplainedLinux Networking Explained
Linux Networking ExplainedThomas Graf
 
LinuxCon 2015 Linux Kernel Networking Walkthrough
LinuxCon 2015 Linux Kernel Networking WalkthroughLinuxCon 2015 Linux Kernel Networking Walkthrough
LinuxCon 2015 Linux Kernel Networking WalkthroughThomas Graf
 
Cilium - API-aware Networking and Security for Containers based on BPF
Cilium - API-aware Networking and Security for Containers based on BPFCilium - API-aware Networking and Security for Containers based on BPF
Cilium - API-aware Networking and Security for Containers based on BPFThomas Graf
 
Offloading TC Rules on OVS Internal Ports
Offloading TC Rules on OVS Internal Ports Offloading TC Rules on OVS Internal Ports
Offloading TC Rules on OVS Internal Ports Netronome
 

La actualidad más candente (20)

2015 FOSDEM - OVS Stateful Services
2015 FOSDEM - OVS Stateful Services2015 FOSDEM - OVS Stateful Services
2015 FOSDEM - OVS Stateful Services
 
DPDK Support for New HW Offloads
DPDK Support for New HW OffloadsDPDK Support for New HW Offloads
DPDK Support for New HW Offloads
 
DPDK Summit 2015 - HP - Al Sanders
DPDK Summit 2015 - HP - Al SandersDPDK Summit 2015 - HP - Al Sanders
DPDK Summit 2015 - HP - Al Sanders
 
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIO
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIOLF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIO
LF_OVS_17_Enabling Hardware Offload of OVS Control & Data plane using LiquidIO
 
TC Flower Offload
TC Flower OffloadTC Flower Offload
TC Flower Offload
 
Open vSwitch - Stateful Connection Tracking & Stateful NAT
Open vSwitch - Stateful Connection Tracking & Stateful NATOpen vSwitch - Stateful Connection Tracking & Stateful NAT
Open vSwitch - Stateful Connection Tracking & Stateful NAT
 
Virtualized network with openvswitch
Virtualized network with openvswitchVirtualized network with openvswitch
Virtualized network with openvswitch
 
The n00bs guide to ovs dpdk
The n00bs guide to ovs dpdkThe n00bs guide to ovs dpdk
The n00bs guide to ovs dpdk
 
Quality of Service Ingress Rate Limiting and OVS Hardware Offloads
Quality of Service Ingress Rate Limiting and OVS Hardware OffloadsQuality of Service Ingress Rate Limiting and OVS Hardware Offloads
Quality of Service Ingress Rate Limiting and OVS Hardware Offloads
 
The Next Generation Firewall for Red Hat Enterprise Linux 7 RC
The Next Generation Firewall for Red Hat Enterprise Linux 7 RCThe Next Generation Firewall for Red Hat Enterprise Linux 7 RC
The Next Generation Firewall for Red Hat Enterprise Linux 7 RC
 
debugging openstack neutron /w openvswitch
debugging openstack neutron /w openvswitchdebugging openstack neutron /w openvswitch
debugging openstack neutron /w openvswitch
 
BPF & Cilium - Turning Linux into a Microservices-aware Operating System
BPF  & Cilium - Turning Linux into a Microservices-aware Operating SystemBPF  & Cilium - Turning Linux into a Microservices-aware Operating System
BPF & Cilium - Turning Linux into a Microservices-aware Operating System
 
Accelerating Neutron with Intel DPDK
Accelerating Neutron with Intel DPDKAccelerating Neutron with Intel DPDK
Accelerating Neutron with Intel DPDK
 
OpenStack networking juno l3 h-a, dvr
OpenStack networking   juno l3 h-a, dvrOpenStack networking   juno l3 h-a, dvr
OpenStack networking juno l3 h-a, dvr
 
LF_OVS_17_State of the OVN
LF_OVS_17_State of the OVNLF_OVS_17_State of the OVN
LF_OVS_17_State of the OVN
 
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)
Taking Security Groups to Ludicrous Speed with OVS (OpenStack Summit 2015)
 
Linux Networking Explained
Linux Networking ExplainedLinux Networking Explained
Linux Networking Explained
 
LinuxCon 2015 Linux Kernel Networking Walkthrough
LinuxCon 2015 Linux Kernel Networking WalkthroughLinuxCon 2015 Linux Kernel Networking Walkthrough
LinuxCon 2015 Linux Kernel Networking Walkthrough
 
Cilium - API-aware Networking and Security for Containers based on BPF
Cilium - API-aware Networking and Security for Containers based on BPFCilium - API-aware Networking and Security for Containers based on BPF
Cilium - API-aware Networking and Security for Containers based on BPF
 
Offloading TC Rules on OVS Internal Ports
Offloading TC Rules on OVS Internal Ports Offloading TC Rules on OVS Internal Ports
Offloading TC Rules on OVS Internal Ports
 

Destacado

LF_OVS_17_Day 1 Opening Remarks
LF_OVS_17_Day 1 Opening RemarksLF_OVS_17_Day 1 Opening Remarks
LF_OVS_17_Day 1 Opening RemarksLF_OpenvSwitch
 
LF_OVS_17_Conntrack + OvS
LF_OVS_17_Conntrack + OvSLF_OVS_17_Conntrack + OvS
LF_OVS_17_Conntrack + OvSLF_OpenvSwitch
 
LF_OVS_17_Riley: Pushing networking to the edge
LF_OVS_17_Riley: Pushing networking to the edgeLF_OVS_17_Riley: Pushing networking to the edge
LF_OVS_17_Riley: Pushing networking to the edgeLF_OpenvSwitch
 
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...LF_OpenvSwitch
 
LF_OVS_17_OVN at Nutanix
LF_OVS_17_OVN at NutanixLF_OVS_17_OVN at Nutanix
LF_OVS_17_OVN at NutanixLF_OpenvSwitch
 
LF_OVS_17_IPSEC and OVS DPDK
LF_OVS_17_IPSEC and OVS DPDKLF_OVS_17_IPSEC and OVS DPDK
LF_OVS_17_IPSEC and OVS DPDKLF_OpenvSwitch
 
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrack
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrackLF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrack
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrackLF_OpenvSwitch
 
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...software
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...softwareLF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...software
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...softwareLF_OpenvSwitch
 
LF_OVS_17_Day 2 Opening Remarks
LF_OVS_17_Day 2 Opening RemarksLF_OVS_17_Day 2 Opening Remarks
LF_OVS_17_Day 2 Opening RemarksLF_OpenvSwitch
 
LF_OVS_17_OVS-DPDK for NFV: go live feedback!
LF_OVS_17_OVS-DPDK for NFV: go live feedback!LF_OVS_17_OVS-DPDK for NFV: go live feedback!
LF_OVS_17_OVS-DPDK for NFV: go live feedback!LF_OpenvSwitch
 
LF_OVS_17_CORD: An open source platform to reinvent the network edge
LF_OVS_17_CORD: An open source platform to reinvent the network edgeLF_OVS_17_CORD: An open source platform to reinvent the network edge
LF_OVS_17_CORD: An open source platform to reinvent the network edgeLF_OpenvSwitch
 

Destacado (11)

LF_OVS_17_Day 1 Opening Remarks
LF_OVS_17_Day 1 Opening RemarksLF_OVS_17_Day 1 Opening Remarks
LF_OVS_17_Day 1 Opening Remarks
 
LF_OVS_17_Conntrack + OvS
LF_OVS_17_Conntrack + OvSLF_OVS_17_Conntrack + OvS
LF_OVS_17_Conntrack + OvS
 
LF_OVS_17_Riley: Pushing networking to the edge
LF_OVS_17_Riley: Pushing networking to the edgeLF_OVS_17_Riley: Pushing networking to the edge
LF_OVS_17_Riley: Pushing networking to the edge
 
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...
LF_OVS_17_OvS-CD: Optimizing Flow Classification for OvS using the DPDK Membe...
 
LF_OVS_17_OVN at Nutanix
LF_OVS_17_OVN at NutanixLF_OVS_17_OVN at Nutanix
LF_OVS_17_OVN at Nutanix
 
LF_OVS_17_IPSEC and OVS DPDK
LF_OVS_17_IPSEC and OVS DPDKLF_OVS_17_IPSEC and OVS DPDK
LF_OVS_17_IPSEC and OVS DPDK
 
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrack
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrackLF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrack
LF_OVS_17_DigitalOcean Cloud Firewalls: powered by OvS and conntrack
 
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...software
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...softwareLF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...software
LF_OVS_17_The birth of SmartNICs -- offloading dataplane traffic to...software
 
LF_OVS_17_Day 2 Opening Remarks
LF_OVS_17_Day 2 Opening RemarksLF_OVS_17_Day 2 Opening Remarks
LF_OVS_17_Day 2 Opening Remarks
 
LF_OVS_17_OVS-DPDK for NFV: go live feedback!
LF_OVS_17_OVS-DPDK for NFV: go live feedback!LF_OVS_17_OVS-DPDK for NFV: go live feedback!
LF_OVS_17_OVS-DPDK for NFV: go live feedback!
 
LF_OVS_17_CORD: An open source platform to reinvent the network edge
LF_OVS_17_CORD: An open source platform to reinvent the network edgeLF_OVS_17_CORD: An open source platform to reinvent the network edge
LF_OVS_17_CORD: An open source platform to reinvent the network edge
 

Similar a LF_OVS_17_Ingress Scheduling

Troubleshooting TCP/IP
Troubleshooting TCP/IPTroubleshooting TCP/IP
Troubleshooting TCP/IPvijai s
 
SDN/OpenFlow #lspe
SDN/OpenFlow #lspeSDN/OpenFlow #lspe
SDN/OpenFlow #lspeChris Westin
 
The analysis of Microburs (Burstiness) on Virtual Switch
The analysis of Microburs (Burstiness) on Virtual SwitchThe analysis of Microburs (Burstiness) on Virtual Switch
The analysis of Microburs (Burstiness) on Virtual SwitchChunghan Lee
 
(NET404) Making Every Packet Count
(NET404) Making Every Packet Count(NET404) Making Every Packet Count
(NET404) Making Every Packet CountAmazon Web Services
 
Pushing Packets - How do the ML2 Mechanism Drivers Stack Up
Pushing Packets - How do the ML2 Mechanism Drivers Stack UpPushing Packets - How do the ML2 Mechanism Drivers Stack Up
Pushing Packets - How do the ML2 Mechanism Drivers Stack UpJames Denton
 
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...The Linux Foundation
 
Summit 16: Achieving Low Latency Network Function with Opnfv
Summit 16: Achieving Low Latency Network Function with OpnfvSummit 16: Achieving Low Latency Network Function with Opnfv
Summit 16: Achieving Low Latency Network Function with OpnfvOPNFV
 
Open stack advanced_part
Open stack advanced_partOpen stack advanced_part
Open stack advanced_partlilliput12
 
NUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioNUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioHajime Tazaki
 
AWS re:Invent 2016: Making Every Packet Count (NET404)
AWS re:Invent 2016: Making Every Packet Count (NET404)AWS re:Invent 2016: Making Every Packet Count (NET404)
AWS re:Invent 2016: Making Every Packet Count (NET404)Amazon Web Services
 
cisco-ws-c2960+48pst-l-datasheet.pdf
cisco-ws-c2960+48pst-l-datasheet.pdfcisco-ws-c2960+48pst-l-datasheet.pdf
cisco-ws-c2960+48pst-l-datasheet.pdfHi-Network.com
 
Redesigning MPTCP in Edge clouds
Redesigning MPTCP in Edge cloudsRedesigning MPTCP in Edge clouds
Redesigning MPTCP in Edge cloudsNitinder Mohan
 
Technical Overview of QUIC
Technical  Overview of QUICTechnical  Overview of QUIC
Technical Overview of QUICshigeki_ohtsu
 
CS4344 09/10 Lecture 10: Transport Protocol for Networked Games
CS4344 09/10 Lecture 10: Transport Protocol for Networked GamesCS4344 09/10 Lecture 10: Transport Protocol for Networked Games
CS4344 09/10 Lecture 10: Transport Protocol for Networked GamesWei Tsang Ooi
 
Nessus scan report using the defualt scan policy - Tareq Hanaysha
Nessus scan report using the defualt scan policy - Tareq HanayshaNessus scan report using the defualt scan policy - Tareq Hanaysha
Nessus scan report using the defualt scan policy - Tareq HanayshaHanaysha
 
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...PROIDEA
 
VYATTAによるマルチパスVPN接続手法
VYATTAによるマルチパスVPN接続手法VYATTAによるマルチパスVPN接続手法
VYATTAによるマルチパスVPN接続手法Naoto MATSUMOTO
 
Bar-BoF session about Simplemux at IETF93, Prague
Bar-BoF session about Simplemux at IETF93, PragueBar-BoF session about Simplemux at IETF93, Prague
Bar-BoF session about Simplemux at IETF93, PragueJose Saldana
 

Similar a LF_OVS_17_Ingress Scheduling (20)

Troubleshooting TCP/IP
Troubleshooting TCP/IPTroubleshooting TCP/IP
Troubleshooting TCP/IP
 
SDN/OpenFlow #lspe
SDN/OpenFlow #lspeSDN/OpenFlow #lspe
SDN/OpenFlow #lspe
 
The analysis of Microburs (Burstiness) on Virtual Switch
The analysis of Microburs (Burstiness) on Virtual SwitchThe analysis of Microburs (Burstiness) on Virtual Switch
The analysis of Microburs (Burstiness) on Virtual Switch
 
(NET404) Making Every Packet Count
(NET404) Making Every Packet Count(NET404) Making Every Packet Count
(NET404) Making Every Packet Count
 
Pushing Packets - How do the ML2 Mechanism Drivers Stack Up
Pushing Packets - How do the ML2 Mechanism Drivers Stack UpPushing Packets - How do the ML2 Mechanism Drivers Stack Up
Pushing Packets - How do the ML2 Mechanism Drivers Stack Up
 
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
 
Summit 16: Achieving Low Latency Network Function with Opnfv
Summit 16: Achieving Low Latency Network Function with OpnfvSummit 16: Achieving Low Latency Network Function with Opnfv
Summit 16: Achieving Low Latency Network Function with Opnfv
 
Open stack advanced_part
Open stack advanced_partOpen stack advanced_part
Open stack advanced_part
 
NUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioNUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osio
 
AWS re:Invent 2016: Making Every Packet Count (NET404)
AWS re:Invent 2016: Making Every Packet Count (NET404)AWS re:Invent 2016: Making Every Packet Count (NET404)
AWS re:Invent 2016: Making Every Packet Count (NET404)
 
cisco-ws-c2960+48pst-l-datasheet.pdf
cisco-ws-c2960+48pst-l-datasheet.pdfcisco-ws-c2960+48pst-l-datasheet.pdf
cisco-ws-c2960+48pst-l-datasheet.pdf
 
Redesigning MPTCP in Edge clouds
Redesigning MPTCP in Edge cloudsRedesigning MPTCP in Edge clouds
Redesigning MPTCP in Edge clouds
 
Technical Overview of QUIC
Technical  Overview of QUICTechnical  Overview of QUIC
Technical Overview of QUIC
 
Polyraptor
PolyraptorPolyraptor
Polyraptor
 
CS4344 09/10 Lecture 10: Transport Protocol for Networked Games
CS4344 09/10 Lecture 10: Transport Protocol for Networked GamesCS4344 09/10 Lecture 10: Transport Protocol for Networked Games
CS4344 09/10 Lecture 10: Transport Protocol for Networked Games
 
Nessus scan report using the defualt scan policy - Tareq Hanaysha
Nessus scan report using the defualt scan policy - Tareq HanayshaNessus scan report using the defualt scan policy - Tareq Hanaysha
Nessus scan report using the defualt scan policy - Tareq Hanaysha
 
Userspace networking
Userspace networkingUserspace networking
Userspace networking
 
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...
PLNOG15: VidMon - monitoring video signal quality in Service Provider IP netw...
 
VYATTAによるマルチパスVPN接続手法
VYATTAによるマルチパスVPN接続手法VYATTAによるマルチパスVPN接続手法
VYATTAによるマルチパスVPN接続手法
 
Bar-BoF session about Simplemux at IETF93, Prague
Bar-BoF session about Simplemux at IETF93, PragueBar-BoF session about Simplemux at IETF93, Prague
Bar-BoF session about Simplemux at IETF93, Prague
 

Más de LF_OpenvSwitch

LF_OVS_17_Day 2 Closing Remarks
LF_OVS_17_Day 2 Closing RemarksLF_OVS_17_Day 2 Closing Remarks
LF_OVS_17_Day 2 Closing RemarksLF_OpenvSwitch
 
LF_OVS_17_OVN and Kelda
LF_OVS_17_OVN and KeldaLF_OVS_17_OVN and Kelda
LF_OVS_17_OVN and KeldaLF_OpenvSwitch
 
LF_OVS_17_OvS manipulation with Go at DigitalOcean
LF_OVS_17_OvS manipulation with Go at DigitalOceanLF_OVS_17_OvS manipulation with Go at DigitalOcean
LF_OVS_17_OvS manipulation with Go at DigitalOceanLF_OpenvSwitch
 
LF_OVS_17_OvS Hardware Offload with TC Flower
LF_OVS_17_OvS Hardware Offload with TC FlowerLF_OVS_17_OvS Hardware Offload with TC Flower
LF_OVS_17_OvS Hardware Offload with TC FlowerLF_OpenvSwitch
 
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.LF_OpenvSwitch
 
LF_OVS_17_Red Hat's perspective on OVS HW Offload Status
LF_OVS_17_Red Hat's perspective on OVS HW Offload StatusLF_OVS_17_Red Hat's perspective on OVS HW Offload Status
LF_OVS_17_Red Hat's perspective on OVS HW Offload StatusLF_OpenvSwitch
 

Más de LF_OpenvSwitch (6)

LF_OVS_17_Day 2 Closing Remarks
LF_OVS_17_Day 2 Closing RemarksLF_OVS_17_Day 2 Closing Remarks
LF_OVS_17_Day 2 Closing Remarks
 
LF_OVS_17_OVN and Kelda
LF_OVS_17_OVN and KeldaLF_OVS_17_OVN and Kelda
LF_OVS_17_OVN and Kelda
 
LF_OVS_17_OvS manipulation with Go at DigitalOcean
LF_OVS_17_OvS manipulation with Go at DigitalOceanLF_OVS_17_OvS manipulation with Go at DigitalOcean
LF_OVS_17_OvS manipulation with Go at DigitalOcean
 
LF_OVS_17_OvS Hardware Offload with TC Flower
LF_OVS_17_OvS Hardware Offload with TC FlowerLF_OVS_17_OvS Hardware Offload with TC Flower
LF_OVS_17_OvS Hardware Offload with TC Flower
 
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.
LF_OVS_17_Enabling hardware acceleration in OVS-DPDK using DPDK Framework.
 
LF_OVS_17_Red Hat's perspective on OVS HW Offload Status
LF_OVS_17_Red Hat's perspective on OVS HW Offload StatusLF_OVS_17_Red Hat's perspective on OVS HW Offload Status
LF_OVS_17_Red Hat's perspective on OVS HW Offload Status
 

Último

Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)wesley chun
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...DianaGray10
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...apidays
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherRemote DBA Services
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?Antenna Manufacturer Coco
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...Neo4j
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessPixlogix Infotech
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 

Último (20)

Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...Workshop - Best of Both Worlds_ Combine  KG and Vector search for  enhanced R...
Workshop - Best of Both Worlds_ Combine KG and Vector search for enhanced R...
 
Advantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your BusinessAdvantages of Hiring UIUX Design Service Providers for Your Business
Advantages of Hiring UIUX Design Service Providers for Your Business
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 

LF_OVS_17_Ingress Scheduling

  • 1. Ingress Scheduling in OvS-DPDK Billy O’Mahony – Intel Jan Scheurich – Ericsson November 16-17, 2017 | San Jose, CA
  • 2. Introduction u Use cases for traffic prioritization in NFV u State of the art in OvS-DPDK datapath u Rx queue prioritization in DPDK datapath u Traffic classification and queue selection on NIC u Next steps
  • 3. Compute Node OvS br-intbr-ctrl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag Compute Node OvS Scenario: NFVI on Converged Data Center VIM control plane sharing physical network with tenant data ToR A ToR B dpdk0 dpdk1 br-intbr-ctl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag VIM = Virtual Infrastructure Manager For example OpenStack component: Nova, Neutron services and their local agents
  • 4. Compute Node OvS br-intbr-ctrl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag Compute Node OvS Use Case 1: In-band OvS Control Plane ToR A ToR B dpdk0 dpdk1 br-intbr-ctl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag LACP bond supervision LACP = Link Aggregation Control Protocol Here: in-band heart-beat between OVS and each ToR
  • 5. Compute Node OvS br-intbr-ctrl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag Compute Node OvS Use Case 1: In-band OvS Control Plane ToR A ToR B dpdk0 dpdk1 br-intbr-ctl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag BFD tunnel monitoring BFD packets are sent inside tunnel BFD = Bidirectional Forwarding Detection Here: Heart-beat between OVS instances connected through tunnel mesh
  • 6. Compute Node OvS br-intbr-ctrl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag Compute Node OvS Use Case 2: VIM Control Plane ToR A ToR B dpdk0 dpdk1 br-intbr-ctl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsNova, Neutron, … Host networking bond br-prv tag tag VIM Control Plane VIM Control Plane In OpenStack: RabbitMQ, REST APIs calls, …
  • 7. Compute Node OvS br-intbr-ctrl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag Compute Node OvS Use Case 2: VIM Control Plane ToR A ToR B dpdk0 dpdk1 br-intbr-ctl Tenant VM Tenant VM vhostuser Local Agents Local Agents Local AgentsVIM Components Host networking bond br-prv tag tag OvS control plane: OpenFlow and OVSDB OpenFlow and OVSDB are special cases of VIM control plane
  • 8. OvS Status Quo in OvS DPDK Datapath NIC PMD 1 PMD 2 ovs-vswitchd thread Tenant VM BFD LACP RSS HW Scheduler VIM Components Host networking br-ctl
  • 9. OvS Scenario: PMD Overload NIC PMD 1 PMD 2 ovs-vswitchd thread Tenant VM br-ctl BFD LACP RSS PMD cycles exhausted through tenant data Rx queues full. Data and control packets dropped. VIM Components Host networking
  • 10. OvS Scenario: Egress Link Overload NIC PMD 1 PMD 2 ovs-vswitchd thread Tenant VM BFD LACP RSS VIM Components Host networking br-ctl PMD Tx queues full. Tenant data being dropped. Separate Tx queue for ovs-vswitchd. HW scheduler can provide fair share of link band- width Egress link band- width exhausted by tenant data
  • 11. Offered load on physical port [Kpps] 2000 2200 2400 2600 2800 3200 3600 4000 Offered load on phy port [Gbit/s] 1.54 1.69 1.84 2.00 2.15 2.46 2.76 3.07 PMD overload factor [%] 0% 9% 18% 27% 45% 64% 82% PMD utilization [%] 99.95% 99.99% 100% 100% 100% 100% 100% 100% Phy port rx drop [%] 0% 0% 8% 15% 21% 31% 39% 45% ping -f average RTT [ms] 0.45 0.50 3.02 3.03 3.15 3.10 3.69 3.95 ping -f packet drop [%] 0 0 10% 16% 21% 37% 45% 49% BFD flappings [1/min] 1.85 3.75 5.66 5.71 Num flaps 0 0 17 17 43 20 OpenFlow connection timeouts in OVS 0 0 0 0 0 0 3 3 Connection closed by peer (ODL) 20 15 Connection reset by peer (ODL) 2 0 37 5.03 PMD polling physical port overloaded with 64B packets u Packet drops in the Rx queue of physical port equally affect tenant data, BFD and OVS control plane packets u “ping –f“ to br-ctl interface to quantify control plane impact u Ping packet drop in line with overall packet drop u RTT jumps from 50 us to 3 ms u BFD flapping occurs already at moderate overload u The rate increases with overload u Above 45% packet drop the OpenFlow control channel breaks due to missed Echo Replies Measurements Impact of PMD Overload source: Ericsson CPU: Dual socket, Xeon CPU E5-2697 v3 @2.60GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x 10Gbit/s; OvS: version 2.6, 1 PMD, 1 phy port, 1 vhostuser port; VM: TRex DPDK traffic source/sink
  • 12. u Egress link overload does not affect the control plane u Outgoing packets are forwarded by the ovs-vswitchd thread, which has its dedicated TX queue in the Fortville NIC u The NIC schedules packets from each of the TX queues in some fair manner, so that the ovs-vswitchd queue gets sufficient bandwidth on the link u Incoming packets are not affected as neither link nor PMDs are overloaded u No BFD flapping Measurements Impact of Egress Link Overload Offered load from VM [Kpps] 800 900 1000 1200 1600 Offered load from VM [Gbit/s] 9.80 11.03 12.26 14.71 19.61 Transmitted load on phy port [Gbit/s] 9.81 9.90 9.88 9.88 9.88 Link overload 0% 11% 24% 49% 99% PMD utilization [%] 41.45% 46.30% 50.20% 56.36% 69.89% ping -f average RTT [ms] 0.109 0.205 0.206 0.210 0.204 ping -f packet drop [%] 0% 0% 0% 0% 0% BFD flappings [1/min] Num flaps 0 0 0 0 0 OpenFlow connection timeouts in OVS 0 0 0 0 0 Connection closed by peer (ODL) Connection reset by peer (ODL) 10G Link from OvS overloaded with outgoing traffic from VM (1500 byte packets) source: Ericsson CPU: Dual socket, Xeon CPU E5-2697 v3 @2.60GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x 10Gbit/s; OvS: version 2.6, 1 PMD, 1 phy port, 1 vhostuser port; VM: TRex DPDK traffic source/sink
  • 13. Use Case 3: QoS for Tenant Data All tenant data traffic is equal!? Well, some packets are more equal than others! u Virtual Network Functions send/receive a large variety of network traffic u Top prio: Critical internal control plane (e.g. cluster membership) u … u Min prio: Bulk user plane u VNFs need prioritization for their critical traffic in the NFVI u How to orchestrate and implement the necessary QoS end-to-end? u Will need additional priority levels and packet marking (e.g. IP Diffserv)
  • 14. Desired Ingress Prioritization on Physical Ports u Priority 1: In-band control plane u Untagged LACP packets u BFD packets inside tunnel based on IP DSCP of outer IP header u Priority 2: VIM control plane u Certain prioritized VLAN tags u Priority 3+: Prioritized tenant data u E.g. based on IP DSCP of outer IP header u Base Priority u Non-prioritized traffic spread through RSS over multiple Rx queues
  • 15. PMD “Schedulers arrange and/or rearrange packets for output.” -- http://www.tldp.org/HOWTO/html_single/Traffic-Control-HOWTO/#e-scheduling Ingress Scheduling RX Queue à TX Queue à ? ovs-vswitchd BFD LACP Priority packet – e.g control plane.
  • 16. Ingress Scheduling - Implementation RX Queue x2 à TX Queue à ? DPDK rte_flowAPI installs rxq assignment filters on supported vendors NICs ovs-vswitchd BFD LACP PMD empties priority queue before reading non- priority queue PMD
  • 17. Ingress Scheduling - Implementation 1. Move packet prioritization decision to the NIC 2. Place prioritized packets on separate RX Queue 3. Read preferentially from “priority” RxQ. Keep it simple: u Read from priority queue until it’s empty u Service other queues u Repeat
  • 18. Ingress Scheduling – Latency effect ~99.9 % of packets already have have a latency <20us There are x10 to x50 less packets in any given latency bucket – good. But worst case latency does not improve. CPU: Dual socket, Xeon(R) CPU E5-2695 v3 @2.30GHz 14 core no-HT, 896K L1, 3584K L2, 35MB L3 cache; NIC: Intel Fortville X710, 4 x 10Gbit/s; OvS: version 2.7.90, 1 PMD, 2 phy port, Hardware trafficsource/sink Source: Intel
  • 19. Ingress Scheduling – Overload protection RX Queue x2 à TX Queue à L ? ovs-vswitchd BFD LACP PMD 1
  • 20. dpdk1 Ingress Scheduling – Traffic Protection u Overload PMD through 64 byte DPDK traffic on dpdk0 à 100% PMD load in pmd-stats-show à 25% rx packet drop on dpdk0 u Add iperf3 UDP traffic (256 bytes) in parallel over dpdk1 u Measurement result: SUT Server OvS VM dpdk testpmd vhostuser iperf3 udp server PMD br-prv ToR dpdk0dpdk1 TGen Server BM dpdk pktgen iperf3 udp client dpdk0 Low priorityHigh/Low priority source: Ericsson Condition 1 dpdk1 low priority Condition 2 dpdk1 high priority iperf3 UDP throughput not measured 1 Gbit/s 460 Kpps 1) iperf3 UDP packet loss 28% 0% 1) iperf3 throughput limited by UDP/IP stack on client side CPU: Dual socket, Xeon CPU E5-2680 v4 @2.40GHz, 14 cores + HT, 896K L1, 3584K L2, 35MB L3 cache NIC: Intel Fortville X710, 4 x 10Gbit/s; OvS: version 2.6, 1 PMD, all ports and VM on NUMA node 0
  • 21. u $ ovs-vsctl set Interface phy1 ingress_sched: eth_type=0x8809 Ingress Scheduling – Configuration Field as per ovs-fields(7) and ofctl add-flow. Not all netdevs/NICs will support all combinations. Single prioritization condition.
  • 22. u $ ovs-vsctl set Interface phy1 ingress_sched: vlan_tci=0x1123/0x1fff,ip,ip_dscp=0x5 Ingress Scheduling – Configuration (future) Several different prioritization conditions. a AND b.
  • 23. u $ ovs-vsctl set Interface phy1 ingress_sched: filter=vlan_tci=0x1123/0x1fff filter=ip,ip_dscp=0x5 Ingress Scheduling – Configuration (future) Several different prioritization conditions. a OR b.
  • 24. u $ ovs-vsctl set Interface phy1 ingress_sched: prio=1, filter,vlan_tci=0x1123/0x1fff, filter,eth_type=0x8809, prio=2, filter,ip,ip_dscp=0x5, Ingress Scheduling – Configuration (future) Traffic Priority Levels Support several levels of prioritization: High and Low but also a Critical level for instance.
  • 25. u $ ovs-vsctl set Interface phy1 ingress_sched: prio=2, filter,ip,ip_dscp=0x5, prio=1, filter=vlan_tci=0x1123/0x1fff, filter,eth_type=0x8809 Ingress Scheduling – Configuration (future) Filter Priority: Filter groups are applied in the order in which they appear on the configuration line.
  • 26. u ovsdb-schema <table name="Interface"… <column name="ingress_sched" key="err"> If the specified ingress scheduling could not be applied, Open vSwitch sets this column to an error description in human readable form. Otherwise, Open vSwitch clears this column. Ingress Scheduling –Error reporting
  • 27. u $ ovs-vsctl set Interface phy1 options:n_rxq=4 u $ ovs-vsctl set Interface phy1 ingress_sched: prio=2, filter,ip,ip_dscp=0x5, prio=1, filter=vlan_tci=0x1123/0x1fff, filter,eth_type=0x8809 Ingress Scheduling – RxQ’s & RSS RSS queues Additional Priority Queues
  • 28. Ingress Scheduling – Next Steps PMD PMD PMD PMD u Avoid poor rxq -> pmd assignment
  • 29. Ingress Scheduling – Next Steps u Use rte_flow API for offload u Extend to several priorities u Priorities of overlapping filters u Multiple traffic priorities u Working with RFC ‘Flow Offload’ feature… u … u Prioritization to the Guest…
  • 30. Summary u OvS-DPDK in NFVI context needs ingress scheduling to protect priority traffic against PMD overload u SW priority queue handling in the PMD loop is effective u Could be upstreamed first, priority configurable per port u Off-loading classification and queue selection to NIC through rte_flow API allows generic solution u Interaction with RFC Flow Classification Offload u Work in progress u Lots left to figure out u We are open for suggestions/collaboration
  • 31. Intel Notices & Disclaimers Intel technologies’ features and benefits depend onsystem configurationand may require enabledhardware, software or service activation. Performance varies dependingon system configuration. Check with your system manufacturer or retailer orlearn more at intel.com. No computer system can be absolutely secure. Tests document performance of components on a particular test, in specific systems. Differences in hardware, software,or configurationwill affect actual performance. Consult othersources of informationto evaluate performance as you consider your purchase. For more complete information about performance andbenchmark results, visit http://www.intel.com/benchmarks . Software and workloads used in performance tests may have been optimizedfor performance onlyon Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measuredusing specific computersystems, components, software,operations and functions. Anychange to any of those factors may cause the results to vary. You should consult otherinformation andperformance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combinedwith other products. For more completeinformation visit http://www.intel.com/benchmarks . Intel's compilers may or may not optimize to the same degree for non-Intel microprocessors for optimizations that are not unique to Intel microprocessors. These optimizations include SSE2, SSE3, and SSSE3 instruction sets and other optimizations. Intel does not guarantee the availability, functionality, or effectiveness of any optimization on microprocessors not manufactured by Intel. Microprocessor-dependent optimizations in this product are intended for use with Intel microprocessors. Certain optimizations not specific to Intel microarchitecture are reserved for Intel microprocessors. Please refer to the applicable product User and Reference Guides for more information regarding the specific instruction sets covered by this notice. Cost reduction scenarios described are intendedas examples of how a given Intel-based product, in the specified circumstances and configurations, may affect future costs and provide cost savings. Circumstances will vary. Intel does not guarantee any costs or cost reduction. Intel does not control or audit third-partybenchmark data or the web sites referencedin this document. Youshouldvisit the referencedweb site and confirm whether referenced data are accurate. © 2017 Intel Corporation. Intel, the Intel logo, and Intel Xeonare trademarks of Intel Corporationin the U.S. and/or other countries. *Other names and brands may be claimedas property of others.
  • 33. References 1. [ovs-dev] [RFC PATCH 0/3] prioritizing latency sensitive traffic u [ovs-dev,RFC,1/3] netdev: Add set_ingress_sched to netdev api u [ovs-dev,RFC,2/3] netdev-dpdk: Apply ingress_sched config to dpdk phy ports u [ovs-dev,RFC,3/3] dpif-netdev: Add rxq prioritization