Brkarc 3452 PDF
Brkarc 3452 PDF
Brkarc 3452 PDF
Switch Architecture
Rohan Grover, Technical Marketing Manager
BRKARC-3452
Session Goal
This session will also examine the Nexus 5600 Multicast Architecture, ACLs and
QOS
This session will not examine NX-OS software architecture or other Nexus
platform architectures
Session Name
BRKDCT-3100
BRKARC-3470
BRKARC-3454
BRKDCT-1980
BRKDCT-3346
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Introduction
Nexus 7000
Nexus 7000
Nexus 5x00
Nexus 6000
FabricPath
CORE
Nexus 7000
L3
AGG.
VPC to Nexus 7000 or 5600
FabricPath
SPINE
Nexus 5x00/6000
Nexus 5x00
Nexus 6000
ACCESS
Nexus 2000
FEX
FabricPath
Nexus 5x00/6k
LEAF
Nexus 5600
Nexus 7000
Nexus 5x00
Nexus 6000
L3
Fabric
with
VXAN
Nexus 5600
LEAF
Nexus 5x00
Nexus 6000
ACCESS
FCoE
FC
MDS
40G
Nexus 6004
40G
Nexus 5624Q, 5648Q
Nexus
5010/5020
Generation 1
2008
Nexus 5500
(10G Optimized)
Generation 2
2010
10G
Nexus 6001
Generation 3
2013
10G
Generation 3.5
2014-2015
Market Share*
Nexus 5600
25 Million
Programmability
Ports Shipped
Network Visibility
20,000+
Customers
Line Rate L3
Bigger buffers and tables
Nexus 5500
125K+
Unified Ports
FEX Architecture
FabricPath
FCoE
Nexus 5500
Flexibility
10G Uplinks
Programmability
Network Visibility
LAN/SAN Convergence
FabricPath
20K+ Customers
Over 5 Years
Nexus 5010/5020
FEX Architecture
CISCO INNOVATION
Density
CUSTOMER VALUE
Nexus 5600/6000
Large Buffers
and Tables
~ 1us
~ 1us
~ 1.8us
L3 Module
32K MAC/ARP
8K (v4) or 4K (v6)
IP Hosts
IP Multicast routes
8K (v4) or 4K (v6)
25MB/12p X 10G)
25MB/12p X 10G)
640KB/10G port
32K*
32K*
8K
VRFs
4K
4K
1K
SPAN
Switch Fabric
Throughput
Port-to-Port Latency
Layer 3 Capability
VXLAN
Unified Ports
MAC Table
IP Unicast Routes
Buffers
IGMP Snooping
Groups
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Architecture
Fabric Connections
10 GE Mode
UPC
448Gbps (32 x 14 Gbps)
224Gbps (4 x 56Gbps)
Three 40 GE
UPC
448Gbps (8 x 56Gbps)
40 GE Mode
In 10 GE mode:
192x384 X-BAR
48x96 X-BAR
Ingress
UPC
192x14Gbps
UCF
(Fabric)
384x14Gbps
Egress
UPC
10 GE Mode
Ingress
UPC
48x56Gbps
UCF
(Fabric)
96x56Gbps
Egress
UPC
40 GE Mode
Fan Side of
Switch
Flexibility to use
4x10G or 40G
Redundant 1+1
AC/DC Power
Supplies
Redundant 1+1
AC/DC Power
Supplies
Choice of
Reversible Airflow
Redundant Fan
Modules
Mgmt 0,
Console, USB
Supervisor
UPC 1
UPC 2
UPC 3
UPC 4
UPC 5
UPC 6
UPC-0
FC PHY
24xSFP+
24xSFP+
Slot 1
6xQSFP+
Slot 2
CPU
Data Plane
PCle
1G MAC
PCle
Forwarding
ASIC
Fabric ASIC
PCle
PCle
CPU Card
Forwarding ASIC
XCon
MGMT
XCon
Console
QSFP/SFP
External Connections
Fabric Card
Line Card
Port Side of
Switch
2 Expansion Slots
Expansion Slots can be 24 port 10G Ethernet/FCOE or 2/4/8 FC with 2
40G QSFP+ Ethernet/FCOE
Base Chassis
4x 40G
QSFP+ Ports
Fan Side of
Switch
Flexibility to use
4x10G or 40G
Redundant 1+1
AC/DC Power
Supplies
Choice of
Reversible Airflow
Redundant Fan
Mgmt 0, Console, Modules
USB
Redundant 1+1
AC/DC Power
Supplies
Expansion Slot 2
Supervisor
Fabric
UPC 1
UPC 2
24x
SFP+
UPC 4
UPC 3
24x
SFP+
Slot 1
CPU
UPC 5
2x
QSFP+
UPC 6
2x
QSFP+
Supervisor
Fabric
UPC 0
CPU
UPC 1
UPC 2
24x
SFP+
UPC 4
UPC 3
24x
SFP+
Slot 1
UPC 5
2x
QSFP+
UPC 6
2x
QSFP+
2 QSFP ports:
Native 40G or 4x 10GE
Ethernet and FCoE
24 UP ports
10G or 1G
Ethernet, FCoE, and FC
Supports OIR *
* shutdown the interfaces on the module and then poweroff module <> before swapping the module.
UPC 1
UPC 2
FC PHY
FC PHY
24x SFP+
UPC Mapping
UPC 3
2x QSFP+
Slot 1
Slot 2
UPC1
10
UPC2
11
UPC3
12
Port Side of
Switch
Expansion Modules
10/40G Ethernet/FCoE
Fan Side of
Switch
10/40G Ethernet/FCoE
Redundant 1+1
AC/DC Power
Supplies
Redundant 1+1
AC/DC Power
Supplies
Choice of
Reversible Airflow
Redundant Fan
Modules
Mgmt 0, Console,
USB
Supervisor
UPC 1
UPC 2
UPC 3
UPC-0
UPC 4
Expansion Slot 2
12x QSFP+ Base Ports
Slot 1
CPU
Back of Switch
Front of Switch
2 Expansion Slots
Flexibility to use 4x10G or 40G
10/40G Ethernet/FCoE
Base Chassis
24x 40G QSFP+ Ports
Redundant n+1
AC/DC Power
Supplies
Redundant n+1
AC/DC Power
Supplies
Choice of
Reversible Airflow
Redundant Fan
Modules
Mgmt 0, Console,
USB
Expansion Slot 4
(Up to 4 UPCs)
Switch Fabric consists of 2
Fabric ASICs
Fabric
Fabric
Supervisor
UPC-0
UPC 1
UPC 2
UPC 3
UPC 4
UPC 5
UPC 6
UPC 7
UPC 8
CPU
24x QSFP+ Base Ports
Slot 1
Slot 2
Supports OIR *
* shutdown the interfaces on the module and then poweroff module <> before swapping the module.
UPC 1
UPC 2
UPC 3
12x QSFP+
UPC 4
Port Side of
Switch
Fan Side of
Switch
Fan Module 3 + 1
Redundancy
Fabric
Supervisor
UPC 1
UPC 2
UPC
UPC 33
UPC-0
UPC-0
UPC
UPC 44
Expansion
Expansion Slot
Slot 22 to
to 88
12x
12x QSFP+
QSFP+ Base
Base Ports
Ports
Slot 1
CPU
CPU
* shutdown the interfaces on the module and then poweroff module <> before swapping the module.
UPC 1
UPC 2
UPC 3
12x QSFP+
UPC 4
Supports 2/4/8G FC
Supports OIR *
* shutdown the interfaces on the module and then poweroff module <> before swapping the module.
UPC 2
UPC 1
FC PHY
FC PHY
20x SFP+
Supports OIR *
* shutdown the interfaces on the module and then poweroff module <> before swapping the module.
UPC 1
UPC 2
UPC 3
UPC 4
100G MAC
100G MAC
100G MAC
100G MAC
CXP1
CXP2
CXP3
CXP4
Chassis
40G
10G
Chassis
6004 LEM
(N6K-C6004-M12Q)
V6.0(2)N1(1)
6004EF LEM
(N6004-M12Q)
V6.0(2)N2(2)
5696Q LEM
(N5696-M12Q)
V7.0(4)N1(1)
6004EF LEM
(N6004X-M20UP)
V7.0(1)N1(1)
N5696-M4C
V7.1(0)N1(1a)
N6004
N6K-C6004-96Q
Yes
Yes
No
Yes
No
No
N6004EF
N6K-C6004
No (Mechanical)
Yes
Yes
Yes
Yes
Yes
N5696Q
No (Mechanical)
No
Yes (VXLAN3)
Yes2 (VXLAN3)
Yes (VXLAN3)
Yes (VXLAN3)
PERFORMANCE
DENSITY
Nexus 56128P
Nexus 5624Q
Nexus 5648Q
Nexus 5696Q
7.0(1)N1(1)
7.0(2)N1(1)
7.1(0)N1(1a)
7.1(1)N1(1)
7.0(4)N1(1)
1.44 Tbps
2.56 Tbps
1.92 Tbps
3.84 Tbps
7.68 Tbps
Port-to-Port Latency
~1.0 us
Layer 3 Capability
Switch Footprint
SCALE
Nexus 5672UP
Integrated Line-Rate
1RU
2RU
1RU
2RU
4RU
10 Gigabit Ethernet
Port
72
128
96
192
384
40 Gigabit Ethernet
Port
24
48
96
Unified Ports
16
48
N/A
N/A
160 *
1152
1152
1152
1152
2304
100
80
60
40
20
0
64
RFC 2544 Unicast L2 Port Pair 100
RFC 2544 Unicast L3 Port Pair 100
80
100
100
96
100
100
128
100
100
256
100
100
512
100
100
Average Latency
(Nano Seconds)
1600
1400
1200
1000
800
600
400
200
0
64
1001
937
128
1021
963
256
1091
1029
512
1111
1048
1024
1105
1047
1280
1104
1046
1518
1104
1045
2240
1103
1044
4096
1101
1043
9216
1098
1040
Switch fabric mode determines the speed of a single fabric link between UPC and Fabric.
Switch fabric mode is independent of front panel port speed
Default switch fabric mode is 10 GE for 5672UP/56128P and 40 GE for 5624Q/5648Q/5696Q
Fabric Mode Configuration is system wide
8x56Gbps
UPC
4x56Gbps
32x14Gbps
UPC
16x14Gbps
10 GE Fabric Mode
40 GE Fabric Mode
Egress
10 GE
40 GE
10 GE
Line-Rate
Line-Rate
40 GE
Line-Rate
Line-Rate
Ingress
Egress
10 GE
40 GE
10 GE
Line-Rate
Line-Rate
40 GE
Line-Rate
<10GE Flow
Throughput depends
on hashing
Ingress
With 40 GE fabric, traffic from 10 GE to 10 GE may fall back to store-and-forward with a heavy traffic
load.
Recommendation is 10 GE fabric when most of the ports are 10 GE and latency is important.
40 GE Fabric Mode
Ingress
Egress
10 GE Fabric Mode
10 GE
40 GE
Ingress
Egress
10 GE
40 GE
10 GE
Store-and-forward
10 GE
Cut-through
Store-and-forward
40 GE
Cut-through
Cut-through
40 GE
Cut-through
Store-and-forward
QSFP BiDi
TX
TX/RX
4 x 10G
12-Fiber
Infrastructure
2 x 20G
12-Fiber
Infrastructure
Duplex
Multimode
Fiber
Duplex
Multimode
Fiber
4 x 10G
2 x 20G
TX
TX/RX
12-fiber MPO
Duplex LC
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Forwarding
What is a VoQ?
Virtual Output Queues (VOQs) - Provides buffering and queuing for ingressbuffered switch architecture
576 ports with 8 class of service: 4608 queues per UPC (or)
4k Unicast VOQ
(8 per egress port)
2
L2/ L3, FP, VXLAN
Forwarding
1, 10 and 40GE
Forwarding
Lookup
1
Retrieve Packet Header
Feed to Forwarding
MAC
Parser
ACL
Ingress Policing
3
Rewrite L3 Packet
headers
Packet
Rewrite
Buffer
Management
4
Allocate Buffer to packet
Signal PFC to MAC if Needed
Buffer
3
Ingress
SPAN
Unified Crossbar
Fabric
Unified Crossbar
Fabric
Egress UPC
4k Unicast Queue
PAUSE Signal
Multicast Drop
Buffer
Management
Scheduling
10
Enqueue packet
Dequeue packet
Egress MultIcast ReplIcatIion
Multicast
Replication
Forwarding Lookup
Rewrite
MAC
Allocate Buffer
Egress Congestion management
for Multicast
Packet Rewrite
Egress SPAN Replication
TX SPAN
Replication
Buffer
Multicast Queue
To Ingress
Buffer Management
1
Retrieve Packet Header
Feed to Lookup Engine
Packet Rewrite
Ingress SPAN Replication
Packets
Allocate Buffer
Signal PFC to MAC if Needed
FWI
(Parser/Rewrite)
Enqueue
Dequeue
System
(Ingress VOQ)
BMI
SPAN
(Buffer management
Ingress)
Fabric
Interface
TX
From FWE
2
Fabric
Interface
Forwarding Lookup
Ingress Policing
MAC
Egress Lookup
Lookup Engine
Access Control
10
Retrieve Packet Header
Feed to Lookup Engine
To
BMI
FWE
Packet Rewrite
Egress SPAN Replication
(Parser/Rewrite)
(service)
6
Allocate Buffer
Egress Congestion management
for Multicast
BME
(Buffer management
Egress)
7
Enqueue packet
Dequeue packet
EgressQueuing
MultIcast ReplIcatIion
System
(Egress Queues)
Enqueue
Dequeue
Fabric
Interface
RX
32,000 Entries
256,000 Entries
Summary Routes
IP Host Region
(ARP/ND/Host
route/(*,G))
* Hardware table size. Please check configuration limit for software scaling
Multicast Table
64,000 Entries
LPM Table
Host Table
MAC Region
(S,G)
MAC Region
00:02:B3:01:02:03
ARP/ND/Host route
10.1.1.1
2001:0:0:0:
0DB8:800:200C:417A
ARP
LPM Table
Host Table
10.1.1.0/24
2001::/64
10.1.1.0
IPv4 Route
2001:0
0:0
0:0
0:0
Two
TwoHardware
Hardware Entries for IPv6 ND
Entries for IPv6 ND
Four Hardware
Entries for IPv6
Route
MAC region
IP region: holds ARP, IPv6 ND, /32 Host Routes
Default sizes are 128k for MAC region and 128k for IP region.
256,000 Entries
Host Table
MAC Region
IP Host Region
(ARP & ND & /32 Host route)
Deployment Scenario
Scalability
L2 switch
256,000 MAC
Description
hrt-128-stm-128
hrt-96-stm-160
hrt-64-stm-192
hrt-32-stm-224
Show commands:
show
show
show
show
template
template default
route resource template
route resource template
10.1.1.1
10.1.1.2
40.1.1.10
{VRF, IP}
ECMP count=2
Adj base=198
LPM Table
(Summary Routes)
20.1.1.64/26
10.1.1.0/24
10.1.2.0/24
30.1.0.0/16
30.2.0.0/16
Adjacency Table(64K)
ECMP count=2
Adj base=200
ECMP count=2
Adj base=200
Unicast VOQ
224Gbps
16MB
Multicast VOQ
Unified
Crossbar
Fabric
Unicast VOQ
Unicast VOQ
224Gbps
Multicast VOQ
Unified
Crossbar
Fabric
Unicast VOQ
Line-Rate Bridging
Line-Rate Routing
Key Differentiator
VXLAN
VXLAN L2
Gateway
VXLAN
VXLAN
VXLAN L3
Gateway
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Multicast
Egress UPC 1
VLAN 10
448 Gbps
VLAN 20
224 Gbps
Ingress UPC
Unified
Crossbar
Fabric
448 Gbps
Egress UPC 2
VLAN 10
VLAN 10
Multicast VoQ
8000 multicast VoQs to eliminate HOLB and help ensure high throughput
Packets with different egress UPC fanouts are assigned to different VoQs
Egress UPC 1
Unicast VOQ
224.1.1.1
Multicast VOQ
Ingress UPC
Unified
Crossbar
Fabric
Egress UPC 2
224.1.1.1
224.1.1.2
Egress UPC 3
224.1.1.2
224.1.1.2
16MB
Multicast VOQ
Ingress UPC
Unicast VOQ
Egress UPC
448
Gbps
224
Gbps
Unified
Crossbar
Fabric
Unicast
Flow control
Unicast
Queue
Port 1
Multicast
Multicast
Queue
Unicast
Queue
Multicast
Queue
Port 2
Egress UPC 1
Ingress UPC
Multicast VOQ
Unicast VOQ
Port 1
Port 3
Selected
Unified
Crossbar
Fabric
Port 2
Egress UPC 2
Port 3
Selected
Port 3
Port 4
Packets that fail RPF check are copied to CPU to generate PIM assert.
A filter is implemented so that only one packet from small flow is copied to CPU
Multicast data packets are copied to CPU to create multicast forwarding state.
1.1.1.10
H1
v3 Report
(224.0.0.22)
IP-Based Forwarding
Vlan10, 10.0.0.1, 224.1.1.1, eth1/1
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Security ACL
MAC, IPv4, and IPv6 ACLs
PACL: ACL enabled under L2 interface
Ingress Processing
PACL
In
VACL
In
RACL
In
Egress Processing
Unified
Crossbar
Fabric
RACL
Out
VACL
Out
RBACL
ACL Type
Attach Point
Direction
PACL
L2 interface, L2 PortChannel
IN
VACL
VLAN
IN OUT
RACL
IN OUT
4096
Entries
Each IPv6 ACL (without port range) requires two TCAM entries.
N5600-TME1# sh platform afm info tcam 0 region ifacl
Default TCAM Partition
ifacl tcam TCAM configuration for asic id 0:
[ vacl tcam]: range
0 - 1023
VACL(1024)
[ifacl tcam]: range 1024 - 2175 *
[ qos tcam]: range 2176 - 2623
IFACL(1152)
[rbacl tcam]: range 2624 - 3775
[ span tcam]: range 3776 - 3839
QoS(448)
[ sup tcam]: range 3840 - 4095
RBACL(1024)
SPAN(64)
Control Traffic(256)
N5600-TME1#
VACL(1024)
4096
Entries
IFACL(1152)
QoS(448)
RBACL(1024)
SPAN(64)
Control Traffic(256)
TCAM Carving
Creating a template:
switch(config-tcam-templ)# ifacl ?
<320-3584> Enter size of ifacl region (in 64 entry increments)
Displaying Template
Ethernet1/10
access-group ip-list-1 in
Ethernet1/11
access-group ip-list-1 in
Ethernet1/12
access-group ip-list-1 in
Label
xyz
eth1/10
xyz
eth1/11
xyz
eth1/12
xyz
ACL Logging
ACL logging feature allows the logging of the packets hitting IPv4/IPv6
ACL.
Logs only packets that hit the deny rule (For mgmt0, logs permit/deny)
10
10002000
11
12
10
11
12
ACE Expansion
Software turns one ACL rule with UDP and TCP port range to multiple ACE entries in TCAM.
Each TCAM field has an associated mask including the UDP or TCP port field.
The mask can be used to support UDP and TCP port ranges.
User-Configured ACL Rule
1
2
3
4
Src IP
Dst IP
Protocol
Src port
Dst port
Content
Permit
10.1.1.0
20.1.1.0
TCP
Mask
0.0.0.255
0.0.0.255
255.255
0.63
Content
Permit
10.1.1.0
20.1.1.0
TCP
64
Mask
0.0.0.255
0.0.0.255
255.255
0.31
Content
Permit
10.1.1.0
20.1.1.0
TCP
96
Mask
0.0.0.255
0.0.0.255
255.255
0.3
Content
Permit
10.1.1.0
20.1.1.0
TCP
100
Mask
0.0.0.255
0.0.0.255
255.255
0.0
Port Range
0-63
Port Range
64-95
Port Range
96-99
Port Range
100
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Quality of Service
Traffic classification
DSCP, CoS, IP Precedence and ACL
Packet marking
DSCP, CoS, and ECN
DCBX 802.1Qaz
Packet Classification
Before:
HTTP
email
vmotion
backup
mgmt
VoIP/Lync
After:
Controlled Bandwidth
Email/HTTP
SAP
Backup
Packet Buffering
25MB packet buffer is shared by every three 40 GE ports or twelve 10 GE ports.
Buffer is 16MB at ingress and 9MB at egress.
Unicast packet can be buffered at both ingress and egress.
16MB
Multicast VOQ
Ingress
UPC
Unicast VOQ
224
Gbps
Unified
Crossbar
Fabric
448
Gbps
9MB
On by default, no configuration
needed
SPAN
Control
Port 2
Dedicated
Port 3
Dedicated
10 GE Port
40 GE Port
64 KB
67.2 KB
38.4 KB
153.6 KB
100 KB
100 KB
Shared buffer
13.2 MB
14.7 MB
queue-limit under network-qos policy specifies the dedicated buffer for each port and each class.
The dedicated buffer can be used by the port for only that class of service.
Without queue-limit each class of service will get 100 KB of dedicated buffer.
The size of dedicated buffer can be different for different classes of service. The policy applies to all
ports in the chassis.
Total ingress buffer minus the dedicated buffer and buffer for control and SPAN will be in the shared
buffer pool.
The following example sets the dedicated buffer for class-default to be 400 KB for all ports.
switch(config)# policy-map type network-qos Policy-buffer
switch(config-pmap-nq)# class type network-qos class-default
switch(config-pmap-nq-c)# queue-limit 400000 bytes
switch(config-pmap-nq-c)# system qos
switch(config-sys-qos)# service-policy type network-qos Policy-buffer
Unicast Buffer
Multicast Buffer
Future software will provide CLI to tune the egress buffer allocation.
10 GE Port
40 GE Port
363 KB
4.3 MB
6.6 MB
switch(config)# hardware random-detect min-thresh 10g 10g-min-threshold 40g 40g-minthreshold max-thresh 10g 10g-max-threshold 40g 40g-max-threshold ecn qos-group qosgroup-number
MAC
Traffic
Classification
Ingress
Cos/DSCP
Marking
Ingress
Policing
Forwarding
Per-class
Buffer Usage
Monitoring
Ingress Buffer
Egress Queues
MAC
Egress UPC
WRED/ECN
Marking
Strict Priority +
DWRR Scheduling
Egress
Scheduling
Egress Buffer
Crossbar
Fabric
VoQs
(8 per egress port)
Agenda
Introduction
Architecture
Forwarding
Multicast
ACL
QOS
Conclusion
Conclusion
High Performance
Line rate L2 and L3 at all
frame sizes
1-microsecond port-to-port
latency
True 40-Gbps flow
40-Gbps FCoE
Cut-through switching for
40 and 10 GE
25-MB buffer per port ASIC
800-MB maximum (on
5696Q)
High Scalability
Up to 224,000 MAC
(Flexible Carving)
Up to 128,000 ARP
(Flexible Carving)
32,000 LPM
16,000 bridge domains
Feature-Rich
Sampled NetFlow
BRKDCT-1980 -Advanced
Analytics in Nexus Switches
Promote your favorite speaker through Twitter and you could win $200 of Cisco
Press products (@CiscoPress)
@mountainviewer
Two hashtags: #CLUS #MyFavoriteSpeaker
You can submit an entry for more than one of your favorite speakers
Table Topics
Related sessions
Thank you