Layer 3 Forwarding and Troubleshooting...
Transcript of Layer 3 Forwarding and Troubleshooting...
Layer 3 Forwarding and Troubleshooting Deep
Dive on Nexus 9000/3000
Ambrish Mehta – Principal Engineer (INSBU Engineering)
Swami Narayanan – Principal Engineer (INSBU Engineering)
Session Abstract
• Closer look at various Layer 3 forwarding options available on Nexus 95xx/93xx/31xx/32xx data center switches.
• Details on Layer 3 forwarding in multi chip and single chip hardware architecture.
• Flexibility offered by Forwarding Engine(ASIC) and how NX-OS is utilising it to meet different deployment needs.
• Empower Network Engineer/Architect to identify ideal routing mode for network deployment.
Glossary of TermsNFE Network Forwarding Engine (Trident2 ASIC)
NFE2 Network Forwarding Engine 2 (Tomahawk ASIC)
DMOD Destination Module (ASIC Instance)
DPORT Destination Port (Port within ASIC instance)
UFT Unified Forwarding Table
LC Line card Module
FM Fabric Module
LPM Longest Prefix Match (e.g IPv4 : /31 to /0, IPv6 /127 to /0)
ALPM Algorithmic Longest Prefix Match
ECMP Equal Cost Multipath
• Introduction to 95xx/93xx/31xx/32xx
• Layer 3 Forwarding Architecture
• NFE/NFE2 Forwarding Architecture
• NX-OS Routing Modes and Deployment Scenarios
• Walk Through of Commonly used Layer 3 Commands
• Q & A
Agenda
Nexus 95xx/93xx Product Family
Nexus 9500 Series Switches
Nexus9500 Series
Nexus 9300 Series Switches
Nexus9504/Nexus9508/Nexus9516N9K-C9332PQ N9K-C9372PX N9K-C9372TX
N9K-C9396PX N9K-C93128TX
Nexus9300 Series
Nexus 31xx/32xx Product Family
Nexus 31xx Series Switches
Nexus 31xx
Nexus 32xx Series Switches
N3K-3132Q N3K-3164Q N3K-3232-C N3K-3264-C
Nexus 32xx
Layer 3 Forwarding Architecture Categories
• Modular Multi-Chip
• Modular switches (95xx) belongs to this category.
• LC and FM Chip operates independently.
• Fixed Multi-Chip
• 3164Q belongs to this category
• Virtual/Pseudo LC and FM operates independently.
• Fixed Single-Chip
• 93xx, 3132Q, 3232-C and 3264-C belongs to this category
• Single Chip is making all forwarding decisions.
Modular Multi-Chip: LC and FM Connectivity
• NFE: Network Forwarding Engine (Forwarding ASIC)
• Each of the 3 NFEs on the line card has 12 x 40 Gbps links to fabric modules
• The above diagram shows the line card fabric connectivity in a Nexus 9508 chassis as an example
• LC and FM are connected via HiGig Trunk (Port-Channel)
• Each Front Panel port in the switch has unique identifier (DMOD/DPORT)
NFE NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 2
NFE NFE
Fabric 3
NFE NFE
Fabric 4
NFE NFE
Fabric 5
NFE NFE
Fabric 6
NFE
HiGig Trunk (Port-Channel)
36-port 40 Gbps QSFP+
Line Card/IO moduleDMOD 1 DMOD 2 DMOD 3
DMOD 100 DMOD 100 DMOD 100 DMOD 100 DMOD 100 DMOD 100
Interface DMOD DPORT
Eth1/1 1 1
Eth1/36 3 12
Fabric
Module
Modular Multi-Chip : Layer 3 Building Blocks
• Separate but identical NFE is present on LC and FM
• We can make use of both to program routes and get higher scale
• Treat LC and FM NFEs as independent forwarding entity
• Treat IPv4 Host, IPv4 LPM, IPv6 Host and IPv6 LPM as independent entity
• Identify programming location of prefixes blocks based on the Configured Routing Mode.
IPv4 Host IPv4 LPM
IPv6 Host IPv6 LPM
Prefix Blocks
Modular Multi-Chip : Layer 3 Building Blocks
IPv4 Host IPv4 LPMIPv6 Host IPv6 LPM
Fabric Module
Line Card
NFE
NFE
IPv4 Host IPv6 Host
IPv6 LPMIPv4 LPM
Fabric Module
Line Card
NFE
NFE
IPv4 Host
IPv6 Host
IPv4 LPM
Fabric Module
Line Card
IPv6 LPM
NX-OS SoftwareOption
1Option
2
Option
3
NFE
NFEIPv4 Host
IPv6 Host IPv6 LPM
IPv4 LPM
Modular Multi-Chip : Hierarchical Routing
NFE NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
36-port 40 Gbps QSFP+
… Fabric Modules
LC and FM both have same NFE
Use 2 Stage Forwarding Decision to efficiently utilise LC and FM tables (Higher Route Prefix Scale)
LC NFE handles 1 set of forwarding lookup, FM NFE handles other set
LC NFE forwarding decision HIT, Egress Point of Exit (DMOD, DPORT) derived
Packet forwarded to Egress Interface directly.
Incase Egress interface in different NFE, then goes through FM NFE (only Hybrid Switching and no
forwarding lookup)
Hierarchical Routing – LC Lookup
NFE NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
…Fabric ModulesHybrid Switching
LC NFE HIT ->
Egress DMOD,
DPORTDMOD 3DMOD 1 DMOD 2
3 12 Data Packet
HiGig
Header
Eth1/1 Eth1/36
• LC NFE – Catch All Entry to punt packet to FM NFE for lookup
• Special DMOD (DMOD 100) in HiGiG Header. FM NFE does forwarding lookup if DMOD is 100
• FM NFE Lookup HIT – Packet Forwarded to Egress Interface (DMOD, DPORT)
• FM NFE Lookup NO HIT – Packet punted from FM NFE to SUP for ICMP Unreachable (Rate policed)
Hierarchical Routing – FM Lookup
NFE NFE NFE
NFE
DMOD 100
NFE NFE
DMOD 100
NFE
Line Card/IO module
…Fabric Modules
LC NFE Miss, Hit
Catch All -> Punt to
FM (DMOD 100)
FM NFE Hit =>
Egress DMOD,
DPORT
100 - Data Packet
HiGig
Header
3 12 Data Packet
HiGig
Header
DMOD 1 DMOD 3DMOD 2
Eth1/1 Eth1/36
• Single Stage Forwarding Decision (only LC NFE)
• Fabric NFE used for Hybrid Switching only
• Scale limited to Single Chip mode setting
• No HIT, packets punted to SUP for ICMP Unreachable (Rate policed)
Non Hierarchical Routing
NFE NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
… Fabric Modules
Hybrid
Switching
LC NFE HIT -> Egress
DMOD, DPORT
Eth1/1 Eth1/36
Fixed Multi-Chip : Pseudo LC and FM Connectivity
Same as 95XX Modular Chip Forwarding Architecture
Use 2 Stage Forwarding Decision to efficiently utilise LC and FM tables (Higher Route Prefix Scale)
LC NFE handles 1 set of forwarding lookup, FM NFE handles other set
All modes supported in Fixed Multi-Chip operates same as Modular Multi Chip
NFE NFE NFE
NFE NFE
Pseudo LC
64-port 40 Gbps QSFP+
Pseudo FM
NFE
HiGig Trunk (Port-Channel)
DMOD 1 DMOD 3DMOD 2
DMOD 100
DMOD 4
Fixed Single-chip
• All IPv4 and IPv6 host and LPM routes are programmed on this Single NFE2
• Single ASIC needs to do all forwarding decision
NFE2
32 x 100 Gbps
IPv4: Host and LPM Routes
IPv6: Host and LPM Routes
N3k-3232-C
Unified Forwarding Table Bank Allocation
Bank 0-1
Dedicated
L2 Bank
Banks 2-3
Configurable
for
MAC/Hosts/
LPM
Bank 6-9
Dedicated
L3 Hosts
(V4 + V6
combined)
Banks 4-5
Configurable
for
MAC/Hosts/
LPM
16K LPM
Table for V4
and V6
LPM routes
UFT Mode 2
Bank 0-1
Dedicated
L2 Bank
Banks 2-3
Carved for
Layer 2
MAC
Bank 6-9
Dedicated
L3 Hosts
(V4 + V6
combined)
Banks 4-5
Carved for
Layer 3 V4
& V6 Hosts
16K LPM
Table for V4
and V6
LPM routes
UFT Mode 3
Bank 0-1
Dedicated
L2 Bank
Bank 2
Carved for
L2
Bank 6-9
Dedicated
L3 Hosts
Hosts(V4 +
V6
combined)
Banks 4-5
Carved for
Layer 3 V4
+ V6 Hosts
16K LPM
Table for V4
and V6
LPM routes
Bank 3
Carved for
L3 V4 + V6
Hosts
NFE TCAM LPM Table
IPv4/IPv6
(<=/64)
IPv6
(/65 to /127)
IPv6 /65 to /127 prefix length carving configuration
N9k(config)# hardware profile ipv6 lpm-entries maximum <256-3072>
Warning: The command will take effect after next reload.
N9k(config)#
16K
LPM Type # TCAM
Entries
LPM Scale
V4 * 1 12K/4K
V6 (<=/64) * 2 6K/2K
V6 (/65 to /127) 4 1K/3K**
* Shared TCAM space between IPv4 LPM and IPv6 LPM <=/64
** Default TCAM allocation allows1K prefixes IPv6 /65 to /127
1
1 1
1 1 1 1
UFT Mode 4
Bank 0-1
Dedicated
L2 Bank
Banks 2-3
Carved for
storing the
V4/V6 LPM
prefixes
Bank 6-9
Dedicated
L3 Hosts
(V4 + V6
combined)
Banks 4-5
Carved for
storing the
V4/V6 LPM
prefixes
16K LPM
Table
(Indexing to
Bank2-5)
ALPM and IPv6 Modes
• Algorithmic LPM - Repurposed UFT table to achieve higher LPM scale for IPv4 and IPv6
• Supports two modes for IPv6 LPM (64B and 128B)
• IPv6 mode setting at NFE level [not individual TCAM banks]
128B mode : All LPM prefix length uses128 bit width entries
64B mode : No /65 to /127 route can be programmed on NFE
• 128B Mode : V6 LPM Scale Tunable with configuration
IPv6 ALPM ModesALPM Mode NFE NFE2
64B V4 LPM : 128K
V6 LPM : 80K
V4 LPM : 128K
V6 LPM : 80K
128B V4 LPM : 128K/64K
V6 LPM : 16K/20K *
V4 LPM : 128K/64K
V6 LPM : 10K/15K *
* Achieved with carve-value configuration knob
N9k(config)# hardware profile ipv6 alpm carve-value <1024-3072>
Warning: The command will take effect after next reload.
N9k(config)#
Route Lookup and Rewrite
• Unicast Route Forwarding has 2 components
Route Lookup
Rewrite packet for next hop
• Rewrite information contains (Next Hop MAC, Egress Interface)
• ARP/Neighbor Discovery provides the information
• NFE stores these details in Rewrite table (Egress Next Hop Table)
Egress NH Table Table Size – 48k/32k (NFE/NFE2)
UFT Scale
Mode L2 L3 Host * LPM *
0 288K 16K 16k
1 224k 56k 16k
2 160k 144k 16k
3 96k 208k 16k
4 32k 16k 128k
* Number of IPv4 Entries
Mode L2 L3 Host * LPM *
0 144K 8k 16k
1 104k 40k 16k
2 72k 72k 16k
3 40k 104k 16k
4 8k 8k 128k
NFE NFE2
Default Routing Mode: Multi-Chip
• Hierarchical Routing on 95xx/3164 switches.
• Suitable for most deployment scenarios.
• Leverages LC and FM NFE for unicast route programming.
• Provides good balanced scale between Host and LPM prefixes for IPv4 and IPv6.
Nexus 950X/3164 Default Routing Mode: Hierarchical
• V4 & V6 Host Routes • Forwarding Lookup performed in LC NFE. Egress
NFE and Port derived
• Packets are sent directly to Egress NFE (FM NFE
acts as bypass and it performs HiGig switching for
this packet)
• V4 & V6 LPM Routes• No host route match in LC NFE
• LC LPM table has catch all default route to redirect
the packets towards FM NFE for LPM lookup
• FM NFE performs the forwarding lookup to
determine Egress NFE and Port
• Data traffic forwarded to Egress NFE
Line Card
UFT Mode 3
Fabric Module
UFT Mode 4
UFT Mode L3 Host LPM
3 208k 16k *
4 N/A 128k
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 LPM
...
* Catch All Default Route (0/0 and 0::/0) is programmed.
Nexus 950X/3164 : Default Mode - Host Lookup
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
…IPv4: /0 to /31
IPv6: /0 to /127Fabric Modules
IPv4: /32
IPv6: /128
IPv4: 0/0
IPv6: 0/0
Host Route Match ->
Dst NFE, Port
HiGiG
Switching
Packet In
Dest : 1.1.1.1Packet Out
N9K1#show ip route 1.1.1.1
1.1.1.1/32, ubest/mbest: 1/0, attached
*via 10.0.25.4, [20/0], 01:44:03…
N9K1#
Nexus 950X/3164 : Default Mode LPM Lookup
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
…IPv4: /0 to /31
IPv6: /0 to /127Fabric Modules
IPv4: /32
IPv6: /128
IPv4: 0/0
IPv6: 0/0
HOST Lookup: Fail
LPM Lookup: System
Default Route to FM
LPM Lookup
=> Dst NFE,
Port
Packet In
Dest: 2.2.2.2Packet Out
Punt to FM for
LPM lookup
N9K1#show ip route 2.2.2.2
2.2.2.0/24, ubest/mbest: 1/0, attached
*via 10.0.25.6, [20/0], 01:44:03…
N9K1#
Scale Number : Hierarchical Routing Mode
ASIC V4 Host V6 Host V4 LPM V6 LPM
NFE 208K (48k )̂ 104K 128K/64k 16k/20K +
NFE2 104K (32k )̂ 52K 128K/64k 10k/15K +
N9k-1# show system routing mode
Configured System Routing Mode: Default
Applied System Routing Mode: Hierarchical (Default)
Configured SVI post-routed unknown-unicast hardware flood mode: enabled
N9k-1#
+ “hardware profile ipv6 alpm carve-value <#>” configuration knob
^ ARP/ND
Background and Requirements
• Default Hierarchical Mode is good for most deployments.
• Enhance NX-OS software to further utilize flexibility of Forwarding ASIC.
• Move around UFT Modes and IPv4/IPv6 prefix blocks to address different deployment needs.
• “My Network has around IPv6 60k (<= /64 prefix length), Is there a way to get higher IPv6 LPM scale ?”
• “My deployment is for L2/L3 boundary role and I have large number of hosts behind these Vlans. Is there a way to get higher ARP/ND scale ?”
• “My Network’s routing table size is very small, is there a way to avoid 2 stage look up ?”
Deployment Scenario 1
• Large scale IPv6 LPM (/64) prefixes with some IPv4 LPM as well.
• Below is prefix scale with some growth room built in.
IPv4: Host: <1k LPM: 2k
IPv6: Host: <1k LPM: 60k (< /64) + 2k (/65 to /127)
Nexus 950X/3164 Hierarchical 64B ALPM
• V4 & V6 Host Routes • Forwarding Lookup performed in LC NFE,
Egress NFE and Port derived
• Packets are sent directly to Egress NFE (FM
NFE acts as bypass and it performs HiGig
switching for this packet)Line Card
UFT Mode 3
Fabric Module
UFT Mode 4
64Bit
UFT Mode L3 Host LPM
3 208k 16k
4 N/A 128k
IPv4 Host
IPv6 Host
...
Nexus 950X/3164 Hierarchical 64B ALPM• V4 LPM & V6 <= /64 prefix length LPM Routes
• LC LPM table has default route to punt the
packets towards FM NFE for LPM lookup
• FM NFE performs the forwarding lookup to
determine Egress NFE and Interface
• Data traffic forwarded to Egress NFE
• V6 LPM prefix /65 to /127
• Programmed in LC NFE
• FM NFE can take only V6 prefix length <= /64
• IPv6 prefix length /65 to /127 programmed in
LC NFE in the LPM table. Hence scale limited
to maximum of 3k
• IPv6 prefixes matching LPM prefixes send
directly to Egress NFE
UFT Mode L3 Host LPM
3 208k 16k
4 N/A 128k
Line Card
UFT Mode 3
Fabric Module
UFT Mode 4
64Bit
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 /64 LPM
...IPv6 /65 to /127 LPM
Nexus 950X/3164 : 64B ALPM - Host Lookup
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
… Fabric Modules
IPv4: 0/0
IPv6: 0/0
Host Route Match ->
Dst NFE, Port
HiGiG
Switching
Packet In
Dest: 1:1:1::1Packet Out
IPv4: /32
IPv6: /128
IPv6: /65 to /127
IPv4: /0 to /31
IPv6: /0 to /64
N9K1#show ipv6 route 1:1:1.:1
1:1:1::1/128, ubest/mbest: 1/0, attached
*via 4001::1, [20/0], 01:44:03…
N9K1#
Nexus 950X/3164 : 64B ALPM LPM (V4 <=/31 and V6 <= /64)
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
…IPv4: /0 to /31
IPv6: /0 to /64 Fabric Modules
IPv4: /32
IPv6: /128, /65 to /127
ALPM
Lookup =>
Dst NFE, Port
Packet In
Dest: 2:2:2::1Packet Out
Punt to FM for
LPM lookup
HOST Lookup: Fail
LPM Lookup: System
Default Route to FM
IPv4: 0/0
IPv6: 0/0
N9K1#show ipv6 route 2:2:2::1
2:2:2::/48, ubest/mbest: 1/0, attached
*via 2001::1, [20/0], 01:44:03…
N9K1#
Nexus 950X/3164 : 64B ALPM – V6 LPM (/65 to /127)
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
… Fabric Modules
IPv4: 0/0
IPv6: 0/0
HOST Lookup: Fail
LPM Lookup: Route
Match -> Dst NFE, Port
HiGiG
Switching
Packet InDest: 3:3:3:3:3::3
Packet Out
IPv4: /32
IPv6: /128, /65 to /127
IPv4: /0 to /31
IPv6: /0 to /64
N9K1#show ipv6 route 3:3:3:3:3::3
3:3:3:3:3::/80, ubest/mbest: 1/0, attached
*via 2002::1, [20/0], 01:44:03…
N9K1#
Scale Number : 64B ALPM ModeASIC V4 Host V6
Host
V4 LPM V6 LPM
(<=64)
V6 LPM
(65-127)
NFE 208K (48k )̂ 104K 128K 80K 1K-3K*
NFE2 104K (32k )̂ 52K 128K 80K 1K-3K*
N9k-1(config)# system routing hierarchical max-mode l3 64b-alpm
Warning: The command will take effect after next reload.
Note: This requires copy running-config to startup-config before switch reload.
N9k-1#
* “hardware profile ipv6 lpm-entries maximum <#>” configuration knob
^ ARP/ND
Deployment Scenario 2
• L2/L3 aggregation role with large scale ARP and ND needs.
• Below is prefix scale with some growth room built in
IPv4: Host(ARP): 60k LPM: 3k
IPv6: Host(ND): 40k LPM: 1k
Nexus 950X Max Host Mode• V4 Host & LPM Routes
• On LC NFE only a V4 default route
programmed. Redirects all V4 hosts & LPM
traffic to FM NFE
• FM NFE does best prefix match for both
host and LPM prefixes
• Packets are sent to Egress NFE Line Card
UFT Mode 2
Fabric Module
UFT Mode 3
IPv4 Host
IPv4 LPM
...
Mode MAC L3 Host LPM
2 160k 144k 16k
3 N/A 208k 16k
Nexus 950X Max Host Mode• V6 Host & LPM Routes
• V6 Hosts and LPM prefixes programmed in
LC NFE
• LC NFE performs the forwarding lookup to
determine Egress NFE and Interface.
• Data traffic forwarded to Egress NFE
directly Line Card
UFT Mode 2
Fabric Module
UFT Mode 3
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 LPM
...
Mode MAC L3 Host LPM
2 160k 144k 16k
3 N/A 208k 16k
Modular Multi-Chip : Max Host Mode – V6 Lookup
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
… Fabric Modules
IPv4: 0/0
V6 Route Match on LC
-> Dst NFE and Port
HiGiG
Switching
Packet In
Dest: 1:1:1::1Packet Out
IPv6: /0 to /128
IPv4: /0 to /32
N9K1#show ipv6 route 1:1:1::1
1:1:1::/48, ubest/mbest: 1/0, attached
*via 2001::1, [20/0], 01:44:03…
N9K1#
Modular Multi-Chip : Max Host Mode – V4 Lookup
NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
…IPv4: /0 to /32Fabric Modules
V4 LPM Route ->
System Default Route
Match on LC
Host/LPM
lookup for V4 ->
Dst NFE, Port
Packet In
Dest: 1.1.1.1Packet Out
Punt to FM for
LPM lookup
IPv6: /0 to /128
IPV4: 0/0 to FM
N9K1#show ip route 1.1.1.1
1.1.1.1/32, ubest/mbest: 1/0, attached
*via 10.0.25.4, [20/0], 01:44:03…
N9K1#
Scale Number : Max Host Mode
ASIC V4 Host V6
Host
V4
LPM
V6 LPM
(<=64)
V6 LPM
(65-127)
NFE 208K (60K )̂ 104K 16K 6K 1K-3K*
N9k-1(config)# system routing max-mode host
Warning: The command will take effect after next reload.
Note: This requires copy running-config to startup-config before switch reload.
N9k-1#
* “hardware profile ipv6 lpm-entries maximum <#>” configuration knob
^ ARP
Deployment Scenario 3
• Network prefixes are hierarchical and aggregated.
• Below is prefix scale with some growth room built in.
IPv4: Host: 16k LPM: 6k
IPv6: Host: 8k LPM: 3k
• If possible, avoid two stage forwarding lookup.
Nexus 950X/3164 Non Hierarchical Routing Mode• All V4/V6 Unicast forwarding decision done
at LC NFE
• V4/V6 Host & LPM Routes• Forwarding Lookup performed in LC NFE,
Egress NFE and Port derived for all L3
Unicast traffic (both Host and LPM routes)
• Packets are sent directly to Egress NFE (FM
NFE acts as bypass and it performs HiGig
switching for this packet)
UFT Mode L3 Host LPM
3 208k 16k
Line Card
UFT Mode 3
Fabric Module
UFT Mode X
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 LPM
...
Nexus 950X/3164 :Non Hierarchical – Any Lookup
NFE NFE NFE
NFE
Fabric 1
NFE NFE
Fabric 6
NFE
Line Card/IO module
… Fabric Modules
IPv4: /0 to /32
IPv6: /0 to /128
HiGiG
Switching
Packet In Packet Out
Host/LPM Route
Match -> Dst NFE,
Port
Scale Number : Non Hierarchical Routing ModeASIC V4 Host V6 Host V4 LPM V6 LPM
(<=64)
V6 LPM
(65-127)
NFE 208K (48k )̂ 104K 12K/4K 6K/2K 1K-3K*
NFE2 104K (32k )̂ 52K 12K/4K 6K/2K 1K-3K*
N9k-1(config)# system routing non-hierarchical
Warning: The command will take effect after next reload.
Note: This requires copy running-config to startup-config before switch reload.
N9k-1#
* “hardware profile ipv6 lpm-entries maximum <#>” configuration knob
^ ARP/ND
Modular Multi-Chip Summary
IPv4 Host IPv4 LPMIPv6 Host IPv6 LPM
Fabric Module
Line Card
NFE
NFE
IPv4 Host IPv6 Host
IPv6 LPMIPv4 LPM
Fabric Module
Line Card
NFE
NFE
IPv4 Host
IPv6 Host
IPv4 LPM
Fabric Module
IPv6 LPM
Default
Mode
Non-Hierarchical
ModeMax-host
Mode
NX-OS Software
Line Card
NFE
NFEIPv4 Host
IPv6 Host IPv6 LPM
IPv4 LPM
Nexus 93xx/31xx/32xx Default Routing
UFT Mode 3
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 LPM
Mode MAC L3 Host LPM
3 96k 208k 16k
Scale Number : 93xx/31xx/32xx – Default Mode
N9k-1# show system routing mode
Configured System Routing Mode: Default
Applied System Routing Mode: Max Host (Default)
N9k-1#
* “hardware profile ipv6 lpm-entries maximum <#>” configuration knob
ASIC V4 Host V6
Host
V4 LPM V6 LPM
(<=64)
V6 LPM
(65-127)
NFE 208K (48k )̂ 104K 12K/4K 6K/2K 1K-3K*
NFE2 104K (32k )̂ 52K 12K/4K 6K/2K 1K-3K*
^ ARP/ND
Nexus 93xx/31xx/32xx Max L3 Mode
Mode MAC L3 Host LPM
4 32k 16k 128k
UFT Mode 4
IPv4 Host
IPv4 LPM
IPv6 Host
IPv6 LPM
Scale Number: 93xx/31xx/32xx – Max L3
ASIC V4 Host V6 Host V4 LPM V6 LPM
(0-127)
NFE 16K 8K 128K 16K-20K +
NFE2 8K 4K 128K 10K-15K +
N9k-1(config)# system routing max-mode l3
Warning: The command will take effect after next reload.
Note: This requires copy running-config to startup-config before switch reload.
N9k-1#
+ “hardware profile ipv6 alpm carve-value <#>” configuration knob
System Routing ModeN9K1# show system routing mode Configured System Routing Mode: Default
Applied System Routing Mode: Hierarchical (Default)
Configured SVI post-routed unknown-unicast hardware flood mode: enabled
N9K1# conf terminal Enter configuration commands, one per line. End with CNTL/Z.
N9K1#(config)# system routing non-hierarchical Warning: The command will take effect after next reload.
Note: This requires copy running-config to startup-config before switch reload.
N9K1#(config)# end
N9K1# show system routing mode Configured System Routing Mode: Non-Hierarchical
Applied System Routing Mode: Hierarchical (Default)
Configured SVI post-routed unknown-unicast hardware flood mode: enabled
N9K1#
Switch is not reloaded after mode change
IPv4 Host Route With Single Nexthop Example
N9K1#show ip route 1.1.1.203
1.1.1.203/32, ubest/mbest: 1/0, attached
*via 10.0.25.4, [20/0], 01:44:03, bgp-300, external, tag 65012
N9K1#show forwarding ip route 1.1.1.203/32 module 1
IPv4 routes for table default/base
------------------+-----------------------------------------+----------------------+-----------------+-----
Prefix | Next-hop | Interface | Labels | Partial Install
------------------+-----------------------------------------+----------------------+-----------------+------
*1.1.1.203/32 10.0.25.4 Ethernet1/1
/32 Host Route
NextHop Physical Interface
Forwarding Routing Modes VerificationN9K1#show forwarding ip route summary module 1
Module Type : Line-Card
Module Mode : Mode-3
Module Route Download-type : Host only (IPv4+IPv6) (1)
…
N9K1#show forwarding ip route summary module 22
Module Type : Fabric-Module
Module Mode : ALPM (Mode-4)
Module Route Download-type : LPM only (IPv4+IPv6) (2)
…
N9K1#
Line Card is running in Mode 3
Fabric Module is running in Mode 4
bcm-shell command reference
Command Usage
l3 l3table show Display IPv4 Host Entries
l3 defip show Display IPv4 LPM Entries
l3 ip6host show Display IPv6 Host Entries
l3 ip6route show Display IPv6 LPM Entries
l3 multipath show Display ECMP object and Nexthop List
l3 egress show Display Egress object with rewrite information
show c rpkt Display Received packets in Hardware
show c tpkt Display Transmitted packets in Hardware
IPv4 Host Route With Single Nexthop Example
N9K1#bcm-shell module 1 "0:l3 l3table show" | inc 1.1.1.203
211300 1 1.1.1.203 00:00:00:00:00:00 100009 0 0 0 n
N9K1#bcm-shell module 1 "0:l3 egress show 100009"Entry Mac Vlan INTF PORT MOD MPLS_LABEL ToCpu Drop RefCount L3MC
100009 60:73:5c:fe:e4:6e 4095 4239 13 2 -1 no no no
N9K1#show system internal ethpm info interface ethernet 1/1 | grep -i STATIC
IF_STATIC_INFO: port_name=Ethernet1/1,if_index:0x38078000,ltl=5952,slot=0,
nxos_port=48,dmod=2,dpid=13,unit=1,queue=2160,xbar_unitbmp=0x2,ns_pid=255,slice_num
=0,port_on_slice=0,src_id=0
Unique hardware
port identifier
Egress Object
Nexthop MAC
IPv4 LPM Route With ECMP Example
N9K1#show ip route 2.2.2.203
2.2.2.0/24, ubest/mbest: 2/0, attached
*via 10.0.25.4, [20/0], 01:44:03, bgp-300, external, tag 65012
*via 10.0.25.6, [20/0], 01:44:04, bgp-300, external, tag 65012
N9K1#show forwarding ip route 2.2.2.0/24 module 21
IPv4 routes for table default/base
------------------+-----------------------------------------+----------------------+-----------------+-----
Prefix | Next-hop | Interface | Labels | Partial Install
------------------+-----------------------------------------+----------------------+-----------------+------
*2.2.2.0/24 10.0.25.4 Ethernet1/1
10.0.25.6 Ethernet1/2
/24 LPM Route
NextHop Physical Interfaces
IPv4 LPM Route With ECMP Example N9K1#bcm-shell module 21 "0:l3 defip show" | inc 2.2.2.0/24
11012 1 2.2.2.0/24 00:00:00:00:00:00 200256 0 0 0 n
N9K1#bcm-shell module 21 "0:l3 multipath show 200256"
Multipath Egress Object 200256
Interfaces: 100009 100010
N9K1#bcm-shell module 21 "0:l3 egress show 100009"Entry Mac Vlan INTF PORT MOD MPLS_LABEL ToCpu Drop RefCount L3MC
100009 60:73:5c:fe:e4:6e 4095 4239 13 2 -1 no no no
N9K1#bcm-shell module 21 "0:l3 egress show 100010"
Entry Mac Vlan INTF PORT MOD MPLS_LABEL ToCpu Drop RefCount L3MC
100010 60:73:5c:fe:e4:6e 4095 4240 14 2 -1 no no no
ECMP Object
Egress Objects
Layer 3 Consistency CheckerN9K1# test forwarding inconsistency module 1
Consistency check started.
N9k-1# show forwarding inconsistency module 1
IPV4 Consistency check : table_id(0x1) slot(1)
Execution time : 15118 ms ()
No inconsistent adjacencies.
No inconsistent routes.
Consistency-Checker: PASS for 1
N9k-1# test forwarding ipv6 inconsistency module 1
V6 Consistency check started.
N9k-1# show forwarding ipv6 inconsistency module 1
IPV6 Consistency check : table_id(0x80000001) slot(1)
Execution time : 3130 ms ()
No inconsistent adjacencies.
No inconsistent routes.
Consistency-Checker: PASS for 1
N9k-1#
Complete Your Online Session Evaluation
Learn online with Cisco Live!
Visit us online after the conference
for full access to session videos and
presentations.
www.CiscoLiveAPAC.com
Give us your feedback and receive a
Cisco 2016 T-Shirt by completing the
Overall Event Survey and 5 Session
Evaluations.– Directly from your mobile device on the Cisco Live
Mobile App
– By visiting the Cisco Live Mobile Site http://showcase.genie-connect.com/ciscolivemelbourne2016/
– Visit any Cisco Live Internet Station located
throughout the venue
T-Shirts can be collected Friday 11 March
at Registration