H3C Data Center Switches DRNI Configuration Guide-6W103

HomeSupportConfigure & DeployConfiguration GuidesH3C Data Center Switches DRNI Configuration Guide-6W103
10-DRNI+EVPN Centralized Gateway Configuration Example

Contents

Example: Deploying DRNI and centralized EVPN gateway· 1

Network configuration· 1

Overlay connectivity models· 7

Applicable product matrix· 7

Configuring S6800 switches as leaf nodes· 8

Procedure summary· 8

Configuring resource modes· 8

Configuring OSPF· 8

Configuring the links towards the spine tier 11

Configuring L2VPN· 11

Configuring DRNI 13

Configuring the links towards the bare metal servers· 19

Configure the spanning tree feature· 20

Configuring an underlay BGP instance· 20

Configuring an EVPN BGP instance· 22

Configuring VSIs and ACs· 23

Configuring S6850 switches as border nodes· 24

Procedure summary· 24

Configure basic settings· 25

Configuring OSPF· 26

Configure the spanning tree feature· 28

Configuring the interfaces connected to the spine nodes· 28

Configuring L2VPN· 29

Configuring DRNI 30

Configuring the routed interfaces connected to the external network· 35

Configuring an underlay BGP instance· 35

Configuring an EVPN BGP instance· 37

Configuring the overlay network· 39

Configuring S12500X-AF switches as spine nodes· 41

Procedure summary· 41

Configuring OSPF· 41

Configuring the downlinks towards the leaf tier 42

Configuring the uplinks towards the border nodes· 43

Configuring an underlay BGP instance· 43

Configuring an EVPN BGP instance· 45

Traffic model 47

About the traffic model 47

Overlay traffic· 47

Convergence performance test results· 49

Failure test results· 49

Verifying the configuration· 50

Verification commands· 50

Procedure· 50

Upgrading the devices· 52

Upgrading a leaf device· 52

Upgrading a spine device· 53

Upgrading a border device· 54

Expanding the network· 55

Adding a leaf device· 55

Replacing hardware· 56

Replacing an interface module· 56

Replacing a switching fabric module· 56


Example: Deploying DRNI and centralized EVPN gateway

Network configuration

As shown in Figure 1, deploy two border devices as a DR system and use it as a centralized gateway. The DR system provides intra-DC connectivity, external connectivity, and DCI.

The following is the network configuration:

·     Use DRNI to build the two border devices into a DR system. Configure the DR system as a centralized EVPN gateway to provide L3 forwarding for VXLANs and as a border node to provide connectivity to the external network.

·     Deploy two S12500X-AF switches at the spine tier. Configure them as route reflectors (RRs) to reflect BGP EVPN routes among border and leaf devices.

·     Use DRNI to deploy two pairs of access switches (ToR) switches as DR systems at the leaf tier. They provide EVPN access services to connect servers (for example, bare metal servers) to their overlay networks.

Figure 1 Network diagram

Device

Interface

IP address

Remarks

Leaf 1

XGE 1/0/7

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected to Server A (bare metal).

XGE 1/0/8

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected to Server C (bare metal).

HGE 1/0/49

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 2/0/49 on Leaf 2.

HGE 1/0/50

N/A

Member port of the IPP.

Connected to HGE 2/0/50 on Leaf 2.

XGE 1/0/47

1.0.0.1/30

Keepalive link between DR member devices.

Connected to XGE 2/0/47 on Leaf 2.

HGE 1/0/53

N/A

IP address borrowed from Loopback 0.

Connected to HGE 1/0/25 on Spine 1.

XGE 1/0/17

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/1 on Spine 2.

Loopback 0

10.182.224.111/32

VTEP IP address for establishing BGP EVPN peering.

Loopback 1

10.182.226.111/32

Virtual VTEP address for the DR system to establish VXLAN tunnels to remote devices.

Vlan-interface 1999

192.168.220.1/30

IP address for establishing Layer 3 connectivity with the peer DR member device.

Leaf 2

XGE 2/0/7

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected Server A (bare metal).

XGE 2/0/8

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected to Server C (bare metal).

HGE 2/0/49

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/49 on Leaf 1.

HGE 2/0/50

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/50 on Leaf 1.

XGE 2/0/47

1.0.0.2/30

Keepalive link between DR member devices.

Connected to XGE 1/0/47 on Leaf 1.

HGE 2/0/54

N/A

IP address borrowed from Loopback 0.

Connected to HGE 1/0/28 on Spine 1.

XGE 2/0/17

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/2 on Spine 2.

Loopback 0

10.182.224.246/32

VTEP IP address for establishing BGP EVPN peering.

Loopback 1

10.182.226.111/32

Virtual VTEP address for the DR system to establish VXLAN tunnels to remote devices.

Vlan-interface 1999

192.168.220.2/30

IP address for establishing Layer 3 connectivity with the peer DR member device.

Leaf 3

XGE 1/0/7

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected Server B (bare metal).

XGE 1/0/21

N/A

Member port of the IPP for IPL establishment between DR member devices.

Connected to XGE 2/0/21 on Leaf 4.

XGE 1/0/22

N/A

Member port of the IPP for IPL establishment between DR member devices.

Connected to XGE 2/0/22 on Leaf 4.

XGE 1/0/17

1.1.0.1/30

Keepalive link.

Connected to XGE 2/0/17 on Leaf 4.

XGE 1/0/3

N/A

IP address borrowed from Loopback 0.

Connected to XGE 2/0/5 on Spine 1.

XGE 1/0/1

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/3 on Spine 2.

Loopback 0

10.182.224.121/32

VTEP IP address for establishing BGP EVPN peering.

Loopback 1

10.182.226.121/32

Virtual VTEP address for the DR system to establish VXLAN tunnels to remote devices.

Vlan-interface 1999

192.168.220.9/30

IP address for establishing Layer 3 connectivity with the peer DR member device.

Leaf 4

XGE 2/0/7

N/A

Member port of a DR interface, on which Ethernet service instances are configured to act as attachment circuits (ACs).

Connected to a bare metal server.

XGE 2/0/21

N/A

Member port of the IPP for IPL establishment between DR member devices.

Connected to XGE 1/0/21 on Leaf 3.

XGE 2/0/22

N/A

Member port of the IPP for IPL establishment between DR member devices.

Connected to XGE 1/0/22 on Leaf 3.

XGE 2/0/17

1.1.0.2/30

Keepalive link between DR member devices.

Connected to XGE 1/0/17 on Leaf 3.

XGE 2/0/3

N/A

IP address borrowed from Loopback 0.

Connected to XGE 2/0/7 on Spine 1.

XGE 2/0/1

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/4 on Spine 2.

Loopback 0

10.182.224.122/32

VTEP IP address for establishing BGP EVPN peering.

Loopback 1

10.182.226.121/32

Virtual VTEP address for the DR system to establish VXLAN tunnels to remote devices.

Vlan-interface 1999

192.168.220.10/30

IP address for establishing Layer 3 connectivity with the peer DR member device.

Spine 1

HGE 1/0/25

N/A

IP address borrowed from Loopback 0.

Connected to HGE 1/0/53 on Leaf 1.

HGE 1/0/28

N/A

IP address borrowed from Loopback 0.

Connected to HGE 2/0/54 on Leaf 2.

XGE 2/0/5

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/3 on Leaf 3.

XGE 2/0/7

N/A

IP address borrowed from Loopback 0.

Connected to XGE 2/0/3 on Leaf 4.

XGE 2/0/1

10.182.221.0/31

Connected to WGE 1/0/53 on Border 1.

XGE 2/0/2

10.182.221.10/31

Connected to WGE 1/0/53 on Border 2.

Loopback 0

10.182.224.90/32

IP address for underlay routing.

Loopback 1

10.182.226.90/32

IP address for overlay routing.

Spine 2

XGE 1/0/1

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/17 on Leaf 1.

XGE 1/0/2

N/A

IP address borrowed from Loopback 0.

Connected to XGE 2/0/17 on Leaf 2.

XGE 1/0/3

N/A

IP address borrowed from Loopback 0.

Connected to XGE 1/0/1 on Leaf 3.

XGE 1/0/4

N/A

IP address borrowed from Loopback 0.

Connected to XGE 2/0/1 on Leaf 4.

XGE 1/0/21

10.182.221.4/31

Connected to WGE 1/0/55 on Border 1.

XGE 1/0/22

10.182.221.14/31

Connected to WGE 1/0/55 on Border 2.

Loopback 0

10.182.224.89/32

IP address for underlay routing.

Loopback 1

10.182.226.89/32

IP address for overlay routing.

Border 1

WGE 1/0/53

10.182.221.1/31

Connected to XGE 2/0/1 on Spine 1.

WGE 1/0/55

10.182.221.5/31

Connected to XGE 1/0/21 on Spine 2.

HGE 1/0/25

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/25 on Border 2.

HGE 1/0/26

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/26 on Border 2.

WGE 1/0/1

2.0.0.1/31

Keepalive link between DR member devices.

Connected to WGE 1/0/1 on Border 2.

WGE 1/0/33

192.101.1.1/31

Connected to the L3 switch.

Loopback 0

10.182.234.1/32

IP address for the device to establish IGP and BGP peering as an edge device (ED).

Loopback 1

10.182.236.1/32

Virtual IP address for the DR system to establish IGP and BGP peering as an ED.

Vlan-interface 1001

192.101.1.101/31

IP address for establishing Layer 3 connectivity with the peer DR member device.

Border 2

WGE 1/0/53

10.182.221.11/31

Connected to XGE 2/0/2 on Spine 1.

WGE 1/0/55

10.182.221.15/31

Connected to XGE 1/0/22 on Spine 2.

HGE 1/0/25

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/25 on Border 1.

HGE 1/0/26

N/A

IPP for IPL establishment between DR member devices.

Connected to HGE 1/0/26 on Border 1.

WGE 1/0/1

2.0.0.2/31

Keepalive link between DR member devices.

Connected to WGE 1/0/1 on Border 2.

WGE 1/0/33

192.101.1.3/31

Connected to the L3 switch.

Loopback 0

10.182.234.2/32

IP address for the device to establish IGP and BGP peering as an ED.

Loopback 1

10.182.236.1/32

Virtual IP address for the DR system  to establish IGP and BGP peering as an ED.

Vlan-interface 1001

192.101.1.100/31

IP address for establishing a peer link with the peer DR member device.

 

Overlay connectivity models

The following are the types of connectivity between bare metal servers and between a bare metal server and the external network:

·     Layer 2 connectivity between bare metal servers attached to the same DR system at the leaf tier.

·     Layer 3 connectivity between bare metal servers attached to the same DR system at the leaf tier.

·     Layer 2 connectivity between bare metal servers attached to different DR systems at the leaf tier.

·     Layer 3 connectivity between bare metal servers attached to different DR systems at the leaf tier.

·     Layer 3 connectivity between bare metal servers and the external network.

Applicable product matrix

IMPORTANT

IMPORTANT:

In addition to running an applicable software version, you must also install the most recent patch, if any.

 

Role

Devices

Software version

Spine

S12500X-AF

This example uses S12500X-AF switches.

R2820

S12500G-AF

R7624P12

Leaf or border

S6800, S6860

This example uses S6800 switches as leaf nodes.

F2717 and higher 27xx versions

Do not use 28xx versions.

S6812, S6813 (only as leaf nodes)

Under development.

Contact H3C Support for version information.

S6805, S6825, S6850, S9850

This example uses S6850 switches as border nodes.

R6635

S6890

Not recommended.

S9820-64H (EVPN gateway not supported).

S9820-8C (EVPN not supported).

Not supported.

SDN controller

Contact H3C Support for version compatibility.

 

Configuring S6800 switches as leaf nodes

This example describes the procedure to deploy nodes Leaf 1 and Leaf 2. The same procedure applies to nodes Leaf 3 and Leaf 4.

Procedure summary

·     Configuring resource modes

·     Configuring OSPF

·     Configuring the links towards the spine tier

·     Configuring L2VPN

·     Configuring DRNI

·     Configuring the links towards the bare metal servers

·     Configure the spanning tree feature

·     Configuring an underlay BGP instance

·     Configuring VSIs and ACs

Configuring resource modes

Leaf 1

Leaf 2

Description

Purpose

Remarks

hardware-resource routing-mode IPv6-128

hardware-resource routing-mode IPv6-128

Enable support for IPv6 routes with prefixes longer than 64 bits.

N/A

Reboot the device for this setting to take effect.

The S6812 and S6813 switches do not support this command.

hardware-resource vxlan l2gw

hardware-resource vxlan l2gw

Set the VXLAN hardware resource mode to Layer 2 gateway mode.

N/A

Reboot the device for this setting to take effect.

The S6812 and S6813 switches do not support this command.

 

Configuring OSPF

Leaf 1

Leaf 2

Description

Purpose

Remarks

ospf 1 router-id 10.182.224.111

ospf 1 router-id 10.182.224.246

Enable an OSPF process and enter its view.

N/A

N/A

spf-schedule-interval 1 10 10

spf-schedule-interval 1 10 10

Set the maximum OSPF SPF calculation interval to 1 second, the minimum OSPF SPF calculation interval to 10 milliseconds, and the incremental OSPF SPF calculation interval to 10 milliseconds.

Shorten the SPF calculation interval to accelerate route convergence.

N/A

lsa-generation-interval 1 10 10

lsa-generation-interval 1 10 10

Set the maximum interval for LSA generation to 1 second, the minimum interval to 10 milliseconds, and the incremental interval to 10 milliseconds.

Enable quicker LSA regeneration upon network topology change to accelerate route convergence.

N/A

area 0.0.0.0

area 0.0.0.0

Create OSPF area 0.

N/A

N/A

fast-reroute lfa

fast-reroute lfa

Enable OSPF FRR and use the LFA algorithm for calculation of the backup next hop.

This feature minimizes service interruption by fast rerouting traffic to the backup path when a link or node fails.

N/A

quit

quit

Return to system view.

N/A

N/A

interface LoopBack0

interface LoopBack0

Create interface Loopback 0 and enter its view.

N/A

N/A

ip address 10.182.224.111 255.255.255.255

ip address 10.182.224.246 255.255.255.255

Assign an IP address to the interface.

VTEP IP address for establishing BGP EVPN peering.

 

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

interface LoopBack1

interface LoopBack1

Create interface Loopback 1 and enter its view.

N/A

N/A

ip address 10.182.226.111 255.255.255.255

ip address 10.182.226.111 255.255.255.255

Assign an IP address to the interface.

Virtual VTEP address for the DR system to establish VXLAN tunnels to remote devices.

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

vlan 1999

vlan 1999

Create the  VLAN for configuring the VLAN interface used for establishing L3 connectivity between the peer DR member devices.

N/A

N/A

interface Vlan-interface1999

interface Vlan-interface1999

Create VLAN-interface 1999 and enter its view.

Specify the IP addresses for establishing L3 connectivity between the peer DR member devices.

When the uplink on one DR member device fails, the uplink traffic that arrives on that member device can traverse the established L3 connectivity to the other DR member device and go outside.

ip address 192.168.220.1 255.255.255.252

ip address 192.168.220.2 255.255.255.252

Assign an IP address to the interface.

N/A

N/A

ospf network-type broadcast

ospf network-type broadcast

Set the OSPF network type of the interface to broadcast.

N/A

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring the links towards the spine tier

Leaf 1

Leaf 2

Description

Purpose

Remarks

interface Ten-GigabitEthernet1/0/17

interface Ten-GigabitEthernet2/0/17

Configure the interface connected to Spine 2.

N/A

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address unnumbered interface LoopBack0

ip address unnumbered interface LoopBack0

Configure the interface to borrow the IP address of Loopback 0.

N/A

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

N/A

ospf network-type p2p

ospf network-type p2p

Set the OSPF network type of the interface to P2P.

N/A

 

interface HundredGigE1/0/53

interface HundredGigE1/0/53

Configure the interface connected to Spine 1.

N/A

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address unnumbered interface LoopBack0

ip address unnumbered interface LoopBack0

Configure the interface to borrow the IP address of Loopback 0.

N/A

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable the OSPF on the interface.

N/A

N/A

ospf network-type p2p

ospf network-type p2p

Set the OSPF network type of the interface to P2P.

N/A

N/A

 

Configuring L2VPN

Leaf 1

Leaf 2

Description

Purpose

Remarks

l2vpn enable

l2vpn enable

Enable L2VPN.

N/A

N/A

vxlan tunnel mac-learning disable

vxlan tunnel mac-learning disable

Disable remote MAC address learning for VXLANs.

This setting avoids the conflict between automatically learned MAC address entries and MAC address entries advertised through BGP EVPN.

N/A

vxlan tunnel arp-learning disable

vxlan tunnel arp-learning disable

Disable remote ARP learning for VXLANs.

This setting avoids the conflict between automatically learned ARP entries and ARP entries advertised through BGP EVPN.

N/A

vxlan tunnel nd-learning disable

vxlan tunnel nd-learning disable

Disable remote ND learning for VXLANs.

This setting avoids the conflict between automatically learned ND entries and ND entries advertised through BGP EVPN.

N/A

mac-address timer aging 3600

mac-address timer aging 3600

Set the aging timer to 3600 seconds for dynamic MAC address entries.

If the DR system has a large number of MAC address entries, increase the MAC aging timer value to ensure complete synchronization of MAC address entries when one of the DR member devices restarts.

This setting must be consistent between the peer member devices in a DR system.

mac-address mac-move fast-update

mac-address mac-move fast-update

Enable ARP fast update for MAC address moves.

This setting helps accelerate VM migration across the network.

N/A

arp timer aging 1440

arp timer aging 1440

Set the aging timer to 1440 minutes for dynamic ARP entries.

If the DR system has a large number of ARP entries, increase the aging timer value for dynamic ARP entries to ensure complete synchronization of ARP entries when one of the DR member devices restarts.

This setting helps accelerate VM migration across the network.

This setting must be consistent between the peer member devices in a DR system.

IPv6 neighbor stale-aging 1440

IPv6 neighbor stale-aging 1440

Set the aging timer to 1440 minutes for ND entries in stale state.

If the DR system has a large number of ND entries, increase the aging timer value for ND entries in stale state to ensure complete synchronization of ND entries when one of the DR member devices restarts.

This setting helps accelerate VM migration across the network.

This setting must be consistent between the peer member devices in a DR system.

 

Configuring DRNI

Leaf 1

Leaf 2

Description

Purpose

Remarks

l2vpn drni peer-link ac-match-rule vxlan-mapping

l2vpn drni peer-link ac-match-rule vxlan-mapping

Enable the device to create frame match criteria based on VXLAN IDs for the dynamic ACs on the Ethernet aggregate link IPL.

N/A

N/A

evpn drni group 10.182.226.111

evpn drni group 10.182.226.111

Enable EVPN distributed relay and set the virtual VTEP address.

The DR member devices (VTEPs) use the virtual VTEP address to establish tunnels with the remote VTEPs.

You must specify the same virtual VTEP address on both VTEPs in the same DR system.

evpn drni local 10.182.224.111 remote 10.182.224.246

evpn drni local 10.182.224.246 remote 10.182.224.111

Specify the IP addresses of the local and peer VTEPs in the EVPN DR system.

You must execute this command if a DR system uses an Ethernet aggregate link as the IPL and has ACs (called single-armed ACs) attached to only one of the member devices. It enables the VTEPs in the DR system to set the next hop of the routes for single-armed ACs to their local VTEP IP addresses when they advertise the routes. This mechanism ensures that the traffic destined for a single-armed AC is forwarded towards its attached VTEP instead of the other VTEP.

The specified local and remote VTEP addresses must each belong to an interface on the local or peer VTEP in the DR system, respectively. Make sure the local VTEP address on one VTEP is the remote VTEP address on the other.

evpn global-mac 00e0-fc00-580a

evpn global-mac 00e0-fc00-580a

Configure an EVPN global MAC address.

N/A

You must specify the same EVPN global MAC address on the devices in the same DR system.

Do not use a reserved MAC address as the EVPN global MAC address.

drni system-mac 00e0-fc00-5800

drni system-mac 00e0-fc00-5800

Set the MAC address of the DR system.

Required.

You must assign the same DR system MAC address to the member devices in a DR system.

drni system-number 1

drni system-number 2

Set the DR system number.

Required.

You must assign different DR system numbers to the member devices in a DR system.

drni system-priority 100

drni system-priority 100

(Optional.) Set the DR system priority.

N/A

You must set the same DR system priority on the member devices in a DR system.

drni standalone enable

drni standalone enable

Enable DRNI standalone mode.

N/A

N/A

interface Ten-GigabitEthernet 1/0/47

interface Ten-GigabitEthernet 2/0/47

Enter the interface view for the keepalive link.

Required.

N/A

port link-mode route

port link-mode route

Configure the interface for keepalive detection to operate in route mode as a Layer 3 interface.

Required.

N/A

ip address 1.0.0.1 24

ip address 1.0.0.2 24

Assign an IP address to the interface as planned.

Required.

N/A

quit

quit

Return to system view.

N/A

N/A

drni keepalive ip destination 1.0.0.2 source 1.0.0.1

drni keepalive ip destination 1.0.0.1 source 1.0.0.2

Configure the source and destination IP addresses of keepalive packets.

Required.

For correct keepalive detection, you must exclude the interfaces that own the IP addresses used for keepalive detection from the shutdown action.

drni mad default-action none

drni mad default-action none

Set the DRNI MAD action to none. When the DR system splits, DRNI MAD will not shut down any network interfaces, except the interfaces configured manually or by the system to be shut down on the secondary device.

N/A

N/A

drni mad include interface HundredGigE1/0/53

drni mad include interface HundredGigE2/0/54

Configure DRNI MAD to shut down the interface upon a DR system split if the device is the secondary DR member device.

N/A

N/A

drni mad include interface Ten-GigabitEthernet1/0/17

drni mad include interface Ten-GigabitEthernet2/0/17

Configure DRNI MAD to shut down the interface upon a DR system split if the device is the secondary DR member device.

N/A

N/A

drni restore-delay 200

drni restore-delay 200

Set the data restoration interval.

This command specifies the maximum amount of time for the secondary DR member device to synchronize data with the primary DR member device during DR system setup. Within the data restoration interval, the secondary DR member device sets all network interfaces to DRNI MAD DOWN state except those excluded from the MAD shutdown action.

To avoid packet loss and forwarding failure, increase the data restoration interval if the amount of data is large, for example, when the device has a large number of routes and interfaces.

N/A

interface Bridge-Aggregation11

interface Bridge-Aggregation11

Create the Layer 2 aggregate interface to be used as the IPP and enter its interface view.

Configure the Layer 2 aggregation interfaces that act as the IPPs at the two ends of the IPL.

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the interface to permit all VLANs to pass through.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

interface HundredGigE 1/0/49

interface HundredGigE 1/0/49

Enter the interface view for the port to be used as a member port of the IPP.

Configure the member ports of the Layer 2 aggregation interfaces that act as the IPPs at the two ends of the IPL.

N/A

port link-type trunk

port link-type trunk

Set the link type of the port to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the port to permit all VLANs to pass through.

N/A

N/A

port link-aggregation group 11

port link-aggregation group 11

Assign the port to the link aggregation group for the IPP (aggregation group 11).

N/A

N/A

interface HundredGigE 1/0/50

interface HundredGigE 1/0/50

Enter the interface view for the port to be used as a member port of the IPP.

Configure the member ports of the Layer 2 aggregation interfaces that act as the IPPs at the two ends of the IPL.

N/A

port link-type trunk

port link-type trunk

Set the link type of the port to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the port to permit all VLANs to pass through.

N/A

N/A

port link-aggregation group 11

port link-aggregation group 11

Assign the port to the link aggregation group for the IPP (aggregation group 11).

N/A

N/A

interface Bridge-Aggregation11

interface Bridge-Aggregation11

Enter the interface view for the IPP (Bridge-Aggregation 11).

Specify the aggregate interface (Bridge-Aggregation 11) as the IPP.

N/A

link-aggregation mode dynamic

link-aggregation mode dynamic

Configure the aggregate interface to operate in dynamic mode.

N/A

N/A

port drni intra-portal-port 1

port drni intra-portal-port 1

Specify the aggregate interface as the IPP.

N/A

N/A

undo mac-address static source-check enable

undo mac-address static source-check enable

Disable the static source check feature on the interface.

This command ensures that the DR member devices can correctly forward the Layer 3 traffic received from each other over the IPL.

You must disable static source check on the IPPs of all leaf nodes and their uplink ports connected to the spine tier.

quit

quit

N/A

N/A

N/A

 

 

NOTE:

If a DR system uses an Ethernet aggregate link as the IPL, each DR member device creates a dynamic AC on the IPL when an AC (Ethernet service instance) is configured on a site-facing interface. The dynamic AC and the site-facing AC have the same frame match criteria and VSI mapping. If two site-facing ACs on different interfaces have the same frame match criteria but different VSI mappings, the dynamic ACs created for the site-facing ACs will conflict with each other. To prevent this issue, use the l2vpn drni peer-link ac-match-rule vxlan-mapping command to enable the DR member devices to create frame match criteria based on VXLAN IDs for the dynamic ACs on the IPL.

 

Configuring the links towards the bare metal servers

Leaf 1

Leaf 2

Description

Purpose

Remarks

interface Bridge-Aggregation1

interface Bridge-Aggregation1

Create an aggregate interface to be used as a DR interface.

Configure the DR interfaces connected to the bare metal servers.

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

link-aggregation mode dynamic

link-aggregation mode dynamic

Configure the aggregate interface to operate in dynamic mode.

N/A

Configure the interface to permit all VLANs to pass through.

port drni group 1

port drni group 1

Assign the aggregate interface to a DR group.

N/A

N/A

interface Ten-GigabitEthernet1/0/7

interface Ten-GigabitEthernet 2/0/7

Enter the view of a member physical interface of the DR interface.

N/A

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

port link-aggregation group 1

port link-aggregation group 1

Add the physical interface to the link aggregation group for the DR interface.

N/A

N/A

quit

quit

N/A

N/A

N/A

 

Configure the spanning tree feature

Leaf 1

Leaf 2

Description

Purpose

stp global enable

stp global enable

Enable the spanning tree feature globally.

N/A

interface Bridge-Aggregation 1

interface Bridge-Aggregation 1

Enter the view of the DR interface connected to the bare metal servers.

N/A

stp edged-port

stp edged-port

N/A

Configure the DR interface as an edge port to exclude the port from spanning tree calculation for rapid state transition.

 

IMPORTANT

IMPORTANT:

·     Make sure the DR member devices are consistent in global, IPP-specific, and DR-interface-specific spanning tree settings. Inconsistent spanning tree settings might cause network flapping.

·     IPPs in the DR system do not participate in spanning tree calculation.

·     After the DR system splits, the DR member devices still use the DR system MAC address to send BPDUs, resulting in incorrect spanning tree calculation. To avoid this issue, enable DRNI standalone mode on the DR member devices.

 

Configuring an underlay BGP instance

Leaf 1

Leaf 2

Description

Purpose

Remarks

bgp 65105

bgp 65105

Enable a BGP instance.

N/A

N/A

router-id 10.182.224.111

router-id 10.182.224.246

Specify a unique router ID for the BGP instance on each BGP device.

To run BGP, a BGP instance must have a router ID.

If you do not specify a router ID for the BGP instance on a device, it uses the global router ID. In this situation, you must make sure a global router ID is set on the device.

N/A

group spines internal

group spines internal

Create an IBGP peer group.

N/A

N/A

peer spines connect-interface LoopBack0

peer spines connect-interface LoopBack0

Specify a source interface for establishing TCP connections to the peer group.

N/A

N/A

peer spines route-update-interval 0

peer spines route-update-interval 0

Specify an interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence.

Execute this command only for IBGP peers.

peer 10.182.226.89 group spines

peer 10.182.226.89 group spines

Add the specified spine device to the peer group.

N/A

N/A

peer 10.182.226.90 group spines

peer 10.182.226.90 group spines

Add the specified spine device to the peer group.

N/A

N/A

address-family IPv4 unicast

address-family IPv4 unicast

Create the BGP IPv4 unicast address family and enter its view.

N/A

N/A

balance 8

balance 8

Enable load balancing and set the maximum number of BGP ECMP routes for load balancing.

 

 

import-route direct

import-route direct

Configure BGP to redistribute direct routes in the BGP instance.

N/A

N/A

peer spines enable

peer spines enable

Enable the device to exchange routes with the peer group.

N/A

N/A

quit

quit

Exit the BGP IPv4 unicast address family view.

N/A

N/A

 

Configuring an EVPN BGP instance

Leaf 1

Leaf 2

Description

Purpose

Remarks

bgp 65001 instance EVPN

bgp 65001 instance EVPN

Enable a BGP instance.

N/A

N/A

router-id 10.182.224.111

router-id 10.182.224.246

Specify a unique router ID for the BGP instance on each BGP device.

To run BGP, a BGP instance must have a router ID.

If you do not specify a router ID for the BGP instance on a device, it uses the global router ID. In this situation, you must make sure a global router ID is set on the device.

N/A

group evpn internal

group evpn internal

Create an IBGP peer group.

N/A

N/A

peer evpn connect-interface LoopBack0

peer evpn connect-interface LoopBack0

Specify a source interface for establishing TCP connections to the peer group.

N/A

N/A

peer evpn route-update-interval 0

peer evpn route-update-interval 0

Specify an interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

Execute this command only for IBGP peers.

peer 10.182.226.89 group evpn

peer 10.182.226.89 group evpn

Add the specified spine device to the peer group.

N/A

N/A

peer 10.182.226.90 group evpn

peer 10.182.226.90 group evpn

Add the specified spine device to the peer group.

N/A

N/A

address-family l2vpn evpn

address-family l2vpn evpn

Create the BGP EVPN address family and enter its view.

N/A

N/A

peer evpn enable

peer evpn enable

Enable the device to exchange routes with the peer group.

N/A

N/A

undo policy vpn-target

undo policy vpn-target

Disable filtering incoming BGP EVPN routes based on route targets.

 

 

quit

quit

Exit the BGP EVPN address family view.

N/A

N/A

 

Configuring VSIs and ACs

Leaf 1

Leaf 2

Description

Remarks

vsi vpn-trusted-7

vsi vpn-trusted-7

Create a VSI and enter its view.

N/A

arp suppression enable

arp suppression enable

Enable ARP flood suppression.

N/A

vxlan 7

vxlan 7

Create a VXLAN and enter its view.

N/A

evpn encapsulation vxlan

evpn encapsulation vxlan

Create a VXLAN EVPN instance on the VSI and enter its view.

N/A

route-distinguisher auto

route-distinguisher auto

Configure the device to automatically generate an RD for the EVPN instance.

N/A

vpn-target auto export-extcommunity

vpn-target auto export-extcommunity

Configure the device to automatically generate an export RT for the EVPN instance.

N/A

vpn-target auto import-extcommunity

vpn-target auto import-extcommunity

Configure the device to automatically generate an import RT for the EVPN instance.

N/A

quit

quit

Return to system view.

N/A

vsi vpn-trusted-522

vsi vpn-trusted-522

Create a VSI and enter its view.

N/A

arp suppression enable

arp suppression enable

Enable ARP flood suppression.

N/A

vxlan 522

vxlan 522

Create a VXLAN and enter its view.

N/A

evpn encapsulation vxlan

evpn encapsulation vxlan

Create a VXLAN EVPN instance on the VSI and enter its view.

N/A

route-distinguisher auto

route-distinguisher auto

Configure the device to automatically generate an RD for the EVPN instance.

N/A

vpn-target auto export-extcommunity

vpn-target auto export-extcommunity

Configure the device to automatically generate an export RT for the EVPN instance.

N/A

vpn-target auto import-extcommunity

vpn-target auto import-extcommunity

Configure the device to automatically generate an import RT for the EVPN instance.

N/A

quit

quit

Return to system view.

N/A

interface Bridge-Aggregation1

interface Bridge-Aggregation1

Enter the view of the site-facing interface on which Ethernet instances will be created.

N/A

undo port trunk permit vlan 1

undo port trunk permit vlan 1

Remove the interface from VLAN 1.

N/A

port trunk permit vlan 7 522

port trunk permit vlan 7 522

Assign the interface to VLANs 7 and 522.

N/A

service-instance 7

service-instance 7

Create an Ethernet service instance and enter its view.

N/A

encapsulation s-vid 7

encapsulation s-vid 7

Configure the Ethernet service instance to match traffic sent from the specified VLAN.

N/A

xconnect vsi vpn-trusted-7

xconnect vsi vpn-trusted-7

Map the Ethernet service instance to the specified VSI.

N/A

service-instance 522

service-instance 522

Create an Ethernet service instance and enter its view.

N/A

encapsulation s-vid 522

encapsulation s-vid 522

Configure the Ethernet service instance to match traffic sent from the specified VLAN.

N/A

xconnect vsi vpn-trusted-522

xconnect vsi vpn-trusted-522

Map the Ethernet service instance to the specified VSI.

N/A

quit

quit

N/A

N/A

 

Configuring S6850 switches as border nodes

Procedure summary

·     Configure basic settings

·     Configuring OSPF

·     Configure the spanning tree feature

·     Configuring the interfaces connected to the spine nodes

·     Configuring L2VPN

·     Configuring DRNI

·     Configuring the routed interfaces connected to the external network

·     Configuring an underlay BGP instance

·     Configuring an EVPN BGP instance

·     Configuring the overlay network

Configure basic settings

Border 1

Border 2

Description

Purpose

Remarks

interface LoopBack0

interface LoopBack0

Create interface Loopback 0 and enter its view.

N/A

N/A

ip address 10.182.234.1 255.255.255.255

ip address 10.182.234.2 255.255.255.255

Assign an IP address to the interface.

VTEP IP

N/A

quit

quit

N/A

N/A

N/A

interface LoopBack1

interface LoopBack1

Create interface Loopback 1 and enter its view.

N/A

N/A

ip address 10.182.236.1 255.255.255.255

ip address 10.182.236.1 255.255.255.255

Assign an IP address to the interface.

Configure the virtual VTEP IP address.

N/A

quit

quit

N/A

N/A

 

ip vpn-instance vpn-trusted

ip vpn-instance vpn-trusted

Create a VPN instance.

N/A

N/A

route-distinguisher 65001:10

route-distinguisher 65002:10

Configure the device to automatically generate an RD for the VPN instance.

N/A

N/A

address-family IPv4

address-family IPv4

Enter VPN instance IPv4 address family view.

N/A

N/A

vpn-target 65001:1 import-extcommunity

vpn-target 65001:1 import-extcommunity

Configure the IPv4 import target for the VPN instance.

N/A

N/A

vpn-target 65001:1 export-extcommunity

vpn-target 65001:1 export-extcommunity

Configure the IPv4 export target for the VPN instance.

N/A

N/A

address-family IPv6

address-family IPv6

Enter VPN instance IPv6 address family view.

N/A

N/A

vpn-target 65001:1 import-extcommunity

vpn-target 65001:1 import-extcommunity

Configure the IPv6 import target for the VPN instance.

N/A

N/A

vpn-target 65001:1 export-extcommunity

vpn-target 65001:1 export-extcommunity

Configure the IPv6 export target for the VPN instance.

N/A

N/A

address-family evpn

address-family evpn

Enter VPN instance EVPN address family view.

N/A

N/A

vpn-target 65001:1 import-extcommunity

vpn-target 65001:1 import-extcommunity

Configure the IPv6 import target for the VPN instance.

N/A

N/A

vpn-target 65001:1 export-extcommunity

vpn-target 65001:1 export-extcommunity

Configure the IPv6 export target for the VPN instance.

N/A

N/A

quit

quit

 

N/A

N/A

 

Configuring OSPF

Border 1

Border 2

Description

Purpose

Remarks

ospf 100 router-id 10.182.234.1 vpn-instance vpn-trusted

ospf 100 router-id 10.182.234.1 vpn-instance vpn-trusted

Create an OSPF process and enter its view.

N/A

N/A

import-route direct

import-route direct

Redistribute direct routes in the OSPF process.

N/A

N/A

spf-schedule-interval 1 10 10

spf-schedule-interval 1 10 10

Set the maximum SPF calculation interval to 1 second, the minimum SPF calculation interval to 10 milliseconds, and the incremental SPF calculation interval to 10 milliseconds for OSPF.

Shorten the SPF calculation interval to accelerate route convergence.

N/A

lsa-generation-interval 1 10 10

lsa-generation-interval 1 10 10

Set the maximum interval for LSA generation to 1 second, the minimum interval to 10 milliseconds, and the incremental interval to 10 milliseconds.

Enable quicker LSA regeneration upon network topology change to accelerate route convergence.

N/A

area 0.0.0.1

area 0.0.0.1

Create OSPF area 1.

N/A

N/A

fast-reroute lfa

fast-reroute lfa

Enable OSPF FRR and use the LFA algorithm for calculation of the backup next hop.

This feature minimizes service interruption by fast rerouting traffic to the precalculated backup next hop when a link or node fails.

N/A

quit

quit

N/A

N/A

N/A

vlan 1001

vlan 1001

Create the VLAN for configuring the VLAN interface used for establishing L3 connectivity between the border nodes.

N/A

N/A

interface Vlan-interface1001

interface Vlan-interface1001

Create VLAN-interface 1001 and enter its view.

Specify the IP addresses for establishing L3 connectivity between the peer DR member devices. When the uplink on one DR member device fails, the uplink traffic that arrives on that member device can traverse the established L3 connectivity to the other DR member device and go outside.

N/A

ip binding vpn-instance vpn-trusted

ip binding vpn-instance vpn-trusted

N/A

N/A

N/A

ip address 192.101.1.101 255.255.255.252

ip address 192.101.1.100 255.255.255.252

Assign an IP address to the interface.

N/A

N/A

ospf network-type broadcast

ospf network-type broadcast

Set the OSPF network type of the interface to broadcast.

N/A

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

 

quit

quit

Return to system view.

N/A

N/A

 

Configure the spanning tree feature

Border 1

(S12508X-AF)

Border 2

(S12508X-AF)

Description

Purpose

Remarks

stp global enable

stp global enable

Enable the spanning tree feature globally.

N/A

N/A

 

IMPORTANT

IMPORTANT:

·     Make sure the DR member devices are consistent in global, IPP-specific, and DR-interface-specific spanning tree settings. Inconsistent spanning tree settings might cause network flapping.

·     IPPs in the DR system do not participate in spanning tree calculation.

·     After the DR system splits, the DR member devices still use the DR system MAC address to send BPDUs, resulting in incorrect spanning tree calculation. To avoid this issue, enable DRNI standalone mode on the DR member devices.

 

Configuring the interfaces connected to the spine nodes

Border 1

Border 2

Description

Purpose

Remarks

interface Twenty-FiveGigE1/0/53

interface Twenty-FiveGigE1/0/53

Configure the interface connected to Spine 1.

N/A

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address 10.182.221.1 255.255.255.254

ip address 10.182.221.11 255.255.255.254

Assign an IP address to the interface.

N/A

N/A

link-delay up 60

link-delay up 60

Configure the interface to suppress link-up events for 60 seconds.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

interface Twenty-FiveGigE1/0/55

interface Twenty-FiveGigE1/0/55

Configure the interface connected to Spine 2.

N/A

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address 10.182.221.5 255.255.255.254

ip address 10.182.221.15 255.255.255.254

Assign an IP address to the interface.

N/A

N/A

link-delay up 60

link-delay up 60

Configure the interface to suppress link-up events for 60 seconds.

N/A

N/A

quit

quit

N/A

N/A

N/A

 

Configuring L2VPN

Border 1

Border 2

Description

Purpose

Remarks

l2vpn enable

l2vpn enable

Enable L2VPN.

N/A

N/A

l2vpn drni peer-link ac-match-rule vxlan-mapping

l2vpn drni peer-link ac-match-rule vxlan-mapping

Enable the device to create frame match criteria based on VXLAN IDs for the dynamic ACs on the Ethernet aggregate link IPL.

N/A

N/A

vxlan tunnel arp-learning disable

vxlan tunnel arp-learning disable

Disable remote ARP learning for VXLANs.

N/A

N/A

vxlan tunnel nd-learning disable

vxlan tunnel nd-learning disable

Disable remote ND learning for VXLANs.

N/A

N/A

vxlan tunnel mac-learning disable

vxlan tunnel mac-learning disable

Disable remote MAC address learning for VXLANs.

N/A

N/A

mac-address timer aging 900

mac-address timer aging 900

Set the aging timer for dynamic MAC address entries to 900 seconds .

If the DR system has a large number of MAC entries, increase the aging timer value for dynamic MAC address entries to ensure complete synchronization of MAC address entries when one of the DR member devices restarts.

This setting helps accelerate VM migration across the network.

This setting must be consistent between the peer member devices in a DR system.

IPv6 neighbor stale-aging 1440

IPv6 neighbor stale-aging 1440

Set the aging timer for ND entries in stale state.

If the DR system has a large number of ND entries, increase the aging timer value for ND entries in stale state to ensure complete synchronization of ND entries when one of the DR member devices restarts.

This setting helps accelerate VM migration across the network.

This setting must be consistent between the peer member devices in a DR system.

arp timer aging 1440

arp timer aging 1440

Set the aging timer for dynamic ARP entries to 1440 minutes.

If the DR system has a large number of ARP entries, increase the aging timer value for dynamic ARP entries to ensure complete synchronization of ARP entries when one of the DR member devices restarts.

This setting helps accelerate VM migration across the network.

This setting must be consistent between the peer member devices in a DR system.

 

IMPORTANT

IMPORTANT:

If a DR system in a VXLAN or EVPN environment uses an Ethernet aggregate link as the IPL, each DR member device creates a dynamic AC on the IPL when an AC (Ethernet service instance) is configured on a site-facing interface. The dynamic AC and the site-facing AC have the same frame match criteria and VSI mapping. If two site-facing ACs on different interfaces have the same frame match criteria but different VSI mappings, the dynamic ACs created for the site-facing ACs will conflict with each other. To prevent this issue, use the l2vpn drni peer-link ac-match-rule vxlan-mapping command to enable the DR member devices to create frame match criteria based on VXLAN IDs for the dynamic ACs on the IPL.

 

Configuring DRNI

Border 1

Border 2

Description

Purpose

Remarks

drni restore-delay 180

drni restore-delay 180

Set the data restoration interval. This parameter specifies the maximum amount of time for the secondary DR member device to synchronize data with the primary DR member device during DR system setup.

This command specifies the maximum amount of time for the secondary DR member device to synchronize data with the primary DR member device during DR system setup. Within the data restoration interval, the secondary DR member device sets all network interfaces to DRNI MAD DOWN state except those excluded from the MAD shutdown action.

To avoid packet loss and forwarding failure, increase the data restoration interval if the amount of data is large, for example, when the device has a large number of routes and interfaces.

N/A

drni role priority 10

drni role priority 20

Set the DR role priority of the device.

N/A

A DR member device is assigned the primary or secondary role based on its DR role priority. The lower the priority value, the higher the priority.

drni system-mac 0002-0002-0001

drni system-mac 0002-0002-0001

Set the MAC address of the DR system.

N/A

You must assign the same DR system MAC address to the member devices in a DR system.

drni system-number 1

drni system-number 2

Set the DR system number.

N/A

You must assign different DR system numbers to the member devices in a DR system.

drni system-priority 10

drni system-priority 10

Set the DR system priority.

N/A

You must set the same DR system priority on the member devices in a DR system.

drni standalone enable

drni standalone enable

Enable DRNI standalone mode.

N/A

N/A

drni mad default-action none

drni mad default-action none

Set the DRNI MAD action to none. When the DR system splits, DRNI MAD will not shut down any network interfaces, except the interfaces configured manually or by the system to be shut down on the secondary device.

Required.

N/A

drni mad include interface Twenty-FiveGigE1/0/33

drni mad include interface Twenty-FiveGigE1/0/33

Configure DRNI MAD to shut down the uplink interface upon a DR system split if the device is the secondary DR member device.

N/A

N/A

drni mad include interface Twenty-FiveGigE1/0/53

drni mad include interface Twenty-FiveGigE1/0/53

Configure DRNI MAD to shut down the downlink interface upon a DR system split if the device is the secondary DR member device.

N/A

N/A

drni mad include interface Twenty-FiveGigE1/0/55

drni mad include interface Twenty-FiveGigE1/0/55

Configure DRNI MAD to shut down the downlink interface when the DR system splits.

N/A

N/A

drni keepalive ip destination 2.0.0.2 source 2.0.0.1

drni keepalive ip destination 2.0.0.1 source 2.0.0.2

Configure the source and destination IP addresses of keepalive packets.

This example uses the IP addresses of management Ethernet interfaces to set up the keepalive link. By default, management Ethernet interfaces are excluded from the DRNI MAD DOWN action.

If the interfaces belong to a VPN instance, specify that VPN instance when you specify source and destination IP addresses for keepalive packets.

If the interfaces that own the IP addresses are not excluded from the DRNI MAD DOWN action by default, manually exclude them from that action.

evpn drni group 10.182.236.1

evpn drni group 10.182.236.1

Enable EVPN distributed relay and specify the virtual VTEP address.

N/A

N/A

evpn global-mac 00e0-fc00-0201

evpn global-mac 00e0-fc00-0201

Configure an EVPN global MAC address.

N/A

You must specify the same EVPN global MAC address on the devices in the same DR system.

Do not use a reserved MAC address as the EVPN global MAC address.

interface Bridge-Aggregation1024

interface Bridge-Aggregation1024

Create the Layer 2 aggregate interface to be used as the IPP, and enter interface view. This example uses interface Bridge-Aggregation 1024 as the IPP on each member device.

N/A

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the interface to permit all VLANs to pass through.

N/A

N/A

link-aggregation mode dynamic

link-aggregation mode dynamic

Configure the aggregate interface to operate in dynamic mode.

N/A

N/A

interface HundredGigE1/0/25

interface HundredGigE1/0/25

Assign the physical port to the aggregation group for the IPP.

N/A

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the interface to permit all VLANs to pass through.

N/A

N/A

port link-aggregation group 1024

port link-aggregation group 1024

Assign the port to the link aggregation group for the IPP.

N/A

N/A

interface HundredGigE1/0/26

interface HundredGigE1/0/26

Assign the physical port to the aggregation group for the IPP.

N/A

N/A

port link-type trunk

port link-type trunk

Set the link type of the interface to trunk.

N/A

N/A

port trunk permit vlan all

port trunk permit vlan all

Configure the interface to permit all VLANs to pass through.

N/A

N/A

port link-aggregation group 1024

port link-aggregation group 1024

Assign the port to the link aggregation group for the IPP.

N/A

N/A

quit

quit

N/A

N/A

N/A

interface Bridge-Aggregation1024

interface Bridge-Aggregation1024

Enter the Layer 2 aggregate interface view for the IPP.

N/A

N/A

port drni intra-portal-port 1

port drni intra-portal-port 1

Specify the aggregate interface as the IPP.

N/A

N/A

quit

quit

N/A

N/A

N/A

 

IMPORTANT

IMPORTANT:

If you use a VXLAN tunnel as the IPL in an EVPN environment, you must retain a large number of logical interfaces (for example, tunnel and loopback interfaces) in up state. To simplify configuration, set the default DRNI MAD action to NONE and execute the drni mad include interface command to specify interfaces that must be shut down by DRNI MAD if they are not in the system-defined included port list for MAD shutdown.

 

Configuring the routed interfaces connected to the external network

Border 1

Border 2

Description

Purpose

interface Twenty-FiveGigE1/0/33

interface Twenty-FiveGigE1/0/33

N/A

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

ip binding vpn-instance vpn-trusted

ip binding vpn-instance vpn-trusted

Associate the interface with the specified VPN instance.

N/A

ip address 192.101.1.1 255.255.255.254

ip address 192.101.1.3 255.255.255.254

Assign an IP address to the interface.

N/A

ospf 100 area 0.0.0.1

ospf 100 area 0.0.0.1

Enable OSPF on the interface.

N/A

ospf peer hold-max-cost duration 300000

ospf peer hold-max-cost duration 300000

On the interface, enable OSPF to advertise the maximum link cost to neighbors within 300000 milliseconds.

N/A

 

Configuring an underlay BGP instance

Border 1

Border 2

Description

Purpose

Remarks

bgp 65031

bgp 65031

Enable a BGP instance.

N/A

N/A

router-id 10.182.234.1

router-id 10.182.234.2

Specify a unique router ID for the BGP instance on each BGP device.

To run BGP, a BGP instance must have a router ID.

If you do not specify a router ID for the BGP instance on a device, it uses the global router ID. In this situation, you must make sure a global router ID is set on the device.

N/A

group spine external

group spine external

Create an EBGP peer group.

N/A

N/A

peer spine as-number 65105

peer spine as-number 65105

Specify an AS number for the EBGP peer group.

N/A

N/A

peer spine route-update-interval 0

peer spine route-update-interval 0

Specify an interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

N/A

peer 10.182.221.0 group spines

peer 10.182.221.10 group spines

Add the specified spine device to the peer group.

N/A

N/A

peer 10.182.221.4 group spines

peer 10.182.221.14 group spines

Add the specified spine device to the peer group.

N/A

N/A

peer 100.5.1.2 as-number 65031

peer 100.5.1.1 as-number 65031

Configure EBGP peering for establishing Layer 3 connectivity between the DR member devices.

N/A

 

peer 100.5.1.2 route-update-interval 0

peer 100.5.1.1 route-update-interval 0

Specify an interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

N/A

address-family IPv4 unicast

address-family IPv4 unicast

Create the BGP IPv4 unicast address family and enter its view.

N/A

N/A

balance 8

balance 8

Enable load balancing and set the maximum number of BGP ECMP routes for load balancing.

N/A

N/A

import-route direct

import-route direct

Configure BGP to redistribute direct routes in the BGP instance.

N/A

N/A

network 10.182.234.1 255.255.255.255

network 10.182.234.2 255.255.255.255

Advertise the IP address of interface Loopback 0.

N/A

N/A

network 10.182.236.1 255.255.255.255

network 10.182.236.1 255.255.255.255

Advertise the IP address of interface Loopback 1.

N/A

N/A

peer spine enable

peer spine enable

Enable the device to exchange routes with the peer group.

N/A

N/A

peer 100.5.1.2 enable

peer 100.5.1.1 enable

Enable the device to exchange routes with the peer group.

N/A

N/A

quit

quit

Exit the BGP IPv4 unicast address family view.

N/A

N/A

 

Configuring an EVPN BGP instance

Border 1

Border 2

Description

Purpose

Remarks

bgp 65001 instance EVPN

bgp 65001 instance EVPN

Enable the specified BGP instance and enter its view.

N/A

N/A

router-id 10.182.234.1

router-id 10.182.234.2

Specify a unique router ID for the BGP instance on the device.

N/A

N/A

group evpn internal

group evpn internal

Create an IBGP peer group.

N/A

The IBGP peer group name must be evpn.

peer evpn connect-interface LoopBack0

peer evpn connect-interface LoopBack0

Specify a source interface for establishing TCP connections to the peer group.

N/A

N/A

peer evpn route-update-interval 0

peer evpn route-update-interval 0

Specify an interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

Execute this command only for IBGP peers.

peer 10.182.226.89 group evpn

peer 10.182.226.89 group evpn

N/A

N/A

N/A

peer 10.182.226.90 group evpn

peer 10.182.226.90 group evpn

N/A

N/A

N/A

address-family l2vpn evpn

address-family l2vpn evpn

Create the BGP EVPN address family and enter its view.

N/A

N/A

peer evpn enable

peer evpn enable

Enable the device to exchange routes with the peer group.

N/A

N/A

nexthop evpn-drni group-address

nexthop evpn-drni group-address

Enable the DR member devices to replace the next hop in advertised BGP EVPN routes with the virtual VTEP address.

This step is required on the EDs in an EVPN DCI environment if the EDs are DR systems.

N/A

quit

quit

Exit the BGP IPv4 unicast address family view.

N/A

N/A

 

Configuring the overlay network

Border 1

Border 2

Description

Purpose

Remarks

interface Vsi-interface7

interface Vsi-interface7

Create a VSI interface and enter its view.

N/A

N/A

ip binding vpn-instance vpn-trusted

ip binding vpn-instance vpn-trusted

Associate the VSI interface with the specified VPN instance.

N/A

N/A

ip address 10.182.7.254 255.255.252.0

ip address 10.182.7.254 255.255.252.0

Assign an IP address to the interface. This IP address will be the gateway IP address.

N/A

N/A

proxy-arp enable

proxy-arp enable

Enable common proxy ARP.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

interface Vsi-interface522

interface Vsi-interface522

Create a VSI interface and enter its view.

N/A

N/A

ip binding vpn-instance vpn-trusted

ip binding vpn-instance vpn-trusted

Associate the VSI interface with the specified VPN instance.

N/A

N/A

ip address 10.180.43.254 255.255.252.0

ip address 10.180.43.254 255.255.252.0

Assign an IP address to the interface. This IP address will be the gateway IP address.

N/A

N/A

proxy-arp enable

proxy-arp enable

Enable common proxy ARP.

N/A

 N/A

quit

quit

Return to system view.

N/A

N/A

interface Vsi-interface10001

interface Vsi-interface10001

Create a VSI interface and enter its view.

N/A

N/A

ip binding vpn-instance vpn-trusted

ip binding vpn-instance vpn-trusted

Bind the VSI interface to the specified VPN instance.

N/A

N/A

IPv6 address auto link-local

IPv6 address auto link-local

Configure the system to automatically generate a link-local address for the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

vsi vpn-trusted-7

vsi vpn-trusted-7

Create a VSI and enter its view.

N/A

N/A

gateway vsi-interface 7

gateway vsi-interface 7

Specify the gateway interface for the VSI.

N/A

N/A

arp suppression enable

arp suppression enable

Enable ARP flood suppression.

N/A

N/A

vxlan 7

vxlan 7

Create a VXLAN and enter its view.

N/A

N/A

evpn encapsulation vxlan

evpn encapsulation vxlan

Create a VXLAN EVPN instance on the VSI and enter its view.

N/A

N/A

route-distinguisher auto

route-distinguisher auto

Configure the device to automatically generate an RD for the EVPN instance.

N/A

N/A

vpn-target auto export-extcommunity

vpn-target auto export-extcommunity

Configure the IPv6 export target for the EVPN instance.

N/A

N/A

vpn-target auto import-extcommunity

vpn-target auto import-extcommunity

Configure the IPv6 import target for the EVPN instance.

N/A

N/A

quit

quit

Return to VSI view.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

vsi vpn-trusted-522

vsi vpn-trusted-522

Create a VSI and enter its view.

N/A

N/A

gateway vsi-interface 522

gateway vsi-interface 522

Specify the gateway interface for the VSI.

N/A

N/A

arp suppression enable

arp suppression enable

Enable ARP flood suppression.

N/A

N/A

vxlan 522

vxlan 522

Create a VXLAN and enter its view.

N/A

N/A

evpn encapsulation vxlan

evpn encapsulation vxlan

Create a VXLAN EVPN instance on the VSI and enter its view.

N/A

N/A

route-distinguisher auto

route-distinguisher auto

Configure the device to automatically generate an RD for the EVPN instance.

N/A

N/A

vpn-target auto export-extcommunity

vpn-target auto export-extcommunity

Configure the IPv6 export target for the EVPN instance.

N/A

N/A

vpn-target auto import-extcommunity

vpn-target auto import-extcommunity

Configure the IPv6 import target for the EVPN instance.

N/A

N/A

quit

quit

Return to VSI view.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring S12500X-AF switches as spine nodes

Procedure summary

·     Configuring OSPF

·     Configuring the downlinks towards the leaf tier

·     Configuring the uplinks towards the border nodes

·     Configuring an underlay BGP instance

·     Configuring an EVPN BGP instance

Configuring OSPF

Spine 1

Spine 2

Description

Purpose

Remarks

ospf 1 router-id 10.182.224.90

ospf 1 router-id 10.182.224.89

Enable an OSPF process and enter its view.

N/A

N/A

spf-schedule-interval 1 10 10

spf-schedule-interval 1 10 10

Set the maximum SPF calculation interval to 1 second, the minimum SPF calculation interval to 10 milliseconds, and the incremental SPF calculation interval to 10 milliseconds for OSPF.

Shorten the SPF calculation interval to accelerate route convergence.

N/A

lsa-generation-interval 1 10 10

lsa-generation-interval 1 10 10

Set the maximum interval for LSP generation to 1 second, the minimum interval to 10 milliseconds, and the incremental interval to 10 milliseconds.

Enable quicker LSP regeneration upon network topology change to accelerate route convergence.

N/A

import-route direct

import-route direct

Configure BGP to redistribute direct routes in the BGP instance.

N/A

N/A

area 0.0.0.0

area 0.0.0.0

Create an OSPF area.

N/A

N/A

quit

quit

Return to OSPF view.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring the downlinks towards the leaf tier

The following information uses one port for example to describe the configuration procedure.

 

Spine 1

Spine 2

Description

Purpose

Remarks

interface HundredGigE1/0/25

interface HundredGigE1/0/25

Enter interface view.

Configure the interface connected to Leaf 1.

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address unnumbered interface LoopBack0

ip address unnumbered interface LoopBack0

Configure the interface to borrow the IP address of Loopback 0.

N/A

N/A

ospf network-type p2p

ospf network-type p2p

Set the OSPF network type of the interface to P2P.

N/A

N/A

ospf 1 area 0.0.0.0

ospf 1 area 0.0.0.0

Enable OSPF on the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring the uplinks towards the border nodes

The following information uses one port for example to describe the configuration procedure. All uplink ports connected to the spine nodes use the same configuration except that their IP addresses must be different.

 

Spine 1

Spine 2

Description

Purpose

Remarks

interface Ten-GigabitEthernet 2/0/1

interface Ten-GigabitEthernet 1/0/21

Enter interface view.

Configure the interface connected to Border 1.

N/A

port link-mode route

port link-mode route

Configure the interface to operate in route mode as a Layer 3 interface.

N/A

N/A

ip address 10.182.221.0 255.255.255.254

ip address 10.182.221.4 255.255.255.254

Assign an IP address to the interface.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring an underlay BGP instance

Border 1

Border 2

Description

Purpose

Remarks

bgp 65105

bgp 65105

Enable a BGP instance and enter its view.

N/A

N/A

router-id 10.182.224.90

router-id 10.182.224.89

Specify a unique router ID for the BGP instance on the device.

N/A

N/A

compare-different-as-med

compare-different-as-med

Enable MED comparison for routes from peers in different ASs.

N/A

N/A

group gw external

group gw external

Create an EBGP peer group for establishing EBGP peering with the border nodes.

N/A

N/A

peer gw as-number 65031

peer gw as-number 65031

Specify an AS number for the EBGP peer group.

N/A

N/A

peer gw route-update-interval 0

peer gw route-update-interval 0

Specify the interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

N/A

group leafs internal

group leafs internal

Create an IBGP peer group for establishing IBGP peering with the leaf nodes.

N/A

N/A

peer leafs connect-interface LoopBack0

peer leafs connect-interface LoopBack0

Specify a source interface for establishing TCP connections to the peer group.

N/A

N/A

peer leafs route-update-interval 0

peer leafs route-update-interval 0

Specify the interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

N/A

peer 10.182.221.1 group gw

peer 10.182.221.5 group gw

Add the specified peer to the peer group.

Establish BGP peering with Border 1.

N/A

peer 10.182.221.11 group gw

peer 10.182.221.15 group gw

Add the specified peer to the peer group.

Establish BGP peering with Border 2.

N/A

peer 10.182.224.111 group leafs

peer 10.182.224.111 group leafs

Add the specified peer to the peer group.

Establish BGP peering with Leaf 1.

N/A

peer 10.182.224.121 group leafs

peer 10.182.224.121 group leafs

Add the specified peer to the peer group.

Establish BGP peering with Leaf 3.

N/A

peer 10.182.224.122 group leafs

peer 10.182.224.122 group leafs

Add the specified peer to the peer group.

Establish BGP peering with Leaf 4.

N/A

peer 10.182.224.246 group leafs

peer 10.182.224.246 group leafs

Add the specified peer to the peer group.

Establish BGP peering with Leaf 2.

N/A

address-family IPv4 unicast

address-family IPv4 unicast

Create the BGP IPv4 unicast address family and enter its view.

N/A

N/A

balance 8

balance 8

Enable load balancing and set the maximum number of BGP ECMP routes for load balancing.

N/A

N/A

balance as-path-neglect

balance as-path-neglect

Enable BGP to ignore the AS_PATH attribute when it performs load balancing.

This enables load balancing across routes with different AS_PATH attribute values.

N/A

import-route direct

import-route direct

Configure BGP to redistribute direct routes in the BGP instance.

N/A

N/A

peer gw enable

peer gw enable

Enable the device to exchange routes with the peer group.

N/A

N/A

peer leafs enable

peer leafs enable

Enable the device to exchange routes with the peer group.

N/A

N/A

peer leafs reflect-client

peer leafs reflect-client

Configure the spine nodes as route reflectors and configure the specified peer group as a client.

N/A

N/A

quit

quit

Return to BGP instance view.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Configuring an EVPN BGP instance

Spine 1

Spine 2

Description

Purpose

Remarks

bgp 65001 instance EVPN

bgp 65001 instance EVPN

Enable a BGP instance and enter its view.

N/A

N/A

router-id 10.182.226.90

router-id 10.182.226.89

Specify a unique router ID for the BGP instance on the device.

N/A

N/A

group evpn internal

group evpn internal

Create an IBGP peer group.

N/A

N/A

peer evpn connect-interface LoopBack1

peer evpn connect-interface LoopBack1

Specify a source interface for establishing TCP connections to the peer group.

N/A

N/A

peer evpn route-update-interval 0

peer evpn route-update-interval 0

Specify the interval for sending the same update to the peer group.

Enable the device to fast send updates to the peer group upon route changes to accelerate route convergence after a DR primary/secondary switchover occurs.

Execute this command only for IBGP peers.

peer 10.182.224.111 group evpn

peer 10.182.224.111 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Leaf 1.

N/A

peer 10.182.224.121 group evpn

peer 10.182.224.121 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Leaf 3.

N/A

peer 10.182.224.122 group evpn

peer 10.182.224.122 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Leaf 4.

N/A

peer 10.182.224.246 group evpn

peer 10.182.224.246 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Leaf 2.

N/A

peer 10.182.234.1 group evpn

peer 10.182.234.1 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Border 1.

N/A

peer 10.182.234.2 group evpn

peer 10.182.234.2 group evpn

Add the specified peer to the peer group.

Establish BGP peering with Border 2.

N/A

address-family IPv4 unicast

address-family IPv4 unicast

Create the BGP IPv4 unicast address family and enter its view.

N/A

N/A

balance as-path-neglect

balance as-path-neglect

Enable BGP to ignore the AS_PATH attribute when it performs load balancing.

This enables load balancing across routes with different AS_PATH attribute values.

N/A

quit

quit

Return to BGP view.

N/A

N/A

address-family l2vpn evpn

address-family l2vpn evpn

Enter BGP EVPN address family view.

N/A

N/A

undo policy vpn-target

undo policy vpn-target

Disable target-based filtering of incoming BGP EVPN routes.

N/A

N/A

peer evpn enable

peer evpn enable

Enable the device to exchange routes with the peer group.

N/A

N/A

peer evpn reflect-client

peer evpn reflect-client

Configure the spine nodes as route reflectors and configure the specified peer group as a client.

N/A

N/A

quit

quit

Return to BGP instance view.

N/A

N/A

quit

quit

Return to system view.

N/A

N/A

 

Traffic model

About the traffic model

The traffic model contains the following information:

·     ID—Traffic ID, in O-X-XXX format. The first segment (O) represents overlay traffic. The second segment (X) represents the IP version (4 for IPv4 and 6 for IPv6). The third segment (XXX) represents a unique number for the traffic.

·     Type—Traffic type, such as known unicast/IPV4 and unicast/L2.

·     Direction—Traffic direction, such as inter-leaf east-west traffic, north-west traffic, and inter-DC traffic.

·     Path—The nodes that the traffic traverses from the source to the destination.

·     Simulation method—Traffic simulation method. Testers are used to simulate the patterns of traffic on the network set up in this example.

·     Load—The network is tested under light load (fewer than 1000 simulation traffic flows).

Overlay traffic

ID

Type

Direction

Path

Simulation method

Load

Description

O-4-001

Unicast/L2

East-west traffic between bare metal servers attached to the same leaf DR system.

Server A > Leaf 1&2 > Server C

Bond4 mode + tester

Light

Layer 2 connectivity between bare metal servers attached to the same DR systems at the leaf tier.

O-4-002

Unicast/L2

East-west traffic between bare metal servers attached to the same leaf DR system.

Server C > Leaf 1&2 > Server A

Bond4 mode + tester

Light

Layer 2 connectivity between bare metal servers.

O-4-003

Known unicast/IPv4

East-west traffic between bare metal servers attached to the same leaf DR system.

Server A > Leaf 1&2 > Spine 1&2 > Border 1&2 > Spine 1&2 > Leaf 1&2 > Server C

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers.

O-4-004

Known unicast/IPv4

East-west traffic between bare metal servers attached to the same leaf DR system.

Server C > Leaf 1&2 > Spine 1&2 > Border 1&2 > Spine 1&2 > Leaf 1&2 > Server A

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers.

O-4-005

Unicast/L2

East-west traffic between bare metal servers attached to different leaf DR systems.

Server A > Leaf 1&2 > Spine 1&2 > Leaf 3&4 > Server B

Bond4 mode + tester

Light

Layer 2 connectivity between bare metal servers.

O-4-006

Unicast/L2

East-west traffic between bare metal servers attached to different leaf DR systems.

Server B > Leaf 3&4 > Spine 1&2 > Leaf 1&2 > Server A

Bond4 mode + tester

Light

Layer 2 connectivity between bare metal servers.

O-4-007

Known unicast/IPv4

East-west traffic between bare metal servers attached to different leaf DR systems.

Server A > Leaf 1&2 > Spine 1&2 > Border 1&2 > Spine 1&2 > Leaf 3&4 > Server B

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers.

O-4-008

Known unicast/IPv4

East-west traffic between bare metal servers attached to different leaf DR systems.

Server B > Leaf 3&4 > Spine 1&2 > Border 1&2 > Spine 1&2 > Leaf 1&2 > Server A

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers.

O-4-009

Known unicast/IPv4

North-south traffic

Server A > Leaf 1&2 > Spine 1&2 > Border 1&2 > public

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers and the external network.

O-4-010

Known unicast/IPv4

North-south traffic

Public > Border 1&2 > Spine 1&2 > Leaf 1&2 > Server A

Bond4 mode + tester

Light

Layer 3 connectivity between bare metal servers and the external network.

 

Convergence performance test results

Failure test results

Table 1 Link failure test results

Device

Failure cause

Traffic downtime

Recovery event

Traffic downtime

Leaf

Single ECMP uplink failure

< 200 ms

Recovery from a single ECMP uplink failure

0 ms

Single member link failure of a DR interface

< 30 ms

Recovery from a single DR interface member link failure

< 20 ms

IPL link failure

< 60 ms

Recovery from IPL link failure

< 60 ms

DRNI keepalive link failure

0 ms

Recovery from DRNI keepalive link failure

0 ms

Device restart

< 200 ms

Device restart upon recovery

< 200 ms

Upgrade

< 500 ms (upgrade the DR member devices one by one)

N/A

N/A

Node expansion

< 500 ms

N/A

N/A

Hardware replacement

Fixed-port device replacement: < 500 ms

Modular device:

·     Device replacement: < 1000 ms

·     Switching fabric module replacement: 0 ms

·     Service module replacement: < 500 ms

N/A

N/A

Spine

Single ECMP uplink failure

< 5 ms

Recovery from a single ECMP uplink failure

0 ms

Single ECMP downlink failure

< 15 ms

Recovery from a single ECMP uplink failure

0 ms

Device restart upon exception

~1 second

Device restart upon recovery

< 10 ms

Upgrade

< 100 ms

N/A

N/A

Border

Single uplink failure

< 30 ms

Recovery from a single uplink failure

0 ms

Single ECMP downlink failure

< 5 ms

Recovery from a single ECMP uplink failure

0 ms

IPL link failure

< 20 ms

Recovery from IPL link failure

< 20 ms

DRNI keepalive link failure

0 ms

Recovery from DRNI keepalive link failure

0 ms

Device restart upon exception

< 20 ms

Device restart upon recovery

0 ms

Upgrade

< 30 ms (upgrade the DR member devices one after another)

N/A

N/A

Node expansion

< 30 ms

N/A

N/A

Hardware replacement

Fixed-port device replacement: < 30 ms

Modular device:

·     Device replacement: < 1000 ms

·     Switching fabric module replacement: 0 ms

·     Service module replacement: < 500 ms

N/A

N/A

 

Verifying the configuration

Verification commands

Leaf 1

Leaf 2

Description

display drni summary

display drni summary

Displays summary information about the IPP and DR interfaces in the DR system.

display drni system

display drni system

Displays DR system settings.

display drni keepalive

display drni keepalive

Displays DR keepalive packet statistics.

display drni role

display drni role

Displays DR role information.

display drni consistency-check status

display drni consistency-check status

Displays the configuration consistency check status.

 

Procedure

# Verify that the DR system that contains Border 1 and Border 2 is operating correctly.

[border1]dis drni summary

Flags: A -- Aggregate interface down, B -- No peer DR interface configured

       C -- Configuration consistency check failed

                                       

IPP: BAGG1024

IPP state (cause): UP

Keepalive link state (cause): UP

# Verify the DR system settings on Border 1.

[border1]display drni system

                     System information

Local system number: 1                      Peer system number: 2

Local system MAC: 0002-0002-0001            Peer system MAC: 0002-0002-0001

Local system priority: 10                   Peer system priority: 10

Local bridge MAC: dcda-8040-0520            Peer bridge MAC: b044-14cc-8270

Local effective role: Primary               Peer effective role: Secondary

Health level: 0

Standalone mode on split: Enabled

In standalone mode: No

                                    

                     System timer information

Timer                      State        Value (s)    Remaining time (s)

Auto recovery             Disabled    -            -

Restore delay             Disabled    90           -

Consistency-check delay Disabled    45           -

Standalone delay          Disabled    0            -

Role to None delay       Disabled    60           -

# Verify that Border 1 can send and receive keepalive packets.

[border1]dis drni keepalive

Neighbor keepalive link status (cause): Up

Neighbor is alive for: 952908 s 870 ms

Keepalive packet transmission status:

  Sent: Successful

  Received: Successful

Last received keepalive packet information:

  Source IP address: 2.0.0.2

  Time: 2001/02/01 05:55:23

  Action: Accept

                       

Distributed relay keepalive parameters:

Destination IP address: 2.0.0.2

Source IP address: 2.0.0.1

Keepalive UDP port : 6400

Keepalive VPN name : N/A

Keepalive interval : 1000 ms

Keepalive timeout  : 5 sec

Keepalive hold time: 3 sec

# Verify that Border 1 and Border 2 are the primary and secondary nodes, respectively, in the DR system.

[border1]dis drni role

                    Effective role information

Factors                  Local                    Peer

Effective role           Primary                  Secondary

Initial role             Primary                  None

MAD DOWN state           No                       No

Health level             0                        0

Role priority            32768                    32768

Bridge MAC               dcda-8040-0520           b044-14cc-8270

Effective role trigger: IPL calculation

Effective role reason: Single None role

                                                                

                    Configured role information

Factors                  Local                    Peer

Configured role          Primary                  Secondary

Role priority            20                       30

Bridge MAC               dcda-8040-0520           b044-14cc-8270

# Verify the configuration consistency status on Border 1.

[border1]dis drni consistency-check status

                 Global Consistency Check Configuration

Local status     : Enabled           Peer status     : Enabled

Local check mode : Strict            Peer check mode : Strict

                                      

                 Consistency Check on Modules

Module           Type1           Type2

LAGG             Check           Check

VLAN             Check           Check

STP              Check           Check

MAC              Not Check       Check

L2VPN            Not Check       Check

                                                                                 

                 Type1 Consistency Check Result

Global consistency check result: SUCCESS

Inconsistent global modules: N/A

Upgrading the devices

Upgrading a leaf device

Verifying that all upgrade requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all upgrade requirements are met.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Upgrading the device

See H3C Switches DR System Upgrade & Replacement & Expansion Guide.

Estimating upgrade downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate downtime when you schedule an upgrade.

When you upgrade the DR member devices one by one while the traffic volume is light, traffic downtime of a member device is less than 200 ms upon failover and 60 ms upon fallback.

Verifying the upgrade result

Execute the commands in "Verification commands" and the following commands to verify that the device is upgraded successfully.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Upgrading a spine device

Verifying that all upgrade requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all upgrade requirements are met.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Upgrading the device

1.     Execute the display version command to verify the current BootWare image version and startup software version.

2.     Use the release notes for the upgrade software version to evaluate the upgrade impact on your network and verify the following items:

¡     Software and hardware compatibility.

¡     Version and size of the upgrade software.

¡     Compatibility of the upgrade software with the current BootWare image and startup software image.

3.     Use the release notes to verify whether the upgrade software images require a license. If licenses are required, check the system for availability of valid licenses. If no valid licenses are available, register and activate licenses for each license-based software image.

4.     Use the dir command to verify that the device has sufficient storage space for the upgrade images. If the storage space is not sufficient, delete unused files by using the delete command. Make sure that all MPUs in the system have sufficient storage space.

5.     Use FTP or TFTP to transfer the upgrade image file to the root directory of a file system.

6.     Upgrade the software. For more information about the software upgrade procedure, see the fundamentals configuration guide for the device.

Estimating upgrade downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate upgrade downtime when you schedule an upgrade.

When you upgrade the DR member devices one by one with reboot while the traffic volume is light, traffic downtime of a member device is less than 500 ms upon failover and 150 ms upon fallback.

Verifying the upgrade result

Execute the commands in "Verification commands" and the following commands to verify that the device is upgraded successfully.

 

Spine 1

Spine 2

Description

display device

display device

Displays summary information about the IPP and DR interfaces in the DR system.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Upgrading a border device

Verifying that all upgrade requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all upgrade requirements are met.

 

Border 1

Border 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Upgrade procedure

See H3C Switches DR System Upgrade & Replacement & Expansion Guide.

Estimating upgrade downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate the downtime when you schedule an upgrade.

When you upgrade the DR member devices one by one while the traffic volume is light, traffic downtime of a member device is less than 30 ms upon failover and 20 ms upon fallback.

Verifying the upgrade result

Execute the commands in "Verification commands" and the following commands to verify that the upgrade has been done correctly.

Table 2 Post-upgrade verification commands

Border 1

Border 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Expanding the network

An expansion operation adds two leaf devices.

Adding a leaf device

Verifying that all expansion requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all requirements are met for an expansion.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Adding the expansion device to the network

1.     Make sure the expansion device is not connected to network management systems.

2.     Upgrade the device to the target software version.

3.     Preconfigure the device.

4.     Connect the device to the network management systems.

5.     Incorporate the device into the controller.

Estimating expansion downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate the downtime when you schedule a node expansion.

Verifying the expansion result

Execute the following commands to verify that the device has been added successfully.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Replacing hardware

Replacing an interface module

Verifying that all replacement requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all requirements are met for a replacement.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Replacing hardware

Before you replace an interface module, make sure the service and management traffic has switched over to other interface modules that are operating correctly.

Replace the interface modules online while the system is operating or power off the system before you do the replacement, depending on the evaluation of the conditions.

For details, see H3C Switches DR System Upgrade & Replacement & Expansion Guide.

Estimated replacement downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate the downtime when you schedule a hardware replacement.

Verifying the replacement result

Use the same commands for pre-replacement verification to verify that the system can operate correctly after the hardware replacement.

Replacing a switching fabric module

Verifying that all replacement requirements are met

Execute the commands in "Verification commands" and the following commands to verify that all requirements are met for a replacement.

 

Leaf 1

Leaf 2

Description

display device

display device

Displays device information.

display boot-loader

display boot-loader

Displays the current software images and startup software images.

display version

display version

Displays system version information.

 

Replacing hardware

Replace the switching fabric module online while the system is operating or power off the system before you do the replacement, depending on the evaluation of the conditions.

Estimating replacement downtime

To minimize the impact on services, use information provided in "Convergence performance test results" to estimate the downtime when you schedule a hardware replacement.

Verifying that the replacement is successful

Use the same commands for pre-replacement verification to verify that the system can operate correctly after the hardware replacement.

  • Cloud & AI
  • InterConnect
  • Intelligent Computing
  • Security
  • SMB Products
  • Intelligent Terminal Products
  • Product Support Services
  • Technical Service Solutions
All Services
  • Resource Center
  • Policy
  • Online Help
All Support
  • Become a Partner
  • Partner Resources
  • Partner Business Management
All Partners
  • Profile
  • News & Events
  • Online Exhibition Center
  • Contact Us
All About Us
新华三官网