NSX-T: Part 9 – T0 Logical Router – Logical Switch & Basic Routing 1/2

 

NSX Replication Modes

Transormers

NSX-V

When two VMs connected to different ESXi hosts need to communicate directly, unicast VXLAN-encapsulated traffic is exchanged between the VTEP IP addresses associated with their associated hypervisors. Traffic originated by a VM may also need to be sent to all the other VMs belonging to the same logical switch. Specific instances of this type of L2 traffic include broadcast, unknown unicast, and multicast. These multi-destination traffic types are collectively referred to using the acronym BUM (Broadcast, Unknown Unicast, Multicast).
In each of these scenarios, traffic originated by a given ESXi host must be replicated to multiple remote. NSX supports three different replications modes to enable multi-destination communication on VXLAN backed Logical Switches: multicast, unicast and hybrid.

  • Multicast Mode (+PIM & +IGM Snooping)

In this mode, a multicast IP address must be associated to each defined VXLAN L2 segment (i.e., Logical Switch). L2 multicast capability is used to replicate traffic to all VTEPs in the local segment (i.e., VTEP IP addresses that are part of the same IP subnet). Additionally, IGMP snooping should be configured on the physical switches to optimize the delivery of L2 multicast traffic. To ensure multicast traffic is also delivered to VTEPs in a different subnet from the source VTEP, the network administrator must configure PIM and enable L3 multicast routing.

Multicast

  • Hybrid Mode (+IGMP Snooping)

Hybrid mode offers operational simplicity similar to unicast mode – IP multicast routing configuration is not required in the physical network – while leveraging the L2 multicast capability of physical switches.
Hybrid mode allows deployment of NSX in large L2 topologies by helping scale multicast at layer 2 with the simplicity of unicast. It also allows scaling in L3 leafspine topologies without requiring PIM to forward multicast (BUM frame) beyond layer 3 ToR boundaries while still allowing multicast replication in physical switch for layer 2 BUM replication.

Hybrid

  • Unicast Mode (-PIM & -IGM Snooping)

Unicast mode replication mode does not require explicit configuration on the physical network to enable distribution of multi-destination VXLAN traffic. This mode can be used for small to medium size environments where BUM traffic is not high and NSX is deployed in an L2 topology where all VTEPs are in the same subnet.

unicast

Source from NSX Reference Design Version 3.0

https://communities.vmware.com/servlet/JiveServlet/download/37591-3-195840/VMware%20NSX-T%20Reference%20Design%20Guide.pdf

 

 

Decepticon

NSX-T

NSX-T does not differentiate between the different kinds of frames replicated to multiple destinations. Broadcast, unknown unicast, or multicast traffic will be flooded in a similar ashion across a logical switch. In the overlay model, the replication of a frame to be flooded on a logical switch is orchestrated by the different NSX-T components. NSX-T provides two different methods for flooding traffic described in the following sections. They can be selected on a per logical switch basis.

  • Head-End Replication Mode

In the head end replication mode, the transport node at the origin of the frame to be flooded on a logical switch sends a copy to each and every other transport node that is connected to thislogical switch.In this mode, the burden of the replication rests entirely on source hypervisor. Seven copies of the tunnel packet carrying the frame are sent over the uplink of “HV1”. This should be taken into account when provisioning the bandwidth on this uplink. Typical use case of this type of replication is necessary when underlay is an L2 fabric.
The source hypervisor transport node knows about the groups based on the information it has received from the NSX-T Controller. It does not matter which transport node is selected to perform replication in the remote groups so long as the remote transport node is up and available.

 

HE

 

  • Two-tier Hierarchical

The two-tier hierarchical mode achieves scalability through a divide and conquest method. In this mode, hypervisor transport nodes are grouped according to the subnet of the IP address of their TEP. Hypervisor transport nodes in the same rack typically share the same subnet for their TEP IPs, though this is not mandatory.
The default two-tier hierarchical flooding mode is recommended as a best practice as it typically performs better in terms of physical uplink bandwidth utilization and reduces CPU utilization.

 

TT

 

Source from VMware NSX-T Reference Design Version 2.0

https://communities.vmware.com/servlet/JiveServlet/downloadBody/27683-102-10-41631/NSX%20Reference%20Design%20Version%203.0.pdf

 

 

Nested vLab Network Environment

The following lab is based on T0 deployment performed on Part 7 .

 

Lab ENV

 

The picture below summarize the environment composed by:

  • A vSphere Compute Cluster (3 Hosts)
  • A KVM Compute Node
  • A vSphere Cluster that hosts the Edge Cluster (2 ESGs) with a T0 Router (DLR-S1-01)
  • 3 Logical Switches routed by T0 DLR
  • 4 Linux VMs (each hosted by an vSphere/KVM Hypervisor)

Setup is performed with NSX-T Advanced workflow

 

NSX-T Routing1

 

 

Create Logical Switches

Under Advanced Networking and Security | Switching | Switches click on +ADD

ls1-1

 

Create 3 new LS (LS-OL-WEB01, LS-OL-APP02 & LS-OL-DB03)

  • Type a LS name
  • Select Overlay Transport Zone
  • Leave Replication Mode ad Two-Tier

ls2-1

LS are created…

ls3-1

 

…and on vSphere Client | Networking view we can verify the new objects attached to Compute Cluster ESXi hosts

ls4-1

 

Attach LS to DLR

Under Advanced Networking and Security | Routers select DLR, click on Configuration | Router Ports | +ADD

ls6-1

 

Add 3 Router Ports for each LS

  • Type a Router Port name
  • Select Type as Downlink
  • Select LS
  • Type a Switch Port name
  • Add an IP Address with a Netmask (this will be the default GW of the network)

 

ls7-1

ls8-1ls9-1

The 3 new RPs are ready

ls10-1

 

Bind vSphere VMs on LS

On vSphere Client, change vNIC parameters for the VMs

web-1-10 –> LS-OL-WEB01
web-1-11 –> LS-OL-WEB01
app-1-20 –> LS-OL-APP02

 

ls11-1

 

ls12-1

 

ADD KVM VM on LS

Download from CentOS site a qcow2 image, scp to KVM host under /var/lib/libvirt/images and copy as db-s1-30

[root@cen-s1-20 images]# ls -lrt
total 8839068
drwx------ 2 root root      16384 Mar 19 14:52 lost+found
-rw-r--r-- 1 root root       1128 Mar 19 23:11 guestinfo.xml
-rw-r--r-- 1 root root 8112766976 Mar 20 00:47 nsx-unified-appliance-2.4.0.0.0.12456291.qcow2
-rw-r--r-- 1 root root  938409984 Apr  4 20:56 CentOS-7-x86_64-GenericCloud.qcow2
[root@cen-s1-20 images]# cp CentOS-7-x86_64-GenericCloud.qcow2 db-s1-30.qcow2

 

Customize CentOS image removing cloud-init and set the root password

[root@cen-s1-20 images]# virt-customize -a db-s1-30.qcow2 --root-password password:centos --uninstall cloud-init
[   0.0] Examining the guest ...
[   2.2] Setting a random seed
[   2.2] Uninstalling packages: cloud-init
[   3.9] Setting passwords
[   4.9] Finishing off

 

Install the KVM VM

[root@cen-s1-20 images]# virt-install --import --vnc --name db-s1-30 --ram 1024 --vcpu 1 --disk path=/var/lib/libvirt/images/db-s1-30.qcow2,format=qcow2 --network bridge:nsx-managed,model=virtio,virtualport_type=openvswitch
WARNING  No operating system detected, VM performance may suffer. Specify an OS with --os-variant for optimal results.
WARNING  Graphics requested but DISPLAY is not set. Not running virt-viewer.
WARNING  No console to launch for the guest, defaulting to --wait -1

Starting install...
Domain installation still in progress. Waiting for installation to complete.

 

Check VM status

[root@cen-s1-20 ~]# virsh list
 Id    Name                           State
----------------------------------------------------
 6     db-s1-30                       running

 

Now we need the VM interface ID in order to create a new NSX-T logical port

[root@cen-s1-20 images]# virsh dumpxml db-s1-30 | grep interfaceid
        <parameters interfaceid='28d256cc-54ff-4f67-bbde-00ef845953f7'/>

 

Under Advanced Networking and Security | Switching | Ports click on +ADD

  • Type a Switch Port name
  • Select the right LS
  • Select VIF as Attachment Type
  • Paste the KVM VM Interfaced ID under Attachment ID

ls15-1

The db-s1-30 KVM VM SP is ready!

ls16-1

 

 

Complete KVM VM setup

 

Login on db-s1-01 console

[root@cen-s1-20 ~]# virsh console db-s1-30
Connected to domain db-s1-30
Escape character is ^]

CentOS Linux 7 (Core)
Kernel 3.10.0-957.1.3.el7.x86_64 on an x86_64

localhost login: root
Password:

 

Setup networking parameter (just for testing)

[root@localhost ~]# ifconfig
eth0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500
inet6 fe80::5054:ff:fe14:d118 prefixlen 64 scopeid 0x20<link>
ether 52:54:00:14:d1:18 txqueuelen 1000 (Ethernet)
RX packets 8 bytes 656 (656.0 B)
RX errors 0 dropped 0 overruns 0 frame 0
TX packets 26 bytes 6812 (6.6 KiB)
TX errors 0 dropped 0 overruns 0 carrier 0 collisions 0

[root@localhost ~]# ifconfig eth0 10.1.3.30 netmask 255.255.255.0 broadcast 10.1.3.255 up

[root@localhost ~]# route add default gw 10.1.3.1

[root@localhost ~]# hostnamectl set-hostname db-s1-30

 

Test reachability and routing

From web-s1-10 (10.1.1.10):

  • ping web-s1-11 (10.1.1.11) on the same LS & different vSphere host
  • ping app-s1-20 (10.1.2.20) on the different LS & different vSphere host
  • ping db-s1-30 (10.1.3.30) on the different LS on KVM host

 

ls17-1

Ben Kenobi

Rispondi

Inserisci i tuoi dati qui sotto o clicca su un'icona per effettuare l'accesso:

Logo di WordPress.com

Stai commentando usando il tuo account WordPress.com. Chiudi sessione /  Modifica )

Google photo

Stai commentando usando il tuo account Google. Chiudi sessione /  Modifica )

Foto Twitter

Stai commentando usando il tuo account Twitter. Chiudi sessione /  Modifica )

Foto di Facebook

Stai commentando usando il tuo account Facebook. Chiudi sessione /  Modifica )

Connessione a %s...