... Here we are, at the final part of our series on how to install a virtual NSX lab.
With our vSphere infra, NSX Manager, and NSX Control Cluster installed, we can now move to the NSX dataplane.

NSX Dataplane

The process for data plane installation is based upon the NSX Manager deploying VXLAN VIBs to each of the ESXi hosts. This results in a VXLAN VTEP (a VMKernel interface) being installed onto each host.

The IP/VLANs that will be used are:

Cluster vDS VLAN Network IP Range
NFV-Resource vDS-Resource 113 172.29.113.0/24 172.29.113.100-150
NFV-Edge vDS-Edge 123 172.29.123.0/24 172.29.123.100-150

image12
Figure 1 - Starting the VXLAN install.
vxlan-install1
Figure 2 - Configuring VXLAN VTEPs.

Verify

Once VXLAN is installed we can verify that the VTEP is present via the following command:

[root@esxi001-nsxlab:~] esxcli network ip interface ipv4 get --netstack=vxlan
Name  IPv4 Address    IPv4 Netmask   IPv4 Broadcast  Address Type  Gateway  DHCP DNS
----  --------------  -------------  --------------  ------------  -------  --------
vmk1  172.29.113.104  255.255.255.0  172.29.113.255  STATIC        0.0.0.0     false
vmk2  172.29.113.106  255.255.255.0  172.29.113.255  STATIC        0.0.0.0     false

From the output above we can see that 2 VTEPs are shown. This is expected due to there being 2 physical uplinks from each ESXi host.

Final Verification

Now that the management, control, and data plane is installed we can run a final verification.

Our final verification will test connectivity between clusters using a logical switch (aka VXLAN tunnel).

Create VMs

2 VMs are deployed - 1 upon the NFV-Control cluster and 1 upon the NFV-Edge (as shown below).

Note: In essence, based on the NFV 3 pod design VMs would only be deployed within the resource pod and not upon the Edge pod. But for sake of testing an L2 network using our overlay (VXLAN) across an underlay spanning different L3 networks, this placement is fine.

image15
Figure 3 - VM layout across clusters.

Define VNI Range

The Segment IDs are defined, aka VNIs are configured. Like so,

image14
Figure 4 - Define the VNI range.

Create Transport Zone

The Transport Zone defines which clusters the logical switch will span. Below shows the steps:

image9
image4-7
Figure 5 and 6 - Create transport zone.

Create Logical Switch

Next the logical switch is created (shown below), with,

  • the previously created transport zone being applied.
  • Enable IP Discovery being enabled.

Note: The Enable IP Discovery option aids with reducing ARP broadcasts. By the NSX discovering the IPs/MAC address mappings from each host. At the point, the VTEP sees an ARP broadcast it intercepts, queries the NSX controller, and then responds to the ARP request locally.

image6
Figure 7 - Create logical switch.

Assign VMs to Logical Switch

Go into each VM, and assign the Network Adapter to the previously created logical switch.

image1-7
Figure 8 - Assign VM to logical switch.

Configure VM IP's

Go into each VM and configure eth0 with IPs in the same network. I.e

  • Server001 - ifconfig eth0 192.168.1.10/24 up
  • Server002 - ifconfig eth0 192.168.1.11/24 up

Ping

With all the steps now done, we can now issue a ping on each VM, and connectivity should be seen. If your pings are successful - good work
- you now have traffic running across your logical switch aka your VXLAN overlay.

Outro

That concludes are series around installing a virtual NSX lab and we hope you have enjoyed the series.

https://docs.vmware.com/en/VMware-NSX-for-vSphere/6.2/com.vmware.nsx.troubleshooting.doc/GUID-18EDB577-1903-4110-8A0B-FE9647ED82B6.html

Ready to Master Network Automation? Start Your Journey Today!
Our membership provides:
  • Full deep-dive course library (inc. Batfish, pyATS, Netmiko)
  • Code repositories inc. full course code, scripts and examples
  • 24x7 multi-vendor labs (Arista, Cisco, Juniper)
  • Private online community
  • Live monthly tech sessions
  • Access to tech session library

Join Now ➜