VCF 4.2 Proof of Concept Guide

POC Guide Overview

The purpose of this document is to act as a simple guide for proof of concepts involving VMware Cloud Foundation 4.2 and associated infrastructure tasks to configure and manage Software Defined infrastructure.

This document is intended for data center cloud administrators who deploy a VMware Cloud Foundation system in their organization's data center. The information in this guide is written for experienced data center cloud administrators.

This document is not a replacement for official product documentation; however, it should be thought of as a guide to augment existing guidance throughout the lifecycle of a proof-of-concept exercise. The guide aims to offer a structured approach during evaluation of VCF features.

Official documentation should supersede guidance documented here if the there is a divergence between this document and product documentation.

When referring to any statements made in this document, verification regarding support capabilities, minimums and maximums should be cross-checked against official VMware Technical product documentation at https://docs.vmware.com/en/VMware-Cloud-Foundation/index.html  and https://configmax.vmware.com/ in case of more recent updates or amendments to what is stated here.

Summary of Changes

The following features have been updated as a part of this document:

This document is laid out into several distinct sections to make the guide more consumable depending on the use case and proof of concept scenario.
 

VCF 4.2 What’s New / Overview
VCF 4.2 BOM updates and new features

Section 1 Planning / Day 0
This section covers off guidance and requirements for VCF bring up and considerations such as external resources and dependencies.

Section 2 VCF Infrastructure Deployment / Day 1
Deployment of VCF infrastructure components such as Mgmt, Workload domains and NSX-T Edge Clusters.

Section 3 vRealize Suite Install
Guidelines to deploy SDDC components such as vRealize Operations and vRealize Log Insight.

Section 4 Solution Deployment guidelines.
Proof of Concept guidelines to deploy SDDC infrastructure using specific features and vSphere with Tanzu, Stretch Cluster, vVOLs or vLCM.

Section 5: Day 2 Use cases
NSX-T Network validation.
Certificate Management.
Password Management
VCF Upgrades
vRealize Suite Monitoring and Alerting configuration and management

Appendices

VCF 4.2 What’s New / Overview

Cloud Foundation Bill of Materials (BOM)

 

For more information, please refer to the release notes in case of updates or amendments.

https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/rn/VMware-Cloud-Foundation-42-Release-Notes.html

Below table lists the full BOM of VCF 4.2.
 

Software Component

Version

Date

Build Number

Cloud Builder VM

4.2

09 FEB 2021

17559673

SDDC Manager

4.2

09 FEB 2021

17559673

VMware vCenter Server Appliance

7.0 Update 1c

17 DEC 2020

17327517

VMware ESXi

7.0 Update 1d

04 FEB 2021

17551050*

VMware NSX-T Data Center

3.1.0

30 OCT 2020

17107167

VMware vRealize Suite Lifecycle Manager

8.2 Patch 2

04 FEB 2021

17513665

Workspace ONE Access

3.3.4

04 FEB 2021

17498518

vRealize Automation

8.2

06 OCT 2020

16980951

vRealize Log Insight

8.2

06 OCT 2020

16957702

vRealize Log Insight Content Pack for NSX-T

3.9.2

n/a

n/a

vRealize Log Insight Content Pack for Linux

2.1

n/a

n/a

vRealize Log Insight Content Pack for Linux  - Systemd

1.0

n/a

n/a

vRealize Log Insight Content Pack for vRealize Suite Lifecycle Manager 8.0.1+

1.0.2

n/a

n/a

vRealize Log Insight Content Pack for VMware Identity Manager

2.0

n/a

n/a

vRealize Operations Manager

8.2

06 OCT 2020

16949153

vRealize Operations Management Pack for VMware Identity Manager

1.1

n/a

n/a

 

VCF 4.2 Summary update

For more information, please review

Release Notes

https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/rn/VMware-Cloud-Foundation-42-Release-Notes.html
 

VCF 4.2 Blog

https://blogs.vmware.com/cloud-foundation/2021/02/04/announcing-vmware-cloud-foundation-4-2/

  • vSAN Data Persistence platform – VMware Cloud Foundation 4.2 supports new partner integrations on the vSAN Data Persistence platform.  With this new integration with VMware Cloud Foundation, customers can manage new S3-compatible object storage for unstructured data with the industry-leading HCI platform, via support for partner offerings from Cloudian and MinIO.  Read more about the partner integrations on the Virtual Blocks blog here.
  • vSAN HCI MeshAnnounced with vSAN 7.0U1 and now supported within VMware Cloud Foundation, HCI Mesh delivers a unique software-based approach for disaggregation of compute and storage resources. HCI Mesh helps customers reduce CAPEX by sharing capacity across vSAN clusters while lowering OPEX as they reduce the amount of storage resources managed by scaling efficiently. 
  • NSX-T 3.1 Federation – Delivered as part of NSX-T 3.1 and now supported with VCF 4.2, NSX-T Federation provides a cloud-like operating model for network administrators by simplifying the consumption of networking and security constructs. Delivered in VMware Cloud Foundation 4.2 via prescriptive guidance, NSX-T Federation includes centralized management, consistent networking and security policy configuration with enforcement and synchronized operational state across large scale federated NSX-T deployments. With NSX-T Federation, VCF customers can leverage stretched networks and unified security policies across multi-region VCF deployments providing workload mobility and simplified disaster recovery.  Read this blog for additional details on NSX-T Federation in VCF.
  • vRealize Automation support for VMware Cloud Foundation: SDDC Manager and VCF workload domains with VMware Cloud Assembly as VMware Cloud Foundation cloud accounts. A VMware Cloud Foundation cloud account can help you facilitate a comprehensive hybrid cloud management solution. For more information, see vRealize Automation 8.2 Release Notes.
  • Static IP Pool for NSX-T TEPs: VMware Cloud Foundation 4.2 introduces additional flexibility to leverage static IP pools for NSX-T Host Overlay (TEP) networks as an alternative to DHCP. This applies to the management domain and VI workload domains with uniform L2 clusters.
  • Release Versions UI: SDDC Manager UI includes a Release Versions page with information on the Bill of Materials, new features, and end of general support dates for each available VMware Cloud Foundation release. This page can help you in upgrade planning.
  • Enhanced skip upgrade experience: You can filter available upgrade bundles by the target release you want to skip to using SDDC Manager UI or public APIs.
  • Improvements to upgrade resiliency: VMware Cloud Foundation 4.2 includes pre-checks for password validation, API performance optimization, and improvements to ESXi error reporting.
  • Advanced Security Add-on for VMware Cloud Foundation: The Advanced Security Add-on for VMware Cloud Foundation now includes advanced threat protection, workload and endpoint security providing the following capabilities*:

Section 1 VCF Deployment Planning & Bring Up / Day 0

To plan for a successful VCF POC, there is a considerable number of external requirements to ensure success.
The key to a successful plan is to use a reasonable hardware configuration that resembles what you plan to use in production.

Physical Network and External services
Certain requirements such as routable VLANS, MTU and DNS and DHCP services are required, these are in summary:

  • Top of Rack switches are configured. Each host and NIC in the management domain must have the same network configuration.
  • IP ranges, subnet mask, and a reliable L3 (default) gateway for each VLAN.
  • At minimum, an MTU of 1600 is required on the NSX-T Host Overlay (Host TEP) and NSX-T Edge Overlay (Edge TEP) VLANs end-to-end through your environment.
  • VLANs for management, vMotion, vSAN, NSX-T Host Overlay (Host TEP), NSX-T Edge Overlay (Edge TEP), and NSX uplink networks are created and tagged to all host ports. Each VLAN is 802.1q tagged. NSX-T Host Overlay (Host TEP) VLAN and NSX-T Edge Overlay (Host TEP) VLAN are routed to each other.
  • Management IP is VLAN-backed and configured on the hosts. vMotion and vSAN IP ranges are configured during the bring-up process.
  • DHCP with an appropriate scope size (one IP per physical NIC per host) is configured for the NSX Host Overlay (Host TEP) VLAN.

 

AVNs or Application Virtual Networks are optional to configure but highly recommended to evaluate vRealize Suite and vSphere Tanzu.

To use Application Virtual Networks (AVNs) for vRealize Suite components you also need:

  • Top of Rack (ToR) switches configured with the Border Gateway Protocol (BGP), including Autonomous System (AS) numbers and BGP neighbor passwords, and interfaces to connect with NSX-T Edge nodes.
  • Two VLANs configured and presented to all ESXi hosts to support the uplink configuration between the (ToR) switches and NSX-T Edge nodes for outbound communication.

Physical Hardware and ESXi Hosts

Refer to the VMware vSAN Design and Sizing Guide for information on design configurations and considerations when deploying vSAN. Be sure the hardware you plan to use is listed on the VMware Compatibility Guide (VCG). BIOS updates, and firmware and device driver versions should be checked to make sure these aspects are updated according to the VCG.

  • Identical hardware (CPU, Memory, NICs, SSD/HDD, and so on) within the management cluster is highly recommended. Refer to vSAN documentation for minimum configuration.
  • Hardware and firmware (including HBA and BIOS) is configured for vSAN.
  • Physical hardware health status is "healthy" without any errors.
  • ESXi is freshly installed on each host.
  • Each ESXi host is running a non-expired license. The bring-up process will configure the permanent license.

 

Software and Licenses

  • The ESXi version matches the build listed in the Cloud Foundation Bill of Materials (BOM). See the VMware Cloud Foundation Release Notes for the BOM.
  • VCF Cloud Builder OVA
  • Adequate licenses for VCF components and number of workload domains that is planned for deployment.

 

Further resources

Step by step overview on how to stand up Enabling Kubernetes on Cloud Foundation on VMware Cloud Foundation
https://core.vmware.com/delivering-developer-ready-infrastructure#step_by_step_guide_to_deploying_developer_ready_infrastructure_on_cloud_foundation_isim_based_demos

 

Management Workload Domain Overview

SDDC Manager and other vSphere, vSAN, and NSX components that form the core of VMware Cloud Foundation are initially deployed to an environment known as the Management workload domain. This is a special-purpose grouping of systems devoted to managing the VMware Cloud Foundation infrastructure.

Each Cloud Foundation deployment begins by establishing the Management workload domain, which initially contains the following components:

  • SDDC Manager
  • vCenter Server with integrated Platform Services Controller
  • vSAN
  • NSX-T

 

Management Workload Domain Logical View:

Graphical user interface, diagram, website</p>
<p>Description automatically generated
 

In addition to the Cloud Foundation components that are provisioned during the bring-up process, additional virtual machine workloads may be deployed to the Management workload domain if required. These optional workloads may include third party virtual appliances or other virtual machine infrastructure workloads necessary to support a particular Cloud Foundation instance.

The vCenter with internal Platform Service Controller instance deployed to the Management workload domain is responsible for SSO authentication services for all other workload domains and vSphere clusters that are subsequently deployed after the initial Cloud Foundation bring-up is completed.

Additional details regarding the conguration and usage of Cloud Foundation workload domains may be found in the following section of this guide, Workload Domain Creation.

Pre-Requisites and Bring-Up Process

Prerequisites and Preparation

VMware Cloud Foundation (VCF) deployment is orchestrated by the Cloud builder appliance, which builds and configures VCF components. To deploy VCF, a parameter file (in the form of an Excel workbook or JSON file) is used to set deployment parameters such as host name, IP address, and initial passwords. Detailed descriptions of

VCF components may be found in the VCF Architecture and Deployment Guide.

The Cloud Builder appliance should be deployed on either an existing vSphere cluster, standalone host, or laptop (requires VMware Workstation or VMware Fusion). The Cloud Builder appliance should have network access to the Management Network segment defined the parameter file to enable connectivity to the ESXi hosts composing the management workload domain.

There are specific requirements that need to be fulfilled before the automated build process or ‘bring-up’ may begin. for instance, DNS records of the hosts, vCenter, NSX Manager, etc. should have been configured. Before starting, download the parameter spreadsheet to support planning and configuration of deployment prerequisites.

The OVA for Cloud Builder appliance and parameter workbook (Cloud Builder Deployment Parameter Guide) for version 4.2 can be found at:
https://my.vmware.com/group/vmware/downloads/details?downloadGroup=VCF420&productId=1121&rPId=60057

 

Alternatively, the parameter workbook may also be downloaded from the Cloud Builder appliance after it has been deployed.
Once the workbook has been completed, the file should be uploaded to the appliance, where upon a script converts the Excel to a JSON file. This JSON file is then validated and used in the bring-up process.

The VMware Cloud Foundation YouTube channel is a useful resource to reference alongside this guide.

Deploy Cloud Builder Appliance

Download the Cloud Builder appliance and import the OVA. Once the OVA has been imported, complete the appliance configuration:Graphical user interface, application</p>
<p>Description automatically generated

The ‘Deployment Architecture’ should be set to ‘vcf’ (default).
Enter credentials for the admin and root accounts; the hostname and IP address of the appliance and gateway, DNS, and NTP details.

Bring-Up Parameters

Parameters required for configuring VCF during the bring-up process are entered into an Excel workbook, which may be downloaded from the Cloud Builder download page or from the appliance itself. Each version of VCF has a specific version of the parameter workbook associated with it.

There are several worksheets within the Excel workbook. Certain fields are subject to validation based on inputs elsewhere in the workbook. Care should be taken not to copy/paste cells, or otherwise alter the structure of the spreadsheet.

'Prerequisite Checklist’: This worksheet lists deployment prerequisites. Mark the ‘Status’ column for each row ‘Verified’ when each prerequisite is satisfied.

Graphical user interface, application</p>
<p>Description automatically generated

‘Management Workloads’: Shows the VMs deployed on the management domain. Only licenses, i.e., column L, should be populated on this worksheet. For the current versions of VCF (4.2), leave the SDDC Manager Appliance license empty. Entering information into the 'Management Workload Domain Calculations' section will help to understand resource usage after deployment.

Graphical user interface, application</p>
<p>Description automatically generated

 

‘Users and Groups’: Enter a password for each service account. Ensure that each password entered meets cell validation requirement

A picture containing application</p>
<p>Description automatically generated

 

‘Hosts and Networks’: VLANs, IP addresses/gateways, and management workload domain hostnames should be entered in this worksheet. If the ‘Validate ESXi Thumbprints’ option is set to ‘No,’ then the respective host SSH ngerprints will be ignored. Any native VLAN should be marked with a zero (0). In many cases, and especially for POC deployments, the vSAN and vMotion networks may be non-routable and not have a gateway. In this case, enter a gateway value within the respective subnet range, but not used by any device (this will produce a warning on bring-up which may be ignored).
 

Graphical user interface</p>
<p>Description automatically generated
Note: The MTU used here is not reflective of a production environment. the MTU was chosen for internal lab restrictions when creating this document. Supported MTU sizes for are 1600 - 9000 for NSX-T based traffic

 

Graphical user interface, application</p>
<p>Description automatically generated

‘Deploy Parameters’: This worksheet contains the bulk of the information required. Here, infrastructure dependencies such as DNS and NTP are specied, along with hostnames and IP addresses of management components.

 

Graphical user interface, text</p>
<p>Description automatically generated
NSX-T Details: More will be covered on Application Virtual Networks section later in this document.

SDDC Manager Details

Specifications related to host network congurations, as well as object names within the vSphere hierarchy are also specied within this worksheet.

To view an interactive demonstration of this process with step-by-step instructions, please visit Deployment Parameters Worksheet in the VCF resource library on core.vmware.com.

Network and VLAN Configuration

There are several VLANs that must be configured for the management domain:

  • Management (for hosts and management VMs) vMotion
  • vSAN
  • NSX-T Host Overlay
  • NSX-T Edge Overlay
  • NSX-T Edge Uplinks (2)
     

Table</p>
<p>Description automatically generated

In addition, for VCF version 4.x, two uplink VLANs are required for BGP peering between the NSX-T Edge VMs and the top of rack switch (see below).

For initial host bring-up, it is important to note that the default ‘VM Network’ port group should be on the same VLAN as the Management port group. The Cloud Builder appliance and SSDC Manager should be deployed to the same VLAN.

Jumbo frames are required for NSX / VTEP (MTU of at least 1600) and recommended for other VLANS (MTU 9000). Configure the network infrastructure to facilitate frames of 9000 bytes.

Note: In the above example, the VLAN was marked zero (0) due to internal lab restrictions. As mentioned previously, native VLAN should be marked with a zero (0).

Also note the MTU used here is not reflective of a production environment. the MTU was chosen for internal lab restrictions when creating this document. Supported MTU sizes for are 1600 - 9000 for NSX-T based traffic.

A change introduced in VCF 4.1 in the UI, is the ability to bring up the management domain with different vSphere Distributed Switch profiles. This allows for the utilization of multiple network interfaces as well as multiple vDS configurations. One example is to use two different network interfaces for vSAN while two other interfaces will carry all other traffic. Multi-pnic capability was added on VCF 4.0 and version 4.0.1 with VCF on VxRail. 

 

 

Finally, a DHCP server is required on the Host Overlay VLAN to issue addresses on each host. If there is no DHCP server available VCF 4.2 can leverage static IP pools for NSX-T Host Overlay (TEP) networks as an alternative. The static IP pool is configured from deploy parameters spreadsheet in the section labeled NSX-T Host Overlay Network. To enable use of the static IP pool, select YES from the “Configure NSX-T Host Overlay Using a Static IP Pool” drop down.

Complete the information required for the static IP pool. The diagram below provides an example.

When planning for IP Pools it is important to ensure the pool is large enough to accommodate the amount of hosts. Each VCF host will require two IP addresses in the host TEP. For example, for a basic workload domain of 4 hosts a static IP pool of 8 IP addresses is required.

 

Application Virtual Networks

To support Application Virtual Networks (AVNs); BGP peering between the NSX-T Edge Service Gateways and upstream network switches is required for the management domain.

The diagram below shows an overview the BGP AS setup between the two NSX-T Edges deployed with VCF and the physical top of rack switches:

A picture containing timeline</p>
<p>Description automatically generated

Inside the rack, the two NSX-T edges form one BGP AS (autonomous system). Upstream, we connect to two separate ToR switches, each in their own BGP AS. The two uplink VLANs connect northbound from each edge to both ToRs.

The BGP conguration is defined in the parameter spreadsheet, in the 'Deploy Parameters' tab, under the section 'Application Virtual Networks'. We define the ToR details (as per the diagram above), with the respective IP address, BGP AS and password:

Graphical user interface, text, application</p>
<p>Description automatically generated

 

To complete the peering, the IP addresses of the two edges, with the ASN should be congured on the ToR (as BGP neighbors).

Note: BGP Password is required and cannot be blank. NSX-T Supports a maximum of 20 characters for the BGP password.

The Cloud builder appliance must be able to resolve and connect to the NSX-T edges in order to validate the BGP setup, etc.

Note that for the purposes of a PoC, virtual routers (such as Quagga or vyos)  could be used to peer with. In this case, make sure that communication northbound for NTP and DNS is available.

Host Configuration

Installing and Configuring ESXi on Host Servers

Hardware components should be checked to ensure they align with the VMware vSphere Compatibility Guide (VCG). Drives and storage controllers must be vSAN certied, and rmware/drivers must be aligned with those specied in the VCG. See section 4.1.1 for a full list of host requirements.

Note that VCF requires identical hardware and software conguration for each ESXi host within a given workload domain, including the Management workload domain.

ESXi should be installed on each host. Hosts must match the ESXi build number specied the VCF Bill of Materials (BOM) for the version of VCF being deployed. Failure to do so may result in failures to upgrade ESXi hosts via SDDC Manager. It is permissible to use a custom image from a hardware vendor as long as the ESXi build number still matches the VCF BOM. The BOM may be located within the Release Notes for each version of VCF.

The release notes for VCF 4.0.x are located at: ttps://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/rn/VMware-Cloud-Foundation-42-Release-Notes.html
From here, we can see that the ESXi build number should be 17551050.

 

Therefore, ensure the correct version and build of ESXi is installed on the hosts:
https://my.vmware.com/en/web/vmware/downloads/info/slug/datacenter_cloud_infrastructure/vmware_vsphere/7_0

After ESXi has been installed, login to the host client on each host and ensure that:

  • The host name should be changed to match the entry in the parameter spreadsheet.
  • The login password is the same as on the parameter spreadsheet.
  • The correct management IP address and VLAN (as per the parameter spreadsheet) has been configured Only one physical adapter is connected to the Standard Switch.
  • No vSAN configuration is present, and all disks (other than the boot disk) have no partitions present.
  • NTP should be configured with the IP address or host name of the NTP server.
  • Both the SSH and NTP service should started and the policy changed to ‘Start and stop with host’ Finally, ensure that the hosts are not in maintenance mode.

DNS Configuration

Every IP address and host name combination defined in the parameter workbook (i.e., hosts, NSX Manager, vCenter, etc.) must have forward and reverse entries in DNS before bring-up.

Ensure entries are correct and accounted for before starting the bring-up process and test each DNS entry for forward and reverse lookup.
Bring up will fail if DNS is not configured correctly.

Post bring-up tasks such as creating new VI Workload domains, new clusters, adding hosts, etc. also require creating forward and reverse DNS lookup entries for associated components.

Regenerate the Self-Signed Certificates on hosts

Once you have configured the ESXi hosts' identity by providing a host name you must regenerate the self-signed certificate to ensure the correct common name is defined. During the installation of ESXi, the installer generates a self-signed certificate for each ESXi host but the process is performed prior to the ESXi identity being configured. This means all ESXi hosts have a common name in their self-signed certificate of localhost.localdomain. All communication between VMware Cloud Builder and the ESXi hosts is performed securely over HTTPS and as a result validates the identify when making a connection by comparing the common name of the certificate against the FQDN provided within the VMware Cloud Builder configuration file.

To ensure that connections attempts during validation do not fail the self-signed certificates must be manually regenerated. To regenerate the certificates follow the steps below.

  • Log in to the ESXi host using an SSH client such as Putty.
  • Regenerate the self-signed certificate by executing the following command:
/sbin/generate-certificates
  • Restart the hostd and vpxa services by executing the following command:
/etc/init.d/hostd restart && /etc/init.d/vpxa restart

SDDC Bring-Up

Once each host has been congured, DNS entries conrmed, and networks setup, verify that the parameter workbook is complete, then begin the bring-up process.

Deploy and Power-on the Cloud Builder appliance. If congured correctly, the appliance will boot to a console displaying the IP address of the appliance:

Graphical user interface, text</p>
<p>Description automatically generated

To start the bring up process, navigate to the Cloud Builder in a web browser and login with the credentials that were provided in the OVA import.

Select ‘VMware Cloud Foundation’ as the platform.
Next, review the bring-up checklist to ensure all steps have been completed:

Graphical user interface, text, application, email</p>
<p>Description automatically generated

On the next page, we are given the option to download the parameter spreadsheet and upload a completed le for validation. If needed, download the Deployment Parameter Spreadsheet.

After the spreadsheet has been completed, upload it to Cloud builder.

Introduced in VCF 3.9.1 Application Virtual Networks (AVN) are the network foundation for supporting workload mobility in applications such as VMware vRealize Automation, VMware vRealize Operations Manager, and VMware vRealize Orchestrator. VMware recommends enabling AVNs from the beginning: conguring AVNs later is possible, but it is a manual process.

The conguration for AVNs can be found in the deployment spreadsheet:

Graphical user interface, table</p>
<p>Description automatically generated with medium confidence

Once the parameter spreadsheet has been uploaded, click on ‘Next’ to begin the validation process.

Once the process has completed, review any errors and warnings. Pay close attention to any password, DNS, or network warnings (note that in many cases, especially for POCs, both vSAN and vMotion networks may not be routable – and therefore the gateway for that network may show as unreachable).

Once satised that the issues have been addressed, click Next:

Graphical user interface</p>
<p>Description automatically generated

Click On “Deploy SDDC: to begin the deployment process:

Graphical user interface, text, application, email</p>
<p>Description automatically generated

During the bring-up process, periodically monitor the running tasks. Filter for ‘In-progress' to see the current task, the Deployment of VCF usually completes in 2-4 hours:

To monitor progress with greater visibility, use tail to display the bring-up logs on the Cloud Builder appliance: open an SSH session to the appliance and log in using the admin account. Run the command below to tail the bring-up logs. Note that there will be a considerable number of messages:

tail -Fqn0 /var/log/vmware/vcf/bringup/* | grep -v "Handling get all"

It may also be useful to login to the deployed vCenter instance (check the status messages to determine when it is available) to monitor bring-up progress.

To view an interactive demonstration of this process with step-by-step instructions, please visit VCF4 Bringup without AVN or VCF4 Bringup with AVN in the VCF resource library.

Once all tasks have been finished, the appliance will indicate that the SDDC setup has been successfully completed:

Graphical user interface, text, application, website</p>
<p>Description automatically generated

Bring-up is complete, and the Cloud Builder appliance may be powered o.

Post Bring-Up Health Check

SDDC Manager: Dashboard

After the bring-up process has nished, login to SDDC Manager. Upon login, a dashboard presenting an overview of the VCF environment is presented.

Graphical user interface, application, Teams</p>
<p>Description automatically generated

All VCF management activities are accomplished through SDDC Manager – no conguration changes should be made to any of the deployed components, including vCenter.

SDDC Manager: User Management

The ‘Users’ panel on the left of the interface shows a list of users inherited from vCenter. To add a user or group, click on '+ User or Group.

Graphical user interface, table</p>
<p>Description automatically generated

Table</p>
<p>Description automatically generated

As such, identity sources from Active Directory, LDAP and Open LDAP added to vCenter will appear here. Note that there are three roles dened in SDDC Manager, ADMIN and OPERATOR and VIEWER.

SDDC Manager: Repository Settings

Once SDDC Manager is setup, users are required to enter ‘My VMware’ account details to enable software bundle downloads. This may require conguration of a proxy in some environments.

Navigate to the ‘Repository Settings’ panel on the left of the interface and enter the account details:

Graphical user interface, application</p>
<p>Description automatically generated

A picture containing graphical user interface</p>
<p>Description automatically generated

 

Once bundles are available to download, the ‘Bundles’ panel will populate:
See the section on ‘LCM Management’ for further information on managing bundles.
Graphical user interface, text, application, email</p>
<p>Description automatically generated

SDDC Manager: Backup Configuration

It is recommended that the NSX managers are backed up to an external destination (currently SFTP is supported). Navigate to ‘Backup Conguration’ on the panel on the left and click on ‘Register External’:

Graphical user interface, application</p>
<p>Description automatically generated

Enter the IP address, port, user credentials, etc. for the external destination:

Graphical user interface, application</p>
<p>Description automatically generated

 

Section 2 VCF Infrastructure Deployment / Day 1.

Workload Domains

Workload Domain Overview

In VMware Cloud Foundation, a “workload domain” (or WLD) is a policy-based resource container with specic availability and performance attributes that combines compute (vSphere), storage (vSAN, NFS, VMFS or vVOLs), and networking (NSX-T) into a single consumable entity. Each workload domain may be created, expanded, and deleted as part of the SDDC lifecycle operations, and may contain one or more clusters of physical hosts.

Every Cloud Foundation deployment begins with provisioning a management workload domain, which hosts SDDC components necessary for Cloud Foundation to function. After the management workload domain is successfully deployed, SDDC Manager may be used to deploy additional Virtual Infrastructure VI) workload domains to host VM and container workloads. Each VI workload domain is managed by a corresponding vCenter instance, which resides within the VCF management domain; other management- related workloads associated with each workload domain instance may also be deployed within the management domain.

Cloud Foundation uses and is validated against vSAN, NFS v3, and VMFS on FC for principal storage. The management domain requires vSAN for principal storage. You can use vSAN, NFS v3, or VMFS on FC for principal storage with VI workload domains. The type of principal storage used by the initial vSphere cluster in a VI workload domain is defined when the VI workload domain is created. For VI workload domains, you can add vSphere clusters that use a different type of principal storage than the type selected for the initial cluster when the VI workload domain was created. After a vSphere cluster has been deployed you cannot change to another storage type.

Each VCF workload domain requires a minimum of three (3) hosts. Exact requirements vary depending on the workload domain type the host resides in. See the table below for details.

Component

Requirements

 

  For vSAN-backed VI workload domains, three (3) compatible vSAN ReadyNodes are required. For information about compatible vSAN ReadyNodes, see

the VMware Compatibility Guide.

  For NFS-backed workload domains, three (3) servers compatible with the vSphere version included with the Cloud Foundation BOM are required. For information about the BOM, see the Cloud Foundation Release Notes. For compatible servers, see the VMware Compatibility Guide.

  For VMFS on Fibre Channel backed workload domains, three (3) servers compatible with the vSphere version included with the Cloud Foundation BOM are required. For information about the BOM, see the Cloud Foundation Release Notes. In addition, the servers must have supported Fibre Channel (FC) cards (Host Bus Adapters) and drivers installed and congured. For compatible servers and Fibre Channel cards, see the VMware Compatibility Guide.

 

Servers within a cluster must be of the same model and type.

 

  For vSAN-backed VI workload domains, supported vSAN congurations are required.

  For NFS-backed VI workload domains, congurations must be compatible with the vSphere version included with the Cloud Foundation BOM. For more information about the BOM, see the Cloud Foundation Release Notes.

  For VMFS on Fibre Channel backed workload domains, congurations must be compatible with the vSphere version included with the Cloud Foundation BOM. For information about the BOM, see the Cloud Foundation Release Notes.

 

  Two 10GbE (or faster) NICs. Must be IOVP certied.

  (Optional) One 1GbE BMC NIC

Servers

CPU, Memory, and Storage

NICs

 

In this proof-of-concept guide, we will focus on conguration of workload domains with vSAN-backed storage. For conguration of NFS or FC-backed storage, please consult the Cloud Foundation documentation in conjunction with documentation from the NFS or FC storage array vendor.

      Commissioning Hosts

      If available hosts that meet requirements are not already in the Cloud Foundation inventory, they must be added to the inventory via the Commission Hosts process. Hosts that are to be commissioned should not be associated with a vCenter and should not be a member of any cluster. Additionally, prior to commissioning, each host must meet certain conguration prerequisites:

      • Hosts for vSAN-backed workload domains must be vSAN compliant and certied per the VMware Hardware Compatibility Guide. BIOS, HBA, SSD, HDD, etc. must match the VMware Hardware Compatibility Guide.
      • Host has a standard virtual switch back by two (2) physical NIC ports with a minimum 10 Gbps speed. NIC numbering should begin with vmnic0 and increase sequentially.
      • Host has the drivers and rmware versions specied in the VMware Compatibility Guide.
      • Host has ESXi installed on it. The host must be preinstalled with supported versions listed in the BOM. SSH and syslog are enabled on the host.
      • Host is congured with DNS server for forward and reverse lookup and FQDN. Hostname should be same as the FQDN.
      • Management IP is congured to rst NIC port.
      • Ensure that the host has a standard switch and the default uplinks with 10Gb speed are congured starting with traditional numbering (e.g., vmnic0) and increasing sequentially.
      • Host hardware health status is healthy without any errors. All disk partitions on HDD / SSD are deleted.
      • Ensure required network pool is created and available before host commissioning.
      • Ensure hosts to be used for VSAN workload domain are associated with vSAN enabled network pool. Ensure hosts to be used for NFS workload domain are associated with NFS enabled network pool.
      • Ensure hosts to be used for VMFS on FC workload domain are associated with NFS or vMotion only enabled network pool.

       

      To commission a host in SDDC manager complete the following steps:

      1. Navigate to the Inventory > Hosts view, and select ‘Commission Hosts’ at the top right of the user interface.
      2. Verify that all host conguration requirements have been met, then click ‘Proceed’.
      3. On the next screen, select one or more hosts to be commissioned. These may be added via the GUI interface, or alternatively may be added through a bulk import process. To add hosts via the GUI, ensure the ‘Add new’ radio button has been selected, and ll in the form. Then, click ‘Add’.
      4. Alternatively, to bulk import hosts, click the ‘JSON’ hyperlink to download a JSON template for entering host information. After entering host details into the. JSON le, save it locally and select the ‘Import’ radio button. Then, click ‘Browse’ to select the. JSON file and click ‘Upload’ at the lower right to upload the le to SDDC Manager.
      5. When all hosts for commissioning are added, conrm the host ngerprints by selecting all hosts in the ‘Hosts Added’ table by clicking the grey circle with a check mark located beside each host ngerprint listed in the ‘Conrm Fingerprint’ column. When the circle turns green, click the ‘Validate All’ button located near the upper right corner of the table.
      6. After clicking ‘Validate All’, wait for the host validation process to complete. This may take some time. When the validation process completes, verify that all hosts have validated successfully, then click ‘Next’ to advance the wizard. On the nal screen of the wizard, review the details for each host, then click ‘Commission’ to complete the process.

      Adding a Workload Domain

      Once the available hosts have been commissioned or added to VCF, a new workload domain can be created. To create a workload domain complete the following steps:

      1. Navigate to the Workload Domains inventory view. Then, at the top right of the screen, click “+Workload Domain”, then select VI – Virtual Infrastructure from the drop down.
      2. Choose the storage type to use for this workload domain, vSAN, NFS, or VMFS on Fiber Channel (this cannot be changed later).
      3. Enter conguration details for the workload domain. Note that information will be used to provision a new instance of vCenter. This instance’s VM resides in within the Management workload domain, and manages the clusters associated with its respective VI workload domain. Please ensure that valid forward and reverse DNS entries for the vCenter FQDN are congured, then click ‘Next.’
      4. On the next screen, choose the version of NSX to deploy, and enter the deployment parameters. Ensure that forward and reverse DNS entries for NSX Manager FQDNs are to be congured, and that the correct NSX software bundles have been downloaded on SDDC Manager. Then, click ‘Next.’
      5. On the fourth screen in the wizard, congure the vSAN default Failures to Tolerate (FTT). Enabling the Dedupe and Compression feature for all-ash clusters is optional. Then, click ‘Next.’
      6. The next step requires selecting available hosts from inventory to add to the workload domain. If there are no hosts available, please follow the instructions above for commissioning hosts within SDDC Manager. VMware recommends deploying no less than 4 hosts per workload domain in order to ensure that compliance with vSAN FTT=1 policy may be maintained if a vSAN cluster host is oine. However, in cases where hosts available for the POC are limited, it is acceptable to construct a workload domain with the minimum three (3) required hosts, then later add an additional host for the purposes of demonstrating workload domain expansion functionality. For clusters supporting vSAN FTT polices greater than one (1) (i.e., FTT=2 or FTT=3), it is recommended to deploy at least one additional host above the minimum required for policy compliance. See the vSAN Design and Sizing guide for additional details. After selecting the hosts to be added, click ‘Next.’
      7. Now, choose licenses available within SDDC Manager for the workload domain. If no applicable licenses are present, please add them to SDDC Manager (Administration > Licensing) then click ‘Next’.
      8. Finally, review the deployment summary, then click ‘Finish’ to launch the deployment.

       

      Workload Domain Creation using multiple physical network interfaces and multiple vSphere Distributed Switches

      Starting with VCF version 4.0, you can create a new Workload Domain using multiple network interfaces; however, currently this option requires a json file to be created and executed via API call. Luckily, VCF now includes the API Explorer, which is a UI driven API command center. The support for VCF on VxRail multi-pnic approach was introduced in VCF version 4.0.1.

      • Within SDDC Manager navigate to Developer center (bottom of the left pane).
      • Select API Explorer.
      • Look for APIs for managing Domains.
      • You will see that each API has a description as to what they do.
      • The POST API for /v1/domains is used to create a Domain.
      • If you click the DomainCreationSpec it will give you a json file that only requires inputs such as names, IP addresses, etc.

      • Download/copy the json file and edit it to add the necessary information.
      • For the purposes of multi-pnic and multi-vDS, we want to focus on the hostNetworkSpec section.
        • This is what tells the automation what network interfaces and vDS will used, so this part is important.

       

       

      • As you can see from the example, you will also need your ESXi license and the host ID for hosts that have already been commissioned in SDDC manager and are ready to be consumed. Please refer to the host commissioning section for details.
      • In order to get the host IDs for the host to be used for the new VI Workload Domain, you can use the API explorer to get this information.
      • Navigate to the API Explorer and then click on APIs for managing Hosts.
      • For status enter UNASSIGNED_USEABLE and click EXECUTE.

       

      • Click on PageofHost and this will expand.
      • You should now see the unassigned hosts that can be used.

       

      • Clicking on each host you will see more information including the ID.
      • After you have completed the json file with all the information needed, including host IDs, ESXi licenses, as well as hostNetworkSpec that includes multiple Nics and vDS settings, paste the contents of the json file into the POST API under APIs for managing Domains and execute.

       

      To view an interactive demonstration of this process with step-by-step instructions, please visit Create Workload Domain (NSX-T and vSAN) in the VCF resource library on TechZone.

      Review Workload Domain Components

      Components deployed during the workload domain creation process may be viewed within SDDC Manager. To view these components, navigate to Inventory > Workload Domains within SDDC Manager, then click the name of the workload domain you would like to inspect.

      To view an interactive walk-through of VCF SDDC components, please visit the Review Workload Domain Components demonstration https://core.vmware.com/?share=isim_demo2129 in the VCF resource center on core.vmware.com.

      Expand Workload Domain Cluster

      To expand the host resources available within a workload domain, the SDDC Manager interface is used to move one or more unused hosts from the SDDC Manager inventory to a workload domain cluster.

      Before attempting to add additional hosts to a workload domain, verify that ‘Unassigned’ hosts are available in the SDDC Manager Inventory. If no hosts are presently ‘Unassigned’, please follow the host commissioning process to make one or more hosts available for use.

      To view an interactive demonstration of this process with step-by-step instructions, please visit the Expand Cluster demonstration https://core.vmware.com/?share=isim_demo2127 in the VCF resource center on core.vmware.com.

       

      Expand Workload Domain using multiple physical network interfaces and multiple vSphere Distributed Switches

      In VCF version 4.2, you can create a new Workload Domain using multiple network interfaces; however, currently this option requires a json file to be created and executed via API call. Luckily, VCF now includes the API Explorer, which is a UI driven API command center.

      • Within SDDC Manager navigate to Developer center (bottom of the left pane).
      • Select API Explorer.
      • Look for APIs for managing Clusters.
      • You will see that each API has a description as to what they do.
      • The PATCH API for /v1/clusters/{id} is used to create a add/remove a host from a cluster.
      • Expand the next to PATCH and click on ClusterUpdateSpec .This will show you the json file that needs to be completed.

       

      • Download/copy the json file and edit it to add the necessary information.
      • For the purposes of multi-pnic and multi-vDS, we want to focus on the hostNetworkSpec section.
        • This is what tells the automation what network interfaces and vDS will used, so this part is important.

       

       

      • As you can see from the example, you will also need your ESXi license and the host ID for hosts that have already been commissioned in SDDC manager and are ready to be consumed. Please refer to the host commissioning section for details.
      • In order to get the host IDs for the host to be used for the new VI Workload Domain, you can use the API explorer to get this information.
      • Navigate to the API Explorer and then click on APIs for managing Hosts.
      • For status enter UNASSIGNED_USEABLE and click EXECUTE.

      • Click on PageofHost and this will expand.
      • You should now see the unassigned hosts that can be used.

       

      • Clicking on each host you will see more information including the ID.
      • After you have completed the json file with all the information needed, including host IDs, ESXi licenses, as well as hostNetworkSpec that includes multiple Nics and vDS settings.
      • After the json file has been completed it is necessary to acquire the id of the cluster that will be expanded
      • To do this, from the APIs for managing Clusters click on GET for /v1/clusters and EXECUTE.
      • This will provide you with a list of clusters and additional information such as cluster ID.
      • Copy the cluster ID for the cluster to be expanded into notepad to keep it handy.
      • The cluster ID will be used for 2 steps.
        • Step 1: Validate json file for the cluster.
        • Step 2: Add host to the cluster.
      • In order to validate the json file go to POST /v1/clusters/{id}/validations under APIs for managing Clusters and expand it.
      • Paste the cluster ID acquired on previous steps.
      • Paste the completed json file and click EXECUTE.

       

      • After validation succeeds, you can now move on to the actual task of expanding the cluster.
      • Click on PATCH API for /v1/clusters/{id}.
      • Paste json file and cluster id and click EXECUTE.
      • You can then see the Task in the task pane at the bottom of the screen and track the progress of each sub-task.

       

      NSX Configuration Overview: VI Workload Domain(s)

      When creating a VI workload domain, NSX-T is deployed to support its networking stack. There are prerequisites for deploying NSX-T; please refer to the VCF Product Documentation for details.

      A cluster of three NSX-T Manager nodes is deployed by default when an NSX-T based workload domain is created. On the workload domain page, select the summary view, and then

      Graphical user interface, application</p>
<p>Description automatically generated

      If an NSX-T Edge Cluster is also created, it will be visible and associated to the workload domain instance of NSX-T.

      Click the FQDN of the NSX-T Cluster. This will open a new browser tab and automatically log into one of the NSX-T Manager instances.

      Graphical user interface, application</p>
<p>Description automatically generated

      Confirm that the NSX-T management cluster is in ‘STABLE’ state. Also verify that the Cluster Connectivity for each node is ‘Up’:

      Graphical user interface, application</p>
<p>Description automatically generated

      To review the Transport Zones congured, Select System > Fabric > Transport zones

      There are two Overlay transport zones and one VLAN transport zone.

      Hosts associated with this workload domain are connected to the default NSX-T overlay for the workload domain, in this case four hosts

      Graphical user interface, application</p>
<p>Description automatically generated

      Select System > Fabric > Nodes and select Host Transport Nodes tab in the “Managed by” drop-down list to show all transport nodes in the cluster associated with the vCenter instance associated with the workload domain.

      Ensure the ‘Conguration’ is set to ‘Success’ and “Node Status” is ‘Up’ for each node:

      Graphical user interface</p>
<p>Description automatically generated with medium confidence

      vCenter

      All NSX Managers for new workload domains are deployed on the Management workload domain resource pool. From the management vCenter, go to Hosts and Clusters and expand the resource pool mgmt-rp

      Graphical user interface</p>
<p>Description automatically generated

       

      vSphere Networking

      With previous versions of NSX-T, installing NSX-T required setting N-VDS and migrating to from vDS. Now it is possible to use a single vSphere Distributed Switch for both NSX-T 3.0 with vSphere 7 networking. When installing NSX-T 3.0 it can run on top of the existing vDS without needing to move pNICs to N-VDS.

      Note: When NSX-T is associated with the vSphere VDS it will be updated on the summary page that it is managed by NSX-T instance.

      NSX-T Edge Cluster Deployment

      You can add multiple NSX-T Edge clusters to workload domains for scalability and resiliency. However, multiple Edge clusters cannot reside on the same vSphere cluster.
      NSX-T Data Centre supports a maximum of 16 Edge clusters per NSX Manager cluster and eight Edge clusters per vSphere cluster.
      The north-south routing and network services provided by an NSX-T Edge cluster created for a workload domain are shared with all other workload domains that use the same NSX Manager cluster.
      For more information, please review VCF documentation. https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/com.vmware.vcf.admin.doc_41/GUID-8FA66DA3-3166-426B-84A8-C45FA7651658.html

      In this POC guide we will have already deployed an additional workload domain.

      The purpose of this document is to walk through the conguration to understand the network requirements and nally to check and validate that the edge(s) were deployed successfully.

      Prerequisites

      As per the documentation
      Note: Please refer to official documentation for detailed steps on https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/com.vmware.vcf.admin.doc_41/GUID-D17D0274-7764-43BD-8252-D9333CA7415A.html. Official documentation should supersede if it differs from guidance documented here.

      Below is a guided deployment with screenshots to augment the deployment.

      • Separate VLANs and subnets are available for NSX-T Host Overlay (Host TEP) VLAN and NSX-T Edge Overlay (Edge TEP) VLAN. A DHCP server must be congured on the NSX-T Host Overlay (Host TEP) VLAN.
      • You cannot use DHCP for the NSX-T Edge Overlay (Edge TEP) VLAN.
      • NSX-T Host Overlay (Host TEP) VLAN and NSX-T Edge Overlay (Edge TEP) VLAN are routed to each other.
      • For dynamic routing, set up two Border Gateway Protocol (BGP) Peers on Top of Rack (ToR) switches with an interface IP, BGP autonomous system number (ASN), and BGP password.
      • Reserve a BGP ASN to use for the NSX-T Edge cluster’s Tier-0 gateway.
      • DNS entries for the NSX-T Edge nodes are populated in the customer-managed DNS server.
      • The vSphere cluster hosting an NSX-T Edge cluster must include hosts with identical management, uplink, host TEP, and Edge TEP networks (L2 uniform).
      • You cannot deploy an Edge cluster on a vSphere cluster that is stretched. You can stretch an L2 uniform vSphere cluster that hosts an Edge cluster.
      • The management network and management network gateway for the Edge nodes must be reachable.
      • Workload Management supports one Tier-0 gateway per transport zone. When creating an Edge cluster for Workload Management, ensure that its overlay transport zone does not have other Edge clusters (with Tier-0 gateways) connected to it.

       

      Deployment Planning

      As a proof of concept, we will deploy a new Edge Cluster to an already deployed workload domain called wld01.

      • We will deploy a small Edge Node Form Factor: 4 GB memory, 2 vCPU, 200 GB disk space. The NSX Edge Small VM appliance size is suitable for lab and proof-of-concept deployments.

           Note: You cannot change the size of the edge form factor after deployment. Only use SMALL for lab or proof of concepts. For Tanzu (Kubernetes) please use LARGE

      • We will deploy two Edges in Active-Active High Availability Mode (In the active-active mode, trac is load balanced across all members and If the active member fails, another member is elected to be active)

       

      We have gathered the following details prior to deployment.

      • ASN number: 65004 for Tier-0 BGP
        • ToR Switch IP addresses and subnets
      • NSX-T Overlay VLAN 1252 (routable to host overlay)
        • Static IP addresses for Edge Overlay VLAN 1252
      • Edge Uplinks VLAN 2081 and 2082. (for connectivity to Top of Rack
        • Switch Static IP addresses for VLAN 2081 and 2082

      NSX-T Edge deployment Step by Step Procedure

      From SDDC manager Click on Workload Domains, select a workload domain, and click on Actions Chose Click Add Edge Cluster
       

      Graphical user interface, application</p>
<p>Description automatically generated

      Click Begin to walk through the wizard.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      The following walk-through demo can be reviewed here to understand the process. Please navigate to Add NSX-T Edge Cluster https://core.vmware.com/?share=isim_demo2119

       

      Validation of NSX-T Edge Cluster

      From SDDC Manager UI the new edge cluster is listed on the Workload Domain summary

      Validation is also explored by reviewing the walk-through demo can be reviewed here to understand the process please navigate to demo Add NSX-T Edge Cluster https://core.vmware.com/?share=isim_demo2119  Once the demo is started, navigate to step 47 of the demo.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      From the SDDC manager shortcut launch the NSX-T web interface

      Click System > Fabric Edge Transport Nodes to see the edge node details. We can see edge01- wld01.vcf.sddc.lab and edge02-wld01.vcf.sddc.lab deployed.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      From the top menu click Networking to view the Tier-0 and Tier-1 dashboards

      We can see from the dashboard. the Tier-0 gateways. This is responsible for North-South Routing. We can see BGP is enabled, and a Peer is congured. The Tier-1 gateway is used for East-West trac.

      To view the topology layout between Tier-1, Tier-0 and the outside physical infrastructure, Select Network Topology

       

      We can see 192.168.17.4/24, 192.168.16.4/24 192.168.17.5/24 and 192.168.16.5/24 represent the IP addresses on the edges that are peered to the of rack AS 65001.
       

      A picture containing diagram</p>
<p>Description automatically generated

      To verify BGP connectivity and peering status to the top of rack switches.

      1. Navigate back to Network Overview, select Tier-0 Gateways, select the Tier-0 Edge, edge01-t0, to expand on the details
      2. Click on BGP to expand
      3. Click on 2 for BGP neighbor details (as we have two neighbors congured)

      We can see the status of both Top of Rack BGP Peers. Status of Success indicates peering has been successfully established.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      vSphere

      The NSX-T Edge Cluster will be deployed on the associated workload domain. And Edge Cluster resource pool is created, and the edges are deployed onto the workload domain cluster, in this case wld01

      Note: the vCenter and NSX-T unied controllers are deployed on the Management Domain vSphere Cluster

      To view the edges from a vSphere perspective, login to the vSphere Client, navigate from host and clusters to the vCenter instance associated with the workload domain, expand the cluster and resource pool to inspect the NSX-T Edges.
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      vSphere Networking and vDS Details

      Two additional vDS port groups are created on the workload domain vDS

      from the vSphere Web client, navigate to vSphere Networking, the workload domain vCenter, and the associated VDS to inspect the edge port-groups.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Edge vNICs

      Each Edge will have a similar VM networking configuration.

      Network adapter 1 is for the management network connectivity (MGMT VLAN 0) Network Adapter 2 is associated with the Edge Uplink (VLAN 2081) Network Adapter 2 is associated with the Edge Uplink (VLAN 2082)

      This conguration can be explored on the summary of the Edge Virtual Machine appliance.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Reusing an existing NSX-T manager for a new workload domain

      If you already have an NSX Manager cluster for a dierent VI workload domain, you can reuse that NSX Manager cluster.

      In order to share an NSX Manager cluster, the workload domains must use the same update manager mechanism. The workload domains must both use vSphere Lifecycle Manager (vLCM), or they must both use vSphere Update Manager (VUM).

      Note: - Do not share an NSX Manager cluster between workload domains catering to dierent use cases that would require dierent NSX-T Edge cluster specications and congurations.

      Please review the click through demo that complements this guide. Add Workload Domain with NSX Manager Reuse https://core.vmware.com/?share=isim_demo2181 The demo first reviews an existing workload domain and then walks through deploying a new workload domain. To quickly go through this scenario, we will go through the main parts of the demo

      From SDDC Manager start the deploy wizard for a new VI - Virtual Infrastructure to deploy workload domain,

      Once new workload domain wizard is launched, add the entries for the workload domain name and new vCenter instance.

      Instead of deploying a brand new NSX-T Instance, we will Re-use the NSX-T Instance associated with an existing workload domain, in our case wld01

      The VLAN ID for the pre-existing NSX-T host overlay needs to be validated.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      All NSX-T entries are greyed out as we are using the NSX-T instance associated with wld01 which SDDC Manager is already aware of

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      The following resources will be pre-xed workload domain name wld01

      • vSphere Distributed Switch
      • Resource Pool
      • Distributed port-group vSAN
      • Distributed port-group vMotion
      • vSAN Datastore

      Graphical user interface, table</p>
<p>Description automatically generated with medium confidence

      Once the Workload domain has been deployed it will simply appear as a new workload domain on SDDC Manager but associated with the NSX-T instance belonging to wld01.

      From a vSphere perspective, a new vCenter Server is deployed, a new datacenter and cluster object is created, and hosts added and configured.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      We can also observe the vCenter server appliance vcenter-wld02.vcf.sddc.lab is hosted on the management workload domain with no further additional NSX-T instances.
       

      Graphical user interface, application, website</p>
<p>Description automatically generated

      vSphere Networking comprises of a vDS and 3 port-groups for mgmt, vSAN and vMotion.
       

      Graphical user interface, application, website</p>
<p>Description automatically generated

       

       

      NSX-T

      Graphical user interface, application</p>
<p>Description automatically generated
      The vCenter Server is registered as an additional compute manager to the existing NSX-T instance (as specied on the new workload domain wizard)

      The vSphere hosts are congured as Host Transport nodes associated with that vCenter

      Table</p>
<p>Description automatically generated

       

      However, they are added to the same transport zone as the transport Nodes in the rst workload domain, wld01, i.e., overlay-tx-nsx-wld01.vcf.sddc.lab
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Section 3 vRealize Suite Deployment

      vRealize Suite 2019

      In version 4.1 VCF began supporting vRealize Suite 2019
      VMware Cloud Foundation 4.1 introduces an improved integration with vRealize Suite Lifecycle Manager. When vRealize Suite Lifecycle Manager in VMware Cloud Foundation mode is enabled, the behavior of vRealize Suite Lifecycle Manager is aligned with the VMware Cloud Foundation architecture.

      Note: Please refer to official documentation for detailed steps on Deploying vRealize Suite Lifecycle Manager in Cloud Foundation. Official documentation should supersede if it differs from guidance documented here.

      Below is a guided deployment with screenshots to augment the deployment.

      Prerequisites

      • You must deploy vRealize Suite Lifecycle Manager before you can deploy other vRealize Suite products on Cloud Foundation
      • You must then deploy Workspace ONE Access before you can deploy the individual vRealize Suite products on Cloud Foundation.

      Once you have vRealize Suite Lifecycle Manager installed, you can deploy the other vRealize Suite products:

      • vRealize Operations
      • vRealize Log Insight
      • vRealize Automation

      Once Deployed you can connect individual workload domains to them.

      Prerequisites

      • You must deploy vRealize Suite Lifecycle Manager before you can deploy other vRealize Suite products on Cloud Foundation
      • You must then deploy Workspace ONE Access before you can deploy the individual vRealize Suite products on Cloud Foundation.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Once you have vRealize Suite Lifecycle Manager installed, you can deploy vRealize Suite products such as

      • vRealize Operations
      • vRealize Log Insight
      • vRealize Automation

      For the purposes of this POC Guide we will cover

      • vRealize Life Cycle Manager
      • vRealize Workspace One Access
      • vRealize Operations
      • vRealize Log Insight

      Deploying vRealize Life Cycle Manager

      vRealize Suite Lifecycle Manager introduces a functionality where you can enable VMware Cloud Foundation mode in vRealize Suite Lifecycle Manager 8.1.

      Any operation triggered through vRealize Suite Lifecycle Manager UI is aligned with the VMware Cloud Foundation architecture design.

      When a VMware Cloud Foundation admin logs in to vRealize Suite Lifecycle Manager, you can perform normal regular operations like any vRealize Suite Lifecycle Manager user. The VMware Cloud Foundation user can view applications like, User Management, Lifecycle Operations, Locker, Marketplace, and Identity and Tenant Management but with some limitations.

      You can perform the same set of operations with limited access to the latest version of the vRealize Suite products. To perform a regular operation, you have to specify the license and certificate settings using the Locker in vRealize Suite Lifecycle Manager UI.

      Some of the features that are used by VMware Cloud Foundation from vRealize Suite Lifecycle Manager.

      • Binary mapping. vRealize Suite Lifecycle Manager in VMware Cloud Foundation mode has a sync binary feature from which you can poll the binaries from the VMware Cloud Foundation repository and maps the source automatically in vRealize Suite Lifecycle Manager.
      • Cluster deployment for a new Environment. You can deploy vRealize Automation, vRealize Operations Manager, vRealize Log Insight in clusters, whereas in VMware Identity Manager, you can only deploy both cluster and single node, and later expand to a cluster.
      • Product Versions. You can only access the versions for the selected vRealize products that are specifically supported by VMware Cloud Foundation itself.
      • Resource Pool and Advanced Properties. The resources in the Resource Pools under the Infrastructure Details are blocked by the vRealize Suite Lifecycle Manager UI, so that the VMware Cloud Foundation topology does not change. Similarly, the Advanced Properties are also blocked for all products except for Remote Collectors. vRealize Suite Lifecycle Manager also auto-populates infrastructure and network properties by calling VMware Cloud Foundation deployment API.

       

      vRSLCM Perquisites

      To Deploy SDDC manager you will need.

      • vRSLCM downloaded via SDDC Manager.
      • AVN networks ensuring routing between AVNs and Management networks is functioning correctly.
      • IP address and DNS record for vRealize Life Cycle Manager.
      • Free IP address in AVN Segment for Tier 1 gateway.
      • DNS and NTP services available from AVN Segments.

      Note: Please refer to official documentation for detailed steps on Deploy vRealize Suite Lifecycle Manager in Cloud Foundation. Official documentation should supersede if it differs from guidance documented here.

      Below is a guided deployment with screenshots to augment the deployment.

      Step by Step Deployment

      From SDDC Manager, select vRealize Suite and click Deploy.
      AVN Network Segment, Subnet, gateway, DNS and NTP settings should be pre-populated by SDDC Manager.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Click Next

      For NSX-T Tier 1 Gateway, enter in free IP Address on AVN Segment. Do not use the same IP address as another IP address on the AVN segment. It must be a free and unused IP address.

      The default System Administrator user ID is vcfadmin@local

       

      Graphical user interface, application</p>
<p>Description automatically generated
       

      vRSLCM Deployment task can be monitored via SDDC Manager Tasks

       

      A picture containing table</p>
<p>Description automatically generated

       

      Once vRSLCM is deployed successfully the next step is to license vRealize Suite

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Add vRealize Suite License key.
       

      Add license key to vRSLCM

      1. Login to vRSLCM with vcfadmin@local (you may have to change password on initial login)
         
      2. Navigate to Locker and Select License.
         

      Graphical user interface, text, application, email</p>
<p>Description automatically generated
      3. Select Add license and validate. Once Validated select Add.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Deploying VMware Identity Manager

      You must deploy Workspace ONE via vRealize Suite Lifecycle Manager

      Requirements are:

      • Workspace ONE Access software bundle is downloaded under Bundle Management on SDDC Manager
      • vRealize Suite License key
      • 5 static IP address with FQDNs (forward and reverse lookup)
      • CA signed certificate or self-signed certificate
      • vcfadmin@local password

       

      Note: Please refer to official documentation for detailed steps on Deploy vRealize Suite Lifecycle Manager in Cloud Foundation. Official documentation should supersede if it differs from guidance documented here.

      In this POC scenario we will deploy a clustered Workspace one instance, so we will require an IP address for

      Cluster VIP, Database IP for 3 IP addresses for each cluster member and a certificate which includes the FQDN names and IP addresses of each member.

      e.g.

      IP (AVN Segment)

      FQDN

      Purpose

      192.168.11.13

      m01wsoa.vcf.sddc.lab

      Cluster VIP

      192.168.11.14

      m01wsoa1.vcf.sddc.lab

      Cluster Node 1

      192.168.11.15

      m01wsoa2.vcf.sddc.lab

      Cluster Node 2

      192.168.11.16

      m01wsoa3.vcf.sddc.lab

      Cluster Node 3

      192.168.11.17

      n/a

      Database IP


      Workspace ONE Binaries

      Ensure binaries are downloaded via SDDC manager. vRSLCM should map product binaries to SDDC repository as part of deployment.

      To verify, in the navigation pane, select Lifecycle management > Bundle management.

      Click the Bundles tab, locate the Workspace ONE Access install bundle. Click Download bundle if not present

       

       

      • Login to vRSLCM with vcfadmin@local
      • Navigate to Lifecycle Operations > Settings > Binary Mapping
      • Ensure Workspace ONE OVA is present. You may have to synch binaries from SDDC Manager if OVA is not present.

       

      Graphical user interface, application, email</p>
<p>Description automatically generated

       

      Add Workspace ONE Certificate

      You can generate a self-signed certificate or Generate CSR for an external CA

      In this scenario we are going to generate a CSR to create a CA signed certificate.

      • Login to vRSLCM with vcfadmin@local
         
      • Navigate to Locker, Select Certificate and Generate.

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      • Add Name, CN name (this is the IP of Workspace ONE Cluster IP)

       

      If Workspace ONE is going to be deployed in cluster mode add Cluster IP, and cluster members (FQDN) and IP addresses for each member.
       

      It is also possible to generate a certificate signing request to submit to an external CA. Click Generate CSR

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      You will be prompted to download and save the .pem file which includes private key and signing request once all fields have been filled out.

      Save the file, so it can be retrieved later.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      The CSR can be signed by an appropriate CA. If using an internal Microsoft CA, paste the CSR PEM file contents to the CA to generate a new certificate request.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

       

      Download Certificate and Certificate chain and use Base 64 Encoding

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Once Certificate chain is generated and downloaded (Base 64 encoded), return to vRSLCM > Lifecycle Operations > Locker > Certificate and click Import

      Graphical user interface, application, Teams</p>
<p>Description automatically generated

       

      To import the generated Certificate. Provide a Name, e.g., Workspace ONE Certificate, paste the Private Key from the CSR request and Paste the certificate chain generated earlier in the procedure.

      Note: The Private Key and Certificate can be combined into a single file to simplify the process.
      Add the Private Key first then append the certificate chain to a file.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

       

      Graphical user interface, application</p>
<p>Description automatically generated

       


      Default Admin Password

      In vRealize Suite Lifecycle Manager 8.X stores all the passwords that are used across the vRealize Suite Lifecycle Manager. You can configure a password at the locker level and are retrieved from the UI.

      Login to vRSLCM with vcfadmin@local

      Navigate to Locker, and Select password, and select Add

      Graphical user interface, application</p>
<p>Description automatically generated

       

      The default password must be a minimum of eight characters.

      Add the details for Password alias, password itself Description and Username

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Install Identity Manager

      • Login to vRSLCM with vcfadmin@local
      • Navigate to Lifecyle Operations, Create Environment. Use the global environment as is already populated with the vCenter details.
      • Add Administrator email and Default Password.

      Note: if Password is not already configured in Locker, a new password can be created by clicking on the “Plus” Sign to add

      • Select Datacenter, which should already be populated from the Management workload domain from drop down list.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      • Opt in or out of VMware Customer Experience Improvement program and click Next
      • Select VMware Identity Manager and in this scenario, we are selecting clustered mode which means we will need 5 IP addresses in AVN Network Segment, ensure corresponding FQDN are created. Click Next to continue  

       

      Graphical user interface, text, application</p>
<p>Description automatically generated
       

      • Accept the EULA, click Next
      • Select the Certificate that was created earlier (or create a new certificate)

       

      Graphical user interface, text, application, email, Teams</p>
<p>Description automatically generated

       

      Infrastructure details should already be populated from SDDC manager. Since we are deploying to vSAN, choose Thin mode, which means the appliances will be deployed thinly provisioned using the default vSAN Storage Policy, click Next.
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Network Details should also be pre-populated from AVN details, click Next

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      VMware Identity Manager Details must be entered for certificate, password, Cluster IP, Database IP, and Cluster members.

      Below are screenshots of each screen to illustrate the number of inputs.

      Graphical user interface, application</p>
<p>Description automatically generated with medium confidence
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Background pattern</p>
<p>Description automatically generated with medium confidence
       

      Background pattern</p>
<p>Description automatically generated with medium confidence

      Table</p>
<p>Description automatically generated with medium confidence

       

      Run Pre-check before deployment to validate inputs and infrastructure.

       

      Graphical user interface</p>
<p>Description automatically generated with medium confidence

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Ensure all pre-checks pass validation

      Graphical user interface, application</p>
<p>Description automatically generated

      The report can be downloaded in PDF format or pre-check can be re-run.

      Click Next.

      At this point you are ready to submit, a json file can be downloaded to deploy programmatically. Pre-check can be re-run. Or progress can be saved now to submit later. Review settings and click Submit.

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Deployment progress can be monitored from vRSLCM, SDDC manager and vCenter.

      vRSLCM

       

      Graphical user interface</p>
<p>Description automatically generated

       

      SDDC Manager
      NSX-T load balancer will be deployed from SDDC manager as part of the deployment, this can be monitored on SDDC Manager tasks
       

      vCenter
      Workspace one OVAs will be deployed on the management cluster
       

      Graphical user interface</p>
<p>Description automatically generated

       

      Once Deployed Successfully the task should be marked as complete from vRSLCM > Life Cycle Operations > Requests

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Verify Workspace ONE Identity Manager

      Navigate to vRSLCM > Lifecycle Operations > Environments

      Click on globalenviroment and View details

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Trigger inventory sync to trigger sync from vRSLCM to VIDM and to SDDC Manager

      This task can be monitored from Requests on SDDC Manager.
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      From SDDC manager Dashboard Navigate to vRealize Suite. Details of vRSLCM and Workspace One access is registered.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Connect to Workspace one access and connect using the credentials specified during install

      Graphical user interface, application</p>
<p>Description automatically generated

       

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Deploying vRealize Operations

      With vRSLCM and Workspace One deployed you are now able to deploy vROPS

      In this POC scenario we will deploy a 3 node vROPs cluster (Master Replica ad Data node)
      Note: Please refer to official documentation for detailed steps. Official documentation should supersede if it differs from guidance documented here.

       
      vROPS Requirements

      • vRealize Operations Manager binaries
      • vRealize Operations Manager bundle synched to vRSLCM product binaries
      • at least 4 IP addresses for VROPS cluster IP, Master, Replica and data node.
      • Appropriate vRealize License Key
      • Certificate (self-signed or signed by CA)
      • Password setup

       

      For example, we need the following IP addresses with FQDN (forward and reverse lookups
       

      IP (AVN Segment)

      FQDN

      Purpose

      192.168.11.18

      m01vrops.vcf.sddc.lab

      Cluster VIP

      192.168.11.19

      m01vropsmaster.vcf.sddc.lab

      Master VROPS Node

      192.168.11.20

      m01vropsreplica.vcf.sddc.lab

      VROPs Replica Node

      192.168.11.21

      m01vropsdata1.vcf.sddc.lab

      Data Node

       

      vROPS Bundle Mapping

      Verify the latest VROPS bundle has been downloaded on SDDC manager

       

       

      If product binaries are displayed on vRSLCM a manual sync maybe necessary

      Connect to vRSLCM and login with vcfadmin@local
      Navigate to Lifecycle Operations > Settings > Binary Mappings
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Similar to Workspace One, we may need to create a default password credential and Certificate for the vROPS Cluster

       vROPS Default Password

      From vRSLCM, Navigate to Locker > Password. Click Add
      Below is a sample value for vROPS Passwords

       

      Setting

      Value

      Password Alias

      vrops-root

      Password

      vrops-root-password

      Confirm Password

      vrops-root-password

      Password Description

      vROPS Root user

      Username

      root

       

      Graphical user interface, application, email</p>
<p>Description automatically generated

       

      vROPs Certificate

      Again, as per Workspace one we can generate a self-signed certificate or a CA signed certificate

      • From vRSLCM, Navigate to Locker > Certificate > Generate for self-signed or Generate CSR for external CA
      • In our case as we already have an external CA, we will generate a CSR
      • Ensure to add the following CN name should match the cluster VIP and add the master, replica and data nodes in hostname and IP fields,

      The graphic below is an example of the completed certificate fields.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      • Click Generate if generating self-signed or Generate CSR
      • In this example we are generating a CSR.
      • Once the CSR is generated, sign with external CA and import certificate
         

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Create Environment

      We are going to setup a new environment for vROPs. This is in addition to the “global environment” already created.

      On vRSLCM dashboard click Lifecycle operations > Create Environment

      In this case we will call the environment VCF-POC with default password of vrops-root we created earlier

      The datacenter will be from the mgmt. workload domain

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Select vROPS, New install with size of medium, and 3 nodes.

      For Product details enter the following, as per VVD guidance
      We will implement.
       

      Setting

      Value

      Disable TLS version

      TLSv1, TLSv1.1

      Certificate

      vROPS Certificate

      Anti-affinity / affinity Rule

      Enabled

      Product Password

      vrops-root

      Integrate with identity Manager

      Selected

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Select and Validate your license

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Select the vROPS Certificate created earlier

       

      vCenter Infrastructure details are pre-filled out and are displayed to be acknowledged, select Disk Mode to Thin and click next

       

      As with Workspace One, networking details are pulled from SDDC Manager to reflect AVN networks, Click Next

       

      Install vRealize Operations
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      For Cluster VIP add vROPS cluster FQDN

      Background pattern</p>
<p>Description automatically generated with medium confidence

       

      For Master Node component add FQDN (m01vropsmaster.vcf.sddc.lab) and IP address details

      The VM name can be changed to match particular naming conventions.

       

      A picture containing application</p>
<p>Description automatically generated

       

      Click on advanced settings (highlighted) to review NTP and time zone settings.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      For Replica Node component add Replica FQDN (m01vropsreplica.vcf.sddc.lab) and IP details

      Table</p>
<p>Description automatically generated

      Click on advanced configuration Icon to add timezone details

      For Data Node component add Data Node FQDN (m01vropsdata1.vcf.sddc.lab) and IP details

      A picture containing table</p>
<p>Description automatically generated

       

      Click on advanced configuration Icon to add or check time zone details.

      Click Next to continue then click RUN PRECHECK

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Address any errors on Pre-check and ensure all validations succeed

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Review Summary and submit vROPS Deployment

       

      Diagram</p>
<p>Description automatically generated

       

      Progress can also be tracked from Life Cycle Operations > Requests

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Progress can also be tracked from SDDC Manager Tasks

      As we can see as part of deployment vROPS will automatically configured to begin monitoring VCF management domain which includes vCenter, vSAN and Workspace One

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Once deployed, the environment can be viewed from Lifecycle Operations > Environments
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Click on view details to see the details

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Clicking on TRIGGER INVENTORY SYNC will rediscover inventory of VCF management Domain.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Deploying vRealize Log Insight

      Similar to vROPS we can now deploy vRealize Log insight in a new environment on vRCLM

      In this POC scenario we will deploy a 3 node vRealize Log Insight (vRLI) Cluster (one vRLI Master and two worker nodes)

      Note: Please refer to official documentation for detailed steps. Official documentation should supersede if it differs from guidance documented here.

      vRealize Log Insight Requirements

      • vRealize Log Insight Binaries downloaded on SDDC Manager
      • vRealize Log Insight bundle synched to vRSLCM product binaries
      • at least 4 IP addresses for vRLI cluster IP, Master, two worker nodes.
      • Appropriate vRealize License Key
      • Certificate (self-signed or signed by CA) added to vRSLCM Locker
      • Password added to vRSLCM locker

       

      Sample IP addresses for vRLI Cluster need the following IP addresses with FQDN (forward and reverse lookups
       

      IP (AVN Segment)

      FQDN

      Purpose

      192.168.11.22

      m01vrli.vcf.sddc.lab

      vRLI Cluster IP

      192.168.11.23

      m01vrlimstr.vcf.sddc.lab

      vRLI Master Node

      192.168.11.24

      m01vrliwrkr01.vcf.sddc.lab

      Worker Node 1

      192.168.11.25

      m01vrliwrkr02.vcf.sddc.lab

      Worker Node 2

       

      vRealize Log Insight Bundle Download

      Ensure install bundle for vRealize Log Insight 8.1.1 is downloaded on SDDC Manager and binaries are synched to vRSLCM

       

       

      From vRealize Suite lifecycle manager, navigate to Lifecyle Operations > Settings > Binary Mappings

      Ensure binaries are synched once vRealize Log Insight 8.1.1 has been downloaded to SDDC manager

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      vRealize Log Insight Default Password.
       

      From vRSLCM, Navigate to Locker Password. Click Add
       

      Setting

      Value

      Password Alias

      vrli-admin

      Password

      vrli-admin-password

      Confirm Password

      vrli-admin-password

      Password Description

      Log Insight admin password

      Username

      admin

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      vRealize Log Insight Certificate
       

      Again, as per Workspace One and vROPS we can generate a self-signed certificate or a CA signed certificate

      Since this is a cluster, we need a certificate for the following hostnames.

      This IP range is based on the “Region A – Logical Segment” as part of VCF bring up using AVNs.

      IP (AVN Segment)

      FQDN

      192.168.10.22

      m01vrli.vcf.sddc.lab

      192.168.10.23

      m01vrlimstr.vcf.sddc.lab

      192.168.10.24

      m01vrliwrkr01.vcf.sddc.lab

      192.168.10.25

      m01vrliwrkr02.vcf.sddc.lab

       

      This maps to Segment in NSX-T Logical Networks for the management domain

      Graphical user interface, application</p>
<p>Description automatically generated

       

      From vRSLCM, Navigate to Locker > Certificate > Generate for self-signed or Generate CSR for external CA

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Either Generate a new certificate or import a certificate

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      vRealize Log Insight Create Environment
       

      From VRSLCM dashboard go to Lifecycle Operations, then Create Environment

      Add VCF POC Log Insight

      Setting

      Value

      Environment name

      VCF POC vRli

      Administrator email

      administrator@vcf.sddc.lab

      Default Password

      Global Admin Password

      Select Datacenter

      m01-dc01

       

       

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

      Select vRLI with deployment type of Cluster

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Click Next and Accept the EULA.

      Select license, click Validate Association, and click Next

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Select vRealize Log Insight Certificate that was created earlier and click next.

      Verify infrastructure details, click next.
      Note: NSX-T Segment should match VCF deployment)
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Verify Network Details

      Graphical user interface, application</p>
<p>Description automatically generated

      Install Log Insight
      For the purposes of this POC document we will select “Small “form factor for Node Size

      Select Certificate, DRS Anti-affinity rule and integrate with Identity Manager

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Add the IP addresses FQDN for Cluster VIP, Master, and two worker nodes

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated
       

       

      Run Precheck once all IP addresses and FQDN have been entered

      Address any issues and re-run pre-check

      Graphical user interface</p>
<p>Description automatically generated with medium confidence

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Once all prechecks are validated, review the configuration and initiate deployment

      Deployment can be monitored by vRSLCM, vCenter and SDDC manager.

      Once vRLI has been deployed, Navigate to SDDC Manager – vRealize Suite and verify vRealize Log Insight has been registered to VCF

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Verify vRealize Log Insight connection to vRealize Operations Integration

       

      1. Using a web browser navigate to vRLI master node FQDN
      2. Login as “admin”
      3. Navigate to Administration > Integration, vRealize Operations
      4. Ensure vROPs hostname and password are pointing to vROPS instance.
      5. Click Test to verify setting

       

       Graphical user interface, application</p>
<p>Description automatically generated

       

      If not already enabled, enable alert management, launch in context and metric calculation and metric calculation.

      Update content packs Navigate to Content Packs and updates as shown below.

      Graphical user interface, application, website</p>
<p>Description automatically generated

       

      Click Update All.

      Section 4 Solution Deployment guidelines.

      Deploying vSphere 7.0 with Tanzu on VCF

      vSphere with Tanzu provides the capability to create upstream compliant Kubernetes clusters within dedicated resource pools by leveraging Tanzu Kubernetes Clusters. Another advantage of vSphere with Tanzu is the ability to run Kubernetes workloads directly on ESXi hosts (vSphere Pods).

      vSphere with Tanzu brings Kubernetes awareness to vSphere and bridges the gap between IT Operations and Developers. This awareness fosters collaboration between vSphere Administrators and DevOps teams as both roles are working with the same objects.

      Graphical user interface</p>
<p>Description automatically generated with low confidence

      IT Operators continue to provision, view and monitor their virtual infrastructure as they have always done, but now with the Kubernetes awareness and insight that has eluded them in the past.

      Developers can now deploy K8s and container-based workloads directly on vSphere using the same methods and tools they have always used in the public cloud. VMware vSphere with Tanzu provides exibility as developers can choose to run pods native to ESXi (native pods) or inside purpose-build Kubernetes clusters hosted on top of namespaces congured on the vSphere clusters (Tanzu Kubernetes Clusters).

      Both teams benet by being able to use their existing tools, nobody has to change the way the work, learn new tools, or make concessions. At the same time, both teams have a consistent view and are able to manage the same objects.

      Benefits of Cloud Foundation

      Running vSphere with Tanzu on VMware Cloud Foundation (VCF) provides a best-in-class modern hybrid cloud platform for hosting both traditional and modern application workloads. VMware Cloud Foundation is a proven, prescriptive approach for implementing a modern VMware based private cloud. One of the key benets of VCF is the advanced automation capabilities to deploy, congure, and manage the full VMware SDDC software stack including products such as vSphere with Tanzu, vSAN, and NSX   among others.

       

      A picture containing graphical user interface</p>
<p>Description automatically generated

       

      Enabling vSphere with Tanzu

      In order to enable vSphere with Tanzu it is necessary to complete a set of tasks. vSphere with Tanzu will be deployed in a Virtual Infrastructure Workload Domain; however, there is also an option to deploy vSphere with Tanzu on a Consolidated VCF deployment (Management Domain). For more information about vSphere with Tanzu supportability on VCF Management Domain please refer to this Blog Post and this White Paper. An NSX-T Edge Cluster will be required as well as tasks including enabling Workload Management, creating a content library, creating a namespace, deploying harbor, obtaining CLI Tools, creating guest clusters and deploying containers.

      vSphere with Tanzu Workflow

      This is a workow overview of the procedure from a two-persona perspective (IT Operator and Developer).

      Timeline</p>
<p>Description automatically generated

      vSphere with Tanzu Requirements

      The requirements are as below; a VI workload domain needs to be created with at least three hosts, backed by an NSX-T edge cluster.

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

      vSphere with Tanzu on Consolidated Architecture Requirements

      This is a special case whereby a K8s cluster can be stood up with just four hosts in total. In order to achieve this, an NSX Edge cluster must be created for the Management domain. Application Virtual Networks (AVNs) is now supported on the management domain together with K8s. The requirements are:

       

      • Cloud Foundation 4.X deployed with one vSphere cluster on the management domain
      • NSX-T congured (edge cluster (large form factor) created, hosts added. etc.) 
      • Enough capacity on the vSAN datastore for all components

       

      NOTE: vSphere with Tanzu on consolidated architecture requires some important steps to be followed. Please refer to this document for step-by-step

      instructions: https://blogs.vmware.com/cloud-foundation/les/2020/05/VMW-WP-vSphr-KUBERNETES-USLET-101-WEB.pdf

      See this blog post for more

      information: https://cormachogan.com/2020/05/26/vsphere-with-kubernetes-on-vcf-4-0-consolidated-architecture/

      Icon</p>
<p>Description automatically generated with medium confidence
      Creating VI Workload Domain

      Creating a VI Workload Domain (VI WLD) falls in the IT Operator persona. The IT Operator will create a new VI WLD from SDDC by following the steps from the that particular POC section. However, there are a few aspects that should be taken into considerations when creating a VI WLD for vSphere with Tanzu use case.

      Note that the VI WLD for Kubernetes should be created should be using VUM (as opposed to vLCM):

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Requirements:

      • Minimum of 3 hosts; 4 or more hosts recommended Licensed for vSphere with Tanzu
      • New NSX-T Fabric
      • VI WLD with VUM enabled (no vLCM)
      • IP subnets for pod networking, service cluster, ingress and egress dened

      Graphical user interface, application, website</p>
<p>Description automatically generated

      Click HERE for a step-by-step demonstration.

      Deploying Edge Cluster

      A picture containing graphical user interface</p>
<p>Description automatically generated
       

      Deploying a NSX Edge Cluster falls in the IT Operator persona. The IT Operator will deploy a new NSX Edge Cluster from SDDC by following the steps below. After creation, the NSX Manager UI can be used to manage such cluster.

      Requirements:

      From SDDC Manager, navigate to the VI Workload Domain, and click on the three vertical dots that appear when hovering on the domain name. Choose the "Add Edge Cluster":

      • One edge cluster per domain
      • Edge cluster type =” Workload Management “
      • Two edge nodes Large form factor congured as active/active

      Graphical user interface, application</p>
<p>Description automatically generated

      Application, table</p>
<p>Description automatically generated with medium confidence

      Verify all the Prerequisites have been met and click Begin:

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Enter all the necessary information for the Edge Cluster.

      Important: make sure that there are no other T0 edge clusters connected for the overlay transport zone of the vSphere cluster

      Graphical user interface, table</p>
<p>Description automatically generated

      Ensure 'Workload Management' is set for the use-case. This is very important to enabling Tanzu

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Add the details for the rst node by lling out the information needed and clicking on 'add edge node'

        This wizard will display Edge Node details such as IP, Edge TEP.
        This can be a significant data entry exercise for both edges.

      After adding the rst node, fill out the information for the second node and click "add edge node". Click 'next' to continue:

      Graphical user interface, application</p>
<p>Description automatically generated

      Double-check the values entered in the summary section and click 'next'

      Click next in the validation section and then Finish after all status shows as succeeded.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Monitor the creation of the Edge Cluster in the Task pane of SDDC Manager.

      Graphical user interface, application, table</p>
<p>Description automatically generated

       

      Once completed, open NSX Manager UI to verify the status of the Edge Cluster.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Result:

      Graphical user interface, website</p>
<p>Description automatically generated

      Click HERE for a step-by-step demonstration.

      Enabling vSphere with Tanzu

       

      A picture containing graphical user interface</p>
<p>Description automatically generated

      The IT Operator can enable vSphere with Tanzu from SDDC Manager by following the steps below.

      Overview:

      • Deploys Workload Management from SDDC Manager
      • Domain and Edge Cluster Validation
      • Hand-o to vSphere Client
      • Installs Kubernetes VIBs on hosts Deploys ’Supervisor’ Pods
      • Instantiates Pod Service

       

      In SDDC Manager click on the "Solutions" section, click on "Deploy":

      Graphical user interface, application</p>
<p>Description automatically generated

      Verify that all the prerequisites have been met, and click "Begin":

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Select the VI Workload Domain and the cluster within the VI WLD to be used for vSphere with Tanzu, then click Next.
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      After Validation is Successful, click Next.

      Review the input information, then click "Complete in vSphere" to go to vCenter to add the remaining information. This button will take you directly to the appropriate location in the correct vCenter server.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      In vCenter UI, select the cluster and click Next.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Select the size for the Control Plane VMs. Click Next.

      Enter the information for K8s ingress and egress, and the management network for the control plane that corresponds to the diagram below.

      Diagram</p>
<p>Description automatically generated

      Graphical user interface, application</p>
<p>Description automatically generated

      Select the Storage where the Control Plane VMs will live. If using vSAN, you are able to select the Storage Policy.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Review all the information for accuracy and click Finish.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Monitor for success in the task pane.

      Once completed, the Supervisor Control Plane VMs will be visible under the Namespaces Resource Pool

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Result:

      Graphical user interface, application</p>
<p>Description automatically generated

      Click HERE for a step-by-step demonstration.

       

      Creating Content Library

      A picture containing graphical user interface</p>
<p>Description automatically generated

       

      IT Operator Persona

      Before creating namespaces, the IT Operator needs to congure a content library. A subscribed or local content library needs to be created on each Supervisor Cluster. For Tanzu Kubernetes, create a content library with the subscription pointing to:

      https://wp-content.vmware.com/v2/latest/lib.json

      To create the Content Library, simply navigate to the Content Library section of the vSphere Client to congure the content library. From vSphere Client, select Menu > Content Libraries > Create
      Provide a Name for the Content Library, and the correct vCenter server. Click Next.
      Choose "Subscribed content library" and provide the subscription URL to be used. Click Next.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      You may get a certicate warning from the subscription source. Click yes if you trust the subscription host.

      Select the storage to be used. Click Next.

      Graphical user interface</p>
<p>Description automatically generated

      Then click Finish to create the Subscribed Content Library.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Graphical user interface, application, website</p>
<p>Description automatically generated

      Result:

      Graphical user interface</p>
<p>Description automatically generated

       

      Creating Namespace

      A picture containing graphical user interface</p>
<p>Description automatically generated

       

      IT Operator Persona

      vSphere with Tanzu introduces a new object in vSphere called a Namespace.

      A namespace sets the resource boundaries where vSphere Pods and Tanzu Kubernetes clusters created by using the Tanzu Kubernetes Grid (TKG) Service can run. When initially created, the namespace has unlimited resources within the Supervisor Cluster. As a vSphere administrator, you can set limits for CPU, memory, storage, as well as the number of Kubernetes objects that can run within the namespace. A resource pool is created for each namespace in vSphere. Storage limitations are represented as storage quotas in Kubernetes.

      To provide access to namespaces, as a vSphere administrator you assign permission to users or user groups available within an identity source that is associated with vCenter Single Sign-On.

      To create namespace, navigate to Workload Management and select the Namespaces tab. Steps:

      In vCenter, navigate to Menu > Workload Management and click Create Namespace.
       

      Graphical user interface, text, application, Teams</p>
<p>Description automatically generated

       

       

      Select the cluster where the namespace will be created and provide a name for the Namespace. Click Create

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Result:

      Graphical user interface, text, email, website</p>
<p>Description automatically generated

      Graphical user interface</p>
<p>Description automatically generated with medium confidence
      Click HERE for a step-by-step demonstration.

       

      Enable Harbor Registry

       

      A picture containing graphical user interface</p>
<p>Description automatically generated

       

       

      IT Operator Persona

      Along with the content library, we must also enable a private image registry on the Supervisor Cluster. DevOps engineers use the registry to push and pull images from the registry as well as deploy vSphere Pods by using these images. Harbor Registry stores, manages, and secures container images.

       

      From the vSphere Cluster, navigate to Conguration and scroll down to Harbor Registry. Simply click the link to enable the harbor registry.

      Click on the vSphere with Tanzu enabled Cluster, select Congure. Under Namespaces, select Image Registry.

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Click Enable Harbor and select the Storage for the Image Registry. The new Harbor Registry will be visible under Namespaces

      Graphical user interface, application, email, website</p>
<p>Description automatically generated

      Result:

      Graphical user interface, website</p>
<p>Description automatically generated
      Click HERE for a step-by-step demonstration.

      Kubernetes CLI Tools

      Icon</p>
<p>Description automatically generated

       

      Developer Persona

      The previous steps from this section of the POC Guide has allowed for a successful deployment and conguration of vSphere with Tanzu. The previous steps were conducted by an IT Operator; however, this step involves the developer side of tasks to complete in order to utilize the deployed environment.

      The namespace has already been created and it is ready to be passed on to the developer by simply providing the name of the namespace along with the Kubernetes Control Plane IP address.

      The developer will be able to access the Control Plane IP address to download the vSphere CLI plugin along with the Docker Credential Helper. This plugin allows the developer to login to the Kubernetes environment and to deploy and manage workloads.

      The link to the CLI Tools can be obtained from the vSphere Client by clicking on the namespace previously created. The link can be copied and provided to the developer or can be opened from the UI.

      Graphical user interface, application, email, website</p>
<p>Description automatically generated

      Select the operating system being used and follow the steps provided to install the kubectl and kubectl-vsphere commands.

      Graphical user interface, text</p>
<p>Description automatically generated

      You can open a terminal window from this location to execute the commands

      Graphical user interface, text</p>
<p>Description automatically generated

       

       

      Deploying Tanzu Kubernetes Cluster (TKG)

      Icon</p>
<p>Description automatically generated

      Developer Persona

      Developers will usually start by deploying a Tanzu Kubernetes (TKG cluster). A Tanzu Kubernetes Cluster is a full distribution of the open-source Kubernetes that is easily provisioned and managed using the Tanzu Kubernetes Grid Service.  Note that TKGs   provides an “opinionated” implementation of Kubernetes optimized for vSphere and supported by VMware.

      Note that there are two Kubernetes environments. The Pod Service, which hosts “native pods” and the TKC cluster with the vSphere optimized Kubernetes pods.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Use the kubectl-vsphere binary downloaded in the previous step to login to the supervisor cluster, e.g.

      kubectl-vsphere login --server <supervisor-cluster IP> --insecure-skip-tls- verify

      Username: administrator@vsphere.local Password:

      Logged in successfully.

      You have access to the following contexts: 172.16.69.1

      mgmt-cluster ns01

      If the context you wish to use is not in this list, you may need to try logging in again later, or contact your cluster administrator.

      To change context, use `kubectl config use-context <workload name>`

       

       

      Here we see that the namespaces 'mgmt-cluster' and 'ns01' are available. We can see the list of nodes, etc. using the standard K8s commands, e.g.

      $ kubectl get nodes

      NAME    STATUS  ROLES   AGE     VERSION

      421923bfdd5501a22ba2568827f1a954        Ready   master  4d21h   v1.16.7-2+bfe512e5ddaaaa

      4219520b7190d95cd347337e37d5b647        Ready   master  4d21h   v1.16.7-2+bfe512e5ddaaaa

      4219e077b6f728851843a55b83fda918        Ready   master  4d21h   v1.16.7-2+bfe512e5ddaaaa

      dbvcfesx01.vsanpe.vmware.com    Ready   agent   4d20h   v1.16.7-sph-4d52cd1

      dbvcfesx02.vsanpe.vmware.com    Ready   agent   4d20h   v1.16.7-sph-4d52cd1

      dbvcfesx03.vsanpe.vmware.com    Ready   agent   4d20h   v1.16.7-sph-4d52cd1

      dbvcfesx04.vsanpe.vmware.com    Ready   agent   4d20h   v1.16.7-sph-4d52cd1

       

      Here we see our three K8 master VMs and the four ESXi servers as agents. At the time of writing, the supervisor cluster runs K8s version 1.16.7

      To get a list of contexts, we can run the following:

      $ kubectl config get-contexts

      CURRENT NAME CLUSTER AUTHINFO NAMESPACE

      * 172.16.69.1 172.16.69.1

      wcp:172.16.69.1:administrator@vsphere.local mgmt-cluster 172.16.69.1

      wcp:172.16.69.1:administrator@vsphere.localmgmt-cluster ns01 172.16.69.1

      wcp:172.16.69.1:administrator@vsphere.local ns01

      Switch to the appropriate context. In this case, 'tkg-guest':

      $ kubectl config use-context ns01 Switched to context "ns01".

      We can see the storage classes by using the following command - in this case we are using vSAN so we can see the default SPBM policy mapped to the storage class:

      $ kubectl get sc

      NAME PROVISIONER AGE

      vsan-default-storage-policy csi.vsphere.vmware.com 4d21h

      Next, we construct a manifest to create the TKG guest cluster - for more details on the various parameters, see https://docs.vmware.com/en/VMware-vSphere/7.0/vmware-vsphere-with-kubernetes/GUID-360B0288-1D24-4698-A9A0-5C5217 C0BCCF.html.

      Create a new le, e.g.

      vi tanzu-deploy.yaml

      First, we need the api-endpoint. At the time of writing it is:

      apiVersion: run.tanzu.vmware.com/v1alpha1 #TKGAPI endpoint

      Next, we set the 'kind' parameter correctly:

      kind: TanzuKubernetesCluster #required parameter

      And then set the name and namespace:

      metadata:

      name: tkgcluster1 namespace: tkg-guest

      Here we set the K8s version to v1.16:

      spec:

      distribution: version: v1.16

      Then we set the topology; rst the controlPane:

      topology:

      controlPane:

      count: 1

      Next, we define the VM class for the Tanzu supervisor cluster. We can see the available classes by using the command:

      $ kubectl get virtualmachineclasses
      NAME AGE
      best-effort-large 4d21h
      best-effort-medium 4d21h
      best-effort-small 4d21h
      best-effort-xlarge 4d21h
      best-effort-xsmall 4d21h
      guaranteed-large 4d21h
      guaranteed-medium 4d21h
      guaranteed-small 4d21h
      guaranteed-xlarge 4d21h
      guaranteed-xsmall 4d21h

      The recommended class is 'guaranteed-small', thus:

      class: guaranteed-small

      Finally, we dene the storage class:

      storageClass: vsan-default-storage-policy

       

      Then we dene the topology for the worker nodes. We create three workers using the same settings as above,

      workers:

      count: 3

      class: guaranteed-small

      storageClass: vsan-default-storage-policy

      Putting it all together, we have:

      apiVersion: run.tanzu.vmware.com/v1alpha1 kind: TanzuKubernetesCluster

      metadata:

      name: tkgcluster1 namespace: tkg-guest

      spec:

      distribution:

      version: v1.16 topology:

      controlPlane: count: 1

      class: guaranteed-small

      storageClass: vsan-default-storage-policy workers:

      count: 3

      class: guaranteed-small

      storageClass: vsan-default-storage-policy

      We can then apply this manifest to create the deployment:

      $ kubectl apply -f tanzu-deploy.yaml

      To monitor we can use the following commands:

      $ kubectl get tkc

      NAME CONTROL PLANE WORKER DISTRIBUTION AGE PHASE

      tkgcluster1 1 3 v1.16.8+vmware.1-tkg.3.60d2ffd 3m7s creating

      $ kubectl describe tkc

      Name: tkgcluster1 Namespace: tkg-guest Labels: <none>
      Annotations: kubectl.kubernetes.io/last-applied-configuration:

      {"apiVersion":"run.tanzu.vmware.com/v1alpha1","kind":"TanzuKubernetesCluster","m etadata":{"annotations":{},"name":"tkgcluster1","namespace...

      API Version: run.tanzu.vmware.com/v1alpha1 Kind: TanzuKubernetesCluster Metadata:

      Creation Timestamp: 2020-06-01T15:54:52Z Finalizers:
      tanzukubernetescluster.run.tanzu.vmware.com Generation: 1
      Resource Version: 2569051

      Self Link: /apis/run.tanzu.vmware.com/v1alpha1/namespaces/tkg- guest/tanzukubernetesclusters/tkgcluster1
      UID: 51408c8e-9096-4139-b52d-ff7e74547e39

      Spec:

      Distribution:

      Full Version: v1.16.8+vmware.1-tkg.3.60d2ffd Version: v1.16

      Settings:

      Network:

      Cni:
      Name: calico Pods:

      Cidr Blocks:
      192.168.0.0/16

      Service Domain: cluster.local Services:

      Cidr Blocks:
      10.96.0.0/12

      Topology:

      Control Plane:
      Class: guaranteed-small
      Count: 1
      Storage Class: vsan-default-storage-policy Workers:
      Class: guaranteed-small
      Count: 3
      Storage Class: vsan-default-storage-policy Status:
      Addons:
      Cloudprovider:
      Name:
      Status: pending Cni:
      Name:
      Status: pending Csi:
      Name:
      Status: pending Dns:
      Name:
      Status: pending Proxy:
      Name:
      Status: pending Psp:
      Name:
      Status: pending Cluster API Status:
      Phase: provisioning Node Status:
      tkgcluster1-control-plane-tdl5z: pending tkgcluster1-workers-lkp87-7d9df77586-9lzdj: pending tkgcluster1-workers-lkp87-7d9df77586-kkjmt: pending tkgcluster1-workers-lkp87-7d9df77586-vqr6g: pending
      Phase: creating
      Vm Status:
      tkgcluster1-control-plane-tdl5z: pending tkgcluster1-workers-lkp87-7d9df77586-9lzdj: pending tkgcluster1-workers-lkp87-7d9df77586-kkjmt: pending tkgcluster1-workers-lkp87-7d9df77586-vqr6g: pending
      Events: <none>

      Under the namespace, the TKC cluster will now be visible

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Navigating to the namespace via  vCenter ( Menu > Workload Management ns01 >  Tanzu Kubernetes)  shows the newly created tkg cluster.

      Graphical user interface, application</p>
<p>Description automatically generated

      Result:Diagram</p>
<p>Description automatically generated

       

      Click HERE for a step-by-step demonstration.

       

      Deploying Containers in TKG

       

       

      Icon</p>
<p>Description automatically generated

       

      Developer Persona

      Once the Tanzu Kubernetes Cluster has been deployed, the developer will manage it just like any other Kubernetes instance. All the Kubernetes and vSphere features and capabilities are available to the developer.
       

      We can now login to the TKG cluster using the following command:

      $ kubectl-vsphere login --server=<ip> --insecure-skip-tls-verify --tanzu- kubernetes-cluster-namespace=<namespace> --tanzu-kubernetes-cluster-name=<tkg cluster>

      In our case,

      $ kubectl-vsphere login --server=https://152.17.31.129 --insecure-skip-tls- verify --tanzu-kubernetes-cluster-namespace=ns01 --tanzu-kubernetes-cluster- name=tkgcluster1

      Username: administrator@vsphere.local Password:

      Logged in successfully.

      You have access to the following contexts: 152.17.31.129
      ns01 tkgcluster1

      If the context you wish to use is not in this list, you may need to try logging in again later, or contact your cluster administrator.
      To change context, use `kubectl config use-context <workload name>`

      We can see that the context 'tkgcluster1', i.e., our TKG cluster is now available. To switch to this context, we issue the command:

      $ kubectl config use-context tkgcluster1

      Now we can issue our usual K8s commands on this context. To see our TKG nodes, we issue the command:

      $ kubectl get nodes

      NAME    STATUS  ROLES   AGE     VERSION
      tkgcluster1-control-plane-g5mgc Ready   master  76m     v1.16.8+vmware.1
      tkgcluster1-workers-swqxc-c86bf7684-5jdth       Ready   <none>  72m v1.16.8+vmware.1
      tkgcluster1-workers-swqxc-c86bf7684-7hfdc       Ready   <none>  72m v1.16.8+vmware.1
      tkgcluster1-workers-swqxc-c86bf7684-g6vks       Ready   <none>  72m v1.16.8+vmware.1

      At this point the developer can deploy application workloads to Tanzu Kubernetes clusters using Kubernetes constructs such as pods, services, persistent volumes, stateful sets, and deployments.

      Graphical user interface, website</p>
<p>Description automatically generated
       

      Deploying Workload Domain with vVOLs

       

      As of VMware Cloud Foundation 4.1, vSphere Virtual Volumes is an applicable option when using storage arrays which support the vSphere Virtual Volume feature.
      For an introduction for Virtual Volumes concepts, please review https://docs.vmware.com/en/VMware-vSphere/7.0/com.vmware.vsphere.storage.doc/GUID-EE1BD912-03E7-407D-8FDC-7F596E41A8D3.html

      In this validation exercise we are going to validate that we can deploy a workload domain with vVOL enabled primary storage

      Requirements

      • Management Domain deployed 
      • vVOL enabled storage
      • at least three spare hosts to configure a new workload domain

       

      Success Criteria

      Workload Domain should be deployed via SDDC manager and vVOL storage is connected automatically to workload cluster.

      Note: Please refer to official documentation for detailed steps. Official documentation should supersede if it differs from guidance documented here

      To prepare for vVOL, follow these guidelines. For additional information, contact your storage vendor.

      • VMware Cloud Foundation Management Domain deployed with VCF/Cloud Builder version 4.1 or later
      • The storage system must be listed on the HCL for vVOLs support https://www.vmware.com/resources/compatibility/search.php?deviceCategory=vvols
      • The storage system or storage array that you use must support vVOL and integrate with the vSphere components through vSphere APIs for Storage Awareness (VASA). The storage array must support thin provisioning and snapshotting.
      • The vVOL storage provider must be deployed.
      • The following components must be configured on the storage side:
        • Protocol endpoints
        • Storage containers
        • Storage profiles
      • Make sure to follow appropriate setup guidelines for the type of storage you use, Fiber Channel, iSCSI, or NFS. If necessary, install and configure storage adapters on your ESXi hosts.
      • If you use iSCSI, activate the software iSCSI adapters on your ESXi hosts. Configure Dynamic Discovery and enter the IP address of your vVOL storage system

      You must configure VMware APIs for Storage Awareness (VASA) provider to manage vVOL storage. VASA provider delivers information from the underlying storage container.

      Three or more ESXi hosts with the following characteristics:

      • vVOL enabled external storage array
      • Fiber Channel / iSCSI or NFS connectivity 
      • ESXi version 7.0.1 or above
      • Hosts should not have any shared VMFS datastores attached
      • vVOL and ESXi hosts must use the same NTP time source

       

      Preparing SDDC Manager for Workload Domains using vVOLs

      The process of building a Workload Domain using vVOLs is as follows

      • Register Storage Array VASA Provider details in SDDC Manager
      • Prepare Storage array for protocol endpoint connectivity (FC, iSCSI or NFS) to ESXi hosts
      • Create network pool with appropriate NFS or iSCSI network settings
      • Commission ESXi Hosts within SDDC Manager for Workload Domain
      • Create the Workload Domain with the vVOL Storage Type

      Register Storage Array VASA Provider details in SDDC Manager

      The following details are required before adding VASA provider to SDDC manager

      Field

      Description

      Name

      Descriptive name of VASA provider

      URL for the VASA provider

      This is the VASA URL a provided by the vVOL Storage array. This detail must be provided by storage admin team

      Username / Password

      Username / Password for vVOL array. This must be provided by storage admin team

      vVOL container name

      This is dependent on a container created on the vVOL storage array. This must be provided by storage admin team

      Container Type

      This is dependent on the storage array protocol FC / NFS or iSCSI can be selected here. This must be verified by storage admin team

      Note: - For the purposes of this guide, we are using a generic “non-vendor specific” vVOL array with NFS protocol for Protocol and thus Protocol endpoints. This is simply to understand the concepts and tasks required.

      Please contact your Storage vendor documentation for specific details pertaining to vendor specific details

      VASA Storage Providers can be added and managed in SDDC Manager by going to the Administration > Storage Settings menu.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Within the Storage Settings menu, the Select +ADD VASA PROVIDER.

      Enter in Name, VASA URL, username and password, protocol (FC, iSCSI or NFS) and container name.

      Graphical user interface, application</p>
<p>Description automatically generated
       

      Once Added, the VASA Provider will be listed under storage settings.

      There is an option to Edit or delete the VASA provider.

      Additionally, multiple containers can be defined on Storage Settings

      This means that an administrator can add more than one container.

      In the screenshot below we show an example of two containers but with different protocols (NFS and iSCSI)

      Depending on the workload domain and storage connectivity chosen, during workload domain creation coke or pepsi can be used.

      This can be highly advantageous to pre-define VVOL containers beforehand as it may aid to automation tasks when deploying multiple workload domains to same storage array.

      Pepsi or Coke can be assigned to different workload domains.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      It’s also worth noting, more storage containers can be added or removed after initial definition on SDDC manager.
       

      Create network pool

      Ensure the Network Pool on SDDC Manager is configured for NFS or iSCSI if vVOL storage is NFS or iSCSI based.

      Since we are using NFS based vVOL storage we need to ensure the network pool has NFS configured.

      From SDDC Manager, navigate to Administration, Network Settings.
      Edit or create a Network pool and ensure NFS or iSCSI settings are present.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Commission ESXi Hosts within SDDC Manager for Workload Domain

      When commissioning hosts for vVOL enabled storage ensure hosts to be used for

      • vVOL FC workload domain is associated with NFS or VMOTION only enabled network pool.
        • NFS is not a hard requirement
      • vVOL NFS workload domain is associated with NFS and VMOTION only enabled network pool.
      • vVOL iSCSI workload domain is associated with iSCSI and VMOTION only enabled network pool.

      In our worked example we are using a network pool called wld01-np which uses NFS and VMOTION. Below screenshot depicts a host being commissioned for vVOLs based on NFS based connectivity. Ensure Storage type is vVOL, protocol is NFS and Network pool has NFS IP settings.

      Graphical user interface, application</p>
<p>Description automatically generated

      A json file can also be created to perform a bulk import

      Here is an equivalent json file that imports three hosts with vVOL storage using NFS as the PE protocol.
       

      {

          "hostsSpec": [

              {

                  "hostfqdn": "esxi-51.vcf.sddc.lab",

                  "username": "root",

                  "storageType": "VVOL",

                  "password": "<password>",

                  "networkPoolName": "wld01-np",

                  "vVolStorageProtocolType": "NFS"

              },

              {

                  "hostfqdn": "esxi-52.vcf.sddc.lab",

                  "username": "root",

                  "storageType": "VVOL",

                  "password": "<password>",

                  "networkPoolName": "wld01-np",

                  "vVolStorageProtocolType": "NFS"

              },

      {

                  "hostfqdn": "esxi-53.vcf.sddc.lab",

                  "username": "root",

                  "storageType": "VVOL",

                  "password": "<password>",

                  "networkPoolName": "wld01-np",

                  "vVolStorageProtocolType": "NFS"

              }

          ]

      }

      Once all hosts are validated click commission to initiate host commissioning.

       

      Graphical user interface, text</p>
<p>Description automatically generated

       

      Table</p>
<p>Description automatically generated

       

      Create the Workload Domain with the vVOLs Storage Type

      From SDDC Manager, select Workload Domains from Inventory menu, click on + Workload Domain, then select, VI -Workload Domains.

      Select vVOL based workload domain for storage selection.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      For vCenter add FQDN. If DNS is configured correctly, then associated IP addresses will be discovered.
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Add NSX-T VLAN overlay, cluster IP and controller IPs.

      Graphical user interface</p>
<p>Description automatically generated

       

      For Virtual Volumes details, we can specify the VASA (storage array) and protocol (NFS, FC, or iSCSI) to associate the Workload Domain vVOL datastore.

      All entries can be selected from drop down list

      Datastore can be entered freehand and is required. Datastore name must have a minimum of 3 characters and a max of 80 characters.

      Below is an example based on NFS based generic vVOL array.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      As mentioned previously storage containers can be added as required on Storage Settings on SDDC manager.
      Add hosts, minimum number of hosts are three for a valid workload domain.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Add license keys for NSX-T and vSphere and click next.

      Review object names for infrastructure and vSphere object names.

      Table</p>
<p>Description automatically generated

       

      Review the summary, paying attention to vVOL storage section.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Once all entries have been validated, click Finish to submit.

      Once the deployment succeeds the workload domain should be active with vVOL storage summary displayed.

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Verification of vVOL storage

      Navigate to the new vVOL backed workload domain in vCenter, from the left-hand navigation pane, Select Datastores and vVOL backed datastore.

      Select Configure. Review summary and protocol endpoints

      ESXi hosts use a logical I/O proxy, called the protocol endpoint, to communicate with virtual volumes and virtual disk files that virtual volumes encapsulate. ESXi uses protocol endpoints to establish a data path on demand from virtual machines to their respective virtual volumes.
      More details here, please review https://docs.vmware.com/en/VMware-vSphere/7.0/com.vmware.vsphere.storage.doc/GUID-EE1BD912-03E7-407D-8FDC-7F596E41A8D3.html

      Since this vVOL datastore we are interested is using NFS, i.e., vVOL container is NFS protocol based, Protocol Endpoints (PE) will be NFS mount points.

      If iSCSI or FC based a PE would be a SCSI device that would need to be mapped or presented beforehand.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      If PE’s are not configured correctly the vVOL datastore will display status of inaccessible to the cluster

      Storage Providers
      For entities represented by storage (VASA) providers, verify that an appropriate provider is registered. After the storage providers are registered, the VM Storage Policies interface becomes populated with information about datastores and data services that the providers represent.

      • Login to workload domain vCenter
      • Navigate to workload domain vCenter > Configure > Storage Providers
      • Verify Storage provider is registered and online.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      vVOL and vSphere HA

      When vSphere HA is enabled a configuration, Virtual Volume is created on each datastore by vCenter Server. In these containers, vSphere HA stores the files it uses to protect virtual machines. vSphere HA does not function correctly if you delete these containers. Only one container is created per Virtual Volume datastore.

      Additionally, vSphere Cluster Services (vCLS) is enabled when you deploy vSphere 7.0 Update 1.

      vCLS uses agent virtual machines to maintain cluster services health. The vCLS agent virtual machines (vCLS VMs) are created when you add hosts to clusters. Up to three vCLS VMs are required to run in each vSphere cluster and are deployed on the vVOL backed datastore.

      Observe below screenshot where we see the vCLS config namespaces on the vVOL datastore.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Virtual Volumes and Storage Policies

      For Virtual Volumes, VMware provides a default storage policy that contains no rules or storage requirements, called VVOL No Requirements Policy. This policy is applied to the VM objects when you do not specify another policy for the virtual machine on the Virtual Volumes datastore.

      However, an administrator can create a new vVOL based Storage policy called vVOL-Storage-Policy.

      • Open the Create VM Storage Policy wizard.
      • Click Menu > Policies and Profiles > Under Policies and Profiles, click VM Storage Policies.
      • Click Create VM Storage Policy.

      Ensure the correct vCenter server that manages the vVOL based datastore is selected

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Depending on the specific storage implementation capabilities will be exposed during VM Storage Policy Configuration. On the Policy structure page under Datastore specific rules, enable rules for a target storage entity, such as Virtual Volumes storage.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Note: Above is a sample representation on what a vVOL capabilities might be exposed from a vVOL array. This document purposely keeps this generic and vendor agnostic.

      Please review to your specific vendor documentation for details.

      To continue with this generic example, Add the specific rule sets exposed by the storage array.

      We see below that the array is advertising the capabilities of QoS of some description.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Verify compatible vVOL based datastores based on the policy rules are available for selection and complete wizard.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Complete the wizard to save the policy.

      Next, we will move on to creating a VM with the policy we just created.

      We will now create a simple VM called Test-VM and specify vVOL based Storage Policy called vVOL-Storage-Policy

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Verify VM is created correctly and is able to power on, browse to the vVOL based datastore to review VM config volumes.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Create a VM snapshot

      With vVOLs, when you right-click on a VM and choose take snapshot, unlike traditional snapshots, a redo-based delta file is not created, but instead vSphere entirely offloads this process to the array. So, the array creates the snapshots while vSphere tracks them.

      Virtual Volumes and snapshot offload integration should be reviewed from storage array vendor management tools. Please review to your specific vendor documentation for details.

      Stretching VCF Management and Workload Domains

      Depending on the POC requirements we may want to show how to stretch a VCF management and workload domain to take advantage for Stretched Cluster topology.

      In this scenario we want to understand process to first stretch the management domain and subsequent workload domain and the integration with SDDC manager and associated workloads

      Pre-requisites

      • VCF Management Domain 4.x deployed
      • Enough hosts to stretch across two availability zones
      • Adequate networking infrastructure

      Success criteria

      An administrator should be able to reconfigure an existing management and workload domains to take advantage of stretch cluster topology and associated high availability benefits.

      Note: Please refer to official documentation for detailed steps. Official documentation should supersede if it differs from guidance documented here.

      Stretched Cluster Prerequisites

      The process of stretching Cloud Foundation workload domains initiates a vSAN stretched cluster task. Rather than running this task within a managed vSAN cluster, this process is initiated by SDDC Manager, allowing SDDC Manager’s aware of this topology change. The same prerequisites that apply to vSAN stretched clusters also apply to Cloud Foundation stretched clusters.

      Stretching Cloud Foundation workload domains allows for the extension of a domain across two availability zones (AZ) running on distinct physical infrastructure. Although there is no distance limitation, key requirements include:

      • Latency below 5ms round trip time (RTT) between each availability zone
      • At least 10Gbps of bandwidth between availability zones

      Additionally, prior to stretching a cluster in a VI workload domain, the management domain cluster must be stretched rst. vCenter Servers for all workload domains are hosted within the management domain. Hence, the management domain must be stretched to protect against availability zone failure, ensuring that supporting SDDC components may continue to manage the workload domain

      Each stretched cluster requires a vSAN witness appliance in a third location. The witness should not share infrastructure dependencies with either availability zone; deployment of the witness to either availability zone it is associated with is not supported. The maximum latency between the vSAN witness appliance and the vSAN hosts is 200ms round trip time (RTT). This appliance is currently not part of SDDC Manager workows; it should be deployed manually and upgraded separately from SDDC LCM process. TCP and UDP ports must be permitted for witness trac between the witness host and the vSAN cluster data nodes; see KB article 52959.

      An in-depth list of requirements may be found on the Deployment for Multiple Availability Zones” document; please review this document prior to any attempt to stretch Cloud Foundation workload domains.

      Each AZ must have an equal number of hosts in order to ensure sucient resources are available in case of an availability zone outage.

      License Verification

      Prior to stretching VCF workload domains, please verify that licenses are not expired and that the correct license type for each product is entered within SDDC Manager.

      vSAN Licensing

      Stretching a workload domain in VCF requires that vSAN Enterprise or Enterprise Plus licensing is present within SDDC Manager in order to stretch vSAN clusters.

      VLAN Configuration Requirements

      The management VLAN, vSAN VLAN, and vMotion VLAN must be stretched between each availability zone. VLAN IDs must be identical at each availability zone.

      Table</p>
<p>Description automatically generated

      VLAN Configuration Requirements

      The management VLAN, vSAN VLAN, and vMotion VLAN must be stretched between each availability zone. VLAN IDs must be identical at each availability zone.

      Availability Zone Network Configurations

      Each availability zone must have its own vSAN, vMotion and VXLAN VLAN networks.

      Any VMs on an external network must be on an NSX virtual wire. If they are on a separate VLAN, that VLAN must be stretched as well.

      Table</p>
<p>Description automatically generated

       

      L3 Routing for vSAN

      vSAN Witness management and vSAN Witness traffic may utilize Layer 3 networks. Additional configuration may be required such as Witness Traffic Separation (WTS) and well as static routing. Please consult https://core.vmware.com/resource/vsan-stretched-cluster-guide for further details.

      Stretching Workload Domains

      The Management workload domain must be stretched prior to stretching any VI workload domains. The vCenter servers for each workload domain are placed within the management domain cluster. Therefore, the management domain must be protected against availability zone failures to ensure management of the workload domains remains available.

      After the Management workload domain has been successfully stretched, it is possible to apply stretched cluster congurations to other VI workload domains that are managed by the Cloud Foundation instance. The process of stretching VI workload domains is the same as the process that was previously used to stretch the Management workload domain.
       

      Network Pool Creation

      Prior to stretching the management domain, a network pool must be created for vMotion and storage networks.
      The subnet in a network pool cannot overlap the subnet of another pool. IP ranges cannot be edited after the network pool has been created, so please ensure the correct IP address range is entered.

      To create the network pool:

      From SDDC Manager Dashboard, click Administration, then Network Settings

      Click ‘+ Create Network Pool’

      A picture containing application</p>
<p>Description automatically generated
      Enter a name for the network pool Select the storage network type

      Provide the following information for vMotion and the selected storage network type

      • VLAN ID between 1-4094
      • MTU between 1500-9216
        Note: Make sure any physical switch trac overhead is accounted for
      • In the Network eld, enter IP address cidr, subnet mask and IP address
      • Enter an IP address range for hosts to be associated with this network pool

      Graphical user interface, application</p>
<p>Description automatically generated

      Commission Hosts

      Hosts are added to the Cloud Foundation inventory via the commissioning workow. Hosts may be added individually or use a JSON template to add multiple hosts at once. For additional details and requirements, refer to section 4.1.1 of the VCF Admin Guide document.

      In order to stretch the VCF management domain, hosts equivalent in number to those presently in the management domain cluster must be commissioned. These hosts will be used to construct the second availability zone (AZ2).

       

      Associate Hosts to Network Pool

      During the commissioning process, the network pool previously created for AZ2 must be associated with the hosts being provisioned for the stretched management domain cluster in AZ2.

       

      Verify Host Health

      Verify that all hosts commissioned are free of errors and are healthy prior to stretching the management domain.

      Deploy vSAN Witness

      Deploying the vSAN witness is a critical dependency supporting stretched management domains. The witness host may be a physical ESXi host, or the VMware-provided virtual witness appliance may be used (preferred). Please refer to vSAN witness information in core.vmware.com for further details.

      The vSAN witness host/appliance must be located in a third location outside of either availability zone it is associated with. Wherever the witness host/appliance is located, it should not share infrastructure dependencies with either availability zone. Due to its relatively relaxed latency requirement of 200ms RTT, the witness may even be hosted in the cloud. Witness trac may utilize either Layer 2 or Layer 3 connectivity. Note that witness trac is not encrypted, as it only contains non-sensitive metadata.
      It is important to highlight that as of the VCF 4.X releases, witness deployment and lifecycle management are currently not part of any SDDC manager workows. Therefore, the witness host/appliance must be deployed and upgraded independently from any SDDC Manager automation or management.
      Please refer to core.vmware.com for detailed instructions for deployment of the witness appliance.

      SDDC Manager Configuration

      In VCF 4.X the stretch cluster operation is completed using the API in the SDDC Manager Developer Center. To perform the stretch cluster operation, complete the following tasks.

      Retrieve the IDs of the hosts in the second network. Host IDs are retrieved by completing the following steps.

      On the SDDC Manager Dashboard, click Developer Center, then API Explorer

      1. Under the APIs for managing Hosts, click GET /v1/hosts.
      2. Click Execute to fetch the hosts information.
      3. Click Download to download the JSON le.
      4. Retrieve the Hosts IDS from the JSON le for hosts.

       

      Retrieve the Cluster ID

      1. On the SDDC Manager Dashboard, click Developer Center, API Explorer
      2. Under APIs for managing Clusters, click GET /v1/clusters.
      3. Click Execute to get the JSON le for the cluster information.
      4. Click Download to download the JSON le.
      5. Retrieve the Cluster ID from the JSON le.

       

      Prepare the JSON file to trigger stretch cluster validation

      1. On the SDDC Manager Dashboard page, click Developer Center, API Explorer
      2. Under APIs for managing Clusters, click POST /v1/clusters/{ID}/validations.
      3. Under the clusterUpdateSpec, click Cluster Update Data ClusterOperationSpecValidation{…}
      4. Update the downloaded update JSON le to keep only stretch related information. Below is an example of the update JSON le.

       

      {

      “clusterUpdateSpec”: { "clusterStretchSpec": {

      "hostSpecs": [ {

      "id": "2c1744dc-6cb1-4225-9195-5cbd2b893be6", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"              },              { "id": "6b38c2ea-0429-4c04-8d2d-40a1e3559714", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"                            }, { "id": "5b704db6-27f2-4c87-839d-95f6f84e2fd0", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"                            }, { "id": "5333f34f-f41a-44e4-ac5d-8568485ab241", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"

      } ],

      "secondaryAzOverlayVlanId": 1624, "witnessSpec": {

      "fqdn": "sfo03m01vsanw01.sfo.rainpole.local", "vsanCidr": "172.17.13.0/24",

      "vsanIp": "172.17.13.201"

      }

      }

      } }

      Execute the validate stretch cluster API

       

      1. From the API explorer under APIs for managing Clusters, select POST /v1/clusters/{id}/validations.
      2. Update the Cluster UID on the ID (required) and Host UID JSON file on the ClusterOperationSpecValidation fields
      3. Click Execute to execute the Stretch Cluster Workow
      4. You will see the Validation result in the Response area.
      5. Make sure the validation result is successful, if unsuccessful, correct any errors and retry.

       

      Prepare the JSON payload to trigger stretch cluster API

      1. Under APIs for managing Clusters, click Patch /v1/clusters/{id}
      2. Under clusterUpdateSpec, click on Cluster Update Data ClusterUpdateSpec{…}
      3. Click the Download arrow to download the JSON le.
      4. Update the Downloaded Patch update JSON le to keep only stretch cluster related information. Below is an example.
      {

      "clusterStretchSpec": { "hostSpecs": [ {

      "id": "2c1744dc-6cb1-4225-9195-5cbd2b893be6", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"              }, { "id": "6b38c2ea-0429-4c04-8d2d-40a1e3559714", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"              }, { "id": "5b704db6-27f2-4c87-839d-95f6f84e2fd0", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"              }, { "id": "5333f34f-f41a-44e4-ac5d-8568485ab241", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"

      } ],

      "secondaryAzOverlayVlanId": 1624, "witnessSpec": {

      "fqdn": "sfo03m01vsanw01.sfo.rainpole.local", "vsanCidr": "172.17.13.0/24",

      "vsanIp": "172.17.13.201"

      }

      }

       

      Execute the Validate Stretch Cluster API

      1. On the SDDC Manager Dashboard page, click Developer Center, API Explorer.
      2. Under APIs for managing Clusters, POST /v1/clusters/{id}/validations.
      3. Update the Cluster UID on id(required) and Host UID JSON le on ClusterOperationSpecValidation elds.
      4. Click Execute, to execute the Stretch Cluster Workow.
      5. You should see the Validation result in the Response area.
      6. Make sure that the validation result is successful, if not, correct the errors and retry.

       

      Prepare the JSON payload to trigger stretch cluster API

      1. On the SDDC Manager Dashboard, click Developer Center, API Explorer.
      2. Under APIs for managing Clusters, click Patch /v1/clusters/{id}.
      3. Under clusterUpdateSpec, click on Cluster Update Data ClusterUpdateSpec{ ... }
      4. Click Download arrow icon, to download the Json le.
      5. Update the Downloaded Patch update Json le to keep only stretch related information, similar to the below sample (replace the actual host id/vSphere license keys)

      {

      "clusterStretchSpec": { "hostSpecs": [ {

      "id": "2c1744dc-6cb1-4225-9195-5cbd2b893be6", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX" }, { "id": "6b38c2ea-0429-4c04-8d2d-40a1e3559714", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"              }, {

      } ],

      "id": "5b704db6-27f2-4c87-839d-95f6f84e2fd0", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX" },              { "id": "5333f34f-f41a-44e4-ac5d-8568485ab241", "licenseKey": "XXXXX-XXXXX-XXXXX-XXXXX-XXXXX"

      "secondaryAzOverlayVlanId": 1624, "witnessSpec": {

      "fqdn": "sfo03m01vsanw01.sfo.rainpole.local", "vsanCidr": "172.17.13.0/24",

      "vsanIp": "172.17.13.201"

      }

      }

      }

      Execute Stretch Cluster API

      1. On the SDDC Manager Dashboard, click Developer Center, API Explorer.
      2. Under APIs for managing Clusters, click Patch /v1/clusters/{id}.
      3. Update the Cluster UID on id(required) and Host UID JSON file on ClusterUpdateSpec elds.
      4. Click Execute, to execute the Stretch Cluster Workow.
      5. You should see the task created in SDDC manager UI.

       

      Check vSAN Health

      While the cluster is being stretched, monitor the state of the task from the SDDC Manager Dashboard. When the task completes successfully, check the health of the vSAN cluster and validate that stretched cluster operations are working correctly by logging in to the vCenter UI associated with the workload domain.

      To check the vSAN Health page:

      • On the home page, click Host and Clusters and then select the stretched cluster.
      • Click Monitor > vSAN > Health
      • Click Retest

      Troubleshoot any warnings or errors

      Refresh vSAN Storage Policies and Check Compliance

      1. It is imperative to check the vSAN storage policy compliance to ensure all objects achieve a state of compliance. To check vSAN storage policies:
      2. On the vCenter home page, click Policies and Profiles > VM Storage Policies > vSAN Default Storage Policy
      3. Select the policy associated with the vCenter Server for the stretched cluster Click Monitor > VMs and Virtual Disks
      4. Click Refresh
      5. Click Trigger VM storage policy compliance check
      6. Check the Compliance Status column for each VM component Troubleshoot any warnings or errors

       

       

      Removing a Workload Domain

      Existing VI workload domains may be removed (deleted) from the SDDC Manager Dashboard.

      When workload domains are deleted, associated clusters are deleted, and associated hosts are returned to the pool of ‘free’ hosts in SDDC Manager inventory. This destroys all VMs within the workload domain and associated data stores.

      Deletion of a workload domain is an irreversible process. Use caution when deleting a workload domain.

      Workload domain deletion also removes components associated with the workload domain that are deployed within the Management workload domain, such as the associated vCenter and NSX cluster instances. However, if an NSX Manager cluster is shared with another VI workload domain, it will not be deleted. Network pools associated with a deleted workload domain will remain within SDDC Manager unless removed manually.

      Prior to deleting a workload domain, ensure the following prerequisites are met:

      • If remote vSAN datastores are mounted on a cluster in the workload domain being deleted, the deletion process will fail. You must first migrate workloads on the remote vSAN datastore to a datastore local to the cluster. Alternatively, delete the workloads on the remote vSAN datastore. Then, the remote vSAN datastore should be unmounted from the cluster(s) being deleted.
      • If any workloads / data on datastores within the workload domain cluster need to be preserved, they should be migrated elsewhere or backed up. Datastores within the workload domain are destroyed during the deletion process.
      • Workloads within the workload domain created outside of Cloud Foundation should be deleted prior to beginning the workload domain deletion process.
      • Any NSX Edge clusters hosted within the workload domain should be deleted. See KB 78635.

       

      Workload Domain Removal (Deletion) Procedure

      1. Navigate to Inventory > Workload Domains
      2. Click the vertical ellipsis (three dots) beside the workload domain to be removed, then click Delete Domain

      Graphical user interface, application, email</p>
<p>Description automatically generated

      1. Review the confirmation dialog that appears on screen
      2. When ready to proceed, enter the workload domain name as prompted, then click Delete Workload Domain on the dialog box

      1. Wait for the workload domain to be deleted (may require up to approximately 20 minutes)
      2. When the removal process is complete, verify that the deleted workload domain is removed from the domains table visible in SDDC Manager
      3. Review the changes in the Management workload domain in vCenter; note removal of vCenter and NSX Manager instances

      Graphical user interface, application</p>
<p>Description automatically generated

       

      vSphere Lifecycle Manager (vLCM) and VCF

      Depending on the POC requirements we may want to show how new vSphere Life Cycle Management capabilities integration with SDDC manager

      In this scenario we want to understand process to configure new workload domains utilizing vSphere Life cycle manager. The admin should be

      Pre-requisites

      • VCF Management Domain 4.x deployed
      • At least 3 hosts to deploy a new workload domain
      • Hardware Support Manager (HSM) integration (Optional)

      Success criteria

      An admin should be able to create import and use vLCM cluster images to SDDC manager with the view to deploy new workload domains.
       

      Note: Please refer to official documentation  on Cluster image management for detailed steps. Official documentation should supersede if it differs from guidance documented here.

      Lifecycle management refers to the process of installing software, maintaining it through updates and upgrades, and decommissioning.

      In the context of maintaining a vSphere environment, your clusters and hosts in particular, lifecycle management refers to tasks such as installing ESXi and firmware on new hosts and updating or upgrading the ESXi version and firmware when required.

      You can use cluster images as an alternative way of performing ESXi host lifecycle operations. A cluster image represents a desired software specification to be applied to all hosts in a vSphere cluster. Software and firmware updates happen simultaneously, in a single workflow.

      vSphere Lifecycle Manager enables you to manage ESXi hosts and clusters with images.

      You use vSphere Lifecycle Manager baselines and baseline groups to perform the following tasks.

      • Upgrade and patch ESXi hosts.
      • Install and update third-party software on ESXi hosts.
      • You use vSphere Lifecycle Manager images to perform the following tasks.
      • Install a desired ESXi version on all hosts in a cluster.
      • Install and update third-party software on all ESXi hosts in a cluster.
      • Update and upgrade the ESXi version on all hosts in a cluster.
      • Update the firmware of all ESXi hosts in a cluster.
      • Generate recommendations and use a recommended image for your cluster.
      • Check the hardware compatibility of hosts and clusters against the VMware Compatibility Guide and the vSAN Hardware Compatibility List.

       

      For more information refer to https://docs.vmware.com/en/VMware-vSphere/7.0/com.vmware.vsphere-lifecycle-manager.doc/GUID-74295A37-E8BB-4EB9-BFBA-47B78F0C570D.html
       

      Cluster Image and vSphere Lifecycle Manager

      Cluster images are made available by the vSphere Lifecycle Manager (vLCM), a vCenter service. This service is now integrated with VMware Cloud Foundation and enables centralized and simplified lifecycle management of ESXi hosts. When a VI workload domain or cluster is created with an image, you can update and upgrade the ESXi software on all hosts in a cluster. You can also install driver add-ons, components, and firmware on the hosts.

      Cluster Image Components

      A cluster image may consist of four elements: ESXi base image, a vendor add-on, a firmware and drivers add-on, and additional components. It is mandatory to add ESXi to a cluster image. Adding the other elements is optional.

      Cluster Images and VCF

      Cluster images must be created in vSphere 7.0 and then imported to VMware Cloud Foundation. Unlike vSphere where cluster images are managed per cluster, VMware Cloud Foundation allows you to manage all cluster images in a single place and re-use them for clusters across workload domains.

      For more information refer to the Cluster image management documentation.

      Hardware Support Manager (HSM)
      a Hardware Support Manager is a plug-in that registers itself as a vCenter Server extension. Each hardware vendor provides and manages a separate hardware support manager that integrates with vSphere.

      If you want to add firmware to the cluster image, you must install the Hardware Support Manager from your vendor. See Firmware Updates

      As of Aug 2020, you can deploy and use hardware support managers from the following vendors.

      Overview

      Cluster images are created in vSphere 7.0. You can create an image either on the management domain vCenter Server, or a vCenter Server external to VMware Cloud Foundation.

      Initial setup will be done on vCenter.

      1. Download the ESXi base image for the VCF version and upload it to vSphere Client
      2. Create an empty cluster in the vCenter Server where you want to create a cluster image. You do not need to add any hosts to this cluster.
      3. During the creation of an image, you define the ESXi version

      optionally add vendor add-ons, components, and firmware.

      1. Import the cluster image to VMware Cloud Foundation

      Note: If you want to add firmware to the cluster image, you must install the Hardware Support Manager (HSM)  from your vendor. See Firmware Updates. Adding and managing HSM is out of scope of this document

      Prepare ESXi Images

      Connect to vCenter UI and connect to  Mgmt or Workload Domain vCenter server

      Navigate to Lifecycle Manager

      Menu > Lifecycle Manager > Import ISOs

       

      Graphical user interface, text, application</p>
<p>Description automatically generated
      ESXi ISOs can be downloaded from myvware.com or from your applicable hardware vendor
      Once ESXi ISO is made available, import to vSphere ISO Depot

      Graphical user interface, text, application</p>
<p>Description automatically generated

      You now have two options to import the Cluster Image into VCF SDDC Manager

      Graphical user interface, text, application, email</p>
<p>Description automatically generated
       

      • Option 1 If an Empty Cluster is created on one of the management or workload domains , the cluster image can be imported directly into vCenter once the empty cluster is created
      • Option 2 If the Cluster image is created on a vCenter Server that is not managed by vCenter, the individual files can be exported and imported into VCF SDDC Manager

      For the purposes of this document, we will walk through Option 1
       

      Create vSphere Cluster for Cluster Image

      We now need to create an empty vSphere Cluster on either a management domain or workload domain vCenter that has an imported ISO
      The purpose of this is to associate an image with a cluster to make ready for import. As mentioned on the documentation the cluster object is temporary and does not need to have any hosts added

      Navigate to vSphere Datacenter in either management domain or workload domain where ISO have been imported

      Right Click on Datacenter object and Create New Cluster

      As per VCF Documentation call this cluster ClusterForImage.”

       

      Ensure to select “Manage all hosts in the cluster with a single image”

      Graphical user interface, website</p>
<p>Description automatically generated

      Select the applicable build of ESXi for the VCF 4.x BOM.

      If there are applicable vendor addons they maybe included during cluster creation
      If a HSM (Hardware Support Manager installed)  firmware image maybe selected at this point

      HSM guidance out of scope on this document. For the purposes of this document, we are going to use a standard vSphere ISO and agnostic server solution

      Click OK to confirm Cluster creation

      We can see now the empty cluster is created with an image associated with it

      Select Cluster and navigate to Updates .

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Importing Cluster Image to VCF

      We are now ready to import the image into SDDC manager

      Connect to the SDDC Manager webpage and navigate to Life-Cycle Management and Image Management

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Select Import Image , we are choosing Option 1 Extract a Cluster Image

      This process is very straight forward, navigate to the workload domain where the Cluster was created, select the cluster, provide a descriptive name and Extract

      Graphical user interface, text, application</p>
<p>Description automatically generated

      SDDC Manager will spawn an upload  task and this can be monitored on the task panel.

      Once image has been uploaded it will be available on Image management to use for new workload domains.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Workload Domain creation using vLCM Cluster Images

      Unlike previous versions of VCF, we now have a new option available to deploy a workload domain cluster with vLCM based images

      A few things to consider

      • Workload Management is not supported for vLCM backed VCF Clusters
      • Stretch Cluster is not supported
      • Firmware Updates are supported ,as long as HSM is configured
      • When you use this image to create a VI workload domain, NSX-T components are added to the default cluster during the domain creation
      • When you create a new domain, only the Cluster Image that matches the VCF BOM will be able to be selected

       

      Similar to creating Workload domains, the process is pretty straight forward

      From SDDC Manager, select Workload Domains from Inventory menu, Click on + Workload Domain VI -Workload Domains.

       

      For illustrative purposes we are using vSAN as the backing storage

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      As mentioned previously, there two choices for vSphere Life-Cycle Management
      Baselines(VUM) or Images (vLCM)

      Baselines are VUM based, while Images are vSphere 7.0 Cluster Image based.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Select Next to continue

      At this point name the workload domain cluster as appropriate and then select the Cluster image imported in SDDC Manager.

      Graphical user interface, website</p>
<p>Description automatically generated

      The remaining VI Workload domain deploy wizard workflow does not change from previous versions.

      The summary screen will summarize that Cluster Image is in use

      Graphical user interface, application</p>
<p>Description automatically generated

      Once Workload Domain has completed, we can see that the Image has been successfully applied to the new cluster

      To verify Login to the new domain vCenter from SDDC manager

      Select applicable workload domain cluster and login to vSphere Client

      Graphical user interface, application</p>
<p>Description automatically generated with medium confidence

      From the vSphere Client, select the vSphere Cluster and select the Updates tab

      We can see listed the ESXi version and if the hosts in the cluster are compliant or not

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      If we click on Show details of the Components within the image, we see NSX-T components have been added to the image during workload domain deployment.
      A component is the smallest unit that can be included in the image. For example, a driver is a component.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Composable Infrastructure (Redfish API) Integration

      Depending on the POC requirements we may want to show how VMware Cloud Foundation interacts with software defined Composable Infrastructure

      In this scenario we want to understand process to configure against composable architecture infrastructure.

      Pre-requisites

      • VCF Management Domain 4.2 deployed
      • HPE Synergy and Dell MX Composable Infrastructure solution deployed.

      Success criteria

      An admin should be able enable integration with various composability solutions on VCF SDDC Manager with 3rd party vendor hardware solutions.

      Beginning with version 3.8, Cloud Foundation supports integration with software defined Composable Infrastructure, allowing for dynamic composition and decomposition of physical system resources via SDDC Manager. This integration currently supports HPE Synergy and Dell MX Composable Infrastructure solutions. This integration leverages each platform’s Redfish API.

      Note: Please refer to official documentation for detailed steps. Official documentation should supersede guidance documented here.
       

      HPE Synergy Integration

      To enable infrastructure composability features, deploy the HPE OneView Connector server.

      Procedure:

      • Deploy Linux server (physical or VM)
      • Install HPE OneView connector for VCF on the Linux server
      • Complete bring-up SDDC Manager if not already done
      • Increase queue capacity for the thread pool
      • Connect to SDDC Manager via SSH using the vcf account Escalate to root privileges with sudo
      • Open the le application-prod.properties:

      vi /opt/vmware/vcf/operationsmanager/config/application-prod.properties

       

      • Update the queue capacity line

       

      om.executor.queuecapacity=300

       

      • Save and close the le
      • If using a self-signed certicate, import the Redsh certicate from the OneView Connector server: SSH in to SDDC Manager using the vcf account
      • Enter su to escalate to root
      • Import certicate from Redsh to SDDC Manager:

      /opt/vmware/vcf/commonsvcs/scripts/cert-fetch-import-refresh.sh --ip=<redfish- ip> --port=<SSL/TLS port> --service-restart=operationsmanager

      • Restart the SDDC Operations Manager service:
        systemctl restart operationsmanager
      • Wait a few minutes for the service to restart
      • From SDDC Manager, click Administration > Composable Infrastructure
      • Enter the URL for the Redsh translation layer

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Dell MX Integration

      Dell MX Composable Infrastructure does not require a separate server instance to be deployed, as the Redsh API translation layer is integrated into the MX management module.

      Certificates

      A signed certificate is necessary in order to establish a connection with the OME Modular interface. The FQDN should be added to DNS as this is included in the certicate. Note that the certicate presented by the MX platform must have a CN that matches the FQDN of the MX management module; VCF will not connect if the default self-signed certificate (CN=localhost) is used.

      The certicate CSR can be generated from the OME Modular Interface on the MX7000.

      1. Log in to the OME interface
      2. Select Application Settings from the main menu
      3. Security -> Certicates
      4. Generate a Certicate Signing Request
      5. Then upload the certicate when it is available

       

      Configure Translation Layer

      The translation layer must be congured prior to connecting the SDDC Manager to the composable infrastructure platform.

      Procedure:
      • Increase queue capacity for the thread pool.
      • Connect to SDDC Manager via SSH using the vcf account.
      • Escalate to root privileges with su
      • Open the le application-prod.properties:

         vi /opt/vmware/vcf/operationsmanager/config/application-prod.properties

      • Update the queue capacity line:

         om.executor.queuecapacity=300

      • Save and close the file
      • If using a self-signed certicate, import the Redsh certicate from the MX MSM to SDDC Manager.
      • SSH in to SDDC Manager using the vcf account
        • Enter su to escalate to root
        • Import certicate from Redfish to SDDC Manager:

      /opt/vmware/vcf/commonsvcs/scripts/cert-fetch-import-refresh.sh --ip=<MSM-ip> --
            port=<SSL/TLS port> --service-restart=operationsmanager

      • Restart the SDDC Operations Manager service:
        systemctl restart operationsmanager
      • Wait a few minutes for the service to restart
      • From SDDC Manager, click Administration > Composable Infrastructure
      • Enter the URL for the Redfish translation layer

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Click Connect

      Composable resources will now be visible within the VCF UI.

      Graphical user interface, table</p>
<p>Description automatically generated

       

      Section 5: Use Case Validation

      Validate AVN Networking and Tier 0 BGP Routing

       

      In this validation exercise we are going to validate Tier-0 BGP Routing between a tier-0 gateway and the outside infrastructure.

      We will configure a DHCP service on existing Cross Region networks used for AVN networks and deploy a VM

      Requirements

      • Management Domain deployed with AVN Networking
      • One or more virtual machines to test

      Success Criteria

      VM should be successfully connected to a segment, acquire a DHCP address, and be able to route out to external infrastructure such as a DNS or NTP server

      Method

      Configure DHCP Server Profile and associate to a AVN segment

      Connect to NSX-T Manager interface associated with Management Domain

      Navigate to Networking > Segments

      Locate segment associated to AVN cross Region Networks, e.g.  xreg-m01-seg01

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Navigate to IP Management and locate DHCP

      Create a DHCP Profile and provide a descriptive name, e.g.  “xreg-dhcp-test”

      Use the same subnet as specified on the segment for AVNS and specify the Edge Cluster

       

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

       

      Click Save once all entries have been completed

      Associate Server Profile to Segment

      Navigate to Networking Segments Edit the Segment name used for AVN, e.g., xreg-m01-seg01
       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Select SET DHCP CONFIG

      • Set DHCP Type to Local DHCP Server
      • Enter in the DHCP Server Profile specified in previous step
      • Enable DHCP config and set DHCP Server IP
      • Define DHCP Scope

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

       

      Click SAVE to apply and CLOSE EDITING

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Deploy A Virtual Machine

      Navigate to Management Domain vCenter Server

      Click Deploy a Virtual Machine

      In this case we are using an Ubuntu VM to test

      Ensure you attach the vNic to the AVN cross Region segment

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Once VM is deployed , configure the interface for DHCP

      Verify that DHCP lease has been acquired on the VM we can see here that the VM has picked up IP address 192.168.11.101

      ifconfig ens160

      ens160    Link encap:Ethernet  HWaddr 00:50:56:a7:b3:ae

                inet addr:192.168.11.101 Bcast:192.168.11.255 Mask:255.255.255.0

                inet6 addr: fe80::250:56ff:fea7:b3ae/64 Scope:Link

                UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

                RX packets:779 errors:0 dropped:0 overruns:0 frame:0

                TX packets:818 errors:0 dropped:0 overruns:0 carrier:0

                collisions:0 txqueuelen:1000

                RX bytes:79509 (79.5 KB) TX bytes:138635 (138.6 KB)

      We can also see the DHCP server IP address of 192.168.11.2

      dhclient -d -nw ens160

      Internet Systems Consortium DHCP Client 4.3.3

      Copyright 2004-2015 Internet Systems Consortium.

      All rights reserved.

      For info, please visit https://www.isc.org/software/dhcp/

      Listening on LPF/ens160/00:50:56:a7:b3:ae

      Sending on   LPF/ens160/00:50:56:a7:b3:ae

      Sending on   Socket/fallback

      DHCPREQUEST of 192.168.11.101 on ens160 to 255.255.255.255 port 67 (xid=0x22b6abe2)

      DHCPACK of 192.168.11.101 from 192.168.11.2

      RTNETLINK answers: File exists

       

      Verify routing works

      The simplest method it to use tracepath from the VM to a routable address on the external network such as a dns server

      In this case we see this go through segment (192.168.11.x) to Tier-1 Router (100.64.112.0)> TOR switch (147.168.66.1) > to external DNS (10.156.169.50)

      #tracepath 10.156.169.50

      1?: [LOCALHOST]                                         pmtu 1500

      1:  192.168.11.1                                          0.270ms asymm 64

      1:  192.168.11.1                                          0.078ms asymm 64

      2:  100.64.112.0                                          0.525ms

      3:  147.168.66.1                                          0.693ms

      4:  dns.vcf.sddc.lab                                      1.025ms reached
       

      Diagram, schematic</p>
<p>Description automatically generated

       

      To verify if DNS traffic is working a simple nslookup can be issued against the configured dns server.

      nslookup dns.vcf.sddc.lab

      Server:         10.156.169.50

      Address:        10.156.169.50#53

      Name:   dns.vcf.sddc.lab

      Address: 10.156.169.50

      NSX-T Trace Flow

      Layer 3 connectivity can also be interrogated from an NSX-T perspective using Traceflow

      Connect to NSX-T Manager, Navigate to Plan and Trouble Shoot -Trace Flow

      Select Source as type VM, choosing the Virtual Interface on the segment and select destination type as IP on the external infrastructure i.e., Layer 3

      In this case we are preforming a trace flow between the VM and an external service such as DNS

      Graphical user interface, application</p>
<p>Description automatically generated

      Once source and destination have been selected, click on trace

      We can now see the successful trace between the VM as it traverses the ESXi hosts to the Edge and out to the outside world infrastructure.

      Graphical user interface, application, table</p>
<p>Description automatically generated

      If you have deployed more than one Virtual machine on another segment, we can also look at the traceflow to interrogate east-west traffic between two segments

      Select VM on source Segment and a VM on Destination segment

      Graphical user interface, application</p>
<p>Description automatically generated

      We can now see the traffic flow between the two VMs and the path it has taken

      Graphical user interface, application</p>
<p>Description automatically generated

      Finishing up
       

      Once testing has completed and results are recorded, the DHCP configuration maybe removed if no longer necessary

       

      SDDC Manager Certificate Management

      Depending on the POC requirements Certificate management may be used to show case security integration with VCF

      In this worked example we will configure a Microsoft Certificate Authority and configure SDDC Manager to use it and replace/update certificate on major infrastructure components.

      Pre-requisites

      • VCF Management Domain deployed
      • Microsoft Certificate Authority server
        • Microsoft Service Account to request certificates

      Success criteria

      We should be able to integrate with an existing Microsoft CA and we should be able to perform an orchestrated certificate replacement of major VCF components such as

      • SDDC Manager
      • NSX-T
      • vCenter  
      • vRSLCM

      All guidance is based on Certificate management

      Out of Scope on this document is Open SSL CA authority and installing certificates with external or third-party certificate authorities
      For more information on third party please refer to the Certificate management documentation.

      Microsoft Certificate Authority server configuration guidance.

      If a Microsoft CA server is already available, below guidance maybe ignored, it is here for only for reference.

      This may be helpful in POC scenarios.

      This guide simply augments the documented process and here as reference only.

      Please refer to official documentation for detailed steps. Official documentation should supersede guidance documented here

      https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-F3D138D7-EBC3-42BD-B922-D3122B491757.html 

      In summary we need to add Certificate Authority and Certificate Authority Web Enrolment roles on a Windows Server to help facilitate automated certificate creation from SDDC Manager

       

      Below is a screenshot of the roles required to enabled on the certificate server

       

      Graphical user interface, text, application, Word</p>
<p>Description automatically generated

      To allow SDDC Manager the ability to management signed certificates, we also need to Configure the Microsoft Certificate Authority with basic authentication.

      How to achieve this task is documented here

      https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-8F72DF53-32D2-4538-90FD-3BCB01EB1811.html

      Configure the Microsoft Certificate Authority with basic authentication. With Role Enablement.
      Click Start > Run, enter Server Manager, and click OK.

      From the Dashboard, click Add roles and features to start the Add Roles and Features wizard

      This is a screenshot of the role required to enable Basic Authentication

      Graphical user interface, text, application, Word</p>
<p>Description automatically generated

      The certificate service template should now be configured for basic authentication (as well as the default web site)
       

      Ensure to Configure the certificate service template and all sites, including default web site, for basic authentication.

      This is documented here https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-8F72DF53-32D2-4538-90FD-3BCB01EB1811.html

       

      Click Start > Run, enter Inetmgr.exe and click OK to open the Internet Information Services Application Server Manager .

      Navigate to your_server > Sites > Default Web Site > CertSrv.

      Under IIS, double-click Authentication.

      On the Authentication page, right-click Basic Authentication and click Enable.

      Graphical user interface, application, Word</p>
<p>Description automatically generated

       

      Enable Basic Authentication

      Graphical user interface, application, table</p>
<p>Description automatically generated

      Create and Add a Microsoft Certificate Authority Template

       

      This process is documented here

      https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-8C4CA6F7-CEE8-45C9-83B4-09DD3EC5FFB0.html

      Attached are some further screenshots for guidance

      Once the VMware certificate is created this is an example of Windows 2016 Compatibility settings
      Ensure to remove Server Authentication from application polices

      Ensure to enable the extension under “Basic Constraints”
      Graphical user interface, application, Word</p>
<p>Description automatically generated
      Ensure to set Signature is proof of origin under “Key Usage”
      Graphical user interface, text, application</p>
<p>Description automatically generated

      Assign Certificate Management Privileges to the SDDC Manager Service Account

      This process is documented here 

      https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-592342B3-DD3A-4172-8BFC-A31A65E940DF.html

      Windows Service Account
      If not already created create a managed Service Account called, for example, svc-vcf-ca

      This is what SDDC Manager will use to request certificates.
      For more information on service accounts please review https://docs.microsoft.com/en-us/windows/security/identity-protection/access-control/service-accounts
       

      For the Certificate CA we will assign least privilege access to the Active Directory service account that SDDC Manger will use

      For the Certificate CA Template, we will also assign least privilege access

      As stated the exact steps are covered here https://docs.vmware.com/en/VMware-Validated-Design/6.1/sddc-deployment-of-the-management-domain-in-the-first-region/GUID-592342B3-DD3A-4172-8BFC-A31A65E940DF.html

      Here are some screenshots to help augment the process from the Microsoft Certificate Authority Utility certsrv.msc

      Least privilege access Microsoft Certificate Authority. (Read and Request)

      Graphical user interface, text</p>
<p>Description automatically generated

       

      Least privilege access Microsoft Certificate Authority Template (Read and Enroll)

      Graphical user interface</p>
<p>Description automatically generated

       

       

      SDDC Manager Certificate Management Procedure
       

      Step 1:

       

      Configure SDDC Manager to use Microsoft Certificate Authority.

      Please following the documentation for guidance
      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/com.vmware.vcf.admin.doc_41/GUID-B83D4273-F280-4978-B8BD-63A283F803A9.html

       

      Below is an example to configure Microsoft CA Authority

      From SDDC Manager, navigate to Administration > Security > Certificate Management
      Enter in CA server URL Service account details and Template, click save.

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Accept the certificate when prompted to complete the configuration
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Step 2:

      Generate Certificate Signing Requests ( CSR)

      In the following worked example, we will install certificates on vCenter, SDDC Manager, NSX-T Manager and vRSLCM (if installed) .

      We will install certificates on vCenter and SDDC manager as an example. This guide augments the documented process located here
      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/com.vmware.vcf.admin.doc_41/GUID-D87B98D3-9F7D-4A34-8056-673C19EDDFD2.html

      Please refer to the documentation for detailed steps, official documentation should supersede guidance documented here

      The steps per component are

      • Generate CSRS
      • Generate Signed Certificates
      • Install Certificates

       

      In this example we will first change the certificate on the vCenter Server.

      From SDDC Manager to Inventory > Workload domain > Management workload domain and select Security tab

      Select the component you wish to change the certificate; in our case we are showing vCenter server

       

      Graphical user interface, website</p>
<p>Description automatically generated

       

      Now we will generate signing request, a new popup wizard will be launched to generate CSR.
      Required details are as follows to generate a CSR request.

      • Algorithm
      • Size
      • OU
      • Org
      • Locality
      • State or Provence
      • Country

       

      Graphical user interface, application</p>
<p>Description automatically generated

      Optionally CSR request can be downloaded at this point

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

      Step 3:

      Request new certificate from Microsoft Certificate Authority Server.

      Once CSR Generation is successful, the next step to request certificate from Microsoft CA.
      Select Generate Signed Certificates, and chose Microsoft as certificate Authority
       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Step 4:
      Install Microsoft Certificate Authority Certificates on Management Domain Components

       

      Once Certificate Generation is successful, we can move on to Installing certificate, once you initiate Install Certificate, the task will be summitted by SDDC Manager immediately

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

       

      This will generate a deployment lock on SDDC manager while certificates are installed. Progress of certificate installation can be monitored from Tasks

       

      A picture containing application</p>
<p>Description automatically generated

       

      For SDDC Manager, certificate replacement,  in general the process is the same, Generate CSRS, Generate Signed Certificates and Install Certificates.

      However, there is one additional step.
      You must manually restart SDDC Manager services to reflect the new certificate and to establish a successful connection between VMware Cloud Foundation services and other resources in the management domain.

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Once certificate installation has been initiated, you may see transient errors such as failed to retrieve tasks if attempting to monitor the progress of the certificate installation.

      To manually restart services, it can be achieved by initiating a ssh session to sddc manager as user vcf then with elevated privileges (su - ) and executing the sddcmanager_restart_services.sh script
      #su -
      #sh /opt/vmware/vcf/operationsmanager/scripts/cli/sddcmanager_restart_services.sh

      Text</p>
<p>Description automatically generated

       

      NSX-T and vRCLM

      Replacing certificates such as NSX-T and vRCLM follows the same process. Verify all tasks completed ok

      Verification

      A simple way to verify is via browser for each of the SDDC components.

      Connect to SDDC Manager URL

      Click the padlock icon next to the URL. Then click the "Details" link.
       

      From here you can see some more information about the certificate and encrypted connection, including the issuing CA and some of the cipher, protocol, and algorithm information. Ensure it matches the expected details supplied

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

      We can also check for the issued certificates on Microsoft CA Authority server.

      Connect to the configured Microsoft CA server (via RDP etc.)

      From the CA Authority server, Click Start > Run, enter certsrv.msc, and click OK.

      Click on Issued Certificates, optionally filter by the requesting service account ID.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

                    
      Summary:

      This task should validate certificate replacement scenarios offered by VCF SDDC manager. The procedure should validate that certificates can be replaced in an orchestrated manner with minimum manual intervention.

      SDDC Manager Password Management

      For security reasons, you can change passwords for the accounts that are used by your VMware Cloud Foundation system. Changing these passwords periodically or when certain events occur, such as an administrator leaving your organization, reduces the likelihood of security vulnerabilities.

      Rotate Passwords

      As a security measure, you can rotate passwords for the logical and physical entities on all racks in your system. The process of password rotation generates randomized passwords for the selected accounts. [Read more]

      Manually Update Passwords

      You can manually change the password for a selected domain account. Unlike password rotation, which generates a randomized password, you provide the new password. [Read more]

      Service account passwords for deployed infrastructure components (e.g., ESXi hosts, NSX Manager) may be changed with SDDC Manager. SDDC Manager either updates passwords with a user-specied password (‘Update’ option), or automatically generates new randomly-generated passwords (‘Rotate’ option).

      For more information refer to documentation

      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/com.vmware.vcf.vxrail.admin.doc/GUID-69B141B8-AEB2-4C31-A4ED-4EB31BC6361D.html?hWord=N4IghgNiBcILJgHZgOYFMAEAFMBnXA7gPYBOAJriAL5A

      Pre-requisites

      • VCF Management Domain 4.x deployed

      Success criteria
      An administrator should be able to rotate or change service account passwords for deployed infrastructure components

      Rotate a password

      From SDDC Manager navigate to the left panel, select Security > Password Management. Then, from the drop-down menu, select the component that will have passwords updated or rotated:
       

      Graphical user interface</p>
<p>Description automatically generated

      To rotate the password with a new, randomly generated password, select the user account(s) that needs to be updated and click ‘Rotate’. This will bring up a window to confirm the change:

      Graphical user interface, text</p>
<p>Description automatically generated with medium confidence

      To update a particular password with a new user-specied password, select only one user account, and click ‘Update’:

       

      Graphical user interface, application</p>
<p>Description automatically generated

      Note:- that the SDDC Manager password must be manually updated using the passwd command.

      Passwords may be viewed by opening an SSH session to SDDC manager and issuing the following command:
      $/usr/bin/lookup_passwords

      Here is a worked example for interrogating vROPs admin passwords

       
      $/usr/bin/lookup_passwords

       

      Password lookup operation requires ADMIN user credentials. Please refer VMware Cloud Foundation Operations and Administration Guide for setting up ADMIN user.

      Supported entity types: ESXI VCENTER PSC NSXT_MANAGER NSXT_EDGE VRSLCM VRLI VROPS VRA WSA BACKUP VXRAIL_MANAGER AD

      Enter an entity type from above list:VROPS

      Enter page number (optional):

      Enter page size (optional, default=50):

      Enter Username: administrator@vsphere.local

      Enter Password:

      VROPS

      identifiers: 192.168.11.18,m01vrops.vcf.sddc.lab

      workload: m01

      username: admin

      password: ###########

      type: API

      account type: SYSTEM

      VROPS

      identifiers: 192.168.11.19,m01vropsmaster.vcf.sddc.lab

      workload: m01

      username: root

      password: ###########

      type: SSH

      account type: SYSTEM

      VROPS

      identifiers: 192.168.11.20,m01vropsreplica.vcf.sddc.lab

      workload: m01

      username: root

      password: ###########

      type: SSH

      account type: SYSTEM

        Page : 1/1, displaying 3 of total 3 entities in a page.

       

      Summary

      An admin should be able to rotate or retrieve SDDC infrastructure passwords from SDDC manager from a centralized console.

      Lifecycle Management - VCF Skip-Level Upgrade Scenarios.

      Depending on the POC requirements we may want to show upgrading VCF from one release to another while skipping various interim patches or minor updates.
      The process involves downloading and applying multiple upgrade bundles as part of a single task. In VCF 4.2 there is no need for the command line utility as you can do the skip level upgrades via UI; however, it is necessary to download all the bundles in between versions.

      For example if you are going from 4.1 to 4.2, you will need any patches, minor upgrades such as 4.1.0.1 and drift bundles as well.

      Pre-requisites

      • VCF Management Domain 4.1 deployed
      • Available Updates to apply or update to, e.g., 4.1.0.1, 4.2, etc.
      • Have a my.vmware login and successfully log in via SDDC manager in order to download bundles.

      Success criteria

      An administrator should be able to upgrade from one release to another without the need to manually perform an interim patch upgrade.

      Executing Skip Level Upgrade

      From SDDC Manager, navigate to the Domain you would like to upgrade. Keep in mind that the Management domain will need to be upgraded first before the updates become available for the Workload Domains.

      Select the updates tab and you will see available bundles.

      At this point the bundles will either show as available to deploy or available to download. Select download if they have not been downloaded yet. Remember that you must download all the bundles shown even if you are doing a skip-level upgrade, as these bundles will also be applied in the process.

      If no bundles are shown, make sure you have successfully signed in to my.vmware from SDDC Manager in order to have access to the bundles.

       

       

      Failure to download all bundles prior to upgrading will cause the upgrade to fail.

       

      Once all the bundles have been downloaded, you can select which version to go to, in this case from 4.1 to 4.2 directly.

       

       

      Monitoring Progress

      Progress can also be monitored by navigating to SDDC manager user interface i.e., https:/sddc-manager FQDN

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      The progress bar can be expanded for SDDC for more details on the individual components, once complete you can click finish

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Once the upgrade has completed, the snapshot will be removed from SDDC Manager VM

      To verify upgrade has completed, login to SDDC manager and review recent SDDC manager tasks.

      After the upgrade has completed, SDDC Manager will now show a release Versions tab on the left that will show the current versions the domains are at and the corresponding BOM for each version.

       

      Summary criteria

      After this POC exercise an administrator should be able to upgrade from one release to another without the need to manually perform an interim patch upgrade of SDDC manager

      vRealize Suite and VCF – vRealize Operations integration

       

      Depending on the POC requirements we may want to show vRealize Suite Monitoring integration with VCF SDDC infrastructure components

      In this scenario we want to verify if vROPs is monitoring and collecting the management workload domain health and statistics.

      An admin should be able to connect a Workload Domain to vROPs and be able to configure and modify various monitoring solutions.

      Pre-requisites

      • VCF Management Domain 4.x deployed
      • vRealize Operations deployed
      • Optionally an additional Workload domain deployed

      Success criteria

      An administrator should be able to verify basic vROPS integration is in place and be able to configure various solutions for vSAN and NSX-T etc.

      Verification

      Once vROPS has been deployed we need to confirm if metrics are being collected from the SDDC Management Cluster

      From SDDC-Manager navigate to vRealize Suite > vRealize Operations card

      click on the hyper-link to connect to vROPS login screen

       

       

      Once logged in, from vROPS dashboard Go to Administration > Cloud Accounts.

      We should see a “Cloud Account” which is another name for vROPS adapter configured with the name of vCenter

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      We should also see the vSAN adapter configured.

      Ensure both have status of OK and are collecting.

      Configure an existing vROPS adapter

      This is an optional step but allows to show you may configure or edit an existing vROPS adapter.

      We will show an example of this by enabling vSAN SMART Monitoring.

      SMART stands for Self-Monitoring, Analysis, and Reporting Technology and is a monitoring system included in hard drives and SSDs that reports on various attributes of the state of a given drive

      For vSAN adapter, it may be useful to collect SMART data on the physical devices (if applicable to the devices in the cluster and SMART statistics are available on the ESXi hosts)

       

      Edit the vSAN Adapter

       

      Expand Advanced Settings and set SMART Monitoring to true and save settings

      Graphical user interface, application</p>
<p>Description automatically generated

       

       

      Dashboards

      Navigate to Home > Operations Overview.

      We should now see the Datacenter(s), Cluster and Virtual Machines deployed

      Graphical user interface, application, table</p>
<p>Description automatically generated

       

      Select Dashboards from top Menu and select vSphere Compute Inventory for compute overview

      Graphical user interface, application</p>
<p>Description automatically generated

      For vSAN views, Select Dashboards drop down list -Operations > vSAN Operations Overview

       

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Enabling NSX-T Monitoring

      In this scenario we want to also monitor and collect data from one of the other critical infrastructure components, NSX-T.

      From VROPS Dashboard, Navigate to Administration > Other Accounts and Click Add Account

      Click on NSX-T Adapter

      Graphical user interface, application, Word</p>
<p>Description automatically generated

      New account wizard will be launched. Review details and click on “+” sign next to credentials to add NSX-T credentials.

      Graphical user interface</p>
<p>Description automatically generated

       

      Add applicable NSX-T Credentials.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Click Add to Add NSX-T Collector

      NSX-T Adapter should now be displayed under “Other Accounts”

      Note: You may have to wait for a short period for warning status to disappear

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Navigate back to Dashboards, select NSX-T and NSX-T Main.

      Environment view and topology view will soon begin to populate.

      Graphical user interface, application</p>
<p>Description automatically generated

      Enabling SDDC Health Management Pack

      The purpose of this optional scenario is to highlight that many management packs can be installed depending on the infrastructure. Below scenario is a simple example using a management pack from the VMware Marketplace.

       

      SDDC Health Management pack can be downloaded and install from VMware Marketplace.

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Download to your workstation.

      https://marketplace.cloud.vmware.com/services/details/vmware-sddc-management-health-solution/?slug=true

      It will be saved as vmware-MPforSDDCHealth-8.1-15995854.pak

      From VROPS Dashboard, Navigate to Administration > Repository > Select Add / Upgrade

      Select the SDDC Health Management by browsing your workstation where you downloaded vmware-MPforSDDCHealth-8.1-15995854.pak.

       

      Select upload to install and click Next.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

       

      SDDC Management Health should now show up as installed under “Other Management Packs”

      Navigate to Dashboards >  SDDC Management Health Overview

      As the data begins populating the dashboard, we will see relationship topology, health and metrics visible on the dashboard.

      Graphical user interface, application</p>
<p>Description automatically generated

      Connect Workload Domain to vRealize Operations

      If a VI Workload domain has been deployed via SDDC Manager, we can connect an existing workload domain to vRealize Operations directly from SDDC manager.

      This shows the integration and simplicity of vRealize Suite integration with VCF SDDC Manager.

       

      From SDDC Manager Dashboard, navigate to vRealize Suite

      Navigate to the vRealize Operations Card

       

      Graphical user interface, text, application, chat or text message</p>
<p>Description automatically generated

       

      Click on CONNECT WORKLOAD DOMAINS

       

      The connect workload domain wizard will begin

       

      From Modify Connection, select your applicable workload domain, and click next
       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Review settings and click finish.

      This will trigger a task in SDDC manager to connect vROPs to the new workload domain

       

       

      Once the task completes successfully navigate to vROPS dashboard to see the new vSphere environment being populated

      When logged on to vROPs dashboard,  navigate to Environment > Environment overview > vSphere Environment > vSphere Hosts and Clusters > vSphere World

      Expand newly discovered Workload Domain vCenter Instance

      Graphical user interface</p>
<p>Description automatically generated

       

       

      Summary

      An administrator should be able to verify basic vROPS integration is in place and be able to configure various solutions for vSAN and NSX-T etc.
      An administrator should also be able to simply initiate monitoring from SDDC manager on newly deployed workload domains.

      vRealize Suite and VCF – vRealize Log Insight integration.

      Depending on the POC requirements we may want to show vRealize Suite Monitoring integration with VCF SDDC infrastructure components

      In this scenario we want to verify if vRealize Log Insight (vRLI)  is monitoring and collecting logs and data from the various infrastructure components.

      An admin should be able to connect a Workload Domain to vRLI and be able to configure and modify various monitoring solutions.

      Pre-requisites

      • VCF Management Domain 4.x deployed
      • vRealize Log Insight
      • Optionally an additional Workload domain deployed

      Success criteria
      An administrator should be able to verify basic vRLI integrations are in place and be able to configure various solutions for vSphere and NSX-T etc.

      Verify Log Insight Connectivity to workload or management domain

      From SDDC Manager Dashboard, navigate to vRealize Suite. Navigate to the vRealize Log Insight card

      Connect to Log Insight from associated link on Log Insight card

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Login with vRLI credentials

       

      vSphere

      From Log Insight, navigate to administration > Integration > vSphere

      Graphical user interface, application, website</p>
<p>Description automatically generated

       

      Confirm the associated vCenter servers for the applicable workload domains are registered.

      Select ESXi hosts by selecting “View Details” to verify ESXi hosts are forwarding events to Log Insight.

      Graphical user interface</p>
<p>Description automatically generated

       

      Close the ESXi details once hosts have been validated

      vSphere Dashboards

      From Log Insight Dashboards, navigate to VMware > vSphere > General > Overview

      Ensure expected number of vCenter servers and ESXi hosts are populated on the vSphere events dashboard

       

       

      Graphical user interface, application, website</p>
<p>Description automatically generated

       

      Enable Launch in Context for vRealize Log Insight in vRealize Operations Manager

      If vROPs has already been installed, you can configure vRealize Log Insight to send alert notifications and metrics to vRealize Operations Manager.  You can configure vRealize Operations Manager to display menu items related to vRealize Log Insight and launch vRealize Log Insight with an object-specific query. 

      For more info, please review
      https://docs.vmware.com/en/vRealize-Log-Insight/8.1/com.vmware.log-insight.administration.doc/GUID-2BD3FBEA-DBD4-4572-8867-5CC5D3675C9E.html

       

      From Log Insight, navigate to Administration > Integration > vRealize Operations

      Ensure vROPS hostname and password are pointing to deployed vROPS instance.

      Click Test to verify setting

       Graphical user interface, application</p>
<p>Description automatically generated

       

      If not already enabled, enable alert management, launch in context and metric calculation.

      As mentioned, this allows vROPs and vRLI to integrate more closely.

      Content packs.

      Content packs contain dashboards, extracted fields, saved queries, and alerts that are related to a specific product or set of logs. You can install content packs from the Content Pack Marketplace without leaving the vRealize Log Insight UI. If your vRealize Log Insight server does not have internet access, you can download and install content packs separately

      vRealize Log Insight comes installed with General, vSphere, VMware vSAN, and vRealize Operations Manager content packs. You can also install content packs from the Content Pack Marketplace or create and export your own content packs for individual or team use.

      Navigate to Content Packs and updates as shown below.

      This allows to update existing content packs if there are updates to apply

       

      Graphical user interface, application, website</p>
<p>Description automatically generated

       

      Click Update All to ensure latest updates are applied

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Redirect NSX-T logs to vRealize Log Insight

      In this scenario we want to show the steps that may be required to ensure critical logs are forwarded from NSX-T infrastructure components.
      We first want to redirect logs to vRLI and verify monitoring is in place
      From an NSX-T perspective there are two methods to configure log forwarding, manually using SSH or using API on NSX-T to redirect logs

      We will briefly cover both here
      For more information please review https://docs.vmware.com/en/VMware-Validated-Design/6.0/sddc-deployment-of-cloud-operations-and-automation-in-the-first-region/GUID-C0931E46-F8ED-48EB-B1C0-AD074A04EF27.html where this procedure is outlined

      Reconfigure syslog forwarding on NSX-T Edges using SSH command line

       

      Connect NSX-T Edge, by using ssh or direct console access, login using admin

      To check existing logging, issue

      m01nsx01a> get logging-servers

      To set syslog, issue

      m01nsx01a> set logging-server m01vrli.vcf.sddc.lab:514 proto tcp level info

      Text</p>
<p>Description automatically generated

      m01nsx01a> get logging-servers

      m01vrli.vcf.sddc.lab:514 proto tcp level info

      Reconfigure syslog forwarding on NSX-T Edges using API method

       

      This can also be accomplished by using a Rest-API client e.g., postman (https://www.postman.com/)

      This is a sample JSON based request

      Identify NSX-T Edge IDs

      Login to NSX-T as admin

      Navigate to System > Fabric > Nodes >Edge Transport Nodes

      Select edge and click on ID Column

      Note down the ID

       

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

       

       

      This will be the URL to interrogate and modify the syslog details, where 667cb761-1655-4106-a611-90409e1cde77 is the NSX-T Edge ID.
      For example
      https://m01nsx01.vcf.sddc.lab/api/v1/transport-nodes/667cb761-1655-4106-a611-90409e1cde77/node/services/syslog/exporters

       

      Using Rest-API client set up Authorization

      Setup Basic auth under Authorization tab, add username and password

      Graphical user interface, application</p>
<p>Description automatically generated

       

       

      To update the configuration, navigate to Body , and select JSON

      Use below as an example to send logging to syslog server called m01vrli.vcf.sddc.lab

       

      {

      “exporter_name” : “syslog1”,

      “level”: “INFO” ,

      “port” : “514”

      “protocol”: “TCP”,

      “server”: “m01vrli.vcf.sddc.lab”

      }

      Once data has been entered issue POST to update.
      Graphical user interface, application, Teams</p>
<p>Description automatically generated

       

      To check, remove text and issue GET to retrieve current configuration.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Repeat for remaining NSX-T Edges

       

      NSX-T Controllers syslog forwarding

      Similarly using API for each NSX-T controller, to update syslog settings for NSX-T Controllers 

      Use the URL

      https://<nsx-t-controller-FQDN-or-IP>/api/v1/node/services/syslog/exporters

       

      Similar to the Edges, with a REST API Client like postman, use POST to update similar to below

       

      As before the content is raw in JSON format
       

      {

      “exporter_name” : “syslog1”,

      “level”: “INFO” ,

      “port” : “514”

      “protocol”: “TCP”,

      “server”: “m01vrli.vcf.sddc.lab”

      }

       

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Once NSX-T controllers and Edges are forwarding events, return to Log Insight

       

      NSX-T log forwarding verification

      From vRLI navigate to the Administration page > Management > Hosts.

      From the list of hosts forwarding events to Log Insight, locate the NSX-T controllers for the applicable VI workload domain and verify they are forwarding events

       

      A picture containing chart</p>
<p>Description automatically generated

       

      Click on an individual NSX-T controller to view interactive analytics and forwarded events via hostname.

      The interactive Analytics dashboard will display events forwarded by NSX-T controllers.

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Trigger an event on NSX-T Cluster e.g.  by powering off an NSX-T controller

      From Log Insight Dashboards, navigate to VMware > NSX-T > Infrastructure

       

       

      Graphical user interface, application</p>
<p>Description automatically generated

      We can clearly see events being triggered on NSX-T manager (controller) dashboard within last 5 minute of data
      Click on Interactive analytics

      We can confirm that one of the controllers cannot communicate to the controller that is currently unavailable.
       

      Graphical user interface, application</p>
<p>Description automatically generated

      Connect VCF Workload Domain to vRealize Log Insight

      If a VI Workload domain has been deployed via SDDC Manager, we can now connect a new workload domain to vRealize Log Insight.

      From SDDC Manager Dashboard, navigate to vRealize Suite

      Navigate to the vRealize Log Insight card

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      Select “CONNECT WORKLOAD DOMAINS

      The connect workload domain wizard will begin

      Select your applicable workload domain, and click next.

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

       

      Click Finish to initiate process,

      Graphical user interface, application</p>
<p>Description automatically generated

       

      A task will be initiated on SDDC manager

       

      Background pattern</p>
<p>Description automatically generated

       

      This will initiate a series of subtasks, click on “Connect domain to vRealize Log Insight” task for more details.

      Below is a list of subtasks that are associated with vRealize Log Insight connect to domain task.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

      To verify From Log Insight that workload domain is connected, navigate to Administration > Integration > vSphere

      Ensure the appropriate number of workload domains or vCenter servers are forwarding events

       

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Another approach is to verify via vSphere, see https://docs.vmware.com/en/VMware-vSphere/7.0/com.vmware.esxi.upgrade.doc/GUID-9F67DB52-F469-451F-B6C8-DAE8D95976E7.html

       

      From one of the workload domains in vCenter browse to the host in the vSphere Client inventory.

      Click Configure.

      Under System, click Advanced System Settings.

      Click Edit, Filter for syslog

       

      A screenshot of a computer</p>
<p>Description automatically generated

       

      Summary

      An administrator should be able to verify basic vRLI integrations are in place and be able to configure various solutions for vSphere and NSX-T etc.
      These scenarios should have proved that a logging was able to be implemented, with various solutions such as NSX-T which forward events to a syslog aggregator such as Log Insight.

      Dashboards and content packs are available to help interpret the log data and integrate with other solutions such as vROPs

      Lifecycle Management - VCF Management Domain Upgrade

      Lifecycle Management Overview

      The Lifecycle Management (LCM) feature of VMware Cloud Foundation enables automatic updating of both the Cloud Foundation software components (SDDC Manager, HMS, and LCM) as well as the VMware SDDC Components such as vCenter Server, ESXi, vSAN and NSX.

      Lifecycle Management in SDDC Manager may be applied to the entire infrastructure or to a specic workload domain. The process is designed to be non-disruptive to tenant virtual machines. As new software updates become available, the SDDC Manager provides notications to VCF Administrators, who may review the update details and, at a time convenient to them, download and schedule the updates.

      This module demonstrates usage of the Cloud Foundation Lifecycle Management feature to upgrade from VMware Cloud Foundation 4.0. to 4.X

      Depending on the POC requirements we may want to show VCF Upgrade from one release to another to highlight Life Cycle Management capabilities.

      In this scenario we want to understand the upgrade process from 4.0.X to 4.1

      Pre-requisites

      • VCF Management Domain 4.0.X deployed.
      • Access to correct binaries.

       

      Success criteria
      An admin should be able download bundles , configure and execute an upgrade of VCF including all managed infrastructure components

      Upgrading the VCF management domain will include the following steps

      1. Downloading the correct bundle
      2. Environment Pre-Check
      3. VCF SDDC manager upgrade
      4. vRealize LCM Upgrade (if installed)
      5. NSX-T Upgrade
        • This includes Edge(s), hosts and manager
      6. vCenter Upgrade
      7. ESXi Host/Cluster Upgrades

      As per official documented guidance on  Upgrade Workload Domains to 4.2
      The management domain in your environment must be upgraded before you upgrade a workload domain

      Ensure to review release notes on known issues, especially around upgrade scenarios
      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/rn/VMware-Cloud-Foundation-41-Release-Notes.html#knownissues

      Bundle Management

      Bundle Types

      Cloud Foundation utilizes two types of bundles for Lifecycle Management: Upgrade Bundles and Install Bundles.

      Upgrade Bundles

      An upgrade bundle contains patches and other software necessary to update VCF software components. In most cases, an upgrade bundle must be applied to the management domain before it may be applied to workload domains.

      Some upgrade bundles are cumulative bundles. In cases where a workload domain is multiple versions behind the target version, cumulative bundles allow Cloud Foundation to directly upgrade to the target version (rather than requiring the installation of multiple bundles in a sequential progression). Cumulative bundles are only available for vCenter Server and ESXi.

      Install Bundles

      Install bundles contain software necessary to deploy new instances of Cloud Foundation components. For instance, VI workload domain install bundles are used to deploy more recent versions of the software components that were not present in the initial Cloud Foundation BOM; these install bundles include software for vCenter Server and NSX-T Data Center.

      Downloading Bundles

      If SDDC Manager is congured with 'My VMware' credentials, Lifecycle Management automatically polls the VMware software depot to access software bundles. SDDC Manager will prompt administrators when a bundle is available and ready for download.

      If SDDC Manager does not have Internet connectivity, software bundles may either be acquired via HTTP(S) proxy, or through a manual download and transfer process.

      This guide demonstrates procedures for automatically downloading bundles, and manually downloading and transferring bundles. For the procedure to download bundles with a proxy server, please refer to the VMware Cloud Foundation Upgrade Guide.

      Configure Credentials

      Login to SDDC Manager, on the left navigation pane navigate to Administration > Repository Settings.
      From the My VMware Account Authentication wizard, enter valid My VMware credentials.
      Once My VMware credentials are validated the Repository settings will display as ‘Active’. In some environments, it may be necessary to congure SDDC Manager to utilize a HTTP(S) proxy.

      Download Bundles

      After registering My VMware credentials, navigate to Repository > Bundle Management.

      Locate and click ‘Schedule for Download’ or ‘Download Now’ to obtain the VMware Software Install Bundle - vRealize Suite Lifecycle Manager.
       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Offline Bundle Download

      If SDDC manager does not have access to the internet offline bundle download task may be achieved using the Bundle Transfer Utility & Skip Level Upgrade Tool.  

      Please refer to official 4.1 Documentation https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/vcf-41-lifecycle/GUID-8FA44ACE-8F04-47DA-845E-E0863094F7B0.html 

      Please refer to official 4.2 Documentation https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/vcf-42-lifecycle/GUID-8FA44ACE-8F04-47DA-845E-E0863094F7B0.html

      For the purposes of this guide we are showing a worked example of how to use the Bundle Transfer Utility for situations where internet access is not available for SDDC Manager
      The Bundle Transfer Utility & Skip Level Upgrade Tool  can be downloaded to a workstation with internet access     https://my.vmware.com/group/vmware/downloads/details?downloadGroup=VCF420_TOOLS&productId=1121

      To download the correct bundles, you will need access to the SDDC manager via SSH and command line access on the machine performing the download.
      In this scenario a Windows workstation with internet access was used to download the bundles

      Step 1: Download and extract the Bundle Transfer Utility & Skip Level Upgrade Tool to your workstation with internet access. Ensure the workstation has adequate space for storing the bundles.

      Use lcm-bundle-transfer-util to test connectivity to my.vmware.com To achieve this use lcm-bundle-transfer-util with --listBundles option

      c:> lcm-bundle-transfer-util --listBundles --depotUser myvmwareUserId
      VMware Cloud Foundation LCM Bundle Transfer Tool, Version: 4.1.0-vcf4100RELEASE-17140097
      VMware Cloud Foundation LCM Tools version : 4.1.0-vcf4100RELEASE-17140097
      Enter Myvmware user password:
      Validating the depot user credentials...
      [WARNING] This operation fetches all the bundles present in the depot...
      ***************************************************************************************************

      Bundle Product Version Bundle Size (in MB) Components

      ****************************************************************************************************

      (truncated list)

      bundle-28605 4.1.0.0 1069.0 MB VRSLCM-8.1.0-16776528-INSTALL

      (truncated list)

      Step 2: 

      Beginning in 4.2 you must download a manifest file first and then upload that to SDDC Manager before downloading any bundles.  From you Windows machine you will use the offline bundle tool to download this from VMware using the following command.  Once complete you should see the lcmManifestv1.json file in the output directory.   

      lcm-bundle-transfer-util --download -manifestDownload --outputDirectory c:\offlinebundle --depotUser user@vmware.com -depotUserPassword userpass

      Step 3:

      Use a transfer utility like WinSCP to move the manifest file to SDDC Manager.  In our example we put the files in /home/vcf.  Change the permissions on the manifest file to an Octal value of 7777.  You can do this from your transfer utility or if you SSH into SDDC Manager run:

      chmod 7777 /home/vcf/lcmManifestv1.json

      Step 4:

      SSH into SDDC Manager and run the following to ingest the manifest file. 

      cd /opt/vmware/vcf/lcm/lcm-tool/bin ./lcm-bundle-transfer-util --updae --sourceManifestDirectory /home/vcf/ --sddcMgrFqdn sddc-manager.vcf.sddc.lab --sddcMgrUser administrator@vsphere.local

      Step 5:

      Connect via ssh to sddc-manager with username “vcf” and change directory to /opt/vmware/vcf/lcm/lcm-tools/bin Use the lcm-bundle-transfer-util to generate a marker file that catalogues all the software on SDDC manager and what bundle IDs needed to download for the version on SDDC Manager
      Here is a worked example

      $ cd /opt/vmware/vcf/lcm/lcm-tools/bin/

      $./lcm-bundle-transfer-util --generateMarker

      $ ls -la /home/vcf/markerFile*

      -rw------- 1 vcf vcf 524 Nov 18 20:25 /home/vcf/markerFile

      -rw------- 1 vcf vcf 32 Nov 18 20:25 /home/vcf/markerFile.md5

      Step 6:
      Download and copy the marker file and marker md5sum from sddc-manager to your workstation using a secure shell transfer client (pscp, winscp etc).
       

      Step 7
      Download the bundle IDs for 4.2 using the marker file and m5sum
      For fresh install of VCF 4.2 we require approx. 20GB free space.  If you only want to download updates for a specific version, you can use the -v switch.  In our example I could have done -v 4.2.0.0 to download only updates for 4.2.0.0.
      Here is a worked example:

      lcm-bundle-transfer-util -download -outputDirectory c:\offlinebundle -depotUser user@vmware.com -markerFile c:\offlinebundle\markerfile -markerMd5File c:\offlinebundle\markerFile.md5

       

      Step 8:
      Copy the output directory in Step 4 to SDDC Manager (using a secure shell client such as pscp /WinSCP) NFS share e.g., /nfs/vmware/vcf/nfs-mount/offlinebundle

      Make sure to change the permissions on the uploaded folder.  From you transfer utility change the octal value to 7777 or from SSH:

      cd /nfs/vmware/vcf/nfs-mount chmod -R 7777 offlinebundle/

      Step 9:

      Ingest the bundles into the VCF LCM repository

      Here is a worked example:

      cd /opt/vmware/vcf/lcm/lcm-tools/bin ./lcm-bundle-transfer-util -upload -bundleDirectory /nfs/vmware/vcf/nfs-mount/offlinebundle/

      This generates SDDC Manager tasks in the SDDC Manager dashboard.

      Once all upload tasks succeed, check the download history under Bundle Download History on the SDDC Manager Dashboard

      A picture containing text</p>
<p>Description automatically generated

      Environment Pre-Check

      Pre-Check environment before upgrade

      As per product documentation https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/vcf-41-lifecycle/GUID-E3F6EEFF-698F-48F0-BCBF-E6CAEF6C1EBD.html 

      Ensure all precheck issues are dealt with prior to upgrade. Any failures should be identified and addressed

      Ensure to re-run pre-check to address and understand all failures and address as appropriate

      Supportability and Serviceability (SoS) Utility

      The SoS utility is a command-line tool that you can use to run health checks, collect logs for Cloud Foundation components. Prior to upgrade it may be advisable to run sos utility as an extra step to verify health of VCF Management domain

       

      To run the SoS utility, SSH in to the SDDC Manager VM using the vcf user account

      Summary

      --get-vcf-summary lists the summary of deployed VCF

      sudo /opt/vmware/sddc-support/sos   --get-vcf-summary    

      Health Check 
      This is equivalent to Pre-Check on SDDC Manager UI
      sudo /opt/vmware/sddc-support/sos  --health-check
      Once the health check runs it will give health status legends of

      GREEN - No attention required, health status is NORMAL
      YELLOW - May require attention, health status is WARNING
      RED - Requires immediate attention, health status is CRITICAL

      Health check can be further modularized
      Health Check:

        --health-check        Perform all available Health Checks

        --connectivity-health

                              Perform Connectivity Health Check

        --services-health     Perform Services Health Check

        --compute-health      Perform Compute Health Check

        --storage-health      Perform Storage Health Check

        --run-vsan-checks     Perform Storage Proactive Checks

        --ntp-health          Perform NTP Health Check

        --dns-health          Perform Forward and Reverse DNS Health Check

        --general-health      Perform General Health Check

        --certificate-health  Perform Certificate Health Check

        --composability-infra-health

      Performs Composability infra Api connectivity

      check(if Composability infra found else skipped)

        --get-host-ips        Get Server Information

        --get-inventory-info  Get Inventory details for SDDC

        --password-health     Check password expiry status

        --hardware-compatibility-report

                              Validates hosts and vsan devices and export the

                              compatibility report

        --json-output-dir JSONDIR

                              Outputs health check results JSON file to given

                              Directory

      Upgrading SDDC Manager

      SDDC Manager upgrade

      Before you can upgrade or update any SDDC management component by using SDDC Manager, you must first upgrade SDDC Manager. You upgrade SDDC Manager by downloading and applying the necessary VMware Cloud Foundation upgrade bundle and configuration drift bundle.

      To upgrade to VMware Cloud Foundation you apply two bundles to the management domain.

      Cloud Foundation Update Bundle

      The VMware Cloud Foundation Update bundle upgrades LCM and VMware Cloud Foundation services, basically SDDC Manager

      Configuration Drift Bundle

      The configuration drift bundle applies configuration changes required for 2nd party software components in the VMware Cloud Foundation Bill of Materials for the target release.

      A picture containing background pattern</p>
<p>Description automatically generated

      Prerequisites
      To download the applicable download bundles complete the following steps:

      1. Navigate to the Updates/Patches tab of the management domain.
      2. Run the upgrade pre-check.
      3. In the Available Updates section, click Update Now or Schedule Update next to the VMware Cloud Foundation Update bundle.

      The Cloud Foundation Update Status window displays the components that will be upgraded and the upgrade status.    

      Click View Update Activity to view the detailed tasks.

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      After the upgrade is completed, a green bar with a check mark is displayed.

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Once logged into SDDC Manager, the recent tasks card on SDDC Dashboard will display the status of the upgrade.
       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Upgrading NSX-T

      Upgrading NSX-T

      Review NSX-T 3.0.2 Updates before launching upgrade
      https://docs.vmware.com/en/VMware-NSX-T-Data-Center/3.0/rn/VMware-NSX-T-Data-Center-302-Release-Notes.html

      Upgrading NSX-T Data Center involves the following components:

      • Upgrade Coordinator
      • Edge clusters (if deployed)
      • Host clusters
      • NSX Manager cluster

      The upgrade wizard provides some flexibility when upgrading NSX-T Data Center for VUM-based workload domains. By default, the process upgrades all Edge clusters in parallel, and then all host clusters in parallel.

       

      Precheck NSX-T health

      Since NSX-T Upgrade from 3.0.1 to 3.0.2 is one of the first infrastructure upgrades, it also wise to interrogate any open alerts or events prior to upgrade.
      From SDDC Manager Dashboard, select Management domain Summary and navigate to NSX-T Manager IP address.
      Login to NSX-T and review Dashboards and Monitoring Dashboards for overall health.
      Address active alarms and health issues prior to upgrade.

       

      Graphical user interface, application</p>
<p>Description automatically generated

      Graphical user interface, application</p>
<p>Description automatically generated

      NSX-T Edge Upgrade

      From SDDC-Manager, Select Management Domain > Updates > Available Updates
      Select VMware Software Update 4.1.0.0 for NSX-T component
      Select Update Now

      Graphical user interface, text, application</p>
<p>Description automatically generated

      A wizard will be launched to guide the admin through the process of upgrade.
      If the management domain has an Edge Cluster, select Upgrade NSX-T edge Clusters only and specify the edge-clusters to upgrade.
      By default, all Edge clusters are upgraded. To select specific Edge clusters, click Enable edge selection. To upgrade only the Edge clusters, select Upgrade NSX-T Edge clusters only.
       

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      If more than one cluster, NSX-T edge clusters will be upgraded in parallel. Once all edge clusters are upgraded,
      NSX-T host clusters can also be upgraded in parallel (except for vLCM enabled host clusters)
      Parallel upgrade can be over-ridden like below if so desired

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Click Finish to launch the upgrade of NSX-T Edges

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      Upgrade for NSX-T Edges Task is now initiated
      From SDDC manager the task can be monitored
      Each subtask can be monitored by selecting the Upgrade DOMAIN task
      As upgrade progresses the overall summary can be viewed on subtask list

      Graphical user interface, application</p>
<p>Description automatically generated

      NSX-T Edge Upgrade can also be monitored by viewing the log files on SDDC manager

      #grep NSX_T_UC /var/log/vmware/vcf/lcm/lcm-debug.log
      2020-11-16T15:45:26.905+0000 DEBUG [vcf_lcm,c373d18774867d19,1b6b,upgradeId=eee02a0c-20ea-40ef-bdc3-09917c2fa183,resourceType=NSX_T_PARALLEL_CLUSTER,resourceId=m01nsx01.vcf.sddc.lab:_ParallelClusterUpgradeElement,bundleElementId=5a3bff6e-0466-4364-b9bb-242d1c1bcad2] [c.v.e.s.l.p.i.n.NsxtParallelClusterPrimitiveImpl,ThreadPoolTaskExecutor-10] All upgrade elements of type NSX_T_UC are COMPLETED_WITH_SUCCESS, thus we proceed to upgrade next batch of type NSX_T_EDGE

      #grep NSX_T_EDGE /var/log/vmware/vcf/lcm/lcm-debug.log
      2020-11-16T15:46:17.331+0000 INFO  [vcf_lcm,c373d18774867d19,1b6b,upgradeId=eee02a0c-20ea-40ef-bdc3-09917c2fa183,resourceType=NSX_T_PARALLEL_CLUSTER,resourceId=m01nsx01.vcf.sddc.lab:_ParallelClusterUpgradeElement,bundleElementId=5a3bff6e-0466-4364-b9bb-242d1c1bcad2] [c.v.e.s.l.p.i.n.s.NsxtEdgeClusterParallelUpgradeStageRunner,ThreadPoolTaskExecutor-10] Performing NSX-T edge cluster upgrade stage NSX_T_EDGE_TYPE_UPGRADE

      NSX-T Edge Upgrade Progress can also be monitored from within NSX-T manager via the NSX-T upgrade coordinator.

      Login to NSX-T Manager and Navigate to System –> Lifecycle Management > Upgrade
      You may then be directed to the upgrade coordinator

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Login to the upgrade Co-Ordinator to monitor progress

      Graphical user interface, text, application</p>
<p>Description automatically generated

      NSX-T Edge Post Upgrade

      Once the NSX-T Edge upgrade is complete return to SDDC Manager UI to start the next phase of NSX-T Upgrade
      Perform Pre-Check to verify health of environment after NSX-T Edge Upgrade.
      There may be errors to be verified on NSX-T, especially around stale edge events that would need to be identified and acknowledged or addressed.

      Once issues have been addressed, they should be marked resolved

      Graphical user interface</p>
<p>Description automatically generated

      NSX-T Host Upgrade

      Once environment is healthy proceed with upgrade of NSX-T Host Clusters

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Since Edge-Clusters have already being upgraded in this scenario, SDDC manager will acknowledge this.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      The next major phase of upgrade is upgrading the NSX-T Host clusters or host transport zones

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

      This option applies to VUM-based workload domains only.

       

      By default, VUM-based workload domains upgrade Edge clusters and host clusters in parallel.

      These options are not available for vLCM-based workload domains, where Edge clusters and host clusters are upgraded sequentially.

       

      Option

      Description

      Enable sequential upgrade of NSX-T Edge clusters

      Upgrades Edge clusters sequentially, instead of in parallel.

      Enable sequential upgrade of NSX-T hosts clusters

      Upgrades host clusters sequentially, instead of in parallel.

       

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Once upgrade steps have been decided, the wizard gives the user the chance to review before submitting task

      Graphical user interface, application</p>
<p>Description automatically generated

      Graphical user interface, text, application, website</p>
<p>Description automatically generatedMonitor NSX host upgrade
      This can be achieved from SDDC manager tasks or from command line on SDDC Manager

      Graphical user interface, application</p>
<p>Description automatically generated

      From SDDC command line, ssh to SDDC manager and review the logs /var/log/vmware/vcf/lcm/lcm-debug.log
      tail -f /var/log/vmware/vcf/lcm/lcm-debug.log

      2020-11-16T21:47:19.631+0000 DEBUG [vcf_lcm,7ea56fea9ec2b9e5,3668] [c.v.evo.sddc.lcm.orch.Orchestrator,pool-6-thread-6] is NSX-T parallel upgrade true
      2020-11-16T21:47:19.631+0000 DEBUG [vcf_lcm,7ea56fea9ec2b9e5,3668] [c.v.evo.sddc.lcm.orch.Orchestrator,pool-6-thread-6] No failed NSX-T parallel cluster upgrade element found
      2020-11-16T21:47:19.631+0000 DEBUG [vcf_lcm,7ea56fea9ec2b9e5,3668] [c.v.evo.sddc.lcm.orch.Orchestrator,pool-6-thread-6] is NSX-T parallel upgrade true
      2020-11-16T21:47:19.631+0000 DEBUG [vcf_lcm,7ea56fea9ec2b9e5,3668] [c.v.evo.sddc.lcm.orch.Orchestrator,pool-6-thread-6] Found an NSX-T parallel cluster upgrade element and is in progress

      Each host cluster will be put into maintenance mode to update NSX-T software

      Monitoring NSX-T Host Upgrades from NSX-T Management
      Once hosts are upgraded the NSX-T upgrade coordinator will reflect the status

      Graphical user interface, application</p>
<p>Description automatically generated

      This should reflect nicely to what SDDC Manager reports

      A picture containing background pattern</p>
<p>Description automatically generated

      NSX-T Manager Upgrade

      The final piece of NSX-T Upgrade is the managers themselves

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Once complete the task should be successful and the Upgrade Coordinator should reflect the same

      SDDC Manager Task

      Graphical user interface, application</p>
<p>Description automatically generated

      From NSX-T, the Upgrade Coordinator should reflect the same

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Upgrading vCenter

      Upgrading vCenter

      The next major piece of the infrastructure us the vCenter in the VCF Management domain

      As per documentation https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/vcf-41-lifecycle/GUID-F9E0A7C2-6C68-45B9-939A-C0D0114C3516.html

      The management domain must be upgrade before upgrading any workload domains.

      Prerequisites and Procedure.
      Download the VMware vCenter upgrade bundle
      from SDDC Manager is to

      • Navigate to the Updates/Patches tab of the domain you are upgrading.
      • Run the upgrade precheck. 

      Then proceed to Update Now or Schedule Update next to the vCenter upgrade bundle.
      If you selected Schedule Update, click the date and time for the bundle to be applied.
      In our case we are choosing to upgrade now

      A picture containing graphical user interface, application</p>
<p>Description automatically generatedThis will initiate a task in SDDC manager

      Graphical user interface, application</p>
<p>Description automatically generated

      As part of the process the vCenter VCSA appliance will be snapshotted

      Graphical user interface, application</p>
<p>Description automatically generated

      Once snapshot is complete, the upgrade will continue.
      Once Upgrade is complete it can be tracked on tasks and events

      A picture containing graphical user interface</p>
<p>Description automatically generated

       

      ESXi Host Upgrades

      Upgrading ESXi Hosts

      Reference  documentation:

      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.1/vcf-41-lifecycle/GUID-10738818-5AD4-4503-8965-D9920CB90D22.html

      By default, the upgrade process upgrades the ESXi hosts in all clusters in a domain in parallel. If you have multiple clusters in the management domain or in a VI workload domain, you can select which clusters to upgrade. You can also choose to update the clusters in parallel or sequentially.

      • Ensure that the domain for which you want to perform cluster-level upgrade does not have any hosts or clusters in an error state. Resolve the error state or remove the hosts and clusters with errors before proceeding.
      • For clusters in a vSphere Lifecycle Manager (vLCM)-enabled workload domain, you must have a cluster image set up that includes the ESXi version that you want to upgrade to. The ESXi version must match the version in the bundle you downloaded. See
      • To add or upgrade the firmware on clusters in a vLCM-enabled workload domain, you must have the vendor Hardware Support Manager installed.
      • To apply firmware updates to hosts in a cluster, you must deploy and configure a vendor provided software module called hardware support manager or HSM. The deployment method and the management of a hardware support manager is determined by the respective OEM.  See https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/com.vmware.vcf.admin.doc_41/GUID-198C2DE5-8DC1-4327-9914-C4677F4964D5.html.
        For detailed information about deploying, configuring, and managing hardware support managers, refer to the vendor-provided documentation.

      For custom ESXi images i.e., with Custom ISO,  you can upgrade ESXi with a custom ISO from your vendor. This feature is available for VMware Cloud Foundation version 3.5.1 

      To Upgrade ESXi with VMware Cloud Foundation Stock ISO and Async Drivers

      You can apply the stock ESXi upgrade bundle with specified async drivers. This feature is available for VMware Cloud Foundation version 3.5.1 and later.

      https://docs.vmware.com/en/VMware-Cloud-Foundation/4.2/vcf-41-lifecycle/GUID-36B94FC4-2018-444B-9AFB-F541CA5B2F99.html

      In the scenario below we are upgrading VUM based images with stock ESXi ISOs and no async drivers
       

      As with vCenter upgrade ESXi Cluster upgrades are initiated from SDDC Manager

      Graphical user interface, text, application</p>
<p>Description automatically generated

      By default, all clusters can be selected in applicable domain, or individual clusters can be selected as per below screenshot

      Graphical user interface, text, application</p>
<p>Description automatically generated

      By default, all clusters are upgraded in parallel. To upgrade clusters sequentially, select Enable sequential cluster upgrade.

      Graphical user interface, text, application</p>
<p>Description automatically generated

       

      Text</p>
<p>Description automatically generated

      Click Enable Quick Boot if desired. Quick Boot for ESXi hosts is an option that allows Update Manager to reduce the upgrade time by skipping the physical reboot of the host. More details on Quick boot can be found on KB https://kb.vmware.com/s/article/52477

      Once the ESXi upgrade task is initiated it can be monitored from SDDC Manager Tasks

      As per normal, sub tasks can be monitored by selecting the main task

      Background pattern</p>
<p>Description automatically generated

      Tasks can also be monitored from Workload domains Update tab

      Click View Status for details
      View status will give a granular view of the upgrade process.

      Graphical user interface, text, application</p>
<p>Description automatically generated

      Update Activity can be further interrogated by clicking View Update Activity

      Graphical user interface, text, application</p>
<p>Description automatically generated

      As each ESXi Node gets upgraded the status will be reflected on update activity.

      Graphical user interface, application</p>
<p>Description automatically generated

      Once upgrade completes the task will be marked successful on SDDC Manager

      Appendix A - NSX-T Fabric Overview

      NSX Configuration Overview: Management Workload Domain

      NSX provides the core networking infrastructure in the software-dened data center stack within VCF. Every workload domain is integrated with and backed by an NSX-T platform.
       

      For more information on NSX-T please review https://docs.vmware.com/en/VMware-NSX-T-Data-Center/index.html

      The Management workload domain is precongured with NSX-T, For VI workload domains, NSX-T can be deployed alongside new workload domains or new workload domains can be added to existing NSX-T deployments. By default, workload domains do not include any NSX-T Edge clusters and as such are isolated.

      During the initial VCF bring-up process, NSX-T is automatically deployed and configured in the management workload domain. Its default components include NSX-T 3.0 instance which is comprised of three controllers with a VIP for management access. Follow the steps below to review the main components of the NSX-T architecture and how it relates to VCF 4.X

      • Log into SDDC Manager
      • In the left panel, navigate to Inventory > Workload Domains
      • Select the workload domain that is of type ‘MANAGEMENT’ (‘ mgmt-domain’ in this example):

      Select the management cluster, (‘mgmt01’ in this example):

      We can see that there is an NSX-T instance (nsx-mgmt-vcf.ssdc.lab) deployed with an associated NSX-T Edge Cluster (mgmt- edge-cluster), as we have chosen to deploy VCF with the option of AVN (Application Virtual Networks).
      Note: If AVN was not chosen as part of bring-up, an Edge Cluster would not be available. as per the following screenshot:

      Accessing NSX-T interface from SDDC Manager

      Click on the hyperlink to launch NSX-T Web interface, and login with administrative privileges defined on bring-up i.e., admin
      Once logged in from the NSX-T Dashboard we are shown four main dashboards Networking, Security, Inventory and System.
       

      In the next section we will focus on System and Networking and how that relates to VCF.

      NSX-T Appliances.

      The NSX-T Data Center Unied Appliance is an appliance included in the installation of NSX-T. It includes the ability to deploy the appliance in the roles of NSX Manager, Policy Manager, or Cloud Service Manager.

      VMware has combined both the NSX Manager and NSX controller into a single virtual appliance called “NSX unied appliance” which can be run in a clustered conguration.

      During initial VCF 4.X bringup NSX-T Appliances are deployed on the management cluster and automatically congured as per the bring-up spec

      This is the screen-shot of the VCF 4.X Excel spread-sheet section relating to NSX-T

       

      To inspect the NSX-T appliances and cluster status

      • Click on System to review the Fabric.
      • On the left-hand navigation pane click on Appliances

       

      We will first inspect the NSX-T Appliances. There are three appliances that are deployed and clustered together. To access the cluster a Virtual IP (in our case IP 10.0.0.20) to automatically congured as per VCF 4.0 bring up spec.

      The NSX-T Cluster Status should be of status stable

      Transport Zones

      In NSX-T Data Center, a transport zone (TZ) is a logical construct that controls which hosts a logical switch can reach. A transport zone denes a collection of hosts that can communicate with each other across a physical network infrastructure. This communication happens over one or more interfaces dened as Tunnel Endpoints (TEPs).

      There are two types of transport zones: Overlay and VLAN. An overlay transport zone is used by ESXi host transport nodes and NSX-T Edge Nodes. When an ESXi host or NSX-T Edge transport node is added to an Overlay transport zone, an N-VDS is installed on the ESXi host or NSX Edge Node.

      VCF 4.0 will automatically congure three transport zones (two if no AVNs are specied)

      To inspect the transport zones automatically congured by VCF, click on Fabric > Transport Zones

      We have three configured transport zones. The VLAN transport zone is used by NSX-T Edge Nodes and ESXi host transport nodes for its VLAN uplinks. When an NSX-T Edge Node is added to a VLAN transport zone, a VLAN N-VDS is installed on the NSX-T Edge Node.

      • Overlay transport zone for host transport nodes and edge nodes
      • VLAN-backed transport zone for host management networks, e.g., vSAN and VMotion VLAN-backed edge transport zone for Edge uplinks

      To inspect host transport zone overlay

      Click on transport zone name, in our example, mgmt-domain-tz-overlay01. The overview will show number of hosts and edges associated, number of switches and switch ports.

      Click on Monitor to review the health and status of the transport nodes, in this case hosts and edge appliances

      You may repeat this procedure for the remaining transport nodes.

      Host Transport Nodes

      In NSX-T Data Center, a Transport Node allows nodes to exchange trac for virtual networks.

      The vSphere hosts were dened on the VCF 4.X Excel Spread-sheet and act as Transport Nodes for NSX-T

      To inspect the host transport nodes from an NSX-T perspective

      • From the system view Click on Fabric -> Nodes
      • From Host Transport Nodes click on drop down pick list next to "Managed by"
      • Select the Compute Manager, in our case vcenter-mgmt.vcf.sddc.lab
      • Expand the Cluster, in our case mgmt-cluster

      We should now see (since this is a management cluster) a minimum of four vSphere hosts from the management cluster prepared successfully and Node status should be Up

      The hosts were dened on the VCF 4.x Excel Spread-sheet as esxi-1 through to esxi-4

      Edge Transport Nodes

      The NSX Edge provides routing services and connectivity to networks that are external to the NSX-T Data Center deployment. An NSX Edge is required if you want to deploy a tier-0 router or a tier-1 router with stateful services such as network address translation (NAT), VPN and so on.

      An NSX Edge can belong to one overlay transport zone and multiple VLAN transport zones. If a Virtual Machine requires access to the outside world, the NSX Edge must belong to the same transport zone that the VM's logical switch belongs to. Generally, the NSX Edge belongs to at least one VLAN transport zone to provide the uplink access.

      We dened NSX-T Edges in the VCF 4.X Excel spread-sheet

      To review the edge transport nodes and clusters.
      Click on Fabric > Nodes > Edge Transport Nodes


      Click on one of the edge-transport nodes for more details


      We can see this edge is associated with 2 transport zones; a VLAN sfo01-m01-edge-uplink-tz and a host overlay transport zone mgmt-domain-tx-overlay01

      Click on Monitor to review the system resources and how each interface on the appliance is associated to each uplink. the interfaces fp-ethX can be mapped to the virtual vNIC interfaces on the edge appliance.

      Compute Manager

      A compute manager, such as vCenter Server manages resources such as hosts and VMs.

      NSX-T Data Center is decoupled from vCenter. When VCF bringup process adds a vCenter Server compute manager to NSX-T, it will use the vCenter Server user's credentials dened in the VCF 4.X bringup specications

      When registered NSX-T polls compute managers to nd out about changes such as, the addition or removal of hosts or VMs and updates its inventory accordingly.

      To inspect the conguration

       

      Click Fabric > Compute Managers

      Click on the registered Compute Manager to gather more details, in this case it is the management vCenter server

      NSX-T Logical Networking Overview

      In this section we will review the logical networking concepts and how they relate to VCF 4.x Management Domain bring-up.

      A few terms to help with this overview, for more information please review the NSX-T 3.0 installation

      guide  https://docs.vmware.com/en/VMware-NSX-T-Data-Center/3.0/installation/GUID-A1BBC650-CCE3-4AC3-A774-92B195183492.html

      Tier-0 Gateway or Tier-0 Logical Router

      The Tier-0 Gateway in the Networking tab interfaces with the physical network. The Tier-0 gateway runs BGP and peers with physical routers.

      Tier-1 Gateway or Tier-1 Logical Router

      The Tier-1 Gateway in the Networking tab connects to one Tier-0 gateway for northbound connectivity and one or more overlay networks for southbound connectivity.

      Graphical user interface, application, email</p>
<p>Description automatically generated

      Segment

      This is also known as a logical switch. A segment provides virtual Layer 2 switching for Virtual Machine interfaces and 
      Gateway interfaces. A segment is a logical entity independent of the physical hypervisor infrastructure and spans many hypervisors, connecting VMs regardless of their physical location.
      Virtual machines attached to the same segments can communicate with each other across transport nodes through   
      encapsulation over tunneling.
      When segments are created, they appear as a port group on vSphere.

      Tier-0 and Tier-1 Gateways

      To review the Tier-0 Gateway deployment

      From the main NSX-T Dashboard, click on Networking > Tier-0 Gateways. We can see that the Tier-0 gateway, mgmt-edge- cluster-t0-gw01, has been deployed as part of VCF 4.0 bring-up.

      Click on Linked -Tier1 Gateways. we now see that the Tier-1 Gateway, mgmt-edge-cluster-t1-gw01 is associated with the Tier-0 gateway

      Click on Networking > Tier-1 Gateways


      Click on Linked Segments, we see the Tier-1 Gateway is associated with 2 segments. These segments were dened on the VCF bring-up spec

      Tier-0 BGP Routing

      To enable access between your VMs and the outside world, you can configure an external or internal BGP (eBGP or iBGP) connection between a tier-0 gateway and a router in your physical infrastructure.

      This is a general review of BGP routing and how it was dened on the VCF 4.X bring-up and what it looks like on the NSX-T Manager

      Here are the deployment parameters for the AVNs on VCF 4.x spreadsheet. This part of the bring-up simply dened how the Edges for the management cluster would be deployed and what should the conguration of Tier-0 (with BGP) and Tier-1 gateways looks like.

      Note the following details,

      • Edge Autonomous System ID 65003
      • Top of Rack Autonomous System ID 65001
      • Top of Rack IPs 192.168.16.10 and 192.168.17.10

      When configuring BGP, you must congure a local Autonomous System (AS) number for the Tier-0 gateway. VCF spec set this value to 65003. Both edges must use the same AS number

      You must also congure the remote AS number on the Top of Rack switches. As per the VCF bringup spec, the physical Top of Rack switch AS number is 65001

      EBGP neighbors must be directly connected and in the same subnet as the tier-0 uplink.

      We can see by the VCF screenshot above both edge node 1 and edge node 2 have Uplinks dened on 192.168.16.0/24 and 192.168.17.0/24.

      We also see both top of rack switches have IP addresses on the same subnet i.e., 192.168.16.10 and 192.168.17.10

      From the main NSX-T Dashboard, click on Networking > Tier-0 Gateways.

      Expand the Tier-0 Gateway for more details.

      Expand BGP Details.

      We can see the local AS number is 65003 which matches the excel spreadsheet entry.
       

      Next, we will look at the BGP Neighbors. we can see the detail by looking at the details behind the BGP Neighbors, in this case 2

      Now we see that there are two neighbors congured, 192.168.16.10 and 192.168.17.10, with AS number 65001 this matches the Top of Rack Switch details dened on the VCF spreadsheet.

       

      Graphical user interface, application</p>
<p>Description automatically generated

      For a graphical representation of Tier-0 BGP configuration, close the BGP Neighbors detail and click on topology view highlighted below in red

      We can see the IP addresses 192.168.16.2, 192.168.16.3, 192.168.17.2 and 192.168.17.3 are congured and peered to the Top Of rack Switches (192.168.16.10 and 192.168.17.10)

      Graphical user interface, application</p>
<p>Description automatically generated

       

      Segments

      This is also known as a logical switch. as per the VCF 4.0 Bring up spec, we have dened two segments as per the AVN setup for Virtual Machine Trac.

      These are Region -A Logical Segment called local-sgement and xRegion Logical Segment xregion-segment

       

      To view the AVN Segments, click on Networking > Segments

      Take note of the two segments highlighted below, these are backed by management domain overlay transport zone.

       

      These Segments are presented as Port Groups on vSphere.

      To view on vSphere

      • Login to vSphere Management vCenter server,
      • Navigate from Home > Networking > Management Networks
      • expand management distributed switch and locate the segments

       

      Edge Segments

      The remaining two Segments are for VLAN backed up-link connectivity for the NSX-Edges
      These VLANs were defined on bring-up on the VCF 4.X excel spreadsheet, see NSX-T Edge Uplink-1 and Edge-Uplink-2

      This is a detailed view of one of the NSX-T Edge uplink segments (Edge Uplink 1)

      NSX-T Edge Overlay

      An NSX-T Edge overlay is also defined for on VCF 4.X bring up excel spreadsheet

      Separate VLANs and subnets are required for NSX-T Host Overlay (Host TEP) VLAN and NSX-T Edge Overlay (Edge TEP) VLAN, as it is a form of isolating the trac for each onto a separate VLAN.

       

      In this way we use separate VLANs for each cluster for the Host TEPs - so if you had three clusters you could have three separate Host TEP VLANs and one Edge TEP VLAN.

      By separating the trac onto dierent VLANs and subnets we remove a potential SPOF. e.g., if there was a Broadcast Storm in the Host TEP VLAN for one cluster it wouldn’t impact the other clusters or Edge Cluster.

      NSX-T Host Overlay (Host TEP) VLAN and NSX-T Edge Overlay (Edge TEP) VLAN must be routed to each other. so, in our case the NSX-T Host Overlay VLAN 0 is routed to NSX-T Edge Overlay VLAN 1252

      You cannot use DHCP for the NSX-T Edge Overlay (Edge TEP) VLAN.

      Note: The NSX Manager interface provides two modes for configuring resources: Policy and Manager View. For more information read the NSX-T 3.0 documentation.

      https://docs.vmware.com/en/VMware-NSX-T-Data-Center/3.0/administration/GUID-BB26CDC8-2A90-4C7E-9331-643D13FEEC4A.html#GUID-BB26CDC8-2A90-4C7E-9331-643D13FEEC4A


      To review the NSX-T Overlay configuration you may have to switch to Manager View. Click on Manager on the top right of the NSX-T Main menu if not already in Manager mode

       

       

      Now click on Logical Switches on Networking Dashboard

      Click on Edge Overlay name as dened on the VCF 4.0 excel spreadsheet, in this case sddc-edge-overlay

      The summary shows this logical switch is associated with the overlay transport zone mgmt-domain-tz-overlay01

       

       

      Click on Transport Zone to view the transport zone and "Where Used"

      To review where the VLAN 1252 is dened, click on System > Fabric > Nodes > Edge Transport nodes.
      Select an edge node and select edit. the Edge node is associated with two transport zones, and a prole that denes VLAN 1252
       

      Graphical user interface, application</p>
<p>Description automatically generated

      Click on System > Fabric > Profiles

      An uplink prole denes policies for the links from hosts to NSX-T Data Centre logical switches or from NSX Edge nodes to top-of- rack switches.
      The settings dened by uplink proles include teaming policies, active/standby links, the transport VLAN ID, and the MTU setting in our case uplink-prole-1252 has the teaming and VLAN settings dened on the uplink prole associated with the Edge transport nodes

       

      Graphical user interface, text, application, email</p>
<p>Description automatically generated

       

       

       

       

       

       

      Filter Tags

      Automation Lifecycle Management Modern Applications Networking Upgrade Cloud Foundation 4 Management Domain Workload Domain Document Proof of Concept Advanced Design Deploy Manage