Skip to content

Commit

Permalink
Additional doc lins checks
Browse files Browse the repository at this point in the history
Signed-off-by: shanrajad23 <[email protected]>
  • Loading branch information
shanrajad23 committed Feb 11, 2025
1 parent fe856e2 commit dd79268
Show file tree
Hide file tree
Showing 7 changed files with 24 additions and 24 deletions.
4 changes: 2 additions & 2 deletions src/deployment-guides/tkg-vsphere-vds-airgap.md
Original file line number Diff line number Diff line change
Expand Up @@ -497,9 +497,9 @@ For information about how to create the user and role for Tanzu Kubernetes Grid,

### Import NSX Advanced Load Balancer in Content Library

Create a content library following the instructions provided in [Create a Library](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere/8-0/vsphere-virtual-machine-administration-guide-8-0/using-content-libraries/create-and-edit-a-content-library.html#GUID-A58AF4FD-6CBE-4210-9E67-27EFBDCC1EF2-en) in VMware vSphere documentation. You will store the NSX Advanced Load Balancer OVA in the library.
Create a content library following the instructions provided in [Create a Library](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere/8-0/create-and-edit-a-content-library.html) in VMware vSphere documentation. You will store the NSX Advanced Load Balancer OVA in the library.

To import the OVA into the content library, see [Import Items to a Content Library](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere/8-0/vsphere-virtual-machine-administration-guide-8-0/using-content-libraries/how-to-populate-libraries-with-content.html#GUID-B413FBAE-8FCB-4598-A3C2-8B6DDA772D5C-en).
To import the OVA into the content library, see [Import Items to a Content Library](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere/8-0/how-to-populate-libraries-with-content.html).

## <a id=configure-alb> </a> Deploy and Configure NSX Advanced Load Balancer

Expand Down
2 changes: 1 addition & 1 deletion src/reference-designs/tko-on-vsphere-nsx.md
Original file line number Diff line number Diff line change
Expand Up @@ -280,7 +280,7 @@ The default installation of Tanzu Kubernetes Grid does not have any default ingr
Contour is an open-source controller for Kubernetes ingress routing. Contour can be installed in the shared services cluster on any Tanzu Kubernetes Cluster. Deploying Contour is a prerequisite if you want to deploy Prometheus, Grafana, and Harbor packages on a workload cluster.
For more information about Contour, see the [Contour](https://projectcontour.io/) website and [Implementing Ingress Control with Contour](https://techdocs.broadcom.com/us/en/vmware-tanzu/cli/tanzu-packages/latest/tnz-packages/packages-contour.htm).
For more information about Contour, see the [Contour](https://projectcontour.io/) website and [Implementing Ingress Control with Contour](https://techdocs.broadcom.com/us/en/vmware-tanzu/cli/tanzu-packages/latest/tnz-packages/packages-contour.html#:~:text=This%20topic%20gives%20an%20overview,the%20Envoy%20reverse%20HTTP%20proxy.).
Another option is to use the NSX Advanced Load Balancer Kubernetes ingress controller that offers an advanced L4-L7 load balancing/ingress for containerized applications that are deployed in the Tanzu Kubernetes workload cluster.
Expand Down
14 changes: 7 additions & 7 deletions src/reference-designs/tko-on-vsphere-with-tanzu-multi-az-nsx.md
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ The following table provides the component versions and interoperability matrix

- **Storage Classes in vSphere with Tanzu** : A StorageClass allows the administrators to describe the classes of storage that they offer. Different storage classes can map to meet quality-of-service levels, to backup policies, or to arbitrary policies determined by the cluster administrators. The policies representing datastore can manage storage placement of such components and objects as control plane VMs, vsphere Pod ephemeral disks, and container images. You might need policies for storage placement of persistent volumes and VM content libraries.

A three-zone Supervisor supports zonal storage, where a datastore is shared across all hosts in a single zone. Storage policies that you create for a Supervisor or for a namespace in a three-zone Supervisor must be topology aware and have the consumption domain enabled. For more information, see [Create Storage Policy for a Three-Zone Supervisor](https://author-techdocs2-prod.adobecqms.net/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/vsphere-iaas-control-plane-services-and-workloads-8-0/using-persistent-storage-with-workloads-in-vsphere-iaas-control-plane/using-persistent-storage-on-a-three-zone-supervisor.html#GUID-46783798-E3FD-49C6-BB62-C5EB09C2B82A-en).
A three-zone Supervisor supports zonal storage, where a datastore is shared across all hosts in a single zone. Storage policies that you create for a Supervisor or for a namespace in a three-zone Supervisor must be topology aware and have the consumption domain enabled. For more information, see [Create Storage Policy for a Three-Zone Supervisor](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/vsphere-supervisor-services-and-workloads-8-0/using-persistent-storage-with-workloads-in-vsphere-iaas-control-plane/using-persistent-storage-on-a-three-zone-supervisor/using-persistent-storage-on-a-three-zone-supervisor.html).

When you prepare storage resources for three-zone Supervisor, consider the following parameters:

Expand Down Expand Up @@ -104,7 +104,7 @@ vSphere with Tanzu supports the following two identity providers:

To authenticate using vCenter Single Sign-On, use vSphere plug-in for kubectl. Once authenticated, use kubectl to declaratively provision and manage the lifecycle of TKG clusters, deploy TKG cluster workloads.

- **External Identity Provider:** You can configure a Supervisor with an external identity provider and support the [OpenID Connect protocol](https://openid.net/connect/). Once connected, the Supervisor functions as an OAuth 2.0 client, and uses the [Pinniped](https://pinniped.dev/) authentication service to connect to Tanzu Kubernetes Grid clusters by using the Tanzu CLI. Each Supervisor instance can support one external identity provider. For more information about the list of supported OIDC providers, see [Configure an External IDP](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/installing-and-configuring-vsphere-iaas-control-plane-8-0/configuring-and-managing-a-supervisor-cluster/configure-an-external-idp-for-use-with-tkg-service-clusters.html).
- **External Identity Provider:** You can configure a Supervisor with an external identity provider and support the [OpenID Connect protocol](https://openid.net/connect/). Once connected, the Supervisor functions as an OAuth 2.0 client, and uses the [Pinniped](https://pinniped.dev/) authentication service to connect to Tanzu Kubernetes Grid clusters by using the Tanzu CLI. Each Supervisor instance can support one external identity provider. For more information about the list of supported OIDC providers, see [Configure an External IDP](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/7-0/configure-an-external-idp-for-use-with-tkg-service-clusters.html).

The Tanzu Kubernetes Grid (TKG) cluster permissions are set and scoped at the vSphere Namespace level. When permissions are set for Namespace, including identity source, users & groups, and roles, all these permissions apply to any TKG cluster deployed within that vSphere Namespace.

Expand Down Expand Up @@ -177,7 +177,7 @@ The following section explains the networking components and services included i
- NSX Container Plugin (NCP) provides integration between NSX and Kubernetes. The main component of NCP runs in a container and communicates with the NSX manager and with the Kubernetes control plane. NCP monitors changes to containers and other resources and manages resources such as logical ports, segments, routers, and security groups for the containers by calling the NSX API.
- By default, NCP creates one shared tier-1 gateway for system namespaces, and a tier-1 gateway and load balancer for each namespace. The tier-1 gateway for namespace is connected to the tier-0 gateway and a default segment.
- System namespaces are namespaces that are used by the core components that are integral to functioning of the Supervisor and Tanzu Kubernetes Grid clusters. The shared network resources that include the tier-1 gateway, load balancer, and SNAT IP are grouped in a system namespace.
- NSX Edge provides connectivity from external networks to the Supervisor resources. An NSX edge cluster normally includes at least two Edge nodes and has a load balancer that provides a redundancy to the Kube-API servers residing on control plane VMs and any application that must be published and be accessible from outside the Supervisor cluster. For more information, see [Install and Configure NSX for vSphere with Tanzu](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/installing-and-configuring-vsphere-iaas-control-plane-8-0/networking-for-vsphere-with-tanzu/install-and-configure-nsx-t-data-center-for-vsphere-with-tanzu.html).
- NSX Edge provides connectivity from external networks to the Supervisor resources. An NSX edge cluster normally includes at least two Edge nodes and has a load balancer that provides a redundancy to the Kube-API servers residing on control plane VMs and any application that must be published and be accessible from outside the Supervisor cluster. For more information, see [Install and Configure NSX for vSphere with Tanzu](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/7-0/vsphere-with-tanzu-configuration-and-management-7-0/networking-for-vsphere-with-tanzu/configuring-supervisor-clusters-with-nsx-t-data-center/install-and-configure-nsx-t-data-center-for-vsphere-with-tanzu.html).

- A tier-0 gateway is associated with the NSX Edge cluster to provide routing to the external network. The uplink interfaces use either the dynamic routing, BGP, or static routing.
- Each vSphere namespace has a separate network and set of networking resources shared by applications inside the namespace, such as tier-1 gateway, load balancer service, and SNAT IP address.
Expand Down Expand Up @@ -339,7 +339,7 @@ The CNI options are open-source software that provide networking for cluster pod

When you deploy a Tanzu Kubernetes cluster using the default configuration, Antrea CNI is automatically enabled in the cluster.

To provision a Tanzu Kubernetes cluster using Calico CNI, see [Deploy Tanzu Kubernetes Clusters with Calico](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/using-tkg-service-with-vsphere-iaas-control-plane-8-0/provisioning-tkg-service-clusters/using-the-cluster-v1beta1-api/v1beta1-example-cluster-with-calico-cni.html).
To provision a Tanzu Kubernetes cluster using Calico CNI, see [Deploy Tanzu Kubernetes Clusters with Calico](https://techdocs.broadcom.com/us/en/vmware-cis/aria/aria-automation/all/v1beta1-example-cluster-with-calico-cni.html).


|**CNI**|**Use Case**|**Pros and Cons**|
Expand All @@ -354,7 +354,7 @@ To provision a Tanzu Kubernetes cluster using Calico CNI, see [Deploy Tanzu Kube
vSphere with Tanzu does not ship a default ingress controller. You can use Any Tanzu-supported ingress controller. For example, Contour, an open-source controller for Kubernetes ingress routing. Contour is part of a Tanzu package and can be installed on any Tanzu Kubernetes cluster. Deploying Contour is a prerequisite for deploying Prometheus, Grafana, and Harbour on a workload cluster. You can also manually deploy AKO on the Tanzu Kubernetes cluster and make use of NSX Advanced Load Balancer as L7 ingress. However, this will require an enterprise license of NSX Advanced Load Balancer.


For more information about Contour, see [Contour](https://projectcontour.io/) and [Ingress Using Contour](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/using-tkg-service-with-vsphere-iaas-control-plane-8-0/deploying-workloads-on-tkg-service-clusters/ingress-using-contour.html).
For more information about Contour, see [Contour](https://projectcontour.io/) and [Ingress Using Contour](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/using-tkg-service-with-vsphere-supervisor/deploying-workloads-on-tkg-service-clusters/ingress-using-contour.html).

[Tanzu Service Mesh](https://tanzu.vmware.com/service-mesh) also offers an Ingress controller based on [Istio](https://istio.io/).

Expand Down Expand Up @@ -383,7 +383,7 @@ You can use one of the following methods to install Harbor:

If you are deploying Harbor without a publicly signed certificate, you must include the Harbor root CA in your Tanzu Kubernetes Grid clusters. For more information, see [Trust Custom CA Certificates on Cluster Nodes](https://techdocs.broadcom.com/us/en/vmware-tanzu/standalone-components/tanzu-kubernetes-grid/2-5/tkg/workload-clusters-secret.html#custom-ca).

To configure TKG cluster with private container registry, see [Integrate TKG 2 cluster with container registry](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/using-tkg-service-with-vsphere-iaas-control-plane-8-0/using-private-registries-with-tkg-service-clusters/integrate-tkg-service-clusters-with-a-private-container-registry.html).
To configure TKG cluster with private container registry, see [Integrate TKG 2 cluster with container registry](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/7-0/integrate-tkg-service-clusters-with-a-private-container-registry.html).

## Scale a Tanzu Kubernetes Grid Cluster

Expand Down Expand Up @@ -412,7 +412,7 @@ There are following two options for backing up and restoring stateless and state
|Velero plug-in for vSphere|<p>Both Kubernetes metadata and persistent volumes can be backed up and restored.</p><p>Velero snapshotting is used for persistent volumes with stateful applications.</p><p>Requires the Velero plug-in for vSphere installed and configured on Supervisor.</p><p></p>|
|Standalone Velero and Restic|<p>Both Kubernetes metadata and persistent volumes can be backed up and restored.</p><p>Restic is used for persistent volumes with stateful applications.</p><p>Use this approach if you require portability.</p><p></p>|

To backup and restore workloads running on TKG Cluster on Zonal Supervisor, create a datastore and install Velero with Restic on Kubernetes cluster. For more information, see [Install and Configure Standalone Velero and Restic](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/8-0/using-tkg-service-with-vsphere-iaas-control-plane-8-0/backing-up-and-restoring-tkg-service-clusters-and-workloads/backup-and-restore-workloads-using-standalone-velero-with-restic/install-and-configure-standalone-velero-and-restic-on-tkg-clusters.html).
To backup and restore workloads running on TKG Cluster on Zonal Supervisor, create a datastore and install Velero with Restic on Kubernetes cluster. For more information, see [Install and Configure Standalone Velero and Restic](https://techdocs.broadcom.com/us/en/vmware-cis/vsphere/vsphere-supervisor/7-0/install-and-configure-standalone-velero-and-restic-on-tkg-clusters.html).

>**Note** Velero plug-in for vSphere runs as a pod which is not supported with Zonal Supervisor, and it requires NSX-T networking.
Expand Down
Loading

0 comments on commit dd79268

Please sign in to comment.