Enabling vSphere with Tanzu using HA-Proxy

In earlier posts, we look at the differences between the original “VCF with Tanzu” offering and the new vSphere with Tanzu offering from VMware. One of the major differences is the use of HA-Proxy to provide a load balancing service, and the deployment steps of the HA-Proxy we covered in detail in a follow-up post. In this post, we are now ready to deploy vSphere with Tanzu, also known as enabling Workload Management. Prerequisites Revisited The prerequisites were covered in detail in the “Getting started” post, and you won’t have been able to successfully deploy the HA-Proxy without following them.…

Getting started with vSphere with Tanzu

With the release of vSphere 7.0U1, vSphere with Kubernetes has been decoupled from VMware Cloud Foundation (VCF). VMware now has two vSphere with Kubernetes offerings, the original VCF based vSphere with Kubernetes offering, now referred to as VCF with Tanzu, and a newer offering outside of VCF, referred to as vSphere with Tanzu. This write-up is to step through the deployment of the new vSphere with Tanzu with HA-Proxy. I won’t cover everything in this single post, but will do a series of 4 posts stepping through the process. Differences: VCF with Tanzu and vSphere with Tanzu I thought it…

Failed to deploy PV to local volume – “No compatible datastore found for storagePolicy”

This is something that I “spun my wheels” on a little bit last week, so I decided I’d write a short article to explain the issue in a bit more detail. This is related to the provisioning of a Persistent Volume on the Supervisor cluster of a vSphere with Kubernetes deployment. I had a local VMFS volume on one of my hosts, so I went ahead and tagged the volume using vSphere Tagging. I then built a tag-based storage policy so that when that policy is selected for provisioning, the objects that get provisioned would be placed on that local,…

New vTopology enhancement – display CSI version

I added a few enhancements to vTopology recently, notably the ability to display the details about the version and status of the vSphere CSI driver that is being used in a Kubernetes (K8s) deployment. I tested this on native K8s, vSphere with Tanzu TKG (Tanzu Kubernetes Grid) guest clusters, standalone TKG clusters as well as TKGI (formerly Enterprise PKS). Using the -csi option, you can now get details regarding the image versions, and also checks on the ready state of the CSI controller and CSI nodes. This is the example take from a standalone TKG deployment. simply ensure that your…

Getting started with the TKGm (multi-cloud) Command Line (Videos)

In this post, I have two short videos demonstrating how to (1) deploy the Tanzu Kubernetes Grid multi-cloud (TKGm) management cluster using the “tkg” command line tool, and then once the TKG management cluster has been deployed, I show how to (2) very simply deploy a subsequent TKG workload cluster using the same “tkg” command. Note that at I have updated this post to use the TKGm acronym, as this is now how we are marketing this particular product. Previously, the term standalone was used. If you wish to know more detail, check out my full post on how to…

Tanzu Kubernetes Grid multi-cloud (TKGm) from the tkg Command Line Interface

After spending quite a bit of time looking at vSphere with Kubernetes, and how one could deploy a Tanzu Kubernetes Grid (TKG) “guest” cluster in a namespace with a simple manifest file, I thought it was time to look at other ways in which customers could deploy TKG clusters on top of vSphere infrastructure. In other words, deploy TKG without vSphere with Kubernetes, or VMware Cloud Foundation (VCF) for that matter.  This post will look at TKG multi-cloud (TKGm) version 1.1.2 and in particular the tkg command line tool to first deploy a TKG management cluster, and once that is…

Integrating embedded vSphere with Kubernetes Harbor Registry with TKG (guest) clusters

A number of readers have hit me up with queries around how they can use the integrated Harbor image repository (that comes integrated with vSphere with Kubernetes) for applications that are deployed on their Tanzu Kubernetes Grid clusters, sometimes referred to as guest clusters. Unfortunately, there is no defined workflow on how to achieve this. The reason for this is that there are a number of additional life-cycle management considerations that we need to take into account before we can fully integrate these components. This includes adding new TKG nodes to the image registry as a TKG cluster is scaled.…