I’m sure most readers will be somewhat familiar with VMware’s Project Pacific at this point. It really is the buzz of VMworld 2019. If I had to describe Project Pacific in as few words as possible, it is a merging of vSphere and Kubernetes (K8s) with the goal of enabling our customers to deploy new, next-gen, distributed, modern applications which may be comprised of container workloads or combined container and virtual machine workloads. Not only that, we also need to provide our customers with a consistent way of managing, monitoring and securing these new modern applications. This is where Project…
I have been doing quite a bit of work on First Class Disks (FCD), also known as Improved Virtual Disks (IVD) over the past number of months. One tool that has been extremely useful in improving my understanding of FCDs has been safekeeping, a tool developed by Max Daneri of VMware and which is now available to download on GitHub. If you did not know, FCDs are used extensively in VMware’s new Cloud Native Storage (CNS) offering that is currently available with vSphere/vSAN 6.7U3. Now, whilst the primary aim of this tool is to help backup vendors become familiar with…
I’ve been looking at ways in which we could query the mappings of objects between the Kubernetes layer and the vSphere layer. One thing that I really wanted to figure out is if I have the VolumeHandle from the Persistent Volume in Kubernetes, could I easily find the datastore and path using PowerCLI. It looks like I can. Let’s begin with a look at the Persistent Volume or PV for short. Note that this is a Kubernetes cluster that is using the new vSphere CSI driver.
Since the release of the vSphere CSI driver in vSphere 6.7U3, I have had a number of requests about how we plan to migrate applications between Kubernetes clusters that are using the original in-tree vSphere Cloud Provider (VCP) and Kubernetes clusters that are built with the new vSphere CSI driver. All I can say at this point in time is that we are looking at ways to seamlessly achieve this at some point in the future, and that the Kubernetes community has a migration design in the works to move from in-tree providers to the new CSI driver as well.…
Last week I had the pleasure of presenting at the French VMware User Group (VMUG) meeting in Paris. The local VMUG team did an awesome job with respect to the logistics, and even had someone on-site to record all of the sessions. The team just sent me the recording of my presentation, which I am delighted to be able to share with you. If you’re interested in learning more about vSAN 6.7U3, Kubernetes and Cloud Native Storage, this recording should help. And don’t worry – only the first few sentences are in my very terrible French 🙂 Encore merci équipe…
I put together a few short (7 – 8 minute) videos to show off some new functionality that we’ve recently added in vSphere 6.7U3, as well as our new Velero v1.1 in action. The first video is on CNS, the VMware Cloud Native Storage feature which we included in vSphere 6.7U3. This demonstration involves the deployment of a Cassandra database on Kubernetes, which incidentally uses the new CSI (Container Storage Interface) driver for persistent volumes. Once the application is deployed, we can see the characteristics of the volumes bubbled up in vSphere. We also see how using CNS, we can…
I’m delighted to be able to share with you that, coinciding with the release of vSphere 6.7 U3, VMware have also announced Cloud Native Storage (CNS). CNS builds on the legacy of the earlier vSphere Cloud Provider (VCP) for Kubernetes, and along with a new release of the Container Storage Interface (CSI) for vSphere and Cloud Provider Interface (CPI) for vSphere, CNS aims to improve container volume management and provide deep insight into how container applications running on top of vSphere infrastructure are consuming the underlying vSphere Storage. Now, there may be a lot of unfamiliar terminology in that opening…