Yes, it’s that time of year again. VMworld 2018 in Las Vegas is taking place next month. Once again, as per previous years, I will be co-presenting on a few sessions. It will be no surprise that, once again, my sessions are focused on storage, hyper-convergence and predominantly vSAN. However, this year I will also be presenting with my CTO and VMware Fellow, Christos Karamanolis, for the very first time. Without further ado, let me go through my sessions in a bit more details, and if they look interesting to you, feel free to sign up. I also included a…
As I am going to be doing some talks around next-gen applications at this year’s VMworld event, I took the opportunity to revisit Pivotal Container Services (PKS) to take a closer look at how we can set persistent volumes on container based applications. Not only that, but I also wanted to leverage the vSphere Cloud Provider feature which is part of our Project Hatchway initiative. I’ve written about Project Hatchway a few times now, but in a nutshell this allows us to create persistent container volumes on vSphere storage, and at the same time set a storage policy on the…
I am in the very fortunate position of having access to a Pure Storage array, and this has been recently updated to support Virtual Volumes. With my new 6.7 vSphere cluster, I finally found some time to take a closer look at Virtual Volume (VVol) snapshots on the Pure array, something that I have been meaning to do for some time. For those of you who are new to Virtual Volumes (VVols), one of the major advantages is the granularity at which certain operations can now be done. In the past, we were always dealing with data services at the…
As part of my preparation work for VMworld 2018, I was looking into how one might be able to automate the deployment of VMs from vRealize Automation 7.4 with an appropriate policy for consuming the under-lying storage. In my case, this underlying storage was vSAN 6.7 (and vSphere 6.7), so I wanted to be able to select a vSAN policy for my VM Home namespace and disks. Fortunately, there is a vRealize Orchestration (vRO) plugin to do just that called the vRA SPBM Integration plugin. But, once I had this plugin installed, and followed the instructions, I hit a snag.…
I wanted to follow-up on my recent Minio S3 post with steps on how to implement a reverse-proxy using Nginx. The purpose of this is to allow an end-user to connect to a single Minio server, and have that connection be redirected in a round-robin fashion to all of my other 16 Minio servers in my Minio S3 deployment. This was surprisingly very straight-forward, and only required a handful of changes to my nginx.conf file. If you want to review the initial deployment steps, you can find these here in my original post. Let’s go through the steps to set…
Some time back, I looked at what it would take to run a container based Minio S3 object store on top of vSAN. This involved using our vSphere Docker Volume Server (aka Project Hatchway, and the details can be found here. However, I wanted to evaluate what it would take to scale out the Minio S3 object store on top of vSAN, paying particular attention to features like distribution and availability, and to examine the various data services that can be provided by both vSAN and Minio. I also wanted to take advantage of the new host-pinning feature in vSAN…
After receiving a number of queries about vSphere Fault Tolerance on vSAN over the past couple of weeks, I decided to take a closer look at how Fault Tolerant VMs behave with different vSAN policies. I wanted to take a look at two different policies. The first is when the “failures to tolerate” (commonly referred to as FTT) is set to 0, and the other is when the “failures to tolerate” is set to 1. The question is whether or not we could deploy VMs without any vSAN protection and allow Fault Tolerant VMs to protect them instead.