Although I didn’t attend EMC World this year, there were a lot of interesting announcements. I managed to catch up with Matt Cowger (who sorts of sits between both the EMC & VMware camps) and ran through some of the main highlights from this year’s conference. There has been a lot written about EMC World already (and I mean a lot) so I’m going to try to keep the highlights to a minimum, and provide links to where you can read more.
In this post, we talk about a particular behaviour with using the default (or None) policy with VSAN. I have stated many times in the past that when a VM is deployed on the VSAN datastore, it behaves like it is thinly provisioned unless the capability ‘Object Space Reservation’ (OSR) is specified in the VM Storage Policy. The OSR will pre-allocate space on the VSAN datastore for the virtual machine’s storage objects, and is specified as a percentage of the actual VMDK size. However, there is a slightly different behaviour when the default policy is used. Once again, I was in a conversation with a customer who stated that when he used the default policy of “None”, he could see space being consumed on the VSAN datastore was equal to the size of the VMDK * FTT (Number of Failures To Tolerate). He wondered why this was the case when the default policy clearly did not contain an Object Space Reservation capability.
Pure Storage are all over the news at the moment. They just secured another round of funding (225 million to be precise), and are now valued at over 3 billion. You can read more about that here. However, even before this announcement, I had already arranged to have a catch up chat with Pure’s primary evangelist (and a good pal of mine), Vaughn Stewart. I was surprised to see that it had been 18 months since I last did a piece on Pure so I really did want to see what changes they had made in the meantime as there were a few vSphere interoperability pieces still to be completed when we last spoke.
Those of you familiar with VSAN will know that one of the capabilities which can be placed in a VM Storage Policy is Number of Disk Stripes Per Object (stripe width for short). I covered this in an earlier post which looked at the various VSAN capabilities. Recently, a customer who had not specified a stripe width in the VM Storage Policy was perplexed to find that his storage objects had indeed been striped across a number of disks. He reached out to me if I could provide an explanation.
Well, VSAN is finally GA today. Check out Duncan’s blog post which has lots of good links about where to get the GA bits.
In this post, I am going to address a question about the VM Home Namespace object on VSAN which has come up a number of times recently and has caused a little bit of confusion. If you’ve been following my series of Virtual SAN articles, you may recall that virtual machines deployed on a VSAN datastore are now made up as a set of objects (as opposed to the set of files that we’ve been used to traditionally). These objects may be virtual machine disk (VMDKs), snapshot deltas, VM swap space and of course the VM Home Namespace. The VM Home Namespace is where we store all the virtual machine configuration files, such as the .vmx, .log, digest files, memory snapshots, etc. Now what a number of folks have noticed is that even though they set a VM Storage Policy with various VSAN capabilities, the VM Home Namespace object does not seem to implement the policy settings when viewed via the vSphere web client. This post will aim to explain why.
I’ve been having lots of fun lately in my new role in Integration Engineering. It is also good to have someone local once again to bounce ideas off. Right now, that person is Paudie O’Riordan (although sometimes I bet he wishes I was in a different timezone ). One of the things we are currently looking at is a VSAN implementation using Fusion-io ioDrive2 cards (which our friends over at Fusion-io kindly lent us). The purpose of this post is to show the steps involved in configuring these cards on ESXi and adding them as nodes to a VSAN cluster. However, even though I am posting about it, Paudie did most of the work, so please consider following him on twitter as he’s got a lot of good vSphere/Storage knowledge to share.
Very exciting day today at VMware – Virtual SAN or VSAN has been officially announced.
If you weren’t able to attend, here are some of the interesting parts of the announcement:
- 32 node support (up from the 16 node support announced at Partner Exchange last month, and up from the 8 nodes which we supported during the beta)
- 2 million IOPS (using IOmeter 100% read, 4KB block size).
- Also 640K IOPS achieved with 70/30 read/write ratio, 4KB block size and 80% random.
- 3200 virtual machines (100 per node)
- 4.4 PB of storage (using 35 disk per host x 32 hosts per cluster)
We were also told that there was linear scaling of performance from a 4 node cluster up to a 32 node cluster.
And we saw interoperability with vSphere Replication for DR. There is also interoperability with vSphere Data Protection for backups, vMotion, DRS, HA, VMware View, etc.
It seems that we’ll GA VSAN during the week of March 10th (next week).
The one thing which wasn’t shared yet was pricing/licensing details. Guess we’ll have to wait for GA to get that info. All very exciting.