I had a meeting a month or so back with an IT Director for one of the large banks in London, he’s always been a big fan of NetApp technology but is also very pragmatic about looking at all options when deciding on an Infrastructure. As with most companies I meet they are heavily invested into virtualisation, in this case with VMware, toward the end of our meeting he stated…
“as soon as VSAN is available then we’ll use it for our virtual environments along with commodity storage”
It really was right at the end of our meeting so I had no time to get into a discussion with him about it, but after the meeting I really started thinking about this and how I could or would respond and I realised very quickly that I really didn’t have a good and relatively brief response for someone at this level, every time I thought about it I ended up digging down into the details. Now that I have thought it through some more, here’s what I’m thinking
I think there are some environments where VSAN might be an interesting fit, although I think it needs significantly more capabilities than it has today, and be more than a version 1 product. However, right now, for most environments you are likely to increase the cost of your storage, not reduce it, you’ll introduce significantly more complexity, not simplify things and above all you will bring additional risk to your data and applications
Cost – If you think that using VSAN will save you money then that simply may not be the case and here’s why,
You’ll obviously need to buy more servers to hold all the data, that’s the OS’s, Applications and all the data that they create, and what you may not initially realise is that you have to create 3 copies of all of the data, maybe even 4 if you actually want to achieve 5 x 9’s of availability. In the first version of VSAN there’s also no deduplication, no compression, in fact basically no data reduction capabilities at all, and don’t forget you’ve got to pay license costs for VSAN.
Yes, you removed that pesky efficient external storage array and in doing so you just exploded your storage requirements, all of the efficiencies gone! and on top of that you also have 3 or 4 copies of this now inefficient data. I’m not convinced that this will make your storage costs go down, especially as you have to find room and power for all these additional disks you’ve just bought.
Complexity – This assumes that you currently think that an external storage array is complex, compared to what? to VSAN? I disagree
In addition to now having a lot more data and a lot more copies of it, therefore significantly more physical disks and servers to manage, you’ve now also got to work out how you’re going to protect the applications and the data they create. The value of our Storage arrays is the way that they integrate Snapshots, SnapRestore and SnapMirror capabilities into the applications in one consistent method, the challenge you now have with VSAN is none of this type of capability exists. You can indeed take a Snapshot of a VM, which consumes huge resources when you do, but it is no more than crash consistent for the application inside, so you will now need to consider how you’re going to protect these applications and not suffer from unacceptable recovery time and point objectives.
Let this sink in for a second…you now have to find a completely new way to protect the Applications and Data that was all previously taken care of by the storage array, and it has to be able to deliver Recovery Time and Point (RTO and RPO) of seconds or minutes. Do you think you can find one tool that can do this for all of your virtualised apps? if not then you’ve just massively increased your complexity and again you have to add this to the cost of the VSAN solution.
Risk – Storage is different than the other layers in the Infrastructure, the servers run the applications that create the data, the networks move it, but it lives on the storage, everywhere else its just in transit but at the storage layer we expect it to be safe for years and that brings many many different considerations.
In the latest IDC storage tracker it showed that we shipped 1,168,018 Terabytes of storage in Q3, I’m not going to do a detailed breakdown here but this obviously means we’ve shipped many many millions of disks over the 20+ years we’ve been in business. Amongst the things we’ve learnt is that drives fail in the most unusual ways, from obvious physical failures to lost writes and torn pages, ensuring that the data is protected as these events occur is a major part of what we do and these lessons learned and technologies developed simply cannot be overlooked or understated.
If a server fails it’s inconvenient, but with VMware you can just bring the virtual servers back on line on a new server, if storage fails or worse still data becomes corrupted then it’s going to have consequences, you must expect your storage environment to reduce all of the possibilities for silent corruption and offer you tools to recover any data, for any application, in a consistent state in seconds or minutes. You simply don’t get this with VSAN today and you have to decide if it’s a risk you’re prepared to take.
You are going to hear the most incredible sale pitches for VSAN, you think the industry spoke a lot about Cloud, it’s just getting warmed up with Software Defined Storage. Don’t get me wrong here, I think that VSAN has the potential to be an interesting technology in the future, but building storage arrays with software and hardware is not a trivial task, so I just wanted to provide my perspective on what it is and what it offers right now.
The next time someone tells me they are going to move to VSAN for their virtual environments, I’ll ask them why? I really want to know this because from my perspective as it stands today
You will not necessarily reduce your costs, you will significantly increase complexity and you will introduce risk, and what do you get? what benefits are you getting that would make you accept this? with the capabilities in VSAN today I would say very very little.