Storagebod Rotating Header Image

Announcement Ennui

Despite my post of approval about IBM’s V7000 announcements; there’s a part of me who wonders who the hell really cares now? The announcements from IBM, HDS, NetApp and the  inevitable EMC announcement later in the year just leave me cold. The storage-array is nearly done as a technology; are we going to see much more in way of genuine innovation?

Bigger and faster is all that is left.

Apart from that, we’ll see incremental improvements in reliability and serviceability. It does seem that the real storage innovation is going to be elsewhere or down in the depths and guts; practically invisible to the end-user and consumer.

So things I expect to see in the traditional array market include a shift from a four-five year refresh cycle for centralised storage arrays to a six-seven year refresh cycle; centralised arrays are getting so large that migration is a very large job and becomes an increasingly large part of an array’s useful life. We will see more arrays offering data-in-place upgrades; replacement of the storage controllers as opposed to the back-end disk.

An Intel-based infrastructure based on common commodity components means that the internal software should be able to run on more generations of any given hardware platform.

We’ll also see more work on alternative to the traditional RAID-constructs; declustered, distributed, micro-RAIDs.

There are going to be more niche devices and some of the traditional centralised storage array market is going to be taken by the AFAs but I am not sure that market is as large as some of the valuations suggest.

AFAs will take off once we have parity pricing with spinning rust but until then they’ll replace a certain amount of tier-1 and find some corner-cases but it is not where the majority of the world’s data is going to sit.

So where is the bulk of the storage market going?

Object storage is a huge market but it is mostly hidden; the margins are wafer-thin when compared to the AFAs and it does not directly replace the traditional centralised array. It is also extremely reliant at the moment on application support. (Yes, I know everyone…I’m a broken record on this!).

The bulk of the storage market is going to be the same as it is now….DAS but in the form of ServerSAN. If this is done right, it will solve a number of problems and remove complexity from the environment. DAS will become a flexible option and no longer just tiny silos of data.

DAS means that I can get my data as close to the compute as possible; I can leverage new technologies in the server such as Diablo Memory Channel Storage but with ServerSAN, I can also scale-out and distribute my data. That East/West connectivity starts to become very important though.

Storage refreshes should be as simple as putting in a new ServerSAN node and evacuating an older node. Anyone who has worked with some of the cluster file-systems will tell you that this has become easy; so much easier than the traditional SAN migration. There is no reason why a ServerSAN migration should not be as simple.

I would hope that we could move away from the artificial LUN construct and just allocate space. The LUN has become increasingly archaic and we should be no longer worrying about queue depths per LUN for example.

There are still challenges; synchronous replication over distance, a key technology for active/active stretched clusters is still a missing technology in most ServerSAN cases. Personally I think that this should move up the stack to the application but developers rarely think about the resilience and other non-functional requirements.

And at some point, someone will come up with a way of integrating ServerSAN and SAN; there’s already discussion on how VSAN might be used with the more traditional arrays.

The storage array is certainly not dead yet and much like the mainframe, it isn’t going away but we are going to see an increasing slowing in the growth of array sales; business appetite for the consumption of storage will continue to grow at a crazy rate but that storage is going to be different.

The challenge for mainstream vendors is how they address the challenge of slowing growth and address the new challenges faced by their customers. How do they address the ‘Internet of Things’? Not with traditional storage arrays…


One Comment

  1. Steven Willson says:

    Martin, I often see clients start to plan migrations when the technology refresh is discussed, and then be forced into spending more on maintenance to complete the migration project. There are seamless migration tools for data (Violin Maestro is one), which will allow a migration of data over a few weeks, then a rapid cutover when 99.9% of the data is moved. For assets to last 6-7 years, the client needs to consider if they can get up front pricing for year 6 and 7 (most struggle to do years 4 and 5 today!). The next consideration is whether the TCO makes sense for those years. If over that period density/performance improves by 4x, and performance gains mean that the existing asset cannot deliver the business SLA, it may be more cost effective to move technology. Not all technologies lend them selves to 7 year cycles. Flash wears out, so the ability of the vendor to manage wear (and guarantee replacement), is critical. At least one of the major vendors this month told our client that after year 1, they were liable for flash SSD replacement costs! (We guarantee wear for all supported systems BTW).

    Server-side SANs are interesting, and where the workload is virtualised the migration of data/VMs is fairly straightforward, the VM will drag the associated data during migration. Interop for these new classes of solution. How does a client ,move from ScaleIO to Ceph? Do you want to migrate storage and server because storage OR server has reached it’s limit? All stuff we will learn, and maybe Openstack primitives will help handle the interop requirement for management.

Leave a Reply

Your email address will not be published. Required fields are marked *