Storagebod Rotating Header Image

Storage

Take Provisions

Provisioning and utilisation seem to be a perennial subject; there still seems to be a massive problem with large Enterprises who appear to have far too much storage for the amount of data that they store. There are many reasons for this and there is certainly fault on both the customer and vendor side.

Firstly, as a customer; you should expect to be able to use all the capacity that you purchase and your internal pricing models should reflect this. Obviously, there is always overhead for data-protection and hot-spares but once this is taken into account, all that capacity should be useable. You need to define what useable capacity means to you.

There should not be a performance hit for utilising all useable capacity. If the vendor states that best practise is only to use 70% of useable capacity; that needs to reflected in your TCO calcuations to give you a true cost of storage.

Secondly, as a customer; you need to ensure that your processes and operational procedures are based around provisioning the right amount of storage and not over-provisioning. Allocating storage for a rainy day is not a great idea; thin provisioning can help with this but it is not a silver bullet.

Thirdly, as a vendor; you need to be up front about the useable capacity; if you can only realistically use 70%, you need to factor this into your pricing models and show the customer exactly what they are getting for their money. Be open and honest. If you want to show a price per IOP, a price per gigabyte or some other measure; that is fine. If you want to show a price based on your assumed dedupe ration; be prepared to put your money where your mouth is.

Fourthly, as a vendor; look at ways of changing how storage is allocated and presented. It is time for us to move away from LUNs and other such archiac notions; provisioning needs to be efficient and simple. And we also need the ability to deallocate as easily as we allocate. This has often been problematic. Obviously this is not just a storage problem; how many companies are spinning up VMs and then not clearing them down properly? But make it easier across the board.

Fifthly, as vendors and users; we need to look at data-mobility. Too often, the reason that an array is under-utilised is because you have reserved capacity for application growth because it is simply ‘too hard’ to move an application’s data once it is in place. This is also a reason why many customers are very wary about thin-provisioning; the rainy day scenario again.

However, large arrays bring their own issues; from change management to refresh cycles. Smaller might be better for many but unless data can be easily moved; there is a tendency to buy arrays that are large and reserving capacity for growth.

A Ball of Destruction…

I’m not sure that EMC haven’t started an unwelcome trend; I had a road-map discussion with a vendor this week where they started to talk about upcoming changes to their architecture..my questioning ‘but surely that’s not just a disruptive upgrade but destructive?’ was met with an affirmative. Of course like EMC; the upgrade would not be compulsory but probably advisable.

The interesting thing with this one is that it was not a storage hardware platform but a software defined storage product. And we tend to be a lot more tolerant of such disruptive and potentially destructive upgrades. Architecturally as we move to more storage as software as opposed to being software wrapped in hardware; this is going to be more common and we are going to have design infrastructure platforms and applications to cope with this.

This almost inevitably means that we will need to purchase more hardware than previously to allow us to build zones of availability to allow upgrades to core systems to be carried out out as non-disruptively as possible. And when we start to dig into the nitty-gritty; we may find that this starts to push costs and complexity up…whether these costs go up so much that the whole commodity storage argument starts to fall to pieces is still open to debate.

I think for some businesses it might well do; especially those who don’t really understand the cloud model and start to move traditional applications into the cloud without a great deal of thought and understanding.

Now this doesn’t let EMC off the hook at all but to be honest; EMC have a really ropey track-record on non-disruptive upgrades in the past…more so than most realise. Major Enginuity upgrades have always come with a certain amount of disruption and my experience has not always been good; the levels of planning and certification required has kept many storage contractors gainfully employed. Clariion upgrades have also been scary in the past and even today, Isilon upgrades are no-where as near as clean as they have you believe.

EMC could have of course got away with the recent debacle if they’d simply released a new hardware platform and everyone would have accepted that this was going to involve data-migration and move data around.

Still, the scariest upgrade I ever had was an upgrade of an IBM Shark which failed half-way and left us with one node at one level of software and one at a different level. And IBM scratching their heads. But recently, the smoothest upgrades have been V7000..so even elephants can learn to dance.

As storage vendors struggle with a number of issues; including the setting of the sun on traditional data protection schemes such as RAID; I would expect the number of destructive and disruptive upgrades to increase. And the marketing spin around them from everyone to reach dizzying heights. As vendors manipulate the data we are storing in more and more complex and clever ways; the potential for disruption and destructive upgrades is going increase.

Architectural mistakes are going to be made; wrong alleys will be followed…Great vendors will admit and support their customers through these changes. This will be easier for those who are shipping software products wrapped with hardware; this is going to be much harder for the software-only vendors. If a feature is so complex that it seems magic; you might not want to use it…I’m looking for simple to manage, operate and explain.

An argument for Public Cloud? Maybe, as this will take the onus away from you to arrange. Caveat Emptor though and this may just mean that disruption is imposed upon you and if you’ve not designed your applications to cope with this…Ho hum!

 

 

 

 

Heady Potential Eventually Means Catastrophe?

Amongst the storage cognoscenti today on Twitter, there’s been quite a discussion about EMC and HP possibly merging. Most people seem to be either negative or at best disbelieving that something like this would bring value or even happen.

But from a technology point of view, the whole thing might make a lot of sense. The storage folks like to point at overlap in the portfolios but I am not convinced that this really matters and the overlap might not be as great as people think. Or at least, the overlap might well finally kill off the weaker products; I’ll let the reader decide those products that deserve to die.

EMC are on a massive push to commoditise and move their technology onto a standard platform; software variants of all their storage platforms exist and just need infrastructure to run on. I’ve mentioned before that HP’s SL4500 range is an ideal platform for many of EMC’s software defined products.

But storage aside; the EMC Federation has a lot of value for HP, it is early days for Pivotal but I suspect Meg can see a lot of potential in it. She’ll see a bit of the eBay in it; she’ll get the value of some of the stuff that they are trying to do. They are still very much a start-up, a well-funded start-up tho’.

VMware, I would expect to continue as it is; it might throw up some questions about EVO-RAIL and HP have pointedly not produced an EVO-RAIL certified stack; despite being invited to. But to fold VMware into the main HP would be rash and would upset too many other vendors. But hey, with IBM pulling out of x86 servers and honestly, who cares about Oracle’s x86 servers; HP might have a decent run at dominating the server marketplace before Lenovo takes a massive bite out of it.

And customers? I’m not sure that they’d be too uncomfortable with a HP/EMC merger; mergers are almost certainly on the agenda and there are less attractive ones on the table.

HP need software to help them build their software-defined data-centre; Openstack will only take them so far today. EMC need a commodity partner to help them build a hardware platform that would be trusted. An HP/EMC stack would be solid and traditional but with potential to grow into the 3rd platform supporting infrastructure as customers move that way.

And they both need a way of fending off Amazon and Google; this might be the way for them to do it.

I know I’ve been talking about this more like a HP take-over of EMC and it’d be closer to a true merger; this makes it harder…true mergers always are but culturally, the companies are less dissimilar than most realise. They both need more rapid cultural change…perhaps a merger might force that on them.

Will it happen, I don’t know…would it be a disaster if it did? I don’t think so. It’d also be good for the industry; lots of hacked-off smart people would leave the new behemoth and build new companies or join some of the pretenders.

A shake up is needed…this might do it. Will the market like it? I’m not especially bothered…I don’t hold shares in either company. I just think it might make more sense than people realise. 

 

ESXi Musings…

VMware need to open-source ESXi and move on; by open-sourcing ESXi, they could start to concentrate on becoming the dominant player in the future delivery of the 3rd platform.

If they continue with the current development model with ESXi; their interactions with the OpenStack community and others will always be treated with slight suspicion. And their defensive moves with regards to VIO to try to keep the faithful happy will not stop larger players abandoning them to more open technologies.

A full open-sourcing of ESXi could bring a new burst of innovation to the product; it would allow the integration of new storage modules for example. Some will suggest that they just need to provide a pluggable architecture but that will inevitably will also leave people with the feeling that they allow preferential access to core partners such as EMC.

The reality is that we are beginning to see more and more companies running multiple virtualisation technologies. If we throw in containerisation into the mix, within the next five years, we will see large companies running three or four virtualisation technologies to support a mix of use-cases and the real headache on how we manage these will begin.

I know it is slightly insane to be even talking about us having more virtualisation platforms than operating systems but most large companies are running at least two virtualisation platforms and probably many are already at three (they just don’t realise it). This ignores those with running local desktop virtualisation by the way.

The battle for dominance is shifting up the stack as the lower layers become ‘good enough’..vendors will need to find new differentiators…

 

Pay it back..

Linux and *BSD have completely changed the storage market; they are the core of so many storage products, allowing start-ups and established vendors to bring new products to the market more rapidly than previously possible.

Almost every vendor I talk to these days have their systems built on top of these and then there are the number of vendors who are using Samba implementations for their NAS functionality. Sometimes they move on from Samba but almost all of version 1 NAS boxen are built on top of Samba.

There is a massive debt owed to the community and sometimes it is not quite as acknowledged as it should be.

So next time you have a vendor in; make sure you ask the question…how many developers do you have submitting code into the core open-source products you are using? What is your policy for improving the key stacks that you use?

Now, I probably wouldn’t reject a product/company that did not have a good answer but I’m going to give a more favourable listen to those who do.

An Opening is needed

As infrastructure companies like EMC try to move to a more software oriented world; they are having to try different things to try to grab our business. A world where tin is not the differentiator and a world where they are competing head-on with open-source means that they are going to have to take a more open-source type approach. Of course, they will argue that they have been moving this way with some of their products for sometime but these have tended to be outside of their key infrastructure market.

The only way I can see products like ViPR in all it’s forms gaining any kind of penetration will be for EMC to actually open-source it; there is quite a need for a ViPR like product, especially in the arena of storage management but it is far too easy for their competitors to ignore it and subtly block it. So for it to gain any kind of traction; it’ll need open-sourcing.

The same goes for ScaleIO which is competing against a number of open-source products.

But I really get the feeling that EMC are not quite ready for such a radical step; so perhaps the first step will a commercial free-to-use license; none of these mealy mouthed, free-to-use for non-production workloads but a proper you can use this and you can put it into production at your own risk type license. If it breaks and you need support; these are the places you can get support but if it really breaks and you *really* need to to pick up the phone and talk to somone, then you need to pay.

It might that if you want the pretty interface that you need to pay but I’m not sure about that either.

Of course, I’m not just bashing EMC; I still want IBM to take this approach with GPFS; stop messing about, the open-source products are beginning to be good enough for much, certainly outside of some core performance requirements. Ceph for example is really beginning to pick-up some momentum; especially now that RedHat have bought Inktank.

More and more, we are living with infrastructure and infrastructure products that are good enough. The pressure on costs continues for many of us and hence good enough will do; we are expected to deliver against tighter budgets and tight timescales. If you can make it easier for me, by for example allowing my teams to start implementing without a huge upfront price negotiation; the long-term sale will have less friction. If you allow customers to all intents and purposes use your software like open-source; because to be frank, most companies who utilise open-source are not changing the code and could care less whether the source is available; you find that this will play well in the long-term.

The infrastructure market is changing; it becomes more a software play every week. And software is a very different play to infrastructure hardware..


			

Not So Cold Fusion?

Can SanDisk make a go of moving into the Enterprise especially with the purchase of Fusion-IO?

It is a very different market to the consumer space and although the margins are considerably higher; it brings with it many challenges. I think it’ll depend on the go-to market strategy that SanDisk take and I fully expect them to carry on the current partnership strategy that appears to work well for them at present.

I don’t see SanDisk moving into the enterprise space as a sales-organisation just yet and I suspect that the ION Accelerator line may well be dropped.

And I expect to see some very close ties being built between SanDisk and HP; HP already resell the ioDrive cards and the new 3Par 7450 array utilises SanDisk  Enterprise drives; it has been heavily pushed as a partnership between them.

Funny really that some people are touting SanDisk as the next EMC; EMC’s first really big partner and arguably the partnership that enabled EMC to grow into the behemoth that it is…was HP. Well until they fell out quite so badly.

What does this mean for the other flash-upstarts? Nothing yet, most of them are in a very different space to Fusion but there’s a whole raft of them who are currently parading themselves in none too subtle ways.

They need to get bought…they can’t all survive. And the list of suitors is rather short at the moment.  But with non-traditional vendors moving such as SanDisk acquiring; I think they’ll be feeling a little more positive.

YMMV

Storage Marketing is one of maddest and craziest parts of the technology industry; so many claims that don’t necessarily stand-up to scrutiny and pretty much all of them need to be caveated with the words

‘It Depends….’

And actually it is very important to understand that it really does depend; for example, when your flash vendor claims that they can supply flash at the price of spinning rust; they may well be making assumptions about deduplication or compression and your data.

If you are in a highly virtualised environment, you might well get a huge amount of deduplication from the operating systems..actually, even if you are not and you utilise SAN-boot, it’ll dedupe nicely. But what if you store your operating system on local disk?

What if you are already utilising compression in your database? What if your data is encrypted or pre-compressed media?

Of course this is obvious but I still find myself explaining this at times to irate sales who seem to assume that their marketing is always true…and not ‘It Depends’.

Understand your data…understand your access patterns…decide what you are trying to achieve…understand true costs…

The problem is that many of us don’t have time to carry out proper engineering tests; so I find it best to be as pessimistic as possible…I’d rather be pleasantly surprised than have an horrible shock. This means at times I am quite horrible to vendors but it saves me being really nasty later.

 

 

 

Peak-y NAS?

So it seems that IBM have finally decided to stop reselling NetApp filers and focus on their own products; I’m also waiting for the inevitable announcement that they will stop selling the rebadged Engenio products as well as there is fairly large cross-over there.

In fact there is more cross-over between IBMs own Storwize range and the Engenio range than there is between IBM’s NAS and NetApp’s NAS. So I guess we’ll probably see V5000u and V3700U announcements in the future. And if IBM really want to be a little bit nasty, they’ll push the external virtualisation capabilities of the Storwize Unified Devices and their much larger current support matrix for external devices.

But have we reached ‘Peak NAS’? I’m beginning to feel that we might have; certainly in the traditional filer-space. Sure there is growth but people simply don’t want to store their oceans of data on expensive filer devices and pretty much all of the filer devices are expensive; they are certainly more expensive than some of the commodity plays but they also are fairly close to the cost of AFAs, especially when you start looking at the inevitable software options that you have to license to make them work well.

NetApp really grew on the back of VMware; it was always amusing that when the VMware sales-team used to turn-up, it would often be with a NetApp salesman in tow and never an EMC salesman; embarrassed by the relationship it seemed. It is only in the last couple of years that there has appeared to be a closer relationship between VMware and EMC..of course VSAN means that they don’t really want to turn up with any storage partner these days.

NetApp’s challenge is going to be how to make themselves relevant again and re-invent themselves; from what I know about FlashRay, this is probably the most promising AFA technology from a mainstream vendor but they need to ship.

And they need to work out how to survive in market that is going to be driven by price…no-one is going to pay $1000s per terabyte to store home-directories and unstructured data; compressed or deduped.

I guess the same challenge that every vendor is currently struggling with…it just feels that NetApp are struggling more than most.

Announcement Ennui

Despite my post of approval about IBM’s V7000 announcements; there’s a part of me who wonders who the hell really cares now? The announcements from IBM, HDS, NetApp and the  inevitable EMC announcement later in the year just leave me cold. The storage-array is nearly done as a technology; are we going to see much more in way of genuine innovation?

Bigger and faster is all that is left.

Apart from that, we’ll see incremental improvements in reliability and serviceability. It does seem that the real storage innovation is going to be elsewhere or down in the depths and guts; practically invisible to the end-user and consumer.

So things I expect to see in the traditional array market include a shift from a four-five year refresh cycle for centralised storage arrays to a six-seven year refresh cycle; centralised arrays are getting so large that migration is a very large job and becomes an increasingly large part of an array’s useful life. We will see more arrays offering data-in-place upgrades; replacement of the storage controllers as opposed to the back-end disk.

An Intel-based infrastructure based on common commodity components means that the internal software should be able to run on more generations of any given hardware platform.

We’ll also see more work on alternative to the traditional RAID-constructs; declustered, distributed, micro-RAIDs.

There are going to be more niche devices and some of the traditional centralised storage array market is going to be taken by the AFAs but I am not sure that market is as large as some of the valuations suggest.

AFAs will take off once we have parity pricing with spinning rust but until then they’ll replace a certain amount of tier-1 and find some corner-cases but it is not where the majority of the world’s data is going to sit.

So where is the bulk of the storage market going?

Object storage is a huge market but it is mostly hidden; the margins are wafer-thin when compared to the AFAs and it does not directly replace the traditional centralised array. It is also extremely reliant at the moment on application support. (Yes, I know everyone…I’m a broken record on this!).

The bulk of the storage market is going to be the same as it is now….DAS but in the form of ServerSAN. If this is done right, it will solve a number of problems and remove complexity from the environment. DAS will become a flexible option and no longer just tiny silos of data.

DAS means that I can get my data as close to the compute as possible; I can leverage new technologies in the server such as Diablo Memory Channel Storage but with ServerSAN, I can also scale-out and distribute my data. That East/West connectivity starts to become very important though.

Storage refreshes should be as simple as putting in a new ServerSAN node and evacuating an older node. Anyone who has worked with some of the cluster file-systems will tell you that this has become easy; so much easier than the traditional SAN migration. There is no reason why a ServerSAN migration should not be as simple.

I would hope that we could move away from the artificial LUN construct and just allocate space. The LUN has become increasingly archaic and we should be no longer worrying about queue depths per LUN for example.

There are still challenges; synchronous replication over distance, a key technology for active/active stretched clusters is still a missing technology in most ServerSAN cases. Personally I think that this should move up the stack to the application but developers rarely think about the resilience and other non-functional requirements.

And at some point, someone will come up with a way of integrating ServerSAN and SAN; there’s already discussion on how VSAN might be used with the more traditional arrays.

The storage array is certainly not dead yet and much like the mainframe, it isn’t going away but we are going to see an increasing slowing in the growth of array sales; business appetite for the consumption of storage will continue to grow at a crazy rate but that storage is going to be different.

The challenge for mainstream vendors is how they address the challenge of slowing growth and address the new challenges faced by their customers. How do they address the ‘Internet of Things’? Not with traditional storage arrays…