All Posts in “SolidFire”

A different take on Storage Archetypes

Lest we throw fuel onto yet another twitter war, I will say thanks to Chad Sakac for inspiring this post, some of what I am about to expound upon was directly lifted from Chad’s blog on understanding storage architectures, while some of this will show disagreement with the way in which he divided the archetypes up I think the essence of what his goal was will still show through.

A different take:
Chad took an approach of 4 storage types, which he has since refined and updated. His take is depicted like this:

I agree that putting storage types in buckets is important both from an industry perspective as well as from customer understanding and decision making viewpoint. The K.I.S.S. principal needs to apply to all things we do. If you are unfamiliar K.I.S.S. stands for Keep It Simple Stupid! So let’s try and make this as simple as possible.

In my view storage breaks into the following categories (admittedly I would reverse Scale Up and Scale Out because of architecture preconceived connotations but I don’t want to stray too far from what is becoming understood by the industry)

1) Storage Islands
2) Scale Up, Dual controller, add shelves for capacity, stuck on performance
3) Scale Up, Hub & Spoke
4) Scale Out, add performance and or capacity linearly as needed
5) Bring your own Platform (BYOP)

StorageArchitypes

Sorry I don’t have a graphics department to help with art.

Let’s get into what these are, Storage Islands are just that, they are dedicated storage solutions for specific use cases, either DASD or dedicated storage appliances purpose built. These storage types tend to require a SATADOM, a Mezzanine card or some other proprietary connection technology at the host\compute layer. Storage islands are seen all over the place, High Performance Compute, High transactional databases, isolated environments that sit in both unnecessarily and totally necessarily air gapped networks.

Scale Up, Dual controllers are the traditional storage arrays that we all have used and know, these solutions are limited by the controller heads as to the way in which you can grow performance, capacity can be gained up to the disk shelf limits. These architectures are versatile and general purpose. Scale Up dual controllers can work for organizations that consume IT in planned growth segments, top down controls and need various protocols.

Scale Up, hub and spoke this is what Chad called scale out tightly coupled, the thing is, this architecture is much closer to traditional scale up. Controller sets are tied to storage, while mix and match is possible there are limits on rapid scale and on the ability to scale capacity or performance, the economics of scale issue is hit hard here which limits use cases at large scale. The way the storage inter-communicates is typically tied to a unique networking plane to that storage type. In this configuration datasets may stripe across multiple arrays and arrays may form clusters, typically we see caps on either storage scale or performance scale which restrict overall use cases. Scale Up hub and spoke work for environments that have a single or small scope use case and need speed and scale architectures that may require scale at off refresh intervals.

Scale Out, true scale out architectures are when performance or capacity are independent and can grow linearly as needed. There are very few architectures that fall into this category without caveats. The use cases tend to fall into organizations that require easy scaling, limited dependencies, global efficiencies and ability to work with multiple workloads. Scale out architectures are leveraged in multi-modal environments where performance is a concern for some workloads, most solutions have protocol limitations but provide other unique characteristics in order to provide a near share nothing architecture that doesn’t limit scaling performance or capacity independently.

Bring your own platform, this is a very broad bucket, but it needs to be this is where we find things like Ceph and VSAN. I have called this true GLOM storage in the past but it’s the concept of software defined storage controls and taking disparate hardware and layering a storage technology on the top of it. This is an emerging market and has it’s own draw backs. BYOP is one that is wide open, this can be for the cost conscious or the tinkerers out there that are playing with a project but don’t want to mess with storage management in the traditional sense, but it’s hardly set it and forget it and requires proper planning to grow at scale and implement for success, a lot of the time BYOP really should be buy your own hardware that has been certified to run “X” solution.

Part 2 is up next explaining who I believe fits in each bucket.

EUC Day 2016

I had an awesome time attending and presenting at the Minnesota EUC Day. Getting the opportunity to finally present my EUC evolution maps and show off mixed workloads on SolidFire was a lot of fun. Check out my presentation:

Thanks to Matt Helstab for putting it all together, check out the other presentations and videos here.

Also there was a great ask the experts panel at the end of the day that I was honored to be a part of.

 

 

The Dawn of the Year of EUC

So I have been a bit quieter than normal lately, it’s because I have been working go figure. The new role is a little different than I expected but challenging and rewarding in it’s own way. Now we get to talk about what I have been working on a little, also upfront apologies for length.

The year of VDI is gone we aren’t doing that anymore so stop talking about it. What we do seePresentation1 in the industry is the upward trend in end user solutions moving to virtualized accessed resources. That means not just VDI, but end user applications, mobile device management, user experience management, and the strengthening of DaaS offerings. All of this combines to into what is deemed EUC right?

Of course you are right.

That’s the easy part, defining what it is that I am working to support the tougher part becomes supporting that the market is really moving beyond VDI. For that I looked towards Simon Wardley and Wardley Mapping tools, this is an awesome tool.

Evolution

What we look for when we map a technology or industry is first to lay out the value chain, this is a high level representation and is still evolving.

Representatively we see the market starting to shift to more cloud based applications and solutions, well less starting and more continuing. Guess who has a sweet spot for cloud service providers? You guessed it the place where I work ;). In addition we see a push towards DaaS, while technologies are somewhat different we do see some similarities in infrastructure design for shared services, so how does that translate for the enterprise not ready to give up their desktops?

Evolution breeds integration success, isolation is what occurs when something is new or has a degree of unknowns. When VMware and Citrix were first introduced to environments their clusters sat in the corner of the datacenter (ok server closets) like ginger headed classmates not quite accepted by the rest of the servers. Overtime these solutions became mainstream and became core to the design and deployment strategies of pretty much everyone. VDI isn’t by any stretch new, but we isolated from our production workloads because of the resource demand and randomness of workload. Even if vSphere and Xen could handle isolating the production server environment from the desktops we had the issue of the draw on our shared storage environment. How could we have both sit together without impacting the performance of either one?

Hold on a tic, we actually should first start by examining what requirements we expect to see in today’s datacenter when it comes to storage solutions that do not require silo’d management and can support a higher operational cost of ownership through ease of use, scalability and flexibility.

Before I turn this into a marketing post, let’s first think about how we manage this on larger deployable storage solutions. Let’s say an EMC VMAX, not a cost effective VDI solution, but feasible with enough SSDs to handle the workload. In a VMAX config multiple virtual pools would be carved out with policies for each to meet the workload requirements. As cool as the VMAX is it’s not the easiest to scale, nor manage, and the policy sets are not quickly adjusted. You may find yourself adding engines and disk shelves to meet requirements. So while feasible we are lacking some of the requirements we laid out earlier.

VNX similarly has to use RAID pools coupled with virtual storage pools to form the data efficiencies necessary to run VDI. With VNX you have to look at total IOPs and what the FAST VP pool has as assigned FAST cache for the EUC workload. It’s possible to do a mixed workload if you ensure there is another dedicated SSD raid group to pin workloads to but that’s not a guaranteed quality of service feature, as you are sizing and creating unique virtual storage pools to meet requirements and scaling this solution requires more disks or potentially a whole new system.

What about the AFA options?

Great question.

Thanks.

Different AFA solutions handle things in different ways, essentially taking metadata and writing it in either fixed or variable block sizes. There is no doubt that the flash technology is capable of delivering performance for EUC, but what about mixed workloads? Well it comes down to sizing like with most of the traditional storage systems. The market solution is pretty much if you want to throw mixed workloads your AFA can handle it just make sure you have enough capacity and IO. But none of that guarantees that you are going to meet your performance requirements for any given workload. I’ve talked about the need to for corporate governance and SLA’s in enough posts now but they are the crux to IT strategic success. While these other solutions may ease of a piece of this stack here or there they all lack the control of IO demand.

SolidFire on the other hand has taken a unique approach as a company the goal was to NewVolprovide cloud service providers with a storage solution that meets the needs of multi-tenant environments to meet performance and reporting requirements that are unique to that market segment. You see the CSPs live and die by SLA’s, that’s how they get paid with uptime and performance being constantly measured. While some solutions have the ability to throttle max IO, SolidFire clusters can also manage the minimum and burst, this means you are ensuring that the minimum IOPs are always being delivered. This means you can run and Oracle DB right next to a EUC workload, right next to a Mongo DB deployment, without any of them impacting the performance of the other.

Sure so the guaranteed QoS is sweet, but we want to scale, scalability is the big thing. Well sure you do, what do you want to scale capacity or performance? Is it a matter of needing to ingest more blocks and write the metadata faster or is it a matter of storing more data? SolidFire scales these independently with mix and match nodes for more performance or denser SSDs. Any node, any cluster kid that’s how you have flexibility and meet todays build DC’s at scale needs.

So we have answered the scalability and the performance questions but what about ease of use?

Well SolidFire has some pretty awesome capabilities here too, with API functionality out the ying-yang. Essentially if you want to automate it we got you covered. Not to mention the ability to leverage a pretty user friendly (for storage admin perspective) HTML based management console. Plugins are also available for service consoles in OpenStack and VMware environments. Microsoft is being considered.

But wait there is one other super cool feature that is something enterprises are going to want to look at and that’s replication. Think of how most data replication works, you take the dataset do a comparison between the primary and secondary location and validate that the data is the same block by block, bit by bit. But what if you are storing data based off of a metadata hash? Do you rehydrate that data to it’s native format and validate it across the WAN. But what if that was just a matter of validating the hash on both sides instead? Think of the reduction in data transmission and the speed at which replication would occur.