Storage Field Day 8 – Wrap-up and Link-o-rama

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

This is a quick post to say thanks once again to Stephen, Claire and the presenters at Storage Field Day 8. I had a fantastic time and learnt a lot. For easy reference, here’s a list of the posts I did covering the event (not necessarily in chronological order).

Storage Field Day – I’ll be at SFD8

Storage Field Day 8 – Day 0

Storage Field Day 8 – (Fairly) Full Disclosure

Cohesity – There’s more to this than just “Secondary Storage”

Violin Memory – Sounds a lot better than it used to

Pure Storage – Orange is the new black, now what?

INFINIDAT – What exactly is a “Moshe v3.0”?

Nimble Storage – InfoSight is still awesome

Primary Data – Because we all want our storage to do well

NexGen Storage – Storage QoS and other good things

Coho Data – Numerology, but not as we know it

Intel – They really are “Inside”

Qumulo – Storage for people who care about their data

 

Also, here’s a number of links to posts by my fellow delegates. They’re all switched-on people, and you’d do well to check out what they’re writing about. I’ll try and update this list as more posts are published. But if it gets stale, the SFD8 landing page has updated links.

 

Ray Lucchesi (@RayLucchesi)

Coho Data, the packet processing squeeze and working set exploits

Primary data’s path to better data storage presented at SFD8

PB are the new TB, GreyBeards talk with Brian Carmody, CTO Inifinidat

 

Mark May (@CincyStorage)

Can Violin Step back From the Brink?

Storage software can change enterprise workflow

Redefining Secondary Storage

 

Scott D. Lowe (@OtherScottLowe)

IT as a Change Agent: It’s Time to Look Inward, Starting with Storage

Overcoming “New Vendor Risk”: Pure Storage’s Techniques

So, What is Secondary Storage Cohesity-Style?

Data Awareness Is Increasingly Popular in the Storage Biz

 

Jon Klaus (@JonKlaus)

Storage Field Day – I will be attending SFD8!

Wow it’s early – Traveling to Storage Field Day 8

Coho Data: storage transformation without disruption

Pure Storage: Non Disruptive Everything

Cohesity is changing the definition of secondary storage

Qumulo: data-aware scale-out NAS

Nimble Storage – InfoSight VMVision

NexGen Storage: All-Flash Arrays can be hybrids too!

Infinidat: Enterprise reliability and performance

 

Alex Galbraith (@AlexGalbraith)

Looking Forward to Storage Field Day 8

Without good Analytics you dont have a competitive storage product

How often do you upgrade your storage array software?

Where and why is my data growing?…

Why are storage snapshots so painful?

 

Jarett Kulm (@JK47TheWeapon)

Storage Field Day 8 – Here I come!

 

Enrico Signoretti (@ESignoretti)

#SFD8, it’s storage prime time!

Analytics, the key to (storage) happiness

We are entering the Data-aware infrastructure era

Has the next generation of monolithic storage arrived?

Juku.beats 25: Qumulo, data-aware scale-out NAS

Infinidat: awesome tech, great execution

Juku.beats 27: NexGen Storage, QoS made easy.

Software defined? No no no, it’s poorly defined storage (and why Primary Data is different)

Juku.beats 28 – Infinidat storage: multiple nine resiliency, high performance, $1/GB

Are you going All-Flash? Nah, the future is hybrid

 

Vipin V.K. (@VipinVK111)

Tech Field Day Calling…! – #SFD8

Infinibox – Enterprise storage solution from Infinidat

Understanding NVMe (Non-Volatile Memory Express)

All-Flash symphony from Violin Memory

Cohesity – Secondary storage consolidation

With FLASH, things are changing ‘in a flash’ !?

 

Josh De Jong (@EuroBrew)

Storage Field Day Here I Come!

Thoughts in the Airport

NexGen Storage – The Future is Hybrid

Pure Storage – Enterprise Ready, Pure and Simple

 

Finally, thanks again to Stephen, Claire (and Tom in absentia). It was an educational and enjoyable few days and I really valued the opportunity I was given to attend.

SFD8_Group

 

NexGen Storage – Storage QoS and other good things

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the NexGen Storage presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the NexGen Storage website (registration required) that covers some of what they presented.

 

NexGen-RGB1-250x63

NexGen Storage presented recently at SFD8 and while the video footage above covered off on some flash basics, when the camera stopped they talked to us about their latest array offering – the N5.

 

Hardware

NexGen claim that the N5 is the first multi-tier, AFA with Storage Quality-of-Service (QoS). They were recently awarded US Patent 9,176,708 for Storage QoS, so they might be on to something. NexGen say that the array can prioritise performance-hungry workloads for more predictable performance while providing enhanced flash management for both performance and endurance

NexGen_Storage

[image courtesy of NexGen Storage]

There are a couple of configuration options:

N5-1500 Head (2.6TB PCIe Flash, 15TB SSD) + you can add up to 3 SSD Capacity Pack(s) with

  • 15TB SSD
  • 4 x 6Gbps SAS

So you can get up to 60TB SSD (RAW).

The N5-3000 offers 2.6TB flash, 30TB RAW (base) and 60TB RAW (max).

The controllers have the following specs:

CPU: 4x 6-core Intel Xeon E5645 2.4GHz, 24x cores, 48x cores with Hyper-Threading
RAM: 96GB
Network: Data (4) 1/10GbE SFP+ -or- (4) 1/10GBT RJ45, iSCSI / mgmt: (4) 1GbE RJ45, http, https.

 

Storage QoS and other good things

Putting a bunch of different types of flash, including PCIe flash (NVMe “ready”), SSDs and RAM in a box is pointless if you can’t do something sensible with it. This is where NexGen’s Dynamic QoS comes into play. As you’re probably aware, QoS is about putting priorities to work targets with automated throttling. The idea is you want things to work a certain way without necessarily having everything suffer because of a noisy neighbour or IOPS hungry VM. The array comes with both preconfigured policies and the ability to manage performance minimums.

The QoS priorities offer the following features:

  • Adaptive BW throttling;
  • Adaptive queueing placement;
  • Real-time, always on; and
  • Prioritised active caching.

 

Closing Thoughts and Further Reading

NexGen Storage have been around a while and certainly have some good pedigree and experience behind them. I’m interested to see how these new arrays perform in the real world, because they certainly look the goods on paper. It was only a matter of time before someone took a “hybrid” approach to AFAs and gave some configuration options back to the end user. I can’t comment on how effective the implementation is, but I think it’s worthy of further investigation. Finally, you can read more about the new product at Storage Review.

Intel – They really are “Inside”

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Intel presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Intel website that covers some of what they presented.

 

1000px-Intel-logo_svg_-196x130

If you’ve spent any time shooting the breeze with me, you’ll probably know I’m into punk music. I may have also gotten into a monologue about how much I enjoy listening to the Dead Kennedys. I have a vinyl copy of Jello Biafra‘s third spoken-word album, “I blow minds for a living“. This is a great album, and I recommend listening to it if you haven’t already. While this is a somewhat tortured segue, what I’m trying to say is that a few of the guys working at Intel seem to also specialise in blowing minds for a living, because I walked out of that presentation at SFD8 with very little understanding of what I’d just seen :)

 

Intel is working hard so you don’t have to

There’s a whole lot to the Intel presentation, and I hearty recommend you watch it for yourself. I found Nate Marushak’s part of the presentation, “Enabling the storage transformation – Intel ISA-L & SPDK” particularly interesting. As I stated previously, I didn’t really keep up with a lot of it. Here are a few of the notes I was able to take.

Intel are keen to address the bottleneck pendulum with a few key pieces of technology:

  • 25/50/100GbE
  • Intel 3D XPoint
  • RDMA

They want to “enable the storage transformation” a couple of ways. The first of these is the Storage Performance Development Kit (SPDK), built on Data Plane Development Kit (DPDK) it provides

  • Software infrastructure to accelerate the packet IO to Intel CPU

Userspace Network Services (UNS)

  • TCP/IP stack implemented as polling, lock-light library, bypassing kernel bottlenecks, and enabling accessibility

Userspace NVMe, Intel Xeon / Intel Atom Processors DMA and Linux AIO drivers

  • optimises back-end driver performance and prevents kernel bottlenecks from forming at the back end of the IO chain

Reference Software and Example Application

  • Intel provides a customer-relevant example application leveraging ISA-L, with support provided on a best-effort basis

SPDK

What is Provided?

  • Builds upon optimised DPDK technology
  • Optimised UNS TCP/IP technology
  • Optimised storage target SW stack
  • Optimised persistent media SW stack
  • Supports Linux OS

How does it help?

  • Avoids legacy SW bottlenecks
  • Removes overhead due to interrupt processing (use polling)
  • Removes overhead due to kernel transitions
  • Removes overhead due to locking
  • Enables greater system level performance
  • Enables lower system level latency

Intel Intelligent Storage Acceleration Library

This is an algorithmic library to address key storage market segment needs:

  • Optimised libraries for Xeon, Atom architectures
  • Enhances performance for data integrity, security / encryption, data protection, deduplication and compression
  • Has available C language demo functions to increase library comprehension
  • Tested on Linux, FreeBSD, MacOS and Windows Server OS

ISA-L Functions include

  • Performance Optimisation
  • Data Protection – XOR (r5), P+Q (r6), Reed-solomon Erasure Code)
  • Data Integrity – CRC-T10, CRC-IEEE (802.3), CRC32-iSCSI
  • Cryptographic Hashing – Multi-buffer: SHA-1, SHA-256, SHA-512, MD5
  • Compression “Deflate” – IGZIP: Fast Compression
  • Encryption

 

Closing Thoughts and Further Reading

As I stated at the start of this post, a lot of what I heard in this presentation went way over my head. I urge you to check out the Intel website and links above to get a feel for just how much they’re doing in this space to make things easier for the various vendors of SDS offerings out there. If you think about just how much Intel is inside everything nowadays, you’ll get a good sense of just how important their work is to the continued evolution of storage platforms in the modern data centre. And if nothing else you might find yourself checking out a Jello Biafra record.

 

IMG_2453

 

Qumulo – Storage for people who care about their data

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Qumulo presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Qumulo website that covers some of what they presented.

 

logo_qumulo_fullcolor_use_on_light

Qumulo is a 3.5 year old company from Seattle, WA that officially launched in March 2015. Interestingly, the founders invented OneFS and Isilon scale-out NAS. They pointed out that nowadays managing storage isn’t the big problem, it’s managing the data. Qumulo want to “[b]uild invisible storage that makes data visible”.

Qumulo_Invisible_Storage

Ostensibly, Qumulo’s QSFS is a software-only layer that can work either with Qumulo’s appliances, roll your own hardware, VMs or cloud.

 

Qumulo sells storage to people that care a great deal about their data

Qumulo stated that their key design goals were to be:

  • Hardware agnostic;
  • Flexible;
  • Fast and scalable; and
  • File and object aware.

They want to achieve this by building “data-aware, primary scale-out storage systems that provide real-time analytics to give visibility into data usage and performance at scale”. Sounds like big ambitions, but Qumulo seem confident they can pull it off.

The solution is ideally suited to commercial HPC and large-scale unstructured data environments, with real-time analytics that provides the ability to curate and manage the data.

Qumulo have taken a flash-first hybrid design approach with their hardware. In their opinion this maximises both price / performance and price / capacity ratio. I tend to agree with this approach, given the relative sluggishness that we’ve seen with regard to the drop in price of flash.

Qumulo also employ a SaaS-type software delivery model, leveraging a subscription model not perpetual licenses to provide pay-as-you-go access to “continual software innovation”.Everything about the solution is 100% programmable, with a public and self-documenting REST API with interactive API explorer built-in to the product. Interestingly, the development and release cycle is tight, with new iterations of the platform being released every 2 weeks.

If you’re into hardware, you might be interested in reading some more about the Qumulo QC24 Hybrid Storage and QC208.

QC24-Hardware

[image courtesy of Qumulo]

You can check the specs for yourself here, but think of the QC24 as the fast one and the QC208 as the big, dense one. Of note, Qumulo state that the supported cluster size is 4 – 1000 nodes. I think someone asked them about that and it hadn’t been tested with physical nodes yet. But the argument from Qumulo was that it was technically supported. Connectivity is also slightly different, with the QC24 Hybrid model sporting 2 * 10GbE SFP+ and the QC208 sporting 4 x 40GbE QSFP+. Cache is also slightly different to support the different capacities. Fine, so they’re not really the same.

 

Closing Thoughts and Further Reading

Qumulo state that they do “100% sales through channel”. I’ll be interested to see how long that model lasts. It’s something I always like to hear, as I work for a VAR, but oftentimes the reality is never quite as expected as the vendor grows and seeks new revenue opportunities.

In any case, as Qumulo pointed out in their presentation, we’re witnessing a shift in enterprise storage, moving from

  • Hardware-based to software-based;
  • Proprietary operating environments to linux-based OS platforms;
  • Hard drives -> flash-first hybrid appliances;
  • Petabytes of data to billions of files / objects; and
  • Storage management to data management.

Qumulo seem well-positioned, on paper at least, to flourish in this new world of enterprise storage. I like their approach to data rather than storage management. I’m really interested to see how they go over the next 12 months, particularly with the channel-only model and the rapid software development cycles.

INFINIDAT – What exactly is a “Moshe v3.0”?

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the INFINIDAT presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the INFINIDAT website that covers some of what they presented.

 

If you’ve been around the traps, you’ve likely heard of Moshe Yanai. He had an awful lot to do with EMC’s Symmetrix for an awfully long time, and went on to fund and lead XIV, which was sold to IBM. He knows a lot about storage, particularly at the big end of town, and he’s back for another crack.

INFINIDAT states that their “modest goal [is to] store humanity’s knowledge, forever”. These are some pretty lofty ambitions, to be sure. Someone at the presentation referred to this as “Moshe v3.0”. I thought that was a nice idea, but wondered what it really meant in light of what I knew about Symm and XIV.

 

Infinibox Architecture

INFINIDAT have stated that their design principles are as follows:

  • provide the right infrastructure for the current and future data explosion;
  • add onboard capabilities to leverage and optimise the use of data;
  • provide superior mechanisms for data protection;
  • overthrow existing paradigms of storage-application integration; and
  • support unlimited use cases by eliminating cost constraints.

Great. So that all seems like a reasonable list of goals, and if you’re going to have a shot, why not go big?

 

INFINIDAT uses commodity hardware to power its offering, with Dell PowerEdge R730s being the choice of platform here. The architecture employs the following key specs:

  • Triple-redundant power (N+2 redundant on everything) and data path
  • Three active-active-active nodes (Linux servers clustered together with InfiniBand point-to-point connections)
  • 24 x 8Gbps FC ports
  • 6/12 x 10 Gbps Ethernet ports
  • Up to 3.2TB RAM
  • Up to 48TB SSD
  • Up to 480 disk drives
  • Support 3, 4 and 6TB NL-SAS drives

Also noteworthy is that the disk enclosures are connected to DC power, not the internal system UPS. There are eight disk enclosures. These have two PSUs, two I/O modules (SAS expander) and house 60 top-loading drives. You can have up to 480 6TB drives deployed with one of these things. This seems like a lot of spinning disks. In addition to this, every node can see every disk.

INFINIDAT_Connectivity

Not only is there a lot of available performance in these systems, they’ve also given some thought to the type of resiliency you’d expect to see in a high-end storage array. You can read more about INFINIDAT’s storage architecture here (registration required).

 

Closing Thoughts and Further Reading

While every man and their dog is working for a storage startup that’s based on an architecture that provides massive scale-out, cloud-optimized, analytic-laden, cheap-as-chips storage services for the masses, INFINIDAT have taken a different approach entirely, and really gone for the high-end of town. Given Moshe’s background, it’s really no surprise that this is the sort of product they’ve come out with. Nonetheless, it will be interesting to see how they compete in the VMAX and XIV space, and whether the approach will pay off. El Reg has a pretty decent write-up on them here that I recommend checking out as well. Stephen also has a post here that is worth your time.

If you’re in the market for a new approach to big-iron storage, INFINIDAT might be just the ticket and would be worth looking into. I’m really looking forward to seeing how they go over the next 12 – 18 months.

 

 

Coho Data – It’s numerology, but not as we know it

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Coho Data presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Coho Data website that covers some of what they presented.

 

coho_logo_final_orangeanddarkgrey-wpcf_100x36

Coho Data wants to be “a data-centric bridge for evolving enterprise infrastructure”, by delivering “storage transformation without disruption for the modern enterprise”.

Andy Warfield (@AndyWarfield) is a really smart dude. And he does a really great presentation. I encourage you to check out the videos from SFD8, because I can’t really do justice to a lot of the content he presented this time around. Indeed, I had the same problem last time I encountered Andy.

IMG_2360

 

It’s numerology, but not as we know it

Andy spent some time talking about the concept of “workload numerology”. No, not that kind of numerology. This is more about the ability to understand workloads hosted on your storage platform – “it’s not just a matter of speeds and feeds”.

Coho Data are right into the concept of being able to manage the placement of both data and network traffic based on detailed workload analysis. They do this in a few different ways, and I recommend you read their architecture white paper for more information.

Andy pointed out that with traditional storage, you were worried about durability, and getting high performance off crappy hardware. However “modern storage design is about solving a connectivity and locality problem, rather than a durability problem”. As such, we need to be taking a new approach to the design and operation of these platforms.

Andy also noted that performance based placement decisions benefit from workload characterisations, but characterising things like working sets are expensive in time and space. You can, however, use counter stacks to efficiently encode the cardinality of uniqueness over time. You can read more about Hyper Log Logs “HLLs” here.

In short, understand the working set characteristics and you’ll get a lot more value from your platform.

 

Closing Thoughts and Further Reading

Ray Lucchesi (@RayLucchesi) did a much better job of covering Andy’s presentation than I ever would, and I strongly encourage you to go read it. If you watch the Coho presentation from SFD6 and compare it to the one by Andy at SFD8, you’ll see that, not only has Andy gotten really good at keeping the discussion on track, but also Coho has made some really good progress with the DataStream platform in general. I’m looking forward to seeing what the next 12 months brings for Coho.

Nimble Storage – InfoSight is still awesome

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Nimble Storage presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Nimble Storage website that covers some of what they presented.

 

400px-Nimble_logo

Nimble Storage have been around for a while now, and I put up an enthusiastic post about their InfoSight product after hearing about it at Storage Field Day 6. The cool thing is they’ve been working hard at improving what was already an impressive offering.

 

InfoSight is still awesome

Rod Bagg provided an overview of InfoSight. Nimble have spent a lot of time working on what they call “Operational Intelligence”. They asked a few pointed questions of their products:

  • “In a connected world why can’t vendors proactively monitor customer deployed systems?”
  • “With modern data analytic tools can vendors predict and prevent problems before they occur?”

Nimble’s design philosophy for InfoSight is as follows:

  • Be intuitive – present use cases and not just data
  • Be prescriptive – provide specific recommendations for immediate action
  • Be predictive – estimate future needs based on current and past learning

Rod then shared some “fun facts” about the statistics they’ve been collecting:

Deep Data

  • 1000s of unique sensors recording operational data each second,
  • 30-70M data points collected from every array every day,
  • >20000000 heartbeats every week

Big Data

  • 200B log events
  • by-day view of every config element of every array
  • lifetime data from day 1

Rich Analytics

  • rich install base
  • data from 1000s of arrays for 5+ years
  • dedicated team of data scientists on support staff
  • advanced analytics techniques,

 

My favourite thing about all of this is the idea that you can use InfoSight as a key part of protecting your investment. If your vendor has access to statistics about 1000s of deployed systems, why wouldn’t you use them to help you with the following exercises?

Workload Sizing

Nimble has access to “1000s of system-years” of real-world data. They also have an understanding of workloads correlated to resource consumption. In this fashion you can aim to understand the exact configuration before you make your purchase.

Predictive Capacity Recommendations

InfoSight also provides you with the ability to perform continuous storage capacity prediction, and plan your storage purchase in advance. This is invaluable when working in environments where budgets are allocated at fixed points in time.

Scale-to-fit Recommendations

As Nimble has a whole bunch of data on workload and behaviour across a number of systems, they can help you with the working-set analysis for cache and CPU optimisation. With this information you’ll also be able to understand the optimal cache, scale-up and scale-out requirements before you deploy the system.

 

Closing Thoughts and Further Reading

I was excited about InfoSight when I first saw it in action, and I remain an enthusiastic advocate for this approach to understanding your storage environment. I love the idea of taking a lot of the guesswork out of platform sizing, in addition to making good use of the available data. While the Nimble Storage hardware isn’t for everyone, I encourage you to have a look at them if you’re in the market for a hybrid array, simply by virtue of the fact that the InfoSight product has the potential to provide a valuable insight into what your storage is doing on a daily basis.

IMG_2457

 

Primary Data – Because we all want our storage to do well

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Primary Data presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Primary Data website that covers some of what they presented.

 

Primary Data presented at Storage Field Day 7 – and I did a bit of a write-up here. At the time they had no shipping product, but now they do. Primary Data were big on putting data in the right place, and they’ve continued that theme with DataSphere.

 

Because we all want our storage to do well

I talk to my customers a lot about the concept of service catalogues for their infrastructure. Everyone is talking about X as a Service and the somewhat weird concept of applying consumer behaviour to the enterprise. But for a long time this approach has been painful, at least with mid-range storage products, because coming up with classifications of performance and availability for these environments is a non-trivial task. In larger environments, it’s also likely you won’t have consistent storage types across applications, with buckets of data being stored all over the place, and accessible via a bunch of different protocols. The following image demonstrates nicely the different kinds of performance levels you might apply to your environment, as not all applications were created equal. Neither are storage arrays, come to think of it.

PD_menu

[image courtesy of Primary Data]

Primary Data say that “every single capability and characteristic of the storage system can be thought of in terms of whether the data needs it or not”. As I said before, you need to look at your application requirements in terms of both:

  • Performance (reads vs writes, IOPS, bandwidth, latency); and
  • Protection (data durability, availability, security).

Of course, this isn’t simple when you then attempt to apply compute requirements and network requirements to the applications as well, but let’s just stick with storage requirements for the time being. Once you understand the client, and the value of the objectives being met on the data, you can start to apply Objectives and “Smart Objectives” (Objectives applied to particular types of data) to the data. With this approach, you can begin to understand the cost of specific performance and protection levels. Everyone wants Platinum, until they start having to pay for it. These costs can then be translated and presented as Service Level Agreements in your organisation’s service catalogue for consumption by various applications.

 

Closing Thoughts and Further Reading

Primary Data has a lot of smart people involved in the product, and they always put on one hell of  whiteboard session when I see them present. To my mind, the key thing to understand with DataSphere isn’t that it will automatically transform your storage environment into a fully optimised, service catalogue enabled, application performance nirvana. Rather, it’s simply providing the smarts to leverage the insights that you provide. If you don’t have a service catalogue, or a feeling in your organisation that this might be a good thing to have, then you’re not going to get the full value out of Primary Data’s offering.

And while you’re at it, check out Cormac’s post for a typically thorough overview of the technology involved.

 

 

Pure Storage – Orange is the new black, now what?

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Pure Storage presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Pure Storage website that covers some of what they presented.

 

Pure

I wrote about Pure Storage’s new hardware platform, Pure1 and Evergreen Storage program when they were announced earlier this year. Pure’s presentation this time around wasn’t earth shattering, but they did cover off on two things that I hadn’t looked at in too much depth previously.

 

Hardware needs to be resilient

While we spend a lot of time in the industry talking about how software-defined storage is the future, a big part of Pure’s success has been the focus on developing a resilient, supporting hardware platform. To that end, Ryan Oler and Larry Touchette did a demo of the FlashArray//m’s hardware resiliency, and took us through some scenarios where components have failed / been removed and what happens as a result. As an aside, here’s a photo of Ryan looking extremely chuffed with the Pure logo silkscreened on a FlashArray//m.

 

IMG_2369

So, when you pull a flash module (one flash module contains two SSDs), you lose about 4 – 5% of your effective capacity (there’re 20 modules in the chassis). To see what’s happening on the environment you can run “puredrive list” to get the status of any drive / module faults. When you put the flash modules back in different slots the array checks what data is still good to use. The good news here is that, in terms of recovery, the array “essentially self heals”.

When an NV-RAM module is pulled, you’ll see short spike in latency. When this happens, it writes DDR data out to internal NAND chips before it loses power.

Another useful command to run is “purehost monitor –balance”. This does a brief I/O sample on the array, which is a good way to look at host traffic balance and check that I/O is being evenly distributed across your available paths.

 

All of this information is pointless if you can’t make sense of it.

In much the same way that Nimble Storage are making some good use of monitoring and hardware analytics, Pure Storage are taking a similar approach with Pure1.

For Pure customers, call home data automatically generates (proactive) support tickets. Approximately 70% of the tickets in Pure are auto-generated. Pure are also making an “[a]ggressive effort to correlate and reduce noise from automatically created alerts”. They do this by processing incoming logs for known “fingerprints” to help prioritise the delivery of fixes.

Pure take the approach that “any manual process at scale is error-prone”. It’s for this reason that they want to automate the upgrade process on the arrays they sell so they can do it at scale and without human error. If you’ve ever been in a data centre at 4 in the morning on a call with a vendor’s support team after a failed array upgrade you’ll appreciate the sentiment behind this approach. This also feeds Pure’s mantra of

  • Never a disruptive upgrade
  • Never required a data migration
  • No customer left behind
  • Rapid innovation cycle

And they seem very keen on upholding this.

One of the coolest things about Pure1 Manage is that it’s SaaS. Pure can keep adding new features and capabilities without the customer needing to go around and update stuff. Pure demoed some new features that had been added since the product went GA, including adding an “Analytics” tab to the dashboard that customers could use for capacity planning.

 

Closing Thoughts and Further Reading

“Other” Scott Lowe had a good article on Pure Storage’s evolution that I urge you to read. I’ve been excited by what Pure has been doing for the last little while, and look forward to seeing them take things to the next level post-IPO. While people seemed to have a bit to say about the success or otherwise of Pure’s recent IPO, I think they’ve hired a lot of smart people and seem to have a solid vision and determination to execute on their current strategy. While this particular presentation didn’t set the world on fire, I think there was enough there to indicate that they’re on the right track. I’d encourage you to check them out if you’re in the market for some flash storage in your DC.

Violin Memory – Sounds a lot better than it used to

Disclaimer: I recently attended Storage Field Day 8.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD8, there are a few things I want to include in the post. Firstly, you can see video footage of the Violin Memory presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Violin Memory website that covers some of what they presented.

 

Violin-Memory-Logo-Full-Color-on-White-Backround-PMS431-PMS369

Violin Memory (NYSE: VMEM) have been around for about 10 years. I’ve never had any real stick time with the kit, but I work in a small part of the world, so I know a few people who’ve been customers or who’ve worked for the company. There was a time when Violin Memory were the toast of the town. While the scuttlebutt in recent years had them tanking spectacularly, they’re still of the opinion that they’re very much in the game, and have spent the last 2 years on a significant overhaul of their core architecture and the way they go about getting things done in what is a pretty competitive flash storage market.

 

Violin Flash Fabric

Violin_FFA-Architecture-Graphic

James Bowen spent some time during the presentation taking us through some of the key design elements of the hardware platform and I thought they bear repeating here. One of the key components of the highly available system design is the “Violin Flash Fabric”, which has

  • Multiple paths between each Violin Intelligent Memory Module (VIMM) and vRAID Control Module (VCM); and
  • The VIMM Tree dynamically reconfigures to handle component failures and / or upgrades.

The Flash Fabric Architecture uses the following protection methods:

  • Multipath VIMM Fabric;
  • VCM Failure Protection; and
  • VIMM Failure protection.

What’s cool about this architecture is that the Fabric can handle the failure of up to 3 VCMs and 4 VIMMs, and with Violin’s vRAID in the mix it can tolerate up to 16 VIMM failures without data loss. In addition to that, Violin have really focussed on performance in the system and are claiming some pretty high numbers, some of which, admittedly, reside in the “4K Vanity Zone“.

 

Closing Thoughts and Further Reading

The full white paper on Flash Fabric Architecture is here. I urge you to read it, as I’ve done a bit of a ham-fisted job covering it in this post. It’s also worth checking out this page on the VIMM, which does a good job of explaining some of the benefits of that particular architecture. Also, check out Mark’s post on Violin here – it provides a nice, balanced view of things.

If you watch the presentation that Violin Memory gave at SFD8, the overwhelming theme was that they’ve got a fair bit of experience in flash architecture, they know how to get decent performance from their hardware, and they’ve learnt a few lessons along the way. In my opinion, Violin have taken a number of steps in recent times to set themselves on the path to success again, and I’m looking forward to seeing them continue on that path in the future.