Back To The Future With Tintri

Disclaimer: I recently attended Storage Field Day 21.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

Tintri recently presented at Storage Field Day 21. You can see videos of the presentation here, and download my rough notes from here.

 

Tintri? 

Remember Tintri? The company was founded in 2008, fell upon difficult times in 2018, and was acquired by DDN. It’s still going strong, and now offers a variety of products under the Tintri brand, including VMstore, IntelliFlash, and NexentaStor. I’ve had exposure to all of these different lines of business over the years, and was interested to see how it was all coming together under the DDN acquisition.

 

Does Your Storage Drive Itself?

Ever since I got into the diskslinger game, self-healing infrastructure has been talked about as the next big thing in terms of reducing operational overheads. We build this stuff, can teach it how to do things, surely we can get it to fix itself when it goes bang? As those of you who’ve been in the industry for some time would likely know, we’re still some ways off that being a reality across a broad range of infrastructure solutions. But we do seem closer than we were a while ago.

Autonomous Infrastructure

Tintri spent some time talking about what it was trying to achieve with its infrastructure by comparing it to autonomous vehicle development. If you think about it for a minute, it’s a little easier to grasp the concept of a vehicle driving itself somewhere, using a lot of telemetry and little computers to get there, than it is to think about how disk storage might be able to self-repair and redirect resources where they’re most needed. Of most interest to me was the distinction made between analytics and intelligence. It’s one thing to collect a bunch of telemetry data (something that storage companies have been reasonably good at for some time now) and analyse it after the fact to come to conclusions about what the storage is doing well and what it’s doing poorly. It’s quite another thing to use that data on the fly to make decisions about what the storage should be doing, without needing the storage manager to intervene.

[image courtesy of Tintri]

If you look at the various levels of intelligence, you’ll see that autonomy eventually kicks in and the concept of supervision and management moves away. The key to the success of this is making sure that your infrastructure is doing the right things autonomously.

So What Do You Really Get?

[image courtesy of Tintri]

You get an awful lot from Tintri in terms of information that helps the platform decide what it needs to do to service workloads in an appropriate fashion. It’s interesting to see how the different layers deliver different outcomes in terms of frequency as well. Some of this is down to physics, and time to value. The info in the cloud may not help you make an immediate decision on what to do with your workloads, but it will certainly help when the hapless capacity manager comes asking for the 12-month forecast.

 

Conclusion

I was being a little cheeky with the title of this post. I was a big fan of what Tintri was able to deliver in terms of storage analytics with a virtualisation focus all those years ago. It feels like some things haven’t changed, particularly when looking at the core benefits of VMstore. But that’s okay, because all of the things that were cool about VMstore back then are still actually cool, and absolutely still valuable in most enterprise storage shops. I don’t doubt that there are VMware shops that have definitely taken up vVols, and wouldn’t get as much out of VMstore as those shops running oldey timey LUNs, but there are plenty of organisations that just need storage to host VMs on, storage that gives them insight into how it’s performing. Maybe it’s even storage that can move some stuff around on the fly to make things work a little better.

It’s a solid foundation upon which to add a bunch of pretty cool features. I’m not 100% convinced that what Tintri is proposing is the reality in a number of enterprise shops (have you ever had to fill out a change request to storage vMotion a VM before?), but that doesn’t mean it’s not a noble goal, and certainly one worth pursuing. I’m a fan of any vendor that is actively working to take the work out of infrastructure, and allowing people to focus on the business of doing business (or whatever it is that they need to focus on). It looks like Tintri has made some really progress towards reducing the overhead of infrastructure, and I’m keen to see how that plays out across the product portfolio over the next year or two.

 

 

Western Digital – The A Is For Active, The S Is For Scale

Disclaimer: I recently attended Storage Field Day 15.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

   

Western Digital recently presented at Storage Field Day 15. You might recall there are a few different brands under the WD umbrella, including Tegile and HGST and folks from both Tegile and HGST presented during Storage Field Day 15. I’d like to talk about the ActiveScale session however, mainly because I’m interested in object solutions. I’ve written about Tegile previously, although obviously a fair bit has changed for them too. You can see their videos from Storage Field Day 15 here, and download a PDF copy of my rough notes from here.

 

ActiveScale, Probably Not What You Thought It Was

ActiveScale isn’t some kind of weight measurement tool for exercise fanatics, but rather the brand of scalable object system that HGST sells. It comes in two flavours: the P100 and X100. Apparently the letters in product names sometimes do mean things, with the “P” standing for Petabyte, and the “X” for Exabyte (possibly in the same way that X stands for Excellent). From a speeds and feeds perspective, the typical specs are as follows:

  • P100 – starts as low as 720TB, goes to 18PB. 17x 9s data durability, 4.6KVA typical power consumption; and
  • X100 – 5.4PB in a rack, 840TB – 52PB, 17x 9s data durability, 6.5KVA typical power consumption.

You can scale out to 9 expansion racks, with 52PB of scale out object storage goodness per namespace. Some of the key capabilities of the ActiveScale platform include:

  • Archive and Backup;
  • Active Data for Analytics;
  • Data Forever Architecture;
  • Versioning;
  • Encryption;
  • Replication;
  • Single Pane Management;
  • S3 Compatible APIs;
  • Multi-Geo Availability Zones; and
  • Scale Up and Scale Out.

They use “BitSpread” for dynamic data placement and you can read a little about their erasure coding mechanism here. “BitDynamics” assures continuous data integrity, offering the following features:

  • Background – verification process always running
  • Performance – not impacted by verification or repair
  • Automatic – all repairs happen with no intervention

There’s also a feature called “GeoSpread” for geographical availability.

  • Single – Distributed erasure coded copy;
  • Available – Can sustain the loss of an entire site; and
  • Efficient – Better than 2 or 3 copy replication.

 

What Do I Use It For Again?

Like a number of other object storage systems in the market, ActiveScale is being positioned as a very suitable platform for:

  • Media & Entertainment
    • Media Archive
    • Tape replacement and augmentation
    • Transcoding
    • Playout
  • Life Sciences
    • Bio imaging
    • Genomic Sequencing
  • Analytics

 

Thoughts And Further Reading

Unlike a lot of people, I find technical sessions discussing object storage at extremely large scale to be really interesting. It’s weird, I know, but there’s something that I really like about the idea of petabytes of storage servicing media and entertainment workloads. Maybe it’s because I don’t frequently come across these types of platforms in my day job. If I’m lucky I get to talk to folks about using object as a scalable archive platform. Occasionally I’ll bump into someone doing stuff with life sciences stuff in a higher education setting, but they’ve invariably built something that’s a little more home-brew than HGST’s offering. Every now and then I’m lucky enough to spend some time with media types who regale me with tales of things that go terribly wrong when the wrong bit of storage infrastructure is put in the path of a particular editing workflow or transcode process. Oh how we laugh. I can certainly see these types of scalable platforms being a good fit for archive and tape replacement. I’m not entirely convinced they make for a great transcode or playout platform, but I’m relatively naive when it comes to those kinds of workloads. If there are folks reading this who are familiar with that kind of stuff, I’d love to have a chat.

But enough with my fascination with the media and entertainment industry’s infrastructure requirements. From what I’ve seen of ActiveScale, it looks to be a solid platform with a lot of very useful features. Coupled with the cloud management feature it seems like they’re worth a look. Western Digital aren’t just making hard drives for your NAS (and other devices), they’re doing a whole lot more, and a lot of it is really cool. You can read El Reg’s article on the X100 here.

Storage Field Day 6 – Wrap-up and Link-o-rama

Disclaimer: I recently attended Storage Field Day 6.  My flights, accommodation and other expenses were paid for by Tech Field Day and their sponsors. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

This is a quick post to say thanks once again to the organisers and sponsors of Storage Field Day 6. I had a great time, learnt a lot, and didn’t get much sleep. For easy reference, here’s a list of the posts I did covering the event (not necessarily in chronological order).

Storage Field Day 6 – Day 0
Storage Field Day 6 – Day 1 – Avere
Storage Field Day 6 – Day 1 – StorMagic
Storage Field Day 6 – Day 1 – Tegile
Storage Field Day 6 – Day 2 – Coho Data
Storage Field Day 6 – Day 2 – Nexenta
Storage Field Day 6 – Day 2 – Pure Storage
Storage Field Day 6 – Day 3 – Nimble Storage
Storage Field Day 6 – Day 3 – NEC
Storage Field Day 6 – (Fairly) Full Disclosure

Also, here’s a number of links to posts by my fellow delegates. They’re all really smart folks, and you’d do well to check out what they’re writing about. I’ll update this list as more posts are published.

 

Eric Shanks
Storage Field Day 6
Local Premises Storage for EC2 Provided by Avere Systems
Nimble Storage Data Analytics – InfoSight

Will All New Storage Arrays be Hybrid?

 

John Obeto
Today at Storage Field Day 6

Day 2 at Storage Field Day 6: Coho Data

Day 2 at Storage Field Day 6: Nexenta Systems

 

Arjan Timmerman
Storage Field Day Starts Today :D

 

Nigel Poulton

Nexenta – Back in da house…

 

Enrico Signoretti
Avere Systems, great technology but…

 

Chin-Fah Heoh

MASSive, Impressive, Agile, TEGILE

 

Jon Klaus
Storage Field Day 6 Day 0 – Sightseeing and Yankee swap
SFD6 Day 1 – Avere, StorMagic, Tegile

 

Finally, thanks again to Stephen, Claire and Tom, it was a great few days and I really valued the opportunity I was given to attend.

Storage Field Day 6 – Day 1 – Tegile

Disclaimer: I recently attended Storage Field Day 6.  My flights, accommodation and other expenses were paid for by Tech Field Day and their sponsors. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

For each of the presentations I attended at SFD6, there are a few things I want to include in the post. Firstly, you can see video footage of the Tegile presentation here. You can also download my raw notes from the presentation here. Finally, here’s a link to the Tegile website that covers some of what they presented.

Before the session we took a quick tour of the lab. Here’s a blurry phone snap of shiny, shiny machines.

SFD6_Tegile

Tegile spent a fair bit of time taking us through their system architecture which I found interesting as I wasn’t overly familiar with their story. You can read about their system hardware in my presentation notes. I thought for this post I’d highlight some of the features in the data services layer.

SFD6_Tegile_IntelliFlash

Data Reduction is offered via:

  • In-line deduplication
    • block level
    • dedupe across media
    • performance multiplier
  • In-line compression
    • block level
    • turn on/off at LUN / share level
    • alogrithm – LZ4, LZJB, GZIP
    • perf multiplier
  • Thin provisioning
    • array-level thin
    • for LUNs and shares
    • supports VMware VAAI “STUN”
    • JIT storage provisioning

Interestingly, Tegile chooses to compress then dedupe, which seems at odds with a few other offerings out there.

 

From a data protection perspective, Tegile offers:

  • Instantaneous thin snapshots
    • point-in-time copies of data
    • space allocated only for changed blocks
    • no reserve space for snapshots
    • unlimited number of snapshots
    • VM-consistent and application-consistent
  • Instantaneous thin clones
    • mountable copies
    • R/W-able copies
    • point-in-time copies
    • Space allocated only for deltas
  • Detect and correct silent corruption
    • checksums all data blocks
    • data and checksum in separate locations
    • match data/checksum for integrity
    • corrupt / mismatched data fixed using blocks from mirrored copy

From a data recovery perspective, the Tegile solution offers:

  • Instantaneous stable Recovery
    • data-consistent VM snapshots
    • hypervisor integrated
    • MSFT VSS co-ordinated data-consistent snapshots
    • VM-consistent and application-consistent snapshots
  • Intelligent data reconstruction
    • no need to rebuild entire drive
    • only portion of data rebuilt
    • accelerated metadata accelerates rebuilds
  • WAN-optimized replication
    • snapshot-based site-to-site replication
    • no need to replicate multiple updates to a block within the replication interval
    • minimizes bandwidth usage
    • one to many / many to one replication

Overall, I found Tegile’s presentation pretty interesting, and will be looking for opportunities to examine their products in more depth in the future. I also recommend checking out Scott D. Lowe’s article that looks into the overall theme of simplicity presented by a number of the vendors at SFD6.