Back To The Future With Tintri

Disclaimer: I recently attended Storage Field Day 21.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

Tintri recently presented at Storage Field Day 21. You can see videos of the presentation here, and download my rough notes from here.



Remember Tintri? The company was founded in 2008, fell upon difficult times in 2018, and was acquired by DDN. It’s still going strong, and now offers a variety of products under the Tintri brand, including VMstore, IntelliFlash, and NexentaStor. I’ve had exposure to all of these different lines of business over the years, and was interested to see how it was all coming together under the DDN acquisition.


Does Your Storage Drive Itself?

Ever since I got into the diskslinger game, self-healing infrastructure has been talked about as the next big thing in terms of reducing operational overheads. We build this stuff, can teach it how to do things, surely we can get it to fix itself when it goes bang? As those of you who’ve been in the industry for some time would likely know, we’re still some ways off that being a reality across a broad range of infrastructure solutions. But we do seem closer than we were a while ago.

Autonomous Infrastructure

Tintri spent some time talking about what it was trying to achieve with its infrastructure by comparing it to autonomous vehicle development. If you think about it for a minute, it’s a little easier to grasp the concept of a vehicle driving itself somewhere, using a lot of telemetry and little computers to get there, than it is to think about how disk storage might be able to self-repair and redirect resources where they’re most needed. Of most interest to me was the distinction made between analytics and intelligence. It’s one thing to collect a bunch of telemetry data (something that storage companies have been reasonably good at for some time now) and analyse it after the fact to come to conclusions about what the storage is doing well and what it’s doing poorly. It’s quite another thing to use that data on the fly to make decisions about what the storage should be doing, without needing the storage manager to intervene.

[image courtesy of Tintri]

If you look at the various levels of intelligence, you’ll see that autonomy eventually kicks in and the concept of supervision and management moves away. The key to the success of this is making sure that your infrastructure is doing the right things autonomously.

So What Do You Really Get?

[image courtesy of Tintri]

You get an awful lot from Tintri in terms of information that helps the platform decide what it needs to do to service workloads in an appropriate fashion. It’s interesting to see how the different layers deliver different outcomes in terms of frequency as well. Some of this is down to physics, and time to value. The info in the cloud may not help you make an immediate decision on what to do with your workloads, but it will certainly help when the hapless capacity manager comes asking for the 12-month forecast.



I was being a little cheeky with the title of this post. I was a big fan of what Tintri was able to deliver in terms of storage analytics with a virtualisation focus all those years ago. It feels like some things haven’t changed, particularly when looking at the core benefits of VMstore. But that’s okay, because all of the things that were cool about VMstore back then are still actually cool, and absolutely still valuable in most enterprise storage shops. I don’t doubt that there are VMware shops that have definitely taken up vVols, and wouldn’t get as much out of VMstore as those shops running oldey timey LUNs, but there are plenty of organisations that just need storage to host VMs on, storage that gives them insight into how it’s performing. Maybe it’s even storage that can move some stuff around on the fly to make things work a little better.

It’s a solid foundation upon which to add a bunch of pretty cool features. I’m not 100% convinced that what Tintri is proposing is the reality in a number of enterprise shops (have you ever had to fill out a change request to storage vMotion a VM before?), but that doesn’t mean it’s not a noble goal, and certainly one worth pursuing. I’m a fan of any vendor that is actively working to take the work out of infrastructure, and allowing people to focus on the business of doing business (or whatever it is that they need to focus on). It looks like Tintri has made some really progress towards reducing the overhead of infrastructure, and I’m keen to see how that plays out across the product portfolio over the next year or two.



VMware – VMworld 2019 – HBI3516BUS – Scaling Virtual Infrastructure for the Enterprise: Truths, Beliefs and the Real World

Disclaimer: I recently attended VMworld 2019 – US.  My flights and accommodation were paid for by Digital Sense, and VMware provided me with a free pass to the conference and various bits of swag. There is no requirement for me to blog about any of the content presented and I am not compensated by VMware for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.

These are my rough notes from “HBI3516BUS – Scaling Virtual Infrastructure for the Enterprise: Truths, Beliefs and the Real World” was a sponsored panel session hosted by George Crump (of Storage Switzerland fame) and sponsored by Tintri by DDN. The panellists were:

JP: Hyper-V is not really for the enterprise. Configuration, and automation were a challenge. Tintri made it easier to deal with the hypervisor.

JD: You put a bunch of disks and connect it up to what you want to. It’s really simple to setup. “Why would you want to go complex if you didn’t have to?”

MB: When we had block storage, we were beholden to the storage team. We’ve never had problems with their [Tintri’s] smallest hybrid arrays.

AA: Back in the ESX 2.5 days – single LUN per VM. We would buy our arrays half-populated – ready to grow. We’re now running 33 – 34 devices. Tintri was great with QoS for VMs. It became a great troubleshooting tool for VMware.

GC: Reporting and analytics with Tintri has always been great.

MB: We use Tintri analytics to create reports for global infrastructure. Tintri will give you per-VM allocation by default. Performance like a Tivo – you can go back and look at analytics at a very granular level.

GC: How did the addition of new arrays go with Global Center?

MB: We manage our purchases based on capacity or projects. 80 – 85% we consider additional capacity. Global Center has a Pools function. It does a storage vMotion “like” feature to move data between arrays. There’s no impact.

JP: We used a UCS chassis, Tintri arrays, and Hyper-V hypervisor. We used a pod architecture. We knew how many users we wanted to host per pod. We have 44000 users globally. VDI is the only thing the bank uses.

AA: We’re more of a compute / core based environment, rather than users.  One of the biggest failings of Tintri is that it just works. When you’re not causing problems – people aren’t paying attention to it.

MB: HCI in general has a problem with very large VMs.

AA: We use a lot of scripting, particularly on the Red Hat (RHV) side of things. Tintri is fixing a lot of those at a different level.

GC: What would you change?

JP: I would run VMware.

MB: The one thing that can go wrong is the network. It was never a standardised network deployment. We had different network people in different regions doing different things.

JP: DR in the cloud. How do you do bank infrastructure in the cloud? Can we DR into the cloud? Tested Tintri replicating into Azure.

AA: We’re taking on different people. Moving “up” the stack.

Consistency in environments. It’s still a hard thing to do.


  • Containers
  • A Virtual Appliance



Some folks get upset about these sponsored sessions at VMworld. I’ve heard it said before that they’re nothing more than glorified advertising for the company that sponsors the session. I’m not sure that it’s really any different to a vendor holding a four day conference devoted to themselves, but some people like to get ornery about stuff like that. One of my favourite things about working with technology is hearing from people out in the field about how they use that technology to do their jobs better / faster / more efficiently.

Sure, this session was a bit of a Tintri fan panel, but I think the praise is warranted. I’ve written enthusiastically in the past about how I thought Tintri has really done some cool stuff in terms of storage for virtualisation. I was sad when things went south for them as a company, but I have hopes that they’ll recover and continue to innovate under the control of DDN.

When everything I’ve been hearing from the keynote speakers at this conference revolved around cloud-native tools and digital transformation, it was interesting to come across a session where the main challenges still involved getting consistent, reliable and resilient performance from block storage to serve virtual desktop workloads to the enterprise. That’s not to say that we shouldn’t be looking at what’s happening with Kubernetes, etc, but I think there’s still room to understand what’s making these bigger organisations tick in terms of successful storage infrastructure deployments.

Useful session. 4 stars.

Tintri Announces Centralised Upgrades


Tintri recently announced centralised upgrades for users of Tintri Global Center (TGC). I normally wouldn’t get too excited about minor innovations from storage vendors, but I do get a little dizzy when I hear about storage vendors making life easier for the hapless storage admin. In this case, if you’re using TGC you can leverage a new feature of TGC that allows you to bulk select storage arrays that you’re managing and set them to upgrade.

This probably isn’t a major issue if you’re running one or two arrays, but if you have 8 or 16 under your watch (or up to 64 per TGC), then this is going to save you a bit of time at the console.



I remember when I started out with midrange storage arrays that the process to upgrade them was tedious at best and oftentimes went pear-shaped thanks to odd behaviour with Java or mis-typed commands at a console. The process to perform the upgrade often ran to tens of pages and involved an awful lot of pre-flight checks. The worst part of broken upgrades was the requirement to trudge to the data centre to find out what was up with that, and if you were lucky, you had a sophisticated enough connectivity solution that your storage vendor could access the array and fix things remotely.

Thankfully, the days of less than seamless array upgrades with 100s of steps are behind us. Instead, most all of the vendors have introduced automated mechanisms to deliver a painless upgrade process that can be performed during the day. Tintri have taken this philosophy a step further and made it easier to do this at scale. I’m all for vendors introducing technology that means I don’t have to perform repetitive tasks, particularly when it comes to mundane operational activities like storage operating environment upgrades. Are A Happy Tintri Customer

Disclaimer: This is a sponsored post and you’ll probably see the content elsewhere on the Internet. Tintri provided no editorial input and the words and opinions in this post are my own.


I recently had the opportunity to speak to Brock Mowry (CTO of about the company’s experience adopting Tintri in their environment. You can read the case study on Tintri’s website, but sometimes it’s nice to get a perspective straight from the source. If you don’t know of, they were established in 2013 and deliver a “[c]ybersecure cloud hosting platform with an emphasis on compliance workloads, [including] HIPAA regulation and PCI regulation”. They have a data centre presence in Las Vegas, NV and Miami, FL and plans to expand that footprint.



I asked Mowry what one of the main challenges was as a growing cloud service provider and he said “[s]torage was one of the challenges”. The problem, it seems, was when they looked at how much time they spent on keeping the environment running, there was a lot of operational overhead with their storage platform, and they “didn’t want to be scaling by head count – [they] wanted to scale by technology”.


What solutions did they look at?

According to Mowry, at they “optimise [the] network for NFS traffic and get really, really good results operating NFS in [the] infrastructure. Again, Tintri being an NFS-based platform, it was really an easy choice from there”. The benefit of deploying an IP-based storage solution was that they were “able to eliminate an entire fibre channel fabric within [the] data centre”. The added benefit of this was that they were able to reduce the number of “employees that are required to operate that platform. That’s a huge cost saving for [them] because at the end of the day head count is typically one of the most expensive things to operate a cloud infrastructure”.


Why not look at hyperconverged solutions then?

It turns out they looked at a number of hyperconverged vendors, including solutions from Nutanix and Cisco. At the time they ran across a problem with the converged nature of the resources in hyperconverged environments. Mowry provided an example where there was a “need to increase […] CPU and RAM capacity to meet a customer’s workload. Well now I’m sitting on a bunch of excess storage that I really don’t want to power, I really don’t want to cool, and I really don’t want to manage, because it’s not needed”. Note that a number of vendors now offer solutions to that problem, with “storage-only” nodes being available to counter the requirement to scale memory, CPU and storage in equal amounts. At the time, however, Mowry felt that it was best to go with what he describes as a “broken-out” architecture, where they “have storage arrays or storage appliances and [they] have UCS blade systems so [they] can increase RAM and increase CPU to the customer’s workloads without having to scale out our storage at the same time where it might not be used”.


Why go All-Flash? have deployed both All-Flash and hybrid arrays, because, as Mowry points out, they “have customers who are demanding that lower tier. And a lot of times they’re trying to hit a price point, they’re not trying to hit a performance point”.


Conclusion are obviously very happy Tintri customers, but not simply because the Tintri arrays they’ve deployed give them per-VM control or nice APIs to use with their own products. Vendors often focus on the technical advantages of the solutions they sell, because they think that’s what demonstrates value to their (potential) customers. But discussions around decreasing operational overhead and improving configuration simplicity by removing fibre channel fabrics are real world examples of how businesses can, in some instances, save money and improve their bottom line by choosing an architecture that aligns well with their operational strengths and experience. People are normally the most expensive part of any type of managed service, so if you can deploy efficient systems that don’t need a lot of people to run them, you’ll be in a good place.

Of interest also was the decision to continue with a decoupled infrastructure architecture that provided them with a solution that scales the way they want it to. In my opinion this a great example of a business choosing a solution that suits them because of a number of reasons, not all of which are technical. Customers like provide a great example of how to understand your requirements (from both a technical and financial perspective), understand your market, and work to your strengths. You can download a full transcript of my chat with Mowry from here.

Tintri FlexDrive Goes GA – Is A Very Handy Feature

Disclaimer: This is a sponsored post and you’ll probably see the content elsewhere on the Internet. Tintri provided no editorial input and the words and opinions in this post are my own.

Tintri recently announced the general availability of FlexDrive, a storage expansion feature for EC6000 all-flash arrays that gives you the ability to increase system capacity to meet a specific storage need by adding drives incrementally. This feature is included at no charge with Tintri’s 4.4.1 release.


What is It?

With FlexDrive, you can purchase a partially populated EC6000 all-flash array (with as few as thirteen drives) and add as little as one drive at a time to increase the capacity. Once the drive’s been added, all you need to do is click on the “Expand” button in the management interface and you’re all set. It’s a non-disruptive activity that you can do yourself, so there’s no need for drawn out change control meetings or extensive planning for support staff to be on-site to deploy the capacity. Imagine working with storage infrastructure that no longer ties you to the requirement to deploy flash drives as if they were spinning disks? Instead of scrambling to deploy additional rack and power capacity that you don’t really need, you can slot in a drive and get what you need out of the equipment, rather than dreading the idea of a capacity upgrade.


Who’s the Estimator?

Many all-flash systems leverage data reduction technologies such as deduplication and compression to provide effective resource utilisation in a dense footprint. So how do you know how the addition of one drive to your array will impact the capacity and performance of the array? Tintri has incorporated an expansion estimator into its standard user interface so that you can model the impact of adding drives. The estimator also uses historical system workload profiles to help you get to the right number of extra drives, ensuring you get the necessary number to meet your target capacity. With the estimator by your side, you not only have clarity around what the future holds, you also get the benefit of some really smart analytics ensuring you get the performance and capacity outcome you need based on what you’ve used the array for previously.


The Benefits are Real

Storage provisioning has been a somewhat clumsy process for years. Anyone familiar with deploying terabytes of disk storage will also be painfully familiar with the various vendors’ requirements to deploy disks in certain, minimum numbers to satisfy configuration and performance requirements. This approach was extremely important when spinning disk ruled the world. Those days are behind us though, with Flash-dominant arrays becoming the norm in a data centres all over the world. When you have Flash as your primary medium, you don’t necessarily have to deploy a lot of it to meet your performance requirements. So why buy five more disks when you only need the capacity of two? It’s a waste of money, not just in terms of the asset, but also the additional costs, such as power, cooling and (possibly) additional rack space. All to meet a requirement that keeps systems from the early part of the century happy. With FlexDrive you don’t need to do this, and nor should you have to. The estimator adds a level of intelligence to the expansion activity that has previously been an exercise in guesswork and hope. Tintri aren’t interested in making you buy trays of disk just because that’s the way it’s always been done. Tintri are interested in you getting the solution that matches your requirements in terms of capacity and performance. This approach yields real benefits when it comes to budget allocations and controlling costs for your storage environment. You no longer have to spend over the odds-on storage to get the performance you need for the next six months.

Tintri ChatOps – Because All I Do Is Hang Out On Slack Anyway


I’m a bit behind the times with my tech news, but Tintri sent me a link to a video they did demonstrating their new “ChatOps” feature. I was going to make fun of it, but it’s actually pretty neat. If you’ve used Slack before, you probably know it’s got a fairly extensible engine that you can use to do a bunch of cool things. With ChatOps, you can send your Tintri arrays commands and things get done. Not only does it do stuff for you, it does them in a sensible / efficient fashion as well. And since I spend a lot of time on Slack in any case, this feature just might take off.

You can read more about this and some other new features from Tintri at El Reg. And I agree with Chris that a focus by Tintri beyond table stakes is a smart move.

Testing Tintri’s Lightning Lab and Pizza

Disclaimer: I was offered a pizza to write this post.  I haven’t taken up the offer yet, but I will be.


I had the opportunity to test drive Tintri’s “Lightning Lab” about six months ago and the nice folks at Tintri thought I might like to post about my experiences. They’ve offered me a pizza for my troubles which, coincidentally, ties in nicely with their current promotion “The Tintri Pizza Challenge“. If you’re in the US or Canada it’s worth checking it out.

In any case, the Lightning Lab is Tintri’s internet accessible lab that showcases a number of its arrays and provides you with an opportunity to take their gear for a spin. From a hardware perspective it’s pretty well provisioned, with T5060, T880, T620 & T540 arrays, along with a Dell R720 host with 128GB of RAM and 2 Dell R610 servers with 48GB of RAM. From a software perspective, the version of the lab I used had VMware vSphere 5.5U2b installed, but I believe this has been since updated. There’s also a functional version of Tintri Global Center, and both the Web Client Plug-in and the vROps plugin configured. Networking wise, management runs overs a 1GbE Dell switch, with Data travelling via a 10GbE Arista switch.


Global Center has a pretty neat login screen. Like all good admins, I use many dots in my password too.


There’s a bunch of stuff I could show from the interface, but one of my favourite bits is the ability to see an aggregated view of your deployed VMstores.


The interface is simple to operate and painfully colourful too. It’s also simple to navigate and makes it really easy to get a quick view of what’s going on in your environment without having to do a lot of digging.



There’s a lot more I could write about Tintri. If you’re aligned with their use case (NFS-only), they have a compelling offering that’s worth checking out. The Lightning Lab is an excellent tool to take their platform for a spin and gain a good understanding of just what you can do with the VMstore and Global Center. I think these kind of offerings are great, and not just because there’s pizza involved. If more storage vendors read this and think that they should be doing something like this, then that’s a great thing. I’ve barely scratched the surface, so you should head over to Andrea Mauro’s blog and check out his thorough write-up of his Lightning Lab experience.

Tintri Keep Doing What They Do, And Well

Disclaimer: I recently attended Storage Field Day 10.  My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event.  Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.


Before I get into it, you can find a link to my notes on Tintri‘s presentation here. You can also see videos of the presentation here.

I’ve written about Tintri recently. As recently, in fact, as a week before I saw them at SFD10. You can check out my article on their most recent product announcements here.


VAS but not AAS (and that’s alright)

Tintri talk a lot about VM-aware Storage (or VAS as they put it). There’s something about the acronym that makes me cringe, but the sentiment is admirable. They put it all over their marketing stuff. They’re committed to the acronym, whether I like it or not. But what exactly is VM-aware Storage? According to Tintri, it provides:

  • VM-level QoS;
  • VM-level analytics;
  • VM data management;
  • VM-level automation with PowerShell and REST; and
  • Supported across multiple hypervisors (Support VMware, Hyper-V, OpenStack, RedHat).

Justin Lauer, Global Evangelist with Tintri, took us through a demo of VAS and the QoS capabilities built in to the Tintri platform.


I particularly liked the fact that I can get a view of end to end latency (host / network / storage (contention and flash) / throttle latency). In my opinion this is something that people have struggled with for some time, and it looks like Tintri have a really good story to tell here. I also liked the look of the “Capacity gas gauge” (petrol for Antipodeans), providing an insight into when you’ll run out of either performance, capacity, or both.

So what’s AAS then? Well, in my mind at least, this is the ability to delve into application-level performance and monitoring, rather than just VM-level. And I don’t think Tintri are doing that just yet. Which, to my way of thinking, isn’t a problem, as I think a bunch of other vendors are struggling to really do this in a cogent fashion either. But I want to know what my key web server tier is doing, for example, and I don’t want to assume that it still lives on the datastore that I tagged for it when I first deployed it. I’m not sure that I get this with VAS, but I still think it’s a long way ahead of where we were a few years ago, getting stats out of volumes and not a lot else.


Further Reading and Conclusion

In the olden days (a good fifteen years ago) I used to struggle to get multiple Oracle instances to play nicely on the same NT4 host. But I didn’t have a large number of physical hosts to play with, and I had limited options when I wanted to share resources across applications. Virtualisation to slice up physical resources in a more concise fashion, And as a result of this it’s made it simple for us to justify running one application per VM. In this way we can still get insights into our applications from understanding what our VMs are doing. This is no minor thing when you’re looking after storage in the enterprise – it’s a challenge at the best of times. Tintri has embraced the concept of intelligent analytics in their arrays in the same way that Nimble and Pure have started really making use of the thousands of data points that they collect every minute.

But what if you’re not running virtualised workloads? Well, you’re not going to get as much from this. But you’ve probably got a whole lot of different requirements you’re working to as well. Tintri is really built from the ground up to deliver insight into virtualised workloads that has been otherwise unavailable. I’m hoping to see them take it to the next level with application-centric monitoring.

Finally, Enrico did some more thorough analysis here that’s worth your time. And Chris’s SFD10 preview post on Tintri is worth a gander as well.


Tintri Announces New Scale-Out Storage Platform

I’ve had a few briefings with Tintri now, and talked about Tintri’s T5040 here. Today they announced a few enhancements to their product line, including:

  • Nine new Tintri VMstore T5000 all flash models with capacity expansion capabilities;
  • VM Scale-out software;
  • Tintri Analytics for predictive capacity and performance planning; and
  • Two new Tintri Cloud offerings.


Scale-out Storage Platform

You might be familiar with the T5040, T5060 and T5080 models, with the Tintri VMstore T5000 all-flash series being introduced in August 2015. All three models have been updated with new capacity options ranging from 17 TB to 308 TB. These systems use the latest in 3D NAND technology and high density drives to offer organizations both higher capacity and lower $/GB.


The new models have the following characteristics:

  • Federated pool of storage. You can now treat multiple Tintri VMstores—both all-flash and hybrid-flash nodes—as a pool of storage. This makes management, planning and resource allocation a lot simpler. You can have up to 32 VMstores in a pool.
  • Scalability and performance. The storage platform is designed to scale to more than one million VMs. Tintri tell me that the  “[s]eparation of control flow from data flow ensures low latency and scalability to a very large number of storage nodes”.
  • This allows you to scale from small to very large with new and existing, all flash and hybrid, partially or fully populated systems.
  • The VM Scale-out software works across any standard high performance Ethernet network, eliminating the need for proprietary interconnects. The VM Scale-out software automatically provides best placement recommendation for VMs.
  • Scale compute and storage independently. Loose coupling of storage and compute provides customers with maximum flexibility to scale these elements independently. I think this is Tintri’s way of saying they’re not (yet) heading down the hyperconverged path.


VM Scale-out Software

Tintri’s new VM Scale-out Software (*included with Tintri Global Center Advanced license) provides the following capabilities:

  • Predictive analytics derived from one million statistics collected every 10 minutes from 30 days of history, accounting for peak loads instead of average loads, providing (according to Tintri) for the most accurate predictions. Deep workload analysis identifies VMs that are growing rapidly and applies sophisticated algorithms to model the growth ahead and avoid resource constraints.
  • Least-cost optimization based on multi-dimensional modelling. Control algorithm constantly optimizes across the thousands of VMs in each pool of VMstores, taking into account space savings, resources required by each VM, and the cost in time and data to move VMs, and makes the least-cost recommendation for VM migration that optimizes the pool.
  • Retain VM policy settings and stats. When a VM is moved, not only are the snapshots moved with the VM, the stastistics,  protection and QoS policies migrate as well using efficient compressed and deduplicated replication protocol.
  • Supports all major hypervisors.


You can check out a YouTube video on Tintri VM Scale-out (covering optimal VM distribution) here.


Tintri Analytics
Tintri has always offered real-time, VM-level analytics as part of its Tintri Operating System and Tintri Global Center management system. This has now been expanded to include a SaaS offering of predictive analytics that provides organizations with the ability to model both capacity and performance requirements. Powered by big data engines such as Apache Spark and Elastic Search, Tintri Analytics is capable of analyzing stats from 500,000 VMs over several years in one second.  By mining the rich VM-level metadata, Tintri Analytics provides customers with information about their environment to help them make better decisions about applications’ behaviours and storage needs.

Tintri Analytics is a SaaS tool that allows you to model storage needs up to 6 months into the future based on up to 3 years of historical data.


Here is a shot of the dashboard. You can see a few things here, including:

  • Your live resource usage for your entire footprint up to 32 VMstores;
  • Average consumption per VM (bottom left); and
  • The types of applications that are your largest consumers of Capacity, Performance and Working Set (bottom center).


Here you can see exactly how your usage of capacity, performance and working set have been trending over time. You can see also when you can expect to run out of these resources (and which is on the critical path). It also provides the ability to change the timeframe to alter the projections, or drill into specific application types to understand their impact on your footprint.

There are a number of videos covering Tintri Analytics that I think are worth checking out:


Tintri Cloud Suites

Tintri have also come up with a new packaging model called “Tintri Cloud”. Aimed at folks still keen on private cloud deployments, Tintri Cloud combines the Tintri Scale-out platform and the all-flash VMstores.

Customers can start with a single Tintri VMstore T5040 with 17 TB of effective capacity and scale out to the Tintri Foundation Cloud with 1.2 PB in as few as 8 rack units. Or they can grow all the way to the Tintri Ultimate Cloud, which delivers a 10 PB cloud-ready storage infrastructure for up to 160,000 VMs, delivering over 6.4 million IOPS in 64 RU for less than $1/GB effective. Both the Foundation Cloud and Ultimate Cloud include Tintri’s complete set of software offerings for storage management, VM-level analytics, VM Scale-out, replication, QoS, and lifecycle management.


Further Reading and Thoughts

There’s another video covering setting policies on groups of VMs in Tintri Global Center here. You might also like to check out the Tintri Product Launch webinar.

Tintri have made quite a big deal about their “VM-aware” storage in the past, and haven’t been afraid to call out the bigger players on their approach to VM-centric storage. While I think they’ve missed the mark with some of their comments, I’ve enjoyed the approach they’ve taken with their own products. I’ve also certainly been impressed with the demonstrations I’ve been given on the capability built into the arrays and available via Global Center. Deploying workload to the public cloud isn’t for everyone, and Tintri are doing a bang-up job of going for those who still want to run their VM storage decoupled from their compute and in their own data centre. I love the analytics capability, and the UI looks to be fairly straightforward and informative. Trending still seems to be a thing that companies are struggling with, so if a dashboard can help them with further insight then it can’t be a bad thing.

Tintri Announces New VMstore T5040

If you’re a vExpert, you might know that Tintri have frequently (and very kindly) provided polo shirts to vExperts. You’re also probably aware that Tintri do a lot more than just free shirts for community vnerds. They also have a bit to do with VM-aware storage. To that end, Tintri recently announced the release of the Tintri VMstoreTM T5040, a new entry-level all-flash storage array.

The data sheet can be found here, but in a nutshell, the specs of the T5040 are as follows:

  • OS 4.1 or higher
  • 1,500 VMs (max)
  • 18TB effective capacity
  • 5.76TB RAW Flash capacity
  • 4 10GbE Network ports
  • 2RU (same as the T5080 and T5060)
  • 455W / 1550BTU (slightly less than the T5080 and T5060)

Additional software includes:


As well as this, Tintri will be hosting a webinar on December 10, 2015 previewing Tintri Analytics with predictive analytics capabilities. According to Tintri, “Tintri Analytics will build on this same native VM-level metadata to be the only solution in the market to leverage the VM Aware Storage insights to provide a realistic profile of your VMs and applications to model growth of applications and additional need for capacity and performance”.


The Tintri VMstore T5040 and Tintri OS 4.1 will be generally available in December 2015. The Tintri VMstore T5040 is available immediately for ordering worldwide. Tintri OS 4.1 will be available as a free upgrade to all Tintri VMstore customers with a current support contract. You can read more about it at The Register and at