Welcome to Random Short Take #86. It’s been a while, and I’ve been travelling a bit for work. So let’s get random.
Let’s get started with three things / people I like: Gestalt IT, Justin Warren, and Pure Storage. This article by Justin digs into some of the innovation we’re seeing from Pure Storage. Speaking of Justin, if you don’t subscribe to his newsletter “The Crux”, you should. I do. Subscribe here.
And speaking of Pure Storage, a survey was conducted and results were had. You can read more on that here.
Switching gears slightly (but still with a storage focus), check out the latest Backblaze drive stats report here.
Oh you like that storage stuff? What about this article on file synchronisation and security from Chin-Fah?
More storage? What about this review of the vSAN Objects Viewer from Victor?
I’ve dabbled in product management previously, but this article from Frances does a much better job of describing what it’s really like.
Edge means different things to different people, and I found this article from Ben Young to be an excellent intro to the topic.
Welcome to Random Short Take #44. A few players have worn 44 in the NBA, including Danny Ainge and Pistol Pete, but my favourite from this list is Keith Van Horn. A nice shooting touch and strong long sock game. Let’s get random.
ATMs are just computers built to give you money. And it’s scary to think of the platforms that are used to deliver that functionality. El Reg pointed out a recent problem with one spotted in the wild in Ngunnawal.
Speaking of computing at the edge, I found this piece from Keith interesting. As much as things change they stay the same. I think he’s spot on when he says “[m]anufacturers and technology companies must come together with modular solutions that enable software upgrades for these assets’ lives”. We need to be demanding more from the industry when it comes to some of this stuff.
I enjoyed this article from Preston about the difference between bunkers and vaults – worth checking out even if you’re not a Dell EMC customer.
Cloud – it can be tough to know which way to go. And a whole bunch of people have an interest in you using their particular solution. This article from Chris Evans was particularly insightful.
DH2i has launched DxOdyssey for IoT – you can read more about that here.
Speaking of news, Retrospectrecently announced Backup 17.5 too. There are some cloud improvements, and support for macOS Big Sur beta.
It’s the 30th anniversary of Vanilla Ice’s “Ice Ice Baby“, and like me you were probably looking for a comprehensive retrospective on Vanilla Ice’s career. Look no further than this article over at The Ringer.
I’ve written in the past about additions to the HC3 Edge Platform. But those things had a rack-mount form factor. The newly announced HE150 runs on Intel NUC devices. Wait, what? That’s right, hyper-converged infrastructure on really small PCs. But don’t you need a bunch of NICs to do HC3 properly? There’s no need for backplane switch requirement, as they use some software-defined networking to tunnel the backplane network across the NIC. The HC3 platform uses less than 1GB RAM per node, and each node has 2 cores. The storage sits on an NVMe drive and you can get hold of this stuff at a retail price of around $5K US for 3 nodes.
[image courtesy of Scale Computing]
Scale at Scale?
How do you deploy these kinds of things at scale then? Conboy tells me there’s full Ansible integration, RESTful API deployment capabilities, and they come equipped with Intel AMT. In short, these things can turn up at the remote site, be plugged in, and be ready to go.
Where would you?
The HE150 solution is 100% specific to multi-site edge implementations. It’s not trying to go after workloads that would normally be serviced by the HE500 or HE1000. Where it can work though, is with:
Oil and Gas exploration – with one in each ship (they need 4-5 VMs to handle sensor data to make command decisions)
Grocery and retail chains
Manufacturing platforms
Telcos – pole-side boxes
In short, think of environments that require some amount of compute and don’t have IT people to support it.
Thoughts
I’ve been a fan of what Scale Computing has been doing with HCI for some time now. Scale’s take on making things simple across the enterprise has been refreshing. While this solution might surprise some folks, it strikes me that there’s an appetite for this kind fo thing in the marketplace. The edge is often a place where less is more, and there’s often not a lot of resources available to do basic stuff, like deploy a traditional, rackmounted compute environment. But a small, 3-node HCI cluster that can be stacked away in a stationery cupboard? That might just work. Particularly if you only need a few virtual machines to meet those compute requirements. As Conboy pointed out to me, Scale isn’t looking to use this as a replacement for the higher-preforming options it has available. Rather, this solution is perfect for highly distributed retail environments where they need to do one or two things and it would be useful if they didn’t do those things in a data centre located hundreds of kilometres away.
The edge presents a number of challenges to enterprises, in terms of both its definition and how to deal with it effectively. Ultimately, the success of solutions like this will hinge on ease of use, reliability, and whether it really is fit for purpose. The good folks at Scale don’t like to go off half-cocked, so you can be sure some thought went into this product – it’s not just a science project. I’m keen to see what the uptake is like, because I think this kind of solution has a place in the market. The HE150 is available for purchase form Scale Computing now. It’s also worth checking out the Scale Computing presentations at Tech Field Day 20.
The HE500 series has been introduced to provide smaller customers and edge infrastructure environments with components that better meet the sizing and pricing requirements of those environments. There are a few different flavours of nodes, with every node offering E-2100 Intel CPUs, 32 – 64GB RAM, and dual power supplies. There are a couple of minor differences with regards to other configuration options.
HE500 – 4x 1,2,4 or 8TB HDD, 4x 1GbE, 4x 10GbE
HE550 – 1x 480GB or 960GB SSD, 3x 1,2, or 4TB HDD, 4x 1GbE, 4x 10GbE
HE500T – 4x 1,2,4 or 8TB HDD, 8 x HDD 4TB, 8TB, 2x 1GbE
HE550TF – 4 x 240GB, 480GB, 960GB SSD, 2x 1GbE
The “T” version comes in a tower form factor, and offers 1GbE connectivity. Everything runs on Scale’s HC3 platform, and offers all of the features and support you expect with that platform. In terms of scalability, you can run up to 8 nodes in a cluster.
Thoughts And Further Reading
In the past I’ve made mention of Scale Computing and Lenovo’s partnership, and the edge infrastructure approach is also something that lends itself well to this arrangement. If you don’t necessarily want to buy Scale-badged gear, you’ll see that the models on offer look a lot like the SR250 and ST250 models from Lenovo. In my opinion, the appeal of Scale’s hyper-converged infrastructure story has always been the software platform that sits on the hardware, rather than the specifications of the nodes they sell. That said, these kinds of offerings play an important role in the market, as they give potential customers simple options to deliver solutions at a very competitive price point. Scale tell me that an entry-level 3-node cluster comes in at about US $16K, with additional nodes costing approximately $5K. Conboy described it as “[l]owering the barrier to entry, reducing the form factor, but getting access to the entire stack”.
Combine some of these smaller solutions with various reference architectures and you’ve got a pretty powerful offering that can be deployed in edge sites for a small initial outlay. People often deploy compute at the edge because they have to, not because they necessarily want to. Anything that can be done to make operations and support simpler is a good thing. Scale Computing are focused on delivering an integrated stack that meets those requirements in a lightweight form factor. I’ll be interested to see how the market reacts to this announcement. For more information on the HC3 Edge offering, you can grab a copy of the data sheet here, and the press release is available here. There’s a joint Lenovo – Scale Computing case study that can be found here.
Disclaimer: I recently attended Storage Field Day 17. My flights, accommodation and other expenses were paid for by Tech Field Day. There is no requirement for me to blog about any of the content presented and I am not compensated in any way for my time at the event. Some materials presented were discussed under NDA and don’t form part of my blog posts, but could influence future discussions.
NGD Systems recently presented at Storage Field Day 17. You can see their videos from Storage Field Day 17 here, and download a PDF copy of my rough notes from here.
Edgy
Storage and compute / processing requirements at the edge aren’t necessarily new problems. People have been trying to process data outside of their core data centres for some time now. NGD Systems have a pretty good handle on the situation, and explained it thusly:
A massive amount of data is now produced at the edge;
AI algorithms demand large amounts of data; and
Moving data to cloud is often not practical.
They’ve taken a different approach with “computational storage” by moving the compute to storage. It then becomes a problem to solve in terms of Power/TB + $/GB + in-situ processing. Their focus has been on delivering a power efficient, low cost, computational storage solution.
A Novel Solution – Move Compute to Storage
Key attributes:
Maintain familiar methodology (no new learning)
Use standard protocols (NVMe) and processes (no new commands)
Minimise interface traffic (power and time savings)
Enhancing limited footprint with maximum benefit (customer TCO)
Moving Computation to Data is Cheaper than moving Data
A computation requested by an application is much more efficient if it is executed near the data it operates on
Minimises network traffic
Increases effective throughput and performance of the system (eg Hadoop Distributed File System)
Enables distributed processing
Especially true for big data (analytics): large sets and unstructured data
Traditional approach: high-performance servers coupled with SAN/NAS storage – Eventually limited by networking bottlenecks
They say that these storage solutions solve the low power, more efficient compute needs without placing strain on the edge and “fog” platforms. I found the CDN use case to be particularly interesting. When you have a bunch of IP-addressable storage sitting in a remote point of presence it can sometimes be a pain to have them talking back to a centralised server to get decryption keys for protected content, for example. In this case you can have the drives do the key handling and authentication, providing faster access to content than would be possible in latency-constrained environments.
It seems silly to quote Gaga Herself when writing about tech, but I think NGD Systems are taking a really interesting approach to solving some of the compute problems at the edge. They’re not just talking about jamming a bunch of disks together with some compute. Instead, they’re jamming the compute in each of the disks. It’s not a traditional approach to solving some of the challenges of the edge, but it seems like it has legs for those use cases mentioned above. Edge compute and storage is often deployed in reasonably rugged environments that are not as well-equipped as large DCs in terms of cooling and power. The focus on delivering processing at storage that relies on minimal power and uses standard protocols is intriguing. They say they can do it at a reasonable price too, making the solution all the more appealing for those companies facing difficulties using more traditional edge storage and compute solutions.
You can check out the specifications of the Newport Platform here. Note the various capacities depend on the form factor you are consuming. There’s also a great paper on computational storage that you can download from here. For some other perspectives on computational storage, check out Max‘s article here, and Enrico’s article here.
It seems like only a few months ago that I was introduced to StorMagic via Storage Field Day 6. You can read my thoughts on that here. I was pretty impressed with StorMagic’s focus on their strengths and the solution’s capacity to solve some difficult problems when it came to virtualised storage at the edge of the network.
In any case, StorMagic announced recently that they’ve officially partnered with VMware as the ROBO storage solution of choice when it comes deploying a VSA at the edge. What that translates to is one SKU from VMware to order the software and licences and one SKU from StorMagic to get your hands on a very solid edge storage VSA solution. Here’s a link to StorMagic’s solution brief on their website. And here’s a picture.
The solution runs on anything that’s in the VMware HCL, can scale down to 2 servers (as opposed to VSAN’s 3-node requirement) and provides edge HA for the large enterprise.
You can also read a great write-up from Amit Panchal here, as well as a typically astute analysis from Jon Klaus here. I think it’s great that StorMagic have been able to make this announcement and look forward to hearing about future developments.
I was making some port-channels between one of our MDS 9513 director switches and a 9124e edge and managed to add the interfaces to the wrong port-channel. Here’re the basic steps on the 9124e end that I took to rectify the issue. I’ve created a pdf file which, while inconvenient, solves the problems related to both my wordpress skills and the age of the theme I use. That is, a 4 page doc was going to look pretty ugly if I tried to insert it in-line. I apologise in advance for the inconvenience you will no doubt experience.
Sometimes, for any number of reasons, you’ll find yourself wanting to downgrade the firmware on your Cisco edge devices to match what you have running in the core. Fortunately, at least for the 9100-series switches, this is basically the same as upgrading the firmware. I’ve included the commands to run here, and also the full output of the process. For the director-class switches, there are a few more things to do, such as clearing out the space on the standby supervisor as well as the active sup card. I’ll try and post something 9500-series specific in the next few weeks.
In short, do this (assuming you’re loading version 3.3(4a) of the code):
show version image bootflash:m9100-s2ek9-mz.3.3.4a.bin
show incompatibility system m9100-s2ek9-mz.3.3.4a.bin
install all system bootflash:m9100-s2ek9-mz.3.3.4a.bin kickstart bootflash:m9100-s2ek9-kickstart-mz.3.3.4a.bin
y
show module
show version
You can also see the full output here. Note that this process works equally well for HP’s 9124e switches (the type you find in the back of c7000 blade chassis for instance), although you should be downloading the firmware from HP’s site, not Cisco’s.
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
This website uses cookies to improve your experience while you navigate through the website. Out of these cookies, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may have an effect on your browsing experience.