How a cloud-native structure handles persistent storage
Cloud-native, or containerised, purposes are actually mainstream. As many as 82% of enterprises now have Kubernetes in manufacturing, in response to the Cloud Native Computing Discussion board (CNCF). That’s up from 66% in 2023. And a full 98% of organisations have at the least some cloud-native purposes, the trade physique says.
However transferring purposes to cloud-native environments doesn’t simply imply creating new code. It additionally means adapting infrastructure. Compute, networking and information storage all must work with container environments. Under no circumstances can all techniques do that out of the field, particularly in terms of on-premise {hardware}.
On the identical time, enterprise IT architects want to contemplate the necessities of legacy purposes and digital machines (VMs) that aren’t being up to date. And enterprises will need to take advantage of environment friendly use of their storage {hardware}, no matter their utility environments.
Transferring to containers means adapting a know-how that was not designed for persistent storage to deal with business-critical information.
Stateless states
Containerised purposes began out as stateless, or ephemeral. The designers by no means supposed containers to carry persistent information. They anticipated that microservices or containerised purposes would use no non-volatile storage and discard the contents of reminiscence, and even their settings, as soon as that they had accomplished their duties.
As a substitute, containerised purposes depend on an exterior information retailer, normally a database or cache.
There are benefits to this method. These embrace less complicated deployment, simpler scaling, fault tolerance and restoration, and utility portability. However most enterprise purposes, if not the bulk, want persistent information.
“Most enterprise purposes require storage. In actuality, until you’re changing Fahrenheit to Celsius and again, you’re storing one thing someplace,” says Dan Ciruli, vice-president and normal supervisor for cloud native at Nutanix.
And the necessity to work with persistent information is all of the extra necessary, as enterprises look to containers as a substitute for standard digital machines.
However this implies rethinking the way in which purposes work. And it requires IT architects to replace their storage techniques to assist modernised, cloud-native purposes. This may be straight, the place array producers assist containers, or by way of a management aircraft comparable to Nutanix or Everpure’s Portworx.
Nearly inevitably, adjustments are being pushed by AI, as enterprises look to assist its data-heavy workloads in fashionable, cloud-native environments. However there are different drivers, too, together with a pattern to maneuver virtualised purposes to containers and the necessity for value controls.
“Kubernetes is likely to be over a decade previous, however it’s persevering with to evolve as AI transforms the way in which we deal with information. Already, Kubernetes has moved past the times when it was constructed just for ephemeral, stateless purposes,” says Michael Cade, world subject chief know-how officer at Veeam Software program.
“Right now, stateful purposes comparable to databases, machine studying pipelines and streaming techniques are actually being handled as first-class residents [in containerised environments] and have been given the specialised instruments they should thrive.”
Storage connections
Connecting storage to Kubernetes, although, depends on assist from each utility builders and {hardware} suppliers.
The primary technique to join storage to container environments is thru the container storage interface (CSI). CSI must be supported straight by the storage supplier, be that the {hardware} producer, a cloud service, or a software-defined storage (SDS) provider.
Because the CNCF’s Kubernetes web page notes: “CSI was developed as a regular for exposing arbitrary block and file storage techniques to containerised workloads on container orchestration techniques like Kubernetes.” CSI permits third-party storage suppliers to write down, and deploy, plug-ins for storage with out altering the core Kubernetes code.
SDS applied sciences, for his or her half, additionally use CSI drivers, however run on commodity {hardware} slightly than devoted storage arrays, in addition to hyper-converged infrastructure. It additionally contains open supply choices, comparable to OpenEBS, Longhorn and Ceph.
“Each setting wants a storage again finish, with a CSI driver that connects it to Kubernetes. It’s as much as the storage supplier to supply the CSI driver,” says Nigel Poulton, an writer and impartial skilled in Kubernetes and containers.
“Most CSI drivers create at the least one StorageClass that maps to a tier of storage and its capabilities. For instance, a CSI driver may create a StorageClass known as ‘fast-replicated’ that maps to high-speed flash storage robotically replication to a distant location. Any utility utilizing this class robotically will get that tier and set of capabilities,” he provides.
This stage of abstraction is extremely helpful for utility builders, as they now not have to fret in regards to the bodily capabilities of the storage system. That’s dealt with by the CSI drivers.
“The CSI drivers allow us to offer entry to storage from the containerised utility, however [for firms to] nonetheless administer the storage the way in which they do the storage that’s operating beneath their VMs,” says Nutanix’s Ciruli. “And that’s an enormous benefit.” He additionally sees clients putting in Kubernetes on naked metallic clusters.
This additionally maintains separation between the Kubernetes workloads and the underlying storage {hardware}. On paper at the least, enterprises can transfer their containerised purposes to a special platform or provider, or new storage {hardware}, with out rewriting code and with minimal disruption.
In follow, large-scale strikes of Kubernetes purposes between platforms are nonetheless comparatively uncommon. Enterprises are inclined to develop purposes to run on Amazon Internet Providers (AWS), Google Cloud Platform (GCP), Microsoft Azure, or native {hardware}, relying on their enterprise necessities.
Utility portability, supported by CSI, is a helpful insurance coverage, even when there are sufficient variations between platforms to recommend warning.
“We actually don’t must turn into an skilled in how EBS [Elastic Block Store] works versus Azure disk, or native SSD [solid-state drives] and the way that works,” says Greg Muscarella, normal supervisor for Portworx at Everpure. “If you must handle these issues, it turns into considerably complicated. Firms are inclined to concentrate on a single cloud setting.”
Few organisations, he suggests, have code the place they might “push a button and transfer it to a special cloud”, not least due to variations between storage architectures from each {hardware} suppliers and cloud suppliers. Nevertheless, enterprises are transferring extra purposes to cloud-native environments. And this more and more contains databases and purposes that beforehand ran in standard digital machines.
New platforms
One of the crucial important developments in utility modernisation is to maneuver each digital machines and database-driven purposes to containers. Value, avoiding provider lock-in and the necessity to consolidate on fewer platforms are all drivers.
“The road between ‘containerised’ and ‘virtualised’ is blurring,” suggests Veeam’s Cade. “For a very long time, containers and VMs had been seen as two separate siloes. However as stateful purposes have developed, and since VMs are basically a typical stateful workload, we’re seeing a major rise in companies operating them straight inside Kubernetes utilizing platforms comparable to Pink Hat OpenShift Virtualization.”
Poulton agrees. He sees extra organisations transferring virtualised workloads to containers, through instruments comparable to KubeVirt. However, though organisations are porting over virtualised purposes, and databases, IT architects must make sure that all the applying’s necessities are met by the storage layer.
“Databases have way more demanding necessities, together with ordered startup, replication, automated failover and backup,” he cautions. “The 2 greatest adjustments are guaranteeing a CSI driver exists for the storage system and doubtlessly deploying an operator.”
A Kubernetes operator offers particulars a couple of database’s particular necessities, and generally storage, too. Operator assist is important to permit databases to ship enterprise workloads over Kubernetes. Once more, the operator helps the fashionable utility aim of separating the code from the storage array or cloud storage service.
Percona, for instance, offers operators for MySQL, PostgreSQL and MongoDB, in addition to Everest. “The operators are mainly the sport changers,” says Kate Obiidykhata, the corporate’s normal supervisor for cloud native. “They encode the human DBA data into the software program, and you’ve got all these most necessary resilience parts, backup, failover, replication and upgrades automated.”
Operators, she provides, assist enterprises to undertake hybrid architectures or multicloud methods, permitting information portability with out the necessity to rewrite purposes. However workloads that function on VMs is not going to robotically run on containers, she says. Companies might want to plan, and check, their deployments with care.
“There are particular playbooks that you need to apply and methodologies which can be clearly completely different from the basic database setup on VMs,” says Obiidykhata. “Nevertheless it’s all doable, and plenty of corporations are actually operating these databases on Kubernetes. They simply have a special playbook to mitigate these points.”
Companies additionally must think about how they run their ported purposes in manufacturing. Improvement, understandably, attracts a lot of the eye. However how techniques run from “day two” onwards is crucial. This contains storage provisioning and tiering, in addition to backup, restoration and safety.
The CSI drivers deal with a lot of the onerous work, however enterprises are more likely to look to put money into new {hardware}, and even storage from suppliers targeted on cloud-native environments, to ease the migration to containers.
“That is normally by deploying new storage architectures, both through new storage merchandise from current distributors, however more and more by partaking with new distributors,” says Poulton. Enterprises, he provides, may nonetheless be operating older {hardware} techniques, however they’re unlikely to make use of them for Kubernetes.

