Disaster Recovery Planning for Hyper-Converged Infrastructure
Most of the chatter these days about Big Data analytics envisions a sprawl of inexpensive server/storage appliances arranged in highly scalable clustered-node configurations. This hyper-converged infrastructure (HCI) is considered well-suited to the challenge of delivering a repository for a large and growing “ocean” (or “lake” or “pool”) of data that is overseen by a distributed network of intelligent server controllers, all operated by a cognitive intelligence application or analytics engine.
HCI has never been well-defined. From a design perspective, it’s pretty straightforward: a commodity server is connected to some storage that’s usually mounted inside the server chassis (for example, internal storage) or externally connected via a bus extension interface (for example, direct-attached storage over Fibre Channel, SAS, eSATA or some other serial SCSI) all glued together with a software-defined storage (SDS) stack implemented to provide control over the connected storage devices.