Disaster Recovery Planning for Hyper-Converged Infrastructure
Most of the chatter these days about #BigData analytics envisions a sprawl of inexpensive server/storage appliances arranged in highly scalable clustered-node configurations. This hyper-converged infrastructure ( #HCI ) is considered well-suited to the challenge of delivering a repository for a large and growing “ocean” (or “lake” or “pool”) of data that is overseen by a distributed network of intelligent server controllers, all operated by a cognitive intelligence application or analytics engine. It all sounds very sci-fi. But, breaking it down, what are we really dealing with? HCI has never been well-defined. From a design perspective, it’s pretty straightforward: a commodity server is connected to some storage that’s usually mounted inside the server chassis (for example, internal storage) or externally connected via a bus extension interface (for example, direct-attached storage over Fibre Channel, SAS, eSATA or some other serial SCSI) all glued together with a software-defined storage ( #SDS ) stack implemented to provide control over the connected storage devices.