"In- Band " VirtualizationHost-based virtualization, as noted above, is only one of at least four discernable variants in the SAN virtualization game. A second group of approaches are referred to collectively as "in-band" or "in-the-wire" virtualization by the trade press and analyst community. This description, while it has the advantage of convenience by capturing several product implementation philosophies under single moniker, also confuses the issue. All SAN virtualization techniques are, by their nature, located in the data path (see Figure 7-5). Figure 7-5. All virtualization is in-band.
In the case of virtualization software located on a host, the software layer is near the beginning of the I/O request, but still in the path. So-called in-band techniques are also in the path because they impact data traffic as it travels between the initiator of the I/O request and the target of that request, the storage platform. Current favorites for this type of virtualization are specialized "appliances" ”single-purpose servers running virtualization software and installed in the wire ”and "fat" switches ”storage switches enhanced with new software functionality for performing virtualization. The third approach, ostensibly called "out-of-band" virtualization, is effectively classified as "in-band" as well. A specialized server appliance sits outside the data path, where it monitors the condition of volumes and determines when capacity must be added. When necessary, it initiates a device driver update process and rewrites the parameters of the device driver on each host in order to change the size of the virtual volumes that applications are able to use. The device driver acted upon by the appliance is in-band. The final approach, virtualization on the controller of an array, is also an in-band approach, since the access made to backend storage is through the "front door" of the controller's virtualization engine. From the perspective of the data path, the distinction between in and out of band is meaningless. Having said that, what we are calling in-band virtualization ”in the data path appliances and enhanced switch functionality ”is currently the hot spot for virtualization engineering. In 2003, it seemed that nearly every storage hardware manufacturer, and many storage management software vendors, were either establishing alliances with in-band appliance vendors or teaming with leading Fibre Channel and IP switch vendors to migrate virtualization software to those platforms. Conceptually, it makes more sense to reduce the complexity of virtualization by vesting the functionality in a single, more easily managed and administered location such as a switch or appliance, rather than on hosts or storage arrays individually. A number of vendors seized upon this model early on, using commodity server hardware and operating system extensions to build appliance products. Concerns were immediately raised about the potential saturation of the commodity server bus by placing it in the line between an expanding number of initiators and an expanding pool of targets. Said one IT manager for a large insurance company, "Queuing theory alone suggests that you can't push traffic from 800-plus servers communicating with 180 TB of disk organized into storage array platforms through what is essentially a PC bus without introducing a significant chokepoint and significant latency." One out-of-band storage appliance vendor used a hypothetical math problem to demonstrate the difficulties "inherent" in its competitors ' in-band approaches:
Thus, consumer speculation, reinforced by vendor infighting, has created an impression of in-band virtualization inefficacy that has, in turn , proven the most difficult hurdle for vendors of in-band appliances to surmount. In fact, virtualization performance is determined by three factors: processor efficiency (how quickly data can be processed ), parallelism (how many processes can be accomplished in parallel or at the same time), and data pathing (how optimized is the path used to move data to where it needs to go, among and between internal components such as cache memory and CPU, and through busses and networks). In the mainframe world, where vendors have enjoyed nearly complete control over the design of both equipment and operating systems, this control has been used to optimize system architecture specifically for virtualization, elevating virtualization in the process to the status of an architectural mainstay. In the realm of open systems, such single-vendor hegemony does not exist, and implementations have been uneven in terms of performance optimization. Products like DataCore Software Corporation's SANSymphony, however, are maturing rapidly . This virtualization engine, which is part of the offerings of other vendors including Hitachi and Fujitsu SOFTEK, is designed to run on multiprocessor Storage Domain Servers (SDS), which do not have the interrupt processing requirements of general-purpose servers. According to the company,
DataCore claims that its caching technique actually reduces the disk I/O latency typically experienced by applications. The vendor attributes this to the fact that SDS caches are fast, yet lower cost and scalable across the network. It also cites its built-in data prefetching scheme and read request sorting as an accelerator, rather than a decelerator of I/O.
In summary, DataCore makes salient observations about its platform intended to answer the many detractors of in-band virtualization in the analyst community, trade press and marketing departments of competitors.
|