White Paper: Virtualization
Server Virtualization and Network Management (ingles)
Server virtualization is the latest major technology trend in the data center. Gartner predicts that virtualization will be the highest-impact trend in IT infrastructure and operations through 2012.
Gartner predicts that virtualization will be the highest-impact trend:
Despite its limits, where server virtualization has been applied rigorously, it has both enormous and positive impacts on corporate power and cooling expenses as well as data center capacity. It can extend the lives of aging data centers and even allow some large organizations to shut down some of their data centers.
And, this is just the first generation of virtualization's impact. A second generation technology, which allows multiple virtual servers to share a single operating system license, promises to allow even further consolidation. This can get to the point where 50-100 virtual servers run on a single physical box which can manage high-compute load applications that are considered poor candidates for first-generation virtualization.
Virtualization works by freeing applications from the constraints of a single physical server. It allows multiple applications to run on one server, but equally important, it allows one application to use resources across the corporate network. One of the promises of virtualization is the ability to allow applications to dynamically move from one physical server to another as the demands and resource availabilities change, without service interruption.
Virtualization also mandates the use of storage-area networks (SANs) and detached storage. This drives the network further into the center of the IT infrastructure and architecture. From a security standpoint, centralizing multiple applications on a single box creates single points of failure, both in the physical server itself, and in its network connection. When a virtualized server crashes, or its network connection slows or breaks, it impacts all the applications on that box. The implication for network planning is that a great deal more traffic will be centralized on a few large servers instead of spread out across a large number of smaller computers on the data center floor.
Also, virtualization works best with detached, rather than attached storage. This requires very fast, dependable network connectivity between servers and the storage devices on the storage area network (SAN). Organizations moving from attached storage, as part of virtualization, will create a large increase in network traffic. All of this increases the need for strong network management. A highly virtualized environment lives or dies on the efficiency and dependability of its data network. Failure of a physical server, connection, switch or router can be very costly when it disconnects knowledge workers, automated factory floors or online retail operations from vital IT functionality.
Network management can provide vital information for planning and testing virtualized environments. For instance, determining which applications are not good candidates for virtualization is vital. The amount and character of network traffic can provide important clues towards identifying those applications.
Why Server Virtualization
Server virtualization has come just in time for IT departments caught between the pressure to cut costs in the face of a developing worldwide recession, steadily increasing energy costs and maxed-out data centers. Virtualization attacks the problem of the low utilization of single application servers. These proliferate in data centers of medium to very large enterprises.
The server populations of many non-virtualized environments average about 20% utilization. The result is a huge waste of power, which is doubled since every kilowatt used has to be balanced with an equal amount of cooling to maintain the servers at optimal operating temperature.
FIGURE: Each application resides
on a seperate physical
server with attached storage
This also has grave implications for the lifespan of data centers as increasing numbers of facilities run short of power, cooling and, in some cases, floor space despite the move to blade servers over the last few years. Virtualization avoids this problem by automating the management issues of stacking multiple applications on a single server and sharing resources among them.
This allows IT shops to increase their server utilization up to 80%. The impact of this on a large organization is demonstrated by the experience of BT in the UK. It achieved a 15:1 consolidation of its 3,000 Wintel servers and saved approximately 2 megawatts of power and $2.4 million in annual energy costs. This helped to reduce server maintenance costs by 90%, and BT disposed 225 tons of equipment (in an ecologically friendly manner) and closed several data centers across the UK.
This saving is even more remarkable because it involved only Wintel servers-BT also has a large population of Unix servers that were not factored into the savings. Not only did BT achieve huge operating savings, partly by avoiding the need to build a new data center at an estimated cost of $120 million, it also received three top European ecology awards for the reductions of its carbon footprint from the program.
During this process, BT continued to grow its business steadily, which puts the lie to the common argument that companies and nations have the choice of only expanding their economies or cutting their environmental impacts. Far from hurting business, BT's project added credibility to the green business consulting practice it launched earlier this decade in both in the European Union and United States. This is not an extreme example of what can be accomplished. Clearly BT is a very large enterprise. Smaller companies will achieve proportionately smaller savings in absolute numbers, but the percentages will hold up for organizations with server populations of 50 or more. This will allow them to extend the effective lives of their data centers as well as saving on energy costs.
Savings through virtulization at BT in the UK:
Even smaller organizations may find benefits to virtualization, particularly when facing the need to upgrade their servers. One large server is less expensive to buy, operate and maintain than even a small population of small servers. The application management
automation that virtualization provides can save a smaller organization one, or more full time equivalent IT employees. These savings are being achieved mostly through hypervisor virtualization as supplied by EMC's VMWare and other vendors.
This technology requires each virtual server to run its own copy of the operating system. This is the preferred solution when applications running different OSs-for instance a mixed population of Windows, Unix and Linux or several different Unix flavors-are being virtualized onto a single physical server. However, this has efficiency implications when a large number of applications on the same OS are virtualized.
An emerging new approach, "application virtualization," offers a more efficient solution in these cases, allowing multiple virtualized applications to share a single copy of an OS. Not only can this allow more applications to run on a given physical server, its proponents claim that it also will allow efficient operation of high compute-load applications which are not good candidates for hypervisor virtualization.
Network Management and Virtualization
At first glance it might seem that virtualization by itself would have little impact on network management. The same applications are running in the same data center, just on a different piece of hardware. However, a closer look shows that it does have network management implications. The extent of those implications varies, depending on the previrtualization architecture. Overall virtualization is one of several forces combining to drive the network into the center of the IT infrastructure and displacing the mainframe. This increases the importance of network operations and the potential impact of interruptions, making strong network management vital to today's business.
How Virtualization Works
Concentration of Applications
First, virtualization centralizes applications on one or comparatively few physical servers, making those, and their "last mile" network connections, single-points-of-failure that can have a major impact on business operations when they fail. Those who remember the green-screen operations of the 1980s know that when the central computer failed, all work stopped in the office. The failure of a network switch or router in a highly virtualized environment could have a similar impact by shutting off access to as many as 50-100 vital applications. A network management tool such as Paessler's PRTG Network Monitor can provide instant notification of the failure of either a Windows server or its network connection, allowing IT to take immediate action to minimize the impact of the crash.
Storage Solutions and Virtualization
Part of the promise of server virtualization is the ability to share and shift resources, including disk space, from one application to another dynamically.
Consolidated Through Virtualization
FIGURE: Three applications run
in a virtual environment
sharing the resources of
the virtualization server
In a multi-server environment, this mandates the migration from direct, attached storage (in which the disk drives are attached directly to the server much as they are on desktop and laptop computers) to separate storage systems running on a storage area network (SAN), which becomes part of the overall network infrastructure. This migration has been ongoing since the late 1990s and has several advantages in terms of flexibility and data protection. It will accelerate as server virtualization works its way into shops still running applications with direct attached storage.
The disadvantage is that all access between applications and data now must run over the network, and even small delays can create issues with many applications. Thus virtualized environments cannot tolerate network overloads or switch failures. This requires strong network management. A strong network monitoring tool such as Paessler's PRTG Network Monitor can provide real-time graphic monitoring of traffic at critical points in the network and can provide long-term usage projections. This helps network managers anticipate traffic increases before they impact service levels.
Unsuitable Applications for Virtualization
Applications with heavy compute or data read/write loads have not proven to be good candidates for hypervisor virtualization. As a result, wholesale virtualization can create serious service level issues. However, identifying which applications should remain on dedicated servers, is not always easy. The volume and character of transmissions to and from these applications provides major clues toward identifying the applications, and a strong network management tool, such as PRTG Network Monitor, is the best source for that vital information.
Virtualizations and Resource Management
Server virtualization promises dynamic reassignment of underlying compute and storage resources to meet spikes in demand on individual applications. This is the real key to increasing efficiency in resource allocation. One major reason that the unvirtualized environment is characterized by such low utilization is that each application needs sufficient computer power, memory and storage to meet its maximum load plus an overhead to accommodate future growth. It is often difficult to predict the speed of growth, which tends to follow a "hockey stick" graph rather than a smooth curve-usage is low while users learn a new application and then may jump quickly as they become comfortable with it and realize its potential for improving their work.
FIGURE: With increasing user acceptance
a new application (red) may
require more resources. This
can easily be achieved at tx
by moving one application
(blue) to another virtual mashine
Often many applications experience major cyclical variations in demand that may be weekly, monthly or annual. Financials are the obvious example of applications with very large monthly and quarterly use cycles, but many others exist. Again, in the unvirtualized environment, each application has to have enough computing resources to deliver promised service levels under the maximum annual load. For some applications, this means a heavy investment in computing resources that will remain idle much of the time. Virtualization removes this problem by providing dynamic reallocation of computer resources.
Demand on the separate applications running in a data center can operate on different cycles. An application experiencing a surge in demand at a given time can "borrow" resources from another that is in a low period of its demand cycle. In the extreme case, virtualization vendors are promising the capability to dynamically move an entire application across the network from one server to another in order to respond to such variations.
Ultimately, this implies that all the servers and storage devices running virtualized environments can, for planning purposes, be treated as one large machine. Their resources be lumped together to meet the total demand of all virtualized software running on them. As a practical matter, this may only work inside a single data center since transmission delays between centers could introduce performance issues. And, it will not work at all unless the network connecting these physical servers is fast and highly fault tolerant. This requires high levels of network management to guarantee performance.
Virtualization and Paessler
Paessler offers three applications that form a complete network monitoring suite for small-to-medium businesses and is used by large and very large enterprises to provide more detailed visibility into critical parts of their networks. Paessler has re-architected all its tools recently, putting them on the optimal technological base and improving their overall functionality, efficiency and usability. Its tool suite consists of: PRTG Network Monitor is an easy-to-use Windows application for monitoring and classifying bandwidth usage and for monitoring network uptime/downtime.
It provides live readings and long-term usage trends for network devices. PRTG makes early detection of network and website problems easy and affordable. It helps organizations monitor critical network resources and detect systems failures or performance problems immediately, minimizing downtime and its economic impact. SNMP Helper enables PRTG Network Monitor to collect detailed performance information on Windows servers and workstations.
Up to several thousand parameters and performance counters on a PC can be monitored with just a few mouse clicks. Webserver Stress Tool is a powerful HTTP-client/server test application designed to pinpoint critical performance issues on a Web site or Web server that may obstruct delivery of an optimal user experience. By simulating simultaneous access by hundreds or thousands of simultaneous users, it tests Web server performance under normal and extreme loads to ensure that critical information and service are available at response times that users expect. All of these tools come in various editions, and include a free edition to allow users to try out their main features before making the purchase decision.