RC2 expected for VMware Workstation 5.5

Three days ago a VMware employee answered a user in the VMTN discussion forums mentioning an upcoming RC2 release for Workstation 5.5.

If so it’s obvious the RTM will not be presented at VMworld 2005 as I was guessing. I’m starting to suspect VMware is waiting for Microsoft Vista beta 2, expected for December 7th, to be sure users can install it flawless as guest OS.

Meanwhile there are issues around the highly expected VMware Player. More datails to come.

Microsoft discloses Virtual Hard Disk format details

Microsoft decided to make its VHD format available to third parties under a royalty-free license.

We already saw PlateSpin, in this post, and Acronis, in this post, moving to gain benefits from this choice.

More vendors will surely adapt their products to interact with the VHD format. The first category doing so will probably be antivirus, now eventually able to scan virtual machines HDs from the host OS.

The specifications disclosed, which are going to be discussed at Microsoft ITforum 05 conference, include:

  • Hard Disk Foot Format
  • Dynamic Disk Header Format
  • Block Allocation Table and Data Blocks
  • Implementing a Dynamic Disk
  • Mapping a Disk Sector to a Sector in the Block
  • Splitting Hard Disk Images
  • Implementing a Differencing Hard Disk
  • CHS Calculation

You can apply to see details by signing the Virtual Hard Disk Image Format Specification License Agreement here.

Update: The license agreement states that any product developed with VHD manipulation capability needs to be distributed in binary form only.

Thanks to Robert Aitchison for the insight.

Microsoft talking about Virtual Server 2005 R2 and Windows Hypervisor at ITforum 05

Did you miss the PDC 05 U.S.A. conference this year? No problem: Microsoft is going to have multiple sessions about its upcoming virtualization technologies at ITforum 05 european conference.

Actual agenda includes the following sessions:

  • CHT013 Automating a Dev/Test Lab with Virtual Server
    Virtualization can significantly reduce the cost of developing and testing complex multi-tier applications. Multiple servers can easily be provisioned and manipulated ‘virtually’ without requiring large amounts of physical hardware. In this Chalk-&-Talk we will discuss how virtualization is changing dev/test and provide some real world examples.
  • CHT014 Architectural Best-Practices for Virtual Server
    Virtualization can provide an effective means for consolidating production servers. In this Chalk-&-Talk we will discuss real-world usage of Virtual Server and how it could best be applied in your IT organization.
  • INF304 Virtual Server – Advanced Scripting and Other Secrets
    Virtual Server provides a powerful selection of COM interfaces for manipulating all facets of a virtual machine. This can support everything from provisioning to day-to-day management operations. In this session we will showcase the COM API and demonstrate a range of scripting examples. We will also demonstrate how partner offerings are using the API to productize key provisioning and management functionality.
  • INF307 Understanding and Working with the Virtual Server Virtual Hard Disk (VHD) Disk Format
    Microsoft has standardized on the VHD file format for the running and management of virtual machines. This file format provides for a range of disk types, including dynamically expanding, linked, undo and differencing disks. In this session we will take a deep dive into the file format, and discuss best practices and partner offerings for disk image management and migration.
  • INF308 Clustering and High Availability with Virtual Server
    Virtual Server 2005 R2 provides new high-availability features such as guest clustering and host-based clustering. These provide the basis of a very robust production server consolidation solution. In this session we will explore these new Virtual Server capabilities in detail, and look at key partner offerings that add additional high-availability functionality.
  • INF318 Windows Hypervisor and Virtualization Futures
    This session provides a technical overview of the Windows Hypervisor. The Windows Hypervisor will provide the foundation for Microsoft’s Windows virtualization solutions in the Longhorn Server timeframe. This session discusses the Hypervisor and related I/O virtualization layers.

Microsoft Virtual Server 2005 review by BentUser

Virtual Server 2005 receives its countless review, this time by BentUser portal.

You could find it interesting since there are some performance comparisons with real hardware in various scenarios.
The test’s results anyway should be taken very carefully since no documentation is provided about how the host OS was configured and the hardware used isn’t the one anybody should use for a virtualization server.

Read it here.

VMware totally revamps web site

VMware just restyled the whole website. I’ve found this new layout much more rational, clearer and simpler to interact with.

An interesting thing is a new comparison between GSX and ESX Server products. In this comparison VMware clearly states something not so evident among customers: GSX Server achieves a best performances rate of 4 VMs/CPU while ESX Server perform a 8 VMs/CPU rate.

It seems VMware is preparing to launch Workstation 5.5 in great style for VMworld 2005 timeframe.

Citrix details itself as Application Virtualization vendor

The keyword of the last period is surely virtualization. Every vendor I see these months is trying to adapt old and new products description to fit the virtualization concept.

Citrix shouldn’t have the need to but Brian Madden reports that Citrix iForum conference brought a more explicit description for its flagship product, Presentation Server, now detailed as a application virtualization solution.

This points me to a 6 months-ago post where I suggested IT industry to adopt a standard naming convention for the various virtualization technologies. Something we probably still need.

PremiTech enhances End-User Systems Management Solution for Citrix and VMware

Quoting from the PremiTech official announcement:

PremiTech today announced the latest version of its proven software for managing end-user quality of IT service, Performance Guard™ 4.1. This new version offers Citrix Presentation Server™ customers enhanced usability and notification features, as well as support for applications virtualized on VMware servers. PremiTech will showcase Performance Guard, developed in close cooperation with Citrix and its customers, in Booth #604 at Citrix® iForum™ 2005, October 9-12, in Las Vegas, Nev.

Performance Guard is utilized by more than 200 global companies including an impressive array of some of the largest Citrix customers in the world, such as Ingersoll-Rand, Maersk Sealand, General Electric, Verizon Wireless, State Street Bank and the U.S. Environmental Protection Agency. The product is supported and distributed by leading Citrix platinum partners, such as Bell Business Solutions, IBM Global Services, RapidApp, DynTek, Vector-MTM and IPM. PremiTech also recently received the DABCC Seal of Approval for its ability to solve customers’ problems and achieve the highest standards of excellence.

With Performance Guard, IT organizations are able to monitor end-user performance and network latency, as well as the service on Presentation Server and supporting back-end servers to obtain consistent, accurate measurement of Citrix application performance in real time. The new release makes it even easier for IT staff to proactively ensure optimal service delivery for their business users while reducing operational costs through early resolution of problems.

New features of Performance Guard 4.1 include:

  • Added performance metrics: Especially important for Presentation Server users, Performance Guard now monitors additional key performance indicators, such as context switches.
  • Improved network analysis: Performance Guard monitors the entire network to pinpoint and document the exact cause of poor response times and broken sessions.
  • Improved distribution of scheduled reporting: Performance alarms and reports can be sent automatically to appropriate personnel inside and outside of IT.
  • Enhanced alarms and auto-baselining: Performance Guard intuitively configures itself to recognize baseline performance and notify IT at any point when the system deviates from normal conditions.
  • Strengthened correlation engine: Performance Guard now maintains server metrics for an extended period of time, providing the ability to correlate application and system performance for improved root cause identification.
  • Active Citrix login measurement: The Performance Guard agent now makes repeated login measurements in order to ensure the Citrix server is up and running and to identify differences in login times.
  • VMware support: For firms consolidating servers and using virtualized applications in their production environments, Performance Guard now performs baseline measurements and provides pre- and post-migration metrics for the VMware virtual infrastructure.
  • “This latest release shows that PremiTech is dedicated to making our jobs easier and helping us reduce support costs through improved efficiencies,” said Dean Matvey, Network Services Division of Ingersoll-Rand’s Infrastructure Sector. “With Performance Guard, we are able to provide a precise explanation on the cause of system issues and an exact account of how we’re going to fix them. You couldn’t get that from any other systems management product.”

About Performance Guard
Performance Guard is an off-the-shelf, standard solution that provides the ability to monitor end-user quality of IT service in Desktop, Citrix and VMware environments. It delivers consistent, accurate measurement of application performance from the end-user perspective in real time. This data is essential for the rapidly growing number of companies using access infrastructure solutions from Citrix to measure mission-critical application performance, identify performance bottlenecks and resolve issues. Using one centralized server with Performance Guard, companies are able to improve system performance, increase user productivity, reduce expenses in infrastructure and help-desk resources, and ensure compliance with Service Level Agreements (SLA).

The virtues of virtualization

Quoting from CIO Asia:

During the past few decades, CIOs have stood at the center of one of the great technological revolutions in history: the replacement of the physical atom by the computational bit as the medium of commerce and culture. The profession might be forgiven for thinking that nothing is left for the next generation but tinkering. What could compare with a transition like that?

Actually, something almost as big might be coming over the horizon: the replacement of the bit with the virtual bit. Virtualization is the substitution of physical computing elements, either hardware or software, with artificial impostors that exactly replicate the originals, but without the sometimes inconvenient need for those originals to actually exist. Need a 1 terabyte hard drive, but only have 10 100GB drives? No problem, virtualization software can provide an interface that makes all 10 drives look and act like a single unit to any inquiring application. Got some data you need from an application you last accessed in 1993 on an aging MicroVAX 2000 that hit the garbage bin a decade ago? A virtual Digital VMS simulator could save your skin.

Stated like that, virtualization can sound like little more than a quick and dirty hack, and indeed, for most of the history of computing, that is exactly how the technique was viewed. Its roots lie in the early days of computing, when it was a means of tricking single-user, single-application mainframe hardware into supporting multiple users on multiple applications. But as every aspect of computing has grown more complex, the flexibility and intelligence that virtualization adds to the management of computing resources have become steadily more attractive. Today it stands on the lip of being the next big thing.

Raising the Dead
The Computer History Simulation Project, coordinated by Bob Supnik at SiCortex, uses virtualization to fool programs of historical interest into thinking that they are running on computer hardware that vanished decades ago. Supnik’s project has a practical end as well: Sometimes old systems are so embedded in the corporate landscape that they must be kept running. If the real hardware is unavailable, the only way to keep the old machines running is to virtualize them.

In a more contemporary example of the power of virtualization, about three years ago J. R. Simplot, a $3 billion food and agribusiness company in Boise, Idaho, found itself in a phase of especially rapid growth in server deployments. Of course, with rapid growth comes the headache of figuring out how to do everything faster. In this case, the company’s IT center concluded that their old server procurement system had to be accelerated.

Servers, of course, are pieces of physical equipment; they come with their own processing, memory, storage resources and operating systems. What the Simplot team did was use virtualization tools from VMware, a virtual infrastructure company, to create software-only servers that interacted with the network just like hardware servers, although they were really only applications. Whenever Simplot needed another server it would just flip the switches appropriate to the server type (Web, application, database, e-mail, FTP, e-commerce and so on). From that point, an automated template generated the virtual machine on a specific VMware ESX host machine.

Virtual Improvements
According to Tony Adams, a technology analyst at Simplot, there were gains all across the board. The time to get a new server up and running on the system went from weeks to hours or less. Uptime also increased, because the servers were programs and could run on any supported x86 hardware anywhere. If a machine failed or needed maintenance, the virtual server could be quickly moved to different hardware.

Perhaps most important were the gains in utilization efficiencies. Servers are built for specific roles. Sometimes demand for a particular role is in sync with available resources, but usually it isn’t. In the case of “real” servers, if there is a mismatch, then there is nothing that you can do about it; you’re stuck with what you have. If you end up with an average utilization rate of 10 percent per server, so be it. (The need to provide for peak demand makes the problem worse, and utilization can often be far below even 10 percent.) Low utilization means IT is stuck with unnecessary maintenance issues, security faces unnecessary access issues (they have to worry about protecting more machines), and facilities must deal with unnecessary heat and power issues.

Virtualization fixes these problems. The power to design any kind and number of servers that you like allows you to align capacity with load continuously and precisely. In the case of Simplot, once Adams’s servers turned virtual, he was able to deploy nearly 200 virtual servers on only a dozen physical machines. And, he says, typical CPU, network, disk and memory utilization on the VMware ESX boxes is greater than 50 percent—compared with utilization of around 5 percent on dedicated server hardware.

Virtualization also makes disaster recovery planning simpler, because it allows you to write server clusters appropriate to whatever infrastructure you have on hand. As Adams points out, conventional disaster recovery schemes force you to have an exact clone of your hardware sitting around doing nothing. “But personally, what I really like,” he says, “is the remote manageability. I can knock out new [servers] or do repairs anywhere on the Net, without even going to the data center.”

Adams wants one machine to look like many machines, but it is just as possible to virtualize the other way: making many machines look like one. Virtualization underlies the well-known RAID storage tricks that allow many disks to be treated as one huge drive for ease of access, and one disk to be treated as many for the purpose of robust backup. Another prime use for virtualization is development. The hardware world is growing much more complex all the time: Product cycles are turning faster, the number of device types is always rising, and the practice of running programs over networks means that any given program might come in contact with a huge universe of hardware. Developers can’t begin to afford to buy all of this hardware for testing, and they don’t need to: Running products on virtualized models of the hardware allows for quality assurance without the capital expense. Virtualizing the underlying hardware also gives developers far more control. Peter Magnusson, CTO of Virtutech, a systems simulation company in San Jose, Calif., points out that you can stop simulated hardware anywhere you like, any time you want to investigate internal details.

Unreal Future
During the next year or two, virtualization is on track to move from its current success in storage, servers and development, to networks and data centers. So CIOs will then be able to build software versions of firewalls, switches, routers, load balancers, accelerators and caches, exactly as needed. Everything that was once embodied in cards, disks and physical equipment of any kind, will be organized around a single point of control. If virtualization vendor promises materialize, changes that once were out of the question, or that at least would have required considerable man-hours and operational risk, will be done in minutes, routinely.

What those changes will mean is very much a topic for current discussion. For instance, all the new knobs and buttons virtualization provides will raise issues of policy, because it will be possible to discriminate among classes of service that once had to be handled together. You will, for instance, be able to write a Web server that gives customers who spend above a certain limit much better service than those who spend only half as much. There will be huge opportunities for automation. Infrastructure may be able to reconfigure itself in response to changes in demand, spinning out new servers and routers as necessary, the way load balancing is done today. (Certainly IBM et al. have been promoting just such a vision of the on-demand computing future.)

Virtualization examples so far have all been hardware-centric, because the inherent inflexibility of hardware means the elasticity advantages of virtualization are greater than with software. However, virtualization can work anywhere in the computing stack. You can virtualize both the hardware and the operating system, which allows programs written for one OS to run on another, and programs written for a virtual OS to run anywhere (similar to how Java maintains its hardware independence through the Java Virtual Machine).

Quite possibly the growth of virtualization predicts a deep change in the responsibilities of CIOs. Perhaps in the not-too-distant future no CIO will ever think about hardware: Raw physical processing and storage will be bought in bulk from information utilities or server farms. Applications will be the business of the departments or offices requiring them. The center of a CIO’s job will be the care and feeding of the execution environment. The very title of CIO might vanish, to be replaced, of course, by CVO.

Taking It All In
In that world, virtualization could graduate into a full-throated simulation of entire systems, the elements of which would not be just computing hardware, as now, but all the motors, switches, valves, doors, engines, vehicles and sensors in a company. The model would run in parallel with the physical company and in real-time. Where now virtualization is used for change management, disaster recovery planning, or maintenance scheduling for networks and their elements, it would in the future do the same for all facilities. Every object or product sold would come with a model of itself that could fit into one of these execution environments. It would be the CVO’s responsibility to make sure that each company’s image of itself was accurate and complete and captured the essentials. And that would not be a virtual responsibility in the least.