Wind River punts homegrown hypervisor
Extra goodies for Intel
AMD and VIA need not apply
But Evensen says these days, about half of Wind River's business is for Power platforms and half is for Intel chips. And when I say Intel, I mean Intel - Wind River does not support x86 and x64 chips from Advanced Micro Devices or VIA Technologies, and has no plans to do so at this time, according to Evensen, because customer demand is not there.
While VxWorks and Wind River Linux applications might not know it, the Wind River Hypervisor is not the same on the four different types of embedded platforms. The hypervisor has to do more things in software on PowerPC chips because IBM and Freescale have not added virtualization electronics to the PowerPC chips - although Power4, Power5, and Power6 chips from IBM have added increasingly sophisticated virtualization features, making the PowerVM hypervisor leaner and meaner on more modern iron.
Intel's Core 2 Duo, Nehalem, and Atom processors have VT-d electronics, and the Wind River Hypervisor takes advantage of those features. If IBM has any brains, it will make sure it has virtualization support in future PowerPC chips so it doesn't lose its game-console business and lots of other embedded controller business to Intel.
Wind River is not announcing specific launch dates for support on MIPS and ARM chips for the new hypervisor, but the company is working on it.
On servers, the advent of multicore and multithreaded processors have enabled companies to use virtualization hypervisors to consolidate multiple workloads and drive up CPU, memory, and I/O utilization on their servers. And the same forces are at work in the embedded systems market - but with a slightly different twist.
Because of security and application-isolation concerns, different applications on an embedded system have had to be run on separate system boards, which can make an embedded system larger, hotter, and more costly than it might otherwise be in the Age of Virtualization. Now, instead of creating custom interconnects between system boards and their dataplanes and custom ASICs, system makers can just use a multicore system and use the memory buses or point-to-point interconnects in the chips to have different processors and their application stacks communicate with each other.
This scheme is sometimes known as asymmetric multiprocessing (AMP), as distinct from symmetric multiprocessor (SMP), which is when multiple processor cores are lashed together and share a single memory space where a single operating system (or these days, a single hypervisor) runs.
According to Evensen, 70 per cent of Wind River's customers are not interested in virtualizing processors to carve up CPUs into skinny slices so much as treating each CPU as an isolated compute engine and virtualizing access to memory and I/O devices. This is another aspect of the scalable hypervisor that Wind River has created. If you don't want to chop up CPUs into skinnier computers, you don't activate and run that part of the hypervisor, which makes it run leaner and meaner.
The first customer for the new hypervisor is Hughes Networks, which has deployed the hypervisor on an unnamed chip inside of the dashboard of an unspecified car. The chip runs the Wind River Hypervisor and has a small, fast-boot Linux kernel that is used to control the engine and ignition system, which you want to turn on as soon as you put the key into the slot. On another partition is a Linux instance that is running the screen and other components of the dashboard screen, which can take a little longer to boot and not keep the car from starting as it boots up.
Wind River did not supply pricing for its hypervisor. Its operating systems and Workbench tools are licensed by developer seats and then customers pay for runtime licenses, depending on how they deploy the hypervisor and operating systems in their embedded devices.
One more thing: Wind River has proven that it is possible to have a single, cross-chip hypervisor - something the commercial IT industry desperately needs but will probably never have because VMware, IBM, Hewlett-Packard, Citrix Systems, and Red Hat have more to gain by keeping control of their hypervisors and related tool stacks than working together to create a common hypervisor and differentiated or similar tool stacks. ®
Sponsored: Benefits from the lessons learned in HPC