Deploy Up to 5X More VMs on Any Storage, Lower Costs by Up to 50%
Atlantis ILIO USX™ is a significant breakthrough in virtualization technology, delivering a solution that unlocks the underutilized capacity of over $50 billion of deployed enterprise storage, similar to what VMware did for server virtualization.
Atlantis ILIO USX software gives IT the flexibility to get more out of their existing storage, even their older arrays, and to create new software-defined storage hybrid arrays, hyper-converged systems, and all-flash arrays by aggregating and pooling existing server SSDs, SAS, flash, and RAM together with shared SAN/NAS arrays.
Atlantis ILIO USX pools and optimizes server RAM, SAS, and/or Flash to create a highly scalable, hyper-converged platform using existing servers. Atlantis ILIO USX provides the flexibility to pool commodity local storage with RAM and/or flash across multiple server farms. By doing so, customers can seamlessly scale out their architecture to create a hyper-converged infrastructure and manage their applications without having to rip out their existing infrastructure.
Atlantis ILIO USX optimizes how storage is consumed by the application or VM by inserting a transparent software layer between the application and storage. The Atlantis ILIO USX software resides on the hypervisor platform as a set of virtual machines that can abstract any storage hardware into pools of virtual storage that can be combined together to form an Application Defined Storage Volume (ADS Volume). There are two types of storage pools:
These Capacity and Memory pools are combined together automatically to create Application Defined Storage Volumes (ADS Volume) with policy-based controls that apply the ideal combination of storage capacity, performance and availability for the application. The ADS Volume provides enterprise-class storage functionality including high-availability, data protection, thin provisioning and cloning. Multiple ADS Volumes can be created from a single Capacity and Performance Pool, enabling true application-centric storage for the first time.
The Atlantis ILIO USX platform applies multiple storage optimization technologies to boost the performance and increase the available storage capacity provided to the application. At the same time, Atlantis ILIO USX dramatically reduces the impact of application IO traffic on storage resource utilization (disk, storage controller, network).
Use with existing Shared Storage
Use Local SAS or SATA
Use RAM as Primary Storage
Optimized All-Flash Arrays
In-Memory Architecture — Atlantis ILIO USX can run VMs completely in server RAM to deliver high-speed, low latency runtime storage without re-architecting applications
IO Processing — Atlantis ILIO USX processes IO operations in real-time at the compute layer to lower latency and reduce network traffic
Inline De-duplication — Atlantis ILIO USX performs inline de-duplication in real-time on-the-wire with microsecond latency, eliminating up to 90% of storage IO traffic
Real-Time Compression — Atlantis ILIO USX compresses the optimized blocks In-Memory with microsecond latency
IO Blender Fix — Atlantis ILIO USX coalesces small random blocks generated by the hypervisor into larger sequential blocks, greatly improving storage access and efficiency
High Availability — Atlantis ILIO USX provides integrated high-availability and data protection to prevent application downtime.
Thin Provisioning — All Atlantis ILIO USX storage volumes are automatically thin provisioned with up to 10:1 consolidation.
Fast Clone — Atlantis ILIO USX can clone full VMs in as little as 4 seconds with no network or storage traffic.
List of new features in VMware vSphere 5.5
• Hot-pluggable SSD PCIe devices
• Support for Reliable Memory Technology
• Enhancements to CPU C-states
• Flash Read Cache (vFRC) or vFLASH – Virtualises server-side flash to provide high performance read cache layer that dramatically lowers application latency.
• vSAN – software based distributed storage solution which is built directly in the hypervisor.
Along with the core vSphere ESXi Hypervisor improvements, vSphere 5.5 provides the following virtual machine–related enhancements:
• Virtual machine compatibility with VMware ESXi 5.5
• Expanded support for hardware-accelerated graphics vendor
• Graphic acceleration support for Linux guest operating systems
In addition, the following vCenter Server enhancements include:
• vCenter Single Sign-On Server security enhancements
• vSphere Web Client platform support and UI improvements
• vCenter Server Appliance configuration maximum increases
• Simplified vSphere App HA application monitoring
• vSphere DRS virtual machine–virtual machine affinity rule enhancements
• vSphere Big Data Extensions, a new feature that deploys and manages Hadoop clusters on vSphere fromwithin vCenter
vSphere 5.5 also includes the following storage-related enhancements:
• Support for 62TB VMDK
• MSCS updates
• vSphere 5.1 enhancements
• 16GB E2E support
• PDL AutoRemove
• vSphere Replication interoperability and multi-point-in-time snapshot retention
vSphere 5.5 also introduces the following networking-related enhancements:
• Improved LACP capabilities
• Traffic filtering
• Quality of Service tagging
• 40 GB nic Support
• Enhanced Host level capture
• SR-IOV – Single Root I/O Virtualisation enhancements.
Adaptive PSP is PSP plug-in that VMware developed to adaptively switch load-balancing strategies based on system load.
Moving forward, VMware are working to further enhance the adaptive logic by introducing a path scoring attribute that ranks different paths based on I/O latency, bandwidth, and other factors.
The score is used to decide whether a specific path should be used for different system I/O load conditions. Further, the logic decides the percentage of I/O requests that should be dispatched to a certain path, and could also combine the path score with I/O priorities, by introducing priority queuing within PSA.
This is the third and final tech preview of storage features which were shown at VMworld 2012. Already we have looked at two tech previews – Distributed Storage and Virtual Volumes (VVOLs). This next feature is vFlash or Virtual Flash, and will look at a project underway here at VMware to integrate local flash devices with vSphere. This will make significant performance improvements and reduce I/O latency for many workloads running in the Virtual Machine.
To date, VMware has done very little around flash. We only have the smartd introduced in vSphere 5.1 to monitor Solid State Drives (SSD) attributes and the swap to SSD feature. Those of you who have read the technical preview article on Distributed Storage will have read about SSD being used as a cache (for both read and write I/O). This post discusses an additional project called vFlash, the purpose of which is to integrate vSphere with local flash devices. What we are looking to do is to enable a new tier of storage for your Virtual Machines. For those of you unfamiliar with flash technology, I recently wrote a blog article on my personal blog which will give you a pretty decent overview.
Please note that since this is a tech preview article, there is no guarantee when this feature will appear (if ever), nor is there any guarantee that the end product will encompass any or all of the attributes discussed in the post. It is simply to give you an idea of what we are working on, and what features the final product might include. vFlash was discussed at VMworld 2012 & my colleague Duncan provides a nice overview of the session on his blog here.
We want our customers to be able to select off the shelf flash devices, be they SSD or PCIe I/O Accelerator cards. Once the flash devices are plugged into vSphere hosts, we have a toolset available in vSphere to manage flash as a resource, just like you currently manage CPU and memory resources. Basically, you will be creating a flash pool, the contents of which will be carved up to provide flash resources to individual VMs using constructs that you are already familiar with, such as reservations, limits & shares.
vFlash is a framework. VMware will be providing our own default software to plugin (vFC) to the framework, but other flash vendors can create their own plugins (vFlash Cache Modules) containing bespoke and proprietary algorithms for utilizing their specific cache/flash devices. The vision is to publish a set of APIs to support this, and share them with potential partners.
The vFlash infrastructure does not specify which Virtual Disk data is to be cached in vFlash. That decision is left up to the vFlash Cache Module. It is envisioned that the vFlash Framework will support multiple vFlash Cache Modules per ESXi host. Different Virtual Machines running on the same host can be configured to use different vFlash Cache Modules. This permits the vFlash caching algorithm used for a given Virtual Disk to be tailored to the storage I/O behavior of the application running in the Virtual Machine using that Virtual Disk.
That caching software can come in two forms – VM-aware caching & VM-transparent caching. With VM-transparent caching, the VMs will share a pool of flash resources and are not aware that they are doing so; they will simply benefit from having flash in their I/O path. With VM-aware caching, chunks of flash can be allocated on a per VM basis, and these will show up as SCSI disks in the Guest OS.
VM-aware Caching (vFlash Memory)
This is where a flash resource is presented directly to the VM. A new virtual hardware device called vFlash Memory is added to the Virtual Machine. The interesting part of this approach is that the caching algorithm needs to be controlled by the VM, and not by the caching software on the hypervisor. This may possibly entail the installation of agents in the Guest OS, in a similar fashion to how some flash vendors currently do things. The cache appears as a disk drive to the Guest OS, which can then be formatted and used appropriately by applications in the Guest OS, and thus can benefit from having access to a flash device.
VM-transparent Caching (vFlash Cache)
This is where the VM is unaware that there is a flash cache in the I/O path, but benefits from it all the same. A new virtual hardware device called vFlash Cache is added to the VM. In this case, cache software on the hypervisor is there to provide a suitable algorithm for the I/O. Options available during the configuration of VM-transparent caching will include reservation size, the choice of vFlash Cache Module, mode of operation (write-back or write-thru), block size (tuned to Guest OS requirements) and what to do with the flash contents during a migration (migrate or drop). The user will have the option of migrating flash contents if the destination host is compatible. The default option is to attempt to migrate the content, but if the destination host is incompatible, then we will drop the flash contents and rebuild it on the destination host. Obviously this will have a performance impact as the flash contents would need to ‘warm up’ on the destination host. VMs with flash cannot be migrated to a destiantion host with insufficient vFlash resources because even if the contents are not migrated, the flash contents will need to be rebuilt at the destination host and we need to ensure that appropriate resources are available to do this. vCenter compatability checks will fail the migration if the destination host does not have sufficient cache resources.
Similarly, when considering vSphere HA, flash will need to be pre-allocated or set aside to ensure that enough resources are available for virtual machines to successfully restart on remaining hosts in the event of a host failure.
Direct attached flash storage is gaining significant momentum right now, and is an ideal reource to provide low latency for latency sensitive applications, enabling even more tier 1 applications to be virtualized. VMware wants to enable our customers to leverage flash resources through well known and well understood vSphere mechanisms.
Get notification of these blogs postings and more VMware Storage information by following me on Twitter: @VMwareStorage
Upgrade vSphere Enterprise Plus to vCloud Suite Standard for Free
Offer expires March 15, 2013
Diskless Virtual Desktop Infrastructure (VDI) is the concept of using local server memory in combination with storage optimization software to store virtual desktop images instead of shared SAN/NAS or local SAS/SSD storage.
By storing virtual desktop images on the local memory of the hypervisor where the desktops execute, response time are faster than even the most expensive local SSD drives (MLC or SLC), cost less when combined with Atlantis ILIO, and increase reliability.
With existing VDI architectures, virtual desktop images are stored on either shared SAN/NAS storage or local SSD disks, which are costly, have limited IOPS for write-intensive VDI workloads, can have a limited lifespan and consume more power than memory.