StorMagic SvSAN: A virtual SAN made simple
StorMagic SvSAN is a virtual SAN - a software-defined solution designed to run on two or more servers and deliver highly available shared storage.
SvSAN simplifies your IT infrastructure. It eliminates the need for a physical SAN, enabling hyperconvergence by virtualizing the internal compute and storage of any x86 server and presenting it via a hypervisor as shared storage.
A typical two-node SvSAN configuration, with a centralized management interface and witness, is shown below.
This SvSAN data sheet is broken down into four sections, covering SvSAN's features, its requirements, hardware and software compatibility and finally support levels.
For a greater technical examination of StorMagic SvSAN, including deployment options and use cases, please refer to the SvSAN Technical Overview white paper.
StorMagic SvSAN has a range of features enabling the storage architect to get the most out of their infrastructure. These features are detailed in the table at the bottom of this page.
All of the features necessary for highly available shared storage are included in an SvSAN license, including the witness. There are also two performance and security-enhancing add-ons available, which are SvSAN's caching and data encryption features. Many of these features are covered in more detail within their own white papers. View the full range on the StorMagic website.
SvSAN provides highly available shared storage on just two nodes using a lightweight witness. The SvSAN witness prevents split-brain scenarios from occurring in two node SvSAN clusters by regularly checking the state of each node in the cluster. The only data passing between the witness and the SvSAN nodes is the ‘heartbeat’ - the witness is not in the data path. This allows it to tolerate significant latency and low bandwidth. For more details, see the SvSAN Witness data sheet here.
SvSAN's caching features include write-back and read-ahead caching, as well as data pinning. Collectively these features are known as Predictive Storage Caching and can dramatically improve an organization's storage performance without requiring significant investment in new hardware.
SvSAN's data encryption feature enables organizations to encrypt the data being mirrored by SvSAN. This allows vulnerable edge sites and the data they hold to be protected. It is FIPS 140-2 compliant, eliminates the need for expensive OS or hypervisor-level solutions, and is compatible with any KMS that uses KMIP, including StorMagic's own key manager, SvKMS.
SvSAN offers a no-cost Container Storage Interface (CSI) driver for container use at the edge. SvSAN with the CSI driver provides persistent, highly available storage for Kubernetes-orchestrated container deployments. SvSAN supports VMs, containers, or both on the same 2-node cluster. For full details of the CSI driver’s system requirements and compatibility, and more information about this deployment option, refer to the SvSAN Container Storage Interface Driver Data Sheet.
StorMagic SvSAN is licensed based on the usable VSA storage capacity. License tiers are set at 2TB, 6TB, 12TB and Unlimited TB.
SvSAN is available as a perpetual or subscription license. After a single payment, perpetual SvSAN licenses can be used forever, with ongoing costs only for support renewal payments. Subscription SvSAN licenses are paid for upfront for a specific time period, one year for example, and then renewed as required. Pricing is based on a single SvSAN license - the number of nodes in the cluster defines how many licenses are required.
For example, two SvSAN licenses are required for a normal 2-node deployment. More information can be found on the SvSAN Pricing webpage.
A free, fully functional evaluation of SvSAN is available to download, enabling organizations to trial and experience the features and benefits of SvSAN, before purchasing.
For more information and to download an evaluation copy, visit: stormagic.com/trial.
SvSAN has very low minimum hardware requirements, detailed in the table opposite.
Witness system requirements
The witness sits separately from the SvSAN nodes and therefore has its own minimum requirements in it's own table opposite. The witness should be installed onto a server separate from the SvSAN VSA.
Hardware and software compatibility
SvSAN works with any x86 server that exists on the VMware vSphere ESXi or Microsoft Hyper-V Hardware Compatibility List (HCL). Furthermore, SvSAN will work with any supported internal server disk storage or JBOD array and supports servers without hardware RAID controllers, due to its software RAID 10 capability.
SvSAN supports VMware vSphere, Microsoft Hyper-V and Linux KVM hypervisors. It is installed as a Virtual Storage Appliance (VSA) requiring minimal server resources to provide the shared storage necessary to enable the advanced hypervisor features.
SvSAN is supported on the versions of VMware vSphere ESXi, Microsoft Windows Server/Hyper-V Server, and Linux KVM distributions listed in the table opposite.
If VMware vSphere is chosen as the hypervisor to deploy with SvSAN, StorMagic recommends vSphere Essentials Plus as a minimum to enable high availability.
For further details of SvSAN’s capabilities on KVM hypervisors, please refer to the SvSAN with KVM data sheet.
VMware vCenter support
With the dedicated plugin, SvSAN can be managed directly from VMware vCenter. SvSAN is compatible with the versions of vCenter in the table below.
VMware vCenter versions compatible with SvSAN:
|VMware vCenter version||SvSAN version|
|VMware vCenter Server 7.0 & updates||•|
|VMware vCenter Server 6.7 & updates||•|
|VMware vCenter Server 6.5 & updates||•||•||•|
|VMware vCenter Server 6.0 & updates||•||•|
StorMagic SvSAN system requirements:
|Servers||2 servers required
• 3 optional for high availability during offline upgrades, maintenance, etc
|CPU||1 x virtual CPU core1
• 2GHz or higher reserved
|Disk||2 x virtual storage devices used by VSA
• 1 x 512MB Boot device
• 1 x 20GB Journal Disk
|Network||1 x 1Gb Ethernet
• Multiple interfaces required for resiliency
• 10Gb Ethernet is supported
• Jumbo frames supported
1When using SvSAN's data encryption to encrypt data, 2+ virtual CPUs are recommended
2Additional RAM may be required when caching is enabled
Requirements for containers
The following software versions are required to deploy containers on StorMagic SvSAN:
|StorMagic SvSAN||6.2 Update 5|
|SvSAN CSI driver||1.0.0|
|Kubernetes||1.19 or 1.20|
|VMware Tanzu Kubernetes Grid1||1.2|
1Optional - for running Kubernetes workloads natively on the hypervisor layer
SvSAN witness requirements:
|CPU||1 x virtual CPU core (1 GHz)|
|Network||1 x 1Gb Ethernet NIC
When using the witness over a WAN link use the following for optimal operation:
• Latency of less than 3000ms, this would allow the witness to be located anywhere in the world
• 9Kb/s of available network bandwidth between the VSA and witness (less than 100 bytes of data is transmitted per second)
|Operating System||The SvSAN witness can be deployed onto a physical server or virtual machine with the following:
• Windows Server 2016 (64-bit)
• Hyper-V Server 2016 (64-bit)
• Raspbian Buster (32-bit)
• vCenter Server Appliance (vCSA)1
• StorMagic SvSAN Witness Appliance
• CentOS 7.7, 7.8, 8.1 & 8.2
• RHEL 7.7, 7.8, 8.1 & 8.2
1VMware vSphere 5.5 and higher
Hypervisors compatible with SvSAN:
|VMware||vSphere 6.0 & updates||•||•|
|vSphere 6.5 & updates||•||•||•|
|vSphere 6.7 & updates||•||•|
|vSphere 7.0 & updates1||•||•|
|Microsoft||Windows Server 2016||•||•||•|
|Hyper-V Server 2016||•||•||•|
|Windows Server 20192||•|
|Hyper-V Server 20192||•|
|Linux KVM3||CentOS 7.5||•||•|
1vSphere 7.0 compatibility is available from SvSAN 6.2 Update 2 Patch 2 onwards.
2For customers looking to use Windows Server 2019 or Hyper-V Server 2019, please contact StorMagic Support at email@example.com for further details and assistance
3Linux KVM compatibility is available for SvSAN 6.2 Patch 5 and onwards.
SvSAN Maintenance & Support
SvSAN Maintenance & Support provides organizations with access to StorMagic support resources, including product updates, knowledgebase access and email support with our technical support staff.
Two levels are available. A summary of each is shown in the table below. More information on SvSAN Maintenance & Support can be found at stormagic.com/svsan/support.
|Gold Support||Platinum Support|
|Hours of operation||8 hours a day1 (Mon - Fri)||24 hours a day2, (7 days a week)|
|Length of service||1, 3 or 5 years||1, 3 or 5 years|
|Access method||Email + Telephone (via platinum engagement form on
|Response method||Email + WebEx||Email + Telephone + WebEx|
|Remote support / WebEx||Yes||Yes|
|Maximum number of support administrators per contract||2||4|
|Response time||4 hours||1 hour|
1Gold Support is only available from 07:00 UTC/DST to 01:00 UTC/DST. If your business hours fall outside this window, you must purchase Platinum Support.
2Global, 24x7 support for Severity 1 - Critical Down & Severity 2 Degraded issues
SvSAN Feature Table
|Synchronous mirroring/high availability
• Data is written to two SvSAN VSA nodes to ensure service uptime
• Write operations only complete once acknowledged on both SvSAN VSAs
• In the event of a failure, applications are failed over to other available resources
|Stretched/metro cluster support - white paper with more information
• Separate nodes geographically to provide an added layer of resiliency
• Different racks, separate rooms or buildings, or even across an entire city
|Volume migration - white paper with more information
• Transparently and non-disruptively migrate volumes from one storage location to another
• Simple and mirrored volumes can be migrated between storage pools on the same SvSAN VSA node or to another SvSAN VSA node entirely
|VMware Fault Tolerance feature
• SvSAN deployed on VMware vSphere hypervisor enables the usage of VMware’s Fault Tolerance feature on clusters of just two nodes
• Fault Tolerance protected VMs see zero downtime or loss of service when one node goes offline
• Keep critical applications online and running in the event of a node failure
|VSA restore (VMware only)
• Automates the recovery process of an SvSAN VSA node following a server failure or replacement
• SvSAN VSA configuration changes are tracked and stored on another SvSAN VSA in the cluster
• Mirror targets are rebuilt and resynchronized, enabling a quick return to optimal service
• Simple targets can be automatically recreated, ready for data recovery from backup
|VMware vSphere Storage API (VAAI) support (VMware only)
• Accelerating VMware I/O operations by offloading them to SvSAN
• Supports the Write Same, Atomic Test & Set (ATS) and UNMAP primitives
|Centralized monitoring and management
• Monitor and manage SvSAN from a single location with multiple options including WebGUI
• Seamless integration with vCenter Web Client enabling alerts to be forwarded and captured on one screen
• Email alert notifications using SMTP, and SNMP integration with support for v2 and v3
|Witness - white paper with more information
• Acts as a quorum or tiebreaker and assists cluster leadership elections to prevent “split-brain”
• Hundreds of locations can share a single witness and it tolerates low bandwidth, high latency WAN links
• Supported configurations include local witness, remote shared witness or no witness
|I/O performance statistics
• Provides granular, historical I/O transaction, throughput and latency statistics for each volume
• Simple, intuitive graphical presentation with minimum, maximum, and average values for daily, monthly, yearly time periods
• Data can be exported to CSV for further analysis
|Multiple VSA GUI deployment and upgrade
• Deploy and upgrade VSAs through a single wizard either immediately or in a staged approach for out-of-hours activity
• SvSAN handles dependencies and performs a health check ensuring there is no impact to environments
|PowerShell script generation
• Deployments over many locations can be handled by generating a custom PowerShell script
• Simplifies the process of upgrading multiple VSAs with full control over date/time and how many/which clusters to upgrade
• Upgrades can be carried out on one or more SvSAN VSAs simultaneously, ensuring storage remains online throughout
• Automated process stages the firmware, checks cluster health, then proceeds with upgrading each VSA in turn
• Install SvSAN on servers without hardware RAID controllers, such as the Lenovo SE350 and HPE Edgeline EL4000
• Configure as RAID 0 (striping) or RAID 10 as required
|Container Storage Interface (CSI) Driver - white paper with more information
• Easily deploy containers at the edge with persistent, highly available storage
• No additional costs to deploy containers in an SvSAN cluster
• One platform at the edge for VMs, containers, or both
|Additional add-on features available|
|Predictive read ahead caching (SSD and memory) - white paper with more information
• Beneficial to sequential read workloads - populates memory with data prior to being requested
• Boosts performance by reducing I/O requests going to disk, instead serving data from low latency memory
|Write back caching (SSD) - white paper with more information
• Utilizes SSDs to improve the performance of all write operations by lowering latencies and increasing the effective IOPS, resulting in faster response times, especially for random write workloads
• All write I/O’s are directed to the SSD allowing completion to be immediately acknowledged back to the server, at a later time the data is written from SSD to the hard disk
|Data pinning - white paper with more information
• Allows data to permanently reside in memory, ensuring that data is always available in the highest performing, lowest latency cache tier, useful for frequently repeated operations such as booting virtual machines
• Intelligent caching algorithms identify 'hot' and 'cold' data, elevating the 'hot' data to the highest performing, lowest latency storage tier (SSD or memory)
|Data encryption - white paper with more information
• Utilizes a FIPS 140-2 compliant algorithm (XTS-AES-256) to deliver encryption to all data handled by SvSAN or just selected volumes
• Allows secure erasure and re-key
• Compatible with any KMIP-compliant key management system including StorMagic SvKMS