NutaNIX
A storage-math waterfall on the left showing 100 TB raw stepping down through RF2 (50 TB) and compression (~85 TB) and a 15-percent reservation to ~72 TB effective usable, alongside a VM-density bar chart on the right showing VDI sessions per node by user profile (Task 180, Knowledge 120, Power user 80, Developer 50). Practical-ceiling and failure-tolerance strips at the bottom.
/nix/nutanix/appendix-f-sizing-rules

Appendix F · Sizing Rules

REFERENCE back-of-envelope ~25% accuracy for early conversations Sizer is the source of truth

These are rules of thumb for initial sizing conversations. They get you within roughly 25% for early planning. Final sizing requires Nutanix Sizer with actual workload data, and that's the number that goes into the proposal. When in doubt, run Sizer.

The discipline at a glance:

Cluster Sizing Fundamentals

Minimum and Recommended Cluster Sizes

For final sizingNutanix Sizer factors workload mix, growth, and failure-tolerance targets into the recommended cluster size.

Failure-Tolerance Math

SchemeFailures ToleratedMinimum NodesNotes
RF213 (N+1)Rebuild at degraded redundancy; second failure during window can cause data loss
RF32 simultaneous5 (N+2)Survives one more failure during rebuild
EC-X 4+116 recommendedStripe technically requires 5 nodes; Nutanix recommends 6+ for rebuild headroom (per TN-2032)
EC-X 4+227+Two-failure tolerance with capacity savings vs RF3
Rule of thumbDon't run RF3 unless you genuinely need it. Most enterprise workloads run RF2 with proper backup. RF3 is for compliance-mandated, mission-critical, or workloads without external backup.

Compute Sizing

VM Density per Node

Typical mid-market mixed VMs (2 vCPU, 8 GB RAM average):

Heavier VMs (4-8 vCPU, 16-32 GB RAM): 20-40 VMs per node typical. More memory-bound than CPU-bound for typical mixed workloads.

VDI density:

For final sizingSizer takes per-VM specs as input and produces node count.

CVM Tax Planning

Every Nutanix node runs a Controller VM. Reserve compute for it:

Workload ProfileCVM vCPUCVM RAM
Light (low IOPS, small cluster)832 GB
Standard (typical enterprise)1248 GB
Heavy (high IOPS, larger cluster)1664 GB
Very heavy (dense IOPS, large cluster)20+96+ GB
Rule of thumbReserve 10-15% of node compute for the CVM. Sizer is more precise.

Headroom Recommendations

ResourcePractical CeilingReasoning
CPU70% peak utilizationSpike absorption, rebalancing, growth
RAM75-85%Memory ballooning is degraded performance
Storage75%Capacity for snapshots, growth, rebalancing
Network bandwidth60-70% peakReplication and rebuild traffic compete
Rule of thumbSize for 18-24 months of growth at the projected rate. Beyond that, plan for cluster expansion (add nodes).

Storage Sizing

Raw to Usable Conversion

The path from "how much disk are we buying" to "how much can the customer actually store":

StepMultiplierExample (100 TB raw)
Raw drive capacity1.00100 TB
Replication Factor 20.5050 TB
Replication Factor 30.3333 TB
EC-X 4+1 (instead of RF2)0.8080 TB
EC-X 4+2 (instead of RF3)0.6767 TB
Compression (typical 1.5-2.0× effective)1.5-2.0×75-100 TB after RF2
Dedup (workload-dependent, 1.0-1.5×)1.0-1.5×additional savings on high-similarity data
Free-space reservation0.85reduce by 15% for headroom
Rule of thumb (mixed enterprise)RF2 + compression + 15% headroom: ~70-80% effective usable vs raw before dedup. RF2 + compression + dedup on VDI: ~110-150% effective vs raw (higher than 100% due to dedup). RF3 + compression: ~50-55% effective vs raw.

Compression Reduction Ratios

Quote ranges, not marketing peaks:

Workload TypeTypical Compression Ratio
OS images (Windows, Linux)1.5-2.5×
General application data1.2-1.8×
Already-compressed (video, images, encrypted)1.0-1.1×
Database (varies)1.3-2.0×
Backup data1.0-1.2× (already-compressed by backup software)
Logs2-4×
Rule of thumbPlan on 1.5-2.0× effective compression for mixed enterprise. Anything above 2× is a bonus. Watch out for customer expectations primed by marketing 4-6× claims; reset early.

Deduplication Ratios

Workload TypeTypical Dedup Ratio
VDI persistent profiles2-3×
VDI non-persistent (similar OS images)3-5×
General-purpose VMs1.0-1.3×
Application data1.0-1.2×
Backup data1.0× (already deduped by backup software)
Rule of thumbEnable on-disk dedup for VDI; leave disabled for general workloads unless POC shows ratio above 1.3×. Dedup adds metadata overhead; not free.

Network Sizing

Cluster ProfileRecommended Link
Small ROBO (1-3 nodes, light workload)10 GbE
Mid-market production (4-8 nodes)25 GbE
Enterprise production (8+ nodes)25 GbE minimum, 100 GbE for high-density
All-NVMe / very-high-IOPS100 GbE
VDI at scale25 GbE acceptable, 100 GbE for >2,000 sessions
Rule of thumb25 GbE is the sweet spot for new deployments. 10 GbE is sufficient for small clusters and many ROBO sites. 100 GbE matters when you're saturating 25 GbE on east-west traffic. 2 NICs minimum for redundancy; 4 NICs typical for separating data, management, replication.

Bond Mode Selection

Bond ModeUse CaseSwitch Coordination
active-backupSimplest; good for most ROBO and small productionNone
balance-slbActive-active without LACP; good throughput, no switch configNone
balance-tcp / LACPBest throughput and load distributionRequired (LACP on switch side)
active-active LACPMaximum performance, MC-LAG capableRequired, advanced switch features
Rule of thumbactive-backup for ROBO and small clusters where simplicity matters. balance-slb for production where switch coordination is hard. LACP for production where the network team controls the switch and can configure it. Nutanix has historically cautioned against LACP without disciplined switch-side validation.

Replication Bandwidth Estimation

For DR planning, estimate WAN bandwidth needed for replication:

Async replication (1-hour RPO):

NearSync (sub-15-min RPO): bandwidth must support continuous change rate plus burst handling. Typically 2-3× the Async sustained number. Plus low-latency requirement (<5 ms RTT typical).

Metro Availability: synchronous; must support write throughput in both directions. Plan for full peak write bandwidth, not average.

Workload-Specific Sizing

VDI

Baseline density assumption: 100 sessions per node for typical knowledge workers, all-NVMe configuration with 768+ GB RAM.

ProfileSessions per Node
Task worker (light apps, browser)120-180
Knowledge worker (Office, browser, light specialty)80-120
Power user (specialty apps, multiple tools)50-80
Developer / heavy workload30-50

Storage: persistent profiles 30-80 GB per user before dedup; non-persistent share base image plus delta. Plan capacity for the full population, not the concurrent count, if persistent.

Boot-storm planningAll-NVMe handles boot storms gracefully; spinning-disk configurations don't. For VDI at scale, all-NVMe is non-negotiable.

Databases

SQL Server / Oracle (typical OLTP):

DB SizeCluster Size Suggestion
Up to 5 TBFits in any production cluster
5-20 TB6+ nodes for I/O distribution
20-50 TB8+ nodes; consider dedicated database cluster
50+ TBDedicated database cluster; specialty design

Containers and Kubernetes

AI / ML Training

Rule of thumbAI training is rarely a Nutanix-only workload at scale; customers often combine on-prem (steady-state datasets) with cloud GPU (burst training). Hybrid is the typical pattern.

Files / Objects / Volumes Sizing

Files

FSVM count and capacity:

File Server SizeFSVM CountNotes
Up to 50 TB3 FSVMsDefault baseline
50-200 TB3-5 FSVMsScale per workload
200 TB+5+ FSVMs, possibly multiple File ServersScale and isolation

Reservations and overhead: DSF storage container backing Files uses the same RF and compression rules as general DSF. Plan 20% overhead for snapshots, file system metadata, growth headroom.

Use cases:

Rule of thumbStart with 3 FSVMs (the HA minimum); scale up if performance metrics show saturation. Single-FSVM File Servers exist for 1-2 node small environments per TN-2041, but give up HA-across-FSVMs.

Objects

By use case:

Volumes

DR Cluster Sizing

Primary-to-DR Sizing Relationship

DR PatternDR Cluster Size
Cold DR (rebuild from backup)None / minimal
Warm DR (Async replication, restart on failover)50-100% of primary capacity
Hot DR (active-passive Metro)100% of primary capacity
Active-Active MetroBoth sites size for full load if other fails
Rule of thumbDR cluster typically sized for 60-80% of primary if running steady-state DR; 100% for active-active. NC2 cloud DR sizes for failover capacity (Tier-1 + Tier-2 typically); use cloud's elastic provisioning to scale up only at failover time.

Replication Bandwidth Quick Calculation

Step 1: Estimate daily change rate. Typical mid-market: 2-5% of total capacity per day.

Step 2: Convert to per-second sustained bandwidth.

Daily change (TB) × 8,000 (Gbits per TB) / 86,400 (seconds per day) = sustained Gbps

Step 3: Apply replication-mode multiplier:

ModeMultiplier on Sustained
Async (1+ hour RPO)1.0× plus 50% catchup headroom
NearSync2.0-3.0× for burst handling
MetroSized for peak write throughput, not average

Step 4: Add WAN compression benefit if applicable. Native compression in Nutanix replication often provides 1.3-1.7× effective bandwidth multiplier.

Worked example:

Rule of thumbFor mid-market, plan 200-500 Mbps WAN for typical Async DR; more for NearSync; full peak write rate for Metro.

Prism Central Sizing

Single PC vs Scale-Out

X-Small PC VM (introduced for smaller environments):

Single PC VM (Small / Standard): typical mid-market through enterprise deployments. Single VM; simpler operations. Single point of failure for management plane (clusters continue running if PC is down).

Scale-out PC (3 VMs): required for the largest deployments and production-grade PC HA. Distributed across hosts. Recommended for enterprise deployments at scale.

PC VM resource specs (verify against current Prism Central sizing guide; numbers evolve):

PC ModevCPURAMStorage
X-Small418 GB100 GB
Small626 GB500 GB
Large (3-VM scale-out)6 per VM (× 3)28 GB per VM (× 3)500 GB per VM
For final sizingVerify against the current Nutanix Prism Central sizing guide at design time. The size tiers and their VM-count ceilings shift between PC versions; exceeding documented limits results in an unsupported configuration.

When to Use These Rules vs Sizer

Use these rules

Switch to formal Sizer when

The Sizer workflow

  1. Customer provides workload data: VM list, resource consumption, growth assumptions.
  2. Sizer produces node count, model recommendations, capacity calculations.
  3. Output is the basis for the proposal BoM.
  4. Sizer output gets validated during POC and refined before order.

The rules in this appendix accelerate the conversation; Sizer produces the number that goes in the contract. Don't conflate the two.

Common Sizing Mistakes to Avoid

  1. Sizing for the average, not the peak. Workloads have spikes; size for peak with headroom.
  2. Ignoring the CVM tax. Reserve compute for the CVM; it's not free.
  3. Forgetting failure-tolerance math. RF2 needs N+1; RF3 needs N+2.
  4. Quoting marketing compression ratios. 1.5-2.5× is the honest range for mixed workloads.
  5. Skipping the headroom reservation. 100% utilization is theoretical; 70-80% is the practical ceiling.
  6. Underestimating replication bandwidth. Async DR is rarely the smallest WAN bill.
  7. Missing growth headroom. Size for 18-24 months; expand cluster after.
  8. Trusting customer's "typical" without validation. Their typical is often a peak; their peak is often catastrophic. Measure during POC.
  9. Forgetting that Files, Objects, Volumes consume the same DSF capacity. Size the cluster for total demand, not per-service.
  10. Skipping PC sizing. Prism Central is part of the design, not an afterthought.

References

The sizing rules in this appendix are heuristics for early conversations; the authoritative source is Nutanix Sizer for binding sizing, and the public sources below for tier-by-tier specifics:

Cross-References

  • Modules: Each section links back to the module where the topic is taught in depth.
  • Glossary: Appendix A defines the terms used.
  • Comparison Matrix: Appendix B helps with competitive sizing comparisons.
  • Reference Architectures: Appendix I has fully-sized reference designs.
  • POC Playbook: Appendix J has the validation steps that confirm sizing during proof-of-concept.
  • Nutanix Sizer: the official tool for binding sizing. Always run Sizer for the proposal.