NutaNIX
A horizontal four-week timeline showing the POC cadence with Week 1 setup, Week 2 core validations, Week 3 advanced validations, Week 4 closeout, terminating in a green DECIDE arrow. An eight-item qualification checklist runs above; ten demo-script cards plus a scope-boundary tile and a final-report tile run below. A five-principle strip closes the bottom.
/nix/nutanix/appendix-j-poc-playbook

Appendix J · POC Playbook

REFERENCE qualify · scope · test · report 10 demo scripts 30-day cadence final-report template

The POC is often the buying decision. Get it right and the customer commits; get it wrong and you've spent weeks confirming a "no." This appendix is the framework BlueAlly SAs use to run POCs that produce decisions.

Five principles for POCs:

  1. Qualify before committing. POCs cost both sides; not every opportunity warrants one.
  2. Define success criteria with the customer, in writing, before the POC starts. Without this, POCs have no end.
  3. Test what matters to the customer, not what showcases the platform.
  4. Get the customer's hands on the platform. Operational validation matters as much as technical validation.
  5. Document everything. The final report is the artifact that converts POC results into a decision.

1POC vs Other Engagement Types

POC (Proof of Concept)

Definition: Deploy actual Nutanix hardware (or NC2 cloud cluster) at the customer site or in BlueAlly's lab; customer team participates in deployment, configuration, workload migration, and operational validation; runs typically 2-6 weeks; produces a documented validation against agreed success criteria.

When right
  • Customer is in active evaluation with budget and timeline
  • Specific technical questions need empirical answers
  • Customer team needs hands-on validation before committing
  • Decision authority is engaged and ready to act on POC results
When wrong
  • Customer is in early discovery without trigger or budget
  • Specific feature questions could be answered in a 2-hour demo
  • Decision authority isn't engaged
  • Success criteria can't be agreed upon

Evaluation

Definition: Lighter than a POC. Often a guided demo plus customer hands-on time in a BlueAlly-hosted lab; days to a week; doesn't require dedicated hardware deployment.

When right
  • Customer wants validation before deciding to invest in a full POC
  • Specific feature questions to answer
  • Customer team wants familiarity, not full operational validation

Proof of Value (POV)

Definition: Combines technical validation with business outcome demonstration. May include TCO analysis, parallel-running of workloads, or measured operational benefits. Often spans longer than a POC.

When right
  • Customer needs financial validation alongside technical
  • CFO or executive sponsor needs business case evidence
  • Multiple stakeholder groups need different forms of validation

Choosing the right engagement type

SignalRecommend
Active eval, budget approved, want hands-onPOC
Early eval, want to see capability before committingEvaluation
Need financial + technical validationPOV
Just shopping; no triggerDemo, build relationship; no POC
Specific feature questionDemo or evaluation
Default rule

When in doubt, recommend an evaluation first; upgrade to POC if customer wants deeper validation. Skip the POC entirely if the customer profile suggests it won't produce a decision.

2POC Qualification Checklist

Before committing to a POC, verify each of these. If 6+ items check, proceed with POC scoping. If fewer, recommend a different engagement type or hold the conversation until conditions improve.

Eight-item qualification checklist
  • Trigger event identified. Refresh, contract end, performance issue, or strategic initiative driving urgency.
  • Budget approved or in approval process. Real budget for the next 12 months covering the in-scope opportunity.
  • Decision authority engaged. The person who will sign the contract is involved and aware of the POC.
  • Success criteria can be agreed. Customer is willing to define what success looks like in writing.
  • Timeline is realistic. 4-6 weeks for typical POC; the customer's timeline accommodates this.
  • Customer team has bandwidth. Resources allocated to participate in the POC, not just observe.
  • Specific technical questions to answer. Generic "evaluate Nutanix" is not specific enough.
  • Disqualifying conditions absent. Customer hasn't already chosen another platform; recent renewal isn't blocking; political dynamics aren't fatal.

3POC Scoping Document

Every POC starts with a written scoping document agreed by both BlueAlly and customer. Both parties sign. This is the contract that bounds the POC.

POC SCOPING DOCUMENT

Customer: <customer name>
BlueAlly SA: <SA name>
Customer Sponsor: <name and role>
Customer Technical Lead: <name and role>
POC Start Date: <date>
POC End Date: <date>
Final Report Due: <date>

OBJECTIVES
1. <specific question to answer>
2. <specific question to answer>
3. <specific question to answer>

SUCCESS CRITERIA
For each objective, define what "success" means:
  - Objective 1: <measurable success definition>
  - Objective 2: <measurable success definition>
  - Objective 3: <measurable success definition>

SCOPE INCLUDED
  - Hardware: <specifics>
  - Software stack: <NCI tier (Pro / Ultimate; formerly AOS Pro / Ultimate),
                     NCM tier (Starter / Pro / Ultimate),
                     Security Add-On for NCI Pro if Flow microsegmentation in scope,
                     Files / Objects / Volumes if in scope>
  - Workloads to test: <specific list>
  - Integrations to validate: <specific list>

SCOPE EXCLUDED
  - <items explicitly not in scope to prevent scope creep>

DELIVERABLES
  - POC environment provisioned and validated
  - Documented test results against success criteria
  - Final POC report with recommendation

CUSTOMER COMMITMENTS
  - Hardware power, cooling, rack space, network drops
  - Customer team time: estimated <hours>/week
  - Test workload data: <specific items>

BLUEALLY COMMITMENTS
  - POC hardware loaner (if applicable)
  - SA on-site or remote support: <hours>/week
  - Final report within <X> days of POC end

DECISION TIMELINE POST-POC
  - Review meeting: <date>
  - Decision target: <date>

4POC Hardware Sourcing

1 · BlueAlly lab POC

POC runs in BlueAlly's lab on BlueAlly hardware. Customer accesses remotely.

Pros
  • No customer hardware logistics
  • Faster to start
  • Consistent environment
Cons
  • Customer can't do real workload migration
  • Limited operational validation
  • Less applicable to specific customer environment
Right for

Evaluations, feature validations, customer team familiarization.

2 · Loaner hardware at customer site

Nutanix-branded NX or OEM partner (Dell, HPE, Lenovo, Cisco) loaner hardware shipped to customer site for the POC duration.

Pros
  • Real customer environment
  • Real workload migration possible
  • Full operational validation
Cons
  • Hardware logistics (rack space, power, cooling, network)
  • Coordination with customer's facilities and network teams
  • Loaner availability constraints
Right for

Full POCs producing buying decisions; customers who need to validate in their environment.

3 · NC2 cloud POC

Spin up a Nutanix cluster in AWS or Azure via NC2 for the POC duration.

Pros
  • Fast deployment (hours, not weeks)
  • No on-site logistics
  • Customer pays cloud bare-metal pricing or BlueAlly absorbs as POC cost
Cons
  • Cloud bare-metal POC costs money
  • Not all customers want their workloads in cloud during POC
  • Some integrations (specific on-prem systems) may not test cleanly from cloud
Right for

Customers evaluating NC2 specifically; customers without lab infrastructure; rapid POCs.

Hardware sizing for POCs

Right-size for the test plan, not the full production deployment:

POCs that size too small don't validate at relevant scale; POCs that size too large waste loaner inventory and add complexity.

5Test Plan Construction

The test plan operationalizes the success criteria.

TEST PLAN

TEST 1: <NAME OF VALIDATION>
  Maps to Objective: <1, 2, or 3>
  Description: <what we're testing>
  Procedure: <step-by-step>
  Success Criteria: <measurable result that defines pass>
  Customer Participation: <what the customer team does>
  Estimated Time: <hours or days>

TEST 2: ...
TEST 3: ...
Tests are not feature demos

A test demonstrates that a specific success criterion is met. "Show snapshots" is a feature demo. "Take an application-consistent snapshot of the customer's SQL Server, restore from snapshot, verify application functionality post-restore" is a test.

A typical POC includes 8-15 tests across:

6Ten Demo Scripts

Starting points. Adapt to the specific customer environment and tools.

Demo

01Cluster Deployment (Foundation)

Objective

Validate the deployment experience.

Procedure

  1. Customer team observes (or participates in) Foundation-based cluster bootstrap on the loaner hardware
  2. Cluster comes up; CVMs initialize; storage pool forms
  3. Validate cluster status: cluster status from any CVM should show all services UP
  4. Run NCC: ncc health_checks run_all; capture output

Success criteria

Cluster bootstrapped within expected timeframe (Foundation typically completes in 2-4 hours for typical clusters); all services healthy; NCC clean.

Customer participation

Customer team watches Foundation run; sees the level of automation; experiences first cluster log-in via Prism.

Demo

02Live VM Migration (vMotion-Equivalent)

Objective

Validate Live Migration works as expected.

Procedure

  1. Create a test VM with a sustained workload (e.g., a database with active queries, or a video stream)
  2. Note the host the VM is running on
  3. Trigger Live Migration to a different host via Prism
  4. Observe: VM continues running; brief network blip during switchover; sustained workload continues uninterrupted
  5. Validate the VM is now on the new host

Success criteria

Live migration completes; workload continues uninterrupted; no data loss; switchover time matches customer expectations.

Customer participation

Customer team selects a VM, triggers the migration, observes the result.

Demo

03Snapshot and Restore

Objective

Validate snapshot operations.

Procedure

  1. Take a snapshot of a test VM (Prism: VMs > select VM > Take Snapshot; or acli vm.snapshot_create)
  2. Modify data on the VM (write some test data)
  3. Restore the snapshot
  4. Verify data state matches the snapshot point

Success criteria

Snapshot completes within expected time (typically seconds for crash-consistent); restore returns VM to snapshot state; no data loss; customer's application functions post-restore.

Customer participation

Customer team takes the snapshot, modifies data, performs the restore.

Demo

04Replication and Failover

Objective

Validate replication to DR (cluster-to-cluster, or cluster-to-NC2).

Procedure

  1. Configure Async replication between primary cluster and DR cluster (or NC2)
  2. Create a Protection Policy with a test VM as a member
  3. Initial replication completes; subsequent replications run on schedule
  4. Trigger a test failover via Recovery Plans (into isolated network)
  5. Verify VM starts at DR site; application reachable
  6. Tear down test failover; verify primary cluster unchanged

Success criteria

Replication completes within RPO target; test failover succeeds without affecting primary; VM operational at DR site.

Customer participation

Customer team configures the protection policy, observes replication, triggers and validates the test failover.

Demo

05Failure Tolerance

Objective

Validate the cluster handles a node failure.

Procedure

  1. Note the cluster's current state via ncli cluster get-domain-fault-tolerance-status type=node (the type= parameter is required in current AOS)
  2. Simulate a node loss (in a POC environment, this might be done by powering off a node)
  3. Verify VMs on the failed node restart on surviving nodes (HA event)
  4. Observe rebuild starting (Curator reconstructing the failed node's data on remaining nodes)
  5. Power the node back on; verify it rejoins the cluster
  6. Verify rebuild completes; cluster returns to full redundancy

Success criteria

VMs restart on surviving nodes within expected time (typically 60-90 seconds); rebuild progresses; cluster recovers full redundancy after rebuild.

Customer participation

Customer team triggers the simulated failure (with appropriate caution), observes the recovery.

Demo

06Workload Migration with Move

Objective

Validate workload migration from existing VMware environment to AHV.

Procedure

  1. Deploy Move VM in customer environment (or use cluster-shipped Move)
  2. Add source environment (vCenter) and target environment (AHV)
  3. Create a migration plan for 1-3 test VMs (representative of customer workloads)
  4. Execute initial replication; verify progress
  5. Schedule cutover; verify VM is shut down at source, started at target
  6. Validate VM operation at target; customer's application accessible

Success criteria

Move executes initial replication and cutover within expected time; VM is operational at target; data integrity preserved; customer's application validates.

Customer participation

Customer team selects the migration candidates, executes the plan, validates the result.

Demo

07Microsegmentation with Flow

Objective

Validate Flow Network Security policy enforcement.

Procedure

  1. Create a category structure (e.g., AppTier: Web, AppTier: App, AppTier: DB)
  2. Tag test VMs with appropriate categories
  3. Create a Flow policy: Web can talk to App; App can talk to DB; Web cannot talk to DB directly
  4. Validate from a Web VM: traffic to App VM succeeds; traffic to DB VM fails
  5. Modify the policy to allow Web-to-DB; revalidate

Success criteria

Flow policies enforce as configured; traffic flows match the policy; policy changes take effect quickly.

Customer participation

Customer team configures the categories and policies, performs the connectivity tests.

Demo

08Files Performance and Operations

Objective

Validate Nutanix Files for the customer's file workloads.

Procedure

  1. Deploy Files cluster with appropriate sizing
  2. Create a File Server; configure a SMB share with AD integration
  3. Mount the share from a customer-owned client (Windows or Mac)
  4. Perform representative file operations (large copy, many-small-files copy, random access)
  5. Take a snapshot of the share; verify Self-Service Restore works (Previous Versions tab in Windows)
  6. Configure replication to a second Files cluster

Success criteria

Files performance matches customer expectations for the workload pattern; AD integration works correctly; SSR works as expected; replication completes.

Customer participation

Customer team configures Files, runs the test workloads, validates results.

Demo

09Objects as Backup Target

Objective

Validate Objects as backup target for customer's backup software.

Procedure

  1. Deploy Objects cluster
  2. Create an Object Store; create a bucket for backup data
  3. Configure customer's backup software (Veeam, Commvault, Rubrik, etc.) to use the bucket as a backup target
  4. Run a test backup
  5. Restore from the backup; verify data integrity
  6. Optionally: configure WORM on the bucket and verify immutability

Success criteria

Backup completes successfully; restore validates data integrity; backup software treats Objects as functionally equivalent to S3 or other supported targets.

Customer participation

Customer's backup admin configures the backup software, runs the test backup and restore.

Demo

10Operational Workflows

Objective

Validate day-2 operations the customer team will perform.

Procedure

Walk the customer team through:

  1. Adding a new VM via Prism
  2. Resizing a VM (add CPU, RAM, disk)
  3. Cloning a VM
  4. Reviewing performance metrics in Prism Element and Prism Central
  5. Capacity reporting (Intelligent Operations in NCM Pro)
  6. Triggering an LCM upgrade dry-run
  7. Reviewing audit logs

Success criteria

Customer team can perform each operation without difficulty; the operations match (or improve on) the customer's current workflow.

Customer participation

Customer team performs the operations themselves; SA observes and provides guidance.

7The 30-Day POC Cadence

Week 1 · Setup and Familiarization
  • Day 1-2: Hardware install, network configuration, cluster bootstrap (Foundation)
  • Day 3-4: Customer team onboarding to Prism; tour of the platform; initial demos
  • Day 5: Test plan walkthrough; refinement of success criteria; agreement on week-by-week milestones
Week 2 · Core Validations
  • Live migration, snapshots, basic operational workflows
  • First workload migration via Move (small test VM)
  • Initial performance baseline establishment
Week 3 · Advanced Validations
  • Failure tolerance testing
  • Replication and DR validation
  • Files / Objects / Volumes specific testing (whichever are in scope)
  • Integration testing (backup software, identity, ServiceNow if applicable)
Week 4 · Operational Validation and Closeout
  • Customer team performs operational tasks independently
  • Review of test results against success criteria
  • Final report drafted and reviewed with customer
  • Decision meeting scheduled
Adjustments

For 6-week POCs: add a week between Week 2 and Week 3 for additional workload migration; add a week at the end for extended operational validation.

For shorter POCs (2-3 weeks): trim Week 3 advanced validations to the customer's specific priorities; Week 4 is closeout.

8Success Criteria Templates

Common patterns. Customize per customer.

Performance

Availability

DR

Operational

Integration

The form

Each criterion is specific, measurable, and binary (passes or fails). Vague criteria like "performance is good" produce arguments at the end of the POC; specific criteria produce decisions.

9The Final POC Report

The report converts POC results into a buying decision document.

POC FINAL REPORT

CUSTOMER: <name>
POC DATES: <start> to <end>
REPORT DATE: <date>
PREPARED BY: <BlueAlly SA name>
REVIEWED WITH CUSTOMER: <date and attendees>

EXECUTIVE SUMMARY
<3-5 sentence summary of POC scope, key findings, and recommendation>

POC OBJECTIVES (from scoping document)
1. <objective 1>: PASSED / PASSED-WITH-CAVEATS / NOT-MET
2. <objective 2>: PASSED / PASSED-WITH-CAVEATS / NOT-MET
3. <objective 3>: PASSED / PASSED-WITH-CAVEATS / NOT-MET

DETAILED FINDINGS
For each objective:
  - What was tested
  - Results (with measurements, screenshots, or output)
  - Comparison to success criteria
  - Any caveats or context

PERFORMANCE OBSERVATIONS
  - Key performance metrics from POC
  - How they compare to customer's current baseline
  - Implications for sizing the production deployment

OPERATIONAL OBSERVATIONS
  - Customer team feedback on platform usability
  - Items that worked well
  - Items that required adjustment or learning

ITEMS NOT VALIDATED IN POC
  - Capabilities or scenarios out of POC scope
  - Recommended next-step validations if applicable

RECOMMENDED NEXT STEPS
  - If POC met objectives: recommended path forward (proposal, sizing, contract)
  - If POC raised concerns: recommended additional validation or alternative paths
  - If POC did not validate Nutanix as the right fit: honest assessment

APPENDICES
  - Raw test results
  - Configuration documentation
  - Photos / screenshots
  - POC team list (BlueAlly + customer)
Audience

The report is written for the buying decision meeting. The customer's executive sponsor reads the executive summary; the technical team reads the detailed findings; the CFO reads the performance and operational observations relevant to TCO.

10Common POC Pitfalls

10 pitfalls to avoid
  1. Skipping the scoping document. POCs without written success criteria drift; the customer remembers different objectives than the SA; results are arguable.
  2. Letting scope creep. Customer asks for "one more test"; SA agrees; week 4 becomes week 8; POC loses momentum. The scoping document is the boundary; in-scope tests run, out-of-scope tests are documented for follow-up.
  3. SA-driven, not customer-driven. SA performs all the tests; customer observes. The customer learns the platform but doesn't develop operational confidence. Wrong outcome. The customer needs hands-on experience to commit.
  4. Testing the platform's strengths, not the customer's needs. SA naturally gravitates to demos that highlight Nutanix's best features. Customer's actual needs may not align. Test plan should target customer needs first.
  5. No failure scenarios. Customers who haven't seen failure recovery in POC discover it in production. Always include node loss, network partition, and similar tests.
  6. Generic workloads, not customer workloads. Synthetic benchmarks don't validate the customer's actual workload. Use customer's data and applications where possible (with appropriate copies, not production).
  7. No final report. POC ends with verbal "looks good"; weeks pass; the buying decision stalls. The written report is the artifact that converts results into decision.
  8. Pushing for decision before report is reviewed. Asking "ready to buy?" right after the POC ends, before the written report is reviewed, comes across as pressure-selling.
  9. Underestimating customer team time. POCs require customer team participation; assuming they have unlimited time produces frustrated stakeholders. Realistic time estimates in the scoping document help both sides plan.
  10. Treating the POC as the sale. The POC validates the platform; the proposal, contract, and decision happen after. Don't confuse them.

11When to Recommend Against a POC

Don't run a POC when

The senior-SA disposition: not every customer should run a POC. Recommend against when:

  • Customer has no specific technical questions (a POC won't surface questions they haven't articulated)
  • Decision authority is not engaged (POC results won't convert to decision without sponsor)
  • Customer's profile genuinely doesn't fit Nutanix (don't run a POC to confirm what should be obvious)
  • Recent platform investment makes a buying decision unlikely in the timeframe
  • Political dynamics make a fair evaluation impossible (incumbent vendor has executive lock; the POC will be evaluated against rigged criteria)
The credibility move

Tell the customer when a POC isn't the right next step. Recommend an alternative (evaluation, demo, peer reference call, design workshop). Customers respect SAs who decline POCs that won't produce decisions.

12POC Variations by Customer Profile

Mid-market POC

Enterprise POC

Compliance-heavy POC

Cloud DR POC (NC2-focused)

VDI POC

A Final Word

The POC is a high-leverage moment in the sales cycle. Customers commit to vendors who run professional, customer-centered POCs. Customers walk away from vendors who treat POCs as feature shows.

The discipline matters more than the specific demos:

  • Qualified scope, not over-broad
  • Customer-defined success criteria, not vendor-imposed
  • Customer hands on the keys, not just observation
  • Honest results, including caveats and items that didn't validate
  • Written final report, not verbal "looks good"
  • Recommendation that respects what the POC actually showed, not what BlueAlly hoped to show

Run POCs that produce decisions. The customers who say yes after a professionally-run POC become 10-year BlueAlly customers.

References

POC engagement methodology (qualification, scoping, demo scripts, success criteria, final-report structure). Technical content underlying the demos is sourced and verified in the per-module References sections plus Appendix G.

  • Module 02 References. Foundation deployment, NCC commands used in Demo 1. Module 2
  • Module 03 References. Live Migration semantics, Move tool used in Demo 2 and Demo 6. Module 3
  • Module 06 References. Flow Network Security used in Demo 7 (NCI Ultimate or Security Add-On for NCI Pro). Module 6
  • Module 07 References. Recovery Plans and replication used in Demo 4. Module 7
  • Module 08 References. Files and Objects used in Demos 8 and 9. Module 8
  • Module 09 References. NCI / NCM / NCP tier structure for the POC scoping document. Module 9
  • Appendix G References. CLI command syntax (ncli, acli, ncc) used across demo procedures, including the corrected ncli cluster get-domain-fault-tolerance-status type=node syntax. Appendix G

Cross-references

  • Modules: Each demo script links to the module where the underlying capability is taught.
  • Glossary: Appendix A defines the terms used.
  • CLI Reference: Appendix G has the commands referenced in demo scripts.
  • Discovery Questions: Appendix E has the questions that surface POC scope.
  • Sizing Rules: Appendix F for POC hardware sizing.
  • Reference Architectures: Appendix I for the production architectures POCs validate against.
  • Objections: Appendix D has responses to objections that often surface during POCs.
← back to home