TL;DR: A good POC validates one critical assumption in 2-6 weeks for $10K-$30K. A bad POC is a mini-MVP that takes 3 months, costs $80K, and still doesn't answer the question you actually needed answered. The difference is clarity of scope: know exactly what you're testing before you write a line of code. Approximately 66% of software projects that skip POC validation fail to deliver expected outcomes.
Why POCs exist (and why most are done wrong)
Here's the scenario: your company is considering a $300K software build. The CTO has concerns about whether the AI classification engine will work with your data quality. The CFO has concerns about whether the integration with the legacy ERP can be done without replacing it. The product team isn't sure users will adopt the new workflow.
A POC is supposed to resolve those concerns before you commit the full budget. Think of it as spending $15K to potentially save $300K.
But most POCs go wrong because they try to answer too many questions at once. Instead of validating one critical assumption, they become compressed development projects — with all the complexity of a real build, a fraction of the timeline, and none of the production infrastructure.
What a POC should (and shouldn't) answer
A POC should answer:
- "Can this technology handle our data?" — Test whether the AI model, the database engine, or the integration protocol works with your actual data volumes and formats.
- "Is this integration feasible?" — Prove that your legacy system's API can exchange data with the new platform reliably enough to build on.
- "Will this approach perform at scale?" — Demonstrate that the proposed architecture handles 10x your current load without falling over.
- "Is the core user workflow viable?" — Show that the critical path (the 2-3 screens or processes that define the product) is technically buildable and reasonably intuitive.
A POC should NOT answer:
- "What should the product look like?" (That's UX research.)
- "What features should we build?" (That's product strategy.)
- "How much will the full product cost?" (That's discovery/estimation.)
- "Will users like it?" (That's beta testing.)
If your POC is trying to answer questions from both columns, it's not a POC. It's an underfunded MVP.
How to structure a POC with an agency
Step 1: Define the hypothesis (week 0)
Before you engage any agency, write down the single question the POC must answer. Not three questions. One.
Good hypotheses: - "Our classification model can process 10,000 documents per hour with >85% accuracy using our actual data." - "The SAP integration can sync inventory data in near-real-time without requiring middleware." - "Users can complete the loan application workflow in under 4 minutes."
Bad hypotheses: - "The product can work." (Too vague — what does "work" mean?) - "Users will love it." (Emotional, not measurable.) - "It's technically feasible." (Feasible at what cost, timeline, and quality level?)
Step 2: Scope ruthlessly (week 0-1)
A POC scope document should fit on one page. If it requires more, you're building too much.
Include: - The hypothesis being tested - The technical components being validated - Success criteria with specific metrics - Data or systems needed from the client - Timeline (2-6 weeks) - What's explicitly out of scope
A real example:
Hypothesis: Our NLP model can classify support tickets into 12 categories with >90% accuracy on real ticket data.
Scope: Train the model on 5,000 historical tickets (provided by client). Deploy to a test environment. Run classification against 500 new tickets. Measure accuracy per category.
Success criteria: Average accuracy >90% across all categories. No single category below 80%. Processing time <2 seconds per ticket.
Out of scope: Production deployment, UI, API integration, user testing, model retraining pipelines.
Timeline: 3 weeks Budget: $12K
Step 3: Agree on deliverables (week 1)
A POC deliverable isn't a product. It's evidence. Specifically:
- Technical report documenting findings, metrics, and recommendations
- Working prototype that demonstrates the tested capability (not polished, not production-ready)
- Decision recommendation — proceed, modify approach, or stop
- Architecture implications — what the POC results mean for the full build (technology choices, risk areas, revised estimates)
Make sure the agency understands that the POC can result in a "don't build this" recommendation. If the agency's incentive is to always recommend proceeding (because the full build is their next revenue opportunity), the POC loses its value as a validation tool.
Step 4: Provide real data (the part most clients skip)
The number one reason POCs produce misleading results: they're tested with sample data instead of real data.
Sample data is clean. Real data has missing fields, inconsistent formats, edge cases, and volumes that sample data never captures. An AI model that performs beautifully on 200 sample records might collapse when fed 20,000 real records with messy inputs.
What to prepare: - Representative production data (anonymized if necessary) - Access to the systems being integrated (even read-only access to a staging environment) - Realistic traffic volumes or transaction patterns - Documentation of data quirks that might affect results
Step 5: Run the POC (weeks 2-5)
The agency builds and tests. Your involvement during this phase should be limited but specific:
- Kickoff meeting to confirm scope, data access, and success criteria
- Weekly check-in (30 minutes) to review progress and unblock issues
- Final review to evaluate results against success criteria
Resist the temptation to expand scope during execution. "While we're at it, can we also test..." is how a 3-week POC becomes a 10-week project. If new questions arise during the POC, write them down for a potential follow-on POC or discovery phase.
Step 6: Evaluate results honestly (week 5-6)
There are three possible outcomes:
Green: The hypothesis is validated. The technology works, the integration is feasible, the performance meets requirements. Proceed to discovery and full development with increased confidence.
Yellow: Partial validation. The approach works but with caveats — lower performance than expected, additional components needed, higher complexity than assumed. The full build is still viable but needs a modified approach, revised timeline, or increased budget.
Red: The hypothesis is invalidated. The technology doesn't work with your data, the integration is impractical, or the performance is fundamentally insufficient. This is actually the most valuable POC outcome. You just saved $300K by spending $15K to discover the approach doesn't work before building it.
POC budget guidelines
| Project Size | POC Budget | POC Duration |
|---|---|---|
| $50K-$100K build | $5K-$10K | 2-3 weeks |
| $100K-$300K build | $10K-$20K | 3-4 weeks |
| $300K-$500K build | $15K-$30K | 4-6 weeks |
| $500K+ build | $25K-$50K | 4-8 weeks |
A POC should cost 3-7% of the total expected project budget. If an agency proposes a POC that's 15-20% of the full build cost, they're building a mini-MVP, not validating a hypothesis.
Mistakes that waste POC budgets
Building a UI for a technical validation
If the question is "can the backend process handle 50,000 concurrent connections?", you don't need a user interface. Build a test harness. Save the UI for the design phase.
Testing with the wrong team
Some agencies assign junior developers to POCs (small budget = small team). But POCs require senior engineers — people who can make sound architecture decisions, identify issues quickly, and produce insights that go beyond "it works/it doesn't."
Not defining success criteria upfront
"We'll know it when we see it" isn't a success criterion. If you can't measure whether the POC succeeded, you can't make an informed go/no-go decision. Define specific, measurable benchmarks before work begins.
Letting the POC scope creep into MVP territory
A POC validates one assumption. An MVP validates market demand. They're different exercises with different scopes and budgets. When stakeholders start adding "and could we also add user authentication and a dashboard?" — you've crossed from POC into product development.
After the POC: what happens next
If the POC validates your hypothesis, the next step is typically a discovery phase (4-8 weeks), not jumping straight into development. The POC proved the concept works. Discovery defines how to build the full product: architecture, scope, timeline, and budget.
Think of the sequence as: POC → Discovery → Development.
Each step reduces risk before the next, larger investment. The POC costs $15K to validate feasibility. Discovery costs $30K-$50K to define the product. Development costs $200K-$500K to build it. At each gate, you have the information to decide whether the next investment makes sense.
How Globalbit runs POCs
We start every POC engagement with a one-page scope document that both sides sign off on. The hypothesis, success criteria, timeline, and budget are locked before any development begins. We assign senior engineers to POCs — the same people who would architect the full build — because POC quality depends on experience, not volume of code. Our POC reports include not just results but architecture recommendations for the full build, so the discovery phase starts with real technical evidence rather than assumptions. And if the POC concludes that the project shouldn't proceed as planned, we say so. Clearly. Tell us about your POC requirements.

