validation
How to Evaluate a Startup Idea: The Framework That Surfaces What You're Missing

The goal isn't to prove you're right. It's to discover where you might be wrong.
42%
of startups fail because they built
something nobody wanted
Source: CB Insights
The Question Most Founders Get Wrong
When founders ask "how do I evaluate my startup idea?", they're usually asking the wrong question.
What they mean is: "How do I confirm this idea is good?"
What they should be asking is: "What would make this idea fail, and can I verify those failure modes don't apply?"
This isn't semantic pedantry. It's the difference between validation (seeking confirmation) and verification (seeking truth). Validation feels good and produces false confidence. Verification is uncomfortable and produces real information.
CB Insights analyzed 111 startup failures and found that 42% died from "no market need." These founders thought they were evaluating their ideas. They were actually confirming their hopes. The market didn't care about their confirmation.
Here's how to actually evaluate a startup idea — including what the evaluation process is designed to surface, and why your instincts will fight you every step of the way.
How It Works
Submit Your Idea
Describe your startup concept in plain language.
AI Research Begins
14 research streams analyze your idea simultaneously.
Kill Vector Scan
Systematic search for the failure modes most likely to destroy this business.
Scoring & Verdict
Quantified assessment across market, competition, economics, and execution.
Receive Your Report
Institutional-grade analysis delivered in minutes, not weeks.
The Mindset Shift: From Advocate to Auditor
Before any framework, you need a mindset shift.
When you conceived your idea, you became its advocate. Every conversation since then has been building the case for why it works. Your brain has been collecting supporting evidence and filtering out contradictions.
Daniel Kahneman called this "theory-induced blindness" — once you have a theory, you stop seeing evidence against it. In Thinking, Fast and Slow, he wrote: "The confidence that individuals have in their beliefs depends mostly on the quality of the story they can tell about what they see."
Your story is compelling. To you.
Evaluation requires temporarily suspending advocacy. Not forever — you'll need to be an advocate again if you proceed. But for the evaluation process to work, you have to become an auditor: someone looking for discrepancies, not someone building a case.
The test: Can you articulate three specific ways this idea might fail? Not vague risks ("maybe the market is small") but concrete failure modes ("the TAM calculation assumes 100% of homeowners consider digital solutions, but historical adoption for similar tools is 3-7%").
If you can't articulate specific failure modes, you haven't evaluated the idea. You've defended it.
Dimension 1: Problem Severity
Not all problems are worth solving. Evaluation starts by assessing whether the problem is severe enough to drive action.
The Core Question
Is this a painkiller or a vitamin?
Painkillers solve urgent problems people will pay to eliminate. Vitamins are nice-to-have improvements that get deprioritized when budgets tighten.
The Tests
The "hair on fire" test: Is the problem urgent enough that people are actively looking for solutions? Or is it something they'd appreciate solving but won't prioritize?
The "current spend" test: What are people paying now to address this problem? Existing spending is demand evidence. No current spending is a warning sign — either the problem isn't severe enough or your framing misses the actual pain.
The "frequency" test: How often do people experience this problem? Daily problems create different urgency than annual problems. A problem that occurs once a year better be extremely painful.
What This Reveals
High-severity problems generate pull — customers seek you out. Low-severity problems require push — you convince customers they should care. Push-based businesses can work, but they're harder and more expensive.
The psychology: Founders often fall in love with clever solutions to minor problems. The solution is elegant; the problem just isn't severe enough to drive adoption. Evaluation forces confrontation with problem severity before you invest in solution elegance.
Dimension 2: Market Size
Market evaluation isn't about finding the biggest number. It's about finding realistic numbers and understanding what they imply.
The Core Question
Is the market large enough to support the kind of business you want to build?
A $20M market can support a lifestyle business. It can't support venture-scale returns. Neither is wrong — but confusing them leads to misaligned strategies.
The Tests
The bottom-up test: Calculate your market from the ground up. Number of potential customers x realistic price x realistic capture rate. This is harder than top-down ("the market is $50B") but more accurate.
The segment specificity test: Can you describe your target customer precisely? "Small businesses" is a demographic. "Restaurants with 20-50 employees that currently use paper scheduling" is a segment. Segments can be counted; demographics can't.
The "access" test: Not just whether the market exists — whether you can reach it. A large market you can't access is worse than a small market you can dominate.
What This Reveals
Market evaluation often reveals that the "big market" is actually many smaller markets, some accessible and some not. The honest TAM is usually smaller than the pitch deck version — but an honest smaller number is more useful than a flattering larger one. For a deeper look at how to size your market correctly, see our guide to TAM SAM SOM.
The psychology: Anchoring bias makes the first large number you encounter sticky. Founders anchor on "$50B travel market" and insufficiently adjust to "$200M golf trip planning market" to "$20M golf trip planning for groups who'll pay for digital tools." Each adjustment is painful. All are necessary.
Dimension 3: Competitive Differentiation
Competition isn't just about who else exists. It's about why customers would choose you — including choosing to change nothing.
The Core Question
Why would customers switch from what they're doing now, and why can't competitors respond?
The Tests
The "current state" test: What is your target customer doing today? The answer is never "nothing." They're using spreadsheets, manual processes, competitors, or workarounds. You're replacing something. Name it.
The "switching cost" test: What does it cost customers to adopt your solution? Time, learning, data migration, risk, relationship disruption. Your improvement has to exceed those costs.
The "one sprint" test: If an incumbent added your core feature in one development sprint, would your differentiation survive? Feature-based differentiation often can't.
The "graveyard" test: Who tried this before? Why did they fail? If you can't answer, you haven't researched. If the answers don't meaningfully distinguish your approach, you might be the next data point. Understanding common startup kill vectors can help you identify patterns that have ended similar ventures.
What This Reveals
Most founders underestimate competition because they define competitors narrowly. The spreadsheet is a competitor. The status quo is a competitor. The "we'll deal with it later" decision is a competitor. Evaluation reveals what you're actually displacing.
The psychology: The curse of knowledge makes your differentiation obvious to you. You can't unsee why you're different. But customers start from their current state, not your positioning. What's obvious to you is invisible to them.
Dimension 4: Unit Economics
Can you make money? Not eventually. Not at scale. With realistic assumptions about what things actually cost.
The Core Question
Do the unit economics work — with assumptions that would survive independent verification?
The Tests
The "realistic CAC" test: Customer acquisition cost based on founder selling to warm networks isn't realistic. What will CAC be when you're running paid campaigns to cold audiences at scale? Usually 2-5x higher.
The "honest churn" test: What retention rate are you assuming, and where does that number come from? Optimistic retention assumptions mask unit economics problems until you've burned through your cohorts.
The "payback reality" test: How long until a customer's revenue exceeds their acquisition cost? Payback periods beyond 18 months require significant capital and patient investors.
The "scale dynamics" test: Do unit economics improve or degrade at scale? Some costs decrease (infrastructure, negotiating power). Others increase (support, complexity, competition). Know which dominates.
What This Reveals
Unit economics evaluation often reveals that the business "works" only if several optimistic assumptions hold simultaneously. Each assumption alone is plausible. All of them together is unlikely.
The psychology: The planning fallacy causes systematic underestimation of costs and overestimation of revenue. Kahneman's research showed this persists even when people have direct experience with similar projects going over budget. Your spreadsheet contains this bias even if you've "been conservative."
Dimension 5: Execution Feasibility
Can you actually build this? Not theoretically. You — with your team, skills, resources, and constraints.
The Core Question
Is there a realistic path from here to a working business, given what you actually have?
The Tests
The "skills inventory" test: What skills does this opportunity require? Which do you have? Which are missing? What's the plan to fill gaps?
The "resource reality" test: What resources — capital, time, tools, partnerships — are required? Which do you have? Which require acquisition? Are the acquisition paths realistic?
The "timeline honesty" test: Most startup timelines are 40-50% optimistic. What does a realistic timeline imply for runway, market timing, and personal sustainability?
The "Plan B" test: If the primary approach doesn't work, what's the pivot? Not having a Plan B isn't confidence — it's blindness.
What This Reveals
Execution evaluation often reveals that founders are planning to do everything — product, sales, operations, fundraising — simultaneously, at speed, without the team or resources to do any of it well. Awareness of constraints isn't defeat. It's strategy.
Dimension 6: Timing
Why now? Not "why is this a good idea" — "why is this the right moment for this idea?"
The Core Question
What's changed that makes this possible or necessary now, when it wasn't before?
The Tests
The "technology shift" test: Is there new technology that enables what wasn't possible? AI, mobile, blockchain, regulatory changes — real shifts that alter the landscape.
The "behavior shift" test: Are people doing something differently now? Behavior shifts (remote work, mobile-first, subscription acceptance) create openings that didn't exist.
The "market window" test: Is there a window that's opening or closing? Some opportunities have timing constraints — regulatory changes, platform transitions, demographic shifts.
The "predecessor autopsy" test: If this idea failed in 2019, what specifically has changed? Not "AI is better now" — specific changes that address specific failure modes.
What This Reveals
Many ideas are too early or too late. Too early: the market isn't ready, customer behavior hasn't shifted, enabling technology isn't mature. Too late: competitors are established, the window has closed, the wave has crested. Timing evaluation reveals where you are on the curve.
Dimension 7: Founder-Market Fit
Are you the right person to build this specific business?
The Core Question
What unique advantage do you have in pursuing this opportunity — and is it an advantage that matters?
The Tests
The "earned insight" test: Do you have insight into this market that others lack? Not "I read about this problem" — "I've lived this problem" or "I've served customers with this problem for years."
The "unfair advantage" test: Do you have access, relationships, or capabilities that competitors can't easily replicate? Domain expertise, customer relationships, technical skills, existing distribution.
The "why you" test: If this is a good idea, why won't someone better positioned execute it? What's your specific claim to this opportunity?
The "10 years" test: Can you see yourself working on this for 10 years? Successful companies take longer than founders expect. Passion matters because persistence does.
What This Reveals
Founder-market fit evaluation often reveals a mismatch: the idea is interesting but the founder lacks specific advantage, or the founder has advantages that point toward different opportunities. Fit isn't everything — but misfit is a headwind you'll fight constantly.
The Meta-Question: Would This Survive External Scrutiny?
After working through all seven dimensions, ask the meta-question:
If someone with no stake in the outcome investigated your answers, would they hold up?
Would your market size survive an analyst's methodology review? Would your competitive analysis hold up against a deeper landscape scan? Would your unit economics survive a CFO's scrutiny?
If the answer is "probably not," that's information. Not about whether to proceed — but about what you're risking and where your vulnerabilities live.
The founders who avoid catastrophic failures aren't smarter. They're the ones who knew what they didn't know, got rigorous analysis before commitment, and made eyes-open decisions. If you want to formalize this process, a startup idea validation framework can impose the discipline that intuition alone can't provide. You can also explore the tools available for validation to match the right instrument to each dimension.
How to Evaluate Startup Ideas FAQs
How do I know if my startup idea is good? You can't "know" — you can only assess probability across dimensions: problem severity, market size, competitive differentiation, unit economics, execution feasibility, timing, and founder-market fit. An idea that clears all dimensions is more likely to succeed, but no evaluation provides certainty.
What's the most important dimension in evaluating a startup idea? Problem severity and market size are prerequisites — without a severe problem in a large enough market, other dimensions don't matter. But unit economics is often the silent killer — ideas that seem good fail because the math never worked.
Can I evaluate my own startup idea objectively? Not reliably. Cognitive biases — optimism, confirmation bias, anchoring — operate below conscious control. External evaluation by parties with no stake in a particular outcome is required to surface what your biases hide.
What's the difference between validation and evaluation? Validation asks "is this idea good?" and tends toward confirmation. Evaluation asks "what would make this idea fail?" and specifically hunts for disconfirming evidence. Validation produces confidence. Evaluation produces information.
How long should startup idea evaluation take? Thorough evaluation requires 1-4 weeks depending on complexity — enough time to research markets, talk to customers, model economics, and investigate competition. Evaluation that takes less than a few days is probably missing critical analysis.
What should I do if my idea fails evaluation? Failed evaluation is valuable information, not wasted time. The findings often point toward pivots — different markets, different positioning, different business models — that address the failure modes. Better to discover this before building than after.
References
- Kahneman, Daniel. Thinking, Fast and Slow. Farrar, Straus and Giroux, 2011.
- CB Insights. "The Top 20 Reasons Startups Fail."
- Fitzpatrick, Rob. The Mom Test. 2013.
- Graham, Paul. "How to Get Startup Ideas." paulgraham.com, 2012.
Verve Intelligence evaluates startup ideas across all seven dimensions using adversarial AI agents whose job is to find what you're missing. Transparent reasoning, data quality scores, GO/PIVOT/NO-GO verdicts. $99. Get your evaluation →