Chapter 01 Β· Problem & Market Validation

The Problem Hunch

"I've been reorganizing the same pile of PDFs for three years. At some point you stop thinking it's just you." β€” Alex's journal, Day 1

Alex had 47 browser tabs open. Three monitors. Sticky notes covering her secondary screen like a patchwork quilt of academic anxiety. Somewhere in this chaos lived the paper she needed, the one about cross-modal attention and working memory that she'd highlighted and annotated and definitely-read-three-months-ago. She'd just spent two hours looking for it.

She found it, finally, inside a folder named "Papers_FINAL_v3_real_ACTUAL." She opened it. The highlighting was there. The annotations were good. Past-Alex had done excellent work. Present-Alex had zero memory of any of it.

She opened a blank notebook and wrote: "What if there was a tool that just... remembered?"

She stared at that sentence for a long time. Then, below it, she wrote a second sentence that would change the next two years of her life: "Is this a me problem or an everyone problem?"

"The best startup ideas feel like secrets, things you know that most people don't yet believe. My secret was scribbled in the margin of a soggy notebook at midnight."
πŸ”

What Is Problem Validation?

Problem validation is the process of confirming, before significant investment, that a real, painful problem exists for a sizable group of people willing to pay for a solution. Most startups fail not because they build poorly, but because they build something nobody wants.

There are two related concepts founders often conflate:

  • Problem Validation: Confirming the problem is real, frequent, and painful enough to motivate action.
  • Market Validation: Confirming that enough people with this problem exist and are reachable, i.e., a viable market.

The key question experienced founders ask isn't "would you use this?" Instead, it's "how do you handle this problem today?" Past behavior reveals real pain far better than hypothetical intent. Customers can't accurately predict future behavior, but they can describe what they actually did last week.

Alex didn't know any of this yet. She just knew she had a hunch, and she wanted to know if it was shared.

Alex googled "research paper management software." The results were extensive: Zotero, Mendeley, Papers, ReadCube, Paperpile, Citavi. She spent 20 minutes with each. Zotero felt like software that had been lovingly maintained since 2009: functional, beloved by academics, utterly without visual ambition. Mendeley crashed twice. Papers was genuinely beautiful, but its organization model didn't match how she actually thought about research. It wanted her to file papers; she wanted to connect them.

She opened G2 and Capterra. She filtered by "research management software" and sorted by lowest rating. The one-star reviews were illuminating. Three different users, different tools, almost the same complaint: "Doesn't handle cross-paper connections." One wrote: "I can find individual papers, but I can't see how they relate to each other. This is literally the only thing I need."

Alex felt a jolt, the specific, electric kind that comes from finding your hypothesis in the wild. She wrote in her notebook: "Competitors exist = demand exists. But they're leaving something on the table."

"Competitors aren't the enemy. They're proof the market is real."
Framework Β· Market Sizing

TAM / SAM / SOM: How Big Is This?

TAM $47B globally
SAM $3.2B
SOM $120M
TAM β€” Total Addressable Market: $47B 180M knowledge workers globally who manage research, documents, or information at work. Everyone who could benefit from the solution.
SAM β€” Serviceable Addressable Market: $3.2B 15M academic researchers reachable via university channels, academic platforms, and research communities. The portion Lumio can realistically reach.
SOM β€” Serviceable Obtainable Market: $120M 500K STEM PhD students writing literature reviews, the wedge Alex can realistically capture in Year 1–2. Bottom-up: 500K users Γ— $20/month Γ— 12 months.

Always size using both approaches: top-down (industry reports β†’ narrow to segment) and bottom-up (# of customers Γ— average revenue per user). If both numbers converge, you have a credible estimate.

Alex had read about smoke tests in a YC blog post. The idea was almost offensively simple: build a landing page for a product that doesn't exist, run ads at it, and measure whether people want it enough to give you their email address. She told herself she'd try it over the weekend. It took her until 11pm on Friday to actually start.

By Saturday morning, she had a Carrd page. Headline: "Lumio β€” Never lose a research insight again." Three bullet points. An email signup form. She spent $50 on Google Ads targeting search terms like "research paper management" and "academic literature tool." She set the campaign live at 3pm and refreshed the analytics dashboard approximately every seven minutes for the rest of the day.

Sunday evening results: 847 impressions, 63 clicks (7.4% CTR), 19 email signups. That was a 30% email capture rate. She read every one of the 19 emails. The third one said: "I have been waiting for something like this for literally years."

Alex screamed into a pillow. Then opened her notebook and wrote, in capital letters: IT'S REAL.

"Don't build it until someone's willing to give you their email address for something that doesn't exist yet."
πŸ’¨

The Smoke Test / Fake Door Test

A smoke test (also called a fake door test) is a lightweight experiment: you launch a landing page or ad describing a product that doesn't yet exist and measure real interest before building anything.

How to set one up:

  • Build a simple landing page (Carrd, Webflow, Unbounce) describing the core value proposition
  • Drive traffic via a small paid ad budget ($50–$200) or organic posts in relevant communities
  • Measure: click-through rate, email capture rate, and any messages you receive

What counts as a signal: An email capture rate above 5% is considered a strong demand signal. Alex hit 30%, which is exceptional. A rate below 1–2% suggests either a messaging problem or a demand problem.

What it doesn't prove: Willingness to pay. "Sounds cool" β‰  credit card. The smoke test validates interest; pre-sales validate intent to pay. Both matter.

Framework Β· Validation

6 Signals That Validate Real Market Demand

These are the signals that separate genuine market pull from polite interest. Alex's score after her first month: 5 out of 6.

βœ“
Customers describe the problem in detail without prompting
Alex's smoke test emails included unprompted paragraphs about paper management frustration.
βœ“
Existing workarounds (spreadsheets, manual processes, duct-tape solutions)
Competitor reviews revealed users building elaborate folder structures and Google Docs systems.
–
Willingness to pay upfront or join a waitlist
Not yet tested. The smoke test captured emails but didn't include a price or pre-sale option.
βœ“
High engagement on a landing page (>5% email capture rate)
30% email capture. Far above the 5% threshold. This is a very strong signal.
βœ“
Active communities, forums, or support groups around the pain point
r/GradSchool, r/PhD, academic Twitter, all full of paper management complaints.
βœ“
Competitors exist (proves demand) but leave a segment underserved
Zotero, Mendeley, Papers, all exist and all miss cross-paper connection and context restoration.
✏️

Write Your Problem Hypothesis

Use Alex's process. Fill in the template below. This is your first validation artifact. Be specific. Vague hypotheses lead to vague findings.

I believe                      [target customer] struggles with                      [problem] when                      [context].
Competitor reviews? Your own experience? Community complaints? Social media?
Set your decision rule BEFORE you run the test. What result would convince you to proceed?
Alex's hypothesis: "I believe PhD students and postdocs struggle with re-establishing context for papers they've already read when returning to them weeks later. I know this is real because 3 competitor reviews mention it explicitly. My threshold: if 25% of landing page visitors sign up, I proceed to interviews."

Chapter 1 Takeaways

1

Validate before building. Most startups fail not because they build badly, but because they build something nobody wants. 15–20 problem interviews and a smoke test can save you months of wasted work.

2

Competitors are your friends. Their existence proves demand. Read their negative reviews to find the exact gap they're leaving open. That gap is your opportunity.

3

Set your threshold before you start. Decide in advance what evidence would convince you to proceed or pivot. "Sounds cool" β‰  willingness to pay. "I've been waiting for this for years" + 30% email capture = something real.