"I've been reorganizing the same pile of PDFs for three years. At some point you stop thinking it's just you." β Alex's journal, Day 1
Alex had 47 browser tabs open. Three monitors. Sticky notes covering her secondary screen like a patchwork quilt of academic anxiety. Somewhere in this chaos lived the paper she needed, the one about cross-modal attention and working memory that she'd highlighted and annotated and definitely-read-three-months-ago. She'd just spent two hours looking for it.
She found it, finally, inside a folder named "Papers_FINAL_v3_real_ACTUAL." She opened it. The highlighting was there. The annotations were good. Past-Alex had done excellent work. Present-Alex had zero memory of any of it.
She opened a blank notebook and wrote: "What if there was a tool that just... remembered?"
She stared at that sentence for a long time. Then, below it, she wrote a second sentence that would change the next two years of her life: "Is this a me problem or an everyone problem?"
Problem validation is the process of confirming, before significant investment, that a real, painful problem exists for a sizable group of people willing to pay for a solution. Most startups fail not because they build poorly, but because they build something nobody wants.
There are two related concepts founders often conflate:
The key question experienced founders ask isn't "would you use this?" Instead, it's "how do you handle this problem today?" Past behavior reveals real pain far better than hypothetical intent. Customers can't accurately predict future behavior, but they can describe what they actually did last week.
Alex didn't know any of this yet. She just knew she had a hunch, and she wanted to know if it was shared.
Alex googled "research paper management software." The results were extensive: Zotero, Mendeley, Papers, ReadCube, Paperpile, Citavi. She spent 20 minutes with each. Zotero felt like software that had been lovingly maintained since 2009: functional, beloved by academics, utterly without visual ambition. Mendeley crashed twice. Papers was genuinely beautiful, but its organization model didn't match how she actually thought about research. It wanted her to file papers; she wanted to connect them.
She opened G2 and Capterra. She filtered by "research management software" and sorted by lowest rating. The one-star reviews were illuminating. Three different users, different tools, almost the same complaint: "Doesn't handle cross-paper connections." One wrote: "I can find individual papers, but I can't see how they relate to each other. This is literally the only thing I need."
Alex felt a jolt, the specific, electric kind that comes from finding your hypothesis in the wild. She wrote in her notebook: "Competitors exist = demand exists. But they're leaving something on the table."
Always size using both approaches: top-down (industry reports β narrow to segment) and bottom-up (# of customers Γ average revenue per user). If both numbers converge, you have a credible estimate.
Alex had read about smoke tests in a YC blog post. The idea was almost offensively simple: build a landing page for a product that doesn't exist, run ads at it, and measure whether people want it enough to give you their email address. She told herself she'd try it over the weekend. It took her until 11pm on Friday to actually start.
By Saturday morning, she had a Carrd page. Headline: "Lumio β Never lose a research insight again." Three bullet points. An email signup form. She spent $50 on Google Ads targeting search terms like "research paper management" and "academic literature tool." She set the campaign live at 3pm and refreshed the analytics dashboard approximately every seven minutes for the rest of the day.
Sunday evening results: 847 impressions, 63 clicks (7.4% CTR), 19 email signups. That was a 30% email capture rate. She read every one of the 19 emails. The third one said: "I have been waiting for something like this for literally years."
Alex screamed into a pillow. Then opened her notebook and wrote, in capital letters: IT'S REAL.
A smoke test (also called a fake door test) is a lightweight experiment: you launch a landing page or ad describing a product that doesn't yet exist and measure real interest before building anything.
How to set one up:
What counts as a signal: An email capture rate above 5% is considered a strong demand signal. Alex hit 30%, which is exceptional. A rate below 1β2% suggests either a messaging problem or a demand problem.
What it doesn't prove: Willingness to pay. "Sounds cool" β credit card. The smoke test validates interest; pre-sales validate intent to pay. Both matter.
These are the signals that separate genuine market pull from polite interest. Alex's score after her first month: 5 out of 6.
Use Alex's process. Fill in the template below. This is your first validation artifact. Be specific. Vague hypotheses lead to vague findings.
Validate before building. Most startups fail not because they build badly, but because they build something nobody wants. 15β20 problem interviews and a smoke test can save you months of wasted work.
Competitors are your friends. Their existence proves demand. Read their negative reviews to find the exact gap they're leaving open. That gap is your opportunity.
Set your threshold before you start. Decide in advance what evidence would convince you to proceed or pivot. "Sounds cool" β willingness to pay. "I've been waiting for this for years" + 30% email capture = something real.