Install
openclaw skills install bookforge-product-market-fit-readiness-gateUse this skill to determine whether a product is ready for scaled growth experimentation BEFORE investing in acquisition, activation, or retention hacks. Run...
openclaw skills install bookforge-product-market-fit-readiness-gateA go/no-go gate that determines whether a product has achieved must-have status before the team invests in scaled growth experimentation. Produces a structured verdict with evidence and, for failing products, a concrete remediation protocol.
Use this skill at the moment a growth team is deciding whether to shift from product development into high-tempo experimentation. Typical triggers:
Do not use as an ongoing monitoring tool once growth is already at scale — at that stage, retention dashboards and North Star metrics replace this gate.
Begin by reading the product brief. Then check whether survey data and cohort retention data are available.
Branch A — Survey data exists (survey-responses.csv):
Proceed directly to Step 3 (scoring). Skip survey template generation.
Branch B — No survey data:
Ask the user: "Have you already sent a must-have survey to active users? If yes, share the response file. If not, I'll generate the survey template now so you can send it before scoring."
Generate must-have-survey-template.md (Step 2), then pause and wait for results before completing the verdict.
Branch C — No retention data (retention-cohorts.csv):
Proceed with survey scoring only. Flag in the verdict that retention analysis is incomplete and recommend pulling cohort data from your analytics tool (Mixpanel, Amplitude, Looker, etc.) before treating the verdict as final.
Read product-brief.md. Extract:
Why: The scoring thresholds and remediation paths differ by product type (high-frequency consumer app vs. annual SaaS vs. marketplace). Without context, the verdict will be generic and the remediation protocol will miss the most likely failure modes.
Output: A one-paragraph context summary to embed in the verdict document.
Write must-have-survey-template.md containing the exact survey question and answer options:
Subject: Quick question about [Product Name]
How would you feel if you could no longer use [Product Name]?
a) Very disappointed
b) Somewhat disappointed
c) Not disappointed (it really isn't that useful)
d) N/A — I no longer use it
Include the following diagnostic questions (send only if initial score is below threshold, or include upfront for efficiency):
What would you likely use as an alternative to [Product Name] if it were no longer available?
What is the primary benefit you have received from [Product Name]?
Have you recommended [Product Name] to anyone?
What type of person do you think would benefit most from [Product Name]?
How can we improve [Product Name] to better meet your needs?
Would it be okay if we followed up by email to clarify one or more of your responses?
Targeting note: Send to active users only (those who have used the product in the past 30 days). Dormant users produce uninformative responses and low completion rates. Aim for at least 200–300 responses before treating the score as reliable.
Why: Survey wording matters. Substituting "love" or "miss" for "very disappointed" changes the emotional register and invalidates comparison to the published threshold. The exact wording is the validated signal.
Output: must-have-survey-template.md
Count responses by category. Calculate:
Very Disappointed % = (count of "Very disappointed" responses) / (total responses excluding "N/A") × 100
Apply the three-band decision rule:
| Score | Verdict | Meaning |
|---|---|---|
| >= 40% | GO | Product has achieved must-have status. Green light for growth experimentation. |
| 25–40% | CONDITIONAL | Product or messaging needs targeted improvement. Limited experiments may run in parallel, but broad acquisition investment is premature. |
| < 25% | NO-GO | Either the wrong audience is being surveyed, or the product requires substantial development. Halt acquisition spend immediately. |
Why: The 40% floor is not arbitrary — it reflects the minimum density of passionate users needed to sustain word-of-mouth and withstand the churn that follows broad acquisition. Scaling below this threshold accelerates user exodus (the BranchOut pattern): viral mechanics bring users who find no compelling reason to stay, CAC permanently exceeds LTV, and the company burns capital on a leaking bucket. The 25–40% band is actionable because the product has real fans — they just need either a better product or better communication of the existing value.
Output: Score summary section in pmf-readiness-verdict.md
If retention-cohorts.csv is available, read it. Evaluate:
Shape: Does the curve flatten and stabilize, or does it continue declining toward zero? A curve that flattens — even at a low absolute rate — indicates a core segment of retained users. A curve continuously declining toward zero means no durable value for any cohort.
Level vs. benchmarks:
Masking signal: Check whether strong new-user acquisition is hiding early-cohort churn. Isolate cohorts by signup month and track each independently.
Verdict mapping:
Why: The survey captures users' stated preference; the retention curve reveals revealed behavior. Both signals need to align. A product can pass the survey (enthusiastic fans who said they'd be very disappointed) but still show declining retention if the aha moment is inconsistently delivered. Retention without survey data is also incomplete — it cannot tell you whether satisfaction is high enough among those who do stay.
Output: Retention analysis section in pmf-readiness-verdict.md
Write pmf-readiness-verdict.md with the following structure:
# PMF Readiness Verdict: [Product Name]
**Date:** [date]
**Verdict:** GO / CONDITIONAL / NO-GO
## Evidence Summary
- Must-have survey score: X% very disappointed (N=total responses)
- Retention curve: [stable at Y% / continuously declining / data unavailable]
- Product category: [consumer mobile / SaaS / marketplace / other]
## Rationale
[2–3 sentences explaining the primary reason for the verdict, using the evidence above]
## Recommended Next Actions
[Specific to verdict — see Step 6 for NO-GO protocol]
Why: A written verdict forces the decision to be explicit and shared. Growth teams that skip this step frequently re-debate readiness at every planning meeting, wasting cycles. The written artifact is the commitment device.
Output: pmf-readiness-verdict.md
For products scoring below 40%, do not guess at solutions in internal planning sessions. The three concurrent methods:
A. Customer interviews (marketing/product design lead) Talk to active users — especially the "very disappointed" sub-segment, however small. Observe them using the product in context, not just in a formal interview. What job are they hiring the product to do? What friction prevents others from reaching that same outcome?
B. Diagnostic analysis of "very disappointed" cohort (data analyst lead) Segment survey respondents by score. Identify behavioral differences: what did the "very disappointed" users do in the product that "somewhat disappointed" or "not disappointed" users did not? That difference often reveals the aha moment path that needs to be broadened.
C. Minimum viable tests on product and messaging (engineering + growth lead) Run the smallest possible experiments that could close the gap — starting with messaging changes (cheapest, fastest) before committing to feature builds. Do not add features without evidence from methods A and B that the feature addresses a confirmed gap.
Caution on feature creep: Adding features is the intuitive response to a failing score, but it is often wrong. The remediation frequently requires removing friction or improving delivery of existing value, not adding new capabilities.
Why: Teams that skip structured remediation and instead hold whiteboard brainstorms burn time and money on unvalidated bets. The three-method approach distributes diagnostic work across specializations and produces convergent evidence before any engineering investment.
Output: Remediation plan appended to pmf-readiness-verdict.md
1. 40% is a floor, not a target. Products that just clear 40% are marginal. Teams should continue working to raise the score even after passing the gate. The threshold is the minimum for viable growth, not evidence of a breakout product.
2. Retention confirms what the survey suggests. The survey captures stated preference under a hypothetical. The retention curve reveals actual behavior. Both signals must align before committing to scaled acquisition. A strong survey score with a declining retention curve means the aha moment is being promised but not reliably delivered.
3. Premature scaling is the number one growth failure mode. Viral and paid acquisition mechanics can temporarily inflate user counts regardless of product quality. The BranchOut case demonstrates the collapse pattern: rapid acquisition before must-have status creates a user pool that churns faster than new users arrive, permanently exceeding cost of acquisition over lifetime value. The growth investment is consumed before the product is ready to convert it.
4. Survey active users only. Targeting dormant users produces response bias and low completion rates. The signal you need comes from people who have experienced the product — even if their verdict is disappointment. Dormant users have already voted with their absence.
5. The 25–40% band is actionable, not a failure. A score in this range means real fans exist. The job is to understand what those fans experience that others do not — and redesign the product or messaging to deliver that experience more consistently or to a better-matched audience.
6. This gate applies once per growth phase, not perpetually. Re-run this gate when the product undergoes significant changes (new ICP, major feature pivot, new market entry). Do not use it as a routine measurement tool once growth is established.
Scenario: A B2B SaaS team has 800 paying users and is considering a $200K acquisition campaign. The Head of Growth asks, "Are we ready to scale paid acquisition?"
Trigger: "Are we ready to scale?" + existing user base large enough to survey.
Process summary:
Output excerpt from pmf-readiness-verdict.md:
Verdict: NO-GO
Score: 18% very disappointed (well below 40% threshold)
Retention: Continuously declining — no stable cohort identified
Rationale: Both signals confirm the product has not achieved must-have status.
Scaling acquisition now would accelerate churn rather than compound growth.
Recommended next actions:
1. Segment the 18% "very disappointed" users — identify their behavioral patterns
2. Conduct 8–10 qualitative interviews with this cohort to identify the aha moment path
3. Run messaging experiments to test whether the core value is being communicated
to the right audience before building any new features
Scenario: A productivity app has been in public beta for 4 months. The founding team is debating whether to hire a growth engineer and launch a referral program.
Trigger: "Should we hire a growth team?" + "do we have product/market fit?"
Process summary:
Output excerpt from pmf-readiness-verdict.md:
Verdict: GO
Score: 52% very disappointed (above 40% threshold)
Retention: Stable at 41% at month 3 — well above consumer mobile average of ~10%
Rationale: Both signals confirm must-have status. A majority of active users would
be significantly impacted by loss of the product. Retention curve has stabilized at
a competitive level. The product is ready for high-tempo growth experimentation.
Recommended next actions:
1. Proceed with referral program design (see: growth-referral-loop-designer)
2. Identify North Star metric before beginning growth experiments (see: north-star-metric-selector)
3. Re-run this gate if the product undergoes a significant pivot or new market entry
Methodology sourced from:
This skill is licensed under CC-BY-SA-4.0. Source: BookForge — Hacking Growth by Sean Ellis and Morgan Brown.
This skill is the foundation of the Hacking Growth skill set. Run it before these dependents:
clawhub install bookforge-north-star-metric-selector — pick the growth equation variable that matters after PMF is confirmedclawhub install bookforge-retention-phase-intervention-selector — use the retention curve from this skill as the diagnostic starting pointBrowse more: bookforge-skills