Customer Discovery Process

v1.0.0

Orchestrate the full customer discovery process — before, during, and after customer conversations — to systematically validate a business idea. This is the...

0· 55·0 current·0 all-time
byHung Quoc To@quochungto

Install

OpenClaw Prompt Flow

Install with OpenClaw

Best for remote or guided setup. Copy the exact prompt, then paste it into OpenClaw for quochungto/bookforge-customer-discovery-process.

Previewing Install & Setup.
Prompt PreviewInstall & Setup
Install the skill "Customer Discovery Process" (quochungto/bookforge-customer-discovery-process) from ClawHub.
Skill page: https://clawhub.ai/quochungto/bookforge-customer-discovery-process
Keep the work scoped to this skill only.
After install, inspect the skill metadata and help me finish setup.
Use only the metadata you can verify from ClawHub; do not invent missing requirements.
Ask before making any broader environment changes.

Command Line

CLI Commands

Use the direct CLI path if you want to install manually and keep every step visible.

OpenClaw CLI

Bare skill slug

openclaw skills install bookforge-customer-discovery-process

ClawHub CLI

Package manager switcher

npx clawhub@latest install bookforge-customer-discovery-process
Security Scan
Capability signals
CryptoCan make purchases
These labels describe what authority the skill may exercise. They are separate from suspicious or malicious moderation verdicts.
VirusTotalVirusTotal
Benign
View report →
OpenClawOpenClaw
Benign
high confidence
Purpose & Capability
The name/description describe an orchestration hub for customer discovery and the SKILL.md only requires checking product/validation artifacts, producing a dashboard, and delegating to sibling discovery skills — all consistent with the stated purpose. There are no unrelated requirements (no cloud credentials or extraneous binaries).
Instruction Scope
Runtime instructions focus on before/during/after discovery tasks and specify looking for local files (product-idea.md, learning-log.md, conversation-notes/) and prompting the user when missing. This scope is appropriate for a process orchestration skill. Note: it does instruct the agent to read and write workspace files, which is expected but worth awareness.
Install Mechanism
This is an instruction-only skill with no install spec, no code files, and no downloads — lowest-risk install profile.
Credentials
The skill declares no required environment variables, credentials, or config paths. It does require generic Read/Write tool access to the agent environment (to inspect and create local discovery artifacts), which is proportionate to its function. There is no request for unrelated secrets.
Persistence & Privilege
The skill is not marked always:true, is user-invocable, and permits normal autonomous invocation (platform default). It does not request system-wide persistence or modify other skills. Its hybrid mode and delegation to sibling skills is appropriate for orchestration.
Assessment
This skill appears coherent and is safe in the sense that it does not request secrets or install code, but it will read and write files in the agent's workspace and invoke other discovery skills. Before installing or running it: (1) ensure there are no sensitive secrets or personally identifiable customer data in workspace files (product-idea.md, conversation-notes/, learning-log.md) — redact if needed; (2) review the sibling skills it depends on (conversation-question-designer, conversation-data-quality-analyzer, etc.) to confirm they do not require credentials or external network access you don't want; (3) be aware the agent needs Read/Write access to create the discovery dashboard and notes — restrict permissions if you want finer control; (4) if you prefer to avoid autonomous invocations, use platform controls to require explicit user approval before the skill runs. If you want further assurance, provide a sample workspace or the sibling skills' manifests so I can inspect those as well.

Like a lobster shell, security has layers — review code before you run it.

Runtime requirements

📚 Clawdis
before-during-aftervk975kcpexb7da8sfc6r29h82m185hs33bookforgevk975kcpexb7da8sfc6r29h82m185hs33customer-discoveryvk975kcpexb7da8sfc6r29h82m185hs33discovery-dashboardvk975kcpexb7da8sfc6r29h82m185hs33latestvk975kcpexb7da8sfc6r29h82m185hs33orchestrationvk975kcpexb7da8sfc6r29h82m185hs33validation-processvk975kcpexb7da8sfc6r29h82m185hs33
55downloads
0stars
1versions
Updated 2d ago
v1.0.0
MIT-0

Customer Discovery Process

When to Use

You need to run a structured customer discovery process to validate a business idea through real conversations. This is the hub skill — it orchestrates the full before/during/after workflow and delegates specialized work to sibling skills. Typical situations:

  • The user has an idea and wants a systematic plan for validating it through customer conversations
  • The user is partway through discovery and needs to assess progress and decide next steps
  • The user suspects their discovery process is broken — lots of meetings but no real learning
  • The user wants to set up discovery from scratch and does not know where to start
  • The user asks "what should I do next?" after a batch of conversations

Mode: Hybrid — The agent orchestrates the process, produces the discovery dashboard, and delegates to specialized skills. The human conducts the actual conversations.

Context & Input Gathering

Required Context (must have — ask if missing)

  • Product idea or problem hypothesis: What is the user building or exploring? A sentence is enough. -> Check prompt for: product descriptions, problem statements, startup ideas -> Check environment for: product-idea.md, README.md -> If still missing, ask: "What product or problem are you exploring? A sentence or two is enough to get started."

  • Current validation stage: Where is the user in their discovery process? -> Check prompt for: "just starting," "have done N conversations," "getting mixed feedback," stage indicators -> Check environment for: learning-log.md, conversation-notes/, customer-segments.md -> If still missing, ask: "Where are you in customer discovery? (a) Have not started yet, (b) Have done a few conversations, (c) Have done many conversations but feel stuck"

Observable Context (gather from environment)

  • Existing discovery artifacts: Check for files that reveal what has already been done. -> Look for: customer-segments.md, question-script.md, learning-log.md, outreach-plan.md, commitment-tracker.md, conversation-notes/ -> If found: assess which process steps are complete and which are pending

  • Team context: Is this a solo founder or a team? -> Look for: mentions of co-founders, team members -> If unavailable: assume solo founder

Default Assumptions

  • If no stage specified -> assume the user is starting from scratch
  • If no team mentioned -> assume solo founder (but flag the importance of having a review partner)
  • If no prior conversations -> begin at the BEFORE phase

Sufficiency Threshold

SUFFICIENT when ALL of these are true:
- Product idea or problem area is known
- Current validation stage is known or can be inferred from environment

PROCEED WITH DEFAULTS when:
- Product idea is approximate ("something for freelancers")
- Stage is vague ("I've talked to a few people")

MUST ASK when:
- No product idea or problem area at all

Process

Step 1: Assess Current Discovery State

ACTION: Determine where the user is in the discovery process by checking for existing artifacts and conversation history. Run the going-through-the-motions diagnostic if they have already started.

WHY: The discovery process is not linear — users enter at different points and may need to loop back. Assessing state prevents repeating completed work and identifies the highest-impact next step. Many users who "feel stuck" are actually going through the motions without real learning, and diagnosing this early saves weeks of wasted conversations.

Going-through-the-motions diagnostic — check for these 9 warning signs:

  1. You are talking more than the customer is
  2. They are complimenting you or your idea
  3. You told them about your idea and do not know what is happening next
  4. You do not have notes from the conversation
  5. You have not reviewed your notes with your team
  6. You got an unexpected answer and it did not change your idea
  7. You were not scared of any of the questions you asked
  8. You are not sure which big question you are trying to answer
  9. You are not sure why you are having the meeting

IF 3 or more warning signs are present -> flag the process as broken. The user needs to fix their approach before having more conversations. Focus on the BEFORE phase regardless of how many conversations they have already done. IF the user has not started -> proceed to Step 2 (BEFORE phase). IF the user has conversation notes to review -> proceed to Step 4 (AFTER phase).

Step 2: BEFORE — Prepare for Conversations

ACTION: Execute the preparation sequence. Each sub-step invokes a specialized skill.

WHY: Most bad conversations happen because of bad preparation. Choosing the wrong segment means talking to the wrong people. Unclear learning goals mean the conversation wanders. Bad questions produce bad data. The BEFORE phase is where you set up the conversation to succeed. Skipping it is the primary cause of "going through the motions."

2a. Segment your customers IF no customer segment is defined or the current one is too broad -> invoke customer-segment-slicer to produce a specific who-where pair. WHY: A conversation without a focused segment produces mixed signals that cannot inform decisions. You are not having 20 conversations with your customers — you are having 1 conversation each with 20 different types of customers.

2b. Plan conversation sourcing IF the user knows WHO but not HOW to reach them -> invoke conversation-sourcing-planner to create an outreach plan with framing templates. WHY: Most founders default to cold email or formal meetings when casual conversations at natural gathering points would produce better data faster with less overhead.

2c. Set learning goals and prepare questions ACTION: Invoke question-importance-prioritizer to classify risks and select the 3 most important learning goals. Then invoke conversation-learning-process (Step 2: pre-conversation prep) to set those goals with the team. Finally invoke conversation-question-designer to produce a question script aligned to those goals. WHY: The minimum viable prep is answering one question: "What do we want to learn from these guys?" Without this, the conversation is aimless. The question script ensures the conversation produces facts instead of compliments.

2d. Choose conversation format IF the user is unsure about meeting format -> invoke conversation-format-selector to recommend casual vs formal vs phone. WHY: Defaulting to 1-hour formal meetings for every conversation is the Meeting Anti-Pattern. A 5-minute casual chat often produces the same learning with a fraction of the time cost.

2e. Define target commitments ACTION: Based on the user's product stage, define what commitment to push for at the end of the meeting.

Product StageTarget Commitment
Problem exploration (no product)Time: next meeting with specific goals. Reputation: intro to others with the same problem.
Wireframes or prototypeTime: sit down to review wireframes. Reputation: intro to decision maker.
Working product or betaTime: non-trivial trial. Reputation: case study. Money: letter of intent.
Live productMoney: purchase or deposit. Reputation: public testimonial.

OUTPUT: A completed preparation checklist ready for the human to execute.

Step 3: DURING — Conversation Execution Guidance

ACTION: Provide the human with a field reference card for real-time use during the conversation.

WHY: Even with perfect preparation, conversations go off-script. The human needs to recognize danger signals in real-time and recover. This step produces a compact reference, not a rigid script — the human leads the conversation while the card keeps them honest.

HANDOFF TO HUMAN — The agent cannot conduct the conversation. Provide this field card:

Keep it casual. If it feels like they are doing you a favor by talking to you, it is probably too formal.

Ask questions that pass 3 rules:

  1. Talk about their life, not your idea
  2. Ask about specifics in the past, not generics about the future
  3. Talk less and listen more

Recover from bad data in real-time:

  • If you hear a compliment -> deflect: "Thanks — but how are you currently handling this?"
  • If you hear fluff ("I would definitely...") -> anchor: "When was the last time that came up?"
  • If you hear a feature request -> dig: "Why do you want that? What would it let you do?"

Press for commitment and next steps. Do not leave without a concrete next step. The meeting either advances or it fails — there is no "went well."

Take notes. Capture exact quotes. Use signal symbols if possible. Without notes, the conversation might as well not have happened.

Step 4: AFTER — Review, Synthesize, and Decide Next Steps

ACTION: Execute the post-conversation sequence. Each sub-step invokes a specialized skill.

WHY: The AFTER phase is where conversations turn into validated learning. Without it, insights stay in one person's head, bad data gets treated as fact, and the team never updates their assumptions. Most teams skip this phase, which is why they feel stuck after dozens of conversations.

4a. Analyze conversation quality ACTION: Invoke conversation-data-quality-analyzer on the conversation notes to classify every statement as FACT, COMPLIMENT, FLUFF, or IDEA. Get the quality rating (STRONG / MIXED / WEAK / EMPTY). WHY: A conversation that "went well" is a warning sign. Quantifying the ratio of facts to noise gives an objective measure instead of a gut feeling driven by compliments.

4b. Evaluate commitment signals ACTION: Invoke commitment-signal-evaluator on the meeting outcomes to classify commitment currencies (time, reputation, money) and detect zombie leads. WHY: Compliments cost nothing and are worth nothing. The only reliable measure of interest is what the prospect gave up. This step separates real leads from false-positive prospects.

4c. Run the team learning review ACTION: Invoke conversation-learning-process (Steps 3-5: categorize notes, run team review, set next learning goals) to transfer learnings to the whole team and update assumptions. WHY: Customer learning is a team sport. When insights stay in one person's head, that person becomes a de-facto dictator wielding "the customer said so" as an unchallengeable trump card. The review is where learning actually transfers.

4d. Update the discovery dashboard ACTION: Update the discovery status dashboard (see Step 5) with new data from this conversation batch.

Step 5: Produce the Customer Discovery Status Dashboard

ACTION: Create or update the discovery dashboard — a single document showing the full state of the user's validation progress.

WHY: Customer discovery produces many artifacts across many conversations. Without a single status view, the user loses track of what has been validated, what remains unknown, and what to do next. The dashboard is the tangible deliverable of this hub skill — it aggregates output from all sibling skills into one actionable summary.

Dashboard format:

# Customer Discovery Status Dashboard

## Product
- **Idea:** [product description]
- **Stage:** [problem exploration / prototype / beta / live]
- **Date Updated:** [today]

## Active Segment
- **WHO:** [specific customer segment]
- **WHERE:** [finding location]
- **Source:** customer-segment-slicer output or user-defined
- **Segment Status:** [needs slicing / defined / validated]

## Learning Goals (Current Batch)
| # | Learning Goal | Status | Evidence |
|---|--------------|--------|----------|

## Conversation Log
| # | Date | Person | Quality | Commitment | Key Insight |
|---|------|--------|---------|------------|-------------|
| 1 | [date] | [who] | [STRONG/MIXED/WEAK] | [currency type] | [one-line insight] |

## Assumption Tracker
| # | Assumption | Status | Evidence | Source |
|---|-----------|--------|----------|--------|

## Commitment Pipeline
| Prospect | Highest Currency | Earlyvangelist Score | Next Step | Status |
|----------|-----------------|---------------------|-----------|--------|
| [name] | [time/reputation/money/none] | [0-4 criteria met] | [action] | [active/zombie/closed] |

## Process Health
- **Conversations this batch:** [N]
- **Going-through-the-motions signs:** [count of 9] — [list any present]
- **Good meeting results:** [count of meetings producing Facts + Commitment + Advancement]
- **Learning bottleneck risk:** [yes/no — is one person hoarding insights?]

## Results of a Good Meeting (Checklist)
For each conversation, verify it produced at least one of:
- [ ] **Facts** — concrete, specific facts about what they do and why (not compliments, fluff, or opinions)
- [ ] **Commitment** — they gave up something of value: time, reputation, or money
- [ ] **Advancement** — they moved to the next step of your real-world funnel, closer to a sale

## Next Actions
1. [Highest-priority next step]
2. [Second priority]
3. [Third priority]

IF the user provided a working directory -> write to discovery-dashboard.md ELSE -> present directly in the conversation

Examples

Scenario: Founder starting from scratch with a SaaS idea

Trigger: "I have an idea for a tool that helps freelance designers manage client feedback. How do I validate this?" Process: Assess state (starting from scratch). BEFORE phase: invoke customer-segment-slicer ("freelance designers" too broad, slice to "freelance UI/UX designers on Dribbble with 3+ simultaneous projects"). Invoke conversation-learning-process for prep (3 goals: feedback loop pain, current tools, past fix attempts). Invoke conversation-question-designer for script. Define target commitments (pre-product: time + reputation). Output: Discovery dashboard with BEFORE phase complete, segment defined, question script ready, zero conversations logged. Action plan: "Talk to 3-5 freelance UI/UX designers this week."


Scenario: Founder stuck after 12 conversations

Trigger: "I've done 12 customer interviews for my restaurant inventory tool but everyone says it sounds great and nobody has signed up for the beta." Process: Run going-through-the-motions diagnostic — detected 3 of 9 signs (compliments without commitments, no scary questions, nothing changed). Invoke conversation-data-quality-analyzer on notes (likely WEAK — mostly compliments and fluff). Invoke commitment-signal-evaluator (zero currencies across 12 meetings, zombie lead pattern). Diagnosis: collecting compliments instead of facts. Output: Dashboard showing 12 conversations, zero validated assumptions, zero commitments. Restart plan: re-slice segment, set scary learning goals, redesign questions to hide the product, define concrete commitment asks.


Scenario: Team reviewing after a productive conversation batch

Trigger: "My co-founder and I finished 5 conversations this week about our invoicing tool. Here are our notes. What next?" Process: Enter AFTER phase. Invoke conversation-data-quality-analyzer on each set of notes. Invoke commitment-signal-evaluator on outcomes. Invoke conversation-learning-process for team review, assumption updates, and next learning goals. Output: Updated dashboard with 5 conversations logged, quality ratings, strongest lead flagged, assumptions updated, next 3 learning goals set.

Key Principles

  • This skill orchestrates, it does not duplicate — Every specialized task (segmenting, questioning, analyzing, evaluating commitments, reviewing learnings) is handled by a sibling skill. This skill sequences them in the right order and maintains the overall discovery state. If you find yourself doing detailed question design or data classification here, you are in the wrong skill.

  • The process is before/during/after, not plan/execute — Customer discovery is not a project plan you execute once. It is a loop: prepare, talk, review, update, repeat. Each cycle through the loop should produce validated or invalidated assumptions. If you have been through the loop multiple times and nothing has changed, the process itself is broken.

  • A meeting that "went well" is a warning sign — Good meetings produce facts, commitment, and advancement. Bad meetings produce compliments and warm feelings. The phrase "it went well" almost always means the latter. Every meeting must be scored against the three results criteria (facts, commitment, advancement), not against how it felt.

  • Conversations are a tool, not an obligation — Having a process does not mean having more meetings. The goal is to learn what you need as quickly as possible and get back to building. Three focused conversations with the right segment can produce more learning than thirty scattered ones with the wrong people. If conversations are not producing learning, fix the process or stop having them.

  • Process without action is worse than no process at all — Having a process is valuable, but do not get stuck in it. Sometimes you can just pick up the phone and hack through the knot. The personal trainer who called the police station instead of agonizing over customer segmentation got a trial session in twenty minutes. Process exists to serve learning, not the other way around.

References

  • Segment customers into who-where pairs -> customer-segment-slicer
  • Create outreach plans and framing templates -> conversation-sourcing-planner
  • Design questions that pass the 3 quality rules -> conversation-question-designer
  • Choose conversation format (casual/formal/phone) -> conversation-format-selector
  • Prioritize which learning goals matter most -> question-importance-prioritizer
  • Analyze conversation notes for bad data -> conversation-data-quality-analyzer
  • Evaluate commitment signals and detect zombie leads -> commitment-signal-evaluator
  • Run team learning review and set next goals -> conversation-learning-process

License

This skill is licensed under CC-BY-SA-4.0. Source: BookForge — The Mom Test by Rob Fitzpatrick.

Related BookForge Skills

Install related skills from ClawhHub:

  • clawhub install bookforge-conversation-question-designer
  • clawhub install bookforge-conversation-data-quality-analyzer
  • clawhub install bookforge-commitment-signal-evaluator
  • clawhub install bookforge-customer-segment-slicer
  • clawhub install bookforge-conversation-learning-process
  • clawhub install bookforge-conversation-sourcing-planner
  • clawhub install bookforge-conversation-format-selector
  • clawhub install bookforge-question-importance-prioritizer

Or install the full book set from GitHub: bookforge-skills

Comments

Loading comments...