Install
openclaw skills install bookforge-product-team-health-diagnosticDiagnose why a product team or organization is slow, not innovative, or delivering poor outcomes. Use when a leader or team observes slow velocity, lack of innovation, poor product quality, feature factory behavior, or team dysfunction — and needs root causes and a prioritized fix list. Also use when a new product leader is assessing an organization, when a CEO or board says teams are too slow, or when someone says 'why are we not shipping faster?', 'engineering and design aren't collaborating', 'we ship but nothing moves the needle', or 'I need to assess team health before proposing changes.' Scores 42 diagnostic criteria across team behaviors, innovation capacity, velocity killers, and design integration. Produces a severity-ranked report with a composite health score and remediation priorities. For culture-level issues (innovation vs. execution quadrant), use product-culture-assessment. For process-level waterfall diagnosis, use product-process-dysfunction-diagnosis.
openclaw skills install bookforge-product-team-health-diagnosticUse this skill when you are:
Preconditions: you have at least one of:
Agent: Clarify the scope before beginning — are you assessing one team, a portfolio of teams, or the entire engineering/product organization? The scoring applies per team; an org-level report aggregates across teams.
Collect evidence for each of the 4 diagnostic categories. WHY: each category targets a distinct failure mode — behavioral dysfunction (Ch64), structural innovation blockers (Ch65), process velocity killers (Ch66), and design integration failures (Ch11). Mixing them obscures root cause.
For each category, ask the assessor (or yourself) the following:
Category A — Team Behaviors (18 criteria) Source: Ch64. These contrast what strong teams do vs. what weak teams do. Ask:
references/diagnostic-criteria.md#category-aCategory B — Innovation Capacity (10 criteria) Source: Ch65. These are organizational attributes that determine whether consistent innovation is even possible. Ask:
references/diagnostic-criteria.md#category-bCategory C — Velocity Killers (10 criteria) Source: Ch66. These are the structural and process causes of slowness. Ask:
references/diagnostic-criteria.md#category-cCategory D — Design Integration (4 criteria) Source: Ch11. These identify whether design is integrated as a discovery partner or treated as a service. Ask:
references/diagnostic-criteria.md#category-dWHY: Scoring converts qualitative observations into a comparable signal, making it possible to prioritize and track improvement over time.
For each of the 42 criteria, assign one of three scores:
| Score | Meaning |
|---|---|
| 2 | Healthy — the team clearly exhibits the strong behavior |
| 1 | Partial — the behavior is inconsistent or only sometimes present |
| 0 | Absent — the weak behavior is the norm |
Scoring rules:
For each category, calculate:
Category score = (sum of item scores) / (max possible score) × 100
| Category | Items | Max Score |
|---|---|---|
| A — Team Behaviors | 18 | 36 |
| B — Innovation Capacity | 10 | 20 |
| C — Velocity Killers | 10 | 20 |
| D — Design Integration | 4 | 8 |
| Composite | 42 | 84 |
WHY: Keeping categories separate prevents a strong score in one area from masking a critical failure in another. A team can be fast (good C score) but consistently build the wrong things (low A score).
WHY: Not all scores below 100% require equal urgency. This classification focuses remediation effort.
Per-category thresholds:
| Score | Severity | Interpretation |
|---|---|---|
| 80–100% | Healthy | Maintain; minor tuning only |
| 60–79% | Caution | Targeted improvements needed |
| 40–59% | Degraded | Structural issues present; prioritize fixes |
| 0–39% | Critical | Fundamental dysfunction; urgent intervention required |
Red flags — any single criterion scoring 0 in these areas triggers automatic Critical classification for that dimension, regardless of category average:
WHY: These specific items represent systemic dysfunctions that a higher average cannot compensate for.
Structure the output as follows:
## Product Team Health Diagnostic Report
**Organization/Team:** [name]
**Assessment Date:** [date]
**Assessor:** [role]
---
### Composite Score: [X/84] — [XX%] — [SEVERITY]
| Category | Score | % | Severity |
|----------|-------|---|----------|
| A — Team Behaviors | X/36 | XX% | [label] |
| B — Innovation Capacity | X/20 | XX% | [label] |
| C — Velocity Killers | X/20 | XX% | [label] |
| D — Design Integration | X/8 | XX% | [label] |
---
### Red Flags (Criteria scoring 0 that require immediate attention)
[List each, with one sentence describing the observed symptom]
---
### Category Findings
#### A — Team Behaviors
**Strengths:** [criteria scoring 2]
**Gaps:** [criteria scoring 0 or 1, with observed evidence]
#### B — Innovation Capacity
[same format]
#### C — Velocity Killers
[same format]
Note: Release cadence benchmark — minimum every 2 weeks; great teams release multiple times per day.
#### D — Design Integration
[same format]
---
### Prioritized Remediation Plan
Ordered by: (1) Critical severity first, (2) within severity by cross-category impact
| Priority | Issue | Category | Current State | Target State | Estimated Effort |
|----------|-------|----------|---------------|--------------|-----------------|
| 1 | ... | ... | ... | ... | ... |
---
### Summary Narrative
[3–5 sentences: what is working, what is broken, and the single most important thing to fix first and why]
Before delivering the report, verify:
WHY: Diagnostic reports are only useful if they drive decisions. Vague findings ("culture needs improvement") create no action. Specific findings ("engineers first see features at sprint planning — exclude from discovery") point to a concrete fix.
Common misread — high velocity, low innovation: A team can score well on Category C (velocity) while scoring poorly on Category B (innovation). This is the "feature factory" pattern — teams ship fast but build the wrong things. The fix is upstream (discovery and vision), not downstream (process acceleration).
Common misread — blaming individuals: Low PM capability scores (B4, C2) often reflect structural issues — PMs who are order-takers because leadership treats them as such. Diagnose whether the PM is weak, or whether the PM has been structurally prevented from being strong.
Common misread — design as optional: Category D scores are frequently rationalized ("we're a B2B company, design matters less"). The book is explicit: strong design is a competitive differentiator in B2B, and companies that treat it as optional are being displaced.
The innovation/velocity relationship: Chapters 65 and 66 share several root causes (weak PMs, engineers excluded, no vision). Fixes to these shared causes yield compound improvements across both dimensions.
Full 42-criterion reference table with good/bad behavior descriptions:
references/diagnostic-criteria.md
This skill is licensed under CC-BY-SA-4.0. Source: BookForge — Inspired How To Create Tech Products by Unknown.
This skill is standalone. Browse more BookForge skills: bookforge-skills