Agent Analytics Autoresearch
v1.0.6Run an autoresearch-style growth loop for landing pages, onboarding, pricing, and experiment candidates. Collect or read analytics snapshots, preserve produc...
Like a lobster shell, security has layers — review code before you run it.
Runtime requirements
Agent Analytics Autoresearch
Use this skill when the user wants a data-informed growth loop for landing pages, onboarding, pricing, CTAs, signup, checkout, activation, or other experiment candidates.
This skill is based on:
- Autoresearch Growth template: https://github.com/Agent-Analytics/autoresearch-growth
- Agent Analytics: https://agentanalytics.sh
- Regular Agent Analytics skill: https://github.com/Agent-Analytics/agent-analytics-skill/tree/main/skills/agent-analytics
Use the regular agent-analytics skill for general setup, tracking installation, ad hoc reporting, and normal experiment operations. Use this skill for structured variant generation and judging from a project brief plus analytics data.
Core Rule
Do not edit production copy, product code, or live experiment setup while running the loop unless the user explicitly asks. Produce reviewable artifacts first.
Default mode is review-only: generate variants, log rounds, and write final_variants.md.
After explicit human approval, continue into the outer experiment loop when requested: implement the approved variant or variants, create the experiment, run it, measure it with Agent Analytics or another analytics source, save the results as the next snapshot, and start the next autoresearch run from evidence.
Inputs
The loop needs:
- target surface
- current control copy
- product truth
- audience
- primary metric
- proxy metric
- guardrails
- analytics snapshot or data brief
- drift constraints
Agent Analytics is preferred, but not required. Accept any evidence source: Agent Analytics CLI/API, PostHog, GA4, Mixpanel, SQL, CSV exports, product logs, dashboard screenshots summarized by the user, or hand-written notes.
When Agent Analytics is the evidence source, use project context as the self-improving product memory for the loop. Read context get <project> before collecting a snapshot, fold project_context into the product truth and metric definitions, and keep activation/event meaning separate per project or domain. After a human correction, scanner result, completed experiment, or repeated measured finding, update context only with durable product truth. Save activation definitions, event meanings, stable goals, and confirmed interpretations; skip weekly numbers, temporary spikes, pasted reports, PII, and unconfirmed guesses.
Quick Start
If the user already has a repo or run folder, work there. Otherwise initialize a run:
bash <skill_dir>/scripts/init_autoresearch_run.sh homepage-signup
Then fill brief.md, collect or paste data, and run the loop:
Read brief.md and run the autoresearch growth loop. Use the latest data snapshot. Run 5 rounds. Append one row per round to results.tsv and write final_variants.md with two distinct variants for review.
When using Agent Analytics, collect a snapshot:
bash <skill_dir>/scripts/collect_agent_analytics_snapshot.sh my-site signup cta_click
If <skill_dir> is not obvious in the runtime, read the script from this skill's scripts/ folder and run an equivalent local command.
References
Load these files only when needed:
references/program.md- exact loop instructions.references/brief-template.md- project brief template.references/final-variants-template.md- final output template.references/results-header.txt- exactresults.tsvheader.
Loop Shape
Inner Autoresearch Loop
- Define the surface, control, audience, product truth, metric, proxy, and guardrails.
- Collect or read a dated analytics snapshot.
- Summarize useful signals and data limitations.
- Generate candidate A.
- Critique A harshly for genericness, drift, unsupported claims, weak conversion intent, and competitor-sayable language.
- Write candidate B from the critique.
- Synthesize AB from the strongest parts of A and B.
- Blind-rank A, B, and AB with Borda scoring.
- Append one TSV-safe row to
results.tsv. - Repeat several rounds.
- Write
final_variants.mdwith two distinct variants and the recommended experiment shape.
Outer Experiment Loop
Only run this phase when the user explicitly approves implementation or experiment setup.
- Implement the approved variant or variants in the target product surface.
- Create the experiment with a control and the approved candidate variants.
- Verify tracking for the primary metric, proxy metric, and guardrails.
- Let the experiment collect real behavior for the requested window.
- Pull experiment results, screenshots or changed-copy notes, funnel movement, guardrails, and data limitations into a new snapshot.
- Start the next inner autoresearch loop from that measured evidence.
The outer loop prevents the LLM panel from becoming the final judge. LLMs generate and criticize, humans approve risk, and users decide what worked.
Agent Analytics Snapshot
Use the official CLI when collecting live Agent Analytics data:
npx --yes @agent-analytics/cli@0.5.20 insights "$PROJECT_SLUG" --period 7d
npx --yes @agent-analytics/cli@0.5.20 pages "$PROJECT_SLUG" --since 7d
npx --yes @agent-analytics/cli@0.5.20 funnel "$PROJECT_SLUG" --steps "page_view,$PROXY_EVENT,$PRIMARY_EVENT" --since 7d
npx --yes @agent-analytics/cli@0.5.20 events "$PROJECT_SLUG" --event "$PROXY_EVENT" --days 7 --limit 50
npx --yes @agent-analytics/cli@0.5.20 events "$PROJECT_SLUG" --event "$PRIMARY_EVENT" --days 7 --limit 50
npx --yes @agent-analytics/cli@0.5.20 experiments list "$PROJECT_SLUG"
If login is needed, prefer the regular agent-analytics skill's browser approval or detached login guidance.
Before interpreting the snapshot, also read the compact project memory:
npx --yes @agent-analytics/cli@0.5.20 context get "$PROJECT_SLUG"
If the autoresearch run reveals durable product truth that should guide future analytics, use the regular agent-analytics skill's project context workflow to read the existing context, merge the compact update, and write it back. Do not store raw round notes or time-bound metric values as project context.
Scoring
Use Borda scoring:
- first place: 2 points
- second place: 1 point
- third place: 0 points
Judge by:
- specificity to the product
- clarity for the target audience
- likely primary-event intent
- preservation of product truth
- low competitor-sayable language
- fit with analytics data
- respect for guardrails
Output
final_variants.md must include:
- candidate_1
- candidate_2
- exact changed copy
- rationale
- risks
- recommended experiment name
- experiment shape
- data limitations
- clear note that the experiment has not been wired yet
Only create or wire an experiment after explicit human approval.
Comments
Loading comments...
