agent-analytics-autoresearch

Run an autoresearch-style growth loop for landing pages, onboarding, pricing, and experiment candidates. Collect or read analytics snapshots, preserve product truth, generate/critique/synthesize variants, blind-rank with Borda scoring, and output two review-ready A/B test variants. Works with any analytics data; best with Agent Analytics CLI/API.

Safety Notice

This listing is from the official public ClawHub registry. Review SKILL.md and referenced scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "agent-analytics-autoresearch" with this command: npx skills add dannyshmueli/agent-analytics-autoresearch

Agent Analytics Autoresearch

Use this skill when the user wants a data-informed growth loop for landing pages, onboarding, pricing, CTAs, signup, checkout, activation, or other experiment candidates.

This skill is based on:

Use the regular agent-analytics skill for general setup, tracking installation, ad hoc reporting, and normal experiment operations. Use this skill for structured variant generation and judging from a project brief plus analytics data.

Core Rule

Do not edit production copy, product code, or live experiment setup while running the loop unless the user explicitly asks. Produce reviewable artifacts first.

Default mode is review-only: generate variants, log rounds, and write final_variants.md.

After explicit human approval, continue into the outer experiment loop when requested: implement the approved variant or variants, create the experiment, run it, measure it with Agent Analytics or another analytics source, save the results as the next snapshot, and start the next autoresearch run from evidence.

Inputs

The loop needs:

  • target surface
  • current control copy
  • product truth
  • audience
  • primary metric
  • proxy metric
  • guardrails
  • analytics snapshot or data brief
  • drift constraints

Agent Analytics is preferred, but not required. Accept any evidence source: Agent Analytics CLI/API, PostHog, GA4, Mixpanel, SQL, CSV exports, product logs, dashboard screenshots summarized by the user, or hand-written notes.

When Agent Analytics is the evidence source, use project context as the self-improving product memory for the loop. Read context get <project> before collecting a snapshot, fold project_context into the product truth and metric definitions, and keep activation/event meaning separate per project or domain. After a human correction, scanner result, completed experiment, or repeated measured finding, update context only with durable product truth. Save activation definitions, event meanings, stable goals, and confirmed interpretations; skip weekly numbers, temporary spikes, pasted reports, PII, and unconfirmed guesses.

Quick Start

If the user already has a repo or run folder, work there. Otherwise initialize a run:

bash <skill_dir>/scripts/init_autoresearch_run.sh homepage-signup

Then fill brief.md, collect or paste data, and run the loop:

Read brief.md and run the autoresearch growth loop. Use the latest data snapshot. Run 5 rounds. Append one row per round to results.tsv and write final_variants.md with two distinct variants for review.

When using Agent Analytics, collect a snapshot:

bash <skill_dir>/scripts/collect_agent_analytics_snapshot.sh my-site signup cta_click

If <skill_dir> is not obvious in the runtime, read the script from this skill's scripts/ folder and run an equivalent local command.

References

Load these files only when needed:

  • references/program.md - exact loop instructions.
  • references/brief-template.md - project brief template.
  • references/final-variants-template.md - final output template.
  • references/results-header.txt - exact results.tsv header.

Loop Shape

Inner Autoresearch Loop

  1. Define the surface, control, audience, product truth, metric, proxy, and guardrails.
  2. Collect or read a dated analytics snapshot.
  3. Summarize useful signals and data limitations.
  4. Generate candidate A.
  5. Critique A harshly for genericness, drift, unsupported claims, weak conversion intent, and competitor-sayable language.
  6. Write candidate B from the critique.
  7. Synthesize AB from the strongest parts of A and B.
  8. Blind-rank A, B, and AB with Borda scoring.
  9. Append one TSV-safe row to results.tsv.
  10. Repeat several rounds.
  11. Write final_variants.md with two distinct variants and the recommended experiment shape.

Outer Experiment Loop

Only run this phase when the user explicitly approves implementation or experiment setup.

  1. Implement the approved variant or variants in the target product surface.
  2. Create the experiment with a control and the approved candidate variants.
  3. Verify tracking for the primary metric, proxy metric, and guardrails.
  4. Let the experiment collect real behavior for the requested window.
  5. Pull experiment results, screenshots or changed-copy notes, funnel movement, guardrails, and data limitations into a new snapshot.
  6. Start the next inner autoresearch loop from that measured evidence.

The outer loop prevents the LLM panel from becoming the final judge. LLMs generate and criticize, humans approve risk, and users decide what worked.

Agent Analytics Snapshot

Use the official CLI when collecting live Agent Analytics data:

npx --yes @agent-analytics/cli@0.5.20 insights "$PROJECT_SLUG" --period 7d
npx --yes @agent-analytics/cli@0.5.20 pages "$PROJECT_SLUG" --since 7d
npx --yes @agent-analytics/cli@0.5.20 funnel "$PROJECT_SLUG" --steps "page_view,$PROXY_EVENT,$PRIMARY_EVENT" --since 7d
npx --yes @agent-analytics/cli@0.5.20 events "$PROJECT_SLUG" --event "$PROXY_EVENT" --days 7 --limit 50
npx --yes @agent-analytics/cli@0.5.20 events "$PROJECT_SLUG" --event "$PRIMARY_EVENT" --days 7 --limit 50
npx --yes @agent-analytics/cli@0.5.20 experiments list "$PROJECT_SLUG"

If login is needed, prefer the regular agent-analytics skill's browser approval or detached login guidance.

Before interpreting the snapshot, also read the compact project memory:

npx --yes @agent-analytics/cli@0.5.20 context get "$PROJECT_SLUG"

If the autoresearch run reveals durable product truth that should guide future analytics, use the regular agent-analytics skill's project context workflow to read the existing context, merge the compact update, and write it back. Do not store raw round notes or time-bound metric values as project context.

Scoring

Use Borda scoring:

  • first place: 2 points
  • second place: 1 point
  • third place: 0 points

Judge by:

  • specificity to the product
  • clarity for the target audience
  • likely primary-event intent
  • preservation of product truth
  • low competitor-sayable language
  • fit with analytics data
  • respect for guardrails

Output

final_variants.md must include:

  • candidate_1
  • candidate_2
  • exact changed copy
  • rationale
  • risks
  • recommended experiment name
  • experiment shape
  • data limitations
  • clear note that the experiment has not been wired yet

Only create or wire an experiment after explicit human approval.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

Revenue Agent Lite

Free daily growth assistant for online businesses. Track basic revenue metrics, get daily action priorities, and monitor progress toward goals.

Registry SourceRecently Updated
1060Profile unavailable
General

Skill Analytics

Monitor ClawHub skill performance with file-based state tracking. Fetches public stats via web_fetch, tracks recommendations and their outcomes, avoids repet...

Registry SourceRecently Updated
2051Profile unavailable
Automation

Marketing OS

AI Agent 营销操作系统 — 包含 Virtual CMO(战略大脑)和 Marketing Operator(执行引擎),提供市场分析、策略制定、Campaign 规划与执行追踪全链路能力。

Registry SourceRecently Updated
1590Profile unavailable
Web3

Growth Engineering Mastery

Design, execute, and measure growth systems — from North Star definition through viral loops, experimentation, and scaling. Complete AARRR+ framework with te...

Registry SourceRecently Updated
4260Profile unavailable