A working library of frameworks, templates, rubrics, and Claude skills for product managers. Built around the artifacts that actually go up the chain: strategy memos, PRDs, launch checklists, metrics dashboards, kill / continue decisions. Not for case-interview prep.
The material covers the full lifecycle, from framing a problem and running customer discovery through launch readiness and post-launch metrics. Two design choices distinguish it from a generic PM checklist library.
Discovery is treated as technique, not philosophy. The repo includes a Mom-Test customer-interview coach (with the rules behind why "would you use this?" produces fluff and "when's the last time that happened?" produces signal), a Lean-Startup-grounded value-hypothesis stress-tester, and substantive framework docs underneath each. The principles are tactical: what to ask, what to ignore, what counts as evidence.
Every primary skill chains into an adversarial second pass. The pm-red-team skill takes the output of any other skill (or any external AI critique, or your own draft) and re-reviews it under a different lens. Most artifacts that "look fine" do so because the first reviewer applied the comfortable lens. The second pass picks a different one. The point matches a working principle from my chief-of-staff setup: critically evaluate AI output, don't defer to it.
The repo organizes around one question: what does a good PM actually do at each stage of the work, and how do we know? The five evaluation criteria below answer it. Everything else (frameworks, decision-making docs, cross-functional traps, rubrics, templates, skills) feeds those five.
If you forked or cloned this repo, do these steps in order. Each one is concrete — you should be able to copy-paste and run.
-
Fork the repo and re-clone your fork. On GitHub, fork
kalyvask/pm-evaluation-framework, thengit clone git@github.com:<you>/pm-evaluation-framework.git. EditLICENSEto your name and update the intro of this README to reflect your perspective. -
Install the Claude skills so they work in every project, not just this repo. The nine skills under
.claude/skills/auto-load when Claude Code is opened inside this repo. To make them available everywhere, copy them into your user-level skills directory:mkdir -p ~/.claude/skills cp -r .claude/skills/* ~/.claude/skills/
Restart Claude Code. Run
/skillsto confirm the ninepm-*skills are listed. If you only want a subset, copy individual skill folders. -
Add your own artifact templates. Put new templates in
templates/alongsideprd-template.md,decision-memo.md,decision-log.md,launch-criteria.md, andblameless-postmortem.md. Match the existing voice (imperative, section-headed, no jargon) so the skills can find and reuse them. -
Customize rubric weights and criteria for your team's bar. The three rubrics in
rubrics/—pm-evaluation-rubric.md,strategy-memo-rubric.md,product-review-rubric.md— are deliberately editable. Re-weight criteria, add team-specific ones (e.g. "addresses regulated-data path"), or replace the five top-level questions in the The five evaluation criteria section below. The skills cite these files by path, so they pick up your edits automatically. -
Extend or override skills with your own voice and style. To add a new skill: copy
SKILL.md.tmplinto a new folder under.claude/skills/<your-skill>/SKILL.mdand fill in the front-matter description (this is what Claude pattern-matches against). To change the voice of an existing skill: edit itsSKILL.mddirectly — tighten the prose, add your own anti-patterns, or point it at frameworks you added in steps above. -
Keep your private material local — it's already gitignored.
.gitignorealready excludesdrafts/,local/,private/,outputs/,.cache/,*.draft.md,*.local.md, CVs, PDFs, DOCX, and.env*. Use these for unredacted notes, in-flight memo drafts, your CV, and any private artifact you want a skill to reason over without committing it. Sanity-check withgit statusbefore every commit. -
Smoke-test that personalization worked. From inside Claude Code, ask: "Use the pm-evaluator skill to grade this draft memo: [paste]" — confirm the skill triggers and cites your edited rubric. Then ask: "Use pm-framework-selector for [your situation]" — confirm the skill routes you to the right framework.
| If you want to... | Start here |
|---|---|
| Pick the right framework for the decision in front of you | frameworks/00-overview.md |
| Sharpen how you frame a problem before jumping to solutions | decision-making/problem-framing.md |
| Think clearly about reversibility before committing | decision-making/risk-and-reversibility.md |
| Decide between user research methods (personas vs. JTBD vs. Kano) | decision-making/research-methods.md |
| Run a customer-discovery interview that actually produces signal | decision-making/customer-interviews.md |
| Stress-test the value hypothesis behind a product before committing | decision-making/value-hypothesis.md |
| Argue defensibility honestly — which moat, with what evidence | decision-making/competitive-moat.md |
| Adversarially re-review an existing critique before acting on it | pm-red-team skill |
| Prioritize a backlog | decision-making/prioritization.md |
| Define what to measure | decision-making/metrics.md |
| Work better with engineering | cross-functional/engineering-partnership.md |
| Run a launch where some failure is expected | cross-functional/failure-management.md |
| Understand how AI is changing the PM role | ai-and-pm.md |
| Evaluate a PM's reasoning on a case | rubrics/pm-evaluation-rubric.md |
| Evaluate a strategy memo before it goes up the chain | rubrics/strategy-memo-rubric.md |
| Show up well in a product / exec review | rubrics/product-review-rubric.md |
| Use a PRD / decision-memo / decision-log / postmortem template | templates/ |
The repo ships with nine Claude Code skills under .claude/skills/. They use the frameworks in this repo as their substrate, and they're organized by where you are in the product lifecycle.
| Stage | Skill | Use when |
|---|---|---|
| Frame | pm-framework-selector |
You have a decision in front of you and don't know which framework to reach for |
| Frame / decide | pm-decision-coach |
You want to be walked through a decision: framing → research → prioritization → risk |
| Discover | pm-customer-interview-coach |
You're prepping or debriefing customer interviews and want them stress-tested against the Mom Test rules |
| Discover | pm-value-hypothesis-tester |
You're about to launch, fundraise, or scale, and you want the underlying what / who / how bet pressure-tested |
| Build | pm-prd-drafter |
You're starting a PRD, have a draft to critique, or are stuck on problem / success / scope |
| Launch | pm-launch-reviewer |
You have a launch coming up and want a gate-by-gate pre-flight against a real readiness bar |
| Measure | pm-metrics-critic |
You're locking success criteria, debating a North Star, or staring at a dashboard that "looks fine" |
| Review | pm-evaluator |
You want a strategy memo, PRD, or analysis graded against the five-criterion rubric before it goes up |
| Challenge | pm-red-team |
You have an existing critique (from another skill, an external AI, or yourself) and want it adversarially re-reviewed before you act on it |
To use them locally, drop the repo in a directory Claude Code can see — the skills will appear automatically. New skills follow the format in SKILL.md.tmpl.
Every framework in this repo ladders up to five questions that travel with you across any product decision:
- Did you identify the right problem before jumping to solutions?
- Did you apply a framework to remove bias and surface stakeholder perspectives?
- Did you account for bundle effects and the full user journey — including procurement, admin, and downstream teams?
- Did you define what success and failure look like upfront?
- Did you prototype or validate before committing real resources?
The full rubric is in rubrics/pm-evaluation-rubric.md.
MIT. Use it, fork it, ship better products.