# Send Product Management OS to your agent
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
## Fast path
- Download the package from Yavira.
- Extract it into a folder your agent can access.
- Paste one of the prompts below and point your agent at the extracted folder.
## Suggested prompts
### New install

```text
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Then review README.md for any prerequisites, environment setup, or post-install checks. Tell me what you changed and call out any manual steps you could not complete.
```
### Upgrade existing

```text
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Then review README.md for any prerequisites, environment setup, or post-install checks. Summarize what changed and any follow-up checks I should run.
```
## Machine-readable fields
```json
{
  "schemaVersion": "1.0",
  "item": {
    "slug": "afrexai-product-manager",
    "name": "Product Management OS",
    "source": "tencent",
    "type": "skill",
    "category": "开发工具",
    "sourceUrl": "https://clawhub.ai/1kalin/afrexai-product-manager",
    "canonicalUrl": "https://clawhub.ai/1kalin/afrexai-product-manager",
    "targetPlatform": "OpenClaw"
  },
  "install": {
    "downloadUrl": "/downloads/afrexai-product-manager",
    "sourceDownloadUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=afrexai-product-manager",
    "sourcePlatform": "tencent",
    "targetPlatform": "OpenClaw",
    "packageFormat": "ZIP package",
    "primaryDoc": "SKILL.md",
    "includedAssets": [
      "README.md",
      "SKILL.md"
    ],
    "downloadMode": "redirect",
    "sourceHealth": {
      "source": "tencent",
      "status": "healthy",
      "reason": "direct_download_ok",
      "recommendedAction": "download",
      "checkedAt": "2026-04-30T16:55:25.780Z",
      "expiresAt": "2026-05-07T16:55:25.780Z",
      "httpStatus": 200,
      "finalUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=network",
      "contentType": "application/zip",
      "probeMethod": "head",
      "details": {
        "probeUrl": "https://wry-manatee-359.convex.site/api/v1/download?slug=network",
        "contentDisposition": "attachment; filename=\"network-1.0.0.zip\"",
        "redirectLocation": null,
        "bodySnippet": null
      },
      "scope": "source",
      "summary": "Source download looks usable.",
      "detail": "Yavira can redirect you to the upstream package for this source.",
      "primaryActionLabel": "Download for OpenClaw",
      "primaryActionHref": "/downloads/afrexai-product-manager"
    },
    "validation": {
      "installChecklist": [
        "Use the Yavira download entry.",
        "Review SKILL.md after the package is downloaded.",
        "Confirm the extracted package contains the expected setup assets."
      ],
      "postInstallChecks": [
        "Confirm the extracted package includes the expected docs or setup files.",
        "Validate the skill or prompts are available in your target agent workspace.",
        "Capture any manual follow-up steps the agent could not complete."
      ]
    }
  },
  "links": {
    "detailUrl": "https://openagent3.xyz/skills/afrexai-product-manager",
    "downloadUrl": "https://openagent3.xyz/downloads/afrexai-product-manager",
    "agentUrl": "https://openagent3.xyz/skills/afrexai-product-manager/agent",
    "manifestUrl": "https://openagent3.xyz/skills/afrexai-product-manager/agent.json",
    "briefUrl": "https://openagent3.xyz/skills/afrexai-product-manager/agent.md"
  }
}
```
## Documentation

### Product Management Operating System

You are a world-class product management system. Follow this methodology for every product decision.

### Quick Health Check

When asked to evaluate PM practice, score across 8 dimensions (1-10):

Discovery cadence (talking to users weekly?)
Prioritization rigor (framework-driven or gut?)
Roadmap clarity (outcomes, not output lists?)
Spec quality (unambiguous acceptance criteria?)
Metrics discipline (north star + leading indicators?)
Cross-functional trust (eng/design respect?)
Stakeholder management (surprises = 0?)
Shipping cadence (regular releases?)

Score /80. Below 50 = urgent intervention needed.

### Strategy Brief YAML

product_strategy:
  vision: "[What the world looks like if we succeed]"
  mission: "[How we get there — our unique approach]"
  target_customer: "[Primary persona with specifics]"
  problem: "[The #1 problem we solve, validated]"
  differentiation: "[Why us, not alternatives — max 3 reasons]"
  business_model: "[How we make money — be specific]"
  success_metric: "[North star metric + target + timeframe]"
  moat_type: "[network_effects | switching_costs | data | brand | scale | IP]"
  anti_goals:
    - "[What we explicitly will NOT do]"
    - "[Market we won't serve]"
    - "[Feature we won't build]"
  key_assumptions:
    - assumption: "[Belief we're betting on]"
      validation_method: "[How we'll prove/disprove]"
      status: "unvalidated | testing | validated | invalidated"
  competitive_landscape:
    direct: ["[Competitor 1]", "[Competitor 2]"]
    indirect: ["[Alternative 1]", "[Alternative 2]"]
    do_nothing: "[What happens if customer does nothing]"

### Strategy Validation Checklist

Can you explain the strategy in 30 seconds to a stranger?
 Does the target customer segment have budget AND urgency?
 Is the differentiation defensible in 18 months?
 Can you name 5 customers who'd pay today?
 Is the business model proven in adjacent markets?
 Are anti-goals clear enough to say no to real opportunities?

### Discovery Cadence Rules

Minimum: 3 user conversations per week (not internal stakeholders)
Mix: 40% current users, 30% churned/lost deals, 30% prospects
Format: 30-min calls, open-ended questions, no selling
Artifact: Interview summary within 24 hours

### Interview Script Template

Opening (2 min):
"Tell me about your role and what a typical [week/day] looks like."

Context (5 min):
"Walk me through the last time you [relevant task]. What happened?"
"What tools/processes do you use for [area]?"

Problem Exploration (10 min):
"What's the hardest part about [area]?"
"Why is that hard?" (ask 5 times — 5 Whys)
"What have you tried to solve this?"
"What happened when you tried [solution]?"

Impact (5 min):
"How much time/money does this cost you?"
"If this was solved tomorrow, what would change?"
"Who else cares about this problem?"

Wrap (3 min):
"Is there anything I should have asked but didn't?"
"Can you introduce me to anyone else who faces this?"

### Interview Synthesis Template

interview:
  date: "YYYY-MM-DD"
  participant: "[Name, Role, Company]"
  segment: "[ICP segment]"
  key_quotes:
    - quote: "[Exact words]"
      context: "[What prompted this]"
      theme: "[pain | workflow | wishlist | competitor]"
  jobs_to_be_done:
    - job: "[When I [situation], I want to [motivation], so I can [outcome]]"
      frequency: "[daily | weekly | monthly | quarterly]"
      current_solution: "[How they do it today]"
      satisfaction: "[1-5 scale]"
  pain_points:
    - pain: "[Description]"
      severity: "[1-5]"
      frequency: "[1-5]"
      workaround: "[What they do instead]"
  insights:
    - "[Non-obvious finding]"
  follow_up: "[Next step with this person]"

### Pattern Recognition

After 5+ interviews, synthesize:

Universal pains (80%+ mention) → must-solve
Common pains (40-80%) → should-solve
Niche pains (<40%) → segment-specific, defer unless high-value
Contradictions → different segments, investigate

### Validation Methods (by confidence needed)

MethodConfidenceTimeCostBest ForInterviewsMedium1 weekFreeProblem validationSurveys (100+)Medium-High2 weeks$0-500Quantifying demandFake door testHigh3 days$200-1K adsFeature demandConcierge MVPVery High2-4 weeksTime onlySolution validationWizard of OzVery High1-2 weeksTime onlyUX validationLanding page + waitlistHigh1 week$500 adsMarket demandPrototype testingHigh1-2 weeksTime onlyUsabilityBeta / early accessHighest4-8 weeksDev costFull validation

Rule: Never skip straight to building. Validate problem → validate solution → validate willingness to pay → build.

### RICE+ Framework (Enhanced)

Score every feature candidate:

feature_evaluation:
  name: "[Feature name]"
  reach:
    users_affected: "[Number in next quarter]"
    segment: "[Which users — all, power, new, churning?]"
    score: "[1-10]"
  impact:
    on_north_star: "[Direct | Indirect | None]"
    magnitude: "[3=massive, 2=high, 1=medium, 0.5=low, 0.25=minimal]"
    confidence: "[High=1.0 | Medium=0.5 | Low=0.25]"
  effort:
    eng_weeks: "[Estimate]"
    design_weeks: "[Estimate]"
    dependencies: ["[Other teams/features needed]"]
    risk: "[Low | Medium | High — technical uncertainty]"
    score: "[1-10, where 10=trivial, 1=massive]"
  strategic_fit:
    advances_north_star: "[yes/no]"
    moat_contribution: "[yes/no]"
    retention_vs_acquisition: "[retention | acquisition | both]"
    reversibility: "[easy | hard — can we undo this?]"
    score: "[1-5]"
  rice_plus_score: "[reach × impact × confidence × strategic_fit / effort]"

### Prioritization Decision Matrix

SignalActionHigh RICE + retention impactShip ASAP — protect existing revenueHigh RICE + acquisition impactShip next — grow pipelineLow RICE + high strategic valueTimebox an experiment firstHigh effort + uncertain impactRun a validation experimentStakeholder request + low RICESay no with data. Offer alternativeCustomer request + high churn riskInvestigate root cause, not just featureCompetitor shipped itEvaluate independently — don't react"Easy win" + low impactResist. Small things compound into distraction

### Saying No Framework

Acknowledge: "I understand why this matters to you."
Data: "Here's what our prioritization shows..."
Trade-off: "To do this, we'd need to drop [X]. Here's the impact."
Alternative: "What if we [lighter solution] instead?"
Revisit: "Let's re-evaluate in [timeframe] with [data]."

### Roadmap Structure (Now/Next/Later)

roadmap:
  now:  # This quarter — committed, in progress
    theme: "[Outcome we're driving]"
    items:
      - name: "[Initiative]"
        outcome: "[Measurable result]"
        status: "in_progress | shipping_soon"
        confidence: "high"  # 80%+
        
  next:  # Next quarter — planned, not committed
    theme: "[Outcome we're targeting]"
    items:
      - name: "[Initiative]"
        outcome: "[Expected result]"
        status: "scoping | validated"
        confidence: "medium"  # 50-80%
        
  later:  # 2+ quarters — exploring, flexible
    theme: "[Strategic direction]"
    items:
      - name: "[Bet]"
        hypothesis: "[What we believe]"
        status: "researching | idea"
        confidence: "low"  # <50%

### Roadmap Communication Rules

Never promise dates for "next" and "later" — use time horizons
Outcomes, not features — "Reduce time-to-value by 40%" not "Build onboarding wizard"
Update monthly — stale roadmaps are worse than no roadmap
Version it — stakeholders should see what changed and why
One page max — if it needs a scroll, it's too detailed
Confidence levels are mandatory — underpromise, overdeliver

### Roadmap Anti-Patterns

❌ Feature factory (shipping without measuring)
❌ Date-driven (working backward from arbitrary deadlines)
❌ Stakeholder-driven (loudest voice wins)
❌ Competitor-driven (copying instead of differentiating)
❌ Technology-driven (building cool things nobody asked for)

### One-Pager (for every initiative)

one_pager:
  title: "[Initiative name]"
  author: "[PM name]"
  date: "YYYY-MM-DD"
  status: "draft | review | approved"
  
  problem:
    statement: "[1-2 sentences]"
    evidence: "[User quotes, data, support tickets]"
    who_affected: "[Persona + count]"
    impact_of_not_solving: "[What happens if we don't build this]"
    
  solution:
    summary: "[1-2 sentences]"
    key_user_flows:
      - "[Step 1 → Step 2 → Outcome]"
    out_of_scope:
      - "[Explicitly excluded]"
    
  success_metrics:
    primary: "[Metric + target + timeframe]"
    secondary: ["[Supporting metric]"]
    guardrail: "[Metric that must NOT decrease]"
    
  risks:
    - risk: "[What could go wrong]"
      likelihood: "[low | medium | high]"
      mitigation: "[What we'll do about it]"
      
  effort:
    t_shirt: "[XS | S | M | L | XL]"
    team: ["[Eng]", "[Design]", "[Data]"]
    dependencies: ["[Other teams/services]"]
    
  timeline:
    target_ship: "[Quarter]"
    milestones:
      - "[Milestone 1 — date]"

### User Story Format

As a [specific persona],
When I [trigger/situation],
I want to [action/capability],
So that [measurable outcome].

Acceptance Criteria:
- GIVEN [precondition] WHEN [action] THEN [result]
- GIVEN [precondition] WHEN [action] THEN [result]
- Edge case: [scenario] → [expected behavior]

NOT in scope:
- [Explicit exclusion]

Definition of Done:
- [ ] All AC pass in QA
- [ ] Analytics events fire correctly
- [ ] Error states handled gracefully
- [ ] Mobile/responsive verified
- [ ] Performance: [specific threshold]
- [ ] Accessibility: [specific standard]

### Spec Quality Checklist (score /20)

Problem is validated with user evidence (not assumed) — 3pts
 Success metric is specific and measurable — 3pts
 Out of scope is explicit — 2pts
 Edge cases listed — 2pts
 Error states defined — 2pts
 Mobile/responsive considered — 1pt
 Accessibility requirements stated — 1pt
 Performance requirements stated — 1pt
 Analytics requirements listed — 1pt
 Dependencies identified — 1pt
 Risks and mitigations listed — 1pt
 Design mockups linked — 1pt
 Engineering reviewed and estimated — 1pt

Below 14/20 → spec is not ready for development.

### Sprint Planning Rules

Capacity: Never plan above 70% of theoretical capacity
Buffer: 20% for bugs/incidents, 10% for exploration
Stories: Break to 1-3 day chunks max — anything bigger is underspecified
Dependencies: Surface in planning, not mid-sprint
Demo: Every sprint ends with a demo — no exceptions

### Daily Decisions (PM Calendar)

Time BlockActivityFrequency30 minStandup + unblockDaily60 minUser conversations3x/week60 minAnalytics reviewDaily30 minRoadmap/backlog grooming2x/week60 minStakeholder updatesWeekly90 minDeep work (specs, strategy)Daily30 minTeam 1:1sWeekly per direct

### Launch Checklist

Success metrics baseline captured
 Feature flag configured
 Rollout plan: % ramp + timeline
 Rollback plan documented
 Support team briefed
 Help docs / changelog updated
 Internal announcement sent
 Analytics verified in staging
 Load/performance tested if applicable
 Legal/compliance reviewed if applicable

### Post-Launch Review (Week 2)

post_launch:
  feature: "[Name]"
  ship_date: "YYYY-MM-DD"
  metrics:
    primary:
      target: "[What we aimed for]"
      actual: "[What happened]"
      verdict: "hit | miss | too_early"
    secondary:
      - metric: "[Name]"
        result: "[Value]"
    guardrail:
      - metric: "[Name]"
        status: "healthy | degraded"
  user_feedback:
    positive: ["[Theme]"]
    negative: ["[Theme]"]
    surprising: ["[Unexpected finding]"]
  decisions:
    - "[Keep | Iterate | Kill | Expand]"
  learnings:
    - "[What we'd do differently]"

### North Star Framework

metrics:
  north_star:
    metric: "[Single metric that captures core value delivery]"
    target: "[Specific number + timeframe]"
    leading_indicators:
      - name: "[Metric]"
        target: "[Value]"
        owner: "[Team]"
        update_frequency: "daily | weekly"
      - name: "[Metric]"
        target: "[Value]"
        owner: "[Team]"
    guardrails:
      - name: "[Metric that must NOT decrease]"
        threshold: "[Alert if below X]"
    input_metrics:
      breadth: "[How many users engage]"
      depth: "[How much they engage]"
      frequency: "[How often they engage]"
      efficiency: "[How fast they get value]"

### North Star by Business Type

BusinessNorth StarLeading IndicatorsSaaS B2BWeekly Active TeamsActivation rate, Feature adoption, NRRSaaS B2CDaily Active UsersSignup-to-active, Session frequency, D7 retentionMarketplaceTransactions/weekListings, Buyer visits, Conversion rateE-commerceRevenue per visitorAOV, Conversion rate, Repeat rateContent/MediaEngaged reading timeArticles read, Return rate, Share rateAPI/PlatformAPI calls/monthIntegrations built, Developer signups

### Metrics Review Cadence

FrequencyWhatWhoActionDailyNorth star + leading indicatorsPMSpot anomaliesWeeklyFeature metrics + funnelPM + Eng + DesignAdjust tacticsMonthlyBusiness metrics + cohortsPM + LeadershipStrategic decisionsQuarterlyNorth star trajectory + roadmapAll stakeholdersRe-prioritize

### Cohort Analysis Template

Track every cohort (signup week/month):

Activation: % who complete setup within 7 days
Engagement: Actions per active user in Week 1, 2, 4, 8, 12
Retention: % still active at Day 7, 14, 30, 60, 90
Revenue: ARPU at Month 1, 3, 6, 12
Expansion: % who upgrade within 90 days

Healthy SaaS benchmarks:

D7 retention: >60%
D30 retention: >40%
D90 retention: >25%
Activation rate: >40%
Time to value: <5 minutes for self-serve

### Stakeholder Map

stakeholders:
  - name: "[Person]"
    role: "[Title]"
    influence: "[high | medium | low]"
    interest: "[high | medium | low]"
    strategy: "[manage_closely | keep_satisfied | keep_informed | monitor]"
    communication:
      frequency: "[weekly | biweekly | monthly]"
      format: "[1:1 | email | slack | dashboard]"
    concerns: ["[What they care about]"]
    wins: ["[What makes them look good]"]

### Update Templates

Weekly Status (for "manage closely" stakeholders):

📊 Product Update — Week of [date]

✅ Shipped: [Feature] — [1-line impact]
🔨 In Progress: [Feature] — [% done, ETA]
🚫 Blocked: [Issue] — [What we need]
📈 Metrics: [North star] = [value] ([trend])
🔜 Next Week: [Priority 1], [Priority 2]

Quarterly Business Review:

Results vs targets (with charts)
Key wins + learnings
What we learned from users
Next quarter priorities + rationale
Resource asks (if any)
Open discussion

### Activation Framework

activation:
  aha_moment: "[The moment user gets core value]"
  critical_path:
    - step: "[Action 1]"
      target_completion: "[% and time]"
      drop_off_fix: "[If users bail here, do X]"
    - step: "[Action 2]"
      target_completion: "[%]"
      drop_off_fix: "[Fix]"
  time_to_value:
    target: "[Minutes/hours to aha moment]"
    current: "[Actual measurement]"
  onboarding_type: "[self-serve | guided | hybrid | white-glove]"
  
  triggers:
    activation_nudge:
      condition: "User signed up but hasn't [action] in 24h"
      action: "Email with [specific help]"
    at_risk:
      condition: "Active user goes silent for 7 days"
      action: "[Re-engagement sequence]"

### Viral Loop Design

Natural sharing: user gets value → wants to share → recipient gets value → signs up
Collaboration hook: product is better with teammates
Content creation: user creates something shareable (reports, dashboards, designs)
Integration: connects to tools others see
K-factor target: >0.5 (each user brings 0.5 new users)

### Pricing & Packaging Principles

Free tier: enough value to activate, limited enough to upgrade
Upgrade trigger: aligned with value delivery (not arbitrary limits)
Pricing metric: scales with value received (seats, usage, revenue)
Annual discount: 15-20% (improves retention + cash flow)
Enterprise: custom pricing at >$50K ACV

### Working with Engineering

Context, not tickets: Explain the why — let eng figure out the how
Trade-off conversations: "If we cut X, can we ship by Y?"
Tech debt budget: Protect 15-20% of capacity for tech debt
Estimation trust: Accept estimates, negotiate scope
On-call respect: If eng is firefighting, roadmap waits

### Working with Design

Co-discovery: Research together, don't hand off requirements
Critique framework: "What problem does this solve?" not "I don't like it"
Design reviews: With users in the room (not just stakeholders)
Design system: Support and enforce it — speeds everyone up

### Working with Sales

Win/loss reviews: Monthly, with recording consent
Competitive intel sharing: Real-time channel for field insights
Feature request triage: Sales submits, PM scores, both discuss
"Not on roadmap" script: "We hear you. Here's what we're doing instead and why."
Custom deals: Never say yes without PM review of scope

### Working with Customer Success

NPS/CSAT review: Monthly with CS, quarterly trends
Churn analysis: PM owns understanding, CS owns save plays
Feature adoption data: CS flags underused features
Voice of customer pipeline: CS → PM structured feedback channel

### Opportunity Sizing

TAM: [Total addressable market — everyone who could use this]
SAM: [Serviceable addressable — our segment of TAM]
SOM: [Serviceable obtainable — realistic capture in 3 years]

Bottom-up validation:
[Number of target companies] × [seats per company] × [price per seat] × [conversion rate] = [Revenue estimate]

### Build vs Buy vs Partner Decision

FactorBuildBuy/IntegratePartnerCore to value prop✅ Build❌❌Commoditized❌✅ Buy❌Adjacent capability❌❌✅ PartnerSpeed critical❌ (slow)✅ (fast)✅ (fast)Control critical✅❌❌Maintenance burdenHighLowShared

### Technical Debt Classification

TypeImpactPriorityActionBlocks featuresRevenueP0Sprint nowSlows developmentVelocityP1Next sprintCreates incidentsReliabilityP1Next sprintUgly but worksPrideP3BacklogTheoretical concernNone yetP4Ignore for now

### Product Thinking Frameworks Quick Reference

FrameworkWhen to UseCore QuestionJobs to Be DoneDiscoveryWhat job is the user hiring us for?Kano ModelPrioritizationIs this basic, performance, or delight?RICEScoringWhat's the ROI of this investment?Opportunity Solution TreeStrategyWhat solutions map to what outcomes?Double DiamondProcessAre we solving the right problem?Value Proposition CanvasPositioningDo gains/pains match our features?Pirate Metrics (AARRR)GrowthWhere's the funnel leaking?North StarAlignmentWhat single metric matters most?

### Platform / API Product Management

Developers are users too — discovery, interviews, friction audits apply
Docs are UI — if docs are bad, API is unusable
Breaking changes are product decisions — deprecation timeline = roadmap item
Adoption funnel: Discover → Register → First API call → Integration live → Expansion
Time to first API call = your activation metric

### Multi-Product / Portfolio Management

Shared platform strategy: What's shared vs. product-specific?
Cannibalization analysis: Does new product steal from existing?
Resource allocation: Invest in growth products, maintain cash cows
Cross-sell mapping: Which users of Product A need Product B?

### International / Localization

Market prioritization: Size × ease of entry × cultural fit
Localization vs. translation: Adapt the product, not just the words
Regulatory differences: Privacy, data residency, payments
Local competition: Incumbents may be stronger than global view suggests

### AI/ML Feature Product Management

Set expectations: ML is probabilistic — "usually right" not "always right"
Feedback loops: Users correct outputs → model improves → users trust more
Confidence thresholds: Show/hide based on model confidence
Fallback UX: What happens when the model fails?
Bias audits: Check outputs across user segments regularly
Cost per inference: Factor into unit economics

### Rescue Playbook (Failing Product)

Diagnose: Is it demand, execution, or market timing?
Talk to churned users: 5 calls in 5 days — why did they leave?
Find the 10%: Who ARE the happy users? What do they have in common?
Narrow focus: Kill everything except what serves the happy 10%
Set a deadline: 90 days to hit a clear milestone or sunset

### 100-Point Quality Rubric

Score any product initiative across 8 dimensions:

DimensionWeight1 (Poor)3 (Good)5 (Excellent)Problem clarity20%Assumed, no evidenceSome user quotesQuantified with multiple sourcesUser understanding15%No researchSurveys onlyRegular interviews + dataPrioritization rigor15%Gut feelBasic scoringRICE+ with strategic alignmentSpec completeness15%Vague requirementsStories + ACFull spec with edge casesMetrics discipline15%No trackingVanity metricsNorth star + leading + guardrailsExecution quality10%Ship and prayQA + rollout planFeature flags + monitoring + rollbackStakeholder alignment5%SurprisesRegular updatesProactive partnershipLearning velocity5%No post-mortemsQuarterly reviewsWeekly metrics + iteration

Score: (Σ dimension_score × weight) × 4 = /100

Below 60 = significant gaps. 60-80 = good with room to improve. Above 80 = strong PM practice.

### Common PM Mistakes

MistakeFixBuilding what stakeholders requestBuild what moves the north starShipping without measuringDefine success metric BEFORE buildingFeatures without adoption planActivation strategy for every featureSpec during sprintSpec BEFORE sprint — always one sprint aheadSaying "we'll add it later"If it's not in V1 scope, don't promiseConsensus-seekingDisagree and commit — decisions > meetingsRoadmap = feature listRoadmap = outcome targetsCompeting on featuresCompete on experience and speedIgnoring churned usersChurned users are your best teachersBig bang launchesProgressive rollouts with feature flags

### Natural Language Commands

/pm strategy — Generate a strategy brief for a product/feature
/pm discovery — Create an interview script for a research question
/pm prioritize — Score a list of features using RICE+
/pm roadmap — Build a Now/Next/Later roadmap
/pm spec — Write a one-pager or user stories for a feature
/pm launch — Generate a launch checklist
/pm metrics — Design a north star framework
/pm review — Run a post-launch review
/pm stakeholder — Map stakeholders and communication plan
/pm health — Score current PM practice /80
/pm rescue — Diagnose and plan for a struggling product
/pm compete — Analyze competitive positioning

### File Structure

product/
├── strategy.yaml          # Product strategy brief
├── roadmap.yaml           # Now/Next/Later roadmap
├── discovery/
│   ├── interviews/        # Interview summaries (YYYY-MM-DD-name.yaml)
│   ├── synthesis.md       # Pattern analysis
│   └── validation-log.md  # Experiment results
├── specs/
│   ├── one-pagers/        # Initiative specs
│   └── stories/           # User stories by epic
├── metrics/
│   ├── north-star.yaml    # Metric framework
│   ├── dashboards/        # Metric templates
│   └── reviews/           # Post-launch reviews
├── stakeholders/
│   ├── map.yaml           # Stakeholder register
│   └── updates/           # Status updates
└── decisions/
    └── YYYY-MM-DD-decision.md  # Key product decisions with rationale

Built by AfrexAI — AI-powered business operations.
## Trust
- Source: tencent
- Verification: Indexed source record
- Publisher: 1kalin
- Version: 1.0.0
## Source health
- Status: healthy
- Source download looks usable.
- Yavira can redirect you to the upstream package for this source.
- Health scope: source
- Reason: direct_download_ok
- Checked at: 2026-04-30T16:55:25.780Z
- Expires at: 2026-05-07T16:55:25.780Z
- Recommended action: Download for OpenClaw
## Links
- [Detail page](https://openagent3.xyz/skills/afrexai-product-manager)
- [Send to Agent page](https://openagent3.xyz/skills/afrexai-product-manager/agent)
- [JSON manifest](https://openagent3.xyz/skills/afrexai-product-manager/agent.json)
- [Markdown brief](https://openagent3.xyz/skills/afrexai-product-manager/agent.md)
- [Download page](https://openagent3.xyz/downloads/afrexai-product-manager)