Amorcer la gestion de produit avec recherche

Transforme un brief produit en artefacts complets et validés via une recherche de domaine exhaustive avant questionnement, réduisant l'interaction utilisateur tout en garantissant des recommandations de qualité professionnelle.

Spar Skills Guide Bot
Business & AdministrationAvancé
21011/03/2026
Claude Code
#product-management#research-synthesis#documentation#artifacts#market-analysis

name: bootstrap-product description: | Transform a product briefing into comprehensive, research-backed product management artifacts.

This skill conducts domain-expert research BEFORE user questioning to enable smarter questions and pre-populated artifacts with validated content. Uses Context7 for technology documentation, WebSearch for market/architecture/security research, and WebFetch for deep-dive analysis.

Generates 4 research-enriched files:

  • product.md (product vision with market research citations)
  • roadmap.md (12-month roadmap with architecture research)
  • architecture.md (technical design with extensive Context7 references)
  • adr.md (architectural decisions with research-justified rationale)

Triggers: "create product vision", "define new product", "product planning", "bootstrap product", "product documentation", "start new product", "product briefing"

allowed-tools:

  • AskUserQuestion
  • WebSearch
  • WebFetch
  • mcp__plugin_context7_context7__resolve-library-id
  • mcp__plugin_context7_context7__query-docs
  • Read
  • Write
  • Edit
  • Grep
  • Glob

model: sonnet

Bootstrap Product Skill

Purpose: Transform minimal product briefings into rich, research-backed product management artifacts that are market-viable and technically sound.

Key Innovation: Domain-expert research agent conducts comprehensive research BEFORE questioning, reducing user burden from 17 questions to typically 8-12 questions while delivering higher-quality, validated recommendations.

Process Overview

1. Accept Product Briefing
   ↓
2. Conduct Domain-Expert Research (NEW)
   ├─ Context7: Technology documentation
   ├─ WebSearch: Market/architecture/security
   ├─ WebFetch: Deep-dive resources
   └─ Domain: Scientific/industry research
   ↓
3. Synthesize Research Report
   ↓
4. Ask Research-Informed Questions (8-12 instead of 17)
   ↓
5. Confirm Understanding (with research context)
   ↓
6. Generate Research-Enriched Artifacts (4 files)
   ↓
7. Update .context/ with Research Summary
   ↓
8. Provide Completion Summary with Citations

Step 1: Accept & Analyze Briefing

Input: Product briefing from user (can be minimal - e.g., "Build a collaborative document editor")

Actions:

  • Parse briefing for core concept, domain, technology hints
  • Extract research keywords: product type, domain, use case, tech stack clues
  • Identify what's missing that research can help fill

Example:

User: "Build a collaborative document editor"
→ Research keywords: "collaborative editing", "document editor", "real-time collaboration"
→ Technology areas: frontend frameworks, WebSocket libraries, rich text editors
→ Domain areas: market size, competitors (Google Docs, Notion), architecture patterns

Step 2: Conduct Domain-Expert Research

CRITICAL: Research happens BEFORE questioning to inform smarter questions and pre-populate artifacts.

2.1 Technology Documentation Research (Context7)

Purpose: Identify best practices and recommended technologies

Process:

  1. Identify 3-5 relevant technology candidates from briefing
  2. For each technology:
    resolve-library-id(
      query="[Technology description]",
      libraryName="[framework name]"
    ) → libraryId
    
    query-docs(
      libraryId="[returned ID]",
      query="best practices for [specific use case]"
    ) → Documentation findings
    
  3. Document findings with library IDs and queries used

Limit: 3-5 Context7 queries maximum

2.2 Architecture Pattern Research (WebSearch + WebFetch)

Purpose: Research proven architecture patterns for this domain

Process:

  1. WebSearch for architecture patterns (5-8 queries):
    • "[domain/use case] architecture patterns 2026"
    • "[domain] scalability best practices 2026"
    • "microservices vs monolith [use case] 2026"
  2. WebFetch 2-3 key resources:
    • Architecture whitepapers
    • Case studies from similar products
    • Implementation guides

Limit: 5-8 WebSearch queries, 2-3 WebFetch resources

2.3 Security & Compliance Research (WebSearch + WebFetch)

Purpose: Identify regulatory requirements and security best practices

Process:

  1. WebSearch for compliance (5-8 queries):
    • "GDPR compliance [domain] applications 2026"
    • "HIPAA requirements [domain] 2026"
    • "SOC2 compliance SaaS applications 2026"
    • "OWASP top 10 [domain] security 2026"
  2. WebFetch official compliance documentation

Limit: 5-8 compliance/security searches

2.4 Domain Knowledge Research (WebSearch + WebFetch)

Purpose: Understand market, competitors, and domain-specific insights

Process:

  1. WebSearch for market intelligence (8-10 queries):
    • "[product type] market size 2026"
    • "[domain] industry trends 2026"
    • "[use case] competitive landscape"
    • "key competitors [product type]"
  2. WebFetch 2-4 resources:
    • Market research reports
    • Academic papers (if applicable)
    • Industry analyses

Limit: 8-10 market/domain searches, 2-4 WebFetch resources

2.5 Research Synthesis

Output: Structured research report containing:

## Research Report

### Technology Research (Context7)
- [Library 1]: [Key findings]
- [Library 2]: [Key findings]
- Recommendation: [Suggested tech stack]

### Architecture Research
- Pattern recommendation: [e.g., Monolith for MVP, microservices later]
- Scalability approach: [Key patterns found]
- Case studies: [Similar products]

### Security & Compliance
- Required standards: [GDPR, HIPAA, SOC2, etc.]
- Security measures: [OWASP compliance, encryption, etc.]

### Domain Knowledge
- Market size: [TAM from research]
- Key competitors: [List with strengths/weaknesses]
- Industry trends: [Relevant trends]

### Research Gaps (Need User Input)
- [Question 1 that research couldn't answer]
- [Question 2 that requires user preference]
- [Question 3 that needs validation]

Step 3: Ask Research-Informed Questions

Strategy:

  • Review research report before asking ANY questions
  • Skip questions where research provides clear answers
  • Ask validation questions to confirm research findings
  • Focus on user preferences, constraints, and goals that research cannot determine
  • Reduce from 17 questions to typically 8-12 questions

Question Categories (see full command file for complete question framework):

  1. Product Essence (4 questions) - May be informed by domain research
  2. Market Context (4 questions) - May have data from market research
  3. Technical Constraints (3 questions) - Research identifies compliance needs
  4. Execution Context (3 questions) - Research informs timeline estimates
  5. Product Scope (3 questions) - Research identifies must-have features

Example (Collaborative Document Editor):

Research found:
- Market size: $5B TAM
- Competitors: Google Docs, Notion, Confluence
- Tech stack: React + WebSocket recommended
- Compliance: GDPR for EU customers
- Architecture: Operational Transform or CRDT patterns

Questions SKIPPED:
✗ "What's the market size?" (research found: $5B)
✗ "Who are competitors?" (research identified 3 major players)
✗ "Technology preferences?" (research suggests React + Socket.io)

Questions ASKED:
✓ "Do you need GDPR compliance?" (validate research finding)
✓ "What's your differentiation vs Google Docs?" (user vision)
✓ "Target scale?" (informs architecture choice)
✓ "MVP timeline?" (user constraint)
✓ "Team size?" (user constraint)

Result: 8 targeted questions instead of 17 generic ones

Step 4: Confirm Understanding

Present research-enhanced confirmation:

Let me confirm what I understand about your product:

**Product**: [Name/description]
**Core Problem**: [2-3 sentences]
**Target Users**: [User persona]
**Market Context**: [Size and competitors FROM RESEARCH]
**Key Differentiation**: [Unique value]
**Technical Approach**: [Architecture informed by Context7 research]
**Compliance Requirements**: [GDPR, HIPAA, SOC2 identified FROM RESEARCH]
**MVP Timeline**: [Timeline]
**Success Metrics**: [2-4 metrics]

**Research Conducted**:
- Technology: [Context7 libraries queried]
- Market: [Key findings]
- Security: [Standards identified]
- Domain: [Insights]

Is this correct? Please confirm or provide corrections.

Step 5: Generate Research-Enriched Artifacts

Generation Order (dependency-driven):

5.1 product.md (150-250 lines)

  • Product vision with market research citations
  • Competitive landscape FROM RESEARCH
  • Success metrics with industry benchmarks FROM RESEARCH

5.2 roadmap.md (200-250 lines)

  • Phases informed by architecture research
  • Timeline realistic based on technology research

5.3 architecture.md (200-300 lines)

  • Technology stack backed by Context7 documentation
  • Architecture pattern from research
  • Security measures from compliance research
  • EXTENSIVE Context7 citations

5.4 adr.md (100-150 lines)

  • ADR-001: Technology Stack (Context7-backed)
  • ADR-002: Architecture Pattern (research-validated)
  • ADR-003: Database Choice (comparative research)
  • ADR-004: Security & Compliance (regulatory research)
  • All ADRs include research citations

Progress Indicators:

Generating research-enriched artifacts...
✓ Created product.md (187 lines) - with market research
✓ Generated roadmap.md (223 lines) - with architecture research
✓ Designed architecture.md (298 lines) - with Context7 references
✓ Documented adr.md (156 lines) - with research-justified decisions

Step 6: Update .context/

notes.md (< 150 lines)

Add Product Bootstrap Summary including:

  • Product overview
  • Research Conducted section
  • Key Research Findings
  • Research Sources Summary
  • Key docs references

changelog.md (< 70 lines)

Add bootstrap entry including:

  • Decisions (7 key decisions)
  • Research Conducted section
  • Artifacts generated WITH research annotations
  • Rationale with research backing

handoff.md

Create comprehensive handoff including:

  • Product artifacts generated
  • Information gathered
  • Research Conducted section (detailed)
  • Important decisions
  • Next steps

Step 7: Provide Summary

Summary Format:

## Product Bootstrapping Complete!

### Product Overview
- **Name**: [Name]
- **Vision**: [One sentence]
- **Target**: [User segment]
- **MVP Timeline**: [Timeline]

### Generated Artifacts
- product.md (X lines) - with market research citations
- roadmap.md (X lines) - with architecture research
- architecture.md (X lines) - with Context7 references
- adr.md (X lines) - 4 ADRs with research justification

### Research Conducted
**Context7**: [X] libraries documented
**WebSearch**: [Y] searches (market/architecture/security)
**WebFetch**: [Z] deep-dive resources

**Impact**:
- Questions reduced from 17 to [actual]
- All decisions research-backed
- Full citation traceability

### Next Steps
1. Review artifacts and research citations
2. Validate findings against domain expertise
3. Begin MVP development planning

Important Guidelines

DO:

  • ✅ Conduct research BEFORE asking questions
  • ✅ Skip questions that research confidently answered
  • ✅ Include research citations in ALL artifacts
  • ✅ Use Context7 for all technology decisions
  • ✅ Cite specific library IDs (/org/project format)
  • ✅ Keep .context/ files under 500 lines
  • ✅ Provide research sources summary

DON'T:

  • ❌ Ask all 17 questions if research answered some
  • ❌ Make technology recommendations without Context7 backing
  • ❌ Skip research phase to save time
  • ❌ Omit research citations from artifacts
  • ❌ Exceed research query limits (causes token bloat)
  • ❌ Generate artifacts without research validation

Research Query Limits (CRITICAL):

  • Context7: 3-5 libraries max
  • WebSearch: 5-8 per category (market, architecture, security)
  • WebFetch: 2-4 deep resources max
  • Enforce these to prevent token bloat and API overuse

Success Criteria

After execution:

  • ✅ 4 comprehensive product files generated (600-1000 lines total)
  • ✅ All artifacts include research citations
  • ✅ Technology decisions backed by Context7 documentation
  • ✅ Architectural decisions validated by industry research
  • ✅ Compliance requirements identified proactively
  • ✅ Questions reduced to 8-12 based on research coverage
  • ✅ .context/ files updated with research summary
  • ✅ All .context/ files under 500 lines
  • ✅ Full citation traceability for all recommendations

Templates

Note: This skill uses abbreviated templates. For complete templates with all sections and examples, see:

  • .claude/commands/bootstrap-product.md (full command file, ~2000 lines)

The full command file contains:

  • Detailed question framework (all 17 questions with research annotations)
  • Complete artifact templates (product.md, roadmap.md, architecture.md, adr.md)
  • Research integration instructions
  • Example execution flows

Command Version: For explicit invocation, use /bootstrap-product [briefing] Skill Version: This file - activated by semantic triggers for product planning conversations

Skills similaires