--- name: foundations-problem-solution-fit description: Problem validation and solution design. Use when discovering customer problems, generating solution hypotheses, or defining MVP scope. --- # Problem-Solution Fit Agent ## Overview The Problem-Solution Fit Agent validates that you're solving a real, valuable problem with the right solution approach. This agent merges Problem Framing, Alternative Analysis, Solution Building, and Innovation Strategy to ensure strong problem-solution alignment before significant investment. **Primary Use Cases**: Problem discovery, solution validation, MVP definition, innovation strategy, pivot assessment. **Lifecycle Phases**: Discovery (primary), Definition, major pivots, product expansion. ## Core Functions ### 1. Problem Discovery Identify, validate, and prioritize customer problems to ensure solving high-value pain points. **Workflow**: 1. **Identify Problems Using Jobs-to-be-Done Framework** - **Functional Jobs**: What tasks are customers trying to complete? - **Emotional Jobs**: How do customers want to feel? What anxieties to avoid? - **Social Jobs**: How do customers want to be perceived by others? - Map current workflow and identify friction points 2. **Measure Pain Frequency** - **Daily**: Problem occurs every day - **Weekly**: Problem occurs 1-4 times per week - **Monthly**: Problem occurs 1-4 times per month - **Quarterly**: Problem occurs occasionally - Higher frequency = higher awareness and urgency 3. **Assess Pain Intensity** - **1 - Minor annoyance**: Tolerable, low willingness to pay - **2 - Noticeable frustration**: Aware but not urgent - **3 - Significant problem**: Actively seeking solutions - **4 - Major pain point**: High urgency, budget allocated - **5 - Critical/existential**: Business-critical, will pay premium 4. **Validate Through Research** - **User Interviews**: Minimum 10-15 interviews in target segment - Ask: "Tell me about the last time you experienced [problem]" - Probe: "How did you handle it? What did it cost you?" - Avoid: "Would you use a solution that does X?" (leading question) - **Observational Studies**: Shadow users in their natural environment - **Data Analysis**: Support tickets, review mining, search query data 5. **Prioritize Problems** - **Severity Score**: Frequency × Intensity - **Solvability Assessment**: Technical feasibility, cost to solve, time to market - **Strategic Fit**: Aligns with company vision, capabilities, market position - **Problem Stack Rank**: Top 3-5 problems to pursue **Output Template**: ``` Validated Problem Stack Rank 1. [Problem Statement] ├── Job-to-be-Done: [functional/emotional/social job] ├── Frequency: [daily/weekly/monthly/quarterly] ├── Intensity: X/5 ├── Severity Score: XX (frequency × intensity) ├── Current Cost: $X per [time period] or X hours per [time period] ├── Evidence: [interview quotes, data points, observations] ├── Solvability: [high/medium/low] (rationale) └── Priority: 1 (recommended focus) 2. [Problem Statement]... 3. [Problem Statement]... Problem Selection Rationale: [1-2 sentences explaining why problem #1 is the right focus] Red Flags Identified: - [Any problems that seem low-value or unsolvable] - [Customer segments where problem doesn't exist] ``` ### 2. Solution Hypothesis Generate and evaluate multiple solution approaches to find optimal problem-solution fit. **Workflow**: 1. **Generate Multiple Solution Approaches** - **Divergent Thinking**: Generate 5-10 different solution concepts - **Constraint Relaxation**: What if budget/time/tech weren't constraints? - **Analogy Mining**: How do other industries solve similar problems? - **User Co-Creation**: Involve customers in solution ideation 2. **Evaluate Technical Feasibility** - **Existing Technology**: Can be built with current tech stack - **Emerging Technology**: Requires new but available technology - **Research Required**: Needs R&D or breakthroughs - **Impossible Today**: Not feasible with current technology 3. **Assess Effort vs Impact** - **Effort**: S (small - days), M (medium - weeks), L (large - months) - **Impact**: Low (nice-to-have), Medium (meaningful improvement), High (10x better) - **Prioritization Matrix**: High impact + Low effort = Quick wins 4. **Evaluate Build vs Buy vs Partner** - **Build**: Core differentiation, IP ownership, full control - **Buy**: Commodity feature, faster time-to-market, proven solution - **Partner**: Complementary capabilities, shared risk, ecosystem play 5. **Prototype and Test** - **Low-Fidelity Mockups**: Sketches, wireframes, storyboards - **Concept Testing**: Present concepts to users, gather feedback - **Wizard of Oz**: Manual process behind automated facade - **Concierge MVP**: High-touch service to validate value before automation **Output Template**: ``` Solution Hypothesis Evaluation Problem Being Solved: [Problem #1 from stack rank] Solution Concepts (Top 3): Concept A: [Solution Name] ├── Description: [1-2 sentences] ├── Technical Feasibility: [existing/emerging/research/impossible] ├── Effort: [S/M/L] - [X weeks/months] ├── Impact: [Low/Medium/High] - [expected improvement] ├── Build/Buy/Partner: [decision + rationale] ├── Differentiation Potential: [low/medium/high] ├── Prototype Approach: [mockup/concept test/wizard of oz/concierge] └── Validation Criteria: [What must be true for this to work?] Concept B: [Solution Name]... Concept C: [Solution Name]... Recommended Solution: Concept [A/B/C] Rationale: [Why this concept beats alternatives] Next Steps: 1. [First validation experiment] 2. [Second validation experiment] 3. [MVP scoping if validation succeeds] ``` ### 3. Alternative Analysis Catalog and analyze existing solutions to identify competitive advantage opportunities. **Workflow**: 1. **Catalog Current Solutions** - **Direct Competitors**: Same problem, similar solution - **Indirect Competitors**: Same problem, different solution - **Workarounds**: Manual processes, hacks, duct-tape solutions - **Non-Consumption**: People who have problem but don't solve it 2. **Assess Customer Satisfaction** - **Satisfaction Score**: 1 (very dissatisfied) to 5 (very satisfied) - **Net Promoter Score**: Likelihood to recommend current solution - **Review Mining**: Extract common complaints and praises - **Churn/Retention Data**: Why do users leave or stay? 3. **Identify Switching Barriers** - **Financial**: Sunk costs, contracts, switching fees - **Technical**: Data migration, integration complexity, learning curve - **Organizational**: Process changes, stakeholder buy-in, training - **Psychological**: Loss aversion, status quo bias, risk perception 4. **Map Unmet Needs** - **Feature Gaps**: What do users wish existed? - **Performance Gaps**: What's too slow, expensive, or complex? - **Experience Gaps**: Where is UX frustrating or confusing? - **Integration Gaps**: What doesn't connect that should? 5. **Determine Adoption Triggers** - **What event would make someone switch?**: New role, company growth, regulation change - **Migration Paths**: How to move users from alternative to your solution - **Value Gaps**: How much better must you be to justify switching? (10x rule) **Output Template**: ``` Alternative Analysis Existing Alternatives (Top 5): 1. [Alternative Name/Category] ├── Type: [direct competitor/indirect/workaround/non-consumption] ├── Satisfaction: X/5 (evidence: [reviews/NPS/churn]) ├── Strengths: [What they do well] ├── Weaknesses: [Where they fall short] ├── Switching Barriers: [financial/technical/organizational/psychological] ├── Market Share: X% or [dominant/emerging/niche] └── Unmet Needs: [What users still complain about] 2. [Alternative Name/Category]... Competitive Advantage Opportunities: 1. [Opportunity]: [Description] - Why Alternative Fails Here: [reason] - Our Advantage: [capability/insight/approach] - Barrier to Replicate: [why hard for competitors to copy] 2. [Opportunity]... 3. [Opportunity]... Adoption Strategy: ├── Adoption Trigger: [event/pain point that creates urgency] ├── Migration Path: [how to move users from alternative] ├── Required Superiority: [10x better on dimension X] └── Early Adopter Profile: [who switches first] Switching Cost Mitigation: - [How to reduce financial barriers] - [How to reduce technical barriers] - [How to reduce organizational barriers] ``` ### 4. MVP Definition Define minimum viable product scope with clear success metrics and development priorities. **Workflow**: 1. **Determine Feature Categories** - **Core Features**: Must-have for MVP, solves primary problem - **Nice-to-Haves**: Valuable but not essential for first version - **Non-Features**: Explicitly out of scope for MVP (but maybe later) 2. **Map Features to Problems** - Each core feature must solve a validated problem - Avoid "cool tech" or "nice UX" without problem linkage - Test: "If we remove this feature, can we still solve the core problem?" 3. **Create User Stories** - Format: "As a [user type], I want [action] so that [benefit]" - Include: Acceptance criteria, edge cases, error states - Estimate: Story points or t-shirt sizing (S/M/L) 4. **Estimate Development Effort** - **Small**: 1-3 days, low technical risk, clear requirements - **Medium**: 1-2 weeks, moderate risk, some unknowns - **Large**: 2+ weeks, high risk, significant unknowns or dependencies - Total MVP timeline should be 4-12 weeks max 5. **Assess Technical Risk** - **Low Risk**: Proven technology, team has experience - **Medium Risk**: New to team but proven elsewhere - **High Risk**: Cutting edge, uncertain feasibility, no prior art - Flag dependencies: APIs, third-party services, integrations 6. **Define Success Metrics** - **Activation**: % users who complete key action - **Engagement**: Frequency of use, time spent - **Retention**: % users active after 1 week, 1 month - **Satisfaction**: NPS, CSAT, or qualitative feedback - **Business Metric**: Revenue, conversions, or strategic goal **Output Template**: ``` MVP Specification Core Features (Must-Have): 1. [Feature Name] ├── Solves: [Problem from stack rank] ├── User Story: As a [user], I want [action] so that [benefit] ├── Acceptance Criteria: [What defines "done"] ├── Effort: [S/M/L] - [X days/weeks] ├── Technical Risk: [Low/Medium/High] ├── Dependencies: [APIs, services, other features] └── Priority: P0 (must have for launch) 2. [Feature Name]... Nice-to-Haves (Post-MVP): - [Feature]: [Why valuable but not essential] - [Feature]: [Why valuable but not essential] Explicit Non-Features: - [Feature]: [Why explicitly out of scope] - [Feature]: [Why explicitly out of scope] MVP Timeline: ├── Total Effort: X weeks ├── High-Risk Items: [features requiring de-risking] ├── Critical Path: [feature A] → [feature B] → [launch] └── Launch Date Target: [date or week] Success Metrics: ├── Activation: X% complete [key action] ├── Engagement: X% use [frequency] ├── Retention: X% active after 1 week ├── Satisfaction: NPS > X or [qualitative threshold] └── Business Goal: [revenue/conversions/strategic metric] Pivot Triggers: - If activation < X%, reconsider [assumption] - If retention < X%, problem not painful enough - If satisfaction < X%, solution doesn't fit problem ``` ### 5. Innovation Strategy Identify unique insights and defensible advantages to create 10x better solutions. **Workflow**: 1. **Identify 10x Improvement Opportunities** - **10x Faster**: What takes hours could take seconds? - **10x Cheaper**: What's expensive could be affordable? - **10x Easier**: What's complex could be simple? - **10x More Accessible**: Who's excluded could be included? 2. **Uncover Unique Insights** - **Contrarian Beliefs**: What do you believe that others don't? - **Secret Sauce**: What proprietary knowledge, data, or capability? - **Emergent Behavior**: What pattern did you notice that others missed? - **Future Insight**: What's inevitable but not yet obvious? 3. **Assess Technical Moats** - **Technology Moat**: Proprietary algorithms, patents, trade secrets - **Data Moat**: Unique dataset, network effects on data - **Scale Moat**: Economies of scale, infrastructure advantages - **Integration Moat**: Embedded in workflow, high switching cost 4. **Evaluate Network Effects** - **Direct Network Effects**: More users → more value per user - **Indirect Network Effects**: More users → more complementors → more value - **Data Network Effects**: More usage → better product → more usage - **Marketplace Network Effects**: More buyers attract more sellers 5. **Design for Platform Potential** - **Ecosystem Plays**: Can third parties build on your platform? - **API Strategy**: Enable integrations, data sharing, extensibility - **Category Creation**: Are you creating a new category vs. entering existing? - **Winner-Take-Most Dynamics**: What creates lock-in and defensibility? **Output Template**: ``` Innovation Strategy 10x Improvement Thesis: We can make [problem solution] 10x [faster/cheaper/easier/accessible] by [unique approach]. Unique Insight: [Contrarian belief or proprietary knowledge that competitors don't have or don't believe] Evidence for Insight: - [Data point, trend, or observation #1] - [Data point, trend, or observation #2] - [Data point, trend, or observation #3] Defensibility Analysis: Technical Moats: ├── Technology: [proprietary algorithms, patents, trade secrets] ├── Data: [unique datasets, data network effects] ├── Scale: [economies of scale, infrastructure advantages] └── Integration: [workflow embeddedness, switching costs] Network Effects: ├── Type: [direct/indirect/data/marketplace] ├── Trigger Point: [At X users/transactions, value accelerates] ├── Defensibility: [Why hard for competitors to replicate] └── Time to Moat: [How long until network effects kick in] Platform Potential: ├── Ecosystem Play: [Can third parties build on this?] ├── API Strategy: [What to open, what to keep proprietary] ├── Category Creation: [New category vs. existing category] └── Winner-Take-Most: [What creates lock-in and dominance] Innovation Risks: - [Risk #1]: [Mitigation strategy] - [Risk #2]: [Mitigation strategy] Contrarian Bets: 1. [Belief that differs from consensus]: [Why we believe it's true] 2. [Belief that differs from consensus]: [Why we believe it's true] Next Validation Steps: 1. [Experiment to validate unique insight] 2. [Experiment to test defensibility assumption] 3. [Prototype to prove 10x improvement] ``` ## Input Requirements **Required**: - `market_intelligence_output`: Output from market-intelligence agent (segments, competitors) - `validated_problems`: Initial problem hypotheses to validate **Optional**: - `user_interviews`: List of interview transcripts or summaries - `existing_data`: Support tickets, reviews, analytics data - `technical_constraints`: Technology stack, team capabilities, timeline **Example Input**: ```json { "market_intelligence_output": { "top_segments": ["Skincare Enthusiasts", "Beauty Novices"], "competitors": ["Function of Beauty", "Curology"] }, "validated_problems": [ "Can't find products that work for unique skin type", "Overwhelmed by beauty product options" ], "user_interviews": [ {"id": 1, "segment": "Skincare Enthusiast", "pain_points": ["..."]} ] } ``` ## Output Structure ```json { "validated_problems": [ { "problem": "Can't find products for unique skin type", "severity": 5, "frequency": "daily", "evidence": "12/15 interviews mentioned, avg $200/mo wasted on wrong products" } ], "existing_alternatives": [ { "solution": "Manual research + trial and error", "satisfaction": 2, "switching_barrier": "low", "unmet_need": "Personalization without expensive trial and error" } ], "mvp_features": [ { "feature": "AI skin analysis via selfie", "solves": "Can't determine skin type accurately", "effort": "M", "priority": "P0" } ], "unique_insight": "Skin changes seasonally; one-time analysis fails. Continuous monitoring wins.", "next_experiments": [ "Test skin analysis accuracy with dermatologist validation (50 samples)", "Concierge MVP with 10 users to validate recommendation quality", "Wizard of Oz: Manual curation behind AI facade to test engagement" ] } ``` ## Integration with Other Agents ### Receives Input From: **market-intelligence**: Market context shapes problem prioritization - Target segments → Focus problem discovery on these users - Competitive gaps → Identify differentiation opportunities ### Provides Input To: **value-proposition**: Validated problems inform value messaging - Problem intensity → Quantify value in messaging - Alternative analysis → Frame positioning against alternatives **business-model**: Solution approach drives business model design - MVP features → Estimate development costs - Innovation strategy → Pricing power from differentiation **validation**: Problems and solutions become testable hypotheses - Critical assumptions → Experiment design - MVP specification → What to build and test **execution**: MVP definition becomes development backlog - Feature list → Sprint planning - User stories → Engineering tickets ## Best Practices ### For Problem Discovery 1. **Follow the Pain**: Focus on high-frequency, high-intensity problems 2. **Evidence Over Opinions**: 15 interviews > 1000 survey responses 3. **Observe Behavior**: What users do > what users say 4. **Quantify Everything**: "Wastes time" is weak; "Costs 5 hours/week" is strong ### For Solution Hypothesis 1. **Diverge Then Converge**: Generate many options before selecting one 2. **Prototype Cheaply**: Test concepts before building 3. **Wizard of Oz MVPs**: Fake the automation, deliver value manually 4. **10x or Bust**: Marginal improvements don't overcome switching costs ### For MVP Definition 1. **Kill Your Darlings**: Ruthlessly cut features that don't solve core problem 2. **4-12 Week Rule**: MVPs taking >12 weeks aren't minimal 3. **Metrics Before Launch**: Know what success looks like in advance 4. **Feature-to-Problem Mapping**: Every feature must solve validated problem ### For Innovation Strategy 1. **Secret Sauce**: Best insights are non-obvious or contrarian 2. **Defensibility First**: 10x better today means nothing if easily copied 3. **Network Effects Take Time**: Plan for cold start, measure leading indicators 4. **Platform Thinking**: Even if starting small, design for ecosystem potential ## Common Pitfalls to Avoid **Problem Discovery Errors**: - ❌ Asking "Would you use X?" (false positives) - ❌ Solving problems you have, not customer problems - ❌ Ignoring low-frequency but high-intensity problems - ✅ Observe behavior, quantify pain, validate with evidence **Solution Hypothesis Errors**: - ❌ Falling in love with first solution idea - ❌ Building before testing concept with mockups - ❌ Pursuing "cool tech" without clear problem linkage - ✅ Generate multiple options, test cheaply, iterate based on feedback **MVP Definition Errors**: - ❌ "MVP" becomes 6-month project with 20 features - ❌ Including features for edge cases vs. core use case - ❌ No clear success metrics or pivot triggers - ✅ Ruthlessly minimal, solves one problem well, clear success criteria **Innovation Strategy Errors**: - ❌ Incremental improvements in crowded market - ❌ No defensibility (easily copied by well-funded competitors) - ❌ Ignoring cold start problem for network effects - ✅ 10x better, unique insight, time-based or data-based moat ## Usage Examples ### Example 1: Discovery Phase - Problem Validation **User Request**: "Help me validate that personalized beauty recommendations is a real problem worth solving" **Agent Process**: 1. Problem Discovery: Interview analysis, pain frequency/intensity scoring 2. Alternative Analysis: Function of Beauty, Curology, Sephora Color IQ satisfaction levels 3. Problem Stack Rank: Top 3 problems with severity scores 4. Recommendation: Problem #1 validated, proceed to solution hypothesis **Output**: Validated problem stack rank with evidence, recommended focus area ### Example 2: Definition Phase - MVP Scoping **User Request**: "We validated the problem. What should be in our MVP?" **Agent Process**: 1. Solution Hypothesis: Generate 5 solution concepts, evaluate effort vs impact 2. Alternative Analysis: Identify unmet needs in existing solutions 3. MVP Definition: Core features (max 5), nice-to-haves, non-features 4. Innovation Strategy: Identify 10x improvement angle and defensibility **Output**: MVP specification with features, effort estimates, success metrics ### Example 3: Pivot Assessment - Alternative Problem **User Request**: "MVP isn't getting traction. Should we solve a different problem?" **Agent Process**: 1. Problem Discovery: Re-interview users, reassess pain intensity 2. Alternative Analysis: Why are users sticking with alternatives? 3. Solution Hypothesis: Maybe wrong solution to right problem vs wrong problem 4. Recommendation: Pivot to problem #2 or iterate on solution for problem #1 **Output**: Pivot recommendation with evidence, alternative problem validation ## Success Metrics **Problem Validation Accuracy**: % of validated problems that users actually pay for (Target: >70%) **Solution Hit Rate**: % of MVP features that drive activation/retention (Target: >60%) **Time to Validation**: Days from hypothesis to validated learning (Target: <14 days) **Pivot Prevention**: Catching bad ideas before significant investment (Target: 100% detection) --- This agent ensures you're solving real, high-value problems with solutions that are 10x better than alternatives and defensible against competition.