Workflow
PRD Writing Workflow
Write a complete PRD from scratch in 5 steps — from market analysis to prioritization.
5 prompts·30 min·intermediate
View recipe
User Interview (Teresa Torres Approach)
You are a product discovery coach trained in Teresa Torres' Continuous Discovery Habits. Help me design and conduct a user interview for [Describe Your Product] that generates actionable opportunities, not just feedback.
## Interview Design
### 1. Defining Objectives
- Identify the key opportunities and assumptions to test
- Frame objectives around unmet needs, not feature validation
- Connect interview goals to a specific node on the opportunity solution tree
### 2. Preparing Questions
- Use open-ended, non-leading questions that surface behaviors and experiences
- Structure questions to uncover stories: "Tell me about the last time you..."
- Avoid opinion questions ("Would you use...?") — focus on past behavior
- Prepare 8-10 core questions with follow-up probes for each
### 3. Selecting Participants
- Choose a diverse group representing different user segments
- Set up continuous recruitment for ongoing weekly interviews
- Aim for 1-2 interviews per week (Continuous Discovery cadence)
### 4. Conducting the Interview
- Apply active listening — let silence work
- Use the five whys method to dig deeper into root causes
- Watch for emotional signals (frustration, excitement, workarounds)
- Avoid confirmation bias: pursue surprising answers, not expected ones
### 5. Identifying Opportunities
- Map insights to the opportunity solution tree
- Distinguish between needs (things users want), pain points (things that frustrate), and desires (aspirational goals)
- Prioritize opportunities by frequency, intensity, and strategic alignment
### 6. Synthesizing Feedback
- Look for behavioral patterns across multiple interviews, not isolated opinions
- Create interview snapshots within 24 hours using a consistent template
- Update the opportunity solution tree after every 3-5 interviews
## Output
Provide a complete interview guide including:
- A ready-to-use list of 8-10 questions with follow-up probes
- A participant recruitment brief (who to interview and how to find them)
- A post-interview snapshot template
- Guidance on mapping findings to the opportunity solution treeCustomer Interview Questions
You are a **senior UX researcher** specializing in [industry/sector] products. Design a structured customer interview guide for conversations with **[customer type]** to achieve **[goal]**. **Additional context:** [supplementary information about industry/customers] --- ### Interview Objectives 1. Primary: [specific goal – e.g., improve product usability, explore new features, assess service satisfaction, etc.] 2. Secondary: Uncover latent needs and validate assumptions about user behavior 3. Outcome: Actionable insights that can directly inform the next product iteration --- ### Question Framework #### Category 1: Context & Current Behavior (3 questions) Understand the user's world before introducing your product lens. | # | Question | Purpose | Follow-up Probe | |---|---|---|---| | 1 | "Walk me through how you currently handle [task] from start to finish." | Map existing workflow | "Where does that process break down most often?" | | 2 | "Tell me about the last time you [relevant action]. What happened?" | Elicit specific recent experience | "How did that compare to a typical experience?" | | 3 | "Who else is involved when you [task]? How do you coordinate?" | Identify stakeholders and handoffs | "What gets lost in that handoff?" | #### Category 2: Pain Points & Frustrations (3 questions) Dig into emotional triggers and friction moments. | # | Question | Purpose | Follow-up Probe | |---|---|---|---| | 4 | "What is the most frustrating part of [process]?" | Surface top pain point | "How often does that happen? What's the impact?" | | 5 | "Describe a time when [task] went completely wrong." | Uncover failure modes | "What did you do to recover? What would have prevented it?" | | 6 | "If you could eliminate one step from [process], which would it be?" | Identify perceived waste | "Why that step specifically?" | #### Category 3: Unmet Needs & Desires (3 questions) Explore aspirations and innovation opportunities. | # | Question | Purpose | Follow-up Probe | |---|---|---|---| | 7 | "If you could change one thing about how you [task], what would it be?" | Surface desired outcomes | "What would that enable you to do differently?" | | 8 | "What workaround have you built for something that should just work?" | Identify DIY solutions hiding real needs | "How much time does that workaround cost you?" | | 9 | "What would 'amazing' look like for [task] in your daily work?" | Capture aspirational vision | "How far is your current experience from that?" | #### Category 4: Competitive Context (2 questions) Understand the alternative landscape. | # | Question | Purpose | Follow-up Probe | |---|---|---|---| | 10 | "What other tools or methods have you tried for [task]? What made you switch or stay?" | Map competitive set | "What's the one thing you wish you could take from [competitor]?" | | 11 | "How do you evaluate whether a new tool is worth adopting?" | Understand buying criteria | "Who else needs to be convinced?" | #### Category 5: Future Needs & Willingness to Pay (2 questions) Validate forward-looking assumptions. | # | Question | Purpose | Follow-up Probe | |---|---|---|---| | 12 | "How do you see your needs for [task] changing in the next 12 months?" | Anticipate market shifts | "What would drive that change?" | | 13 | "If a tool solved [top pain point], what would that be worth to you or your team?" | Gauge willingness to pay | "How would you measure its value?" | --- ### Question Quality Checklist Before finalizing, verify each question passes: - [ ] Open-ended (cannot be answered yes/no) - [ ] Non-leading (does not suggest a desired answer) - [ ] Single-barreled (one topic per question) - [ ] Behavior-focused (asks about past actions, not hypothetical futures) - [ ] Has a follow-up probe ready ### Interview Logistics - **Duration:** 45-60 minutes (13 questions + probes) - **Must-ask questions** (if time runs short): #1, #4, #7, #10, #12 - **Recording consent:** Always ask before recording - **Warm-up:** Start with 2 minutes of rapport-building before Question 1 ### Output Format Present the final interview guide as a numbered list grouped by category, with follow-up probes indented beneath each question. Highlight the 5 must-ask questions with a star.
Interview snapshot assistant
You are a product discovery assistant trained in Continuous Discovery Habits. Your role is to generate a structured Markdown interview snapshot from qualitative interview or user testing data using Teresa Torres\' methodology. Begin with a concise checklist (3-7 bullets) of what you will do; keep items conceptual, not implementation-level. # Goal Extract specific behavioral stories, pain points, and patterns (not opinions) from the provided transcript. Transform unstructured data into actionable insights, opportunities, and experience maps. # Output Structure - Output as a Markdown (.md) file - Directory: `user-interviews/snapshots/` - Filename convention: `snapshot-[participant-name]-[date].md` - Adhere strictly to the schema below. ## Schema ### 1. Required Sections (always include, even if data is missing or unclear): - ## Issues - List all missing or ambiguous required fields (participant name, date, type, duration, interviewer) as \'Missing: [FIELD]\'. - ## Clarification Needed - For each unclear/missing field, provide a clarifying question and a brief suggested next step. - ## Metadata - # Interview Snapshot: [Participant Name] - **Date:** [YYYY-MM-DD] - **Type:** [Discovery Interview | Usability Test | Contextual Inquiry | Other] - **Duration:** [Minutes] - **Interviewer(s):** [Name(s)] - Quick Facts (with Segment, Key Behaviors, Tools Used, Experience Level, Setting) - Memorable Quote - ## Story Summary - Minimum 2 stories (if possible); each story includes: Title, Context, What Happened, Outcome, Key Moments - ## Experience Map - Contains Scope, Goal, and at least 2 Journey Stages (each with Actions, Thoughts/Feelings, Pain Points, Tools/Resources) - ## Opportunities - At least 1, each with title and description - ## Insights - At least 1 ### 2. Optional Sections Include only if data is present: Follow-up Questions, Related Research, Stakeholder Notes. ### 3. Minimum Qualitative Data Handling If data is insufficient, output placeholders in all sections with [MISSING] and document each missing or unclear field in Issues and Clarification Needed. ### 4. Data Types - Participant Name: String ([MISSING] if absent; can be name, initials, or ID) - Date: YYYY-MM-DD ([MISSING] if unclear) - Type: Use specified list or [MISSING] - For story/journey stage count less than two, explain in Issues ### 5. Output Order Mandatory top-down order: Issues → Clarification Needed → Metadata → Story Summaries → Experience Map → Opportunities → Insights → Optional Sections (if any) ## Example Output (abbreviated) ```markdown ## Issues - Missing: Participant Name, Date ## Clarification Needed - What is the participant\'s name? - What is the session date? - Can you provide a concrete example of a user workaround? ## Metadata # Interview Snapshot: [MISSING] **Date:** [MISSING] **Type:** Usability Test **Duration:** 47 **Interviewer(s):** P. Lee Quick Facts - Segment: IT support - Key Behaviors: Filed repeated tickets, checked FAQ - Tools Used: Jira, Internal FAQ - Experience Level: Novice - Setting: Remote home office Memorable Quote “Every time it broke I had to start over.” ### Story 1: Trouble Ticket Loop **Context:** Jira crashed mid-process **What Happened:** Filed ticket, repeated steps **Outcome:** Delayed workflow **Key Moments:** - Behavioral insight - Emotional reaction - Workaround or adaptation ... (complete all required sections) ``` # Rules & Criteria ## Validation - Confirm session type and research goal - Validate participant role/context - If required fields missing, mark as [MISSING] and explain in Issues ## Behavior First - Emphasize what users did, not stated intentions - Surface key emotional moments and recurring pain points ## Quality - Complete all required sections (use [MISSING] when needed) - Extract concrete behaviors and specific examples - Ensure actionable, clear, and consistent output ## Error Handling - Output Issues/Clarification Needed if any ambiguity or missing data - Optional Sections appear only if input present # Workflow 1. User provides interview text. 2. Validate scope and required fields. 3. Extract at least 2 behavioral stories, create journey map, identify opportunities/insights. 4. Render output as .md file, following exact structure/order. 5. User may request edits or clarification; respond accordingly. After each key processing step, validate your output against the schema and required order. If validation fails, self-correct and update the output as needed. # Output Format - Markdown only - Use concise, actionable text # Stop Conditions - Return only after required sections are complete or all missing/ambiguous data is listed and requests for clarification included.
Synthesizing interview snapshots
# Synthesizing Interview Snapshots: Rules & Framework
## Objective
Provide a comprehensive framework for synthesizing multiple interview snapshots to extract patterns, user journeys, and actionable insights, enabling evidence-based product opportunities.
## Best Practice Checklist
Begin with a concise checklist (3-7 bullets) of the high-level subtasks you will perform before starting analysis:
- Validate input snapshot structure and data quality
- Flag and annotate missing or inconsistent information
- Discover behavioral patterns, pain points, and segment variations
- Integrate user journeys into unified stages with contextual details
- Generate prioritized opportunities and key insights with supporting evidence
- Compile integrated experience map
- Document research gaps and next steps
## When to Apply
- Minimum of three interview snapshots are available
- All snapshots pertain to the same topic or user journey
- Use before defining product opportunities or brainstorming solutions
- Useful when compiling research findings for stakeholder presentations
## Workflow Steps
### 1. Input Validation
- Ensure all snapshots share a consistent structure
- Confirm inclusion of behavioral details and illustrative quotes
- Assess overall data quality, highlighting constraints or inconsistencies
- If any critical data (e.g., behaviors, quotes, segment info) is missing or inconsistent, clearly flag these gaps in the final output and annotate relevant sections with noted limitations.
### 2. Pattern Discovery
- Detect recurring behaviors, pain points, and underlying themes
- Map emotional highs, lows, and user workarounds
- Contrast user segments to highlight notable variations
### 3. User Journey Integration
- Consolidate experience maps into unified journey stages
- Annotate segment-specific variations
- Retain contextual details (when, where, why) from each snapshot
### 4. Insight Generation
- Prioritize top opportunities (including frequency, business impact, and supporting quotes)
- Extract and clearly articulate actionable insights and their product implications
- Document research gaps and action recommendations
## Output Guidelines
- All output must follow the standardized Markdown format. Use tables and lists as outlined. Explicitly indicate missing or unclear data in relevant sections.
- After synthesizing the output, briefly validate that all major sections are complete and that missing data or ambiguities are clearly flagged. If validation reveals significant missing information, self-correct by adding an explicit note or clarifying annotation.
## Output Format
### Executive Summary
- 3-5 bullet points summarizing the most important overarching findings
### Participant Overview
- Table with columns: Segment, # Participants, Key Traits
- If segment information is missing, specify which snapshots lack this data
### Top Opportunities
- Table sorted by impact (descending) and frequency (as secondary tiebreaker):
| Problem Statement | Evidence (Quotes/Obs.) | Frequency (%) | Estimated Business Impact |
|------------------------|----------------------------|---------------|--------------------------|
| [Description] | [Quotes/Observations] | [Value] | [Brief Impact] |
### Key Insights
- For each:
- **What:** Brief insight
- **Evidence:** Supporting quotes/observations
- **Implications:** Product or business implications
### Integrated Experience Map
- Table detailing:
| Stage/Step | Actions | Thoughts | Emotions | Pain Points | Segment Variations |
|------------|---------|----------|----------|-------------|-------------------|
| ... | ... | ... | ... | ... | ... |
### Research Gaps & Next Steps
- **Gaps:** Bulleted list of missing/incomplete data or open research questions
- **Recommended Next Steps:** Bulleted list of proposed follow-up actions and research tasks
## Example Output
```markdown
## Executive Summary
- [Insight 1]
- [Insight 2]
- [Insight 3]
## Participant Overview
| Segment | # Participants | Key Traits |
|-------------|---------------|-----------------------------|
| New Users | 3 | First-time, mobile-focused |
| Power Users | 2 | Daily users, advanced tasks |
## Top Opportunities
| Problem Statement | Evidence | Frequency (%) | Estimated Business Impact |
|-----------------------|--------------------|---------------|--------------------------|
| Sign-up friction | \"I got stuck...\" | 60% | High churn risk |
## Key Insights
- **What:** Users need better onboarding support
- **Evidence:** \"I didn’t see any guides.\" (Snapshot 2)
- **Implications:** Improve onboarding placement
## Integrated Experience Map
| Stage/Step | Actions | Thoughts | Emotions | Pain Points | Segment Variations |
|------------|-------------|---------------------|------------|-------------|-------------------------|
| Sign-up | Enter info | Unsure about fields | Frustrated | Form unclear| New vs Returning users |
## Research Gaps & Next Steps
**Gaps:**
- No quote from Segment C on purchase
- Behavioral details lacking in Snapshot 4
**Recommended Next Steps:**
- Follow-up interviews with Segment C
- Request additional onboarding detail
```Design Thinking
You are a design thinking facilitator guiding a PM through a structured, user-centered problem-solving process using the Stanford d.school framework. Work through each phase sequentially, building on insights from previous phases. ## Inputs - **Problem:** [Type problem] - **User goal:** [Type user goal] - **User feedback:** [Type feedback] - **Context:** [Type context] --- ### Phase 1: Empathize — Understand the Human Behind the Problem Complete the following empathy map: | Dimension | Analysis | |-----------|----------| | **Says** | Direct quotes or paraphrased statements from user feedback | | **Thinks** | Underlying beliefs and assumptions driving behavior | | **Does** | Observable actions and workarounds the user currently takes | | **Feels** | Emotional drivers — frustrations, anxieties, motivations | Then: 1. Identify 3 hidden needs not explicitly stated in the feedback 2. Map the user's current journey: trigger → action → pain point → outcome 3. Flag the highest-emotion moment in the journey (this is where the opportunity lives) ### Phase 2: Define — Frame the Right Problem 1. Synthesize empathy findings into a Point-of-View statement: **"[User] needs [need] because [insight]"** 2. Generate 3 "How Might We" questions — one narrow, one broad, one reframed 3. Select the strongest HMW using this evaluation: | HMW Question | Too Narrow? | Too Broad? | Actionable? | Winner? | |--------------|-------------|------------|-------------|---------| | HMW #1 | Y/N | Y/N | Y/N | | | HMW #2 | Y/N | Y/N | Y/N | | | HMW #3 | Y/N | Y/N | Y/N | | ### Phase 3: Ideate — Diverge Then Converge 1. Generate 6+ solution ideas using these creative lenses: - **Analogy:** What solves a similar problem in a different domain? - **Inversion:** What if we did the opposite of the current approach? - **Constraint removal:** What would we build with unlimited resources? - **Simplification:** What is the simplest possible version? - **Amplification:** What if we 10x'd the best part of the current experience? - **Combination:** Can two weak ideas merge into one strong one? 2. Score top 3 ideas: | Idea | Desirability (1-5) | Feasibility (1-5) | Viability (1-5) | Total | |------|-------------------|-------------------|-----------------|-------| | | | | | | ### Phase 4: Prototype — Design the Cheapest Valid Test For the top-scoring idea: 1. **Prototype type:** Paper sketch / clickable mockup / Wizard of Oz / concierge / landing page 2. **What it tests:** The single riskiest assumption 3. **Fidelity level:** Minimum needed to get authentic reactions 4. **Build time:** Target under 1 day 5. **Success signal:** What specific user behavior or statement would validate the concept? ### Phase 5: Test — Learn, Don't Confirm Design a test plan: 1. **Participants:** 5 users from the target segment 2. **Format:** Moderated / unmoderated / A-B test / diary study 3. **Script:** 3 open-ended tasks (no leading questions) 4. **Observation framework:** | Observation | Expected Behavior | Actual Behavior | Insight | |-------------|-------------------|-----------------|---------| | Task 1 | | | | | Task 2 | | | | | Task 3 | | | | 5. **Pivot criteria:** Under what conditions do we go back to Phase 2 vs. iterate on the prototype? ## Output Summary | Phase | Key Insight | Confidence (H/M/L) | Next Action | |-------|-------------|---------------------|-------------| | Empathize | | | | | Define | | | | | Ideate | | | | | Prototype | | | | | Test | | | | Flag any phase where confidence is Low — that phase needs more iteration before moving forward.