-
Notifications
You must be signed in to change notification settings - Fork 125
Description
Overview
Create dt-method-08-testing.instructions.md — the method-tier instruction file for Method 8: User Testing. Users test their hi-fi prototypes with real or representative users to gather evidence for refinement decisions. Method 8 is critical for supporting non-linear loops — testing results may trigger returns to Method 2 (Research), Method 4 (Brainstorming), or any earlier method when test findings invalidate core assumptions.
Target File
.github/instructions/dt-method-08-testing.instructions.md
Frontmatter
---
description: 'Design Thinking Method 8: User Testing — evidence-based evaluation, test protocols, and non-linear iteration support'
applyTo: '**/.copilot-tracking/dt/**/method-08*'
---Required Content
Method Purpose
Method 8 puts prototypes in front of users and gathers evidence. Unlike Method 6's feedback planning (which was lo-fi), Method 8 conducts structured testing with functional prototypes. Test results drive one of three outcomes: iterate (Method 9), revisit (return to an earlier method), or validate (proceed to Implementation Space exit).
Three Sub-Methods
| Sub-Method | Phase | Coach Behavior |
|---|---|---|
| 8a — Test Planning | Planning | Help user design test protocols — who, what, how, and what constitutes success/failure. Coach asks: "What would the test results need to show for you to be confident this works?" |
| 8b — Test Execution | Execution | Guide testing sessions. Coach helps maintain neutral observation: "Let's watch what users do, not what they say they'll do." |
| 8c — Results Analysis | Documentation | Analyze test data and determine next steps. Coach facilitates honest interpretation: "The data suggests users struggled with X — what does that mean for the concept?" |
Two Specialized Hats
| Hat | Role | When Activated |
|---|---|---|
| Test Designer | Helps plan testing methodology — participant selection, success metrics, protocol design, and bias mitigation | During 8a |
| Evidence Analyst | Facilitates objective interpretation of test results — separates signal from noise, identifies patterns, and connects findings to decisions | During 8b and 8c |
Non-Linear Iteration Loops
Method 8 is the primary trigger for non-linear navigation. Test results may reveal:
| Finding | Action | Target |
|---|---|---|
| Missing user data | Return to research | → Method 2 |
| Concept invalidated | Return to brainstorming | → Method 4 |
| Wrong fidelity level | Revisit prototyping | → Method 6 or 7 |
| Minor usability issues | Iterate | → Method 9 |
| Core assumptions validated | Proceed | → Method 9 or exit |
The coach helps users interpret findings honestly: "Testing revealed that users don't understand the core value proposition. That's a Method 4 finding, not a Method 9 tweak — should we revisit our concepts?"
Test Protocol Design
Structured testing approaches the coach can guide:
- Task-based testing: Users attempt specific tasks while the coach helps observe completion rate, time, and error patterns
- A/B comparison testing: Two prototype variants tested against the same criteria
- Think-aloud testing: Users narrate their thinking during interaction
- Wizard of Oz testing: Parts of the prototype are simulated by a human — tests concept viability before full build
- Longitudinal testing: Extended use over days/weeks — reveals adoption patterns (when feasible)
Method 8 Artifacts
Outputs stored at .copilot-tracking/dt/{project-slug}/method-08-testing/:
test-protocol.md— Testing methodology, participant profiles, success criteriatest-sessions/— Per-session notes, observations, and raw dataresults-analysis.md— Synthesized findings, patterns, and statistical observationsdecision-log.md— Iterate, revisit, or validate decisions with supporting evidence
Coaching Approach
The coach in Method 8:
- Designs rigorous tests: "A yes/no question won't tell us much — what open-ended tasks would reveal how users actually behave?"
- Prevents confirmation bias: "I notice we're focusing on the positive results — what about the 3 users who abandoned the task?"
- Supports difficult decisions: "The data suggests the concept needs significant changes. That's not a failure — it's evidence that saves you from building the wrong thing."
- Facilitates loop decisions: "This finding goes deeper than a UI tweak. Should we take this back to Method 4 and reconsider the concept?"
Token Budget
Target: ~1,500-2,000 tokens (method tier)
How to Build This File
This is an .instructions.md file — use the prompt-builder agent (not task-implementor) for the authoring phase. The prompt-builder includes built-in Prompt Quality Criteria validation and sandbox testing specific to AI artifacts (.instructions.md, .prompt.md, .agent.md, SKILL.md).
Workflow: /task-research → /task-plan → /prompt-build → /task-review
Between each phase, run /clear to reset context.
Phase 1: Research
Gather source material for the user testing method instruction file.
Source Material:
design-thinking-for-hve-capabilities/guidance/08-user-testing.md
This file lives in the DT4HVE repository. If you don't have local access, ask the user to provide it or useread_fileif the repo is cloned nearby.
Steps:
- Read the source material above.
- Read
.github/instructions/prompt-builder.instructions.mdfor authoring standards. - Read any existing
dt-method-*instruction files for structural precedent. - Gather content on testing methodologies, non-linear loop patterns, bias mitigation in testing, and evidence interpretation frameworks.
Starter prompt:
/task-research
Research for IS034: dt-method-08-testing.instructions.md
Read the DT4HVE source material at design-thinking-for-hve-capabilities/guidance/08-user-testing.md. Extract:
- Testing methodologies (usability testing, A/B testing, cognitive walkthrough, heuristic evaluation)
- Non-linear loop patterns — how testing results feed back to earlier phases
- Bias mitigation in testing — confirmation bias, leading questions, observer effects
- Evidence interpretation frameworks for making go/no-go decisions
- Test protocol types and when to apply each
Also read .github/instructions/prompt-builder.instructions.md for authoring standards and any existing dt-method-*.instructions.md files for structural precedent.
Output: research summary from Phase 1 above
Phase 2: Plan
Plan the file structure and content organization using the research output.
Steps:
- Review the research output from Phase 1.
- Plan the instruction file structure — method purpose, three sub-methods, two specialized hats, non-linear loop table, test protocol types, decision framework.
- Define section ordering, token allocation, and applyTo targeting.
Starter prompt:
/task-plan
Plan for IS034: dt-method-08-testing.instructions.md
Using the Phase 1 research output, plan the instruction file:
- Method purpose: validating prototypes through structured user testing
- Three sub-methods: 8a Test Planning, 8b Test Execution, 8c Evidence Interpretation
- Two hats: Test Facilitator, Evidence Interpreter
- Non-linear loop table mapping test findings to earlier DT phases
- Test protocol types with selection guidance
- Decision framework for go/iterate/pivot based on test results
- Bias mitigation coaching integrated throughout
- Section ordering and token budget allocation (~1,500-2,000 tokens)
- applyTo: '**/.copilot-tracking/dt/**/method-08*'
Output: plan at .copilot-tracking/plans/{date}-is034-dt-method-08-plan.md
Phase 3: Build
Author the instruction file using the prompt-builder agent.
Steps:
- Review the plan from Phase 2.
- Author the instruction file using
/prompt-build. - Non-linear iteration support is the defining feature — ensure the loop table and coaching guidance empower honest interpretation.
Starter prompt:
/prompt-build file=.github/instructions/dt-method-08-testing.instructions.md
Build IS034 using the plan at .copilot-tracking/plans/{date}-is034-dt-method-08-plan.md.
This is a method-tier instruction file for Method 8: User Testing. Key authoring notes:
- applyTo targets Method 8 artifact paths: '**/.copilot-tracking/dt/**/method-08*'
- Non-linear iteration support is the defining feature — the loop table and coaching guidance must empower honest interpretation
- Three sub-methods with consistent table structure
- Two specialized hats with activation triggers
- Non-linear loop table connecting test findings back to earlier DT phases
- Bias mitigation coaching woven throughout — not a separate section
- Test protocol types with clear selection criteria
- Decision framework: go/iterate/pivot with evidence thresholds
- Coaching examples demonstrate honest evidence interpretation over confirmation bias
Phase 4: Review
Validate the instruction file against the plan and prompt-builder standards.
Steps:
- Review the built file against prompt-builder standards and the issue requirements.
- Validate non-linear loop support, testing rigor, bias mitigation, and structural compliance.
Starter prompt:
/task-review
Review IS034: .github/instructions/dt-method-08-testing.instructions.md
Validate against:
- prompt-builder.instructions.md authoring standards
- Non-linear loop support — loop table maps findings to earlier phases
- Testing rigor — protocol types are well-differentiated with selection criteria
- Bias mitigation — coaching addresses confirmation bias, leading questions, observer effects
- Decision framework clarity with evidence thresholds
- Token budget: ~1,500-2,000 tokens
- Frontmatter applyTo correctness
- Three sub-methods and two hats with consistent table structure
After Review
- Pass: Mark IS034 complete.
- Iterate: Address review findings, rebuild, re-review.
- Escalate: If blocked by missing DT4HVE source material or architectural questions, raise to the user.
Authoring Standards
Follow .github/instructions/prompt-builder.instructions.md:
applyTotargets Method 8 artifact paths- Three sub-methods with consistent table structure
- Two specialized hats with activation triggers
- Non-linear iteration loops are first-class concerns
Success Criteria
- File created at
.github/instructions/dt-method-08-testing.instructions.md - Frontmatter
applyTotargets Method 8 artifact paths - Three sub-methods defined (planning, execution, analysis)
- Two specialized hats with clear activation triggers
- Non-linear iteration loop table mapping findings to revisit targets
- Test protocol design covering task-based, A/B, think-aloud, and Wizard of Oz approaches
- Coaching examples support honest interpretation and difficult decisions to revisit earlier methods
- Token count within ~1,500-2,000 target
- Passes task-reviewer validation against prompt-builder standards
- Each prompt, instructions, or agent file registered in
collections/design-thinking.collection.ymlwithpathandkindfields - Each prompt, instructions, or agent file registered in
collections/hve-core-all.collection.ymlwithpathandkindfields -
npm run plugin:generatesucceeds after collection manifest updates
Metadata
Metadata
Assignees
Labels
Type
Projects
Status