1. Reframe “speed” as iteration speed, not output speed
Focus on iteration, not volume
- Replace “deliverables per sprint” metrics with “validated improvements per sprint”.
- Use AI to generate multiple quick variations, then run short feedback loops to learn fast.
- Encourage teams to show learning velocity, not just production output.
2. Keep humans for judgement
Designate a quality reviewer role in each sprint - someone responsible for checking alignment with user needs, tone, and ethics.
- Use human review checkpoints before publishing or releasing AI content.
- Encourage cross-disciplinary review (e.g., a designer reviews copy, a researcher reviews UX flow).
3. Build reflection pauses
Schedule short “reflection breaks” (10-15 min) at the end of a working block.
4. Use quality guardrails
Co-create an AI quality checklist with your team - clarity, accuracy, usefulness, tone, ethics. Build quality prompts into your workflows:
- “Check this for factual accuracy and fairness.”
- “Rephrase for inclusive language.”
- Automate some checks (grammar, factual verification), but review nuance manually.
5. Shift from done to ready for iteration
Label outputs in progress boards as “Draft”, “For review”, or “Validated”.
- Train teams to treat first drafts as starting points, not final assets.
- Encourage visible iteration — show version histories and what changed between versions.
6. Foster a culture of iteration
Celebrate revisions that improve clarity or user value rather than praising first-pass brilliance.
- Include “iteration highlights” in sprint reviews.
- Normalise feedback — make it routine and safe, not personal or punitive.
7. Measure learning, not throughput
Add metrics such as:
- Number of insights gained from user feedback
- Percentage of improvements validated by testing
Visualise progress over cycles — show learning curves, not just deadlines met. Encourage post-mortems focused on what was learned, not what went wrong.
8. Guard against automation bias
Run “challenge the AI” sessions - team members must find at least one thing the model got wrong or incomplete.
9. Keep craftsmanship central
Create “craft review” sessions where teams discuss why one version is better - focus on human judgement and nuance.
- Document examples of high-quality outcomes and why they work.
- Give credit for human refinement and taste, not just speed.
10. Use speed for learning loops
Use AI to accelerate hypothesis testing, not content production.
Run quick experiments (A/B, mock user feedback) on AI outputs.
Reflect after each: “What did we learn about users or quality?” Archive learnings in a shared “AI playbook” or Miro board for ongoing refinement.