A practical playbook for PMs and POs worldwide. Eight modules. The complete lifecycle of product delivery — accelerated by AI at every phase.
Answer 9 questions. Get a personalised breakdown of PM/PO savings, development savings, and rework avoided — with your ROI on AI tooling investment.
Interactive ToolHow to extract, structure, and understand any client input using AI tools — and produce a Discovery Summary, Discovery Questionnaire, and Concept Note ready for Module 2.
Product Idea Formulation is the first phase of any engagement. A client comes with a problem, a vision, an RFP, or a rough idea. Your job is to decode what they actually need — not just what they said.
The core principle: A client says what they want. Your job is to understand what they need. These are almost never the same thing — and the gap between them is where projects fail.
Clients often describe a symptom, not the root cause. "We need a dashboard" often means "our doctors need to make faster treatment decisions without missing critical patient history."
"We need a dashboard to track our revenues."
Our pricing decisions depend on 2 people and the knowledge dies with them. We're losing revenue daily.
Your question: "What happens today when neither of those two people is available?"
⚠️ Watch for clients who jump straight to features. When someone says "I want a mobile app with push notifications and a leaderboard" — pause. Ask: "What problem are you trying to solve for the person using this?"
A feature request: "I want an AI chatbot." A business problem: "Our support team is overwhelmed and response times are hurting customer retention." Features are solutions. Problems are opportunities.
When a client names a feature, ask "why" five times in gentle succession. You'll almost always arrive at the real business problem by the third or fourth why. It feels uncomfortable at first. It works every time.
This is more common than you think. A client comes in excited with vague energy: "something like Uber but for laundry" or "we want to use AI somehow." Don't panic. Don't fill the silence with your own assumptions.
Fill the gap with your own assumptions. Nod along. Leave thinking you understood when you didn't.
"Help me understand — who is the person whose life gets easier when this product exists? Walk me through their typical day."
The real cost of getting this phase wrong: A requirement misunderstood here costs 1× to fix. Caught in development: 10×. Caught in production: 100×. This phase is the cheapest place to get it right.
Before gathering information, identify what type of engagement this is. The questions you ask differ significantly.
Nothing exists yet. The client wants to build a new product from scratch. Focus: understanding the problem space, target users, and what success looks like from zero.
A product already exists. The client needs evolution, new features, or AI added. Focus: what's broken, what's working, and what the constraints are.
In real life, you won't always get the same type of information. Here are the five input types and what to look for in each.
By the end of this phase, you need answers — or documented unknowns — across all 9 categories.
You won't get answers to everything in the first conversation. What matters is knowing what you still need to find out — and going back for it before moving to Module 2.
Here is how you work through this phase every time, regardless of the input type you receive.
Don't just skim. Read looking for four specific things: the pain word (the emotional trigger), the trigger event (why now), the named solution (what they think they want), and the gap (what is conspicuously missing).
Before any client conversation, map your current knowledge against all 9 categories. Mark each: Known ✓, Partially Known ~, or Unknown ?. The unknowns drive your questionnaire.
Using your knowledge map, build a targeted questionnaire. Focus on your unknowns. Don't ask about what you already know — it wastes the client's time and signals you didn't read their brief. Business context first, technical constraints last.
The discovery call is not a Q&A session — it's a guided conversation. Ask one question, listen fully, probe one layer deeper, then move on. End with: "Is there anything you expected me to ask that I haven't?"
The Concept Note is your deliverable from this phase. 1-2 pages capturing everything you now know. Clear enough for a non-technical client to validate. Detailed enough for your team to understand what comes next.
What a good Concept Note contains: Project working title · Problem statement (1-2 sentences) · Engagement type (Build or Support) · Primary users and their core need · Product type and platform · Top 5-8 MVP features · Key unknowns still to resolve · Recommended next step
Use these four prompts in sequence. Fill in the highlighted placeholders with your specific context. Always review and correct the output before using it.
AI processes what is written. It cannot know what the client said between the lines, what your presales colleague told you off the record, or what your experience tells you about this type of client. Every AI output here requires your review, your correction, and your sign-off. You are accountable — not the tool.
Does the core problem match what you actually understood? Add context the AI couldn't know — things said verbally, client history, relationships. Correct anything that feels off before moving to Prompt 2.
Remove questions already answered. Add questions only you know to ask based on conversations so far. Reorder by importance. You know this client — the AI does not.
Read it as the client: would you validate this description of your own problem? Read it as a developer: could you understand what you're being asked to build? Fix what doesn't pass both tests. This document is yours — not the AI's.
Each example shows the actual raw input, what each prompt produces, and what the final Concept Note looks like. This is exactly what you'll produce in real work.
Example 1 · Mobile Product · EdTech · UK Market
From: Presales Team — "We are talking to a prospect who wants to build an educational maths application for the UK primary school market. They want something similar to existing gamified learning platforms — avatars, coins, reward systems, and an engaging experience for children aged 6-11. Can you help understand what skills, tools, and approach would be required so we can include this in our proposal?"
Compliance & Regulatory Flags (critical):
Specialist skills needed: EdTech-specialist UX/UI designer · 2D character illustrator with modular asset experience · UK educational consultant · QA engineer with accessibility background
Project Title: LearnQuest UK — Gamified Primary Maths Platform
Problem Statement: UK primary school children lack a motivating, curriculum-aligned digital maths platform they actually want to use. Existing tools are either too rigid or too entertainment-focused.
Platform: Web-first (school compatibility), mobile app (iOS + Android) for home use.
MVP Features: Modular avatar builder and reward economy · UK National Curriculum question bank (KS1/KS2) · Progress tracking and milestone badges · Teacher dashboard with assignment controls · Parent view · Audio hints for KS1 learners
Next Step: Schedule 60-minute discovery call with the client's product owner. The compliance awareness conversation must happen before any effort estimate is prepared.
AI surfaced the UK Age Appropriate Design Code (AADC) compliance requirement — a legal obligation the PM didn't initially know to look for. Missing this in the proposal would have added weeks of unscoped work mid-project. This is where AI's domain breadth protects you.
Example 2 · AI Product · Hospitality · Enterprise
"Their revenue management expertise is concentrated in 2-3 senior individuals. Their existing software produces unexplained recommendations the team doesn't trust and overrides manually. They want an AI system that encodes their pricing logic, explains every recommendation in plain language, and can scale across all their properties without additional headcount."
Project Title: ARIA — AI Revenue Intelligence & Automation Platform
Problem Statement: A hotel group's pricing expertise is locked inside 2-3 individuals. Their existing software produces recommendations nobody trusts — so they're overriding it manually every day, losing revenue. They need an AI system that encodes that expertise, explains every recommendation in plain English, and scales across their portfolio.
MVP Features: Demand forecasting engine (arrival-based, segment-level) · BAR and segment pricing recommendations · Plain-English AI narration for every recommendation · Competitor rate and market signal integration · Performance dashboard · Multi-property architecture (12 properties, scalable to 30)
★ Next Step: Schedule an Algorithm Scoping Session with the client's senior revenue managers and our data science lead. Effort estimation cannot be finalised until this session is complete.
AI reframed the project correctly — from "build a new RMS" to "encode 30 years of human expertise into an explainable AI system." That reframe alone changes the estimation by months and hundreds of thousands in cost. The algorithm scoping session became the single critical next step. Without it, any estimate would be a guess.
Module 2 builds directly on what you produce here. If you can't check a box, go back and fill the gap.
I can state the client's core problem in one sentence — without using the word "solution" or "system"
I have identified the engagement type — BUILD (greenfield) or SUPPORT (existing product) — and I can justify it
I know who the primary user is and what their single biggest daily frustration looks like in practice
I have mapped all 9 knowledge categories — and documented which are Known, Partially Known, or Unknown
I have a discovery questionnaire ready — tailored to this project type, not a generic list
I have identified the compliance and regulatory requirements relevant to this project and region
I have a Concept Note that a non-technical client can read and validate — and that my team can brief from
I have reviewed every AI output — corrected what's wrong, added what AI couldn't know, and I own the result
I have a clear complexity signal — Low / Medium / High / Very High — and I can explain why
I know the recommended next step — what needs to happen before Module 2 begins
This is the one-page summary of everything Module 1 asks you to do.
✅ Module 1 complete. You now have a Discovery Summary, a tailored questionnaire, and a Concept Note. Take these into Module 2 — Know the Battlefield, where you validate the product idea against the real market before committing to a proposal.
How to use AI for market research, competitor analysis, and feature prioritisation — so you can validate the product idea before committing to a proposal.
You have a Discovery Summary from Module 1. A client has described their problem. Now you need to answer a harder question: should this be built at all — and if so, what should it actually do?
The output of this module is not a research report. It is a Validated PRD v0.1 — a document that says: here is the problem, here is the market reality, here are the features that will win, and here is why I believe this. Without that conviction, you will build the wrong thing or price it wrong or both.
In most delivery teams, structured market research before scoping is almost universally skipped. The standard flow is: receive brief → clarify scope → estimate effort → write proposal. Market validation is not part of it.
This happens for a practical reason: research takes time that no one has budgeted for, and AI was not fast enough to make it worthwhile. That equation has changed. With the right prompt sequence, a PM can run a credible competitive landscape, feature benchmark, and PRD draft in under two hours.
This module introduces a new step into standard delivery practice. It is not documenting what most teams do. It is showing what the best teams now do — and what every PM can do once AI makes it fast enough to be practical.
Someone needs to want this product enough to pay for it. Your research either confirms this or challenges it. Both are valuable.
There is almost always a competitor — direct or indirect. Understanding what they do well and where they fall short tells you what the winning position is.
Not every feature the client mentioned is worth building. Your job is to separate the features that drive adoption from the ones that sound good in a meeting.
⚠ Common trap — Confirmation bias. You come in believing the client's idea is right. Your research looks for evidence that confirms this. Every finding gets interpreted to support the answer you already expect. Fight this actively. Ask AI to argue against the idea. Look for signals that the market is already saturated or that users have tried this before and abandoned it.
| Delivery context | How to use Module 2 |
|---|---|
| New product build | Run in full — all six research areas apply |
| Significant new feature or capability | Run partially — competitor benchmarking and feature gap analysis. Skip market sizing. |
| Enhancement or iteration on live product | Use only feature gap analysis and user sentiment sections |
| Support or maintenance engagement | Does not apply — proceed directly to Module 3 |
| Team augmentation | Does not apply — proceed directly to Module 3 |
Gaps are fine at the start — the point is to know which areas are incomplete so you can go back and fill them before the PRD is written.
| Research Area | What You Are Looking For |
|---|---|
| 1 — Market Size & Growth | Is the market big enough? Is it growing or shrinking? What is the addressable segment for this client's product — not the whole industry TAM? |
| 2 — Customer Segments & Pain Points | Who actually buys products like this? What specific pains are they describing in reviews, forums, and surveys? What language do they use to describe their problem? |
| 3 — Competitor Landscape | Who are the direct competitors? Who are the indirect ones? What do users love about them? What do users complain about? What is each competitor's pricing model? |
| 4 — Feature Benchmarks | What features exist across the market? Which are table stakes? Which are differentiators? Which are unmet gaps nobody has addressed? |
| 5 — Regulatory & Technical Context | Are there compliance requirements this product must meet? What technical standards, certifications, or data handling rules apply in this domain? |
| 6 — Industry Trends | What is changing in this market over the next 12–24 months? New entrants, AI disruption, regulatory changes, or shifts in user behaviour that will affect this product? |
Not every area will be equally important for every project. A B2B compliance tool needs deep regulatory research. A consumer app needs deep user sentiment research. Use the six areas as a framework and weight your effort accordingly.
Each step has a clear input, a clear AI action, and a clear output. Follow this sequence every time.
Do not start fresh. Feed the AI your Discovery Summary from Module 1 as the first message. This anchors all subsequent research to the right context and prevents the AI from giving you generic industry answers.
Ask the AI to identify market research, analyst reports, and industry data for the specific domain. Prompt it to give you estimates with sources, caveats on data reliability, and the key drivers of market growth or decline. Cross-check any significant data point with a second source.
Give the AI the product description and ask it to identify direct competitors, indirect competitors, and adjacent players. For each, ask it to summarise: what they do, what users say about them, and what pricing model they use. Then do your own spot-check on 2–3 of the most important ones.
For key competitors, ask the AI to analyse publicly available user reviews (G2, Capterra, App Store, Reddit, Product Hunt) and extract the top recurring praises and complaints. Pattern these into themes — not individual comments.
List all features mentioned in the Discovery Summary. Add competitor features from your research. Then classify each feature: Table Stakes (must-have), Differentiator (competitive advantage), or Gap (unmet by existing products). This becomes the foundation of your feature prioritisation.
Synthesise everything into the PRD template. Use AI to help structure and articulate — but every assertion in the PRD must be one you believe and can defend. This document goes to the client. It represents your credibility.
Human in the loop: Before you finalise the PRD v0.1, do a 20-minute review with someone who knows the domain — a senior PM, a Solutions Architect, or a colleague who has worked in that industry. Ask: "What am I missing?" This is the fastest way to catch a blind spot before it becomes a proposal error.
Run these in a single AI session in sequence. Prompt 1 anchors your context. Prompts 2–5 build the research. Prompt 6 assembles the PRD. Prompt 7 converts it into the Scope Definition Document.
The Scope Definition Document feeds directly into Module 3 as the primary estimation input. Every assumption listed here becomes a risk if violated. Read it as the estimation team — can they start scoping without asking you any clarifying questions? If not, add more detail.
One product build scenario and one data/support scenario. Each shows the starting brief, key prompt outputs, and what the feature gap analysis looks like in practice.
Example A · Product Build · Field Service Management
Mid-sized HVAC maintenance company. All dispatch managed via WhatsApp and spreadsheets. Losing approximately 30% of bookings to competitor response time. 40 field technicians, 5 office dispatchers. No existing software.
The global field service management software market is estimated at $6.2B (2023) growing at 11.4% CAGR through 2028. Key growth drivers: labour shortage forcing automation, IoT-connected equipment enabling predictive dispatch, post-COVID acceleration of mobile-first operations.
The "$6.2B market" is not the client's addressable market. The PM's job is to narrow this to the relevant segment — mid-market HVAC SMBs, specific geographies. This is the judgement AI cannot make for you.
| Competitor | Key Notes |
|---|---|
| ServiceTitan | Enterprise FSM for home services. Dominant in US HVAC. Expensive ($500+/user/mo). Praised for job tracking. Criticised for steep learning curve and being overkill for small teams. |
| Jobber | SMB-focused. Simpler UI, cheaper pricing ($50–150/mo). Strong in landscaping and cleaning. HVAC users find it lacks technician routing intelligence. |
| FieldEdge | HVAC-specific. Good equipment history tracking. Limited mobile UX. Often described as "clunky but functional." |
| HouseCall Pro | Strong mobile-first for technicians. Good customer communication. Dispatcher dashboard is weak. |
| Feature | Classification | Rationale |
|---|---|---|
| Job scheduling calendar | TABLE STAKES | Every FSM tool has this. Expected by default. |
| Technician mobile app | TABLE STAKES | Standard across all competitors. |
| Real-time technician location tracking | DIFFERENTIATOR | Offered by enterprise tools only. Mid-market gap. |
| Intelligent job routing (proximity + skill match) | DIFFERENTIATOR | Poorly executed across most competitors. Strong win area. |
| Customer SMS notifications (booking, ETA) | DIFFERENTIATOR | Available but not universal. High user demand in reviews. |
| Equipment service history per site | GAP | HVAC-specific. Most FSM tools handle this poorly. High switching cost once it exists. |
| WhatsApp integration for legacy transition | GAP — Additional | Not in client brief, but migration from WhatsApp is their exact pain. Reduces adoption friction significantly. |
Example B · Data / Support AI · B2B SaaS
B2B SaaS company selling accounting software to SMBs. Support volume grown 3× in 18 months while team grew 20%. Support team of 12 handling 4,000 tickets/month. Target: deflect 40% of tickets without adding headcount. Existing stack: Zendesk, Slack, proprietary billing system.
The conversational AI and chatbot market for customer support is $1.3B (2023), growing at 23% CAGR. Differentiation now comes from quality of knowledge base integration, not NLP sophistication.
ROI on support AI is highly sensitive to ticket complexity distribution. If over 60% of tickets are genuinely Tier-1, deflection targets of 40–50% are achievable. If the mix is more complex, 20–25% is more realistic. The PM must validate ticket type distribution before committing to a deflection target in the proposal.
| Feature | Classification | Rationale |
|---|---|---|
| FAQ / knowledge base Q&A chatbot | TABLE STAKES | Every support AI tool does this. |
| Zendesk integration for ticket creation and escalation | TABLE STAKES | Standard integration — must-have given client stack. |
| Contextual escalation: conversation summary passed to human agent | DIFFERENTIATOR | Consistently missing or poorly executed in reviews. High CSAT impact. |
| Ticket category and deflection analytics dashboard | DIFFERENTIATOR | Strong demand in reviews. Critical for the client to justify ROI internally. |
| Knowledge base gap detection (questions AI couldn't answer) | DIFFERENTIATOR | Not widely implemented. High operational value. |
| Proactive in-app nudges based on user behaviour patterns | GAP — Additional | Not in client brief, but platform has usage data. Could reduce ticket volume upstream. |
Every item must be confirmed before you treat the PRD v0.1 as ready. If you can't check an item, go back and fill the gap.
I have an estimate of addressable market size with at least one credible source cited
I know whether the market is growing, flat, or shrinking — and why
I have confirmed the client's target segment is distinct enough to be viable
I have mapped at least 4 direct competitors and 2 indirect competitors
I know what users consistently complain about across the competitive set
I know what 1–2 unmet needs exist that no current competitor addresses well
Every feature from the Discovery Summary has been classified: table stakes, differentiator, or gap
I have identified at least 1 feature not in the original brief that market data supports
I have a clear out-of-scope list for v1 — and a reason for each item
The PRD has all 7 sections: Problem Statement, Market Context, User Profile, Competitive Position, Feature Set, Out of Scope, Open Questions
Every market data point in the PRD has a source or a stated level of confidence
I have done a 20-minute peer review with someone who knows the domain
I have a deliverables list that names concrete hand-over items — not features, not capabilities, but actual things the team produces
Every assumption the estimate will rest on is written down
I can hand this document to the estimation team and they can begin scoping without asking me clarifying questions
| Tool | Best Used For |
|---|---|
| Claude / ChatGPT | Primary research synthesis, competitor summaries, PRD drafting. Use with web browsing enabled for up-to-date market data. |
| Perplexity AI | Market research with live citations. Reduces hallucination risk on market size data. Good for regulatory context. |
| Google Trends | Validate whether search interest in your product category is growing or declining. Free and fast. |
| G2 / Capterra | Primary source for competitor user reviews. Read the 3-star reviews — they are the most useful. |
| Product Hunt | Understand recent launches in the space and early user reactions. Good for spotting gaps. |
| LinkedIn / Reddit | User communities surface real pain points more honestly than formal review sites. |
| Mistake | How to Avoid It |
|---|---|
| Using total industry TAM instead of the addressable segment | Narrow the market to the segment the client can realistically win. A $10B market is irrelevant if the client is targeting a $50M niche. |
| Treating AI competitor research as ground truth | AI can hallucinate product features and pricing. Always spot-check 2–3 competitors manually on their actual website or G2 page. |
| Listing features without classification | A feature list without table stakes / differentiator / gap labelling is just a wish list. It does not help with prioritisation or pricing. |
| Skipping the Open Questions section | What you still don't know is as important as what you do. Include it in every PRD v0.1. |
| Writing a PRD you cannot defend | If someone asks "why is this feature in the MVP?" you must have an answer beyond "the client asked for it." |
✅ Ready for Module 3 — Price it Right, where you take the PRD v0.1 and Scope Definition Document and use them to build a defensible effort estimate covering timeline, team composition, cost, and AI savings.
How to build a defensible effort estimate — covering workstreams, team composition, timeline phasing, AI savings, and proposal pricing — for any project type and any engagement type.
A senior stakeholder asks for a ballpark. You give a number. Two weeks later that number has become the client's expectation, the internal budget, and the project success criterion — even though you had almost no information when you said it.
A defensible estimate is not the same as an accurate estimate. Accuracy is a function of information you don't yet have. Defensibility is a function of the rigour you applied to the information you do have. You are responsible for defensibility. You are not responsible for information that doesn't exist yet — as long as you named the gap as an assumption.
Every estimate contains facts and assumptions. Good estimators make assumptions explicit. Bad estimators treat assumptions as facts and get surprised when they are wrong.
The most accurate estimates come from the people who will actually do the work — not from a senior PM estimating on their behalf. Get the relevant specialists involved early, even for a few hours.
A single total number hides everything. A phased estimate — who is involved when, at what intensity — forces you to think through the work properly and gives the client a picture they can actually evaluate.
⚠ The most dangerous words in estimation: "It should be roughly similar to the last one." Analogical thinking is a valid starting point. It is not a finishing point. Use it to sanity-check, not to replace bottom-up analysis.
| Document | How it is used in Module 3 |
|---|---|
| Validated PRD v0.1 | Tells you what the product must do and which features are in MVP scope. Informs workstream identification. |
| Scope Definition Document | The primary estimation input. Every workstream, role, and hour estimate maps to a deliverable in this document. |
If Module 2 was skipped (support or augmentation engagement): use your Discovery Summary from Module 1 as the input. Prompt 1 below handles both cases.
Estimation requires knowledge about the work itself and the context it will be delivered in. Missing either produces an estimate that will not survive contact with reality.
Primary workstreams: Architecture · Backend API development · Frontend/UI · DevOps & infrastructure · QA · Security
Specialist roles: Solutions Architect · Backend Engineers · Frontend Engineers · DevOps Engineer · QA Engineers · UI/UX Designer
Primary workstreams: UX design · Mobile development (native or cross-platform) · Backend API · QA (device matrix) · App store release
Specialist roles: Mobile Developers (iOS/Android or React Native/Flutter) · Backend Engineers · UI/UX Designer · QA Engineers
Primary workstreams: Data ingestion & pipeline · Data modelling & warehousing · BI & dashboard layer · Data quality & governance
Specialist roles: Data Engineers · Data Architect · BI Developer · Analytics Engineer · QA Engineer
Primary workstreams: Data ingestion & pipeline · Feature engineering · Model development & training · Model serving & MLOps · Monitoring
Specialist roles: Data Engineers · Data Scientists · ML Engineers · MLOps Engineer · Data Architect
Primary workstreams: LLM integration & prompt engineering · RAG / knowledge base · Evaluation & safety layer · API & product layer · Fine-tuning if required
Specialist roles: AI/LLM Engineers · Prompt Engineers · Backend Engineers · Data Engineers (for RAG) · QA/Eval Engineers
Primary workstreams: API design & integration layer · Data mapping & transformation · Error handling & monitoring · Testing across systems
Specialist roles: Integration Architects · Backend Engineers · QA Engineers · DevOps
Primary workstreams: Infrastructure design · Cloud setup & IaC · CI/CD pipeline · Monitoring & alerting · Security & compliance
Specialist roles: Cloud Architect · DevOps/Platform Engineers · Security Engineer
| Complexity driver | Effect on estimate | Most common in |
|---|---|---|
| Third-party API integrations | Each integration adds design, development, error handling, and testing time. Poorly documented APIs multiply this. | Cloud Native, Mobile, AI/LLM |
| Legacy system dependencies | Undocumented systems, inconsistent data formats, and no test environments create significant hidden work. | Enterprise Integration, Enhancement |
| Data quality issues | Dirty, incomplete, or inconsistent source data adds pipeline cleaning, validation, and reprocessing work. | Data Platform, AI/ML |
| Regulatory / compliance requirements | GDPR, HIPAA, SOC 2 add security architecture, audit logging, and review cycles. | All types in regulated industries |
| Multi-platform delivery | Web + iOS + Android multiplies QA effort significantly. Each platform needs independent testing. | Mobile, Cloud Native |
| Model training data requirements | If labelled training data does not exist, creating it is a significant workstream often missed in early estimates. | Data Platform with AI, LLM |
| Greenfield architecture decisions | When there is no existing system, architecture decisions take longer and have more downstream impact. | New builds across all types |
AI-assisted discovery, research, spec writing, and handoff (Modules 1, 2, 4, 5, 6). Saves time on the PM's hours before and during the build phase. Well-supported by McKinsey research on knowledge work acceleration.
AI-assisted coding tools (Cursor, GitHub Copilot, Claude Code) applied to eligible workstreams. GitHub and DORA research shows 18–26% productivity gains on well-structured coding tasks. Not applicable to architecture, complex state management, or integration debugging.
Better specs and structured handoff (Modules 5, 6, 7) reduce mid-sprint change requests. Accenture and MIT research shows structured requirements reduce rework by 40–60%. This layer grows the more rigorously the spec process is followed.
Steps 1–5 build the raw estimate. Steps 6–8 layer in AI savings and convert the estimate into a client-facing proposal. Do not skip ahead to pricing before the raw estimate is complete.
| Method | When to use | Accuracy | Time to produce |
|---|---|---|---|
| Analogical | Early-stage, quick ballpark needed. Use a similar past project as reference and apply adjustments. | Low–Medium | 1–2 hours |
| Bottom-up | Final proposal before SOW is signed. Break scope into deliverables, estimate each with specialists, then aggregate. | High | 4–8 hours with team |
| Parametric | You have historical data on similar project types. Apply known productivity rates to the scope. | Medium–High | 2–3 hours |
Read your Scope Definition Document. Name both explicitly. Write them at the top of your estimation worksheet. Everything downstream flows from this.
Based on the project type, list every workstream this engagement requires. A workstream is a distinct body of work with its own roles and deliverables. Do not merge workstreams — each needs its own estimate.
For each workstream, identify: which roles are needed, at what seniority level, and how many hours each role requires. Involve the relevant specialists — a senior backend engineer estimating backend hours is more accurate than a PM doing it alone.
Review the complexity drivers from Part 2. For each one that applies, add a justified percentage uplift to the affected workstream hours. Do not apply a blanket uplift to everything — complexity is specific to specific workstreams.
Well-defined enhancement: 10–15%. New build with moderate complexity: 15–25%. Greenfield with significant unknowns or integrations: 25–35%. Write down what the buffer is covering. Named risks from the Scope Definition Document belong in the buffer.
Convert the hours matrix into a phased timeline showing who is involved when. Architects and UX designers are front-loaded, engineers dominant in build, QA ramping up as features complete, DevOps active at start and launch. Map this as a Gantt or phase table — this is the artefact the client sees.
Identify which parts of the estimate are affected by AI tooling. Express each saving layer as a reduction in hours with a stated rationale. The calculator carries the specific percentages — your job is to tell the story clearly.
Convert hours into cost using your rate card. Show gross estimate, AI savings, AI tool costs, net cost, and ROI on the AI investment. Every number must trace back to a line in your estimation worksheet in under 60 seconds.
Human in the loop: Before the estimate leaves the building, do a 30-minute internal review with the delivery lead or solutions architect. Ask them to challenge your workstream coverage, seniority mix, and buffer reasoning. An estimate not challenged internally will be challenged externally — at a much worse time.
Run these in a single AI session in sequence. Prompt 1 loads context from Module 2. Prompts 2–4 build the estimate progressively. Prompt 5 assembles the client-facing proposal narrative.
Clients will ask: "How do I know the savings are real?" Your answer must not be a percentage from a report. It must be a specific, honest explanation of what your team does differently because of AI, where that saves time, and what it does not affect. Know all three layers before you walk into the room.
Example A is a Cloud Native new build. Example B is a Data Platform with AI/ML project. Each shows the full flow from Scope Definition Document to priced proposal with AI savings.
Example A · Cloud Native · New Build · B2B SaaS
Core features: project scheduling, document management, subcontractor coordination, mobile app for field teams. Integrations: accounting system (client API) + weather data feed. 6-city pilot before full launch.
| Workstream | Roles | Seniority |
|---|---|---|
| UX Design | UI/UX Designer | Senior — multi-user, complex info architecture |
| Backend API | Backend Engineers ×2, Solutions Architect | Senior Architect, Mid Engineers |
| Web Frontend | Frontend Engineers ×2 | Mid-level — well-defined component structure |
| Mobile (React Native) | Mobile Developer | Senior — offline sync is complex |
| QA | QA Engineers ×2 | Mid-level |
| DevOps | DevOps Engineer | Senior — multi-tenant infra requires careful design |
| Project Management | PM | Senior PM |
| Workstream | Base | Complexity add | Total | Note |
|---|---|---|---|---|
| UX Design | 180h | +20h | 200h | Multi-tenant UX, role-based views |
| Backend API | 480h | +80h | 560h | Multi-tenancy + 2 integrations |
| Web Frontend | 320h | +40h | 360h | Admin portal scope larger than standard |
| Mobile | 240h | +60h | 300h | Offline sync adds complexity |
| QA | 200h | +40h | 240h | Cross-platform device matrix + regression |
| DevOps | 100h | +20h | 120h | Multi-tenant infra, two environments |
| PM | 180h | — | 180h | Standard PM overhead |
| Buffer (20%) | 392h | Greenfield, 2 integrations, offline mobile | ||
| TOTAL | 2,352h | |||
ROI on AI investment: 7.6× — $27,200 total saving on $3,600 AI tool investment. Layer 2 (development) is the largest saving. Layer 3 (rework) grows as spec discipline improves across the team.
Example B · Data Platform + AI/ML · Retail · Enterprise
Goal: reduce stockouts by 30%, reduce overstock by 20%. 3 years of historical sales data in a legacy ERP. Product master in a separate system. External signals (weather, local events) not currently used.
| Workstream | Realistic | Complexity add | Total | Note |
|---|---|---|---|---|
| Data Engineering | 400h | +120h | 520h | Legacy ERP unknown quality. Largest risk in estimate. |
| Feature Engineering | 160h | +40h | 200h | External data integration adds pipeline complexity |
| ML Modelling | 280h | +40h | 320h | Time-series with external signals — tuning intensive |
| MLOps & Serving | 180h | +20h | 200h | Retraining pipeline + drift monitoring non-trivial |
| BI & Dashboard | 120h | — | 120h | Well-defined once model outputs are stable |
| QA & Validation | 140h | +20h | 160h | Data quality testing adds specialist QA time |
| PM | 160h | — | 160h | Standard PM overhead |
| Buffer (25%) | 370h | High: legacy data risk, ML uncertainty, new tech stack | ||
| TOTAL | 2,050h | |||
AI savings note for Data & AI/ML projects: AI coding tool savings are lower on data engineering work (pipeline logic is bespoke and requires domain understanding) but higher on dashboard and API layer work. The most significant saving on this project type is in PM/PO phase work and rework reduction. The single largest cost risk is data quality in the legacy ERP — surfaced by better upfront requirements, not discovered after the first pipeline run.
Every item below must be confirmed before the estimate leaves your hands. Gaps surface during delivery — at ten times the cost.
I have confirmed the project type and engagement type explicitly — not assumed
Every deliverable in the estimate maps to a named item in the Scope Definition Document
I have not estimated deliverables that are in the out-of-scope list
I have mapped every workstream — including DevOps, QA, and Project Management (commonly missed)
I have involved at least one specialist in validating hours for the most complex or highest-risk workstream
I have checked the complexity drivers list and applied adjustments where they apply — not as a blanket uplift
My buffer percentage is justified with named reasons — not picked by habit
I can explain to a client what the buffer covers and why
I have a phased timeline showing who is involved when — not just a total duration
The Gantt or phase table is clear enough that a client can read it without my verbal explanation
I have applied AI savings to the correct layers — not as a blanket discount on the total
I can explain the three-layer savings story (PM/PO phase, development, rework) in plain language to a client
AI tool costs are included in the net cost calculation
I have anticipated the top 3 client pushback questions on AI savings and have honest answers ready
Every number in the proposal traces back to a line in my estimation worksheet in under 60 seconds
| Project type | Typical core team |
|---|---|
| Cloud Native Product | Solutions Architect · Backend Engineers (×2–3) · Frontend Engineers (×1–2) · DevOps Engineer · QA Engineers (×1–2) · PM |
| Mobile Product | Mobile Developers (×1–2) · Backend Engineers (×1–2) · UI/UX Designer · QA Engineer · PM |
| Data Platform + Analytics | Data Architect · Data Engineers (×2) · BI Developer · Analytics Engineer · QA Engineer · PM |
| Data Platform + AI/ML | Data Architect · Data Engineers (×2) · Data Scientists (×2) · ML Engineer · QA Engineer · PM |
| AI / LLM Product | AI/LLM Engineers (×1–2) · Backend Engineers (×1–2) · Data Engineer (for RAG) · QA/Eval Engineer · PM |
| Enterprise Integration | Integration Architect · Backend Engineers (×2) · QA Engineer · PM |
| Platform / Infrastructure | Cloud Architect · DevOps/Platform Engineers (×2–3) · Security Engineer · PM |
| Engagement type | Typical buffer | What the buffer covers |
|---|---|---|
| New build — well-defined | 15–20% | Normal variation, minor scope creep, integration friction |
| New build — complex / greenfield | 25–35% | Architecture unknowns, technology risk, significant integrations, client dependency delays |
| Enhancement on existing system | 10–15% | Codebase ramp-up time, regression risk, hidden dependencies |
| Support and maintenance | 10% | Ticket volume variance, escalation handling time |
| Team augmentation | N/A | Buffer is in the rate, not the hours |
✅ Ready for Module 4 — Draw the Journey, where you move from proposal to delivery — building user flows, UX blueprints, and screen-level designs using AI tools.
How to use AI to build user flows, UX blueprints, and screen-level designs — turning a validated PRD into a design-ready artefact that engineering can actually implement.
Design in a delivery context is about reducing ambiguity. Every screen you produce, every flow you map, every edge case you document reduces the number of decisions engineering has to make without you.
Design that leaves interpretation to engineering is not finished design. It is deferred decision-making with a higher price tag. Every ambiguous wireframe note that says "TBD" or "as discussed" is a future sprint disruption waiting to happen.
You provide the PRD v0.1 and Scope Definition Document. You run a brief alignment session with the designer. You confirm the screen inventory and user personas before a single wireframe is opened.
You review the UX Blueprint for completeness — are all user journeys covered, are edge cases addressed, does this match the PRD scope? You do not tell the designer how to lay out a screen. You tell them when something is functionally missing.
No design artefact goes to engineering without PM sign-off on functional completeness. Visual polish is the designer's responsibility. Functional completeness is yours.
| Engagement type | Does Module 4 apply? |
|---|---|
| New product build | Always applies — full UX journey from IA through to Figma-ready designs |
| Significant new feature | Usually applies — new screens and changed flows need UX design work |
| Enhancement or iteration | Partially applies — focus on wireframe and handoff sections only |
| Support or maintenance | Does not apply — skip to Module 5 |
| Team augmentation | Does not apply — client team owns design. Skip to Module 5 |
Design cannot begin productively until the PM and designer share a clear, aligned picture of what is being built. The inputs exist — they came from Modules 1 through 3.
| What the designer needs | Where it comes from |
|---|---|
| Who the user is | Not in general terms — specifically. Role, context, technical literacy, device environment, frequency of use. From Module 1 Discovery Summary and PRD v0.1. |
| What the user is trying to accomplish | The primary jobs to be done. What does a successful session look like from the user's perspective? Drives flow design and information architecture decisions. |
| What screens or surfaces exist | The complete list of pages, screens, views, and states in scope — not features, screens. The screen inventory comes from the Scope Definition Document deliverables list. |
| What is explicitly out of scope | The designer must know what not to design. Out-of-scope items create scope creep when they appear in wireframes without agreement. |
| What constraints exist | Technology stack (affects what interactions are possible), brand guidelines, accessibility requirements, responsive breakpoints, and any third-party component libraries in use. |
A map of how a user moves through the product — screens, decisions, entry and exit points, error paths, and edge cases. Produced before wireframes. The skeleton that everything else hangs from.
A structured document listing every screen by name and ID, its purpose, primary user action, key components, entry points, exit paths, and edge cases. No visual design — structure only.
Lo-fi visual representation of each screen — layout, hierarchy, content placement. No colour, no brand. Produced in Figma. Reviewed and approved before hi-fi design begins.
Clickable Figma file or Lovable-generated prototype showing real interactions and navigation. Used for client validation, usability testing, and engineering briefing.
✓ Must contain
✗ Must not contain
This is the standard process a PM and designer follow from receiving the PRD to handing a design-ready artefact pack to engineering. AI accelerates several steps significantly — but none can be skipped.
PM and designer meet for 60–90 minutes. Walk through the PRD v0.1 and Scope Definition Document together. Output: confirmed screen inventory, confirmed persona list, and a list of open design questions that need answers before work begins. Do not skip this session.
Before any user flow is drawn, produce a complete list of every screen in scope. Name each one, give it a screen ID (S-01, S-02...), and write a one-sentence description of its purpose. If a screen is not on this list, it should not be designed. AI is excellent at this — Prompt 1 does it.
For each primary user journey, map the complete flow: entry point, happy path, decision points, alternate paths, error states, empty states, and exit points. Start with the happy path — then add alternate flows, then edge cases. AI generates edge cases designers often miss under time pressure — Prompt 2 covers this.
For every screen, document: screen ID and name, entry points, primary user action, key UI components (structurally, not visually), exit paths, and edge cases specific to this screen. PM reviews and signs off on the Blueprint before wireframes begin. This is Gate 1.
With the Blueprint signed off, the designer produces lo-fi wireframes for every screen. No colour, no brand at this stage — layout, hierarchy, content placement only. Wireframes reference Blueprint screen IDs so every wireframe traces back to a documented screen purpose.
Every screen has multiple states: default, hover, focus, active, disabled, loading, empty, error, success. All states must be designed before the wireframe is considered complete. This is the most commonly missed step — and the most common cause of rework during engineering.
Two paths: (A) Figma prototype — link wireframe screens together for a clickable demo. (B) Lovable / v0 / Emergent — feed the Blueprint and wireframes into a tool that generates an interactive HTML prototype. Best for client demos before engineering begins.
Before anything goes to engineering, the PM reviews against three criteria: (1) Does every screen in the screen inventory have a wireframe? (2) Does every wireframe have all required states? (3) Does the design match the PRD scope — nothing missing, nothing added? This is Gate 2. Design changes after this require change control.
The complete design artefact pack: UX Blueprint, Figma file with all wireframes organised by screen ID, prototype link if applicable, and a design QA checklist for engineering. The designer walks the engineering lead through the Figma file. Non-obvious design decisions are annotated with rationale notes.
Human in the loop: The PM signs off at Step 4 (Blueprint) and Step 8 (wireframes). These are the two gates where scope drift is most likely — features added silently, screens out of scope quietly included. Review with the PRD and Scope Definition Document open side by side.
Each prompt maps to a specific step in the process. Prompt 1 feeds directly from your PRD v0.1 and Scope Definition Document from Module 2. Carry your prior outputs forward — do not start fresh.
Review the Blueprint against the PRD v0.1 and Scope Definition Document before wireframes begin. Every screen must trace back to an agreed deliverable. Nothing added, nothing missing.
No single project needs all of these. Choose based on what the engagement requires. Each tool has a specific job — using the right tool for the right step matters.
| Tool | What it does | When to use it |
|---|---|---|
| Figma | The industry standard design tool. All wireframes, hi-fi designs, and interactive prototypes live here. Every design engagement uses Figma. Non-negotiable. | Steps 5, 6, 9 — wireframes, states, handoff |
| Figma AI | AI-assisted features within Figma: auto-layout suggestions, content generation for realistic mockups, design critique. Use for speed inside Figma — not a replacement for design thinking. | Step 5 — wireframe production |
| Claude / ChatGPT | Primary AI tool for this module. Use for all six prompts: screen inventory, user flow mapping, Blueprint generation, design critique, state inventory, handoff spec. | Steps 2, 3, 4, 5, 6, 9 |
| Whimsical AI | Fast AI-assisted flowcharting. Generates user flow diagrams and sitemaps from text descriptions. Useful for quickly visualising flows before translating them into Figma. | Step 3 — user flow mapping |
| Mermaid | Text-based diagram syntax that generates flow diagrams. Good for flows that can be version-controlled and embedded in documentation. Claude can generate Mermaid diagrams directly. | Step 3 — technical teams |
| Lovable / v0 / Emergent | Generative prototype tools. Convert a Blueprint or wireframe description into a working interactive HTML prototype. Best for client demos and stakeholder validation before engineering begins. | Step 7 — interactive prototype |
| Uizard | Converts hand-drawn sketches or screenshots into editable wireframes. Useful for rapid wireframe generation from rough sketches in early alignment sessions. | Step 5 — rapid wireframing |
| Dovetail AI | Research synthesis tool. Analyses user interview transcripts and feedback to extract themes, pain points, and insights. Relevant if the engagement includes user research before design. | Step 1 — if user research exists |
On using Lovable and generative tools: Feed your UX Blueprint and screen inventory into the tool and iterate on the generated output. Treat the output as a client communication tool, not an engineering artefact. The Figma wireframes remain the source of truth for implementation.
Example A is a Cloud Native mobile app. Example B is a data-heavy AI dashboard. Example C is a real Ideas2IT story — a PO who used AI no-code tools to build client mockups without a dedicated designer.
Example A · Cloud Native + Mobile · Field Service Management
6-person team · 6-month build · 40 field technicians + 5 supervisors · iOS and Android · Offline capability required
| State | What the user sees | Trigger |
|---|---|---|
| Default | Job details fully loaded | Technician opens a job with complete data available |
| Loading | Skeleton screen with loading indicators | App is fetching job details from server |
| Offline — cached | Full details with offline banner at top | Technician offline but job was cached |
| Offline — uncached | Partial data with "Full details unavailable offline" | Technician offline and job was not cached |
| Error — load failed | "Unable to load job details. Tap to retry." | Server error or timeout on data fetch |
| Completed | Read-only view with completion timestamp and photos | Job has been marked complete |
| Reassigned | Banner: "This job has been reassigned" | Supervisor reassigned the job while technician was viewing it |
Example B · Data Platform + AI/ML · Retail · Enterprise
200+ stores · AI-generated recommendations · Design must prioritise information hierarchy, data trust signals, and clear recommended actions over visual complexity.
When designing interfaces that display AI-generated recommendations, the design must explicitly address three things:
The AI's recommended action must be the most prominent element on the screen — not buried in a data table.
Every recommendation links to an explanation: what signals drove it — seasonal trend, recent sales velocity, external event? This is what separates a trusted AI tool from a black box.
S-04 must include a mandatory reason code field when overriding. This is the training signal that improves the model over time. Not a UX nicety — a functional requirement.
Three wireframe requirements unique to AI output screens: (1) Confidence indicators on every recommendation — high/medium/low badge or percentage. (2) Override with mandatory reason code capture — this trains the model. (3) Explainability panel linked from every recommendation. These must be in scope and in the wireframes before engineering starts.
Written by Mithun Chandar V, PO at Ideas2IT. This is a live example of the Module 4 mindset applied in practice — and proof that AI no-code tools have genuinely changed what a PO can produce independently.
Mithun needed to produce UI mockups for a client engagement but did not have a dedicated designer available at that stage of the project. Rather than waiting or producing rough sketches, he used a combination of AI no-code tools to produce realistic, interactive-feeling mockups that the client could review and validate — before engineering started.
The output: client-validated screens in a fraction of the time it would have taken through a traditional design process. The approach has since been shared internally as a reusable playbook.
The core insight from Mithun's playbook is the same mindset this module teaches — but applied to the tooling layer:
"Traditional design tools require you to place every element manually. AI no-code tools require you to describe what you want clearly and guide the output. You are the director — not the camera operator."
The shift matters: a PO who understands this can produce client-reviewable mockups in hours, not days. A PO who does not understand it produces generic, unusable AI output and concludes the tools do not work.
| Tool | Used for | Best for |
|---|---|---|
| Lovable | Building realistic-looking screens with interactivity. Primary tool for screens that need to feel like a real product. | When stakeholders need to click through something that feels real |
| v0 by Vercel | Specific UI components — tables, cards, modals — that would be handed off to a developer. Clean React + Tailwind output. | When the output needs to be close to production-ready |
| Emergent Mind | Early concept exploration. Rapid generation of multiple screen concepts to identify direction before committing to one approach. | Ideation — when you have a rough idea and want to see it visualised fast |
| Bolt | Full app prototype with linked flows — not just individual screens but a connected user journey. | When you need to demo a working UI to stakeholders or investors |
| Framer AI | High-polish landing page or marketing site for a client presentation. Design quality and visual polish matter more than functionality. | When you want to publish a live URL quickly |
Rather than writing one massive prompt, Mithun built mockups in layers — the same way the six prompts in this module are structured:
Start with the big picture — dashboard, form, list, onboarding. Give the tool just enough to understand the shape of the screen.
What must appear on this screen? What can the user do? Feed the user story or Blueprint entry — not the full PRD.
Minimal, bold, data-heavy, friendly — give the tool a direction. Reference a known product if it helps: "similar to Stripe's dashboard" or "clean like Notion".
Do not start fresh when something is wrong. Tell the tool exactly what to change and what to leave alone: "Keep the overall layout. Move the chart to the top. Remove the sidebar."
Mithun's playbook includes a pre-flight check before opening any tool. This maps directly to the design brief alignment session in Step 1 of this module's process:
When is the mockup done? When stakeholders can understand the screen without your verbal explanation. When developers know what to build — even if exact styling is TBD. A mockup is a communication tool, not a finished product. Ship it when it answers the question, not when it looks perfect.
When NOT to use these tools: When you need pixel-perfect final designs for engineering handoff — use Figma. When you have zero clarity on the user or the goal — do discovery first (Module 1). When a simple whiteboard sketch would communicate the idea just as well. AI no-code tools produce the most value when you have a clear Brief and need to move fast. They produce the least value when the problem is still unclear.
Design that goes to engineering incomplete is more expensive to fix than design that takes one more day to finish.
A design brief alignment session has been run between PM and designer before wireframes started
The designer received the PRD v0.1 and Scope Definition Document as their primary inputs
All open design questions from the alignment session have been answered
Every screen in scope has been identified, named, and given a screen ID
Every screen traces back to a feature or deliverable in the PRD or Scope Definition Document
No screens are included that are not in the agreed scope
Every primary user journey has a complete user flow documented
Each flow includes the happy path, at least two alternate flows, error states, and empty states
The Blueprint has an entry for every screen in the inventory
Every Blueprint entry has: entry points, primary action, key components, exit paths, and edge cases
The PM has reviewed the Blueprint and signed off before wireframes began
A wireframe exists for every screen in the screen inventory
All interaction states are designed: default, loading, empty, error, success, disabled
Mobile and responsive states are designed for any screen that will be used on mobile
For AI/ML dashboards: confidence indicators, override capability, and explainability panels are in the wireframes
Figma file is organised by screen ID — no unnamed or unlabelled frames
All spacing, states, and interactions are annotated in Figma
An engineering walkthrough has been scheduled or completed
How to convert a UX Blueprint into atomic, testable specifications and Functional Specification Documents (FSDs) that engineering can implement directly — without interpreting, assuming, or asking.
The most common mistake in spec writing is treating it as documentation — a record of what was discussed. A spec is a contract between the product team and the engineering team. It defines exactly what will be built, exactly how it will behave in every situation, and exactly what done looks like.
The test of a good spec is simple: can a developer who has never spoken to you about this feature implement it correctly from the document alone? If the answer is no — the spec is not finished.
Owns the writing of the FSD. Takes the UX Blueprint from Module 4 and converts each screen into a structured specification. Owns the accuracy of every field, rule, and behaviour described.
Does not write the FSD but is accountable for its completeness and correctness. Reviews every FSD before it is locked, confirms it matches the PRD scope, and signs off before engineering sees it.
Reviews FSDs for technical feasibility before lock, not before they are written. The spec defines what. Engineering defines how. If engineering is writing the spec, something has gone wrong upstream.
Should receive FSDs as they are written — not at the end. Test cases are written from specs. The earlier QA reads the spec, the earlier they catch ambiguity that would surface as a bug.
"The email field validates on blur. If the format is invalid, an inline error appears: 'Please enter a valid email address.' The submit button remains disabled until all required fields pass validation."
"The form validates user input."
What happens when input is invalid, the network fails, permission is denied, or the state is empty. These fill sprint reviews with bugs when missing.
Describes only what happens when everything works. Ignores error states, edge cases, and permission scenarios.
Each line covers one behaviour, one rule, one element. Each is independently testable. QA can write one test case per line.
Tries to describe multiple behaviours together. Cannot be tested individually. QA cannot trace a test case to a specific line.
✓ AI is excellent at
✗ AI cannot replace
An FSD is structured around screens, not features. One FSD section per screen, using the screen IDs from the Module 4 Blueprint. Everything inside each section follows a consistent seven-part structure.
| # | Section | What it contains |
|---|---|---|
| 1 | Screen overview | Screen ID, screen name, purpose in one sentence, user persona, feature area. The orientation block — tells anyone reading the spec where they are and why this screen exists. |
| 2 | Entry conditions | What must be true for a user to arrive at this screen. Authenticated? Specific role? A prior action completed? Defines the precondition for every test case written against this screen. |
| 3 | UI component inventory | Every UI element on the screen: labels, input fields, buttons, dropdowns, data tables, navigation elements, icons. Named consistently with the wireframe and Figma file. No visual design detail. |
| 4 | Functional behaviour per component | For each interactive component: what it does, its states, what triggers each state, what happens on interaction, what validation applies. Atomic — one behaviour per line. |
| 5 | Validation rules | All field-level and form-level validation rules: which fields are required, accepted format, character limits, when validation fires, exact error message for each failure. |
| 6 | Business rules | Rules that govern the feature beyond UI behaviour. Marked as CONFIRMED, INFERRED, or ASSUMPTION. Every ASSUMPTION must be confirmed before spec lock. |
| 7 | Exit conditions and navigation | What happens after the user completes the primary action. Where do they go? What state does the system enter? Is there a confirmation? An undo? What if they navigate away mid-task? |
Written in Given / When / Then format. Each criterion covers exactly one testable behaviour. If a QA engineer cannot write one test case directly from one criterion, it is not atomic.
Given the user is on the login screen
When they enter an email without the @ symbol
Then an inline error appears: "Please enter a valid email address"
And the Login button remains disabled
Specific · testable · one behaviour · exact copy specified
"The form should validate all fields correctly before submission."
"Correctly" is not testable · which fields? · what validation? · what happens on failure?
| Missing element | Why it matters |
|---|---|
| Empty states | What does the screen show when there is no data? Every list, table, and data view needs an empty state message and a suggested action. A blank white screen is a bug, not a design decision. |
| Permission states | What does a user without the required role see when they arrive at a restricted screen? An error page? A redirect? A disabled view? Almost always missing from first-draft specs. |
| Loading states | What does the user see while data is being fetched? How long before a timeout? What happens on timeout? |
| Concurrent action conflicts | What happens when two users act on the same record simultaneously? These cause the worst production bugs and are almost never in the spec. |
| Mobile-specific behaviour | Touch targets, collapsed navigation, swipe interactions, keyboard behaviour when a field is focused — all spec items, not design items. |
Steps 1–4 produce the draft FSD. Steps 5–7 validate it, fill gaps, and lock it. Do not lock a spec until all seven steps are complete.
One FSD per product area or feature group — not one per screen (too many documents). Within each FSD, one section per screen using the screen IDs from the Module 4 Blueprint. The screen IDs become your section headings. Figma screen names must match FSD section names exactly — no synonyms, no paraphrasing.
Paste the Blueprint entry for each screen into AI and run Prompt 1. AI generates a structured first-draft FSD section covering UI components, functional behaviour, states, and validation rules. This first draft will be approximately 60–70% complete. It will have the structure right and obvious behaviours covered. It will miss business rules and anything requiring client knowledge.
Review the AI draft and identify every place where a rule depends on business logic that only the client or domain expert knows. Make a list of every unanswered business rule question. Go back to the client or client-side PO and get explicit answers. Do not assume. Do not infer from past projects.
For every functional behaviour in the spec, write one or more acceptance criteria in Given/When/Then format. Use Prompt 3 to generate a first set, then review and edit. If you cannot write a testable criterion for a behaviour, the behaviour is not specified precisely enough.
Share the draft FSD with the engineering lead before it is locked. Ask one specific question: "Is there anything in this spec that is technically impossible or significantly more complex than the estimate accounts for?" Engineering flags constraints — they do not rewrite the spec.
Share the draft FSD with QA before lock. Ask QA to write one test case for each acceptance criterion. If they cannot write a test case, the criterion is not testable — which means the spec is not complete. This is the most powerful quality check in the entire process.
PM reviews against three things: (1) Does every section trace back to a Blueprint screen and PRD feature? (2) Are all business rules explicitly confirmed — not assumed? (3) Does the acceptance criteria set cover all scenarios the client will use to evaluate the product? When PM signs off, the spec is locked with a version number and lock date.
Spec lock is a formal event, not a formality. Once locked, any change requires: a written change request, impact assessment on timeline and estimate, PM and client sign-off, and a version number bump on the FSD. Teams that treat spec lock casually end up with engineering building from different versions of the truth.
Prompt 1 generates the FSD first draft. Prompts 2–4 fill specific gaps. Prompt 5 generates the QA test case set. Run them in sequence for each screen.
Prompt 5 saves one to two days of QA planning time per feature. Use it as the QA team's starting point — they review, extend, and own the final test case set. Do not hand it directly to engineering without QA review.
The five prompts in Part 4 are the core AI layer — these tools are the environment they run in. No single project needs all of them.
| Tool | How it is used in Module 5 | When to use it |
|---|---|---|
| Claude / ChatGPT | Primary tool for all five prompts: FSD first-draft, acceptance criteria, business rules extraction, edge case finding, QA test case generation. Use with web browsing disabled when working with client data — paste only anonymised content into public AI tools. | All five prompts — throughout |
| Notion / Confluence | Document management for FSDs. FSD sections are living documents until lock — they need version history, comment threads, and review workflows. Notion AI can assist with reformatting and consistency checking across multiple sections. | Drafting, review, and lock management |
| Jira / Linear | If acceptance criteria are linked to engineering tickets, write them directly in the ticket description. Many teams paste individual acceptance criteria into the corresponding Jira story — this creates traceability between spec and implementation. | Acceptance criteria — linked to tickets |
| Google Sheets | Simple, effective tool for the UI component inventory and validation rules table. A shared spreadsheet where every component, its states, and validation rules are listed in rows is easier to review than a Word table. Engineering and QA can comment directly. | UI component inventory · validation rules |
| Zeplin / Supernova | Design-to-spec handoff tools. Connect Figma screens to spec annotations. When the FSD references a screen component, the spec can link directly to the Figma frame — showing and telling simultaneously. | Linking FSD components to Figma screens |
| Dovetail AI | If the engagement included user research or usability testing, Dovetail analyses session feedback and surfaces patterns that should inform spec decisions — particularly acceptance criteria around user behaviour and error recovery. | If user research is available |
AI tool hygiene when writing specs: FSDs often contain client-specific business rules, proprietary workflows, and domain data. Before pasting any spec content into a public AI tool, replace client names, product names, and any identifying information with generic placeholders. Use the anonymised version for prompt input. Restore the actual content in your local document. This applies to all five prompts in this module.
Example A specs the Job Status Update screen (S-04) from the field service app. Example B specs the Order Accept / Modify / Override screen (S-04) from the demand forecasting dashboard.
Example A · Mobile · Field Service App
Field technician updates job status through its lifecycle. States: En Route → On Site → Complete. Photo required for completion. Offline queuing required.
Screen ID: S-04 | Screen: Job Status Update | User: Field Technician | Feature area: Job management
Purpose: Allows a field technician to update the status of an assigned job through its lifecycle from assignment to completion.
Given the technician is viewing a job with status "Assigned"
When they tap "En Route"
Then the job status changes to "En Route"
And a timestamp is recorded
And the supervisor dashboard reflects the new status within 5 seconds
Given the technician is viewing a job with status "On Site"
When they tap "Complete" and no photo is attached
Then the system displays: "A photo is required before marking this job complete"
And the Complete button remains disabled
Given the technician has no network connectivity
When they tap any status update button
Then the action is queued locally
And a banner displays: "Status update saved — will sync when connected"
And the status in the app updates immediately to reflect the queued state
Given the technician tapped a status update while offline
When connectivity is restored
Then the queued status update is posted to the server automatically
And the banner disappears
And the technician receives a silent confirmation that sync was successful
Example B · Data Platform + AI/ML · Demand Forecasting
Purchasing manager accepts, modifies, or overrides an AI-generated order recommendation. Three critical spec requirements unique to AI output screens: confidence display, modification constraints, override reason capture.
| Missing scenario | Action |
|---|---|
| User modifies quantity then navigates away without submitting — is the modification discarded? Is there an "unsaved changes" warning? | CONFIRM WITH CLIENT |
| User's session expires while the Override reason dropdown is open — unsaved changes discarded, redirect to login, on return land on S-03 not S-04 | ADD TO SPEC |
| Recommendation already acted on by another user (multi-buyer scenario) — does S-04 show "This recommendation has already been processed"? | CONFIRM WITH CLIENT |
| Model returns no confidence value (null) — display "Confidence unavailable" in place of the badge | ADD TO SPEC |
Why override reason capture is non-negotiable in AI products: The reason code is not a UX nicety — it is the training signal that improves the model over time. If the override reason is not captured, the model cannot learn from user corrections. This is a business requirement that must be in the spec and in the acceptance criteria before engineering starts.
A spec locked with known gaps is not a locked spec — it is a deferred problem. Every item below must be confirmed before the FSD is marked as locked and handed to engineering.
The FSD is organised by screen ID using the same IDs as the Module 4 Blueprint and Figma file
Every screen in the Blueprint has a corresponding FSD section — no screens missing
Screen names are consistent across Blueprint, FSD, and Figma — no synonyms or paraphrasing
Every screen section has all seven required parts: overview, entry conditions, UI inventory, functional behaviour, validation rules, business rules, exit conditions
Every interactive component has its functional behaviour documented to the atomic level
All interaction states are covered: default, loading, empty, error, success, disabled
The five commonly missed elements have been addressed: empty states, permission states, loading states, concurrent conflicts, mobile-specific behaviour
Every business rule is marked as CONFIRMED, INFERRED, or ASSUMPTION
Every INFERRED and ASSUMPTION rule has been reviewed with the client or client-side PO
No ASSUMPTION-marked rule remains in the spec at lock — all have been confirmed or removed
Every functional behaviour has at least one acceptance criterion in Given / When / Then format
Every acceptance criterion is independently testable — no compound criteria
No criterion contains the words "correctly", "appropriately", "properly", or any vague qualifier
Engineering feasibility review completed — all technical constraints flagged and resolved
QA has attempted to write test cases from the acceptance criteria — any criteria QA could not test have been rewritten
PM has reviewed the complete FSD against the PRD and Scope Definition Document
PM has signed off — spec is formally locked with a version number and lock date
How to run cross-team validation with design, engineering, QA, and compliance — lock the spec, manage change control, and produce a signed-off Spec Pack that everyone has agreed to.
Most teams think alignment means running a review meeting and asking "does everyone agree?" That is not alignment — that is a show of hands. Real alignment means every stakeholder has read the spec, raised their concerns, had those concerns resolved or documented, and signed off on a specific version.
Alignment is not consensus. You are not trying to get everyone to love the spec. You are trying to get everyone to agree that this is what we are building, that their concerns have been heard, and that they will not raise these same concerns again mid-delivery.
The cost of misalignment is not discovered at the review meeting. It is discovered three sprints into development when engineering builds something the client never agreed to, or QA tests against different acceptance criteria than engineering implemented, or a compliance requirement surfaces that nobody flagged in the spec. By that point the cost of fixing it is five to ten times what it would have cost to align properly upfront.
Is the spec consistent with the wireframes? Does every screen in the FSD match the Blueprint? Are there any spec decisions that conflict with design decisions already made?
Is every specified behaviour technically feasible within the estimated effort? Are there hidden dependencies, integration constraints, or performance implications the spec does not account for?
Does every acceptance criterion have a clear pass/fail condition? Are test environments and test data available for all scenarios? Are there spec behaviours that are impossible to test in the current setup?
Does the product as specified meet all applicable regulatory requirements? GDPR, HIPAA, accessibility (WCAG), financial regulation, data residency — any requirement flagged in the Discovery Summary must be verified against the locked spec.
The Spec Pack is not just the FSD. It is a complete, version-controlled artefact set that any team member — design, engineering, QA, or compliance — can use as their single source of truth.
All FSD sections from Module 5, version-numbered, with lock date. One version for each feature group or product area in scope.
The signed-off Blueprint from Module 4. Confirms that every spec screen traces back to a designed and approved Blueprint entry.
A consolidated view of all acceptance criteria across all screens in one document. Used by QA as their master test planning reference.
A record of every change request raised — what was requested, who raised it, what the decision was, and whether the spec was updated as a result.
A table showing every PRD feature maps to a Blueprint screen, FSD section, acceptance criteria, and QA test case. Gaps in this chain indicate coverage issues.
Written confirmation from each stakeholder group. An email thread is acceptable. A signature page is better. A verbal "yes" in a meeting is not sufficient.
Every deliverable in this course traces forward and backward. A broken link means something was built that was not agreed, or something was agreed that was not built.
The traceability matrix makes this chain visible. It is not bureaucracy — it is the mechanism that allows a PM to answer "was this feature tested?" or "where did this requirement come from?" in under two minutes.
Steps 1 and 2 prepare for alignment. Steps 3 and 4 run the reviews. Step 5 locks and packages.
Before any review meeting, assemble a complete draft of the Spec Pack: locked FSD sections, Blueprint, acceptance criteria register, and a traceability matrix showing coverage. Share with all reviewers at least 48 hours before the review session. Reviewers who have not read the spec before the meeting cannot contribute a meaningful review. A review session where people are reading the spec for the first time is not a review — it is a first read with an audience.
Before any synchronous session, ask each stakeholder group to review the spec independently and log their comments. Comments should be logged as: (A) Blocking — must be resolved before implementation, (B) Non-blocking — discuss but will not hold up the spec, or (C) Noted — acknowledged, no change needed. This pre-review step makes the synchronous session dramatically more efficient.
A focused working session — not a presentation. Walk through blocking comments only. For each: what is the issue, what are the options, what is the decision, who owns the update. Time-box aggressively — a medium-sized product alignment session should take 90 minutes, not a full day. If you are running over, the spec needed more work before the session.
Every blocking comment that results in a spec change is a change request. Log it: what changed, in which FSD section, from which version to which version. Update the FSD, bump the version number, and circulate the updated section to the reviewers who raised the blocking comment. Do not circulate the full spec again — only the changed sections.
When all blocking comments are resolved and confirmed, the PM formally locks the spec. The lock means: no further changes without a formal change request, a new version number, and PM sign-off. Assemble the complete Spec Pack with the final version of all artefacts, the sign-off record, and the change control log.
On change control after spec lock: every post-lock change has a cost — not just the spec update, but the ripple. A spec change may require a Blueprint update, new test cases, re-review by the affected stakeholder group, and an impact assessment on the effort estimate. The PM's job is not to block change — it is to make the cost of change visible before anyone agrees to it.
AI is most useful in this module for preparing alignment materials, synthesising review feedback, and managing change control documentation. The alignment conversation itself is human — but the before and after can be significantly accelerated.
Example A shows a clean alignment run with one blocking issue surfaced and resolved. Example B shows a post-lock change request arriving mid-engineering and how the PM manages it.
Example A · Field Service App · Cross-Team Alignment
Engineering raised one blocking issue. Design confirmed Blueprint alignment. QA confirmed all acceptance criteria are testable.
BLOCKING — 1 item
Conflict resolution on offline sync not specified
Engineering cannot implement without a decision. Section affected: S-04 functional behaviour, offline queuing rules.
Options: (A) Last write wins — most recent timestamp takes precedence · (B) Server wins — supervisor update always takes precedence · (C) Conflict alert — both parties notified, manual resolution required
NON-BLOCKING — 2 items
QA requests test data specification — what job states need to exist in the test environment before offline sync tests can run.
Design notes that the offline banner copy differs between S-04 and S-06 — suggest standardising to one string.
Alignment Session Agenda
Item 1 (15 min): Offline conflict resolution — PM to present three options, team to decide.
Item 2 (5 min): Offline banner copy — design to propose one string, PM to confirm.
Estimated total: 20 minutes. No further items require synchronous discussion.
Decision made in session: Server wins — the supervisor's update takes precedence. The technician's offline action is discarded if a conflict exists, and the technician receives a notification: "This job was updated while you were offline. Your action was not applied."
FSD updated with this rule. Version bumped from 1.0 to 1.1. Engineering confirmed version 1.1. Spec locked. Total alignment session: 20 minutes.
Example B · Demand Forecasting Dashboard · Post-Lock Change
Three weeks into engineering, the client's Head of Supply Chain requests a new feature: the ability to bulk-accept all recommendations in a category with one click.
Change accepted for next sprint with a formal change order signed by the client confirming the additional 24–30 hours are in scope and the delivery timeline adjusts by one week. Spec updated to v2.0. Change control log updated. Client confirmation received in writing before engineering begins.
The principle demonstrated here: The PM did not agree to the change in the meeting. The PM assessed the full impact using Prompt 3 first — then presented it to the client with the cost made explicit. The client then made an informed decision. This is what good change control looks like.
Every item must be confirmed before the Spec Pack is handed to engineering. A Spec Pack with known gaps is not a Spec Pack — it is a deferred problem.
The Spec Pack contains all six required artefacts: locked FSD set, Blueprint, acceptance criteria register, traceability matrix, change control log, sign-off record
Every FSD section has a version number and lock date
The traceability matrix shows full coverage — no PRD feature without a spec section, no spec section without a PRD feature
All blocking review comments have been resolved and confirmed by the reviewer who raised them
All post-lock change requests are logged with decision and outcome — even if the decision was to reject the change
Written sign-off has been received from: design, engineering lead, QA lead, and any applicable compliance or legal reviewer
Engineering has confirmed they have received and read the final locked version — not a draft
A shared location exists where all team members can access the current Spec Pack version — no spec documents distributed only by email
Module 6 output → Module 7: Signed-off Spec Pack · Locked FSD set · Signed-off UX Blueprint · Acceptance criteria register · Traceability matrix · Change control log · Written sign-off from all stakeholder groups
How to package all finalised artefacts into a single, version-controlled, development-ready source of truth — so engineering can start building with zero ambiguity and zero missing context.
The most damaging myth in product delivery is that the PM's job is done when the spec is signed off. If the handoff is clean, engineering starts fast, builds confidently, and produces what was specified. If the handoff is messy, the first two weeks of the build are spent reconstructing context the PM already had.
The test of a clean handoff: can a new developer who joins the team in week three of the build, having missed every conversation up to that point, get fully up to speed from the Dev-Ready Package alone in half a day? If not — the package is not complete.
The spec is in Notion, the wireframes in Figma, the API docs in a Confluence page, and the architecture diagram in a Slack message from three weeks ago. Engineering spends the first sprint finding things, not building.
Three different versions of the FSD exist. Nobody is sure which one is current. The wireframes in Figma do not match the spec that was signed off. Engineering builds from the wrong version and discovers the mismatch during QA.
The spec says what to build. It does not say why a decision was made a particular way. When engineering hits a constraint and needs to deviate, they have no context to make a good judgment call. They either build it wrong or stop and wait for the PM.
The Dev-Ready Package is not a new document — it is a curated, version-controlled assembly of everything produced in Modules 1 through 6, organised for engineering consumption.
2–3 pages written specifically for engineers. What the product is, who the user is, the tech stack, key integrations, and the top decisions that were made and why. The context layer no other document provides.
The complete locked Spec Pack from Module 6: FSD set, Blueprint, acceptance criteria register, traceability matrix, change control log, sign-off record. Version numbers must match.
Final Figma file organised by screen ID matching the FSD headings exactly. All states designed. All assets exported. Prototype link included. Engineering-readable annotation layer for non-obvious UI behaviours.
Endpoint list, authentication method, request/response format, error codes, and rate limits for every integration. Engineering cannot build integrations without this.
Key technical decisions: tech stack, cloud platform, database approach, auth method — and any decisions made for non-obvious reasons. These are the decisions a developer might otherwise undo without realising they were intentional.
What test environments exist, how to access them, what test data is available, and what needs to be set up before QA begins. Without this, the first week of testing involves setting up environments instead of running tests.
Every product-specific, client-specific, or domain-specific term used in the spec, with a plain-language definition. Terms that mean different things to the client and to the engineering team are the source of more bugs than almost any other cause.
Naming convention: Every artefact named consistently — [Product]-[Artefact]-[Version]-[Date]. No files named "final", "final v2", or "use this one". One location. All team members can access it. Nothing distributed only by email.
Before packaging, audit every artefact produced across Modules 1 through 6. For each: is it the final version, is it in the agreed location, does its version number match the sign-off record, and does it contain any placeholder content that was never filled. A Dev-Ready Package that contains a TBD is not ready for development. Fix every gap before packaging.
Write it last, once everything else is confirmed final. It references the other artefacts rather than duplicating them. Keep it to two or three pages. An engineer who reads this brief should be able to answer: what am I building, who is it for, what does done look like, where do I find the detailed spec, and what are the most important decisions I need to be aware of.
All artefacts in one location — one shared drive folder, one Notion workspace, or one Confluence space. Structure mirrors the delivery sequence: Onboarding Brief at the top, then Spec Pack, then Figma, then API specs, then Architecture decisions, then Test guide, then Glossary.
The PM walks the engineering lead through the Dev-Ready Package in a 60–90 minute session. Not a presentation — a walkthrough. The PM asks engineering: "What would stop you from starting tomorrow?" Any answer that is not "nothing" is a gap to fix. The walkthrough is not optional.
After the walkthrough, the Dev-Ready Package is formally handed over. The PM's role shifts: from builder of artefacts to guardian of scope. During the development phase, the PM answers spec questions, manages change requests, reviews progress against the traceability matrix, and ensures that what engineering builds matches what was agreed.
AI is most useful for writing the Engineering Onboarding Brief and auditing the package for completeness. The organisation and walkthrough are human work.
Example A shows the Engineering Onboarding Brief structure for the field service app. Example B is a real Ideas2IT story — how clean delivery process enabled the build of ReleasePulse, an AI tool that now runs inside the company.
Example A · Field Service App · Engineering Onboarding Brief
This is the document an engineer reads on day one. It gives them full context without any other conversation.
A mobile app for field technicians and a web dashboard for supervisors at a mid-sized HVAC maintenance company. The product replaces a WhatsApp-and-spreadsheet dispatch process. Core problem: dispatchers have no real-time visibility of job status, causing booking loss and technician idle time. MVP scope: job assignment, status updates with photo, supervisor dashboard, offline capability, iOS and Android.
| Component | Technology | Reason for choice |
|---|---|---|
| Mobile | React Native | iOS + Android from single codebase — client cannot afford separate native builds |
| Backend | Node.js + PostgreSQL on AWS | Team competency and client's existing AWS relationship |
| Real-time updates | WebSockets | Supervisor dashboard requires live job status without polling |
| Offline | SQLite local storage | Field technicians frequently work in low-connectivity environments |
| Auth | JWT with refresh token rotation | Stateless, mobile-friendly, standard for this stack |
These three decisions alone prevent three potential mid-sprint incidents. An engineer who reads this brief will not accidentally implement last-write-wins, store photos locally, or start building a Flutter component. This is what the Onboarding Brief is for.
ReleasePulse is a real internal product. Built by the Ideas2IT team. It runs today. This is the story of how it was conceived, built, and handed over — and why the Module 7 handoff process made it possible.
Every software release at Ideas2IT required manual preparation of release communications across three platforms — Jira, Confluence, and Slack. The process was time-consuming, error-prone, and inconsistent across projects and PMs. Release notes got missed. Confluence pages were out of date. Slack announcements varied in quality and completeness.
The team identified this as a perfect candidate for AI automation: a structured, repeatable process with clear inputs (release details) and clear outputs (three types of communications across three platforms).
ReleasePulse uses a multi-agent AI framework (Crew.ai) with three sequential agents, triggered by a single Slack command.
PM types /trigger-release in the dedicated Release channel. A form appears collecting: version number, release highlights, and important notes. This is the single input that drives all three agents.
Publishes formal release details into the Confluence Wiki. Creates a structured release page with version, highlights, linked Jira tickets, and release notes. Replaces the manual Confluence update that PMs previously did by hand.
Updates release content and links in Jira. Creates a Release ticket with the description of the release and an epic ticket holding all the stories worked. Links tickets back to the Confluence page. Replaces manual Jira bookkeeping that was often skipped entirely.
Publishes a formatted release announcement to the designated Slack channel(s). Consistent format every time — version, highlights, links to Confluence and Jira. No more inconsistent announcements written differently by different PMs.
ReleasePulse was built by an internal team, not a client. But the same handoff principles applied. When the architecture was designed and the agent logic was defined, the team produced:
Without this documentation, a new engineer joining the team mid-build would have spent days reconstructing the agent logic and integration architecture from the code alone. With it — half a day. That is the test of a clean handoff.
Before triggering a release, the PM runs two setup commands in Slack:
Then /trigger-release fires all three agents sequentially. The entire release communication process — which previously took 30–60 minutes of manual work per release — now takes under 2 minutes.
What this proves: When a team follows a structured delivery process — clear spec, clean handoff, documented decisions — they can build and ship internal AI tools at the same speed and quality as client products. ReleasePulse is not a side project. It is a production system that runs on every Ideas2IT release. It was built the right way because the team handed it over clean.
The Dev-Ready Package contains all seven required artefacts
Every artefact has a version number and date that matches the sign-off record from Module 6
No artefact contains placeholder text, TBD markers, or unresolved gaps
The Engineering Onboarding Brief exists and covers all seven required sections
Figma file is organised by screen ID, all states are present, all assets are exported
API and integration specs are complete for every integration in scope
The test environment guide confirms environments are accessible and test data is available
The glossary covers all domain-specific and client-specific terms in the spec
The engineering kickoff walkthrough has been completed — not just the package delivered
Engineering has confirmed in writing that they have everything needed to begin
A process exists for engineers to ask spec questions during the build phase — and the PM has committed to a response time
How to use AI to track product performance after launch, measure against the success metrics defined in the PRD, identify what to improve, and turn a delivered project into an ongoing engagement.
Most PMs treat launch as the end of their job. It is not. Everything before launch was hypothesis — the PRD, the user flows, the feature decisions, the effort estimate. Launch is the moment those hypotheses meet reality.
Every delivered product contains the seeds of the next engagement. The PM's job in this module is to find those seeds, cultivate them with data, and present them in a way that makes the next phase feel like the obvious next step — not a new sale.
A client whose PM disappears after launch is a client who will not return for the next phase. A client whose PM comes back with a structured performance brief, a clear view of what is working and what is not, and a credible proposal for the next iteration — that client is an account that grows.
Honest comparison of PRD targets against actual post-launch data. The credibility foundation for everything that follows.
Prioritised, evidence-backed list of improvements. Each item has evidence, user impact, business impact, and estimated effort.
Structured proposal for the next phase. The document that converts a completed delivery into a growing account relationship.
Product analytics alone is not enough. User feedback alone is not enough. Technical metrics alone are not enough. The Performance Brief is built from all three.
| Metric | What it tells you |
|---|---|
| Feature adoption rate | What % of active users have used each feature at least once? Features below 20% adoption in the first 30 days are at risk of being considered low-value by the client. |
| User retention | Day 1, Day 7, Day 30 retention rates. A product that looks healthy on launch day but shows 60% drop-off by day 7 has a fundamental engagement problem. |
| Flow completion rates | For every primary user flow mapped in Module 4, what % of users who start it complete it? Drop-off points indicate UX or spec issues not caught in testing. |
| Error rates | How often are users hitting error states? High error rates indicate a spec gap, an implementation bug, or a user mental model mismatch. |
| Time on task | Significantly longer than the design assumption indicates friction. Significantly shorter may indicate shortcuts that bypass intended flows. |
| Feedback type | What it tells you |
|---|---|
| Support ticket analysis | What are users asking for help with? Ticket themes reveal where the product is unclear, where onboarding failed, and where spec assumptions were wrong about user behaviour. |
| NPS or CSAT scores | Overall satisfaction score and the verbatim comments attached to low scores. Low scores with no comments are data. Low scores with comments are insights. |
| Usability test sessions | If budget allows, 2–3 moderated usability sessions post-launch reveals behaviours analytics cannot show — confusion, workarounds, and unmet needs. |
| Feature request patterns | What features are users asking for that do not exist? Requests clustering around the same need are a signal for the next iteration roadmap. |
| Metric | What it tells you |
|---|---|
| System uptime and reliability | Is the product available when users expect it? Downtime events and their causes. |
| API response times | Slow API responses create UX problems that analytics may surface as user drop-off without indicating the root cause. |
| Error log patterns | Error patterns that do not surface in user-facing analytics often indicate deeper reliability issues. |
| Infrastructure cost vs estimate | Cost overruns at this stage indicate either scope creep in the build or underestimation of usage volume. |
The success metrics in the PRD v0.1 are your baseline. Before launch, confirm with the client which metrics matter most, what the target values are, and what the measurement window is (first 30 days, first 90 days, first full quarter). Measuring without a baseline is just collecting numbers.
At the agreed measurement window, collect data from product analytics, user feedback, and technical monitoring. Use Prompt 1 to synthesise support ticket themes and user feedback at speed. Use Prompt 2 to generate a structured analysis of usage data. You are looking for patterns, not individual data points.
Compare actual metrics against the PRD baseline. For each metric: what was the target, what is the actual, what is the delta, and what does the delta tell you. Be honest — a Performance Brief that only reports metrics that look good is a document no one will trust for the next phase.
From the Performance Brief, identify every issue, gap, or opportunity. For each item: evidence, who is affected, estimated impact, and approximate effort. Prioritise using impact vs effort. Tier 1 = fix immediately. Tier 2 = next iteration. Tier 3 = backlog.
From top-priority backlog items and any new features requested, build the roadmap for the next phase. Frame it as: objectives (what outcomes the next phase targets), scope (what features deliver those outcomes), and commercial case (why investing in the next phase makes business sense). This is the document that drives the next engagement.
AI is highly effective at synthesising large volumes of unstructured feedback and usage data into structured insights. The judgment about what to prioritise is human. The pattern-finding in raw data is AI.
Example A: Field service app at 30 days. Example B: AI forecasting dashboard at 90 days with model drift. Example C: ReleasePulse — Ideas2IT's own AI tool, measured post-launch.
Example A · Field Service App · 30-Day Performance Brief
40 technicians active, 5 supervisors. Measurement window: days 1–30.
Objectives:
Scope — Must have:
Commercial case: Estimated effort: 3 weeks, 3-person team. Expected outcome: photo compliance to 95%+, support tickets from 18/month to under 7/month. Recommend proceeding immediately while team context is fresh.
Example B · Demand Forecasting Dashboard · 90-Day AI Model Issue
The forecasting model's accuracy has declined from 82% at launch to 71%. Buyers are overriding 34% of AI recommendations — against a target of under 15%.
What happened: The demand forecasting model was trained on 3 years of historical data. In the 90 days since launch, buying behaviour shifted due to a promotional event that was outside the training data. The model did not adapt because the retraining pipeline was configured for manual trigger only.
| Metric | At launch | Now (90 days) | Target | Status |
|---|---|---|---|---|
| Model accuracy | 82% | 71% | 80% | BELOW |
| Override rate | 18% | 34% | <15% | BELOW |
Objective: Restore model accuracy to 80%+ and reduce override rate below 20%.
Scope:
Commercial case: The current 34% override rate means buyers are spending significant time manually reviewing AI recommendations — the ROI promised in the proposal is not being delivered at this accuracy level. Phase 2 restores the promised saving. This is the honest conversation every PM must be prepared to have.
ReleasePulse was built and handed over (Module 7). Now it has been running. This is the post-launch story — what was measured, what the data showed, and what it means for the next iteration.
The baseline success metrics defined before launch:
Executive summary: ReleasePulse launched across all active Ideas2IT projects. The headline finding is that manual release documentation time has been reduced by approximately 80%, exceeding the 70% target. The recommended next step is extending the tool to capture release metrics over time — so PMs can see trends across releases, not just per-release outputs.
What is working well: The single Slack command trigger is frictionless — adoption was immediate because the workflow required no behaviour change. Consistency across channels has eliminated the inconsistency that previously caused stakeholder confusion. The Jira Release ticket and Confluence page are now always in sync.
What needs improvement: The tool currently produces communications but does not capture them for retrospective analysis. PMs cannot look back and see how release scope has grown over time, or compare release frequency across projects. This is the gap the next iteration should address.
The principle this demonstrates: Every internal tool built with a proper delivery process becomes a product that can be measured, improved, and extended. ReleasePulse did not end at launch — it entered the same measure and improve cycle as any client product. That is AI-powered delivery culture in practice.
A measurement baseline was established from the PRD success metrics before or at launch
Data has been collected from all three sources: product analytics, user feedback, and technical performance
The Performance Brief compares actual metrics against PRD targets — not just reports actuals
The brief is honest about underperformance — no metrics omitted because they look bad
Every claim in the brief is backed by a data point — no "users seem to like" without a number
The Improvement Backlog contains at least one Tier 1 item (immediate action required)
Every backlog item has evidence, user impact, business impact, estimated effort, and a priority tier
The Next Iteration Roadmap has stated objectives tied to measurable metrics
The commercial case for the next phase is explicit — not implied
The roadmap includes a clear next step the client needs to take to proceed
The Performance Brief has been reviewed internally before being presented to the client
Eight modules. The complete lifecycle of AI-powered product delivery — from the first client conversation through to post-launch performance and the next engagement.
| Module 1 — Decode the Brief | Understanding the client's real problem. Output: Discovery Summary + Concept Note |
| Module 2 — Know the Battlefield | Market research, competitor analysis, feature prioritisation. Output: Validated PRD v0.1 + Scope Definition Document |
| Module 3 — Price it Right | Effort estimation, team composition, timeline phasing, AI savings. Output: Effort Estimate + Proposal Pricing |
| Module 4 — Draw the Journey | UX Blueprint, user flows, wireframes, prototype. Output: UX Blueprint + Figma Package |
| Module 5 — Make it Testable | FSD writing, acceptance criteria, business rules, QA test cases. Output: Locked FSD Set |
| Module 6 — Get Everyone Aligned | Cross-team validation, spec lock, change control. Output: Signed-off Spec Pack |
| Module 7 — Hand it Over Clean | Dev-Ready Package, engineering onboarding, kickoff. Output: Dev-Ready Package |
| Module 8 — Measure & Improve | Post-launch performance, improvement backlog, next iteration roadmap. Output: Performance Brief + Roadmap |
The through-line across all eight modules:
Every module produces an output that feeds the next. Every prompt builds on the documents produced in prior modules. Every quality gate checks that what you are handing forward is complete enough for the person receiving it to do their job without guessing.
This is not eight separate skills — it is one integrated system. AI accelerates every phase. Human judgment governs every decision. The combination is what produces delivery that clients trust and return to.
Answer 9 questions. Get a personalised ROI breakdown — across your PM/PO phase, dev phase, and rework avoided.