Workshops16 min read

Prioritization Workshop: Help Your Team Make Better Trade-offs

A hands-on prioritization workshop using RICE scoring on your actual backlog. Includes scoring calibration, debate rounds, and final stack ranking exercises.

By Tim Adair• Published 2026-02-12

Overview

Every product team says they prioritize. Most do not. They negotiate. The loudest stakeholder wins, the most recent customer request jumps the queue, or the HiPPO (Highest Paid Person's Opinion) overrides the data. The result is a backlog that reflects politics, not strategy.

This workshop fixes that by putting the team's actual backlog through a structured scoring process. Not a theoretical exercise — real features, real scores, real debates about trade-offs. By the end, you will have a stack-ranked backlog that the team built together and can defend to anyone who asks "why is my feature not on the roadmap?"

Who this is for: Product managers facilitating prioritization with their cross-functional team (PM, engineering leads, design, and optionally one senior stakeholder).

Time required: 90 minutes

What participants will walk away with:

  • A RICE-scored backlog of 10-15 candidate features
  • A calibrated understanding of what "high impact" and "high effort" actually mean for your team
  • A final stack rank with clear rationale for each position
  • A documented decision log that you can share with stakeholders who were not in the room

  • Pre-Work (Required — Send 72 Hours Before)

    The workshop fails without preparation. Send this to all participants at least 3 days before:

    For the facilitator (PM):

  • Select 10-15 candidate features from the backlog. Aim for a mix: 3-4 obvious priorities, 3-4 "maybe" items, and 3-4 items you suspect will score low but have vocal advocates.
  • Write a one-line description and a two-sentence "user story" for each candidate. Keep it consistent — if some features have detailed specs and others have nothing, the scoring exercise becomes a spec-quality exercise instead.
  • Pre-fill the Reach column using actual data: number of users/accounts affected per quarter. This is the one dimension where the PM should come in with a number, not an opinion.
  • Set up a shared spreadsheet with columns: Feature | Reach | Impact | Confidence | Effort | RICE Score. Leave Impact, Confidence, and Effort blank.
  • For participants:

  • Review the candidate features list.
  • For each feature, think about: Who is asking for this? What problem does it solve? What is the rough effort (your gut, not an estimate)?
  • Come ready to score, debate, and commit.

  • Materials Needed

  • Shared spreadsheet (Google Sheets works well) projected on screen or shared via screen share
  • RICE calculator — open in a browser tab for real-time score computation
  • Timer visible to the room
  • Printed or digital copy of the RICE framework scoring definitions
  • The pre-filled candidate feature list (10-15 items)
  • Sticky dots or a voting mechanism for the final round
  • Spreadsheet template setup:

    Create a Google Sheet with these columns before the workshop:

    ColumnPre-filled?Notes
    Feature NameYesOne-line description
    User StoryYesTwo sentences max
    Reach (users/quarter)YesPM fills from analytics
    Impact (1-5)NoTeam scores in workshop
    Confidence (%)NoTeam scores in workshop
    Effort (person-months)NoTeam scores in workshop
    RICE ScoreFormula=(Reach Impact Confidence) / Effort
    NotesNoCapture debate flags and rationale

    Share the spreadsheet link with all participants 30 minutes before the workshop so they can familiarize themselves with the feature list.


    Part 1: Scoring Calibration (20 minutes)

    What this accomplishes

    If you skip calibration and jump straight into scoring, you will spend the entire workshop arguing about whether "High Impact" means 2 or 3 on your scale. Calibration gets the team aligned on what the numbers mean before you apply them to real features.

    Facilitator instructions

    Explain the RICE scoring system (5 minutes):

    Walk through each dimension using the RICE framework definitions:

  • Reach: How many users/accounts will this affect per quarter? Use real numbers from your analytics. This is the most objective dimension — lean on data.
  • Impact: How much will this move the needle for each user who encounters it? Use a 1-5 scale: 1 = minimal, 2 = low, 3 = medium, 4 = high, 5 = massive. Most features are a 2 or 3. A 5 is rare — reserve it for features where users literally cannot do their job without it.
  • Confidence: How sure are we about the Reach and Impact scores? 100% = high confidence (we have data). 80% = we have some evidence. 50% = it is a guess. This dimension penalizes wishful thinking.
  • Effort: Person-months of work. Include design, engineering, QA, and launch effort. Round to the nearest 0.5.
  • Calibration exercise (15 minutes):

    Pick two features from the candidate list — one that everyone intuitively feels is "high priority" and one that feels "low priority."

    For each feature, do a simultaneous reveal:

    "Everyone, write down your Impact score for Feature A. Do not show anyone. 3, 2, 1 — hold up your number."

    If scores vary by more than 1 point, pause and discuss:

    "We have a 2 and a 4 on Impact. [Person with 2], what are you seeing? [Person with 4], what makes this high for you?"

    Run through Impact, Confidence, and Effort for both calibration features. By the end, the team should have a shared sense of what each number means in the context of your product.

    Key facilitation tip: The most common calibration failure is Impact inflation. Everyone wants their feature to be "high impact." Counter this by asking: "Compared to the last feature we shipped that you would call genuinely high impact — say, [name a specific recent win] — how does this feature compare?"

    Calibration reference card (post this on the wall or share in chat):

    ScoreImpactConfidenceEffort
    5 / 100% / 0.5moMassive — users cannot do their job without itHigh confidence — validated by data from 50+ users or controlled experimentTiny — one person, one week
    4 / 80% / 1moHigh — noticeably improves daily workflowSome evidence — 10+ customer requests, qualitative interviewsSmall — one person, one month
    3 / 50% / 2moMedium — nice to have, solves a real but non-critical painEducated guess — based on market research or analogous productsMedium — 2 people, one month
    2Low — marginal improvement most users would not noticeLow confidence — PM intuition onlyLarge — one quarter of a team
    1Minimal — affects edge cases onlySpeculation — no supporting evidenceVery large — full team for a quarter

    Print this card for every participant. It eliminates 80% of the "what does a 3 mean?" debates.


    Part 2: Batch Scoring (30 minutes)

    What this accomplishes

    This is the core of the workshop: scoring every candidate feature using the calibrated RICE framework. The batch approach prevents the common trap of spending 20 minutes on the first feature and rushing the last five.

    Facilitator instructions

    Set the pace (2 minutes per feature):

    "We have [X] features to score in 30 minutes. That is roughly 2 minutes per feature. I will keep us on time. For each feature, I will read the description, we will do a simultaneous reveal on each dimension, and I will capture the consensus score. If there is a disagreement of more than 1 point, we will flag it for the debate round."

    Scoring process for each feature:

  • Facilitator reads the one-line description and user story
  • Reach: PM shares the pre-filled number (no debate needed — this is data)
  • Impact: Simultaneous reveal → capture score (if spread > 1 point, flag for debate)
  • Confidence: Quick round — PM proposes a number, team agrees or adjusts
  • Effort: Engineering lead proposes, team adjusts
  • Capture in the spreadsheet in real time. The RICE calculator auto-computes the final score.

    Flagging disagreements:

    When a score has a spread greater than 1, do not resolve it now. Write "DEBATE" next to the feature and move on. You will have a dedicated round for these.

    Expect 3-5 features to get flagged. If more than half are flagged, your calibration was insufficient — pause and recalibrate on Impact specifically.

    Facilitator tips for this section:

  • Resist the urge to discuss feasibility or design details. This is scoring, not solutioning.
  • If someone says "it depends," ask: "On what? Give me the most likely scenario and we will score that."
  • Track effort carefully — engineering leads tend to sandbag effort estimates in group settings. Ask: "If this were the only thing on the team's plate, how long would it actually take?"
  • If the group gets stuck on a feature, set a hard 3-minute cap and move on. Scoring paralysis on individual items is the number one time killer. Features where the group cannot agree in 3 minutes go straight to the debate round.
  • Example scoring output after Part 2:

    FeatureReach (users/qtr)Impact (1-5)Confidence (%)Effort (person-mo)RICE ScoreNotes
    Self-serve data import3,200480%25,120
    Dashboard redesign8,000350%43,000
    SSO for enterprise4005100%3667DEBATE (Impact split: 3 vs 5)
    Mobile notifications5,000250%15,000
    Custom reporting1,200350%3600DEBATE (Effort split: 2 vs 5)

    Notice that the raw RICE score sometimes produces surprises. Mobile notifications — which nobody championed loudly — scores nearly as high as the self-serve data import because of its combination of high reach and low effort. That is the framework doing its job: revealing priorities the room's intuition would have missed.


    Part 3: Debate Round (20 minutes)

    What this accomplishes

    The flagged features from Part 2 are the interesting ones. They are interesting because reasonable people looked at the same feature and reached different conclusions — which means there is information asymmetry in the room. This round surfaces that information.

    Facilitator instructions

    Triage the flagged features (2 minutes):

    Look at the flagged items. If a flagged feature scores so low overall that resolving the disagreement would not change its rank position, skip it. Focus debate time on features near the cut line where the disagreement matters.

    Structured debate format (4 minutes per feature, up to 4-5 features):

    For each flagged feature:

  • State the disagreement (30 seconds): "Feature X has Impact scores ranging from 2 to 4. That is a meaningful spread."
  • Low scorer speaks first (60 seconds): "Why did you score this a 2?"
  • High scorer responds (60 seconds): "Why did you score this a 4?"
  • Open floor (60 seconds): Anyone else can add context — customer data, competitive intel, technical considerations.
  • Re-score (30 seconds): "Based on what you just heard, everyone write down your new Impact score."
  • Take the median of the new scores. Update the spreadsheet.

    What to watch for in debates:

  • If the disagreement is about customer value, the PM should cite data — support tickets, usage analytics, customer interview quotes.
  • If the disagreement is about effort, the engineering lead should reference comparable past projects. "The last time we built something like this, it took X."
  • If the disagreement is philosophical ("we should invest in the platform" vs. "we should ship features"), that is a strategy question, not a prioritization question. Flag it for a separate strategy conversation.

  • Part 4: Final Stack Rank and Commitments (20 minutes)

    What this accomplishes

    The RICE scores produce a rank. But scores alone do not account for dependencies, sequencing, or strategic themes. This final section takes the mathematical output and applies human judgment to produce the actual priority order.

    Facilitator instructions

    Review the scores (5 minutes):

    Sort the spreadsheet by RICE score, highest to lowest. Share the screen.

    "Here is what the math says. The top 5 features by RICE score are: [read them]. Before we finalize, I want to flag two things: first, are there any dependency chains where Feature B cannot ship without Feature A? Second, are there any strategic reasons to override the score?"

    Adjustment round (5 minutes):

    Allow the room to propose adjustments. Each adjustment requires a stated reason:

  • "I want to move Feature X above Feature Y because X is a prerequisite for Y" — valid
  • "I want to move Feature X up because the CEO asked for it" — valid only if you treat CEO input as a Reach/Impact signal and rescore accordingly
  • "I just feel like Feature X should be higher" — not valid without a reason
  • Limit adjustments to 2-3 moves. If the team wants to rearrange the entire list, the scoring was not rigorous enough and you should schedule a follow-up.

    Document the final rank (5 minutes):

    For each feature in the top 5 (or however many fit in the next quarter):

    FeatureRICE ScoreOwnerTarget QuarterKey Assumption
    Feature A42[Name]Q2 2026Users will adopt self-serve onboarding

    The "Key Assumption" column is borrowed from the strategy alignment workshop. It keeps the team honest about why something is prioritized.

    Close the loop (5 minutes):

    "For every feature below the cut line: this does not mean 'never.' It means 'not this quarter, given what we know today.' If new data changes the Reach or Impact score, we will re-evaluate. For stakeholders who asked for specific features that did not make the cut, I will share the RICE score and the rationale. Any questions?"

    Get explicit acknowledgment from each participant that they support the final rank.


    Adapting for Different Team Sizes and Contexts

    For small teams (3-5 people)

    Skip the small-group split in Part 2. Instead, score as a full group with the spreadsheet projected. This is faster and eliminates the share-out step. You can comfortably score 15 features in 25 minutes with a small group. Cut total workshop time to 75 minutes.

    For large teams (10+ people)

    Split into two scoring groups of 5, each with their own spreadsheet. Each group scores all features independently. Then merge the spreadsheets by averaging scores. Where the two groups diverge by more than 20% on a feature's total score, that feature goes to the debate round. This parallel approach keeps the scoring phase tight even with large groups.

    For teams using frameworks other than RICE

    The workshop structure works with ICE (Impact, Confidence, Ease), weighted scoring, or MoSCoW with minor modifications. For ICE, replace the Reach dimension with a combined "Impact" score and rename "Effort" to "Ease" (invert the scale — higher is easier). For MoSCoW, replace the scoring phase with a sorting exercise: each participant independently places features into Must/Should/Could/Won't, then the group resolves conflicts category by category. The RICE vs ICE vs MoSCoW comparison covers when each framework works best.

    For distributed teams running this quarterly

    After the first workshop, subsequent quarterly sessions get faster. Create a "rolling backlog" spreadsheet where new feature requests are pre-scored by the PM as they come in during the quarter. In the quarterly workshop, the team validates the PM's pre-scores and only debates items where the team disagrees with the PM's assessment. This cuts the batch scoring phase from 30 to 15 minutes.


    Next Steps for the Facilitator

    Within 24 hours of the workshop:

  • Send the scored spreadsheet to all participants and stakeholders. Transparency is the best defense against "but I thought we agreed to build X."
  • Write a brief summary for leadership: top 5 priorities, key trade-offs, and features that were deprioritized with rationale.
  • Update the backlog in your project management tool to reflect the new stack rank.
  • Schedule the monthly re-score — a 30-minute standing meeting to score any new candidates and rescore anything near the cut line.
  • For teams considering different scoring approaches, the RICE vs ICE vs MoSCoW comparison can help you decide if RICE remains the right framework for your team's maturity level.


    Facilitator Tips

    On managing debates:

  • Time-box rigorously. The debate round can easily expand to fill the entire session if you let it. Use a visible timer and cut discussions at the 4-minute mark even if the group wants more time.
  • If two people are going back and forth, invite a third voice: "We have heard from engineering and product. Design, what is your read on the impact here?"
  • Never let a debate become personal. If you hear "you always overestimate effort" or "you do not understand customers," redirect: "Let's talk about this specific feature, not patterns. What data do we have?"
  • On effort estimation:

  • Effort is the most contentious RICE dimension because it is the easiest to manipulate. An engineer who does not want to build something will estimate high. A PM who wants something built will estimate low.
  • Counter this by using relative sizing: "Compared to Feature Y, which we agreed was 2 person-months, is Feature X smaller, the same, or bigger?"
  • If you cannot agree on effort within 30 seconds, use the higher estimate. Effort surprises are almost always upward.
  • On remote facilitation:

  • Use the spreadsheet as the single source of truth. Do not ask people to hold numbers in their heads.
  • For the simultaneous reveal, use a Slack thread or chat: "Type your Impact score for Feature X. Do not hit send until I say 'go.'"
  • If latency is an issue, batch the scoring: have everyone fill in their scores for all features in a private copy, then average the results.
  • On stakeholder management after the workshop:

  • The scored spreadsheet is your shield. When a stakeholder says "why is my feature not on the roadmap?" you can show them exactly where it ranked and why. This is dramatically more effective than saying "we decided it was lower priority."
  • If a stakeholder disagrees with a score, invite them to the next monthly re-score with new data. This channels their energy productively rather than into lobbying.
  • Common pitfalls:

  • Scoring without calibration: Results in meaningless scores and wasted debate time.
  • Too many features: More than 15 candidates makes the scoring round feel like a slog. Be selective in the pre-work.
  • Skipping the pre-work: If participants have not reviewed the feature list, they cannot score meaningfully. Delay the workshop rather than run it unprepared.
  • Treating RICE scores as sacred: The score is a tool for structured conversation, not a divine ranking. If the math says Feature X beats Feature Y but every person in the room disagrees, investigate why. The scores may be wrong, or the team may have information that the framework does not capture.
  • Frequently Asked Questions

    Should we use RICE, ICE, or MoSCoW for the workshop?+
    RICE works best for most product teams because it forces you to estimate Reach and Effort separately, which prevents the common trap of confusing 'exciting' with 'impactful.' ICE is simpler but less rigorous — good for early-stage teams. MoSCoW works better for scope negotiations than ongoing prioritization. See the full comparison at /compare/rice-vs-ice-vs-moscow.
    What do we do when stakeholders disagree on RICE scores?+
    That disagreement is the point. When two people score the same feature differently, it means they have different information or assumptions. Use the debate round to surface those differences explicitly. The goal is not consensus on the score — it is shared understanding of the trade-off. After discussion, the PM makes the final call on the score.
    How often should we run this prioritization exercise?+
    Run the full 90-minute workshop once per quarter at the start of your planning cycle. Between workshops, use a lightweight 30-minute version monthly where you re-score only new candidates and anything that scored within 10% of the cut line. Scores decay — a feature scored 6 months ago may have different Reach or Effort today.
    Free Resource

    Want More Guides Like This?

    Subscribe to get product management guides, templates, and expert strategies delivered to your inbox.

    No spam. Unsubscribe anytime.

    Want instant access to all 50+ premium templates?

    Start Free Trial →

    Put This Guide Into Practice

    Use our templates and frameworks to apply these concepts to your product.