Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Objective

Evaluate each initiative using standardised criteria to enable objective comparison and ensure alignment with strategic goals.

Evaluation criteria

Reach (R): The number of users/customers affected within a specific time frame

Impact (I): The degree of benefit each user/customer receives from the initiative

Confidence (C): The level of certainty in the estimates of reach and impact

Effort (E): The total amount of work required to complete the initiative

Actions

Assign reach scores:

Estimate the number of users/customers impacted over a set period (e.g., per month or quarter). Use actual data if at all possible.

Assign impact scores:

Rate the benefit per user on a standardised scale:

  • 3 (Massive impact): Significant positive change in user behaviour or satisfaction

  • 2 (High impact): Notable improvement in user experience or key metrics

  • 1 (Medium impact): Moderate enhancement to existing features or processes

  • 0.5 (Low impact): Minor improvements with limited effect

  • 0.25 (Minimal impact): Minimal noticeable change to users

Assign confidence scores:

Rate your certainty about reach and impact estimates:

  • 100% (High confidence): Based on solid data and testing

  • 80% (Medium confidence): Based on reliable but incomplete data

  • 50% (Low confidence): Based on assumptions or limited information

Estimate effort:

Approximate the total work required. Include all resources needed from discovery through to launch (engineering, design, QA, compliance, etc).

Score strategic alignment:

Rate alignment with strategic goals on a scale:

  • 3 (High alignment): Directly supports key strategic objectives

  • 2 (Moderate Medium alignment): Supports strategic objectives but not directly

  • 1 (Low alignment): Indirectly supports or has minimal alignment

...