Scoreboard for process measurement (Mota, 2009)
Based on: Mota, Pedro João. “Scoreboard: a support for management information needs.” _MSE Reflection Paper (2009)._
1. Core Idea
The Scoreboard is a lightweight tool to support project and team managers by collecting qualitative insights directly from team members.
-
Instead of relying only on quantitative metrics (budget, code size, defect counts), it captures perceptions and opinions—often the only way to assess issues like morale, cohesion, or trust.
-
It’s especially useful for distributed teams or in areas where hard data is hard to define or too costly to measure
2. How It Works
-
Design questions – general (“How do you evaluate the Project Manager?”) + specific (“How accurate were the weekly task estimates?”).
-
Collect answers – usually via a web questionnaire, short (≤10 min), ideally anonymous to encourage honesty.
-
Evaluate answers – semantic 5-point scale (Very Bad → Very Good) with optional comments.
-
Analyze results – highlight worst 5 issues and downward trends.
-
Discuss in reflection meetings – the data itself doesn’t solve problems, but triggers conversations and commitments.
-
Maintain scoreboard – add/remove questions to stay relevant, avoid fatigue.
flowchart LR
%% Phases
subgraph T[Think]
GQM(Define goals & info needs<br/>use GQM if possible)
QDEF(Design questions<br/>• General roles/areas<br/>• Specific issues/risks)
SCALE(Define scale & inputs<br/>• 5-level semantic scale<br/>• Comments<br/>• Optional “Don’t know”)
GQM --> QDEF --> SCALE
end
subgraph A[Act]
DIST(Distribute questionnaire<br/>web form, short)
RESP(Collect responses)
DIST --> RESP
end
subgraph R[Reflect]
AGG(Aggregate data<br/>• Averages by question<br/>• Trends over time)
SELECT(Select focus<br/>• Worst ~5 items<br/>• Downward trends)
MEET(Reflection meeting<br/>Discuss: what/why/why missed/how prevent)
ACTION(Plan actions & owners)
AGG --> SELECT --> MEET --> ACTION
end
subgraph M[Maintain]
UPDATE(Update questions<br/>• Add for new risks/actions<br/>• Remove stable/high items<br/>• Tune wording/scales)
BALANCE(Balance with quantitative metrics)
CHECK(Monitor adherence & trust)
UPDATE --> BALANCE --> CHECK
end
%% Cross-phase flow
SCALE --> DIST
RESP --> AGG
ACTION --> UPDATE
CHECK -.-> GQM
%% Cadence (as a node, not a note)
WEEK[[Weekly cadence:<br/>Thu collect → Fri reflect]]
DIST -. aligns .-> WEEK
MEET -. aligns .-> WEEK
sequenceDiagram
autonumber
actor TM as Team Members
actor MM as Metrics Manager
actor PM as Project Manager
participant S as Survey System
%% Think
MM->>MM: Prepare questionnaire<br/>(add/remove/tune questions)
MM->>S: Configure survey (scale, comments, anonymity)
%% Act
MM->>S: Distribute survey (Thu morning)
S-->>TM: Email with survey link
TM->>S: Submit answers (~10 min, honest, anonymous if possible)
S-->>MM: Collect responses
%% Reflect
MM->>MM: Aggregate results (averages & trends)
MM->>MM: Select focus (worst 5 & downward trends)
MM-->>PM: Prepare report for reflection
PM->>TM: Facilitate reflection meeting (Fri)
PM->>PM: Discuss problems, causes, prevention
PM->>PM: Define corrective actions & assign owners
%% Maintain
PM-->>MM: Share action items
MM->>MM: Update questions (add for new risks, remove stable items)
MM->>MM: Balance qualitative with quantitative metrics
MM->>MM: Monitor adherence & trust
3. Strengths
-
Cost-efficient: quick for members, rich data for managers.
-
Focus: highlights most important issues for team discussion.
-
Promotes reflection: helps individuals step back and assess (“reflection-on-action”).
-
Flexible: can extend to risk management, training needs, or team suggestions.
4. Weaknesses
-
Can overshadow quantitative metrics if overused.
-
Lack of shared meaning in scales may skew results.
-
Averages can hide dissent (outliers lost in mean).
-
If results aren’t acted on, team motivation to answer drops.
-
Danger of being seen as evaluating individuals instead of roles/activities.
-
Requires continuous trust and maintenance
5. Lessons from the Mappers Project
-
Weekly cycle: answer Thursday → analyze same day → discuss Friday. Short gap between answering and discussion is crucial.
-
Around 35 questions balanced between roles and general team issues.
-
Questions that stabilized (consistently good for 4+ weeks) were removed to prevent fatigue.
-
Checklists (Google Docs) ensured process discipline.
-
The Scoreboard became the main source of reflection input, more than quantitative metrics.
Disclaimer: AI is used for text polishing and explaining. Authors have verified all facts and claims. In case of an error, feel free to file an issue.