Lesson 015: Borda Count for Ranked Voting

Lesson 015: Borda Count for Ranked Voting

Problem

The Artemis category voting mode asks voters to rank their top 3 images within a category. We need to convert these partial rankings into numeric scores that can be aggregated across voters and combined with batch and pairwise preference signals.

Why It Matters

Ranked voting captures more information than binary choice — a voter who ranks image A first, B second, and C third is telling us more than a voter who just picks A. The scoring method determines whether this extra information is preserved or lost.

What Happened

  1. Had 250 category ranking submissions (synthetic), each ranking the voter's top 3 images within a category. Needed to convert these partial ordinal rankings into numeric scores that aggregate across voters and combine with batch and pairwise signals.
  2. Considered Plackett-Luce as the probabilistic model for ranked data (analogous to BTL for pairwise). Rejected for the same reason as BTL — requires connectivity in the comparison structure, and with only 250 rankings across 8 categories, coverage is too sparse.
  3. Chose standard Borda count: rank 1 = 3 points, rank 2 = 2 points, rank 3 = 1 point. Simple, transparent, and aggregates naturally by summation.
  4. Decided to use total Borda score (not mean) as the primary metric. An image ranked #1 in 10 submissions (30 points) carries stronger signal than one ranked #1 in 2 submissions (6 points), even though both have mean = 3. Total rewards both high ranking and frequent ranking.
  5. Identified the exposure-set problem: when a voter ranks top 3, we don't know which other images they saw and rejected. Unranked images are treated as missing data, which may slightly overestimate scores for images that were seen but not ranked. No clean fix without logging the full exposure set.
  6. Borda enters the composite formula at 10% weight via quantile rank — the most modest of the three signals. The low weight reflects both the sparse coverage (only 75 images have any Borda score) and the exposure-set uncertainty.

Design Choice: Standard Borda Count with 3/2/1 Scoring

Key terms

Aggregation approach

For each image, we sum the Borda scores across all ranking submissions where it appeared:

The borda_score (total) is used in the composite because it rewards images that are both highly ranked and frequently ranked. An image ranked #1 in 10 submissions (30 points) is stronger than one ranked #1 in 2 submissions (6 points), even though both have the same mean.

Alternatives Considered

  1. Average rank: Use mean rank position instead of Borda points. Loses the distinction between "ranked once as #1" and "ranked ten times as #1."
  2. Plackett-Luce model: Full probabilistic model for ranked data. Deferred — same connectivity and sparsity issues as BTL.
  3. Category-specific models: Fit separate preference models within each category. Would produce more precise within-category rankings but doesn't help with cross-category comparison.
  4. Dowdall system (harmonic): Scores of 1, 1/2, 1/3 for ranks 1, 2, 3. Gives even more weight to first place. Considered unnecessary with only 3 ranks.

What Was Learned

Borda count is the right simple default for partial ranking data. Its main advantage is transparency — it's easy to explain that "rank 1 = 3 points, rank 2 = 2 points, rank 3 = 1 point" and the aggregation is just a sum. The exposure-set problem is the biggest limitation and there's no clean fix without logging which images each voter actually saw. For the composite score, Borda's contribution is modest (10% weight via quantile rank) precisely because of this uncertainty.