ParetoRank assigns each system a rank based on Pareto dominance across two
dimensions: quality (e.g. mean_score) and cost (e.g. cost_of_pass). A system
with rank 1 is Pareto-optimal — no other system is simultaneously better on
both dimensions.
What is Pareto dominance?
System B dominates system A when:- B’s quality is at least as high as A’s, and
- B’s cost is at most A’s, and
- B is strictly better on at least one of the two dimensions.
1
means no system dominates it — it sits on the Pareto frontier.
| Rank | Meaning |
|---|---|
1 | Pareto-optimal: best trade-off available |
2 | One system dominates it |
3 | Two systems dominate it |
Constructor parameters
Summary key to use as the quality dimension.
Summary key to use as the cost dimension. Lower is better.
rank_systems() static method
Because Pareto ranking requires comparing all systems simultaneously,
ParetoRank exposes a static method that operates on the full
EvalResult.summary dict rather than on per-system rows.
The
EvalResult.summary dict mapping system name → metric values.Key in each system’s summary dict to use as the quality axis.
Key in each system’s summary dict to use as the cost axis.
dict[str, int] — system name → rank (1 = best).
Usage
When it is enabled
The CLI automatically runsParetoRank.rank_systems() and writes pareto_rank
into each system’s summary when two or more --proxy flags are provided.
rank_systems() manually after evaluate().
The per-system
compute() method returns {"pareto_rank": 0.0} as a
placeholder. Always use rank_systems() for the actual ranking — it is
the only method that sees all systems at once.