Scoring
Overview
Each server receives a single composite score in the range , where is the top performer in the run. The algorithm is inspired by TechEmpower’s Framework Benchmarks: tests are normalised to a common throughput unit so that operations with naturally high RPS (simple lookups) don’t dominate over operations with naturally low RPS (large expansions), then weighted by a per-test bias coefficient before being summed.
Composite score
Step 1 — Effective RPS
For each (server, test) pair, collapse the three VU levels into a single number by taking the maximum effective RPS across levels:
Multiplying by penalises servers that sustain throughput by returning errors.
Step 2 — Normalising weight
Tests differ in natural throughput — a simple lookup returns results orders of magnitude faster than a large intensional expansion. A normalising weight scales every test’s contribution to the same RPS unit as the reference test (LK01 — simple SNOMED CT lookup):
where is the mean eRPS for test computed over servers that actually ran the test (real results only — imputed values are excluded to prevent feedback loops between imputation and normalisation).
LK01’s own weight is always by definition.
Step 3 — Weighted RPS
Apply the normalising weight and a per-test bias coefficient :
Bias down-weights operations that are less representative of real production workloads. See the Tests page for the full bias table.
Step 4 — Raw score
Sum weighted RPS across all tests:
Step 5 — Normalise to 0–100
The top server always scores 100. All others are expressed as a percentage of that.
Imputation
Servers that did not participate in a test (failed preflight or did not run it) receive an imputed eRPS rather than zero. This prevents a server from being unfairly penalised for a capability gap when the gap is already reflected in the capability matrix.
The imputed value is the 15th percentile of the participating servers’ eRPS for that test, anchored at zero:
Anchoring at zero means: the fewer real participants there are, the lower the floor. A server absent from a test it could have won receives less benefit of the doubt than one present with real results.
Imputed values are used only in the weighted sum (Step 4). They are excluded from the normalising weight calculation (Step 2).
Exclusions
A server is excluded from a test if it:
- returned HTTP
4xxor5xxfor the operation during preflight - returned a semantically incorrect response during preflight (wrong resource type, missing required fields, incorrect values)
Exclusions are recorded in the capability matrix. Excluded tests are handled via imputation rather than being scored as zero.