This article walks the practitioner through the design of a labeling strategy, the measurement of
Why labeling is a first-class readiness concern
In a supervised model, the labels are the ground truth the model tries to reproduce. If the labels are wrong, the model learns the wrong thing and no amount of model-side tuning will undo the damage. In a preference-tuned model (including most modern large language models in their RLHF phases), the preference labels shape the model’s behavior in ways that are hard to audit after training. The Time investigation published in January 2023 on Sama-operated content-moderation labeling for OpenAI described the human conditions in which labels were produced and raised questions about label quality and labeler welfare that the practitioner cannot ignore in a readiness assessment.3
A readiness practitioner’s labeling audit covers four questions:
- Is there a written rubric, and can three independent labelers produce the same labels from it?
- Is there a measurement of inter-annotator agreement, and does it meet a threshold appropriate to the risk tier?
- Is there an adjudication process for disagreements, and does it produce an auditable record?
- Is there an ongoing quality program, or was the labeling done once as a launch-gate event?
Weak answers to any of the four are readiness findings. Strong answers to all four are the baseline from which higher-risk use cases can proceed.
The labeling rubric
A
A good rubric has five properties:
- Complete — covers the expected input distribution without substantial gaps.
- Testable — a labeler can apply it and produce a label for every instance.
- Calibrated — worked examples reflect the actual distribution, not just clean edge cases.
- Versioned — rubric changes are dated, documented, and associated with the labels produced after each change.
- Auditable — the rubric and its version history are retained as evidence.
A rubric that fails any of these properties will produce label noise the practitioner will find only after the model starts misbehaving in production. Rubric authoring is the first line of defense.
The gold set
A
A gold set should cover the class distribution, should include difficult cases, and should be refreshed when the rubric changes. A labeling operation with no gold set cannot measure its own quality.
Reviewer plurality
Most labels should be produced by more than one labeler. Plurality — the number of labelers per instance — is a primary quality lever.
- Single-labeler — one labeler per instance. Cheap but noisy. Appropriate only where the task is easy, the risk is low, and ongoing quality sampling is strong.
- Double-labeler with tiebreaker — two labelers per instance; disagreements are sent to a third (senior) labeler. The common workhorse regime for medium-risk tasks.
- Triple-labeler with majority rule — three labelers per instance; majority wins on disagreements. Strong for moderately difficult tasks and allows direct measurement of Fleiss’s kappa.
- Triple-labeler with adjudication — three labelers, and all disagreements (not just majority failures) are sent to a senior adjudicator. Strong but expensive; appropriate for high-risk tasks.
The readiness practitioner should not default to the highest plurality. The right regime is the one matched to the use case risk, the task difficulty, and the rubric’s measured reliability. A high-risk use case with a stable rubric may be well-served by double-labeler; a medium-risk use case with a volatile rubric may need triple-labeler adjudication.
[DIAGRAM: MatrixDiagram — plurality-adjudication-regimes — 2x2 of “Reviewer plurality (low / high)” against “Adjudication rigor (low / high)”, mapping the four quadrants to named labeling regimes: light-touch (low-low), double-with-tiebreaker (high-low), spot-audit (low-high), triple-with-adjudication (high-high); each cell annotated with typical use-case risk tier and cost multiplier]
Measuring inter-annotator agreement
Agreement is measured with named statistics. The practitioner should understand which statistic fits which task.
Cohen’s kappa (κ) — two annotators on a categorical task. Ranges from −1 to 1, with 0 meaning chance agreement and 1 meaning perfect agreement. κ ≥ 0.6 is usually considered substantial; κ ≥ 0.8 is usually considered near-perfect. The practitioner should never accept κ < 0.4 as evidence of a usable dataset without intensive rubric and training remediation.
Fleiss’s kappa — three or more annotators on a categorical task. Interpretation is similar to Cohen’s κ. Fleiss’s formulation requires that each instance be labeled by the same number of annotators.2
Krippendorff’s alpha (α) — more general agreement measure that handles missing data, multiple annotators per instance with varying number, and ordinal, interval, or ratio data. For many enterprise labeling tasks Krippendorff’s α is the right statistic because it handles the messiness of real operations.1
Intersection-over-union (IoU) — for bounding-box and segmentation tasks. Agreement is computed geometrically rather than categorically. Thresholds are task-dependent (IoU ≥ 0.5 is common for object detection; IoU ≥ 0.75 is demanding).
The readiness scorecard records the statistic used, the threshold, the observed value, and the sample size. Agreement measured on a tiny sample is not evidence; the practitioner should require a minimum of 50 instances per class for categorical tasks, adjusted upward for rare classes.
The labeling operation lifecycle
A labeling operation is a program, not an event. Six stages:
- Rubric draft — the data-science team, with domain experts, drafts the rubric and the gold set.
- Pilot — a small cohort of labelers applies the rubric to a pilot set. Disagreements surface the rubric gaps.
- Calibration — the pilot findings drive rubric revision and labeler training.
- Production — labeling proceeds at volume with the calibrated rubric; gold-set instances are seeded throughout.
- Audit — an independent reviewer samples labeled instances to detect drift, with statistics recomputed on the audit sample.
- Refresh — the rubric is updated, labelers are retrained, and prior labels may be rescored against the new rubric.
The operation’s governance record is the audit artifact. It should show dates, participant counts, agreement statistics at each stage, disagreement rates, and the decisions that followed.
[DIAGRAM: StageGateFlow — labeling-operation-lifecycle — six-stage horizontal flow (rubric draft → pilot → calibration → production → audit → refresh), each stage annotated with the gate criterion (is the rubric complete? is agreement above the threshold? is the gold-set pass rate above the threshold?) and the evidence produced at each gate]
Third-party and platform-assisted labeling
Most enterprise labeling is not done in-house. The practitioner must assess third-party and platform-assisted labeling with distinct readiness controls.
Third-party labeling services — firms that operate dedicated labeler workforces. The readiness controls are: labeler vetting, rubric transfer, agreement measurement on a pilot batch, ongoing audit access, and labeler-welfare evidence (especially for content-moderation and sensitive-content tasks; the Time investigation on Sama is the cautionary anchor).3 The readiness practitioner should require a right-to-audit clause in the vendor contract and should exercise it.
Platform-assisted labeling — automated systems that pre-label instances using a prior model or a rules engine, with humans confirming or correcting. Platform-assisted labeling can substantially reduce cost at the risk of inheriting the prior model’s biases. The readiness practitioner should require: measurement of the platform’s error rate against a gold set, human-review coverage of at least the cases where the platform is uncertain, and a periodic audit where humans re-label from scratch to compare against the platform-assisted pipeline.
RLHF / preference labeling — the practitioner should treat preference labeling as a specialized case. Preference labels shape model behavior in ways that are harder to audit than categorical labels. Public coverage of large RLHF programs at scale (including the operations described by Scale AI and peers in 2022-2023 industry coverage) documents the governance challenges.4 The readiness controls are: documented rubric for preference rankings, plurality of at least three rankers per preference pair, measurement of rank-correlation agreement (Kendall’s τ or Spearman’s ρ), and ongoing audit.
Cost modeling and budgeting for labeling operations
Labeling cost is the primary friction against label quality. Higher plurality, stronger adjudication, and better-trained labelers all cost money. The readiness practitioner should be able to estimate labeling costs so that quality tradeoffs can be made explicitly rather than by accident.
A reasonable first-cut model: cost-per-label = (labeler-hourly-rate / labels-per-hour) × plurality × (1 + adjudication-rate × adjudicator-rate-multiplier) + platform-fee-per-label + quality-audit-overhead.
For concreteness, a medium-difficulty categorical task at triple-labeler plurality with 10% adjudication rate and 30% quality-audit overhead typically costs three to five times the single-labeler baseline. A high-risk task with expert annotators can cost ten to twenty times the baseline. The practitioner should present the cost range alongside the quality regime in the readiness scorecard so the sponsor’s approval of a regime is also an approval of its cost.
Where cost pressure is forcing a regime below the quality threshold the risk tier requires, the practitioner has three options: re-scope the use case to a lower risk tier with laxer requirements; raise the budget with a business-case defense; or recommend the use case does not proceed at the current budget. Quietly accepting sub-threshold labeling is not an option.
Consistency across labeling refresh cycles
Labels produced in one quarter may not be comparable to labels produced in another quarter even against the same rubric. Annotator drift, rubric version changes, gold-set evolution, and operational context all introduce inconsistency across time. The practitioner’s check:
- Cross-cycle gold-set pass rate. Seed the same gold-set instances across cycles. If the pass rate changes across cycles, agreement has drifted.
- Reviewer-turnover impact. Track the labeler roster; significant turnover requires accelerated calibration for new labelers.
- Historical label re-scoring. When a rubric changes materially, old labels should be re-scored under the new rubric, or the training set should be partitioned by rubric version.
A labeling operation without these controls accumulates hidden inconsistency that the model inherits as noise.
Bias-introduction risk in labeling
Labelers bring their own context to the task. A rubric that does not close this down introduces bias from the labeling process itself. Recurring patterns:
- Cultural bias in content moderation — what counts as hate speech, acceptable political content, or a medical claim varies across cultures. A single-culture labeling team will produce labels that reflect their culture’s norms, and the resulting model will reflect those norms regardless of deployment context.
- Acquiescence bias — labelers defer to an implicit expectation in the rubric. If the rubric examples skew toward one class, labelers’ calibration drifts toward that class.
- Fatigue bias — agreement degrades late in a shift, especially on difficult tasks. Shift-level agreement monitoring catches the drift.
- Experience-skew bias — senior labelers apply implicit criteria junior labelers cannot see. Rubric refresh after each audit catches the implicit drift.
The readiness practitioner should expect a labeling program to have a written plan for each of these patterns. Where the plan is missing, the scorecard captures a gap.
Cross-references
- COMPEL Core — Data governance for AI (
EATF-Level-1/M1.5-Art07-Data-Governance-for-AI.md) — labeling is governed as a data-producing process. - COMPEL Practitioner — Data quality and technology assessment deep dive (
EATP-Level-2/M2.2-Art06-Data-Quality-and-Technology-Assessment-Deep-Dive.md) — the practitioner-level treatment of quality, of which labeling quality is a subset. - AITM-DR Article 7 (
./Article-07-Bias-Relevant-Variables-and-Subgroup-Coverage.md) — labels are one of the primary vectors through which bias enters a training set. - AITM-DR Article 11 (
./Article-11-The-Readiness-Scorecard.md) — labeling agreement is one of the ten scored dimensions.
Summary
Labeling is a first-class readiness concern because labels shape what the model learns. A labeling strategy includes a written rubric, a gold set, a chosen plurality regime, a measured agreement statistic, a documented adjudication process, and an ongoing audit cadence. Third-party and platform-assisted labeling add specific controls around vendor oversight and automated-labeler error rates. The readiness scorecard records the agreement statistic, the threshold, the observed value, and the sample size. A labeling operation without these controls is a noise source the downstream model will inherit.
Footnotes
-
K. Krippendorff, Content Analysis: An Introduction to Its Methodology, 2nd ed., Sage (2004). ↩ ↩2
-
J. L. Fleiss, Measuring nominal scale agreement among many raters, Psychological Bulletin 76, no. 5 (1971): 378–382. ↩ ↩2
-
B. Perrigo, Exclusive: OpenAI Used Kenyan Workers on Less Than $2 Per Hour to Make ChatGPT Less Toxic, Time, January 18, 2023. https://time.com/6247678/openai-chatgpt-kenya-workers/ ↩ ↩2
-
Scale AI, Case Studies, https://scale.com/case-studies (industry coverage of RLHF program operations, 2022–2024). ↩