Skip to main content
Insight Article4 min read

The UK Buyer's Guide to Data Annotation Outsourcing

How UK AI teams scale data annotation projects without hiring. Costs, tools, quality frameworks & outsourcing strategy.

Insight ArticleTTreba Research4 min read

What Is Data Annotation and Why Do AI Teams Struggle with It?

Data annotation is the process of labelling raw data—images, text, audio, or video—so machine learning models can learn from it. Without annotated data, AI models cannot be trained effectively.

The types of annotation tasks include: image labelling (assigning objects to categories), bounding boxes (drawing rectangles around objects), semantic segmentation (pixel-level classification), text classification (assigning sentiment, intent, or topic to sentences), named entity recognition (identifying people, places, organisations), and audio transcription (converting speech to text with timestamps).

Why do UK AI teams struggle? Volume is the first challenge. A single model training project can require 10,000–100,000+ annotated samples. In-house teams cannot scale fast enough. Cost is the second challenge. Hiring full-time annotators at UK salary rates (£25,000–£30,000 annually) is expensive for short-term or variable-volume projects. Quality consistency is the third. Maintaining inter-annotator agreement across a growing team is labour-intensive.

Real-World Volume Scenario

A UK fintech company building a fraud-detection model needed 50,000 labelled transaction screenshots. At UK salaries, this would cost £8,000+ in labour alone (at 16p per annotation). Outsourcing to Kenya reduced costs to £2,000 (at 4p per annotation) whilst maintaining 92% inter-annotator agreement.

Annotation Types and Complexity Levels

Annotation tasks range from simple to highly technical. Understanding the complexity tier helps you estimate timelines and vendor capability.

Comparison

Line ItemUK (London)Treba (Nairobi)Saving
Annotation TypeComplexityUK Cost per 1,000Kenya Cost per 1,000
Image ClassificationLow£180£45
Bounding Box (1–3 objects)Medium£400£100
Semantic SegmentationHigh£800£200
Text Classification (sentiment)Low£150£40
Named Entity Recognition (NER)Medium£350£85
Audio Transcription + TimestampsHigh£600£150

Cost differences reflect not just wages, but productivity. Kenyan annotators trained by Treba average 35–40% higher throughput due to structured workflows and tool familiarity.

Quality Control Frameworks and Inter-Annotator Agreement

Outsourcing does not mean sacrificing quality. Professional providers use structured QA frameworks to maintain consistency.

Four-Layer Quality Control

Layer 1: Annotation Guidelines. Detailed written rules (e.g., "label all cars in profile view; omit toy cars and drawings") prevent ambiguity. Guidelines should include 10–20 visual examples.

Layer 2: Gold Standard Sets. A small subset (5–10%) of annotated samples validated by a subject-matter expert. New annotators calibrate against these before production work.

Layer 3: Inter-Annotator Agreement (IAA) Measurement. Cohen's Kappa or Fleiss' Kappa measures agreement across annotators. Industry standard: Kappa > 0.80 (80% agreement). Treba maintains 0.85–0.92 Kappa across most projects.

Layer 4: Spot-Check Audits. Random 5–10% sample review by a senior annotator or QA lead. Failed samples trigger retraining or task reassignment.

Example: Image Bounding Box QA

A UK insurance company needed 40,000 vehicle damage photos labelled with bounding boxes. Treba implemented: (1) 15-page annotation guide with 40 example images, (2) 500-sample gold standard set, (3) bi-weekly IAA audits (target Kappa > 0.88), (4) 10% spot-check audits. Result: 98% first-pass accuracy, project completed in 6 weeks.

Annotation Tools: CVAT, Labelbox, Scale AI, Prodigy

Your choice of tool impacts cost, speed, and quality. Here's a quick comparison:

Comparison

Line ItemUK (London)Treba (Nairobi)Saving
ToolBest ForStrengthsUK Cost/mo
CVAT (Open Source)Image, video, 3DFree, on-prem, full control£0–£200
LabelboxImage, text, multi-modalCollaborative, auto-labelling£400–£2,000
Scale AIEnterprise, managed serviceOutsourced QA, complex tasks£3,000+
ProdigyNLP, text, active learningFast iteration, weak supervision£50–£300/mo

For UK teams outsourcing to Kenya, we recommend CVAT or Prodigy deployed on a shared server accessible from Nairobi. This gives you cost control and transparency.

Team Structure and Cost Breakdown

A typical outsourced annotation project follows this structure:

Comparison

RoleResponsibilityTypical Cost (Kenya)
RoleResponsibilityUK Annual Cost
Annotation Lead / QA ManagerGuideline creation, QA audits, vendor management, retraining£35,000–£45,000
Data Annotators (team of 5)Label data per guidelines, flagging edge cases£125,000–£150,000 (5 × £25–£30k)
Project Manager (part-time, 0.5 FTE)Timeline tracking, communication with vendor, sample review£20,000–£25,000 (0.5 FTE)
Subject Matter Expert (ad hoc)Gold standard validation, difficult cases, process improvement£15,000–£20,000

Total cost (5 annotators, UK in-house): £195,000–£240,000 annually. Total cost (5 annotators, Kenya outsourced): £39,000–£57,000 annually. Saving: 75–80%.

Vendor Selection and SLA Frameworks

Not all outsourcing providers are equal. Look for:

Key Selection Criteria

  • 1. Domain expertise. Have they annotated for computer vision, NLP, or your specific use case before? Ask for 2–3 references.
  • 2. Scalability proof. Can they reliably deliver 5,000+ annotations per week? Request a small pilot project (500–1,000 samples) before committing to full volume.
  • 3. Quality guarantees. Do they commit to a minimum Kappa score? Offer penalties for falling below target?
  • 4. Pricing transparency. Fixed per-task rates are better than hourly billing. Example: £0.04 per image classification, £0.12 per bounding box.
  • 5. Security and compliance. GDPR data processing agreement (DPA), ISO 27001 certification, and encrypted communication channels.

Sample SLA Framework

  • Turnaround time: 70% of tasks delivered within 5 business days of submission.
  • Quality: Cohen's Kappa ≥ 0.85 for 95% of projects.
  • Availability: Minimum 95% team attendance month-on-month.
  • Escalation: Edge cases escalated to senior annotator within 24 hours.

Key takeaways

1

• Data annotation is labour-intensive; outsourcing cuts costs by 60–75% whilst maintaining quality. • Quality control relies on annotation guidelines, gold standard sets, inter-annotator agreement (Kappa > 0.80), and spot-check audits. • Tool choice (CVAT, Labelbox, Prodigy) impacts cost and speed; consider shared-server deployments for overseas teams. • Team structure: 1 QA lead, 5 annotators, 0.5 project manager = £39,000–£57,000/year outsourced vs. £195,000–£240,000 in-house. • Vendor selection should prioritise domain expertise, scalability proof (pilot projects), quality guarantees, and GDPR compliance.

T

Written by

Treba Research

Treba editorial team — expert analysis on outsourcing, compliance, and building distributed UK–Kenya teams.


FAQ

Frequently Asked Questions

WE ARE TREBA

Need Help Scaling Annotation? Our team has delivered 2M+ annotations across 50+ AI projects.

Let's discuss your requirements.