Skip to content

Revision Rate

  • What revision rate is and how it is calculated
  • How revision rate differs from precision, recall, and accuracy
  • How to interpret high and low revision rates
  • Why unreviewed annotations matter for this metric

Revision rate measures the overall human correction effort required for a given model. While metrics like precision and accuracy focus on specific types of errors, revision rate gives you a single number that captures how much work humans are doing to fix what the model produces.

RevisionRate = RevisedAnnotations / TotalAnnotations x 100

Where:

  • RevisedAnnotations = the number of distinct ML annotations that have at least one non-Confirmed revision (such as Deleted, BoxResized, SideChanged, AnimalIdChanged, or HumanAdded).
  • TotalAnnotations = all IsMachine annotations attributed to this model, whether they have been reviewed or not.

How revision rate differs from other metrics

Section titled “How revision rate differs from other metrics”

The key distinction is in the denominator. Precision, recall, and accuracy only count reviewed annotations — they tell you about model quality among the annotations that humans have actually looked at. Revision rate counts all ML annotations, including those not yet reviewed.

This matters because it reflects the real-world burden on your team:

MetricDenominatorWhat it tells you
PrecisionConfirmed + DeletedHow often the model detects something real (among reviewed)
RecallConfirmed + HumanAddedHow often the model finds everything (among reviewed)
Side AccuracyConfirmedHow often the side is correct (among reviewed)
ID AccuracyConfirmedHow often the ID is correct (among reviewed)
Revision RateAll ML annotationsHow much of the model’s total output required human correction
  • Low revision rate (e.g., 5-10%) — The model’s output is largely correct. Most annotations either pass through review without changes or have not yet needed review.
  • Moderate revision rate (e.g., 15-25%) — A meaningful portion of annotations require correction. This is common for actively reviewed populations where the team is thorough.
  • High revision rate (e.g., 30%+) — Humans are spending significant effort correcting the model. This suggests the model may benefit from retraining, or that the population’s data has characteristics the model was not trained to handle.

Because TotalAnnotations includes unreviewed annotations, revision rate is influenced by how much of your data has been reviewed. Consider two scenarios:

Scenario A: 1,000 total ML annotations, 500 reviewed, 50 revised. Revision rate = 50 / 1,000 = 5%.

Scenario B: 1,000 total ML annotations, 900 reviewed, 90 revised. Revision rate = 90 / 1,000 = 9%.

Both populations have the same underlying correction rate among reviewed annotations (10%), but Scenario B shows a higher revision rate because more annotations have been reviewed. As you review more data, revision rate converges toward the true correction rate.

Revision rate is particularly valuable for:

  • Comparing models over time — If revision rate trends upward across daily snapshots, the model may be degrading relative to incoming data.
  • Estimating review workload — A revision rate of 20% means roughly one in five annotations will need human correction. This helps plan reviewer time.
  • Justifying retraining — A consistently high revision rate provides concrete evidence that a model retrain would save human effort.