CaraComp
Log inTry Free
CaraComp
Forensic-Grade AI Face Recognition for:
Start Free Trial
facial-recognition

What 99% Accurate Really Means for Face Recognition

What "99% Accurate" Really Means in Facial Recognition

Here's a number that should make any investigator set down their coffee: a facial comparison system can score 99% accuracy in a published benchmark test and still be wrong on 1 out of every 10 genuine matches it encounters in the field. Not because the vendor lied. Not because the algorithm is broken. But because of something far more uncomfortable — the number itself was never telling you what you thought it was.

TL;DR

A single accuracy percentage conceals three metrics that actually matter for investigators: False Accept Rate, False Reject Rate, and demographic consistency — and most vendors are hoping you never ask about any of them.

This isn't a niche technical complaint. It's the difference between a lead that holds up under cross-examination and one that quietly unravels it. Understanding how accuracy is actually calculated — and where it hides its failures — is one of the most practically useful things an investigator can learn about facial recognition technology right now.

So let's get into it.


The Benchmark Trap: Why Lab Numbers Don't Survive Contact With Reality

Most accuracy claims in facial recognition trace back to a handful of respected evaluation programs, with NIST's Face Recognition Vendor Testing (FRVT) being the gold standard. These evaluations are genuinely rigorous and genuinely valuable. They create a consistent playing field, and top performers — companies like Regula and NEC — compete fiercely for ranking positions because the results carry real credibility.

But here's the part nobody puts in the press release: NIST evaluations are conducted primarily on high-quality, frontal, well-lit imagery. Think controlled mugshot conditions. Controlled datasets. Controlled everything. NIST itself explicitly cautions that benchmark rankings do not translate directly to operational performance in real-world deployments. That caveat tends to get lost somewhere between the algorithm lab and the marketing department.

What happens when that same algorithm meets real-world footage? Faces turned 30 degrees. Grainy CCTV captures from 40 feet away. Subjects wearing hats, sunglasses, scarves, or just the natural disguise of ten years of aging. Under those conditions, documented accuracy degradation is consistent and measurable — systems that score above 99% in controlled benchmarks routinely drop to the 70–80% range in genuinely uncontrolled environments. This article is part of a series — start with Deepfake Detection Accuracy Gap Investigator Workf.

Think of it this way: quoting a single benchmark accuracy number for facial comparison is like advertising a car's fuel economy using only highway driving data. Technically true. Completely misleading the moment you hit city traffic, a rainstorm, or a steep hill. Benchmark conditions are the highway. Real investigations are city traffic — and the hills are everywhere.

"Reaching the highest accuracy in the NIST evaluation proves the strength of our forensic-driven approach and biometric verification expertise. Just as important, the results confirm that Regula performs consistently across a wide range of real-world conditions, making our solution the most universal on the market." — Ihar Kliashchou, CTO, Regula, via Biometric Update

Notice what Kliashchou emphasizes: consistency across real-world conditions. That's not an accident. It's exactly the right thing to highlight — because it's exactly what a single accuracy number cannot tell you.


Metric #1 and #2: The Two Failure Modes Hidden Inside One Number

Here's the math problem nobody explains clearly enough. When vendors report "99% accuracy," they're typically reporting a composite figure — a blend of how often the system correctly identifies true matches and how often it correctly rejects non-matches. In most real-world comparison datasets, genuine non-matches vastly outnumber genuine matches. Which means the composite score is dominated by how well the system rejects strangers — not by how well it finds your suspect.

A system could correctly reject 99.9% of non-matching pairs and still be catastrophically wrong on the match side. And you'd never see it in the headline number.

This is why investigators need to demand two separate rates:

The Two Failure Modes That Matter

  • False Accept Rate (FAR) — How often the system says "match" when it isn't one. This is the wrongful accusation risk. A high FAR means you're generating false leads, potentially pursuing innocent people, and building a case on sand.
  • 🔍 False Reject Rate (FRR) — How often the system says "no match" when there actually is one. This is the missed perpetrator risk. A high FRR means your actual subject is walking away clean while the system waves them through.
  • ⚖️ The Threshold Trade-Off — FAR and FRR are not independent. They're locked in a seesaw relationship controlled by a single variable: the system's match threshold. Tighten it to reduce false positives, and false negatives automatically increase. Loosen it to catch more true matches, and false positives climb. Every vendor has made a choice about where to set that dial — and almost none of them volunteer which direction they've tuned it, or why.

That last point is worth sitting with. The threshold setting isn't just a technical parameter — it's a policy decision disguised as a technical one. A system tuned for high-security access control (where false accepts are catastrophic) will behave very differently than one tuned for investigative triage (where missing a match is the bigger problem). Same algorithm. Completely different error profile. The accuracy number won't tell you which one you're dealing with.

For investigators who want to understand how to push for better results from their existing tools, this breakdown of practical techniques for improving facial comparison outcomes covers how threshold settings and image quality interact in ways most vendor documentation never mentions. Previously in this series: Deepfake Laws Changed Evidence Standards Investiga.


Trusted by Investigators Worldwide
Run Forensic-Grade Comparisons in Seconds
2 free forensic comparisons with full reports. Results in seconds.
Run My First Search →

Metric #3: The Demographic Consistency Problem

This is the one that tends to make vendor representatives suddenly very interested in checking their phones.

100x
The factor by which false positive rates varied across demographic groups on some algorithms tested in NIST's landmark 2019 FRVT report
Source: NIST Face Recognition Vendor Testing (FRVT), 2019

One hundred times. Not 10% worse. Not twice as bad. One hundred times higher false positive rates for certain demographic subgroups compared to others, on the same algorithm, in the same evaluation. An aggregate accuracy number hides this entirely — because the errors aren't distributed evenly. They concentrate.

What this means practically: a system that reports 99% overall accuracy might be performing at 99.8% for one demographic group and 92% for another. Those two numbers average out to something that sounds impressive on a slide deck. In an actual investigation involving individuals from underrepresented groups in the training data, that system is operating with significantly less reliability — and the investigator has no way of knowing that from the headline figure alone.

The right question to ask any vendor isn't "what's your accuracy?" It's "what's your false positive rate broken down by age, gender, and ethnicity — and at what threshold?" If they can't answer that, or suddenly need to "follow up with the technical team," you have your answer.


What "Court-Ready" Actually Requires

Look, nobody's saying benchmark testing is meaningless. NIST's FRVT program produces genuinely useful comparative data, and strong performance in those evaluations is a legitimate signal of algorithmic quality. The problem isn't that vendors test in controlled conditions — all standardized testing requires controlled conditions. The problem is treating the result as the complete story.

For evidence that needs to survive adversarial scrutiny — depositions, cross-examination, defense challenges — investigators need to be able to answer three specific questions about any facial comparison result: Up next: How Facial Recognition Accuracy Is Really Measured.

First: At what threshold was this match generated, and what is the documented false accept rate at that threshold? Second: Has this system's performance been validated on image quality and demographic conditions comparable to the evidence in this case? Third: Is there documented demographic consistency data, or is the published accuracy figure an aggregate that obscures subgroup variation?

If the answer to any of those is "I don't know" or "the vendor didn't provide that," the match is a lead — not evidence. That's a meaningful distinction in any proceeding where someone's freedom is at stake.

Key Takeaway

The accuracy number vendors publish measures how their system performs under ideal conditions on a balanced dataset. The three numbers investigators actually need — False Accept Rate, False Reject Rate, and demographic consistency across subgroups — are almost never in the headline, and almost always available if you know to demand them.

Here's the real punchline, though — the thing worth remembering long after you've closed this tab.

The metric that sounds best in a press release is mathematically guaranteed to be the least useful metric for an investigator. Overall accuracy is high because true non-matches dominate real-world datasets, and systems are good at rejecting strangers. The hard problem — finding a real match in messy, real-world conditions, reliably, across all the people you might encounter — is exactly where the number stops telling the truth.

When a vendor hands you a 99% accuracy figure, the only question that matters is this: 99% of what, exactly? Because depending on the answer, that number might be the most confident way anyone has ever told you almost nothing.

Ready for forensic-grade facial comparison?

2 free comparisons with full forensic reports. Results in seconds.

Run My First Search