← Retour au Blog
EvaluationKYCAccess Control

How to Evaluate Face Recognition Models for KYC and Access Control

# How to Evaluate Face Recognition Models for KYC and Access Control

Picking a face recognition model is not just a benchmark exercise. KYC and access control workloads have very different operating points, error budgets, and downstream consequences. This article gives enterprise teams a structured way to evaluate face recognition candidates, including InsightFace.

Step 1: Define the operating point

Before running a single comparison, write down the operating point you actually need:

  • Verification (1:1) — comparing a live capture against a stored reference, e.g. KYC document selfie matching.
  • Identification (1:N) — searching against a gallery, e.g. employee or member access.
  • Watchlist (1:K, large) — searching against a much larger gallery with low base-rate matches.

The threshold, the false match rate, and the false non-match rate that matter to your business are not the same across these three.

Recommended targets

WorkloadTypical FMR targetTypical FNMR target
KYC verification1e-4 to 1e-6< 2% at the chosen FMR
Employee access control1e-5 to 1e-6< 1% at the chosen FMR
Watchlist (1:N large)1e-7 or stricterdepends on tolerance for misses

Treat these as a starting point. Your real targets come from your risk model.

Step 2: Build a representative validation set

Public benchmarks are useful but rarely match production. Build a private validation set that includes:

  • Demographics representative of your user base.
  • The capture devices and lighting your users actually have.
  • Edge cases you care about — masks, glasses, off-angle, partial occlusion, low-light.
  • Spoofing attempts if you are evaluating presentation-attack-detection separately.

If you cannot share data outside your environment, plan to run the evaluation on-premise or against an SDK.

Step 3: Run the comparison fairly

A few practical tips:

  • Run the candidate models on the same set, the same hardware, and the same preprocessing.
  • Report FMR and FNMR at the operating points you actually care about, not just AUC or accuracy.
  • Measure latency at the batch sizes you will deploy at.
  • Test cross-ethnicity and cross-age stability — not just headline numbers.

Step 4: Stress-test on the failures

Once the candidate models pass the headline metrics, look at the actual mistakes. Are they failing on the demographic groups that matter most? Are they failing on the device class your customers use most? A model that is 0.5% better on average but 5% worse on a specific cohort is usually not the right pick.

Step 5: Validate the deployment fit

Accuracy is necessary but not sufficient. Confirm:

  • Latency on your target hardware (cloud GPU, on-premise CPU, or device).
  • Memory footprint, especially for SDK and edge deployments.
  • The deployment model (Cloud API, on-premise, SDK) you will actually use in production.
  • Acceptable use boundaries set by the vendor.

Step 6: Decide and lock in

Pick the model that wins on your operating point and your data, not on the public leaderboard. Lock in the threshold, the deployment, and the licensing in the same conversation so you are not re-doing this work in six months.

Where InsightFace fits

InsightFace recognition models consistently rank among the top performers in NIST FRVT, but that is not why most enterprise teams pick them. The reason is that the private model evaluation process is designed to support exactly the workflow above: test on your own data, at your own operating point, on the deployment model you intend to use, before signing a commercial agreement.

Next steps

To start a structured evaluation against your own data, submit an enterprise inquiry with your use case (KYC verification, access control, or watchlist), expected volume, and target deployment.