Why Yoti’s facial age estimation is not facial recognition

profile picture Amba Karsondas 9 min read
Image of woman's face being estimated for age, which is not the same as facial recongition

There’s quite a bit of confusion about the differences between facial age estimation and facial recognition. While both types of technology work with images of faces, they’re used for different reasons and are trained in different ways.

To help clear up some of these misconceptions, we’ve explained some of the key ways that our facial age estimation is not facial recognition.

 

Facial age estimation vs. facial recognition: designed to give two different outcomes.

Facial age estimation delivers an estimated age result. Facial recognition delivers a match (or no match) between images of a person.

Facial age estimation delivers an estimated age result. Facial recognition delivers a match (or no match) between images of a person.

The objective of facial age estimation technology is completely different and separate from the objective of facial recognition.

In its simplest form, the purpose of facial recognition is to either identify or confirm a person’s identity from their face. Facial recognition is able to match one face (from either a digital image or video) with another face. This could be from a photo on an ID document or from images stored within a database.

With this in mind, facial recognition technology is built to deliver one of two results: a match or no match. If facial recognition is used to match a face to one in a database, it may pull up all the information stored within that database about the person it has recognised. This could range from personal information such as a name or date of birth, to browsing habits or social media accounts. Basically, whatever information has been previously inputted into that database.

On the other hand, the purpose of facial age estimation is to estimate a person’s age from a facial image. The result delivered by facial age estimation differs from facial recognition.

Instead of delivering a match or no match, facial age estimation technology is built to deliver a numerical value. Depending on regulatory requirements, these could be one of three results. If we take the example of a 15-year-old, these three results could be:

  • 15 years old (i.e. an exact estimated age)
  • between 13 and 17 years old (i.e. an estimated age range)
  • over 13 or under 18 (i.e. an “over” or “under” age result)

Once this age result is delivered, this will indicate whether the person (whose identity is unknown) will be able to access the relevant age-restricted goods, services or experiences.

 

Facial age estimation is trained in a different way to facial recognition.

“Facial age estimation is trained on photos accompanied by known-age values. Facial recognition is trained on pairs of photographs with identity labels.” – NIST, 2024

Facial age estimation is trained on photos accompanied by known-age values. Facial recognition is trained on pairs of photographs with identity labels.

Facial age estimation and facial recognition both rely on a modern technique called “deep learning” and they both start with a face. But that’s where their similarities end. The two models are trained with entirely different objectives and on different data.

Our facial age estimation doesn’t have the ability to match or identify faces because it simply hasn’t been taught how to do so.

In fact, if a facial age estimation model was able to identify the people it had been trained on, it wouldn’t be suitable for the task of estimating age. This is because it would always return the same age for these people even as they continue to age. For example, if someone was 16 when the model was trained, it would always estimate that person as being 16. This wouldn’t be useful if that same person wanted to, for instance, buy age-restricted goods a few years later, and needed to prove that they were over 18.

Image showing how facial age estimation works: 1) Face is detected and reduced to pixels then numbers for AI to understand; 2) numbers are computed by a neural network that has been trained to recognise age by looking at millions of images of faces; 3) the AI finds a pattern in the numbers and produces an age. Note at the bottom to be clear that no images are stored during the estimation process, that the technology is scalable to tens of millions of checks a day, and that the process is instantaneous.

Facial age estimation is performed by a ‘neural network’. The algorithm has learned to estimate age in the same way humans do – by looking at faces.

It works by converting the pixels of a facial image into numbers and comparing the pattern of numbers to patterns associated with known ages. The technology detects a live human face, analyses the pixels in the image and gives an age estimate.

Facial recognition works in a different way to facial age estimation. The National Institute of Standards and Technology (NIST) states that facial recognition “compares identity information extracted from two photos with the goal of determining if they are from the same person. The first sample is stored in a passport, or phone, or database. The second sample is collected for verification or identification”. The purpose of facial recognition is to match the new facial image with one that has been previously captured, and therefore with all the data associated with that image.

Unlike facial recognition, facial age estimation technology is not trained to match an image of one face to the image of another. It simply analyses the pixels of the facial image, delivers the age result, such as ‘over 18’ and then deletes the image. It’s not able to identify people or recognise faces.

 

Facial recognition relies on existing images. Facial age estimation has no existing images.

For facial recognition to work, it needs to match a new face to an existing image. Facial age estimation cannot match faces as no images are saved or stored in the first place.

Image showing that facial recognition relies on existing images. Facial age estimation has no existing images.

For facial recognition to take place, there needs to be at least one other facial image for the technology to match the original image to. This is so the system can compare the image of the face presented to it with all existing images. Otherwise, there’s no way of a person “being recognised”.

When facial age estimation is used in practice, facial recognition cannot take place as there’s no database to begin with. There’s nothing to compare the facial image to. This means that one face cannot be matched to another.

As soon as an age estimate is delivered, we permanently delete the facial images in real time. We do not use them for our own learning or training purposes. This is externally reviewed as part of our SOC 2 and PAS1296 assessments.

The images are not stored, shared, sold on or used for any other purpose. And since no data is saved, there’s no way for us to create a database of users.

 

During development, facial recognition models require different training images than facial age estimation models.

During the training and testing stages, facial recognition models require multiple facial images of each person, taken at different angles, with different cameras, in different environments. In contrast, facial age estimation models only require a single facial image of each person with a corresponding month and year of birth.

Image showing that during development, facial recognition models require different training images than facial age estimation models.

The first step in developing our facial age estimation model is what we call “the training stage”. This is when the algorithm is trained with accurate age data. The model is trained with lots of images, paired with their corresponding age values, until it arrives at a processing formula that works best.

At this stage, the only information we require is a single facial image and the corresponding month and year of birth (but not the exact day they were born).

After this, the second step is “the testing stage”. Here, the algorithm is tested on a new dataset to evaluate how well it performs. For this stage, the only information required is, again, a single facial image and the corresponding month and year of birth. We need to know how old the person in the photo is, so that we can check that the algorithm is correctly estimating their age.

Although facial recognition systems are trained in a similar way, the types of images required are vastly different.

Since the objective of facial recognition systems is to be able to match one face to another, it needs multiple images of the same person. These images should ideally be taken at different angles, with different cameras, in different environments. Preferably, it would also be trained on the same people over an extended period of time too. Otherwise, the model will only be able to recognise that person if they were in the same environment, position and lighting conditions as the original training image.

Therefore, even if an attempt was made to recognise a face using a facial age estimation model, it would perform terribly. This is because these two models are trained on different – and “orthogonal” – objectives. A facial recognition model is trained to consider images of the same person to be “similar”, regardless of their age. On the other hand, facial age estimation is trained to group together images of people of the same age, regardless of who they are.

 

Yoti’s facial age estimation is not facial recognition.

Our facial age estimation is not facial recognition as they’re designed to do fundamentally different things. Facial recognition systems are designed to match faces to other faces within an existing image. On the other hand, facial age estimation technology is trained to give an age estimate of the unknown face that is presented to it, and then delete the image.

As a result, facial recognition systems can’t estimate age, and similarly, our facial age estimation system can’t recognise faces.

If you’d like to take a deeper dive into exactly how our facial age estimation works, take a look at our latest white paper or get in touch.

Keep reading

An image of a girl lying on a bed and using a mobile phone.

Keep calm and carry on - what to expect ahead of Australia's age checks for social media

We’ve been talking to a number of media outlets ahead of Australia’s upcoming social media ban for under 16’s – set to kick in on 10th December – and thought it was worth sharing our insights more widely. After all, there’s plenty of speculation about what will actually happen: Will systems cope? Will there be issues accessing sites? Will everything grind to a halt? After supporting a number of countries across the world as they’ve rolled out age assurance measures (including the UK Online Safety Act’s launch in July this year), we can say with confidence that the infrastructure

7 min read
An image of a young woman wearing a Santa hat and smiling whilst she's using her mobile phone. An illustrative icon of a shield with a checkmark is next to her, representing the theme of security.

Give scammers the cold shoulder this Christmas with your Yoti Digital ID

The Christmas lights are glowing and the treats are flowing. But with all the festive fun comes a flurry of extra online shopping and last-minute plans. So it’s no surprise that scams, fraud, identity risks and in-person safety concerns tend to spike right alongside the mince pies. That’s where your Digital ID can come in. A trusted, privacy-friendly way to prove who you are (and check who others are) can make all the difference. Here’s how your Digital ID can help you keep things merry and bright this Christmas.   Verify someone before you meet or trade Hunting for

6 min read
An image of Robin with accompanying text that reads "Thoughts from our CEO, Robin Tombs, November 2025".

Thoughts from our CEO

In this blog series, our CEO Robin Tombs will be sharing his experience, whilst focusing on major themes, news and issues in the world of identity verification and age assurance. This month, Robin talks about Yoti’s latest MyFace model, buffers in both Germany and Australia and recent government discussions around digital ID.   Yoti’s latest MyFace model achieves a perfect score for Level 1 and Level 2 liveness detection I’m delighted that Yoti’s latest MyFace passive liveness model has been independently tested by iBeta and is compliant with the ISO 30107-3 standard for Presentation Attached Detection (PAD) at both

9 min read