Within the field of computer vision, facial recognition is an area of research and development which deals with giving machines the ability to recognize and verify human faces. Researchers primarily work on creating face recognition technology that can improve businesses and better human lives. To help strengthen your understanding of the technology, this guide will explain what facial recognition is, how it works, its various applications, and how accurate it is today.
What is Facial Recognition?
Facial recognition software has countless applications in consumer markets, as well as the security and surveillance industries. In fact, facial recognition technology is already being used to improve security protocols and payment procedures in China, and it is possible that the rest of the world will follow suit.
There are two main tasks that facial recognition models perform. The first is verification, which is the task of comparing a new input face to a known identity. A good example of this is the unlocking of smartphones with facial identification. When setting up the system, the phone will register your face as the phone’s owner. Therefore, the only task when unlocking is to compare new input faces to your registered face on the device.
The second is recognition, which is the task of comparing an input face to a database of multiple face identities. This task is often used for security and surveillance systems. A good example is facial recognition in law enforcement. On the INTERPOL website, there is a forensics section which explains how they use facial recognition to identify persons of interest at airports and border crossings.
How Does Face Recognition Work?
With a lot of interest in the field, data scientists develop new approaches to face recognition every year. This section will briefly discuss the basics of how face recognition models work and the main differences between two methods of creating facial embeddings.
At the very basic level, face recognition models follow these steps:
1. An input image is fed to the algorithm.
2. The algorithm creates a facial embedding for the input image.
3. The algorithm compares the input image’s facial embedding to the embeddings of known faces in the database.
Every approach has a different training method and researchers often adjust or add elements to established methods in the field. However, most systems employ triplet loss for the training of the algorithm. In regards to face recognition, triplet loss works by feeding the algorithm three images (see below).
Two of the images are of person A and the remaining image is of person B. The algorithm creates a facial embedding of each image and then compares them.
After the comparison, the network will be adjusted slightly so that embeddings of person A are more similar to each other than they are to the embedding of person B. Subsequently, this teaches the algorithm to use facial measurements that allow it to accurately classify images of the same person as being similar to each other. This process is then repeated hundreds of thousands or even millions of times. Finally, the network should then be able to produce accurate facial embeddings for faces it has never seen before.
Before the algorithm can compare faces, we must convert the face images into data that the algorithm can understand. To do this, the system calculates measurements based on facial features and landmarks. Fig. 2 is a visualization of 68 facial landmarks, also known as facial keypoints.
Depending on the neural network, these landmarks may or may not be used in creating the embedding. Sometimes the landmarks are just used to crop the face image and remove noise in the background of the image.
Deterministic Face Embedding
Essentially, by creating face embeddings you are converting a face image into numerical data. That data is then represented as a vector in a latent semantic space. The closer the embeddings are to each other in the latent space, the more likely they are of the same person.
However, the accuracy of models which use deterministic face embeddings depends on the clarity of the input images. Furthermore, such models are often tested under constrained settings. In practice, input images (from surveillance video for example) are often taken in unconstrained or uncontrolled settings. The image quality may be low or portions of the face may be covered in the image. In such cases, approaches that use deterministic face embeddings suffer in performance.
Probabilistic Face Embedding (PFE)
In a paper published in August of 2019, Yichun Shi and Anil K. Jain proposed an alternative approach to face recognition, PFEs. Probabilistic face embeddings represent face images as a distribution rather than a point estimation.
The image above provides a good visualization of the difference between deterministic and probabilistic embeddings. On the left, the facial embedding in the latent space is represented as a point estimation with no indicators to the uncertainty of the embeddings. Whereas, with PFEs, the mean of the distribution represents the best estimation of the facial features in the image. The variance in the distribution represents the uncertainty.
Of course, with blurry images where the facial features are missing or ambiguous, there will be a higher degree of uncertainty.
How Accurate is Face Recognition?
We’ve discussed what facial recognition is and how it works. However, perhaps the more important question is can we rely on it? How accurate is the facial recognition tech that we have today?
There is likely no single concrete answer to that question. Obviously, face recognition accuracy varies depending on the algorithm and the conditions we test it in. The table below presents what some of the current models are able to achieve. The data shows three well-known systems, the dataset they were tested on, and the resulting accuracy percentage.
Face Recognition Accuracy Table
|FaceNet||Labeled Faces in the Wild (LFW)||99.63%|
Data via Shi and Jain
Developed by Google, FaceNet achieves great results on LFW, a face image dataset with highly constrained settings. On the other hand, AM-Softmax was only able to achieve 84.69% (87.58% with PFEs) accuracy on IJB-A, a dataset of face images taken in unconstrained settings. Many of the images in the IJB-A dataset vary in pose, image conditions, and angle.
The data shows that many of the state-of-the-art methods perform relatively well under constrained settings. On the other hand, there is a substantial drop in accuracy when tested in unconstrained settings.
But when it comes to facial recognition, how good is good enough? A 99.63% accuracy may sound pretty high. However, acceptable accuracy levels are all a matter of scale and context. For example, when we look at that number in the context of airport security, it’s difficult to decide on an acceptable margin of error.
The Hartsfield–Jackson Atlanta International Airport receives over 260,000 passengers daily. At such a scale, most security officials would not accept even a 0.37% error rate. However, researchers are constantly exploring new approaches to face recognition. As computing power increases, face recognition accuracy should continue to improve as well.
What is Facial Recognition Used For?
Law enforcement and smartphone developers use facial recognition to improve security. However, those are not the only use cases. In fact, the applications of the technology are vast and varied. The following examples are just a few of the most interesting ways many businesses use face recognition today.
Augmented reality (AR)
Numerous popular smartphone applications rely on face recognition. Some famous examples would be the face filters on Instagram, Snapchat, and LINE. By locating the user’s facial landmarks, the AR app can accurately superimpose image filters over the user’s face in real time.
While not yet available in most countries, there are numerous merchants that now accept facial recognition payments in China. Moreover, on October 16th, 2019, SnapPay announced their launch of face recognition payment technology in North America.
Another application of this technology would be security gates or doors. Whether it be the entrance to your apartment complex, the front lobby of your office, or even train station ticket barriers, facial recognition technology could be used to allow or deny entry. While this technology isn’t that common yet in most countries, many businesses in China seem to be warming to the technology rather quickly.
The video above shows a quick demo of the face recognition security gates at Baidu, one of the largest tech companies in China.
As you can see, there are numerous beneficial applications of facial recognition. As the accuracy of the models increase, more and more countries will likely adopt face recognition technology into their infrastructure.
Hopefully, the information above helped strengthen your basic understanding of what facial recognition is and how it works. Due to the numerous use cases for businesses and our everyday lives, facial recognition continues to be a popular area of machine learning research. New studies in the field are being done year after year by some of the world’s largest tech companies. For more guides about facial recognition please see the related resources below.
For other machine learning guides and AI news, please subscribe to our newsletter.