News Release

Facial recognition ID with a twist: Smiles, winks and other facial movements for access

New algorithm makes ID verification more secure by tracking facial motion

Business Announcement

Brigham Young University

Photo illustration: Faces showing movements

image: A student makes different facial expressions and facial movements for a photo illustration. view more 

Credit: BYU Photo

Using your face to unlock your phone is a pretty genius security protocol. But like any advanced technology, hackers and thieves are always up to the challenge, whether that's unlocking your phone with your face while you sleep or using a photo from social media to do the same.

Like every other human biometric identification system before it (fingerprints, retina scans) there are still significant security flaws in some of the most advanced identity verification technology. Brigham Young University electrical and computer engineering professor D.J. Lee has decided there is a better and more secure way to use your face for restricted access.

It's called Concurrent Two-Factor Identity Verification (C2FIV) and it requires both one's facial identity and a specific facial motion to gain access. To set it up, a user faces a camera and records a short 1-2 second video of either a unique facial motion or a lip movement from reading a secret phrase. The video is then input into the device, which extracts facial features and the features of the facial motion, storing them for later ID verification.

"The biggest problem we are trying to solve is to make sure the identity verification process is intentional," said Lee, a professor of electrical and computer engineering at BYU. "If someone is unconscious, you can still use their finger to unlock a phone and get access to their device or you can scan their retina. You see this a lot in the movies -- think of Ethan Hunt in Mission Impossible even using masks to replicate someone else's face."

To get technical, C2FIV relies on an integrated neural network framework to learn facial features and actions concurrently. This framework models dynamic, sequential data like facial motions, where all the frames in a recording have to be considered (unlike a static photo with a figure that can be outlined).

Using this integrated neural network framework, the user's facial features and movements are embedded and stored on a server or in an embedded device and when they later attempt to gain access, the computer compares the newly generated embedding to the stored one. That user's ID is verified if the new and stored embeddings match at a certain threshold.

"We're pretty excited with the technology because it's pretty unique to add another level of protection that doesn't cause more trouble for the user," Lee said.

In their preliminary study, Lee and his Ph.D. student Zheng Sun recorded 8,000 video clips from 50 subjects making facial movements such as blinking, dropping their jaw, smiling or raising their eyebrows as well as many random facial motions to train the neural network. They then created a dataset of positive and negative pairs of facial motions and inputted higher scores for the positive pairs (those that matched). Currently, with the small dataset, the trained neural network verifies identities with over 90% accuracy. They are confident the accuracy can be much higher with a larger dataset and improvements on the network.

Lee, who has filed a patent on the tech already, said the idea is not to compete with Apple or have the application be all about smartphone access. In his opinion, C2FIV has broader application, including accessing restricted areas at a workplace, online banking, ATM use, safe deposit box access or even hotel room entry or keyless entry/access to your vehicle.

"We could build this very tiny device with a camera on it and this device could be deployed easily at so many different locations," Lee said. "How great would it be to know that even if you lost your car key, no one can steal your vehicle because they don't know your secret facial action?"

###


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.