Subscribe
By subscribing you agree to with our Privacy Policy and provide consent to receive updates from our company.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

The FBI are cracking down on people using deepfakes to apply for remote jobs

Companies looking to hire somebody with IT abilities might want to ask themselves whether the individual they are interviewing is real or not.

The FBI has uncovered numerous cases involving serious technology crimes, making it essential for companies to remain vigilant against impostors. These fraudsters use images, videos, and voice recordings to secure positions in technology, programming, database, and software firms. By exploiting stolen data, they can convincingly impersonate anyone they choose.

Many of these fake job applicants may have gained access to sensitive consumer or corporate information, including financial and proprietary data, suggesting that their intentions extend beyond simple fraud. They may aim to steal valuable information while deceiving the company. The extent of successful fake job applications remains unclear compared to those that have been detected and reported.

A more alarming possibility is that some of these impostors may have accepted job offers, received salaries, and later been arrested. Instances have been reported where voice spoofing techniques were used during online interviews, with the candidate’s lip movements not matching the audio. In some cases, applicants coughed or sneezed, yet the video spoofing software failed to detect these discrepancies.

In May, the FBI issued a warning to businesses about North Korean government operatives seeking remote IT and other technical jobs. These impostors often use fake documents and credentials to secure remote work through platforms like Upwork and Fiverr. The federal agency's report detailed how some fraudsters used multiple shell companies to conceal their identities, making detection even more challenging.

While deepfake technology has advanced significantly, some of the more rudimentary attempts still result in mismatches between fake voices and the speaker’s mouth movements. Detecting these fake videos can be difficult, especially if one is not actively searching for them. Creating a lifelike human in a video is more complex than it might appear, and unnoticed fakes can slip through the cracks.

Researchers at Carnegie Mellon University have developed artificial intelligence capable of recognising edited videos with an accuracy ranging from 30% to 97%. Those familiar with identifying certain visual inconsistencies, such as incorrect shadows or unnatural skin textures, can detect phoney videos more effectively.

September 5, 2022
How to Make the Most Out of Your Recruitment Consultant (Especially in Crypto and Web3)
June 27, 2025

Looking to get more out of your recruitment consultant in the crypto space? Learn how to build a successful partnership and unlock better results with Priority Crypto.

Read more
Priority Crypto x IdeaFoundry: Unlocking the Future of Work
June 25, 2025

Discover how Priority Crypto and IdeaFoundry are reshaping the future of work by connecting Web3 talent with AI-powered digital Agents.

Read more
Bitget and UNICEF Are Training 300,000 Girls in Blockchain - Here’s Why That Matters
June 17, 2025

Bitget and UNICEF partner to train 300,000 girls in blockchain across 8 countries. Find out why this matters for diversity in Web3 and the future of crypto talent.

Read more