Top

Can you deceive an AI lie detector? Not for long

AI lie detector: Historically, polygraphs have been a popular tool for legal authorities seeking to determine whether someone is lying. This test measures physical responses—such as breathing rate, sweating, and blood pressure—to assess truthfulness. While polygraphs have served as lie detectors since the dawn of modern industry, their reliability has long been questioned.

Critics argue that the accuracy of polygraph tests may hinge more on anxiety than on actual dishonesty. The underlying assumption of the polygraph is that humans exhibit certain physiological changes when they lie. However, this premise has faced significant scrutiny from psychologists and legal experts alike.

In light of these concerns, the United States Congress passed a law in 1988 banning the use of polygraphs in hiring processes, and in 1998, the Supreme Court ruled that polygraph results cannot be accepted as evidence in court. Despite these legal setbacks, the polygraph remains one of the most popular deception detection tools worldwide, generating around US$2 billion annually. Its widespread use is indicative of society’s ongoing fascination with the concept of truth and deception.

AI Deception Detectors at the Forefront

As technology advances, several developers are pioneering AI lie detection technologies that promise to enhance accuracy and efficiency. One leading tool is Coyote, created by Vancouver-based Prodigy Intelligence. This deception detection tool has been trained on a diverse set of transcripts labelled as true or false. According to the company, Coyote can assess whether a textual statement is a lie, boasting an accuracy rate of around 80%.

Coyote is user-friendly, requiring no complicated tutorials. Users simply input text or media files—such as audio or video—into the tool, generating a comprehensive report featuring multiple parts. Each report provides a lie detection rating, such as “Truth likely” or “Deception likely”, along with the AI’s confidence level. This ease of use positions Coyote as a promising tool for individual users and businesses.

Another noteworthy contender is Deceptio.AI, developed by a team in Florida. This software can analyse up to 5,000 words and deliver a detailed scoring report on truth probability within seconds. If the truthfulness percentage falls below 85%, the tool is likely to classify it as “deceptive”. Deceptio.AI relies on proprietary data that does not include user’s personal information, which may alleviate some privacy concerns associated with other tools. Co-founder Mark Carson has stated that the company has gathered extensive “human behavioural data” to map how people lie, enhancing the software’s predictive capabilities.

AI lie detector

A recent entrant making headlines is Q, developed by CyberQ. The American company prides itself on being backed by deception detection experts, including former CIA officer Phil Houston, who has 25 years of experience as an investigator and polygrapher. Like other lie detectors, Q is trained on a variety of transcript materials, including police interviews with high-profile criminals. Upon receiving a prompt and text input, Q categorises the analysis into three options: no deceptive behaviour found, deceptive behaviour found, and follow-up required. Though not yet available for public use, the company envisions applications for environments such as online video calls, potentially transforming how we communicate digitally.

Can you deceive an AI lie detector? Not for long
Can you deceive an AI lie detector? Not for long

Lying in the Digital Age

While AI detectors may be useful in combating misinformation or safeguarding personal data from cyber threats, they have flaws. Despite their reliance on advanced technology, the effectiveness of AI lie detectors still depends on human input during the training process, which may introduce perceptual biases. Given that lying is a complex human behaviour influenced by personality and context, there remains considerable room for improvement in AI analysis.

Moreover, these advancements prompt significant ethical considerations. The development of AI-driven lie detectors raises fundamental questions about the nature of lying itself. Is lying inherently bad, or can it sometimes serve a greater purpose, such as preserving feelings or maintaining social harmony? The integration of these systems into everyday communication could create a troubling landscape of public distrust. If people begin to rely on AI assessments in daily interactions, even benign lies might be scrutinised, leading to a dystopian future where suspicion prevails over understanding.

The implications extend beyond individual interactions; businesses, law enforcement, and even relationships could be impacted by a society that places undue faith in AI assessments. What happens when an innocent comment is flagged as deceptive? Such scenarios could undermine trust among colleagues, friends, and loved ones, ultimately affecting social cohesion.

Sunny Um is a Seoul-based journalist working with 4i Magazine. She writes and talks about policies, business updates, and social issues around the Korean tech industry. She is best known for in-depth explanations of local issues for readers who need a better understanding of the Korean context. Sunny’s works appeared in prominent Korean news outlets, such as the Korea Times and Wired Korea. She currently makes regular writing contributions to newsrooms worldwide, such as Maritime Fairtrade, a non-profit media organization based in Singapore. She also works as a content strategist at 1021 Creative. A person who holds a Master’s degree in Political Economy from King’s College London, she loves to follow up on news of Korean politics and economy when she’s not writing.