acm-header
Sign In

Communications of the ACM

ACM Careers

No Lie! Researcher Developing Online Polygraph


View as: Print Mobile App Share:
polygraph lines with face in background, illustration

Can you spot a liar?

It's tricky enough in face-to-face conversations that offer facial expressions, gestures, and tone of voice because those physical cues add context. Spotting a liar gets even tougher in blind computer conversations.

Florida State University researcher Shuyuan Ho wants to shed those blinders by creating a revolutionary online polygraph.

"The future of my research is an online polygraph that could be used many different ways," says Ho, an associate professor in the College of Communication and Information. "You could use it for online dating, Facebook, Twitter—the applications are endless. I think the future is unlimited for an online polygraph system."

Ho envisions a future where technology can identify liars and truthtellers based on the words they write in electronic messages. Her latest research dives into the murky depths of Internet deception where trolling, identity theft, and phishing for credit card numbers snag an increasing number of online users.

The research is described in "Context In a Bottle: Language-Action Cues in Spontaneous Computer-Mediated Deception," published in the journal Computers in Human Behavior, which details the findings of an online game Ho created to measure truthful and deceptive communications between two people.

Ho parsed the words in those conversations, hoping to extract context from millions of bits of data in many messages—described as language-action cues—just as people get context from seeing physical cues that indicate whether someone is telling the truth or lying.

The results were startling.

The experiments revealed a person could spot lies in messages about 50 percent of the time, while a machine-learning approach could identify deception with an accuracy rate ranging from 85 to 100 percent.

The Computers in Human Behavior article is co-authored by Jeffrey T.Hancock of Stanford University.

Ho is excited about the potential for this research.

"I want to get the world's attention on this research so we can hopefully make it into a commercial product that could be attached to all kinds of online social forums," Ho says. "This basic research offers great potential to develop an online polygraph system that helps protect our online communication."

Ho supervises the iSensor Lab on FSU's campus where researchers conduct experiments to better understand deception in online communications.

To facilitate those conversations, she created an online game designed to identify language cues that unmask deceivers and truthtellers.

The game randomly assigned players to play the roles of "The Saint" and "The Sinner." As sinners and saints interacted via computers, researchers in the iSensor Lab captured those conversations and used machine-learning technology to scrutinize patterns of words and writing.

Some fascinating language tendencies emerged from that analysis. The lying sinners were found to be less expressive, but they used more decorative words per message. They displayed more negative emotions and appeared more anxious when they communicated with truthtellers.

Deceivers also took less time to respond and used more words of insight, such as "think" and "know," and they tended to use more words of certainty, including "always" or "never."

Conversely, truthtellers used more words of speculation, such as "perhaps" and "guess," and they took longer to respond to inquiries. These saints provided more reasoned ideas by using words of causation—"because"—and they expressed more reflective thinking with words like "should" and "could."

"Truthtellers tend to say 'no' a lot. Why?" Ho asks. "They like to provide emphasis when they explain their reasons. If you ask them, 'Is this true?' They tend to say 'no' because there is another true reason."

Researchers also calculated time lags between every sentence, and even parts of a sentence, by placing time stamps on the words. That precise breakdown clearly showed how much a person paused during interactions—another language-action cue.

Those pauses might have been so slight that they would not necessarily be noticeable to a person, but machine-learning technology could spot it.

Ho says she hopes her research will eventually give people better protection when they're online.

"I think we all have good common sense about the people we meet face to face, but how much common sense do we have with the strangers we encounter online where you can meet a lot of people very fast," Ho says. "This research is so important because it can provide another reference point offering more protection. All of society can benefit."

This research was funded by the U.S. National Science Foundation, Florida Center for Cybersecurity Collaborative Seed Grant, and Florida State University Council for Research and Creativity Planning Grant.


 

No entries found

Sign In for Full Access
» Forgot Password? » Create an ACM Web Account