acm-header
Sign In

Communications of the ACM

ACM News

Cell Phone Technology Allows Deaf People To Communicate Via Sign Language


View as: Print Mobile App Share:
Mobile ASL

The Mobile ASL project is developing compression technology to support the transmission of real-time video of American Sign Language over existing wireless networks.

Credit: University of Washington; Cornell University

For those who are deaf or hard of hearing, cell phone use has largely been limited to text messaging. But technology is catching up: Cornell researchers and colleagues have created a real-time video compression scheme for cell phones that allows deaf people to communicate in sign language — the same way hearing people use phones to talk.

"We completely take cell phones for granted," says Sheila Hemami, Cornell professor of electrical and computer engineering, who leads the research with Eve Riskin and Richard Ladner of the University of Washington. "Deaf people can text, but if texting were so fabulous, cell phones would never develop.There is a reason that we like to use our cell phones. People prefer to talk."

The technology, Hemami says, is about much more than convenience. It allows deaf people "untethered communication in their native language" — exactly the same connectivity available to hearing people, she said. The real-time compression will transmit video within the existing wireless network in real time while maintaining video quality that allows for intelligible sign language.

Since the project, Mobile ASL (American Sign Language), started four years ago, the researchers have published several academic papers on their technology and given talks around the world. The first phone prototypes were created last year and are now in the hands of about 25 deaf people in the Seattle area.

Standard videoconferencing is used widely in academia and industry, for example, in distance-learning courses. But the Mobile ASL team designed their video compression software specifically with ASL users in mind, with the goal of sending clear, understandable video over existing limited bandwidth networks. They also faced such constraints as phones' battery life and their ability to process real-time video at enough frames per second. They solved the battery life problem by writing software smart enough to vary the frames per second based on whether the user is signing or watching the other person sign.

Because ASL requires efficient motion capture, the researchers had to make video compression software that could deliver video at about 10 frames per second. They also had to work within the standard wireless 2G network, which only allows transmission of video at about 15-to-20 kilobits per second.

This is a relatively small amount of information when compared with a YouTube video, which travels at about 600 kilobits per second. For further comparison, high-definition digital television images come in at 6-to-10 megabits per second.

Researching how ASL developed gave the team clues on how people use it, says Frank Ciaramello, a graduate student working on the project. They learned that deaf people often use only one hand to sign, depending on the situation, and that they're very good at adapting as needed.

And they found that when two people are talking to each other, they spend almost the entire time focused on the other person's face.

"The facial expressions are really important in ASL, because they add a lot of information," Ciaramello says. They concluded that their cell phone video would have to be clearest in the face and hands, while they could spare some detail in the torso and in the background. Studies with deaf people who rated different videos on an intelligibility scale helped the researchers hone in on the best areas to focus in their video.

The researchers are now perfecting their intelligibility metrics while also looking for ways to bring down the cost of integrating the software into the phones. Making the phones as user friendly as possible is a key goal of the project, Hemami says.

"We don't want people to use the technology and say, 'This is annoying,' " Hemami says. "We want it to be really technology transparent. We want them to call their mother and have a nice conversation."

This research is funded by the U.S. National Science Foundation.

View a video of the Mobile ASL video compression project.
 


Comments


melissa beard mejia

MY NAME IS MELISSA AND AM DEAF HALF. I WANT CELLPHONE NOW AND EXCITE TO BUY IT NOW I CANT WAIT TO HEAR ABOUT IT

THANK MUCH


dalmiris toribio

HELL0 EVERYONE,

MY NAME IS DALMIRIS TORIBIO AND AM FULL DEAF. I NOFICE THAT NEW ASL BY CELL PHONE IS COOL THEN I NEED TO PHONE NOW AND I WANT TO BUY WHEN IT COMING SOON!


luis A veras

HELLO EVERY ONE
MY NAME IS LUIS AND IM FULL DEAF. I WANT PHONE NOW AND BUY FOR PHONE THEN HOPE SO HEARD ABOUT IT


Anonymous

hi my name is Andrea im 16 and i use my cellphone all the time during work school and to chat with my friends over things... i think deaf and hard of hearing should use their phones more than hearing people because we cant talk like they can


Anonymous

Hi there.I am 16 and i think that this is a really good idea!! just using txting isn't enough for most people to get enough information! I love the idea of video talking with a cell phone!


Displaying all 5 comments

Sign In for Full Access
» Forgot Password? » Create an ACM Web Account