亚洲AV

Students program computers to interpret sign language

Body

With a twist or shake of your wrist, your smartphone can interpret motion to take a picture, turn on a light, and more.聽Last year, 亚洲AV computer science professors and were brainstorming how similar technology could help society in even greater ways. Their idea? To automatically translate sign language into text or speech.

鈥淭here are some products that can do gesture recognition, but they鈥檙e very preliminary. And it鈥檚 very different from ASL [American Sign Language], which is not just a few gestures鈥攊t鈥檚 thousands of words,鈥 said Pathak, principal investigator on the Summer Team Impact Project funded by Mason鈥檚 (OSCAR).

This summer, nine Mason undergraduates joined in the research that could help make the technology a reality.

A student does sign language in front of a computer camera while two other students on the other side of the table look at the data on their laptops.
Frederick Olson (from left), Sai Gurrapu and Dom Huh are part of a summer research project on automatic multimodal sign language recognition. Photo by Lathan Goumas/Office of Communications and Marketing.

鈥淭he goal would be to deliver a readable message to a device so that it鈥檚 bridging the gap between ASL users and non-users,鈥 said senior Riley Wilkerson, 鈥渁n easier, more effective, and more personal way of communicating.鈥

Close up of Riley Wilkerson's hands as she signs in front of a radar sensor.
Riley Wilkerson signs for a radar sensor as part of a summer research project on automatic multimodal sign language recognition. Photo by Lathan Goumas/Office of Communications and Marketing.

Three teams of students are experimenting with different sensors: a wireless radar, a camera, and an inertial measurement unit (a wearable motion sensor used in smartphones and Fitbits). Each sensor offers certain opportunities, but also challenges including privacy and ease of use, said Pathak, who is guiding the students on the project along with Mason computer science professor and Mason鈥檚 Helen A. Kellar Institute for Human disabilities director , and graduate student Panneer Selvam Santhalingham.

On each team, a student familiar with ASL signs in front of a sensor that collects data about the motion or the environment. and students refine the data to find patterns and write machine learning algorithms鈥攃ode that allows them to interpret the computer鈥檚 recognition of the signs.

So far, the undergraduates have 鈥渢aught鈥 their machines to recognize about 20 signs with accuracy rates ranging from 70-97 percent. The fluctuations in accuracy are due to the machine learning process, said senior computer science major Yuanqi Du. 聽

Diverse data helps the computer recognize the signs with increased accuracy, Du said. In initial trials with one student, accuracy rates were higher. When a new ASL user was introduced, the accuracy diminished, Du said. Once the new ASL user鈥檚 data was included in the algorithms, accuracy rates rose again.

Cell phones are strapped on a student's wrists as signs.
Ariana Havens wears cell phones as she signs as part of a summer research project on automatic multimodal sign language recognition. Photo by Lathan Goumas.

As the multi-year project continues, Pathak said the team plans to increase the number of signs the computer can recognize using data from many diverse users. They will also scale it to interpret full sentences and pick up other gestures used in ASL such as body tilts and micro expressions like raising an eyebrow, he said.

鈥淏eing able to communicate instantly would hopefully remove issues [the ASL community experiences],鈥 said Frederick Olson, a senior IT major who said both his parents are deaf. That includes being able to ask a question at a store, socializing, 聽communicating with doctors easily during appointments, or being able to land better job opportunities. The technology could be life-changing, he said.

It could also be applied beyond the deaf community, the students said, helping people with autism or developmental and learning disabilities for whom communicating using spoken words is challenging, Wilkerson said.

鈥淚t could be applicable to other industries and disciplines in the future [that will work with similar technology], too,鈥 said junior computer science major Sai Gurrapu.

And, the project pushes student learning to the next level, Pathak said.

鈥淭hey鈥檙e not given a fixed task here鈥攖hey鈥檙e given a problem and they have to find a solution,鈥 Pathak said.

鈥淭his project is one of a variety of opportunities [Mason] has presented to me that goes beyond just taking 15 credits each semester,鈥 Wilkerson said. 鈥淵ou can only learn so much in a classroom鈥攜ou have to apply it.鈥澛

One student does sign language in front of a radar sensor and two other students on the other side of the table view the data on their computers.
Seniors Yuanqi (from left), Nguyen Dang and Riley Wilkerson are part of a summer research project on automatic multimodal sign language recognition. Photo by Lathan Goumas/Office of Communications and Marketing.