Wearable-tech glove translates sign language into speech in real time

A digital rendering of the system that helps convert sign language into speech. @ Jun Chen Lab/UCLA

UCLA bioengineers have designed a glove-like device that can translate American Sign Language into English speech in real time though a smartphone app. Their research is published in the journal Nature Electronics.

"Our hope is that this opens up an easy way for people who use sign language to communicate directly with non-signers without needing someone else to translate for them," said Jun Chen, an assistant professor of bioengineering at the UCLA Samueli School of Engineering and the principal investigator on the research. "In addition, we hope it can help more people learn sign language themselves."

The system includes a pair of gloves with thin, stretchable sensors that run the length of each of the five fingers. These sensors, made from electrically conducting yarns, pick up hand motions and finger placements that stand for individual letters, numbers, words and phrases.

The device then turns the finger movements into electrical signals, which are sent to a dollar-coin-sized circuit board worn on the wrist. The board transmits those signals wirelessly to a smartphone that translates them into spoken words at the rate of about a one word per second.

The researchers also added adhesive sensors to testers' faces -- in between their eyebrows and on one side of their mouths -- to capture facial expressions that are a part of American Sign Language.

Previous wearable systems that offered translation from American Sign Language were limited by bulky and heavy device designs or were uncomfortable to wear, Chen said.

The device developed by the UCLA team is made from lightweight and inexpensive but long-lasting, stretchable polymers. The electronic sensors are also very flexible and inexpensive.

In testing the device, the researchers worked with four people who are deaf and use American Sign Language. The wearers repeated each hand gesture 15 times. A custom machine-learning algorithm turned these gestures into the letters, numbers and words they represented. The system recognized 660 signs, including each letter of the alphabet and numbers 0 through 9.

Sign-to-speech translation using machine-learning-assisted stretchable sensor arrays

Zhihao Zhou, Kyle Chen, Xiaoshi Li, Songlin Zhang, Yufen Wu, Yihao Zhou, Keyu Meng, Chenchen Sun, Qiang He, Wenjing Fan, Endong Fan, Zhiwei Lin, Xulong Tan, Weili Deng, Jin Yang & Jun Chen

Nature Electronics (2020)

DOI: 10.1038/s41928-020-0428-6

Contact information:

Jun Chen

UCLA assistant professor of bioengineering


Phone: (310) 794-5550

Wearable Bioelectronics Research Group

University of California, Los Angeles (UCLA)




FEFU scientists helped design a new type of ceramics for laser applications

Reviewing multiferroics for future, low-energy data storage

Researchers develop a plant-based thermotherapy patch

A new all-2-D light-emitting field-effect transistor

Altering the properties of 2-D materials at the nanometer scale

Study examines spontaneous symmetry breaking in twisted double bilayer graphene

Lineage tracing of direct astrocyte-to-neuron conversion for brain repair