SignGest: Sign Language Recognition Using Acoustic Signals on Smartphones

Sign language is a bridge for communication be-tween people with normal hearing and impaired hearing. How-ever, only a few people with normal hearing have an understanding of sign language. Thus, Sign Language Recognition (SLR) has attracted significant interest from both industry and academia. Prio...

Full description

Saved in:
Bibliographic Details
Published in2022 IEEE 20th International Conference on Embedded and Ubiquitous Computing (EUC) pp. 60 - 65
Main Authors Wang, Haoyu, Zhang, Junbao, Li, Yue, Wang, Lin
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.12.2022
Subjects
Online AccessGet full text
DOI10.1109/EUC57774.2022.00019

Cover

Abstract Sign language is a bridge for communication be-tween people with normal hearing and impaired hearing. How-ever, only a few people with normal hearing have an understanding of sign language. Thus, Sign Language Recognition (SLR) has attracted significant interest from both industry and academia. Prior traditional methods have certain limitations such as light condition and extra hardware. To address these issues, this study aims to recognize sign languages using acoustic signals on a smartphone. We design an SLR system called SignGest, which captures user's sign language gestures with built-in microphones. Afterwards, we build a Convolutional Neural Network(CNN) model to extract features of different gestures to distinguish them. Furthermore, we use a Deep Convolutional Generative Adversarial Network(DCGAN) to generate abundant training data that look no different from the real samples. We handle various challenges including effective gesture segmentation and training data collection. Finally, SignGest is implemented on a server and an Android smartphone with built-in microphones and speakers without any extra hardware or infrastructure equipment. When users perform the sign language gestures before the smartphone, SignGest can recognize each of them and print the result. Through extensive experiments, our results show that SignGest can achieve robust and satisfactory performance.
AbstractList Sign language is a bridge for communication be-tween people with normal hearing and impaired hearing. How-ever, only a few people with normal hearing have an understanding of sign language. Thus, Sign Language Recognition (SLR) has attracted significant interest from both industry and academia. Prior traditional methods have certain limitations such as light condition and extra hardware. To address these issues, this study aims to recognize sign languages using acoustic signals on a smartphone. We design an SLR system called SignGest, which captures user's sign language gestures with built-in microphones. Afterwards, we build a Convolutional Neural Network(CNN) model to extract features of different gestures to distinguish them. Furthermore, we use a Deep Convolutional Generative Adversarial Network(DCGAN) to generate abundant training data that look no different from the real samples. We handle various challenges including effective gesture segmentation and training data collection. Finally, SignGest is implemented on a server and an Android smartphone with built-in microphones and speakers without any extra hardware or infrastructure equipment. When users perform the sign language gestures before the smartphone, SignGest can recognize each of them and print the result. Through extensive experiments, our results show that SignGest can achieve robust and satisfactory performance.
Author Li, Yue
Wang, Lin
Wang, Haoyu
Zhang, Junbao
Author_xml – sequence: 1
  givenname: Haoyu
  surname: Wang
  fullname: Wang, Haoyu
  email: 2020107236@zut.edu.cn
  organization: School of Computer Science Zhongyuan University of Technology,Zhengzhou,China
– sequence: 2
  givenname: Junbao
  surname: Zhang
  fullname: Zhang, Junbao
  email: junbao.zhang@zut.edu.cn
  organization: School of Computer Science Zhongyuan University of Technology,Zhengzhou,China
– sequence: 3
  givenname: Yue
  surname: Li
  fullname: Li, Yue
  email: 2020107227@zut.edu.cn
  organization: School of Computer Science Zhongyuan University of Technology,Zhengzhou,China
– sequence: 4
  givenname: Lin
  surname: Wang
  fullname: Wang, Lin
  email: 2020007208@zut.edu.cn
  organization: School of Computer Science Zhongyuan University of Technology,Zhengzhou,China
BookMark eNotTs1OwzAYCxIcYOwJ4JAXaPnyJW36cZuqMSZVQtrW85QmaYkE6bR0B96e8nOyZVu279h1HKNn7EFALgTQ07qtC621yhEQcwAQdMWWpKmSBUgqZVHcsu0-DHHj0_TMfxhvTBwuZvB85-04xDCFMfI2hTjwlR0vaQr2N2g-Ep-d_ac5T6f3eTjds5t-Vv3yHxesfVkf6tesedts61WTBQQ1ZUYr1ISGKgIjrXJCqsqV0AtL5DX5yjqDIGRfdh2Cm592FjtnFYrCOS0X7PGvN3jvj6dzmC98HQVAVSIK-Q1Ig0og
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/EUC57774.2022.00019
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEL
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 9798350396355
EndPage 65
ExternalDocumentID 10086221
Genre orig-research
GroupedDBID 6IE
6IL
CBEJK
RIE
RIL
ID FETCH-LOGICAL-i204t-a742792a9890a3c4d1348d60f1c99e79e8cda2013f6bb20d039bc2bdc4215dd73
IEDL.DBID RIE
IngestDate Thu Jan 18 11:14:30 EST 2024
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i204t-a742792a9890a3c4d1348d60f1c99e79e8cda2013f6bb20d039bc2bdc4215dd73
PageCount 6
ParticipantIDs ieee_primary_10086221
PublicationCentury 2000
PublicationDate 2022-Dec.
PublicationDateYYYYMMDD 2022-12-01
PublicationDate_xml – month: 12
  year: 2022
  text: 2022-Dec.
PublicationDecade 2020
PublicationTitle 2022 IEEE 20th International Conference on Embedded and Ubiquitous Computing (EUC)
PublicationTitleAbbrev EUC
PublicationYear 2022
Publisher IEEE
Publisher_xml – name: IEEE
Score 1.8468941
Snippet Sign language is a bridge for communication be-tween people with normal hearing and impaired hearing. How-ever, only a few people with normal hearing have an...
SourceID ieee
SourceType Publisher
StartPage 60
SubjectTerms acoustic sensing
Acoustics
Assistive technologies
Auditory system
CNN
Feature extraction
GAN
Gesture recognition
sign language recognition
Training data
Ubiquitous computing
Title SignGest: Sign Language Recognition Using Acoustic Signals on Smartphones
URI https://ieeexplore.ieee.org/document/10086221
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8MwGA66kycVJ36Tg9duaZrmw5uMzSk6xDnYbTRfMsRWtLv4632TdiqC4C00oQ1J6PMmeZ7nRejcAWxL6mViAB4SxiVPpMxN4gsG4bP32pMgcL6b8PGM3czzeStWj1oY51wkn7leKMa7fFuZVTgq66cxAA-y8U0heSPWap2EUqL6w9kgFxDOwK6PRhvOYJ_zI2dKhIzRNpqsP9YwRZ57q1r3zMcvH8Z_92YHdb_Vefj-C3d20YYr99D1dPlUXsEv_gKHEr5tzyHxw5ohVJU48gPwpaliCq_YEFYfhprpCyyhQFN37100Gw0fB-OkTZOQLClhdVLA7lYoWiipSJEZZtOMScuJT41STignjS0A5zPPtabEkkxpQ7U1DODeWpHto04J7z9AWEinMoiITJCcMkEKyrzOCfeCEqGFOUTdMA6L18YJY7EegqM_nh-jrTAXDf3jBHXqt5U7BRCv9VmcvE-ISpym
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwGA0yD3pSceJvc_DaLU3TJvEmY7rpNsRtsNtofskQO9Hu4l_vl7RTEQRvoQltSELfl-S99yF0aQG2BXUi0gAPEctEFgmR6sjlDMJn55QjXuA8HGW9KbubpbNarB60MNbaQD6zLV8Md_lmqVf-qKwdhwDcy8Y3U8ZYWsm1ai-hmMh2d9pJOQQ0sO-jwYjTG-j8yJoSQONmB43Wn6u4Is-tVala-uOXE-O_-7OLmt_6PPzwhTx7aMMW-6g_XjwVt_CTv8K-hAf1SSR-XHOElgUODAF8rZchiVdoCOsPQ834BRaRJ6rb9yaa3nQnnV5UJ0qIFpSwMsphf8slzaWQJE80M3HChMmIi7WUlksrtMkB6ROXKUWJIYlUmiqjGQC-MTw5QI0C3n-IMBdWJhATaS86ZZzklDmVksxxSrji-gg1_TjMXysvjPl6CI7_eH6BtnqT4WA-6I_uT9C2n5eKDHKKGuXbyp4BpJfqPEzkJ-6Mn_M
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2022+IEEE+20th+International+Conference+on+Embedded+and+Ubiquitous+Computing+%28EUC%29&rft.atitle=SignGest%3A+Sign+Language+Recognition+Using+Acoustic+Signals+on+Smartphones&rft.au=Wang%2C+Haoyu&rft.au=Zhang%2C+Junbao&rft.au=Li%2C+Yue&rft.au=Wang%2C+Lin&rft.date=2022-12-01&rft.pub=IEEE&rft.spage=60&rft.epage=65&rft_id=info:doi/10.1109%2FEUC57774.2022.00019&rft.externalDocID=10086221