Home > Research > Publications & Outputs > AudioGest

Links

Text available via DOI:

View graph of relations

AudioGest: Enabling fine-grained hand gesture detection by decoding echo signal

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
  • Wenjie Ruan
  • Quan Z. Sheng
  • Lei Yang
  • Tao Gu
  • Peipei Xu
  • Longfei Shangguan
Close
Publication date12/09/2016
Host publicationUbiComp 2016 - Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing
Place of PublicationNew York
PublisherAssociation for Computing Machinery, Inc
Pages474-485
Number of pages12
ISBN (electronic)9781450344616
<mark>Original language</mark>English
Event2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp 2016 - Heidelberg, Germany
Duration: 12/09/201616/09/2016

Conference

Conference2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp 2016
Country/TerritoryGermany
CityHeidelberg
Period12/09/1616/09/16

Conference

Conference2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, UbiComp 2016
Country/TerritoryGermany
CityHeidelberg
Period12/09/1616/09/16

Abstract

Hand gesture is becoming an increasingly popular means of interacting with consumer electronic devices, such as mobile phones, tablets and laptops. In this paper, we present AudioGest, a device-free gesture recognition system that can accurately sense the hand in-air movement around user's devices. Compared to the state-of-the-art, AudioGest is superior in using only one pair of built-in speaker and microphone, without any extra hardware or infrastructure support and with no training, to achieve fine-grained hand detection. Our system is able to accurately recognize various hand gestures, estimate the hand in-air time, as well as average moving speed and waving range. We achieve this by transforming the device into an active sonar system that transmits inaudible audio signal and decodes the echoes of hand at its microphone. We address various challenges including cleaning the noisy reflected sound signal, interpreting the echo spectrogram into hand gestures, decoding the Doppler frequency shifts into the hand waving speed and range, as well as being robust to the environmental motion and signal drifting. We implement the proof-of-concept prototype in three different electronic devices and extensively evaluate the system in four real-world scenarios using 3,900 hand gestures that collected by five users for more than two weeks. Our results show that AudioGest can detect six hand gestures with an accuracy up to 96%, and by distinguishing the gesture attributions, it can provide up to 162 control commands for various applications.