Your browser doesn't support javascript.
loading
End-to-End Ultrasonic Hand Gesture Recognition.
Fertl, Elfi; Nguyen, Do Dinh Tan; Krueger, Martin; Stettinger, Georg; Padial-Allué, Rubén; Castillo, Encarnación; Cuéllar, Manuel P.
Afiliación
  • Fertl E; Infineon Technologies AG, 85579 Neubiberg, Germany.
  • Nguyen DDT; Department of Electronics and Computer Technology, University of Granada, 18071 Granada, Spain.
  • Krueger M; Infineon Technologies AG, 85579 Neubiberg, Germany.
  • Stettinger G; Infineon Technologies AG, 85579 Neubiberg, Germany.
  • Padial-Allué R; Infineon Technologies AG, 85579 Neubiberg, Germany.
  • Castillo E; Department of Electronics and Computer Technology, University of Granada, 18071 Granada, Spain.
  • Cuéllar MP; Department of Electronics and Computer Technology, University of Granada, 18071 Granada, Spain.
Sensors (Basel) ; 24(9)2024 Apr 25.
Article en En | MEDLINE | ID: mdl-38732843
ABSTRACT
As the number of electronic gadgets in our daily lives is increasing and most of them require some kind of human interaction, this demands innovative, convenient input methods. There are limitations to state-of-the-art (SotA) ultrasound-based hand gesture recognition (HGR) systems in terms of robustness and accuracy. This research presents a novel machine learning (ML)-based end-to-end solution for hand gesture recognition with low-cost micro-electromechanical (MEMS) system ultrasonic transducers. In contrast to prior methods, our ML model processes the raw echo samples directly instead of using pre-processed data. Consequently, the processing flow presented in this work leaves it to the ML model to extract the important information from the echo data. The success of this approach is demonstrated as follows. Four MEMS ultrasonic transducers are placed in three different geometrical arrangements. For each arrangement, different types of ML models are optimized and benchmarked on datasets acquired with the presented custom hardware (HW) convolutional neural networks (CNNs), gated recurrent units (GRUs), long short-term memory (LSTM), vision transformer (ViT), and cross-attention multi-scale vision transformer (CrossViT). The three last-mentioned ML models reached more than 88% accuracy. The most important innovation described in this research paper is that we were able to demonstrate that little pre-processing is necessary to obtain high accuracy in ultrasonic HGR for several arrangements of cost-effective and low-power MEMS ultrasonic transducer arrays. Even the computationally intensive Fourier transform can be omitted. The presented approach is further compared to HGR systems using other sensor types such as vision, WiFi, radar, and state-of-the-art ultrasound-based HGR systems. Direct processing of the sensor signals by a compact model makes ultrasonic hand gesture recognition a true low-cost and power-efficient input method.
Asunto(s)
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Redes Neurales de la Computación / Aprendizaje Automático / Gestos / Mano Límite: Humans Idioma: En Revista: Sensors (Basel) Año: 2024 Tipo del documento: Article País de afiliación: Alemania Pais de publicación: Suiza

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Asunto principal: Redes Neurales de la Computación / Aprendizaje Automático / Gestos / Mano Límite: Humans Idioma: En Revista: Sensors (Basel) Año: 2024 Tipo del documento: Article País de afiliación: Alemania Pais de publicación: Suiza