这是用户在 2025-4-10 16:33 为 https://www.mdpi.com/1424-8220/23/4/1790 保存的双语快照页面,由 沉浸式翻译 提供双语支持。了解如何保存?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article  开放获取文章

UltrasonicGS: A Highly Robust Gesture and Sign Language Recognition Method Based on Ultrasonic Signals
基于超声波信号的超声 GS:一种高度鲁棒的姿态和手语识别方法

by 1,   作者:王悦娇 1,2,*,
1 ,郝占军
1,2,
1,2,* ,邓晓超
1 and    1,2 ,张振毅 1   1 和 李梦桥
1
College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China
西北师范大学计算机科学与工程学院,兰州 730070,中国
2
Gansu Province Internet of Things Engineering Research Center, Lanzhou 730070, China
甘肃省物联网工程研究中心,兰州 730070,中国
*
Author to whom correspondence should be addressed.
请将回信地址寄给作者。
Sensors 2023, 23(4), 1790; https://doi.org/10.3390/s23041790
Submission received: 5 January 2023 / Revised: 1 February 2023 / Accepted: 2 February 2023 / Published: 5 February 2023
投稿已收到:2023 年 1 月 5 日 / 修订:2023 年 2 月 1 日 / 录用:2023 年 2 月 2 日 / 出版:2023 年 2 月 5 日
(This article belongs to the Topic Internet of Things: Latest Advances)
(本文属于主题:物联网:最新进展)

Abstract  摘要

With the global spread of the novel coronavirus, avoiding human-to-human contact has become an effective way to cut off the spread of the virus. Therefore, contactless gesture recognition becomes an effective means to reduce the risk of contact infection in outbreak prevention and control. However, the recognition of everyday behavioral sign language of a certain population of deaf people presents a challenge to sensing technology. Ubiquitous acoustics offer new ideas on how to perceive everyday behavior. The advantages of a low sampling rate, slow propagation speed, and easy access to the equipment have led to the widespread use of acoustic signal-based gesture recognition sensing technology. Therefore, this paper proposed a contactless gesture and sign language behavior sensing method based on ultrasonic signals—UltrasonicGS. The method used Generative Adversarial Network (GAN)-based data augmentation techniques to expand the dataset without human intervention and improve the performance of the behavior recognition model. In addition, to solve the problem of inconsistent length and difficult alignment of input and output sequences of continuous gestures and sign language gestures, we added the Connectionist Temporal Classification (CTC) algorithm after the CRNN network. Additionally, the architecture can achieve better recognition of sign language behaviors of certain people, filling the gap of acoustic-based perception of Chinese sign language. We have conducted extensive experiments and evaluations of UltrasonicGS in a variety of real scenarios. The experimental results showed that UltrasonicGS achieved a combined recognition rate of 98.8% for 15 single gestures and an average correct recognition rate of 92.4% and 86.3% for six sets of continuous gestures and sign language gestures, respectively. As a result, our proposed method provided a low-cost and highly robust solution for avoiding human-to-human contact.
随着新型冠状病毒在全球范围内的传播,避免人与人之间的接触已成为切断病毒传播的有效方式。因此,无接触式手势识别成为降低疫情预防和控制中接触感染风险的有效手段。然而,识别一定人群的日常行为手语对传感技术来说是一个挑战。无处不在的声学为感知日常行为提供了新的思路。低采样率、慢传播速度和易于获取设备等优势使得基于声学信号的姿态识别传感技术得到广泛应用。因此,本文提出了一种基于超声波信号的无接触式手势和手语行为传感方法——UltrasonicGS。该方法使用基于生成对抗网络(GAN)的数据增强技术,无需人工干预即可扩展数据集并提高行为识别模型的性能。 此外,为了解决连续手势和手语手势输入输出序列长度不一致和难以对齐的问题,我们在 CRNN 网络之后添加了连接主义时序分类(CTC)算法。此外,该架构能够更好地识别某些人的手语行为,填补了基于声学感知的中文手语感知的空白。我们已经在各种真实场景中对 UltrasonicGS 进行了广泛的实验和评估。实验结果表明,UltrasonicGS 在 15 个单个手势上实现了 98.8%的综合识别率,对于六组连续手势和手语手势,分别实现了 92.4%和 86.3%的平均正确识别率。因此,我们提出的方法提供了一种低成本且高度鲁棒的避免人与人接触的解决方案。
Keywords:
ultrasonic sensing; gesture recognition; sign language recognition; GAN; CTC
关键词:超声波传感;手势识别;手语识别;生成对抗网络(GAN);CTC

1. Introduction  1. 引言

The world has suffered from a sudden outbreak of a new coronavirus that has had a widespread impact on people’s lives. In particular, in recent times, a number of countries and regions around the world have seen a recurrence of the situation. The situation of epidemic prevention and control is still serious. In the face of this massive epidemic, the World Health Organization (WHO) states in its guidance article that avoiding human-to-human contact can effectively cut off the spread of the virus [1]. Therefore, contactless gesture recognition becomes an effective means to reduce the risk of contact infection in outbreak prevention and control. However, especially in the face of daily behavior recognition for certain populations, such as the deaf, the labor cost of hiring a sign language teacher is high. Therefore, how to correctly and efficiently recognize sign language gestures and perform human–computer interaction has become a problem that needs to be solved.
世界突然爆发了一种新型冠状病毒,对人们的生活产生了广泛的影响。特别是,在最近一段时间,世界上许多国家和地区都出现了疫情的反复。疫情防控形势依然严峻。面对这场大规模的疫情,世界卫生组织(WHO)在其指导文章中指出,避免人与人之间的接触可以有效切断病毒的传播 [1]。因此,无接触式手势识别成为降低疫情预防控制中接触感染风险的有效手段。然而,特别是在面对某些人群的日常行为识别时,例如聋人,雇佣手语教师的劳动成本很高。因此,如何正确、高效地识别手语手势并实现人机交互,已成为需要解决的问题。
Past research work on gesture recognition was divided into three main categories: sensor-based [2], vision-based [3], and Wi-Fi-based [4,5]. In sensor-based systems, limb motion features are captured by body-worn sensors. In vision-based systems, limb motion features are captured by optical cameras. In Wi-Fi-based systems, extracting channel state information (CSI) can recognize limb motion. By collecting human behavior information, different data processing processes, and classification learning, all of the above methods can identify people’s behaviors. However, there are certain limitations to these techniques. Vision-based sensing technology is highly influenced by light and has poor privacy and high energy consumption requirements for long-term detection. Sensor-based sensing technology causes a lot of inconvenience to users because they need to wear external devices for a long time. For Wi-Fi-based sensing technology, recognition accuracy is affected because Wi-Fi signals are susceptible to interference from electromagnetic waves.
过去关于手势识别的研究工作分为三大类:基于传感器的[2]、基于视觉的[3]和基于 Wi-Fi 的[4, 5]。在基于传感器的系统中,通过穿戴式传感器捕捉肢体运动特征。在基于视觉的系统中,通过光学摄像头捕捉肢体运动特征。在基于 Wi-Fi 的系统中,通过提取信道状态信息(CSI)来识别肢体运动。通过收集人类行为信息、不同的数据处理过程和分类学习,上述所有方法都可以识别人们的行为。然而,这些技术存在一定的局限性。基于视觉的传感技术受光线影响很大,对长期检测有较差的隐私保护和较高的能耗要求。基于传感器的传感技术会给用户带来很多不便,因为他们需要长时间佩戴外部设备。对于基于 Wi-Fi 的传感技术,由于 Wi-Fi 信号容易受到电磁波的干扰,识别精度受到影响。
To compensate for the limitations of traditional techniques, the use of acoustic waves for human activity perception is gradually gaining attention. Due to the advantages of slow propagation speed, low sampling rate, and easy access to equipment, in recent years, relevant research based on ultrasonic signals has also made great progress in smart homes [6], location tracking [7], gesture recognition [8], and facial recognition [9]. Research work in gesture recognition includes: Gao et al. [10] captured gestures using lightweight MobileNet by using dual speakers and microphones in smartphones. LLAP [11] obtained the accurate motion direction and distance by measuring the phase change of the received signal to realize two-dimensional gesture tracking. Strata [12] achieved more accurate recognition of gestures by estimating the Channel Impulse Response (CIR) of the reflected signal. In this paper, we focus on human gesture recognition, especially extending to sign language recognition for certain groups, such as deaf people [13], and providing higher perceptual accuracy.
为了弥补传统技术的局限性,使用声波进行人类活动感知逐渐受到关注。由于传播速度慢、采样率低、设备易于获取等优势,近年来,基于超声波信号的相关研究在智能家居[6]、位置跟踪[7]、手势识别[8]和面部识别[9]等方面也取得了重大进展。手势识别的研究工作包括:高翔等[10]利用智能手机中的双扬声器和麦克风,通过轻量级 MobileNet 捕获手势。LLAP[11]通过测量接收信号的相位变化,实现了精确的运动方向和距离测量,以实现二维手势跟踪。Strata[12]通过估计反射信号的信道脉冲响应(CIR),实现了更精确的手势识别。在本文中,我们专注于人类手势识别,特别是扩展到某些群体(如聋人[13])的手语识别,并提供更高的感知精度。
Due to the complexity of gesture movements, implementing acoustic-based fine-grained, and highly robust gesture and sign-language-recognition methods have two challenges. The first challenge is insufficient training data. The approach in this paper involves three tasks: single gesture recognition, continuous gesture recognition, and sign language gesture recognition. It takes time and effort to collect sufficient data for each task. Past work either did not use data augmentation methods or used traditional data augmentation methods based on geometric transformations and image manipulation. Although it can alleviate the problem of neural network overfitting and improve the generalization ability to a certain extent, the method used lacks flexibility and covers more limited situations. The second challenge is to solve the problem of inconsistent length and difficult alignment of input and output sequences of continuous and sign language gestures. Because most of the previous perception-based research work [14] can only recognize a single gesture, or several consecutive individual actions, especially since there is no research work using acoustic perception for Chinese sign language recognition. Continuous gesture and sign language recognition is an indeterminate length sequence prediction problem. Traditional sequence prediction networks usually only produce fixed-length outputs and can not determine the length of the prediction sequence adaptively.
由于手势动作的复杂性,实现基于声学的高精度和高度鲁棒的手势和手语识别方法有两个挑战。第一个挑战是训练数据不足。本文提出的方法涉及三个任务:单个手势识别、连续手势识别和手语手势识别。为每个任务收集足够的数据需要花费时间和精力。过去的工作要么没有使用数据增强方法,要么使用了基于几何变换和图像操作的传统数据增强方法。尽管这可以在一定程度上缓解神经网络过拟合的问题并提高泛化能力,但使用的方法缺乏灵活性,覆盖的情况更为有限。第二个挑战是解决连续和手语手势输入和输出序列长度不一致和难以对齐的问题。 因为大多数以前基于感知的研究工作[14]只能识别单个手势,或者几个连续的单独动作,尤其是在没有使用声学感知进行中国手语识别的研究工作。连续手势和手语识别是一个不确定长度的序列预测问题。传统的序列预测网络通常只产生固定长度的输出,不能自适应地确定预测序列的长度。
For this purpose, a highly robust gesture and sign language recognition method based on ultrasonic signals are proposed in this paper. First, we use the ultrasonic device Acoustic Software Defined Radios Platform (ASDP) to capture the gesture movement data and the amplitude information is used as the feature value for denoising and smoothing. Then we use short-time Fourier transform (STFT) to extract the Doppler shift of the movement data. To address the challenge of insufficient training data, we use GAN to automatically generate data. Then ResNet34 is used to extract the feature values and the bi-directional long short-term memory (Bi-LSTM) algorithm is used to classify the single gesture. For continuous gestures and sign language gestures, the CTC algorithm is added after the Bi-LSTM network. We use the dynamic programming method to find the output result with the highest probability as the final output result of the model. The main contributions of this paper are as follows:
本文提出了一种基于超声波信号的高度鲁棒的姿态和手语识别方法。首先,我们使用超声波设备 Acoustic Software Defined Radios Platform (ASDP)来捕捉手势运动数据,并将振幅信息用作去噪和平滑的特征值。然后,我们使用短时傅里叶变换(STFT)提取运动数据的多普勒频移。为了解决训练数据不足的挑战,我们使用 GAN 自动生成数据。然后,使用 ResNet34 提取特征值,并使用双向长短期记忆(Bi-LSTM)算法对单个手势进行分类。对于连续手势和手语手势,在 Bi-LSTM 网络之后添加了 CTC 算法。我们使用动态规划方法找到概率最高的输出结果作为模型的最终输出结果。本文的主要贡献如下:
1.
We propose a data augmentation method based on GAN. Due to the randomness of GAN itself, it makes the generated samples more diverse and can cover more real situations, while it can reduce the classification model error and improve the performance of the model.
我们提出了一种基于生成对抗网络(GAN)的数据增强方法。由于 GAN 本身的随机性,它使得生成的样本更加多样化,可以覆盖更多真实情况,同时可以降低分类模型误差并提高模型性能。
2.
We feed the multi-scale semantic features extracted by the residual neural network into the Bi-LSTM algorithm. The algorithm enables the classification network to fuse the information of feature dimension and temporal dimension to achieve high-precision gesture recognition. Meanwhile, in order to fill the gap of acoustic perception recognition of continuous gestures and Chinese sign language gestures and solve the problem of inconsistent length and difficult alignment of continuous gesture and sign language gesture input and output sequences, we add the CTC algorithm after the Bi-LSTM network. It enables the model to achieve good results for continuous gesture recognition and sign-language-recognition problems as well.
我们将残差神经网络提取的多尺度语义特征输入到 Bi-LSTM 算法中。该算法使分类网络能够融合特征维度和时间维度的信息,以实现高精度的手势识别。同时,为了填补连续手势和中文手势语对声学感知识别的空白,并解决连续手势和手势语输入输出序列长度不一致和难以对齐的问题,我们在 Bi-LSTM 网络之后添加了 CTC 算法。这使得模型在连续手势识别和手势语识别问题上取得了良好的效果。
3.
In this paper, we obtain real data on gestures from multiple groups of volunteers and form an open-source database. Through two real scene tests, it is verified that the proposed method has high robustness, the accuracy of single gesture recognition reaches 98.8%, and the recognition distance is 0.5 m. At the same time, the sign language data collected can provide data support for education professionals to study the daily interaction behavior of certain groups, such as the deaf.
本文从多组志愿者中获取了手势的真实数据,并形成了一个开源数据库。通过两次真实场景测试,验证了所提出的方法具有高度鲁棒性,单手势识别准确率达到 98.8%,识别距离为 0.5 米。同时,收集到的手语数据可以为教育专业人士研究某些群体(如聋人)的日常互动行为提供数据支持。
The remaining sections of this paper are organized as follows. Section 2 summarizes the existing work related to gesture and sign language recognition. Section 3 explains the implementation process of the UltrasonicGS method. In Section 4, we experiment and evaluate the performance of the UltrasonicGS method. Finally, Section 5 summarizes the work of this paper and explains the next research directions.
本文剩余部分组织如下。第 2 节总结了与手势和手语识别相关的研究工作。第 3 节解释了 UltrasonicGS 方法的实现过程。在第 4 节中,我们实验并评估了 UltrasonicGS 方法的表现。最后,第 5 节总结了本文的工作并解释了未来的研究方向。

2. Related Work  2. 相关工作

In this section, we present the current research related to single gesture recognition, continuous gesture recognition, and sign language gesture recognition in terms of Inertial Measurement Unit (IMU) sensors, vision, and acoustic. A single gesture is the execution of one action at a time, and a continuous gesture is the execution of multiple actions at a time. Additionally, a sign language gesture is the execution of all the gestures contained in a complete sentence at a time.
本节中,我们以惯性测量单元(IMU)传感器、视觉和声学为依据,介绍了与单手势识别、连续手势识别和手语手势识别相关的当前研究。单手势是指一次执行一个动作,连续手势是指一次执行多个动作。此外,手语手势是指一次执行一个完整句子中包含的所有手势。
IMU sensor: IMU sensor is composed of a gyroscope (GYRO) and an accelerometer (ACC). It is usually placed on the user’s arm to capture the movement of the arm. The IMU sensor-based recognition of single gestures works as follows. Trong et al. [15] used the accelerometer and gyroscope in a smartwatch to collect data and combined a one-dimensional convolutional neural network with a bi-directional long short-term memory (1D-CNN-BiLSTM) to analyze and learn the signal features from the sensor signals. The proposed model could achieve a 90% correct rate. Rinalduzzi et al. [16] proposed a machine learning method in conjunction with a magnetic positioning system for recognizing the static gestures associated with the sign language alphabet. The proposed machine learning method is based on a support vector machine, which demonstrated good generalization properties and resulted in a classification accuracy of approximately 97%. There is no related work on continuous gesture recognition, but more on recognition of sign language gestures based on IMU sensors. Hou et al. [17] designed the SignSpeaker system using the IMU sensor of a smartwatch. The SignSpeaker system provided an isolated fine-grained fingerspelling recognition model and a continuous sign language recognition model. Additionally, the system used LSTM and CTC to recognize sign language gestures, but it could not use a smartwatch to recognize two-handed movements. In a sensor-based system, gesture behavior is captured by the wearable sensor. Although it can accurately capture fine-grained behavior characteristics, wearable sensors will bring great inconvenience to daily life, and the cost is high, which can only be used in a few fixed places.
IMU 传感器:IMU 传感器由陀螺仪(GYRO)和加速度计(ACC)组成。它通常放置在用户的胳膊上以捕捉手臂的运动。基于 IMU 传感器的单手势识别工作如下。Trong 等人[15]在智能手表中使用了加速度计和陀螺仪来收集数据,并将一维卷积神经网络与双向长短期记忆(1D-CNN-BiLSTM)相结合,从传感器信号中分析和学习信号特征。所提出的模型可以达到 90%的正确率。Rinalduzzi 等人[16]提出了一种结合磁定位系统的机器学习方法,用于识别与手语字母相关的静态手势。所提出的机器学习方法基于支持向量机,表现出良好的泛化特性,并导致分类准确率约为 97%。关于连续手势识别没有相关研究,但更多关于基于 IMU 传感器的手势识别。Hou 等人[17]使用智能手表的 IMU 传感器设计了 SignSpeaker 系统。 SignSpeaker 系统提供了一个独立的细粒度手语识别模型和一个连续的手语识别模型。此外,该系统使用 LSTM 和 CTC 来识别手势,但不能使用智能手表来识别双手动作。在基于传感器的系统中,手势行为是通过可穿戴传感器捕捉的。尽管它可以准确地捕捉细粒度行为特征,但可穿戴传感器会给日常生活带来很大不便,成本也很高,只能用于少数固定地点。
Vision: vision-based systems typically use optical cameras to capture human behavioral features. After the research, vision-based technologies are mainly used to implement continuous gesture and sign language recognition. For continuous gestures, Liu et al. [18] proposed a few-shot continuous gesture recognition scheme based on RGB video. The scheme used Mediapipe to detect the key points of each frame in the video stream, decomposed the basic components of gesture features based on certain human palm structures, and then extracted and combined the above basic gesture features by a lightweight autoencoder network. Mahmoud et al. [19] presented a robust deep learning approach for characterizing, segmenting, and classifying isolated and continuous gesture sequences using depth, RGB, and grayscale input data. The proposed process was suitable for both full human action and gesture recognition. For sign language recognition and sign language translation work, Guo et al. [20] proposed a hierarchical-LSTM framework for sign language translation, which builds a high-level visual semantic embedding model for SLT. However, unseen sentence translation was still a challenging problem with limited sentence data and unsolved out-of-order word alignment. Tang et al. [21] proposed a graph-based multimodal sequential embedding graph (MSeqGraph) network to solve sign language translation with multimodal cues. Experiments on two benchmarks demonstrated the effectiveness of the proposed MSeqGraph and showed that exploiting multimodal cues contributes to a better representation and improved performance. GEN-OBT [22] was proposed to solve the task of sign language translation. Additionally, it designed a CTC-based reverse decoder to convert the generated poses backward into glosses, which guaranteed semantic consistency during the processes of gloss-to-pose and pose-to-gloss. Vision-based sign-language-recognition technology is already mature, and the technology not only considers sign language movements but also incorporates facial expressions, lip-synthesis, etc., which has improved recognition accuracy to a certain extent. Additionally, many sign language translation efforts have been proposed in order to reduce the differences between natural language and sign language recognition. However, the technology is susceptible to light, some infringement of user privacy, and high energy demand for long-term monitoring.
愿景:基于视觉的系统通常使用光学摄像头来捕捉人类行为特征。经过研究,基于视觉的技术主要用于实现连续手势和手语识别。对于连续手势,刘等人[18]提出了一种基于 RGB 视频的少样本连续手势识别方案。该方案使用 MediaPipe 检测视频流中每帧的关键点,基于某些人类手掌结构分解手势特征的基本组成部分,然后通过轻量级自动编码器网络提取和组合上述基本手势特征。Mahmoud 等人[19]提出了一种鲁棒的深度学习方法,用于使用深度、RGB 和灰度输入数据对孤立和连续的手势序列进行表征、分割和分类。该过程适用于完整的人类动作和手势识别。对于手语识别和手语翻译工作,郭等人[20]提出了一种用于手语翻译的分层-LSTM 框架,该框架为 SLT 构建了一个高级视觉语义嵌入模型。 然而,未见句子翻译仍然是一个具有挑战性的问题,因为句子数据有限且存在未解决的单词顺序错位问题。Tang 等人[21]提出了一种基于图的多模态序列嵌入图(MSeqGraph)网络来解决带有多模态提示的手语翻译问题。在两个基准测试上的实验证明了所提出的 MSeqGraph 的有效性,并表明利用多模态提示有助于更好的表示和性能提升。GEN-OBT[22]被提出用于解决手语翻译任务。此外,它设计了一个基于 CTC 的反向解码器,将生成的姿态反向转换为词素,保证了在词素到姿态和姿态到词素的过程中的语义一致性。基于视觉的手语识别技术已经成熟,该技术不仅考虑手语动作,还结合了面部表情、唇形合成等,这在一定程度上提高了识别精度。此外,还提出了许多手语翻译方法,以减少自然语言和手语识别之间的差异。 然而,这项技术容易受到光线影响,可能侵犯用户隐私,以及长期监控对能源需求较高。
Acoustic: acoustic-based systems typically use speakers and microphones embedded in electronic devices such as smartphones, headphones, and smart bracelets to obtain gesture information. Acoustic gesture recognition can solve the problem of wearable sensors inconvenient high cost but also based on the visual sensitivity to light, the user privacy impact of the problem. Acoustic technology only requires the use of speakers and microphones embedded in smart devices to collect data, reducing device collection costs, expanding the scope of everyday use, and slowing propagation characteristics to enable more accurate recognition. Some recent research works on acoustic gesture recognition have appeared. For single gestures, Mao et al. [23] proposed a system to measure the propagation distance and angle of arrival (AOA) of reflected signals using a four-element microphone array and dual speakers. The system did not allow for finger-level gesture recognition because the user need to hold the phone. Wang et al. [24] solved the frequency selective fading problem caused by multipath effects by periodically transmitting acoustic signals of different frequencies. Additionally, they solved the challenge of insufficient data by automatically generating data based on the correlation between CIR measurements and gesture changes, achieving a breakthrough in the limitations of acoustic gesture recognition in terms of accuracy and robustness. However, this research work can only recognize single gestures and can not handle the case of continuous gestures. For continuous gestures, FingerIO [25] analyzed the echo signal changes caused by finger movements by transmitting orthogonal frequency division multiplexing (OFDM) modulated acoustic signals to achieve accurate tracking of moving objects. However, it only captured finger movements in the two-dimensional plane and could not capture arm movements. The work most similar to ours is the work of Jin’s team. Jin et al. [26] used the speaker and microphone in a commercial headset to send and receive signals for real-time dynamic recognition of sign language gestures, and the system achieved 93.8% recognition for 42 words and 90.6% recognition for 30 sentences. However, the system is dependent on a wearable device (headset) to operate, making it a poor experience to use. Unlike Jin’s team, we did not rely on any wearable device and proposed the first acoustic-based Chinese continuous gesture and sign language recognition system with state-of-the-art results.
声学:基于声学的系统通常使用嵌入在智能手机、耳机和智能手环等电子设备中的扬声器和麦克风来获取手势信息。声学手势识别可以解决可穿戴传感器不便、成本高以及基于视觉对光敏感、对用户隐私影响的问题。声学技术只需要使用嵌入在智能设备中的扬声器和麦克风来收集数据,降低设备收集成本,扩大日常使用范围,并减缓传播特性,以实现更准确的识别。一些关于声学手势识别的最近研究工作已经出现。对于单个手势,Mao 等人[23]提出了一种系统,使用四元素麦克风阵列和双扬声器来测量反射信号的传播距离和到达角(AOA)。该系统不允许进行手指级手势识别,因为用户需要握住手机。Wang 等人[24]通过周期性地传输不同频率的声学信号,解决了由多径效应引起的频率选择性衰落问题。 此外,他们通过根据 CIR 测量值和手势变化之间的相关性自动生成数据,解决了数据不足的挑战,在准确性方面实现了突破,提高了声学手势识别的鲁棒性。然而,这项研究工作只能识别单个手势,无法处理连续手势的情况。对于连续手势,FingerIO [25] 通过传输正交频分复用(OFDM)调制的声学信号来分析由手指运动引起的回波信号变化,以实现移动对象的精确跟踪。然而,它只能捕捉二维平面上的手指运动,无法捕捉手臂运动。与我们工作最相似的是金团队的研究。金等人[26]使用商业头戴式耳机中的说话人和麦克风发送和接收信号,以实时动态识别手语手势,该系统对 42 个单词的识别率达到 93.8%,对 30 个句子的识别率达到 90.6%。然而,该系统依赖于可穿戴设备(耳机)来运行,使用体验不佳。 与金队不同,我们没有依赖任何可穿戴设备,并提出了第一个基于声学的具有最先进结果的中文连续手势和手语识别系统。

3. System Design  3. 系统设计

3.1. Overview  3.1. 概述

The system proposed in this paper is divided into four main parts: data collection, data pre-processing, feature extraction and gesture classification, and the system flow is shown in Figure 1. In the data collection and processing phase, two speakers are used as transmitters to send a single 20 kHz audio signal, a microphone is used as a receiver, and the receiving device records and stores the original echo signal. The raw echo signal is processed and converted to Doppler shift. Firstly, the images are filtered using a Butterworth bandpass filter and STFT, followed by a Gaussian filter to smooth the images. Finally, the dataset is expanded using GAN. In the feature extraction phase, the features of the spectrogram are extracted using the Resnet34 algorithm to generate feature vectors. The gesture classification phase feeds feature vectors into a Bi-LSTM network for classification and recognition. For the sequence prediction problem where the input and output sequences of continuous gestures and sign language gestures are of inconsistent length and difficult to align, we add the CTC algorithm after the Bi-LSTM network, which can convert the feature vector into an indeterminate length gesture sequence or sign language sequence.
本文提出的系统分为四个主要部分:数据收集、数据预处理、特征提取和手势分类,系统流程如图 1 所示。在数据收集和处理阶段,使用两名说话者作为发射器发送单个 20 kHz 音频信号,使用麦克风作为接收器,接收设备记录并存储原始回声信号。对原始回声信号进行处理并转换为多普勒频移。首先,使用巴特沃斯带通滤波器和短时傅里叶变换(STFT)对图像进行滤波,然后使用高斯滤波器平滑图像。最后,使用生成对抗网络(GAN)扩展数据集。在特征提取阶段,使用 Resnet34 算法从频谱图中提取特征,生成特征向量。在手势分类阶段,将特征向量输入到双向长短期记忆网络(Bi-LSTM)中进行分类和识别。 对于输入和输出序列长度不一致且难以对齐的连续手势和手语序列预测问题,我们在 Bi-LSTM 网络之后添加了 CTC 算法,该算法可以将特征向量转换为不定长手势序列或手语序列。
Figure 1. Overview of UltrasonicGS (In the output result module, “我是教师” is a Chinese sentence, which means “I am a teacher” in English. Where “我”“是”“教师” correspond to “I”, “am” and “teacher” respectively).
图 1. UltrasonicGS 概述(在输出结果模块中,“我是教师”是中文句子,在英语中意为“I am a teacher”。其中“我”“是”“教师”分别对应“I”,“am”和“teacher”)。

3.2. Data Collection and Pre-Processing
3.2. 数据收集与预处理

Data collection and pre-processing. The frequency of living noise is usually located at [1000, 4000] Hz [27]. In order to ensure that the signal frequency used in the experiment does not conflict with the frequency of living noise, this paper sets the speaker to send a single audio signal of 20 kHz. The single audio signal has the advantage of low complexity and high resolution in terms of Doppler shift [28]. Figure 2, Figure 3 and Figure 4 show the schematic diagrams of the Doppler effect corresponding to 15 single gestures, six sets of continuous gestures, and six sets of sign language gesture data after pre-processing, respectively. To better describe the gesture under test, in Figure 2 we use X→ to indicate the hand motion along the X-axis and double arrows (e.g., X ) to indicate the back and forth motion of the hand along the X-axis.
数据收集和预处理。生活噪音的频率通常位于[1000, 4000] Hz [27]。为了确保实验中使用的信号频率不与生活噪音的频率冲突,本文将扬声器设置为发送 20 kHz 的单音频信号。单音频信号在多普勒频移方面具有低复杂度和高分辨率的优势[28]。图 2、图 3 和图 4 分别显示了 15 个单个手势、六组连续手势和六组手语手势数据预处理后的多普勒效应示意图。为了更好地描述待测手势,在图 2 中,我们用 X→表示手沿 X 轴的运动,用双箭头(例如,X )表示手沿 X 轴的往返运动。
Figure 2. Single gesture spectrogram.
Figure 3. Continuous gesture spectrogram.
Figure 4. Sign language gesture spectrogram.
Hand gesture data processing. A Butterworth bandpass filter with a frequency of [19,000, 21,000] Hz is first used to eliminate the interference of background noise, followed by an STFT to extract the Doppler shift caused by the gesture motion. STFT is the most commonly used method for time-frequency analysis, but the time resolution and frequency resolution are difficult to balance. To balance real-time and frequency resolution, we set the frame length to 8192 and the window step size to 1024. The frequency change of the signal after reflection is estimated by calculating the Doppler shift, and the image shown in Figure 5a is obtained.
Δf=f0×|1vs±vfvsvf|
where f0 is the frequency of the signal sent by the speaker (20 kHz), vs is the speed of sound (340 m/s), vf is the speed of gesture movement (maximum movement speed 4 m/s). So the synthesized frequency shift is about 470.6 Hz, and the effective frequency range should be within [19,530, 20,470] Hz.
Figure 5. Single gesture action data processing process. (a) Bandpass filtering data; (b) Gaussian smoothing data.
To eliminate the effect of isolated noise generated by sudden hardware noise on the signal, the point where the STFT value changes most dramatically, 0.15, is set as the threshold value, and any isolated noise less than this threshold is set to 0. After we use a Gaussian filter to smooth the image. For two-dimensional images, the following Gaussian functions are used for smoothing.
G(x,y)=12πσ2exp(x2+y22σ2)
where x is the distance of the horizontal axis from the origin, y is the distance of the vertical axis from the origin, σ is the standard deviation of the Gaussian distribution, and the processed image is shown in Figure 5b.

3.3. Data Augmentation

Traditional data augmentation [29] generates new data from limited data by synthesis or transformation. Traditional data augmentation techniques in the image domain are based on a series of known affine transformations, such as rotation, scaling, displacement, etc., and some simple image processing tools, such as light color transformation, contrast transformation, noise addition, etc. This method of data augmentation based on geometric transformation and image manipulation can alleviate the overfitting problem of neural networks and improve the generalization ability to a certain extent, but the addition of new data does not fundamentally solve the problem of insufficient data compared with the original data. The recent emergence of GAN [30] can also be used for data augmentation. This network-based synthesis method is more complex than traditional data enhancement techniques, but the generated samples are more diverse and can be applied to various scenarios, such as image editing and image denoising.
GAN consists of a discriminator network and a generator network. Discriminators are two-category classification networks that distinguish whether x comes from the true distribution or the generative model. Unlike the fully connected neural network-based discriminator in the original GAN network, we use CNN as a discriminator to better extract features in gesture images. The generation needs to make the discriminator network distinguish its own generated samples from real samples. First, the generator randomly initializes a latent vector, and then continuously performs convolution and upsampling operations to transform the latent vector to the size of the actual image. The basic structure of GAN is shown in Figure 6.
Figure 6. Overview of the GAN.
X represents the real data, z represents the noise of the generator network, G(z) means unreal data generated by the generator network, and D(x) represents the probability that x belongs to the real sample distribution, where D[0,1]. The optimization principle of GAN is simply that the generator network, G, generates G(z) through continuous training and learning and makes the discriminator network, D, unable to distinguish the difference between G(z) and X. D is to improve their discriminant ability through continuous training and learning, that is, to recognize that X and G(z) are different.
The optimization function of the whole GAN network can be summarized by Equation (3):
minGmaxDV(D,G)=ExPdata(x)[logD(x)]+EZPZ(Z)[log(1D(G(z)))]
The main meaning of this equation is that one is the G remains constant and the D wants to distinguish the real samples from the training samples. Additionally, the other is the D remains constant and by adjusting the G it wants the D to make a mistake and not let it distinguish as much as possible. The training process of generators and discriminators is iterated alternately. First, optimize the discriminator D. The purpose of the discriminator is to be able to correctly distinguish between G(z) and X. When optimizing the discriminator network, it is necessary to give D and G in advance and try to increase D(x) and decrease D(G(z)), i.e., the optimization objective of the discriminator network is maxDV(D,G). When optimizing the generator network, it is also necessary to give D and G in advance and optimize minGV(D,G).
Specifically, we set the set of input images P={p1,p2,,pm}. To train the discriminator model, for each small batch, m samples are sampled from the prior noise distribution pg(z) as {z(1),,z(m)}, and m samples are obtained from the real data distribution pdata(x) as {x(1),,x(m)}, and the discriminator is updated by boosting the random gradient Equation (4). When training the generator model, for each small batch, again m samples are sampled from the prior noise distribution pg(z) and the generator is updated by reducing the random gradient Equation (5).
θd1mi=1m[logD(x(i))+log(1D(G(z(i))))]
θg1mi=1mlog(1D(G(z(i))))
In practice, we build a GAN network for each category of data separately. As shown in Figure 6, the generated images are basically the same as the original images, and it is difficult to distinguish the difference between the real samples and the generated samples. Therefore, by means of GAN, a large amount of high-quality data can be expanded in a short time and used for the training of subsequent gesture recognition models.

3.4. Feature Extraction and Gesture Classification

3.4.1. Feature Extraction

In this paper, we use ResNet34 [31] to extract features, and its structure is shown in Figure 7. The ResNet34 model has 34 convolutional layers, including a total of 16 residual learning units, where all convolutional operations use a convolutional kernel of size 3 × 3. The spectrogram obtained from data augmentation is used as the input to ResNet34, ensuring that the input images are all 64 × 64 pixels in size. After each convolutional layer and before the activation function (ReLU), batch normalization is used to accelerate the convergence. Performing reshapes and flatten operations on the output of the last residual block, we can obtain the feature vector y=[y1,y2,,yT], the total number of feature vectors T=512, and the length of each feature vector is 16.
Figure 7. Overview of the ResNet34 and Bi-LSTM.

3.4.2. Gesture Classification

Bi-LSTM. Traditional LSTM can only encode information from front to back, not from back to front, but information from back to front is also important for determining activity. Bi-LSTM [32] can better capture the semantic dependencies in both directions. The Bi-LSTM network computation is usually divided into the following four steps:
Step 1: from the forgetting gate ft, determine the information to be discarded from the cell state. The forgetting gate can read the output ht1 of the previous sequence, the input xt of the current sequence and perform the Sigmoid operation:
ft=σ(Wf·[ht1,xt]+bf)
Step 2: determine what new information will be stored in the sequence state. First of all, the Sigmoid layer determines which values we will update. Subsequently, a new vector of candidate values C˜t is created using the tanh layer.
it=σ(Wi·[ht1,xt]+bi)
C˜t=tanh(Wc·[ht1,xt]+bc)
Step 3: update sequence status.
Ct=ftCt1+itC˜t
Step 4: determine the output values based on the updated sequence states. First of all, the Sigmoid layer is used to determine which sequence states can be output. Then the sequence states Ct obtained in the third step are mapped to between −1 and 1 using tanh and multiplied with the Sigmoid gate ot to obtain the final output ht.
ot=σ(W0·[ht1,xt]+b0)
ht=ottanh(Ct)
where ht1 denotes the output of the spectrogram sequence at the previous moment, xt denotes the input of the spectrogram sequence at the current moment, W and b are the weight term and bias term to be learned, respectively, σ denotes the Sigmoid operation, ft denotes the output of the forgotten gate at time t, it denotes the information of the spectrogram sequence to be activated at the moment t, Ct1, and Ct denote the state of the spectrogram feature sequence at the moment t1 and moment t, respectively, ht is the output result of the output gate at time t.
Specifically, the feature vectors y extracted by the residual neural network are passed to two LSTM layers, each of which has T(T=512) LSTM storage units. To improve the generalization ability of the model set the dropout of the model to 0.8. These two layers perform sequence feature extraction in opposite directions, and each LSTM memory cell will be computed by three gating units. After calculation, the output Hforward of the forward LSTM and the output Hbackward of the reverse LSTM can be obtained. After that, we concatenate and flatten Hforward and Hbackward to obtain the vector P. In the single-category gesture recognition task, since the classifier eventually needs to recognize 15 gestures, we design a fully connected neural network with 15 output neurons. Finally, softmax operations are performed on the output of the fully connected layer to accurately classify and recognize different gestures. In the case of continuous gesture or sign language recognition tasks, it is necessary to input the vector p to the CTC algorithm for processing, and we will describe this process in detail in the next section.
CTC. In this paper, we use the CTC [33] algorithm as a classifier for the continuous gesture and sign-language-gesture recognition. CTC is an algorithm commonly used in speech recognition, text recognition, and other fields to solve the problem of unaligned input and output sequences of different lengths. Unlike single gesture prediction, after the Bi-LSTM network obtains the feature vector pRc×n (c represents the length of the feature vector and n represents the number of classes of gestures or sign language), the fully connected layer is no longer designed, but p is input into the CTC algorithm. Algorithm 1 shows the steps of the CTC method.
First, the CTC layer receives the output sequence p from the Bi-LSTM and then computes the probability pctc(Y|p) between p and the true label Y on any alignment π, where π[t] is the character ID aligned to the tth frame in p, as follows:
C=softmax(pWctc+bctc)
p(B(π)=Y|p)=t=1nsubC[t,π[t]]
pctc(Y|p)=πB1(Y)p(B(π)=Y|p)
where WctcRn×char and bctcRchar are learnable parameters, CRc×char is the output of CTC, C[t,π[t]] is the probability that the output character π[t] is aligned with the tth frame. The many-to-one mapping B(π) is used to remove redundant symbols from the alignment π, for example, B(aaØb)=ab, where Ø is a blank character and the one-to-many mapping B1 projects the sequence of characters into a set of character sequences with redundant symbols.
B1(Y)={π|Y=B(π)}
In the training phase, we train the entire set of models using the CTC loss function.
Lctc=logctc(Y|p)
In the prediction phase, we need to use the Beam Search Decoding algorithm to convert the feature vectors predicted by Bi-LSTM into the final sign language sequence prediction results. In the sequence prediction problem, the model prediction process is essentially a spatial search process, the core of which is to calculate the probability of expanding nodes at each step. The sequence with the highest probability the last time is taken as the final output of the model.
Algorithm 1 Steps of CTC
Input: Sequence of strings L, Number of nodes in each expansion W
Output: The sequence Q with the maximum probability at time T
1: for t=1 to T do
2:       Set B^ = the W most probable sequences in B (L when t=1)
3:       Set B={ }
4:       for pB^ do
5:             if pØ then
6:                   r+(p,t)=r+(p, t1)ypet
7:                   if p^B^ then
8:                         r+(p, t)+=Probability(pe, p^, t)
9:             r(p, t)=r(p, t1)ybt
10:             add p to B
11:             for k=1 to K do
12:                   r(p+k, t)=0
13:                   r+(p+k, t)=Probability(k, p, t)
14:                   add (p+k) to B
15: return  argmaxpB r(p, T)1|p|

4. Experimentation and Evaluation

4.1. Experiment Setting

Experimental platform. In the experimental phase, ASDP equipped with one microphone and two speakers were chosen as the data collection tool. Two speakers are transmitters (Tx) and one microphone is a receiver (Rx). ASDP is an acoustic software-defined radio platform, a multi-functional communication and sensing platform. The ASDP is mainly composed of hardware, such as Raspberry Pi, INMP411, TPS54332, WM8731, etc. The platform is shown in Figure 8a. Set the speaker to emit a 20 kHz continuous single audio signal and set the microphone sampling rate to 44.1 kHz.
Figure 8. Experimental equipment and environment. (a) Data collection equipment; (b) Laboratory environment; (c) Corridor environment.
Dataset. We collected data in two scenarios, laboratory, and corridor, and the real scenario was shown in Figure 8b,c. We invited 6 male volunteers and 6 female volunteers to perform 15 single gestures. Additionally, we collected 720 sets of data under 4 practical influencing factors of distance, speed, noise, and angle. Then we invited 2 male volunteers and 2 female volunteers to perform 6 continuous gestures and 6 sign language gestures, and 120 sets of data were collected for each. All of the above actions were performed by the volunteer while keeping the body stationary and within a distance of 0.2 m to 0.5 m from the device. The open source address for the dataset is: https://github.com/yuejiaowang/database (accessed on 31 December 2022).
Implementation details. In our experiments, the input image for single gesture recognition is resized to 512 × 512, and the input image for the continuous gesture and sign language gesture recognition is resized to 620 × 462. For data augmentation, we use the method mentioned in Section 3.3 for 20× data augmentation with the addition of random scaling and random rotation. In the experiments for single gesture recognition, continuous gesture recognition, and sign language recognition, we use 80% of the data as the training set and the remaining 20% as the test set. Additionally, the results reported in the experiments are all 5-fold cross-validation results. Our network architecture is implemented in PyTorch. In single gesture recognition experiment, we use Adam optimizer with a learning rate 1×103 and set the batch size to 16. A total of 60 epochs are trained. In the experiments of continuous gesture and sign language gesture, we use the Adam optimizer with an initial learning rate of 1×104 and set the batch size to 2. A total of 100 epochs are trained and the learning rate is reduced by a factor of 10 in the 60th and 80th epochs, respectively. All recognition models are not loaded with any pre-training weights and experiments are conducted on NVIDIA Tesla P40 GPU.

4.2. Ablation Study

4.2.1. Impact of Different Influencing Factors

In order to evaluate the UltrasonicGS method in terms of different influencing factors, this paper designed experiments in three aspects: distance between gesture and transceiver, angle of arrival, and gesture speed in laboratory and corridor environments, respectively. (1) Five experimenters were asked to execute the gesture at 5 cm, 15 cm, 25 cm, 35 cm, and 50 cm from the transceiver position. (2) Five experimenters were asked to execute the gestures at 30°, 60°, 90°, 120°, and 150° with the equipment. (3) Five experimenters were asked to perform gestures of duration 0.5 s, 1 s, 1.5 s, 2 s, and 2.5 s, respectively. The results of the experiment are shown in Figure 9.
Figure 9. Impact of different distance, angles and speeds. (a) Impact of different distance; (b) impact of different angles; and (c) impact of different speeds.
Figure 9a shows the impact of environment and distance on the correct gesture recognition rate. From the perspective of the environment, it can be seen that the recognition result of the corridor environment is higher than that of the laboratory environment at the same distance from the transceiver. This is due to the fact that the laboratory contains regularly distributed equipment with tables and chairs, so the multipath effect is more disturbing. From the perspective of distance, it can be seen that when the distance between the hand and the device is 15 cm, the correct gesture recognition rate reaches up to 98%. As the distance between the hand and the device increases, the correct gesture recognition rate gradually decreases. When the distance is 50 cm, the correct gesture recognition rate is close to 88%. The reason for this phenomenon is that when the distance is too small, the signal reflected by the hand is not completely received by the microphone. When the distance is too large, the interference of the multipath effect on the reflected signal increases.
Figure 9b shows the impact of environment and angle of arrival on the correct gesture recognition rate. As can be seen from the figure, when the experimenter performs the gesture at 90° to the device, the gesture recognition rate is 99% correct. When the experimenter is at 30°, 60°, 120°, and 150° to the device, the gesture recognition rate does not differ much, fluctuating around 96%. This is because when the angle of arrival is 90°, the direction of hand motion is perpendicular to the signal domain, which has a greater impact on the signal. Additionally, when the experimenter is at other angles to the device, the hand motion generates a horizontal motion component with a smaller signal amplitude. Overall, UltrasonicGS is able to maintain high performance specifications in all directions.
Figure 9c shows the impact of environment and speed on the correct rate of gesture recognition. The figure shows that when the gesture duration is 1.5 s, the highest correct gesture recognition rate can reach 98.7%. As the duration of the gesture increases or decreases, the correct gesture recognition rate decreases. This is because the gesture duration is too long, the gesture speed is too slow, and the signal change caused by the Doppler shift is not obvious. The gesture duration is too short, the gesture speed is too fast, and the microphone fails to receive the complete signal in a short period.
The experimental results demonstrate that UltrasonicGS maintains good recognition performance within a distance of 50 cm between the hand and the transceiver, in all directions, and within a hand gesture duration of 2.5 s.

4.2.2. Impact of Noise and Personnel Interference

To evaluate the impact of the UltrasonicGS method on ambient noise, line-of-sight (LOS), non-line-of-sight (NLOS), and personnel interference, we designed the following two experiments. (1) Experimenters were asked to perform 15 gestures at 15 cm from the device position in the no noise, low-frequency noise, and 19 kHz ultrasonic noise of LOS and NLOS environments, respectively. (2) Experimenters were asked to perform 15 gestures in four situations of interference: no human interference, human static interference (experimenter standing still), human light interference (experimenter walking back and forth), and human heavy interference (experimenter executing disturbance gestures while walking).
The results in Figure 10a show that the correct gesture recognition rate stays above 98% in the LOS environment and fluctuates around 91.2% in the NLOS environment. This is due to the better signal quality and higher throughput in the LOS channel model, however, the multipath effect in the NLOS channel model leads to frequency selective fading. From the perspective of noise, it can be seen that low-frequency noise and ultrasonic noise have basically no effect on the experimental results, which further verifies that the data pre-processing method proposed in this paper can remove noise interference well.
Figure 10. Impact of different environments and interference states. (a) Impact of different environments and (b) impact of different interference states.
The cumulative distribution functions (CDF) of the error rate for different interference states are given in Figure 10b. The x-axis represents the recognition error rate and the y-axis represents the CDF percentage. At a CDF of 0.8, the error rates corresponding to no human interference, human static interference, human light interference, and human heavy interference are 0.09, 0.11, 0.14, and 0.18, respectively. The highest accuracy is achieved in an environment without human interference, and the worst recognition performance is achieved in an environment with human heavy interference. However, the error rate of about 80% of the test data is less than 18%, which indicates that the method proposed in this paper has some anti-interference capability.

4.2.3. Impact of Dataset Size

To evaluate whether data augmentation helps to improve the performance of the gesture recognition model, we conducted experiments in three tasks: single gestures, continuous gestures, and sign language gestures, respectively. Figure 11 shows the ROC curves with and without data augmentation in turn.
Figure 11. Impact on recognition performance of single gesture, continuous gesture and sign language gesture when data augmentation is used or not. (a) Single gesture; (b) continuous gesture; and (c) sign language gesture.
In Figure 11, the blue curve and the area surrounded by the x-axis are the Area Under Curve (AUC) when the data augmentation method is used in the UltrasonicGS method and the red curve and the area surrounded by the x-axis are the AUC when the data augmentation method is not used. We can observe that, whether it is a single gesture, continuous gesture, or sign language gesture, when we use the GAN data augmentation method, the receiver operating characteristic (ROC) curve rises faster and the area occupied by AUC will be larger, and the recognition effect will be better than without the method. Therefore, data augmentation techniques can extend the dataset and help to improve the performance of the gesture recognition model. We will use data augmentation techniques in a series of subsequent experiments.

4.3. Comparison with the State-Of-The-Art Methods

In order to verify the superiority of our proposed method in gesture recognition, we compared it with the classical methods of acoustic sensing gesture recognition in recent years. Table 1 details the differences between the five methods with respect to the five aspects of sending signal, device, application, algorithm, and feature extraction for the word level. Table 2 compares with SonicASL, which is based only on acoustics for sign language sensing.
Table 1. Comparison with the word level methods.
Table 2. Comparison with the sentence level methods.
In Table 1, it can be observed that the recognition accuracy of our proposed method reaches 98.8%, which is the best performance among all methods. AudioGest and SoundWave are suitable for recognizing whole-hand gestures, while our dataset contains fine-grained finger-level gestures, resulting in poor recognition of the above two methods, with recognition accuracies of 89.1% and 88.6%, respectively. Thanks to the multiscale semantic features extracted by our CNN fed into the Bi-LSTM algorithm, we can make the classification network fuse the information of feature dimension and temporary dimension. Additionally, the recognition performance is significantly better than that of other finger-level recognition methods UltraGesture and Push. In Table 2, both SonicASL and our method can recognize word-level and sentence-level gesture activities. Additionally, our proposed method recognizes individual gestures with a 5% higher correct rate than SonicASL but recognizes sign language gestures with 4.3% lower than the comparison method. The reason for this situation is that we perform Chinese sign language recognition, while SonicASL performs English sign language recognition, which is a more complex situation involving homophones and split words. After experiments, our method increases the recognition correct rate when recognizing continuous sentences in English. Therefore, our proposed method can meet the demand for action recognition in general perceptual space and can ensure stable recognition accuracy.

4.4. Overall Performance

4.4.1. Overall Accuracy of Single Gestures

In order to evaluate the accuracy of 15 single gestures, the experimenters were asked to perform this experiment in different environments (multipath-rich and multipath-not-rich rooms) and with different influencing factors (distance angle and speed) in this section. The results of the experiment are shown in Figure 12.
Figure 12. Overall performance of single gestures.
Figure 12 shows the overall confusion matrix for performing 15 single gestures in different environments and with different influencing factors. The results of the confusion matrix show that the UltrasonicGS method has a combined recognition rate of 98.8%. Among them, 10 gestures, such as “1, 2, pinch, pull, push” can achieve 100% correct recognition rate. In order to ensure the authenticity and expandability of the dataset, each experimenter can perform the gestures “3” and “OK” according to their own habits when actually collecting data. This resulted in similar gestures for “3” and “OK”, with a small difference in the Doppler effect. The recognition rate of the above two gestures is slightly lower, but the correct rate is 93%. In summary, the UltrasonicGS method is able to distinguish the 15 single gesture actions well.

4.4.2. Performance Evaluation of Continuous Gesture

To evaluate the performance of the UltrasonicGS method for continuous gesture recognition, four classification models were selected. ResNet34 extracted feature values, Bi-LSTM, and CTC-classified gestures. VGG16 [37] extracted feature values, Bi-LSTM and CTC classified gestures. ResNet34 extracted feature values, LSTM [38], and CTC classified gestures. VGG16 extracted feature values, LSTM, and CTC-classified gestures. The six groups of continuous gestures selected in the experiment were: Spread and Pinch; Push and Pull; Hover and OK; Around Left and Around Right; One, Two, and Three; and Slide Up, Slide Down, Slide Left, and Slide Right. The experimental results are shown in Figure 13 and Figure 14.
Figure 13. Impact of classification model on continuous gesture performance. (a) Spread and Pinch; (b) Push and Pull; (c) Hover and OK; (d) Around Left and Around Right; (e) One, Two, and Three; and (f) Slide Up, Slide Down, Slide Left, and Slide Right.
Figure 14. Impact of different models on accuracy of continuous gestures.
The CDF of error rates for different classification algorithms are given in Figure 13. The six CDF figures represent six different continuous gestures, where the first four CDF figures are continuous gestures composed of two gestures, the fifth is a continuous gesture composed of three gestures, and the sixth is a continuous gesture composed of four gestures. Globally, the six CDF plots of error rates for each classification algorithm vary essentially uniformly. Using ResNet34 to extract feature values, Bi-LSTM and CTC achieve the highest accuracy for classification of continuous gestures, where approximately 89% of the tested data have an error rate of less than 10%. Using ResNet34 to extract feature values, LSTM and CTC gesture classification have similar recognition rates as using VGG16 to extract feature values, with approximately 80% of the test data having an error rate of less than 20%.
Figure 14 shows the accuracy of six continuous gestures with different classification models. C1, C2, C3, C4, C5, and C6 correspond to each of the six gestures in Figure 13. For each gesture using ResNet34 to extract the feature values, both Bi-LSTM and CTC classification achieved the highest accuracy, with an average accuracy of 92.4%. Using VGG16 to extract the feature values, LSTM and CTC achieved the lowest accuracy, with an average accuracy of 90.97%. This shows that the method used in this paper can recognize not only single gestures but also continuous gestures. Additionally, the method incorporates the information of feature dimension and temporary dimension, which effectively improves the accuracy of gesture recognition.

4.4.3. Performance Evaluation of Sign Language Gesture

In order to evaluate the performance of the UltrasonicGS method for sign language gesture recognition, we also chose the same four classification models as in the previous experimental continuous gesture performance evaluation for “I am a teacher.” “I am fine, thanks.” “What day is today?” “Sorry, I am late.” “What do you do?” “What is your name?” six groups of Chinese sign language carried out the experiment, and the experimental results are shown in Figure 15 and Figure 16.
Figure 15. Impact of classification model on sign language gesture performance. (a) I am a teacher. (b) I am fine, thanks. (c) What day is today? (d) Sorry, I am late. (e) What do you do? (f) What is your name?
Figure 16. Impact of different models on accuracy of sign language gestures.
The ROC curves of different classification models are given in Figure 15. The x-axis represents the false positive case rate, the y-axis represents the true case rate, and the six ROC plots represent the six different sign language gestures. Globally, there is almost no difference in ROC curves and similar AUC areas for the six different sentence descriptions, which indicates that the same model is similarly effective in recognizing six different sets of sign language sentences. Using ResNet34 to extract feature values, the Bi-LSTM and CTC algorithms are used to classify sign language gestures with the fastest ROC curve change and the largest AUC area, while the other three classification models have a slightly slower ROC curve change and smaller corresponding AUC areas.
Figure 16 shows the accuracy of the six sign language gestures under different classification models. S1, S2, S3, S4, S5, and S6 correspond to the six gestures in Figure 15. For each gesture using ResNet34 to extract the feature values, both Bi-LSTM and CTC classification achieved the highest accuracy with an average accuracy of 86.3%. Using VGG16 to extract feature values, LSTM and CTC achieved the lowest correct classification rate of 84.2% for gestures. This shows that the method used in this paper can recognize not only continuous gestures but also sign language gestures. The method incorporates the information on feature dimension and temporary dimension, which effectively improves the accuracy of gesture recognition.

5. Conclusions

In this study, we propose the UltrasonicGS, a highly robust gesture and sign language recognition method based on ultrasonic signals. The method can recognize 15 single gestures with high accuracy and robustness. Additionally, in order to satisfy more audience groups, especially special groups, such as the deaf, we extend the method to recognize continuous gestures and sign language gestures. To achieve fine-grained gesture recognition, the extraction of feature values using ResNet34 and the classification of single gestures by Bi-LSTM. For continuous gestures and sign language gestures, we add CTC algorithm after Bi-LSTM network to solve the problem of inconsistent length and difficult alignment of input and output sequences of continuous gestures and sign language gestures. To further improve the robustness of UltrasonicGS, automatic data generation using GAN can alleviate the problem of neural network overfitting and improve the generalization ability to a certain extent. Finally, a dataset containing three categories of gestural behavior is constructed and open sourced. The experimental results show that the method recognize a distance of 0.5m, and the overall correct rate of single gestures reach 98.8%, and the average correct rates of recognition for six groups of continuous gestures and sign language gestures are 92.4% and 86.3%, respectively.
In future work, we will further investigate (1) improving the recognition accuracy of this model for sign language datasets and (2) replacing the collection device with a cell phone to achieve sign language gesture speech conversion and text conversion functions to improve human–computer interaction.

Author Contributions

Conceptualization, Y.W. and Z.H.; methodology, Y.W.; software, Y.W.; validation, Z.Z. and M.L.; data curation, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, Z.H. and X.D.; supervision, Z.H.; project administration, Z.H.; funding acquisition, Z.H. and X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant 62262061, Grant 62162056, Grant 62261050), Key Science and Technology Support Program of Gansu Province (Grant 20YF8GA048), 2019 Chinese Academy of Sciences “Light of the West” Talent Program, Science and Technology Innovation Project of Gansu Province (Grant CX2JA037, 17CX2JA039), 2019 Lanzhou City Science and Technology Plan Project (2019-4-44), 2020 Lanzhou City Talent Innovation and Entrepreneurship Project (2020-RC-116, 2021-RC-81), and Gansu Provincial Department of Education: Industry Support Program Project (2022CYZC-12).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Considerations for Quarantine of Contacts of COVID-19 Cases: Interim Guidance, 25 June 2021; Technical Report; World Health Organization: Geneva, Switzerland, 2021.
  2. Savoie, P.; Cameron, J.A.; Kaye, M.E.; Scheme, E.J. Automation of the timed-up-and-go test using a conventional video camera. IEEE J. Biomed. Health Inform. 2019, 24, 1196–1205. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, Y.; Ma, J.; Li, X.; Zhong, A. Hierarchical multi-classification for sensor-based badminton activity recognition. In Proceedings of the 2020 15th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 6–9 December 2020; Volume 1, pp. 371–375. [Google Scholar]
  4. Li, J.; Yin, K.; Tang, C. SlideAugment: A Simple Data Processing Method to Enhance Human Activity Recognition Accuracy Based on WiFi. Sensors 2021, 21, 2181. [Google Scholar] [CrossRef]
  5. Zhou, S.; Zhang, W.; Peng, D.; Liu, Y.; Liao, X.; Jiang, H. Adversarial WiFi sensing for privacy preservation of human behaviors. IEEE Commun. Lett. 2019, 24, 259–263. [Google Scholar] [CrossRef]
  6. Wang, W.; Li, J.; He, Y.; Guo, X.; Liu, Y. MotorBeat: Acoustic Communication for Home Appliances via Variable Pulse Width Modulation. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2022, 6, 1–24. [Google Scholar] [CrossRef]
  7. Zhuang, Y.; Wang, Y.; Yan, Y.; Xu, X.; Shi, Y. ReflecTrack: Enabling 3D Acoustic Position Tracking Using Commodity Dual-Microphone Smartphones. In Proceedings of the 34th Annual ACM Symposium on User Interface Software and Technology, Virtual, 10–14 October 2021; pp. 1050–1062. [Google Scholar]
  8. Xu, X.; Gong, J.; Brum, C.; Liang, L.; Suh, B.; Gupta, S.K.; Agarwal, Y.; Lindsey, L.; Kang, R.; Shahsavari, B.; et al. Enabling hand gesture customization on wrist-worn devices. In Proceedings of the CHI Conference on Human Factors in Computing Systems, New Orleans, LA, USA, 29 April–5 May 2022; pp. 1–19. [Google Scholar]
  9. Xu, X.; Shi, H.; Yi, X.; Liu, W.; Yan, Y.; Shi, Y.; Mariakakis, A.; Mankoff, J.; Dey, A.K. Earbuddy: Enabling on-face interaction via wireless earbuds. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; pp. 1–14. [Google Scholar]
  10. Gao, Y.; Jin, Y.; Li, J.; Choi, S.; Jin, Z. EchoWhisper: Exploring an Acoustic-based Silent Speech Interface for Smartphone Users. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–27. [Google Scholar] [CrossRef]
  11. Wang, W.; Liu, A.X.; Sun, K. Device-free gesture tracking using acoustic signals. In Proceedings of the 22nd Annual International Conference on Mobile Computing and Networking, New York, NY, USA, 3–7 October 2016; pp. 82–94. [Google Scholar]
  12. Yun, S.; Chen, Y.C.; Zheng, H.; Qiu, L.; Mao, W. Strata: Fine-grained acoustic-based device-free tracking. In Proceedings of the 15th Annual International Conference on Mobile Systems, Applications, and Services, Niagara Falls, NY, USA, 19–23 June 2017; pp. 15–28. [Google Scholar]
  13. Wang, P.; Jiang, R.; Liu, C. Amaging: Acoustic Hand Imaging for Self-adaptive Gesture Recognition. In Proceedings of the IEEE INFOCOM 2022-IEEE Conference on Computer Communications, London, UK, 2–5 May 2022; pp. 80–89. [Google Scholar]
  14. Hao, Z.; Duan, Y.; Dang, X.; Liu, Y.; Zhang, D. Wi-SL: Contactless fine-grained gesture recognition uses channel state information. Sensors 2020, 20, 4025. [Google Scholar] [CrossRef] [PubMed]
  15. Nguyen-Trong, K.; Vu, H.N.; Trung, N.N.; Pham, C. Gesture recognition using wearable sensors with bi-long short-term memory convolutional neural networks. IEEE Sens. J. 2021, 21, 15065–15079. [Google Scholar] [CrossRef]
  16. Rinalduzzi, M.; De Angelis, A.; Santoni, F.; Buchicchio, E.; Moschitta, A.; Carbone, P.; Bellitti, P.; Serpelloni, M. Gesture Recognition of Sign Language Alphabet Using a Magnetic Positioning System. Appl. Sci. 2021, 11, 5594. [Google Scholar] [CrossRef]
  17. Hou, J.; Li, X.Y.; Zhu, P.; Wang, Z.; Wang, Y.; Qian, J.; Yang, P. Signspeaker: A real-time, high-precision smartwatch-based sign language translator. In Proceedings of the 25th Annual International Conference on Mobile Computing and Networking, Los Cabos, Mexico, 21–25 October 2019; pp. 1–15. [Google Scholar]
  18. Liu, Z.; Pan, C.; Wang, H. Continuous Gesture Sequences Recognition Based on Few-Shot Learning. Int. J. Aerosp. Eng. 2022, 2022, 7868142. [Google Scholar] [CrossRef]
  19. Mahmoud, R.; Belgacem, S.; Omri, M.N. Towards an end-to-end isolated and continuous deep gesture recognition process. Neural Comput. Appl. 2022, 34, 13713–13732. [Google Scholar] [CrossRef]
  20. Guo, D.; Zhou, W.; Li, H.; Wang, M. Hierarchical lstm for sign language translation. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  21. Tang, S.; Guo, D.; Hong, R.; Wang, M. Graph-based multimodal sequential embedding for sign language translation. IEEE Trans. Multimed. 2021, 24, 4433–4445. [Google Scholar] [CrossRef]
  22. Tang, S.; Hong, R.; Guo, D.; Wang, M. Gloss Semantic-Enhanced Network with Online Back-Translation for Sign Language Production. In Proceedings of the 30th ACM International Conference on Multimedia, New York, NY, USA, 10–14 October 2022; pp. 5630–5638. [Google Scholar]
  23. Mao, W.; He, J.; Qiu, L. Cat: High-precision acoustic motion tracking. In Proceedings of the 22nd Annual International Conference on Mobile Computing and Networking, New York, NY, USA, 3–7 October 2016; pp. 69–81. [Google Scholar]
  24. Wang, Y.; Shen, J.; Zheng, Y. Push the limit of acoustic gesture recognition. IEEE Trans. Mob. Comput. 2020, 21, 1798–1811. [Google Scholar] [CrossRef]
  25. Nandakumar, R.; Iyer, V.; Tan, D.; Gollakota, S. Fingerio: Using active sonar for fine-grained finger tracking. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 1515–1525. [Google Scholar]
  26. Jin, Y.; Gao, Y.; Zhu, Y.; Wang, W.; Li, J.; Choi, S.; Li, Z.; Chauhan, J.; Dey, A.K.; Jin, Z. Sonicasl: An acoustic-based sign language gesture recognizer using earphones. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 1–30. [Google Scholar] [CrossRef]
  27. Basner, M.; Babisch, W.; Davis, A.; Brink, M.; Clark, C.; Janssen, S.; Stansfeld, S. Auditory and non-auditory effects of noise on health. Lancet 2014, 383, 1325–1332. [Google Scholar] [CrossRef] [PubMed]
  28. Cai, C.; Pu, H.; Hu, M.; Zheng, R.; Luo, J. Acoustic software defined platform: A versatile sensing and general benchmarking platform. IEEE Trans. Mob. Comput. 2021, 22, 647–660. [Google Scholar] [CrossRef]
  29. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
  30. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Kawakami, K. Supervised Sequence Labelling with Recurrent Neural Networks. Ph.D. Thesis, Technical University of Munich, Munich, Germany, 2008. [Google Scholar]
  33. Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 369–376. [Google Scholar]
  34. Ruan, W.; Sheng, Q.Z.; Yang, L.; Gu, T.; Xu, P.; Shangguan, L. AudioGest: Enabling fine-grained hand gesture detection by decoding echo signal. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; pp. 474–485. [Google Scholar]
  35. Gupta, S.; Morris, D.; Patel, S.; Tan, D. Soundwave: Using the doppler effect to sense gestures. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Austin, TX, USA, 5–10 May 2012; pp. 1911–1914. [Google Scholar]
  36. Ling, K.; Dai, H.; Liu, Y.; Liu, A.X.; Wang, W.; Gu, Q. Ultragesture: Fine-grained gesture sensing and recognition. IEEE Trans. Mob. Comput. 2020, 21, 2620–2636. [Google Scholar] [CrossRef]
  37. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
Figure 1. Overview of UltrasonicGS (In the output result module, “我是教师” is a Chinese sentence, which means “I am a teacher” in English. Where “我”“是”“教师” correspond to “I”, “am” and “teacher” respectively).
Sensors 23 01790 g001
Figure 2. Single gesture spectrogram.
Sensors 23 01790 g002
Figure 3. Continuous gesture spectrogram.
Sensors 23 01790 g003
Figure 4. Sign language gesture spectrogram.
Sensors 23 01790 g004
Figure 5. Single gesture action data processing process. (a) Bandpass filtering data; (b) Gaussian smoothing data.
Sensors 23 01790 g005
Figure 6. Overview of the GAN.
Sensors 23 01790 g006
Figure 7. Overview of the ResNet34 and Bi-LSTM.
Sensors 23 01790 g007
Figure 8. Experimental equipment and environment. (a) Data collection equipment; (b) Laboratory environment; (c) Corridor environment.
Sensors 23 01790 g008
Figure 9. Impact of different distance, angles and speeds. (a) Impact of different distance; (b) impact of different angles; and (c) impact of different speeds.
Sensors 23 01790 g009
Figure 10. Impact of different environments and interference states. (a) Impact of different environments and (b) impact of different interference states.
Sensors 23 01790 g010
Figure 11. Impact on recognition performance of single gesture, continuous gesture and sign language gesture when data augmentation is used or not. (a) Single gesture; (b) continuous gesture; and (c) sign language gesture.
Sensors 23 01790 g011
Figure 12. Overall performance of single gestures.
Sensors 23 01790 g012
Figure 13. Impact of classification model on continuous gesture performance. (a) Spread and Pinch; (b) Push and Pull; (c) Hover and OK; (d) Around Left and Around Right; (e) One, Two, and Three; and (f) Slide Up, Slide Down, Slide Left, and Slide Right.
Sensors 23 01790 g013
Figure 14. Impact of different models on accuracy of continuous gestures.
Sensors 23 01790 g014
Figure 15. Impact of classification model on sign language gesture performance. (a) I am a teacher. (b) I am fine, thanks. (c) What day is today? (d) Sorry, I am late. (e) What do you do? (f) What is your name?
Sensors 23 01790 g015
Figure 16. Impact of different models on accuracy of sign language gestures.
Sensors 23 01790 g016
Table 1. Comparison with the word level methods.
ProjectSignalDevice FreeApplicationAlgorithmFeatureAccuracy
AudioGest [34]UltrasoundYesWhole-hand Gesture/Doppler Effect89.1%
SoundWave [35]UltrasoundYesWhole-hand GestureCNNDoppler Effect88.6%
UltraGesture [36]UltrasoundYesFinger-level GestureCNNCIR93.5%
Push [24]UltrasoundYesFinger-level GestureCNN+LSTMCIR95.3%
OursUltrasoundYesFinger-level GestureCNN+Bi-LSTMDoppler Effect98.8%
Table 2. Comparison with the sentence level methods.
ProjectSignalApplicationAlgorithmSingleContinuousSign Language
SonicASL [26]UltrasoundWord and SentenceCNN+LSTM+CTC93.8%/90.6%
OursUltrasoundWord and SentenceCNN+Bi-LSTM+CTC98.8%92.4%86.3%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Hao, Z.; Dang, X.; Zhang, Z.; Li, M. UltrasonicGS: A Highly Robust Gesture and Sign Language Recognition Method Based on Ultrasonic Signals. Sensors 2023, 23, 1790. https://doi.org/10.3390/s23041790

AMA Style

Wang Y, Hao Z, Dang X, Zhang Z, Li M. UltrasonicGS: A Highly Robust Gesture and Sign Language Recognition Method Based on Ultrasonic Signals. Sensors. 2023; 23(4):1790. https://doi.org/10.3390/s23041790

Chicago/Turabian Style

Wang, Yuejiao, Zhanjun Hao, Xiaochao Dang, Zhenyi Zhang, and Mengqiao Li. 2023. "UltrasonicGS: A Highly Robust Gesture and Sign Language Recognition Method Based on Ultrasonic Signals" Sensors 23, no. 4: 1790. https://doi.org/10.3390/s23041790

APA Style

Wang, Y., Hao, Z., Dang, X., Zhang, Z., & Li, M. (2023). UltrasonicGS: A Highly Robust Gesture and Sign Language Recognition Method Based on Ultrasonic Signals. Sensors, 23(4), 1790. https://doi.org/10.3390/s23041790

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Citations

Crossref
 
Scopus
 
Web of Science
 
ads
 
PMC
 
Google Scholar

Article Access Statistics

Created with Highcharts 4.0.4Chart context menuArticle access statisticsArticle Views11. Jan12. Jan13. Jan14. Jan15. Jan16. Jan17. Jan18. Jan19. Jan20. Jan21. Jan22. Jan23. Jan24. Jan25. Jan26. Jan27. Jan28. Jan29. Jan30. Jan31. Jan1. Feb2. Feb3. Feb4. Feb5. Feb6. Feb7. Feb8. Feb9. Feb10. Feb11. Feb12. Feb13. Feb14. Feb15. Feb16. Feb17. Feb18. Feb19. Feb20. Feb21. Feb22. Feb23. Feb24. Feb25. Feb26. Feb27. Feb28. Feb1. Mar2. Mar3. Mar4. Mar5. Mar6. Mar7. Mar8. Mar9. Mar10. Mar11. Mar12. Mar13. Mar14. Mar15. Mar16. Mar17. Mar18. Mar19. Mar20. Mar21. Mar22. Mar23. Mar24. Mar25. Mar26. Mar27. Mar28. Mar29. Mar30. Mar31. Mar1. Apr2. Apr3. Apr4. Apr5. Apr6. Apr7. Apr8. Apr9. Apr10. Apr0k1k2k3k4k
For more information on the journal statistics, click here.
Multiple requests from the same IP address are counted as one view.
Back to Top  返回顶部Top