Next Article in Journal
A Study of the Effect of Temperature on the Capacitance Characteristics of a Metal-μhemisphere Resonant Gyroscope
Previous Article in Journal
TEA-GCN: Transformer-Enhanced Adaptive Graph Convolutional Network for Traffic Flow Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lower Limb Motion Recognition Based on sEMG and CNN-TL Fusion Model

by
Zhiwei Zhou
1,
Qing Tao
1,*,
Na Su
1,2,
Jingxuan Liu
1,
Qingzheng Chen
1 and
Bowen Li
1
1
College of Intelligent Manufacturing Modern Industry, Xinjiang University, Urumqi 830017, China
2
The First Affiliated Hospital, Xinjiang Medical University, Urumqi 830017, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(21), 7087; https://doi.org/10.3390/s24217087
Submission received: 17 September 2024 / Revised: 22 October 2024 / Accepted: 30 October 2024 / Published: 4 November 2024
(This article belongs to the Section Sensor Networks)

Abstract

:
To enhance the classification accuracy of lower limb movements, a fusion recognition model integrating a surface electromyography (sEMG)-based convolutional neural network, transformer encoder, and long short-term memory network (CNN-Transformer-LSTM, CNN-TL) was proposed in this study. By combining these advanced techniques, significant improvements in movement classification were achieved. Firstly, sEMG data were collected from 20 subjects as they performed four distinct gait movements: walking upstairs, walking downstairs, walking on a level surface, and squatting. Subsequently, the gathered sEMG data underwent preprocessing, with features extracted from both the time domain and frequency domain. These features were then used as inputs for the machine learning recognition model. Finally, based on the preprocessed sEMG data, the CNN-TL lower limb action recognition model was constructed. The performance of CNN-TL was then compared with that of the CNN, LSTM, and SVM models. The results demonstrated that the accuracy of the CNN-TL model in lower limb action recognition was 3.76%, 5.92%, and 14.92% higher than that of the CNN-LSTM, CNN, and SVM models, respectively, thereby proving its superior classification performance. An effective scheme for improving lower limb motor function in rehabilitation and assistance devices was thus provided.

1. Introduction

The recognition of lower limb motion holds significant application value across various fields. In rehabilitation treatment, it allows doctors to objectively assess a patient’s physical condition, thereby optimizing the rehabilitation plan. Additionally, lower limb motion recognition facilitates the effective integration of the human body with wearable power-assisted exoskeleton robots, enhancing both weight-bearing capacity and work efficiency. It is also reported that identifying lower limb movements improves the quality of daily life in elderly, frail, and motor-impaired patients. Composed of superimposed action potentials generated by multiple motor units, sEMG contains biological information closely related to human behavior. The acquisition of sEMG signals, when compared to intramuscular EMG signal acquisition, offers significant advantages, including high safety, strong reliability, and non-invasiveness [1]. Widely used in rehabilitation robots, wearable exoskeleton control, and human–computer interactions, surface electromyography signals directly reflect muscle activity patterns.
The preprocessing research of surface electromyography (sEMG) and dynamic electroencephalogram (EEG) signals primarily focuses on feature extraction methods and classification algorithms, with the aim of achieving a high recognition rate for various human limb movement patterns [2]. For example, Totty et al. effectively classify daily life activities in the functional arm activity behavior observation system by adopting the K-nearest neighbor (KNN) algorithm combined with muscle activation and movement data [3]. Wang et al. analyze the shortcomings of the traditional threshold function in sEMG signal denoising and propose an improved threshold function. The classification performance of LSTM is then compared with that of CNN, SVM, and other classification algorithms [4]. Cui et al. combine the sEMG feature space construction method using time-domain and time–frequency-domain features and propose an SVM classifier based on sEMG, an AFSA-optimized SVM classifier, and a deep learning CNN classification model to recognize upper limb gestures [5]. Shi proposes a lower limb rehabilitation training mode based on surface electromyography (sEMG) signals. By extracting features from the sEMG signal and using a BP neural network to identify movement intention, the result is then used as the driving signal for the rehabilitation robot to facilitate active training [6]. Zheng et al. establish a mapping model between sEMG signals and motion angles by extracting features and utilizing a back propagation neural network. This approach predicts joint single-degree-of-freedom motions and combined motion angles with high accuracy [7]. Huang et al. propose an improved deep forest model for hand action recognition using sEMG data. By leveraging the complementarity between three classifier algorithms, they enhance the classification and recognition accuracy of the deep forest model, achieving an average recognition accuracy of 94% for 16 commonly used hand actions [8]. Ai et al. extract time-domain features and wavelet coefficients from surface EMG signals and utilize dynamic time warping (DTW) distance to extract features from acceleration signals. They then apply linear discriminant analysis (LDA) and SVM to classify five lower limb movements [9]. Liu et al. utilize the improved RelieF algorithm to select the optimal time-domain feature vector after data preprocessing and apply a support vector machine based on a balanced decision tree to layer the identification of five common dumbbell actions. This approach lays the foundation for personalized dumbbell action guidance [10]. Zhong et al. propose a multi-scale time-frequency information fusion representation method (MTFIFR) to obtain time-frequency features of multi-channel sEMG signals. They design a multi-feature fusion network (MFFN) and introduce a deep belief network (DBN) as the classification model for MFFN, aiming to improve the generalization performance for a broader range of upper limb movements [11]. Zhao et al. adopt a data-enhanced EFM as the input sample and integrate an ECA mechanism into the CNN architecture to assign higher weights to key feature information. Through modular design, they further adjust the layers of the deep feature extraction module within the network, effectively improving the accuracy of gesture recognition [12]. Xi et al. apply wavelet transform to process sEMG signals, calculate wavelet coherence coefficients, and use SVM to classify six types of daily activities [13]. Wei et al. proposed a multi-stream convolutional neural network (CNN) framework to improve the recognition accuracy of gestures by using a “divide and conquer” strategy to learn the correlation between a single muscle and a specific gesture [14]. Liu et al. propose a muscle fatigue classification method based on electromyography signals, which incorporates the crossover mutation of a genetic algorithm and the improved fruit fly optimization algorithm. This approach is combined with a neural network to identify muscle fatigue, enabling accurate detection and classification of muscle fatigue [15]. Sui et al. decompose the sEMG signal using wavelet packet transform (WPT) to extract wavelet packet coefficients and feed the calculated variance and energy as inputs to the improved SVM classifier, achieving an accuracy of 90.66% in recognizing six upper limb actions [16]. Liu et al. improve the accuracy of hand movement intention classification by enhancing the activity segment detection technology of sEMG. They employ a two-stage discriminative adaptive threshold technique to detect the active segment of sEMG, using the feature matrix corresponding to the motion intention and its label as the input and output of the LSTM hand motion intention classification model. This approach increases the average classification accuracy of six motion intentions to 91.7% [17]. Gupta et al. record dual-channel EMG signals from 18 subjects across three lower limb motion modes and evaluate the influence of window size, feature vector type, and classifier type on recognition performance. They find that selecting a 256 ms window size, 32 ms overlap, LDA classifier, and temporal feature vector yields excellent performance [18].
According to Refs. [19,20], current motion recognition technology primarily relies on EMG and EEG signals. The use of surface EMG signals for lower limb action recognition achieves satisfactory results under objective conditions. Existing research on lower limb action classification methods generally focuses on extracting sEMG features, which are then identified and classified using machine learning or deep learning algorithms to recognize lower limb actions. However, this approach to manual feature extraction addresses only a segment of the information, failing to encapsulate the complete characteristics of the input data and missing essential features. As a result, certain sEMG information is lost in the feature extraction process, which adversely affects the classification accuracy. To address this issue of information loss in traditional feature extraction, this study proposes a CNN-TL lower limb motion recognition model based on surface EMG signals. The model effectively captures spatio-temporal features in EMG signals, enhancing the recognition accuracy and robustness of lower limb actions, and provides a more reliable and efficient solution for lower limb motion rehabilitation.

2. Materials and Methods

2.1. Experimental Objects and Equipment

In this study, sEMG data from 20 healthy subjects are collected, with the basic body parameters displayed in Table 1. None of the subjects suffer from a fracture, sprain, muscle strain, or any other injury that could affect motor function prior to the start of the experiment. No strenuous exercise is performed one week before the experiment, effectively avoiding muscle soreness or discomfort. The PLUX wireless EMG acquisition device is used as the experimental equipment, with a sampling frequency of 1000 Hz selected. Additionally, silver chloride (AgCl) electromyography electrode sheets and 75% alcohol wipes are used.

2.2. Experiment Procedure

The experiment is conducted in the Intelligent Medical Rehabilitation Robotics Laboratory. To reduce impedance, the hair on the skin surface of the subject’s _target areas is removed before the experiment, and the area is wiped with 75% alcohol to remove surface oils. After the alcohol dries, electrodes are placed at the highest point of the _target muscle bulge, ensuring consistent spacing between electrodes. In this study, selected muscles for analysis include the rectus femoris, vastus lateralis, vastus medialis, semitendinosus, tibialis anterior, lateral gastrocnemius, and medial gastrocnemius. These muscles are involved in coordinated force generation during walking, stair ascent and descent, and squatting. By selecting these muscles, comprehensive capture of muscle activation patterns during common lower limb movements is achieved, providing support for _targeted rehabilitation planning, ensuring coordinated muscle development, and effectively guiding movement quality and functional recovery, as shown in Table 2:
As shown in Figure 1, subjects perform each of the four lower limb movements: ascending stairs, squatting, walking, and descending stairs. For the two lower limb movement modes of ascending and descending stairs, each subject performs 8 sets of each movement, with each set requiring 5 repetitions, and each movement cycle lasting 2 s. Each subject performs three sets of 20 squats, with each exercise cycle lasting 3 s. Five sets of walking exercises are performed, with each exercise cycle lasting 60 s, yielding 30 sets of data. A rest period of 3–5 min is taken between sets, and 10–15 min of rest is observed between different movements to avoid muscle fatigue affecting the results.

2.3. Data Preprocessing

During sEMG acquisition, environmental noise and other bioelectric signals, such as ECG signals, can easily interfere with the results, making noise reduction essential, as shown in Figure 2. The frequency range containing useful information in sEMG is primarily concentrated between 5 and 200 Hz. Therefore, in this study, a fourth−order Butterworth bandpass filter ranging from 30 to 300 Hz, along with a 50 Hz notch filter, is applied to filter the raw sEMG signals. Additionally, the db2 wavelet basis is used for 4−level wavelet decomposition, and the wavelet threshold technique is employed to further reduce noise.
The processed data are subsequently segmented according to the initial point of each action cycle, with the action cycles for walking and ascending and descending stairs set to 2 s, and the cycle for squatting set to 3 s. The overlapping window technique is subsequently employed to further divide the sEMG data within each action cycle. A sliding window of 1024 ms with a step size of 512 ms is utilized to break down the sEMG from each channel into serialized data windows. After expanding the data from 7 channels using this sliding time window, 8640 signal data samples, each of size 7 × 1024, are obtained. Figure 3 illustrates the process of segmenting EMG signals from a single channel using sliding window technology. After screening, the final number of samples obtained is as follows: 2160 samples for ascending stairs, 1920 samples for squatting, 1680 samples for walking, and 2160 samples for descending stairs.

2.4. Feature Extraction and Analysis

To extract effective information from the processed EMG data for use as feature input in the machine learning model, this study extracts both time-domain and frequency-domain features from the collected EMG signals and normalizes these features. In the time-domain analysis, key metrics such as the mean absolute value (MAV) and root mean square (RMS) of the EMG time series x i are considered. For frequency-domain analysis, mean power frequency (MPF) and median frequency (MF) are further extracted as essential frequency-domain features by analyzing the power spectral density function P ( f ) of the EMG signals. The calculation formulas are presented in Table 3.

2.5. Lower Limb Action Recognition Model

2.5.1. CNN

sEMG is considered a mixed signal, originating from the temporal and spatial superposition of multiple muscle activities, leading to its high complexity [21]. Traditional convolutional neural networks, primarily used for image processing, struggle to effectively capture features in one-dimensional time series data. In contrast, a one-dimensional convolutional neural network (1DCNN) directly handles time series data, allowing the model to have fewer parameters while effectively learning features from these sequences. Therefore, this paper selects 1 1DCNN as the model to process sEMG, providing robust support for action recognition. The model architecture is illustrated in Figure 4.
In this paper, a 1DCNN model is employed, consisting of three convolutional layers, pooling layers, batch normalization (BN) layers, three fully connected layers, and a Dropout layer. The input signal in the convolutional layer is processed through one-dimensional convolution and calculated as follows:
h j l = i = 1 n x i l k i j l + b j l
In the formula, l represents the number of layers, j denotes the ordinal number of elements, n is the length of the input feature vector, h is the output feature vector, x represents the input feature vector, k is the convolution kernel, and b is the bias vector.
To enhance the training efficiency and stability of the model, batch normalization (BN) is applied after each convolutional layer. The calculation process is as follows:
First, the mean and variance of the same channels are calculated along the batch dimension, with the calculation expressed by the following formula:
μ ß = 1 n i = 1 n x i
σ ß 2 = 1 n i = 1 n ( x i μ ß ) 2
Then, the mean and variance along the same batch dimension are normalized, with the calculation expressed by the following formula:
x ^ i = x i μ ß σ ß 2 + ε
Next, any necessary scaling and shifting to restore the eigenvalue are performed by learning two parameters, γ (scaling factor) and β (translation factor), calculated as follows:
y i = γ x ^ i + β BN γ , β ( x i ) y
Finally, the ReLU activation function is applied after each BN, with the calculation expressed by the following formula:
R e L U ( x ) = max ( 0 , x ) = x x 0 0 x < 0
To improve the model training speed, Max pooling is used for downsampling. During the training phase, the Cross Entropy Loss function is employed to evaluate the deviation between the probability distribution of the model output and the true label, calculated as follows:
L ( θ ) = 1 n i = 1 n k = 1 K I { y i = k } log exp ( θ K T x ) j = 1 K exp ( θ j T x )
Additionally, to optimize the learning process and minimize the loss function, the Adadelta optimizer is adopted in this study to automatically adjust the learning rate. Two Dropout layers are inserted between the three fully connected layers to enhance the model’s generalization ability and prevent overfitting. The network model configuration is shown in Table 4.

2.5.2. LSTM

The long short-term memory (LSTM) network is a specialized recurrent neural network. By introducing a gating unit to control data flow, it effectively addresses the issues of gradient vanishing and gradient explosion commonly encountered by traditional recurrent neural networks when processing long sequence data [22]. EMG signals often contain complex temporal dynamics that may exhibit correlations over extended time scales, and the characteristic architecture of LSTMs enables them to learn these long-term dependencies. The structure of the LSTM network is shown in Figure 5.
The working mechanism of LSTM includes processing the current input X t and the hidden state H t 1 from the previous time step through the sigmoid function to generate the forgetting gate output F t , which varies between 0 and 1. An output of 0 indicates “completely forgotten”, while a value of 1 indicates “fully retained”. The calculation formula is as follows:
F t = σ ( X t W x f + H t 1 W h f + b f )
In the formula, W f represents the weight, σ is the sigmoid activation function, and b f denotes the bias.
The input gate plays a crucial role in updating the cell state. First, the portion of information that needs to be updated is determined by the sigmoid function, generating the output I t of the input gate. Next, a new candidate value vector C t ~ is generated through the tanh layer, representing new information that may be added to the state. Finally, the output of the forget gate F t is multiplied with the previous cell state C t 1 to discard unnecessary information, while the output of the input gate I t is multiplied with the candidate value C t ~ to update the new cell state C t . The formula is given as follows:
I t = σ ( X t W x i + H t 1 W h t + b i )
C ˜ t = tanh ( X t W x c + H t 1 W h c + b c )
C t = F t C t 1 + I t C ˜ t
In the formula, C t ~ represents candidate vectors, F t C t 1 denotes the selective forgetting of irrelevant information, and I t C t ~ signifies the retention of useful information.
In LSTM, the output gate processes the cell state through a sigmoid layer to determine what should be output. The cell state is then adjusted through a tanh layer and multiplied by the output of the sigmoid layer to produce the final output H t . The formula is as follows:
O t = σ ( X t W x o + H t 1 W h o + b o )
H t = O t tanh ( C t )
In the formula, O t represents the activation vector of the output gate.
The parameter configuration of the LSTM network used in this study is presented in Table 5.

2.5.3. Transformer

The transformer structure includes an encoder and a decoder, with the encoder used for classification [23]. It primarily utilizes core components such as the multi-head attention mechanism, positional encoding, and feed-forward networks to capture long-term dependencies and complex signal patterns. The complete structure is illustrated in Figure 6.
(1) Positional encoding: In the transformer network, since the self-attention mechanism is used to extract information without a recursive structure, positional encodings are added to provide the model with information about the signal order. The formula is as follows:
P E ( p o s , 2 i ) = sin ( p o s / 10000 2 i / d )
P E ( p o s , 2 i + 1 ) = cos ( p o s / 10000 2 i / d )
(2) The self-attention mechanism, the core of the transformer network, enables the model to weight each element based on its relationship to other elements in the sequence when processing sequential data. The formula is given as follows:
A t t e n t i o n Q , K , V = S o f t m a x Q K T / D K V
In the formula, Q is the query matrix, K is the key matrix, V is the value matrix, and D K represents the matrix dimension. Transformer converts single-head self-attention to multi-head self-attention through concatenation, with the calculation expressed by the following formula:
M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 ; ; h e a d h ) W o
h e a d i = S o f t max ( Q W i Q ( K W i K ) T d k ) V W i V
In the formula, M u l t i H e a d represents the multi-head attention mechanism function, and h e a d i denotes the output of the ith attention head.
(3) Layer normalization and residual connections: Layer normalization stabilizes training by standardizing input features [24]. Residual connections facilitate the direct flow of gradients through the network, helping to prevent the vanishing or exploding gradient problem in deep networks. The calculation formula is as follows:
X o u t = L a y e r N o r m = x i j μ j / σ j 2 + ε
In the formula, μ j represents the mean, and σ j 2 denotes the variance.

2.5.4. CLT Model

Building on the above analysis, this study constructs a CNN-Transformer-LSTM (CNN-TL) lower limb motion recognition model based on sEMG signals, with its overall framework illustrated in Figure 7. First, the preprocessed sEMG input is sent to the CNN layer for feature extraction. After passing through two fully connected layers, the data are successively processed by the Transformer and LSTM layers. Finally, the fully connected layer generates the final classification decision. For multi-channel EMG signals, CNN analyzes the interactions between different muscles and their spatial layout. The transformer effectively captures long-range dependencies within sequences through its self-attention mechanism, while LSTM handles time series data and retains long-term dependencies. By integrating these three models, sEMG signals are identified efficiently and accurately.

3. Results

3.1. Model Evaluation Index

In evaluating classification task performance, this paper adopts four core statistical measures as evaluation criteria: accuracy, precision, recall, and F1 score [25]. These metrics comprehensively reflect multiple dimensions of the classification model’s performance, providing a basis for the overall evaluation of the model. The calculation formulas are as follows:
A c c u r a r y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 S c o r e = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n
In the formula, T N represents the count of negative samples correctly identified as the negative class, T P represents the count of positive samples accurately classified as the positive class, F N denotes the number of cases where a positive sample is incorrectly classified as the negative class, and F P denotes the number of instances where a negative sample is incorrectly classified as the positive class.

3.2. Analysis of Experimental Results

During 300 rounds of training on the CNN-TL model, changes in the model’s loss value are tracked, and the fitting quality is evaluated by analyzing the training and test loss curves, as shown in Figure 8. The training results indicate that the model’s loss value is reduced to just 1.7%, while the recognition accuracy reaches 96.13%, further confirming the excellent convergence of the CNN-TL model.
Figure 9 presents a comparative analysis of the results from the proposed four classification model recognition schemes. For the nonlinear and multiclass challenges of sEMG feature vectors, a one-vs-all classification method is employed. Specifically, multiple four-class support vector machine (SVM) models are trained, with each model designating one class as the positive class and treating the others as negative classes. In the SVM models, the kernel selected is the Gaussian kernel, with the penalty parameter C set to 1 and the γ (gamma) value set to 0.1. The data reveal that the CNN-TL model demonstrates remarkable performance across various indicators. Specifically, the model achieves an accuracy of 96.13%, a precision of 95.71%, a recall of 95.60%, and an F1 score of 95.65%. Notably, the accuracy of the CNN-TL model is 3.76% higher than that of the CNN-LSTM model, 5.92% higher than that of the CNN model, and 14.92% higher than that of the SVM model. Compared to the CNN-LSTM model, the CNN-TL model’s precision, recall, and F1 score increase by 4.61%, 3.37%, and 3.99%, respectively. When compared to the CNN model, these increases are 6.77%, 8.64%, and 7.71%, respectively. Compared to the SVM model, the improvements are 13.33%, 16.63%, and 15.01%, respectively. These results demonstrate that the CNN-TL model exhibits higher adaptability and superiority in lower limb movement recognition than the CNN-LSTM, CNN, and SVM models.
To evaluate the performance of various lower limb motion recognition models, CNN-TL, CNN-LSTM, CNN, SVM, and other models are tested using the same dataset. When dealing with different lower limb movements, the probability of the test sample being classified into a specific category is first calculated. Based on these probabilities, the ratio of false positives and true positives under each threshold condition are determined, and the corresponding ROC curve is generated. For each lower limb movement, ROC curves are generated in each test set. By averaging the ROC curves generated by each action recognition model, comprehensive ROC curves for each model in the lower limb action recognition task are obtained, as shown in Figure 10.
As shown in Figure 11, the confusion matrix of the four different models displays the classification results for four lower limb movements, clearly demonstrating that the proposed CNN-TL model accurately identifies various lower limb movements based on the surface EMG signals from the seven channels of the lower limb. In the confusion matrix, the diagonal elements represent the quantity of samples correctly classified for each action, while the non-diagonal elements indicate the quantity of samples incorrectly classified. The categories are defined as follows: 0 for walking, 1 for ascending stairs, 2 for descending stairs, and 3 for squatting.

3.3. Ablation Experiment

In this section, to validate the effectiveness of the CNN-TL model, the influence of various components on model performance is examined. During the training process, the convolutional neural network (CNN) serves as the benchmark model to explore the effectiveness of the CNN-TL model and the specific influence of each component on its performance. The models are decomposed into CNN, CNN-LSTM, and CNN-Transformer for comparative analysis. The experimental results are presented in Table 6.
Analyzing the experimental data in Table 6 reveals that, compared to standalone CNN and LSTM models, the CNN-LSTM configuration exhibits superior performance. This improvement is attributed to the synergy between CNN and LSTM, enabling not only the extraction of spatio-temporal features but also the capture of dynamic data changes along the time dimension. The LSTM-Transformer fusion model combines LSTM’s proficiency in processing time-series data with transformer’s capability to capture long-range dependencies. Owing to its self-attention mechanism, the CNN-Transformer model demonstrates a natural aptitude for handling sequences of varying lengths, markedly surpassing the classification efficacy of singular models. Notably, the CNN-TL model achieves a classification accuracy of 96.13%, outperforming the other configurations. This superiority arises from the introduction of the Transformer, which effectively addresses long-term dependency challenges, while the LSTM component enhances the model’s comprehension of short-term dependencies in time series. This amalgamation not only optimizes the model’s processing of sequential data but also augments its ability to discern intrinsic data patterns from multiple dimensions, thereby enhancing its generalization performance on novel, previously unseen data.

4. Conclusions

Due to the CNN model’s restricted capacity to capture global information from extended series of EMG signals, the CNN-TL model is proposed by integrating Transformer and LSTM. This combined model effectively captures global information and sequence relationships, addressing the issue of missing local feature information in the CNN model. In this investigation, four distinct motion recognition models were developed and assessed to classify four specific movements during lower limb rehabilitation exercises. From the sEMG data collected from 20 participants, four key features—MAV, RMS, MPF, and MDF—were extracted and applied as inputs to the machine learning models. The preprocessed sEMG data were then used for the learning and training of the CNN, LSTM, CNN-LSTM, CNN-Transformer, and CNN-TL models. This study compares the performance of each model on the lower limb motion recognition task. The experimental results indicate that the CNN-TL model achieves an average accuracy of 96.13% on the test data.
The CNN-TL model proposed in this study demonstrates significant advantages over other models, confirming its excellent performance in the field of lower limb motion recognition. In future reseaarch, the study plans to expand the participant group from healthy individuals to patients with impaired lower limb function, further standardizing the EMG acquisition process to enhance the action recognition accuracy of the lower limb rehabilitation robot based on sEMG across different patient groups. Additionally, different training models will be explored and refined to optimize action recognition performance.

Author Contributions

Individual contributions were distributed as follows: conceptualization: Z.Z. and Q.T.; methodology: Z.Z. and J.L.; data collection: Z.Z, Q.C., N.S., J.L. and B.L.; software: Z.Z.; project administration: Q.T.; review and editing: Z.Z. and Q.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (52365039); Tianshan Talent Training Program (2023TSYCLJ0051).

Institutional Review Board Statement

All experimental procedures involved in this study were evaluated and approved by the Ethics Committee of the School of Mechanical Engineering, Xinjiang University.

Informed Consent Statement

Written informed consent has been obtained from the participants involved in this study.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

We would like to thank Qing Tao, Jingxuan Liu, Na Su and Qingzheng Chen for their help in guiding the project.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, C.; Xi, X.G.; Chen, S.J.; Miran, S.M.; Hua, X.; Luo, Z.Z. SEMG-based multifeatures and predictive model for knee-joint-angle estimation. AIP Adv. 2019, 9, 095042. [Google Scholar] [CrossRef]
  2. Simão, M.; Mendes, N.; Gibaru, O.; Neto, P. A review on electromyography decoding and pattern recognition for human-machine interaction. IEEE Access 2019, 7, 39564–39582. [Google Scholar] [CrossRef]
  3. Totty, M.S.; Wade, E. Muscle activation and inertial motion data for noninvasive classification of activities of daily living. IEEE Trans. Biomed. Eng. 2018, 65, 1069–1076. [Google Scholar] [PubMed]
  4. Wang, J.; Sun, S.; Sun, Y. A Muscle Fatigue Classification Model Based on LSTM and Improved Wavelet Packet Threshold. Sensors 2021, 21, 6369. [Google Scholar] [CrossRef]
  5. Cui, B.; Deng, J.; Zhang, X. Hand Gesture Recognition Based on Surface Electromyography. Sci. Technol. Eng. 2023, 23, 15133–15141. [Google Scholar]
  6. Shi, X.H.; Lu, H.; Liao, Z.Y. Active training of lower limb rehabilitation based on surface electromyography signal. Sci. Technol. Eng. 2018, 18, 61–66. [Google Scholar]
  7. Zheng, Y.; Zheng, G.; Zhang, H.; Zhao, B.; Sun, P. Mapping Method of Human Arm Motion Based on Surface Electromyography Signals. Sensors 2024, 24, 2827. [Google Scholar] [CrossRef]
  8. Huang, S.; Mao, J. Recognition method of sEMG gesture based on improved deep forest. J. Shanghai Univ. Eng. Sci. 2023, 37, 190–197. [Google Scholar]
  9. Ai, Q.; Zhang, Y.; Qi, W.; Liu, Q.; Chen, K. Research on lower limb motion recognition based on fusion of sEMG and accelerometer signals. Symmetry 2017, 9, 147. [Google Scholar] [CrossRef]
  10. Liu, G.P.; Wang, N.X.; Zhou, Y.; Wang, W.; Tang, M. Recognition of Dumbbell Movements Based on an Improved ReliefF Algorithm. Sci. Technol. Eng. 2019, 19, 219–224. [Google Scholar]
  11. Zhong, T.; Li, D.; Wang, J.; Xu, J.; An, Z.; Zhu, Y. Fusion Learning for sEMG Recognition of Multiple Upper-Limb Rehabilitation Movements. Sensors 2021, 21, 5385. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, S.H.; Zhou, J.H.; Fu, Y.F. Investigation of gesture recognition using attention mechanism CNN combined electromyography feature matrix. J. Electron. Meas. Instrum. 2023, 37, 59–67. [Google Scholar]
  13. Xi, X.G.; Yang, C.; Shi, J.H.; Luo, Z.Z.; Zhao, Y.B. Surface electromyography-based daily activity recognition using wavelet coherence coefficient and support vector machine. Neural Process. Lett. 2019, 50, 2265–2280. [Google Scholar] [CrossRef]
  14. Wei, W.; Wong, Y.; Du, Y.; Hu, Y.; Kankanhalli, M.; Wang, G. A multi-stream convolutional neural network for sEMG-based gesture recognition in muscle-computer interface. Pattern Recognit. Lett. 2019, 119, 131–138. [Google Scholar] [CrossRef]
  15. Liu, G.D.; Xu, L.Y.; Xiao, R.L.; Sun, J.; Cai, J.; Zhang, S. Muscle Fatigue Classification Method of sEMG Signal Based on FGNN Algorithm. Sci. Technol. Eng. 2022, 22, 8370–8377. [Google Scholar]
  16. Sui, X.W.; Wan, K.X.; Zhang, Y. Pattern recognition of sEMG based on wavelet packet transform and improved SVM. Optik 2019, 176, 228–235. [Google Scholar] [CrossRef]
  17. Liu, S.Z.; Xu, D.Z. A New Method for Hand Motion Intent Recognition Based on Improved sEMG Signal Active Segment Detection. J. Jiaxing Univ. 2022, 34, 105–112. [Google Scholar]
  18. Gupta, R.; Agarwal, R. Electromyographic signal-driven continuous locomotion mode identification module design for lower limb prosthesis control. Arabian J. Sci. Eng. 2018, 43, 7817–7835. [Google Scholar] [CrossRef]
  19. Reaz, M.; Hussain, M.; Mohd-Yasin, F. Techniques of EMG signal analysis: Detection, processing, classification and applications (Correction). Biol. Proced. Online 2006, 8, 11–35. [Google Scholar] [CrossRef]
  20. Xu, H.; Xiong, A. Advances and disturbances in sEMG-based intentions and movements recognition: A review. IEEE Sens. J. 2021, 21, 13019–13028. [Google Scholar] [CrossRef]
  21. Chang, J.; Wang, Z.J.; Mckeown, M.A. Hidden Markov, multivariate autoregressive (HMM-mAR) network framework for analysis of surface EMG (sEMG) data. IEEE Trans. Signal Process. 2008, 56, 4069–4081. [Google Scholar] [CrossRef]
  22. Wang, R.; Yang, J.R. Study on the applicability of vulnerability analysis based on long short-term memory neural network. Sichuan Build. Sci. 2024, 50, 9–15. [Google Scholar]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  24. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  25. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In Proceedings of the Advances in Information Retrieval: 27th European Conference on IR Research, ECIR 2005, Santiago de Compostela, Spain, 21–23 March 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
Figure 1. Diagram of the experimental scenario.
Figure 1. Diagram of the experimental scenario.
Sensors 24 07087 g001
Figure 2. Elimination of extraneous information signals from sEMG.
Figure 2. Elimination of extraneous information signals from sEMG.
Sensors 24 07087 g002
Figure 3. Sliding window segmentation diagram of single-channel sEMG signal.
Figure 3. Sliding window segmentation diagram of single-channel sEMG signal.
Sensors 24 07087 g003
Figure 4. 1DCNN model structure.
Figure 4. 1DCNN model structure.
Sensors 24 07087 g004
Figure 5. LSTM model structure.
Figure 5. LSTM model structure.
Sensors 24 07087 g005
Figure 6. Transformer encoder model structure.
Figure 6. Transformer encoder model structure.
Sensors 24 07087 g006
Figure 7. CNN-TL overall architecture.
Figure 7. CNN-TL overall architecture.
Sensors 24 07087 g007
Figure 8. The loss curve of the CNN-TL model.
Figure 8. The loss curve of the CNN-TL model.
Sensors 24 07087 g008
Figure 9. Comparison of the performance of the four classification models.
Figure 9. Comparison of the performance of the four classification models.
Sensors 24 07087 g009
Figure 10. ROC curves for four classification models.
Figure 10. ROC curves for four classification models.
Sensors 24 07087 g010
Figure 11. Confusion matrices for four models.
Figure 11. Confusion matrices for four models.
Sensors 24 07087 g011aSensors 24 07087 g011b
Table 1. Subjects’ basic information.
Table 1. Subjects’ basic information.
PeopleAgeHeight/(cm)Weight/(kg)
2022~28160~18960~85
Table 2. Muscle, electrode placement, recording channels.
Table 2. Muscle, electrode placement, recording channels.
Muscle NamePatch PositionAcquisition Channel
Rectus femorisSensors 24 07087 i001EMG-ch1
Vastus lateralisEMG-ch2
Vastus medialisEMG-ch3
Semitendinosus muscleEMG-ch4
Tibialis anterior EMG-ch5
Gastrocnemius lateralisEMG-ch6
Medial gastrocnemiusEMG-ch7
Table 3. Feature calculation formula.
Table 3. Feature calculation formula.
Feature Value NameFormula of Calculation
Time-Domain FeaturesMAV T M A V = 1 N i = 1 N x i
RMS T R M S = 1 N i = 1 N x i 2
Frequency-
Domain Features
MPF F M P F = 0 + f P ( f ) d f / 0 + P ( f ) d f
MF 0 F M F P f d f = 1 2 0 + P ( f ) d f
Table 4. 1DCNN model parameter settings.
Table 4. 1DCNN model parameter settings.
Network LayerKernel SizeKernels NumberOutput
Convolutional layer 13 × 16464 × 1020
BN1+ReLU64-64 × 1020
Max-Pool 12 × 26464 × 510
Convolutional layer 23 × 13232 × 508
BN2+ReLU32-32 × 508
Max-Pool 22 × 23232 × 254
Convolutional layer 33 × 11010 × 252
BN3+ReLU10-10 × 252
Max-Pool 32 × 21010 × 126
Fully connected layer 1126011260 × 1
Fully connected layer 26001600 × 1
Fully connected layer 21001100 × 1
Table 5. LSTM model parameter settings.
Table 5. LSTM model parameter settings.
HyperparametersValue
loss functionCross Entropy
optimizerAdadelta
layers3
LSTM unit100
batch size40
learning rate0.001
Table 6. Comparison of model results (%).
Table 6. Comparison of model results (%).
ModelAccPreRecF1
SVM81.2182.3878.9780.64
CNN90.2188.9486.9687.94
LSTM84.0386.6482.3784.45
CNN-LSTM92.3791.1092.2391.66
CNN-Transformer92.0691.2891.1491.26
CNN-TL96.1395.7195.6095.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Z.; Tao, Q.; Su, N.; Liu, J.; Chen, Q.; Li, B. Lower Limb Motion Recognition Based on sEMG and CNN-TL Fusion Model. Sensors 2024, 24, 7087. https://doi.org/10.3390/s24217087

AMA Style

Zhou Z, Tao Q, Su N, Liu J, Chen Q, Li B. Lower Limb Motion Recognition Based on sEMG and CNN-TL Fusion Model. Sensors. 2024; 24(21):7087. https://doi.org/10.3390/s24217087

Chicago/Turabian Style

Zhou, Zhiwei, Qing Tao, Na Su, Jingxuan Liu, Qingzheng Chen, and Bowen Li. 2024. "Lower Limb Motion Recognition Based on sEMG and CNN-TL Fusion Model" Sensors 24, no. 21: 7087. https://doi.org/10.3390/s24217087

APA Style

Zhou, Z., Tao, Q., Su, N., Liu, J., Chen, Q., & Li, B. (2024). Lower Limb Motion Recognition Based on sEMG and CNN-TL Fusion Model. Sensors, 24(21), 7087. https://doi.org/10.3390/s24217087

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop
  NODES
admin 3
Association 4
COMMUNITY 1
Idea 2
idea 2
innovation 4
INTERN 31
Note 16
Project 3
twitter 1