
Citation: Vikas Kumar, Nitu Kumari. Controlling chaos in three species food chain model with fear effect[J]. AIMS Mathematics, 2020, 5(2): 828-842. doi: 10.3934/math.2020056
[1] | Wenbo Yang, Wei Liu, Qun Gao . Prediction of dissolved oxygen concentration in aquaculture based on attention mechanism and combined neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 998-1017. doi: 10.3934/mbe.2023046 |
[2] | Chongyi Tian, Longlong Lin, Yi Yan, Ruiqi Wang, Fan Wang, Qingqing Chi . Photovoltaic power prediction based on dilated causal convolutional network and stacked LSTM. Mathematical Biosciences and Engineering, 2024, 21(1): 1167-1185. doi: 10.3934/mbe.2024049 |
[3] | Mohanad A. Deif, Ahmed A. A. Solyman, Mehrdad Ahmadi Kamarposhti, Shahab S. Band, Rania E. Hammam . A deep bidirectional recurrent neural network for identification of SARS-CoV-2 from viral genome sequences. Mathematical Biosciences and Engineering, 2021, 18(6): 8933-8950. doi: 10.3934/mbe.2021440 |
[4] | Xiaotong Ji, Dan Liu, Ping Xiong . Multi-model fusion short-term power load forecasting based on improved WOA optimization. Mathematical Biosciences and Engineering, 2022, 19(12): 13399-13420. doi: 10.3934/mbe.2022627 |
[5] | Hui Li, Xintang Liu, Dongbao Jia, Yanyan Chen, Pengfei Hou, Haining Li . Research on chest radiography recognition model based on deep learning. Mathematical Biosciences and Engineering, 2022, 19(11): 11768-11781. doi: 10.3934/mbe.2022548 |
[6] | Zhangjie Wu, Minming Gu . A novel attention-guided ECA-CNN architecture for sEMG-based gait classification. Mathematical Biosciences and Engineering, 2023, 20(4): 7140-7153. doi: 10.3934/mbe.2023308 |
[7] | Sakorn Mekruksavanich, Wikanda Phaphan, Anuchit Jitpattanakul . Epileptic seizure detection in EEG signals via an enhanced hybrid CNN with an integrated attention mechanism. Mathematical Biosciences and Engineering, 2025, 22(1): 73-105. doi: 10.3934/mbe.2025004 |
[8] | Liguo Zhang, Liangyu Zhao, Yongtao Yan . A hybrid neural network-based intelligent body posture estimation system in sports scenes. Mathematical Biosciences and Engineering, 2024, 21(1): 1017-1037. doi: 10.3934/mbe.2024042 |
[9] | Jing Wang, Jiaohua Qin, Xuyu Xiang, Yun Tan, Nan Pan . CAPTCHA recognition based on deep convolutional neural network. Mathematical Biosciences and Engineering, 2019, 16(5): 5851-5861. doi: 10.3934/mbe.2019292 |
[10] | Tao Yang, Qicheng Yang, Yibo Zhou, Chuanbiao Wen . Glucose trend prediction model based on improved wavelet transform and gated recurrent unit. Mathematical Biosciences and Engineering, 2023, 20(9): 17037-17056. doi: 10.3934/mbe.2023760 |
Handwriting is a complex and essential communication skill for humans that is often disregarded. Nonetheless, handwriting (both manuscript and cursive) is a critical motor skill for children's cognitive development and academic success [1]. Considering as an example the case of a musical composition, experts in this area show a preference for handwriting editions over the use of specialized, user-friendly interfaces. The latter due to the lack of confidence in using a computer for artistic applications. Despite this fact, the automatic detection of hand-written musical notation in digital media is highly recommended and profitable in such an application, similar to characters recognition in text analysis. Calvo-Zaragoza and Oncina proposed a music notation recognition method based on an electronic pen and a finite-state machine to detect user strokes competitively [2].
In hand gesture analysis, Mendes et al. studied the interaction human and robot based on gesture recognition by developing an Electromyography sensor to capture and discriminate valid gestures using a Deep Learning technique [3].
Analogously, Słapekand and Paskiel proposed a method based on Bèzier curves to extract predefined hand gestures without using a marker at the start or end in the sequence of points [4]. Such type of methods tries to detect the handwriting style to develop improved methodologies in smart human-computer interfaces and handwriting recognition.
On the other hand, the development of brain-computer interfaces offers an alternative for clinical and assisted applications involving motor biosignals. For instance, the communication abilities of people with or without disabilities can be enhanced through devices interfaced with Electroencephalography (EEG) and/or Electromyography (EMG) signals.
There are two kinds of EMG [5]: Surface EMG and intramuscular EMG. Intramuscular EMG is a technique that is performed by inserting electrodes into a skeletal muscle. In contrast, surface EMG is performed by placing electrodes on the skin surface above a muscle. Compared to intramuscular EMG, surface EMG has the advantage of being a non invasive technique which is clinically superior in the diagnosis, rehabilitation, and testing of new therapies linked to motor activity. Surface EMG (sEMG) signals form an intricate interference pattern of bioelectrical signals generated mainly from the dynamical activity of skeletal muscles. As a consequence, sEMG signals from individual muscles cannot be reliably distinguished.
Additionally, sEMG has found several new applications supported by Machine Learning and Deep Learning methods as movement analysis, muscular fatigue, sports therapies, wearable devices, and particularly in handwriting recognition [6]. sEMG data acquisition and the corresponding waveforms are affected by many factors such as the muscles' physiological properties, added noise in the instrumentation used for sensing, as well as the chosen location of the electrodes related to the active muscle units. Hence, conventional features extraction methods involving sEMG generally require robust pre-processing steps and filtering. Moreover, the sEMG signal complexity can also be explained by the fact that there are various ways in which a motor task can be performed (motor redundancy) [7,8]. Simao et al. made a recent review of the sEMG-based systems using pattern recognition techniques and describing sensitive processes as signal acquisition and filtering. Their study focused on new human-machine interaction modalities [9].
Recently, advances in biomedicine have made possible the invention of myoelectric prostheses capable of performing real arm or leg movements [10]. Neural prostheses strive to restore the mobility of limbs, and the ability to communicate of subjects with disabilities, through the interconnection of brain potentials, actuators, or artificial devices in functional electrical stimulation. A significant innovation was the implementation of a device to predict a premature birth several weeks in advance. In this context, bioelectric interfaces allow extracting handwriting patterns for diagnostic purposes in patients with Parkinson's disease, and more recently, dysgraphia disability. Therefore, sEMG signals generated during the handwriting can be used to characterize neurological disorders [11]. In handwriting, the sEMG signals are generated and captured from the hand and forearm muscles during the handwriting activity, to be decoded by the specific purpose recognition systems [12]. In early studies, Linderman et al. developed a reliable system using sEMG signals to recognize handwriting characters to diagnose or measure disease progression in medical treatments [13].
In general, handwriting recognition from sEMG signals is addressed using the following approaches: Template Matching [13], Dynamic Time Warping (DTW) [12,14], Radial Basis Functions (RBF) Neural Networks [11]. It is noteworthy that sEMG signals convey a great amount of information about limb movements. Following this reasoning, several applications have been proposed in the literature for the recognition, modeling, and reconstruction of handwriting movements. Additionally, the reconstruction of handwriting from sEMG signals has been addressed using the Wiener filters [13], Recursive Least Squares [15], RBF neural networks [16], and Kalman filters [17]. Therefore, these advances are not only limited to medical purposes, but also toys and video games can be controlled through the musculature’s bioelectric activity. In a related work, Wei et al. devised a gesture recognition method by using multi-view features for both unimodal and multimodal sEMG-based gestures driven by a Deep convolutional neural network (DCNN) model [18].
Dash et al. [19] designed an approach called AirScript, for visualizing and recognizing characters written in the air. In this case, IMU (Inertial Measurement Unit) data from the Myo armband were applied to obtain a 2D coordinates sequence to visualize a handwritten digit as an image. For handwritten digit recognition, a convolutional neural network (CNN), and two GRU (Gated Recurrent Unit) neural networks were analyzed. The images were used as inputs to the CNN, the generated sequence of 2D coordinates was utilized as input to the first GRU neural network, and the raw IMU signal as input to the second GRU neural network. Subsequently, the outputs of those three neural networks were fused to yield the class of the recognized digit. The system reached an accuracy of 91.7 and 96.7% for independent and dependent tasks, respectively. In the method implemented by Roy et al. [20], single stroke digits were written in the air using a marker of uniform color. A video camera recorded the handwriting trajectory. After segmenting each video frame, the marker tip position was used to approximate the handwriting trajectory and to obtain an image of the handwritten digit. Next, a CNN was used for recognizing the image of each handwritten digit reporting an accuracy of 97.7, 95.4 and 93.7% for the English, Bengali, and Devanagari digits, respectively.
In some studies, the recognition of handwriting characters was limited to use small databases, imposing several restrictions on the placement of the electrodes, and in the number and complexity of handwriting strokes. Most handwriting recognition methods using sEMG signals were mainly based on pre-processing and feature extraction [13], while some others use the DTW method after pre-processing [14].
The approach discussed in this paper addreses handwriting character recognition problem from a more general point of view. To cope with the variability of the sEMG signals, the use of CNNs is suggested for the signal processing and feature extraction [21,22]. Hence, a Recurrent Neural Network (RNN) based on the Long Short-Term Memory (LSTM) and GRU networks is used to recognize sEMG signals, considering that LSTMs have also been previously applied successfully to gesture recognition [23]. Implemented CNN architectures are based on several studies demonstrating the efficacy and robustness of CNNs for extracting discriminant features directly from raw data. CNNs use a multi-level representation to learn and decode original data from specialized layers applied to an optimal recognition task [24].
A convolutional neural network (CNN), is a well-known Deep Learning architecture originally inspired by findings in visual perception. In 1959, Hubel & Wiesel discovered the existence of simple cells in the visual cortex of cats [25]. Simple cells are neurons that respond to bars or edges directed to a specific orientation. Hubel & Wiesel also recommended a hierarchical model to explain the behavior of simple cells. According to Marr [26], the receptive fields of simple cells are either bar o edge-shaped, and they compute a linear convolution with a bar or edge-shaped mask. Inspired on these findings, Fukushima suggested the Neocognitron in 1980 [27], which perhaps constitutes one of the first CNN architectures. The Neocognitron introduced two basic layer types: Convolutional layer and subsampling layer. Weng et al. proposed the Cresceptron architecture, which introduced max-pooling [28] and a variant called blurring [29].
In 1990, LeCun et al. created the LeNet architecture to classify handwritten digits [30]. The LeNet architecture and variants were the first to be trained with the backpropagation algorithm [31]. This architecture obtained effective representations directly from the image, which made it possible to recognize visual patterns from raw pixels with little-to-none pre-processing. At the same time, Zhang et al. used a Shift-Invariant Artificial Neural Network (SIANN) to recognize characters in an image [32].
In the beginning, despite the success of the LeNet-5 for digit recognition, CNNs were not successful for more complex tasks. The main reasons included the low computing power available, the small number of training samples, and diverse problems that arose especially for Deep CNNs (DCNN) [33]. After several efforts to surmount the difficulties for training DCNNs and thanks to the availability of new powerful computing resources, Krizhevsky et al. presented the AlexNet architecture, which won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC2012) [34]. The overall AlexNet architecture [35] is basically the same as the LeNet-5 architecture but considerably deeper, with stacked convolutions and with more filters per layer. Among the innovations incorporated in AlexNet, were the use of the Rectified Linear Unit (ReLU) activation functions [36], max-pooling layers, dropout [37], data augmentation [38], and ad-hoc initialization algorithms, Jiuxiang Gu et al. detail the recent improvements of CNNs related to the aforementioned innovations [33].
Several years later, He et al. proposed the ResNet architecture, which won the ILSVRC 2015 [39]. It was several times deeper than AlexNet and incorporated innovations such as residual blocks, and batch-normalization layers [40].
Recurrent neural networks are a class of neural networks originally conceived as en effort to incorporate time in neural networks. In 1986 Jordan et al. [41] proposed a neural network with dynamic memory which incorporated recurrent connections from the output layer to the hidden layer. Several years later, Elman et al. [42] proposed another recurrent network based on the Jordan network. This model incorporated recurrent connections directly from the output of the hidden layer allowing the network to maintain a sort of state or memory in the hidden layer. This allowed the network to perform difficult tasks such as temporal sequence prediction, sequence-to-sequence mapping, etc.
Recurrent neural networks are trained by using Backpropagation-Through-Time (BPTT). First, the network is unfolded in time. As a result, a very deep shared weights feedforward neural network is obtained. Then training is performed by using the back-propagation algorithm. At that time, recurrent neural networks had a limited success because they presented learning problems when dealing with very long sequences [43,44]. This was called the vanishing gradient problem because the backpropagated gradients used to update the weights vanished rapidly to zero. As a result, the network was not able to learn patterns contained at the beginning of the sequence. This means that the network was not able to learn long-range dependencies.
In 1997 Hochreiter & Schmidhuber [45] proposed Long Short-Term Memory (LSTM) networks as solution to the vanishing gradient problem. The basic structure of an LSTM network is composed of a chain of units that can decide when to remember, forget, and acquire new information. Consequently, an LSTM unit is composed of one memory cell and three multiplicative gates: Forget, input, and output gates. Figure 1 illustrates the basic structure of an LSTM unit, including in essence sigmoid and tanh units, adders, and multipliers.
Mathematically, the equations governing the operation of an LSTM unit are described next.
The forget gate output is given by Eq (2.1):
ft=σ(Wxfxt+Whfht−1+bf), | (2.1) |
where σ(x)=11+e−x∈[0,1] is the sigmoid activation function, xt is the current input, ht is the current output of the LSTM unit, W and b are weight and bias matrices, respectively.
The output of the input gate it is given by Eq (2.2):
it=σ(Wxixt+Whiht−1+bi), | (2.2) |
where ht−1 is the previous output, W and b are respectively, the weight matrix and bias parameters that must be learned. Besides, the current output of the cell state ct, is obtained by Eq (2.3):
ct=ftct−1+ittanh(Wxcxt+Whcht−1+bc), | (2.3) |
where tanh(x) is the hyperbolic tangent activation function. Finally, ot is the output of the output gate defined as Eq (2.4):
ot=σ(Wxoxt+Whoht−1+bo). | (2.4) |
LSTM networks have found countless applications since their invention, which involve speech recognition, text analysis, language processing, machine translation, handwriting generation, among others.
Gated Recurrent Unit (GRU) networks were introduced by Cho [46] as an alternative to LSTM networks with fewer parameters. A GRU unit utilizes only two gates to decide when to acquire new information and its hidden state is also its output. Basically, a GRU unit is composed of one memory cell and two multiplicative gates: Update, and output gates. Figure 2 illustrates the structure of a fundamental GRU unit.
The computations performed by a GRU unit are described next. The reset gate's output rt, which decides the amount of information to forget, is defined as Eq (2.5)
rt=σ(Wxrxt+Whrht−1+br), | (2.5) |
where σ is a sigmoid activation function, xt is the current input, ht−1 is the hidden state of the previous time step, W and b are the weight matrix and bias parameters, respectively. The output of the update gate for time step t, zt, is described by Eq (2.6)
zt=σ(Wxzxt+Whzht−1+bi). | (2.6) |
Finally, the current hidden state (or output) ht of the GRU unit is formally given by Eq (2.7)
ht=(1−zt)ht−1+zt˜ht, | (2.7) |
where ˜ht is simply defined as,
˜ht=tanh(Wxhxt+rtWhhht−1+bh), | (2.8) |
which uses the reset gate output to feedback past relevant information.
Most previous approaches to handwriting character recognition from sEMG signals were based on pre-processing and feature extraction followed by a linear classifier such as LDA [13]. Whereas some other approaches were based on pre-processing followed by the DTW method. Nevertheless, all these approaches are highly restrictive, in the sense that they require the electrodes to be carefully placed on the muscles to facilitate classification. Additionally, all those approaches were only designed for single-stroke characters.
In this paper, a more general and reliable approach is proposed. This method disposes of highlighted features as multi-user, free-style, multi-stroke handwriting character recognition. Each test-subject wears a Myo armband device on the forearm, and no restriction is made regarding a predefined position of the electrodes. Figure 3 shows the global components of the database construction of the characters recognition approach.
The main contributions are focused on the database creation and the development of a discriminant and versatile characters recognition system. The four blocks are dedicated to the character's acquisition system and especially to the construction of the database using software programmed to control the captured patterns (see details in Section 3). To cope with the variability of sEMG signals, powerful techniques for feature extraction, and sequence processing, such as CNNs and RNNs (LSTM, GRU) networks, were respectively used. From a multi-channel sEMG signal approach, four neural network architectures were designed and implemented, which were named DCNN-LSTM, DCNN-2LSTM, DCNN-GRU, and DCNN-2GRU architectures.
Figure 4 shows the internal layers of the proposed CNN-LSTM architecture.
Such a DCNN-LSTM architecture comprises two convolutional layers, one pooling layer, one LSTM layer, and one fully connected layer. Each convolutional layer has a ReLU activation function and 32 filters of size 3×1. A max-pooling layer follows the first convolutional layer. Thereunto, a dropout layer is added to avoid over-training and improve generalization in each convolutional layer. Besides, the LSTM layer is composed of 180 units. Conversely, the fully connected layer has 36 outputs and a SoftMax activation function.
The neural architectures were implemented in Python using Keras with Tensorflow backend with: The loss function based on categorical-cross-entropy, the optimizer known as Nadam, the fundamental metrics defined as loss and accuracy.
Apart from the above, the internal layers of the proposed CNN-GRU architecture are shown in Figure 5.
It is worth mentioning that the DCNN-GRU architecture is similar to the DCNN-LSTM architecture. The main difference relies on using a GRU layer instead of an LSTM layer. Whereas, the DCNN-2LSTM and DCNN-2GRU architectures use two stacked LSTM and GRU layers, respectively.
Figure 6 presents the sequence of processes involved in the identification of the character class.
The most important steps are armband calibration, handwriting task, sEMG capture (Myo armband: Raw sEMG signals generation, filtering and sampling), magnitude and sequence length normalization to finally send the 8-channel sEMG data to the neural architecture for recognition. In concrete terms, the input to the Neural Network is an array of size (900,8) with eight-channel sEMG signals, one pattern obtained from the armband at each capture, the number of outputs are 36. Each output represents an estimation of a class probability.
The raw sEMG data used in the numerical tests were collected from a MyoTM armband. The MyoTM armband is a commercial electronic device developed by Thalmic LabsⓇ. Such a system is a low-cost, non-invasive, user-friendly wireless eight-channel device with dry electrodes working at a sampling frequency of 200 Hz.
A capture application was developed in Python, under Windows 10. Synthetically, the capture procedure operates as follows, first, the application to capture sEMG signals is started from a command prompt. The application displays a GUI with an option of terminating the application at any time. Next, a subject wears a Myo armband on the left or right forearm (see Figure 7). Posteriorly, the armband is turned on and connected to the PC via a Bluetooth dongle that is plugged into a USB port. Once the device has been correctly detected, a personal calibration profile must be created and selected by using the Myo armband Manager software. Then, the user types a character (letter or digit) into a text field in the GUI. If the typed character (or class) is wrongly chosen, the software still provides the possibility to cancel the capture. After that, the user clicks on Start button to begin the capture process. At this time the subject starts writing with the index finger (or a stylus pen) on the touchscreen. The capture process finishes when the subject clicks on Stop button. Now, the subject must click on Save button to write the captured data to a CSV file. The captured data comprises eight signals corresponding to the eight sensors of the Myo armband. Such characters are written in a freehand style with the index finger or a stylus pen. Additionally, the (x, y) coordinates of the handwriting movements are visualized on the screen. Thereunto, remember that before each handwriting task, the Myo armband is calibrated so that the device always started with the same personal configuration for different users.
Three healthy subjects seated in a comfortable position participated in the creation of the database.
Each task consisted of drawing with the index finger (or stylus pen), 26 lowercase characters ('a'-'z'), and 10 decimal digits ('0'-'9'). The subjects were free to use their particular handwriting style with one or more strokes.
Because a total of 3 subjects participated in creating the database, the number of classes of characters is 36 (26 letters and 10 digits), and each subject wrote each character class one thousand times, the total number of patterns is 108,000 (i.e., 3×36×1000). Ergo, the database contains 108,000 patterns of multi-channel sEMG signals divided into 36 classes and 3,000 patterns per class. Additionally, the database also provides the trajectory of each handwritten character along with the stroke number. This can be useful for visualization or research purposes (i.e., handwriting trajectory reconstruction). Figure 9 shows the representative (average) pattern over all classes. The amplitudes are small in the right side because there are fewer patterns with lengths above 2.5 seconds.
The database was split into training and testing sets to assess the performance of the proposed architectures. The testing set contains 10% of the samples, whereas the rest of the samples formed the training set. Thus, 97,200 samples were used for training and 10,800 samples were used for testing.
All architectures were trained and tested on a workstation with an IntelⓇ CoreTM i7 processor, 16 GB RAM, an NvidiaⓇ GeForce RTXTM 2080 GPU, and Windows 10 using Keras with a TensorflowTM backend. For all cases, the Nesterov-accelerated ADAptive Moment (NADAM) algorithm [47] was used for training using the cross-entropy loss function and a batch size of 1024.
Firstly, each architecture was trained with similar tuning parameters for 8,000 epochs using a cyclical learning rate schedule [48]. Thus, the minimum and maximum learning rates were set to 0.000001 and 0.001, respectively. Furthermore, the step size was defined to five times the number of iterations in each epoch. Secondly, each training essay was repeated several times with different initial conditions to help avoid local minima. Moreover, the dropout rate was varied from 0.8 to 0.6 to help reduce the effects of overfitting. Likewise, the number of units in the LSTM and GRU layers varied from 80 to 200.
Table 1 shows the results obtained for the four architectures using only samples for one subject.
Architecture | Accuracy (%) |
CNN-LSTM | 96.37 |
CNN-2LSTM | 96.48 |
CNN-GRU | 97.03 |
CNN-2GRU | 96.29 |
As can be seen, the best accuracy (97.03%) is achieved by the CNN-GRU architecture, whereas the CNN-2GRU architecture achieves the lowest accuracy (96.29%).
Table 2 shows the results for the CNN-GRU architecture for each and all subjects.
Subjects | Accuracy (%) |
Subject1 | 97.04 |
Subject2 | 95.05 |
Subject3 | 94.89 |
All | 94.85 |
Note that the highest accuracy was obtained for subject1 and the lowest for subject3. The accuracy of all subjects was 94.85%, which is slower than the accuracy of a single subject. This effect is due to the higher variability of the sEMG signals of all subjects. Table 3 shows a quantitative comparison with representative state of the art studies.
Method | Samples | Subjects | Classes | Accuracy (%) |
Template Matching [13] | 350 | 6 | 10 | 97 |
DTW [12] | 520 | 3 | 26 | 84.29 |
DTW [14] | 780 | 4 | 26 | 78.24 to 92.42 |
Proposed | 108k | 3 | 30 | 94.85 |
From the numerical results, it can be observed and concluded that the proposed approach is highly competitive, and in general, compares well with the already-existing methods in the state of the art. Figure 10 shows the confusion matrix obtained for the CNN-GRU architecture.
It is noteworthy that all the classes are recognized with high accuracy for samples from all test subjects. Figure 11 shows the ROC curve obtained for class a by using the CNN-GRU architecture.
As can be seen, the ROC area is very close to one, which indicates an excellent separation capability between classes. A similar result was obtained for the other 35 classes.
The numerical results showed that the four architectures proposed presented comparable accuracies. However, the fastest architecture (CNN-GRU) also presented the highest average accuracy. It is worth mentioning that these results yielded a high accuracy, although no pre-processing was performed. For these reasons, the CNN-GRU architecture was chosen for the task of multi-user, free-style, handwriting character recognition. Notwithstanding, there are always options for amending the actual systems. For instance, an opportunity of improvement is represented by the possibility of testing our approach using a more extensive multi-user database of sEMG signals.
A new methodology for recognizing multi-stroke handwriting characters from sEMG signals using an MyoTM armband device has been presented. The proposed approach consists of using robust Deep Learning architectures for feature extraction, and sequence recognition algorithms such as CNNs and RNNs. One advantage of the discussed approach is that characters can be written either by using a pen or a finger. Another advantage is that characters can also be written in the air.
A CNN-GRU architecture achieved average accuracy of 94.85% for the task of multi-user handwriting character recognition. In comparison with previous works, the results obtained in this work are promising even though sEMG signals presented complex variations on their shapes due to the differences in the placement of the armband device. Therefore, this study is an essential contribution to the area of brain-computer interfaces. Brain-computer interfaces offer an alternative to restore or enhance the communications skills of people with and without disabilities.
This project was fully supported by the Electronics Department of the Universidad de Guanajuato under the Program POA 2020. Especially, authors would like to give special thanks to Perla Villegas-Torres for her support in the edition of this manuscript. Authors are also grateful to the Mexican Council of Science and Technology (CONACyT), for granting the scholarships (CVU/SGN: 495556/455203 and 495754/284009) to fund their doctoral studies.
The authors declare that there is no conflict of interests regarding the publication of this paper.
[1] |
A. Hastings, T. Powell, Chaos in a three-species food chain, Ecol., 72 (1991), 896-903. doi: 10.2307/1940591
![]() |
[2] |
J. N. Eisenberg, D. R. Maszle, The structural stability of a three-species food chain model, J. Theor. Biol., 176 (1995), 501-510. doi: 10.1006/jtbi.1995.0216
![]() |
[3] |
B. Sahoo, S. Poria, The chaos and control of a food chain model supplying additional food to top-predator, Chaos. Solitons Fractals., 58 (2014), 52-64. doi: 10.1016/j.chaos.2013.11.008
![]() |
[4] | P. Panday, N. Pal, S. Samanta, et al. Stability and bifurcation analysis of a three-species food chain model with fear, Internat. J. Bifur. Chaos Appl. Sci. Engrg., 28 (2018), 1850009. |
[5] | B. Nath, N. Kumari, V. Kumar, et al. Refugia and allee effect in prey species stabilize chaos in a tri-trophic food chain model, Differ. Equ. Dyn. Syst., (2019), 1-27. |
[6] | R. J. Taylor, Predation, New York: Chapman and Hall Press, (1984). |
[7] |
S. L. Lima, L. M. Dill, Behavioral decisions made under the risk of predation: A review and prospectus, Can. J. Zool., 68 (1990), 619-640. doi: 10.1139/z90-092
![]() |
[8] |
O. J. Schmitz, A. P. Beckerman, K. M. O'Brien, Behaviorally mediated trophic cascades: Effects of predation risk on food web interactions, Ecol., 78 (1997), 1388-1399. doi: 10.1890/0012-9658(1997)078[1388:BMTCEO]2.0.CO;2
![]() |
[9] |
L. Y. Zanette, A. F. White, M. C. Allen, et al. Perceived predation risk reduces the number of offspring songbirds produce per year, Sci., 334 (2011), 1398-1401. doi: 10.1126/science.1210908
![]() |
[10] |
W. Cresswell, Predation in bird populations, J. Ornitho., 152 (2011), 251-263. doi: 10.1007/s10336-010-0638-1
![]() |
[11] |
K. B. Altendorf, J. W. Laundré, C. A. López González, et al. Assessing effects of predation risk on foraging behavior of mule deer, J. Mammal., 82 (2001), 430-439. doi: 10.1644/1545-1542(2001)082<0430:AEOPRO>2.0.CO;2
![]() |
[12] |
X. Wang, L. Zanette, X. Zou, Modelling the fear effect in predator-prey interactions, J. Math. Biol., 73 (2016), 1179-1204. doi: 10.1007/s00285-016-0989-1
![]() |
[13] |
S. K. Sasmal, Population dynamics with multiple allee effects induced by fear factors-a mathematical study on prey-predator interactions, Appl. Math. Model., 64 (2018), 1-14. doi: 10.1016/j.apm.2018.07.021
![]() |
[14] |
H. Zhang, Y. Cai, S. Fu, et al. Impact of the fear effect in a prey-predator model incorporating a prey refuge, Appl. Math. Computation, 356 (2019), 328-337. doi: 10.1016/j.amc.2019.03.034
![]() |
[15] | K. Kundu, S. Pal, S. Samanta, et al. Impact of fear effect in a discrete-time predator-prey system, Bull. Calcutta Math. Soc., 110 (2018), 245-264. |
[16] |
S. Mondal, A. Maiti, G. Samanta, Effects of fear and additional food in a delayed predator-prey model, Biophys. Rev. Lett., 13 (2018), 157-177. doi: 10.1142/S1793048018500091
![]() |
[17] |
D. Duan, B. Niu, J. Wei, Hopf-hopf bifurcation and chaotic attractors in a delayed diffusive predator-prey model with fear effect, Chaos. Solitons Fractals., 123 (2019), 206-216. doi: 10.1016/j.chaos.2019.04.012
![]() |
[18] | X. Wang, X. Zou, Pattern formation of a predator-prey model with the cost of anti-predator behaviors, Math. Biosci. Eng., 15 (2017), 775-805. |
[19] | S. Chen, Z. Liu, J. Shi, Nonexistence of nonconstant positive steady states of a diffusive predatorprey model with fear effect, J. Nonlinear Model. Anal., 1 (2019), 47-56. |
[20] |
A. Sha, S. Samanta, M. Martcheva, et al. Backward bifurcation, oscillations and chaos in an ecoepidemiological model with fear effect, J. Biol. Dyn., 13 (2019), 301-327. doi: 10.1080/17513758.2019.1593525
![]() |
[21] | S. Pal, S. Majhi, S. Mandal, et al. Role of fear in a predator-prey model with beddington-deangelis functional response, Z. für Naturforsch. A., 74 (2019), 581-595. |
[22] |
V. Rai, R. K. Upadhyay, Chaotic population dynamics and biology of the top-predator, Chaos, Solitons Fractals., 21 (2004), 1195-1204. doi: 10.1016/j.chaos.2003.12.065
![]() |
[23] |
R. Upadhyay, R. Naji, N. Kumari, Dynamical complexity in some ecological models: Effect of toxin production by phytoplankton, Nonlinear Anal. Model Control., 12 (2007), 123-138. doi: 10.15388/NA.2007.12.1.14726
![]() |
[24] | M. A. Aziz-Alaoui, Study of a Leslie-Gower-type tritrophic population model. Chaos Solitons Fractals., 14 (2002), 1275-1293. |
1. | Shashank Kumar Singh, Amrita Chaturvedi, Leveraging deep feature learning for wearable sensors based handwritten character recognition, 2023, 80, 17468094, 104198, 10.1016/j.bspc.2022.104198 | |
2. | Ayush Tripathi, A. P. Prathosh, Suriya Prakash Muthukrishnan, Lalan Kumar, SurfMyoAiR: A Surface Electromyography-Based Framework for Airwriting Recognition, 2023, 72, 0018-9456, 1, 10.1109/TIM.2023.3248084 | |
3. | William Villegas-Ch, Isabel Urbina-Camacho, Joselin García-Ortiz, Detection of Abnormal Patterns in Children’s Handwriting by Using an Artificial-Intelligence-Based Method, 2023, 10, 2227-9709, 52, 10.3390/informatics10020052 | |
4. | Sike Ni, Mohammed A.A. Al-qaness, Ammar Hawbani, Dalal Al-Alimi, Mohamed Abd Elaziz, Ahmed A. Ewees, A survey on hand gesture recognition based on surface electromyography: Fundamentals, methods, applications, challenges and future trends, 2024, 166, 15684946, 112235, 10.1016/j.asoc.2024.112235 | |
5. | Jayakanth Kunhoth, Somaya Al-Maadeed, Suchithra Kunhoth, Younes Akbari, Moutaz Saleh, Automated systems for diagnosis of dysgraphia in children: a survey and novel framework, 2024, 27, 1433-2833, 707, 10.1007/s10032-024-00464-z | |
6. | Andrea Tigrini, Simone Ranaldi, Alessandro Mengarelli, Federica Verdini, Mara Scattolini, Rami Mobarak, Sandro Fioretti, Silvia Conforto, Laura Burattini, 2024, Automatic Handwriting Recognition with a Minimal EMG Electrodes Setup: A Preliminary Investigation, 979-8-3503-0799-3, 1, 10.1109/MeMeA60663.2024.10596883 | |
7. | Ayush Tripathi, Prathosh A.P., Suriya Prakash Muthukrishnan, Lalan Kumar, TripCEAiR: A multi-loss minimization approach for surface EMG based airwriting recognition, 2023, 85, 17468094, 104991, 10.1016/j.bspc.2023.104991 | |
8. | William Villegas-Ch, Jaime Govea, Angel Jaramillo-Alcazar, IoT Anomaly Detection to Strengthen Cybersecurity in the Critical Infrastructure of Smart Cities, 2023, 13, 2076-3417, 10977, 10.3390/app131910977 | |
9. | Andrea Tigrini, Federica Verdini, Mara Scattolini, Federico Barbarossa, Laura Burattini, Micaela Morettini, Sandro Fioretti, Alessandro Mengarelli, Handwritten Digits Recognition From sEMG: Electrodes Location and Feature Selection, 2023, 11, 2169-3536, 58006, 10.1109/ACCESS.2023.3279735 | |
10. | Ines Boutabia, Abdelmadjid Benmachiche, Ali Abdelatif Betouil, Chaouki Chemam, 2024, A Survey in the Use of Deep Learning Techniques in The Open Classroom Approach, 979-8-3503-5026-5, 1, 10.1109/PAIS62114.2024.10541268 | |
11. | Awang Karisma As’Ad Adi Asta, Eko Mulyanto Yuniarno, Supeno Mardi Susiki Nugroho, Cries Avian, 2023, Handwriting Classification based on Hand Movement using ConvLSTM, 979-8-3503-1395-6, 341, 10.1109/ISITIA59021.2023.10221037 | |
12. | Shashank Kumar Singh, Amrita Chaturvedi, An efficient multi-modal sensors feature fusion approach for handwritten characters recognition using Shapley values and deep autoencoder, 2024, 138, 09521976, 109225, 10.1016/j.engappai.2024.109225 | |
13. | Shashank Kumar Singh, Amrita Chaturvedi, 2025, Chapter 21, 978-3-031-81403-7, 288, 10.1007/978-3-031-81404-4_21 | |
14. | Mohammed A. A. Al-qaness, Sike Ni, TCNN-KAN: Optimized CNN by Kolmogorov-Arnold Network and Pruning Techniques for sEMG Gesture Recognition, 2025, 29, 2168-2194, 188, 10.1109/JBHI.2024.3467065 |
Architecture | Accuracy (%) |
CNN-LSTM | 96.37 |
CNN-2LSTM | 96.48 |
CNN-GRU | 97.03 |
CNN-2GRU | 96.29 |
Subjects | Accuracy (%) |
Subject1 | 97.04 |
Subject2 | 95.05 |
Subject3 | 94.89 |
All | 94.85 |
Architecture | Accuracy (%) |
CNN-LSTM | 96.37 |
CNN-2LSTM | 96.48 |
CNN-GRU | 97.03 |
CNN-2GRU | 96.29 |
Subjects | Accuracy (%) |
Subject1 | 97.04 |
Subject2 | 95.05 |
Subject3 | 94.89 |
All | 94.85 |
Method | Samples | Subjects | Classes | Accuracy (%) |
Template Matching [13] | 350 | 6 | 10 | 97 |
DTW [12] | 520 | 3 | 26 | 84.29 |
DTW [14] | 780 | 4 | 26 | 78.24 to 92.42 |
Proposed | 108k | 3 | 30 | 94.85 |