Processing math: 100%
Research article

Comparison of effectiveness of enhanced infection countermeasures in different scenarios, using a dynamic-spread-function model

  • When formulating countermeasures to epidemics such as those generated by COVID-19, estimates of the benefits of a given intervention for a specific population are highly beneficial to policy makers. A recently introduced tool, known as the "dynamic-spread" SIR model, can perform population-specific risk assessment. Behavior is quantified by the dynamic-spread function, which includes the mechanisms of droplet reduction using facemasks and transmission control due to social distancing. The spread function is calibrated using infection data from a previous wave of the infection, or other data felt to accurately represent the population behaviors. The model then computes the rate of spread of the infection for different hypothesized interventions, over the time window for the calibration data. The dynamic-spread model was used to assess the benefit of three enhanced intervention strategies – increased mask filtration efficiency, higher mask compliance, and elevated social distancing – in four COVID-19 scenarios occurring in 2020: the first wave (i.e. until the first peak in numbers of new infections) in New York City; the first wave in New York State; the spread aboard the Diamond Princess Cruise Liner; and the peak occurring after re-opening in Harris County, Texas. Differences in the efficacy of the same intervention in the different scenarios were estimated. As an example, when the average outward filtration efficiency for facemasks worn in New York City was increased from an assumed baseline of 67% to a hypothesized 90%, the calculated peak number of new infections per day decreased by 40%. For the same baseline and hypothesized filtration efficiencies aboard the Diamond Princess Cruise liner, the calculated peak number of new infections per day decreased by about 15%. An important factor contributing to the difference between the two scenarios is the lower mask compliance (derivable from the spread function) aboard the Diamond Princess.

    Citation: Gavin D'Souza, Jenna Osborn, Shayna Berman, Matthew Myers. Comparison of effectiveness of enhanced infection countermeasures in different scenarios, using a dynamic-spread-function model[J]. Mathematical Biosciences and Engineering, 2022, 19(9): 9571-9589. doi: 10.3934/mbe.2022445

    Related Papers:

    [1] Holly Gaff, Elsa Schaefer . Optimal control applied to vaccination and treatment strategies for various epidemiological models. Mathematical Biosciences and Engineering, 2009, 6(3): 469-492. doi: 10.3934/mbe.2009.6.469
    [2] Han Ma, Qimin Zhang . Threshold dynamics and optimal control on an age-structured SIRS epidemic model with vaccination. Mathematical Biosciences and Engineering, 2021, 18(6): 9474-9495. doi: 10.3934/mbe.2021465
    [3] Yinggao Zhou, Jianhong Wu, Min Wu . Optimal isolation strategies of emerging infectious diseases with limited resources. Mathematical Biosciences and Engineering, 2013, 10(5&6): 1691-1701. doi: 10.3934/mbe.2013.10.1691
    [4] Erika Asano, Louis J. Gross, Suzanne Lenhart, Leslie A. Real . Optimal control of vaccine distribution in a rabies metapopulation model. Mathematical Biosciences and Engineering, 2008, 5(2): 219-238. doi: 10.3934/mbe.2008.5.219
    [5] Kento Okuwa, Hisashi Inaba, Toshikazu Kuniya . Mathematical analysis for an age-structured SIRS epidemic model. Mathematical Biosciences and Engineering, 2019, 16(5): 6071-6102. doi: 10.3934/mbe.2019304
    [6] Eunha Shim . Optimal strategies of social distancing and vaccination against seasonal influenza. Mathematical Biosciences and Engineering, 2013, 10(5&6): 1615-1634. doi: 10.3934/mbe.2013.10.1615
    [7] Toshikazu Kuniya, Taisuke Nakata, Daisuke Fujii . Optimal vaccine allocation strategy: Theory and application to the early stage of COVID-19 in Japan. Mathematical Biosciences and Engineering, 2024, 21(6): 6359-6371. doi: 10.3934/mbe.2024277
    [8] Tingting Xue, Long Zhang, Xiaolin Fan . Dynamic modeling and analysis of Hepatitis B epidemic with general incidence. Mathematical Biosciences and Engineering, 2023, 20(6): 10883-10908. doi: 10.3934/mbe.2023483
    [9] Yoon-gu Hwang, Hee-Dae Kwon, Jeehyun Lee . Feedback control problem of an SIR epidemic model based on the Hamilton-Jacobi-Bellman equation. Mathematical Biosciences and Engineering, 2020, 17(3): 2284-2301. doi: 10.3934/mbe.2020121
    [10] Tingting Ding, Tongqian Zhang . Asymptotic behavior of the solutions for a stochastic SIRS model with information intervention. Mathematical Biosciences and Engineering, 2022, 19(7): 6940-6961. doi: 10.3934/mbe.2022327
  • When formulating countermeasures to epidemics such as those generated by COVID-19, estimates of the benefits of a given intervention for a specific population are highly beneficial to policy makers. A recently introduced tool, known as the "dynamic-spread" SIR model, can perform population-specific risk assessment. Behavior is quantified by the dynamic-spread function, which includes the mechanisms of droplet reduction using facemasks and transmission control due to social distancing. The spread function is calibrated using infection data from a previous wave of the infection, or other data felt to accurately represent the population behaviors. The model then computes the rate of spread of the infection for different hypothesized interventions, over the time window for the calibration data. The dynamic-spread model was used to assess the benefit of three enhanced intervention strategies – increased mask filtration efficiency, higher mask compliance, and elevated social distancing – in four COVID-19 scenarios occurring in 2020: the first wave (i.e. until the first peak in numbers of new infections) in New York City; the first wave in New York State; the spread aboard the Diamond Princess Cruise Liner; and the peak occurring after re-opening in Harris County, Texas. Differences in the efficacy of the same intervention in the different scenarios were estimated. As an example, when the average outward filtration efficiency for facemasks worn in New York City was increased from an assumed baseline of 67% to a hypothesized 90%, the calculated peak number of new infections per day decreased by 40%. For the same baseline and hypothesized filtration efficiencies aboard the Diamond Princess Cruise liner, the calculated peak number of new infections per day decreased by about 15%. An important factor contributing to the difference between the two scenarios is the lower mask compliance (derivable from the spread function) aboard the Diamond Princess.



    Voice pathology influences the patient's speaking ability, affecting disorders, including agitation, voice tiredness, trouble pronunciation the phrases. There are, in total, three instances of associated illnesses used in this work: dysphonia, laryngitis, and vocal cord paralysis [1]. The trouble in sustaining the voice is linked to dysphonia and primarily to the heartbeat. Chronic laryngitis often includes hoarseness effects, triggered by larynx irritation for a lengthy period of time. Vocal cord paralysis impacts the patient's speech and breathing and frequently contributes to a lack of vocal capacity as all cords are paralyzed. The diagnostic test to classify these language pathologies is rather intrusive for the patient. Thus, the research and production of machine learning approaches have risen in recent years [2]. In the case of various pathologies, most works have created, using audio with a steady voice, mostly vowel /a/ [3] or other vows such as /i/ and /u/ [4,5], as well as obtaining satisfactory results. However, there are still very few continuous voice databases with medical diagnostic annotations accessible to study, notwithstanding the scientific interests. Based on the previous record the specialized doctors or physician detects that whether it's a pathological or normal voice in subjective detection. Now the objective detection is getting more attention in voice diseases detection in medical diagnosis. The risk of voice pathologies is increasing in individuals [6]. In medical science, engineering and programming are induced to construct a system to identify voice disorders at the beginning of these diseases. One of the open problems is voice disease identification and detection. It is about detecting how to sever the voice disease because when the severity of the disease is low, the patient can easily pronounce the vowel.

    Vocal folds are affected during phonation in voice disorders, due to which the vocal folds produce asymmetrical vibrations. The asymmetrical vibrations may be due to the malfunctioning factors responsible for producing vibrations [7]. There is not any kind of tracheobronchial effect on the vocal tract when the vowels are produced. The vocal tract is influenced by the structure of infra laryngeal and articulatory interaction in the vocal tract and articulatory in vocal folds [8]. Information related to the vocal tract is predictable and allows vocal folds identification characteristics, especially in phonation [9]. It depends on the type and location of voice disease in vocal folds, affecting the vibration production during phonation. There are certain factors upon which the vibration of vocal folds depends, and they're as follow.

    ● Presence of mucous on the tissue of vocal fold.

    ● The stiffness of vocal fold

    ● The tension created in the vocal fold

    ● The muscles of the larynx

    ● Folds opening and closing

    As certain different types of voice diseases affect these factors differently. The closure of vocal folds is different in vibration because of the disease's size and location; due to this, there is variation in vibrations the diseases are differentiated. The research fields which are present in speech are speech and speaker recognition, synthesis of speech, coding of speech. The automatic identification of the speaker has the following objectives: feature extraction, feature characterization, and recognition of speaker identity. Dysphonia and other voice pathological problems were increasing drastically in the US; about 7.5 million people were having voice disease [10]. The main aim and main concern are to establish a system that will identify the disease of voice correctly and accurately. The acoustic analysis is not dependent upon the interventions of humans and will help in making the decision to the clinicians. However, the doctors know the decision according to the diagnosis, and this system will act as an assistive tool to them. The proposed methodology is the continuity of our research work [11] that was the meta-analysis of classifiers used for voice disorders. The gaps that we have identified in [11] tried to full fill in the proposed methodology. In our research, we have used the free access database of voice pathology, SVD. This database is recorded by the institute of phonetics of Saarland University. The noninvasive technique which is based on digital processing of given audio signals is acoustic analysis. It is an efficient technique used in the diagnosis and identification of voice diseases. Two types of analysis are used in automatic voice disorder identification i.e., short-term analysis and long-term analysis. The long-term analysis takes the parameters which are yield through acoustic analysis of an audio signal [12] whereas the parameters of short-term analysis are yield from MFCC (Mel-frequency cepstral coefficients), LPCC, and LPC [8]. MFCC is based on short time spectral observations. Mel frequency cepstral coefficient act on the human auditory system. That's why it is used to extract the Features from the given audio signal and is very useful in identifying voice diseases [13].

    In [14], the key goal of Nasheri et al. works is to develop accurate and robust feature extraction to classify and differentiate voice disease by autocorrelation and entropy analysis of several frequency bands. From each frame, maximal peak values and corresponding lag values were obtained, utilizing autocorrelation as a function for the identification and differentiation of pathological samples. After normalizing the values to be used as functions, we have obtained entropy for and frame of the talk signal. These features were tested in various frequency bands for the recognition and classification schemes to assess the contributions of each unit. Various continuous vocal examples were obtained for both regular and irregular voices in English, German and Arabic from three distinct datasets. As a classifier, the support vector machine was used. 92% for SVD is the best-recorded accuracy [15]. In [15], the paper's main objective is to evaluate multidimensional voice parameters (MDVP) such that voice pathologies can be automatically defined and separated in multiple data sets before deciding the parameters of the two processes, which are well-compared. The experimental results indicate that the utility of the MDPV parameters using these databases is clearly different. Substantially rated parameters differ from database to database. The three top-rated MDVPs organized with the fisher prejudice ratio were used to obtain optimum accuracy: 99.98 percent for SVD. The article [16] utilizes a correlation approach in order to define and identify pathological materials to extract optimum pick values and their corresponding lag values from each frame of the spoken signal. In different frequency bands, these features are tested to assess the contribution of each band to processes of recognition and classification. The most contributing bands are between 1000 and 8000 Hz for detection and classification. Maximum accurate consistency in Massachusetts Eye and Ear Infirmary, the Saarbrücken Expression Database, and the Arabic Voice Pathology Database was 99.809, 90.979 and 91.16%. However, 99.255, 98.941 and 95.188% respectively in three datasets reached full accuracy by cross-correlation. Teixeira proposed the voice recognition device and, in both, his publications retained the same features but modified the classifications. In [17] SVM was used with Jitter, Glitter, and HNR, and the accuracy recorded was 71%. The recorded accuracy was 100% but only for female voices in the [18] MLP-ANN used for jitter, glitter, and HNR. In [19], the recorded accuracy of Fonseca et al. using SVM with SE, ZCRs, SH was 95%. Many automated disorders detection systems were established by applying various types of traditional voice characteristics such as linear prediction coefficients, linear cepstral coefficients, and cepstral coefficients of Mel-frequency (MFCCs). The aim of the research is to see whether traditional voice features accurately identify voice pathology and can compare it with voice quality [20].In order for this research to be investigated, an automated MFCC method of detection was developed, and three separate databases of voice impairment were used. The findings of the experiment indicate that the exactness of the MFCC dependent method differs between the database and the database. The intra-database detections vary from 72 to 95%, but for the SVD database, the highest reported accuracy was 80% [20]. In [21], while assessing voice pathologies, when 13 conventional MFCC features were extracted, the highest reported accuracy is 72% using the SVM as a machine learning classifier.

    SVD stands for Saarbrücken Voice Database. Basically, SVD is a freely accessible archive, a compilation of voice recordings from more than 2000 individuals. Vocal register provided at normal, high, and low pitch. The reality has been documented in a recording session-Voice documentation for increasing pitch. The voice signal and the EGG signal have been stored in individual files for the specified components [22]. The database has a text file containing all applicable dataset material. These features make it a reasonable alternative for the experimenters to use. Both captured SVD voices were sampled at a resolution of 16-bit at 50 kHz. The Saarbruecken Voice Server is accessible via this web interface. It includes several internet pages that are used to choose the parameters for the database program, to play directly and to record, and to select the listening session files that are to be exported from the SVD database to the required parameter [14]. From the SVD index, the diseases we picked are laryngitis, cyst, non-fluency syndrome, and dysphonia and we have also used audio samples of vowels (a, e, i, o, u) [22].

    SVM is a good way to build a classifier. It seeks to establish a decision limit between two groups, which will enable labels to be predicted from one or more vectors. This choice, called the hyper-plane, is so directed that from the nearest data points in each class, it is as far as possible. The nearest points are regarded as support vectors [23]. An SVM classifier generates the most advanced hyperplane in the transformed entrance space and separates the outstanding groups, and maximizes the gap to cleanly separated instances nearby. A quadratic optimization problem is the parameters of the hyperplane approach [24]. So to label a dataset:

    (x1,y1),,(xn,yn),xiRdandyi(1,+1)

    where xi is a characteristic vector representation and yi is a class mark (negative or positive) of a training formula i. The optimum hyperplane can then be described as:

    wxT+b=0

    where w is the weight matrix, x is the input vector, and b is the bias. For all the elements of the training collection, w and b must fulfill the following inequalities:

    wxT+b+1ifyi=1
    wxT+b1ifyi=1

    The main aim is to find w and b so that we can form a hyperplane and increase the margin 1/w2. Vectors xi for which |yi|(wxiT+b)=1 will be termed support-vector. In the proposed methodology, as shown in graph 1, linear SVM, quadratic SVM, cubic SVM, fine Gaussian SVM, medium Gaussian SVM, and coarse Gaussian have been used to check the trained model.

    In the Naive Bayes classification, the conditional likelihood is centered on the features in one class after the collection of features utilizing current processes. It identifies functions by an established system of selection and selects a supplementary functionality that can reclassify the class space for the features selected [25]. Denote a vector of variables D=(di),i=1,2,,n, where di is corresponding to a letter, a word, or other attributes about some text in reality, and a set of C={C1,C2,,Ck} is predefined classes. Text classification is to assign a class label Cj,j=1,2,,k from C to a document. In essence, the Bayes Classifier is a hybrid probability model parameter:

    P(cj|D)=P(cj)P(D|cj)P(D)

    where P(cj) is previous details on the obvious likelihood of a class is the information from the observations, which is the experience of the text itself to be graded, and P(D|cj) is the probability of the distribution of document D in class space. As the naive Bayes assigns the most likelihood text to the class, the Naive Bayes is perfect in the context of probability. As speech recognition has the problem of multiple classes' classification so, Naïve Bayes can handle such problems and works well. Naïve Bayes works on Bayesian theorem along with naïve assumption, i.e., the pair of features are independent of each other. Naïve Bayes classifier works well in real world conditions. Naïve Bayes classifier has good performance and also works fast. The benefits of using naïve Bayes classifier are situational unconventional assumption that helps to get fast classification results, probabilistic assumption, and high accuracy [26,27]. By the trained data, the Naïve Bayes classifier is made while on the basis of the dataset, the algorithm is formed. The trained data is classified under four classes' dysphonia, cyst, laryngitis, and non-fluency syndrome. Gaussian and kernel approaches have been used to test the model.

    A decision tree categorizes data objects by answering a range of questions about their characteristics. Each query is embedded in an internal node, and with each potential answer to its question, each inner node points to a child node. The questions shape a tree-encoded hierarchy. In the simplest type, we pose questions of yes or no, and there is a yes child and no child in an inner node. An object is sorted into a class in conjunction with the answers to the item in question by following the way from the top node, the root, to a child-loss node. The class associated with the leaf that reaches is allocated an object. In certain combinations, each leaf includes a distribution of probabilities around the groups, which calculates the likelihood that an object that enters the leaf is of a specific class. However, it may be challenging to quantify impartial odds [28].

    Ensemble methods are algorithms that build up a group of classifiers and then categories new data points via a (weighted) majority of their correlations. The initial ensemble approach is averaging in Bayesian but modern algorithms provide error correction, bagging, and boosting performance [29]. Ensemble learning is about multiple classifiers systems. Training of distinguish classifier is included, and then predictions are combined to get better accuracy result of classification. The method of ensemble tries to gather a set of the learner as compared to ordinary approaches of learning, and this helps to make predictions based on a single learner [30]. Outstanding results of the stable classifier are received by the Bagged tree [31]. It deals with artificial and also real-world problems and also helps to improve the accuracy of some classifiers.

    Speech signals are split into frames because speech signals are constant or non-varying for a short period of time, so the characteristics of the frame in the form of features are extracted where the speech signals remain unchanged [32]. We have used a supervised machine learning algorithm for which we have taken the audio signals from the SVD dataset and extracted the features that can be used in the classifier learner app as an input signal. Then the feature vector is constructed from each. The popular audio features are using the technique of Mel Frequency Cepstral coefficient. Other than 13 MFCC and pitch, we have six other features, which are spectral flux, spectral centroid, ZCR, roll-off, short-term energy, and spectral entropy. It has been observed in a literature review [20] and [21] that when MFCC was used alone, the reported accuracy was not better, i.e., 72 and 80%. So to increase the accuracy and for better detection of voice pathologies, we have added 7 more features and extracted them on audio features along with 13 conventional MFCC features.

    Mel frequency cepstral coefficient works on the principle of the human auditory system. It is used to extract the features from the audio signal datasets. In MFCC, the bands of frequency using the Mel scale parameter there is the equal spacing that is very close to the human auditory system as compared to Cepstrum, which has linear spacing in bands of frequency. According to the psychophysical study, the human auditory system doesn't have a linear scale that it follows. Every incoming audio signal has an original frequency 'f', which is measured in Hz, and a pitch is measured on the Mel scale [33]. MFCC is said to be coefficients that are derived from audio signals. The speech input is actually the audio signal input that undergoes the process of framing. Before framing the audio signal undergoes the process of pre-emphasis, this is beneficial to achieve good accuracy and good efficiency level. Using this process, the compensation in higher frequency is achieved that is curbed in the human auditory system during sound production.

    c2(n)=c(n)dc(n1)

    Here c2(n) shows the signal of output and d recommended value is 0.9 and 1 [33]. The filter's z transform is as follow,

    H(z)=1dz(1)

    After the process of pre-emphasis now the aim is to split the entire audio signal into a number of frames to easily analyze and interpret every frame signal. The audio signal is split into a frame of around 10 ms while the Standard framing size is 25 ms [34]. It shows that the length of the frame for the 50 kHz audio signal will be 50 k * 0.01 = 500 samples. The step of framing allows the overlapping of the frames. The step of the frame is 10 ms for the first 500 samples. The 500 sample frame starts at 0 samples, and then another 500 sample frame will begin at sample 501 and so on until the last audio signal is reached. The reason for taking a 10 ms signal is that audio recordings in the SVD database [22] are very small, i.e., vowel recordings, so the suitable recording time as per the database is 10 ms. There is some effect of spectral artifacts after framing; these effects are reduced by applying Hamming window function. The combination of short-term spectrum and the transfer function of window (hamming) becomes convolution in the frequency domain. In order to maintain continuity between the first and last marks in the frame, each frame has to be multiplied with the function of hamming window [35]. The efficient function of the window has a narrow main lobe and low side lobe. The hamming window function is given below,

    K(n)=0.540.46cos(2πn/N1),0nN1

    K(n) represents the window, and the output is given by Q (n) below, whereas X (n) shows the input frame signal.

    Q(n)=X(n)w(n)

    Now the signal needs to be converted into the frequency domain signal from the time domain signal. For that, FFT (Fast Fourier transformation) is applied to the signal to prepare it for the next stage of the Mel filter bank [35]. The mathematical expression is given below. The period gram of audio signal (CYSTE1.WAV) before the MFCC is shown below in Figure 1.

    Q(ω)=FFT[k(t)X(t)]=k(ω)X(ω)
    Figure 1.  Period gram before MFCC.

    Now the original frequency of the signal is converted into the Mel frequency using the Mel filter bank. The Mel filter bank consists of a group of triangular filters. There is not any uniform spacing between the filters, and fewer filters are present in the higher frequency region while a greater number of filters are present in the lower frequency region [34]. The unique property of Filter bank is that it can be applicable on time domain and also in frequency domain signals. In the processing of Mel frequency cepstral coefficient (MFCC), it is applicable in frequency domain signals.

    In Figure 2, the filter applied in the region of lower and higher frequency range showing the change in frequency. The perceived frequency, the pitch of a tone, is related with the Mel scale to its original frequency measured. The human auditory system is capable of distinguishing the slight changes in pitch at low frequency as compared to a higher frequency. So the main reason for using the Mel scale is to do feature extracting on the basis of the human auditory system [32]. The mathematical formula that is used to convert the original frequency into the Mel Scale is as follow:

    M(f)=1125xln(1+f/700)
    Figure 2.  Showing number of filters in the lower and higher region of the frequency.

    Here f represents the actual frequency of the audio signal.

    The method of zero paddings is used if the frames of audio signals are not in even numbers. The upcoming steps are applicable on every frame. 13 MFC coefficients are extracted from every frame. The time-domain signal B(n) after framing we get Bi(n), here n shows the range from 1 to 500 (we have 500 samples for 50 kHz frequency audio signal) [36]. When the complex discrete Fourier transform (DFT) is calculated so, we haveBi(k), here 'i' represents a number of the frame corresponding to the time domain.Zi(k) is then the frame power spectrum. In order to apply DFT, we have the following equation,

    Bi(k)=Nn=1Bi(n)h(n)ej2πkn/N

    here h(n) represents the Hamming window function, while K shows the length of DFT. The frame's Bi(n) power spectral estimation based on the period gram is given by:

    zi(k)=1N|Bi(k)|2

    It is called spectral power estimation based on the period gram. Take the absolute value of the complex Fourier transform, then the square of the result. Generally, N point FFT is performed, and first C coefficients are kept. Calculate Mel filter bank, which is the group of 20 to 40 triangular filters. These triangular filters are applied to spectral power estimation. In order to get energies of the filter bank, we need to multiply every bank of the filter with power spectrum and then add coefficients [34].

    Signal changes from positive to negative at a specific rate; that rate is called zero crossing rates (ZCR). ZCR is about the extent of no. of times in a frame that audio signal's amplitude passes over zero value as shown in the equation below,

    ZCR=1/(T1)(t=1)(T1)

    Before computation of zero crossing rate, the waveform is shifted to distinguish between unvoiced audio signals and environment & noise. When the noise is small, it is very useful. Zero-crossing rate has a high frequency of the signal and lowers for voiced audio signals than unvoiced signals. It is a useful feature used to detect endpoints and to classify voiced and unvoiced signals. Zero-crossing rate doesn't notice noise which is its drawback [37].

    The spectral roll-off is the frequency under which cutoff frequency, i.e., some percent of total energy, is contained. Spectral roll-off is used to differentiate between noise signals and harmonic signals. (Noise is above roll off and harmonic is below roll-off). Similarly, in our project, we have used spectral roll-off to differentiate between voiced and unvoiced signals (normal and pathological signals). The frequency below 85% of spectrum magnitude distribution is concentrated in the spectral roll-off. In digital signal processing, the spectral centroid is widely used [38]. The spectrum's gravity is the spectral centroid, i.e., and it shows the center of the mass of the spectrum. Spectral flux is the square of the difference between a successive spectral distribution and successive frames of signal. It measures the rate of change of the power spectrum of a signal. It can be calculated by comparing the power spectrum of one frame with the previous frame. In audio signal classification, short-term energy is widely used. It helps to differentiate between voiced signals and unvoiced signals. In a high-quality audio signal case, it is used to distinguish between audio and silence. It allows the calculating amount of energy in the audio signal at a specific instance in time.

    The spectral complexity is quantified of a system by entropy information, i.e., spectral entropy. An uncertain system assumes a variable Y as a system state [39]. Through Fast Fourier Transform, signals of time series become power spectrum, while the entropy information of power spectrum is power spectral entropy.

    In the section on material and method, now we are going to cover the method we used for the extraction of features from audio signals. The feature extraction and classification of then those extracted features have been carried out on MATLAB. The MATLAB version used to implement the proposed methodology is R2018b. The extraction of features is the method to convert the audio signals into the sequence of the vector of features; it carries the information of characteristics in audio signals. Then these vectors are used in the analysis for the different algorithms. It is difficult for those analysis algorithms are based on features extracted from window sources. The features which are based on the window show the description of the signal for a short time. MATLAB has the function of MFCC (Mel-frequency cepstral coefficients), which is used to extract pitch and 13 MFCC features from the imported dataset. After the process of feature extraction, the classification is carried out using the classification learner App in MATLAB. The classification learner app is easily available in the version of MATLAB, which is used to implement the methodology. No extra downloads were required overall. Classification learner is used to training the model of classifiers including; linear discriminant analysis, support vector machine, k nearest neighbor, decision tree, etc., using the classification learner application several other functions as an exploration of data, feature selection, specification of validation scheme, and results from the evaluation. In order to use the model for other data also the model can be exported to the workspace. The model can be trained using the classification learner app on MATLAB. The procedure is as follows:

    1) Open MATLAB and perform the feature extraction on the dataset and save the extracted feature in an excel file.

    2) Open classification leaner app in MATLAB, create a new session and upload your excel file (extracted features).

    3) After successfully uploading the file, select the workspace variable 'FEATURES', and in response, you need to select 'LABELS'.

    4) Make sure that all the features are selected. Here we are using 20 audio features, which include 13 MFCC (Mel-frequency cepstral coefficients), pitch, ZCR, energy Entropy, Rolloff, Spectral Centroid, Spectral Flux, and Energy.

    5) Provide the validation scheme the value of cross-validation. In the proposed methodology, we are opting for 5-fold validation to avoid error-based results and to train the model to avoid overfitting.

    6) Now to train your model on SVM, KNN, LDA, and other models, you need to click on the MODEL TYPE drop-down and select the particular model you want to train. For example: if you want to train your data on SVM so, select SVM, you will get multiple options to click on 'ALL SVM'.

    7) The types of models are enabled according to the data we have as we are using the audio dataset, so we are going to apply the Decision tree and its kernel, the Naïve Bayes classifier, and its kernel SVM and its kernel, and the ensemble classifier and its kernel.

    8) After the selection of the particular type of MODEL, you need to click on 'TRAIN'. It depends on the data that how much time it will take to be trained.

    9) After the training of the model, you can export the trained model in the workspace so that it can be used for other data as well. Before exiting the MATLAB, make sure that you have saved it as the trained model on your PC. This trained data is helpful in the real-time system also.

    The audio features were extracted from the audio database; i.e., 20 features were extracted. After the feature extraction, the classification of features was carried out using the classification learner app. Among the several classifiers, we have used four classifiers, and their kernels which are support vector machine classifier (SVM), Naïve Bayes classifier, decision tree classifier, and ensemble classifier, and the results that we have generated through the classification app can be observed in Table 1. The reported accuracy of the individual classifier is same for all the samples of laryngitis, cyst, non-fluency syndrome, and dysphonia.

    Table 1.  Average classification accuracy rate of different classifier.
    Classifiers Accuracy (%) Misclassification rate (%)
    SVM 93.18 6.82
    Decision Tree 100 0
    Naïve bayes 99.45 0.55
    Ensemble 51 49

     | Show Table
    DownLoad: CSV

    Among all voice data, 5-fold cross-validation has been used for testing. After pre-processing, the voice signals are disrupted into the short stream by implementing a moderate windowing technique, hamming window. Every segment is processed separately. Hamming window is commonly used in narrow-band applications, e.g., telephone signal spectrum. Fast Fourier transform is computed for every segmented signal. Correlation of these small segmented signals is carried. De-correlation is performed to reconstruct these short segmented signals. The cross-correlating values are eliminated. A permutation is applied to find out the relation between these independent segmented signals. These segmented values are combined on the basis of relation. The features that we extract from each frame for classification. Combining these futures forming a support vector. Table 2 shows the reported accuracy of types of SVM, decision tree, Naïve Bayes, and ensemble, which we have used in our intercomparison approach.

    Table 2.  Comparison of accuracies of different classifiers with their kernel.
    Classifiers Accuracy rate
    SVM kernel
    Linear SVM 100%
    Quadratic SVM 100%
    Cubic SVM 100%
    Fine Gaussian SVM 72.7%
    Medium Gaussian SVM 97.8%
    Coarse Gaussian SVM 88.6%
    Naïve Bayes
    Gaussian Naive Bayes 99.8%
    Kernel Naive Bayes 99.1%
    Decision Tree
    Fine Tree 100%
    Medium Tree 100%
    Coarse Tree 100%
    Ensemble
    Bagged Tree 100%
    Bootstrapping 26.5%
    RUSBoosted Tree 26.5%

     | Show Table
    DownLoad: CSV

    In Figure 3 shows the comparison between six kernels of the SVM classifier. Among all of them, linear, quadratic, and cubic SVM gave an accuracy of 100% with 0% misclassification rate while fine Gaussian, medium Gaussian, and coarse Gaussian gave 72.70, 97.80 and 88.60% accuracy rare respectively, with a misclassification rate of 27.3, 2.2 and 11.4%. In Figure 4 shows the reported accuracy of the Naïve Bayes classifier. This graph clearly shows that the reported accuracy rate is 99.80 and 99.10% when classified by Gaussian Naïve -Bayes and kernel Naïve Bayes, respectively, with the misclassification rate of 0.2 and 0.9%. In Figure 5 shows the accuracy of the decision tree classifier with its three kernels. Fine decision tree, medium decision tree, and coarse decision tree gave an accuracy rate of 100% with 0% misclassification rate when applied to our audio database. This shows that the proposed technique of MFCC gives the best accuracy results when classified by Decision Tree. In Figure 6 shows the reported accuracy from the Ensemble classifier and its kernels. Bagged trees, Boosted trees, and RUSBoosted trees gave the accuracy of 100, 26.50 and 26.50%, respectively, with the misclassification rate of 0, 73.5 and 73.5%.

    Figure 3.  Comparison of accuracy of different kernels of the SVM classifier.
    Figure 4.  Comparison between different classifiers of Naïve Bayes.
    Figure 5.  Comparison of accuracy of different classifiers of decision tree.
    Figure 6.  Comparison between kernels of ensemble classifiers.

    There are several studies based on the detection of voice pathologies. Vocal cord paralysis primarily impacts the patient's speech and breathing and often contributes to a lack of vocal capacity as all cords are paralyzed. The diagnostic test in order to classify these language pathologies is rather intrusive for the patient and, thus, the research and production of approaches to machine learning has risen in recent years. Similar to our approach but with different features, SVM was used as a classifier for SVD, and the reported accuracy 92% [15]. In [21], while assessing voice pathologies, when 13 conventional MFCC features were extracted, the highest reported accuracy is 72% using the SVM as a machine learning classifier. In the proposed methodology, we have used the freely accessible database of voice pathology, Saarbrucken Voice database (SVD). This database is recorded by the institute of phonetics of Saarland University, and it contains normal and pathological voice signals. We worked on four different voice disorders, which are laryngitis, cyst, non-fluency syndrome, and dysphonia, available on the SVD database. According to the database, we have extracted features that were required to perform training and testing have been extracted. As the voice signals are continuous signals, so we have split the signals into the frame of 10ms because the audio signal is itself is so small, so to increase the frames, we split them in 10ms and considered the voice signals to be stationary statistical. Twenty features were extracted from each frame of the voice signal. The features which were assumed to be extracted are 13 MFCC and pitch, spectral flux, spectral centroid, spectral roll-off, spectral entropy, short term energy, and zero crossing rates (ZCR). After the extraction of the feature, we applied support vector machine, decision tree, Naïve Bayes, and ensemble classifiers on the database by classification learner app on MATLAB software. A classification learner app was used to classify the extracted features. We applied each of the classifiers to know the results of every classifier on the publicly available database. The main focus of the study is to extract the seven new features along with 13 conventional MFCC features and to make such a combination that could give the best 10 ms audio signal as compared to the results discussed in the literature review. The inter classifier comparison concluded that the decision tree and Naïve Bayes give the highest accuracies, which to the best of our knowledge, have not been used for the selected set of features. The accuracies that classifiers show that using these twenty features, SVM gave 93.18% accuracy, Naïve Bayes gave 99.45% accuracy, decision tree gave 100% accuracy, and Ensemble Classifier gave 51% accuracy. The intercomparison approach provides accuracies that will enable future researchers to opt for the best fit classifier. Hence it is concluded that machine-based detection of voice disorders can be a breakthrough with such promising results. In the future, researchers can opt for artificial neural networks, and deep learning classifies with the same set of features used in the proposed methodology.

    There is no conflict of interest between authors.



    [1] R. O. J. H. Stutt, R. Retkute, M. Bradley, C. A. Gilligan, J. A. Colvin, Modelling framework to assess the likely effectiveness of facemasks in combination with 'lock-down' in managing the COVID-19 pandemic, Proceed. Royal Soc. A Math. Phys. Eng. Sci., 476 (2020), 20200376. https://doi.org/10.1098/rspa.2020.0376 doi: 10.1098/rspa.2020.0376
    [2] G. Giordano, F. Blanchini, R. Bruno, P. Colaneri, A. Di Filippo, A. Di Matteo, et al., Modelling the COVID-19 epidemic and implementation of population-wide interventions in Italy, Nat. Med., 26 (2020), 855–860. https://doi.org/10.1038/s41591-020-0883-7 doi: 10.1038/s41591-020-0883-7
    [3] I. Cooper, A. Mondal, C. Antonopoulos, A SIR model assumption for the spread of COVID-19 in different communities, Chaos Soliton. Fract., 139 (2020), 110057. https://doi.org/10.1016/j.chaos.2020.110057 doi: 10.1016/j.chaos.2020.110057
    [4] A. L. Bertozzi, E. Franco, G. Mohlerd, M. B. Shorte, D. Sledge, The challenges of modeling and forecasting the spread of COVID-19, Proc. Nat. Acad. Sci., 117 (2020), 16732–16738. https://doi.org/10.1073/pnas.2006520117 doi: 10.1073/pnas.2006520117
    [5] C. N. Ngonghala, E. Iboi, S. Eikenberry, M. Scotch, C. R. MacIntyre, M. H. Bonds, et al., Mathematical assessment of the impact of non-pharmaceutical interventions on curtailing the 2019 novel Coronavirus, Math. Biosci., 325 (2020), 108364. https://doi.org/10.1016/j.mbs.2020.108364 doi: 10.1016/j.mbs.2020.108364
    [6] S. E. Eikenberry, M. Mancuso, E. Iboi, T. Phan, K. Eikenberry, Y. Kuang, et al., To mask or not to mask: Modeling the potential for face mask use by the general public to curtail the COVID-19 pandemic, Infect. Disease Model., 5 (2020), 293–308. https://doi.org/10.1016/j.idm.2020.04.001 doi: 10.1016/j.idm.2020.04.001
    [7] J. Fernández-Villaverde, C. I. Jones, Estimating and simulating a SIRD Model of COVID-19 for many countries, states, and cities, J. Econom. Dynam. Control., (2022), 104318. https://doi.org/10.1016/j.jedc.2022.104318 doi: 10.1016/j.jedc.2022.104318
    [8] J. Osborn, S. Berman, S. Bender-Bier, G. D'Souza, M. Myers, Retrospective analysis of interventions to epidemics using dynamic simulation of population behavior, Math. Biosci., 341 (2021), 108712. https://doi.org/10.1016/j.mbs.2021.108712 doi: 10.1016/j.mbs.2021.108712
    [9] N. I. Stilianakis, Y. Drossinos, Dynamics of infectious disease transmission by inhalable respiratory droplets. J. Royal Soc. Interf., 7 (2010), 1355–1366. https://doi.org/10.1098/rsif.2010.0026 doi: 10.1098/rsif.2010.0026
    [10] M. Myers, J. Yan, P. Hariharan, S. Guha, A mathematical model for assessing the effectiveness of protective devices in reducing risk of infection by inhalable droplets, Math. Med. Biol., 35 (2018), 1–23. https://doi.org/10.1093/imammb/dqw018 doi: 10.1093/imammb/dqw018
    [11] J. Yan, P. Hariharan, S. Guha, M. Myers, Modeling the effectiveness of respiratory protective devices in reducing influenza outbreak, Risk Analysis, 39 (2019), 647–661. https://doi.org/10.1111/risa.13181 doi: 10.1111/risa.13181
    [12] F. Zhou, T. Yu, R. Du, G. Fan, Y. Liu, Z. Liu, et al., Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: A retrospective cohort study, Lancet, 395 (2020), 1054–1062. https://doi.org/10.1016/S0140-6736(20)30566-3 doi: 10.1016/S0140-6736(20)30566-3
    [13] A. Kerr, A Historical Timeline of COVID-19 in New York City, Investopedia, 2021.
    [14] C. Francescani, Timeline: The first 100 days of New York Gov. Andrew Cuomo's COVID-19 response, ABC News, 2020.
    [15] National Institute of Infectious Diseases, Japan, (2020). https://www.niid.go.jp/niid/en/2019-ncov-e/9407-covid-dp-fe-01.html
    [16] E. Nakazawa, H. Ino, A. Akabayashi, Chronology of COVID-19 cases on the diamond princess cruise ship and ethical considerations: A report from Japan, Disaster Med. Public Health Prep., 14 (2020), 506–513. https://doi.org/10.1017/dmp.2020.50 doi: 10.1017/dmp.2020.50
    [17] Harris County Commissioners Court Agenda, Harris County, Texas, (2020). https://agenda.harriscountytx.gov/COVID19Orders.aspx
    [18] Johns Hopkins Coronavirus Resource Center, (2020). https://coronavirus.jhu.edu/map.html
    [19] NYC Health, (2020). https://www1.nyc.gov/site/doh/covid/covid-19-data.page
    [20] The New York Times, (2020). https://www.nytimes.com/interactive/2020/us/new-york-coronavirus-cases.html
    [21] K. Mizumoto, G. Chowell, Transmission potential of the novel coronavirus (COVID-19) onboard the diamond Princess Cruises Ship, 2020, Infect. Disease Model., 5 (2021), 264–270. https://doi.org/10.1016/j.idm.2020.02.003 doi: 10.1016/j.idm.2020.02.003
    [22] Harris County/City of Houston COVID-19 Data Hub, (2020). https://covid-harriscounty.hub.arcgis.com
    [23] J. Howard, A. Huang, Z. Lik, Z. Tufekci, V. Zdimal, H. M. van der Westhuizen, et al., Face masks against COVID-19: An evidence review, Proc. Nat. Acad. Sci. U.S.A., 118 (2021), e2014564118. https://doi.org/10.1073/pnas.2014564118 doi: 10.1073/pnas.2014564118
    [24] A. Newman, Are New Yorkers wearing masks? Here's what we found in each borough, The New York Times, 2020. https://www.nytimes.com/2020/08/20/nyregion/nyc-face-masks.html
    [25] H. Enten, The Northeast leads the country in mask-wearing, CNN, June 26, 2020. https://www.cnn.com/2020/06/26/politics/maskwearing-coronavirus-analysis/index.html
    [26] M. Brenan, Americans' Face Mask Usage Varies Greatly by Demographics, Gallup, 2020. https://news.gallup.com/poll/315590/americans-face-mask-usage-varies-greatly-demographics.aspx
    [27] S. Guha, A. Herman, I. A. Carr, D. Porter, R. Natu, S. Berman, et al., Comprehensive characterization of protective face coverings made from household fabrics, PLoS One, 16 (2021), e0244626. https://doi.org/10.1371/journal.pone.0244626 doi: 10.1371/journal.pone.0244626
    [28] A. R. Ives, C. Bozzuto, Estimating and explaining the spread of COVID-19 at the county level in the USA, Commun. Biol., 4 (2021), 60. https://doi.org/10.1038/s42003-020-01609-6 doi: 10.1038/s42003-020-01609-6
    [29] S. Zhang, M. Diao, W. Yu, L. Pei, Z. Lin, D. Chen, Estimation of the reproductive number of novel coronavirus (COVID-19) and the probable outbreak size on the Diamond Princess cruise ship: A data-driven analysis, Int. J. Infect. Diseases, 93 (2020), 201–204. https://doi.org/10.1016/j.ijid.2020.02.033 doi: 10.1016/j.ijid.2020.02.033
    [30] K. T. L. Sy, L. F. White, B. E. Nichols, Population density and basic reproductive number of COVID-19 across United States counties, PLoS One, 16 (2021), e0249271. https://doi.org/10.1371/journal.pone.0249271 doi: 10.1371/journal.pone.0249271
    [31] American Society of Mechanical Engineers, ASME V & V 20-2009 - Standard for Verification and Validation in Computational Fluid Dynamics and Heat Transfer, ASME, New York, 2009.
  • This article has been cited by:

    1. Rinaldo M. Colombo, Mauro Garavello, Well Posedness and Control in a NonLocal SIR Model, 2020, 0095-4616, 10.1007/s00245-020-09660-9
    2. Rinaldo M. Colombo, Mauro Garavello, Francesca Marcellini, Elena Rossi, An age and space structured SIR model describing the Covid-19 pandemic, 2020, 10, 2190-5983, 10.1186/s13362-020-00090-4
    3. Hiroya Kikuchi, Hiroaki Mukaidani, Ramasamy Saravanakumar, Weihua Zhuang, 2020, Robust Vaccination Strategy based on Dynamic Game for Uncertain SIR Time-Delay Model, 978-1-7281-8526-2, 3427, 10.1109/SMC42975.2020.9283389
    4. Giacomo Albi, Lorenzo Pareschi, Mattia Zanella, Modelling lockdown measures in epidemic outbreaks using selective socio-economic containment with uncertainty, 2021, 18, 1551-0018, 7161, 10.3934/mbe.2021355
    5. Giacomo Albi, Lorenzo Pareschi, Mattia Zanella, Control with uncertain data of socially structured compartmental epidemic models, 2021, 82, 0303-6812, 10.1007/s00285-021-01617-y
    6. Christian John Hurry, Alexander Mozeika, Alessia Annibale, Vaccination with partial transmission and social distancing on contact networks, 2022, 2022, 1742-5468, 033302, 10.1088/1742-5468/ac50ae
    7. Martin Kröger, Mustafa Turkyilmazoglu, Reinhard Schlickeiser, Explicit formulae for the peak time of an epidemic from the SIR model. Which approximant to use?, 2021, 425, 01672789, 132981, 10.1016/j.physd.2021.132981
    8. Emanuele Bernardi, Lorenzo Pareschi, Giuseppe Toscani, Mattia Zanella, Effects of Vaccination Efficacy on Wealth Distribution in Kinetic Epidemic Models, 2022, 24, 1099-4300, 216, 10.3390/e24020216
    9. Reinhard Schlickeiser, Martin Kröger, Analytical Modeling of the Temporal Evolution of Epidemics Outbreaks Accounting for Vaccinations, 2021, 3, 2624-8174, 386, 10.3390/physics3020028
    10. Erivelton G. Nepomuceno, Márcia L. C. Peixoto, Márcio J. Lacerda, Andriana S. L. O. Campanharo, Ricardo H. C. Takahashi, Luis A. Aguirre, Application of Optimal Control of Infectious Diseases in a Model-Free Scenario, 2021, 2, 2662-995X, 10.1007/s42979-021-00794-3
    11. Alejandro Carballosa, José Balsa-Barreiro, Pablo Boullosa, Adrián Garea, Jorge Mira, Ángel Miramontes, Alberto P. Muñuzuri, Assessing the risk of pandemic outbreaks across municipalities with mathematical descriptors based on age and mobility restrictions, 2022, 160, 09600779, 112156, 10.1016/j.chaos.2022.112156
    12. Candy Sonveaux, Joseph J. Winkin, State feedback control law design for an age-dependent SIR model, 2023, 158, 00051098, 111297, 10.1016/j.automatica.2023.111297
    13. Fernando Córdova-Lepe, Juan Pablo Gutiérrez-Jara, A Dynamic Reaction-restore-type Transmission-rate Model for COVID-19, 2024, 21, 2224-2902, 118, 10.37394/23208.2024.21.12
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2695) PDF downloads(68) Cited by(3)

Figures and Tables

Figures(16)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog