Loading [MathJax]/jax/element/mml/optable/GeneralPunctuation.js
Research article

A new hybrid approach based on genetic algorithm and support vector machine methods for hyperparameter optimization in synthetic minority over-sampling technique (SMOTE)

  • Received: 10 October 2022 Revised: 30 January 2023 Accepted: 30 January 2023 Published: 15 February 2023
  • MSC : 68Txx, 68w01, 68w40, 68wxx, 82-08

  • The crucial problem when applying classification algorithms is unequal classes. An imbalanced dataset problem means, particularly in a two-class dataset, that the group variable of one class is comparatively more dominant than the group variable of the other class. The issue stems from the fact that the majority class dominates the minority class. The synthetic minority over-sampling technique (SMOTE) has been developed to deal with the classification of imbalanced datasets. SMOTE algorithm increases the number of samples by interpolating between the clustered minority samples. The SMOTE algorithm has three critical parameters, "k", "perc.over", and "perc.under". "perc.over" and "perc.under" hyperparameters allow determining the minority and majority class ratios. The "k" parameter is the number of nearest neighbors used to create new minority class instances. Finding the best parameter value in the SMOTE algorithm is complicated. A hybridized version of genetic algorithm (GA) and support vector machine (SVM) approaches was suggested to address this issue for selecting SMOTE algorithm parameters. Three scenarios were created. Scenario 1 shows the evaluation of support vector machine SVM) results without using the SMOTE algorithm. Scenario 2 shows that the SVM was used after applying SMOTE algorithm without the GA algorithm. In the third scenario, the results were analyzed using the SVM algorithm after selecting the SMOTE algorithm's optimization method. This study used two imbalanced datasets, drug use and simulation data. After, the results were compared with model performance metrics. When the model performance metrics results are examined, the results of the third scenario reach the highest performance. As a result of this study, it has been shown that a genetic algorithm can optimize class ratios and k hyperparameters to improve the performance of the SMOTE algorithm.

    Citation: Pelin Akın. A new hybrid approach based on genetic algorithm and support vector machine methods for hyperparameter optimization in synthetic minority over-sampling technique (SMOTE)[J]. AIMS Mathematics, 2023, 8(4): 9400-9415. doi: 10.3934/math.2023473

    Related Papers:

    [1] Yunfeng Shi, Shu Lv, Kaibo Shi . A new parallel data geometry analysis algorithm to select training data for support vector machine. AIMS Mathematics, 2021, 6(12): 13931-13953. doi: 10.3934/math.2021806
    [2] Bao Ma, Yanrong Ma, Jun Ma . Adaptive robust AdaBoost-based kernel-free quadratic surface support vector machine with Universum data. AIMS Mathematics, 2025, 10(4): 8036-8065. doi: 10.3934/math.2025369
    [3] Thavavel Vaiyapuri, Prasanalakshmi Balaji, S. Shridevi, Santhi Muttipoll Dharmarajlu, Nourah Ali AlAseem . An attention-based bidirectional long short-term memory based optimal deep learning technique for bone cancer detection and classifications. AIMS Mathematics, 2024, 9(6): 16704-16720. doi: 10.3934/math.2024810
    [4] Wahida Mansouri, Amal Alshardan, Nazir Ahmad, Nuha Alruwais . Deepfake image detection and classification model using Bayesian deep learning with coronavirus herd immunity optimizer. AIMS Mathematics, 2024, 9(10): 29107-29134. doi: 10.3934/math.20241412
    [5] Hongxin Xue, Lingling Zhang, Haijian Liang, Liqun Kuang, Huiyan Han, Xiaowen Yang, Lei Guo . Influenza trend prediction method combining Baidu index and support vector regression based on an improved particle swarm optimization algorithm. AIMS Mathematics, 2023, 8(11): 25528-25549. doi: 10.3934/math.20231303
    [6] Peng Wang, Jiajun Huang, Weijia He, Jingqi Zhang, Fan Guo . Maximum likelihood DOA estimation based on improved invasive weed optimization algorithm and application of MEMS vector hydrophone array. AIMS Mathematics, 2022, 7(7): 12342-12363. doi: 10.3934/math.2022685
    [7] Farah Liyana Azizan, Saratha Sathasivam, Nurshazneem Roslan, Ahmad Deedat Ibrahim . Logic mining with hybridized 3-satisfiability fuzzy logic and harmony search algorithm in Hopfield neural network for Covid-19 death cases. AIMS Mathematics, 2024, 9(2): 3150-3173. doi: 10.3934/math.2024153
    [8] Nuha Alruwais, Hayam Alamro, Majdy M. Eltahir, Ahmed S. Salama, Mohammed Assiri, Noura Abdelaziz Ahmed . Modified arithmetic optimization algorithm with Deep Learning based data analytics for depression detection. AIMS Mathematics, 2023, 8(12): 30335-30352. doi: 10.3934/math.20231549
    [9] Almudena P. Márquez, Francisco Javier García-Pacheco, Míriam Mengibar-Rodríguez, Alberto Sánchez-Alzola . Supporting vectors vs. principal components. AIMS Mathematics, 2023, 8(1): 1937-1958. doi: 10.3934/math.2023100
    [10] Senyuan Yang, Bo Yu, Jianxin Pan, Wuliang Yin, Hua Wang, Kai Yang, Qingtai Xiao . Application of a hybrid nonlinear algorithm driven by machine learning and feature importance identification for temperature control prediction of the bath smelting process. AIMS Mathematics, 2025, 10(6): 13104-13129. doi: 10.3934/math.2025588
  • The crucial problem when applying classification algorithms is unequal classes. An imbalanced dataset problem means, particularly in a two-class dataset, that the group variable of one class is comparatively more dominant than the group variable of the other class. The issue stems from the fact that the majority class dominates the minority class. The synthetic minority over-sampling technique (SMOTE) has been developed to deal with the classification of imbalanced datasets. SMOTE algorithm increases the number of samples by interpolating between the clustered minority samples. The SMOTE algorithm has three critical parameters, "k", "perc.over", and "perc.under". "perc.over" and "perc.under" hyperparameters allow determining the minority and majority class ratios. The "k" parameter is the number of nearest neighbors used to create new minority class instances. Finding the best parameter value in the SMOTE algorithm is complicated. A hybridized version of genetic algorithm (GA) and support vector machine (SVM) approaches was suggested to address this issue for selecting SMOTE algorithm parameters. Three scenarios were created. Scenario 1 shows the evaluation of support vector machine SVM) results without using the SMOTE algorithm. Scenario 2 shows that the SVM was used after applying SMOTE algorithm without the GA algorithm. In the third scenario, the results were analyzed using the SVM algorithm after selecting the SMOTE algorithm's optimization method. This study used two imbalanced datasets, drug use and simulation data. After, the results were compared with model performance metrics. When the model performance metrics results are examined, the results of the third scenario reach the highest performance. As a result of this study, it has been shown that a genetic algorithm can optimize class ratios and k hyperparameters to improve the performance of the SMOTE algorithm.



    As pointed out in Claessens et al. (2014), global financial crises underline the importance of innovative modelling approaches to financial markets. The quantum mechanics approach suggests an alternative way to describe the unpredictable stock market behaviour (see e.g., Baaquie (2009)).

    This paper is motivated by the quantum mechanics approach to the actuarial modelling and risk analysis as initiated in Tamturk and Utev (2018) and developed in Lefèvre et al. (2018), Tamturk and Utev (2019). We modify and adapt their methods to derive and analyze a quantum-type financial modelling. Specifically, the Schrödinger, Heisenberg, Feynman, Dirac approaches in quantum mechanics (see, e.g., Griffiths and Schroeter (2018); Parthasarathy (2012); Plenio (2002)) will be applied to the well-known option pricing models of Black and Scholes (1973) and Cox et al. (1979).

    Moreover, motivated by the data analysis of the quantum reserve process proposed in Lefèvre et al. (2018), we model the financial data as eigenvalues of certain 1 or 2-step observable operators. These data then are analyzed to identify price jumps using supervised machine learning tools such as k-fold cross-validation techniques (see, e.g., Bishop (2006); Hastie et al. (2009); Wittek (2014)).

    Option pricing models using quantum techniques discussed, for example, in Baaquie(2004, 2014); Bouchaud and Potters (2003); Haven (2002) are often based on the Schrödinger wave function with Hamiltonian operator H and are mainly oriented to continuous-time markets. For discrete-time markets as considered here, following Chen(2001, 2004), , we choose the discrete-time formalism and analyze the quantum version of the Cox-Ross-Rubinstein binomial model. Then, we establish the limit of the spectral measures providing the convergence to the geometric Brownian motion model. We also identify the limit of the N-step non-self adjoint bond market as a planar Brownian motion.

    The paper is organized as follows. Section 2 deals with heterogeneous quantum binomial markets. In Section 3, two convergence properties to continuous-time quantum markets are obtained. In Section 4, discrete and continuous-time quantum mechanics techniques are applied to the problem of option pricing. Section 5 is devoted to an analysis of stock market data.

    The motivation to our models come from the quantum mechanics approach to insurance proposed in Tamturk and Utev (2018, 2019), the insurance claim data analysis via quantum tools introduced in Lefèvre et al. (2018) and the non-traditional financial modelling initiated in Ma and Utev (2012) and developed in Karadeniz and Utev(2015, 2018). The quantum modelling approach is partly inspired by Baaquie (2004); Chen(2001, 2004), and Parthasarathy (2012).

    We begin by outlining some standard arguments for a quantum-type modelling (see e.g., Lefèvre et al. (2018); Parthasarathy (2012)). An observable is a linear operator (matrix) on a certain Hilbert space. The quantum product of two independent observables A,B is implemented by the tensor product of the observables AB. So, ln(AB) acts as a quantum sum of two independent observables; in particular, AA is the quantum product of two independent identical observables.

    The 1-step quantum geometric random walk is defined as a 2×2 matrix A with eigenvalues eu,ed. Thus, AN (N0) models the N-step geometric random walk.

    The quantum type modification of the classical Cox-Ross-Rubinstein model is originated in Chen(2001, 2004). The dynamics per period is defined by two moves: eu (the share price goes up) and ed (the share price goes down) with d<0<u.

    The quantum binomial model over N periods is then represented by the share price operator

    HSN=S0HN, (1)

    where the main 1-step observable H=A is a 2×2 self-adjoint (hermitian) matrix with eigenvalues eu,ed and representation

    A=UDU=U(eu00ed)U, (2)

    where U is a 2×2 unitary matrix.

    In the sequel, the quantum binomial model discussed will be heterogeneous with share price operator

    HSN=S0H1HN. (3)

    The motivation to this relatively new financial model comes from Lefèvre et al. (2018) quantum mechanics approach to non-life insurance and the Lamplighter group approximation to the financial modelling suggested in Ma and Utev (2012). Based on the approach to the financial data analysis developed in Ma and Utev (2012); Karadeniz and Utev(2015, 2018) (see also references therein), we treat the data as having big jumps, say eu, small jumps, say ed or no jump. Moreover, the financial stock is observed at fixed times Δk but the number of jumps occurred during the time period ((k1)Δ,kΔ] is not observed. For simplicity and following Lefèvre et al. (2018), we then assume there are at most two jumps per period.

    In this circumstance, the main observable operator is given, similarly to the insurance case Lefèvre et al. (2018), by

    H3=S0(P0I4+P1(AI2)+P2(A2)), (4)

    where In is a n×n identity matrix, A is the 2×2 matrix representing the 1-step geometric move operator (2) with eigenvalues eu,ed, and P0,P1,P2 are 3×3 matrix projection operators corresponding to the 0,1,2 claim occurrence operators and defined by

    Pi=WDi+1|3W,i=0,1,2, (5)

    where Di+1|3 is a 3×3 diagonal matrix which has all its elements equal to 0 except the (i+1,i+1)-th with value 1, and W is a 3×3 unitary matrix.

    The share price operator over N periods is then defined as

    HSN=S0HN3=S0(P0I4+P1(AI2)+P2(A2))N. (6)

    Furthermore, when constructing the density operator, we consider the following two cases.

    Maxwell-Boltzmann statistics. In this case, the jump sizes are i.i.d. with a two-point distribution. More precisely, there are 0,1,2 jumps with probabilities δ0,δ1,δ2 given via a Poisson process, and each jump size has two possible values ed,eu with probabilities q,p (see also later in Section 4.1).

    Bose-Einstein statistics. In this case, the claim sizes are dependent but remain independent of the claim occurrences.

    This model makes a bridge between the traditional binomial model and the actuarial type model. In this case, the dynamics per period is defined by three moves: no change, down and up. The trinomial type financial modelling is a well established topic in finance (see e.g., Boyle (1986); Tian (1993); Leisen and Reimer (1996)).

    Now, the 1-step observable operator H2=B is the 3×3 self-adjoint matrix with eigenvalues eu,1,ed corresponding to these three moves and is given by

    H2=B=UDU=U(eu0001000ed)U, (7)

    where U is now a 3×3 unitary matrix.

    Then, the share price operator over N periods is defined by

    HSN=S0HN2=S0(U)NDNUN. (8)

    Remarks. Notice that the heterogeneous versions of the trinomial model and the actuarial type model are available with the representation similar to (3). However, they are not treated in this paper because the main purpose for considering these two models is the non-standard data analysis (to be presented in Section 5). Although the data analysis of time dependent models is a fascinating topic, it is out of the scope of this paper.

    As mentioned in above, the quantum binomial model (1) is originated in Chen(2001, 2004). Our presentation is somewhat different and based on the algebraic tensor product properties. In addition, since heterogeneity is an important topic in option pricing (see e.g., Benninga and Mayshar (2000)), we treat a slightly more general time dependent or heterogeneous quantum binomial model (3).

    A quantum state ρ is defined as a positive self-adjoint operator with trace tr(ρ)=1. We recall the following properties.

    Lemma 2.1. Let A,B,C,D be self-adjoint operators, U a unitary matrix, f a function of observables and ρ a quantum state. Then,

    tr(ABC)=tr(CAB),tr(AB)=tr(A)tr(B),(AB)(CD)=(AC)(BD),(AB)=AB,A,B0AB0,A=UDUf(A)=Uf(D)U,UρU=quantum state. (9)

    For simplicity, we choose a quantum state ρ as a tensor product, i.e.

    ρ=ρ1ρN, (10)

    where each ρi is a self-adjoint non-negative 2×2 matrix such that tr(ρi)=1. From Lemma 2.1, we see that ρ is a proper quantum state.

    The risk-neutral world of the quantum Black-Scholes model consists of self-adjoint non-negative 2×2 matrices ρi that satisfy

    tr(ρiHi)=1+ri,i=1,,N, (11)

    where ri is the risk-free interest rate for the period i.

    Let us define

    Hi=UiDiUi,and let~ρi=UiρiUi,i=1,,N. (12)

    Note that the ~ρi are also quantum states. In addition, ˜ρi has non-negative diagonal elements q(i)u,q(i)d and it can be shown to have the representation

    ˜ρi=(quixˉxqdi), (13)

    where

    qui=1+ridiuidi,qdi=1qui. (14)

    The transformed operator ˜ρ is then obtained from Lemma 2.1 (third property) as

    ˜ρ=U(ρ1...ρN)U=(U1UN)(ρ1...ρN)(U1UN)=(U1ρ1U1)(UNρNUN)=˜ρ1˜ρN, (15)

    after using (12), and ˜ρ is again a quantum state. For the classical probability case, that is when all matrices are commutative, this transform is the form of the change of measure.

    Moreover, the quantum no arbitrage condition (11) is satisfied for the transformed density, i.e.

    tr(~ρiDi)=1+ri,i=1,,N. (16)

    Consider a payoff function f and a discount factor

    dN=1/(1+r1)(1+rN). (17)

    From (3), the price of the general option without arbitrage OP(f(HSN)) for the N-step quantum binomial model is then defined by

    OP(f(HSN))=dNtr(ρf(HSN))=dNtr(ρ1...ρNf(S0H1...HN)). (18)

    Applying properties given in Lemma 2.1, we then obtain

    OP(f(HSN))=tr(ρ1...ρNf(S0H1...HN))(1+r1)(1+rN)=tr(ρ1...ρNUf(S0D1...DN)U)(1+r1)(1+rN)=tr([U(ρ1...ρN)U]f(S0D1...DN))(1+r1)(1+rN)=tr(˜ρ1˜ρNf(S0D1...DN))(1+r1)(1+rN), (19)

    after using (15). Since ˜ρ is a tensor product and f(S0D1...DN) is a diagonal matrix, we deduce the option pricing formula (20) below.

    Theorem 2.2. For the heterogeneous quantum binomial model,

    OP(f(HSN))=σf(S0yσ)qσ(1+r1)(1+rN), (20)

    where the index σ denotes any feasible path yσ of the form

    yσ=eσ1eσNwithσi{ui,di}, (21)

    which occurs with the probability qσ=qσ1qσN where qσi{qui,qdi}.

    In particular, for the homogeneous case (the quantum Cox-Ross-Rubinstein model) where for all i, ri=r,ui=u,di=d with qui=qu,qdi=qd, the formula (20) reduces to

    OP(f(HSN))=1(1+r)NNn=0f(S0euned(Nn))qndqNnu. (22)

    Non-self adjoint quantum binomial market. Assume that Hi are invertible 2×2 matrices with two different eigenvalues ed and eu but no more self-adjoint, in general. In this case the transformed density ˜ρi are no more proper states, in general. However, they still have same diagonal elements qui and qdi and so the transformed matrix ˜ρ defined in (15) again has same diagonal elements as in the self-adjoint case, but is not a proper state, in general. By inspecting the proof, we see that Theorem 2.2 holds true in this case as well.

    In this section, we consider the homogeneous quantum binomial model, and we discuss two examples on the convergence to continuous-time markets, namely the Black-Scholes model and the planar Brownian motion.

    Theorem 3.1. Let μN be the measure of the eigenvalues of HN with respect to the quantum state ρN. Suppose that r=λ/N and u=d=σN1/2. Then, as N,

    f(S0x)dμN(x)f(S0x)dμ(x), (23)

    where μ is a lognormal distribution (i.e. μ(x)=P(ea+σZx) for suitable constants a,σ and Z a standard normal variable).

    Proof. We begin with the representation via the spectral measure μN. Observe that

    Ef(HSN)=tr(ρNf(S0HN))=f(S0x)μN(dx), (24)

    where μN is the measure of the eigenvalues λσ of HN with respect to the quantum probability tr(ρNHN). Since λσ=eσ1eσN with σi{ui,di}, we get

    μN(x)=σ:λσxqσ=σ:eσ1eσNxqσ1qσN, (25)

    where qσi{qu,qd} respectively.

    Let us move on to the weak convergence desired. We can write that

    qσ1qσN=P(Y1=eσ1)P(YN=eσN)=P(Y1=eσ1,,YN=eσN), (26)

    where the Yi are i.i.d. variables with P(Yi=eu)=qu,P(Yi=ed)=qd. Thus,

    μN(x)=P(Y1YNx)=P(lnY1++lnYNlnx)=P(TNlnx), (27)

    with TNlnY1++lnYN. As N, we obtain from the central limit theorem that

    μN(x)=P(TNlnx)P(ea+σZx)=μ(x), (28)

    for certain constants a,b and Z a standard normal. This gives the limit result (23).

    Remarks. (ⅰ) The limiting measure μ corresponds to that of the geometric Brownian motion St=S0e(ρσ2/2)t+σBt when S0=1, t=1 and ρσ2/2=a.

    (ⅱ) It would be interesting to compare the technique with the semi-classical approximation, such as expanding the action around the classical path (see (48) in Contreras et al. (2010)). Another interesting question is to analyse connection with arbitrage as discussed in Haven (2002) and Contreras et al. (2010).

    (ⅲ) The Cauchy transform is an alternative approach to deal with quantum probabilities (see Mudakkar and Utev (2013)). In particular,the convergence of spectral measures is reduced to the convergence of the Cauchy transforms. We recall that the Cauchy transform for the measure μ is defined by

    Sμ(z)=Rμ(dt)tz,whereμ((a,b))=limϵ01πbaSμ(x+iϵ)dx, (29)

    for all zCR={zC:z0} and open intervals (a,b) with μ({a,b})=0. The goal is then to show that Sμn(z)Sμ(z). However,the common approach of the moment expansion does not work in this case since

    dμ(x)zx=k=01zk+1xkdμ(x)=k=01zk+1E(ek(a+σZ))=k=01zk+1eka+k2σ2/2=. (30)

    In this part,by treating the relatively simple bond market,we show that non-commutative markets provide a richer class of financial models (see also Haven (2002); Contreras et al. (2010); Herscovich (2016)).

    Bond market. Assume that returns are non-risky,that is the outcomes are equal. In the quantum setup,we suppose that matrix H has a single eigenvalue,eu say.

    Self-adjoint quantum bond market. Let ρN denote a quantum state defined as before. Assume that in addition operator H is self-adjoint. Then H=euI2 and

    HSN=S0HN=S0euNIN2andf(HSN)=f(S0euN)IN2. (31)

    In particular,any option claim f(HSN) is commutative with the state ρN which implies that the self-adjoint quantum bond market is commutative,that is equivalent to the simple classical probability financial market with two non-risky assets. The no-arbitrage condition becomes global to give eu=tr(ρH)=1+r since tr(ρ)=1 which restates that under no-arbitrage the non-risky returns are equal. Moreover,the option price for f(HSN) is provided by

    OP(f(HSN))=tr(ρNf(S0HN))(1+r)N=euNf(S0euN). (32)

    Non self-adjoint quantum bond market. Now,assume that H is no longer self-adjoint,for example a product of two noncommutative self adjoint observables. Thus,non self-adjoint quantum bond market is noncommutative,in general. In this case,we show that the limit is sensitive to the density state ρ and the representation of H. For simplicity,we assume that the basic observable H and the state ρ are defined by

    Hu=(eu100eu100eu),ρ=(ρ11yδyρ220δ0ρ33), (33)

    i.e.,Hu is now a 3×3 Jordan matrix. The bond price process is then defined by

    HSN=S0HNu. (34)

    To satisfy the no-arbitrage condition,we ask that

    tr(ρHu)=1+r,thusy=1+reu. (35)

    Theorem 3.2. Suppose that r=λ/N,u=a/N and δ=Δ/N with Δ0. Then,for any positive integer k,as N,

    OP([HSN]k)eλSk0E[ek(λ+Δ/2+iBΔ)], (36)

    regardless of a, where i is the imaginary unit and BΔ is a Brownian motion at time Δ.

    Proof. Notice that

    Hku=(ekue(k1)uke(k2)uk(k1)/20ekue(k1)uk00eku). (37)

    From the definition of HSN and ρ, we then get

    OP([HSN]k)=OP([S0(HN)]k)=(1+r)Ntr(ρN[S0(HN)]k)=(1+r)Ntr(Sk0ρN[(Hk)N])=Sk0(1+r)Ntr([ρHk]N)=Sk0(1+r)N(ρ11euk+ρ22euk+ρ33euk+yeu(k1)k+δe(k2)uk(k1)/2)N=Sk0(1+r)N(euk+yeu(k1)k+δe(k2)uk(k1)/2)N=Sk0(1+r)N[1+ak/N+(λa)k/NΔk(k1)/2N+O(1/N2)]NSk0eλe(λ+Δ/2)kk2Δ/2 (38)

    after using the assumptions made on r,u,δ. This provides the limit result (36).

    Remarks. (ⅰ) This representation is useful in computing the European call option OP([HSNK)+]) via the Fourier techniques combined with the Monte Carlo approximation. For the distribution of this process, we view R2 as the complex plane and the planar Brownian motion ~Bt=(B1(t),B2(t)) is then interpreted as a complex-valued Brownian motion.

    (ⅱ) Although observables are traditionally considered to be self-adjoint, the non self-adjoint data is modelled by considering an observable such as H=AB, i.e. the product of two self-adjoint matrices A,B where A and B may represent the non-trade and trade time changes, respectively.

    (ⅲ) Similarly to the convergence to the Black-Scholes model, the limit does not depend on the shift parameter a. However, now the limit depends on the mysterious characteristic Δ.

    In this section, we will apply some methods of quantum mechanics to finance. Discrete and continuous-time markets must be treated separately because of different stochastic behaviors. To simplify the presentation, we assume that the interest rate is 0 and that the risky processes for share prices are martingales.

    The use of Dirac-Feynman quantum mechanics techniques for insurance risk modelling was initiated in Tamturk and Utev (2018) and then developed in Lefèvre et al. (2018); Tamturk and Utev (2019). We adapt this approach to the problem of option pricing in finance, in particular for the pricing of path-dependent options.

    In the Dirac formalism, bra-ket notation is a standard way of describing quantum states. Consider a class of n×n matrices treated as C algebra. A column vector x corresponds to a ket-vector |x>. An associated bra-vector <x| is a row vector defined as its Hermitian conjugate. The usual inner product is denoted by <x|y>, while the outer product |x><y| is the operator/matrix defined by

    |x><y||z>=<y|z>|x>(abc=bca rule). (39)

    In the Feynman path integral methods, the transition probability P(xjxj+1) is computed as the propagator <xj|Aj+1|xj+1> when Aj is a Markovian operator. Thus, the typical path is written as |x0>→|x1>→|xn>, and its probability is given by

    P(|x0>→|x1>→|x2|xn>)=<x0|A1|x1><x1|A2|x2><xn1|An|xn>. (40)

    The main ingredient is then the path calculation formula that calculates the probability P(x0xn) via the sum of the probabilities on all the appropriate paths, i.e.

    P(x0xn)=<x0|A1A2An|xn>=x1,,xn1<x0|A1|x1><x1|A2|x2><xn1|An|xn>. (41)

    It remains to define a suitable propagator for discrete time. For simplicity, we take the operators Aj all equal to A and the time intervals Δti all equal to Δt. In a similar way to e.g. Baaquie (2004); Tamturk and Utev (2018), we assume that the operator A is defined via an Hamiltonian operator H such as A=eΔtH, where H is a Markovian generator called Markovian Hamiltonian. Thus, P(xixi+1)=<xi|eΔtH|xi+1> which can be computed applying the Fourier transform to the momentum space (e.g., Griffiths and Schroeter (2018); Tamturk and Utev (2018)). Specifically, let |p> be a basis in that space, and write <x|p>=eipx and <p|x>=eipx. Then, we get

    <xi|eΔtH|xi+1>=2π0dα2π<xi|eΔtH|α><α|xi+1>=2π0dα2π<xi|α><α|xi+1>eΔtKα=12π2π0(eixiαeixi+1α)eΔtKαdα, (42)

    where {|α>,Kα} is the set of eigenvectors and eigenvalues of the Hamiltonian operator H (i.e., H|α>=Kα|α>). Therefore, we deduce from (41), (42) that

    P(x0xn)=x1,,xn1<x0|eΔtH|x1><x1|eΔtH|x2><xn1|eΔtH|xn>=x1,,xn11(2π)n2π02π02π0[n1i=0(eixiαieixi+1αi)eΔtKαi]dα0dα1dαn1. (43)

    For a more detailed overview of the theory, we refer the reader to the books by Feynman and Hibbs (2010); Griffiths and Schroeter (2018); Parthasarathy (2012) and Plenio (2002).

    Option pricing formula. Consider a claim of the form C=f(S0,S1,...,SN), using the notation of Section 2. From (43), we obtain for the corresponding option price

    OP(C)=dNS0,S1,...,SNf(S0,S1,...,SN)<S0|A1|S1><SN1|AN|SN>. (44)

    The discrete-time approach followed to derive (43) can then be easily applied to the current formula (44).

    A modified Cox-Ross-Rubinstein model. Consider a discrete-time market in which, during each i-th time interval Δt, the share price Si can

    (.) have 1 jump giving Si+1=Sieu,Sied with probabilities p,q=1p, or

    (.) have 2 jumps giving Si+1=Sieu+d,Sie2u,Sie2d with probabilities 2pq,p2,q2, or

    (.) remain the same giving Si+1=Si,

    where d,u are integers with d<0<u. Furthermore, the possible jumps arrive according to a Poisson process of parameter λ so that

    (.) δ1P[N(Δt)=1]=eλΔt(λΔt),

    (.) δ2P[N(Δt)=2]=eλΔt(λΔt)2/2,

    (.) δ0P[N(Δt)=0]1δ1δ2.

    Define xi=ln(Si). The transition probabilities P(SiSi+1) are equivalent to P(xixi+1). Set Δt=1, say. We can now apply (42) with A=eH where the set {|α>,Kα} is found by solving the Schrödinger equation eH|α>=eKα|α> in which

    eH|α>(x)=Ex(eiαxi+1)=eiαx(δ0+eiαupδ1+eiαdqδ1+eiα2up2δ2+eiαx2dq2δ2+eiα(u+d)2pqδ2). (45)

    Notice that the Hamiltonian H is not Markovian, but (42) is still applicable to the discrete times kΔt. Moreover, by construction, the martingale probabilities p=qu and q=qd for the Cox-Ross-Rubinstein model (without interest) yield a martingale in the present situation too since we have

    E(Si+1|Si)=Si[δ0+δ1(queu+qded)+δ2(e2uq2u+e2dq2d+eq+d2quqd)]=Si(δ0+δ1+δ2)=Si. (46)

    We have illustrated numerically the option pricing results obtained for the model. The tables and figures are however too large to be included here.

    This short part is mostly a review of the application of quantum mechanics approach to the continuous time markets and closely follows Baaquie (2004).

    The continuous time formalism is based on the Fourier transform of tempered distributions on the basis |p> in the momentum space. Consider a risky asset price St that evolves in function of an Hamiltonian operator H. First, the method is applied to compute the pricing kernel

    p(x,τ;x)=<x|eτH|x>=dp2π<x|eτH|p><p|x>. (47)

    Then, the option price at time t for the claim QQ(ST), T>t, given St=x is defined by

    OP(Q(ST)|St=x)=<x|e(Tt)H|Q>. (48)

    Black-Scholes model. In this classical approach, the stock price St follows a geometric Brownian motion, i.e. St=S0e˜Bt where ˜Bt=μt+σBt (μ is the drift, σ the volatility and Bt a standard Brownian motion).

    We apply the quantum mechanics approach. Motivated by Baaquie (2004), we work with ˜Bt rather than with St. The corresponding Hamiltonian for the Brownian motion is Hf=(1/2)f (computed, for example, via the Itô formula). The Brownian motion kernel is then defined from (47) by the normal density function

    \begin{equation} p(u, \tau ;u') = f_{N(\tau \mu, \tau\sigma^2)}(u'-u) = \frac{1}{\sqrt{2 \pi \tau \sigma^2}} \exp[-(1/2\tau\sigma^2)(u'-u -\tau\mu)^2]. \end{equation} (49)

    To obtain the option price (48), we set x = S_0e^u and Q(x) = Q(S_0e^u)\equiv g(u) . Then, the Feynman-Kac formula yields

    \begin{align} OP(Q(S_T)|S_t = x) = \lt u|e^{-(T-t) H}|g \gt & = \int^{\infty}_{-\infty} \lt u|e^{-(T-t)H}|u' \gt g(u')du' \\ & = \int^{\infty}_{-\infty} p(u, T-t ;u') Q(S_0e^{u'})du'. \end{align} (50)

    Via path integrals. To find the pricing kernel p(x, \tau; x') = < x|e^{-\tau H}|x' > for \tau = T-t , an alternative method consist in using path integral methods. For this, we discretize the time in N intervals of length \Delta and consider the x_i = x(t_i) where t_i = i \Delta . We then proceed as in the situation in discrete-time. The pricing kernel for (x, x') = (x_0, x_N) becomes

    \begin{equation} p(x, N\Delta;x') = \int \int \ldots \int dx_1 dx_2 \ldots dx_{N-1} \, \, \prod\limits^N_{i = 1} \lt x_{i-1}|e^{-\Delta H}|x_{i} \gt . \end{equation} (51)

    Applications to non-life insurance. Consider Feynman's modification \tilde{H} of the Brownian motion Hamiltonian H by adding the potential V , i.e. for the Hamiltonian \tilde{H} = H+V . For example, choose a ruin level B and take V(x) = +\infty for x < B . Then, the path calculation formula (43) allows us to compute ruin probabilities when B = 0 (Tamturk and Utev (2018, 2019)) and exotic options with barriers when B > 0 .

    Finally, let us mention that numerically, the binomial model formula (20) and the path integral approach (51) were found to give results close to the Black-Scholes formula, even for relatively small values of N (of order 40 ).

    We are going to analyse a generated set of financial data by choosing two different quantum models described in Section 2.1. First, we consider the quantum actuarial-type model (see Section 2.1.3 and in 4.1) with the N -step observable defined in (6). Then, we consider the quantum trinomial model (see Section 2.1.2) with the N -step observable defined in (8). Note that matching 1 -step for the actuarial case to two steps for the trinomial model is natural, since in the actuarial case, we choose at most two jumps.

    We apply supervised machine learning methods such as developed e.g. in the books by Bishop (2006); Hastie et al. (2009) and Wittek (2014).

    Overall approach. The dataset is supposed to come from a non-ordered class of randomly perturbed observables. Each data is the observation of an eigenvalue \lambda of the observable perturbed by i.i.d. error terms. The observable is the 1 -step operator for the actuarial-type model and the 2 -step oparator for the trinomial model. Matching 1 -step in the actuarial case to 2 -step in the trinomial case is natural since there are at most two jumps in the actuarial case considered.

    First, the data are classified in classes G_\lambda with respect to the eigenvalues \lambda . Then, the probabilities p_\lambda are estimated by maximum likelihood using Maxwell-Boltzmann or Bose-Einstein statistics. Finally, the \lambda are estimated via the weighted L_1 -norm risk error function.

    Let us explain in more detail for the actuarial-type model, for example.

    Step 1. An initial (u = u_0, u = d_0) is chosen randomly.

    Step 2. For a given (u, d) , the data is classified and labeled against the eigenvalues of the observable using a nearest neighbor algorithm. This leads to the classes G_{\lambda} .

    Step 3. For the same (u, d) , the estimates \hat{p} and \hat{q} are obtained by maximizing the likelihood function L(p, q).

    Step 4. The (u, d) is updated by minimizing a weighted L_1 -norm risk error function F(u, d) \equiv F(\lambda) defined by

    \begin{equation} F(\lambda) = \|\beta-\lambda\| = \sum\limits_\lambda p_\lambda\sum\limits_{\beta_i\in G_\lambda} |\beta_i-\lambda|. \end{equation} (52)

    Step 5. The loop of steps 2 to 4 is repeated until the relative error becomes smaller than a selected difference M , i.e. when

    \begin{equation} |F(u_{i+1}, d_{i+1})-F(u_{i}, d_{i})| \lt M. \end{equation} (53)

    k- fold cross-validation. To reduce the risk of error, we use a k- fold cross-validation strategy. The dataset is randomly divided into k subsets of equal size. One of the subsets is chosen as the training set and the others as test sets. The process is repeated k times, each subset constituting a training element. At each iteration, steps 1 to 5 above are applied to the training data and the results obtained are then checked in the test data. Finally, the estimates used are an average of those obtained on the k iterations.

    Numerical example. As a simple illustration, we will consider the following dataset

    \begin{equation} V = \{86, 63, 35, 52, 41, 8, 24, 12, 19, 24, 42, 5, 91, 95, 50, 49, 34, 91, 37, 11\}. \end{equation} (54)

    From the assumptions of the model, the 1 -step observable H has the eigenvalues

    \begin{equation} \{\lambda\} = \{1, e^{d}, e^{u}, e^{d+u}, e^{2d}, e^{2u}\}, \end{equation} (55)

    with probabilities respectively given by

    \begin{equation} \{p_\lambda\} = \{\delta_0, q\delta_1, p\delta_1, p_{d+u}\delta_2, p_{2d}\delta_2, p_{2u}\delta_2\}. \end{equation} (56)

    For the probabilities p_{d+u}, q_{2d}, p_{2u} , we consider two possible statistics often used in the analysis of quantum observables.

    Maxwell-Boltzmann independence. This case yields the binomial model since

    \begin{equation} p_{d+u} = 2pq, \, p_{2d} = q^2, \, p_{2u} = p^2. \end{equation} (57)

    Bose-Einstein dependence. In this case, corresponds to probabilities

    \begin{equation} p_{d+u} = Cpq, \, p_{2d} = Cq^2, \, p_{2u} = Cp^2, \quad \mbox{where } C(pq+q^2+p^2) = 1. \end{equation} (58)

    As pointed out in Lefèvre et al. (2018), both statistics admit a formal construction, via the proper choice of the density operator \rho for the number of occurrences and the density projection operators \rho_\lambda , such that

    \begin{equation} tr(\rho \otimes \rho_{\lambda}) = p_\lambda. \end{equation} (59)

    Likelihood functions. Denote by \#x the number of x observed in the data set. For the Maxwell-Boltzmann statistics, the likelihood is defined by the probabilities p , q , \delta_0 , \delta_1 and \delta_2 such that

    \begin{equation} L(p, q) = L(p, q, \delta_0, \delta_1, \delta_2) = (\delta_0)^{\#0}(q\delta_1)^{\#d}(p\delta_1)^{\#u}(2pq\delta_2)^{\#u+d}(q^2\delta_2)^{\#2d}(p^2\delta_2)^{\# 2u}. \end{equation} (60)

    For the Bose-Einstein statistics, the likelihood function is modified as

    \begin{equation} L(p, q) = L(p, q, \delta_0, \delta_1, \delta_2) = (\delta_0)^{\#0}(q\delta_1)^{\#d}(p\delta_1)^{\#u}(Cpq\delta_2)^{\#u+d}(Cq^2\delta_2)^{\#2d}(Cp^2\delta_2)^{\# 2u}. \end{equation} (61)

    Risk functions. From (55), the scaled share price spectrum of H_{S_1} is given by S_0\{\lambda\} = \{S_0, S_0e^{u}, S_0 e^{d}, S_0e^{u+d}, S_0e^{2d}, S_0e^{2u}\} . Thus, using the risk function (52), we get for the Maxwell-Boltzmann case

    \begin{align} F_1(u, d)& = \delta_0 \sum\limits_{\beta_i \in G_{0}}|\beta_i-S_0|+q\delta_1\sum\limits_{\beta_i \in G_{d}}|\beta_i-S_0e^{d}|p\delta_1\sum\limits_{\beta_i \in G_{u}}|\beta_i-S_0e^{u}| \\ &+2pq\delta_2\sum\limits_{\beta_i \in G_{u+d}}|\beta_i-S_0e^{u+d}|+q^2\delta_2\sum\limits_{\beta_i \in G_{2d}}|\beta_i-S_0e^{2d}|+p^2\delta_2\sum\limits_{\beta_i \in G_{2u}}|\beta_i-S_0e^{2u}|, \end{align} (62)

    and for the Bose-Einstein case,

    \begin{align} F_2(u, d)& = \delta_0 \sum\limits_{\beta_i \in G_{0}}|\beta_i-S_0|+q\delta_1\sum\limits_{\beta_i \in G_{d}}|\beta_i-S_0e^{d}|+p\delta_1\sum\limits_{\beta_i \in G_{u}}|\beta_i-S_0e^{u}| \\ &+Cpq\delta_2\sum\limits_{\beta_i \in G_{u+d}}|\beta_i-S_0e^{u+d}|+Cq^2\delta_2\sum\limits_{\beta_i \in G_{2d}}|\beta_i-S_0e^{2d}|+Cp^2\delta_2\sum\limits_{\beta_i \in G_{2u}}|\beta_i-S_0e^{2u}|. \end{align} (63)

    Numerical illustration. The data (54) is treated as the eigenvalues of the 1 -step observable H defined in (6) for N = 1 . Let us assume that the Poisson process rate is 1 and the length of time is \Delta t = 1 .

    We estimate the values d, u and the probabilities q, p by applying the algorithm of Section 5.1 and using the Maxwell-Boltzmann statistics. Choose, for instance, S_0 = 70 , (u_0, d_0) = (0.8, -0.1) and M = 0.000001 . The results obtained from (60), (62) are presented in Table 1.

    Table 1.  Estimation for N = 1 in the Maxwell-Boltzmann case.
    Given (u, d)Maximum likelihoodOptimum (p, q)Optimum (u, d)Risk error F(u, d) |F(u_i, d_i)- F(u_{i+1}, d_{i+1})|
    (0.8, -0, 1)4.6063e-14(0.01, 0.99)(0.3, -0.4)43.310743.3107
    (0.3, -0.4)2.2681e-18(0.16, 0.84)(0.3, -0.5)22.873420.4373
    (0.3, -0.5)5.1730e-19(0.23, 0.77)(0.3, -0.6)15.86547.0080
    (0.3, -0.6)1.7631e-18(0.24, 0.76)(0.3, -0.7)14.47411.3913
    (0.3, -0.7)1.2479e-19(0.29, 0.71)(0.3, -0.7)11.19763.2765
    (0.3, -0.7)1.2479e-19(0.29, 0.71)(0.3, -0.7)11.19760

     | Show Table
    DownLoad: CSV

    We then also apply a 4 -foldcross-validation procedure. Let V_i , 1\leq i \leq 4 , before randomly chosen subsets of V such that \cup_{i = 1}^{4} V_i = V . For each iteration, \{V\setminus V_i\} and V_i are treated as the training set and the test set, respectively. First, the maximum likelihood estimation and the risk error computation are executed for the training data. Then, the obtained estimates are implemented in the test data. This gives the results of Table 2.

    Table 2.  Using a 4 -fold cross-validation strategy.
    Training dataTest data
    Training & Test dataOptimum (u, d) Risk error F(u, d) Optimum (u, d) Risk error F(u, d) Total risk error function
    \{V\setminus V_1\}, V_1 (0.2, -0.4)9.5462(0.2, -0.5)0.08429.6304
    \{V\setminus V_2\}, V_2 (0.2, -0.3)11.1223(0.2, -0.1)0.192811.3151
    \{V\setminus V_3\}, V_3 (0.2, -0.3)8.0269(0.3, -0.4)0.20428.2311
    \{V\setminus V_4\}, V_4 (0.2, -0.6)9.2466(0.4, -0.7)0.06299.3095

     | Show Table
    DownLoad: CSV

    We observe that the risk errors are small for the test data but are larger for the training data. Thus, the total risk errors are significant. Similar numerical calculations have also been performed under the Bose-Einstein assumption.

    The spectrum of H^{\otimes N}_2 is given by

    \begin{equation} \{\lambda\} = \{e^{(N-i-j)u+id}, \, 0\leq j \leq N, 0\leq i \leq N-j\}. \end{equation} (64)

    Thus, for the case N = 2 , the set of observables is exactly the same as (55). However, the associated probabilities differ from (56) and are equal to

    \begin{equation} \{p_\lambda\} = \{p_2^2, 2p_2p_3, 2p_1p_2, 2p_1p_3, p_3^2, p_1^2\}. \end{equation} (65)

    Likelihood and risk functions. Using the Maxwell-Boltzmann statistics, the likelihood is defined by the probabilities p_1 , p_2 , p_3 as

    \begin{equation} L(p_1, p_2, p_3) = (p_2^2) ^{\#0}(2p_2p_3) ^{\#d}(2p_1p_2) ^{\#u}(2p_1p_3) ^{\#d+u}(p_3^2) ^{\#2d}(p_1^2) ^{\#2u}. \end{equation} (66)

    Note that for the Cox-Ross-Rubinstein model, p_2 = 0 so that the likelihood is simplified to L(p_1, p_3) = (2p_1p_3) ^{\#d+u}(p_3^2) ^{\#2d}(p_1^2) ^{\#2u} .

    The corresponding risk function is then defined by

    \begin{align} F_3(u, d)& = p_2^2\sum\limits_{\beta_i \in G_{0}}|\beta_i-S_0| + 2p_2p_3\sum\limits_{\beta_i \in G_{d}}|\beta_i-S_0e^{d}|+ 2p_1p_2\sum\limits_{\beta_i \in G_{u}}|\beta_i-S_0e^{u}| \\ &+ 2p_1p_3\sum\limits_{\beta_i \in G_{d+u}}|\beta_i-S_0e^{u+d}| + p_3^2\sum\limits_{\beta_i \in G_{2d}}|\beta_i-S_0e^{2d}| + p_1^2\sum\limits_{\beta_i \in G_{2u}}|\beta_i-S_0e^{2u}|. \end{align} (67)

    Numerical illustration. We process the data (54) again with S_0 = 70 , (u_0, d_0) = (0.8, -0.1) and M = 0.000001 . The results obtained using the algorithm of Section 5.1 with the functions (66), (67) are given in Table 3.

    Table 3.  Estimation for N = 2 in the Maxwell-Boltzmann case.
    Given (u, d)Maximum likelihoodOptimum (p_1, p_2, p_3) Optimum (u, d)Risk error F(u, d) |F(u_i, d_i)- F(u_{i+1}, d_{i+1})|
    (0.8, -0.1)7.3122e-10(0.01, 0.22, 0.77)(0.5, -0.4)0.03210.0321
    (0.5, -0.4)2.2847e-11(0.1, 0.18, 0.72)(0.8, -0.5)0.32960.2975
    (0.8, -0.5)7.4285e-12(0.1, 0.23, 0.67)(0.8, -0.5)0.33020.0006
    (0.8, -0.5)7.4285e-12(0.1, 0.23, 0.67)(0.8, -0.5)0.33020

     | Show Table
    DownLoad: CSV

    Then, we apply a 4 -fold cross-validation procedure as previously done. The obtained results are shown in Table 4. Note that this method allows to reduce the risk errors.

    Table 4.  Using a 4 -fold cross-validation strategy.
    Training dataTest data
    Training & Test dataOptimum (u, d) Risk error F(u, d) Optimum (u, d) Risk error F(u, d) Total risk error function
    \{V\setminus V_1\}, V_1 (0.3, -0.1)0.1093(0.3, -0.2)0.08600.1953
    \{V\setminus V_2\}, V_2 (0.3, -0.1)0.1843(0.7, -0.6)0.00570.1900
    \{V\setminus V_3\}, V_3 (0.8, -0.5)0.1591(0.8, -0.6)0.01380.1729
    \{V\setminus V_4\}, V_4 (0.7, -0.4)0.2575(0.8, -0.6)0.07210.3296

     | Show Table
    DownLoad: CSV

    The observable operators of the quantum actuarial-type and trinomial models give us different results as expected. Which model to choose? One possible approach might be to consider a mixture of quantum models via the mixture of Hamiltonians (see, e.g., Wittek (2014)).

    Several quantum type financial models are constructed that benefit from the physical interpretation of the unpredictable stock market behaviour and associated dependences. The models provide a general physical type framework for pricing of derivatives and a possibility to construct quantum trading strategies. Moreover, it is revealed that certain quantum type models are applied both in actuarial and financial sciences.

    C. Lefèvre received support from the Chair DIALog sponsored by CNP Assurances. S. Utev benefited from a Scientific Mission granted by the Belgian FNRS. We thank the referees for the careful reading of the paper which greatly helped to improve the presentation.

    All authors declare no conflicts of interest in this paper.



    [1] A. Fernández, S. García, F. Herrera, Addressing the classification with imbalanced data: open problems and new challenges on class distribution, In: Lecture Notes in Computer Science, Heidelberg: Springer, 6678 (2011). https://doi.org/10.1007/978-3-642-21219-2_1
    [2] M. Liuzzi, P. A. Pelizari, C. Geiß, A. Masi, V. Tramutoli, H. Taubenböck, A transferable remote sensing approach to classify building structural types for seismic risk analyses: the case of Val d'Agri area (Italy), Bull. Earthq. Eng., 17 (2019), 4825–4853.
    [3] D. Devarriya, C. Gulati, V. Mansharamani, A. Sakalle, A. Bhardwaj, Unbalanced breast cancer data classification using novel fitness functions in genetic programming, Expert Syst. Appl., 140 (2020), 112866. https://doi.org/10.1016/j.eswa.2019.112866 doi: 10.1016/j.eswa.2019.112866
    [4] S. Katoch, S. S. Chauhan, V. Kumar, A review on genetic algorithm: past, present, and future, Multimed. Tools Appl., 80 (2021), 8091–8126. https://doi.org/10.1007/s11042-020-10139-6 doi: 10.1007/s11042-020-10139-6
    [5] Y. L. Yuan, J. J. Ren, S. Wang, Z. X. Wang, X. K. Mu, W. Zhao, Alpine skiing optimization: A new bio-inspired optimization algorithm, Adv. Eng. Softw., 170 (2022), 103158 https://doi.org/10.1016/j.advengsoft.2022.103158 doi: 10.1016/j.advengsoft.2022.103158
    [6] J. F. Goycoolea, M. Inostroza-Ponta, M. Villalobos-Cid, M. Marín, Single-solution based metaheuristic approach to a novel restricted clustering problem, 2021. https://doi.org/10.1109/SCCC54552.2021.9650429
    [7] J. H. Holland, Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence, Bradford Books, 1992.
    [8] S. N. Sivanandam, S. N. Deepa, Introduction to Genetic Algorithms, Heidelberg: Springer Berlin, 2010.
    [9] F. Ortiz, J. R. Simpson, J. Pignatiello, A. Heredia-Langner, A genetic algorithm approach to multiple-response optimization, J. Qual. Technol., 36 (2004), 432–450. https://doi.org/10.1080/00224065.2004.11980289 doi: 10.1080/00224065.2004.11980289
    [10] H. I. Calvete, C. Gale, P. M. Mateo, A new approach for solving linear bilevel problems using genetic algorithms, European J. Oper. Res., 188 (2008), 14–28 https://doi.org/10.1016/j.ejor.2007.03.034 doi: 10.1016/j.ejor.2007.03.034
    [11] S. S. Nimankar, D. Vora, Designing a model to handle imbalance data classification using SMOTE and optimized classifier, In: Data Management, Analytics and Innovation, Singapore: Springer, 2020,323–334.
    [12] K. Jiang, J. Lu, K. L. Xia, A novel algorithm for imbalance data classification based on genetic algorithm improved SMOTE, Arab. J. Sci. Eng., 41 (2016), 3255–3266. http://doi.org/10.1007/s13369-016-2179-2 doi: 10.1007/s13369-016-2179-2
    [13] R. Obiedat, R. Qaddoura, A. M. Al-Zoubi, L. Al-Qaisi, O. Harfoushi, M. Alrefai, et al., Sentiment analysis of customers' reviews using a hybrid evolutionary SVM based approach in an imbalanced data distribution, IEEE Access, 10 (2022), 22260–22273. https://doi.org/10.1109/ACCESS.2022.3149482 doi: 10.1109/ACCESS.2022.3149482
    [14] L. Wang, Imbalanced credit risk prediction based on SMOTE and multi-kernel FCM improved by particle swarm optimization, Appl. Soft Comput., 114 (2022), 108153. https://doi.org/10.1016/j.asoc.2021.108153 doi: 10.1016/j.asoc.2021.108153
    [15] L. Demidova, I. Klyueva, SVM classification: Optimization with the SMOTE algorithm for the class imbalance problem, 2017. https://doi.org/10.1109/MECO.2017.7977136
    [16] S. Sreejith, H. K. Nehemiah, A. Kannan, Clinical data classification using an enhanced SMOTE and chaotic evolutionary feature selection, Comput. Biol. Med., 126 (2020), 103991. https://doi.org/10.1016/j.compbiomed.2020.103991 doi: 10.1016/j.compbiomed.2020.103991
    [17] Y.-C. Wang, C.-H. Cheng, A multiple combined method for rebalancing medical data with class imbalances, Comput. Biol. Med., 134 (2021), 104527. https://doi.org/10.1016/j.compbiomed.2021.104527 doi: 10.1016/j.compbiomed.2021.104527
    [18] B. Zorić, D. Bajer, G. Martinović, Employing different optimisation approaches for SMOTE parameter tuning, 2016. https://doi.org/10.1109/SST.2016.7765657
    [19] E. Sara, C. Laila, I. Ali, The impact of SMOTE and grid search on maintainability prediction models, 2019. https://doi.org/10.1109/AICCSA47632.2019.9035342
    [20] J. J. Ren, Z. X. Wang, Y. Pang, Y. L. Yuan, Genetic algorithm-assisted an improved AdaBoost double-layer for oil temperature prediction of TBM, Adv. Eng. Inform., 52 (2022), 101563. https://doi.org/10.1016/j.aei.2022.101563 doi: 10.1016/j.aei.2022.101563
    [21] Y. L. Yuan, X. K. Mu, X. Y. Shao, J. J. Ren, Y. Zhao, Z. X. Zhao, Optimization of an auto drum fashioned brake using the elite opposition-based learning and chaotic k-best gravitational search strategy based grey wolf optimizer algorithm, Appl. Soft Comput., 123 (2022), 108947. https://doi.org/10.1016/j.asoc.2022.108947 doi: 10.1016/j.asoc.2022.108947
    [22] M. L. Shi, S. Wang, W. Sun, L. Y. Lv, X. G. Song, A support vector regression-based multi-fidelity surrogate model, 2019.
    [23] D. E. Goldberg, Genetic Algorithms in Search, Optimization and Machine Learning, Addison-Wesley Professional, 1989.
    [24] S. Panda, N. P. Padhy, Comparison of particle swarm optimization and genetic algorithm for FACTS-based controller design, Appl. Soft Comput., 8 (2008), 1418–1427. https://doi.org/10.1016/j.asoc.2007.10.009 doi: 10.1016/j.asoc.2007.10.009
    [25] D. Orvosh, L. Davis, Using a genetic algorithm to optimize problems with feasibility constraints, IEEE World Congress on Computational Intelligence, 1994. https://doi.org/10.1109/ICEC.1994.350001 doi: 10.1109/ICEC.1994.350001
    [26] E. C. Gonçalves, A. Plastino, A. A. Freitas, A genetic algorithm for optimizing the label ordering in multi-label classifier chains, In: 2013 IEEE 25th International Conference on Tools with Artificial Intelligence, 2013. https://doi.org/10.1109/ICTAI.2013.76
    [27] J. Han, M. Kamber, J. Pei, Data Mining: Concepts and Techniques, 2011.
    [28] V. Vapnik, Principles of risk minimization for learning theory, In: Proceedings of the 4th International Conference on Neural Information Processing Systems, 1991,831–838.
    [29] T. Koc, P. Akın, Estimation of high school entrance examination success rates using machine learning and beta regression models, J. Intell. Syst. Theory Appl., 5 (2022), 9–15. http://doi.org/10.38016/jista.922663 doi: 10.38016/jista.922663
    [30] D. Guleryuz, Estimation of soil temperatures with machine learning algorithms-Giresun and Bayburt stations in Turkey, Theor. Appl. Climatol., 147 (2022), 109–125.
    [31] Q. Quan, Z. Hao, X. F. Huang, J. C. Lei, Research on water temperature prediction based on improved support vector regression, Neural Comput. Appl., 2020, 1–10. https://doi.org/10.1007/S00521-020-04836-4 doi: 10.1007/S00521-020-04836-4
    [32] N. V. Chawla, K. W. Bowyer, L. O. Hall, W. P. Kegelmeyer, SMOTE: synthetic minority over-sampling technique, J. Artif. Intell. Res., 16 (2002), 321–357. https://doi.org/10.5555/1622407.1622416 doi: 10.5555/1622407.1622416
    [33] J. Brandt, E. Lanzén, A comparative review of SMOTE and ADASYN in imbalanced data classification, In: Bachelor's Thesis, Uppsala: Uppsala University, 2021.
    [34] H. Al Majzoub, I. Elgedawy, Ö. Akaydın, M. K. Ulukök, HCAB-SMOTE: A hybrid clustered affinitive borderline SMOTE approach for imbalanced data binary classification, Arab. J. Sci. Eng., 45 (2020), 3205–3222.
    [35] P. Akin, Y. Terzi, Comparison of unbalanced data methods for support vector machines, Turkiye Klinikleri J. Biostat., 13 (2021), 138–146. http://doi.org/10.5336/biostatic.2020-80268 doi: 10.5336/biostatic.2020-80268
    [36] S. Uğuz, Makine öğrenmesi teorik yönleri ve Python uygulamaları ile bir yapay zeka ekolü, Nobel Yayıncılık Ankara, 2019.
    [37] R. E. Wright, Logistic regression, In: Reading and Understanding Multivariate Statistics, 1995,217–244.
    [38] T. Koc, H. Koc, E. Ulas, Üniversite öğrencilerinin kötü alışkanlıklarının bayesci ağ yöntemi ile belirlenmesi, Çukurova Üniversitesi Sosyal Bilimler Enstitüsü Dergisi, 26 (2017), 230–240.
    [39] S. V. Buuren, K. Groothuis-Oudshoorn, Mice: Multivariate imputation by chained equations in R, J. Statist. Softw., 45 (2011), 1–68.
  • This article has been cited by:

    1. Yuqi Fan, Huimin Yang, Yaping Wang, Zunshan Xu, Daoxiang Lu, A Variable Step Crow Search Algorithm and Its Application in Function Problems, 2023, 8, 2313-7673, 395, 10.3390/biomimetics8050395
    2. Debashis Roy, Utathya Aich, Anandarup Roy, Utpal Roy, 2024, Chapter 19, 978-3-031-47220-6, 209, 10.1007/978-3-031-47221-3_19
    3. Dadi Zhang, Kwok-Wai Mui, Massimiliano Masullo, Ling-Tim Wong, Application of Machine Learning Techniques for Predicting Students’ Acoustic Evaluation in a University Library, 2024, 6, 2624-599X, 681, 10.3390/acoustics6030037
    4. Ying Li, Keyue Yan, Prediction of bank credit customers churn based on machine learning and interpretability analysis, 2025, 5, 2769-2140, 19, 10.3934/DSFE.2025002
    5. Zhimin Rao, Yicheng Li, Jiandong Mao, Hu Zhao, Xin Gong, Prediction of Atmospheric Bioaerosol Number Concentration Based on PKO–AGA–SVM Fusion Algorithm and Fluorescence Lidar Telemetry, 2025, 16, 2073-4433, 638, 10.3390/atmos16060638
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2506) PDF downloads(142) Cited by(5)

Figures and Tables

Figures(2)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog