Processing math: 47%
Research article Special Issues

Modified Marine Predators Algorithm hybridized with teaching-learning mechanism for solving optimization problems


  • Received: 18 June 2022 Revised: 24 August 2022 Accepted: 01 September 2022 Published: 29 September 2022
  • Marine Predators Algorithm (MPA) is a newly nature-inspired meta-heuristic algorithm, which is proposed based on the Lévy flight and Brownian motion of ocean predators. Since the MPA was proposed, it has been successfully applied in many fields. However, it includes several shortcomings, such as falling into local optimum easily and precocious convergence. To balance the exploitation and exploration ability of MPA, a modified marine predators algorithm hybridized with teaching-learning mechanism is proposed in this paper, namely MTLMPA. Compared with MPA, the proposed MTLMPA has two highlights. Firstly, a kind of teaching mechanism is introduced in the first phase of MPA to improve the global searching ability. Secondly, a novel learning mechanism is introduced in the third phase of MPA to enhance the chance encounter rate between predator and prey and to avoid premature convergence. MTLMPA is verified by 23 benchmark numerical testing functions and 29 CEC-2017 testing functions. Experimental results reveal that the MTLMPA is more competitive compared with several state-of-the-art heuristic optimization algorithms.

    Citation: Yunpeng Ma, Chang Chang, Zehua Lin, Xinxin Zhang, Jiancai Song, Lei Chen. Modified Marine Predators Algorithm hybridized with teaching-learning mechanism for solving optimization problems[J]. Mathematical Biosciences and Engineering, 2023, 20(1): 93-127. doi: 10.3934/mbe.2023006

    Related Papers:

    [1] Jingjing Yang, Jianqiu Lu . Stabilization in distribution of hybrid stochastic differential delay equations with Lévy noise by discrete-time state feedback controls. AIMS Mathematics, 2025, 10(2): 3457-3483. doi: 10.3934/math.2025160
    [2] Guangyang Liu, Yang Chang, Hongyan Yan . Uncertain random problem for multistage switched systems. AIMS Mathematics, 2023, 8(10): 22789-22807. doi: 10.3934/math.20231161
    [3] Hui Sun, Zhongyang Sun, Ya Huang . Equilibrium investment and risk control for an insurer with non-Markovian regime-switching and no-shorting constraints. AIMS Mathematics, 2020, 5(6): 6996-7013. doi: 10.3934/math.2020449
    [4] Jiaojiao Li, Yingying Wang, Jianyu Zhang . Event-triggered sliding mode control for a class of uncertain switching systems. AIMS Mathematics, 2023, 8(12): 29424-29439. doi: 10.3934/math.20231506
    [5] Qiang Yu, Na Xue . Asynchronously switching control of discrete-time switched systems with a Φ-dependent integrated dwell time approach. AIMS Mathematics, 2023, 8(12): 29332-29351. doi: 10.3934/math.20231501
    [6] Gengjiao Yang, Fei Hao, Lin Zhang, Lixin Gao . Stabilization of discrete-time positive switched T-S fuzzy systems subject to actuator saturation. AIMS Mathematics, 2023, 8(6): 12708-12728. doi: 10.3934/math.2023640
    [7] Zhengqi Zhang, Huaiqin Wu . Cluster synchronization in finite/fixed time for semi-Markovian switching T-S fuzzy complex dynamical networks with discontinuous dynamic nodes. AIMS Mathematics, 2022, 7(7): 11942-11971. doi: 10.3934/math.2022666
    [8] Qiang Yu, Yuanyang Feng . Stability analysis of switching systems with all modes unstable based on a Φ-dependent max-minimum dwell time method. AIMS Mathematics, 2024, 9(2): 4863-4881. doi: 10.3934/math.2024236
    [9] P. K. Lakshmi Priya, K. Kaliraj, Panumart Sawangtong . Analysis of relative controllability and finite-time stability in nonlinear switched fractional impulsive systems. AIMS Mathematics, 2025, 10(4): 8095-8115. doi: 10.3934/math.2025371
    [10] Yanmei Xue, Jinke Han, Ziqiang Tu, Xiangyong Chen . Stability analysis and design of cooperative control for linear delta operator system. AIMS Mathematics, 2023, 8(6): 12671-12693. doi: 10.3934/math.2023637
  • Marine Predators Algorithm (MPA) is a newly nature-inspired meta-heuristic algorithm, which is proposed based on the Lévy flight and Brownian motion of ocean predators. Since the MPA was proposed, it has been successfully applied in many fields. However, it includes several shortcomings, such as falling into local optimum easily and precocious convergence. To balance the exploitation and exploration ability of MPA, a modified marine predators algorithm hybridized with teaching-learning mechanism is proposed in this paper, namely MTLMPA. Compared with MPA, the proposed MTLMPA has two highlights. Firstly, a kind of teaching mechanism is introduced in the first phase of MPA to improve the global searching ability. Secondly, a novel learning mechanism is introduced in the third phase of MPA to enhance the chance encounter rate between predator and prey and to avoid premature convergence. MTLMPA is verified by 23 benchmark numerical testing functions and 29 CEC-2017 testing functions. Experimental results reveal that the MTLMPA is more competitive compared with several state-of-the-art heuristic optimization algorithms.



    Over the past decades, the use of biochemical reactors and correlation techniques has increased greatly because of their fruitful application in converting biomass or cells into pharmaceutical or chemical products, such as vaccines [1], antibiotics [2], beverages [3], and industrial solvents [4]. Among various classes or operation regions of bioreactors, the fed-batch modes have extensively used in the biotechnological industry due to its considerable economic profits [5,6,7]. The main objective of these reactors is to achieve a given or maximum concentration of production at the end of the operation, which can be implemented by using some suitable feed rates [8,9,10]. Thus, in order to ensure economic benefit and product quality of the fed-batch processes, the process control of this units is an very important topic for the engineers [11,12,13].

    Switched dynamical systems provide a flexible modeling method for a variety of different types of engineering systems, such as financial system [14], train control system [15], hybrid electric vehicle [16], chemical process system [17], and biological system [18,19,20,21]. Generally speaking, switched dynamical systems are formed by some continuous-time or discrete-time subsystems and a switching rule [22]. There usually exist four types of switching rules as follows: time-dependent switching [23], state-dependent switching [24], average dwell time switching [25], and minimum dwell time switching [26]. Recently, switched dynamical system optimal control problems are becoming increasingly attractive due to their significance in theory and industry production [27,28,29,30]. Because of the discrete nature of switching rules, it is very challenging that switched dynamical system optimal control problems are solved by directly using the classical optimal control approaches such as the maximum principle and the dynamic programming method [31,32,33,34]. In additions, analytical methods also can not be applied to obtain an solution for switched dynamical system optimal control problems due to their nonlinear nature [35,36,37]. Thus, in recent work, two kinds of well-known numerical optimization algorithms are developed for switched dynamical system optimal control problems to obtain numerical solutions. One is the bi-level algorithm [38,39]. The other is the embedding algorithm [40,41]. Besides above two kinds of well-known numerical optimization algorithms, many other available numerical optimization algorithms are also developed for obtaining the solution of switched dynamical system optimal control problems [42]. Unfortunately, most of these numerical optimization algorithms depend on the following assumption: the time-dependent switching strategy is used to design the switching rules, which implies that the system dynamic must be continuously differentiable with respect to the system state [43,44,45]. However, this assumption is not reasonable, since some small perturbations of the system state may lead to the dynamic equations being changed discontinuously. Thus, the solution obtained is usually not optimal. In additions, although these approaches have demonstrated to be effective by solving many practical problems, they only obtaining an open loop control [46,47,48,49,50,51,52,53]. Unfortunately, such open loop controls are not usually robust in practice. Thus, an optimal feedback controller is more and more popular.

    In this paper, we consider an optimal feedback control problem for a class of fed-batch fermentation processes by using switched dynamical system approach. Our main contributions are as follows. Firstly, a dynamic optimization problem for a class of fed-batch fermentation processes is modeled as a switched dynamical system optimal control problem, and a general state-feedback controller is designed for this dynamic optimization problem. Unlike the existing works, the state-dependent switching method is applied to design the switching rule, and the structure of this state-feedback controller is not restricted to a particular form. In generally, the traditional methods for obtaining an optimal feedback control require solving the well-known Hamilton-Jacobi-Bellman partial differential equation, which is a very difficult issue even for unconstrained optimal control problems. Then, in order to overcome this difficulty, this problem is transformed into a mixed-integer optimal control problem by introducing a discrete-valued function. Furthermore, each of these discrete variables is represented by using a set of 0-1 variables. Then, by using a quadratic constraint, these 0-1 variables are relaxed such that they are continuous on the closed interval [0,1]. Accordingly, the original mixed-integer optimal control problem is transformed into a nonlinear parameter optimization problem, which can be solved by using any gradient-based numerical optimization algorithm. Unlike the existing works, the constraint introduced for these 0-1 variables are at most quadratic. Thus, it does not increase the number of locally optimal solutions of the original problem. During the past decades, many iterative approaches have been proposed for solving the nonlinear parameter optimization problem by using the information of the objective function. The idea of these iterative approaches is usually that a iterative sequence is generated such that the corresponding objective function value sequence is monotonically decreasing. However, the existing algorithms have the following disadvantage: if an iteration is trapped to a curved narrow valley bottom of the objective function, then the iterative methods will lose their efficiency due to the target with objective function value monotonically decreasing may leading to very short iterative steps. Next, in order to overcome this challenge, an improved gradient-based algorithm is developed based on a novel search approach. In this novel search approach, it is not required that the objective function value sequence is always monotonically decreasing. And a large number of numerical experiments shows that this novel search approach can effectively improve the convergence speed of this algorithm, when an iteration is trapped to a curved narrow valley bottom of the objective function. Finally, an optimal feedback control problem of 1, 3-propanediol fermentation processes is provided to illustrate the effectiveness of this method developed by this paper. Numerical simulation results show that this method developed by this paper is low time-consuming, has faster convergence speed, and obtains a better result than the existing approaches.

    The rest of this paper is organized as follows. Section 2 presents the optimal feedback control problem for a class of fed-batch fermentation processes. In Section 3, by introducing a discrete-valued function and using a relaxation technique, this problem is transformed into a nonlinear parameter optimization problem, which can be solved by using any gradient-based numerical optimization algorithm. An improved gradient-based numerical optimization algorithm are developed in Section 4. In Section 5, the convergence results of this numerical optimization algorithm are established. In Section 6, an optimal feedback control problem of 1, 3-propanediol fermentation processes is provided to illustrate the effectiveness of this algorithm developed by this paper.

    In this section, a general state-feedback controller is proposed for a class of fed-batch fermentation process dynamic optimization problems, which will be modeled as an optimal control problem of switched dynamical systems under state-dependent switching.

    Let α1=[α11,,α1r1]TRr1 and α2=[α21,,α2r2]TRr2 be two parameter vectors satisfying

    a_iα1riˉai,i=1,,r1, (2.1)

    and

    b_jα2rjˉbj,j=1,,r2, (2.2)

    respectively, where a_i, ˉai, i=1,,r1; b_j, ˉbj, j=1,,r2 present given constants. Suppose that tf>0 presents a given terminal time. Then, a class of fed-batch fermentation process dynamic optimization problems can be described as choose two parameter vectors α1Rr1, α2Rr2, and a general state-feedback controller

    u(t)=Υ(x(t),ϑ),t[0,tf], (2.3)

    to minimize the objective function

    J(u(t),α1,α2)=ϕ(x(tf)), (2.4)

    subject to the switched dynamical system under state-dependent switching

    {Subsystem1:dx(t)dt=f1(x(t),t),ifg1(x(t),α1,t)=0,Subsystem2:dx(t)dt=f2(x(t),u(t),t),ifg2(x(t),α2,t)=0,t[0,tf], (2.5)

    with the initial condition

    x(0)=x0, (2.6)

    where x(t)Rn presents the system state; x0 presents a given initial system state; u(t)Rm presents the control input; ϑ=[ϑ1,,ϑr1]TRr3 presents a state-feedback parameter vector satisfying

    c_kϑkˉck,k=1,,r3, (2.7)

    c_k and ˉck, k=1,,r present given constants. Υ:Rn×RrRm; ϕ:RnR, f1:Rn×[0,tf]Rn, f2:Rn×Rm×[0,tf]Rn, g1:Rn×Rr1×[0,tf]Rn, g2:Rn×Rr2×[0,tf]Rn present five continuously differentiable functions. For convenience, this problem is called as Problem 1.

    Remark 1. In the switched dynamical system (2.5), Subsystem 1 presents the batch mode, during which there exists no input feed (i.e., control input) u(t), and Subsystem 2 presents the feeding mode, during which there exists input feed (i.e., control input) u(t). This fed-batch fermentation process will oscillate between Subsystem 1 (the batch mode) and Subsystem 2 (the feeding mode), and g1(x(t),α1,t)=0 and g2(x(t),α2,t)=0 present the active conditions of Subsystems 1 and 2, respectively.

    Remark 2. Note that an integral term, which is used to measure the system running cost, can be easily incorporated into the objective function (2.4) by augmenting the switched dynamical system (2.5) with an additional system state variable (see Chapter 8 of this work [54]). Thus, it is not a serious restriction that the integral term does not appear in the objective function (2.4).

    Remark 3. The structure for this general state-feedback controller (2.3) can be governed by the given continuously differentiable function Υ, and the state-feedback parameter vector ϑ is decision variable vector, which will be chosen optimally. For example, the linear state-feedback controller described by u(t)=Kx(t) is a very common state-feedback controller, where KRm×n presents a state-feedback gain matrix to be found optimally.

    In Problem 1, the state-dependent switching strategy is adopted to design the switching rule, which is unlike the existing switched dynamical system optimal control problem. Then, the solution of Problem 1 can not be obtained by directly using the existing numerical computation approaches for switched dynamical systems optimal control problem, in which the switching rule is designed by using time-dependent strategy. In order to overcome this difficulty, by introducing a discrete-valued function, the problem will be transformed into a equivalent nonlinear dynamical system optimal control problem with discrete and continuous variables in this subsection.

    Firstly, by substituting the general state-feedback controller (2.3) into the switched dynamical system (2.5), Problem 1 can be equivalently written as the following problem:

    Problem 2. Choose (α1,α2,ϑ)Rr1×Rr2×Rr3 to minimize the objective function

    ˉJ(α1,α2,ϑ)=ϕ(x(tf)), (3.1)

    subject to the switched dynamical system under state-dependent switching

    {Subsystem1:dx(t)dt=f1(x(t),t),ifg1(x(t),α1,t)=0,Subsystem2:dx(t)dt=ˉf2(x(t),ϑ,t),ifg2(x(t),α2,t)=0,t[0,tf], (3.2)

    and the three bound constraints (2.1), (2.2) and (2.7), where ˉf2(x(t),ϑ,t)=f2(x(t),Υ(x(t),ϑ),t).

    Next, note that the solution of Problem 1 can not be obtained by directly using the existing numerical computation approaches for switched dynamical systems optimal control problem, in which the switching rule is designed by using time-dependent strategy and not state-dependent strategy. In order to overcome this difficulty, a novel discrete-valued function y(t) is introduced as follows:

    y(t)={1,ifg1(x(t),α1,t)=0,2,ifg2(x(t),α2,t)=0,t[0,tf]. (3.3)

    Then, Problem 2 can be transformed into the following equivalent optimization problem with discrete and continuous variables:

    Problem 3. Choose (α1,α2,ϑ,y(t))Rr1×Rr2×Rr3×{1,2} to minimize the objective function

    ˜J(α1,α2,ϑ,y(t))=ϕ(x(tf)), (3.4)

    subject to the nonlinear dynamical system

    dx(t)dt=(2y(t))y(t)f1(x(t),t)+(y(t)1)ˉf2(x(t),ϑ,t),t[0,tf], (3.5)

    the equality constraint

    (2y(t))y(t)g1(x(t),α1,t)+(y(t)1)g2(x(t),α2,t)=0,t[0,tf], (3.6)

    and the three bound constraints (2.1), (2.2), and (2.7).

    Note that standard nonlinear numerical optimization algorithms are usually developed for nonlinear optimization problems only with continuous variables, for example the sequential quadratic programming algorithm, the interior-point method, and so on. Thus, the solution of Problem 3, which has discrete and continuous variables, can not be obtained by directly using these existing standard algorithms. In order to overcome this difficulty, this subsection will introduce a relaxation problem, which has only continuous variables.

    Define

    P(σ(t))=2i=1i2σi(t)(2i=1iσi(t))2, (3.7)

    where σ(t)=[σ1(t),σ2(t)]T. Then, a theorem can be established as follows.

    Theorem 1. If the nonnegative functions σ1(t) and σ2(t) satisfy the following equality:

    σ1(t)+σ2(t)=1,t[0,tf], (3.8)

    then two results can be obtained as follows:

    (1) For any t[0,tf], the function P(σ(t)) is nonnegative;

    (2) For any t[0,tf], P(σ(t))=0 if and only if σi(t)=1 for one i{1,2} and σi(t)=0 for the other i{1,2}.

    Proof. (1) By using the equality (3.8) and the Cauchy-Schwarz inequality, we have

    2i=1iσi(t)=2i=1(iσi(t))σi(t)2i=1(i2σi(t))2i=1σi(t)=2i=1(i2σi(t)), (3.9)

    Note that the functions σ1(t) and σ2(t) are nonnegative. Then, squaring both sides of the inequality (3.9) yields

    2i=1(i2σi(t))(2i=1iσi(t))2,

    which implies that for any t[0,tf], the function P(σ(t)) is nonnegative.

    (2) The correctness of the second part for Theorem 1 only need to prove the following result: for any t[0,tf], P(σ(t))=0 has solutions σi(t)=1 for one i{1,2} and σi(t)=0 for the other i{1,2} and ii.

    Define

    v1(t)=[σ1(t),2σ2(t)],v2(t)=[σ1(t),σ2(t)].

    Then, the inequality (3.9) can be equivalently transformed into as follows:

    v1(t)v2(t)v1(t)v2(t), (3.10)

    where and present the vector dot product and the Euclidean norm, respectively. Note that the equality

    v1(t)v2(t)=v1(t)v2(t) (3.11)

    holds if and only if there exists a constant βR such that

    v1(t)=βv2(t). (3.12)

    By using the equality (3.8), one obtain v1(t)0 and v2(t)0, where 0 presents the zero vector. Then, β is a nonzero constant and the equality (3.12) implies

    (1β)σ1(t)=0, (3.13)
    (2β)σ2(t)=0. (3.14)

    Furthermore, the constant β can be set equal to one integer i{1,2}, and for the other integer i{1,2}, one have

    σi(t)=0,ii, (3.15)

    From the two equalities (3.8) and (3.15), we obtain σi(t)=1. This completes the proof of Theorem 1.

    Now, Problem 3 can be rewritten as a relaxation problem as follows:

    Problem 4. Choose (α1,α2,ϑ,σ(t))Rr1×Rr2×Rr3×R2 to minimize the objective function

    Jrelax(α1,α2,ϑ,σ(t))=ϕ(x(tf)), (3.16)

    subject to the nonlinear dynamical system

    dx(t)dt=(2ˉy(t))ˉy(t)f1(x(t),t)+(ˉy(t)1)ˉf2(x(t),ϑ,t),t[0,tf], (3.17)

    the two equality constraints

    (2ˉy(t))ˉy(t)g1(x(t),α1,t)+(ˉy(t)1)g2(x(t),α2,t)=0,t[0,tf], (3.18)
    P(σ(t))=0,t[0,tf], (3.19)

    the bound constraint

    0σi(t)1,i=1,2,t[0,tf], (3.20)

    the equality constraint (3.8), and the three bound constraints (2.1), (2.2), and (2.7), where

    ˉy(t)=1×σ1(t)+2×σ2(t). (3.21)

    By using Theorem 1, one can derive that Problems 3 and 4 are equivalent.

    Note that the bound constraint (3.20) is essentially some continuous-time inequality constraints. Thus, the solution of Problem 4 can not also be obtained by directly using the existing standard algorithms. In order to obtain the solution of Problem 4, this subsection will introduce a nonlinear parameter optimization problem, which has some continuous-time equality constraints and several bound constraints.

    Suppose that τi presents the ith switching time. Then, one have

    0=τ0τ1τ2τM1τM=tf, (3.22)

    where M1 presents a given fixed integer. It is important to note that the switching times are not independent optimization variables, whose values can be obtained indirectly by using the state trajectory of the switched dynamical system (2.5). Then, Problem 4 can be transformed into an equivalent optimization problem as follows:

    Problem 5. Choose (α1,α2,ϑ,ξ)Rr1×Rr2×Rr3×R2M to minimize the objective function

    ˉJrelax(α1,α2,ϑ,ξ)=ϕ(x(tf)), (3.23)

    subject to the nonlinear dynamical system

    dx(t)dt=Mi=1((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)f1(x(t),t)+((ξ1i+2ξ2i)1)ˉf2(x(t),ϑ,t))χ[τi1,τi)(t),
    t[0,tf], (3.24)

    the equality constraints

    Mi=1((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)g1(x(t),α1,t)+((ξ1i+2ξ2i)1)g2(x(t),α2,t))χ[τi1,τi)(t)=0,
    t[0,tf], (3.25)
    ˉP(ξ,t)=0,t[0,tf], (3.26)
    ξ1i+ξ2i=1,i=1,,M, (3.27)

    the bound constraint

    0ξji1,j=1,2,i=1,,M, (3.28)

    and the three bound constraints (2.1), (2.2), and (2.7), where ξi1 and ξi2 present, respectively, the values of σ1(t) and σ2(t) on the ith subinterval [τi1,τi), i=1,,M; ξ=[(ξ1)T,(ξ2)T]T, ξ1=[ξ11,,ξ1M]T, ξ2=[ξ21,,ξ2M]T; ˉP(ξ,t)=Mi=1(2j=1j2ξji(2j=1jξji)2)χ[τi1,τi)(t); and χI(t) is given by

    χI(t)={1,iftI,0,otherwise, (3.29)

    which is a function defined on the subinterval I[0,tf].

    Due to the switching times being unknown, it is very challenging to acquire the gradient of the objective function (3.23). In order to overcome this challenge, the following time-scaling transformation is developed to transform variable switching times into fixed times:

    Suppose that the function t(s):[0,M]R is continuously differentiable and is governed by the following equation:

    dt(s)ds=Mi=1θiχ[i1,i)(s), (3.30)

    with the boundary condition

    t(0)=0, (3.31)

    where θi is the subsystem dwell time on the ith subinterval [i1,i)[0,tf]. In general, the transformation (3.30)–(3.31) is referred to as a time-scaling transformation.

    Define θ=[θ1,,θM]T, where

    0θitf,i=1,,M. (3.32)

    Then, by using the time-scaling transform (3.30) and (3.31), we can rewrite Problem 5 as the following equivalent nonlinear parameter optimization problem, which has fixed switching times.

    Problem 6. Choose (α1,α2,ϑ,ξ,θ)Rr1×Rr2×Rr3×R2M×RM to minimize the objective function

    ˆJrelax(α1,α2,ϑ,ξ,θ)=ϕ(ˆx(M)), (3.33)

    subject to the nonlinear dynamical system

    dˆx(s)ds=Mi=1θi((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)f1(ˆx(s),s)+((ξ1i+2ξ2i)1)ˉf2(ˆx(s),ϑ,s))χ[i1,i)(s),
    s[0,M], (3.34)

    the continuous-time equality constraints

    Mi=1θi((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)g1(ˆx(s),α1,s)+((ξ1i+2ξ2i)1)g2(ˆx(s),α2,s))χ[i1,i)(s)=0,
    s[0,M], (3.35)
    ˆP(ξ,s)=0,s[0,M], (3.36)

    the linear equality constraint (3.27), the three bound constraints (2.1), (2.2), (2.7), (3.28), and (3.32), where ˆx(s)=x(t(s)) and ˆP(ξ,s)=Mi=1θi(2j=1j2ξji(2j=1jξji)2)χ[i1,i)(s).

    In this section, an improved gradient-based numerical optimization algorithm will be proposed for obtaining the solution of Problem 1.

    In order to handle the continuous-time equality constraints (3.35) and (3.36), by adopting the idea of l1 penalty function [55], Problem 6 will be written as a nonlinear parameter optimization problem with a linear equality constraint and several simple bounded constraints in this subsection.

    Problem 7. Choose (α1,α2,ϑ,ξ,θ)Rr1×Rr2×Rr3×R2M×RM to minimize the objective function

    Jγ(α1,α2,ϑ,ξ,θ)=ϕ(ˆx(M))+γM0L(ˆx(s),α1,α2,ϑ,ξ,θ,s)ds, (4.1)

    subject to the nonlinear dynamical system (3.34), the linear equality constraint (3.27), the three bound constraints (2.1), (2.2), (2.7), (3.28) and (3.32), where

    L(ˆx(s),α1,α2,ϑ,ξ,θ,s)
    =ˆP(ξ,s)+Mi=1θi((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)g1(ˆx(s),α1,s)+((ξ1i+2ξ2i)1)g2(ˆx(s),α2,s))χ[i1,i)(s),

    where γ>0 presents the penalty parameter.

    The idea of l1 penalty function [47] indicates that any solution of Problem 7 is also a solution of Problem 6. In additions, it is straightforward to acquire the gradient of the linear function in the equality constraint (3.27), and the gradient of the objective function (4.1) will be presented in Section 4.2. Thus, the solution of Problem 7 can be achieved by applying any gradient-based numerical computation method.

    In order to acquire the solution of Problem 7, the gradient formulae of this objective function (4.1) will be presented by the following theorem in this subsection.

    Theorem 2. For any s[0,M], the gradient formulae of the objective function (4.1) with respect to the decision variables α1, α2, ϑ, ξ, and θ are given by

    Jγ(α1,α2,ϑ,ξ,θ)α1=M0H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))α1ds, (4.2)
    Jγ(α1,α2,ϑ,ξ,θ)α2=M0H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))α2ds, (4.3)
    Jγ(α1,α2,ϑ,ξ,θ)ϑ=M0H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))ϑds, (4.4)
    Jγ(α1,α2,ϑ,ξ,θ)ξ=M0H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))ξds, (4.5)
    Jγ(α1,α2,ϑ,ξ,θ)θ=M0H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))θds, (4.6)

    where H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s)) denotes the Hamiltonian function defined by

    H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))=L(ˆx(s),α1,α2,ϑ,ξ,θ,s)+(λ(s))Tˉf(ˆx(s),α1,α2,ϑ,ξ,θ,s), (4.7)
    ˉf(ˆx(s),α1,α2,ϑ,ξ,θ,s)
    =Mi=1θi((2(ξ1i+2ξ2i))(ξ1i+2ξ2i)f1(ˆx(s),s)+((ξ1i+2ξ2i)1)ˉf2(ˆx(s),ϑ,s))χ[i1,i)(s), (4.8)

    and the function λ(s) presents the costate satisfying the following system:

    (dλ(s)ds)T=H(ˆx(s),α1,α2,ϑ,ξ,θ,λ(s))ˆx(s) (4.9)

    with the terminal condition

    (λ(M))T=ϕ(ˆx(M))ˆx(M). (4.10)

    Proof. Similarly to the discussion of Theorem 5.2.1 described in [56], the gradient formulae (4.2)–(4.6) can be obtained. This completes the proof of Theorem 2.

    For simplicity of notation, let g(η)=˜Jγ(η) presents the gradient of the objective function Jγ described by (4.1) at η, where η=[(α1)T,(α2)T,ϑT,ξT,θT]T. In additions, let and present, respectively, the Euclidean norm and the infinity norm, and suppose that the subscript k presents the function value at the point ηk or in the kth iteration, for instance, gk and (Jγ)k. Then, based on the above discussion, an improved gradient-based numerical optimization algorithm will be provided to acquire the solution of Problem 1 in this subsection.

    Algorithm 1. An improved gradient-based numerical optimization algorithm for solving Problem 1.
    01. Initial: η0Rr1+r2+r3+3M, 0<μ<1, 0<ϖ<1, ρmaxρmin>0, 0<NminN0Nmax, ε>0;
    02. begin
    03.   calculate the objective function (Jγ)0 and the gradient g0 at the point η0;
    04.   (ˆJγ)p(0):=Jγ(η0), ρ0:=1, k:=0;
    05.   while gkε do
    06.     dk:=ρkgk, ωk:=1, ˆηk:=ηk+ωkdk;
    07.     while Jγ(ˆηk)>(ˆJγ)p(k)+μωk(gk)Tdk do
    08.     ωk:=ϖωk, ˆηk:=ηk+ωkdk;
    09.     end
    10.     ηk+1:=ˆηk, (Jγ)k+1:=Jγ(ˆηk);
    11.     calculate δk by using the following equality:
                    δk=(zk1)Tek1(ek1)Tek1,          (4.11)
            where zk1=ηkηk1, ek1=gkgk1;
    12.     if δk<0 then
    13.     ρk:=1ρmax;
    14.     otherwise
    15.     ρk+1:=min{ρmax,max{ρmin,1ρk}};
    16.     end
    17.     calculate gk+1;
    18.     calculate Nk by using the following equality:
                    Nk={Nk1+1,ifgk0.1,Nk1,if0.001gk0.1,Nk11,otherwise,             (4.12)
          in [Nmin,Nmax];
    19.     update (ˆJγ)p(k) by using the following equality:
                    (ˆJγ)p(k)=max0imin(k,Nk){(Jγ)ki},k=0,1,2,,               (4.13)
           which satisfying the following inequality:
                    Jγ(ηk+ωkdk)(ˆJγ)p(k)+μωk(gk)Tdk;                (4.14)
    20.     k := k +1;
    21.   end
    22. η:=ηk, Jγ:=Jγ(ηk);
    23. end
    24. Output: η, Jγ.
    25. construct the optimal solution and optimal value of Problem 1 by using η and Jγ.

     | Show Table
    DownLoad: CSV

    Remark 4. During the past decades, many iterative approaches have been proposed for solving the nonlinear parameter optimization problem by using the information of the objective function [57]. The idea of these iterative approaches is usually that a iterative sequence is generated such that the corresponding objective function value sequence is monotonically decreasing. However, the existing algorithms have the following disadvantage: if an iteration is trapped to a curved narrow valley bottom of the objective function, then the iterative methods will lose their efficiency due to the target with objective function value monotonically decreasing may leading to very short iterative steps. Then, in order to overcome this challenge, an improved gradient-based algorithm is developed based on a novel search approach in Algorithm 1. In this novel search approach, it is not required that the objective function value sequence is always monotonically decreasing. In additions, an improved adaptive strategy for the memory element Nk described by (4.12), which is used in (4.13), is proposed in iterative processes in Algorithm 1. The corresponding explanation on the equality (4.12) is as follows. If the 1st condition described by (4.12) holds, then it implies that the iteration is trapped to a curved narrow valley bottom of the objective function. Thus, in order to avoid creeping along the bottom of this narrow curved valley, the value of the memory element Nk should be increased. If the 2nd condition described by (4.12) holds, then the value of the memory element Nk is better to remain unchanged. If the 3rd condition described by (4.12) holds, then it implies that the iteration is in a flat region. Thus, in order to decrease the objective function value, the value of the memory element N_k will be decreased. Above discussions imply that the novel search approach described in Algorithm 1 is also an adaptive method.

    Remark 5. The sufficient descent condition is extremely important for the convergence of any gradient-based numerical optimization algorithm. Thus, the goal of lines 12–16 described in Algorithm 1 is avoiding uphill directions and keeping \{ \rho_k \} uniformly bounded. As a matter of fact, for any k , \rho _{\min } \leqslant \rho _k \leqslant \rho _{\max } and d_k = - \rho _k g_k ensure that there are two constants l_1 > 0 and l_2 > 0 such that d_k satisfies the following two conditions:

    \begin{equation} \left( {g_k } \right)^{\rm T} d_k \leqslant - l_1 \left\| {g_k } \right\|^2, \end{equation} (4.15)
    \begin{equation} \left\| {d_k } \right\| \leqslant l_2 \left\| {g_k } \right\|. \end{equation} (4.16)

    This section will establish the convergence results of Algorithm 1 developed by Section 4. In order to establish the convergence results of this algorithm, we suppose that the following two conditions hold:

    Assumption 1. J_\gamma is a continuous differentiable function and bounded below on R^{r_1 } \times R^{r_2 } \times R^{r_3 } \times R^{2M} \times R^M .

    Assumption 2. For any \eta_1 \in \Omega and \eta_2 \in \Omega , there is a constant l_3 such that

    \begin{equation} \left\| {g\left( {\rho _1 } \right) - g\left( {\rho _2 } \right)} \right\| \leqslant l_3 \left\| {\rho _1 - \rho _2 } \right\|, \end{equation} (5.1)

    where \Omega presents a open set and g\left(\eta \right) presents the gradient of J_\gamma \left(\eta \right) .

    Theorem 3. Suppose that Assumptions 1 and 2 hold. Let \{ \eta_k \} be a sequence obtained by using Algorithm 1. Then, there is a constant \varsigma > 0 such that the following inequality holds:

    \begin{equation} \left( {J_\gamma } \right)_{k + 1} \leqslant \left( {\hat J_\gamma } \right)_{p\left( k \right)} - \varsigma \left\| {g_k } \right\|^2. \end{equation} (5.2)

    Proof. Let \varsigma _0 be defined by \varsigma _0 = \mathop {\inf }\limits_{\forall k} \left\{ {\omega _k } \right\} \geq 0 .

    If \varsigma _0 > 0 , then by using the inequalities (4.14) and (4.15), one can obtain

    \begin{equation} \left( {J_\gamma } \right)_{k{\rm{ + }}1} \leqslant \left( {\hat J_\gamma } \right)_{p\left( k \right)} - l_1 \varsigma _0 \left\| {g_k } \right\|^2. \end{equation} (5.3)

    Let \varsigma be defined by \varsigma = l_1 \varsigma _0 . Then, the proof of Theorem 1 is complete for \varsigma _0 > 0 .

    If \varsigma _0 = 0 , then there is a subset \Lambda \subseteq \left\{ {0, 1, 2, \cdots } \right\} such that the following equality holds:

    \begin{equation} \mathop {\lim }\limits_{k \in \Lambda , \;k \to \infty } \omega _k = 0, \end{equation} (5.4)

    which indicates that there exists a \hat k such that the following inequality holds:

    \begin{equation} \frac{{\omega _k }} {\varpi } \leqslant 1, \end{equation} (5.5)

    for any k > \hat k and {k \in \Lambda } . Let \omega = \omega _k \varpi . Then, the inequality (4.14) does not hold. That is, one can obtain

    \begin{equation} J_\gamma \left( {\eta _k + \omega d_k } \right) \leqslant \left( {\hat J_\gamma } \right)_{p\left( k \right)} + \mu \omega \left( {g_k } \right)^{\rm T} d_k, \end{equation} (5.6)

    which implies

    \begin{equation} \left( {J_\gamma } \right)_k - J_\gamma \left( {\eta _k + \omega d_k } \right) < - \mu \omega \left( {g_k } \right)^{\rm T} d_k. \end{equation} (5.7)

    Applying the mean value theorem to the left-hand side of the inequality (5.7) yields

    \begin{equation} - \omega \left( {g\left( {\eta _k + \zeta _k \omega d_k } \right)} \right)^{\rm T} d_k < - \mu \omega \left( {g_k } \right)^{\rm T} d_k, \end{equation} (5.8)

    where 0 \leqslant \zeta _k \leqslant 1 . From the inequality (5.8), one obtain

    \begin{equation} \left( {g\left( {\eta _k + \zeta _k \omega d_k } \right)} \right)^{\rm T} d_k > \mu \left( {g_k } \right)^{\rm T} d_k. \end{equation} (5.9)

    By using Assumption 2 and Cauchy-Schwartz inequality, from (4.15) and (5.9), we have

    l_3 \omega \left\| {d_k } \right\|^2 \geqslant \left\| {g\left( {\eta _k + \omega \zeta _k d_k } \right) - g_k } \right\|\left\| {d_k } \right\| \geqslant \left( {g\left( {\eta _k + \omega \zeta _k d_k } \right) - g_k } \right)^{\rm T} d_k
    \begin{equation} \geqslant - \left( {1 - \mu } \right)\left( {g_k } \right)^{\rm T} d_k \geqslant l_1 \left( {1 - \mu } \right)\left\| {g_k } \right\|^2. \end{equation} (5.10)

    Furthermore, applying \omega = \omega _k \varpi and the inequality (4.16) to the inequality (5.10), one obtain

    \begin{equation} \omega _k \geqslant \frac{{l_1 \left( {1 - \mu } \right)\left\| {g_k } \right\|^2 }} {{\varpi l_3 \left\| {d_k } \right\|^2 }} \geqslant \frac{{l_1 \left( {1 - \mu } \right)}} {{\left( {l_2 } \right)^2 \varpi l_3 }} > 0, \end{equation} (5.11)

    for any k > \hat k and {k \in \Lambda } . Clearly, the inequalities (5.4) and (5.11) are contradictory. Thus, \varsigma _0 > 0 . This completes the proof of Theorem 3.

    Lemma 1. Suppose that Assumptions 1 and 2 hold. Let \{ \eta_k \} be a sequence obtained by using Algorithm 1. Then, the following inequalities

    \begin{equation} \mathop {\max }\limits_{1 \leqslant j \leqslant A} J_\gamma \left( {\eta _{Ap + j} } \right) \leqslant \mathop {\max }\limits_{1 \leqslant j \leqslant A} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + j} } \right) - \varsigma \mathop {\min }\limits_{1 \leqslant j \leqslant A} \left\| {g_{Ap + j - 1} } \right\|^2, \end{equation} (5.12)
    \begin{equation} \sum\limits_{p = 1}^\infty {\mathop {\min }\limits_{1 \leqslant j \leqslant A} } \left\| {g_{Ap + j - 1} } \right\|^2 < + \infty, \end{equation} (5.13)

    are true, where A = N_{\max } .

    Proof. Note that if the following inequality

    \begin{equation} J_\gamma \left( {\eta _{Ap + j} } \right) \leqslant \mathop {\max }\limits_{1 \leqslant j \leqslant A} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + j} } \right) - \varsigma \left\| {g_{Ap + j - 1} } \right\|^2, j = 1, 2, \cdots, A, \end{equation} (5.14)

    is true, then the inequality (5.12) also holds. Here, the inequality (5.14) will be proved by using mathematical induction.

    Firstly, Theorem 3 indicates

    \begin{equation} J_\gamma \left( {\eta _{Ap + 1} } \right) \leqslant \mathop {\max }\limits_{1 \leqslant j \leqslant q\left( {Ap} \right)} J_\gamma \left( {\eta _{Ap + j} } \right) - \varsigma \left\| {g_{Ap} } \right\|^2, \end{equation} (5.15)

    where q\left({Ap} \right) = \min \left\{ {Ap, N_{Ap} } \right\} . By using 0 \leqslant q\left({Ap} \right) \leqslant A and the inequality (5.15), one can derive that the inequality (5.14) is true for j = 1 .

    Suppose that the inequality (5.14) is true for 1 \leqslant j \leqslant A - 1 . Note that \varsigma > 0 and the term \left\| {g_{Ap + j - 1} } \right\|^2 described in (5.14) is nonnegative. Then, one can obtain

    \begin{equation} \mathop {\max }\limits_{1 \leqslant i \leqslant j} J_\gamma \left( {\eta _{Ap + i} } \right) \leqslant \mathop {\max }\limits_{1 \leqslant i \leqslant A} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + i} } \right), \end{equation} (5.16)

    for 1 \leqslant j \leqslant A - 1 .

    Next, by using 0 \leqslant q\left({Ap} \right) \leqslant A , the inequality (5.2), and the inequality (5.16), one can derive

    J_\gamma \left( {\eta _{Ap + j + 1} } \right) \leqslant \mathop {\max }\limits_{1 \leqslant i \leqslant q\left( {Ap + j} \right)} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + j + 1} } \right) - \varsigma \left\| {g_{Ap + j} } \right\|^2
    \leqslant \max \left\{ {\mathop {\max }\limits_{1 \leqslant i \leqslant A} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + i} } \right), \mathop {\max }\limits_{1 \leqslant i \leqslant j} J_\gamma \left( {\eta _{Ap + i} } \right)} \right\} - \varsigma \left\| {g_{Ap + j} } \right\|^2
    \begin{equation} \leqslant \mathop {\max }\limits_{1 \leqslant i \leqslant A} J_\gamma \left( {\eta _{A\left( {p - 1} \right) + i} } \right) - \varsigma \left\| {g_{Ap + j} } \right\|^2, \end{equation} (5.17)

    which implies that the inequality (5.14) is also true for j+1 . Then, the inequality (5.14) is true for 1 \leqslant j \leqslant A by using mathematical induction. Thus, the inequality (5.12) holds.

    In additions, Assumption 1 shows J_\gamma being a continuous differentiable function and bounded below on R^{r_1 } \times R^{r_2 } \times R^{r_3 } \times R^{2M} \times R^M , which indicates that

    \begin{equation} \mathop {\max }\limits_{1 \leqslant i \leqslant A} J_\gamma \left( {\eta _{Ap + i} } \right) > - \infty. \end{equation} (5.18)

    Then, summing the inequality (5.12) over p yields

    \sum\limits_{p = 1}^\infty {\mathop {\min }\limits_{1 \leqslant j \leqslant A} } \left\| {g_{Ap + j - 1} } \right\|^2 < + \infty.

    Thus, the inequality (5.13) holds. This completes the proof of Lemma 1.

    Theorem 4. Suppose that these conditions of Theorem 3 are true. Then, the following equality holds:

    \begin{equation} \mathop {\lim }\limits_{k \to \infty } \left\| {g\left( {\eta _k } \right)} \right\| = 0, \end{equation} (5.19)

    where g\left({\eta _k } \right) presents the gradient of the objective function J_\gamma described by (4.1) at the point \eta_k .

    Proof. Firstly, the following result will be proved: there is a constant l_4 such that

    \begin{equation} \left\| {g\left( {\eta _{k + 1} } \right)} \right\| \leqslant l_4 \left\| {g\left( {\eta _k } \right)} \right\|. \end{equation} (5.20)

    By using Assumptions 1 and 2, one can obtain

    \left\| {g\left( {\eta _{k + 1} } \right)} \right\| \leqslant \left\| {g\left( {\eta _{k + 1} } \right) - g\left( {\eta _k } \right) + g\left( {\eta _k } \right)} \right\|
    \leqslant \left\| {g\left( {\eta _{k + 1} } \right) - g\left( {\eta _k } \right)} \right\| + \left\| {g\left( {\eta _k } \right)} \right\|
    \leqslant l_3 \omega _k \left\| {d_k } \right\| + \left\| {g\left( {\eta _k } \right)} \right\|
    \begin{equation} \leqslant \left( {1 + l_2 l_3 \omega _k} \right)\left\| {g\left( {\eta _k } \right)} \right\|. \end{equation} (5.21)

    Let the constant l_4 be defined by l_4 = 1 + l_2 l_3 \omega _k . Then, the inequality (5.21) implies that the inequality (5.20) is true.

    Define the function \psi \left(p \right) by

    \begin{equation} \psi \left( p \right) = \mathop {\arg \min }\limits_{0 \leqslant j \leqslant A - 1} \left\| {g\left( {\eta _{Ap + j} } \right)} \right\|. \end{equation} (5.22)

    Then, Lemma 1 indicates that the following equality holds:

    \begin{equation} \mathop {\lim }\limits_{p \to \infty } \left\| {g\left( {\eta _{Ap + \psi \left( p \right)} } \right)} \right\| = 0. \end{equation} (5.23)

    By using the inequality (5.20), one can obain

    \begin{equation} \left\| {g\left( {\eta _{A\left( {p + 1} \right) + j} } \right)} \right\| \leqslant l_4^{2A} \left\| {g\left( {\eta _{Ap + \psi \left( p \right)} } \right)} \right\|, \quad j = 0, 1, \cdots , A - 1. \end{equation} (5.24)

    Thus, from (5.23) and (5.24), one can deduce that the equality (5.19) is true. This completes the proof of Theorem 4.

    In this section, an optimal feedback control problem of 1, 3-propanediol fermentation processes is provided to illustrate the effectiveness of the approach developed by Sections 2–5, and the numerical simulations are all implemented on a personal computer with Intel Pentium Skylake dual core processor i5-6200U CPU(2.3GHz).

    The 1, 3-propanediol fermentation process can be described by switching between two subsystem: batch subsystem and feeding subsystem. There exists no input feed during the batch subsystem, while alkali and glycerol will be added to the fermentor during the feeding subsystem. In generally, the subsystem switching will happen, if the glycerol concentration reaches the given upper and lower thresholds. By using the result of the work [58], the 1, 3-propanediol fermentation process can be modeled as the following switched dynamical system under state-dependent switching:

    \begin{equation} \left\{ \begin{array}{l} Subsystem\;1: \ \frac{{dx\left( t \right)}} {{dt}} = f_1 \left( {x\left( t \right), t} \right), \quad \quad \quad \quad if \; x_3 \left( t \right) - \alpha _1 = 0, \hfill \cr Subsystem\;2: \ \frac{{dx\left( t \right)}} {{dt}} = f_1 \left( {x\left( t \right), t} \right) + f_2 \left( {x\left( t \right), u\left( t \right), t} \right), \, \, \, if \; x_3 \left( t \right) - \alpha _2 = 0, \hfill \cr \end{array} \right. t \in \left[ {0, \;t_f } \right], \end{equation} (6.1)

    where t_f denotes the given terminal time; the system states x_1 \left(t \right) , x_2 \left(t \right) , x_3 \left(t \right) , x_4 \left(t \right) denote the volume of fluid ( L ), the concentration of biomass ( gL^{-1} ), the concentration of glycerol ( mmolL^{-1} ), the concentration of 1, 3-propanediol ( mmolL^{-1} ), respectively; the control input u\left(t \right) denotes the feeding rate ( Lh^{-1} ); x\left(t \right) = \left[{x_1 \left(t \right), x_2 \left(t \right), x_3 \left(t \right), x_4 \left(t \right)} \right]^{\rm T} denotes the system state vector; Subsystem 1 and Subsystem 2 denote the batch subsystem and the feeding subsystem, respectively; \alpha _1 and \alpha _2 (two parameters that need to be optimized) denote the upper and lower of the glycerol concentration, respectively; and the functions f_1 \left({x\left(t \right), t} \right) , f_2 \left({x\left(t \right), u\left(t \right), t} \right) are given by

    \begin{equation} f_1 \left( {x\left( t \right), t} \right) = \left( \begin{array}{l} 0 \\ {\varphi \left( {x_3 \left( t \right), x_4 \left( t \right)} \right)x_2 \left( t \right)} \\ { - \Delta _1 \left( {x_3 \left( t \right), x_4 \left( t \right)} \right)x_2 \left( t \right)} \\ {\Delta _2 \left( {x_3 \left( t \right), x_4 \left( t \right)} \right)x_2 \left( t \right)} \\ \end{array} \right), \end{equation} (6.2)
    \begin{equation} f_2 \left( {x\left( t \right), u\left( t \right), t} \right) = \frac{{u\left( t \right)}} {{x_1 \left( t \right)}}\left( \begin{array}{l} {x_1 \left( t \right)} \\ { - x_2 \left( t \right)} \\ {l_5 l_6 - x_3 \left( t \right)} \\ { - x_4 \left( t \right)} \\ \end{array} \right). \end{equation} (6.3)

    Subsystem 1 is essentially a natural fermentation process due to no input feed. The functions \varphi , \Delta_1 , and \Delta_2 are defined by

    \begin{equation} \varphi \left( {x_3 \left( t \right), x_4 \left( t \right)} \right) = \frac{{h_1 x_3 \left( t \right)}} {{x_3 \left( t \right) + Y_1 }}\left( {1 - \frac{{x_3 \left( t \right)}} {{x_3^ * }}} \right)\left( {1 - \frac{{x_4 \left( t \right)}} {{x_4^ * }}} \right), \end{equation} (6.4)
    \begin{equation} \Delta _1 \left( {x_3 \left( t \right), x_4 \left( t \right)} \right){\rm{ = }}l_7 + Z_1 \varphi \left( {x_3 \left( t \right), x_4 \left( t \right)} \right) + \frac{{h_2 x_3 \left( t \right)}} {{x_3 \left( t \right) + Y_2 }}, \end{equation} (6.5)
    \begin{equation} \Delta _2 \left( {x_3 \left( t \right), x_4 \left( t \right)} \right) = - l_8 + Z_2 \varphi \left( {x_3 \left( t \right), x_4 \left( t \right)} \right) + \frac{{h_3 x_3 \left( t \right)}} {{x_3 \left( t \right) + Y_3 }}, \end{equation} (6.6)

    which denote the growth rate of cell, the consumption rate of substrate, and the formation rate of 1, 3-propanediol, respectively. In the equality (6.4), the parameters x_3^ * and x_4^ * denote the critical concentrations of glycerol and 1, 3-propanediol, respectively; h_1 , h_2 , h_3 , Y_1 , Y_2 , Y_3 , Z_1 , Z_2 , l_7 , and l_8 are given parameters.

    Note that the feeding subsystem doesn't only consist of the natural fermentation process. Thus, the function f_2 \left({x\left(t \right), u\left(t \right), t} \right) is provided to describe the process dynamics because of the control input feed in Subsystem 2. In the equality (6.3), the given parameters l_5 and l_6 denote the proportion and concentration of glycerol in the control input feed, respectively.

    In generally, as the increase of the biomass, the consumption of glycerol also increases. Then, during Subsystem 1 (batch subsystem), the concentration of glycerol will eventually become too low due to no new glycerol being added. Thus, Subsystem 1 will switch to Subsystem 2 (feeding subsystem), when the equality x_3 \left(t \right) - \alpha _2 = 0 (the active condition of Subsystem 2) satisfies. On the other hand, during Subsystem 2 (feeding subsystem), the concentration of glycerol will eventually become too high due to new glycerol being added. This will inhibit the growth of cell. Thus, Subsystem 2 will switch to Subsystem 1 (batch subsystem), when the equality x_3 \left(t \right) - \alpha _1 = 0 (the active condition of Subsystem 1) satisfies.

    Suppose that the feeding rate u \left(t \right) , the upper of the glycerol concentration \alpha _1 , and the lower of the glycerol concentration \alpha _2 satisfy the following bound constraints:

    \begin{equation} 1.0022 \leqslant u\left( t \right) \leqslant 1.9390, \end{equation} (6.7)
    \begin{equation} 295 \leqslant \alpha _1 \leqslant 605, \end{equation} (6.8)
    \begin{equation} 45 \leqslant \alpha _2 \leqslant 265, \end{equation} (6.9)

    respectively.

    The model parameters of the dynamic optimization problem for the 1, 3-propanediol fermentation process are presented by

    h_1 = 0.8041, \quad h_2 = 7.8296, \quad h_3 = 20.2518, \quad Y_1 = 0.4901, \quad Y_2 = 9.4628, \quad Y_3 = 38.6596,
    Z_1 = 144.9216, Z_2 = 80.8538, l_5 = 0.5698, l_6 = 10759.0000 \; mmolL^{-1}, l_7 = 0.2981, l_8 = 12.2603,
    x_3^* = 2040.0000 \; mmolL^{-1}, \quad x_4^* = 1035.0000 \; mmolL^{-1}, \quad t_f = 25.0000 \; hours, \quad M = 9,
    x_0 = \left[ {5.0000, \; 0.1113, \;496.0000, \; 0.0000} \right]^{\rm T}.

    Suppose that the control input u\left(t \right) takes the piecewise state-feedback controller u\left(t \right) = \sum\limits_{i = 1}^{M} {k_i x\left(t \right)\chi _{\left[{\tau _{i - 1}, \tau _i } \right)} \left(t \right)} . Our main objective is to maximize the concentration of 1, 3-propanediol at the terminal time t_f . Thus, the optimal feedback control problem of 1, 3-propanediol fermentation processes can be presented as follows: choose a control input u\left(t \right) to minimize the objective function J \left(u\left(t \right) \right) = - x_4 \left(t \right) subject to the switched dynamical system described by (6.1) with with the initial condition x \left(0 \right) = x_0 and the bound constraints (6.7–6.9). Then, the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) is adopted to solve the optimal feedback control problem of 1, 3-propanediol fermentation processes by using Matlab 2010a. The optimal objective function value is J^* = -x_4 \left(t_f \right) = -1265.5597 and the optimal values of the parameters \alpha_1 and \alpha_2 are 584.3908 and 246.5423 , respectively. The optimal feedback gain matrixes K_i^* , i = 1, \cdots, 9 are presented by

    K_1^* = \left[ {0, \; 0, \; 0, \; 0} \right], \quad K_2^* = \left[ {0.0140, \; 0.0039, \; 1.1300, \; 0.4786} \right], \quad K_3^* = \left[ {0, \; 0, \; 0, \; 0} \right],
    K_4^* = \left[ {0.0095, \; 0.0058, \; 0.7149, \; 0.6392} \right], K_5^* = \left[ {0, \; 0, \; 0, \; 0} \right], \\ K_6^* = \left[ {0.0082, \; 0.0069, \; 0.6080, \; 0.8297} \right],
    K_7^* = \left[ {0, \; 0, \; 0, \; 0} \right], \quad K_8^* = \left[ {0.0084, \; 0.0073, \; 0.5711, \; 1.0615} \right], \quad K_9^* = \left[ {0, \; 0, \; 0, \; 0} \right],

    and the corresponding numerical simulation results are presented by Figures 14.

    Figure 1.  The optimal volume ( L ) of fluid: x_1(t) .
    Figure 2.  The optimal concentration ( gL^{-1} ) of biomass: x_2(t) .
    Figure 3.  The optimal concentration ( mmolL^{-1} ) of glycerol: x_3(t) .
    Figure 4.  The optimal concentration ( mmolL^{-1} ) of 1, 3-propanediol: x_4(t) .

    Note that Problem 6 is an optimal control problem of nonlinear dynamical systems with state constraints. Thus, the finite difference approximation approach developed by Nikoobin and Moradi [59] can also be applied for solving this dynamic optimization problem of 1, 3-propanediol fermentation processes. In order to compare with the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3), the finite difference approximation approach developed by Nikoobin and Moradi [59] is also adopted for solving this dynamic optimization problem of 1, 3-propanediol fermentation process with the same model parameters under the same condition, and the numerical comparison results are presented by Figure 5 and Table 1.

    Figure 5.  Convergence rates for the finite difference approximation approach developed by Nikoobin and Moradi [59] and the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3).
    Table 1.  The comparison results between the finite difference approximation approach developed by Nikoobin and Moradi [59] and the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3).
    Algorithm Computation time (second) x_4(t_f)
    The finite difference approximation approach developed by Nikoobin and Moradi [59] 1165.3872 1052.9140
    The improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) 439.1513 1265.5597

     | Show Table
    DownLoad: CSV

    Figure 5 shows that the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) takes only 67 iterations to obtain the satisfactory result x_4(t_f) = 1265.5597 , while the finite difference approximation approach developed by Nikoobin and Moradi [59] takes 139 iterations to achieve the satisfactory result x_4(t_f) = 1052.9140 . That is, the iterations of the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) is reduced by 51.7986\% . In additions, Table 1 also shows that the result x_4(t_f) = 1052.9140 obtained by using the finite difference approximation approach developed by Nikoobin and Moradi [59] is not superior to the result ( x_4(t_f) = 1265.5597 ) obtained by using the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) with saving 60.4695\% computation time.

    In conclusion, the above numerical simulation results show that the improved gradient-based numerical optimization algorithm (Algorithm 1 described by Section 4.3) is low time-consuming, has faster convergence speed, and can obtain a better numerical optimization than the finite difference approximation approach developed by Nikoobin and Moradi [59]. That is, an effective numerical optimization algorithm is presented for solving the dynamic optimization problem of 1, 3-propanediol fermentation process.

    In this paper, the dynamic optimization problem for a class of fed-batch fermentation processes is modeled as an optimal control problem of switched dynamical systems under state-dependent switching, and a general state-feedback controller is designed for this dynamic optimization problem. Then, by introducing a discrete-valued function and using a relaxation technique, this problem is transformed into a nonlinear parameter optimization problem. Next, an improved gradient-based algorithm is developed based on a novel search approach, and a large number of numerical experiments show that this novel search approach can effectively improve the convergence speed of this algorithm, when an iteration is trapped to a curved narrow valley bottom of the objective function. Finally, an optimal feedback control problem of 1, 3-propanediol fermentation processes is provided to illustrate the effectiveness of this method developed by this paper, and the numerical simulation results show that this method developed by this paper is low time-consuming, has faster convergence speed, and obtains a better result than the existing approaches. In the future, we will continue to study the dynamic optimization problem for a class of fed-batch fermentation processes with uncertainty constraints.

    The authors express their sincere gratitude to the anonymous reviewers for their constructive comments in improving the presentation and quality of this manuscript. This work was supposed by the National Natural Science Foundation of China under Grant Nos. 61963010 and 61563011, and the Special Project for Cultivation of New Academic Talent and Innovation Exploration of Guizhou Normal University in 2019 under Grant No. 11904-0520077.

    The authors declare no conflicts of interest.



    [1] J. Kennedy, R. Eberhart, Particle swarm optimization, in Proceedings of ICNN'95 - International Conference on Neural Networks, 4 (1995), 1942–1948. https://doi.org/10.1109/ICNN.1995.488968
    [2] A. H. Gandomi, A. H. Alavi, Krill herd: a new bio-inspired optimization algorithm, Commun. Nonlinear Sci. Numer. Simul., 17 (2012), 4831–4845. https://doi.org/10.1016/j.cnsns.2012.05.010 doi: 10.1016/j.cnsns.2012.05.010
    [3] A. H. Gandomi, X. Yang, A. H. Alavi, Cuckoo search algorithm: a metaheuristic approach to solve structural optimization problems, Eng. Comput., 29 (2013), 17–35. https://doi.org/10.1007/s00366-011-0241-y doi: 10.1007/s00366-011-0241-y
    [4] S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Adv. Eng. Software, 69 (2014), 46–61. https://doi.org/10.1016/j.advengsoft.2013.12.007 doi: 10.1016/j.advengsoft.2013.12.007
    [5] R. V. Rao, V. J. Savsani, D. P. Vakharia, Teaching–learning-based optimization: a novel method for constrained mechanical design optimization problems, Comput.-Aided Des., 43 (2011), 303–315. https://doi.org/10.1016/j.cad.2010.12.015 doi: 10.1016/j.cad.2010.12.015
    [6] S. Mirjalili, A. Lewis, The whale optimization algorithm, Adv. Eng. Software, 95 (2016), 51–67. https://doi.org/10.1016/j.advengsoft.2016.01.008 doi: 10.1016/j.advengsoft.2016.01.008
    [7] A. Faramarzi, M. Heidarinejad, S. Mirjalili, A. Gandomi, Marine predators algorithm: a nature-inspired metaheuristic, Expert Syst. Appl., 152 (2020), 113377. https://doi.org/10.1016/j.eswa.2020.113377 doi: 10.1016/j.eswa.2020.113377
    [8] J. Sarvaiya, D. Singh, Selection of the optimal process parameters in friction stir welding/processing using particle swarm optimization algorithm, Mater. Today: Proc., 62 (2022), 896–901. https://doi.org/10.1016/j.matpr.2022.04.062 doi: 10.1016/j.matpr.2022.04.062
    [9] Z. Hu, H. Norouzi, M. Jiang, S. Dadfar, T. Kashiwagi, Novel hybrid modified krill herd algorithm and fuzzy controller based MPPT to optimally tune the member functions for PV system in the three-phase grid-connected mode, ISA trans., 2022 (2022). https://doi.org/10.1016/j.isatra.2022.02.009 doi: 10.1016/j.isatra.2022.02.009
    [10] Q. Bai, H. Li, The application of hybrid cuckoo search-grey wolf optimization algorithm in optimal parameters identification of solid oxide fuel cell, Int. J. Hydrogen Energy, 47 (2022), 6200–6216. https://doi.org/10.1016/j.ijhydene.2021.11.216 doi: 10.1016/j.ijhydene.2021.11.216
    [11] C. Song, X. Wang, Z. Liu, H. Chen, Evaluation of axis straightness error of shaft and hole parts based on improved grey wolf optimization algorithm, Measurement, 188 (2022), 110396. https://doi.org/10.1016/j.measurement.2021.110396 doi: 10.1016/j.measurement.2021.110396
    [12] H. Abaeifar, H. Barati, A. R. Tavakoli, Inertia-weight local-search-based TLBO algorithm for energy management in isolated micro-grids with renewable resources, Int. J. Electr. Power Energy Syst., 137 (2022), 107877. https://doi.org/10.1016/j.ijepes.2021.107877 doi: 10.1016/j.ijepes.2021.107877
    [13] V. K. Jadoun, G. R. Prashanth, S. S. Joshi, K. Narayanan, H. Malik, F. García Márquez, Optimal fuzzy based economic emission dispatch of combined heat and power units using dynamically controlled Whale Optimization Algorithm, Appl. Energy, 315 (2022), 119033. https://doi.org/10.1016/j.apenergy.2022.119033 doi: 10.1016/j.apenergy.2022.119033
    [14] M. Al-qaness, A. Ewees, H. Fan, L. Abualigah, M. Elaziz, Boosted ANFIS model using augmented marine predator algorithm with mutation operators for wind power forecasting, Appl. Energy, 314 (2022), 118851. https://doi.org/10.1016/j.apenergy.2022.118851 doi: 10.1016/j.apenergy.2022.118851
    [15] M. Al-qaness, A. Ewees, H. Fan, A. Airassas, M. Elaziz, Modified aquila optimizer for forecasting oil production, Geo-spatial Inf. Sci., 2022 (2022), 1–17. https://doi.org/10.1080/10095020.2022.2068385 doi: 10.1080/10095020.2022.2068385
    [16] A. Dahou, M. Al-qaness, M. Elaziz, A. Helmi, Human activity recognition in IoHT applications using Arithmetic Optimization Algorithm and deep learning, Measurement, 199 (2022), 111445. https://doi.org/10.1016/j.measurement.2022.111445 doi: 10.1016/j.measurement.2022.111445
    [17] M. Elaziz, A. Ewees, M. Al-qaness, L. Abualigah, R. Ibrahim, Sine–Cosine-Barnacles Algorithm Optimizer with disruption operator for global optimization and automatic data clustering, Expert Syst. Appl., 207 (2022), 117993. https://doi.org/10.1016/j.eswa.2022.117993 doi: 10.1016/j.eswa.2022.117993
    [18] X. Chen, X. Qi, Z. Wang, C. Cui, B. Wu, Y. Yang, Fault diagnosis of rolling bearing using marine predators algorithm-based support vector machine and topology learning and out-of-sample embedding, Measurement, 176 (2021), 109116. https://doi.org/10.1016/j.measurement.2021.109116 doi: 10.1016/j.measurement.2021.109116
    [19] P. H. Dinh, A novel approach based on three-scale image decomposition and marine predators algorithm for multi-modal medical image fusion, Biomed. Signal Process. Control, 67 (2021), 102536. https://doi.org/10.1016/j.bspc.2021.102536 doi: 10.1016/j.bspc.2021.102536
    [20] M. A. Sobhy, A. Y. Abdelaziz, H. M. Hasanien, M. Ezzat, Marine predators algorithm for load frequency control of modern interconnected power systems including renewable energy sources and energy storage units, Ain Shams Eng. J., 12 (2021), 3843–3857. https://doi.org/10.1016/j.asej.2021.04.031 doi: 10.1016/j.asej.2021.04.031
    [21] A. Faramarzi, M. Heidarinejad, S. Mirjalili, A. Gandomi, Marine predators algorithm: a nature-inspired metaheuristic, Expert Syst. Appl., 152 (2020), 113377. https://doi.org/10.1016/j.eswa.2020.113377 doi: 10.1016/j.eswa.2020.113377
    [22] M. A. Elaziz, D. Mohammadi, D. Oliva, K. Salimifard, Quantum marine predators algorithm for addressing multilevel image segmentation, Appl. Soft Comput., 110 (2021), 107598. https://doi.org/10.1016/j.asoc.2021.107598 doi: 10.1016/j.asoc.2021.107598
    [23] M. Ramezani, D. Bahmanyar, N. Razmjooy, A new improved model of marine predator algorithm for optimization problems, Arabian J. Sci. Eng., 46 (2021), 8803–8826. https://doi.org/10.1007/s13369-021-05688-3 doi: 10.1007/s13369-021-05688-3
    [24] M. Abdel-Basset, D. El-Shahat, R. K. Chakrabortty, M. Ryan, Parameter estimation of photovoltaic models using an improved marine predators algorithm, Energy Convers. Manage., 227 (2021), 113491. https://doi.org/10.1016/j.enconman.2020.113491 doi: 10.1016/j.enconman.2020.113491
    [25] K. Zhong, G. Zhou, W. Deng, Y. Zhou, Q. Luo, MOMPA: multi-objective marine predator algorithm, Comput. Methods Appl. Mech. Eng., 385 (2021), 114029. https://doi.org/10.1016/j.cma.2021.114029 doi: 10.1016/j.cma.2021.114029
    [26] R. Sowmya, V. Sankaranarayanan, Optimal vehicle-to-grid and grid-to-vehicle scheduling strategy with uncertainty management using improved marine predator algorithm, Comput. Electr. Eng., 100 (2022), 107949. https://doi.org/10.1016/j.compeleceng.2022.107949 doi: 10.1016/j.compeleceng.2022.107949
    [27] E. H. Houssein, I. E. Ibrahim, M. Kharrich, S. Kamel, An improved marine predators algorithm for the optimal design of hybrid renewable energy systems, Eng. Appl. Artif. Intell., 110 (2022), 104722. https://doi.org/10.1016/j.engappai.2022.104722 doi: 10.1016/j.engappai.2022.104722
    [28] D. Yousri, A. Ousama, Y. Shaker, A. Fathy, T. Babu, H. Rezk, et al., Managing the exchange of energy between microgrid elements based on multi-objective enhanced marine predators algorithm, Alexandria Eng. J., 61 (2022), 8487–8505. https://doi.org/10.1016/j.aej.2022.02.008 doi: 10.1016/j.aej.2022.02.008
    [29] Y. Ma, X. Zhang, J. Song, L. Chen, A modified teaching–learning-based optimization algorithm for solving optimization problem, Knowledge-Based Syst., 212 (2020), 106599. https://doi.org/10.1016/j.knosys.2020.106599 doi: 10.1016/j.knosys.2020.106599
    [30] N. E. Humphries, N. Queiroz, J. Dyer, N. Pade, M. Musyl, K. Schaefer, et al., Environmental context explains Lévy and Brownian movement patterns of marine predators, Nature, 465 (2010), 1066–1069. https://doi.org/10.1038/nature09116 doi: 10.1038/nature09116
    [31] D. W. Sims, E. J. Southall, N. E. Humphries, G. Hays, C. Bradshaw, J. Pitchford, et al., Scaling laws of marine predator search behaviour, Nature, 451 (2008), 1098–1102. https://doi.org/10.1038/nature06518 doi: 10.1038/nature06518
    [32] G. M. Viswanathan, E. P. Raposo, M. Luz, Lévy flights and superdiffusion in the context of biological encounters and random searches, Phys. Life Rev., 5 (2008), 133–150. https://doi.org/10.1016/j.plrev.2008.03.002 doi: 10.1016/j.plrev.2008.03.002
    [33] F. Bartumeus, J. Catalan, U. L. Fulco, M. Lyra, G. Viswanathan, Optimizing the encounter rate in biological interactions: Lévy versus Brownian strategies, Phys. Rev. Lett., 88 (2002), 097901. https://doi.org/10.1103/PhysRevLett.88.097901 doi: 10.1103/PhysRevLett.88.097901
    [34] A. Einstein, Investigations on the theory of the brownian movement, DOVER, 35 (1956), 318–320. https://doi.org/10.2307/2298685 doi: 10.2307/2298685
    [35] J. D. Filmalter, L. Dagorn, P. D. Cowley, M. Taquet, First descriptions of the behavior of silky sharks, Carcharhinus falciformis, around drifting fish aggregating devices in the Indian Ocean, Bull. Mar. Sci., 87 (2011), 325–337. https://doi.org/10.5343/bms.2010.1057 doi: 10.5343/bms.2010.1057
    [36] D. Yousri, H. M. Hasanien, A. Fathy, Parameters identification of solid oxide fuel cell for static and dynamic simulation using comprehensive learning dynamic multi-swarm marine predators algorithm, Energy Convers. Manage., 228 (2021), 113692. https://doi.org/10.1016/j.enconman.2020.113692 doi: 10.1016/j.enconman.2020.113692
    [37] M. Abdel-Basset, R. Mohamed, S. Mirjalili, R. Chakrabortty, M. Ryan, An efficient marine predators algorithm for solving multi-objective optimization problems: analysis and validations, IEEE Access, 9 (2021), 42817–42844. https://doi.org/10.1109/ACCESS.2021.3066323 doi: 10.1109/ACCESS.2021.3066323
    [38] T. Niknam, R. Azizipanah-Abarghooee, M. R. Narimani, A new multi objective optimization approach based on TLBO for location of automatic voltage regulators in distribution systems, Eng. Appl. Artif. Intell., 25 (2012), 1577–1588. https://doi.org/10.1016/j.engappai.2012.07.004 doi: 10.1016/j.engappai.2012.07.004
    [39] T. Niknam, F. Golestaneh, M. S. Sadeghi, θ-Multiobjective teaching–learning-based optimization for dynamic economic emission dispatch, IEEE Syst. J., 6 (2012), 341–352. https://doi.org/10.1109/JSYST.2012.2183276 doi: 10.1109/JSYST.2012.2183276
    [40] R. V. Rao, V. Patel, An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems, Sci. Iran., 20 (2013), 710–720. https://doi.org/10.1016/j.scient.2012.12.005 doi: 10.1016/j.scient.2012.12.005
    [41] P. K. Roy, S. Bhui, Multi-objective quasi-oppositional teaching learning based optimization for economic emission load dispatch problem, Int. J. Electr.Power Energy Syst., 53 (2013), 937–948. https://doi.org/10.1016/j.ijepes.2013.06.015 doi: 10.1016/j.ijepes.2013.06.015
    [42] H. Bouchekara, M. A. Abido, M. Boucherma, Optimal power flow using teaching-learning-based optimization technique, Electr. Power Syst. Res., 114 (2014), 49–59. https://doi.org/10.1016/j.epsr.2014.03.032 doi: 10.1016/j.epsr.2014.03.032
    [43] M. Liu, X. Yao, Y. Li, Hybrid whale optimization algorithm enhanced with Lévy flight and differential evolution for job shop scheduling problems, Appl. Soft Comput., 87 (2020), 105954. https://doi.org/10.1016/j.asoc.2019.105954 doi: 10.1016/j.asoc.2019.105954
    [44] D. Tansui, A. Thammano, Hybrid nature-inspired optimization algorithm: hydrozoan and sea turtle foraging algorithms for solving continuous optimization problems, IEEE Access, 8 (2020), 65780–65800. https://doi.org/10.1109/ACCESS.2020.2984023 doi: 10.1109/ACCESS.2020.2984023
    [45] K. Zhong, Q. Luo, Y. Zhou, M. Jiang, TLMPA: teaching-learning-based marine predators algorithm, AIMS Math., 6 (2021), 1395–1442. https://doi.org/10.3934/math.2021087 doi: 10.3934/math.2021087
  • This article has been cited by:

    1. Jibril Abdullahi Bala, Taliha Abiodun Folorunso, Majeed Soufian, Abiodun Musa Aibinu, Olayemi Mikail Olaniyi, Nimat Ibrahim, 2022, Fuzzy Logic based Fed Batch Fermentation Control Scheme for Plant Culturing, 978-1-6654-7978-3, 1, 10.1109/NIGERCON54645.2022.9803140
    2. Ricardo Aguilar‐López, Pablo A. López‐Pérez, Ricardo Femat, Eduardo Alvarado‐Santos, Improved bioethanol production from cocoa agro‐industrial waste optimization based on reaction rate rules, 2025, 0268-2575, 10.1002/jctb.7815
    3. Dadang Rustandi, Mersi Kurniati, Sensus Wijonarko, Siddiq Wahyu Hidayat, Tatik Maftukhah, , A Control System for Pepper Submersion Tub Actuators, 2025, 2973, 1742-6588, 012009, 10.1088/1742-6596/2973/1/012009
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2535) PDF downloads(171) Cited by(5)

Figures and Tables

Figures(38)  /  Tables(11)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog