Loading [MathJax]/jax/output/SVG/jax.js
Research article Special Issues

Metastability of solitary waves in diatomic FPUT lattices

  • Received: 14 December 2018 Accepted: 25 February 2019 Published: 30 April 2019
  • It is known that long waves in spatially periodic polymer Fermi-Pasta-Ulam-Tsingou lattices are well-approximated for long, but not infinite, times by suitably scaled solutions of Korteweg-de Vries equations. It is also known that dimer FPUT lattices possess nanopteron solutions, i.e., traveling wave solutions which are the superposition of a KdV-like solitary wave and a very small amplitude ripple. Such solutions have infinite mechanical energy. In this article we investigate numerically what happens over very long time scales (longer than the time of validity for the KdV approximation) to solutions of diatomic FPUT which are initially suitably scaled (finite energy) KdV solitary waves. That is we omit the ripple. What we find is that the solitary wave continuously leaves behind a very small amplitude "oscillatory wake." This periodic tail saps energy from the solitary wave at a very slow (numerically sub-exponential) rate. We take this as evidence that the diatomic FPUT "solitary wave" is in fact quasi-stationary or metastable.

    Citation: Nickolas Giardetti, Amy Shapiro, Stephen Windle, J. Douglas Wright. Metastability of solitary waves in diatomic FPUT lattices[J]. Mathematics in Engineering, 2019, 1(3): 419-433. doi: 10.3934/mine.2019.3.419

    Related Papers:

    [1] Nurul Akhma Zakaria, A.A. Kutty, M.A. Mahazar, Marina Zainal Abidin . Arsenic acute toxicity assessment on select freshwater organism species in Malaysia. AIMS Environmental Science, 2016, 3(4): 804-814. doi: 10.3934/environsci.2016.4.804
    [2] Patricia Morcillo, Maria Angeles Esteban, Alberto Cuesta . Mercury and its toxic effects on fish. AIMS Environmental Science, 2017, 4(3): 386-402. doi: 10.3934/environsci.2017.3.386
    [3] Antonio Zanin, Ivonez Xavier de Almeida, Francieli Pacassa, Fabricia Silva da Rosa, Paulo Afonso . Maturity level of environmental management in the pulp and paper supply chain. AIMS Environmental Science, 2021, 8(6): 580-596. doi: 10.3934/environsci.2021037
    [4] Djordje Vilimanovic, Gangadhar Andaluri, Robert Hannah, Rominder Suri, A. Ronald MacGillivray . Occurrence and aquatic toxicity of contaminants of emerging concern (CECs) in tributaries of an urbanized section of the Delaware River Watershed. AIMS Environmental Science, 2020, 7(4): 302-319. doi: 10.3934/environsci.2020019
    [5] Zeynab Rezazadeh Salteh, Saeed Fazayeli, Saeid Jafarzadeh Ghoushchi . Evaluation and prioritization of barriers to the implementation of the eco-regenerative supply chains using fuzzy ZE-numbers framework in group decision-making. AIMS Environmental Science, 2024, 11(4): 516-550. doi: 10.3934/environsci.2024026
    [6] Francesco Teodori . Health physics calculation framework for environmental impact assessment of radiological contamination. AIMS Environmental Science, 2021, 8(4): 403-420. doi: 10.3934/environsci.2021026
    [7] P. U. Singare, J.P. Shirodkar . Persistent and carcinogenic polycyclic aromatic hydrocarbons in the north-western coastal marine environment of India. AIMS Environmental Science, 2021, 8(2): 169-189. doi: 10.3934/environsci.2021012
    [8] Melanie Voigt, Indra Bartels, Anna Nickisch-Hartfiel, Martin Jaeger . Elimination of macrolides in water bodies using photochemical oxidation. AIMS Environmental Science, 2018, 5(5): 372-388. doi: 10.3934/environsci.2018.5.372
    [9] Fernanda P. Maciel, Joshua M. Peschel . A GIS-based tool for bioaccumulation risk analysis and its application to study polychlorinated biphenyls in the Great Lakes. AIMS Environmental Science, 2018, 5(1): 1-23. doi: 10.3934/environsci.2018.1.1
    [10] Jamshid Ali . Environmental resilience: transition to regenerative supply chain management. AIMS Environmental Science, 2024, 11(2): 107-128. doi: 10.3934/environsci.2024007
  • It is known that long waves in spatially periodic polymer Fermi-Pasta-Ulam-Tsingou lattices are well-approximated for long, but not infinite, times by suitably scaled solutions of Korteweg-de Vries equations. It is also known that dimer FPUT lattices possess nanopteron solutions, i.e., traveling wave solutions which are the superposition of a KdV-like solitary wave and a very small amplitude ripple. Such solutions have infinite mechanical energy. In this article we investigate numerically what happens over very long time scales (longer than the time of validity for the KdV approximation) to solutions of diatomic FPUT which are initially suitably scaled (finite energy) KdV solitary waves. That is we omit the ripple. What we find is that the solitary wave continuously leaves behind a very small amplitude "oscillatory wake." This periodic tail saps energy from the solitary wave at a very slow (numerically sub-exponential) rate. We take this as evidence that the diatomic FPUT "solitary wave" is in fact quasi-stationary or metastable.


    With the continuous development and progress of computing technology over the past decade, deep learning has taken a big step on the road of evolution. Deep learning can be found in many fields, such as imaging and natural language. Scholars have begun to apply deep learning to solve some complex partial differential equations (PDEs), which include PDEs with high order derivatives [1], high-dimensional PDEs [2,3], subdiffusion problems with noisy data [4] and so on. Based on the deep learning method, Raissi et al. [5] proposed a novel algorithm called the physics informed neural network (PINN), which has made excellent achievements for solving forward and inverse PDEs. It integrates physical information described by PDEs into a neural network. In recent years, the PINNs algorithm has attracted extensive attention. To solve forward and inverse problems of integro-differential equations (IDEs), Yuan et al. [6] proposed auxiliary physics informed neural network (A-PINN). Lin and Chen [7] designed a two-stage physics informed neural network for approximating localized wave solutions, which introduces the measurement of conserved quantities in stage two. Yang et al. [8] developed Bayesian physics informed neural networks (B-PINNs) that takes Bayesian neural network and Hamiltonian Monte Carlo or the variational inference as a priori and posteriori estimators, respectively. Scholars have also presented other variant algorithms of PINN, such as RPINNs [9] and hp-VPINNs [10]. PINN has also achieved good performance for solving physical problems, including high-speed flows [11] and heat transfer problems [12]. In addition to integer-order differential equations, authors have studied the application of PINN in solving fractional differential equations such as fractional advection-diffusion equations (see Pang et al. [13] for fractional physics informed neural network (fPINNs)), high dimensional fractional PDEs (see Guo et al. [14] for Monte Carlo physics-informed neural networks (MC-PINNs)) and fractional water wave models (see Liu et al. [15] for time difference PINN).

    In recent decades, fractional differential equations (FDEs) have been concerned and studied in many fields, such as image denoising [16] and physics [17,18,19]. The reason why fractional differential equations have attracted wide attention is that they can more clearly describe complex physical phenomena. As an indispensable part of fractional problems, distributed-order differential equations are difficult to solve due to the complexity of distributed-order operators. To solve the time multi-term and distributed-order fractional sub-diffusion equations, Gao et al. [20] proposed a second order numerical difference formula. Jian et al. [21] derived a fast second-order implicit difference scheme of time distributed-order and Riesz space fractional diffusion-wave equations and analyzed the unconditional stability and second-order convergence. Li et al. [22] applied the mid-point quadrature rule with finite volume method to approximate the distributed-order equation. For the nonlinear distributed-order sub-diffusion model [23], the distributed-order derivative and the spatial direction were approximated by the FBN-θ formula with a second-order composite numerical integral formula and the H1-Galerkin mixed finite element method, respectively. In [24], Guo et al. adopted the Legendre-Galerkin spectral method for solving 2D distributed-order space-time reaction-diffusion equations. For the two-dimensional Riesz space distributed-order equation, Zhang et al. [25] used Gauss quadrature to calculate the distributed-order derivative and applied an alternating direction implicit (ADI) Galerkin-Legendre spectral scheme to approximate the spatial direction. For the distributed-order fourth-order sub-diffusion equation, Ran and Zhang [26] developed new compact difference schemes and proved their stability and convergence. In [27,28], authors developed spectral methods for the distributed-order time fractional fourth-order PDEs.

    As we know, distributed-order fractional PDEs can be regarded as the limiting case of multi-term fractional PDEs [29]. Moreover, Diethelm and Ford [30] have observed that small changes in the order of a fractional PDE lead to only slight changes in the final solution, which gives initial support to the employed numerical integration method. In view of this, we develop the FBN-θ [31,32] with a second-order composite numerical integral formula combined with a multi-output neural network for solving 1D and 2D nonlinear time distributed-order models. Based on the idea of using a single output neural network combined with the discrete scheme of fractional models [13], we also use a single output neural network combined with a time discrete scheme to solve the nonlinear time distributed-order models. However, the accuracy of the prediction solution calculated by the single output neural network scheme is low and the training progress takes a lot of time. Therefore, we introduce a multi-output neural network to obtain the numerical solution of the time discrete scheme. Compared with the single output neural network scheme, the proposed multi-output neural network scheme has two main advantages as follows:

    ● Saving more computing time. The multi-output neural network scheme makes the sampling domain of the collocation points from the spatiotemporal domain to the spatial domain, which decreases the number of the training dataset and thus reduces the training time.

    ● Improving the accuracy of predicted solution. Due to the discrete scheme of the distributed-order derivative, the n-th output item of the multi-output neural network will be constrained by the previous n1 output items.

    The remainder of this article is as follows: In Section 2, we show what the components of neural network are and how to construct a neural network. In Section 3, we give the lemmas used to approximate the distributed-order derivative and the process of building the loss function. In Section 4, we provide some numerical results to confirm the capability of our proposed method. Finally, we make some conclusions in Section 5.

    In face of different objectives in various fields, scholars have developed many different types of neural networks, such as feed-forward neural network (FNN) [6], recurrent neural network (RNN) [33] and convolutional neural network (CNN) [34]. The FNN considered in this article can effectively solve most PDEs. Input layer, hidden layer and output layer are three indispensable components of FNN, which can be given, respectively, by

    input layer:Φ0(x)=xRdin,hidden layers:Φk(x)=σ(WkΦk1(x)+bk)Rλk,1kK1,output layer:ΦK(x)=WKΦK1(x)+bKRdout.

    WkRλk×λk1 and bkRλk represent the weight matrix and the bias vector in the kth layer, respectively. We define δ={Wk,bk}1kK, which is the trainable parameters of FNN. λk represents the number of neurons included in the kth layer. σ is a nonlinear activation function. In this article, the hyperbolic tangent function [3,6] is selected as the activation function. There are many other functions that can be considered as activation functions, such as the rectified linear unit (ReLU) σ(x)=max{x,0} [4] and the logistic sigmoid σ(x)=11+ex [35].

    In this article, we consider a nonlinear distributed-order model with the following general form:

    {Dwtu+N(u)=f(x,t),(x,t)Ω×J,u(x,0)=u0(x),xˉΩ, (3.1)

    where ΩRd(d2) and J=(0,T]. N[] is a nonlinear differential operator. Dwtu represents the distributed-order derivative and has the following definition:

    Dwtu(x,t)=10ω(α)C0Dαtu(x,t)dα, (3.2)

    where ω(α)0, 10ω(α)dα=c0>0 and C0Dαtu(x,t) is the Caputo fractional derivative expressed by

    C0Dαtu(x,t)={1Γ(1α)t0uη(x,η)(tη)αdη,0<α<1,ut(x,t),α=1. (3.3)

    The specific boundary condition is determined by the practical problem.

    For simplicity, choosing a mesh size Δα=12I, we denote the nodes on the interval [0,1] with coordinates αi=iΔα for i=0,1,2,,2I. The time interval [0,T] is divided as the uniform mesh with the grid points tn=nΔt(n=0,1,2,,N) and Δt=T/N is the time step size. We denote vnun=u(x,tn), un+12t=un+1unΔt+O(Δt2) and un+12:=un+1+un2. vn is defined as the approximation solution of the time discrete scheme. The following lemmas are introduced to construct the numerical discrete formula of (3.1):

    Lemma 3.1. (See [23]) Supposing ω(α)C2[0,1], we can get

    10ω(α)dα=Δα2Ik=0ciω(αi)Δα212ω(2)(γ),γ(0,1), (3.4)

    where

    ci={12,i=0,2I,1,otherwise. (3.5)

    Lemma 3.2. From [23,31,32], the discrete formula of the Caputo fractional derivative (3.3) can be obtained by

    C0Dαtu(x,tn+12)=C0Dαtun+1+C0Dαtun2+O(Δt2)=Δtαn+1s=0˜κ(α)n+1sus+O(Δt2), (3.6)

    where

    ˜κ(α)n+1s={κ(α)02,s=n+1,κ(α)ns+κ(α)n+1s2,otherwise. (3.7)

    The parameters κ(α)i(i=0,1,,n+1) that are the coefficients of FBN-θ (θ[12,1]) can be given by

    κ(α)i={2α(1+αθ)(32θ)α,i=0,ϕ0κ(α)0ψ0,i=1,12ψ0[(ϕ0ψ1)κ(α)1+ϕ1κ(α)0],i=2,1iψ03j=1[ϕj1(ij)ψj]κ(α)ij,i3, (3.8)

    where

    ϕi={2α(θ1)(αθ+1)+αθ(θ32),i=0,α(2θ23αθ+4αθ21),i=1,αθ(12θ+α2αθ),i=2, (3.9)

    and

    ψi={12(32θ)(1+αθ),i=0,αθ2(32θ)2(1θ)(αθ+1),i=1,12(2θ1)(αθ+1)2αθ(θ1),i=2,12αθ(12θ). (3.10)

    Lemma 3.3. (See [23]) The distributed-order term Dωtu at t=tn+12 can be calculated by the following formula:

    Dωtu(x,tn+12)=Dωtun+1+Dωtun2+O(Δt2)=12n+1s=0βn+1sus+O(Δt2+Δα2), (3.11)

    where

    βn+1s={ˆκns+ˆκn+1s,0s<n+1,ˆκ0,s=n+1, (3.12)

    and

    ˆκns=2Ii=0φiΔtαiκ(αi)ns,φi=Δαω(αi)ci. (3.13)

    Based on the above lemmas, the discrete scheme of the distributed-order model (3.1) at t=tn+12(n=0,1,2,,N1) can be expressed by the following equality:

    12n+1s=0βn+1svs+N(vn+12)=fn+12. (3.14)

    Then we can obtain the system of equations as follows:

    {121s=0β1svs+N(v12)=f12,122s=0β2svs+N(v1+12)=f1+12,12Ns=0βNsvs+N(vN12)=fN12. (3.15)

    The system of Eq (3.15) can be rewritten as the following matrix form:

    v(x)M+N(v(x))+v0(x)ρ0=f(x), (3.16)

    where the symbols ρ0, v(x), f(x) and N(v(x)) are vectors, which are given, respectively, by

    ρ0=[12β10,12β20,12β30,,12βN0],v(x)=[v1(x),v2(x),,vN(x)],f(x)=[f12(x),f1+12(x),,fN12(x)],N(v(x))=[N(v12(x)),N(v1+12(x)),,N(vN12(x))].

    The symbol M is an N×N matrix that has the following definition:

    M=(12β1112β2112βN1012β2212βN20012βNN).

    Now, we introduce a multi-output neural network v(x;δ)=[v1(x;δ),v2(x;δ),,vN(x;δ)] into Eq (3.16), which takes x as an input and is used to approximate time discrete solutions v(x)=[v1(x),v2(x),,vN(x)]. This will result in a multi-output PINN (x)=[1(x),2(x),,N(x)]:

    (x)=v(x;δ)M+N(v(x;δ))+v0(x)ρ0f(x), (3.17)

    where n+1(x) is denoted as residual error of the discrete scheme (3.14), which is given by

    n+1(x)=12n+1s=1βn+1svs(x;δ)+N(vn+12(x;δ))+12βn+10v0(x)fn+12(x),n=0,1,2,,N1. (3.18)

    The loss function is constructed in the form of mean square error. Combined with the boundary condition loss, the total loss function can be expressed by the following formula:

    MSEtotal=MSE+MSEb, (3.19)

    where

    MSE=1N×NxNj=1Nxi=1|j(xi)|2, (3.20)

    and boundary condition loss

    MSEb=1N×NbNj=1Nbi=1|vj(xib;δ)uj(xib)|2. (3.21)

    Here, {xi}Nxi=1 corresponds to the collocation points on the space domain Ω and {xib}Nbi=1 denotes the boundary training data. The schematic diagram of using the multi-output neural network scheme to solve nonlinear time distributed-order models is shown in Figure 1.

    Figure 1.  A multi-output neural network framework to solve nonlinear time distributed-order models, where MSE represents the loss function.

    In this section, we consider two nonlinear time distributed-order equations to verify the feasibility and effectiveness of our proposed method. The performance is evaluated by calculating the relative L2 error between the predicted and exact solutions. The definition of relative L2 error is given by

    ||uv||L2=Nj=1Nxi=1|uj(xi)vj(xi)|2Nj=1Nxi=1|uj(xi)|2. (4.1)

    Table 1 indicates which optimizer is selected for each example to minimize the loss function.

    Table 1.  The hyperparameters configured in numerical examples.
    Example Optimizer Learning rate Iterations
    1 Adam + L-BFGS 0.001 20000
    2 Adam + L-BFGS 0.001 20000
    3 Adam + L-BFGS 0.001 20000
    4 L-BFGS - -

     | Show Table
    DownLoad: CSV

    We use Python to code our algorithms and all codes run on a Lenovo laptop with AMD R7-6800H CPU @ 3.20 GHz and 16.0GB RAM.

    Here, we solve the following distributed-order sub-diffusion model:

    ut+DωtuΔuΔut+G(u)=f(x,t),(x,t)Ω×J, (4.2)

    with boundary condition

    u(x,t)=0,xΩ,tˉJ, (4.3)

    and initial condition

    u(x,0)=u0(x),xˉΩ, (4.4)

    where the nonlinear term G(u)=u2. The symbol Δ is the Laplace operator. Based on Eqs (3.14)–(3.21), the loss function MSEtotal can be obtained by

    MSEtotal=MSE+MSEb,

    where

    MSE=1N×NxNj=1Nxi=1|j(xi)|2,MSEb=1N×NbNj=1Nbi=1|vj(xib;δ)0|2.

    Example 1.

    For this example, we set space domain Ω=(0,1) and time interval J=(0,12]. The training set consists of the boundary points and Nx=200 collocation points randomly selected in the space domain Ω. Choosing ω(α)=Γ(3α) and the source term

    f(x,t)=2tsin(2πx)+Γ(3)t(t1)lntsin(2πx)+4t2π2sin(2πx)+8tπ2sin(2πx)+(t2sin(2πx))2,

    then we can obtain the exact solution u(x,t)=t2sin(2πx).

    To evaluate the performance of our proposed method, the exact solution and the predicted solution solved by a multi-output neural network that consists of 6 hidden layers with 40 neurons in each hidden layer are showed in Figure 2.

    Figure 2.  Example 1: the exact solution and predicted solution with Δα=1500, θ=1 and N=20 at t=0.5.

    The influence of different network structures on our proposed method to solve Example 1 is presented in Table 2. The accuracy of the predicted solution fluctuates with different network architectures, and shows a fluctuating growth with the increase of the number of hidden layers. Based on the three network architectures, Figure 3 shows the behavior performance of the proposed method with gradual decrease of the time step size. With the increase of the number of grid points in the time interval, we observe that the behavior of relative L2 error generally presents an upward trend for fixed network architecture and expanding the depth of the hidden layer can effectively improve the accuracy of the predicted solution.

    Table 2.  The relative L2 error between the predicted solution with parameters N=20, Δα=1500, θ=1 and exact solution for different numbers of hidden layers and different numbers of neurons per layer.
    20 30 40 50 60
    2 3.583482e-02 7.850782e-02 5.238403e-02 1.143600e-01 3.761048e-02
    4 3.815002e-02 5.144618e-02 3.622479e-02 2.317038e-02 3.721057e-02
    6 2.433509e-02 1.216386e-01 2.638224e-02 2.837553e-02 2.682958e-02

     | Show Table
    DownLoad: CSV
    Figure 3.  Example 1: the variation trend of relative L2 error between the predicted solution with θ=1, Δα=1500 and the exact solution.

    Numerical results calculated by the single output neural network and multi-output neural network schemes are presented in Table 3, where we select 200 collocation points in the given spatial domain by random sampling method and set N=10, θ=1 and Δα=1500. It is easy to find that the accuracy of the predicted solution calculated by the multi-output neural network scheme is higher than that solved by the single output neural network scheme and replacing single output neural network with multi-output neural network can save a lot of computing time.

    Table 3.  The relative L2 error and computing time given by the single output and multi-output neural network schemes.
    Neural network Layers Neurons Relative L2 error CPU time (s)
    multi-output 4 20 1.948281e-02 32.80
    40 1.858120e-02 44.45
    6 20 9.724722e-03 51.40
    40 1.024869e-02 65.88
    single output 4 20 8.741336e-02 661.99
    40 2.702814e-02 742.17
    6 20 2.937448e-02 748.31
    40 2.278774e-02 840.09

     | Show Table
    DownLoad: CSV

    Example 2.

    In this numerical example, considering the space domain Ω=(0,1)×(0,1), the time interval J=(0,12], ω(α)=Γ(3α) and the exact solution u(x,y,t)=t2sin(2πx)sin(2πy), the source term can be given by

    f(x,y,t)=2tsin(2πx)sin(2πy)+Γ(3)t(t1)lntsin(2πx)sin(2πy)+8t2π2sin(2πx)sin(2πy)+16tπ2sin(2πx)sin(2πy)+(t2sin(2πx)sin(2πy))2.

    The training dataset is shown in Figure 5 and the collocation and boundary points are selected by random sampling method.

    To better illustrate the behavior of the predicted solution, Figure 4 portrays the contour plot of the exact and predicted solutions, where the training set consists of 961 collocation points and 124 boundary points selected by the equidistant uniform sampling method and the network architecture consists of 12 hidden layers with 60 neurons per layer.

    Figure 4.  Example 2: the contour plot of the exact (a) and predicted (b) solutions with N=10, θ=1 and Δα=1500 at t=0.5.
    Figure 5.  Example 2: distribution of the collocation points (a) sampled in the domain Ω and the boundary training dataset (b).

    Table 4 shows the impact of depth and width of the network on the accuracy of the predicted solution. In Figure 6, we present the behavior of how relative L2 error changes with respect to different grid points N. Combined with Table 4 and Figure 6, we observe that increasing the number of hidden layers or neurons has a positive effect on reducing relative L2 error in general.

    Table 4.  The relative L2 error between the predicted solution with parameters N=20, θ=1, Δα=1500 and the exact solution for different number of hidden layers and neurons per layer.
    20 30 40 50 60
    4 1.155984e-01 1.101123e-01 7.228454e-02 7.787707e-02 4.492646e-02
    6 6.273340e-02 6.088659e-02 9.802474e-02 6.536637e-02 6.245142e-02
    8 6.163051e-02 7.424834e-02 6.844676e-02 5.500003e-02 5.169303e-02

     | Show Table
    DownLoad: CSV
    Figure 6.  Example 2: the variation trend of relative L2 error between the predicted solution with θ=1, Δα=1500 and the exact solution.

    The results shown in Table 5 reveal the performance of the single output neural network and multi-output neural network schemes, where we select 40 boundary points and 200 collocation points in the given spatial domain by random sampling method and set N=10, θ=1 and Δα=1500. One can see that using multi-output neural network effectively improves the precision and reduces the computing time.

    Table 5.  The relative L2 error and computing time given by the single output and multi-output neural network schemes.
    Neural network Layers Neurons Relative L2 error CPU time (s)
    multi-output 4 20 1.080369e-01 52.84
    40 7.472573e-02 66.05
    6 20 6.654115e-02 72.71
    40 7.937601e-02 93.77
    single output 4 20 3.443264e-01 731.22
    40 7.054495e-01 728.11
    6 20 2.636401e-01 772.08
    40 1.410780e-01 892.94

     | Show Table
    DownLoad: CSV

    Further, we consider the following distributed-order fourth-order sub-diffusion model:

    ut+DωtuΔu+Δ2u+G(u)=f(x,t),(x,t)Ω×J, (4.5)

    with boundary condition

    u(x,t)=Δu(x,t)=0,xΩ,tˉJ,

    and initial condition

    u(x,0)=u0(x),xˉΩ,

    where the nonlinear term G(x)=u2.

    Similarly, the corresponding loss function MSEtotal can be calculated by

    MSEtotal=MSE+MSEb,

    where

    MSE=1N×NxNj=1Nxi=1|j(xi)|2,MSEb=1N×NbNj=1Nbi=1|vj(xib;δ)0|2+1N×NbNj=1Nbi=1|Δvj(xib;δ)0|2.

    Example 3.

    Here, we define the space-time domain Ω×J=(0,1)×(0,12]. Considering ω(α)=Γ(3α) and the source term

    f(x,t)=2tsin(πx)+Γ(3)t(t1)lntsin(πx)+t2π2sin(πx)+t2π4sin(πx)+(t2sin(πx))2, (4.6)

    the exact solution can be given by u(x,t)=t2sin(πx). Similar to Example 1, we also randomly sample 200 collocation points in the space domain Ω.

    In order to conveniently observe the capability of our proposed method, Figure 7 shows the change in the trajectory of the predicted and exact solutions with respect to the space point x, where the parameters are set as N=20, Δα=1500, θ=1 and the network consists of 6 hidden layers with 50 neurons per layer. Figure 8 portrays the trajectory of relative L2 error with three different network architectures. Table 6 shows the impact of expanding depth or width of the network on the accuracy of the predicted solutions. Based on Figure 8 and Table 6, it is easy to observe that increasing the number of hidden layer plays a positive role in improving the accuracy of the predicted solutions.

    Figure 7.  Example 3: the predicted and exact solutions at t=0.5.
    Figure 8.  Example 3: the relative L2 error between the predicted solution with θ=1, Δα=1500 and exact solution for different number of grid points.
    Table 6.  The relative L2 error between the predicted solution with parameters N=20, θ=1, Δα=1500 and the exact solution for different number of hidden layers and neurons per layer.
    20 30 40 50 60
    2 8.994465e-03 8.835675e-03 7.976196e-03 1.121301e-02 1.284193e-02
    4 6.698531e-03 6.623791e-03 1.012124e-02 7.129064e-03 6.760145e-03
    6 5.238653e-03 3.178300e-03 4.705057e-03 7.221552e-03 5.002096e-03

     | Show Table
    DownLoad: CSV

    The relative L2 error and CPU time obtained by the multi-output neural network and single output neural network schemes are presented in Table 7, where we select 200 collocation points in the given spatial domain by random sampling method and set N=10, θ=1 and Δα=1500. The error of the proposed multi-output neural network scheme is smaller than that of the single output neural network scheme. For this 1D system, the multi-output neural network scheme is more efficient than the single output neural network scheme.

    Table 7.  The relative L2 error and computing time given by the single output and multi-output neural network schemes.
    Neural network Layers Neurons Relative L2 error CPU time (s)
    multi-output 4 20 8.027377e-03 334.07
    40 8.656712e-03 366.52
    6 20 8.315435e-03 557.78
    40 6.788274e-03 601.19
    single output 4 20 2.412753e-02 994.62
    40 2.365781e-02 1317.16
    6 20 2.428662e-02 1344.76
    40 2.139562e-02 1795.75

     | Show Table
    DownLoad: CSV

    Example 4.

    Now we take space domain Ω=(0,1)×(0,1) and time interval J=(0,12]. Let ω(α)=Γ(3α) and the exact solution u(x,y,t)=t2sin(πx)sin(πy). Then we arrive at the source term

    f(x,y,t)=2tsin(πx)sin(πy)+Γ(3)t(t1)lntsin(πx)sin(πy)+2t2π2sin(πx)sin(πy)+4t2π4sin(πx)sin(πy)+(t2sin(πx)sin(πy))2. (4.7)

    Here, we apply the training data set shown in Figure 5.

    In order to more intuitively demonstrate the feasibility of our proposed method for solving this 2D system, Figure 9 shows the contour plot of u and |uv|, where the training set consists of 900 collocation points and 120 boundary points selected by the equidistant uniform sampling method and the network is composed of 6 hidden layers with 20 neurons per layer.

    Figure 9.  Example 4: the contour plot of u and |uv| with θ=1, Δα=1500 and N=40 at t=0.5.

    From the behavior of relative L2 error in Figure 10, one can see that the accuracy of the predicted solutions with the fixed network first presents a increasing trend and then gradually presents a downward trend. This is because the approximation ability of neural network reaches saturation with the increase of grid points N. Table 8 shows the relative L2 error calculated by different network architectures. On the whole, the relative L2 error slightly decreases with expanding depth of the network, while it first slightly decreases and then increases with expanding width of the network. To show the precision and efficiency of the multi-output neural network scheme for this 2D system, the relative L2 error and CPU time obtained by the multi-output neural network and single output neural network schemes are shown in Table 9, where we select 40 boundary points and 200 collocation points in the given spatial domain by random sampling method and set N=10, θ=1 and Δα=1500. It illustrates that the multi-output neural network scheme is more accurate and efficient than the single output neural network scheme.

    Figure 10.  Example 4: the relative L2 error between the predicted solution with θ=1, Δα=1500 and the exact solution for different number of grid points.
    Table 8.  The relative L2 error between the predicted solution with parameters N=40, θ=1, Δα=1500 and the exact solution for different number of hidden layers and neurons per layer.
    } 20 30 40 50 60
    2 7.593424e-02 4.556132e-02 2.926755e-02 5.850593e-02 4.956921e-02
    4 3.716823e-02 2.498269e-02 3.388639e-02 4.073820e-02 5.282780e-02
    6 2.690082e-02 3.039336e-02 2.769315e-02 4.331023e-02 6.456580e-02

     | Show Table
    DownLoad: CSV
    Table 9.  The relative L2 error and computing time given by the single output and multi-output neural network schemes.
    Neural network Layers Neurons Relative L2 error CPU time (s)
    multi-output 4 20 4.549939e-02 345.03
    40 5.847813e-02 339.46
    6 20 5.680785e-02 466.02
    40 6.110534e-02 528.10
    single output 4 20 1.455785e-01 902.99
    40 1.330922e-01 1309.05
    6 20 1.496581e-01 1469.85
    40 1.322566e-01 1990.96

     | Show Table
    DownLoad: CSV

    In this article, a multi-output physics informed neural network combined with the Crank-Nicolson scheme including the FBN-θ method and the composite numerical integral formula was constructed to solve 1D and 2D nonlinear time distributed-order models. The calculation process is described in detail. Numerical experiments are provided to prove the effectiveness and feasibility of our algorithm. Compared with the results calculated by a single output neural network combined with the FBN-θ method and the Crank-Nicolson scheme, one can clearly see that the proposed multi-output neural network scheme is more efficient and accurate. Moreover, some numerical methods, such as finite difference or finite element method, need to linearize the nonlinear term, which will give rise to extra costs. The process of linearization can be directly omitted by PINN. Further work will investigate the application of the proposed methodology in high-dimensional problems and practical problems [36,37,38,39,40].

    The authors declare that they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors would like to thank the editor and all the anonymous referees for their valuable comments, which greatly improved the presentation of the article. This work is supported by the National Natural Science Foundation of China (12061053, 12161063), Natural Science Foundation of Inner Mongolia (2021MS01018), Young innovative talents project of Grassland Talents Project, Program for Innovative Research Team in Universities of Inner Mongolia Autonomous Region (NMGIRT2413, NMGIRT2207), and 2023 Postgraduate Research Innovation Project of Inner Mongolia (S20231026Z).

    The authors declare that they have no conflict of interest.



    [1] Faver TE, Wright JD (2018) Exact diatomic Fermi-Pasta-Ulam-Tsingou solitary waves with optical band ripples at infinity. SIAM J Math Anal 50: 182–250. doi: 10.1137/15M1046836
    [2] Lustri CJ, Porter MA (2018) Nanoptera in a period-2 Toda chain. SIAM J Appl Dyn Syst 17: 1182–1212. doi: 10.1137/16M108639X
    [3] Porter M, Daraio C, Szelengowicz I, et al. (2009) Highly nonlinear solitary waves in heterogeneous periodic granular media. Phys D 238: 666–676. doi: 10.1016/j.physd.2008.12.010
    [4] Gaison J, Moskow S, Wright JD, et al. (2014) Approximation of polyatomic FPU lattices by KdV equations. Multiscale Model Simul 12: 953–995. doi: 10.1137/130941638
    [5] Qin WX (2015) Wave propagation in diatomic lattices. SIAM J Math Anal 47: 477–497. doi: 10.1137/130949609
    [6] Betti M, Pelinovsky DE (2013) Periodic traveling waves in diatomic granular chains. J Nonlinear Sci 23: 689–730. doi: 10.1007/s00332-013-9165-6
    [7] Chirilus-Bruckner M, Chong C, Prill O, et al. (2012) Rigorous description of macroscopic wave packets in infinite periodic chains of coupled oscillators by modulation equations. Discrete Contin Dyn Syst Ser S 5: 879–901. Available from: https://doi.org/10.3934/dcdss.2012.5.879. doi: 10.3934/dcdss.2012.5.879
    [8] Brillouin L (1953) Wave Propagation in Periodic Structures. Electric Filters and Crystal Lattices,2Eds., New York: Dover Publications, Inc.
    [9] Tabata Y (1996) Stable solitary wave in diatomic Toda lattice. J Phys Soc Jpn 65: 3689–3691. doi: 10.1143/JPSJ.65.3689
    [10] Okada Y, Watanabe S, Tanaca H (1990) Solitary wave in periodic nonlinear lattice. J Phys Soc Jpn 59: 2647–2658. Available from: https://doi.org/10.1143/JPSJ.59.2647. doi: 10.1143/JPSJ.59.2647
    [11] Hoffman A, Wright JD (2017) Nanopteron solutions of diatomic Fermi-Pasta-Ulam-Tsingou lattices with small mass-ratio. Phys D 358: 33–59. doi: 10.1016/j.physd.2017.07.004
    [12] Vainchtein A, Starosvetsky Y, Wright JD, et al. (2016) Solitary waves in diatomic chains. Phys Rev E 93: 042210. doi: 10.1103/PhysRevE.93.042210
    [13] Schneider G, Wayne CE (1999) Counter-propagating waves on fluid surfaces and the continuum limit of the Fermi-Pasta-Ulam model, In: International Conference on Differential Equations,Vol. 1, 2 (Berlin, 1999), 390–404, World Sci. Publ., River Edge, NJ, 2000.
    [14] Friesecke G, Pego RL (1999) Solitary waves on FPU lattices: I. Qualitative properties, renormalization and continuum limit. Nonlinearity 12: 1601–1627.
    [15] Friesecke G, Pego RL (2002) Solitary waves on FPU lattices: II. Linear implies nonlinear stability. Nonlinearity 15: 1343–1359.
    [16] Friesecke G, Pego RL (2004) Solitary waves on Fermi-Pasta-Ulam lattices: III. Howland-type Floquet theory. Nonlinearity 17: 207–227.
    [17] Friesecke G, Pego RL (2004) Solitary waves on Fermi-Pasta-Ulam lattices: IV. Proof of stability at low energy. Nonlinearity 17: 229–251.
    [18] Mizumachi T (2009) Asymptotic stability of lattice solitons in the energy space. Commun Math Phys 288: 125–144. doi: 10.1007/s00220-009-0768-6
    [19] Boyd JP (1998) Weakly Nonlocal Solitary Waves and Beyond-All-Orders Asymptotics Generalized Solitons and Hyperasymptotic Perturbation Theory , In Series: Mathematics and its Applications. Dordrecht: Kluwer Academic Publishers, vol. 442. Available from: https://doi.org/10.1007/978-1-4615-5825-5.
    [20] Faver T (2017) Nanopteron-stegoton traveling waves in spring dimer Fermi-Pasta-Ulam-Tsingou lattices, in press. Available from: https://arxiv.org/abs/1710.07376.
    [21] Lombardi E (2000) Oscillatory Integrals and Phenomena Beyond All Algebraic Orders with Applications to Homoclinic Orbits in Reversible Systems, In series: Lecture Notes in Mathematics. Berlin: Springer-Verlag, vol. 1741. Available from: https://doi.org/10.1007/BFb0104102.
    [22] Sun SM (1999) Non-existence of truly solitary waves in water with small surface tension. Proc Math Phys Eng Sci 455: 2191–2228. doi: 10.1098/rspa.1999.0399
    [23] Martínez AJ, Kevrekidis PG, Porter MA (2016) Superdiffusive transport and energy localization in disordered granular crystals. Phys Rev E 93: 022902. Available from: https://doi.org/10.1103/physreve.93.022902. doi: 10.1103/PhysRevE.93.022902
    [24] Hairer E, Lubich C, Wanner G (2006) Geometric Numerical Integration: Structure-Preserving Algorithms for Ordinary Differential Equations, 2Eds., In series: Springer Series in Computational Mathematics. Springer, Heidelberg, 2010, vol. 31.
    [25] Beyn WJ, Thümmler V (2004) Freezing solutions of equivariant evolution equations. SIAM J Appl Dyn Syst 3: 85–116. doi: 10.1137/030600515
    [26] Beyn WJ, Otten D, Rottmann-Matthes J (2018) Computation and stability of traveling waves in second order evolution equations. SIAM J Numer Anal 56: 1786–1817. doi: 10.1137/16M108286X
    [27] Beale JT (1991) Exact solitary water waves with capillary ripples at infinity. Commun Pure Appl Math 44: 211–257. doi: 10.1002/cpa.3160440204
    [28] Sun SM (1991) Existence of a generalized solitary wave solution for water with positive Bond number less than 1/3. J Math Anal Appl 156: 471–504. doi: 10.1016/0022-247X(91)90410-2
    [29] LeVeque RJ, Yong DH (2003) Solitary waves in layered nonlinear media. SIAM J Appl Math 63: 1539–1560. doi: 10.1137/S0036139902408151
    [30] LeVeque RJ, Yong DH (2003) Phase plane behavior of solitary waves in nonlinear layered media, In: Hyperbolic Problems: Theory, Numerics, Applications. Berlin: Springer, 43–51.
    [31] Kevrekidis PG, Stefanov AG, Xu H (2016) Traveling waves for the mass in mass model of granular chains. Lett Math Phys 106: 1067–1088. doi: 10.1007/s11005-016-0854-6
    [32] Pnevmatikos S, Flytzanis N, Remoissenet M (1986) Soliton dynamics of nonlinear diatomic lattices. Phys Rev B 33: 2308–2321. doi: 10.1103/PhysRevB.33.2308
  • This article has been cited by:

    1. Laura Lomba, David Lapeña, Natalia Ros, Elena Aso, Mariachiara Cannavò, Diego Errazquin, Beatriz Giner, Ecotoxicological study of six drugs in Aliivibrio fischeri, Daphnia magna and Raphidocelis subcapitata, 2020, 27, 0944-1344, 9891, 10.1007/s11356-019-07592-8
    2. Laura Lomba, M. Pilar Ribate, Estefanía Zuriaga, Cristina B. García, Beatriz Giner, Acute and subacute effects of drugs in embryos of Danio rerio. QSAR grouping and modelling, 2019, 172, 01476513, 232, 10.1016/j.ecoenv.2019.01.081
    3. Carlos Castillo-Zacarías, Mario E. Barocio, Enrique Hidalgo-Vázquez, Juan Eduardo Sosa-Hernández, Lizeth Parra-Arroyo, Itzel Y. López-Pacheco, Damià Barceló, Hafiz N.M. Iqbal, Roberto Parra-Saldívar, Antidepressant drugs as emerging contaminants: Occurrence in urban and non-urban waters and analytical methods for their detection, 2021, 757, 00489697, 143722, 10.1016/j.scitotenv.2020.143722
    4. Melanie M. Marshall, Kevin E. McCluney, Mixtures of co-occurring chemicals in freshwater systems across the continental US, 2021, 268, 02697491, 115793, 10.1016/j.envpol.2020.115793
    5. Harveer S. Srain, Karen F. Beazley, Tony R. Walker, Pharmaceuticals and personal care products and their sublethal and lethal effects in aquatic organisms, 2021, 29, 1181-8700, 142, 10.1139/er-2020-0054
    6. Kaiwen Hu, Weifeng Li, Weixia Zhang, Kuankuan Yuan, Chenxin Gong, Yang Shu, Yingying Yu, Conghui Shan, Yan Gao, Xunyi Zhang, Haibin Yu, Wei Shi, Guangxu Liu, Diltiazem disrupts Ca2+-homeostasis and exerts immunotoxic effects on a marine bivalve mollusc, the blood clam (Tegillarca granosa), 2025, 217, 0025326X, 118055, 10.1016/j.marpolbul.2025.118055
  • Reader Comments
  • © 2019 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(6735) PDF downloads(1653) Cited by(12)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog