
We propose a physics-informed neural network (PINN)-based method to solve the heat transfer equation under imperfect contact conditions. A major challenge arises from the discontinuity of the solution across the interface, where the exact jump is unknown and implicitly determined by the Kapitza thermal resistance condition. Since the neural network function is smooth on the entire domain, conventional PINN could be inefficient to capture such discontinuities without certain modifications. One remedy is to extend a piecewise continuous function on Rd to a continuous function on Rd+1. This is achieved by applying a Sobolev extension for the solution within each subdomain and introducing, additional coordinate variable that labels the subdomains. This formulation enables the design of neural network functions in the augmented variable, which retains the universal approximation property. We define the PINN in an augmented variable by the minimizer of the loss functional, which includes the implicit interface conditions. Once the loss functional is minimized, the solution obtained by the axis-augmented PINN satisfies the implicit jump conditions. In this way, our method offers a user-friendly way to solve heat transfer equations with imperfect contact conditions. Another advantage of using a continuous representation of solutions in augmented variables is that it allows error analysis in the space of smooth functions. We provide an error analysis of the proposed method, demonstrating that the difference between the exact solution and the predicted solution is bounded by the physics-informed loss functional. Furthermore, the loss functional can be made small by increasing the parameters in the neural network such as the number of nodes in the hidden layers.
Citation: Hansaem Oh, Gwanghyun Jo. Physics-informed neural network for the heat equation under imperfect contact conditions and its error analysis[J]. AIMS Mathematics, 2025, 10(4): 7920-7940. doi: 10.3934/math.2025364
[1] | M. A. El-Shorbagy, Waseem, Mati ur Rahman, Hossam A. Nabwey, Shazia Habib . An artificial neural network analysis of the thermal distribution of a fractional-order radial porous fin influenced by an inclined magnetic field. AIMS Mathematics, 2024, 9(6): 13659-13688. doi: 10.3934/math.2024667 |
[2] | Jihahm Yoo, Haesung Lee . Robust error estimates of PINN in one-dimensional boundary value problems for linear elliptic equations. AIMS Mathematics, 2024, 9(10): 27000-27027. doi: 10.3934/math.20241314 |
[3] | Zhi-Ying Feng, Xiang-Hua Meng, Xiao-Ge Xu . The data-driven localized wave solutions of KdV-type equations via physics-informed neural networks with a priori information. AIMS Mathematics, 2024, 9(11): 33263-33285. doi: 10.3934/math.20241587 |
[4] | Fujun Cao, Dongfang Yuan . A class of HOC finite difference method for elliptic interface problems with imperfect contact. AIMS Mathematics, 2023, 8(3): 5789-5815. doi: 10.3934/math.2023292 |
[5] | Khalil Ur Rehman, Wasfi Shatanawi, Zeeshan Asghar, Haitham M. S. Bahaidarah . Neural networking analysis for MHD mixed convection Casson flow past a multiple surfaces: A numerical solution. AIMS Mathematics, 2023, 8(7): 15805-15823. doi: 10.3934/math.2023807 |
[6] | Hyun Geun Lee, Youngjin Hwang, Yunjae Nam, Sangkwon Kim, Junseok Kim . Benchmark problems for physics-informed neural networks: The Allen–Cahn equation. AIMS Mathematics, 2025, 10(3): 7319-7338. doi: 10.3934/math.2025335 |
[7] | Shanhao Yuan, Yanqin Liu, Yibin Xu, Qiuping Li, Chao Guo, Yanfeng Shen . Gradient-enhanced fractional physics-informed neural networks for solving forward and inverse problems of the multiterm time-fractional Burger-type equation. AIMS Mathematics, 2024, 9(10): 27418-27437. doi: 10.3934/math.20241332 |
[8] | Kottakkaran Sooppy Nisar, Muhammad Shoaib, Muhammad Asif Zahoor Raja, Yasmin Tariq, Ayesha Rafiq, Ahmed Morsy . Design of neural networks for second-order velocity slip of nanofluid flow in the presence of activation energy. AIMS Mathematics, 2023, 8(3): 6255-6277. doi: 10.3934/math.2023316 |
[9] | Asifa, Poom Kumam, Talha Anwar, Zahir Shah, Wiboonsak Watthayu . Analysis and modeling of fractional electro-osmotic ramped flow of chemically reactive and heat absorptive/generative Walters'B fluid with ramped heat and mass transfer rates. AIMS Mathematics, 2021, 6(6): 5942-5976. doi: 10.3934/math.2021352 |
[10] | Khalil Ur Rehman, Wasfi Shatanawi, Zead Mustafa . Artificial intelligence (AI) based neural networks for a magnetized surface subject to tangent hyperbolic fluid flow with multiple slip boundary conditions. AIMS Mathematics, 2024, 9(2): 4707-4728. doi: 10.3934/math.2024227 |
We propose a physics-informed neural network (PINN)-based method to solve the heat transfer equation under imperfect contact conditions. A major challenge arises from the discontinuity of the solution across the interface, where the exact jump is unknown and implicitly determined by the Kapitza thermal resistance condition. Since the neural network function is smooth on the entire domain, conventional PINN could be inefficient to capture such discontinuities without certain modifications. One remedy is to extend a piecewise continuous function on Rd to a continuous function on Rd+1. This is achieved by applying a Sobolev extension for the solution within each subdomain and introducing, additional coordinate variable that labels the subdomains. This formulation enables the design of neural network functions in the augmented variable, which retains the universal approximation property. We define the PINN in an augmented variable by the minimizer of the loss functional, which includes the implicit interface conditions. Once the loss functional is minimized, the solution obtained by the axis-augmented PINN satisfies the implicit jump conditions. In this way, our method offers a user-friendly way to solve heat transfer equations with imperfect contact conditions. Another advantage of using a continuous representation of solutions in augmented variables is that it allows error analysis in the space of smooth functions. We provide an error analysis of the proposed method, demonstrating that the difference between the exact solution and the predicted solution is bounded by the physics-informed loss functional. Furthermore, the loss functional can be made small by increasing the parameters in the neural network such as the number of nodes in the hidden layers.
Composite media often contain imperfect contacts at the material interfaces, resulting in discontinuous temperatures across the interface. Since the heat transfer and energy dissipation through imperfectly matched interfaces can significantly influence the temperature profile in internal domains [1,2,3,4], neglecting the effects of thermal resistance occurring from imperfect contact in modeling can deteriorate the predictions of thermal behavior in the domain of interest. Kapitza's thermal resistance is one of the widely adopted heat transfer models in material science [5,6,7], where the amount of jump in the temperature variable along the interface is proportional to the normal heat flux.
One of the difficulties in predicting the thermal profile in imperfectly contacted materials is that the temperature jump across interfaces is implicitly given. In [8,9], the jump is treated as independent variables; the jump profiles are updated in an iterative way, where in each iteration the elliptic interface problem is solved with the approximate jump. On the other hand, some people applied finite element method (FEM)-based algorithms, where interface conditions are implicitly contained in bilinear forms [10,11]. In addition, there are extended approaches based on the finite element method (XFEM) to solve the imperfect contact problem where additional degrees of freedom appear near the interface to fit implicit jump conditions [7,12,13,14]. However, the numerical methods discussed above are inherently complex to implement, as they require additional degrees of freedom in the case of XFEM [13,14] or involve intricate modifications of the bilinear form [10,11] or demand iterative updates for the jump profile [8,9]. Therefore, there arises a need for convenient and efficient methods for the imperfect contact heat transfer equation.
Recently, neural network-based methods have emerged as successful alternatives for solving partial differential equations (PDEs) in the scientific computing community. In the physics-informed neural network (PINN) method, neural network functions serve as a surrogate model for solving PDEs, with its parameters optimized by minimizing objective functions that include the strong form of the governing equations. For an excellent review of PINNs, we refer to [15]. With the advancement of highly efficient automatic differentiation techniques such as those provided by PyTorch [16], PINNs have been successfully applied to a wide range of problems, including Peridynamics [17], thermal-fluid dynamics [18,19], electromagnetism [20], and Poisson-Boltzmann [22].
In this work, we develop a new PINN-based method for the heat equation under the imperfect contact condition. One of the restrictions of conventional PINN in solving PDE lies in the fact that neural networks belong to a continuous family of functions. Therefore, when PDE contains some interface with possibly discontinuous solutions along the interface, it is hard to expect an accurate reconstruction of solutions by PINN. One remedy is to extend the piecewise continuous function to a continuous function of higher dimension. In [21,22], the axis augmentation techniques are proposed, where an additional variable labels the locations of subdomains. Following the way of [21,22], we first employ continuous Sobolev extensions for the piecewise H1-temperature variable. Then, together with the auxiliary variable that labels the locations of subregions, we develop a continuous representation of the heat variable in an augmented variable. Finally, we define the physics-informed neural network in an augmented variable by the minimizer of the loss functional, which includes the implicit jump condition across the interface. In this way, our method offers a user-friendly way to solve heat transfer equations with imperfect contact conditions, without the need for mesh generation or iterative methods to determine jump profiles across the interface.
Aside from the ease of implementation, another advantage of using a continuous representation of solutions in augmented variables is that it allows error analysis in the space of smooth functions. We provide an error analysis of the proposed method. Our main theorem (see Theorem 3.3) states that the difference between the exact solution and the neural network approximation in an energy-like norm is bounded by the physics-informed loss functional. Here, the loss functional can be made arbitrarily small by increasing the parameters in the neural network, such as the number of nodes and number of training samples (see Theorem 3.6). The numerical tests in the result section support the error analysis. To the best of the authors' knowledge, the proposed method is the first successful application of PINN for solving interface problems with implicit jump conditions, accompanied by an error analysis. Let us summarize the advantages of the proposed work: 1) by establishing the continuous representation of solutions in the augmented variable, our method benefits from the universal approximation property of the neural networks; 2) since our methods are formulated on smooth function space, we were able to derive the error estimates for the proposed PINN; 3) our method does not require mesh generation or linearization procedures, in contrast to the conventional FEM-based method.
The rest of the paper is organized as follows. In the next section, we describe the governing equation and develop our version of PINN. The error estimates are carried out in Section 3. In Section 4, we document the performance of the proposed algorithm. The conclusion follows in Section 5.
In this section, we develop a PINN method for a heat equation involving imperfect contact. We employ continuous Sobolev extensions for the temperature variable to represent the solution as a continuous function in an augmented variable. The model equation is described in Subsection 2.1, and our version of the PINN method is proposed in Subsection 2.2.
We introduce some definitions. Suppose G is any domain in Rd. For integer d≥1, s≥0, and 1≤p≤∞, we denote by Ws,p(G) the usual Sobolev space associated with norm ‖⋅‖Ws,p(G) and seminorm |⋅|Ws,p(G). In particular, if p=2, we then write Ws,2(G)=Hs(G), with ‖⋅‖Ws,2(G)=‖⋅‖s,G and |⋅|Ws,2(G)=|⋅|s,G. We also write W0,p(G)=Lp(G) and ‖⋅‖W0,p(G)=‖⋅‖Lp(G). For the inner product on G, we use the notation (⋅,⋅)G. Given an integer k≥0, let Ck(G) denote the space of all Ck functions in the closure of G. Given a real-valued function ψ, its support is denoted by supp(ψ). We denote by χG the indicator function of G.
We consider a (bounded) composite material domain Ω⊂Rd (d=2,3) divided by material interface Γ, i.e., Ω=Ω+∪Ω−∪Γ (see Figure 1). The heat flux variable is defined as q=−k∇T, where T and k are the temperature variable and thermal conductivity constant, respectively. In an imperfect contact situation, the temperature jump across the material interface is proportional to the heat flux in the normal direction of the interface:
T|Ω+−T|Ω−=−αq|Ω−⋅nΓonΓ, |
where α the is Kapitza constant [5,6,7]. Now, let us summarize the steady-state heat equation on Ω under the imperfect condition as follows:
2−div(k∇T)=fin Ω+∪Ω−, | (2.1a) |
[T]Γ=αk−∇T|Ω−⋅nΓon Γ, | (2.1b) |
[k∇T⋅nΓ]Γ=0on Γ, | (2.1c) |
T=gon ∂Ω, | (2.1d) |
where f is the source term and nΓ=nΩ−. Also, [⋅]Γ denotes a jump operator across the interface, i.e.,
[T]Γ:=T|Ω+−T|Ω−,[k∇T⋅nΓ]Γ=(k∇T)|Ω+⋅nΓ−(k∇T)|Ω−⋅nΓ. |
Here, we assume that k is piecewise constant, that is, k=k+χΩ++k−χΩ− for some positive constants k+ and k−. The problem (2.1) has a unique solution [23,24].
In this subsection, we describe the PINN method for the governing equation. The overall process is illustrated in Figure 2.
One challenge is that the solution and its derivative are discontinuous across Γ, whereas conventional neural network functions are inherently smooth on the entire domain. As a result, traditional neural network functions may be inefficient in capturing the discontinuities in the governing equation. To remedy this, we adopt the axis augmentation technique introduced in [21]. The key idea is to introduce an augmented variable z to extend the discontinuous function T to an augmented function Taug that is continuous over Ω×[−1,1]. Let us assume that T|Ω− and T|Ω+ are sufficiently smooth. We apply the Sobolev extension theorem (see, e.g., [25,26]), which guarantees the existence of smooth extensions Ts, (s=±), satisfying
Ts=TonΩs,s=±. |
With the additional variable z, we define axis-augmented function
Taug(x,z):=1+z2T+(x)+1−z2T−(x),∀(x,z)∈Ω×[−1,1]. | (2.2) |
Then, it follows that
Taug(x,z=1)=T(x),x∈Ω+,Taug(x,z=−1)=T(x),x∈Ω−, |
In this way, we have derived a continuous function Taug in an augmented variable, where the variable z labels the locations of subregions, i.e., z=1 for Ω+ and z=−1 for Ω−.
The advantage of introducing such a continuous representation is that it enables the design of neural network surrogate functions in the augmented variable space, which retains the universal approximation property [27]. The neural network functions in the augmented variable are defined as follows: Given integer n≥1, we introduce the set Naugn of all augmented neural network functions ϕaugθ having n hidden layers:
Naugn:={ϕaugθ:ϕaugθ(x,z)=(Ln∘σ∘⋯∘σ∘L0)(x,z)}, | (2.3) |
where Li's are linear affine functions and σ is a tanh activation function. Here, θ denotes the collection of all training parameters. The advantage of defining such an augmented neural network function is that, by restricting z to subdomains, we can obtain the piecewise continuous function, i.e.,
ϕθ(x)=ϕaugθ(x,1)χΩ+(x)+ϕaugθ(x,−1)χΩ−(x),x∈Ω, | (2.4) |
which can efficiently approximate the desired piecewise continuous solution T. It remains to impose loss functionals for the functions in
N±n:={ϕθ:ϕθ satisfies (2.4) for some ϕaugθ∈Naugn}. |
Since the residual of the governing equation can be evaluated on discrete points, we need to define the collection of training points in the variable (x,z). Let {(x+i,1)}M+i=1 and {(x−i,−1)}M−i=1 be the collections of training points in Ω+×{+1} and Ω−×{−1}, respectively. Let {xΓi}MΓi=1 be the collection of training points on Γ. We also let {x∂i}M∂i=1 be the collection of points on ∂Ω. Some choices of the training points can be found in [21,22]. Motivated by the governing Eqs. (2.1a)–(2.1d), the loss functionals are defined as follows: For any piecewise smooth function ϕθ∈N±n,
LΩ(ϕθ):=1M+M+∑i=1|k+Δϕθ(x+i)+f(xi)|2+1M−M−∑i=1|k−Δϕθ(x−i)+f(xi)|2, | (2.5) |
LΓ,1(ϕθ):=1MΓMΓ∑i=1|[ϕθ]Γ(xΓi)−αk−∇ϕθ|Ω−(xΓi)⋅nΓ|2, | (2.6) |
LΓ,2(ϕθ):=1MΓMΓ∑i=1|[k∇ϕθ⋅nΓ]Γ(xΓi)|2, | (2.7) |
L∂(ϕθ):=1M∂M∂∑i=1|ϕθ(x∂i)−g(x∂i)|2, | (2.8) |
L(ϕθ):=LΩ(ϕθ)+LΓ,1(ϕθ)+LΓ,2(ϕθ)+L∂(ϕθ). | (2.9) |
The detailed calculations of loss functionals are provided in Appendix A. Finally, our version of the PINN method for (2.1) is formulated as follows: Find Tθ∈N±n such that
Tθ=argminϕθ∈N±nL(ϕθ). | (2.10) |
Once Tθ is obtained, the function inherently satisfies the implicit jump conditions (2.1b)–(2.1c). Consequently, we introduce an efficient and user-friendly method to solve the imperfect contact problem without requiring mesh generation or iterative procedures to determine the jump profile.
In this section, we derive the error estimates of the proposed PINN method. We define the residuals RΩ,RΓ, and R∂ as follows: Given a real-valued function ϕ on Ω,
RΩ(ϕ):=‖k+Δϕ+f‖20,Ω++‖k−Δϕ+f‖20,Ω−,RΓ(ϕ):=‖[ϕ]Γ−αk−∇ϕ|Ω−⋅nΓ‖20,Γ+‖[k∇ϕ⋅n]Γ‖20,Γ,R∂(ϕ):=‖ϕ−g‖20,∂Ω. |
Also, let R:=RΩ+RΓ+R∂. Then the main result in this section is summarized as follows.
● (Theorem 3.3) If T is the solution of (2.1) which is sufficiently smooth in Ω+ and Ω−, then for any Tθ∈N±n, the piecewise H1-error between T and Tθ can be bounded above by the residual R(Tθ).
● (Theorem 3.6) If T is sufficiently smooth in Ω+ and Ω−, then the residual can be made arbitrarily small. For any ε>0 there exists Tθ∈N±n such that R(Tθ)<ε.
Note that, for a sufficiently large number of training points in the definition of the loss functional L, the errors between the loss functional L and the residual R can be made arbitrarily small (see, e.g., [28,29,30,31]). Therefore, we assume that the following holds.
Assumption 3.1. It holds that Ls(ϕ)≈Rs(ϕ) for s=Ω,Γ,∂.
This assumption, together with the results in Theorem 3.3 and Theorem 3.6, implies that the error between the exact solution and the neural network solution can be made arbitrarily small by controlling the parameters in the neural network.
Finally, we introduce some broken function spaces
Hk(Ω±)={v∈L2(Ω):v|Ωs∈Hk(Ωs), s=±},k≥1,C1(Ω±)={v∈L2(Ω):v|Ωs∈C1(Ωs), s=±}, |
and associated norms (respectively)
‖v‖2k,Ω±=∑s=+,−‖v‖2k,Ωs,v∈Hk(Ω±),‖v‖C1(Ω±)=‖v|Ω+‖C1(Ω+)+‖v|Ω−‖C1(Ω−),v∈C1(Ω±). |
To prove Theorem 3.3, we need the Poincaŕe inequality for the piecewise H1-functions which can be found in [22, Lemma 1].
Lemma 3.2. It holds that
‖v‖0,Ω≤C(|v|1,Ω±+‖[v]Γ‖0,Γ+‖v‖0,∂Ω),∀v∈H1(Ω±), |
where C is a positive constant depending only on Ω and Γ.
Theorem 3.3. Suppose that the solution T of (2.1) satisfies T∈H2(Ω±)∩C1(Ω±). Then we have
‖T−Tθ‖21,Ω±+‖[T−Tθ]Γ‖20,Γ≤C(R(Tθ)+(RΓ(Tθ))12+(R∂(Tθ))12),∀Tθ∈N±n, |
where C is a positive constant depending only on Ω, Γ, k, α, ‖T‖C1(Ω±) and ‖Tθ‖C1(Ω±).
Proof. Let v=T−Tθ. By Lemma 3.2 and by the fact that k−,k+, and α are positive constants, we have
‖v‖21,Ω±+‖[v]Γ‖20,Γ≤C0(|v|21,Ω±+‖[v]Γ‖20,Γ+‖v‖20,∂Ω)≤˜C(|(k+)1/2∇v|21,Ω++|(k−)1/2∇v|21,Ω−+‖α−1/2[v]Γ‖20,Γ+‖v‖20,∂Ω), | (3.1) |
where ˜C=C0max{(k−)−1,(k+)−1,α−1}. By the boundary condition (2.1d), we have
‖v‖20,∂Ω=‖Tθ−g‖20,∂Ω=R(Tθ). |
Using integration by parts and the fact that −kΔT=f on each Ω−∪Ω+, we have
|(k+)1/2∇v|21,Ω++|(k−)1/2∇v|21,Ω−+‖α−1/2[v]Γ‖20,Γ=(k+∇(T−Tθ),∇v)0,Ω++(k−∇(T−Tθ),∇v)0,Ω−+(α−1[T−Tθ]Γ,[v]Γ)0,Γ=(k+ΔTθ+f,v)0,Ω++(k−ΔTθ+f,v)0,Ω−+(k∇(T−Tθ)⋅nΩ,T−Tθ)0,∂Ω+(k+∇(T−Tθ)|Ω+⋅nΩ+,v|Ω+)0,Γ+(k−∇(T−Tθ)|Ω−⋅nΩ−,v|Ω−)0,Γ+(α−1[T−Tθ]Γ,[v]Γ)0,Γ | (3.2) |
We remark that nΓ=nΩ−=−nΩ+. Using the identity that
a+b+−a−b−=(a++a−2)(b+−b−)+(a+−a−)(b++b−2), |
we can rewrite the last three terms of (3.2) as
−(k+∇(T−Tθ)|Ω+⋅nΓ,v|Ω+)0,Γ+(k−∇(T−Tθ)|Ω−⋅nΓ,v|Ω−)0,Γ+(α−1[T−Tθ]Γ,[v]Γ)0,Γ=−({k∇(T−Tθ)⋅nΓ}Γ,[v]Γ)0,Γ−([k∇(T−Tθ)⋅nΓ]Γ,{v}Γ)0,Γ+α−1[T−Tθ]Γ,[v]Γ)0,Γ=−({k∇(T−Tθ)⋅nΓ}Γ−α−1[T−Tθ]Γ,[v]Γ)0,Γ−([k∇(T−Tθ)⋅nΓ]Γ,{v}Γ)0,Γ. | (3.3) |
Here, by the interface condition (2.1b)–(2.1c), we have
{k∇T⋅nΓ}Γ−α−1[T]Γ=0,[k∇T⋅nΓ]Γ=0. |
Therefore, (3.3) becomes
({k∇Tθ⋅nΓ}Γ−α−1[Tθ]Γ,[v]Γ)0,Γ+([k∇Tθ⋅nΓ]Γ,{v}Γ)0,Γ=(k−∇Tθ|Ω−⋅nΓ−α−1[Tθ]Γ,[v]Γ)0,Γ+12([k∇Tθ⋅nΓ]Γ,[v]Γ)0,Γ+([k∇Tθ⋅nΓ]Γ,{v}Γ)0,Γ | (3.4) |
By (3.2)–(3.5), we obtain
|(k+)1/2∇v|21,Ω++|(k−)1/2∇v|21,Ω−+‖α−1/2[v]Γ‖20,Γ | (3.5) |
=(k+ΔTθ+f,v)0,Ω++(k−ΔTθ+f,v)0,Ω−+(k∇(T−Tθ)⋅nΩ,v)0,∂Ω+(k−∇Tθ|Ω−⋅nΓ−α−1[Tθ]Γ,[v]Γ)0,Γ+12([k∇Tθ⋅nΓ]Γ,[v]Γ)0,Γ+([k∇Tθ⋅nΓ]Γ,{v}Γ)0,Γ=:I1+⋯+I6. | (3.6) |
We estimate each of the terms. By the Cauchy-Schwarz and Young's inequalities, we have for I1, I2,
|I1|+|I2|=|(k−ΔTθ+f,v)0,Ω+|+|(k+ΔTθ+f,v)0,Ω−|≤‖k−ΔTθ+f‖0,Ω+‖v‖0,Ω++‖k+ΔTθ+f‖0,Ω−‖v‖0,Ω−≤δ12‖k−ΔTθ+f‖20,Ω++12δ1‖v‖20,Ω++δ22‖k+ΔTθ+f‖20,Ω−+12δ2‖v‖20,Ω−=12max(δ1,δ2)RΩ(Tθ)+(12δ1+12δ2)‖v‖21,Ω±. |
Here, δ1 and δ2 are positive constants related to the Young's inequality, which will be determined later.
Next, let us estimate I3. By the fact that T,Tθ∈C1(Ω±) and T=g on ∂Ω, we have
|I3|≤‖k∇(T−Tθ)⋅n‖0,∂Ω‖Tθ−g‖0,∂Ω≤C(R∂(Tθ))12. |
By the Cauchy-Schwarz and Young's inequalities (with δ3,δ4>0), we have, for I4 and I5,
|I4|+|I5|=|(k−∇Tθ|Ω−⋅nΓ−α−1[Tθ]Γ,[v]Γ)0,Γ|+12|([k∇Tθ⋅n]Γ,[v]Γ)0,Γ|≤‖k−∇Tθ|Ω−⋅nΓ−α−1[Tθ]Γ‖0,Γ⋅‖[v]Γ‖0,Γ+12‖[k∇Tθ⋅n]Γ‖0,Γ⋅‖[v]Γ‖0,Γ≤δ32‖k−∇Tθ|Ω−⋅nΓ−α−1[Tθ]Γ‖20,Γ+12δ3‖[v]Γ‖20,Γ+δ42‖[k∇Tθ⋅n]Γ‖20,Γ+18δ4‖[v]Γ‖20,Γ≤max{δ32α2,δ42}RΓ(Tθ)+(12δ3+18δ4)‖[v]Γ‖20,Γ. |
Since T,Tθ∈C1(Ω±) and [k∇⋅nΓ]=0, we have, for I6,
|I6|≤‖[k∇Tθ⋅n]Γ‖0,Γ‖{T−Tθ}‖0,Γ≤C(RΓ(Tθ))12. |
Then, plugging the estimates of I1, ⋯, I6 into (3.6), together with (3.1), we have
[1−˜C(12δ1+12δ2)]‖v‖21,Ω±+(1−˜C2δ3−˜C8δ4)‖[v]Γ‖20,Γ≤˜C[12max(δ1,δ2)RΩ(Tθ)+R∂(Tθ)+C(R∂(Tθ))12+max{δ32α2,δ42}RΓ(Tθ)+C(RΓ(Tθ))12] |
Now, by the choice
δ1=δ2=2˜C,δ3=2˜C,δ4=˜C2, |
we obtain the desired inequality.
Remark 3.4. As stated in Theorem 3.3, the constant C may depend on the piecewise C1-norm of the neural network Tθ. In practice, to prevent this norm from becoming excessively large, we can include an L2-regularization term in the loss functional L, following the approach in [29,31].
As proved in the previous subsection, the piecewise H1-error can be controlled by the residuals RΩ, R∂ and RΓ. Given some regularity assumption, we will prove that if n>1, then the residuals can be made arbitrarily small, i.e., for any ε>0, there exists Tθ∈N±n such that R(Tθ)<ε.
Lemma 3.5. Suppose that n>1 and the solution T of (2.1) satisfies T|Ωs∈W3,∞(Ωs) for s=±. For any ε>0, there exists a neural network T±θ∈N±n such that
‖T−T±θ‖W2,∞(Ω+)+‖T−T±θ‖W2,∞(Ω−)≤ε. |
Proof. By the argument given in the beginning of Subsection 2.2, for sufficiently large R>0, one can construct an extension Taug∈W3,∞([−R,R]d+1) of T such that Ω×[−1,1]⊂[−R,R]d+1 and
Taug(x,1)=T(x),∀x∈Ω+,Taug(x,−1)=T(x)∀x∈Ω−. |
Then, according to the approximation theorem of neural network functions with the tanh activation function [32], there exists Taugθ∈Naugn such that
‖Taug−Taugθ‖W3,∞([−R,R]d+1)≤ε. |
Now the conclusion follows from the Sobolev embedding theorem (cf. [26]).
Using the lemma above, we obtain the desired theorem.
Theorem 3.6. Suppose that n>1 and the solution T of (2.1) satisfies T|Ωs∈W3,∞(Ωs) for s=±. For any ε>0, there exists Tθ∈N±n such that
R(Tθ)≤Cε, |
where C is a positive constant depending only on Ω, Γ, k, α and T.
Proof. Let ε>0 be chosen arbitrarily. By Lemma 3.5, there exists Tθ∈N±n such that
‖T−T±θ‖W2,∞(Ω+)+‖T−T±θ‖W2,∞(Ω−)≤ε. |
Therefore, we have
‖T−Tθ‖22,Ω++‖T−Tθ‖22,Ω−≤|Ω+|‖T−T±θ‖W2,∞(Ω+)+|Ω−|‖T−T±θ‖W2,∞(Ω−)≤|Ω|ε. | (3.7) |
Since T is the solution of (2.1), we obtain from (3.7) and the trace inequality that
RΩ(Tθ)=‖k+Δ(T−Tθ)‖20,Ω++‖k−Δ(T−Tθ)‖20,Ω−≤max{(k−)2,(k+)2}|Ω|ε,RΓ(Tθ)≤C(‖T−Tθ‖22,Ω++‖T−Tθ‖22,Ω−)≤C|Ω|ε,R∂(Tθ)=‖T−Tθ‖20,∂Ω≤Cε. |
This completes the proof.
Remark 3.7. In the proof of Lemma 3.5, the approximation property of the tanh neural network functions [32] was employed. This validates our choice of tanh activation functions in (2.3).
In this section, we present some examples. The domain Ω=[−1,1]2 is separated by zeros of a level set function L(x,y), i.e., Ω−={(x,y)∈Ω:L(x,y)<0} and Ω+={(x,y)∈Ω:L(x,y)>0}. For all examples, we assume that n=2 (that is, the number of hidden layers is two) and the tangent hyperbolic function is adopted for the activation function. In addition, we assign N-number of nodes for each hidden layer. The weights in the neural network are initialized using the Kaiming uniform method [33], with biases set to zero.
We present three examples, each featuring a different interface shape: circle, line, and perturbed circle. In Example 4.1–4.2, the L2- and L∞-errors against the exact solutions are computed, where we observe reasonable convergence as the parameters increase. In Example 4.3, a unidirectional heat flux generated by some boundary condition is considered. Since the analytic solution is unknown in this case, we validate the PINN-predicted solution by comparing it with the FEM solution. All examples are performed on a single core of an Intel(R) Core(TM) i7-13700 CPU.
In this subsection, we justify our choice of optimizer and the training sampling strategy. We compare the results obtained by two optimizers in Example 4.1. The first is the Levenberg–Marquardt (LM) optimizer [34] with 3,000 epochs. The LM optimizer is configured with an initial damping parameter of 105, which is updated at every iteration. The damping parameter is divided by 1.3 when the process is successful and is multiplied by 3 otherwise. The second is the Limited-memory Broyden–Fletcher–Goldfarb–Shanno (L-BFGS) optimizer [35] with 10,000 epochs, a learning rate of 0.1, and a strong Wolfe line search option.
To describe the training points, let us introduce the order-related parameter m, which determines the total number of training points. The inner points (m2-number) are determined either by Chebyshev–Gauss points, Gauss–Legendre points, or uniformly spaced grid points. For example, Chebyshev–Gauss points of order m is defined by
xij=(cos((2i+1)π/2m),cos((2j+1)π/2m)),0≤i,j<m |
The interface (4m-number) and boundary points (4m-number) are uniformly sampled at the interface and boundary, respectively. In this way, the total number of training points corresponding to the parameter m is m2+8m. We compare the results obtained by different types of training points in Example 4.1.
Example 4.1 (Circle shaped interface). In this example, the level set function is L(x,y)=x2+y2−r20 and the exact solution is
T={x2+y22k−−r0α+(12k+−12k−)r20,if (x,y)∈Ω−,x2+y22k+,if (x,y)∈Ω+, |
where r0=0.5 and α=1. We consider two different thermal conductivity contrasts: (k−,k+)=(1,10), or (100,1).
We compare loss curves with increasing training epochs for the LM and L-BFGS optimizer in Figure 3. The loss function decreases to the lower level by the LM optimizer. Consequently, we adopted the LM optimizer for all experiments in this study. Next, we compare the L2 and L∞ errors computed using Chebyshev–Gauss, Gauss–Legendre, and uniform grids type training points. We report errors in Figure 4. Overall, the smallest errors are achieved with the choice of Chebyshev–Gauss points and with N=20. Therefore, we use Chebysev–Gauss type training points for the rest of the Section.
Now, let us present numerical results by the proposed PINN method. Figure 5 presents the numerical solutions for the two conductivity cases. Notably, no spurious oscillation appears near the interface in any case. We report the errors L2 and L∞ that are computed at 10,000 uniformly sampled test points. The errors and CPU time for the two conductivity contrasts, that is, (k+,k−)=(1,10) or (100,1), are documented in Tables 1 and 2, respectively.
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 4.55E−6 | 1.02E−6 | 75.40 |
16 | 9.70E−7 | 2.17E−7 | 90.16 | |
32 | 4.82E−6 | 1.08E−6 | 156.61 | |
20 | 8 | 5.63E−6 | 1.11E−6 | 114.57 |
16 | 2.87E−6 | 3.89E−7 | 157.11 | |
32 | 3.70E−7 | 1.44E−7 | 261.38 |
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 8.99E−4 | 2.54E−4 | 75.90 |
16 | 9.58E−6 | 1.98E−6 | 88.90 | |
32 | 4.46E−6 | 6.29E−7 | 157.78 | |
20 | 8 | 1.79E−3 | 6.74E−4 | 113.07 |
16 | 3.53E−5 | 3.94E−6 | 157.24 | |
32 | 3.02E−6 | 3.61E−7 | 273.93 |
Let us discuss the computational complexities with respect to N and m. One of the bottlenecks of the LM optimizer is the inversion of the Hessian matrix associated with the loss functional. In our case, the number of entries in the Hessian matrix is O(N2m2). As a results, the CPU time may dramatically increases when N and m increase. Fortunately, the smallest errors were achieved at relatively small parameters, specifically N=20 and m=32. In particular, with this parameter choice, the L2 errors remain below 4⋅10−7 in Tables 1 and 2 with corresponding CPU times of 261 (s) and 274 (s), respectively.
Before closing this example, we compare the proposed method with the FEM-type method. One of the authors has the experience of implementing FEM for elliptic interface problems with implicit jump conditions, as described in [10], where the L2 error is of O(h−2). To achieve comparable accuracy, the FEM approach requires solving an algebraic system with over 105 unknowns, resulting in substantial computational cost (see e.g., Table 2 in [10]). Although it is difficult to compare directly, our method achieves a comparable accuracy with less CPU time.
In this subsection, we compare the proposed PINN with other methods. In Example 4.2, we compare our method with the piecewise PINN method [36], while in Example 4.3, we compare it with a FEM-type method [11].
Example 4.2 (Line interface). The level set function is L(x,y)=x+2y+r0 and the exact solution is
T={x+2y+r0k−−√5α,if (x,y)∈Ω−,x+2y+r0k+,if (x,y)∈Ω+, |
where α=0.5 and r0=0.1. The thermal conductivity is (k−,k+)=(1,10). The graphs of Tθ and |T−Tθ| are presented in Figure 6. The L2 and L∞ errors are reported on the left side of Table 3. Notably, the L2 errors are below 2×10−7 when N=10 and m=16. Therefore, the proposed PINN successfully approximates the exact solution.
N | m | Proposed PINN | Piecewise PINN [36] | ||||
‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ||
10 | 8 | 3.98E−5 | 1.09E−5 | 77.76 | 7.38E−5 | 2.29E−5 | 288.85 |
16 | 6.32E−7 | 2.34E−7 | 96.07 | 1.10E−6 | 2.72E−7 | 345.40 | |
32 | 2.88E−6 | 6.54E−7 | 136.73 | 1.33E−5 | 1.76E−6 | 400.98 | |
20 | 8 | 2.14E−4 | 5.93E−5 | 173.40 | 9.65E−4 | 1.20E−4 | 778.21 |
16 | 1.56E−6 | 4.21E−7 | 261.18 | 3.11E−5 | 3.70E−6 | 858.13 | |
32 | 2.38E−5 | 2.91E−6 | 405.33 | 2.32E−5 | 3.06E−6 | 1149.51 |
We compare our version of PINN with piecewise PINN introduced in [36]. In [36], the authors propose two separate neural network functions for each subdomain. As a result, the total number of parameters is approximately doubled when the same number of nodes is used in each hidden layer. We report the L2 and L∞ errors and CPU time on the right side of Table 3. We observe that L2 and L∞ errors obtained by the proposed PINN and piecewise PINN are similar. However, in terms of CPU time, our PINN is more efficient, which is due to the increased number of parameters in piecewise PINN.
Example 4.3 (Perturbed-circle shaped interface). In this example, we consider the unidirectional heat flow given by the homogeneous outer source (f=0) and the boundary condition that
T=1,when x=−1,T=0,when x=1,∇T⋅nΩ=0,when y=−1ory=1, |
where (k−,k+)=(1,2) and α=0.5. We consider the perturbed circle shape interface:
r=12(1+17sin5θ). |
Since the exact solution is unknown, we compare the solution obtained by PINN with that by FEM [11], denoted as TFEM. In Appendix B, we provide a brief description of FEM for the completeness of the presentation.
Figure 7 presents the comparison between TFEM and Tθ, showing the overall agreement between the two solutions. We also calculate L2 and L∞ differences between Tθ and TFEM:
‖Tθ−TFEM‖L2(Ω)=3.31×10−3,‖Tθ−TFEM‖L∞(Ω)=1.08×10−2, |
which are reasonably small. Thus, we conclude that the solutions obtained by PINN are comparable to those from the well-established FEM. However, in terms of implementation, the proposed method is simpler than FEM type method, without the needs for the mesh generation or iteration techniques to determine jump profiles.
In this subsection, we provide some discussion regarding the numerical experiments. We begin by examining the relation between the choice of the parameter N (number of nodes in a hidden layer) and its influences on the accuracy. In Figure 8, we plot the L2 errors of the reconstructed solutions obtained by our methods with varying number of N, with m=16 fixed both for the Example 4.1 and 4.2. Here, we do not observe clear decaying trends in errors as N increases. This could result from difficulties in finding the global minimizer using optimization algorithms, as discussed in [37,38]. The convergence of the optimizer, closly related to the optimization landscape, depends on various factors including the interface shapes, PDE-related parameters (e.g., k±), etc. Notably, in Example 4.2, where both the interface shape and the solutions are relatively simple, the smallest L2 error (2.3E−7) was observed with just N=10. A similar phenomenon is observed for the m parameter, which determines the number of training points. In Figure 4, we do not observe the clear decay of errors as m increases. Based on our parameter study, we suggest selecting moderate values such as N=10,20 and m=16 rather than significantly large values.
We now turn to the discussion of the stability of the optimization process. Through our experiments, we observed that the optimization process becomes unstable when there is a large contrast of k+ and k− or the parameters N and m are too large. In particular, when the damping parameter in LM becomes close to zero, the ill-conditioned Hessian matrix may cause the loss function to diverge. To remedy this, we modified the LM optimizer slightly by introducing an additional rule: if current loss exceeds twice the previous loss, increase the damping parameter by a factor of 105. This adjustment improved the robustness of the optimization process.
Finally, let us discuss the regularity assumptions used in our error estimates. In the statements of Lemma 3.5 and Theorem 3.6, we assumed that T|Ωs∈W3,∞(Ωs) for s=±, which is a strong regularity condition. However, this assumption may not hold in practical applications, especially when material interfaces induce singularities. In such cases, the errors may not be controlled by increasing the number of parameters since the universal approximation fails under low regularity conditions.
In this work, we developed a new PINN method to solve heat transfer equations involving imperfect contact conditions. There are mainly two difficulties in solving imperfect contact problems: 1) the solutions are discontinuous across the interfaces, and 2) the jumps are implicitly determined. To remedy these, we adopted the axis-augmentation techniques introduced in [21] to establish a continuous representation of the solution in the augmented variable. Then, we defined the neural network function in an augmented variable with the physics-informed loss functional, which includes the implicit jump conditions. After the loss function is minimized, the resulting solutions obtained by PINN naturally satisfy the desired jump conditions. One of the advantages of such a continuous representation formulation was the availability of error analysis. We carried out the error analysis; our main theorem states that given a sufficient number of parameters, the loss between exact and PINN solutions can be made arbitrarily small. The numerical experiments that support our analysis are reported in the result section. Let us discuss the limitations of the proposed method. First, the number of parameters in the neural network increases with the introduction of a new axis. Second, while our analysis assumes strong regularity conditions, these conditions may not always hold in the real-world applications.
We provide details regarding the calculation of loss functionals presented in (2.5)–(2.9). Recall first that ϕθ(x)=ϕaugθ(x,1)χΩ+(x)+ϕaugθ(x,−1)χΩ−(x). While the loss functionals are written in the function ϕθ, the augmented neural network function ϕaugθ(x,z) is employed during the implementation stage. Specifically, the parameter z∈{1,−1}, indicates whether the sampling points belong to Ω− or Ω+, respectively. Below, we summarize how operators such as ∇, ∇, and [⋅] acting on ϕθ are calculated using the augmented neural network functions:
Δϕθ(x±i)=Δxϕaugθ(x±i,z=±1),[ϕθ]Γ(xΓi)=ϕaugθ(xΓi,z=1)−ϕaugθ(xΓi,z=−1),[k∇ϕθ⋅nΓ]Γ(xΓi)=k+∇xϕaugθ(xΓi,z=1)⋅nΓ−k−∇xϕaugθ(xΓi,z=−1)⋅nΓ, |
where
∇xϕaugθ=(∂ϕaugθ∂x,∂ϕaugθ∂y)Δxϕaugθ=∂2ϕaugθ∂x2+∂2ϕaugθ∂y2. |
We briefly describe the FEM introduced in [11], where the so-called immersed finite element (IFE) method was employed. In IFE methodology, uniform grids are employed for the interface problem, allowing the interface to cut through the element, which contrasts with the usage of interface-fitted grids. Instead, the basis function is modified so that the they satisfy the interface conditions.
Let Th be a triangulation of the domain Ω. When E∈Th is not cut through by the interface, the local space is a standard linear space, denoted by Sh(E). Suppose E is cut through by the interface; the basis function is defined as a piecewise polynomial of the form:
ϕ(x)={ϕ+(x)=a+x+b+y+c+,when x∈E∩Ω+ϕ−(x)=a−x+b−y+c−,when x∈E∩Ω− | (B.1) |
Here, the coefficients in (B.1) are determined by the interface conditions (2.1b)-(2.1c), i.e.,
{ϕ+−ϕ−=αk−∇ϕ−⋅nΓk+∇ϕ+⋅nΓ−k−∇ϕ−⋅nΓ=0. |
In this way, the local space ˆSh(E) is defined with nodes and degrees of freedom. The global space ˆSh(Ω) is defined by patching local spaces, imposing the continuity at the nodes. Let us remark that the dimension of ˆSh(Ω) corresponds to the number of nodes in Th. Finally, FEM solution is obtained by solving the weak problem : find Th∈ˆSh such that satisfies
ah(Th,ϕ)=(f,ϕ),∀ϕ∈ˆSh. |
where
ah(Th,ϕ)=∑E∈Th∫Ek∇Th⋅∇ϕdx+1α∫Γ[Th]Γ[ϕ]Γds. |
The detailed implementation can be found in [11].
Hansaem Oh: Formal analysis, Methodology, Writing-original draft; Gwanghyun Jo: Conceptualization, Methodology, Writing-review & editing.
The first author (H. Oh) is supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NO. RS-2024-00463063). We thank Dr. Hyeokjoo Park for his valuable advice on the error analysis.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
All authors declare no conflicts of interest in this paper.
[1] |
L. J. Challis, Kapitza resistance and acoustic transmission across boundaries at high frequencies, J. Phys. C: Solid State Phys., 7 (1974), 481. https://doi.org/10.1088/0022-3719/7/3/005 doi: 10.1088/0022-3719/7/3/005
![]() |
[2] |
E. T. Swartz, R. O. Pohl, Thermal resistance at interfaces, Appl. Phys. Lett., 51 (1987), 2200–2202. https://doi.org/10.1063/1.98939 doi: 10.1063/1.98939
![]() |
[3] |
R. J. Stoner, H. J. Maris, Kapitza conductance and heat flow between solids at temperatures from 50 to 300 K, Phys. Rev. B, 48 (1993), 16373–16387. https://doi.org/10.1103/PhysRevB.48.16373 doi: 10.1103/PhysRevB.48.16373
![]() |
[4] |
D. G. Cahill, W. K. Ford, K. E. Goodson, G. D. Mahan, A. Majumdar, H. J. Maris, et al., Nanoscale thermal transport, J. Appl. Phys., 93 (2003), 793–818. https://doi.org/10.1063/1.1524305 doi: 10.1063/1.1524305
![]() |
[5] |
P. Kapitza, Heat transfer and superfluidity of helium II, Phys. Rev., 60 (1941), 354–355. https://doi.org/10.1103/PhysRev.60.354 doi: 10.1103/PhysRev.60.354
![]() |
[6] |
C. W. Nan, R. Birringer, Determining the kapitza resistance and the thermal, Phys. Rev. B, 57 (1998), 8264–8268. https://doi.org/10.1103/PhysRevB.57.8264 doi: 10.1103/PhysRevB.57.8264
![]() |
[7] |
J. Yvonnet, Q. C. He, Q. Z. Zhu, J. F. Shao, A general and efficient computational procedure for modelling the Kapitza thermal resistance based on XFEM, Comput. Mater. Sci., 50 (2011), 1220–1224. https://doi.org/10.1016/j.commatsci.2010.02.040 doi: 10.1016/j.commatsci.2010.02.040
![]() |
[8] |
J. Weisz, On an iterative method for the solution of discretized elliptic problems with imperfect contact condition, J. Comput. Appl. Math., 72 (1996), 319–333. https://doi.org/10.1016/0377-0427(96)00003-9 doi: 10.1016/0377-0427(96)00003-9
![]() |
[9] |
S. Hüeber, B. I. Wohlmuth, Thermo-mechanical contact problems on non-matching meshes, Comput. Method. Appl. M., 198 (2009), 1338–1350. https://doi.org/10.1016/j.cma.2008.11.022 doi: 10.1016/j.cma.2008.11.022
![]() |
[10] |
G. H. Jo, D. Y. Kwak, Enriched P1-Conforming Methods for Elliptic Interface Problems with Implicit Jump Conditions, Adv. Math. Phys., 2018 (2018), 9891281. https://doi.org/10.1155/2018/9891281 doi: 10.1155/2018/9891281
![]() |
[11] | D. Y. Kwak, S. Lee, Y. Hyon, A new finite element for interface problems having robin type jump, Int. J. Numer. Anal. Model., 14 (2017), 532–549. |
[12] |
J. Yvonnet, Q. C. He, C. Toulemonde, Numerical modelling of the effective conductivities of composites with arbitrarily shaped inclusions and highly conducting interface, Compos. Sci. Technol., 68 (2008), 2818–2825. https://doi.org/10.1016/j.compscitech.2008.06.008 doi: 10.1016/j.compscitech.2008.06.008
![]() |
[13] |
J. T. Liu, S. T. Gu, E. Monteiro, Q. C. He, A versatile interface model for thermal conduction phenomena and its numerical implementation by XFEM, Comput. Mech., 53 (2014), 825–843. https://doi.org/10.1007/s00466-013-0933-9 doi: 10.1007/s00466-013-0933-9
![]() |
[14] |
T. Xue, X. Zhang, K. K. Tamma, Generalized heat conduction model involving imperfect thermal contact surface: Application of the GSSSS-1 differential-algebraic equation time integration, Int. J. Heat Mass Tran., 116 (2018), 889–896. https://doi.org/10.1016/j.ijheatmasstransfer.2017.09.081 doi: 10.1016/j.ijheatmasstransfer.2017.09.081
![]() |
[15] |
M. Raissi, P. Perdikaris, G. E. Karniadakis, Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, J. Comput. Phys., 378 (2019), 686–707. https://doi.org/10.1016/j.jcp.2018.10.045 doi: 10.1016/j.jcp.2018.10.045
![]() |
[16] | Automatic Differentiation in PyTorch, Adv. Neural Inf. Process. Syst., 2017. Available from: https://openreview.net/forum?id = BJJsrmfCZ |
[17] |
E. Haghighat, A. C. Bekar, E. Madenci, R. Juanes, A nonlocal physics-informed deep learning framework using the peridynamic differential operator, Comput. Method. Appl. Mech. Eng., 385 (2021), 114012. https://doi.org/10.1016/j.cma.2021.114012 doi: 10.1016/j.cma.2021.114012
![]() |
[18] |
H. Bararnia, M. Esmaeilpour, On the application of physics informed neural networks (PINN) to solve boundary layer thermal-fluid problems, Int. Commun. Heat Mass Tran., 132 (2022), 105890. https://doi.org/10.1016/j.icheatmasstransfer.2022.105890 doi: 10.1016/j.icheatmasstransfer.2022.105890
![]() |
[19] |
L. Ning, Z. Cai, H. Dong, Y. Liu, W. Wang, Physics-informed neural network frameworks for crack simulation based on minimized peridynamic potential energy, Comput. Method. Appl. Mech. Eng., 417 (2023), 116430. https://doi.org/10.1016/j.cma.2023.116430 doi: 10.1016/j.cma.2023.116430
![]() |
[20] |
M. Baldan, P. D. Barba, D. A. Lowther, Physics-Informed Neural Networks for Inverse Electromagnetic Problems, IEEE T. Magn., 59 (2023), 1–5. https://doi.org/10.1109/TMAG.2023.3247023 doi: 10.1109/TMAG.2023.3247023
![]() |
[21] |
W. F. Hu, T. S. Lin, M. C. Lai, A discontinuity capturing shallow neural network for elliptic interface problems, J. Comput. Phys., 469 (2022), 111576. https://doi.org/10.1016/j.jcp.2022.111576 doi: 10.1016/j.jcp.2022.111576
![]() |
[22] |
H. J. Park, G. H. Jo, A physics-informed neural network based method for the nonlinear Poisson-Boltzmann equation and its error analysis, J. Comput. Phys., 522 (2025), 113579. https://doi.org/10.1016/j.jcp.2024.113579 doi: 10.1016/j.jcp.2024.113579
![]() |
[23] |
P. Hansbo, M. G. Larson, Discontinuous Galerkin methods for incompressible and nearly incompressible elasticity by Nitsche's method, Comput. Method. Appl. Mech. Eng., 191 (2002), 1895–1908. https://doi.org/10.1016/S0045-7825(01)00358-9 doi: 10.1016/S0045-7825(01)00358-9
![]() |
[24] | D. Leguillon, E. Sanchez-Palencia, Computation of singular solutions in elliptic problems and elasticity, New York: John Wiley & Sons, Inc., 1987. http://doi.org/10.5555/39281 |
[25] | Evans, C. Lawrence, Partial differential equations, 2 Eds., Providence, RI: Am.Math.Soc., 2010. http://doi.org/10.1090/gsm/019 |
[26] | J. Wloka, Partial differential equations, Cambridge: Cambridge Univ.Press, 1987. http://doi.org/10.1017/CBO9781139171755 |
[27] |
K. Hornik, M. Stinchcombe, H. White, Multilayer feedforward networks are universal approximators, Neural Netw., 2 (1989), 359–366. https://doi.org/10.1016/0893-6080(89)90020-8 doi: 10.1016/0893-6080(89)90020-8
![]() |
[28] |
R. E. Caflisch, Monte Carlo and quasi-Monte Carlo methods, Acta Numer., 7 (1998), 1–49. https://doi.org/10.1017/S0962492900002804 doi: 10.1017/S0962492900002804
![]() |
[29] |
S. Mishra, R. Molinaro, Estimates on the generalization error of physics-informed neural networks for approximating PDEs, IMA J. Numer. Anal., 43 (2022), 1–43. https://doi.org/10.1093/imanum/drab093 doi: 10.1093/imanum/drab093
![]() |
[30] |
S. Mishra, R. Molinaro, Estimates on the generalization error of physics-informed neural networks for approximating a class of inverse problems for PDEs, IMA J. Numer. Anal., 42 (2021), 981–1022. https://doi.org/10.1093/imanum/drab032 doi: 10.1093/imanum/drab032
![]() |
[31] |
T. D. Ryck, A. D. Jagtap, S. Mishra, Error estimates for physics-informed neural networks approximating the Navier-Stokes equations, IMA J. Numer. Anal., 44 (2023), 83–119. https://doi.org/10.1093/imanum/drac085 doi: 10.1093/imanum/drac085
![]() |
[32] |
T. D. Ryck, S. Lanthaler, S. Mishra, On the approximation of functions by tanh neural networks, Neural Netw., 143 (2021), 732–750. https://doi.org/10.1016/j.neunet.2021.08.015 doi: 10.1016/j.neunet.2021.08.015
![]() |
[33] |
K. He, X. Zhang, S. Ren, J. Sun, Delving deep into rectifiers: Surpassing human-level performance on ImageNet classification, Proc. IEEE Int. Conf. Comput. Vis., (2015), 1026–1034. https://doi.org/10.1109/ICCV.2015.123 doi: 10.1109/ICCV.2015.123
![]() |
[34] |
D. W. Marquardt, An Algorithm for Least-Squares Estimation of Nonlinear Parameters, J. Soc. Ind. Appl. Math., 11 (1963), 431–441. https://doi.org/10.1137/0111030 doi: 10.1137/0111030
![]() |
[35] |
D. C. Liu, J. Nocedal, On the limited memory BFGS method for large scale optimization, Math. Program., 45 (1989), 503–528. https://doi.org/10.1007/BF01589116 doi: 10.1007/BF01589116
![]() |
[36] |
C. He, X. Hu, L. Mu, A mesh-free method using piecewise deep neural network for elliptic interface problems, J. Comput. Appl. Math., 412 (2022), 114358. https://doi.org/10.1016/j.cam.2022.114358 doi: 10.1016/j.cam.2022.114358
![]() |
[37] |
J. F. Urbán, P. Stefanou, J. A. Pons, Unveiling the optimization process of Physics Informed Neural Networks: How accurate and competitive can PINNs be?, J. Comput. Phys., 523 (2025), 113656. https://doi.org/10.1016/j.jcp.2024.113656 doi: 10.1016/j.jcp.2024.113656
![]() |
[38] |
C. Liu, L. Zhu, M. Belkin, Loss landscapes and optimization in over-parameterized non-linear systems and neural networks, Appl. Comput. Harmon. Anal., 59 (2022), 85–116. https://doi.org/10.1016/j.acha.2021.12.009 doi: 10.1016/j.acha.2021.12.009
![]() |
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 4.55E−6 | 1.02E−6 | 75.40 |
16 | 9.70E−7 | 2.17E−7 | 90.16 | |
32 | 4.82E−6 | 1.08E−6 | 156.61 | |
20 | 8 | 5.63E−6 | 1.11E−6 | 114.57 |
16 | 2.87E−6 | 3.89E−7 | 157.11 | |
32 | 3.70E−7 | 1.44E−7 | 261.38 |
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 8.99E−4 | 2.54E−4 | 75.90 |
16 | 9.58E−6 | 1.98E−6 | 88.90 | |
32 | 4.46E−6 | 6.29E−7 | 157.78 | |
20 | 8 | 1.79E−3 | 6.74E−4 | 113.07 |
16 | 3.53E−5 | 3.94E−6 | 157.24 | |
32 | 3.02E−6 | 3.61E−7 | 273.93 |
N | m | Proposed PINN | Piecewise PINN [36] | ||||
‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ||
10 | 8 | 3.98E−5 | 1.09E−5 | 77.76 | 7.38E−5 | 2.29E−5 | 288.85 |
16 | 6.32E−7 | 2.34E−7 | 96.07 | 1.10E−6 | 2.72E−7 | 345.40 | |
32 | 2.88E−6 | 6.54E−7 | 136.73 | 1.33E−5 | 1.76E−6 | 400.98 | |
20 | 8 | 2.14E−4 | 5.93E−5 | 173.40 | 9.65E−4 | 1.20E−4 | 778.21 |
16 | 1.56E−6 | 4.21E−7 | 261.18 | 3.11E−5 | 3.70E−6 | 858.13 | |
32 | 2.38E−5 | 2.91E−6 | 405.33 | 2.32E−5 | 3.06E−6 | 1149.51 |
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 4.55E−6 | 1.02E−6 | 75.40 |
16 | 9.70E−7 | 2.17E−7 | 90.16 | |
32 | 4.82E−6 | 1.08E−6 | 156.61 | |
20 | 8 | 5.63E−6 | 1.11E−6 | 114.57 |
16 | 2.87E−6 | 3.89E−7 | 157.11 | |
32 | 3.70E−7 | 1.44E−7 | 261.38 |
N | m | ‖T−Tθ‖L∞ | ‖T−Tθ‖L2 | CPU time |
10 | 8 | 8.99E−4 | 2.54E−4 | 75.90 |
16 | 9.58E−6 | 1.98E−6 | 88.90 | |
32 | 4.46E−6 | 6.29E−7 | 157.78 | |
20 | 8 | 1.79E−3 | 6.74E−4 | 113.07 |
16 | 3.53E−5 | 3.94E−6 | 157.24 | |
32 | 3.02E−6 | 3.61E−7 | 273.93 |
N | m | Proposed PINN | Piecewise PINN [36] | ||||
‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ‖Tθ−T‖L∞ | ‖Tθ−T‖L2 | CPU time | ||
10 | 8 | 3.98E−5 | 1.09E−5 | 77.76 | 7.38E−5 | 2.29E−5 | 288.85 |
16 | 6.32E−7 | 2.34E−7 | 96.07 | 1.10E−6 | 2.72E−7 | 345.40 | |
32 | 2.88E−6 | 6.54E−7 | 136.73 | 1.33E−5 | 1.76E−6 | 400.98 | |
20 | 8 | 2.14E−4 | 5.93E−5 | 173.40 | 9.65E−4 | 1.20E−4 | 778.21 |
16 | 1.56E−6 | 4.21E−7 | 261.18 | 3.11E−5 | 3.70E−6 | 858.13 | |
32 | 2.38E−5 | 2.91E−6 | 405.33 | 2.32E−5 | 3.06E−6 | 1149.51 |