Loading [MathJax]/jax/output/SVG/jax.js
Review

Artificial intelligence and machine learning in aerodynamics

  • With the increasing availability of flow data from simulation and experiment, artificial intelligence and machine learning are revolutionizing the research paradigm in aerodynamics and related disciplines. The integration of machine learning with theoretical, computational, and experimental investigations unlocks new possibilities for solving cutting-edge problems. In this paper, we review the status of artificial intelligence and machine learning in aerodynamics, including knowledge discovery, theoretical modeling, numerical simulation, and multidisciplinary applications. Representative techniques and successful applications are summarized. Finally, despite successful applications, challenges still remain, which are discussed in the conclusion.

    Citation: Jiaqing Kou, Tianbai Xiao. Artificial intelligence and machine learning in aerodynamics[J]. Metascience in Aerospace, 2024, 1(2): 190-218. doi: 10.3934/mina.2024009

    Related Papers:

    [1] Kangqun Zhang . Existence and uniqueness of positive solution of a nonlinear differential equation with higher order Erdélyi-Kober operators. AIMS Mathematics, 2024, 9(1): 1358-1372. doi: 10.3934/math.2024067
    [2] Weerawat Sudsutad, Chatthai Thaiprayoon, Aphirak Aphithana, Jutarat Kongson, Weerapan Sae-dan . Qualitative results and numerical approximations of the (k,ψ)-Caputo proportional fractional differential equations and applications to blood alcohol levels model. AIMS Mathematics, 2024, 9(12): 34013-34041. doi: 10.3934/math.20241622
    [3] Apassara Suechoei, Parinya Sa Ngiamsunthorn . Extremal solutions of φCaputo fractional evolution equations involving integral kernels. AIMS Mathematics, 2021, 6(5): 4734-4757. doi: 10.3934/math.2021278
    [4] Hadjer Belbali, Maamar Benbachir, Sina Etemad, Choonkil Park, Shahram Rezapour . Existence theory and generalized Mittag-Leffler stability for a nonlinear Caputo-Hadamard FIVP via the Lyapunov method. AIMS Mathematics, 2022, 7(8): 14419-14433. doi: 10.3934/math.2022794
    [5] Snezhana Hristova, Antonia Dobreva . Existence, continuous dependence and finite time stability for Riemann-Liouville fractional differential equations with a constant delay. AIMS Mathematics, 2020, 5(4): 3809-3824. doi: 10.3934/math.2020247
    [6] Reny George, Sina Etemad, Ivanka Stamova, Raaid Alubady . Existence of solutions for [p,q]-difference initial value problems: application to the [p,q]-based model of vibrating eardrums. AIMS Mathematics, 2025, 10(2): 2321-2346. doi: 10.3934/math.2025108
    [7] Özge Arıbaş, İsmet Gölgeleyen, Mustafa Yıldız . On the solvability of an initial-boundary value problem for a non-linear fractional diffusion equation. AIMS Mathematics, 2023, 8(3): 5432-5444. doi: 10.3934/math.2023273
    [8] Thanin Sitthiwirattham, Rozi Gul, Kamal Shah, Ibrahim Mahariq, Jarunee Soontharanon, Khursheed J. Ansari . Study of implicit-impulsive differential equations involving Caputo-Fabrizio fractional derivative. AIMS Mathematics, 2022, 7(3): 4017-4037. doi: 10.3934/math.2022222
    [9] Kamal Shah, Muhammad Sher, Muhammad Sarwar, Thabet Abdeljawad . Analysis of a nonlinear problem involving discrete and proportional delay with application to Houseflies model. AIMS Mathematics, 2024, 9(3): 7321-7339. doi: 10.3934/math.2024355
    [10] Naimi Abdellouahab, Keltum Bouhali, Loay Alkhalifa, Khaled Zennir . Existence and stability analysis of a problem of the Caputo fractional derivative with mixed conditions. AIMS Mathematics, 2025, 10(3): 6805-6826. doi: 10.3934/math.2025312
  • With the increasing availability of flow data from simulation and experiment, artificial intelligence and machine learning are revolutionizing the research paradigm in aerodynamics and related disciplines. The integration of machine learning with theoretical, computational, and experimental investigations unlocks new possibilities for solving cutting-edge problems. In this paper, we review the status of artificial intelligence and machine learning in aerodynamics, including knowledge discovery, theoretical modeling, numerical simulation, and multidisciplinary applications. Representative techniques and successful applications are summarized. Finally, despite successful applications, challenges still remain, which are discussed in the conclusion.



    The objective in RL is to teach an agent that sequentially interacts with an environment to choose "good" actions. For each action, the agent receives an immediate real-valued reward. The rewards are accumulated over time, resulting in the so-called return, which describes the overall performance of the agent. Randomness may be involved at all levels of this interaction: in choosing actions, the environment reacting to actions, and/or in the rewards received by the agent. Hence, the return is to be considered random as well. In more classical approaches to RL problems, the randomness is averaged out, and only the expected return is considered when evaluating the performance of an agent. In [1], not only the expectation but the complete distribution of the return was considered, introducing what is now known as distributional RL, see [2].

    Mathematically, an RL problem is typically modeled by a {Markov decision process} (MDP), that is, as a particular type of a discrete-time stochastic control problem. An overview of the classical MDP theory and its applications is presented in [3]. For more details on distributional RL using notations similar to here, we refer to [2,4].

    For any measurable space X, we write P(X) for the set of probability distributions on X. The distribution (law) of a X-valued random variable (RV) X is denoted by L(X)=P[X]P(X). We also write short Xμ, if L(X)=μP(X). In case X is countable, discrete distributions νP(X) can be identified with functions ν:X[0,1] satisfying xXν(x)=1. For a random variable X and an event A with P[A]>0, we write L(X|A)P(X) for the conditional distribution of X given A.

    Let S and A be non-empty finite sets. A MDP on states S and actions A is a function

    ρ:S×AP(S×R),(s,a)ρ(s,a).

    The interpretation is as follows: S is the set of states an environment can occupy and A the set of possible actions the agent can perform. If in state sS the agent performs action aA, the environment reacts with a (possibly random) successor state S together with a (possibly random) real-valued reward R having joint distribution (S,R)ρ(s,a)P(S×R). How an agent chooses its actions can be modeled by a (stationary) policy

    π:SP(A),sπs.

    An agent that is in state s and acts according to policy π chooses a (possibly random) action A distributed as AπsP(A).

    Suppose an agent acts according to π. Starting at time t=0 from a possibly random state S(0), the following dynamic is defined inductively: at time tN0, the agent finds itself in state S(t), chooses action A(t)πS(t) (independent of the past given S(t)) and the environment reacts with the next state S(t+1) and immediate reward R(t) jointly distributed as (S(t+1),R(t))ρ(S(t),A(t)) (independent of the past given (S(t),A(t))). The resulting stochastic process

    (S(t),A(t),R(t))tN0 (1.1)

    is called a Markov reward process in [2] or a full MDP in [4]. To emphasize that the distribution of the stochastic process (1.1) depends on π, we write Pπ instead of P.

    A suitable way to judge the overall performance of the agent is to choose a discount factor γ(0,1) and to consider the discounted accumulated rewards, the so-called return:

    G=R(0)+γR(1)+γ2R(2)+=t=0γtR(t).

    The return G is defined as an infinite series of random variables and its existence as a R-valued random variable is not automatically guaranteed, such as if rewards are (extremely) heavy-tailed. Theorem 1 below provides a complete characterization of when the return G exists almost surely as a R-valued random variable starting from any state S(0)=s. The relation of Theorem 1 to earlier results is discussed in Remarks 2, 3 and 4. In various applications, heavy-tailed reward distributions are not encountered. For instance, if rewards are obtained as a deterministic function of states and actions, the rewards become bounded, and the existence of the return G easily follows from a geometric series argument. However, unbounded and even heavy-tailed reward distributions are of interest in various applications: fields in which RL approaches have been considered and heavy-tailed distributions play a crucial role include insurance and finance. See [5] and [6] for RL approaches to pricing and trading and [7] for the role of heavy-tailed distributions in that field. Recently, another research area in which heavy-tailed reward distributions have been considered is the study of multi-armed bandit models, which corresponds to an MDP with only a single state, |S|=1. In [8], the authors report on several studies in this direction, including linear bandits [9,10], pure exploration [11], Lipschitz-bandits [12], Bayesian optimization [13], and Thompson sampling [14]. Notably, in the latter work, a particular emphasis is placed on α-stable reward distributions, which is an assumption also covered by our analysis, see Theorem 4 and Example 3. It is worth noting that all of these works consider finite horizon non-discounted returns rather than infinite discounted returns. Our analysis offers a theoretical justification for the inclusion of heavy-tailed reward distributions in infinite discounted reward scenarios.

    Remark 1. The case γ=0 is sometimes considered in RL literature, resulting in G=R(0) and trivializing many of the research questions we investigate here. Although we do not discuss this case further, it should be noted that several parts of our results remain applicable even when γ=0.

    Suppose the return G exists as a R-valued RV starting from any state s. Of interest in distributional policy evaluation are the distributions of the return starting from given states. The state(-action) return distributions are defined by

    ηπs=Pπ[G|S(0)=s],sS,ηπ(s,a)=Pπ[G|S(0)=s,A(0)=a],(s,a)S×A.

    In [2], the collection ηπ=(ηπs)sS is called a return distribution function. The state(-action) return distributions can be found as solutions to the distributional Bellman equations, which we now explain. For rR,γ(0,1) let fr,γ:P(R)P(R) be the map that sends ν=L(X) to fr,γ(ν)=L(r+γX). Using the recursive structure of the return, G=R(0)+γG with G=t=0γtR(t+1), and Markov properties of (1.1), it is seen that state and state-action return distributions are related by

    ηπs=aAπs(a)ηπ(s,a),sS,ηπ(s,a)=S×Rfr,γ(ηπs)dρ(s,a)(s,r),(s,a)S×A.

    Substituting the formulas in one another yields the distributional Bellman equations, which comes in two forms: one for states and one for state-actions:

    ηπs=aAS×Rπs(a)fr,γ(ηπs)dρ(s,a)(s,r),sS,ηπ(s,a)=S×RaAπs(a)fr,γ(ηπ(s,a))dρ(s,a)(s,r),(s,a)S×A. (1.2)

    The distributional Bellman equations are a system of |S| resp. |S×A| one-dimensional distributional equations in R. The right hand side of the equations can be used to introduce the distributional Bellman operator, which for the state return distributions is defined as:

    Tπ:P(R)SP(R)S,η=(ηs)sSTπ(η)=(Tπs(η))sSwith s -th component functionTπs(η)=aAS×Rπs(a)fr,γ(ηs)dρ(s,a)(s,r). (1.3)

    No assumption on the MDP ρ nor the policy π is needed to define (1.8), and the relation to the state return distributions is as follows: if the return G, starting from any state s, converges almost surely in R then the return distribution function ηπ=(ηπs)sS is a fixed point of Tπ, that is ηπ=Tπ(ηπ). We explore this connection in more detail.

    We write log+(x)=log(x) for x1 and log+(x)=0 for x<1. The definition of "essential state" is given before Theorem 2. One result of this note is the following:

    Theorem 1. The following are equivalent:

    (i) Tπ has a fixed point ηP(R)S,

    (ii) G=t=0γtR(t) converges Pπ[|S(0)=s]-almost surely in R for every initial state sS,

    (iii) S×Rlog+(|r|)dρ(s,a)(s,r)< for every pair (s,a)S×A that is essential with respect to the law of the Markov chain (S(t),A(t))tN0 under Pπ,

    (iv) For every νP(R)S as n the sequence (Tπ)n(ν) of n-th iterations of Tπ converges weakly in the product space P(R)S.

    If these hold the fixed point of Tπ is unique, given by the return distribution function ηπ and (Tπ)n(ν)ηπ weakly as n for every νP(R)S.

    A theorem that reads analogously can be formulated for the distributional Bellman operator Tπ:P(R)S×AP(R)S×A defined in terms of state-actions. We cover both cases simultaneously by introducing simplified notations in Section 1.2 that allow to analyze the distributional Bellman operator, its fixed point equations and its solutions more conveniently. In Section 2, we present our main results in these notations, Theorem 1 will follow from the more general Theorem 2 presented there. For connections to a model and results in [15] see Remark 2.

    Besides giving necessary and sufficient conditions for the existence of solutions to the distributional Bellman equations, we also study properties of their solutions, that is, of the return distribution function ηπ (and also state-action return distributions). We consider tail probability asymptotics ηπs((,x)) and ηπs((x,)) as x. In Theorem 3, we identify cases of exponential decay and existence of p-th moments and in Theorem 4 we cover regular variation. A possible application of the latter is in distributional dynamic programming (see Chapter 5 in [2]), see Remark 6 of the present note.

    Building upon our simplified notation, we explore the connection of the distributional Bellman equations to multivariate distributional equations of the form Xd=AX+B in Section 2.1. This connection seems to have not been noticed in the literature so far, and can be used to apply available results in that field to the distributional RL setting; we do that when proving Theorem 4 by applying results presented in [16].

    Before we switch to the simplified notation and present the main results, we shortly present the ordinary Bellman equations for convenience.

    The following is well-known and the standard setting in much of the distributional RL literature: if reward distributions ρ(s,a)(S×)P(R) have finite expectations for all (s,a), the return converges almost surely and has finite expectations as well; this also follows from Theorems 2 and 3 presented in this note. In many classical approaches to RL, only expected returns are used for policy evaluation. The state values and state-action values are defined by

    vπs=Eπ[G|S(0)=s]=Rgdηπs(g),qπ(s,a)=Eπ[G|S(0)=s,A(0)=a]=Rgdηπ(s,a)(g).

    The collection vπ=(vπs)sS is called a state value function and qπ a state-action value function. They can be found as solutions to the (ordinary) Bellman equations, which can be derived from their distributional counterparts (1.2) using linearity of expectation. For action a in state s, let r(s,a)R be the expected reward and p(s,a)P(S) the distribution of the next state, that is

    r(s,a)=rdρ(s,a)(s,r)=Eπ[R(t)|S(t)=s,A(t)=a],p(s,a)=ρ(s,a)(×R)=Pπ[S(t+1)|S(t)=s,A(t)=a].

    The ordinary Bellman equations are a system of |S| (resp. |S×A|) linear equations in the same number of unknowns, and read as follows:

    vπs=aAπs(a)[r(s,a)+γsSp(s,a)(s)vπs],sS,qπ(s,a)=r(s,a)+γsSp(s,a)(s)aAπs(a)qπ(s,a),(s,a)S×A. (1.4)

    As in the distributional setting, one can use the right-hand side of these equations to define the (ordinary) Bellman operators (RSRS for states) such that (1.4) are equivalent to the associated fixed point equation of these operators.

    When judging a policy based on expected returns, π is considered to be at least as good as some other policy ˜π if vπv˜π pointwise at each s. Classical MDP theory shows that there always exists a policy at least as good as every other policy. Being able to calculate expected returns is not just of interest for evaluating π, but also to improve it: the policy π defined by πs=δargmaxaAqπ(s,a) is at least as good as π. This leads to an iterative algorithm known as policy iteration: starting from an initial policy π(0) and letting π(k+1)=(π(k)) it can be shown that vπ(k) converges as k to the state value function of an optimal policy.

    We introduce simplified notations and define the distributional Bellman operator in this setting. Let dN and (R1,J1),,(Rd,Jd) be random pairs such that each Ri is real-valued and Ji is discrete taking values in [d]={1,,d}. For each i[d], it is L(Ri,Ji)P(R×[d]) the joint distribution of the random pair (Ri,Ji). The random pairs (R1,J1),,(Rd,Jd) together with some constant γ(0,1) define (a version of) the distributional Bellman operator as follows:

    T:P(R)dP(R)d,η=(η1ηd)T(η)=(T1(η)Td(η)),

    the i-th coordinate function Ti:P(R)dP(R) being

    Ti(η)=Ti(η1,,ηd)=L(Ri+γGJi), (1.5)

    where on the right hand side L(Gj)=ηj for each j[d] and G1,,Gd,(Ri,Ji) are independent.

    Example 1 (State return). Let ρ be a MDP and π a stationary policy. Let d=|S| and choose an arbitrary enumeration i=s to identify [d] with S. Let i=s,j=s and BR be measurable. Consider the joint distribution

    P[RiB,Ji=j]=Pπ[R(t)B,S(t+1)=s|S(t)=s]=aAπs(a)ρ(s,a)({s}×B).

    In this case T in (1.5) is the same as Tπ in (1.3).

    Example 2 (State-action return). Let ρ be a MDP and π a stationary policy. Let d=|S×A| and choose an arbitrary enumeration i=(s,a) to identify [d] with S×A. Let i=(s,a),j=(s,a) and BR measurable. Consider the joint distribution

    P[RiB,Ji=j]=Pπ[R(t)B,S(t+1)=s,A(t+1)=a|S(t)=s,A(t)=a]=ρ(s,a)({s}×B)πs(a).

    In this case, the distributions L(Ri,Ji),i[d] completely encode both ρ and π and, provided existence, state-action return distributions (ηπ(s,a))(s,a)S×A are fixed points of T.

    The connection to distributional RL shows that it is of interest to understand the fixed points of T, that is probability distributions ηP(R)d satisfying

    η=T(η). (1.6)

    Let G1,,Gd be real-valued random variables with ηi=L(Gi). Then, (1.6) is equivalent to the distributional Bellman equations, which is a system of d one-dimensional distributional equations

    Gid=Ri+γGJi,i[d], (1.7)

    where d= (also denoted by =d) denotes equality in distribution and in the i-th equation G1,,Gd,(Ri,Ji) are independent, i[d]. For i,j[d] let

    pij=P[Ji=j]andμij=L(Ri|Ji=j),

    where in case pij=0 we set μij=δ0, the Dirac measure in 0. It holds P[Ri,Ji=j]=pijμij(). Let Fi(x)=P[Gix] be the cumulative distribution function (cdf) of real-valued random variable Gi. A third equivalent formulation of the distributional Bellman equations (1.7) in terms of cdf's is

    Fi()=dj=1pijFj(rγ)dμij(r),i[d].

    The distributional Bellman equations in terms of cdfs appeared in the literature before, for instance, in [17] or, in a more specific form, in [18].

    We study the following aspects regarding solutions to the distributional Bellman equations, i.e., fixed points of T:

    ● Existence and uniqueness for which conditions being necessary and sufficient are given in Theorem 2, see also Remark 3.

    ● Tail probability asymptotics, that is, if η is a fixed point of T with i-th component ηi=L(Gi) we study the asymptotic behavior of P[Gi>x] and P[Gi<x] as x. Depending on properties of the distributions L(R1,J1),,L(Rd,Jd), we identify cases of

    ● exponential decay and existence of p-th moments, see Theorem 3.

    ● regular variation, see Theorem 4.

    In Section 2.1, we explain how our findings relate to the area of multivariate distributional fixed point equations of the form

    Xd=AX+B, (1.8)

    in which X,B are random d-dimensional vectors, A a random d×d matrix, and X and (A,B) are independent. Such equations arise in many different applications, such as in the context of perpetuities and random difference equations. In the above notation, the distributional Bellman equations (1.7) are a system of d one-dimensional distributional equations which is less restrictive as a single distributional equation in Rd such as (1.8), but more complicated then a single one-dimensional distributional equation. However, using Theorem 2, we show how general results from the multivariate setting become applicable to study distributional Bellman equations, see Corollary 1.

    In case d=1, the random pairs (R1,J1),,(Rd,Jd) reduce to a single real-valued random variable (R1,J1) with J11 and the distributional Bellman equations (1.7) reduce to a single distributional equation in R:

    G1d=γG1+R1,

    with G1 independent of R1. This is a simple special case of a one-dimensional distributional equation X=dAX+B, also called perpetuity equation, for which the existence and uniqueness of solutions as well as their properties have been studied in great detail, see, e.g., [19,20] or [21]. A comprehensive discussion is given in Chapter 2 in [16].

    Let γ(0,1) and (R1,J1),,(Rd,Jd) be the R×[d]-valued random pairs defining the Bellman operator T:P(R)dP(R)d as in (1.12). Some results can be formulated and proven within a special construction based on random variables

    (It)tN0,(Rijt)(i,j)[d]2,tN0, (2.1)

    in which

    (It)tN0 is a [d]-valued homogeneous Markov chain having transition probabilities (pij)(i,j)[d]2 and a uniform starting distribution, that is, for any path i0,i1,,iT[d] it holds that

    P[I0=i0,,IT=iT]=d1pi0i1pi1i2piT1iT,

    (Rijt)(i,j)[d]2,tN0 forms an array of independent real-valued random variables with

    L(Rijt)=L(Ri|Ji=j)=μij,

    in particular, for each pair (i,j), the sequence (Rijt)tN0 is independent identically distributed (iid) with distribution μij,

    (It)tN0 and (Rijt)(i,j)[d]2,tN0 are independent.

    From now on, we omit quantification of indices in array-like quantities once the index-ranges have been introduced.

    Using RVs (2.1), an explicit description of the n-th iteration of T defined inductively as Tn=T(n1)T can be given: the i-th component of Tn is the map Tni:P(R)dP(R) defined by

    Tni(η1,,ηd)=L(n1t=0γtRIt,It+1,t+γnGIn|I0=i),

    with L(Gj)=ηj and G1,,Gd,(Rijt)ijt,(It)t independent.

    To introduce some basic definitions from Markov chain theory, let i,j[d] be states. We write ij if there exists tN0 with P[It=j|I0=i]>0. State i is essential if for every state j the implication ijji holds. Since the state space [d] is finite, a state i is essential if and only if it is recurrent, that is, P[It=ifor somet>0|I0=i]=1, which implies P[It=ifor  many  t|I0=i]=1.

    Theorem 2. The following conditions are equivalent:

    (i) T has a (unique) fixed point,

    (ii) E[log+|Ri|]< for all essential i,

    (iii) The infinite series G=t=0γtRIt,It+1,t is almost surely (absolutely) convergent,

    (iv) Tni(ν) converges weakly as n for each i[d] for some (all) νP(R)d.

    If one and hence all of (i)–(iv) hold, then the unique fixed point ηP(R)d of T has i-th component

    ηi=L(G|I0=i)=L(t=0γtRIt,It+1,t|I0=i).

    Remark 2. In [15], a model of iterations of random univariate affine linear maps within a Markovian environment is introduced and its stability is studied. This model is similar to our version of the distributional Bellman operator in (1.5). In [15], the Markov chain is allowed to have countable state space but required to be ergodic whereas the present Markov chain in the distributional RL setting has finite spate space and no further restrictions. The results in Section 3 of [15] partly imply claims of the present Theorem 2 with similar underlying techniques based on [19,21].

    Note that for a real-valued random variable X, the following much stronger properties each imply E[log+|X|]<:

    1) |X| is bounded: |X|K for some constant K0,

    2) |X| has a finite exponential moment: E[exp(β|X|)]< for some β>0,

    3) |X| has a finite p-th moment: E[|X|p]< for some p[1,).

    Of course, we have 1)2)3). Hence, by Theorem 2, a sufficient condition for the existence of a fixed point of T is that every Rj,j[d] satisfies one of the latter three properties. Moreover, supposing T has a fixed point η with i-component ηi=L(Gi), these properties are transferred from Rj,j[d] to Gi, the subsequent theorem presents a concise summary of these well-known facts:

    Theorem 3. Let i[d]. If for all j with ij

    1) |Rj|K then |Gi|11γK,

    2) E[exp(β|Rj|)]< then E[exp(β|Gi|)]<,

    3) E[|Rj|p]< then E[|Gi|p]<.

    The first property transfer is fundamental in MDP theory, while property transfers of the second type are well-established in perpetuity equation theory (cf. Theorem 2.4.1 in [16]). The third is common throughout the Distributional RL literature (cf. Remark 3). For completeness, a short proof of Theorem 3 is presented in Section 3.2.

    Remark 3. For p1, let Pp(R)P(R) be the subset of p-integrable distributions on R, that is, Pp(R)={μP(R)|R|x|pdμ(x)<}. Let pri:RdR be the i-th projection. The p-th Wasserstein distance on Pp(R) is defined as

    dp(μ1,μ2)=inf{R×R|xy|pdψ(x,y)|ψP(R2)withψpr1i=μi,i=1,2}1/p.

    The space (Pp(R),dp) is a complete metric space and so is the product (Pp(R)d,dp) with dp(η,ν)=maxi[d]dp(ηi,νi). In [1], the following was shown by extending techniques of [22]: if for all j[d] it holds that E[|Rj|p]<, that is L(Rj)Pp(R), then the restriction T|Pp(R)d of T to Pp(R)d maps to Pp(R)d and is a γ-contraction with respect to dp. Banach's fixed point theorem yields that the operator T|Pp(R)d:Pp(R)dPp(R)d has a unique fixed point ηPp(R)d. The latter also follows from our results: if E[|Rj|p]< for all j then E[log+|Rj|]< for all j and, by Theorem 2, there exists a unique fixed point ηP(R)d. Now L(Rj)Pp(R) for all j implies ηPp(R)dP(R)d by Theorem 3. Note, however, that our Theorem 2 improves upon results of [1] based on the dp metric twofold: First, Theorem 2 has necessary and sufficient conditions for the existence and uniqueness of fixed points of T without moment assumption such as E[|Rj|p]< for some p1. Second, the uniqueness of the fixed point is shown in the full space P(R)d, whereas the earlier results show uniqueness only in the subspace Pp(R)dP(R)d.

    Remark 4. Theorem 2 shows that, if it exists, the fixed point η of T can be obtained as the weak limit of the iterates Tn(ν) as n. This was observed in the distributional RL literature before, see Proposition 4.34 in Chapter 4 of [2], and the proof of this part of the theorem shares many similarities with their arguments.

    Next, we focus on asymptotic behavior of tail probabilities P[X>x] and P[X<x] as x. Each of the three properties stated in Theorem 3 above directly yields asymptotic bounds, in the latter two cases by applying Markov's inequality:

    1) if |X|K then P[X>x]=P[X<x]=0 for all x>K,

    2) if E[exp(β|X|)]< for some β>0 then

    lim supxlogP[X>x]xβandlim supxlogP[X<x]xβ,

    3) if E[|X|p]< for some p[1,) then P[X>x],P[X<x] are O(xp) as x.

    With this at hand, in each of the three cases presented in Theorem 3, one can obtain asymptotic bounds for the tail probabilities P[Gi>x] and P[Gi<x] as x.

    Besides the three classical properties discussed so far, a further interesting and important property of (the distribution of) a random variable X implying E[log+|X|]< is that of regular variation. Regularly varying distributions represent a broad and adaptable class of heavy-tailed distributions, containing several important distribution families as specific cases, see Example 3. As mentioned in the introduction, previous applied works have investigated scenarios that involve heavy-tailed reward distributions. Consequently, an in-depth understanding of how heavy-tailed reward behavior is reflected in the returns is of interest, a question that Theorem 4 answers in the case of regular variation.

    Regular variation is directly defined in terms of the asymptotic behavior of the tail probabilities P[X>x] and P[X<x] as x. First, a function f:(0,)[0,) with f(x)>0 for all x>x0 is called regularly varying with index βR if it is measurable and

    limxf(tx)f(x)=tβfor everyt>0.

    A function L:(0,)[0,) is called slowly varying if it is regularly varying with index β=0, so L(tx)/L(x)1 for every t>0. Every regularly varying function f with index β is of the form f(x)=xβL(x) for some slowly varying function L, see [23].

    Following [16], a real-valued random variable X is called regularly varying with index α>0 if

    a) the function xP[|X|>x] is regularly varying with index α and

    b) left and right tail probabilities are asymptotically balanced: for some q[0,1]

    limxP[X>x]P[|X|>x]=qandlimxP[X<x]P[|X|>x]=1q.

    An equivalent definition reads as follows: a real-valued random variable X is regularly varying with index α>0 if and only if there exists a slowly varying function L and some q[0,1] such that

    limxP[X>x]xαL(x)=qandlimxP[X<x]xαL(x)=1q. (2.2)

    In this situation, we have P[|X|>x]=P[X>x]+P[X<x]xαL(x) as x, so P[|X|>x] is regularly varying of index α, and the tail probabilities are asymptotically balanced.

    Similar to the other three properties, regular variation transfers from rewards to returns in the following sense:

    Theorem 4. Let i[d], α>0 and L a slowly varying function. Suppose for all j with ij, there exist qj[0,1] and cj[0,), with cj>0 for at least one j, such that

    limxP[Rj>x]xαL(x)=qjcjandlimxP[Rj<x]xαL(x)=(1qj)cj.

    Then with

    wij=t=0(1γα)γαtP[It=j|I0=i],

    that is wij=P[IN=j|I0=i] with NGeo(1γα) independent of (It)t, it holds that

    limxP[Gi>x]xαL(x)=dj=1wijqjcj1γαandlimxP[Gi<x]xαL(x)=dj=1wij(1qj)cj1γα. (2.3)

    In particular, Gi is regularly varying with index α.

    Remark 5. For d=1 Theorem 4 is equivalent to Theorem 2.4.3 (2) in [16]. For d=1, the tail asymptotic of P[|G1|>x] as x could also be directly obtained from [24], Theorem 2.3.

    Example 3. A regularly varying random variable R is called Pareto-like if P[R>x]qcxα and P[R<x](1q)cxα for some constants c>0,q[0,1],α>0, that is, if in (2.4) one can choose L to be constant. Examples of Pareto-like distributions include Pareto, Cauchy, Burr, and α-stable distributions, α<2, see [7]. If there exists a non-empty subset J[d] and α>0 such that for each jJ Rj is Pareto-like with index α and for each jJ P[|Rj|>x]=o(xα), then Theorem 4 applies by choosing L1 and yields that Gi is also Pareto-like.

    Remark 6 (Application in distributional dynamic programming). The goal in distributional dynamic programming is to perform distributional policy evaluation on a computer. A major problem to solve is that probability distributions ηsP(R) are infinite dimensional objects and hence some form of approximation has to be applied in practice. Several approaches are discussed in Chapter 5 of [2]. Additional issues arise in cases of Pareto-like tail behaviors with unbounded support as in Example 3: it is then known that (some of) the true state-value distributions are Pareto-like, which may be an information desirable to include in evaluating the policy. Theorem 4 justifies modeling the tails of the state-value distributions a priori as a Pareto-like distribution, the correct choice of asymptotic parameters is given by (2.3). We plan to report on such issues in future work.

    A prevalent method for examining probabilistic behaviors of systems involves coupling techniques, wherein multiple dependent versions of a system of interest are constructed and their collective behavior is analyzed. In this section, we adapt this approach to our specific context, demonstrating how it allows for the application of results from the well-established theory of multivariate distributional fixed point equations in the study of distributional Bellman equations.

    A first key insight is that the operator T:P(R)dP(R)d, defined in (1.5), depends on the random pairs (R1,J1),,(Rd,Jd) only through their marginal laws

    (L(R1,J1),,L(Rd,Jd))P(R×[d])d,

    recall the definition of T=(T1,,Td) being

    T:P(R)dP(R)d,T(L(G1),,L(Gd))=(L(R1+γGJ1),,L(Rd+γGJd)),

    where in the i-th component the random variables G1,,Gd,(Ri,Ji) are independent. However, the second key insight is that the independence of G1,,Gd is not necessary for the definition: any random vector (˜G1,,˜Gd) having the same marginal laws as (G1,,Gd) and being independent of (Ri,Ji) satisfies L(Ri+γ˜GJi)=L(Ri+γGJi).

    By a coupling of (R1,J1),,(Rd,Jd) we understand any way the random pairs could be defined on a common probability space, thus having a joint distribution

    L((R1,J1),,(Rd,Jd))P((R×[d])d). (2.4)

    A coupling (2.4) induces an operator S on P(Rd) given by

    S:P(Rd)P(Rd),S(L(G1,,Gd))=L(R1+γGJ1,,Rd+γGJd),

    where on the right hand side (G1,,Gd) and ((R1,J1),,(Rd,Jd)) are independent. The two key insights mentioned above yield that S and T are related by

    S(ζ)pr1i=Ti(ζpr11,,ζpr1d),i[d],ζP(Rd), (2.5)

    with pri:RdR being the i-th coordinate projection and hence ζpr1iP(R) the i-th marginal law of ζP(Rd). An immediate consequence of (2.5) of interest to us is

    L(G1,,Gd)=ζP(Rd)is a fixed point of   S(L(G1),,L(Gd))=(ζpr11,,ζpr1d)P(R)dis a fixed point of   T. (2.6)

    Corollary 1 below implies the following converse of this implication: if T has a (unique) fixed point (L(G1),,L(Gd)) then for any coupling (2.4) the induced operator S has a unique fixed point which, due to (2.6), takes the form of a coupling L(G1,,Gd) of G1,,Gd. This justifies the following approach to study (fixed points of) T: choose a convenient coupling (2.4) and study its induced operator S, in particular the marginal laws of a fixed point of S. We use this approach proving Theorem 4, where it is most convenient to consider (R1,J1),,(Rd,Jd) to be independent.

    One advantage of this coupling approach is that it enables access to numerous results from the field of multivariate fixed point equations, as we explain in the following.

    Let S be induced by a coupling (2.4). Define the random d-dimensional vector

    R=(R1,,Rd)T

    and the random d×d matrix

    J=(Jij)(i,j)[d]×[d]withJij=γ1(Ji=j).

    The coupling is then completely encoded in the joint distribution L(J,R)P(Rd×d×Rd). Moreover, for any random vector G=(G1,,Gd)T it holds

    JG+R=(R1+γGJ1,,Rd+γGJd)T

    and hence, if G is independent of (J,R), it holds that

    S(L(G))=L(JG+R)

    and L(G)P(Rd) is a fixed point of S if and only if

    Gd=JG+R. (2.7)

    Note that (2.7) is a single distributional equation in Rd, whereas the distributional Bellman equations, see (1.7), are a system of d distributional equations in R. Multivariate fixed point equations like (2.7) have long been studied for general joint laws L(J,R)P(Rd×d×Rd) under various types of assumptions, some of them directly applicable to the situation presented here. A comprehensive overview of the theory of multivariate distributional fixed point equations and its applications, in particular to stochastic difference equations and many important stochastic time-series models, is presented in [16].

    Theorem 2 can be used to show the following results, in which (J(t),R(t))tN0 are iid copies of (J,R).

    Corollary 1. For any operator S induced by a coupling (2.4), the following statements are equivalent

    (i) S has a (unique) fixed point,

    (ii) Sn(ζ) converges weakly as n for some (all) ζP(Rd),

    (iii) The infinite series G=t=0[t1s=0J(s)]R(t) is almost surely (absolutely) convergent,

    (iv) T has a (unique) fixed point,

    (v) E[log+|Ri|]< for all essential i.

    If one, and hence all, of these hold, then the unique fixed point ζP(Rd) of S is given by the law of the infinite vector-valued series L(G) in (iii) and the unique fixed point of T is the vector of marginal laws of ζ.

    Before we present a short proof, we explain how this result fits the known theory of multivariate fixed points equations. Let || be the euclidean norm on Rd. We also write || for the induced operator norm on Rd×d. A result by [25] shows that for any joint law L(J,R)P(Rd×d×Rd) such that E[log+|J|],E[log+|R|]< and such that (the distribution of) J has a strictly negative top Lyapunov exponent, that is

    inft11tE[log|t1s=0J(s)|]<0,

    the sum in (iii) converges almost surely and its law is the unique solution to (2.7). We refer to Appendix E in [16] for more information about top Lyapunov exponents. In our case, having Jij=γ1(Ji=j), and hence |t1s=0J(s)|=γt, the top Lyapunov exponent of J equals log(γ)<0, so it is strictly negative, and |J|=γ implies E[log+|J|]=0<. Thus, the implication (v)(iii) in Corollary 1 says that the condition E[log+|R|]<, which is equivalent to E[log+|Ri|]< for all i[d], can be relaxed to E[log+|Ri|]< for all essential i to obtain almost sure convergence of the infinite series in (iii) in our case.

    The implication (ii)(iii) can directly be obtained from Theorem 2.1 in [26] since |t1s=0J(s)|=γt0 (almost surely) in our case. To deduce almost sure convergence from weak convergence in Theorem 2 (implication (ii)(iii) there), we present Proposition 1 in Section 3, the proof of which is based on ideas presented in [27] to prove a version of Kolmogorov's three Series theorem also involving distributional convergence.

    Besides the top Lyapunov exponent another important notion in multivariate fixed point equations is that of irreducibility: the joint law L(J,R)P(Rd×d×Rd) is called irreducible if the only affine linear subspace HRd that fulfills P[JH+RH]=1 is the complete subspace H=Rd. In case (J,R) is irreducible, the implication (i)(iii) can be obtained directly from Theorem 2.4 in [28]. In our situation, irreducibility of L(J,R) is not given in any situation and Corollary 1 shows that it is also not needed to obtain the implication (i)(iii).

    Proof of Corollary 1. (iv)(v), see Theorem 2.

    (iv)(iii). By Theorem 2, the infinite series t=0γtRIt,It+1,t is almost surely (absolute) convergent. For each i[d], the i-th component of the vector-valued infinite series t=0[t1s=0J(s)]R(t) has the same law as L(t=0γtRIt,It+1,t|I0=i), hence every component converges almost surely (absolutely).

    (iii)(ii). Let ζ=L(G0) with G0 independent of (J(t),R(t))t. The n-th iteration of S can be represented as

    Sn(ζ)=L(n1t=0[t1s=0J(s)]R(t)+[n1s=0J(s)]G0).

    Now [n1s=0J(s)]G00 almost surely and n1t=0[t1s=0J(s)]R(t) converges almost surely by assumption. So Sn(ζ) converges in distribution, namely to the law of the infinite sum in (iii).

    (ii)(i). Assume Sn(ζ)ζ0 converges weakly for some ζ. The map S:P(Rd)P(Rd) is continuous with respect to the topology of weak convergence, hence S(ζ0)=limnS(Sn(ζ))=limnS(n+1)(ζ)=ζ0, so ζ0 is a fixed point of S.

    (i)(iv). See (2.5).

    Let mN and γ(0,1) be fixed. Consider random variables

    (Kt)tN0,(Xlt)l[m],tN0 (3.1)

    such that (Kt)t is an [m]={1,,m}-valued homogeneous Markov chain with an arbitrary starting distribution, (Xlt)lt is an array of independent real-valued random variables such that for each l[m] the variables (Xlt)t are iid and (Xlt)lt and (Kt)t are independent. A state l is called accessible if P[Kt=l]>0 for some tN0. Note that we allow any starting distribution in our formulation, hence this is not automatically satisfied. A state l is accessible essential if and only if P[Kt=lfor infinitely manytN0]>0 and this probability equals 1 if conditioned on the event of at least one visit in l. In light of Theorem 2, which we proof in the next subsection, any result shown within the setting (3.1) can be applied to (2.1) by considering

    [m][d]2,l=(i,j),Xlt=Rijt,Kt=(It,It+1)

    and in this case working under the condition P[|I0=i] is equivalent to consider the starting distribution P[K0=(i,j)]=1(i=i)pij.

    The proof of Theorem 2 is based on the following proposition, which shares similarities with Kolmogorov's three-series theorem presented Theorem 5.18 in [27].

    Proposition 1. For nN let Sn=n1t=0γtXKt,t. Then the following are equivalent

    (i) Sn converges in distribution as n.

    (ii) Sn converges almost surely as n.

    (iii) E[log+|Xlt|]< for each accessible essential state l[L].

    Proof. Every finite state space Markov chain with probability one finally takes values in one of its irreducible components, hence to show (iii)(ii), we can reduce to the case that (Kt)t is irreducible, in particular, every state is accessible essential in this case.

    (iii)(ii): We assume E[log+|Xlt|]< for all l. Let 0<c<log(1/γ) and ˜c=log(1/γ)c>0. For each tN0, let

    At={γt|XKt,t|exp(ct)}={log+|XKt,t|˜ct}.

    We show t=0P[At]< and apply the Borel–Cantelli lemma. Conditioning on Kt=l and bounding P[Kt=l]1 yields

    t=0P[At]=P[A0]+t=1P[log+|XKt,t|˜ct]P[A0]+Ll=1t=1P[log+|Xl,t|˜ct]P[A0]+Ll=1E[log+|Xl,t|˜c]<.

    The Borel–Cantelli lemma implies that Act={γt|XKt,t|<exp(ct)} occurs for all but finitely many t almost surely, and hence, almost surely t=0γt|XKt,t|<, since the tail of the infinite sum is a.s. finally dominated by that of the geometric series texp(ct)<. Hence, Sn is absolutely convergent and thus is convergent almost surely.

    (ii)(iii): We assume there is an accessible essential state l such that E[log+|Xlt|]= and show that Sn diverges almost surely. We closely follow the idea in the proof of [16], Theorem 2.1.3 (also [19], proof of Lemma 1.7) and apply the root test, which states that for any real-valued sequence (at)tN0

    lim suptN0|at|1/t>1t=0atdiverges,

    where "diverges" means not having a finite limit. Let ˜C>1, C=˜C/γ>1 and

    Bt={|γtXKt,t|1/t>˜C}={|XKt,t|1/t>C}={log+|XKt,t|>tlog(C)}.

    If we can show P[lim suptN0Bt]=1, the root test applies and yields the almost sure divergence of Sn and hence not (ii). We want to apply the second Borel–Cantelli lemma, but since the events Bt,tN0 are not independent, we have to take a little higher effort. For l[m],jN, let Tlj be the time of the j-th visit of (Kt)t in l with Tlj= in case no such visit occurs. With this, it holds that

    lim suptN0Bt=ml=1lim supjN{log+|Xl,Tlj|>Tljlog(C)}.

    For calculating the probability of the right-hand side, we can replace Xl,Tlj with Xlj since (Xlt)lt and (Tlj)lj are independent and (Xlt)t is iid for each l. Therefore, by defining the event

    Clj={log+|Xlj|>Tljlog(C)}

    we have

    P[lim suptN0Bt]=P[Ll=1lim supjNClj]P[lim supjNClj],

    where on the right hand side the state l is such that E[log+|Xlt|]= which exists by assumption. We show that the right-hand side is one. Using independence of (Xlt)t and (Tlj)j Fubini's theorem implies

    P[lim supjNClj]=P[lim supjN{log+|Xlj|>tljlog(C)}]dP(Tlj)j((tlj)j), (3.2)

    where the integral on the right hand side is with respect to the distribution of (Tlj)j under P. Since (Kt)t is assumed to be an irreducible Markov chain it holds that 1jTljcl>0 almost surely as j (where cl is the inverse of the unique stationary distribution probability for state l). Let (tlj)lj be a realization of (Tlj)lj with tljjcl as j. If we can show that for any such realization it holds that j=1P[log+|Xlj|>tljlog(C)]=, we can apply the second Borel–Cantelli lemma to conclude P[lim supjN{log+|Xlj|>tljlog(C)}]=1. Formula (3.2) and the fact that almost every realization (tlj)j of (Tlj)j satisfies tljjcl, then would yield P[lim supjNCj]=1 as desired.

    Since tljjcl for every ε>0 there is some j0 such that tlj<(1+ε)jcl for every j>j0 which allows to conclude

    j=1P[log+|Xlj|>(1+ε)jcllog(C)]=j=1P[log+|Xlj|>tljlog(C)]=

    With C=(1+ε)cllog(C)>0, we have

    j=1P[log+|Xlj|>(1+ε)jcllog(C)]E[log+|Xlj|C]=,

    since by assumption E[log+|Xlj|]=.

    (ii)(i). Obvious

    (i)(iii). We show that if there is some accessible essential state l with E[log+|Xl,t|]=, then Sn does not converge in distribution. For that, we use two technical Lemmas from [27] which were used to prove a version of Kolmogorov's three series theorem. Let (Xlt)lt be an array of independent RVs, independent of (Xlt)lt,(Kt)t and distributed as (Xlt)lt. Let ˜Xlt=XltXlt. In [27] ˜Xlt is called symmetrization of Xlt. Let Sn=n1t=0γtXKt,t and ˜Sn=SnSn=n1t=0γt˜XKt,t. Note that ˜Sn is not a symmetrization of Sn, because Sn,Sn are not independent as they both depend on (Kt)t. But for any fixed realization (kt)t we have n1t=0γt˜Xkt,t a symmetrization of n1t=0γtXkt,t, which is enough for our reasoning.

    From E[log+|Xl,t|]=, one can conclude that E[log+|˜Xl,t|]=, and hence the already established equivalence (ii)(iii) shows that ˜Sn does not converge almost surely, in fact we saw that ˜Sn diverges almost surely. Thus, using independence of (Kt)t and (˜Xkt)kt, Fubini's theorem implies

    1=P[˜Sndiverges as  n]=P[n1t=0γt˜Xkt,tdiverges as  n]dP(Kt)t((kt)t),

    hence for almost every realization (kt)t of (Kt)t the sequence n1t=0γt˜Xkt,t diverges almost surely. Theorem 5.17 in [27] about sums of independent symmetric random variables applies and shows that |n1t=0γt˜Xkt,t| in probability for almost every (kt)t. Hence, for almost every realization (kt)t and any constant C>0 it holds that

    limnP[|n1t=0γt˜Xkt,t|>C]=1.

    Since n1t=0γt˜Xkt,t is a symmetrization of n1t=0γtXkt,t, Lemma 5.19 in [27] applies and yields the following bound

    P[|n1t=0γt˜Xkt,t|>C]2P[|n1t=0γtXkt,t|>C].

    Hence, for almost every realization (kt)t of (Kt)t it holds that

    lim infnP[|n1t=0γtXkt,t|>C]12.

    Fatou's lemma together with Fubini's theorem yields

    lim infnP[|Sn|>C]12.

    This holds for every C>0 and hence the sequence L(Sn),nN is not tight which shows that it does not converge weakly.

    Proof of Theorem 2. Using the special construction let

    Sn=n1t=0γtRIt,It+1,t.

    For any νP(R)d with i-th component νi=L(Wi), let W1,,Wd,(Rijt)ijt,(It)t be independent. Then the representation

    Tni(ν)=L(Sn+γnWIn|I0=i)

    holds. (i)(iv). Let ηP(R)d be a fixed point of T, hence with i-th component ηi=L(Gi) it holds that

    ηi=Tni(η)=L(Sn+γnGIn|I0=i).

    Since γnGIn0 almost surely, this implies that conditioned on I0=i the sequence Sn converges to ηi in distribution. In particular, if a fixed point exists, its i-th component has to be the distributional limit of Sn under I0=i, and thus, fixed points are unique. If a fixed point exists, then Tn(ν) converges to the unique fixed point η for any ν, since γnWIn0 almost surely holds for any random variables W1,,Wd. Now suppose that Tn(ν)η holds for some ν. The map T is continuous in the product topology on P(R)d and hence T(η)=T(limnTn(ν))=limnT(Tn(ν))=limnT(n+1)(ν)=η. So a fixed point exists which finished the equivalence (i)(iv).

    (iv)(iii). Supposing (iv) yields that conditioned on I0=i the sequence Sn converges in distribution. Applying Proposition 1 by considering l=(i,j) and Xlt=Rijt and Kt=(It,It+1) with P[K0=(i,j)]=1(i=i)pij yields the almost sure convergence. Since it holds conditioned for every i, it also holds unconditioned, hence (iii).

    (iii)(iv). Almost sure convergence holds also conditioned on I0=i, one can now apply Proposition 1.

    (iii)(ii). Again, we apply the equivalence of (ii) and (iii) from Proposition 1. We only note that in the situation l=(i,j) and Kt=(It,It+1) a state l=(i,j) is essential with respect to (Kt)t iff i is essential with respect to (It)t and pij>0. Also any state l with pij>0 is accessible when starting in I0=i. Hence, one obtains that the infinite sum (iii) in Theorem 2 converges almost surely iff E[log+|Rijt|]< for each essential pair (i,j), that is for each essential i and j with pij>0. Since we defined μij=δ0 for pij=0 this is equivalent to E[log+|Ri|]< for every essential i.

    Proposition 2. Let β>0. If E[exp(β|Xlt|)]< for each accessible state l then the infinite sum S=t=0γtXKt,t exists almost surely and it holds that E[exp(β|S|)]<.

    Proof. The infinite sum exists almost surely due to Proposition 1, as the existence of exponential moments for each accessible state clearly yields the existence of the logarithmic moment. Using the triangle inequality, that xexp(βx) is continuous non-decreasing and monotone convergence of expectations yields

    E[exp(β|S|)]limnE[exp(βnt=0γt|XKt,t|)].

    Conditioning on (K0,,Kn) and using independence of (Kt)t and (Xlt)lt yields

    E[exp(βnt=0γt|XKtt|)]E[nt=0E[exp(β|Xlt|)γt]|l=Kt].

    For each tN0, we have γt(0,1), and hence xx(γt) is a concave function on [0,). For each accessible state l we have

    E[exp(β|Xlt|)γt]E[exp(β|Xlt|)]γt<.

    Let c(β)=max{E[exp(β|Xl|)]|l[m]accessible}. We have c(β)<. One can bound the right hand side of (3.10) by

    nt=0c(β)γt=c(β)nt=0γt.

    Putting all together yields

    E[exp(β|S|)]limnc(β)nt=0γt=c(β)1/(1γ)<.

    Proof of Theorem 3. Consider Gi=tγtRIt,It+1,t starting with I0=i. With probability one (It,It+1) realizes to a pair (j,j) with ij and pjj>0.

    1) We have P[|Rj|K]=jpjjP[|Rjjt|K]. Hence, if P[|Rj|K]=1 for all j with ij, then P[|Rjjt|K]=1 for all (j,j) with ij and pjj>0. This yields P[|RIt,It+1,t|K|I0=i]=1 for every t and hence |Gi|tγt|RIt,It+1,t|K/(1γ) almost surely.

    2) Due to Theorem 2, we can consider the situation of Proposition 2 via l=(i,j), Xlt=Rijt and Kt=(It,It+1) having starting distribution P[K0=(i,j)]=pij. In this situation, a state l=(i,j) is accessible with respect to (Kt)t if and only if ii and pij>0. So, Gi has the same distribution as the almost sure limit t=0γtXKt,t and one obtains E[exp(β|Gi|)]<.

    3) We have E[|Rj|p]=jpjjE[|Rjjt|p] hence E[|Rj|p]< for all j with ij implies E[|Rjjt|p]< for all (j,j) with ij and pjj>0. Let ||Rjjt||p=E[|Rjjt|p]1/p and K=max(j,j)||Rjjt||p where the maximum runs over all relevant pairs (j,j), hence K<. Let the Markov chain (It)t be started in state I0=i. Then Gi,T=Tt=0γtRIt,It+1,tGi almost surely as T and the process (It,It+1)t only visits pairs (j,j) with ||Rjjt||pK<. Since ||||p is a norm, for T>T

    ||Gi,TGi,T||p=Tt=T+1γtRIt,It+1,tpTt=T+1||γtRIt,It+1,t||pKTt=T+1γt0asmin{T,T},

    hence (Gi,T)TN is a Cauchy sequence with respect to ||||p. The associated Lp-space is complete, hence there is a random variable ˜Gi in Lp, that is ||˜Gi||p<, such that ||Gi,T˜Gi||p0. This mode of convergence also implies Gi,T˜Gi in probability. The almost sure convergence Gi,TGi implies that Gi,TGi in probability. Limits with respect to convergence in probability are almost surely unique, hence ˜Gi=Gi with probability one and ||Gi||p=||˜Gi||p<.

    Our second main theorem is the result about regular variation of fixed points of T, see Theorem 4. As noted before, our proof makes use of the well-developed theory for multivariate fixed point equations

    Gd=JG+R,G   and  (J,R)independent. (3.4)

    In particular, a notion of regular variation for multivariate distributions, extending the one-dimensional case, has been explored in the context of equations such as (3.4) under various assumptions on the joint law L(J,R)P(Rd×d×Rd). Below, we use the coupling approach explained in Section 2.1 to apply some of the available results in proving Theorem 4.

    First, we introduce the multivariate notion of regular variation in random vectors and state some facts we use in the proof. We refer the reader to Appendix C in [16] for more details on the basic properties of multivariate regular variation. Let ˉR=R{,} be the extended real numbers and let ˉRd0=ˉRd{0}. A Rd-valued random vector X is called regularly varying if there exists a non-null Radon measure μ on ˉRd0, that does not charge infinite points, such that

    limxP[x1XC]P[|X|>x]=μ(C)for every measurable  μcontinuity set   CˉRd0, (3.5)

    note that || denotes the euclidean norm on Rd. The measure μ is called the limit measure of regular variation of the random vector X. The following are well-known facts about regularly varying random vectors, again we refer to Appendix C in [16]

    ● If μ is the limit measure of regular variation of a random vector X, then there is a unique α>0, called the index of regular variation, such that μ(tC)=tαμ(C) for every μ-continuity set C and t>0.

    ● If (3.5) is satisfied with index of regular variation α>0, then the set C={xRd||x|>1} is a μ-continuity set, and hence, because {x1XC}={|X|>x}, it holds μ(C)=1 and

    xP[|X|>x]

    is a regularly varying function with index α. Further, the set {xRd|wTx>1} is a μ-continuity set for each wRd,

    X is regularly varying with index α>0 if and only if there exists a distribution ΞP(Sd1) on the unit sphere Sd1={xRd||x|=1} such that for every t>0

    P[|X|>tx,X|X|]P[|X|>x]wtαΞ()as   x,

    where w is weak convergence of finite measures.

    The following theorem can be obtained easily as a special case of Theorem 4.4.24 in [16]; note that the empty product of (random) matrices is the identity matrix.

    Theorem 5. Let (R1,J1),,(Rd,Jd) be a coupling, R=(R1,,Rd)T and J=(Jij)ij with Jij=γ1(Ji=j), see Section 2.1. Suppose R is regularly varying with index α>0 and limit measure μR. Then the multivariate distributional fixed point equation (3.4) has a unique solution G which is regularly varying with index α. In particular, with (J(s))sN0 being iid copies of J, it holds

    limxP[x1GC]P[|R|>x]=t=0E[μR({xRd|(t1s=0J(s))xC})] (3.6)

    for every μR-continuity set C and

    P[|G|>x]cP[|R|>x]asx, (3.7)

    where c>0 equals the right hand side of (3.6) evaluated at C={xRd||x|>1}.

    Proof. Since C={xRd||x|>1} is a μ-continuity set that satisfies P[|G|>x]=P[x1|G|C] the asymptotic expression (3.7) follows from (3.6). The remaining part of the theorem is a direct consequence of Theorem 4.4.24 from [16] by noticing that the additional necessary assumptions stated in that theorem, E[|J|α]<1 and E[|J|α+δ]< for some α>0 and δ>0, are satisfied in our situation of interest as Jij=γ1(Ji=j) and hence |J|=γ<1.

    In Theorem 5, an arbitrary coupling (R1,J1),,(Rd,Jd) is considered and the (multivariate) regular variation of R=(R1,,Rd)T, which is an assumption there, transfers over to the (multivariate) regular variation of G=(G1,,Gd)T, whose marginal laws are the solutions to the distributional Bellman equations of interest in Theorem 4. Moreover, multivariate regular variation of G can be used to show univariate regular variation of G1,,Gd by testing (3.5) on sets of the form C={xRd|±eTjx>1} with ejRd being unit vectors. Note that Theorem 4 concerns solutions of the distributional Bellman operator T. Hence, the result of that theorem only depends on the marginal laws (L(R1,J1),,L(Rd,Jd)) and hence Theorem 5, which assumes presence of a coupling satisfying certain properties, is not directly applicable there. However, as explained in Section 2.1, results about coupled situations can become useful by choosing a convenient coupling L((R1,J1),,(Rd,Jd)) that makes certain arguments work out. In the following proof, this will be the case when an independence coupling is considered: independence of R1,,Rd together with the assumptions of Theorem 4 implies that the random vector R=(R1,,Rd)T is regularly varying in the multivariate sense, so Theorem 5 becomes applicable. We now present technical details:

    Proof of Theorem 4. Let (R1,J1),,(Rd,Jd) be independent. Let ejRd be the j-th unit vector and

    R=(R1,,Rd)T=dj=1Rjej,

    that is, we express R as a sum of independent random vectors. In the following, we rely on well-established results from the theory of (multivariate) regular variation which show that regular variation is transferred to sums of independent random variables (vectors) each having this property.

    First, we determine the asymptotic behavior of P[|R|>x] as x: let i[d] be fixed. Only those pairs (Rj,Jj) with ij are relevant for the distribution of Gi, hence we assume w.l.o.g. that ij holds for all j, that is for every j[d] there exists constants qj[0,1] and cj[0,) such that

    limxP[Rj>x]xαL(x)=qjcjandlimxP[Rj<x]xαL(x)=(1qj)cj, (3.8)

    where α>0 and L is a slowly varying function. This implies for every j

    cj=limxP[|Rj|>x]xαL(x)=limxP[R2j>x2]xαL(x)=limyP[R2j>y]yα/2˜L(y)with˜L(y)=L(y).

    The function ˜L is slowly varying which shows that R2j is regularly varying with index α/2 in case cj>0 and P[R2j>x]=o(xα/2) as x in case cj=0. In such a situation, and since R21,,R2d are independent by choice of coupling, standard results from the theory of regular variation apply, see Section 1.3.1 in [29], and show

    P[R21++R2d>y]P[R21>y]++P[R2d>y]asy

    and hence, by re-substituting x2 for y,

    P[R21++R2d>x2]P[R21>x2]++P[R2d>x2](c1++cd)xαL(x)asx,

    that is

    limxP[|R|>x]xαL(x)=c1++cd. (3.9)

    Next, we give an argument to show that R is regularly varying with index α (in the multivariate sense): let sgn(Rj){1,1} be the sign of Rj, so that |Rjej|=|Rj| and Rjej|Rjej|=sgn(Rj)ej in case Rj0. If cj>0 then Rjej is regularly varying at index α>0 in the multivariate sense, the spectral measure ΞjP(Sd1) is given by

    Ξj=qjδej+(1qj)δej=limxP[sgn(Rj)ej||Rj|>x].

    Applying standard arguments, see Lemma C.3.1 in [16], it is easily seen that R=dj=1Rjej (a sum of independent random vectors) is regularly varying at index α. Let μR be the associated measure. We are now able to apply Theorem 5 to the solution G=(G1,,Gd)T of the distributional equation (3.4).

    First, we evaluate the limit measure μR at some sets of interest: Combining (3.8) and (3.9) yields for every j[d]

    μR({xRd|eTjx>1})=limxP[Rj>x]P[|R|>x]=qjcjc1++cd.

    As in Theorem 5, let (J(s))sN0 be a sequence of iid copies of J. For every tN0 the entry of the random matrix t1s=0J(s) at position (i,j)[d]2 equals γt1(Iit=j), where (Iit)i[d],tN0 are [d]-valued random variables such that for every i[d] the sequence (Iit)t has the same law as (It)t under P[|I0=i].

    By Theorem 5 for every i[d] it holds that

    limxP[Gi>x]P[|R|>x]=t=0E[μR({xRd|eTi(t1s=0J(s))x>1})].

    Using μR(tC)=tαμR(C) for every tN0 it holds

    E[μR({xRd|eTi(t1s=0J(s))x>1})]=E[μR({xRd|γteTIitx>1})]=γαtj[d]P[It=j|I0=i]μR({xRd|eTjx>1})=γαtj[d]P[It=j|I0=i]qjcjc1++cd

    and hence

    limxP[Gi>x]P[|R|>x]=t=0γαtj[d]P[It=j|I0=i]qjcjc1++cd=11γαj[d]P[IN=j|I0=i]qjcjc1++cd,

    where NGeo(γα) is independent of (It)t. With wij=P[IN=j|I0=i] and (3.9) we have

    limxP[Gi>x]xαL(x)=j[d]wijqjcj1γα.

    The tail asymptotic of P[Gi<x]xαL(x) can be calculated the same way using

    μR({xRd|eTjx<1})=limxP[Rj<x]P[|R|>x]=(1qj)cjc1++cd.

    To see that Gi is regularly varying with index α note that P[|Gi|>x]=P[Gi>x]+P[Gi<x](dj=1wijcj)xαL(x) is a regularly varying function of index α and the balance equation hold:

    limxP[Gi>x]P[|Gi|>x]=dj=1qjwijcjdj=1wijcjandlimxP[Gi<x]P[|Gi|>x]=dj=1(1qj)wijcjdj=1wijcj=1dj=1qjwijcjdj=1wijcj.

    The authors have not used Artificial Intelligence (AI) tools in the creation of this article.

    We thank three anonymous referees for their careful reading, many constructive remarks and suggestions. The referees' comments led to a significant improvement of the manuscript.

    The first author was partially supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) - 502386356.

    The authors declare there is no conflicts of interest.



    [1] Anderson J (2011) EBOOK: Fundamentals of Aerodynamics (SI units), McGraw hill, 2011.
    [2] Wikipedia (2023) Shadowgraph Images of Re-entry Vehicles — Wikipedia, the free encyclopedia, https://commons.wikimedia.org/wiki/File: Shadowgraph_Images_of_Re-entry_Vehicles_-_GPN-2000-001938.jpg.
    [3] Kou J, Zhang W (2021) Data-driven modeling for unsteady aerodynamics and aeroelasticity. Prog Aerosp Sci 125: 100725. https://doi.org/10.1016/j.paerosci.2021.100725 doi: 10.1016/j.paerosci.2021.100725
    [4] Xiao T, Frank M (2021) Using neural networks to accelerate the solution of the boltzmann equation. J Computat Phys 443: 110521. https://doi.org/10.1016/j.jcp.2021.110521 doi: 10.1016/j.jcp.2021.110521
    [5] Xiao T, Frank M (2023) Relaxnet: A structure-preserving neural network to approximate the boltzmann collision operator. J Comput Phys 490: 112317. https://doi.org/10.1016/j.jcp.2023.112317 doi: 10.1016/j.jcp.2023.112317
    [6] Xiao T, Schotthfer S, Frank M (2023) Predicting continuum breakdown with deep neural networks. J Computat Phys 489: 112278. https://doi.org/10.1016/j.jcp.2023.112278 doi: 10.1016/j.jcp.2023.112278
    [7] Xiao T, Frank M (2021) A stochastic kinetic scheme for multi-scale flow transport with uncertainty quantification. J Computat Phys 437: 110337. https://doi.org/10.1016/j.jcp.2021.110337 doi: 10.1016/j.jcp.2021.110337
    [8] Xiao T, Frank M (2021) A stochastic kinetic scheme for multi-scale plasma transport with uncertainty quantification. J Computat Phys 432: 110139. https://doi.org/10.1016/j.jcp.2021.110139 doi: 10.1016/j.jcp.2021.110139
    [9] Kou J, Zhang W (2017) An improved criterion to select dominant modes from dynamic mode decomposition. European J Mech-B/Fluids 62: 109–129. https://doi.org/10.1016/j.euromechflu.2016.11.015 doi: 10.1016/j.euromechflu.2016.11.015
    [10] Kou J, Zhang W, Liu Y, et al. (2017) The lowest reynolds number of vortex-induced vibrations. Phys Fluids 29.https://doi.org/10.1063/1.4979966 doi: 10.1063/1.4979966
    [11] Kou J, Ning C, Zhang W (2022) Transfer learning for flow reconstruction based on multifidelity data. AIAA J 60: 5821–5842. https://doi.org/10.2514/1.J061647 doi: 10.2514/1.J061647
    [12] Kou J, Zhang W (2017) Multi-kernel neural networks for nonlinear unsteady aerodynamic reduced-order modeling. Aerosp Sci Technol 67: 309–326. https://doi.org/10.1016/j.ast.2017.04.017 doi: 10.1016/j.ast.2017.04.017
    [13] Li K, Kou J, Zhang W (2019) Deep neural network for unsteady aerodynamic and aeroelastic modeling across multiple mach numbers. Nonlinear Dynam 96: 2157–2177. https://doi.org/10.1007/s11071-019-04915-9 doi: 10.1007/s11071-019-04915-9
    [14] Kou J, Zhang W (2019) Multi-fidelity modeling framework for nonlinear unsteady aerodynamics of airfoils. Appl Math Model 76: 832–855. https://doi.org/10.1016/j.apm.2019.06.034 doi: 10.1016/j.apm.2019.06.034
    [15] Kou J, Zhang W (2017) Layered reduced-order models for nonlinear aerodynamics and aeroelasticity. J Fluid Struct 68: 174–193. https://doi.org/10.1016/j.jfluidstructs.2016.10.011 doi: 10.1016/j.jfluidstructs.2016.10.011
    [16] Kou J, Zhang W (2019) A hybrid reduced-order framework for complex aeroelastic simulations. Aerosp Sci Technol 84: 880–894. https://doi.org/10.1016/j.ast.2018.11.014 doi: 10.1016/j.ast.2018.11.014
    [17] Kou J, Botero-Bolívar L, Ballano R, Eet al. (2023) Aeroacoustic airfoil shape optimization enhanced by autoencoders. Expert Syst Appl 217 : 119513. https://doi.org/10.1016/j.eswa.2023.119513 doi: 10.1016/j.eswa.2023.119513
    [18] Sprent P (2019) Data driven statistical methods. Routledge.
    [19] Jordan MI, Mitchell TM (2015) Machine learning: Trends, perspectives, and prospects. Science 349: 255–260. https://doi.org/10.1126/science.aaa8415 doi: 10.1126/science.aaa8415
    [20] Brunton SL, Noack BR, Koumoutsakos P (2020) Machine learning for fluid mechanics. Ann Rev Fluid Mech 52 : 477–508. https://doi.org/10.1146/annurev-fluid-010719-060214 doi: 10.1146/annurev-fluid-010719-060214
    [21] Vinuesa R, Brunton SL (2022) Enhancing computational fluid dynamics with machine learning. Nall 2: 358–366. https://doi.org/10.1038/s43588-022-00264-7 doi: 10.1038/s43588-022-00264-7
    [22] Vinuesa R, Brunton SL, McKeon BJ (2023) The transformative potential of machine learning for experiments in fluid mechanics. Nat Rev Phys 5: 536–545. https://doi.org/10.1038/s42254-023-00622-y doi: 10.1038/s42254-023-00622-y
    [23] Zhang W, Kou J, Liu Y (2021) Prospect of artificial intelligence empowered fluid mechanics. Acta Aeronaut et Astronaut Sin 42: 524689.
    [24] Zhang W, Wang X, Kou J (2023) Prospects of multi-paradigm fusion methods for fluid mechanics research. Adv Mech 53: 433–467. https://doi.org/10.6052/1000-0992-22-050 doi: 10.6052/1000-0992-22-050
    [25] Sirovich L (1987) Turbulence and the dynamics of coherent structures. i. coherent structures. Q Appl Math 45: 561–571.
    [26] Taira K, Brunton SL, Dawson ST (2017) Ukeiley, Modal analysis of fluid flows: An overview. Aiaa J 55: 4013–4041. https://doi.org/10.2514/1.J056060 doi: 10.2514/1.J056060
    [27] Noack BR (2016) From snapshots to modal expansions–bridging low residuals and pure frequencies. J Fluid Mech 802: 1–4. https://doi.org/10.1017/jfm.2016.416 doi: 10.1017/jfm.2016.416
    [28] Berkooz G, Holmes P, Lumley JL (1993) The proper orthogonal decomposition in the analysis of turbulent flows. Ann Rev Fluid Mech 25: 539–575.
    [29] Schmid PJ (2010) Dynamic mode decomposition of numerical and experimental data. J Fluid Mecha 656: 5–28. https://doi.org/10.1017/S0022112010001217 doi: 10.1017/S0022112010001217
    [30] Schmid PJ (2022) Dynamic mode decomposition and its variants. Ann Rev Fluid Mech 54: 225–254. https://doi.org/10.1146/annurev-fluid-030121-015835 doi: 10.1146/annurev-fluid-030121-015835
    [31] Sieber M, Paschereit CO, Oberleithner K (2016) Spectral proper orthogonal decomposition. J Fluid Mech 792: 798–828. https://doi.org/10.1017/jfm.2016.103 doi: 10.1017/jfm.2016.103
    [32] Towne A, Schmidt OT, Colonius T (2018) Spectral proper orthogonal decomposition and its relationship to dynamic mode decomposition and resolvent analysis. J Fluid Mech 847: 821–867. https://doi.org/10.1017/jfm.2018.283 doi: 10.1017/jfm.2018.283
    [33] Kaiser E, Noack BR, Cordier L, et al. (2014) Cluster-based reduced-order modelling of a mixing layer. J Fluid Mech 754: 365–414. https://doi.org/10.1017/jfm.2014.355 doi: 10.1017/jfm.2014.355
    [34] Torres P, Clainche S Le, Vinuesa R (2021) On the experimental, numerical and data-driven methods to study urban flows. Energies 14: 1310. https://doi.org/10.3390/en14051310 doi: 10.3390/en14051310
    [35] Garuma GF (2018) Review of urban surface parameterizations for numerical climate models. Urban Clim 24: 830–851. https://doi.org/10.1016/j.uclim.2017.10.006 doi: 10.1016/j.uclim.2017.10.006
    [36] Masson V, Heldens W, Bocher E, et al. (2020) City-descriptive input data for urban climate models: Model requirements, data sources and challenges. Urban Clim 31: 100536. https://doi.org/10.1016/j.uclim.2019.100536 doi: 10.1016/j.uclim.2019.100536
    [37] Rowley CW, Mezić I, Bagheri S, et al. (2009) Spectral analysis of nonlinear flows. J Fluid Mech 641: 115–127. https://doi.org/10.1017/S0022112009992059 doi: 10.1017/S0022112009992059
    [38] Mezić I (2013) Analysis of fluid flows via spectral properties of the koopman operator. Ann Rev Fluid Mech 45: 357–378. https://doi.org/10.1146/annurev-fluid-011212-140652 doi: 10.1146/annurev-fluid-011212-140652
    [39] Franz T, Zimmermann R, Grtz S, et al. (2014) Interpolation-based reduced-order modelling for steady transonic flows via manifold learning. Int J Comput Fluid Dynam 28: 106–121. https://doi.org/10.1080/10618562.2014.918695 doi: 10.1080/10618562.2014.918695
    [40] Farzamnik E, Ianiro A, Discetti S, et al. (2023) From snapshots to manifolds–a tale of shear flows. J Fluid Mech 955: A34. https://doi.org/10.1017/jfm.2022.1039 doi: 10.1017/jfm.2022.1039
    [41] Milano M, Koumoutsakos P (2002) Neural network modeling for near wall turbulent flow. J Comput Phys 182: 1–26. https://doi.org/10.1006/jcph.2002.7146 doi: 10.1006/jcph.2002.7146
    [42] Eivazi H, Clainche S Le, Hoyas S (2022) Towards extraction of orthogonal and parsimonious non-linear modes from turbulent flows. Expert Syst Appl 202: 117038. https://doi.org/10.1016/j.eswa.2022.117038 doi: 10.1016/j.eswa.2022.117038
    [43] Fukami K, Fukagata K, Taira K (2019) Super-resolution reconstruction of turbulent flows with machine learning. J Fluid Mech 870: 106–120. https://doi.org/10.1017/jfm.2019.238 doi: 10.1017/jfm.2019.238
    [44] Wipf D, Nagarajan S (2007) A new view of automatic relevance determination. Adv Neur Inf Proc Syst 20.
    [45] Rudy SH, Sapsis TP (2021) Sparse methods for automatic relevance determination. Physica D 418: 132843. https://doi.org/10.1016/j.physd.2021.132843 doi: 10.1016/j.physd.2021.132843
    [46] Brunton SL, Proctor JL, Kutz JN (2016) Discovering governing equations from data by sparse identification of nonlinear dynamical systems. P Natl Acad Sci 113: 3932–3937. https://doi.org/10.1073/pnas.1517384113 doi: 10.1073/pnas.1517384113
    [47] Rudy SH, Brunton SL, Proctor JL, et al. (2017) Data-driven discovery of partial differential equations. Sci Adv 3: e1602614. https://doi.org/10.1126/sciadv.1602614 doi: 10.1126/sciadv.1602614
    [48] Quade M, Abel M, Nathan Kutz J, et al. (2018) Sparse identification of nonlinear dynamics for rapid model recovery. Chaos 28.https://doi.org/10.1063/1.5027470 doi: 10.1063/1.5027470
    [49] Li S, Kaiser E, Laima S, et al. (2019) Discovering time-varying aerodynamics of a prototype bridge by sparse identification of nonlinear dynamical systems. Physical Review E 100: 022220. https://doi.org/10.1103/PhysRevE.100.022220 doi: 10.1103/PhysRevE.100.022220
    [50] Sun C, Tian T, Zhu X et al. (2021) Sparse identification of nonlinear unsteady aerodynamics of the oscillating airfoil, Proceedings of the Institution of Mechanical Engineers. P G J Aerospace Eng 235: 809–824. https://doi.org/10.1177/0954410020959873 doi: 10.1177/0954410020959873
    [51] Ma T, Cui W, Gao T, et al. (2023) Data-based autonomously discovering method for nonlinear aerodynamic force of quasi-flat plate. Phys Fluids 35.https://doi.org/10.1063/5.0133526 doi: 10.1063/5.0133526
    [52] Kocijan J, Girard A, Banko B, et al. (2005) Dynamic systems identification with gaussian processes. Math Comput Model Dyn Syst 11: 411–424. https://doi.org/10.1080/13873950500068567 doi: 10.1080/13873950500068567
    [53] Wenk P, Gotovos A, Bauer S, et al. (2019) Fast gaussian process based gradient matching for parameter identification in systems of nonlinear odes, in: The 22nd International Conference on Artificial Intelligence and Statistics, PMLR, 1351–1360.
    [54] Raissi M, Karniadakis GE (2018) Hidden physics models: Machine learning of nonlinear partial differential equations. J Comput Phys 357: 125–141. https://doi.org/10.1016/j.jcp.2017.11.039 doi: 10.1016/j.jcp.2017.11.039
    [55] Grauer JA, Morelli EA (2015) A new formulation of the filter-error method for aerodynamic parameter estimation in turbulence. in: AIAA Atmospheric Flight Mechanics Conference, 2704. https://doi.org/10.2514/6.2015-2704
    [56] Qin T, Wu K, Xiu D (2019) Data driven governing equations approximation using deep neural networks. J Comput Phys 395: 620–635. https://doi.org/10.1016/j.jcp.2019.06.042 doi: 10.1016/j.jcp.2019.06.042
    [57] Long Z, Lu Y, Ma X, et al. (2018) Pde-net: Learning pdes from data, in: International Conference on Machine Learning PMLR, 3208–3216.
    [58] Cranmer M, Sanchez Gonzalez A, Battaglia P, et al. (2020) Discovering symbolic models from deep learning with inductive biases. Adv Neur Inform Process Syst 33: 17429–17442.
    [59] Berg J, Nystrm K (2019) Data-driven discovery of pdes in complex datasets. J Comput Phys 384: 239–252. https://doi.org/10.1016/j.jcp.2019.01.036 doi: 10.1016/j.jcp.2019.01.036
    [60] Xu H, Zhang W, Wang Y (2021) Explore missing flow dynamics by physics-informed deep learning: The parameterized governing systems. Phys Fluids 33.https://doi.org/10.1063/5.0062377 doi: 10.1063/5.0062377
    [61] Duraisamy K, Iaccarino G, Xiao H (2019) Turbulence modeling in the age of data. Ann Rev Fluid Mech 51: 357–377. https://doi.org/10.1146/annurev-fluid-010518-040547 doi: 10.1146/annurev-fluid-010518-040547
    [62] Ling J, Kurzawski A, Templeton J (2016) Reynolds averaged turbulence modelling using deep neural networks with embedded invariance. J Fluid Mech 807: 155–166. https://doi.org/10.1017/jfm.2016.615 doi: 10.1017/jfm.2016.615
    [63] Wang JX, Wu JL, Xiao H (2017) Physics-informed machine learning approach for reconstructing reynolds stress modeling discrepancies based on dns data. Phys Rev Fluids 2: 034603. https://doi.org/10.1103/PhysRevFluids.2.034603 doi: 10.1103/PhysRevFluids.2.034603
    [64] Parish EJ, Duraisamy K (2016) A paradigm for data-driven predictive modeling using field inversion and machine learning. J Comput Physics 305: 758–774. https://doi.org/10.1016/j.jcp.2015.11.012 doi: 10.1016/j.jcp.2015.11.012
    [65] Weatheritt J, Sandberg R (2016) A novel evolutionary algorithm applied to algebraic modifications of the rans stress–strain relationship. J Comput Phys 325: 22–37. https://doi.org/10.1016/j.jcp.2016.08.015 doi: 10.1016/j.jcp.2016.08.015
    [66] Zhao Y, Akolekar HD, Weatheritt J, et al. (2020) Rans turbulence model development using cfd-driven machine learning. J Comput Phys 411: 109413. https://doi.org/10.1016/j.jcp.2020.109413 doi: 10.1016/j.jcp.2020.109413
    [67] Zhu L, Zhang W, Kou J (2019) Machine learning methods for turbulence modeling in subsonic flows around airfoils. Phys Fluids 31.https://doi.org/10.1063/1.5061693 doi: 10.1063/1.5061693
    [68] Volpiani PS, Bernardini RF, Franceschini L (2022) Neural network-based eddy-viscosity correction for rans simulations of flows over bi-dimensional bumps. Int J Heat Fluid Fl 97: 109034. https://doi.org/10.1016/j.ijheatfluidflow.2022.109034 doi: 10.1016/j.ijheatfluidflow.2022.109034
    [69] Singh AP, Medida S, Duraisamy K (2017) Machine-learning-augmented predictive modeling of turbulent separated flows over airfoils. AIAA J 55: 2215–2227. https://doi.org/10.2514/1.J055595 doi: 10.2514/1.J055595
    [70] Maulik R, San O (2017) A neural network approach for the blind deconvolution of turbulent flows. J Fluid Mech 831: 151–181. https://doi.org/10.1017/jfm.2017.637 doi: 10.1017/jfm.2017.637
    [71] Maulik R, San O, Rasheed A, et al. (2019) Subgrid modelling for two-dimensional turbulence using neural networks. J Fluid Mech 858: 122–144. https://doi.org/10.1017/jfm.2018.770 doi: 10.1017/jfm.2018.770
    [72] Beck A, Flad D, Munz CD (2019) Deep neural networks for data-driven les closure models. J Comput Phys 398: 108910. https://doi.org/10.1016/j.jcp.2019.108910 doi: 10.1016/j.jcp.2019.108910
    [73] Yang X, Zafar S, Wang JX, et al. (2019) Predictive large-eddy-simulation wall modeling via physics-informed neural networks. Phys Rev Fluids 4: 034602. https://doi.org/10.1103/PhysRevFluids.4.034602 doi: 10.1103/PhysRevFluids.4.034602
    [74] Zhou Z, He G, Yang X (2021) Wall model based on neural networks for les of turbulent flows over periodic hills. Phys Rev Fluids 6: 054610. https://doi.org/10.1103/PhysRevFluids.6.054610 doi: 10.1103/PhysRevFluids.6.054610
    [75] Bae HJ, Koumoutsakos P (2022) Scientific multi-agent reinforcement learning for wall-models of turbulent flows. Nat Commun 13: 1443. https://doi.org/10.21203/rs.3.rs-573667/v1 doi: 10.21203/rs.3.rs-573667/v1
    [76] Dowell EH, Hall KC (2001) Modeling of fluid-structure interaction. Ann Rev Fluid Mech 33: 445–490. https://doi.org/10.1146/annurev.fluid.33.1.445 doi: 10.1146/annurev.fluid.33.1.445
    [77] Lucia DJ, Beran PS, Silva WA (2004) Reduced-order modeling: new approaches for computational physics. Prog Aerosp Sci 40: 51–117. https://doi.org/10.1016/j.paerosci.2003.12.001 doi: 10.1016/j.paerosci.2003.12.001
    [78] Theodorsen T (1949) General theory of aerodynamic instability and the mechanism of flutter. Tech rep.
    [79] Leishman JG, Beddoes T (1989) A semi-empirical model for dynamic stall. J Am Helicopter Soc 34: 3–17. https://doi.org/10.4050/JAHS.34.3.3 doi: 10.4050/JAHS.34.3.3
    [80] Dowell EH (2014) A modern course in aeroelasticity. 217, Springer.
    [81] Noack BR, Morzynski M, Tadmor G (2011) Reduced-order modelling for flow control. 528, Springer Science & Business Media. https://doi.org/10.1007/978-3-7091-0758-4
    [82] Gao C, Zhang W (2020) Transonic aeroelasticity: A new perspective from the fluid mode. Prog Aerosp Sci 113: 100596. https://doi.org/10.1016/j.paerosci.2019.100596 doi: 10.1016/j.paerosci.2019.100596
    [83] Ljung L (1998) System identification, in: Signal analysis and prediction, Springer, 163–173. https://doi.org/10.1007/978-1-4612-1768-8_11
    [84] Theofilis V (2011) Global linear instability. Ann Rev Fluid Mech 43: 319–352. https://doi.org/10.1146/annurev-fluid-122109-160705 doi: 10.1146/annurev-fluid-122109-160705
    [85] Juang JN, Pappa RS (1985) An eigensystem realization algorithm for modal parameter identification and model reduction. J Guid Control Dynam 8: 620–627.
    [86] Cowan TJ, Arena Jr AS, Gupta KK (2001) Accelerating computational fluid dynamics based aeroelastic predictions using system identification. J Aircraft 38: 81–87. https://doi.org/10.2514/2.2737 doi: 10.2514/2.2737
    [87] Leishman J (1988) Validation of approximate indicial aerodynamic functions for two-dimensional subsonic flow. J Aircraft 25: 914–922. https://doi.org/10.2514/3.45680 doi: 10.2514/3.45680
    [88] Skujins T, Cesnik CE (2014) Reduced-order modeling of unsteady aerodynamics across multiple mach regimes. J Aircraft 51: 1681–1704. https://doi.org/10.2514/1.C032222 doi: 10.2514/1.C032222
    [89] Silva W (2005) Identification of nonlinear aeroelastic systems based on the volterra theory: progress and opportunities. Nonlinear Dynam 39: 25–62. https://doi.org/10.1007/s11071-005-1907-z doi: 10.1007/s11071-005-1907-z
    [90] Glaz B, Liu L, Friedmann PP (2010) Reduced-order nonlinear unsteady aerodynamic modeling using a surrogate-based recurrence framework. AIAA J 48: 2418–2429. https://doi.org/10.2514/1.J050471 doi: 10.2514/1.J050471
    [91] Linse DJ, Stengel RF (1993) Identification of aerodynamic coefficients using computational neural networks. J Guid Control Dynam 16: 1018–1025.
    [92] Huang R, Hu H, Zhao Y (2014) Nonlinear reduced-order modeling for multiple-input/multiple-output aerodynamic systems. AIAA J 52: 1219–1231. https://doi.org/10.2514/1.J052323 doi: 10.2514/1.J052323
    [93] Benner P, Gugercin S, Willcox K (2015) A survey of projection-based model reduction methods for parametric dynamical systems. SIAM Rev 57: 483–531. https://doi.org/10.1137/13093271 doi: 10.1137/13093271
    [94] Noack BR, Afanasiev K, MORZYŃSKI M, et at. (2003) A hierarchy of low-dimensional models for the transient and post-transient cylinder wake. J Fluid Mech 497: 335–363. https://doi.org/10.1017/S0022112003006694 doi: 10.1017/S0022112003006694
    [95] Balajewicz MJ, Dowell EH, Noack BR (2013) Low-dimensional modelling of high-reynolds-number shear flows incorporating constraints from the navier–stokes equation. J Fluid Mech 729: 285–308. https://doi.org/10.1017/jfm.2013.278 doi: 10.1017/jfm.2013.278
    [96] Wei M, Rowley CW (2009) Low-dimensional models of a temporally evolving free shear layer. J Fluid Mech 618: 113–134. https://doi.org/10.1017/S0022112008004539 doi: 10.1017/S0022112008004539
    [97] Deng N, Noack BR, Morzyński M, et al. (2020) Low-order model for successive bifurcations of the fluidic pinball. J Fluid Mech 884: A37. https://doi.org/10.1017/jfm.2019.959 doi: 10.1017/jfm.2019.959
    [98] Yu J, Yan C, Guo M (2019) Non-intrusive reduced-order modeling for fluid problems: A brief review. P I Mech Eng G J Aerosp Eng 233: 5896–5912. https://doi.org/10.1016/j.jcp.2019.07.053 doi: 10.1016/j.jcp.2019.07.053
    [99] Xiao D, Fang F, Buchan AG, et al. (2015) Non-intrusive reduced order modelling of the navier–stokes equations. Comput Method Appl Mech Eng 293: 522–541. https://doi.org/10.1016/j.cma.2015.05.015 doi: 10.1016/j.cma.2015.05.015
    [100] Hesthaven JS, Ubbiali S (2018) Non-intrusive reduced order modeling of nonlinear problems using neural networks. J Comput Phys 363: 55–78. https://doi.org/10.1016/j.jcp.2018.02.037 doi: 10.1016/j.jcp.2018.02.037
    [101] Wang Z, Xiao D, Fang F, et al. (2018) Model identification of reduced order fluid dynamics systems using deep learning. Int J Numer Method Fluids 86: 255–268. https://doi.org/10.1002/fld.4416 doi: 10.1002/fld.4416
    [102] Xu J, Duraisamy K (2020) Multi-level convolutional autoencoder networks for parametric prediction of spatio-temporal dynamics. Comput Method Appl Mech Eng 372: 113379. https://doi.org/10.1016/j.cma.2020.113379 doi: 10.1016/j.cma.2020.113379
    [103] Kramer B, Peherstorfer B, Willcox KE (2024) Learning nonlinear reduced models from data with operator inference. Ann Rev Fluid Mech 56.https://doi.org/10.1146/annurev-fluid-121021-025220 doi: 10.1146/annurev-fluid-121021-025220
    [104] Kato H, Yoshizawa A, Ueno G, et al. (2015) A data assimilation methodology for reconstructing turbulent flows around aircraft. J Computat Phys 283: 559–581. https://doi.org/10.1016/j.jcp.2014.12.013 doi: 10.1016/j.jcp.2014.12.013
    [105] Mons V, Chassaing JC, Gomez T, et al. (2016) Reconstruction of unsteady viscous flows using data assimilation schemes. J Comput Phys 316: 255–280. https://doi.org/10.1016/j.jcp.2016.04.022 doi: 10.1016/j.jcp.2016.04.022
    [106] Han ZH, Grtz S (2012) Hierarchical kriging model for variable-fidelity surrogate modeling. AIAA J 50: 1885–1896. https://doi.org/10.2514/1.J051354 doi: 10.2514/1.J051354
    [107] Wang X, Kou J, Zhang W, et al. (2022) Incorporating physical models for dynamic stall prediction based on machine learning. AIAA J 60: 4428–4439. https://doi.org/10.2514/1.J061210 doi: 10.2514/1.J061210
    [108] Li K, Kou J, Zhang W (2022) Deep learning for multifidelity aerodynamic distribution modeling from experimental and simulation data. AIAA J 60: 4413–4427. https://doi.org/10.2514/1.J061330 doi: 10.2514/1.J061330
    [109] Ribeiro MD, Stradtner M, Bekemeyer P (2023) Unsteady reduced order model with neural networks and flight-physics-based regularization for aerodynamic applications. Comput Fluids 105949.https://doi.org/10.1016/j.compfluid.2023.105949 doi: 10.1016/j.compfluid.2023.105949
    [110] Raissi M, Yazdani A, Karniadakis GE (2020) Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations. Science 367: 1026–1030. https://doi.org/10.1126/science.aaw4741 doi: 10.1126/science.aaw4741
    [111] Thuerey N, Weißenow K, Prantl L, et al. (2020) Deep learning methods for reynolds-averaged navier–stokes simulations of airfoil flows. AIAA J 58: 25–36. https://doi.org/10.2514/1.J058291 doi: 10.2514/1.J058291
    [112] Kochkov D, Smith JA, Alieva A, et al., (2021) Machine learning–accelerated computational fluid dynamics, P Natl Acad Sci 118: e2101784118. https://doi.org/10.1073/pnas.2101784118 doi: 10.1073/pnas.2101784118
    [113] Raissi M, Perdikaris P, Karniadakis GE, (2019) Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J Comput Phys 378: 686–707. https://doi.org/10.1016/j.jcp.2018.10.045 doi: 10.1016/j.jcp.2018.10.045
    [114] Pang G, Lu L, Karniadakis GE (2019) fpinns: Fractional physics-informed neural networks. SIAM J Sci Comput 41: A2603–A2626. https://doi.org/10.1137/18M1229845 doi: 10.1137/18M1229845
    [115] Raissi M, Wang Z, Triantafyllou MS, et al. (2019) Deep learning of vortex-induced vibrations. J Fluid Mech 861: 119–137. https://doi.org/10.31224/osf.io/fnwjy doi: 10.31224/osf.io/fnwjy
    [116] Sun L, Gao H, Pan S, et al. (2020) Surrogate modeling for fluid flows based on physics-constrained deep learning without simulation data. Comput Method Appl M Eng 361: 112732. https://doi.org/10.1016/j.cma.2019.112732 doi: 10.1016/j.cma.2019.112732
    [117] Jin X, Cai S, Li H, et al. (2020) Nsfnets (navier-stokes flow nets): Physics-informed neural networks for the incompressible navier-stokes equations. arXiv preprint arXiv: 2003.06496.
    [118] Wang S, Yu X, Perdikaris P (2020) When and why pinns fail to train: A neural tangent kernel perspective. arXiv preprint arXiv:2007.14527.
    [119] Wang S, Teng Y, Perdikaris P (2020) Understanding and mitigating gradient pathologies in physics-informed neural networks. arXiv preprint arXiv:2001.04536.
    [120] Chen RT, Rubanova Y, Bettencourt J, et al. (2018) Neural ordinary differential equations, in: Advances in Neural Information Processing Systems, 6571–6583.
    [121] Wang S, Wang H, Perdikaris P (2021) On the eigenvector bias of fourier feature networks: From regression to solving multi-scale pdes with physics-informed neural networks. Comput Method Appl Mech Eng 384: 113938. https://doi.org/10.1016/j.cma.2021.113938 doi: 10.1016/j.cma.2021.113938
    [122] Lu L, Jin P, Pang G, et al. (2021) Learning nonlinear operators via deeponet based on the universal approximation theorem of operators. Nat Mach Intell 3: 218–229. https://doi.org/10.1038/s42256-021-00302-5 doi: 10.1038/s42256-021-00302-5
    [123] Xu W, Lu Y, Wang L (2023) Transfer learning enhanced deeponet for long-time prediction of evolution equations, in: Proceedings of the AAAI Conference on Artificial Intelligence 37: 10629–10636. https://doi.org/10.1609/aaai.v37i9.26262
    [124] He J, Kushwaha S, Park J, et al. (2024) Sequential deep operator networks (s-deeponet) for predicting full-field solutions under time-dependent loads. Eng Appl Artif Intell 127: 107258. https://doi.org/10.1016/j.engappai.2023.107258 doi: 10.1016/j.engappai.2023.107258
    [125] Li Z, Kovachki N, Azizzadenesheli K, et al. (2020) Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895.
    [126] Kovachki N, Lanthaler S, Mishra S (2021) On universal approximation and error bounds for fourier neural operators. J Mach Learn Res 22: 13237–13312.
    [127] Guibas J, Mardani M, Li Z, et al. (2021) Adaptive fourier neural operators: Efficient token mixers for transformers. arXiv preprint arXiv:2111.13587.
    [128] Li Z, Huang D, Liu B, et al. (2022) Fourier neural operator with learned deformations for pdes on general geometries. arXiv preprint arXiv:2207.05209.
    [129] Lu L, Meng X, Cai S, et al. (2022) A comprehensive and fair comparison of two neural operators (with practical extensions) based on fair data. Comput Method Appl Mech Eng 393: 114778. https://doi.org/10.1016/j.cma.2022.114778 doi: 10.1016/j.cma.2022.114778
    [130] Di Leoni PC, Lu L, Meneveau C, et al. (2023) Neural operator prediction of linear instability waves in high-speed boundary layers. J Computat Phys 474: 111793. https://doi.org/10.1016/j.jcp.2022.111793 doi: 10.1016/j.jcp.2022.111793
    [131] Mao Z, Lu L, Marxen O, et al. (2021) Deepm & mnet for hypersonics: Predicting the coupled flow and finite-rate chemistry behind a normal shock using neural-network approximation of operators. J Comput Phys 447: 110698. https://doi.org/10.1016/j.jcp.2021.110698 doi: 10.1016/j.jcp.2021.110698
    [132] Yin M, Ban E, Rego BV, et al. (2022) Simulating progressive intramural damage leading to aortic dissection using deeponet: an operator–regression neural network. J R Soc Interface 19: 20210670. https://doi.org/10.1098/rsif.2021.0670 doi: 10.1098/rsif.2021.0670
    [133] Goswami S, Yin M, Yu Y, et al. (2022) A physics-informed variational deeponet for predicting crack path in quasi-brittle materials. Comput Meth Appl Mech Eng 391: 114587. https://doi.org/10.1016/j.cma.2022.114587 doi: 10.1016/j.cma.2022.114587
    [134] Zhu M, Zhang H, Jiao A, et al. (2023) Reliable extrapolation of deep neural operators informed by physics or sparse observations. Comput Meth Appl Mech Eng 412: 116064. https://doi.org/10.1016/j.cma.2023.116064 doi: 10.1016/j.cma.2023.116064
    [135] Sekar V, Jiang Q, Shu C, et al. (2019) Fast flow field prediction over airfoils using deep learning approach. Phys Fluids 31.https://doi.org/10.1063/1.5094943 doi: 10.1063/1.5094943
    [136] Han R, Wang Y, Zhang Y, et al. (2019) A novel spatial-temporal prediction method for unsteady wake flows based on hybrid deep neural network. Phys Fluids 31.https://doi.org/10.1063/1.5127247 doi: 10.1063/1.5127247
    [137] Bhatnagar S, Afshar Y, Pan S, et al. (2019) Prediction of aerodynamic flow fields using convolutional neural networks. Comput Mech 64: 525–545. https://doi.org/10.1007/s00466-019-01740-0 doi: 10.1007/s00466-019-01740-0
    [138] Hui X, Bai J, Wang H, et al. (2020) Fast pressure distribution prediction of airfoils using deep learning. Aerosp Sci Technol 105: 105949. https://doi.org/10.1016/j.ast.2020.105949 doi: 10.1016/j.ast.2020.105949
    [139] Wang J, He C, Li R, et al. (2021) Flow field prediction of supercritical airfoils via variational autoencoder based deep learning framework. Phys Fluids 33.https://doi.org/10.1063/5.0053979 doi: 10.1063/5.0053979
    [140] Hu JW, Zhang WW (2022) Mesh-conv: Convolution operator with mesh resolution independence for flow field modeling. J Comput Phys 452: 110896. https://doi.org/10.1016/j.jcp.2021.110896 doi: 10.1016/j.jcp.2021.110896
    [141] Zuo K, Ye Z, Zhang W, et al. (2023) Fast aerodynamics prediction of laminar airfoils based on deep attention network. Phys Fluids 35.https://doi.org/10.1063/5.0140545 doi: 10.1063/5.0140545
    [142] Hirsch C (2007) Numerical computation of internal and external flows: The fundamentals of computational fluid dynamics, Elsevier.
    [143] Chen X, Li T, Wan Q (2022) Mgnet: a novel differential mesh generation method based on unsupervised neural networks. Eng Comput 38: 4409–4421. https://doi.org/10.1007/s00366-022-01632-7 doi: 10.1007/s00366-022-01632-7
    [144] Gao W, Wang A, Metzer G, et al. (2022) Tetgan: A convolutional neural network for tetrahedral mesh generation. arXiv preprint arXiv:2210.05735.
    [145] Zhang Z, Wang Y, Jimack PK, et al. (2020) Meshingnet: A new mesh generation method based on deep learning, in: International Conference on Computational Science, Springer, 186–198. https://doi.org/10.1007/978-3-030-50420-5_14
    [146] Chen X, Liu J, PangY, et al. (2020) Developing a new mesh quality evaluation method based on convolutional neural network. Eng Appl Comput Fluid Mech 14: 391–400. https://doi.org/10.1080/19942060.2020.1720820 doi: 10.1080/19942060.2020.1720820
    [147] Lu Y, Zhong A, Li Q, et al. (2018) Beyond finite layer neural networks: Bridging deep architectures and numerical differential equations, in: International Conference on Machine Learning, PMLR, 3276–3285.
    [148] Guo L, Li M, Xu S, et al. (2019) Study on a recurrent convolutional neural network based fdtd method, in: 2019 International Applied Computational Electromagnetics Society Symposium-China (ACES), 1: IEEE, 1–2. https://doi.org/10.23919/ACES48530.2019.9060707
    [149] Guo X, Li W, orio FI (2016) Convolutional neural networks for steady flow approximation, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 481–490. https://doi.org/10.1145/2939672.2939738
    [150] de Lara FM, Ferrer E (2022) Accelerating high order discontinuous galerkin solvers using neural networks: 1d burgers' equation. Comput Fluids 235: 105274. https://doi.org/10.1016/j.compfluid.2021.105274 doi: 10.1016/j.compfluid.2021.105274
    [151] de Lara FM, Ferrer E (2023) Accelerating high order discontinuous galerkin solvers using neural networks: 3d compressible navier-stokes equations. J Comput Phys 112253.https://doi.org/10.1016/j.jcp.2023.112253 doi: 10.1016/j.jcp.2023.112253
    [152] Belbute-Peres FDA, Economon T, Kolter Z (2020) Combining differentiable pde solvers and graph neural networks for fluid flow prediction, in: International Conference on Machine Learning, PMLR, 2402–2411.
    [153] Liu W, Yagoubi M, Schoenauer M (2021) Multi-resolution graph neural networks for pde approximation, in: Artificial Neural Networks and Machine Learning–ICANN 2021: 30th International Conference on Artificial Neural Networks, Bratislava, Slovakia, 14–17, Proceedings, Part Ⅲ 30, Springer, 151–163. https://doi.org/10.1007/978-3-030-86365-4_13
    [154] Chen J, Hachem E, Viquerat J (2021) Graph neural networks for laminar flow prediction around random two-dimensional shapes. Phys Fluids 33.https://doi.org/10.1063/5.0064108 doi: 10.1063/5.0064108
    [155] Wang ZJ, Fidkowski K, Abgrall R, et al. (2013) High-order cfd methods: current status and perspective. Int J Numer Meth Fluids 72: 811–845. https://doi.org/10.1002/fld.3767 doi: 10.1002/fld.3767
    [156] Deng X, Mao M, Tu G, et al. (2012) High-order and high accurate cfd methods and their applications for complex grid problems. Commun Computat Phys 11: 1081–1102. https://doi.org/10.4208/cicp.100510.150511s doi: 10.4208/cicp.100510.150511s
    [157] Xiao T (2021) A flux reconstruction kinetic scheme for the boltzmann equation. J Comput Phys 447: 110689. https://doi.org/10.1016/j.jcp.2021.110689 doi: 10.1016/j.jcp.2021.110689
    [158] Beck AD, Zeifang J, Schwarz A (2020) A neural network based shock detection and localization approach for discontinuous galerkin methods. J Comput Phys 423: 109824. https://doi.org/10.1016/j.jcp.2020.109824 doi: 10.1016/j.jcp.2020.109824
    [159] Sun Z (2020) Convolution neural network shock detector for numerical solution of conservation laws. Commun Comput Phys 28.https://doi.org/10.4208/cicp.OA-2020-0199 doi: 10.4208/cicp.OA-2020-0199
    [160] Morgan NR, Tokareva S, Liu X (2020) A machine learning approach for detecting shocks with high-order hydrodynamic methods. AIAA Scitech 2020 Forum. https://doi.org/10.2514/6.2020-2024
    [161] Yin X, Chen X, Chang H (2019) Experimental study of atmospheric turbulence detection using an orbital angular momentum beam via a convolutional neural network. IEEE Access 7: 184235–184241. https://doi.org/10.1109/ACCESS.2019.2960544 doi: 10.1109/ACCESS.2019.2960544
    [162] Li J, Zhang M, Wang D, et al. (2018) Joint atmospheric turbulence detection and adaptive demodulation technique using the cnn for the oam-fso communication. Opt Express 26: 10494–10508. https://doi.org/10.1364/OE.26.010494 doi: 10.1364/OE.26.010494
    [163] Levermore CD (1996) Moment closure hierarchies for kinetic theories. J Stat Phys 83: 1021–1065. https://doi.org/10.1007/BF02179552 doi: 10.1007/BF02179552
    [164] Sadr M, Torrilhon M, Gorji MH (2020) Gaussian process regression for maximum entropy distribution. J Comput Phys 418: 109644. https://doi.org/10.1016/j.jcp.2020.109644 doi: 10.1016/j.jcp.2020.109644
    [165] Schotthfer S, Xiao T, Frank M, et al. (2021) A structure-preserving surrogate model for the closure of the moment system of the boltzmann equation using convex deep neural networks, in: AIAA Aviation 2021 Forum 2895.https://doi.org/10.2514/6.2021-2895
    [166] Schotthfer S, Xiao T, Frank M, et al. (2022) Structure preserving neural networks: A case study in the entropy closure of the boltzmann equation, in: Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, USA, 17–23.
    [167] Kusch J, Schotthfer S, Stammer P, et al. (2022) Kit-rt: An extendable framework for radiative transfer and therapy. arXiv preprint arXiv:2205.08417. https://doi.org/10.1145/3630001
    [168] Mohamed L, Calderhead B, Filippone M, et al. (2012) Population mcmc methods for history matching and uncertainty quantification. Comput Geosci 16: 423–436. https://doi.org/10.1007/s10596-011-9232-8 doi: 10.1007/s10596-011-9232-8
    [169] Rafiee J, Reynolds AC (2018) A two-level mcmc based on the distributed gauss-newton method for uncertainty quantification, in: ECMOR XVI-16th European Conference on the Mathematics of Oil Recovery. European Association of Geoscientists & Engineers, 1–33. https://doi.org/10.3997/2214-4609.201802140
    [170] Mohamed L, Christie M, Demyanov V (2010) Comparison of stochastic sampling algorithms for uncertainty quantification. SPE J 15: 31–38. https://doi.org/10.2118/119139-MS doi: 10.2118/119139-MS
    [171] Zhang J (2021) Modern monte carlo methods for efficient uncertainty quantification and propagation: A survey, Wiley Interdisciplinary Reviews. Comput Stat 13: e1539. https://doi.org/10.1002/wics.1539 doi: 10.1002/wics.1539
    [172] Izmailov P, Vikram S, Hoffman MD, et al. (2021) What are bayesian neural network posteriors really like? in: International conference on machine learning, PMLR, 4629–4640.
    [173] Jospin LV, Laga H, Boussaid F (2022) Hands-on bayesian neural networks—a tutorial for deep learning users. IEEE Comput Intell Mag 17: 29–48. https://doi.org/10.1109/MCI.2022.3155327 doi: 10.1109/MCI.2022.3155327
    [174] Goan E, Fookes C (2020) Bayesian neural networks: An introduction and survey, Case Studies in Applied Bayesian Data Science, in: CIRM Jean-Morlet Chair, Fall 2018, 45–87. https://doi.org/10.1007/978-3-030-42553-1_3
    [175] Fortuin V, Garriga-Alonso A, Ober SW, et al. (2021) Bayesian neural network priors revisited. arXiv preprint arXiv:2102.06571.
    [176] Springenberg JT, Klein A, Falkner S, et al. (2016) Bayesian optimization with robust bayesian neural networks. Adv Neur Inf Process Sys 29.
    [177] Chan S, Elsheikh AH (2018) A machine learning approach for efficient uncertainty quantification using multiscale methods. J Comput Phys 354: 493–511. https://doi.org/10.1016/j.jcp.2017.10.034 doi: 10.1016/j.jcp.2017.10.034
    [178] Rahaman R (2021) Uncertainty quantification and deep ensembles. Adv Neur Inf Process Syst 34: 20063–20075.
    [179] Lakshminarayanan B, Pritzel A, Blundell C (2017) Simple and scalable predictive uncertainty estimation using deep ensembles. Adv Neur Inf Process Syst 30.
    [180] Egele R, Maulik R, Raghavan K, et al. (2022) Autodeuq: Automated deep ensemble with uncertainty quantification, in: 2022 26th International Conference on Pattern Recognition (ICPR), IEEE, 1908–1914. https://doi.org/10.1109/ICPR56361.2022.9956231
    [181] Moya C, Zhang S, Lin G, et al. (2023) Deeponet-grid-uq: A trustworthy deep operator framework for predicting the power grid's post-fault trajectories. Neurocomputing 535: 166–182. https://doi.org/10.1016/j.neucom.2023.03.015 doi: 10.1016/j.neucom.2023.03.015
    [182] Zhang J, Zhang S, Lin G (2022) Multiauto-deeponet: A multi-resolution autoencoder deeponet for nonlinear dimension reduction, uncertainty quantification and operator learning of forward and inverse stochastic problems. arXiv preprint arXiv:2204.03193.
    [183] Silva WA, Bartels RE (2004) Development of reduced-order models for aeroelastic analysis and flutter prediction using the cfl3dv6. 0 code. J Fluids Struct 19: 729–745. https://doi.org/10.1016/j.jfluidstructs.2004.03.004 doi: 10.1016/j.jfluidstructs.2004.03.004
    [184] Amsallem D, Farhat C (2008) Interpolation method for adapting reduced-order models and application to aeroelasticity. AIAA J 46: 1803–1813. https://doi.org/10.2514/1.35374 doi: 10.2514/1.35374
    [185] Winter M, Breitsamter C (2016) Neurofuzzy-model-based unsteady aerodynamic computations across varying freestream conditions. Aiaa J 54: 2705–2720. https://doi.org/10.2514/1.J054892 doi: 10.2514/1.J054892
    [186] Zhang W, Li X, Ye Z (2015) Mechanism of frequency lock-in in vortex-induced vibrations at low reynolds numbers. J Fluid Mech 783: 72–102. https://doi.org/10.1017/jfm.2015.548 doi: 10.1017/jfm.2015.548
    [187] Yao W, Jaiman R (2017) Model reduction and mechanism for the vortex-induced vibrations of bluff bodies. J Fluid Mech 827 : 357–393. https://doi.org/10.1017/jfm.2017.525 doi: 10.1017/jfm.2017.525
    [188] Gao C, Zhang W, Li X, (2017) Mechanism of frequency lock-in in transonic buffeting flow. J Fluid Mech 818: 528–561. https://doi.org/10.1017/jfm.2017.120 doi: 10.1017/jfm.2017.120
    [189] Balajewicz M, Dowell E (2012) Reduced-order modeling of flutter and limit-cycle oscillations using the sparse volterra series. J Aircraft 49: 1803–1812. https://doi.org/10.2514/1.C031637 doi: 10.2514/1.C031637
    [190] Zhang W, Wang B, Ye Z, wt al. (2012) Efficient method for limit cycle flutter analysis based on nonlinear aerodynamic reduced-order models. AIAA J 50: 1019–1028. https://doi.org/10.2514/1.J050581 doi: 10.2514/1.J050581
    [191] Mannarino A, Mantegazza P (2014) Nonlinear aeroelastic reduced order modeling by recurrent neural networks. J Fluids Struct 48: 103–121. https://doi.org/10.1016/j.jfluidstructs.2014.02.016 doi: 10.1016/j.jfluidstructs.2014.02.016
    [192] Chen G, Zuo Y, Sun J, et al. (2012) Support-vector-machine-based reduced-order model for limit cycle oscillation prediction of nonlinear aeroelastic system. Math Probl Eng 2012.
    [193] Liu Z, Han R, Zhang M, et al. (2022) An enhanced hybrid deep neural network reduced-order model for transonic buffet flow prediction. Aerosp Sci Technol 126: 107636. https://doi.org/10.1016/j.ast.2022.107636 doi: 10.1016/j.ast.2022.107636
    [194] Huang R, Li H, Hu H, et al. (2015) Open/closed-loop aeroservoelastic predictions via nonlinear, reduced-order aerodynamic models. AIAA J 53: 1812–1824. https://doi.org/10.2514/1.J053424 doi: 10.2514/1.J053424
    [195] Mannarino A, Dowell EH (2015) Reduced-order models for computational-fluid-dynamics-based nonlinear aeroelastic problems. Aiaa J 53: 2671–2685. https://doi.org/10.2514/1.J053775 doi: 10.2514/1.J053775
    [196] Huang R, Liu H, Yang Z, et al. (2018) Nonlinear reduced-order models for transonic aeroelastic and aeroservoelastic problems. AIAA J 56: 3718–3731. https://doi.org/10.2514/1.J056760 doi: 10.2514/1.J056760
    [197] Yang Z, Huang R, Liu H, et al. (2020) An improved nonlinear reduced-order modeling for transonic aeroelastic systems. J Fluids Struct 94: 102926. https://doi.org/10.1016/j.jfluidstructs.2020.102926 doi: 10.1016/j.jfluidstructs.2020.102926
    [198] Jameson A (2003) Aerodynamic shape optimization using the adjoint method, Lectures at the Von Karman Institute, Brussels.
    [199] Martins JR, Lambe AB (2013) Multidisciplinary design optimization: a survey of architectures. AIAA J 51: 2049–2075. https://doi.org/10.2514/1.J051895 doi: 10.2514/1.J051895
    [200] Jameson A (1988) Aerodynamic design via control theory. J Sci Comput 3: 233–260. https://doi.org/10.1007/BF01061285 doi: 10.1007/BF01061285
    [201] Queipo NV, Haftka RT, Shyy W, et al. (2005) Surrogate-based analysis and optimization. Prog Aerosp Sci 41: 1–28. https://doi.org/10.1016/j.paerosci.2005.02.001 doi: 10.1016/j.paerosci.2005.02.001
    [202] Han ZH, Zhang KS (2012) Surrogate-based optimization. Real-world applications of genetic algorithms. InTech 343: 343–362. https://doi.org/10.5772/36125 doi: 10.5772/36125
    [203] Yondo R, Andrés E, Valero E (2018) A review on design of experiments and surrogate models in aircraft real-time and many-query aerodynamic analyses. Prog Aerosp Sci 96: 23–61. https://doi.org/10.1016/j.paerosci.2017.11.003 doi: 10.1016/j.paerosci.2017.11.003
    [204] Li J, Du X, Martins JR (2022) Machine learning in aerodynamic shape optimization. Prog Aerosp Sci 134: 100849. https://doi.org/10.1016/j.paerosci.2022.100849 doi: 10.1016/j.paerosci.2022.100849
    [205] Xu M, Song S, Sun X, et al. (2021) Machine learning for adjoint vector in aerodynamic shape optimization. Acta Mech Sinica 37: 1416–1432. https://doi.org/10.1007/s10409-021-01119-6 doi: 10.1007/s10409-021-01119-6
    [206] Xu M, Song S, Sun X, et al. (2021) A convolutional strategy on unstructured mesh for the adjoint vector modeling. Phys Fluids 33.
    [207] Wu X, Zhang W, Peng X, et al. (2019) Benchmark aerodynamic shape optimization with the pod-based cst airfoil parametric method. Aerosp Sci Technol 84: 632–640. https://doi.org/10.1016/j.ast.2018.08.005 doi: 10.1016/j.ast.2018.08.005
    [208] Sun G, Wang S (2019) A review of the artificial neural network surrogate modeling in aerodynamic design. P I Mech Eng G J Aer 233: 5863–5872. https://doi.org/10.1177/0954410019864485 doi: 10.1177/0954410019864485
    [209] Chen LW, Cakal BA, Hu X, et al. (2021) Numerical investigation of minimum drag profiles in laminar flow using deep learning surrogates. J Fluid Mech 919: A34.
    [210] Yan X, Zhu J, Kuang M, et al. (2019) Aerodynamic shape optimization using a novel optimizer based on machine learning techniques. Aerosp Sci Technol 86: 826–835. https://doi.org/10.1016/j.ast.2019.02.003 doi: 10.1016/j.ast.2019.02.003
    [211] Li R, Zhang Y, Chen H (2021) Learning the aerodynamic design of supercritical airfoils through deep reinforcement learning. AIAA J 59: 3988–4001. https://doi.org/10.2514/1.J060189 doi: 10.2514/1.J060189
    [212] Gad-el Hak M (1989) Flow control. https://doi.org/10.1115/1.3152376
    [213] Bewley TR (2001) Flow control: new challenges for a new renaissance. Prog Aerosp Sci 37: 21–58. https://doi.org/10.1001/archneur.58.1.21 doi: 10.1001/archneur.58.1.21
    [214] Brunton SL, Noack BR (2015) Closed-loop turbulence control: Progress and challenges. Appl Mech Rev 67: 050801.
    [215] Fouatih OM, Medale M, Imine O, et al. (2016) Design optimization of the aerodynamic passive flow control on naca 4415 airfoil using vortex generators. Eur J Mech-B/Fluids 56: 82–96. https://doi.org/10.1016/j.euromechflu.2015.11.006 doi: 10.1016/j.euromechflu.2015.11.006
    [216] Kral LD (2000) Active flow control technology. ASME Fluids Eng Tech Brief 1–28.
    [217] Collis SS, Joslin RD, Seifert A, et al. (2004) Issues in active flow control: theory, control, simulation, and experiment. Prog Aerosp Sci 40: 237–289. https://doi.org/10.1016/j.paerosci.2004.06.001 doi: 10.1016/j.paerosci.2004.06.001
    [218] Cattafesta Ⅲ LN, Sheplak M (2011) Actuators for active flow control. Annu Rev Fluid Mech 43: 247–272. https://doi.org/10.1146/annurev-fluid-122109-160634 doi: 10.1146/annurev-fluid-122109-160634
    [219] Lee C, Hong G, Ha Q, et al. (2003) A piezoelectrically actuated micro synthetic jet for active flow control. Sensors Actuat A Phys 108: 168–174. https://doi.org/10.1016/S0924-4247(03)00267-X doi: 10.1016/S0924-4247(03)00267-X
    [220] Gao C, Zhang W, Ye Z (2016) Numerical study on closed-loop control of transonic buffet suppression by trailing edge flap. Comput Fluids 132 32–45. https://doi.org/10.1016/j.compfluid.2016.03.031 doi: 10.1016/j.compfluid.2016.03.031
    [221] Faisal KM, Salam M, Ali MT, et al. (2017) Flow control using moving surface at the leading edge of aerofoil. J Mech Eng 47: 45–50. https://doi.org/10.3329/jme.v47i1.35420 doi: 10.3329/jme.v47i1.35420
    [222] Rowley CW, Dawson ST (2017) Model reduction for flow analysis and control. Ann Rev Fluid Mech 49: 387–417. https://doi.org/10.1146/annurev-fluid-010816-060042 doi: 10.1146/annurev-fluid-010816-060042
    [223] Ahuja S, Rowley CW (2010) Feedback control of unstable steady states of flow past a flat plate using reduced-order estimators. J Fluid Mech 645: 447–478. https://doi.org/10.1017/S0022112009992655 doi: 10.1017/S0022112009992655
    [224] Duriez T, Brunton SL, Noack BR (2017) Machine learning control-taming nonlinear dynamics and turbulence, Springer.
    [225] Gao C, Zhang W, Kou J, et al. (2017) Active control of transonic buffet flow. J Fluid Mech 824: 312–351. https://doi.org/10.1017/jfm.2017.344 doi: 10.1017/jfm.2017.344
    [226] Ren K, Chen Y, GaoC, et al. (2020) Adaptive control of transonic buffet flows over an airfoil. Phys Fluids 32.
    [227] Nair NJ, Goza A (2020) Leveraging reduced-order models for state estimation using deep learning. J Fluid Mech 897: R1. https://doi.org/10.1017/jfm.2020.501 doi: 10.1017/jfm.2020.501
    [228] Rabault J, Kuchta M, Jensen A, et al. (2019) Artificial neural networks trained through deep reinforcement learning discover control strategies for active flow control. J Fluid Mech 865: 281–302. https://doi.org/10.1017/jfm.2019.62 doi: 10.1017/jfm.2019.62
    [229] Paris R, Beneddine S, Dandois J (2021) Robust flow control and optimal sensor placement using deep reinforcement learning. J Fluid Mech 913: A25. https://doi.org/10.1017/jfm.2020.1170 doi: 10.1017/jfm.2020.1170
    [230] Rabault J, Ren F, Zhang W, et al. (2020) Deep reinforcement learning in fluid mechanics: A promising method for both active flow control and shape optimization. J Hydrodyn 32: 234–246. https://doi.org/10.1007/s42241-020-0028-y doi: 10.1007/s42241-020-0028-y
    [231] Ren F, Hu Hb, Tang H (2020) Active flow control using machine learning: A brief review. J Hydrodyn 32: 247–253. https://doi.org/10.1007/s42241-020-0026-0 doi: 10.1007/s42241-020-0026-0
    [232] Li Y, Chang J, Kong C, et al. (2022) Recent progress of machine learning in flow modeling and active flow control. Chinese J Aeronaut 35: 14–44. https://doi.org/10.1016/j.cja.2021.07.027 doi: 10.1016/j.cja.2021.07.027
    [233] Brodnik NR, Carton S, Muir C, et al. (2023) Perspective: Large language models in applied mechanics. J Appl Mech 90: 101008. https://doi.org/10.1115/1.4062773 doi: 10.1115/1.4062773
  • This article has been cited by:

    1. Xueying Yang, Chen Li, Xu Li, Zhonghua Lu, A Parallel Monte Carlo Algorithm for the Life Cycle Asset Allocation Problem, 2024, 14, 2076-3417, 10372, 10.3390/app142210372
    2. Minfeng Yu, Bo Li, Shuaiyu Zhao, Nitin Roy, Bin Zhang, PPO-based resilient control framework for safer operation of exothermic CSTR, 2024, 09575820, 10.1016/j.psep.2024.11.059
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3983) PDF downloads(522) Cited by(0)

Figures and Tables

Figures(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog