Research article Special Issues

New remarks on the Kolmogorov entropy of certain coarse-grained deterministic systems

  • Unless an appropriate dissipation mechanism is introduced in its evolution, a deterministic system generally does not tend to equilibrium. However, coarse-graining such a system implies a mesoscopic representation which is no longer deterministic. The mesoscopic system should be addressed by stochastic methods, but they lead to practically infeasible calculations. However, following the pioneering work of Kolmogorov, one finds that such mesoscopic systems can be approximated by Markov processes in relevant conditions, mainly, if the microscopic system is ergodic. So, the mesoscopic system tends to stationarity in specific situations, as expected from thermodynamics. Kolmogorov proved that in the stationary case, the instantaneous entropy of the mesoscopic process, conditioned by its past trajectory, tends to a finite limit at infinite times. Thus, one can define the Kolmogorov entropy. It can be shown that in certain situations, this property remains true even in the nonstationary case. We anticipated this important conclusion in a previous article, giving some elements of a justification, whereas it is precisely derived below in relevant conditions and in the case of a discrete system. It demonstrates that the Kolmogorov entropy is linked to basic aspects of time, such as its irreversibility. This extends the well-known conclusions of Boltzmann and of more recent researchers and gives a general insight to the fascinating relation between time and entropy.

    Citation: Michel Moreau, Bernard Gaveau. New remarks on the Kolmogorov entropy of certain coarse-grained deterministic systems[J]. AIMS Mathematics, 2023, 8(11): 26328-26342. doi: 10.3934/math.20231343

    Related Papers:

    [1] Muhammad Sheraz, Vasile Preda, Silvia Dedu . Non-extensive minimal entropy martingale measures and semi-Markov regime switching interest rate modeling. AIMS Mathematics, 2020, 5(1): 300-310. doi: 10.3934/math.2020020
    [2] Andrey Borisov . Filtering of hidden Markov renewal processes by continuous and counting observations. AIMS Mathematics, 2024, 9(11): 30073-30099. doi: 10.3934/math.20241453
    [3] Zelin Zhang, Zhengtao Xiang, Yufeng Chen, Jinyu Xu . Fuzzy permutation entropy derived from a novel distance between segments of time series. AIMS Mathematics, 2020, 5(6): 6244-6260. doi: 10.3934/math.2020402
    [4] Meijiao Wang, Qiuhong Shi, Maoning Tang, Qingxin Meng . Stochastic differential equations in infinite dimensional Hilbert space and its optimal control problem with Lévy processes. AIMS Mathematics, 2022, 7(2): 2427-2455. doi: 10.3934/math.2022137
    [5] Mashael A. Alshehri, Mohamed Kayid . Cumulative entropy properties of consecutive systems. AIMS Mathematics, 2024, 9(11): 31770-31789. doi: 10.3934/math.20241527
    [6] Liu Yang, Lijun Yin, Zuicha Deng . Drift coefficient inversion problem of Kolmogorov-type equation. AIMS Mathematics, 2021, 6(4): 3432-3454. doi: 10.3934/math.2021205
    [7] Haifeng Zheng, Dan Wang . A study of value iteration and policy iteration for Markov decision processes in Deterministic systems. AIMS Mathematics, 2024, 9(12): 33818-33842. doi: 10.3934/math.20241613
    [8] Minyu Wu, Xizhong Yang, Feiran Yuan, Xuyi Qiu . Averaging principle for two-time-scale stochastic functional differential equations with past-dependent switching. AIMS Mathematics, 2025, 10(1): 353-387. doi: 10.3934/math.2025017
    [9] Mansour Shrahili . Some new results involving residual Renyi's information measure for k-record values. AIMS Mathematics, 2024, 9(5): 13313-13335. doi: 10.3934/math.2024649
    [10] Zengming Feng, Tingwen Cao . FOA-BDNet: A behavior detection algorithm for elevator maintenance personnel based on first-order deep network architecture. AIMS Mathematics, 2024, 9(11): 31295-31316. doi: 10.3934/math.20241509
  • Unless an appropriate dissipation mechanism is introduced in its evolution, a deterministic system generally does not tend to equilibrium. However, coarse-graining such a system implies a mesoscopic representation which is no longer deterministic. The mesoscopic system should be addressed by stochastic methods, but they lead to practically infeasible calculations. However, following the pioneering work of Kolmogorov, one finds that such mesoscopic systems can be approximated by Markov processes in relevant conditions, mainly, if the microscopic system is ergodic. So, the mesoscopic system tends to stationarity in specific situations, as expected from thermodynamics. Kolmogorov proved that in the stationary case, the instantaneous entropy of the mesoscopic process, conditioned by its past trajectory, tends to a finite limit at infinite times. Thus, one can define the Kolmogorov entropy. It can be shown that in certain situations, this property remains true even in the nonstationary case. We anticipated this important conclusion in a previous article, giving some elements of a justification, whereas it is precisely derived below in relevant conditions and in the case of a discrete system. It demonstrates that the Kolmogorov entropy is linked to basic aspects of time, such as its irreversibility. This extends the well-known conclusions of Boltzmann and of more recent researchers and gives a general insight to the fascinating relation between time and entropy.



    Following the work of Kolmogorov (see [1] and references herein), the stochastic theory of coarse-grained deterministic, ergodic systems has been studied in [2]. In particular, it was shown that such mesoscopic systems can be approximated by Markov processes, which may explain why Markov models are so widely used in the literature. Here we present new remarks on such systems, in particular concerning their Kolmogorov entropy, which has been introduced for stationary processes in the work of this author [1]. In particular, we will show that the Kolmogorov entropy can be defined for a nonstationary process obeying simple properties which should be satisfied for generic, realistic systems. Here, we will focus on finite systems, which significantly simplifies the reasoning.

    Our definitions and notations are identical with those of [2]. Nevertheless, for the sake of clarity, we summarize them below in Section 2, as well as the known results. New outcomes are presented in Section 3, with simplified demonstrations. Conclusion and discussion are given in Section 4. Detailed derivations are postponed to Appendix A. A more complete and rigorous theory will be presented elsewhere (see Section 4).

    It is known [1] that a coarse-grained deterministic system S can be represented by a non-Markovian stochastic process. One has to define this stochastic process on the space M of the observable mesoscopic states and during all the period of observation we will assume that this period begins at time t = 0, without assigning it a finite end.

    For the sake of simplicity, we consider a finite microscopic dynamical system: The space X of microscopic states is finite and contains N microscopic states x, each of them corresponding to the ultimate possible description of the system, according to the usual conventions of statistical mechanics. Furthermore, time will be discretized: t = 0, 1, …k, …, the elementary time step τ being taken as time unit.

    A probability measure μ(t) is defined on the finite microscopic space X at time t ≥ 0, including N microscopic states. The probability of a set A of microstates states at time t is μ(A, t). The system obeys a deterministic stationary process which transfers an initial microscopic state x into the microscopic state φt(x) after time t, where the evolution function φt satisfies the standard property of such functions (see for instance the book by Arnold and Avez [1] and references therein). So, we assume that φt is measure-preserving, i.e., for any measurable subspace A of X, μ(A,t)=μ(φtA,0). From now on, we also assume that μ is stationary: μ(A,t)=μ(A,0).

    We adopt the current hypothesis that the microscopic dynamical system considered here is ergodic [1]. There is no measure-invariant subspace Y of the microscopic space X, except X itself and the empty space ∅. It is well-known [1,3] that if the microscopic system is ergodic, the stationary measure is unique.

    In the absence of any microscopic information before time 0, it can be assumed that the initial microscopic probability distribution μ is uniform in the whole space X: μ(x) ≡ μ(x, 0) = 1/N, the uniform law being obviously stationary. So, in this article we suppose that the stationary law μ is uniform, although the following reasoning can in many cases be extended to more general stationary measures.

    Because the possible observations are limited and because the measure accuracies are finite, these microscopic states are not directly observable. On the other hand, the limited accuracy of actual, available experiments allows one to define M observable mesostates i, constituting the mesospace M, in such a way that each microscopic state x belongs to one and only one mesostate i. On the other hand, a mesostate i corresponds to ni different microstates, with ni ≥ 1. Clearly, these are the usual conventions of classical statistical mechanics discussed in all textbooks.

    The initial mesoscopic stationary distribution of a mesostate i is proportional to the number ni of microstates included in the mesostate i:

    p0(i,0)=μ(i)=ni/M

    where the upper index 0 denotes the stationary case, and the probability of i at time k is

    p0(i,k)=μ(φki)=p0(i,0) (1)

    for all k > 0.

    The stochastic process representing the coarse-grained states i0, i1, …ik, … at the respective times 0, 1, …k, …. is defined by the probability p0k(i0,0;i1,1;...;ik1,k1) of any k-times trajectory, for all k > 0. The complete stationary probability law, for all k, is denoted p0.

    It is easily seen [2] that the probability of a (k+1)-times trajectory from time 0 is

    p0k+1(i0,0;i1,1;...;ik,k)=μ(φkitk...φ1i1i0). (2)

    To simplify the notations, we will now omit the lower index k+1 in the probability p0k+1 when it is possible without confusion, for instance when the variables are explicitly mentioned.

    With this convention, conditional probabilities can be defined and written in the usual elementary way. For instance, if p0(i0,0;i1,1;...;ik1,k1)>0

    p0(ik,k|ik1,k1;...;i0,0)=p0(i0,0;i1,1;...;ik,k)p0(i0,0;i1,1;...;ik1,k1) (3)

    If p0(i0,0;i1,1;...;ik1,k1)=p0(i0,0;i1,1;...;ik,k)=0, it is well known that the conditional probability is not defined by (3), but this indetermination has no influence on the following calculations.

    The system can be prepared in order that the initial probability of any mesostate i obeys some arbitrary distribution p(i). Then, the microscopic dynamics of the system and the initial probability law p(i) determine the law of the stochastic coarse-grained process over the space of mesostates M ≡ (im), m = 1, …M. Since no available observation can distinguish two microscopic states inside the same mesostate i, it can be logically assumed that the microscopic initial distribution is piecewise uniform, being uniform in each mesostate i. Then, if the initial probability of i is p(i) ≠μ(i), it can be shown [2] that the mesoscopic k-times law is

    pk(i0,0;i1,1;...;ik1,k1)=p(i0)μ(i0)μ(φk+1ik1...φ1i1i0) (4)

    where μ is again the stationary measure on X. This formula allows one to obtain all probabilities concerning finite mesoscopic trajectories, such as the probability that the system is in some mesostate ik at time k, as well as all relevant conditional probabilities.

    If, for instance, the microscopic system is prepared to be initially localized inside some initial mesostate i0, it will no stay concentrated in i0 at the next step (this is forbidden by ergodicity) but it will generally be distributed between several mesostates states i'1, i"1; …. Only in very specific cases, all the microscopic cases included in i0 at time 0, are transferred to the same mesostate i1 at time 1 and to a mesostate state i2 at time 2, etc. Since φ is measure- preserving, μ(i0) = μ(i1) and the microscopic states are uniformly distributed inside i1 at time 1, then transferred to i2 at time 2, etc. In this special case, the mesoscopic trajectory i0, i1, i2, … is periodic, as well as the microscopic trajectories, because the system is supposed to be finite. We will discard such an exceptional situation, which presents no interest and is not realized in current phenomena. On the contrary, we will assume that the coarse-graining is such that, after a relatively small number of steps, the microstates initially concentrated in some i0, are essentially distributed between all the mesostates of M. This is a usual hypothesis, adopted in most textbooks of statistical thermodynamics. It provides an intuitive justification of the memory erasing precisely defined in Section 3.4, Eq (12) and derived previously [2].

    In these conditions, following Kolmogorov [1] and using intuitive extensions of his methods, we will present some remarks on the non-stationary case in Section 3, mainly concerning the entropy of these processes.

    We call n-times entropy Sn(p) of the process the Shannon entropy [2,5,6,7,8] S(pn) of the n-times trajectory (i)n = (i0, …, in-1) in the phase space

    Sn(p)=i0,...in1pn(i0,0;...;in1,n1)lnpn(i0,0;...;in1,n1)S(pn). (5)

    So, among other interpretations [9], this quantity measures the uncertainty, or disorder contained in the n-times probability pn. Equivalently, according to Shannon [5], this is the information recovered after an experiment where an actual trajectory is observed, whereas before the experiment one only knew the probability of this trajectory. Clearly, this entropy vanishes if the trajectory is deterministic.

    It is well known that suppressing correlations between the different states increases the disorder in the stochastic system and increases its entropy. So, the maximum n-times entropy Sn occurs when all the states are statistically independent. In this very special case, the entropy of the n-times process is

    ¯Sn=n1k=0[Nik=1p1(ik,k)lnp1(ik,k)]n1k=0sk(p1(t=k)) (6)

    where sk(p1(t=k)) is the one-time entropy at time k (see below).

    In case the process is stationary, the one-time probability p1 is time invariant and the one-time entropy s1 as well, so that the maximum n-times entropy is ¯Sn=ns1.

    The new information obtained by observing the system in the mesoscopic state in at time n, knowing that it was in the respective states i0, …in−1 at the prior times 0, …n−1, will be called the (average) instantaneous entropy sn at time n [1]

    sn(p)=Sn+1(p)Sn(p)=i0,...inp(i0,0;....;in,n)lnp(in,n|in1,n1;...;i0,0)0=i0,...in1p(i0,0;....;in1,n1)S(p(,n|in1,n1;...;i0,0)). (7)

    Here,

    S(p(,n|in1,n1;...;i0,0))=inp(in,n|in1,n1;...;i0,0)lnp(in,n|in1,n1;...;i0,0) (8)

    is the entropy of the conditional probability at time n, conditioned by the past trajectory. It generally differs from the usual 1-time entropy S(p(., n)), which is often used in physics [10,11] when one does not know the previous states of the system. This 1-time entropy is

    S(p(.,n))=inp(in,n)lnp(in,n). (9)

    S(p(.,n)) is a state function, as defined in thermodynamics. It is seen that

    S(p(.,n))sn(p)=i0,...np(i0,0;...;.in,n)lnp(in,n|in1,n1;...;i0,0)p(in,n)0 (10)

    the equality holding only if the state of S at time n is independent of its prior trajectory.

    The properties of S(pn) and sn(p) have been extensively studied by Kolmogorov and other authors in the case of the stationary process [1,2]. In particular, Kolmogorov [1] showed that if p is the stationary process p0, sn(p0)s0n decreases with time n

    s0n+1s0n0.

    As a result, s0n tends to a non-negative limit ˉs when n → ∞ and

    Sn(pˊa)nandsn(p0)¯s(p0)[0,s0(p0)]if n. (11)

    With some simplification [1,2], ˉs(p0) is the Kolmogorov entropy of the stationary process p0.

    It may be noticed that since p0 is stationary from time 0, the state entropy S(p0(., n)) is clearly a constant s0, whereas sn(p0) decreases from s0 to ˉs when n increases from 0 to infinity.

    It has been shown recently [2] that the memory of the stationary mesoscopic distribution p0 can be approximately limited to the n last past events, n depending of the accuracy required for the approximation. More precisely, for any positive number ε, it is possible to find a positive integer n(ε) such that for any integer k > n(ε)

    0<ikp0(ik,k|ik1,k1;...;i0,0)lnp0(ik,k|ik1,k1;...;i0,0)p0(ik,k|ik1,k1;...;ikn,knp0(0,...k1)<ε. (12)

    Here Ap0(0,...k1) denotes the average of A with respect to the k-times stationary probability p0(i0,0;....;ik1,k1).This property implies [2] that if n is large enough, with overwhelming probability (in the space Mk ≡ {i0, …, ik-1}) the absolute distance(1) between the complete conditional probability p0(ik,k|...;i0,0) and the truncated conditional probability p0(ik,k|...;ikn,kn) is less than ε.

    Thus,

    p0(ik,k;ik1,k1;...;i0,0)=p0(ik,k|ik1,k1;...;i0,0)p(ik1,k1;...;i0,0)p0(ik,k|ik1,k1;..;ikn,kn)p(ik1,k1;...;i0,0)w(ik|ik,...,ikn)p(ik1,k1;...;i0,0) (13)

    where, because of the stationarity of p0, w is defined by

    p0(ik,k|ik1,k1;...;ikn,kn)=p0(ik,n|ik1,n1;...;ikn,0)w(ik|ik1;...,ikn). (14)

    Equation (14) defines the transition probability w.

    (1)Note. The absolute distance [2] between two probabilities p and q on the same discrete space (j) is

    d(p,q)=12j|pjqj|.

    It can be shown [2] that the following Pinsker inequality holds

    2(d(p,q))2<S(p|q)jpjlnpjqj

    which implies (13).

    Because the memory of the process can be approximately limited to the first n past times, we can define an approximation ˉp0 of the stationary process p0 which is a n-Markov process, as defined below.

    (1) n-Markov process

    We say that a process q is a n-Markov process if it satisfies the following property:

    For any integer N > n, one can define a function w(iN,N|iN1,n1;...;iNn,Nn) of the partial trajectory i(Nn,iNn1,...iN) between times N-n and N, such that

    q(iN,N;...;i0,0)=w(iN;N|iN1,N1;...;iNn,Nn)q(iN1,N1;...;i0,0) (15)

    or equivalently

    q(iN,N|iN1,N1;...;i0,0)=w(iN,N|iN1,N1;...;iNn,Nn) (15')

    Summing (15) on i0, … iN-n+1, one sees that, for N > n

    q(iN,N;...;iNn,Nn)=w(iN,N|iN1,N1;...;iNnNn)q(iN1,N1;...;iNn,Nn)

    and consequently

    q(iN,N|iN1,N1;...;iNn,Nn)=w(iN,N|iN1,N1;...;iNn,Nn)=q(iN,N|iN1,N1;...;i0,0). (16)

    This is the characteristic property of a n-Markov process: the conditional probability of any state at time N > n, conditioned by its complete past trajectory from time 0, is identical to the conditional probability of this state, conditioned by its past trajectory during the n previous times only.

    Note that if q is stationary, formula (15) shows that w is invariant if time N is replaced by N+ h, where h is any positive integer, so w is independent of N

    w(iN,N|iN1,N1;...;iNn,Nn)=w(iN,n|iN1,n1;...;i0,0)w(iN|iN1,...,iNn) (17)

    It is seen that, by the approximate equality (13), the coarse-grained stationary distribution p0 of Section 2 is almost an n-Markov process. We now state this property more rigorously.

    (2) Truncated stationary process

    The truncated process ˉp0 is defined from p0 by

    ¯p0(ik,k;...;i0,0)=p0(ik,k;...;i0,0)forkn (18)

    and for k > n, it is obtained by repeated applications of the following iterative formula simulating Eq (13)

    ¯p0(ik,k;ik1,k1;...i0,0)=p0(ik,k|ik1,k1;...;ikn,kn)¯p0(ik1,k1;...;i0,0)=w(ik|ik,...,ikn)¯p0(ik1,k1;...;i0,0). (19)

    All probabilities for fragmentary trajectories between time N-n and N are obtained by summing (19) on irrelevant states. The conditions of Kolmogorov theorem [3] are then satisfied and the n-Markov process ˉp0 is defined by its probability law.

    The fact thatˉp0 is an approximation of p0, as described in 3.4, was derived in [2].

    Equation (19) can easily be generalized to (writing now times in decreasing order)

    ¯p0(ik+n1,k+n1;...;ik,k;...;i0,0)=w(ik+n1,...ik|ik1,....,ikn)x¯p0(ik1,...;ikn,kn;...;i0,0) (20)

    with

    w(ik+n1,...,ik|ik1,....ikn)=w(k+n1|ik+n2,..i,k1)...w(,ik|ik1,...,ikn). (21)

    Summing (20) on i0, …ik--1, we obtain

    ¯p0(ik+n1,k+n1;...;ik,k)=ik1,...knw(ik+n1,...ik|ik1,k1;....ikn)x¯p0(ik1,k1;...;ikn,kn) (22)

    which is clearly a kind of master equation. It is written more easily in the following formalism of partial trajectories.

    We consider integers K ≥ 0 and we define the partial n-steps trajectories (again writing times in decreasing order) IK=(i(K+1)n1),...,iKn+1,iKn) at the n corresponding, decreasing times TK=((K+1)n1,...,Kn+1,Kn).

    Here, n is the integer determined by the accuracy needed for the approximation, according to formula (12).

    So, in abbreviated notations, we can write

    (IK,TK)(i(K+1)n1,(K+1)n1;...;IKn+1,Kn+1;iKn,Kn)

    and

    ¯P0(IK,TK)¯p0(iK+n1,K+n1;...;iK,K). (23)

    Choosing k = Kn in (22) for some positive integer K, this equation takes the condensed form

    ¯P0(IK,TK)=IK1W(IK|IK1)¯P0(IK1,TK1). (24)

    Because of the stationarity of p0, W(IK|IK1)P0(IK,TK|IK1,TK1) is independent of TK, TK-1.

    So, for K ≥ 1, (24) is a generalized master equation [3,12] for the n-steps partial trajectories and the generalized transition rate W(IK|IK1) can be explicitly computed from (21).

    At this step, Eq (24) essentially has a formal interest, since the solution ¯P0(IK,TK) is known from (19). However, it will prove to be very useful in Section 3.1.2.

    Assuming that the microscopic probability at the initial time is piecewise uniform, we saw that the mesoscopic k-times law is given by (3) or

    p(ik1,k1;...;i1,1;i0,0)=p(i0)μ(i0)μ(φk+1itk1...φ1i1i0). (25)

    Thus,

    p(ik|ik1,k1;...;i1,1;i0,0)=p0(ik|ik1,k1;...;i1,1;i0,0). (26)

    Many similar equalities can be found between conditional probabilities of the non-stationary law p and the similar conditional probabilities of the stationary law p0, provided that the conditioning trajectory includes the initial mesostate i0 at time 0. In fact, although the following notations may be heavy, one can easily deduce from (26) the following property:

    If A = {ih}, h, = 1, 2, … k, is any set of k≥ 1 mesostates, and if T = {th}, h = 1, 2, …k, is any set of k positive integer times th, then the conditional probability of the partial trajectory (A, T) = {ih, th}, conditioned by some partial trajectory at times not included in T, but including state i0 at time 0, is identical to the corresponding conditional probability calculated with the stationary law p0k.

    These very simple properties allow one to show [2] that the nonstationary process obeys an approximate generalized master equation which holds on the partial trajectories (IK,TK) of length n defined in 3.5.2.

    They have been evoked in previous papers [2,13]. New results will now be presented and justified. Detailed derivations are postponed to Appendix A.

    Thanks to (12), we can write for k > n

    p(ik,k;...;i1,1;i0,0)=p(ik,k|ik1,k1;...;i1,1;i0,0)p(ik1,k1;...;i1,1;i0,0)=p0(ik,k|ik1,k1;...;i1,1;i0,0)p(ik1,k1;...;i1,1;i0,0)p0(ik,k|ik1,k1;...;ikn,kn)p(ik1,k1;...;i1,1;i0,0). (27)

    Thus, summing on i0, …in− k−1 we obtain

    p(ik,k;...;ikn,kn)p0(ik,k|ik1,k1;...;ikn,kn)p(ik1,k1;...;ikn,kn). (28)

    Comparing (28) and (29), it is seen that the memory of the non-stationary process p at time k is approximately limited to the first n past times k−1, … k− n, as for the memory of the stationary process

    p(ik,k|ik1,k1;...;ikn,kn)p(ik,k|ik1,k1;...;i0,0)=p0(ik,k|ik1,k1;...;ikn,kn)=w(ik|ik1,...,ikn). (29)

    Similar to the reasoning of paragraph 2.2.5.2, Eq (29) implies

    p(ik+n1,k+n1;...;ik,k)ik1,...knw(ik+n1,...,ik|ik1,....,ikn)xp(ik1,k1;...;ikn,kn). (30)

    Taking k = Kn for K ≥ 1 and using the condensed notations defined in 3.5.2, it can be concluded that, with a high probability

    P(IK,TK)IK1W(IK|IK1)P(IK1,TK1). (31)

    So, the non-stationary probability P(IK,TK) approximately satisfies the master Eq (24). Assume that the (nM)x(nM) stochastic matrix W=(W(IK|IK1)) is regular. Then, the exact stationary solution P0 of Eq (31) is unique.

    From the theory of stochastic matrices [14] it is expected that, for any n-steps partial trajectory

    J=j0,j1....,jn1 at the successive times TKn,Kn+1;....,(K+1)n1, we have for any I0

    P(J,TI0,T0)P0(J) when K (32)

    This can be proved with relevant assumptions (see [2] and remarks below). Consequently, if K→∞ (again writing times in decreasing order)

    p(jn1,(K+1)n1;...;j0,Kn|in1,n1:...:i0,0)μ(jn1,,...j(0). (33)

    Renumbering the times and summing on appropriate indexes, (33) implies that, for any positive integers h, k, m and for m < k

    p(ik+h,k+h;...;ik,k|im,m;...;i0,0)μ(ik+h,k+h;...;ik,k) when k. (34)

    So, the non-stationary process is mixing [1,3], as well as the stationary process.

    The meaning of approximation 3.1.2 is further discussed in Appendix A.

    Remarks. The reasoning from Eq (31) to Eq (34) applies to the truncated approximation ¯p that can be defined from p as ¯p0 is defined from the stationary distribution p0 in Section 2.2.5.1. It is shown in [2] that in the notation of partial trajectories, ¯P(IK,TK) satisfies (31). So, it results [2] from the matrix theory that if the matrix W(IK|IK1) is regular, ¯P(IK,TK) tends to the stationary solution of (31) for any initial partial trajectory I0 when K→∞

    ¯P(IK,TK|I0,T0) P0(IK for any I0 if K. (35)

    We now present our main, new result: Under certain, reasonable conditions, the instantaneous entropy sn(p) tends to a finite limit s(p) when n→∞, thus defining the Kolmogorov entropy s(p) of the mesoscopic non-stationary process p. So, with relevant assumptions the basic result established by Kolmogorov for the stationary situation p0 is extended to non-stationarity. The discussion of this assertion needs some detailed calculations which are discussed in Appendix A. The proof can be summarized as follows.

    With a given accuracy ε, the memory of the process can be neglected at times larger than n(ε), in the sense of Section 2.2.4. Then, it is probable that the instantaneous entropy (6) at time N = k + n(ε) is, for any k > 0,

    sk+n(p)=i0p(i0)i1,...ik+nμ(ik+n,k+n;...;i0,0)μ(i0)lnμ(ik+n,k+n|ik+n1,k+n1;...;i0,0)i0p(i0)i1,...ik+nμ(ik+n,k+n;...;i1,1;i0,0)μ(i0)lnμ(ik+n,k+n|ik+n1,k+n1;...;ik,k)=i0p(i0)ik,...ik+nμ(ik+n,k+n;...;ik,k;i0,0)μ(i0)lnμ(ik+n,k+n|ik+n1,k+n1;...;ik,k) (36)

    Here, p(i0) is the initial, non-stationary probability of i0 and μ(.ik,k;...;i0,0) is the stationary probability of a trajectory at times k. …, 0.

    Considering that by (34), the stationary process is mixing [1,3], we have

    μ(ik+n,k+n;...;ik,k|i0,0)μ(ik+n,k+n;....;ik,k;i0,0)μ(i0)μ(ik+n,k+n;...;ik,k)ifk. (37)

    Thus, if k and n are large enough,

    sk+n(p)ik,...ik+nμ(ik,k;....;.ik+n,k+n)lnμ(ik+n,k+n|ik+n1,k+n1;...;ik,k)=s0n (38)

    where we used the stationarity of the measure μ. It follows from (38) and (11) that sN(p) and sN(p0)s0N have a common limit ˉs when N →∞

    sN(p)s(p0)=¯sifN. (39)

    So, in appropriate conditions the Kolmogorov entropy ˉs is defined even for a non-stationary mesoscopic process. This is our main result: Its detailed justification is commented in Appendix A.

    We have proved that with relevant hypothesis, even in the non-stationary situation, the instantaneous entropy of the mesoscopic coarse-grained process tends to a finite limit, depending only on its stationary measure μ. Thus, we have completed the analysis presented in [2], which essentially proved that the partial trajectories traveled during a mesoscopic time interval n can be approximated by a n-times Markov process if n is large enough. This was done with relatively simple methods, although the notations and the calculations may be somewhat heavy. A more general theory, using the formalism of martingales, will be presented in a further publication (B. Gaveau, M. Moreau, Coarse-graining a deterministic system: Martingale theory, unpublished work). A complete discussion of our approach, including a comparison with other points of view (see for instance [24] and references herein) should profitably be performed at the light of the forthcoming article.

    From the present results, the entropy introduced by Kolmogorov can suggest new remarks concerning time and its relations with physics and probabilities. This subject has been addressed in a vast literature and over the course of centuries innumerable philosophers tried to analyze time [14]. Of course, we do not intend to discuss or even to evoke all these works. We would only like to point out that the Kolmogorov entropy presents an innovative point of view on two basic aspects of time: Its irreversibility and its (apparently) regular progress. Since Boltzmann [16,17], time irreversibility (the arrow of time) is linked to the growth of entropy for isolated systems. This principle stems from classical thermodynamics [10,11,18], but it received a first analytical basis thanks to Boltzmann [16,17], who not only gave a theoretical definition of entropy but also proved, by his celebrated H theorem, that the one-time entropy of a non-equilibrium isolated system increases with time, within the collision model of low density gases. However, this is a very specific model. In spite of some possible extensions, it gives no assurance on the generality of this conclusion. More recently, the relation between time and irreversibility again attracted the attention of many scientists (see for instance [19,20,21,22,23,24,25]). In particular it was given an original form by I. Prigogine [20,21,22,23]. However, despite their interest, it seems that many of these works are restricted to special examples and can hardly represent a general approach.

    Things are clearer if, with Kolmogorov, one considers the entropy of a stochastic process. On the one hand, this entropy is directly associated with its time evolution, which is included in the very definition of stochastic processes [3]. Furthermore, it is based on the memory of the events. Not only does this point agree with current observation, but it also meets the opinion of philosophers who pointed out that subjective time is related to human memory (see for instance [16] and references therein). On the other hand, the trajectory entropy clearly increases with time in all circumstances, whereas the average, instantaneous entropy (conditioned by the past) does not necessarily increase with time: It is even non-increasing for a stationary trajectory, as proved by Kolmogorov [1].

    Eventually, the instantaneous entropy tends to a finite limit for a stationary process and in certain conditions for a non-stationary process as well, as shown in this article. This fact appears to be linked with the regular flow of time commonly experienced. In the Kolmogorov approach, the time scale of the process is related to the rate of information creation due to the possible observations, according to Shannon [5], or it arises from the rate of disorder production due to the stochastic evolution, according to Boltzmann.

    We think that these simple remarks, concerning a very old and complex problem, could deserve to be developed in the framework of Kolmogorov entropy.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors declare no conflict of interest.

    Although the memory erasing expressed by formula (12) only holds in the stationary situation, it has important consequences for the instantaneous entropy sn of a nonstationary mesoscopic process, defined by (6). This is due to the fact that the nonstationary conditional probabilities are equal to the stationary conditional probabilities in the specific case considered here and described in Section 3.1.1. So, the instantaneous entropy sN at time N can be written by (6)

    sN=i0p(i0)i1,...iNμ(iN,N;...;i0,0)μ(i0)lnμ(iN,N|iN1,N1;...;iN,Nn)i0p(i0)i1,...iNμ(iN,N;...;i0,0)μ(i0)lnμ(iN,N|iN1,N1;...;i0,0)μ(iN,N|iN1,N1;...;iNn,Nn)s(1)N+s(2)N(therespectivesumsappearingin the 1stlineandin the 2ndlineabose). (A.1)

    Let N = k + n, n and k be positive integers. Assuming that the stationary process μ is mixing in the sense defined by (32), it is found that if k is large enough,

    s(1)n+ks0n

    where s0n is the instantaneous entropy of the stationary process at time n (see Eq (37)). So, if k and n→∞

    s(1)n+kˉs (A.2)

    and s is the Kolmogorovb entropy (39) of the stationary process.

    On the other hand, it results from the memory erasing property (12) that, in s(2)N, the ratio

    λμ(iN,N|iN1,N1;...;i0,0)μ(iN,N|iN1,N1;...;iNn,Nn) (A.3)

    is (probably) very close to 1 if n is large enough. As a consequence, it can be proved, with relevant assumptions (see below) that

    s(2)N0ifN. (A.4)

    From (A.1-3), we can conclude with some additional hypotheses that

    sN¯sifN. (A.5)

    Then, the instantaneous entropy of a nonstationary process tends to the same limit ˉs as the stationary process corresponding to the stationary measure μ.

    It should be pointed out that supplementary hypotheses are necessary to prove (A.5). In fact, the memory erasing relation (12), which is essential for the derivation, is most probably satisfied but with a very low probability it can fail to be verified. A sufficient condition for obtaining (A.5) is that the ratio λ of (A.3) has finite upper and lower bounds, independent of n and N.

    With this assumption, it is not difficult to derive the previous results, but the calculations are lengthy. Rather than detailing them we prefer to summarize and discuss the main hypotheses used in the reasoning.

    It has been shown that the mesoscopic process approximately satisfies a generalized Markov process, whose probability tends to the stationary probability at infinite times. This fact does not necessarily imply that the mesoscopic probability has the same property. We gave arguments in this sense elsewhere [2,13].

    This is a general problem in modeling, when the evolution of some actual system is shown to obey approximate, theoretical equations: It is difficult to know whether their formal asymptotics still represent the natural system correctly. This point is overlooked in some physical publications. Even if mathematical conditions are found to ensure the relevance of the model at large times, it is difficult to check whether these conditions are verified in practice. The present study does not avoid this difficulty completely.

    Another basic point is that the stationary, mesoscopic probability μ is supposed to be mixing (see (32)). This property seems to be a natural extension of the memory erasing, proved for the mesoscopic process. In fact, this assertion is not obvious, but it is true for a process with a finite n-steps memory, or generalized Markov process. According to the previous discussion, this should be true also for the stationary probability μ, since we have shown that, with any accuracy ε, μ can be approximated by a process ˉp0 with a finite memory of n(ε), steps.

    Eventually, the assumption that the ratio λ, defined by (A.3), has finite, nonzero upper and lower bounds seems reasonable because it just completes and reinforces the fact that λ is almost everywhere close to 1. One can notice that the ratio λ should only be considered if μ(iN,N|iN1,N1;...;iNn,Nn)>0. The case μ(iN,N|iN1,N1;...;iNn,Nn)=0 implies μ(iN,N;iN1,N1;...;iNn,Nn)=0 and μ(iN,N;iN1,N1;...;i0,0)=0. In this situation, the partial trajectory (i0,0;...;iN,N) does not contribute to the entropy sN and it can be ignored for calculating λ. Because the space M of mesostates i is finite and consists in M elements, the number of partial trajectories iN-n, …iN is M n and the upper and lower bounds of λ are finite and independent of N. Their values can only be estimated in specific cases or for some academic models.



    [1] V. I. Arnold, A. Avez, Ergodic problems of classical mechanics, Mathematical Physics Monographs, Benjamin, 1968.
    [2] B. Gaveau, M. Moreau, On the stochastic representation and Markov approximation of Hamiltonian systems, Chaos, 30 (2020), 083104. https://doi.org/10.1063/5.0001435 doi: 10.1063/5.0001435
    [3] J. Doob, Stochastic processes, Wiley, New York, 1953.
    [4] P. Levy, Théorie de l'addition des variables aléatoires, Gauthier-Villars, Paris, 1937.
    [5] C. E. Shannon, A mathematical theory of communication, Bell Syst. Tech. J., 27 (1948), 623–656. https://doi.org/10.1002/j.1538-7305.1948.tb00917.x doi: 10.1002/j.1538-7305.1948.tb00917.x
    [6] A. I. Khinchin, Mathematical foundations of information theory, Dover, New York, 1957.
    [7] R. MacMillan, The basic theorems of information theory, Ann. Math. Stat., 24 (1953), 193. https://doi.org/10.1214/aoms/1177729028 doi: 10.1214/aoms/1177729028
    [8] T. Cover, J. Thomas, Elements of information theory, Wiley, New York, 1991.
    [9] L. Brillouin, Science an information theory, Academic Press, New York, 1956.
    [10] L. D. Landau, E. M. Lifshitz, Statistical physics, 3 Eds., Pergamon Press, Oxford, 1980.
    [11] F. Reif, Fundamentals of statistical and thermal physics, Mc Graw Hill, New York, 1965.
    [12] W. Feller, An introduction to probability theory and its applications, Wiley, New York, 1971.
    [13] M. Moreau, B. Gaveau, Stochastic theory of coarse-grained deterministic systems: Martingales and Markov applications, In: Advances in Dynamical Systems Theory, Models, Algorithms and Applications, 2021. https://doi.org/10.5772/intechopen.95903
    [14] G. Akemann, J. Baik, P. Di Francesco, The Oxford handbook of random matrix theory, Oxford University Press, 2015. https://doi.org/10.1093/oxfordhb/9780198744191.001.0001
    [15] C. Hoerl, T. McCormack, Time and memory: Issues in philosophy and psychology, Clarendon, Oxforf, 2001.
    [16] L. Boltzmann, Vorlesungenüber gastheorie, Part 2, J. A. Barth, Leipzig, 1898.
    [17] L. Boltzmann, Reply to Zermelo's remarks on the theory of heat, In: History of modern physical sciences: The kinetic theory of gases, Imperial College Press, 57 (1896), 567.
    [18] H. B. Callen, Thermodynamics and an introduction to thermostatistics, John Wiley and Sons, New York, 1985.
    [19] L. S. Schulman, Time's arrows and quantum measurement, Cambridge University Press, Cambridge, 1997. https://doi.org/10.1017/CBO9780511622878
    [20] L. S. Schulman, arXiv.cond-mat/991101, 1999.
    [21] I. Prigogine, From being to becoming, Freeman, 1980.
    [22] I. Prigogine, Time and change, 2003.
    [23] I. Prigogine, Time, dynamics and chaos: Integrating Poincare's non-integrable systems, Center for Studies in Statistical Mechanics and Complex Systems at the University of Texas-Austin, United States Department of Energy-Office of Energy Research, Commission of the European Communities, 1990.
    [24] J. B. Gao, Y. H. Cao, W. W. Tung, J. Hu, Multiscale analysis of complex time series-integration of chaos and random fractal theory, and beyond, Wiley, New York, 2007.
    [25] J. B. Gao, F. Y. Liu, J. F. Zhang, J. Hu, Y. H. Cao, Information entropy as a basic building block of complexity theory, Entropy, 15 (2013), 3396. https://doi.org/10.3390/e15093396 doi: 10.3390/e15093396
  • This article has been cited by:

    1. Bernard Gaveau, Michel Moreau, Generalized kinetic theory of coarse-grained systems. II. Comparison of various approximations and coarse-grainings, 2025, 194, 09600779, 116093, 10.1016/j.chaos.2025.116093
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1376) PDF downloads(52) Cited by(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog