
Local discovery plays an important role in Bayesian networks (BNs), mainly addressing PC (parents and children) discovery and MB (Markov boundary) discovery. In this paper, we considered the problem of large local discovery. First, we focused on an assumption about conditional independence (CI) tests: We explained why it was unreasonable to assume all CI tests were reliable in large local discovery, studied how the power and reliability of CI tests changed with the data size and the number of degrees of freedom, and then modified the assumption about CI tests in a more reasonable way. Second, we concentrated on improving local discovery algorithms: We posed the problem of premature termination of the forward search, analyze why it arose frequently in large local discovery when implementing the existing local discovery algorithms, put forward an idea of preventing the premature termination of forward search called information connection (IC), and used IC to build a novel algorithm called ICPC; the theoretical basis of ICPC was detailedly presented. In addition, a more steady incremental algorithm as the subroutine of ICPC was proposed. Third, the way of breaking ties among equal associations was considered and optimized. Finally, we conducted a benchmarking study by means of six synthetic BNs from various domains. The experimental results revealed the applicability and superiority of ICPC in solving the problem of premature termination of the forward search that arose frequently in large local discovery.
Citation: Jianying Rong, Xuqing Liu. Local discovery in Bayesian networks by information-connecting[J]. AIMS Mathematics, 2024, 9(8): 22743-22793. doi: 10.3934/math.20241108
[1] | Yuezhen Ren, Ruihu Li, Guanmin Guo . New entanglement-assisted quantum codes constructed from Hermitian LCD codes. AIMS Mathematics, 2023, 8(12): 30875-30881. doi: 10.3934/math.20231578 |
[2] | Yang Liu, Ruihu Li, Qiang Fu, Hao Song . On the minimum distances of binary optimal LCD codes with dimension 5. AIMS Mathematics, 2024, 9(7): 19137-19153. doi: 10.3934/math.2024933 |
[3] | Hatoon Shoaib . Double circulant complementary dual codes over F4. AIMS Mathematics, 2023, 8(9): 21636-21643. doi: 10.3934/math.20231103 |
[4] | Bodigiri Sai Gopinadh, Venkatrajam Marka . Reversible codes in the Rosenbloom-Tsfasman metric. AIMS Mathematics, 2024, 9(8): 22927-22940. doi: 10.3934/math.20241115 |
[5] | Irwansyah, Intan Muchtadi-Alamsyah, Fajar Yuliawan, Muhammad Irfan Hidayat . Generalized Reed-Solomon codes over number fields and exact gradient coding. AIMS Mathematics, 2024, 9(4): 9508-9518. doi: 10.3934/math.2024464 |
[6] | Wei Qi . The polycyclic codes over the finite field Fq. AIMS Mathematics, 2024, 9(11): 29707-29717. doi: 10.3934/math.20241439 |
[7] | Xuesong Si, Chuanze Niu . On skew cyclic codes over M2(F2). AIMS Mathematics, 2023, 8(10): 24434-24445. doi: 10.3934/math.20231246 |
[8] | Ismail Aydogdu . On double cyclic codes over Z2+uZ2. AIMS Mathematics, 2024, 9(5): 11076-11091. doi: 10.3934/math.2024543 |
[9] | Guanghui Zhang, Shuhua Liang . On the construction of constacyclically permutable codes from constacyclic codes. AIMS Mathematics, 2024, 9(5): 12852-12869. doi: 10.3934/math.2024628 |
[10] | Adel Alahmadi, Tamador Alihia, Patrick Solé . The build up construction for codes over a non-commutative non-unitary ring of order 9. AIMS Mathematics, 2024, 9(7): 18278-18307. doi: 10.3934/math.2024892 |
Local discovery plays an important role in Bayesian networks (BNs), mainly addressing PC (parents and children) discovery and MB (Markov boundary) discovery. In this paper, we considered the problem of large local discovery. First, we focused on an assumption about conditional independence (CI) tests: We explained why it was unreasonable to assume all CI tests were reliable in large local discovery, studied how the power and reliability of CI tests changed with the data size and the number of degrees of freedom, and then modified the assumption about CI tests in a more reasonable way. Second, we concentrated on improving local discovery algorithms: We posed the problem of premature termination of the forward search, analyze why it arose frequently in large local discovery when implementing the existing local discovery algorithms, put forward an idea of preventing the premature termination of forward search called information connection (IC), and used IC to build a novel algorithm called ICPC; the theoretical basis of ICPC was detailedly presented. In addition, a more steady incremental algorithm as the subroutine of ICPC was proposed. Third, the way of breaking ties among equal associations was considered and optimized. Finally, we conducted a benchmarking study by means of six synthetic BNs from various domains. The experimental results revealed the applicability and superiority of ICPC in solving the problem of premature termination of the forward search that arose frequently in large local discovery.
Even though the roots of fractional calculus are in the late 17th century [1], the subject has received significant attention from researchers only in recent decades. The importance of fractional calculus and fractional differential equations (FDEs) is evidenced by the emergence of fractional-order models for real-world phenomena, which has lead to a plethora of applications of FDEs [2]. A short review of the areas in which FDEs are applied is given below.
The advantage of fractional-order models is that they can be used to model systems with the memory effect, under which the current state of the system at time t=t0 is affected by an interval of states at time t<t0. Thus, such models lend themselves naturally to the description of viscoelastic materials [3,4]. It can be noted that most biological tissues possess viscoelastic properties [5] which has lead to the use of FDEs to describe the evolution of various biomedical systems [6,7,8]. In physics, fractional-order models are widely used to study dielectric materials [9,10,11] and Bose-Einstein condensates [12]. FDEs have also been used to describe the memory effect in economic models [13,14].
Due to the applicability of fractional-order models, a wide variety of methods have been developed to construct solutions to FDEs. The Q-homotopy analysis transform is applied in order to obtain analytical solutions to the fractional coupled Ramani equation in [15]. The solutions of integer-order differential equations are used to construct the solutions to the models of cooling and spread of epidemic diseases [16]. An Adams-type predictor-corrector method for the construction of numerical solutions to FDEs is discussed in [17]. A new family of predictor-corrector methods for FDEs is discussed in [18]. A Legendre operational matrix approach is used to numerically solve FDEs in [19,20].
The homotopy analysis method is adapted to FDEs for the computation of approximate solutions to various equations, including fractional Korteweg-de-Vries, Burgers and Boussinesq models in [21]. A class of rational Krylov methods for the construction of numerical solutions to partial fractional differential equations is considered in [22]. The traditional reproducing kernel method is adapted to be applicable to FDEs in [23]. A restricted transform technique is used to solve irrational order FDEs in [24]. Artificial neural networks are used to construct approximate solutions to FDEs in [25].
The main objective of this paper is to present a novel framework for the extension of the solutions to fractional differential equation (FDEs) to negative argument values. The main idea of this procedure is to construct the ordinary differential equation (ODE) that has a solution which can be transformed to the solution of the considered FDE. Since the solution of the ODE also exists for negative argument values, it can be used to extend the solution of the FDE for such argument values as well. This results in complex-valued solutions to FDEs for negative argument values. Such extension is a novel viewpoint into the solutions to FDEs, as they can be analyzed not only in the set x∈R+, but in the entire real line x∈R. This goes beyond the current state-of-the-art, as with current algorithms the solutions to FDEs cannot be considered in the negative x-axis.
In this paper we consider functions defined via power series:
f(x)=+∞∑j=0vjw(n)j, | (2.1) |
where vj∈R are coefficients of the series and the base functions w(n)j;n∈N,j=0,1,… are defined as:
w(n)j(x)=xjnΓ(jn+1);n∈N,j=0,1,…. | (2.2) |
The usual notation Γ(x) is used to denote the gamma function [26].
The number n which denotes the order of the basis of fractional power series is selected with respect to the fractional derivative order which is analyzed. If the fractional derivative is of order α=km, and gcd(k,m)=1, it is taken that n=m [27].
Consider x>0 and let n√x∈R. Using the substitution t:=n√x, (2.1) can be rewritten as:
f(x)=ˆf(t)=+∞∑j=0γj(n√x)j=+∞∑j=0γjtj,γj=vjΓ(jn+1). | (2.3) |
For subsequent computations it is required that ˆf(t) be analytic (except for some finite number of singularity points) and non-singular at t=0. Thus, there exists a convergence radius |t|<T0;T0>0. The function ˆf will be referred to as the characteristic function of f.
For any given n∈N, the set of all functions defined by (2.1) and meeting the requirements stated above is denoted as CFn. Conventional addition and multiplication operations on functions belonging to CFn are defined. For example, given two series f=∑+∞j=0vjw(n)j and g=∑+∞j=0bjw(n)j, the product is defined as:
f⋅g=+∞∑j=0(j∑k=0(jk)vkbj−k)w(n)j. | (2.4) |
Note that the sum above obeys the finite summation principle: all coefficients of the above series are given by finite sums, not series [28].
The set CFn with addition and multiplication operations forms an algebra over C. The properties of algebra CFn are discussed in detail in [27] and [29].
Let f=∑+∞j=0vjw(n)j∈CFn. Caputo differentiation operator of order 1n is then defined as [30]:
CD(1/n)f=+∞∑j=0vj+1w(n)j. | (2.5) |
Note the Caputo derivatives of basis functions [27]:
CD(1/n)w(n)j={0,j=0w(n)j−1,j=1,2,… | (2.6) |
Derivatives of rational order kn;k∈N are defined by higher powers of operator CD(1/n) [29]:
(CD(1/n))kf=+∞∑j=0vj+kw(n)j=+∞∑j=0Γ(j+kn+1)γj+kΓ(jn+1)(n√x)j. | (2.7) |
In the special case of k=m⋅n;m∈N, (2.7) reads:
(CD(1/n))m⋅nf=+∞∑j=0Γ(jn+m+1)Γ(jn+1)γj+mn(n√x)j=+∞∑j=0(m∏k=1(jn+k))γj+mn(n√x)j. | (2.8) |
It is well-known that analytic functions can be extended beyond their convergence radius [31]. The same idea is adapted for series of the type (2.1). Let f=∑+∞j=0γj(n√x)j and choose n√x0 such that x0 is nonzero and in the characteristic function's convergence radius, 0<x0<Tn0. Then (2.1) can be rearranged:
f=+∞∑j=0γj((n√x−n√x0)+n√x0)j=+∞∑j=0j∑k=0(jk)γj(n√x0)j−k(n√x−n√x0)k=+∞∑k=0(+∞∑j=k(jk)γj(n√x0)j−k)(n√x−n√x0)k=+∞∑j=0ˆγj(n√x−n√x0)j. | (2.9) |
Coefficients ˆγj are defined as:
ˆγj=+∞∑k=j(kj)γk(n√x0)k−j. | (2.10) |
Note that coefficients (2.10) are finite, because the non-extended series f converges in its convergence radius 0<x0<Tn0. The convergence radius for (2.9) is |n√x−n√x0|<T1,T1>0. Furthermore, if the same procedure is repeated for x1≠x0 that also satisfies 0<x1<Tn0, then:
+∞∑j=0ˆγj(n√x−n√x0)j=+∞∑j=0˜γj(n√x−n√x1)j, | (2.11) |
as long as x is such that series on both sides of the equation converge.
This procedure can be applied to extend a function from the set CFn between two singularity points (or singularity and infinity, if no other singularities are present) by rewriting the function in different basis (n√x−n√xk)j,k=0,1,…. Note that these extended series coincide for x that lies in all of their convergence radii, thus the extensions are unique.
Note that computational operations, such as the Caputo differentiation operator are only defined for basis (n√x)j (in the neighbourhood of x=0). For this reason, all computations are first performed in this neighbourhood and then extended throughout the entire function domain. Thus, the zero neighbourhood of x is called the algebraic operation neighbourhood. The concept of this neighbourhood is introduced in [28].
In the subsequent sections of this paper, constructing series solutions to ordinary differential equations (ODEs) is necessary. The generalized differential operator technique is used for this task. A short description of this method is given in this section. A detailed overview can be found in [32,33].
Let us consider an ordinary differential equation in the explicit form:
dz dt=P(t,z);z(c)=s;c,s∈R, | (2.12) |
where P is an arbitrary bivariate analytic function. The following generalized differential operator is associated with the ODE (2.12) [34]:
Dcs=Dc+P(c,s)Ds, | (2.13) |
where Dλ denotes the partial differentiation operator with respect to variable λ.
The series solution to (2.12) can then be written in the form [32]:
z(t,c,s)=+∞∑j=0(t−c)jj!pj(c,s), | (2.14) |
where pj(c,s)=Djcss.
The following fractional differential equation is considered in this section:
(CD(1/n))nyn=Qm(yn);yn=yn(x), | (3.1) |
where yn∈CFn and Qm is an arbitrary mth order polynomial:
Qm(y)=m∑k=0akyk;m∈N,am≠0. | (3.2) |
In the remainder of this section, solutions in the operation neighbourhood of x=0 will be constructed directly and later extended into different neighbourhoods using the procedure described in previous sections.
Consider a series yn=∑+∞j=0vjw(n)j=∑+∞j=0γj(n√x)j∈CFn. Inserting the function into (3.1) yields:
+∞∑j=0(jn+1)γj+n(n√x)j=m∑k=1(ak(+∞∑j=0(∑k1+⋯+km=jk∏l=1γkl)(n√x)j))+a0. | (3.3) |
Let n√x=t and ˆyn=∑+∞j=1γjtj. Then, (3.3) yields:
+∞∑j=0(j+n)γj+ntj=nQm(ˆyn). | (3.4) |
Rearranging the sums in (3.4) yields:
+∞∑j=njγjtj−1=ntn−1Qm(ˆyn). | (3.5) |
It can be noted that adding ∑n−1j=1jγjtj−1 to both sides of (3.5) transforms the right side to an ordinary derivative of ˆyn:
dˆyndt=ntn−1Qm+n−1∑j=1jγjtj−1. | (3.6) |
Noting that γj=vjΓ(jn+1) yields a different form of (3.6):
dˆyndt=n(tn−1Qm(ˆyn)+n−1∑j=1vjΓ(jn+1)tj−1). | (3.7) |
Adding the initial condition yn(c)=s, Equation (3.7) can be solved using the method described in 2.4. The generalized differential operator for (3.7) reads:
Dcs=Dc+n(cn−1Qm(s)+n−1∑j=1vjΓ(jn+1)cj−1)Ds. | (3.8) |
The solution to (3.7) (taking values of t for which the series converges) has the form:
ˆyn=ˆyn(t,c,s)=+∞∑j=0(t−c)jj!pj(c,s), | (3.9) |
where pj(c,s)=Djcss. Noting that yn(x)=ˆyn(n√x) yields Theorem 3.1.
Theorem 3.1. The fractional differential equation (3.1) admits the following general solution for any parameter values u0;v1,v2,…,vn−1;x0≥0:
yn(x)=+∞∑j=0(n√x−n√x0)jj!pj(n√x0,u0), | (3.10) |
where
yn(n√x0)=u0, | (3.11) |
and
(CD(1/n))kyn|x=0=vk;k=1,…,n−1, | (3.12) |
if x satisfies |n√x−n√x0|<Tx0. Here Tx0>0 denotes the convergence radius of (3.10).
Note that the results of this section can be extended to analytic functions. Taking Q(x)=∑+∞k=0akxk instead of Qm(x) and following the steps outlined above would yield a theorem that is analogous to Theorem 3.1. However, the analysis of such equations is not the focus of this paper, thus we will consider only polynomial Qm(x) in the remainder of the text.
Based on Theorem 3.1, the Cauchy initial value problem for (3.1) can be formulated at point x=0:
(CD(1/n))nyn=Qm(yn),yn=yn(x,u(0)0); | (3.13) |
(CD(1/n))kyn|x=0=vk;k=1,…,n−1; | (3.14) |
yn|x=0=u(0)0. | (3.15) |
Note that since n√x in general possesses n branches, in these computations we select the branch with a minimum value of argx.
By the results obtained in the previous section, in the neighbourhood of x=0, the problem (3.13)–(3.15) has the solution:
yn=+∞∑j=0(n√x)jj!pj(0,u(0)0). | (3.16) |
This solution can be extended by applying the algorithm discussed in section 2.3. First, choose a sequence x0=0<x1<x2<… and freely choose u(0)0∈R. Then, compute parameters u(1)0,u(2)0,… as:
u(k+1)0=ˆyn(n√xk+1,n√xk,u(k)0),k=0,1,… | (3.17) |
The general solution to (3.13)–(3.15) can then be written for any k:
yn=yn(x,u(0)0)=ˆyn(n√x,n√xk,u(k)0), | (3.18) |
for |n√x−n√xk|<Txk, Txk>0,xk≥0.
The rearrangements described in the previous section can be applied to construct a more general Cauchy initial value problem, where the first initial condition can be formulated at nonzero x:
(CD(1/n))nyn=Qm(yn),yn=yn(x,u(0)0); | (3.19) |
(CD(1/n))kyn|x=0=vk;k=1,…,n−1; | (3.20) |
yn|x=x0=u(0)0;x0≠0.. | (3.21) |
Analogously to problem (3.13)–(3.15), the solution reads:
yn=yn(x,x0,u(0)0)=ˆyn(n√x,n√x0,u(0)0). | (3.22) |
Note that initial conditions that correspond to fractional derivative values still must be formulated at x=0 for solution (3.22) to hold true.
As stated earlier, the main objective of this paper is to extend the solutions to FDEs so that they would exist for negative values of argument x. This extension results in complex-valued fractional power series, which are defined in the following subsection. Note that this extension is not possible when considering fractional derivatives in the traditional Caputo sense, as they are defined via integrals only for non-negative values of x.
Let us consider an extension to the power series basis presented in section 2.1:
(w(n)j)k:=(n√|x|)jΓ(jn+1)exp(ijargx+2πkn), | (4.1) |
where x∈R, k=0,…,n−1;j=0,1,… and n√|x|∈R denotes the root with the lowest value of argx. The basis with k=0 corresponds to w(n)j presented in section 2.1, while k=1,2,…,n−1 result in complex-valued functions (w(n)j)k.
Using (4.1), the standard fractional power series (2.1) can be extended into n complex-valued series:
fk(x)=+∞∑j=0vj(w(n)j)k;k=0,…,n−1. | (4.2) |
By Theorem 3.1, the solution to
(CD(1/n))nyn=Qm(yn);yn=yn(x), | (4.3) |
reads:
yn(x)=+∞∑j=0(n√x−n√x0)jj!pj(n√x0,u0), | (4.4) |
where coefficients pj are generated via the generalized differential operator technique as described in previous sections. Using the basis defined in the previous section, the solution (4.4) can extended into the complex plane:
(yn(x))k=+∞∑j=0(n√|x|−n√|x0|)jj!exp(ijargx0+2πkn)pj(n√|x0|exp(iargx0+2πkn),u0). | (4.5) |
Denoting α:=argx0 and β(k)n(α):=exp(iα+2πkn), series (4.5) can be rewritten as:
(yn(x))k=+∞∑j=0(n√|x|−n√|x0|)jj!(λ(k)j(|x0|,α,u0)+iμ(k)j(|x0|,α,u0)), | (4.6) |
where
λ(k)j(|x0|,α,u0):=Re((β(k)n(α))jpj(n√|x0|β(k)n(α),u0)); | (4.7) |
μ(k)j(|x0|,α,u0):=Im((β(k)n(α))jpj(n√|x0|β(k)n(α),u0)). | (4.8) |
Expression (4.6) allows the consideration of solutions for x<0, however, these solutions are multi-valued (a total of n branches corresponding to the number of distinct roots n√x) and take complex values.
A refinement of the procedure described in section 3.2 is needed to construct particular solutions to FDEs (4.3). Two sequences of numbers are chosen: ⋯<x−r<⋯<x−1<0; 0<x1<x2<… and x0=0. For any k=0,…,n−1, the initial condition s(k)0∈R can be taken freely.
Sequences s(k)−r and s(k)r are computed using relations:
s(k)r+1:=ˆy((n√xr+1)k,(n√xr)k,s(k)r);r=0,1,…; | (4.9) |
s(k)−r−1:=ˆy((n√x−r−1)k,(n√x−r)k,s(k)−r);r=0,1,…. | (4.10) |
Here ˆy denotes the solution of the ODE that the FDE (4.3) is transformed to, as given by Theorem 3.1.
Then the particular solution to (4.3) corresponding to the kth branch of the root n√x reads:
(yn(x))k=ˆy((n√x)k;(n√xr)k;s(k)r), | (4.11) |
where r=0,±1,±2,….
Note that the sequence ⋯<x−1<x0<x1<… must be chosen in such a way that the series generated by ˆy are convergent in both R and C. Also, the point x=0 is a branching point for the solution to the FDE, thus the functions (yn(x))k are non-differentiable at this point.
Let us consider the non-fractional Riccati equation:
dy dx=2y2−5y−3; | (5.1) |
y(x0)=u0. | (5.2) |
It is well-known that (5.1) admits the kink solitary solution:
y(x,x0,u0)=α2(u0−α1)exp(2α1(x−x0))−α1(u0−α2)exp(2α2(x−x0))(u0−α1)exp(2α1(x−x0))−(u0−α2)exp(2α2(x−x0)), | (5.3) |
where α1=3,α2=−12 are the roots of polynomial 2y2−5y−3. Solution (5.3) is depicted in Figure 1.
Suppose that n=2, x0=0 and let us consider the fractional Riccati equation:
(CD(1/2))2y=2y2−5y−3. | (5.4) |
As stated earlier, the initial conditions read:
CD(1/2)y|x=0=v1;y(0)=u(0)0. | (5.5) |
The solution to (5.4), (5.5) is constructed using the algorithm described in section 4.3. Note that since √x possesses two branches, two solutions exist. The real and imaginary parts of these solutions are shown in Figure 2. Note that solution is purely real for x≥0 – the imaginary parts appears only for negative values of x.
It can be seen in Figure 3 that as v1 approaches zero, the solutions to (5.4) approach the kink solitary solution (5.3).
Taking n=3 results in the fractional Riccati equation:
(CD(1/3))3y=2y2−5y−3. | (5.6) |
The number of initial conditions for (5.6) increases to three:
CD(1/3)y|x=0=v1;(CD(1/3))2y|x=0=v2;y(0)=u(0)0. | (5.7) |
Riccati equation (5.6) has three solutions: two of them are complex-valued and one is a real-valued solution. The real and imaginary parts of these solutions are depicted in Figure 4. A spatial plot for two different sets of initial conditions is shown in Figure 5.
Note that the as x→±∞, the fractional solitary solutions approach the non-fractional kink solitary solution. This is due to the properties of the characteristic equation (3.6). Thus, the limits of the fractional and non-fractional coincide as x→±∞.
An analytical framework for the extension of solutions to FDEs into the negative half-line is presented in this paper. The main idea of this extension is the construction of a characteristic differential equation for a given FDE. The solution to the characteristic equation is used to construct a fractional power series solution to the FDE. This solution is then extended using classical Riemann extension techniques to enable the consideration of the solution at a neighbourhood different than the origin x=0. Furthermore, this extension is valid for both positive and negative values of x, which extends the solution to the entire real line.
As demonstrated by computational experiments, the solution to the FDE is complex-valued for negative values of x. Furthermore, there exist n branches of solutions, where n is the denominator of the fractional differentiation order. If the initial conditions (3.20) that correspond to fractional derivatives tend to zero, the solution to the FDE tends to the non-fractional solution of the same equation.
The presented technique provides a solid foundation for the construction of solutions to FDEs on the negative half-line. In particular, this allows to travel backwards in time from the initial conditions of a FDE – such a possibility has not been reported previously. It is well-known that fractional order derivatives help to model memory effects in dynamical systems. Therefore it would be tempting to extend the available coronavirus pandemic models [35,36] by introducing fractional derivatives. However, the results of this paper show that such an extension would eliminate the possibility of the unique backwards extrapolation. These questions as well as the physical interpretation of the extensions to the negative half-line remain a definite objective of future research.
This research was funded by a grant (No. S-COV-20-8) from the Research Council of Lithuania.
All authors declare no conflicts of interest in this paper.
[1] | J. Pearl, Probabilistic reasoning in intelligent systems: Networks of plausible inference, San Francisco: Morgan Kaufmann, 1988. |
[2] | R. E. Neapolitan, Learning bayesian networks, Upper Saddle River: Prentice Hall, 2004. |
[3] |
R. Daly, Q. Shen, S. Aitken, Learning bayesian networks: Approaches and issues, Knowl. Eng. Rev., 26 (2011), 99–157. https://doi.org/10.1017/S0269888910000251 doi: 10.1017/S0269888910000251
![]() |
[4] | P. Parviainen, M. Koivisto, Finding optimal bayesian networks using precedence constraints, J. Mach. Learn. Res., 14 (2013), 1387–1415. https://www.jmlr.org/papers/volume14/parviainen13a/parviainen13a.pdf |
[5] | L. W. Zhang, H. P. Guo, Introduction to bayesian networks, Beijing: Science Press, 2006. |
[6] | N. Friedman, I. Nachman, D. Peér, Learning Bayesian network structure from massive datasets: The "sparse candidate" algorithm, arXiv Preprint, 2013. |
[7] | I. Tsamardinos, L. E. Brown, C. F. Aliferis, The max-min hill-climbing Bayesian network structure learning algorithm, Mach. Learn., 65 (2006), 31–78. https://doi.org/10.1007/s10994-006-6889-7 |
[8] | C. F. Aliferis, A. Statnikov, I. Tsamardinos, S. Mani, X. D. Koutsoukos, Local causal and Markov blanket induction for causal discovery and feature selection for classification part Ⅰ: Algorithms and empirical evaluation, J. Mach. Learn. Res., 11 (2010), 171–234. https://www.jmlr.org/papers/volume11/aliferis10a/aliferis10a.pdf |
[9] | C. F. Aliferis, A. Statnikov, I. Tsamardinos, S. Mani, X. D. Koutsoukos, Local causal and Markov blanket induction for causal discovery and feature selection for classification part Ⅱ: Analysis and extensions, J. Mach. Learn. Res., 11 (2010), 235–284. https://www.jmlr.org/papers/volume11/aliferis10b/aliferis10b.pdf |
[10] |
S. R. de Morais, A. Aussem, A novel Markov boundary based feature subset selection algorithm, Neurocomputing, 73 (2010), 578–584. https://doi.org/10.1016/j.neucom.2009.05.018 doi: 10.1016/j.neucom.2009.05.018
![]() |
[11] | S. Fu, M. C. Desmarais, Markov blanket based feature selection: A review of past decade, In: Proceedings of the World Congress on Engineering, 2010,321–328. |
[12] |
F. Schlüter, A survey on independence-based Markov networks learning, Artif. Intell. Rev., 42 (2014), 1069–1093. https://doi.org/10.1007/s10462-012-9346-y doi: 10.1007/s10462-012-9346-y
![]() |
[13] | J. P. Pellet, A. Elisseeff, Using Markov blankets for causal structure learning, J. Mach. Learn. Res., 9 (2008), 1295–1342. https://www.jmlr.org/papers/volume9/pellet08a/pellet08a.pdf |
[14] |
A. R. Masegosa, S. Moral, A Bayesian stochastic search method for discovering markov boundaries, Knowl.-Based Syst., 35 (2012), 211–223. https://doi.org/10.1016/j.knosys.2012.04.028 doi: 10.1016/j.knosys.2012.04.028
![]() |
[15] | I. Tsamardinos, C. F. Aliferis, Towards principled feature selection: Relevancy, filters and wrappers, In: International Workshop on Artificial Intelligence and Statistics, 2003,300–307. |
[16] | A. Statnikov, N. I. Lytkin, J. Lemeire, C. F. Aliferis, Algorithms for discovery of multiple Markov boundaries, J. Mach. Learn. Res., 14 (2013), 499–566. https://www.jmlr.org/papers/volume14/statnikov13a/statnikov13a.pdf |
[17] |
X. Q. Liu, X. S. Liu, Swamping and masking in Markov boundary discovery, Mach. Learn., 104 (2016), 25–54. https://doi.org/10.1007/s10994-016-5545-0 doi: 10.1007/s10994-016-5545-0
![]() |
[18] | X. Q. Liu, X. S. Liu, Markov blanket and markov boundary of multiple variables, J. Mach. Learn. Res., 19 (2018), 1–50. https://www.jmlr.org/papers/volume19/14-033/14-033.pdf |
[19] |
N. K. Kitson, A. C. Constantinou, Z. G. Guo, Y. Liu, K. Chobtham, A survey of Bayesian network structure learning, Artif. Intell. Rev., 56 (2023), 8721–8814. https://doi.org/10.1007/s10462-022-10351-w doi: 10.1007/s10462-022-10351-w
![]() |
[20] | J. Lemeire, Learning causal models of multivariate systems and the value of it for the performance modeling of computer programs, ASP/VUBPRESS/UPA, 2007. |
[21] |
J. Lemeire, S. Meganck, F. Cartella, T. T. Liu, Conservative independence-based causal structure learning in absence of adjacency faithfulness, Int. J. Approx. Reason., 53 (2012), 1305–1325. https://doi.org/10.1016/j.ijar.2012.06.004 doi: 10.1016/j.ijar.2012.06.004
![]() |
[22] | F. Bromberg, D. Margaritis, Improving the reliability of causal discovery from small datasets using argumentation, J. Mach. Learn. Res., 10 (2009), 301–340. https://www.jmlr.org/papers/volume10/bromberg09a/bromberg09a.pdf |
[23] |
J. M. Peña, R. Nilsson, J. Björkegren, J. Tegnér, Towards scalable and data efficient learning of Markov boundaries, Int. J. Approx. Reason., 45 (2007), 211–232. https://doi.org/10.1016/j.ijar.2006.06.008 doi: 10.1016/j.ijar.2006.06.008
![]() |
[24] |
J. Cheng, R. Greiner, J. Kelly, D. Bell, W. R. Liu, Learning Bayesian networks from data: An information-theory based approach, Artif. Intell., 137 (2002), 43–90. https://doi.org/10.1016/S0004-3702(02)00191-1 doi: 10.1016/S0004-3702(02)00191-1
![]() |
[25] | H. Cramér, Mathematical methods of statistics, New Jersey: Princeton University Press, 1999. |
[26] | S. Kullback, Information theory and statistics, New York: Dover Publications, 1997. |
[27] | L. M. de Campos, A scoring function for learning Bayesian networks based on mutual information and conditional independence tests, J. Mach. Learn. Res., 7 (2006), 2149–2187. https://www.jmlr.org/papers/volume7/decampos06a/decampos06a.pdf |
[28] |
W. G. Cochran, Some methods for strengthening the common χ2 tests, Biometrics, 10 (1954), 417–451. https://doi.org/10.2307/3001616 doi: 10.2307/3001616
![]() |
[29] |
D. N. Lawley, A general method for approximating to the distribution of likelihood ratio criteria, Biometrika, 43 (1956), 295–303. https://doi.org/10.2307/2332908 doi: 10.2307/2332908
![]() |
[30] |
B. S. Hosmane, Improved likelihood ratio tests and pearson chi-square tests for independence in two dimensional contingency tables, Commun. Stat.-Theor. M., 15 (1986), 1875–1888. https://doi.org/10.1080/03610928608829224 doi: 10.1080/03610928608829224
![]() |
[31] |
B. S. Hosmane, Improved likelihood ratio test for multinomial goodness of fit, Commun. Stat.-Theor. M., 16 (1987), 3185–3198. https://doi.org/10.1080/03610928708829566 doi: 10.1080/03610928708829566
![]() |
[32] |
B. S. Hosmane, Smoothing of likelihood ratio statistic for equiprobable multinomial goodness-of-fit, Ann. Inst. Stat. Math., 42 (1990), 133–147. https://doi.org/10.1007/BF00050784 doi: 10.1007/BF00050784
![]() |
[33] |
S. Brin, R. Motwani, C. Silverstein, Beyond market baskets: Generalizing association rules to correlations, Proceedings of the 1997 ACM SIGMOD International Conference on Management of Data, 26 (1997), 265–276. https://doi.org/10.1145/253260.253327 doi: 10.1145/253260.253327
![]() |
[34] |
C. Silverstein, S. Brin, R. Motwani, Beyond market baskets: Generalizing association rules to dependence rules, Data Min. Knowl. Disc., 2 (1998), 39–68. https://doi.org/10.1023/A:1009713703947 doi: 10.1023/A:1009713703947
![]() |
[35] | S. Yaramakala, Fast Markov blanket discovery, Iowa State University, 2004. |
[36] | P. Spirtes, C. Glymour, R. Scheines, Causation, prediction, and search, Cambridge: MIT Press, 2001. |
[37] | S. K. Fu, M. Desmarais, Local learning algorithm for Markov blanket discovery, Advances in Artificial Intelligence, 2007, 68–79. |
[38] |
W. Khan, L. F. Kong, S. M. Noman, B. Brekhna, A novel feature selection method via mining Markov blanket, Appl. Intell., 53 (2023), 8232–8255. https://doi.org/10.1007/s10489-022-03863-z doi: 10.1007/s10489-022-03863-z
![]() |
[39] | D. Koller, M. Sahami, Toward optimal feature selection, In: Thirteen International Conference in Machine Learning, Stanford InfoLab, 1996,284–292. |
[40] | D. Margaritis, S. Thrun, Bayesian network induction via local neighborhoods, Carnegie Mellon University, 1999. |
[41] | D. Margaritis, S. Thrun, Bayesian network induction via local neighborhoods, In: Advances in Neural Information Processing Systems, Morgan Kaufmann, 1999,505–511. |
[42] | I. Tsamardinos, C. F. Aliferis, A. Statnikov, Algorithms for large scale Markov blanket discovery, In: Proceedings of the Sixteenth International Florida Artificial Intelligence Research Society Conference (FLAIRS), 2003,376–381. |
[43] |
X. L. Yang, Y. J. Wang, Y. Ou, Y. H. Tong, Three-fast-inter incremental association Markov blanket learning algorithm, Pattern Recogn. Lett., 122 (2019), 73–78. https://doi.org/10.1016/j.patrec.2019.02.002 doi: 10.1016/j.patrec.2019.02.002
![]() |
[44] |
H. R. Liu, Q. R. Shi, Y. B. Cai, N. T. Wang, L.Y. Zhang, D. Y. Liu, Fast shrinking parents-children learning for markov blanket-based feature selection, Int. J. Mach. Learn. Cyber., 15 (2024), 3553–3566. https://doi.org/10.1007/s13042-024-02108-4 doi: 10.1007/s13042-024-02108-4
![]() |
[45] | K. P. Murphy, Bayes Net Toolbox for matlab, Version: FullBNT-1.0.7, 2007. Available from: https://github.com/bayesnet/bnt |
[46] |
T. Gao, Q. Ji, Efficient score-based Markov blanket discovery, Int. J. Approx. Reason., 80 (2017), 277–293. https://doi.org/10.1016/j.ijar.2016.09.009 doi: 10.1016/j.ijar.2016.09.009
![]() |
[47] | T. Niinimäki, P. Parviainen, Local structure disocvery in Bayesian network, arXiv Preprint, 2012. |
[48] | T. Silander, P. Myllymäki, A simple approach for finding the globally optimal bayesian network structure, arXiv Preprint, 2012. |
[49] |
J. Cussens, M. Bartlett, E. M. Jones, N. A. Sheehan, Maximum likelihood pedigree reconstruction using integer linear programming, Genet. Epidemiol., 37 (2013), 69–83. https://doi.org/10.1002/gepi.21686 doi: 10.1002/gepi.21686
![]() |
[50] | G. Brown, A. Pocock, M. J. Zhao, M. Luján, Conditional likelihood maximisation: A unifying framework for information theoretic feature selection, J. Mach. Learn. Res., 13 (2012), 27–66. https://www.jmlr.org/papers/volume13/brown12a/brown12a.pdf |
[51] | K. T. Fang, J. L. Xu, Statistical distributions, Beijing: Science Press, 1987. |
[52] | N. L. Johnson, S. Kotz, Distributions in statistics: Continuous univariate distributions-2, Boston: John Wiley & Sons, 1970. |
[53] | G. Schwarz, Estimating the dimension of a model, Ann. Stat., 6 (1978), 461–464. https://www.jstor.org/stable/2958889 |
1. | Inga Timofejeva, Zenonas Navickas, Tadas Telksnys, Romas Marcinkevicius, Minvydas Ragulskis, An Operator-Based Scheme for the Numerical Integration of FDEs, 2021, 9, 2227-7390, 1372, 10.3390/math9121372 | |
2. | R. Marcinkevicius, I. Telksniene, T. Telksnys, Z. Navickas, M. Ragulskis, The construction of solutions to CD(1/n) type FDEs via reduction to (CD(1/n))n type FDEs, 2022, 7, 2473-6988, 16536, 10.3934/math.2022905 |