Processing math: 33%
Research article Special Issues

A multi-dimension information fusion-based intelligent prediction approach for health literacy


  • Received: 09 July 2023 Revised: 23 August 2023 Accepted: 05 September 2023 Published: 21 September 2023
  • Health literacy refers to the ability of individuals to obtain and understand health information and use it to maintain and promote their own health. This paper manages to make predictions toward its development degree in society with use of a big data-driven statistical learning method. Actually, such results can be analyzed by discovering latent rules from massive public textual contents. As a result, this paper proposes a deep information fusion-based smart prediction approach for health literacy. Specifically, the latent Dirichlet allocation (LDA) and convolutional neural network (CNN) structures are utilized as the basic backbone to understand semantic features of textual contents. The feature learning results of LDA and CNN can be then mapped into prediction results via following multi-dimension computing structures. After constructing the CNN model, we can input health information into the model for feature extraction. The CNN model can automatically learn valuable features from raw health information through multi-layer convolution and pooling operations. These characteristics may include lifestyle habits, physiological indicators, biochemical indicators, etc., reflecting the patient's health status and disease risk. After extracting features, we can train the CNN model through a training set and evaluate the performance of the model using a test set. The goal of this step is to optimize the parameters of the model so that it can accurately predict health information. We can use common evaluation indicators such as accuracy, precision, recall, etc. to evaluate the performance of the model. At last, some simulation experiments are conducted on real-world data collected from famous international universities. The case study analyzes health literacy difference between China of developed countries. Some prediction results can be obtained from the case study. The proposed approach can be proved effective from the discussion of prediction results.

    Citation: Xiaoyan Zhao, Sanqing Ding. A multi-dimension information fusion-based intelligent prediction approach for health literacy[J]. Mathematical Biosciences and Engineering, 2023, 20(10): 18104-18122. doi: 10.3934/mbe.2023804

    Related Papers:

    [1] Xiaoping Zhang, Jin Li . Composite trapezoidal quadrature for computing hypersingular integrals on interval. AIMS Mathematics, 2024, 9(12): 34537-34566. doi: 10.3934/math.20241645
    [2] Iqbal M. Batiha, Shameseddin Alshorm, Iqbal Jebril, Amjed Zraiqat, Zaid Momani, Shaher Momani . Modified 5-point fractional formula with Richardson extrapolation. AIMS Mathematics, 2023, 8(4): 9520-9534. doi: 10.3934/math.2023480
    [3] Yong Zhang, Xiaobing Bao, Li-Bin Liu, Zhifang Liang . Analysis of a finite difference scheme for a nonlinear Caputo fractional differential equation on an adaptive grid. AIMS Mathematics, 2021, 6(8): 8611-8624. doi: 10.3934/math.2021500
    [4] Li Wen, Feng Yin, Yimou Liao, Guangxin Huang . A greedy average block Kaczmarz method for the large scaled consistent system of linear equations. AIMS Mathematics, 2022, 7(4): 6792-6806. doi: 10.3934/math.2022378
    [5] Wen-Ning Sun, Mei Qin . On maximum residual block Kaczmarz method for solving large consistent linear systems. AIMS Mathematics, 2024, 9(12): 33843-33860. doi: 10.3934/math.20241614
    [6] Tingting Ma, Yuehua He . An efficient linearly-implicit energy-preserving scheme with fast solver for the fractional nonlinear wave equation. AIMS Mathematics, 2023, 8(11): 26574-26589. doi: 10.3934/math.20231358
    [7] Shuqi He, Kun Wang . Exponential integrator method for solving the nonlinear Helmholtz equation. AIMS Mathematics, 2022, 7(9): 17313-17326. doi: 10.3934/math.2022953
    [8] Hassanein Falah, Parviz Darania, Saeed Pishbin . Study of numerical treatment of functional first-kind Volterra integral equations. AIMS Mathematics, 2024, 9(7): 17414-17429. doi: 10.3934/math.2024846
    [9] Cuijie Zhang, Zhaoyang Chu . New extrapolation projection contraction algorithms based on the golden ratio for pseudo-monotone variational inequalities. AIMS Mathematics, 2023, 8(10): 23291-23312. doi: 10.3934/math.20231184
    [10] Duangdaw Rakjarungkiat, Nimit Nimana . An extrapolated fixed-point optimization method for strongly convex smooth optimizations. AIMS Mathematics, 2024, 9(2): 4259-4280. doi: 10.3934/math.2024210
  • Health literacy refers to the ability of individuals to obtain and understand health information and use it to maintain and promote their own health. This paper manages to make predictions toward its development degree in society with use of a big data-driven statistical learning method. Actually, such results can be analyzed by discovering latent rules from massive public textual contents. As a result, this paper proposes a deep information fusion-based smart prediction approach for health literacy. Specifically, the latent Dirichlet allocation (LDA) and convolutional neural network (CNN) structures are utilized as the basic backbone to understand semantic features of textual contents. The feature learning results of LDA and CNN can be then mapped into prediction results via following multi-dimension computing structures. After constructing the CNN model, we can input health information into the model for feature extraction. The CNN model can automatically learn valuable features from raw health information through multi-layer convolution and pooling operations. These characteristics may include lifestyle habits, physiological indicators, biochemical indicators, etc., reflecting the patient's health status and disease risk. After extracting features, we can train the CNN model through a training set and evaluate the performance of the model using a test set. The goal of this step is to optimize the parameters of the model so that it can accurately predict health information. We can use common evaluation indicators such as accuracy, precision, recall, etc. to evaluate the performance of the model. At last, some simulation experiments are conducted on real-world data collected from famous international universities. The case study analyzes health literacy difference between China of developed countries. Some prediction results can be obtained from the case study. The proposed approach can be proved effective from the discussion of prediction results.



    We consider the following hypersingular integral equation on a circle:

    c+2πc=   f(t)sin2ts2dt=g(s),s(c,c+2π). (1.1)

    Equation (1.1), considered the natural integral equation for harmonic problems, aligns with the form derived by Yu [28,29]. While natural integral equations themselves hold limited direct significance for solving boundary value problems (BVPs), they become instrumental in unbounded domain BVPs [8]. By introducing artificial boundaries such as circles or ellipses and applying corresponding natural integral equations on these boundaries, domain decomposition methods and coupled algorithms can be effectively constructed [29].

    Integrals of this kind possess several definitions and we present our definition as follows:

    c+2πc=   f(t)sin2ts2dt=limϵ0{(sϵc+c+2πs+ϵ)f(t)sin2ts2dt8f(s)ϵ} (1.2)

    and

    c+2πc=   f(t)sin2ts2dt=limϵ0{(sϵc+c+2πs+ϵ)f(t)sin2ts2dt4f(s)cotϵ2}. (1.3)

    The equivalent of (1.2) and (1.3) can be similarly obtained in [32].

    The composite middle rectangle rule remains applicable when the singular point coincides with the midpoint of each subinterval, leading to a superconvergence phenomenon. This allows us to derive the error expansion of the error functional. We leverage these asymptotic expansion results to develop an algorithmic scheme for solving hypersingular integral equations.

    The hypersigular integral defined on the integral

    ba=f(t)(ts)2dt=limε0{sεaf(t)(ts)2dt+bs+εf(t)(ts)2dt2f(s)ε},s(a,b), (1.4)

    has been paid much attention recently.

    Hypersingular integrals frequently arise in boundary element methods and various physical problems, including fracture mechanics, elasticity, acoustics, and electromagnetic scattering [31]. Numerous methods have been devised to handle these integrals, such as Gaussian quadrature [9,10], Newton-Cotes rules [3,12,16,17,19,25,26,27,28], transformation methods [5,7,11], and others [2,4,11,24,30]. These studies rigorously analyze trapezoidal-type quadrature formulas for weakly singular, singular (Cauchy principal value), and hypersingular integrals, providing full asymptotic expansions for error analysis [33,34,35,36,37]. Reference [38] explored product integration using a variant of the generalized Euler-Maclaurin summation formula, while [39] applied this formula to study the convergence of weakly singular Fredholm and Volterra integral equations. Hypersingular integral equations in potential theory was studied in [40], and modal analysis of a submerged elastic disk was studied by [43]. Superconvergence results for the hypersingular integral equation was presented in [41]. Singular integral operators [45], finite-part integrals of highly oscillatory functions [42], and hypersingular integrals [44] have been paid much attention in recent years.

    The extrapolation method[13,21,22] is based on the error function as

    T(h)a0=a1h2+a2h4+a3h6+,

    where T(0)=a0 and aj are constants independent of h. In [14], a generalized trapezoidal rule for numerical computation of hypersingular integrals on intervals was introduced, with asymptotic error expansion proven. In [15], the composite trapezoidal rule for Hadamard finite-part integrals with the hypersingular kernel 1/sin2(xs) was discussed, obtaining the main part of the asymptotic error expansion. In [18], two extrapolation algorithms were presented for hypersingular integrals on intervals, proving their convergence rates, which match those of classical middle rectangle rule approximations. In [20], an extrapolation algorithm for supersingular integrals was introduced.

    Before presenting our main results, we first let

    ϕik(t)={1(ki)!11=Mik(τ,t)(τt)2dτ,|t|<1,1(ki)!11Mik(τ,t)(τt)2dτ,|t|>1, (1.5)

    where τ[1,1] and

    Mik(τ,t)=τi(τt)ki=Fi(τ)(τt)ki. (1.6)

    Set J:=()(1,1)(1,), W:C(J)C(1,1) and let

    Tik(τ):=ϕik(τ)+j=1ϕik(2j+τ)+j=1ϕik(2j+τ), (1.7)

    be the linear operator.

    Compared with Gaussian quadrature and Newton-Cotes rules, the extrapolation method is more effective. In the following, the asymptotic error expansion of the middle rectangle rule for Hadamard finite-part integrals on a circle is given, based on the asymptotic error expansion, and an algorithm to solve the hypersingular integrals equation is presented. The asymptotic error expansion is

    En(f,s)=i=0hi2i+1f(i+1)(s)ai(τ), (1.8)

    where ai(τ) are functions independent of h defined as

    ai(τ)={ik=i1(1)k(k+1)!Tki(τ),i1,12Tki(τ),i=0. (1.9)

    This paper provides the asymptotic error expansion of the middle rectangle rule for Hadamard finite-part integrals on circles. Based on this expansion, an algorithm for solving hypersingular integral equations is proposed. To avoid computing ai(τ), an extrapolation algorithm is suggested for a given τ. A series of sj values approximate the singular point s with mesh refinement. Using extrapolation techniques, we achieve higher-order accuracy and obtain a posteriori error estimates. Additionally, a collocation scheme is constructed to solve hypersingular integral equations via extrapolation methods, with proven convergence rates.

    Let c=t0<t1<<tn1<tn=c+2π be a uniform partition of the interval [c,c+2π] with mesh size h=2π/n and set

    ti=t0+(i1)h,i=1,2,,n, (2.1)
    ˆti=ti+h2,i=1,2,,n. (2.2)

    We define fC(t), the constant interpolation for f(t), as

    fC(t)=f(ˆti),  i=1,,n, (2.3)

    and a linear transformation

    t=ˆti(τ):=(τ+1)(ti+1ti)/2+ti,i=1,,n1,  τ[1,1], (2.4)

    from the reference element [1,1] to the subinterval [ti,ti+1].

    Replacing f(t) in (1.1) with fC(t) produces the composite rectangle rule:

    In(f,s):=c+2πc=   fC(t)sin2ts2dx=ni=1ωi(s)f(ti)=I(f;s)En(f,s), (2.5)

    where ωi(s) denotes the Cotes coefficients given by

    ωi(s)=2cot(ti1s2)2cot(tis2) (2.6)

    and En(f,s) is the error functional.

    Let

    γ(τ)=γ(h,s)=min1in|sti|h=1|τ|2 (2.7)

    and

    In,i(s)={titi1tˆtisin2ts2dt,im,tmtm1=      tˆtmsin2ts2dt,i=m. (2.8)

    We also define

    Fi(τ)=τi (2.9)

    and

    ϕi,i+1(t)={1211  Fi(τ)τtdτ,|t|<1,1211Fi(τ)τtdτ,|t|>1. (2.10)

    If Fi(τ) is replaced by the Legendre polynomial, it is known that ϕi,i+1 defines the Legendre function of the second kind [1]. Let

    ϕii(t)={1211= Fi(τ)(τt)2dτ,|t|<1,1211Fi(τ)(τt)2dτ,|t|>1 (2.11)

    and

    ϕik(t)=1211Fi(τ)(τt)ki2(ki)!dτ,  k>i+1. (2.12)

    Define Ks(t) as

    Ks(t)={(ts)2sin2ts2ts,4,t=s. (2.13)

    Theorem 2.1. Assume f(t)C[c,c+2π]. For the middle rectangle rule In(f;s) defined in (2.5) and ai(τ) defined in (1.9), we have

    En(f,s)=i=0hi2i+1f(i+1)(s)ai(τ) (2.14)

    where s=tm1+(1+τ)h/2,  m=1,2,,n.

    Then the error function can be written as

    En(f,s)=f(s)2a0(τ)+hf(s)22a1(τ)++hi2i+1f(i+1)(s)ai(τ)+ (2.15)

    From the error functional, we know that it is not convergence when the first term is a0(τ). By the first part of a0(τ)=0, we have the convergence rate O(h) which is called the superconvergence phenomenon.

    Define

    Mjik(t,s)=(ttj1/2)i(ts)ki=Fji(t)(ts)ki,  ki, (3.1)

    where

    Fji(t)=(ttj1/2)i. (3.2)

    By (2.4), we have

    Mjik(t,s)=hk+22k+2τi(τcj)ki=hk+22k+2Mik(τ,cj)=hk+22k+2Fi(τ)(τcj)ki, (3.3)

    where

    Mik(τ,cj)=τi(τcj)ki=Fi(τ)(τcj)ki, (3.4)
    cj=2(stj1)/h1 (3.5)

    and Fi(τ) is defined as in(2.9).

    Lemma 3.1. Let Ks(t) be defined as in (2.13). For t(tj1,tj), by linear transformation (2.6), we have

    Ks(t)=Kcj(τ),  τ(1,1) (3.6)

    where

    Kcj(τ)=4+4l=1(τcj)2(τcj2ln)2+4l=1(τcj)2(τcj+2ln)2 (3.7)

    and cj is defined as in (3.5).

    Proof. By the identity in [1],

    π2sin2πt=l=l=1(t+l)2, (3.8)

    and then we get

    1sin2ts2=4(ts)2+4l=11(ts2lπ)2+4l=11(ts+2lπ)2 (3.9)

    and

    (ts)2sin2ts2=4+4l=1(ts)2(ts2lπ)2+4l=1(ts)2(ts+2lπ)2. (3.10)

    Then

    Ks(t)=(ts)2sin2ts2=4+4l=1(τcj)2(τcj4lπ/h)2+4l=1(τcj)2(τcj+4lπ/h)2=4+4l=1(τcj)2(τcj2ln)2+4l=1(τcj)2(τcj+2ln)2=Kcj(τ),

    which completes the proof.

    Lemma 3.2. (See [18, Lemma 3.2]) Let Pn(t),n=0,1,, be the Legendre function [1] defined as

    Pn(t)=12n[n2]r=0(1)r(2n2r)!r!(nr)!(n2r)!tn2r. (3.11)

    For the polynomial tn,n=0,1,, it can be expanded by the Legendre function as

    tn=n!2n[n2]k=0(2n4k+1)Pn2k(t)k!(32)nk, (3.12)

    where (a)n=(a)(a+1)(a+n1).

    Lemma 3.3. Let ϕi,i+1(t) and ϕii(t) be defined as in (2.10) and (2.11), respectively. Then

    ϕi,i+1(t)={i1+1j=1ω2j1Q2j1(t),i=2i1,i1j=0ω2jQ2j(t),i=2i11 (3.13)

    and

    ϕii(t)={Q0(t)+i1j=1ajQ2j(t),i=2i1,Q1(t)+i1j=1bjQ2j1(t),i=2i11, (3.14)

    where

    ωj=2j+1211Fi(τ)Pj(τ)dτ (3.15)

    and

    aj=(4j+1)jk=1ω2k1,
    bj=(4j1)jk=1ω2k2.

    Proof. For i=2i1,

    Fi(τ)=τi

    leads to

    Fi(τ)=i1+1j=1ω2j1P2j1(τ), (3.16)

    where ω2j1 is defined as in (3.15) and Pj(τ) are Legendre polynomials. The first part of (3.13) follows immediately from the definition of ϕi,i+1(τ). Since

    i1+1j=1ω2j1=i1+1j=1ω2j1P2j1(1)=Fi(1)=1,

    then we have

    ϕi,i+1(t)=i1j=1aj4j+1[Q2j+1(t)Q2j1(t)]

    with aj=(4j+1)jk=1ω2k1, which leads to the first part of (3.14) by using the recurrence relation [1].

    Pl+1(t)Pl1(t)=(2l+1)Pl(t),  l=1,2,, (3.17)

    which completed the proof.

    Lemma 3.4. Assume s(tm1,tm), for some m, cj defined as in (3.5). Then, we have

    ψik(cj)={2khk+1tmtm1=      Mjik(t,s)sin2ts2dt,j=m,2khk+1tjtj1Mjik(t,s)sin2ts2dt,jm. (3.18)

    Proof. By the equation of (1.2),

    tmtm1=      Mjik(t,s)sin2ts2dt (3.19)
    =tmtm1=      Fmi(t)Ks(t)(ts)2dt=limε0{sεtm1Fmi(t)Ks(t)(ts)2dt+tms+εFmi(t)Ks(t)(ts)2dt2Fmi(s)Ks(s)ε}=hk+12k+1limε0{(cm2εh1+1cm+2εh)Fi(τ)Ks(τ)(τcm)2dτhFi(cm)Ks(cm)ε}=hk+12k+111= Fi(τ)Ks(τ)(τcm)2dτ=hk+12kψii(cm). (3.20)

    The second identity can be obtained similarly.

    Lemma 3.5. Under the assumption of Lemma 3.4, then we have

    ψik(cj)={2khk+1tmtm1=      Mmik(t,s)sin2ts2dt,j=m,2khk+1tjtj1Mjik(t,s)sin2ts2dt,jm. (3.21)

    Proof. We have:

    tmtm1=      Mmik(t,s)sin2ts2dt=tmtm1       Fmi(t)Ks(t)tsdt=limε0{sεtm1Fmi(t)Ks(t)tsdt+tms+εFmi(t)Ks(t)tsdt}=hk+12k+1limε0{(cmε1+1cm+ε)Fi(τ)Ks(τ)τcmdτ}=hk+12k+111 Fi(τ)Ks(τ)τcmdτ=hk+12kψi,i+1(cm). (3.22)

    The second part of this Lemma 3.4 can be obtained similarly.

    Lemma 3.6. For k>i+1, it holds that

    ψik(cj)=2khk+1tjtj1=   Mjik(t,s)(ki)!sin2ts2dt. (3.23)

    Proof. The proof of (3.23) can be obtained similarly to Lemma 3.4 or Lemma 3.5.

    Lemma 3.7. Suppose f(t)C[c,c+2π]. If stj, for any j=1,2,,n, then it holds that

    f(t)fC(t)=i=0k=i(1)i+1f(k)(s)i!Mjik(t,s)(ki)!. (3.24)

    Proof. By taking the Taylor expansion for f(tj1/2), we have

    f(tj1/2)=f(t)+i=0f(i)(t)i!(tj1/2t)i

    and thus,

    f(t)fC(t)=i=1(1)i+1f(i)(t)i!(ttj1/2)i (3.25)

    and

    f(i+1)(t)=k=if(k)(s)(ts)ki(ki)!. (3.26)

    Combining (3.25) and (3.26) leads to (3.24).

    Define

    Hj(t)=f(t)fC(t)i=0k=i(1)i+1f(k+1)(s)h(i+1)!Mjik(t,s)(ki)!,  t(tj1,tj). (3.27)

    Proof.

    (tm1c+c+2πtm  )f(t)fC(t)sin2ts2dt=nj=1,jmtjtj1f(t)fC(t)sin2ts2dt=i=0k=i(1)i+1f(k+1)(s)h(i+1)!(ki)!nj=1,jmtjtj1Mjik(t,s)sin2ts2dt. (3.28)

    By (3.27) of Hj(t), we have

    tmtm1=      f(t)fC(t)sin2ts2dt=tmtm1=      Hm(t)sin2ts2dt+i=0k=i(1)i+1f(k+1)(s)h(i+1)!(ki)!tmtm1=      Mmik(t,s)sin2ts2dt. (3.29)

    Putting (3.28) and (3.29) together yields

    c+2πc=     f(t)fC(t)sin2ts2dt=i=1k=i(1)i+1hkf(k+1)(s)(i+1)!(ki)!nj=1ψik(τ)=i=0hi2i+1f(i+1)(s)ai(τ) (3.30)

    where

    ai(τ)={ik=i1(1)k+1(k+1)!Tki(τ),i>0,12Tki(τ),i=0. (3.31)

    The proof is complete.

    For i=1,

    a0(τ)=ni=1titi1=   tˆtisin2ts2dt=2hk=1ni=1{sin[k(tis)]+sin[k(ti1s)]}4k=11kni=1{cos[k(tis)]cos[k(ti1s)]}=4hk=1ni=1sin[k(tis)]=4hj=1nsin[nj(t1s)]=8πj=1sin[j(1+τ)π]=4πcot(1+τ)π2=4πtanτπ2, (3.32)

    where

    ni=1sin[k(tis)]={nsin[k(t1s)],k=nj,0,otherwise

    has been used. When τ=0, we get a0(τ)=0 [6]. This is not the same as the case with a singular point located on the interval. The reason is that on the circle each subinterval is equal so we just consider the case of s located at the middle of the interval.

    At last, we suggest the modified composite middle rectangle rule, denoting by ˜In(f,s), defined by

    ˜In(f,s)=In(f,s)4πf(s)tanτπ2. (3.33)

    Following asymptotic expansion,

    En(f,s)=i=0hi2i+1f(i+1)(s)ai(τ). (4.1)

    For the given s and positive integer n0, we have

    m0:=n0(sc)2π.

    First, we partition [c,c+2π] into n0 equal subintervals to get a mesh denoted by Π1 with mesh size h1=2π/n0. Then Π1 is refined to get mesh Π2 with mesh size h2=h1/2. In this way, a series of meshes {Πj}(j=1,2,) is obtained in which Πj is refined from Πj1 with mesh size denoted by hj. The extrapolation scheme is presented in Table 1.

    Table 1.  Extrapolation scheme of T(j)i.
    T(h1)=T(1)1
    T(h2)=T(2)1 T(1)2
    T(h3)=T(3)1 T(2)2 T(1)3
    T(h4)=T(4)1 T(3)2 T(2)3 T(1)4
    T(h5)=T(5)1 T(4)2 T(3)3 T(2)4 T(1)5

     | Show Table
    DownLoad: CSV

    Define

    sj=s+τ+12hj,  j=1,2,, (4.2)

    and

    T(hj)=I2j1n0(f,sj). (4.3)

    The extrapolation algorithm takes the form of:

    Step one: Compute  T(j)1=T(hj),  j=1,,m.
    Step two:Compute  T(j)i=T(j+1)i1+T(j+1)i1T(j)i12i11,  i=2,,m,    j=1,,mi.

    Theorem 4.1. Under the same condition of Theorem 2.1, for a given τ and (4.2), we have

    |I(f,s)T(j)i|Chi (4.4)

    and

    |T(j+1)i1T(j)i12i11|Chi1.

    Proof. By Eq (4.1), we get

    I(f,s)T(hj)=I(f,s)I(f,sj)+I(f,sj)T(hj)=I(f,s)I(f,sj)+i=0hij2iai(τ)f(i+1)(sj). (4.5)
    I(f;sj)=I(f;s)+I(f;s)τ+12hj+I(f;s)(τ+12hj)2++I(l2)(f;s)(τ+12hj)l2+, (4.6)

    and

    f(i+1)(sj)=f(i+1)(s)+f(i+2)(s)τ+12hj+f(i+3)(s)2!(τ+12hj)2++f(l)(s)(li1)!(τ+12hj)li1+. (4.7)

    Putting (4.5)–(4.7) together, we have

    I(f,s)T(hj)=i=0bi(s,τ)hij, (4.8)

    where

    bi(s,τ)=f(i+1)(s)ik=1ak(τ)2k(τ+12)ik1(ik)!(τ+1)i2ii!I(i)(f,s), (4.9)

    where bi(s,τ) is a constant for a given τ. By (4.8), we also have

    I(f,s)T(hj+1)=i=0bi(s,τ)hij+1. (4.10)

    By (4.8) and (4.10), with hj=2hj+1, we have

    I(f,s)=2T(hj+1)T(hj)+i=2bi(s,τ)(12i11)hij+O(hl1j)=T(j)2+i=2bi(s,τ)(12i11)hij, (4.11)

    which implies

    I(f,s)T(j)2=i=2bi(s,τ)(12i11)hij (4.12)

    and

    T(j)2=2T(hj+1)T(hj). (4.13)

    The accuracy O(h3) can be obtained after performing the extrapolation process.

    In order to directly use the error expansion of (4.1), we presented the following the new partition as c=t00<t01<<t0,n1<t0n=c+2π with t0i=ti+h2,i=0,1,,n1 and

    f0C(t)=f(t0i). (4.14)

    Then we have the approximate formula:

    ˜In(f;s):=c+2πc=    f0C(t)sin2ts2dt (4.15)
    =nj=0˜ωj(s)f(tj)=c+2πc=   f(t)sin2ts2dt˜En(f,s), (4.16)

    and get

    ˜ωi(s)={2cot(t0s2)2cot(t01s2),i=0,2cot(t0is2)2cot(t0,i+1s2),0<i<n,2cot(t0ns2)2cot(tn+1s2),i=n, (4.17)

    and then we find the following theorem similarly as with Theorem 2.1.

    Theorem 4.2. Assume f(t)C[c,c+2π]. For the rectangle rule ˜In(f;s) defined in (4.15) and a2i+1(τ) defined in (1.9), independent of h and s, such that

    ˜En(f,s)=c+2πc=     f(t)f0C(t)sin2ts2dt=i=0h2i22i+1f(i+1)(s)a2i+1(τ) (4.18)

    where s=tm1+(1+τ)h/2,  m=1,2,,n.

    Similarly as extrapolation scheme T(hj), new extrapolation scheme is presented in Table 2.

    ˜T(hj)=˜I2j1n0(f,sj). (4.19)
    Table 2.  Extrapolation scheme of ˜T(j)i.
    ˜T(h1)=˜T(1)1
    ˜T(h2)=˜T(2)1 ˜T(1)2
    ˜T(h3)=˜T(3)1 ˜T(2)2 ˜T(1)3
    ˜T(h4)=˜T(4)1 ˜T(3)2 ˜T(2)3 ˜T(1)4
    ˜T(h5)=˜T(5)1 ˜T(4)2 ˜T(3)3 ˜T(2)4 ˜T(1)5

     | Show Table
    DownLoad: CSV

    Theorem 4.3. Under the asymptotic expansion of Theorem 4.2, we have

    |I(f,s)˜T(j)i|Ch2i (4.20)

    and the posteriori asymptotic error estimate is given by

    |˜T(j+1)i1˜T(j)i14i11|Ch2i2.

    The proof of Theorem 4.3 can be similarly obtained as with Theorem 4.1, so we omitted it here.

    Consider

    14π2π0=   f(t)sin2ts2dt=g(s),  s(0,2π), (5.1)

    with

    2π0g(t)dt=0. (5.2)

    As in [29], under the compatibility condition (5.2), there exists a unique solution up to an additive constant for (5.1). With periodical condition

    2π0f(t)dt=0, (5.3)

    there is a unique solution of (5.1).

    Applying ˜In(f,s) to approximate the hypersingular integral in (5.1) and using ˆtk=tk1+h/2(k=1,2,,n) of each subinterval to be a collocation point, we get

    12πnm=1(cotˆtktm2cotˆtktm12)fm=g(ˆtk),  k=1,2,,n, (5.4)

    denoted by

    AnFan=Gen, (5.5)

    where

    An=(akm)n×n,akm=12π(cotˆtkxm2cotˆtkxm12),k,m=1,2,,n,Fan=(f1,f2,,fn)T,Gen=(g(ˆt1),g(ˆt2),,g(ˆtn))T, (5.6)

    and fk=f(ˆtk)(k=1,2,,n). Obviously, An is a circulant matrix and also a symmetric Toeplitz matrix. Since,

    nm=1akm=12πnm=1(cotˆtktm2cotˆtktm12)=0, (5.7)

    we see that An is singular.

    The regularizing factor γ0n (see Reference [23]) in (5.4) is introduced, which leads to

    {γ0n+12πnm=1(cotˆtktm2cotˆtktm12)fm=g(ˆtk),k=1,2,,n,nm=1fm=0, (5.8)

    where γ0n has the form

    γ0n=12πnk=1g(ˆtk)h. (5.9)

    System (5.8) is denoted as

    An+1Fan+1=Gen+1, (5.10)

    where

    An+1=(0eTnenAn),Fan+1=(γ0nFan),Gen+1=(0Gen), (5.11)

    and en=(1,1,,1n).

    Linear system (5.8) can be rewritten as

    {γ0n+12πnm=1fm+1fmhcotˆtktm2h=g(ˆtk),  k=1,2,,n,12πnm=1fm+1fmhh=0, (5.12)

    where f1=fn+1 has been used. Let vm=(fm+1fm)/h, and we get

    {γ0n+12πnm=1cotˆtktm2vmh=g(ˆtk),k=1,2,,n,12πnm=1vmh=0. (5.13)

    The following lemma has been proved in [23].

    Lemma 5.1. ([23, Theorem 6.2.1, §6.2, Chapter 6]) For Eq (5.13), its solution is

    \begin{equation} v_m = -\frac{h}{2\pi}\sum\limits_{k = 1}^n \cot \frac{t_m-\hat{t}_k}{2} f(\hat{t}_k). \end{equation} (5.14)

    Lemma 5.2. ([6, Lemma 4.2]) Let \mathcal{B}_{n+1} = (b_{ik})_{(n+1)\times (n+1)} be the inverse matrix of \mathcal{A}_{n+1} , defined in (5.10). Then,

    (1) \mathcal{B}_{n+1} is expressed as

    \begin{equation} \mathcal{B}_{n+1} = \left ( \begin{array}{ll} b_{00} & \textbf{B}_1 \\ \textbf{ B}_2& \mathcal{B}_n \end{array} \right), \end{equation} (5.15)

    where

    \begin{eqnarray} \textbf{B}_1& = &(b_{01}, b_{02}, \cdots, b_{0n}), \textbf{B}_2 = (b_{10}, b_{20}, \cdots, b_{n0})^{T}, \end{eqnarray} (5.16)
    \begin{eqnarray} b_{i0}& = &b_{0i} = \frac 1n, 1\leq i \leq n, \end{eqnarray} (5.17)
    \begin{eqnarray} b_{ik}& = & \frac{h^2}{2\pi} \left[ \sum\limits_{m = i}^{n-1} \cot \frac{\hat{t}_k-t_m} 2 - \frac1n \sum\limits_{m = 1}^{n-1} m \cot \frac{\hat{t}_k-t_m}2\right], \end{eqnarray} (5.18)
    \begin{eqnarray} && 1\leq i \leq n-1, 1\leq k \leq n, \\ b_{nk}& = &-\frac{h^2}{2n\pi} \sum\limits_{m = 1}^{n-1} m \cot \frac{\hat{t}_k-t_m}2, 1\leq k \leq n. \end{eqnarray} (5.19)

    (2) \mathcal{B}_n is a Toeplitz and also circulant matrix.

    (3) for i = 1, 2, \cdots, n , we have

    \begin{equation} \sum\limits_{k = 1}^n |b_{ik}| \leq C. \end{equation} (5.20)

    Circulant matrices have unique advantages in matrix operations, such as matrix multiplication. It can greatly save the calculation time and improve the calculation efficiency. The circular matrix only needs to store the elements of the first row (or the first column) to completely determine the entire matrix, because the remaining row (or column) elements are obtained from the cyclic shift of the first row (or column) elements. Circulant matrices have many good mathematical properties, such as the fact that their eigenvalues and eigenvectors can be calculated analytically, which makes it easier to analyze and deal with circulant matrices than general matrices in solving linear equations and eigenvalue problems.

    We present the main theorem of this section.

    Theorem 5.1. The solution of linear system (5.8) or (5.10)is

    \begin{equation} f(\hat{t}_i)-f_i = \sum\limits_{i = 0}^{\infty}\frac{h^{2i}}{2^{2i+1}}f^{(i+1)}(s)a_{2i+1}(\tau). \end{equation} (5.21)

    Proof. Let {\bf F}_{n+1}^e = (0, f(\hat{t}_1), f(\hat{t}_2), \cdots, f(\hat{t}_n))^T be the exact vector. Then, from (5.10), we get

    \begin{equation} {\bf F}_{n+1}^e - {\bf F}_{n+1}^a = \mathcal{B}_{n+1}(\mathcal{A}_{n+1} {\bf G}_{n+1}^e - {\bf G}_{n+1}^e), \end{equation} (5.22)

    which implies

    \begin{equation} \begin{array}{ll} f(\hat{t}_i)-f_i& = b_{i0} \sum\limits_{m = 1}^n g(\hat{t}_m)+\sum\limits_{k = 1}^n b_{ik}\tilde{E}_{n}(f, \hat{t}_k) \\ & = \frac1{2\pi} \sum\limits_{m = 1}^n g(\hat{t}_m) h + \sum\limits_{k = 1}^nb_{ik} \tilde{E}_{n}(f, \hat{t}_k) , \end{array} \end{equation} (5.23)

    where \{b_{ik}\} are the entries of \mathcal{B}_{n+1} and \tilde{E}_{n}(f, s) is defined in (4.15).

    For the first part of \frac1{2\pi} \sum_{m = 1}^n g(\hat{t}_m) h of the rectangle rule to compute the definite integral \frac1{2\pi} \int_{0}^{2\pi} g(t) dt .

    Then we have

    \begin{equation} \begin{array}{ll} f(\hat{t}_i)-f_i& = \sum\limits_{k = 1}^nb_{ik} \tilde{E}_{n}(f, \hat{t}_k) \\ & = \sum\limits_{k = 1}^nb_{ik} \sum\limits_{i = 0}^{\infty}\frac{h^{2i}}{2^{2i+1}}f^{(i+1)}(\hat{t}_k)a_{2i+1}(\tau) \\ & = \sum\limits_{i = 0}^{\infty}\frac{h^{2i}}{2^{2i+1}}a_{2i+1}(\tau)\sum\limits_{k = 1}^nb_{ik}f^{(i+1)}(\hat{t}_k). \end{array} \end{equation} (5.24)

    By (5.24) and (5.20), we obtain

    b_{ik} = \frac{c}{n}+O(h^3)

    and

    \sum\limits_{k = 1}^nb_{ik}f^{(i+1)}(\hat{t}_k) = C+O(h^2).

    Then we have

    \begin{array}{ll} f(\hat{t}_i)-f_i& = \sum\limits_{i = 0}^{\infty}\frac{h^{2i}}{2^{2i+1}}a_{2i+1}(\tau), \end{array}

    where (5.19) and (5.20) has been used.

    Example 6.1. We consider (1.1) with density function f(t) = 1+2\cos(t)+ 2\cos(2t), c = -\pi, c+2\pi = \pi .

    Numerical results are presented in Table 3 for s = t_{[n/4]}+(1+\tau)h/2 . The right half of Table 3 shows that the accuracy of \tilde{I}_{n}(s, f) is always O(h^2) . However, from the left half, the accuracy is O(h^2) if s is located at the superconvergence point (\tau = 0) , and if \tau\ne 0 , the middle rectangle rule is divergent.

    Table 3.  Errors of I_{n}(s, f) and \tilde{I}_{n}(s, f) with s = t_{[n/4]}+(1+\tau)h/2.
    I_{n}(f, s) \hat{I}_{n}(f, s)
    n \tau=0 \tau=2/3 \tau=-2/3 \tau=-2/3 \tau=2/3
    32 5.3695e-004 -1.2446e+000 1.0651e+000 1.8350e-003 -7.6074e-004
    64 1.2851e-004 -1.1295e+000 1.0421e+000 4.5684e-004 -1.9980e-004
    128 3.1416e-005 -1.0741e+000 1.0310e+000 1.1397e-004 -5.1140e-005
    512 7.7655e-006 -1.0470e+000 1.0256e+000 2.8464e-005 -1.2933e-005
    1024 1.9303e-006 -1.0336e+000 1.0229e+000 7.1123e-006 -3.2516e-006

     | Show Table
    DownLoad: CSV

    Example 6.2. We still consider f (t) = 1+2\cos(t)+ 2\cos(2t), c = -\pi, c+2\pi = \pi . For s = -\pi/2 , and g(s) = 5.026548245744136e+001 , we use s = t_{[n/4]}+(\tau+1) h/2 with \tau = 0 to compute s = -\pi/2 .

    For s = 0 and g(s) = -7.539822368615504e+01 , we also use s = t_{[n/2]}+(\tau+1) h/2 with \tau = 0 to compute s = 0 .

    In Tables 4 and 6, we present the error of the middle rectangle rule as h^2 , h^{4} , h^{6} , and h^{8} , respectively. In Tables 5 and 7, the convergence rates h^2 , h^{4} , and h^{6} also agree with the theoretic analysis of Theorem 4.2.

    Table 4.  Error estimate of the mid-rectangle rule s = -\pi/2.
    0 h^4-extra h^6-extra h^8-extra
    8 5.0106e+000
    16 1.2820e+000 1.9666 3.9121e-002
    32 3.2236e-001 1.9917 2.4790e-003 3.9801 3.6167e-005
    64 8.0707e-002 1.9979 1.5547e-004 3.9950 5.6990e-007 5.9878 4.8600e-009
    128 2.0184e-002 1.9995 9.7253e-006 3.9988 8.9188e-009 5.9977 1.4339e-011 8.4156
    256 5.0464e-003 1.9999 6.0796e-007 3.9997 1.3873e-010 6.0088 -6.3238e-013 4.0491

     | Show Table
    DownLoad: CSV
    Table 5.  A posteriori estimate of the mid-rectangle rule s = -\pi/2.
    0 h^4-extra h^6-extra h^8-extra
    8
    16 1.2429e+000
    32 3.1988e-001 1.9581 2.4428e-003
    64 8.0551e-002 1.9896 1.5490e-004 3.9791 5.6504e-007
    128 2.0174e-002 1.9974 9.7164e-006 3.9948 8.9044e-009 5.9877 1.9003e-011
    256 5.0458e-003 1.9993 6.0782e-007 3.9987 1.3937e-010 5.9976 5.8710e-014 8.3268

     | Show Table
    DownLoad: CSV
    Table 6.  Error estimate estimate of the mid-rectangle rule s = 0.
    0 h^4-extra h^6-extra h^8-extra
    8 -5.6517e+00
    16 -1.4432e+00 1.9694 -4.0361e-02
    32 -3.6271e-01 1.9924 -2.5567e-03 3.9806 -3.6452e-05
    64 -9.0799e-02 1.9981 -1.6033e-04 3.9951 -5.7436e-07 5.9879 -4.8673e-09
    128 -2.2707e-02 1.9995 -1.0029e-05 3.9988 -9.0075e-09 5.9947 -3.3666e-11 7.1899
    256 -5.6773e-03 1.9999 -6.2699e-07 3.9996 -1.6669e-10 5.7493 -2.6361e-11 -

     | Show Table
    DownLoad: CSV
    Table 7.  A posteriori estimate of the mid-rectangle rule s_j = s+(\tau+1) h_j/2.
    0 h^4-extra h^6-extra h^8-extra
    8
    16 -1.4028e+00
    32 -3.6016e-01 1.9616 -2.5203e-03
    64 -9.0638e-02 1.9904 -1.5976e-04 3.9796 -5.6949e-07
    128 -2.2697e-02 1.9976 -1.0020e-05 3.9949 -8.9738e-09 5.9878 -1.8955e-11
    256 -5.6766e-03 1.9994 -6.2682e-07 3.9987 -1.4033e-10 5.9990 -2.8645e-14 9.6078

     | Show Table
    DownLoad: CSV

    In Tables 8 and 10, we present the error of the middle rectangle rule as h , h^{2} , h^{3} , and h^{4} , respectively. In Tables 9 and 11, the convergence rates h , h^{2} , and h^{3} also agree with the theoretic analysis of Theorem 2.1.

    Table 8.  Error estimate of the mid-rectangle rule s = -\pi/2.
    0 h^2-extra h^3-extra h^4-extra
    32 8.1836e+00
    64 3.7710e+00 1.1178 -6.4164e-01
    128 1.7978e+00 1.0687 -1.7536e-01 1.8714 -1.9940e-02
    256 8.7613e-01 1.0370 -4.5544e-02 1.9450 -2.2699e-03 3.1350 2.5439e-04
    512 4.3227e-01 1.0192 -1.1587e-02 1.9747 -2.6881e-04 3.0779 1.7053e-05 3.8989
    1024 2.1467e-01 1.0098 -2.9214e-03 1.9879 -3.2661e-05 3.0410 1.0752e-06 3.9874

     | Show Table
    DownLoad: CSV
    Table 9.  A posteriori estimate of the mid-rectangle rule s_j = s+(\tau+1) h_j/2.
    0 h^2-extra h^3-extra h^4-extra
    32
    64 4.4126e+00
    128 1.9732e+00 1.1611 -1.5542e-01
    256 9.2167e-01 1.0982 -4.3274e-02 1.8447 -2.5243e-03
    512 4.4386e-01 1.0542 -1.1319e-02 1.9348 -2.8587e-04 3.1425 1.5822e-05
    1024 2.1760e-01 1.0284 -2.8887e-03 1.9702 -3.3736e-05 3.0830 1.0652e-06 3.8927

     | Show Table
    DownLoad: CSV
    Table 10.  Error estimate of the mid-rectangle rule s = -\pi.
    0 h^2-extra h^3-extra h^4-extra
    32 -4.5518e+00
    64 -1.9715e+00 1.2071 6.0874e-01
    128 -9.0311e-01 1.1263 1.6531e-01 1.8807 1.7498e-02
    256 -4.3016e-01 1.0700 4.2802e-02 1.9494 1.9665e-03 3.1535 -2.5230e-04
    512 -2.0964e-01 1.0369 1.0874e-02 1.9768 2.3104e-04 3.0894 -1.6885e-05 3.9013
    1024 -1.0345e-01 1.0190 2.7394e-03 1.9889 2.7950e-05 3.0472 -1.0634e-06 3.9890

     | Show Table
    DownLoad: CSV
    Table 11.  A posteriori estimate of the mid-rectangle rule s_j = s+(\tau+1) h_j/2.
    0 h^2-extra h^3-extra h^4-extra
    32
    64 -2.5803e+00
    128 -1.0684e+00 1.2720 1.4781e-01
    256 -4.7296e-01 1.1757 4.0835e-02 1.8559 2.2188e-03
    512 -2.2051e-01 1.1008 1.0643e-02 1.9400 2.4793e-04 3.1618 -1.5694e-05
    1024 -1.0619e-01 1.0542 2.7115e-03 1.9727 2.9013e-05 3.0951 -1.0548e-06 3.8952

     | Show Table
    DownLoad: CSV

    Example 6.3. In this example, we will not only test our method on a hypersingular integral, but also solve a hypersingular integral equation g(s) = -2\cos(2s)- 2\sin(2s), c = -\pi, c+2\pi = \pi , while the analysis solution is f(t) = \cos(2t)+ \sin(2t) . The good numerical results indicate that our quadrature is efficient and accurate, which matches the theoretical analysis.

    In Table 12, we list the numerical results at s = 1.45122657606971 . Numerical results show that the convergence rate of the hypersingular integral equation has the order O(h^2) . With extrapolation methods, the order can reach O(h^4) , O(h^6) , and O(h^8) in Table12.

    Table 12.  Error estimate of a hypersingular integral equation.
    h^2- h^4-extra h^6-extra h^8-extra
    16 3.70E-02
    32 9.13E-03 1.67E-04
    64 2.27E-03 1.03E-05 -1.70E-07
    128 5.68E-04 6.39E-07 -2.61E-09 4.19E-11
    256 1.42E-04 3.99E-08 -4.06E-11 1.72E-13

     | Show Table
    DownLoad: CSV

    The article mainly focuses on a class of hypersingular integral equations in the boundary element method. A solution centered on the extrapolation algorithm is proposed, which is based on the asymptotic expansion of the error of the composite rectangle rule. By leveraging the asymptotic characteristics of the error of the composite rectangle rule, an extrapolation algorithm is constructed, providing a new approach to solving hypersingular integral equations. The algorithm is verified through in-depth theoretical and numerical examples. It is confirmed that the main part of the error function of the composite rectangle rule has an asymptotic expansion. Numerical experiments show that special functions significantly affect the convergence rate of the algorithm. This discovery has important guiding significance for subsequent algorithm optimization and improvement of computational efficiency. The adoption of this algorithm has two prominent advantages. First, it can obtain high-precision calculation results, meeting the precision requirements of practical applications; second, it is convenient for deriving a posteriori error estimates.

    \int_{a}^{b} \hskip-.22in = — —- —- hypersingular integrals

    \int_{a}^{b} \hskip-.22in - — —- —- Cauchy principal integral

    {I}_{n}(f; s) — —- —- quadrature for hypersingular integrals

    {E}_{n}(f; s) — —- —- error functional of the quadrature for {I}_{n}(f; s)

    \gamma(\tau) — —- —- distance of a singular point to the mesh point

    \eta(y) — —- —- distance of singular point s to the boundary point

    \omega_i(s) — —- —- Cote coefficients of {I}_{n}(f; s)

    \phi_{ik}(t) — —- —- special function of {E}_{n}(f; s)

    a_{i}(\tau) — —- —- special function in {E}_{n}(f; s)

    \gamma_{0n} — —- —- the regularization factor

    T(h_j) — —- —- extrapolation value

    \mathcal{H}_j(t) — —- —- errors of

    \begin{equation*} \mathcal{H}_j(t) = f(t)-f_C(t) - \sum\limits_{i = 0}^{\infty} \sum\limits_{k = i}^{\infty} \frac{(-1)^{i+1}f^{(k+1)}(s)}{h (i+1)!}\frac{M_{ik}^j(t, s)}{(k-i)!}, \ \ t\in (t_{j-1}, t_j), \end{equation*}

    subinterval [t_{m}, t_{m+1}] .

    Qian Ge performed the data analysis; Jin Li performed the formal analysis and wrote the manuscript. All authors have read and approved the final version of the manuscript for publication.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors gratefully acknowledge the helpful comments and suggestions of the reviewers, which have improved the presentation.

    The work of Jin Li was supported by the Natural Science Foundation of Shandong Province (Grant No. ZR2022MA003).

    The authors declare no conflicts of interest.



    [1] Z. Guo, Y. Shen, S. Wan, W. Shang, K. Yu, Hybrid intelligence-driven medical image recognition for remote patient diagnosis in internet of medical things, IEEE J. Biomed. Health Inf., 26 (2022), 5817–5828. https://doi.org/10.1109/JBHI.2021.3139541 doi: 10.1109/JBHI.2021.3139541
    [2] Q. Li, L. Liu, Z. Guo, P. Vijayakumar, F. Taghizadeh-Hesary, K. Yu, Smart assessment and forecasting framework for healthy development index in urban cities, Cities, 131 (2022), 103971. https://doi.org/10.1016/j.cities.2022.103971 doi: 10.1016/j.cities.2022.103971
    [3] N. C. Oleck, A. R. Johnson, B. N. N. Tran, H. S. Ayyala, E. S. Lee, B. T. Lee, A multimetric health literacy analysis of online information for gluteal augmentation with fat grafting, Ann. Plast. Surg., 85 (2020), S97–S101. https://doi.org/10.1097/SAP.0000000000002425 doi: 10.1097/SAP.0000000000002425
    [4] L. Yang, Y. Li, S. X. Yang, Y. Lu, T. Guo, K. Yu, Generative adversarial learning for intelligent trust management in 6G wireless networks, IEEE Network, 36 (2022), 134–140. https://doi.org/10.1109/MNET.003.2100672 doi: 10.1109/MNET.003.2100672
    [5] L. Zhao, Z. Yin, K. Yu, X. Tang, L. Xu, Z. Guo, et al., A fuzzy logic based intelligent multi-attribute routing scheme for two-layered SDVNs, IEEE Trans. Network Serv. Manage., 19 (2022), 4189–4200. https://doi.org/10.1109/TNSM.2022.3202741 doi: 10.1109/TNSM.2022.3202741
    [6] G. Kampouroglou, V. S. Velonaki, I. Pavlopoulou, E. Drakou, M. Kosmopoulos, N. Kouvas, et al., Parental anxiety in pediatric surgery consultations: The role of health literacy and need for information, J. Pediatric Surg., 55 (2020), 590–596. https://doi.org/10.1016/j.jpedsurg.2019.07.016 doi: 10.1016/j.jpedsurg.2019.07.016
    [7] C. C. Cutilli, Health Literacy, Health Disparities and Sources of Health Information in US Older Adults, Ph.D thesis, Duquesne University, 2015.
    [8] Z. Guo, K. Yu, N. Kumar, W. Wei, S. Mumtaz, M. Guizani, Deep distributed learning-based poi recommendation under mobile edge networks, IEEE Internet Things J., 10 (2023), 303–317. https://doi.org/10.1109/JIOT.2022.3202628 doi: 10.1109/JIOT.2022.3202628
    [9] J. Zhang, L. Zhao, K. Yu, G. Min, A. Y. Al-Dubai, A. Y. Zomaya, A novel federated learning scheme for generative adversarial networks, IEEE Trans. Mobile Comput., 2023 (2023). https://doi.org/10.1109/TMC.2023.3278668
    [10] B. St. Jean, N. Greene Taylor, C. Kodama, M. Subramaniam, Assessing the health information source perceptions of tweens using card-sorting exercises, J. Inf. Sci., 44 (2018), 148–164. https://doi.org/10.1177/0165551516687728 doi: 10.1177/0165551516687728
    [11] B. Parlak, A. K. Uysal, The effects of globalisation techniques on feature selection for text classification, J. Inf. Sci., 47 (2021), 727–739. https://doi.org/10.1177/0165551520930897 doi: 10.1177/0165551520930897
    [12] K. Saho, K. Sugano, M. Kita, K. Uemura, M. Matsumoto, Classification of health literacy and cognitive impairments using higher-order kinematic parameters of the sit-to-stand movement from a monostatic doppler radar, IEEE Sensors J., 21 (2021), 10183–10192. https://doi.org/10.1109/JSEN.2021.3060050 doi: 10.1109/JSEN.2021.3060050
    [13] Y. J. Yi, B. Hwang, H. Yoon, H. Jeong, Health literacy and health information-seeking behavior of immigrants in south korea, Library Inf. Sci. Res., 43 (2021), 101121. https://doi.org/10.1016/j.lisr.2021.101121 doi: 10.1016/j.lisr.2021.101121
    [14] Q. He, Z. Feng, H. Fang, X. Wang, L. Zhao, Y. Yao, et al., A blockchain-based scheme for secure data offloading in healthcare with deep reinforcement learning, IEEE/ACM Trans. Networking, 2023 (2023). https://doi.org/10.1109/TNET.2023.3274631
    [15] A. L. Neves, L. Freise, L. Laranjo, A. W. Carter, A. Darzi, E. Mayer, Impact of providing patients access to electronic health records on quality and safety of care: A systematic review and meta-analysis, BMJ Quality Saf., 29 (2020), 1019–1032. https://doi.org/10.1136/bmjqs-2019-010581 doi: 10.1136/bmjqs-2019-010581
    [16] A. McNeil, R. Arena, The evolution of health literacy and communication: introducing health harmonics, Prog. Cardiovasc. Dis., 59 (2017), 463–470. https://doi.org/10.1016/j.pcad.2017.02.003 doi: 10.1016/j.pcad.2017.02.003
    [17] Z. Guo, D. Meng, C. Chakraborty, X. R. Fan, A. Bhardwaj, K. Yu, Autonomous behavioral decision for vehicular agents based on cyber-physical social intelligence, IEEE Trans. Comput. Soc. Syst., 10 (2022), 2111–2122. https://doi.org/10.1109/TCSS.2022.3212864 doi: 10.1109/TCSS.2022.3212864
    [18] J. Zhang, Q. Yan, X. Zhu, K. Yu, Smart industrial iot empowered crowd sensing for safety monitoring in coal mine, Digital Commun. Networks, 9 (2023), 296–305. https://doi.org/10.1016/j.dcan.2022.08.002 doi: 10.1016/j.dcan.2022.08.002
    [19] A. De Thurah, Sp0008 the future for health professionals in rheumatology, Ann. Rheum. Dis., 76 (2017), 3. https://doi.org/10.1136/annrheumdis-2017-eular.7155 doi: 10.1136/annrheumdis-2017-eular.7155
    [20] Z. Zhou, Y. Su, J. Li, K. Yu, Q. M. J. Wu, Z. Fu, et al., Secret-to-image reversible transformation for generative steganography, IEEE Trans. Dependable Secure Comput., 2022 (2022). https://doi.org/10.1109/TDSC.2022.3217661
    [21] S. S. Coughlin, J. L. Stewart, L. Young, V. Heboyan, G. De Leo, Health literacy and patient web portals, Int. J. Med. Inf., 113 (2018), 43–48. https://doi.org/10.1016/j.ijmedinf.2018.02.009 doi: 10.1016/j.ijmedinf.2018.02.009
    [22] R. Garcia-Retamero, E. T. Cokely, Designing visual aids that promote risk literacy: A systematic review of health research and evidence-based design heuristics, Hum. Factors, 59 (2017), 582–627. https://doi.org/10.1177/0018720817690634 doi: 10.1177/0018720817690634
    [23] Y. Zhang, Y. Sun, Y. Kim, The influence of individual differences on consumer's selection of online sources for health information, Comput. Hum. Behav., 67 (2017), 303–312. https://doi.org/10.1016/j.chb.2016.11.008 doi: 10.1016/j.chb.2016.11.008
    [24] H. Wang, N. Wang, M. Li, S. Mi, Y. Shi, Student physical health information management model under big data environment, Sci. Program., 2021 (2021), 1–10. https://doi.org/10.1155/2021/5795884 doi: 10.1155/2021/5795884
    [25] Z. Zhou, X. Dong, R. Meng, M. Wang, H. Yan, K. Yu, et al., Generative steganography via auto-generation of semantic object contours, IEEE Trans. Inf. Forensics Secu., 18 (2023), 2751–2765. https://doi.org/10.1109/TIFS.2023.3268843 doi: 10.1109/TIFS.2023.3268843
    [26] H. Xue, D. Chen, N. Zhang, H. Dai, K. Yu, Integration of blockchain and edge computing in internet of things: A survey, Future Gener. Comput. Syst., 144 (2023), 307–326. https://doi.org/10.1016/j.future.2022.10.029 doi: 10.1016/j.future.2022.10.029
    [27] M. Grene, Y. Cleary, A. Marcus-Quinn, Use of plain-language guidelines to promote health literacy, IEEE Trans. Prof. Commun., 60 (2017), 384–400. https://doi.org/10.1109/TPC.2017.2761578 doi: 10.1109/TPC.2017.2761578
    [28] T. P. Liang, Y. H. Liu, Research landscape of business intelligence and big data analytics: A bibliometrics study, Expert Syst. Appl., 111 (2018), 2–10. https://doi.org/10.1016/j.eswa.2018.05.018 doi: 10.1016/j.eswa.2018.05.018
    [29] F. Huang, L. Wang, H. Jia, Research trends for papillary thyroid carcinoma from 2010 to 2019: A systematic review and bibliometrics analysis, Medicine, 100 (2021), e26100. https://doi.org/10.1097/MD.0000000000026100 doi: 10.1097/MD.0000000000026100
    [30] M. M. Mirończuk, J. Protasiewicz, A recent overview of the state-of-the-art elements of text classification, Expert Syst. Appl., 106 (2018), 36–54. https://doi.org/10.1016/j.eswa.2018.03.058 doi: 10.1016/j.eswa.2018.03.058
    [31] S. Burkhardt, S. Kramer, Online multi-label dependency topic models for text classification, Mach. Learn., 107 (2018), 859–886. https://doi.org/10.1007/s10994-017-5689-6 doi: 10.1007/s10994-017-5689-6
    [32] F. Lei, X. Liu, Z. Li, Q. Dai, S. Wang, Multihop neighbor information fusion graph convolutional network for text classification, Math. Prob. Eng., 2021 (2021), 1–9. https://doi.org/10.1155/2021/6665588 doi: 10.1155/2021/6665588
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1747) PDF downloads(49) Cited by(0)

Figures and Tables

Figures(9)  /  Tables(3)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog