Research article

Forecasting the monthly retail sales of electricity based on the semi-functional linear model with autoregressive errors

  • Received: 03 October 2024 Revised: 04 January 2025 Accepted: 14 January 2025 Published: 23 January 2025
  • MSC : 62J05, 62P12

  • In many statistical applications, data are collected sequentially over time and exhibit autocorrelation characteristics. Ignoring this autocorrelation may lead to a decrease in the model's prediction accuracy. To this end, assuming that the error process is an autoregressive process, this paper introduced a semi-functional linear model with autoregressive errors. Based on the functional principal component analysis and the spline method, we obtained the estimators of the slope function, nonparametric function, and autoregressive coefficients. Under some regular conditions, we found the convergence rate of the proposed estimators. A simulation study was conducted to investigate the finite sample performance of the proposed estimators. Finally, we applied our model to forecast the monthly retail sales of electricity, which illustrates the validity of our model from a predictive perspective.

    Citation: Bin Yang, Min Chen, Jianjun Zhou. Forecasting the monthly retail sales of electricity based on the semi-functional linear model with autoregressive errors[J]. AIMS Mathematics, 2025, 10(1): 1602-1627. doi: 10.3934/math.2025074

    Related Papers:

    [1] Lijie Zhou, Liucang Wu, Bin Yang . Estimation and diagnostic for single-index partially functional linear regression model with $ p $-order autoregressive skew-normal errors. AIMS Mathematics, 2025, 10(3): 7022-7066. doi: 10.3934/math.2025321
    [2] Gaosheng Liu, Yang Bai . Statistical inference in functional semiparametric spatial autoregressive model. AIMS Mathematics, 2021, 6(10): 10890-10906. doi: 10.3934/math.2021633
    [3] Juxia Xiao, Ping Yu, Zhongzhan Zhang . Weighted composite asymmetric Huber estimation for partial functional linear models. AIMS Mathematics, 2022, 7(5): 7657-7684. doi: 10.3934/math.2022430
    [4] Yanping Liu, Juliang Yin . B-spline estimation in varying coefficient models with correlated errors. AIMS Mathematics, 2022, 7(3): 3509-3523. doi: 10.3934/math.2022195
    [5] Rabia Noureen, Muhammad Nawaz Naeem, Dumitru Baleanu, Pshtiwan Othman Mohammed, Musawa Yahya Almusawa . Application of trigonometric B-spline functions for solving Caputo time fractional gas dynamics equation. AIMS Mathematics, 2023, 8(11): 25343-25370. doi: 10.3934/math.20231293
    [6] Jin Xie, Xiaoyan Liu, Lei Zhu, Yuqing Ma, Ke Zhang . The C3 parametric eighth-degree interpolation spline function. AIMS Mathematics, 2023, 8(6): 14623-14632. doi: 10.3934/math.2023748
    [7] Fatimah Alshahrani, Wahiba Bouabsa, Ibrahim M. Almanjahie, Mohammed Kadi Attouch . Robust kernel regression function with uncertain scale parameter for high dimensional ergodic data using $ k $-nearest neighbor estimation. AIMS Mathematics, 2023, 8(6): 13000-13023. doi: 10.3934/math.2023655
    [8] Dewang Li, Daming Xu, Meilan Qiu, Shuiping Yang . Forecasting the public financial budget expenditure in Dongguan with an optimal weighted combination Markov model. AIMS Mathematics, 2023, 8(7): 15600-15617. doi: 10.3934/math.2023796
    [9] Salvatore Mancha Gonzales, Hasnain Iftikhar, Javier Linkolk López-Gonzales . Analysis and forecasting of electricity prices using an improved time series ensemble approach: an application to the Peruvian electricity market. AIMS Mathematics, 2024, 9(8): 21952-21971. doi: 10.3934/math.20241067
    [10] Jairo A. Angel, Francisco M.M. Rocha, Jorge I. Vélez, Julio M. Singer . A new test for detecting specification errors in Gaussian linear mixed-effects models. AIMS Mathematics, 2024, 9(11): 30710-30727. doi: 10.3934/math.20241483
  • In many statistical applications, data are collected sequentially over time and exhibit autocorrelation characteristics. Ignoring this autocorrelation may lead to a decrease in the model's prediction accuracy. To this end, assuming that the error process is an autoregressive process, this paper introduced a semi-functional linear model with autoregressive errors. Based on the functional principal component analysis and the spline method, we obtained the estimators of the slope function, nonparametric function, and autoregressive coefficients. Under some regular conditions, we found the convergence rate of the proposed estimators. A simulation study was conducted to investigate the finite sample performance of the proposed estimators. Finally, we applied our model to forecast the monthly retail sales of electricity, which illustrates the validity of our model from a predictive perspective.



    Functional regression modeling has been extensively studied as an important research field in functional data analysis (FDA). For example, to investigate the linear relationship between a scalar response and a functional explanatory variable, comprehensive studies focused on functional linear regression models have been conducted by Manteiga and Calvo [1] and Bande et al. [2]. However, the linear assumption is sometimes too restrictive to describe the data structure comprehensively. To this end, the functional nonparametric regression model, as an extension of the nonparametric model in the functional data setting, has also been proposed (see Ferraty and Vieu [3] and Belarbi et al. [4]). Furthermore, to improve the power of prediction and interpretation of functional regression models, various functional semiparametric models have been proposed by introducing some additional real-valued predictors. For example, Pérez and Vieu [5] proposed a semi-functional partial linear regression model by combining a traditional linear model with a nonparametric treatment of functional data. Shin [6] and Shin and Lee [7] studied the problem of estimation and prediction of the partial functional linear regression model, respectively. Yao et al. [8] extended the partial functional linear regression to quantile regression settings and established a more flexible and robust method. Kong et al. [9] investigated a class of partial function linear regression models for processing multiple functional and non-functional explanatory variables and automatically identifying essential risk factors through appropriate regularization. In addition, Zhou and Chen [10] first introduced the semi-functional linear model (SFL) by combining the features of a functional linear regression model and a traditional nonparametric model and studied its spline estimation. Peng et al. [11] proposed a varying coefficient partially functional linear regression model, which combines a varying coefficient regression model and a functional linear regression model, and investigated the estimation problem based on the polynomial spline method. Novo et al. [12] developed an automatic and location-adaptive estimation procedure for the functional single-index model. For a recent survey on this field, we refer to some literature like Cuevas [13], Goia and Vieu [14], Ling and Vieu [15,16], and Aneiros et al. [17].

    Meanwhile, given that electricity cannot be stored, modeling and forecasting electricity demand and price are of primary interest to electric market agents. Traditional prediction models, like simple time series models, often fail to address the complexities of electricity demand, so functional regression models have been applied to electricity demand forecasting. For example, Vilar et al. [18] utilized functional nonparametric and semi-functional partial linear models to forecast next-day electricity demand and price. Based on the robust functional principal component analysis and nonparametric models with both functional response and covariate, Aneiros et al. [19] proposed two functional regression models to forecast daily curves of electricity demand and price. In addition, using residual-based bootstrap algorithms, Vilar et al. [20] provided two procedures to obtain prediction intervals for electricity demand and price based on the functional nonparametric autoregressive model and the partial linear semiparametric model. Peláez et al. [21] also proposed three model-based procedures to construct bootstrap prediction regions for daily curves of electricity demand and price.

    All the models mentioned above are based on a critical assumption that the random errors are independently and identically distributed. However, the independence assumption may be inappropriate in many applications. For example, the electricity consumption dataset that has been extensively studied by Pérez and Vieu [22], Yu et al. [23], and Yang et al. [24] exhibits a specific autocorrelation structure. If this autocorrelation structure is ignored, it may lead to inaccurate prediction. To this end, many works have been developed to deal with the autocorrelation structure in functional regression models. For example, Niang and Guillas [25] studied parametric estimation of a functional semiparametric partially linear model with autoregressive errors. Zhang et al. [26] proposed a functional polynomial regression with autoregressive errors. Yu et al. [23] considered composite quantile estimation for the partial functional linear regression model with random errors from a short-range dependent and strictly stationary linear process. Xiao and Wang [27] also studied the estimation problem of the partial functional linear model with autoregressive errors. Wang et al. [28] studied the multiple functional linear model with autoregressive errors. Although these functional regression models with autoregressive errors have improved prediction accuracy, the linear assumption on the covariates is sometimes inappropriate. Therefore, it is necessary to develop more general models to fit the more complex data structures. To achieve this goal, by relaxing the linear assumption on the partial functional linear model, a semi-functional linear model with autoregressive errors is developed in this paper. Based on the functional principal component analysis and the spline method, we obtain the estimators of the slope function, nonparametric function, and autoregressive coefficients, respectively.

    The rest of this paper is organized as follows. Section 2 describes the semi-functional linear model with autoregressive errors and presents the estimation method and algorithm. In Section 3, we study the asymptotic properties of the proposed estimators under some regular conditions. A Monte Carlo simulation is presented in Section 4 to illustrate the finite sample performance. The proposed method is illustrated by electricity consumption data in Section 5. Sections 6 and 7 present future prospects and summarize the paper, respectively. All technical details and proofs are provided in the appendix.

    Suppose that (Xt,Yt,Zt),1tn, is a random sample from the semi-functional linear model with autoregressive errors (SFLAR)

    Yt=Tβ(s)Xt(s)ds+g(Zt)+εt,εt=ql=1alεtl+et, (2.1)

    where Yt is a real-valued random response variable and Zt is a real-valued covariate defined on a compact interval Z=[z_,¯z]. The functional covariate Xt(s) defined on an interval TR is a zero mean, second-order (i.e., E|X(s)|2< for all sT) stochastic process defined on (Ω,B,P) with sample paths in the Hilbert space L2(T). The inner product of L2(T) is defined by u,v=Tu(s)v(s)ds, for any u,vL2(T) and norm u=u,u1/2. The unknown slope function β(s) belongs to L2(T). For simplicity, throughout this paper, we assume T=[0,1]. The nonparametric function g() is an unknown smooth function. εt is a random error with zero mean and finite variance independent of (Xt,Zt). et's are independent and identically distributed with mean 0 and variance σ2. Note that model (2.1) includes the partial functional linear model with autoregressive errors studied by Xiao and Wang [27].

    Since the unknown functions β() and g() are infinite-dimensional parameters, it is impossible to obtain their estimators based on finite observation data. Therefore, we need to apply dimension reduction methods. Specifically, if we denote the covariance function of the functional variable X by CX(s,u)=Cov[X(s),X(u)], and assume that CX(s,u) is continuous on the interval [0,1]2, according to the Mercer theorem, we have

    CX(s,u)=i=1λiϕi(s)ϕi(u),

    where λ1>λ2>>0 is the ordered eigenvalue sequence and ϕ1,ϕ2, is the corresponding orthonormal eigenfunction sequence of the covariance operator with kernel CX(s,u). By the Karhunen-Loève representation, functional variables Xt(s) and the slope function β(s) can be respectively expanded as:

    Xt(s)=i=1ξtiϕi(s),β(s)=i=1biϕi(s), (2.2)

    where ξti=10Xt(s)ϕi(s)ds are called the principle component scores satisfying Eξti=0 and Eξtiξtk=λiI(i=k), and coefficients bi=10β(s)ϕi(s)ds. {ϕi(s)}i=1 forms a standard orthogonal basis of L2(T), commonly referred to as the functional principal component (FPC) basis. Truncate Eqs (2.2) at M, and the model (2.1) can be approximated as

    YtMi=1biξti+g(Zt)+εt. (2.3)

    Furthermore, we can approximate the nonparametric function g() using polynomial splines. Specifically, let Gp,Jn be the space of polynomial splines defined on the interval [z_,¯z] with degree p1 and knot sequence z_<z1<<zJn<¯z. The space Gp,Jn is a J-dimensional linear space, where J=J(n)=Jn+p. Following the arguments of de Boor [29], we can conclude that, if it is sufficiently smooth, the nonparametric function g(Zt) can be approximately expressed as

    g(Zt)Jj=1cjBj(Zt),

    where Bj,j=1,2,,J, are the B-spline basis functions in Gp,Jn. Substituting the approximation of g(Zt) into the model (2.3) yields

    YtMi=1biξti+Jj=1cjBj(Zt)+εt. (2.4)

    However, since the covariance function CX(s,u) is unknown, the eigenfunctions ϕi and variables ξti are also unknown and unobservable. To address this issue, we define the empirical version of the covariance function CX(s,u) by

    ˆCX(s,u)=1nnt=1Xt(s)Xt(u).

    By Mercer's theorem, we also have

    ˆCX(s,u)=i=1ˆλiˆϕi(s)ˆϕi(u),

    where (ˆλi,ˆϕi) are the eigenvalue-eigenfunction pairs for the linear operator with kernel ˆCX(s,u), ordered such that ˆλ1ˆλ20. We consider (ˆλi,ˆϕi) and ˆξti=10Xt(s)^ϕi(s)ds as estimators of (λi,ϕi) and ξti, respectively. Then, the estimators ˆβ(s)=Mi=1ˆbiˆϕi(s) and ˆg(Zt)=Jj=1ˆcjBj(Zt) can be obtained by minimizing the following loss function:

    Ln(a,b,c)=1nqnt=q+1[(YtˆξtbBtc)ql=1al(YtlˆξtlbBtlc)]2, (2.5)

    where ^ξt=(ˆξt1,ˆξt2,,ˆξtM), Bt=(B1(Zt),B2(Zt),,BJ(Zt)), a=(a1,a2,,aq), b=(b1,b2,,bM), c=(c1,c2,,cJ), and t=q+1,q+2,,n.

    To implement the proposed estimation method, we must choose the degrees of the spline functions, the positions, and the number of knots and eigenfunctions. This paper uses a B-spline basis with equally spaced knots and the fixed degree 2. Then, we only need to choose the numbers of B-spline functions (J) and eigenfunctions (M). Various methods, such as the Akaike information criterion (AIC) and Bayesian information criterion (BIC), can be used to select the truncation parameters J and M. This paper uses the BIC criterion to choose the number of J and M. The selection of J and M is determined by minimizing the following BIC criteria:

    BIC(M,J)=log(nt=q+1[ˆεtql=1ˆalˆεtl]2nq)+log(nq)nq(M+J+q), (2.6)

    where ˆεt=YtMi=1ˆbiˆξtiJj=1ˆcjBj(Zt).

    This subsection mainly introduces the algorithm to estimate the unknown parameters. For ease of description, we rewrite the loss function (2.5) as

    Ln(a,θ)=1nqnt=q+1[(YtˆUtθ)ql=1al(YtlˆUtlθ)]2=1nqnt=q+1[(Ytql=1alYtl)(ˆUtql=1alˆUtl)θ]2, (2.7)

    where ˆUt=(ˆξt,Bt), ˆξt=(ˆξt1,ˆξt2,,ˆξtM). Since there is no closed-form expression of the minimizer of the loss function (2.7), we adopt the two-step iteratively least-squares (TSILS) algorithm to estimate the unknown parameters. Specifically, we first set a(0)=(0,0,,0) and compute θ(0) by minimizing Ln(a(0),θ). Afterward, we substitute θ(0) into the loss function (2.7) to get a(1) by Ln(a,θ(0)), and substitute a(1) into (2.7) to get θ(1)=argminLn(a(1),θ). Repeat this process until convergence. The TSILS algorithm is summarized as follows:

    Algorithm 1. The TSILS algorithm initialization: Given the initial value: a(0)=(0,0,,0), compute θ(0)=(ˆUˆU)1ˆUY based on the least squares method, where Y=(Y1,Y2,,Yn) and ˆU=(ˆU1,ˆU2,,ˆUn).

    Step 1. Update ˜V(k), ˜H(k):

    ˜V(k)=(ˆε(k)q+1ˆε(k)n),˜H(k)=(ˆε(k)qˆε(k)1ˆε(k)n1ˆε(k)nq),

    where ˆε(k)t=YtˆUtˆθ(k1), t=1,2,,n.

    Step 2. Compute a(k):

    a(k)=(˜H(k)˜H(k))1˜H(k)˜V(k).

    Step 3. Update V(k)t, H(k)t:

    V(k)t=Ytql=1a(k)lYtl,H(k)t=ˆUtql=1a(k)lˆUtl.

    Step 4. Compute θ(k):

    θ(k)=(H(k)H(k))1H(k)V(k),

    where

    V(k)=(V(k)q+1V(k)n),H(k)=(H(k)q+1H(k)n).

    Step 5. Set k=k+1, and repeat the above steps 1–4 until convergence. Then, the estimation of the parameter vectors ˆa=(ˆa1,ˆa2,,ˆaq), ˆb=(ˆb1,ˆb2,,ˆbM), and ˆc=(ˆc1,ˆc2,,ˆcJ) can be obtained.

    In this section, we investigate the theoretical properties of the proposed estimators. First, we assume that the true slope function, nonparametric function, and autoregressive coefficient of the model are β0(t),g0(z), and a0=(a01,a02,,a0q), respectively. For convenience, the following notation is needed. For two sequences of positive numbers an and bn, anbn signifies that an/bn is uniformly bounded and anbn if anbn and bnan. Throughout this paper, the constant Ci may change from line to line. In order to establish the theoretical properties of the estimators, the following assumptions need to be introduced:

    Assumption 1. All roots of the polynomial A(z)=1ql=1alzl lie outside the unit circle in the complex plane.

    Assumption 2. {Xt,Zt}nt=1 and {et}nt=1 are independent random sequences and {Xt,et}nt=1 has the finite fourth-order moment, respectively, i.e. E(X4t(s)ds)<, E(e4t)<, for any 1tn, and random variable ξti satisfy E(ξ4ti)C1λ2i,i1. The density function fZ(z) of the random variable Z is a continuous bounded function defined on the interval [z_,¯z]. Meanwhile, the random variable et is independent of (Xt,Zt) and E(X|Z)=0.

    Assumption 3. The eigenvalues λi(i=1,2,) of the covariance function CX and the score coefficients bi satisfy the following conditions, respectively.

    (1) There exist some strictly positive constants C2 and α>1 such that C12iαλiC2iα and λiλi+1C2iα1,i1.

    (2) There exist some strictly positive constants C3 and γ>α/2+3 such that |bi|C3iγ,i1.

    Assumption 4. Let 0<ν1. The nonparametric function g(z) is a continuous differentiable function of order r, and for z1,z2[z_,¯z], there are |g(r)(z1)g(r)(z2)|C4|z1z2|ν. Define p=r+ν as a smoothness measure of the function g(z), which satisfies pγ+(α1)/2.

    Assumption 5. Define the matrix

    Λ=E{(Utql=1a0lUtl)(Utql=1a0lUtl)},
    V=E(εt1,εt2,,εtq)(εt1,εt2,,εtq),

    where Ut=(ξt,Bt), ξt=(ξt1,ξt2,,ξtM), Bt=(B1(Zt),B2(Zt),,BJ(Zt)), and the matrix V is a positive definite matrix. Denote the minimum eigenvalue of Λ by ρmin(Λ), which satisfies ρmin(Λ)=Op(Mα).

    Assumption 6. The tuning parameter M, J satisfies MJn1/(α+2γ).

    Remark 1. Assumption 1 is an invertibility condition on the autoregression process and ensures εt is a stationary series. Assumption 2 is standard in the literature about the least squares method, which ensures the consistency of CX(t,s). Assumption 3 is established for the consistency of the estimated principal component scores. Assumption 3(1) prevents the spacings among eigenvalues from being too small for identification of the slope function bi. Assumption 3(2) makes the slope function sufficiently smooth relative to the covariance function CX(t,s). Assumption 4 ensures that the nonparametric function g(z) is sufficiently smooth so that the spline function can approach it. Assumption 5 is a necessary condition for the model consistency to be established. Assumption 6 gives the order of the truncation parameters M and J to obtain the convergence speed of ˆβ(s) and ˆg(z), and Assumption 6 requires that the tuning parameters M and J do not grow too fast.

    Theorem 1. Under the Assumptions 1–6, we have:

    ˆaa022=Op(n(α+2γ1)/(α+2γ)),
    ˆβ(s)β0(s)2=Op(n(2γ1)/(α+2γ)),
    ˆg(z)g0(z)2=Op(n(α+2γ1)/(α+2γ)).

    Remark 2. The theorem shows that the convergence rate of the autoregressive coefficients cannot reach the optimal convergence rate Op(n1/2) (such as Wang et al. [30]), which may be due to the existence of functional infinite dimensional explanatory variables. The convergence rates of ˆβ(s) and ˆg(z) are the same as that of the case where the random error is independent and identically distributed (such as Tang [31]), indicating that the autocorrelation of the random error does not affect the convergence rate of the slope function and the nonparametric function.

    In this section, we conduct a simulation under different settings to investigate the finite sample performance of our proposed method. The simulation data {(Xt,Zt,Yt),1tn} are generated from the following SFLAR model:

    Yt=10β(u)Xt(u)du+g(Zt)+εt,εt=ql=1alεtl+et,

    where Zt is distributed uniformly on [0, 1], and the functional predictor Xt=50j=1ξtjϕj, where ξtj are distributed as independent normal with a mean 0 and variance λj=((j0.5)π)2 and ϕj=2sin((j0.5)πt). For the unknown functions β() and g(), and autoregressive coefficients (a1,,aq), two scenarios have been considered as follows:

    Scenario 1. β(u)=2sin(0.5πu)+32sin(1.5πu), g(z)=sin(2πz), εt=aεt1+et, and the autoregressive coefficient a takes each of the values in the set (–0.5, 0.1, 0.8).

    Scenario 2. β(u)=2sin(0.5πu)+22sin(1.5πu), g(z)=cos(2πz), εt=a1εt1+a2εt2+et, and the autoregressive coefficients (a1,a2) take each of the values in the set {(0.1,0.8),(0.5,0.8)}.

    The error variables et are N(0,σ2). In both scenarios, σ takes 0.5 and 1, corresponding to different signal-to-noise ratios. To implement the proposed method, we apply the BIC criterion of Eq (2.6) to choose the number J of B-spline functions and the number M of eigenfunctions. Specifically, J and M are obtained by selecting the minimum BIC value within the preset range. In this simulation, J is chosen between 3 and 8, and M is selected between 1 and 6. In addition, the errors of the estimators of function β(u), g(z) and the model prediction error were evaluated by the integral of square errors ISE1=10[ˆβ(u)β(u)]2du, ISE2=10[ˆg(z)g(z)]2dz and the mean square prediction error (MSPE): MSPE=1nnt=q+1(^Ytg(Zt)10β(u)Xt(u)du)2, respectively.

    In our simulation, we considered sample sizes of n=400, 800, and 1200, with each setting repeated 500 times. Tables 1 and 2 summarize the performance of various estimators regarding bias (Bias), root mean square error (RMSE) for the estimated a and σ, and integrated square errors (ISE) for the estimated β() and g() under normal error conditions. From these tables, we can draw the following observations:

    (1) Given the sample size, as the value of σ increases, the bias (Bias) and root mean square error (RMSE) of ˆa and ˆσ increase, and the mean integral square error (ISE) of the functions ˆβ(u) and ˆg(z) increase, indicating that with the rise of variance, that is, the signal-to-noise ratio decreases, the estimation accuracy of the model parameters will decrease. The mean square prediction error (MSPE) of the model will increase.

    (2) As the sample size n increases from n=400 to 800 and 1200, we see a decrease in the root mean square error (RMSE) and the meawe see n integral square error (ISE), and the performance of the estimation method becomes better and better. When the sample size n increases from n=400 to 800 and 1200, the estimation of the autoregressive coefficient is asymptotically unbiased. The results show that the estimation of the autoregressive coefficient a, and the model parameters ˆβ(u) and ˆg(z), are consistent estimators. The estimation curves for the slope and nonparametric function of the model in Scenario 1 (AR(1)) with parameters n=400, a=0.5, and σ=0.5, and in Scenario 2 (AR(2)) with n=400, a=(0.5,0.8), and σ=0.5, were generated from 500 repetitions (see Figures 1 and 2). In the top left of Figures 1 and 2, the true slope function curve β() (red) and the median estimation curve ˆβ() (blue) are displayed for Scenarios 1 and 2, respectively. The top right figure shows a comparison between the true curve (red) and the median estimation curve ˆg() (blue) of the nonparametric function g() in both scenarios. The bottom of Figures 1 and 2 displays the comparison of the estimation curves for the slope and the nonparametric function, repeated 500 times in Scenarios 1 and 2, respectively. Figures 1 and 2 show that when n=400 and σ=0.5, the slope β(), the nonparametric function g(), and the median estimation curve (in blue) closely align with the true curve (in red) for both AR(1) and AR(2). For σ=0.5, a=0.5, and a=(0.5,0.8), the 95% pointwise confidence bands for β() and g() are displayed for sample sizes n=400, 800, and 1200 in Scenarios 1 and 2, respectively.

    Table 1.  The bias and RMSE of ˆa, ˆσ, the ISE of ˆβ(u), ˆg(z), and the MSPE under Scenario 1.
    a n σ ˆa ˆσ ˆβ(u) ˆg(z) Model
    Bias RMSE Bias RMSE ISE1 ISE2 MSPE
    400 0.5 0.003 0.049 -0.004 0.018 0.053 0.003 0.005
    1 -0.002 0.051 -0.016 0.077 0.118 0.013 0.018
    0.1 800 0.5 -0.001 0.036 -0.001 0.013 0.027 0.002 0.003
    1 -0.004 0.033 -0.012 0.050 0.055 0.006 0.009
    1200 0.5 0.000 0.028 0.000 0.010 0.019 0.001 0.002
    1 0.000 0.030 -0.007 0.043 0.039 0.005 0.006
    400 0.5 0.002 0.043 -0.003 0.018 0.052 0.003 0.004
    1 -0.004 0.044 -0.016 0.072 0.111 0.009 0.013
    -0.5 800 0.5 0.000 0.032 -0.001 0.012 0.025 0.002 0.002
    1 -0.001 0.031 -0.006 0.048 0.052 0.005 0.007
    1200 0.5 0.000 0.024 0.000 0.010 0.016 0.001 0.002
    1 -0.002 0.025 -0.006 0.042 0.033 0.003 0.005
    400 0.5 -0.006 0.032 -0.003 0.017 0.050 0.018 0.019
    1 -0.007 0.033 -0.020 0.071 0.124 0.066 0.069
    0.8 800 0.5 -0.005 0.022 -0.002 0.012 0.024 0.009 0.010
    1 -0.003 0.022 -0.005 0.055 0.044 0.034 0.035
    1200 0.5 -0.003 0.017 0.000 0.010 0.017 0.007 0.007
    1 -0.002 0.017 -0.006 0.040 0.031 0.021 0.022

     | Show Table
    DownLoad: CSV
    Table 2.  The bias and RMSE of (ˆa1,ˆa2), ˆσ, the ISE of ˆβ(u), ˆg(z), and the MSPE under Scenario 2.
    (a1,a2) n σ ˆa1 ˆa2 ˆσ ˆβ(u) ˆg(z) Model
    Bias RMSE Bias RMSE Bias RMSE ISE1 ISE2 MSPE
    400 0.5 -0.002 0.033 -0.016 0.036 -0.003 0.019 0.027 0.068 0.068
    1 -0.005 0.033 -0.011 0.035 -0.017 0.068 0.085 0.284 0.284
    (0.1, 0.8) 800 0.5 -0.004 0.022 -0.005 0.022 -0.001 0.013 0.013 0.034 0.034
    1 -0.002 0.022 -0.007 0.022 -0.011 0.054 0.033 0.109 0.110
    1200 0.5 -0.004 0.018 -0.004 0.017 0.001 0.011 0.009 0.022 0.022
    1 -0.001 0.018 -0.003 0.018 -0.008 0.04 0.021 0.085 0.085
    400 0.5 -0.002 0.031 0.000 0.030 -0.005 0.019 0.028 0.003 0.004
    1 -0.002 0.030 -0.001 0.030 -0.025 0.073 0.066 0.011 0.013
    (0.5, -0.8) 800 0.5 -0.001 0.022 0.001 0.021 -0.001 0.012 0.013 0.002 0.002
    1 0.002 0.021 -0.001 0.021 -0.010 0.050 0.037 0.005 0.006
    1200 0.5 -0.002 0.018 0.001 0.018 0.001 0.011 0.009 0.002 0.002
    1 0.001 0.017 0.000 0.017 -0.008 0.042 0.020 0.004 0.004

     | Show Table
    DownLoad: CSV
    Figure 1.  Estimation curves of the slope and nonparametric function under S cenario 1.
    Figure 2.  Estimation curves of the slope and nonparametric function under Scenario 2.

    When n=400, 800 and 1200, and σ=0.5, Figures 3 and 4 display the 95% pointwise confidence bands for Scenarios 1 and 2, respectively. Regardless of the changes in the model parameters β(u) and g(z) or the errors from AR(1) or AR(2), the proposed estimation method yields estimates for β(u) and g(z) that closely align with the true curve. Moreover, as the sample size increases, ˆβ(u) and ˆg(z) remain consistent estimates, demonstrating that the proposed model and estimation method perform well in finite samples.

    Figure 3.  The 95% pointwise confidence band of the slope (top) and nonparametric function (bottom) under Scenario 1.
    Figure 4.  The 95% pointwise confidence band of the slope (top) and nonparametric function (bottom) under Scenario 2.

    According to the suggestions of an anonymous reviewer, we run a new simulation to compare the predictive ability of the proposed model with that of existing models. Specifically, the simulation data {(Xt,Zt,Yt),1tn} are generated according to the following two scenarios:

    Scenario 3. It is the same as Scenario 1 with autoregressive coefficient a=0.5 and standard deviation σ=0.5.

    Scenario 4. It is the same as Scenario 3 except for g(Z)=2Z.

    For each simulation case, we set the sample size to n=400, and the simulation data is divided into the training data (Xt,Zt,Yt)320t=1 and the test data (Xt,Zt,Yt)400t=321. Similar to Aneiros and Vieu [22] and Yu et al. [23], we utilize the following two error criteria: mean quadratic error (MQE):

    MQE=179400t=322(YtˆYt)2,

    and mean relative quadratic error (MRQE):

    MRQE=179400s=322(YtˆYt)2Var(Y),

    to evaluate the prediction ability of the model, where Var(Y) is the empirical variance of {Yt}400t=321. To demonstrate the superiority of the proposed model, we compare its prediction results with three standard models: the semi-functional linear model (SFL) from Zhou and Chen [10], the partial functional linear regression model (PFL) by Zhang et al. [26], and the partial functional linear regression model with autoregressive error (PFLAR) by Yu et al. [23]. Each simulation experiment is repeated 200 times, and the results are summarized in Table 3. From Table 3, we can observe that:

    (ⅰ) When the true model is SFLAR, the proposed model has the best prediction accuracy.

    (ⅱ) Using our proposed SFLAR to fit PFLAR data has the same performance as using PFLAR to fit PFLAR data.

    Table 3.  The MQE and MRQE of different models.
    Scenarios Criterions Models
    SFLAR SFL PFLAR PFL
    Scenario 3 MQE 0.258 0.346 0.629 0.759
    MRQE 0.160 0.214 0.389 0.469
    Scenario 4 MQE 0.256 0.344 0.255 0.341
    MRQE 0.175 0.234 0.174 0.233

     | Show Table
    DownLoad: CSV

    In this section, the model and method proposed in this paper will be applied to study the power data set, which can be downloaded from https://www.eia.gov. This paper aims to forecast the monthly retail sales of electricity Ci kilowatt hour (KWH) in the commercial sector. For this purpose, it is considered that the US monthly electricity consumed by commercial sectors which includes the power consumption data from January 1973 to January 2016 (517 months) (Figure 5) and their annual average retail price Pt (per KWH, including tax) (43 years).

    Figure 5.  ln (power data) from January 1973 to January 2016.

    As shown in Figure 5, the time series shows some linear trends and some heterogeneity in the variance structure. Similar to Aneiros and Vieu [22] and Yu et al. [23], we eliminate the heteroscedasticity and the linear trend of the electricity retail sales data by differencing the ln(data), and obtain the time series (see Figure 6): Xi=ln(Ci+1)ln(Ci),i=1,2,,516. Let Xt(s)={X12(t1)+s,t=1,2,,43,s=1,2,,12} be the monthly ln(difference electricity retail sales data).

    Figure 6.  The differenced ln (power data to the commercial sector) from January 1973 to January 2016.

    There is only one observation per month, so we use ten cubic B-spline basis functions to convert the discrete monthly power data into functional continuous smooth annual curve predictors {Xt(u),u[1,12],t=1,2,,43} (see Figure 7).

    Figure 7.  Annual curves of power data to the commercial sector from January 1973 to January 2016.

    The response variable is Yt(s)=X12t+s, which represents the power consumption data in the s-month of the t-year, where t=1,2,,42,s=1,2,,12.

    Similar to Aneiros and Vieu [22] and Yu et al. [23], the power data set is divided into the training sample {Yt(s),Pt,Xt(u)}41t=1 and the test sample {Y42(s),P42,X42(u)}. According to Yang et al.[24], we consider the following semi-functional linear model (SFLAR) with autoregressive error to fit the training set:

    Yt(s)=121β(u)Xt(u)du+g(Pt)+εt,εt=ql=1alεtl+et,t=1,2,,41.

    Then, the test set verifies the model's predicted value Y42(s). To demonstrate the superiority of the proposed model, we also compare its prediction results with three standard models: The semi-functional linear model (SFL), the partial functional linear regression model (PFL), and the partial functional linear regression model with autoregressive error (PFLAR). In addition, similar to Aneiros and Vieu [22] and Yu et al.[23], we also utilize the mean quadratic error (MQE):

    MQE=11212s=1(Y42(s)ˆY42(s))2,

    and mean relative quadratic error (MRQE):

    MRQE=11212s=1(Y42(s)ˆY42(s))2Var(Y(s))

    to evaluate the prediction ability of the model, where Var(Y(s)) is the empirical variance of {Yi(s)}41i=1.

    The prediction effects of the four models under the test set are summarized in Table 4 and Figure 8. It is observed from Figure 8 that the four models exhibit consistency with the changing trend of the curve in Figure 7. Additionally, the prediction results from the test set in Table 4 and Figure 8 indicate that the SFL is closer to the actual value than the PFL. The MQE and MRQE of the SFL are also smaller than those of the PFL, suggesting that the prediction effect of the SFL is superior to the PFL, regardless of whether the autoregressive error is considered. Furthermore, the model with autoregressive error demonstrates a better prediction effect than the model without, implying that the former is more reasonable. The SFLAR model exhibits the smallest MQE and MRQE among the four models, signifying that the SFLAR model proposed in this paper yields the best prediction effect. Therefore, using the SFLAR model for this data is more reasonable.

    Table 4.  The MQE and MRQE of different models.
    Criterions Models
    SFLAR SFL PFLAR PFL
    MQE 0.00017 0.00025 0.00027 0.00030
    MRQE 0.34483 0.45411 0.47235 0.48765

     | Show Table
    DownLoad: CSV
    Figure 8.  Monthly power data forecast of different models in 2015.

    The results of our study on forecasting monthly retail sales of electricity using an SFLAR model provide important insights into the dynamics of electricity consumption, which captures the influence of both functional and non-functional predictors of electricity sales. Incorporating autoregressive errors allowed the model to effectively capture temporal dependencies within the monthly sales data, thereby enhancing forecasting accuracy. However, our model also has some limitations. For example, we only consider one-dimensional covariate Z in our model. Although our model can be easily generalized to multidimensional situations, it may encounter the curse of dimensionality and computational burden. As one of the anonymous reviewers said, a single index model could provide much more efficient results. Hence, we can extend our model to the semi-functional single index model in future research. At the same time, deep neural networks have been extensively used for nonparametric regression. This also constitutes a topic for our future research.

    This study focused on the SFLAR model, incorporating a random error sequence with an AR(p) structure. By utilizing Mercer's theorem, the linear component of the function is dimensionally reduced using the method of functional principal component basis expansion. Meanwhile, the nonparametric function is approximated using the B-spline function. The slope function β(), nonparametric function g(), and autoregressive coefficient a are estimated using a two-step iterative algorithm. Theoretical properties of the estimated parameters are provided under specific regular conditions. Based on the simulation results, the proposed model and estimation method perform well even with a limited sample size. The real data analysis also indicates that the SFLAR model has the best prediction accuracy among these four models when autoregressive errors exist. Therefore, the SFLAR model proposed in this paper is more suitable for this power data.

    Bin Yang: Investigation, funding acquisition, formal analysis, validation, software, data curation, writing-original draft, writing-review and editing; Min Chen: Conceptualization, methodology, supervision, project administration, validation; Jianjun Zhou: Conceptualization, funding acquisition, investigation, resources, validation, data curation, writing-original draft, writing-review and editing. All authors have read and approved the final version of the manuscript for publication.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors are grateful to the two anonymous referees and the editor for their constructive remarks that greatly improved the manuscript. The research is partially supported by a grant from the National Key R & D Program of China, Grant/Award Number: 2022YFA1003701, National Nature Science Foundation of China (No.11861074, 12261051), Applied Basic Research Project of Yunnan Province (2019FB138), Key Project of Yunnan Provincial Basic Research (Grant No.202401AS070061), and the Yunnan Provincial Department of Education Science Research Fund (Grant No.2024J0087).

    The authors declare no conflict of interest.



    [1] W. G. Manteiga, A. M. Calvo, Bootstrap in functional linear regression, J. Stat. Plan. Infer., 141 (2011), 453–461. https://doi.org/10.1016/j.jspi.2010.06.027 doi: 10.1016/j.jspi.2010.06.027
    [2] M. F. Bande, P. Galeano, M. G. Manteiga, Functional principal component regression and functional partial least-squares regression: An overview and a comparative study, Int. Stat. Rev., 85 (2017), 61–83. https://doi.org/10.1111/insr.12116 doi: 10.1111/insr.12116
    [3] F. Ferraty, P. Hall, P. Vieu, Most-predictive design points for functional data predictors, Biometrika, 97 (2010), 807–824. http://www.jstor.org/stable/29777138
    [4] F. Belarbi, S. Chemikh, A. Laksaci, Local linear estimate of the nonparametric robust regression in functional data, Stat. Probab. Lett., 134 (2018), 128–133. https://doi.org/10.1016/j.spl.2017.11.003 doi: 10.1016/j.spl.2017.11.003
    [5] G. A. Pérez, P. Vieu, Semi-functional partial linear regression, Stat. Probabil. Lett., 76 (2006), 1102–1110. https://doi.org/10.1016/j.spl.2005.12.007 doi: 10.1016/j.spl.2005.12.007
    [6] H. Shin, Partial functional linear regression, J. Stat. Plan. Infer., 139 (2009), 3405–3418. https://doi.org/10.1016/j.jspi.2009.03.001 doi: 10.1016/j.jspi.2009.03.001
    [7] H. Shin, M. Lee, On prediction rate in partial functional linear regression, J. Multivariate Anal., 103 (2012), 93–106. https://doi.org/10.1016/j.jmva.2011.06.011 doi: 10.1016/j.jmva.2011.06.011
    [8] F. Yao, S. S. Chee, F. Wang, Regularized partially functional quantile regression, J. Multivariate Anal., 156 (2017), 39–56. https://doi.org/10.1016/j.jmva.2017.02.001 doi: 10.1016/j.jmva.2017.02.001
    [9] D. H. Kong, K. J. Xue, F. Yao, H. H. Zhang, Partially functional linear regression in high dimensions, Biometrika, 103 (2016), 147–159. https://doi.org/10.1093/biomet/asv062 doi: 10.1093/biomet/asv062
    [10] J. J. Zhou, M. Chen, Spline estimators for semi-functional linear model, Stat. Probabil. Lett., 82 (2012), 505–513. https://doi.org/10.1016/j.spl.2011.11.027 doi: 10.1016/j.spl.2011.11.027
    [11] Q. Y. Peng, J. J. Zhou, N. S. Tang, Varying coefficient partially functional linear regression models, Stat. Pap., 57 (2016), 827–841. https://doi.org/10.1007/s00362-015-0681-3 doi: 10.1007/s00362-015-0681-3
    [12] S. Novo, G. Aneiros, P. Vieu, Automatic and location-adaptive estimation in functional single-index regression, J. Nonparametr. Stat., 31 (2019), 364–392. https://doi.org/10.1080/10485252.2019.1567726 doi: 10.1080/10485252.2019.1567726
    [13] A. Cuevas, A partial overview of the theory of statistics with functional data, J. Stat. Plann. Inference., 147 (2014), 1–23. https://doi.org/10.1016/j.jspi.2013.04.002 doi: 10.1016/j.jspi.2013.04.002
    [14] A. Goia, P. Vieu, An introduction to recent advances in high/infinite dimensional statistics, J. Multivariate Anal., 146 (2016), 1–6. https://doi.org/10.1016/j.jmva.2015.12.001 doi: 10.1016/j.jmva.2015.12.001
    [15] N. X. Ling, P. Vieu, Nonparametric modelling for functional data: Selected survey and tracks for future, Statistics, 52 (2018), 934–949. https://doi.org/10.1080/02331888.2018.1487120 doi: 10.1080/02331888.2018.1487120
    [16] N. X. Ling, P. Vieu, On semiparametric regression in functional data analysis, Wires. Comput. Stat., 13 (2020), e1538. https://doi.org/10.1002/wics.1538 doi: 10.1002/wics.1538
    [17] G. Aneiros, R. Cao, R. Fraiman, C. Genest, P. Vieu, Recent advances in functional data analysis and high-dimensional statistics, J. Multivariate Anal., 170 (2019), 3–9. https://doi.org/10.1016/j.jmva.2018.11.007 doi: 10.1016/j.jmva.2018.11.007
    [18] J. Vilar, R. Cao, G. Aneiros, Forecasting next-day electricity demand and price using nonparametric functional methods, Int. J. Elec. Power, 39 (2012), 48–55. https://doi.org/10.1016/j.ijepes.2012.01.004 doi: 10.1016/j.ijepes.2012.01.004
    [19] G. Aneiros, J. Vilar, P. Raña, Short-term forecast of daily curves of electricity demand and price, Int. J. Elec. Power, 80 (2016), 96–108. https://doi.org/10.1016/j.ijepes.2016.01.034 doi: 10.1016/j.ijepes.2016.01.034
    [20] J. Vilar, G. Aneiros, P. Raña Prediction intervals for electricity demand and price using functional data, Int. J. Elec. Power, 96 (2018), 457–472. https://doi.org/10.1016/j.ijepes.2017.10.010 doi: 10.1016/j.ijepes.2017.10.010
    [21] R. Peláez, G. Aneiros, J. Vilar, Bootstrap prediction regions for daily curves of electricity demand and price using functional data, Int. J. Elec. Power, 162 (2024), 110244. https://doi.org/10.1016/j.ijepes.2024.110244 doi: 10.1016/j.ijepes.2024.110244
    [22] G. A. Pérez, P. Vieu, Nonparametric time series prediction: A semi-functional partial linear modeling, J. Multivariate Anal., 99 (2008), 834–857. https://doi.org/10.1016/j.jmva.2007.04.010 doi: 10.1016/j.jmva.2007.04.010
    [23] P. Yu, T. Li, Z. Y. Zhu, Z. Z. Zhang, Composite quantile estimation in partial functional linear regression model with dependent errors, Metrika, 82 (2019), 633–656. https://doi.org/10.1007/s00184-018-0699-3 doi: 10.1007/s00184-018-0699-3
    [24] B. Yang, M. Chen, J. J. Zhou, Testing for error correlation in semi-functional linear models, J. Syst. Sci. Complex., 36 (2023), 1697–1716. https://doi.org/10.1007/s11424-023-1431-6 doi: 10.1007/s11424-023-1431-6
    [25] S. D. Niang, S. Guillas, Functional semiparametric partially linear model with autoregressive errors, J. Multivariate Anal., 101 (2010), 307–315. https://doi.org/10.1016/j.jmva.2008.06.008 doi: 10.1016/j.jmva.2008.06.008
    [26] T. Zhang, P. Dai, Q. Zhang, Joint detection for functional polynomial regression with autoregressive errors, Commun. Stat.-Theory M., 46 (2017), 7837–7854. https://doi.org/10.1080/03610926.2015.1096384 doi: 10.1080/03610926.2015.1096384
    [27] P. X. Xiao, G. C. Wang, Partial functional linear regression with autoregressive errors, Commun. Stat.-Theory M., 51 (2022), 4515–4536. https://doi.org/10.1080/03610926.2020.1818097 doi: 10.1080/03610926.2020.1818097
    [28] M. Wang, M. L. Shu, J. J. Zhou, S. X. Wu, M. Chen, Least square estimation for multiple functional linear model with autoregressive errors, Acta Math. Appl. Sin.-E., 41 (2022), 84–98. https://doi.org/10.1007/s10255-024-1143-2 doi: 10.1007/s10255-024-1143-2
    [29] C. D. Boor, A practical guide to spline, New York: Springer, 1978. https://doi.org/10.2307/2006241
    [30] H. S. Wang, G. D. Li, T. C. Ling, Regression coefficient and autoregressive order shrinkage and selection via the Lasso, J. R. Stat. Soc. B., 69 (2007), 63–78. https://doi.org/10.1111/j.1467-9868.2007.00577.x doi: 10.1111/j.1467-9868.2007.00577.x
    [31] Q. G. Tang, Estimation for semi-functional linear regression, Statistics., 49 (2015), 1262–1278. https://doi.org/10.1080/02331888.2014.979827 doi: 10.1080/02331888.2014.979827
    [32] L. Schumaker, Spline functions: Basic theory, 1 Eds., New York: John Wiley & Sons, 1981. https://doi.org/10.2307/2007301
    [33] P. Hall, J. L. Horowitz, Methodology and convergence rates for functional linear regression, Ann. Stat., 35 (2007), 70–91. https://doi.org/10.1214/009053606000000957 doi: 10.1214/009053606000000957
    [34] H. Lian, Shrinkage estimation and selection for multiple functional regression, Stat. Sinica, 23 (2013), 51–74. https://doi.org/10.5705/ss.2011.160 doi: 10.5705/ss.2011.160
  • This article has been cited by:

    1. Lijie Zhou, Liucang Wu, Bin Yang, Estimation and diagnostic for single-index partially functional linear regression model with $ p $-order autoregressive skew-normal errors, 2025, 10, 2473-6988, 7022, 10.3934/math.2025321
  • Reader Comments
  • © 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(463) PDF downloads(47) Cited by(1)

Figures and Tables

Figures(8)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog