Loading [MathJax]/jax/output/SVG/jax.js
Research article

Statistical inference in functional semiparametric spatial autoregressive model

  • Received: 26 January 2021 Accepted: 15 July 2021 Published: 28 July 2021
  • MSC : 62G05, 62J05, 62M30

  • Semiparametric spatial autoregressive model has drawn great attention since it allows mutual dependence in spatial form and nonlinear effects of covariates. However, with development of scientific technology, there exist functional covariates with high dimensions and frequencies containing rich information. Based on high-dimensional covariates, we propose an interesting and novel functional semiparametric spatial autoregressive model. We use B-spline basis function to approximate the slope function and nonparametric function and propose generalized method of moments to estimate parameters. Under certain regularity conditions, the asymptotic properties of the proposed estimators are obtained. The estimators are computationally convenient with closed-form expression. For slope function and nonparametric function estimators, we propose the residual-based approach to derive its pointwise confidence interval. Simulation studies show that the proposed method performs well.

    Citation: Gaosheng Liu, Yang Bai. Statistical inference in functional semiparametric spatial autoregressive model[J]. AIMS Mathematics, 2021, 6(10): 10890-10906. doi: 10.3934/math.2021633

    Related Papers:

    [1] Bin Yang, Min Chen, Jianjun Zhou . Forecasting the monthly retail sales of electricity based on the semi-functional linear model with autoregressive errors. AIMS Mathematics, 2025, 10(1): 1602-1627. doi: 10.3934/math.2025074
    [2] Rabia Noureen, Muhammad Nawaz Naeem, Dumitru Baleanu, Pshtiwan Othman Mohammed, Musawa Yahya Almusawa . Application of trigonometric B-spline functions for solving Caputo time fractional gas dynamics equation. AIMS Mathematics, 2023, 8(11): 25343-25370. doi: 10.3934/math.20231293
    [3] W. B. Altukhaes, M. Roozbeh, N. A. Mohamed . Feasible robust Liu estimator to combat outliers and multicollinearity effects in restricted semiparametric regression model. AIMS Mathematics, 2024, 9(11): 31581-31606. doi: 10.3934/math.20241519
    [4] Emre Kırlı . A novel B-spline collocation method for Hyperbolic Telegraph equation. AIMS Mathematics, 2023, 8(5): 11015-11036. doi: 10.3934/math.2023558
    [5] Xin Liang, Xingfa Zhang, Yuan Li, Chunliang Deng . Daily nonparametric ARCH(1) model estimation using intraday high frequency data. AIMS Mathematics, 2021, 6(4): 3455-3464. doi: 10.3934/math.2021206
    [6] M. J. Huntul, Kh. Khompysh, M. K. Shazyndayeva, M. K. Iqbal . An inverse source problem for a pseudoparabolic equation with memory. AIMS Mathematics, 2024, 9(6): 14186-14212. doi: 10.3934/math.2024689
    [7] Salim Bouzebda . Weak convergence of the conditional single index $ U $-statistics for locally stationary functional time series. AIMS Mathematics, 2024, 9(6): 14807-14898. doi: 10.3934/math.2024720
    [8] Jingjing Yang, Weizhong Tian, Chengliang Tian, Sha Li, Wei Ning . Empirical likelihood method for detecting change points in network autoregressive models. AIMS Mathematics, 2024, 9(9): 24776-24795. doi: 10.3934/math.20241206
    [9] Hasim Khan, Mohammad Tamsir, Manoj Singh, Ahmed Hussein Msmali, Mutum Zico Meetei . Numerical approximation of the time-fractional regularized long-wave equation emerging in ion acoustic waves in plasma. AIMS Mathematics, 2025, 10(3): 5651-5670. doi: 10.3934/math.2025261
    [10] Qi Xie, Yiting Huang . A class of generalized quadratic B-splines with local controlling functions. AIMS Mathematics, 2023, 8(10): 23472-23499. doi: 10.3934/math.20231193
  • Semiparametric spatial autoregressive model has drawn great attention since it allows mutual dependence in spatial form and nonlinear effects of covariates. However, with development of scientific technology, there exist functional covariates with high dimensions and frequencies containing rich information. Based on high-dimensional covariates, we propose an interesting and novel functional semiparametric spatial autoregressive model. We use B-spline basis function to approximate the slope function and nonparametric function and propose generalized method of moments to estimate parameters. Under certain regularity conditions, the asymptotic properties of the proposed estimators are obtained. The estimators are computationally convenient with closed-form expression. For slope function and nonparametric function estimators, we propose the residual-based approach to derive its pointwise confidence interval. Simulation studies show that the proposed method performs well.



    Spatial autoregressive model (SAR) and its derivatives have been widely used in many areas such as economics, political science, public health and so on. There are lots of literatures concerning about spatial autoregressive models such as Anselin [1], LeSage [19], Anselin and Bera [2], Lee and Yu [20], LeSage and Pace [21], Lee [22], Dai, Li and Tian [9]. In particular, Lee [22] utilised generalized method of moments to make inference about spatial autoregressive model. Xu and Li [25] investigated the instrumental variable (Ⅳ) and maximum likelihood estimators for spatial autoregressive model using a nonlinear transformation of dependent variable. However, spatial autoregressive model may not be flexible enough to capture nonlinear impact of some covariates since its parametric structure. In order to enrich model adaptability and flexibility, some semiparametric spatial autoregressive models have been proposed. For example, Su [31] studied a semiparametric SAR, which includes nonparametric covariates. Su and Jin [32] proposed partially linear SAR with both linear covariates and nonparametric explanatory variables. Sun et al. [33] studied a semiparametric spatial dynamic model with a profile likelihood approach. Wei and Sun [36] derived semiparametric generalized method of moments estimator. Hoshino [35] proposed a semiparametric series generalized method of moments estimator and established consistency and asymptotic normality of the proposed estimator.

    However, with development of economic and scientific technology, huge amounts of data can be easily collected and stored. In particular, some types of data are observed in high dimensions and frequencies containing rich information. We usually call them functional data. When those types data are included in a model as covariates, it is common to use functional linear model (FLM). There exist vast literatures on estimation and prediction for FLM (See, for example, Reiss et al.[26], Ramsay and Dalzell [27], Delaigle and Hall [11], Aneiros-Pˊerez and Vieu [3]). Many methods were proposed to estimate the slop function such as Cardot et al. [5], Hall and Horowitz [14], Crambes et al. [8], Shin [28]. In particular, Hall and Horowitz [14] established minimax convergence rates of estimation. Cai and Hall [6] proposed functional principle components method and a reproducing kernel Hilbert space approach was used in Yuan and Cai [7].

    In many applications of spatial data, there are often covariates with nonlinear effect and functional explanatory variables. This motivates us to propose an interesting and novel functional semiparametric spatial autoregressive model. The model is relatively flexible because it utilises functional linear model to deal with functional covariate and semiparametric SAR model to allow spatial dependence and nonlinear effect of scalar covariate. Recently, some models consider both functional covariates and spatial dependence. For instance, Pineda-Rˊios [24] proposed functional SAR model and used least squares and maximum likelihood methods to estimate parameters. The functional SAR model considers spatial effect for error term instead of spatial effect for response variable. Huang et al. [12] considered spatial functional linear model and they developed an estimation method based on maximum method and functional principle component analysis. Hu et al. [13] developed generalized methods of moments to estimate parameters in spatial functional linear model. In the paper, we proposed a generalized method of moments estimator which is heteroskedasticity robust and takes a closed-form written explicitly.

    The rest of the paper is organized as follows. Section 2 introduces the proposed model and the estimation procedure. The asymptotic properties of the proposed estimators are established in Section 3. Section 4 conducts simulation studies to evaluate the empirical performance of the proposed estimators. Section 5 gives some discussions about the model. All technical proofs are provided in appendix.

    Consider the following novel functional semiparametric spatial autoregressive model,

    Y=ρWY+Zθ+10X(t)β(t)dt+g(U)+ε, (2.1)

    where Y is a response variable, ρ is an unknown coefficient of the spatial neighboring effect, W is the constant spatial weight matrix with a zero diagonal, Z=(Z1,...,Zp) is a p-dimensional covariate and θ is its coefficient. {X(t):t[0,1]} is a zero-mean and second-order (i.e. E|X(t)|2<,t[0,1]) stochastic process defined on (Ω,B,P) with sample paths in L2[0,1], the Hilbert space containing integrable functions with inner product x,y=10x(t)y(t)dt,x,yL2[0,1] with norm x=x,x1/2. The slope function β(t) is a square integrable function on [0,1], U is a random variable, g() is an unknown function on its support [0,1] without loss of generality. We assume E[g()]=0 to ensure the identifiability of the nonparametric function. ε is a random error with zero mean and finite variance σ2, independent of Z,U and X(t).

    Remark 1. The model (2.1) is more flexible to take different models. It generalizes both semiparametric spatial autoregressive model [32] and functional partial linear model [28] which correspond to the cases β(t)=0 and ρ=0, respectively. The model can be represented by Y=(IρW)110X(t)β(t)dt+(IρW)1Zθ+(IρW)1g(U)+(IρW)1ε. We assume IρW could be inverted to make the presentation valid. Thus Yi is also influenced by its neighbours' covariates Xj(t) as ji. Parameter ρ indicates the basic impact of the neighbours. Greater absolute value of ρ means that the response variable is more likely to be affected by its neighbours.

    In the section, we give a method to estimate unknown parameter ρ and θ, slope function β() and nonparametric function g(). We use B-spline basis function to approximate g() and β(). Let 0=u0<u1...<uk1+1=1 be a partition of interval [0,1]. Using ui as knots, we have N1=k1+l1+1 normalized B-spline basis functions of order l1+1 that from a basis function for the linear spline space. Put the basis function as a vector B1(t)=(B11(t),...,B1N1(t)) and then the slope function β() is approximated by B1()γ. Similarly, let B2(u)=(B21(u),...,B2N2(u)) be normalized B-spline basis function vector determined by k2 interior knots in [0,1] and the order l2+1 to approximate g(), where N2=k2+l2+1. Then it follows that

    β(t)B1(t)γ,   g(u)B2(u)ζ,

    where γ=(γ1,...,γN1) and ζ=(ζ1,...,ζN2).

    Let D=X(t),B1(t)=(10X(t)B11(t)dt,...,10X(t)B1N1(t)dt),Di=Xi(t),B1(t). Then the model can be rewritten as

    YρWY+Zθ+Dγ+B2(U)ζ+ε.

    Let P=Π(ΠΠ)1Π denote the projection matrix onto the space by Π, where Π=(D,B2(U)). Similar to Zhang and Shen [39], profiling out the functional approximation, we obtain

    (IP)Yρ(IP)WY+(IP)Zθ+(IP)ε.

    Let Q=(WY,Z) and η=(ρ,θ). Applying the two stage least squares procedure proposed by Kelejian and Prucha [17], we propose the following estimator

    ˆη=(Q(IP)M(IP)Q)1Q(IP)M(IP)Y,

    where M=H(HH)1H and H is matrices of instrumental variables. Moreover,

    (ˆγ,ˆζ)=(ΠΠ)1Π(YQˆη).

    Consequently, we use ˆβ(t)=B1(t)ˆγ,ˆg(u)=B2(u)ˆζ as the estimator of β(t) and g(u).

    For statistical inference based on ˆη, consistent estimators of the asymptotic covariance matrices are needed. Define the following estimator

    ˆσ2=1nYˆρWYZˆθDˆγB2(U)ˆζ2,

    and

    ˆΣ=1nQ(IP)M(IP)Q,

    where is the L2 norm for a function or the Euclidean norm for a vector. In order to make statistical inference about σ2, it need to get the value ω=E[(ε21σ2)2]. Therefore, we use the following estimator ˆω to estimate ω

    ˆω=1nni=1(ˆε2iˆσ2)2.

    Similar to Zhang and Shen [39], we use an analogous idea for the construction of instrument variables. In the first step, the following instrumental variables are obtained ˜H=(W(I˜ρW)1(Z,D˜γ,B2(U)˜ζ),Z,Π), where ˜ρ, ˜γ and ˜ζ are obtained by simply regressing Y on pseudo regressor variables WY,Π. In the second step, instrumental variables ˜H are used to obtain the estimators ˉη, ˉγ and ˉζ, which are necessary to construct the following instrumental variables H=(W(IˉρW)1(Zˉθ+Dˉγ+B2(U)ˉζ),Z). Finally, we use the instrumental variables H to obtain the final estimators ˆρ and ˆθ.

    In this section, we derive the asymptotic normality and rates of convergence for the estimators defined in previous section. Firstly, we introduce some notations. For convenience and simplicity, c denote a generic positive constant, which may take different values at different places. Let β0() and go() be the true value of function β() and g() respectively. K(t,s)=Cov(X(t),X(s)) denotes the covariance function of X(). anbn means that an/bn is bounded away from zero and infinity as n. In the paper, we make the following assumptions.

    C1 The matrix IρW is nonsingular with |ρ|<1.

    C2 The row and column sums of the matrices W and (IρW)1 are bounded uniformly in absolute value for any |ρ|<1.

    C3 For matrix S=W(IρW)1, there exists a constant λc such that λcISS is positive semidefinite for all n.

    C4 Matrix 1n˜Q(IP)M(IP)˜QΣ in probability for some positive definite matrix, where ˜Q=(W(IρW)1(Zθ+10X(t)β(t)dt+g(U)),Z).

    C5 For matrix ˜Q, there exsits a constant λc such that λcI˜Q˜Q is positive semidefinite for all n.

    C6 X(t) has a finite fourth moment, that is, EX(t)4c.

    C7 K(t,s) is positive definite.

    C8 The nonparametric function g() has bounded and continuous derivatives up to order r(2) and the slope function β(t)Cr[0,1].

    C9 The density of U, fU(u), is bounded away from 0 and on [0,1]. Furthermore, we assume that fU(u) is continuously differentiable on [0,1].

    C10 For the knots number kj,(j=1,2), it is assumed k1k2k.

    Assumptions C1C3 are required in the setting of spatial autoregressive model (see, for example, Lee [23], Kelejian and Prucha [18], Zhang and Shen [39]). They concern the restriction of spatial weight matrix and SAR parameter. Assumption C4 (see Du et al.[10]) is used to represent the asymptotic covariance matrix of ˆη. Moreover, assumption C4 requires implicitly that the generated regressors CZ and Z, deviated from their functional part of projection onto Π, are not asymptotically multicollinear. Assumption C5 is required to ensure the identifiability of parameter η. Assumptions C6C7 are commonly assumed in functional linear model [14]. Assumption C6 is a mild restriction to prove the convergence of our estimator. Assumption C7 guarantees the identifiability of β(t). Assumption C8 ensures that β() and g() are sufficiently smoothed and can be approximated by basis functions in the spline space. Assumption C9 requires a bounded condition on the covariates. It is often assumed in asymptotic analysis of nonparametric regression problems (see, for example [15,37]). Assumption C10 is required to achieve the optimal convergence rate of ˆβ() and ˆg().

    Let

    Δn=E(DD)E{E(DB2(u)|U)[E(B2(u)B2(u)|U)]1E(B2(u)D|U)},Ωn=E(B2B2)E{E(B2(u)D|V)[E(DD|V)]1E(DB2(u)|V)},

    where V=X(t),β0(t). The following theorems state the asymptotic properties of the estimators for parameter and nonparametric function.

    Theorem 1. Suppose assumptions C1-C10 hold, then

    n(ˆηη)dN(0,σ2Σ1).

    Theorem 2. Assumptions C1-C10 hold and kn12r+1, then

    ˆβ()β0()2=Op(n2r2r+1).ˆg()g0()2=Op(n2r2r+1).

    Remark 2. Theorem 2 gives the consistency of function estimators. The slope function estimator ˆβ() and nonparametric function estimator ˆg() have the same optimal global convergence rate established by Stone [29].

    Theorem 3. Suppose assumptions C1-C10 hold, and E(|ε1|4+r)< for some r>0, then

    ˆσ2pσ2,    ˆωpω,    and    ˆΣpΣ.

    Remark 3. From the proof of Theorem 3, if trace(S)/n=o(1), it can be shown that

    n(ˆσ2σ2)dN(0,ω).

    Theorem 4. Suppose assumptions C1-C10 hold and n/(k2r+11)=n/(k2r+12)=o(1), for any fixed points t,u(0,1), as n,

    nk1(ˆβ(t)β(t))dN(0,Ξ(t)),
    nk2(ˆg(u)g(u))dN(0,Λ(u)).

    where β(t)=B1(t)γ0, g(u)=B2(u)ζ0, Ξ(t)=limnσ2k1B1(t)ΔnB1(t), Λ(u)=limnσ2k2B2(u)ΩnB2(u), γ0 and ζ0 are defined in Lemma 1 of appendix.

    Remark 4. The above conclusions is similar to those of Yu et al. [38], which gave the asymptotic normality for spline estimators in single-index partial functional linear regression model. Note that ˆβ(t)β0(t)=(β(t)β0(t))+(ˆβ(t)β(t)). We obtain that β(t)β0(t)=O(kr1) by Lemma 1 on Appendix and ˆβ(t)β(t) dominates β(t)β0(t). Therefore we can use the asymptotic behaviors of ˆβ(t)β(t) to describe the asymptotic behaviors of ˆβ(t)β0(t).

    The variance Ξ(t) and Λ(u) are involved in basic function and knots. Different basis functions and knots can get different variance estimators. Moreover, the variance expression contains unknown quantities. Replacing them by consistent estimators can lead to approximation errors. What's more, there may exist heteroscedasticity in error term and then the estimator ˆσ2 is not consistent. Consequently, we propose the following residual-based method to construct piecewise confidence interval in practice.

    It is crucial that spatial structure must be preserved during data resampling in models with spatial dependence [1]. Therefore, we employ the residual-based bootstrap procedure to derive the empirical pointwise standard error of ˆβ(t) and ˆg(). The procedure can be described as follows:

    (1) Based on the data sets {Y,Z,X(t),U} and spatial matrix W, one fits the proposed model and obtain the residual vector ˆε1=(ˆε11,...,ˆεn1). Then, we derive the centralized residual vector ˆε.

    (2) Draw a bootstrap sample ˆε with replacement from the empirical distribution function of ˆε and generate Y=(IˆρW)1(Zθ+Dˆγ+B2(U)ˆζ+ˆε).

    (3) Based on the new data sets {Y,Z,X(t),U} and spatial matrix W, we fit the proposed model again to derive the estimator ˆβ(t) and ˆg(u). Repeat the process many times. Thus, for given t and u, calculate empirical variance of ˆβ(t) and ˆg(u) respectively. Consequently, we use the empirical variance to construct its confidence interval.

    In this section, we use simulation examples to study the properties of the proposed estimators. The data is generated from the following model:

    Yi=ρni=1wijYj+Z1iθ1+Z2iθ2+10β(t)Xi(t)dt+g(Ui)+εi,i=1,...,n,

    where ρ=0.5, β(t)=2sin(πt/2)+32sin(3πt/2) and X(t)=50j=1γjϕj(t), where γj is distributed as independent normal with mean 0 and variance λj=((j0.5)π)2 and ϕj(t)=2sin((j0.5)πt). Zi1 and Zi2 are independent and follow standard normal distribution, θ1=θ2=1, UiU(0,1), g(u)=sin(π(uA)CA),A=321.65412,C=32+1.65412. The spatial weight matrix W=(wij)n×n is generated based on mechanism that wij=0.3|ij|I(ij),1i,jn with wii=0,i=1,...,n. A standardized transformation is used to convert the matrix W to have row-sums of unit. We set the following three kinds of error term: (1) εiN(0,σ2); (2) εi0.75t(3); (3) εi(1+0.5Ui)N(0,σ2), where σ2=1. In order to compare the different situations for magnitudes of ρ, we set ρ={0.2,0.7} with error term N(0,σ2). Simulation results are derived based on 1000 replications.

    To achieve good numerical performances, the order l1 and l2 of splines and the number of interior knots k1 and k2 should be chosen. To reduce the burden of computation, we use the cubic B-spline with four evenly distributed knots (i.e., k1=k2=2) for slope function β() and nonparametric function g() respectively. These choices of k1 and k2 are small enough to avoid overfitting in typical problem with sample size not too small and big enough to flexibly approximate many smooth function. We use the square root of average square errors (RASE) to assess the performance of estimators ˆβ() and ˆg() respectively

    RASE1={n1i=1(ˆβ(ti)β(ti))2n1}1/2,
    RASE2={n2i=1(ˆg(ui)g(ui))2n2}1/2,

    where {ti,i=1,...,n1}, {ui,i=1,...,n2} and n1=n2=200 are grid points chosen equally spaced in the domain of β() and g() respectively.

    Tables 13 show simulation results with different kinds of error terms. Table 4 presents different magnitudes of ρ with error term N(0,1). They show the bias (Bias), standard deviation (SD), standard error (SE) and coverage probability (CP) with nominal level of 95% for estimator and the mean and standard deviation (SD) of RASEj(j=1,2) for ˆβ() and ˆg(). The simulation results can be summarized as follows:

    Table 1.  Simulation results for ρ=0.5 with error term N(0,1).
    n Est Bias SD SE CP
    100 ˆρ -0.0091 0.0789 0.0799 0.9430
    ˆθ1 -0.0065 0.1058 0.1030 0.9500
    ˆθ2 -0.0012 0.1052 0.1078 0.9430
    ˆσ2 -0.1100 0.1379 0.1232 0.7900
    RASE1 1.4361 0.7046
    RASE2 0.1950 0.0683
    300 ˆρ -0.0031 0.0441 0.0444 0.9520
    ˆθ1 -0.0011 0.0594 0.0594 0.9580
    ˆθ2 0.0027 0.0595 0.0595 0.9480
    ˆσ2 -0.0291 0.0834 0.0785 0.8970
    RASE1 0.7932 0.3728
    RASE2 0.1108 0.0392
    500 ˆρ -0.0019 0.0339 0.0332 0.9600
    ˆθ1 -0.0039 0.0456 0.0442 0.9610
    ˆθ2 -0.0004 0.0455 0.0461 0.9410
    ˆσ2 -0.0212 0.0653 0.0616 0.9100
    RASE1 0.6253 0.2847
    RASE2 0.0838 0.0303
    Note: It shows that the bias (Bias), standard deviation (SD), standard error (SE) and coverage probability (CP) with nominal level of 95% for estimator and the mean and standard deviation (SD) of RASE for ˆβ() and ˆg() from 1000 repetitions.

     | Show Table
    DownLoad: CSV
    Table 2.  Simulation results for ρ=0.5 with error term 0.75t(3).
    n Est Bias SD SE CP
    100 ˆρ -0.0152 0.1010 0.1092 0.9620
    ˆθ1 -0.0059 0.1332 0.1297 0.9540
    ˆθ2 0.0028 0.1332 0.1376 0.9350
    RASE1 1.8018 1.0485
    RASE2 0.2440 0.1053
    300 ˆρ -0.0093 0.0567 0.0600 0.9530
    ˆθ1 0.0062 0.0761 0.0805 0.9460
    ˆθ2 -0.0002 0.0759 0.0793 0.9400
    RASE1 0.9992 0.5281
    RASE2 0.1393 0.0591
    500 ˆρ -0.0017 0.0429 0.0431 0.9500
    ˆθ1 -0.0008 0.0577 0.0571 0.9540
    ˆθ2 -0.0008 0.0578 0.0596 0.9410
    RASE1 0.8034 0.4053
    RASE2 0.1073 0.0427
    Note: It shows that the bias (Bias), standard deviation (SD), standard error (SE) and coverage probability (CP) with nominal level of 95% for estimator and the mean and standard deviation (SD) of RASE for ˆβ() and ˆg() from 1000 repetitions.

     | Show Table
    DownLoad: CSV
    Table 3.  Simulation results for ρ=0.5 with error term (1+0.5Ui)N(0,1).
    n Est Bias SD SE CP
    100 ˆρ -0.0143 0.0979 0.0984 0.9490
    ˆθ1 -0.0065 0.1331 0.1465 0.9500
    ˆθ2 0.0048 0.1328 0.1339 0.9540
    RASE1 1.8170 0.8823
    RASE2 0.2431 0.0922
    300 ˆρ -0.0037 0.0556 0.0540 0.9530
    ˆθ1 0.0012 0.0750 0.0735 0.9500
    ˆθ2 -0.0032 0.0752 0.0755 0.9470
    RASE1 1.0316 0.4720
    RASE2 0.1411 0.0536
    500 ˆρ -0.0017 0.0431 0.0430 0.9470
    ˆθ1 0.0026 0.0577 0.0555 0.9570
    ˆθ2 -0.0026 0.0577 0.0574 0.9440
    RASE1 0.8035 0.3615
    RASE2 0.1058 0.0385
    Note: It shows that the bias (Bias), standard deviation (SD), standard error (SE) and coverage probability (CP) with nominal level of 95% for estimator and the mean and standard deviation (SD) of RASE for ˆβ() and ˆg() from 1000 repetitions.

     | Show Table
    DownLoad: CSV
    Table 4.  Simulation results for different magnitudes of ρ with error term N(0,1).
    ρ n Est Bias SE SD CP
    0.2 100 ˆρ -0.0114 0.1036 0.1046 0.9410
    ˆθ1 -0.0036 0.1046 0.1064 0.9390
    ˆθ2 -0.0019 0.1052 0.1009 0.9550
    ˆσ2 -0.1141 0.1334 0.1228 0.7500
    RASE1 1.4329 0.6579
    RASE2 0.1922 0.0705
    300 ˆρ 0.0013 0.0565 0.0532 0.9620
    ˆθ1 -0.0029 0.0585 0.0594 0.9500
    ˆθ2 -0.0011 0.0586 0.0586 0.9480
    ˆσ2 -0.0389 0.0817 0.0779 0.8820
    RASE1 0.8313 0.3826
    RASE2 0.1115 0.0397
    500 ˆρ -0.0012 0.0434 0.0435 0.9540
    ˆθ1 -0.0006 0.0452 0.0457 0.9460
    ˆθ2 -0.0021 0.0452 0.0439 0.9520
    ˆσ2 -0.0210 0.0658 0.0617 0.9090
    RASE1 0.6203 0.2789
    RASE2 0.0863 0.0307
    0.7 100 ˆρ -0.0059 0.0569 0.0553 0.9600
    ˆθ1 0.0028 0.1075 0.1101 0.9400
    ˆθ2 -0.0011 0.1068 0.1125 0.9340
    ˆσ2 -0.0990 0.1396 0.1239 0.7740
    RASE1 1.4460 0.6813
    RASE2 0.1935 0.0679
    300 ˆρ -0.0012 0.0317 0.0319 0.9500
    ˆθ1 0.0008 0.0597 0.0606 0.9480
    ˆθ2 0.0010 0.0597 0.0578 0.9520
    ˆσ2 -0.0324 0.0840 0.0782 0.8860
    RASE1 0.7987 0.3837
    RASE2 0.1109 0.0415
    500 ˆρ -0.0017 0.0242 0.2382 0.9560
    ˆθ1 -0.0035 0.0459 0.0443 0.9580
    ˆθ2 0.0006 0.0459 0.0488 0.9410
    ˆσ2 -0.0199 0.0651 0.0617 0.9030
    RASE1 0.6124 0.2717
    RASE2 0.0853 0.0311
    Note: It shows that the bias (Bias), standard error (SE), standard deviation (SD) and coverage probability (CP) with nominal level of 95% for estimator and the mean and standard deviation (SD) of RASE for ˆβ() from 1000 repetitions.

     | Show Table
    DownLoad: CSV

    (1) The estimators ˆρ,ˆθ1,ˆθ2,ˆσ2 are approximately unbiased and the estimated standard errors are close to sample standard deviations in normal error distribution. The empirical coverage probabilities approximate the nominal level of 95% well.

    (2) Figure 1 gives an example of the estimated function curve ˆβ() and ˆg() and its empirical 95% confidence interval with sample size n=300 for error term N(0,1). From the mean and standard deviation (SD) of RASEj(j=1,2), combined with Figure 1, we conclude that the proposed function estimators ˆβ() and ˆg() perform well.

    Figure 1.  It displays the true curve β(t) and g(u) (red solid line), the estimated curve ˆβ(t) and ˆg(u) (green dotted line) and ponitwise 2.5 and 97.5 percentile of the estimated function ˆβ() and ˆg(u) (light green line) in 500 replications with sample size n=300 respectively. In the firgure, the left one shows estimator ˆβ(t) and the right one shows estimator ˆg(u) with error term N(0, 1).

    (3) For error term 0.75t(3) and (1+0.5Ui)N(0,1), the estimators ˆρ,ˆθ1,ˆθ2 are approximately unbiased and the estimated standard errors are close to sample standard deviations. In addition, the mean and standard deviation for RASE of estimated coefficient function ˆβ() and ˆg() are decreasing. It indicates that parametric and non parametric estimators perform well in non-normal error term.

    (4) From Table 1 and Table 4, as basic spatial effect ρ increases, the SE and SD of ˆρ decrease. For the different magnitudes of ρ, the Bias and SD of parametric estimators for ˆθ1 and ˆθ2, and the mean of RASE for ˆβ() and ˆg() remain stable. It means that the magnitudes of ρ do not affect the other parametric and nonparametric estimators.

    In this paper, an interesting and novel functional semiparametric spatial autoregressive model is proposed. The model considers functional covariates based on semiparametric spatial autoregressive model. The slope function and nonparametric function are approximated by B-spline basis function. Then generalized method of moments is proposed to estimate parameters. Under mild conditions, we establish the asymptotic properties for proposed estimators.

    In order to use our model in practical applications, firstly, response variable needs spatial dependence. Secondly, there are covariates with nonlinear effect and functional variables. A problem of practical interest is to extend our model to take into account functional covariates and single index function simultaneously. What's more, making a test about spatial dependence and nonlinear effect of covariates is an important issue. Those topics are left for future work.

    We would like to thank the referees for their helpfull suggestions and comments which lead to the improvement of this article. Bai's work was supported by the National Natural Science Foundation of China (No.11771268).

    No potential conflict of interest was reported by the authors.

    Lemma 1. Assume condition C8 holds for g0(u) and β0(t), there exits γ0 and ζ0 such that

    supt(0,1)β0(t)B1(t)γ0c1kr1,   supu(0,1)g0(u)B2(u)ζ0c2kr2,

    where γ=(γ01,...,γ0N1), ζ=(ζ01,...,ζ0N2) and c1>0, c2>0 depend only on l1 and l2, respectively.

    Proof of Lemma 1. It can be followed by spline's approximation property ([4,16,34]).

    Proof of Theorem 1. The proof is similar to Theorem 1 in [10] and we omit here.

    Proof of Theorem 2. Let δ=nr2r+1, T1=δ1(γγ0), T2=δ1(ζζ0) and T=(T1,T2). We then prove that for any given ϵ>0, there exits a sufficient large constant L=Lϵ such that

    p{infT=Ll(ϕ0+δT)>l(ϕ0)}1ϵ,

    where ϕ0=(γ0,ζ0), l(γ,ζ)=ni=1(YiQiˆηDiγB2(ui)ζ)2. This implies that with the probability at least 1ϵ that there exits a local minimizer in the ball {ϕ0+δT:TL}. By Taylor expansion and simple calculation, it holds that

    {l(ϕ0+δT)l(ϕ0)}2δni=1(εi+R1i+R2i+Qi(ηˆη))Vi+δ2ni=1V2i+op(1)=A1+A2+op(1),

    where R1i=Xi(t),β0(t)B1(t)γ0,R2i=g0(ui)B2(ui)ζ0, Vi=DiT1+B2(ui)T2. By assumption C6, Lemmas 1 and 8 of Stone [30], we derive that R1ickr1,R2i=Op(kr2). Then by simple calculation, we obtain that

    ni=1R1iVi=ni=1R1i(DiT1+B2(ui)T2)=Op(nkr)T.

    Similarly, it hold that ni=1εiVi=Op(n)T,ni=1R2iVi=Op(nkr)T,ni=1V2i=Op(n)T2. Similar to the proof of Theorem 2 in Du et al. [10], we get (ηˆη)QQ(ηˆη)=Op(1). Then it holds that ni=1Qi(ηˆη)Vi=Op(n)T. Consequently, we show that A1=Op(nδ2)T, A2=Op(nδ2)T2. Then through choosing a sufficiently large L, A2 dominates A1 uniformly in T=L. Thus, there exits local minimizers ˆγ,ˆζ such that γˆγ=Op(δ),ζˆζ=Op(δ).

    Let R1k1(t)=β0(t)B1(t)γ0. Then we get

    ˆβ(t)β0(t)2=10(B1(t)ˆγβ0(t))2dt=10(B1(t)ˆγB1(t)γ0+R1k1(t))2dt210{B1(t)(ˆγγ0)}2dt+210R21k1(t)dt=2(ˆγγ0)10B1(t)B1(t)dt(ˆγγ0)+210R21k1(t)dt.

    Since γˆγ=Op(δ) and 10B1(t)B1(t)dt=O(1), then we have

    (ˆγγ0)10B1(t)B1(t)dt(ˆγγ0)=Op(δ2).

    In addition, by Lemma1, it holds that 10R21k1(t)dt=Op(δ2). Thus, we obtain ˆβ(t)β0(t)2=Op(δ2). Similarly, we get ˆg(u)g0(u)2=Op(δ2).

    Proof of Theorem 3. The proof is similar to Theorem 3 in [10] and we omit here.

    Proof of Theorem 4. By the definition of l(γ,ζ) in the proof of Theorem 2, we have

    12nl(ˆγ,ˆζ)γ=1nni=1[YiQiˆηDiγB2(ui)ζ]Di=1nni=1[˜eiDi(ˆγγ0)B2(ui)(ˆζζ0)]Di+op(1)=0, (A.1)

    where ˜ei=εi+R1i+R2i,R1i=Xi(t),β0(t)B1(t)γ0,R2i=g0(ui)B2(ui)ζ0. The remainder is op(1) because 1nni=1Qi(ˆηη)=op(1) by Theorem 1. In addition, we have

    12nl(ˆγ,ˆζ)ζ=1nni=1[˜eiDi(ˆγγ0)B2(ui)(ˆζζ0)]B2(ui)+op(1)=0. (A.2)

    It follows from (A.2) that

    ˆζζ0=[1nni=1B2(ui)B2(ui)]1{1nni=1˜eiB2(ui)1nni=1B2(ui)Di(ˆγγ0)+op(1)}. (A.3)

    Let

    ˉΛn=1nni=1DiDi1nni=1DiB2(ui)[1nni=1B2(ui)B2(ui)]11nni=1B2(ui)Di.

    By substituting (A.3) into (A.1), we obtain

    ˆγγ0=(ˉΛn)1{1nni=1˜ei[Di(1nDiB2(ui))×[1nni=1B2(ui)B2(ui)]1B2(ui)]+op(1)}.

    Since ˆβ(t)β(t)=B1(t)(ˆγγ0) and for any t(0,1), as n, by the law of large numbers, the slutsky's theorem and the property of multivariate normal distribution, we obtain that

    nk1(ˆβ(t)β(t))dN(0,Ξ(t)),

    where Ξ(t)=limnσ2k1B1(t)ΔnB1(t). Similar arguments hold for ˆg(u).



    [1] L. Anselin, Spatial econometrics: methods and models, The Netherlands: Kluwer Academic Publishers, 1988.
    [2] L. Anselin, A. K. Bera, Spatial dependence in linear regression models with an introduction to spatial econometrics, In: Handbook of Applied Economic Statistics, New York: Marcel Dekker, 1998.
    [3] G. Aneiros-Pˊerez, P. Vieu, Semi-functional partial linear regression, Stat. Probabil. Lett., 76 (2006), 1102–1110. doi: 10.1016/j.spl.2005.12.007
    [4] C. de Boor, A practical guide to splines, New York: Springer-Verlag, 2001.
    [5] H. Cardot, F. Ferraty, P. Sarda, Spline estimators for the functional linear model, Stat. Sinica, 13 (2003), 571–592.
    [6] T. Cai, P. Hall, Prediction in functional linear regression, Ann. Statist., 34 (2006), 2159–2179.
    [7] T. Cai, M. Yuan, Minimax and adaptive prediction for functional linear regression, J. Am. Stat. Assoc., 107 (2012), 1201–1216. doi: 10.1080/01621459.2012.716337
    [8] C. Crambes, A. Kneip, P. Sarda, Smoothing splines estimators for functional linear regression, Ann. Statist., 37 (2009), 35–72.
    [9] X. Dai, S. Li, M. Tian, Quantile regression for partially linear varying coefficient spatial autoregressive models, 2016, arXiv: 1608.01739.
    [10] J. Du, X. Sun, R. Cao, Z. Zhang, Statistical inference for partially linear additive spatial autoregressive models, Spat. Stat., 25 (2018), 52–67. doi: 10.1016/j.spasta.2018.04.008
    [11] A. Delaigle, P. Hall, Methodology and theory for patial least squares applied to functional data, Ann. Statist., 40 (2012), 322–352.
    [12] T. Huang, S. Gilbert, H. Wang, S. Wang, Spatial functional linear model and its estimation method, 2018, arXiv: 1811.00314.
    [13] Y. Hu, S. Wu, S. Feng, J. Jin, Estimation in partial functional linear spatial autoregressive model, Mathematics, 8 (2020), 1–12.
    [14] P. Hall, J. L. Horowitz, Methodology and convergence rates for functional linear regression, Ann. Statist., 35 (2007), 70–91.
    [15] J. Huang, Efficient estimation of the partly linear additive Cox model, Ann. Statist., 27 (1999), 1536–1563.
    [16] J. Z. Huang, Local asymptotics for polynomial spline regression, Ann. Statist., 31 (2003), 1600–1635.
    [17] H. H. Kelejian, I. R. Prucha, A generalized spatial two-stage least squares procedure for estimating a spatial autoregressive model with autoregressive disturbances, J. Real Estate Finan. Econ., 17 (1998), 99–121. doi: 10.1023/A:1007707430416
    [18] H. H. Kelejian, I. R. Prucha, A generalized moments estimator for the autoregressive parameter in a spatial model, International Economic Review, 40 (1999), 509–533. doi: 10.1111/1468-2354.00027
    [19] J. P. LeSage, The theory and practice of spatial econometrics, Ohio: University of Toledo, 1999.
    [20] L. F. Lee, J. H. Yu, Estimation of spatial autoregressive panel data models with fixed effects, J. Econometrics, 154 (2010), 165–185. doi: 10.1016/j.jeconom.2009.08.001
    [21] J. LeSage, R. K. Pace, Introduction to spatial econometrics, Boca Raton: Chapman and Hall/CRC, 2009.
    [22] L. F. Lee, GMM and 2SLS estimation of mixed regressive, spatial autoregressive models, J. Econometrics, 137 (2007), 489–514. doi: 10.1016/j.jeconom.2005.10.004
    [23] L. F. Lee, Best spatial two-stage least squares estimators for a spatial autoregressive model with autoregressive disturbances, Econometric Rev., 22 (2003), 307–335. doi: 10.1081/ETC-120025891
    [24] W. Pineda-Rios, R. Giraldo, E. Porcu, Functional SAR models: With application to spatial econometrics, Spat. Stat., 29 (2019), 145–159. doi: 10.1016/j.spasta.2018.12.002
    [25] X. Qu, L. F. Lee, A spatial autoregressive model with a nonlinear transformation of the dependent variable, J. Econometrics, 184 (2015), 209–232. doi: 10.1016/j.jeconom.2014.08.008
    [26] P. T. Reiss, J. Goldsmith, H. L. Shang, R. T. Ogden, Methods for scalar-on-function regression, Int. Stat. Rev., 85 (2017), 228–249. doi: 10.1111/insr.12163
    [27] J. O. Ramsay, C. J. Dalzell, Some tools for functinal data analysis (with discussion), J. R. Stat. Soc. B, 53 (1991), 539–572.
    [28] H. Shin, Partial functional linear regression, J. Stat. Plan. Infer., 139 (2009), 3405–3418. doi: 10.1016/j.jspi.2009.03.001
    [29] C. J. Stone, Optimal rates of convergence for nonparametric estimators, Ann. Statist., 8 (1980), 1348–1360.
    [30] C. J. Stone, Additive regression and other nonparametric models, Ann. Statist., 13 (1985), 689–705.
    [31] L. J. Su, Semiparametric GMM estimation of spatial autoregressive models, J. Econometrics, 167 (2012), 543–560. doi: 10.1016/j.jeconom.2011.09.034
    [32] L. J. Su, S. N. Jin, Profile quasi-maximum likelihood estimation of partially linear spatial autoregressive models, J. Econometrics, 157 (2010), 18–33. doi: 10.1016/j.jeconom.2009.10.033
    [33] Y. Sun, H. Yan, W. Zhang, Z. Lu, A semiparametric spatial dynamic model, Ann. Statist., 42 (2014), 700–727.
    [34] L. Schumaker, Spline functions: basic theory, Cambridge University Press, 2007.
    [35] H. Tadao, Semiparametric spatial autoregressive models with endogenous regressors: with an application to crime data, J. Bus. Econ. Stat., 36 (2018), 160–172. doi: 10.1080/07350015.2016.1146145
    [36] H. Wei, Y. Sun, Heteroskedasticity-robust semi-parametric GMM estimation of a spatial model with space-varying coefficients, Spatial Economic Analysis, 12 (2017), 113–128. doi: 10.1080/17421772.2017.1250940
    [37] L. Wang, X. Liu, H. Liang, R. Carroll, Estimation and variable selection for generalized additive partial linear models, Ann. Statist., 39 (2011), 1827–1851.
    [38] P. Yu, J. Du, Z. Zhang, Single-index partial functional linear regression model, Stat. Papers, 11 (2018), 1–17.
    [39] Y. Q. Zhang, D. M. Shen, Eseimation of semi-parametric varying-coefficient spatial panel data models with random effects, J. Statist. Plann. Infer., 159 (2015), 64–80. doi: 10.1016/j.jspi.2014.11.001
  • This article has been cited by:

    1. Yu Liu, Adaptive lasso variable selection method for semiparametric spatial autoregressive panel data model with random effects, 2022, 0361-0926, 1, 10.1080/03610926.2022.2119088
    2. Xinrong Tang, Peixin Zhao, Xiaoshuang Zhou, Weijia Zhang, Robust estimation for semiparametric spatial autoregressive models via weighted composite quantile regression, 2024, 0361-0926, 1, 10.1080/03610926.2024.2395881
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3157) PDF downloads(160) Cited by(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog